sha
null | last_modified
null | library_name
stringclasses 154
values | text
stringlengths 1
900k
| metadata
stringlengths 2
348k
| pipeline_tag
stringclasses 45
values | id
stringlengths 5
122
| tags
sequencelengths 1
1.84k
| created_at
stringlengths 25
25
| arxiv
sequencelengths 0
201
| languages
sequencelengths 0
1.83k
| tags_str
stringlengths 17
9.34k
| text_str
stringlengths 0
389k
| text_lists
sequencelengths 0
722
| processed_texts
sequencelengths 1
723
| tokens_length
sequencelengths 1
723
| input_texts
sequencelengths 1
61
| embeddings
sequencelengths 768
768
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | Basha738/llama2-supervised-ft-5epochs | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"4-bit",
"region:us"
] | 2024-02-08T06:30:17+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
59,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.049007222056388855,
0.16460949182510376,
-0.005271392408758402,
0.021910345181822777,
0.09685911983251572,
0.01403510570526123,
0.07018975168466568,
0.11002060770988464,
-0.02425350993871689,
0.11399492621421814,
0.03344893455505371,
0.09780009090900421,
0.11368958652019501,
0.1498505026102066,
-0.002398149576038122,
-0.23227156698703766,
0.04924226179718971,
-0.1249755248427391,
-0.03746527433395386,
0.1159956082701683,
0.15001481771469116,
-0.10170940309762955,
0.07611104100942612,
-0.029819702729582787,
-0.008722295984625816,
-0.032589927315711975,
-0.056551046669483185,
-0.04997202008962631,
0.051094699651002884,
0.07382578402757645,
0.06793182343244553,
0.004094683099538088,
0.09450557827949524,
-0.2669448256492615,
0.0197003111243248,
0.0730973482131958,
-0.002068581758067012,
0.07547242939472198,
0.054895199835300446,
-0.07525460422039032,
0.09282654523849487,
-0.0507965162396431,
0.1469351053237915,
0.08020289987325668,
-0.09152709692716599,
-0.19188682734966278,
-0.0887833908200264,
0.10164182633161545,
0.18469172716140747,
0.045696184039115906,
-0.022488808259367943,
0.09940612316131592,
-0.08621317893266678,
0.011039474047720432,
0.05154034495353699,
-0.06937182694673538,
-0.05223534256219864,
0.06355299055576324,
0.08018788695335388,
0.07678371667861938,
-0.12301702797412872,
-0.02094447799026966,
0.008637533523142338,
0.00831096712499857,
0.08201737701892853,
0.023290244862437248,
0.1510206013917923,
0.03883988782763481,
-0.12744688987731934,
-0.050009194761514664,
0.10665731877088547,
0.041741468012332916,
-0.04784774035215378,
-0.25138479471206665,
-0.030326439067721367,
-0.027732934802770615,
-0.029999805614352226,
-0.03873695060610771,
0.04263332113623619,
-0.0072723389603197575,
0.0826614573597908,
-0.008116158656775951,
-0.07679495960474014,
-0.03798604756593704,
0.06191713735461235,
0.060809630900621414,
0.026244111359119415,
-0.011753023602068424,
0.010934822261333466,
0.1174238994717598,
0.10631082952022552,
-0.12367359548807144,
-0.051516905426979065,
-0.06431761384010315,
-0.07867198437452316,
-0.04216236248612404,
0.03455616533756256,
0.041060756891965866,
0.049376390874385834,
0.2486443817615509,
0.017620395869016647,
0.05382118001580238,
0.03803925961256027,
0.010167144238948822,
0.06406087428331375,
0.11435336619615555,
-0.061582546681165695,
-0.09715550392866135,
-0.025186026468873024,
0.08966731280088425,
0.01176387071609497,
-0.04024789482355118,
-0.05783011019229889,
0.06293477863073349,
0.016524890437722206,
0.1202789843082428,
0.09223750233650208,
0.003793274285271764,
-0.07138240337371826,
-0.06413803994655609,
0.1937950700521469,
-0.1626761257648468,
0.04747059941291809,
0.034180231392383575,
-0.038511235266923904,
-0.0016249394975602627,
0.008853171020746231,
0.024325255304574966,
-0.021725021302700043,
0.08937039971351624,
-0.05618007108569145,
-0.041590798646211624,
-0.10983981937170029,
-0.035744234919548035,
0.03192625194787979,
0.009910091757774353,
-0.03217151761054993,
-0.031847331672906876,
-0.08444786816835403,
-0.06831640005111694,
0.09424425661563873,
-0.07356466352939606,
-0.053753651678562164,
-0.016938211396336555,
-0.07437273859977722,
0.024786023423075676,
0.01960081420838833,
0.07747352123260498,
-0.02004585787653923,
0.042900070548057556,
-0.05549933388829231,
0.06014169380068779,
0.10937028378248215,
0.033117540180683136,
-0.05445994809269905,
0.0621645413339138,
-0.2418462336063385,
0.0997670441865921,
-0.06829129904508591,
0.05325306951999664,
-0.15072302520275116,
-0.02465333603322506,
0.04913770779967308,
0.008168290369212627,
-0.010590006597340107,
0.13754788041114807,
-0.21924975514411926,
-0.027699807658791542,
0.1631394773721695,
-0.09464818984270096,
-0.07676627486944199,
0.05986984074115753,
-0.052457790821790695,
0.10692904144525528,
0.04047565534710884,
-0.026259733363986015,
0.06162377819418907,
-0.13397987186908722,
0.0005626814090646803,
-0.045883387327194214,
-0.01928110048174858,
0.15731419622898102,
0.07587230950593948,
-0.06994020938873291,
0.07348526269197464,
0.023750323802232742,
-0.023168303072452545,
-0.046913031488657,
-0.017583578824996948,
-0.1088033989071846,
0.010729904286563396,
-0.061985816806554794,
0.01937699131667614,
-0.025795195251703262,
-0.09332547336816788,
-0.028493179008364677,
-0.17521639168262482,
-0.020266273990273476,
0.08516935259103775,
-0.009352635592222214,
-0.01925206556916237,
-0.11787936836481094,
0.015734510496258736,
0.03501737862825394,
0.002549536293372512,
-0.1319509893655777,
-0.05043373629450798,
0.02751830592751503,
-0.16075198352336884,
0.033688947558403015,
-0.05403051897883415,
0.0491553395986557,
0.03133281692862511,
-0.031412381678819656,
-0.028679344803094864,
0.022094380110502243,
0.004997676704078913,
-0.014611656777560711,
-0.24550160765647888,
-0.026604164391756058,
-0.02145342156291008,
0.16796952486038208,
-0.21640902757644653,
0.0374150350689888,
0.07194960117340088,
0.15254895389080048,
0.008589224889874458,
-0.038006994873285294,
0.002335198922082782,
-0.075041763484478,
-0.03255171701312065,
-0.06050482019782066,
-0.009038056246936321,
-0.03572068363428116,
-0.05482286959886551,
0.04863523691892624,
-0.16824471950531006,
-0.029467429965734482,
0.1015508770942688,
0.06473538279533386,
-0.13604550063610077,
-0.019663551822304726,
-0.03585261106491089,
-0.042308371514081955,
-0.05517838895320892,
-0.05935737490653992,
0.10260266810655594,
0.05827045813202858,
0.04566904529929161,
-0.06485172361135483,
-0.0747392401099205,
0.0017082487465813756,
-0.019673427566885948,
-0.022536588832736015,
0.09213293343782425,
0.07581926137208939,
-0.12331884354352951,
0.09213830530643463,
0.10402927547693253,
0.08686267584562302,
0.0966128259897232,
-0.023164015263319016,
-0.08361977338790894,
-0.049845483154058456,
0.02228725142776966,
0.017598064616322517,
0.13447505235671997,
-0.007804518099874258,
0.05406574159860611,
0.04160919412970543,
-0.013909573666751385,
0.009752067737281322,
-0.09242741018533707,
0.032518286257982254,
0.03427431732416153,
-0.01857241988182068,
0.041615914553403854,
-0.039849672466516495,
0.019975949078798294,
0.09018522500991821,
0.046917494386434555,
0.04021155461668968,
0.014107138849794865,
-0.04660527780652046,
-0.11187547445297241,
0.16612006723880768,
-0.12780359387397766,
-0.23512837290763855,
-0.1463187336921692,
0.0034277087543159723,
0.03630480915307999,
-0.009390040300786495,
0.0017278295708820224,
-0.06397698074579239,
-0.11876852810382843,
-0.09194197505712509,
0.010153552517294884,
0.04896695911884308,
-0.0851091742515564,
-0.0603698305785656,
0.05686335638165474,
0.04057794436812401,
-0.14546048641204834,
0.019262617453932762,
0.04933769255876541,
-0.09224124997854233,
-0.009894786402583122,
0.08289197087287903,
0.06857553124427795,
0.18091025948524475,
0.013082148507237434,
-0.02271466888487339,
0.03428078070282936,
0.21755947172641754,
-0.13586747646331787,
0.11420658230781555,
0.1426045000553131,
-0.09194567799568176,
0.08309654146432877,
0.19839057326316833,
0.04078111797571182,
-0.10157861560583115,
0.032499175518751144,
0.018653791397809982,
-0.030491048470139503,
-0.24355553090572357,
-0.07171683013439178,
0.00034942623460665345,
-0.057900771498680115,
0.07530075311660767,
0.09018687158823013,
0.09155713021755219,
0.01583298109471798,
-0.0946493074297905,
-0.07830986380577087,
0.05305508151650429,
0.10324970632791519,
0.020061472430825233,
-0.013236436992883682,
0.09051742404699326,
-0.03375976160168648,
0.017617853358387947,
0.09066354483366013,
0.0011531224008649588,
0.17065346240997314,
0.05820678174495697,
0.18275249004364014,
0.07604338973760605,
0.07338658720254898,
0.01378361415117979,
0.01180104911327362,
0.019032908603549004,
0.02708563208580017,
-0.004741039127111435,
-0.08538748323917389,
-0.01599922962486744,
0.12008915096521378,
0.07424698024988174,
0.015674617141485214,
0.014355434104800224,
-0.04089333862066269,
0.08203015476465225,
0.17435193061828613,
-0.001506963511928916,
-0.1824604868888855,
-0.06271602213382721,
0.08220411837100983,
-0.09449198096990585,
-0.10147359222173691,
-0.02445729449391365,
0.03089604340493679,
-0.17088350653648376,
0.023070847615599632,
-0.016430631279945374,
0.11182350665330887,
-0.13931094110012054,
-0.019696295261383057,
0.0640200525522232,
0.07118809968233109,
-0.00031885437783785164,
0.05944213643670082,
-0.16128569841384888,
0.10404066741466522,
0.013166810385882854,
0.06712377816438675,
-0.09715772420167923,
0.10046469420194626,
-0.006883090827614069,
-0.013416164554655552,
0.13275203108787537,
0.008256223052740097,
-0.07161599397659302,
-0.07921489328145981,
-0.09379399567842484,
-0.009093280881643295,
0.12668752670288086,
-0.14835532009601593,
0.08585991710424423,
-0.035368360579013824,
-0.04256736859679222,
0.0022144275717437267,
-0.10755012929439545,
-0.12217973172664642,
-0.1874755620956421,
0.05520224943757057,
-0.1321607530117035,
0.039849888533353806,
-0.10649667680263519,
-0.03462952747941017,
-0.029491933062672615,
0.1882491409778595,
-0.22971367835998535,
-0.06835493445396423,
-0.15157760679721832,
-0.09785088151693344,
0.14553189277648926,
-0.04969761520624161,
0.08694402873516083,
-0.005991519894450903,
0.18016821146011353,
0.022223925217986107,
-0.021585633978247643,
0.09859558939933777,
-0.09382225573062897,
-0.1963716447353363,
-0.08180448412895203,
0.15751656889915466,
0.13459575176239014,
0.03521031513810158,
-0.0027760460507124662,
0.037876322865486145,
-0.01856307126581669,
-0.12259240448474884,
0.021658578887581825,
0.17797763645648956,
0.0652514174580574,
0.02310643345117569,
-0.026529761031270027,
-0.11104881763458252,
-0.06772379577159882,
-0.033685971051454544,
0.03064778819680214,
0.18449479341506958,
-0.0722544714808464,
0.18419069051742554,
0.143813356757164,
-0.05867353826761246,
-0.1976030021905899,
0.008879725821316242,
0.03365374729037285,
0.007196295075118542,
0.03445420414209366,
-0.20255140960216522,
0.0841677114367485,
0.00034181843511760235,
-0.05190233513712883,
0.13343381881713867,
-0.17106693983078003,
-0.15042030811309814,
0.07339101284742355,
0.03619921952486038,
-0.19460853934288025,
-0.11963265389204025,
-0.08913769572973251,
-0.05391303077340126,
-0.18051348626613617,
0.10290905088186264,
0.03496568650007248,
0.008035079576075077,
0.03376363217830658,
0.028494013473391533,
0.01669638603925705,
-0.03928735852241516,
0.1920013129711151,
-0.026591487228870392,
0.029855716973543167,
-0.08456290513277054,
-0.06990274786949158,
0.04655740037560463,
-0.05482156574726105,
0.0760476216673851,
-0.027013001963496208,
0.011612839996814728,
-0.10561433434486389,
-0.042526841163635254,
-0.029051896184682846,
0.013453613966703415,
-0.0963861495256424,
-0.08940120041370392,
-0.0490599125623703,
0.09310506284236908,
0.09519506990909576,
-0.035876575857400894,
-0.03684677556157112,
-0.07069114595651627,
0.039579302072525024,
0.18676936626434326,
0.17657315731048584,
0.04523694887757301,
-0.0789421945810318,
-0.005537794437259436,
-0.011924253776669502,
0.04352729767560959,
-0.21637341380119324,
0.06442029029130936,
0.05013522133231163,
0.017847778275609016,
0.11767403781414032,
-0.02045002020895481,
-0.1556767225265503,
-0.07006701827049255,
0.06328949332237244,
-0.06132598593831062,
-0.1951322853565216,
0.005576360039412975,
0.054395273327827454,
-0.16848263144493103,
-0.048018258064985275,
0.04364382475614548,
-0.004054433200508356,
-0.0402018167078495,
0.01867259293794632,
0.08977478742599487,
0.003425614908337593,
0.0704059898853302,
0.05869606137275696,
0.08224445581436157,
-0.10246741771697998,
0.07471306622028351,
0.08622124791145325,
-0.07954994589090347,
0.026619622483849525,
0.09149482846260071,
-0.05819176882505417,
-0.02969011478126049,
0.02704544924199581,
0.0793747529387474,
0.011502381414175034,
-0.042540501803159714,
0.011518802493810654,
-0.10228829830884933,
0.06203006953001022,
0.08760257810354233,
0.03265642002224922,
0.015443529933691025,
0.03219176456332207,
0.045628782361745834,
-0.07176384329795837,
0.1219232901930809,
0.028246978297829628,
0.015991143882274628,
-0.04067446291446686,
-0.04898078367114067,
0.024271609261631966,
-0.0303955040872097,
-0.006366716232150793,
-0.03475780412554741,
-0.0729878842830658,
-0.0171539094299078,
-0.16714228689670563,
-0.016664555296301842,
-0.04662061110138893,
0.009329318068921566,
0.03086909092962742,
-0.03788549080491066,
0.008464637212455273,
0.007407912518829107,
-0.07459274679422379,
-0.06477426737546921,
-0.022905457764863968,
0.09289900958538055,
-0.16393527388572693,
0.02335011027753353,
0.08690579235553741,
-0.12064014375209808,
0.09392421692609787,
0.01837589405477047,
-0.0037578048650175333,
0.028480252251029015,
-0.14924435317516327,
0.038928523659706116,
-0.03113253228366375,
0.014821149408817291,
0.04454975947737694,
-0.2236335128545761,
0.0009650349384173751,
-0.033828526735305786,
-0.06339430809020996,
-0.009390673600137234,
-0.036760155111551285,
-0.11370383948087692,
0.10629112273454666,
0.007970798760652542,
-0.08916810154914856,
-0.031690530478954315,
0.032128699123859406,
0.08206479996442795,
-0.0239556971937418,
0.15763959288597107,
-0.0023972811177372932,
0.0736590027809143,
-0.1675432026386261,
-0.019303109496831894,
-0.011248460970818996,
0.020926566794514656,
-0.018098697066307068,
-0.01251189224421978,
0.04078914225101471,
-0.02225574664771557,
0.18437865376472473,
-0.023570427671074867,
0.023348741233348846,
0.06592654436826706,
0.027775658294558525,
-0.025002485141158104,
0.10530006885528564,
0.05339968949556351,
0.021854043006896973,
0.02036798559129238,
0.00273964018560946,
-0.04241073876619339,
-0.023610878735780716,
-0.1998770385980606,
0.06446972489356995,
0.14037446677684784,
0.09086652100086212,
-0.017234215512871742,
0.08257289230823517,
-0.1004219725728035,
-0.11521948128938675,
0.11568495631217957,
-0.05446505919098854,
-0.004037478007376194,
-0.0672159418463707,
0.12938179075717926,
0.1446845531463623,
-0.19097456336021423,
0.06995914876461029,
-0.06848131865262985,
-0.049033988267183304,
-0.11654651165008545,
-0.1963350623846054,
-0.05714293569326401,
-0.05161691829562187,
-0.01663723587989807,
-0.046969223767519,
0.07560921460390091,
0.05719533935189247,
0.007424132898449898,
-0.0017566849710419774,
0.06332923471927643,
-0.026077456772327423,
0.00009585227962816134,
0.026813751086592674,
0.06610306352376938,
0.013093758374452591,
-0.02985633723437786,
0.017491595819592476,
-0.012147722765803337,
0.042048826813697815,
0.06357792019844055,
0.04670548066496849,
-0.030032360926270485,
0.016853880137205124,
-0.03863191977143288,
-0.10680584609508514,
0.041318636387586594,
-0.028504958376288414,
-0.08043242245912552,
0.1491626501083374,
0.02454165369272232,
0.008750278502702713,
-0.0205967016518116,
0.2416755110025406,
-0.0737907737493515,
-0.09567341208457947,
-0.1479424238204956,
0.10524045675992966,
-0.04420987144112587,
0.06244929879903793,
0.045180387794971466,
-0.10425344854593277,
0.016717668622732162,
0.12817999720573425,
0.16302813589572906,
-0.044200748205184937,
0.020526019856333733,
0.027614353224635124,
0.004152800887823105,
-0.03678637370467186,
0.0514480359852314,
0.06988705694675446,
0.1595088243484497,
-0.048713311553001404,
0.09546878933906555,
-0.0016016386216506362,
-0.09618084132671356,
-0.03802286460995674,
0.11709540337324142,
-0.018092934042215347,
0.017691975459456444,
-0.055210161954164505,
0.11857418715953827,
-0.06138255074620247,
-0.2316483110189438,
0.06108921393752098,
-0.06591550260782242,
-0.13765475153923035,
-0.02143050730228424,
0.08041442185640335,
-0.013238796964287758,
0.02708347514271736,
0.07207029312849045,
-0.07533451914787292,
0.20003929734230042,
0.037636954337358475,
-0.05420409142971039,
-0.05360380560159683,
0.08255447447299957,
-0.10376271605491638,
0.27565470337867737,
0.016520937904715538,
0.04948882386088371,
0.10317612439393997,
-0.012690499424934387,
-0.13475549221038818,
0.02108365297317505,
0.09600389003753662,
-0.0946137085556984,
0.04216265305876732,
0.19903649389743805,
0.0003853837260976434,
0.1207512691617012,
0.0790785402059555,
-0.07618726044893265,
0.049590613692998886,
-0.0941753089427948,
-0.07070460170507431,
-0.09001081436872482,
0.09455035626888275,
-0.07685617357492447,
0.14261877536773682,
0.1292559802532196,
-0.053739987313747406,
0.010677514597773552,
-0.028576120734214783,
0.04638256877660751,
0.0034859003499150276,
0.1005801111459732,
0.010024284943938255,
-0.18460705876350403,
0.02157641015946865,
0.01203901320695877,
0.1056026741862297,
-0.16518552601337433,
-0.09804878383874893,
0.042120642960071564,
0.0014211505185812712,
-0.060778699815273285,
0.12909291684627533,
0.06027422100305557,
0.04478219151496887,
-0.04292554408311844,
-0.020403601229190826,
-0.009860116057097912,
0.13677826523780823,
-0.10241927951574326,
0.0014122816501185298
] |
null | null | null |
Finetune mT5-large with LoRA for English-Vietnamese translation
# How to use it
```python
import torch
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
base_model = AutoModelForSeq2SeqLM.from_pretrained("google/mt5-large")
tokenizer = AutoTokenizer.from_pretrained("google/mt5-large")
# load Peft model
peft_model = PeftModel.from_pretrained(base_model, "adapter")
peft_model.save_pretrained("inference")
del peft_model, base_model
# reload with mixed precision
device = 'cuda' if torch.cuda.is_available() else 'cpu'
model = AutoModelForSeq2SeqLM.from_pretrained(
"inference", torch_dtype=torch.float16
).to(device)
```
# Inference
```python
def preprocess(input_text:str = None, prefix:str = 'vi'):
return prefix + input_text + "<END>"
def postprocess(output_text):
return output_text[4:].split("<END>")[0]
def translate(text:str = None, prefix:str = 'vi'):
text = preprocess(text, prefix)
tokenized = tokenizer(
text,
padding=True, truncation=True,
return_tensors="pt"
).to(model.device)
with torch.no_grad():
outputs = model(**tokenized, beam_search=4, use_cache=True)
output = tokenizer.decode(outputs[0], skip_special_tokens=True)
return postprocess(output)
```
## Translate sentences
For translate English to Vietnamese
```
en2vi = 'VnExpress provides latest Vietnam news, regional, business, financial, industries, travel news and views to policy makers'
prefix = 'en: '
result = translate(text = en2vi, prefix = prefix)
# VnExpress cung cấp tin tức mới nhất về Việt Nam, khu vực, kinh doanh, tài chính, công nghiệp, tin tức du lịch và quan điểm cho các nhà hoạch định chính sách
```
For translate Vietnamese to English
```
vi2en = 'VnExpress là một tờ báo tại Việt Nam được thành lập bởi tập đoàn FPT, ra mắt vào ngày 26 tháng 2 năm 2001'
prefix = 'vi: '
result = translate(text = vi2en, prefix = prefix)
# VnExpress is a newspaper in Vietnam founded by FPT Corporation, launched on February 26, 2001.
```
# Hyperparameters Training
```
- batch_size = 2
- gradient_acc = 14
- total_steps = 100k
- warm_up = 0.1
- max_len = 512
- learning_rate = 3e-5
- scheduler = cosine
- deepspeed = v1
- gpus = 2x3090
- dataset_train = presencesw/hash_v3
- lora_target = ['q', 'wi_1', 'k', 'wi_0', 'v', 'wo', 'o', 'lm_head']
- trainable parameter: 0.38228893842642636%
- LORA_R=8
- LORA_ALPHA=8
- LORA_DROPOUT = 0.1
```
- Total time: ~168 hours
- Total dataset: ~5.6M samples
- Curve loss

| {"language": ["en", "vi"], "license": "apache-2.0", "pipeline_tag": "translation"} | translation | pythera/translator-v1 | [
"safetensors",
"translation",
"en",
"vi",
"license:apache-2.0",
"region:us"
] | 2024-02-08T06:31:33+00:00 | [] | [
"en",
"vi"
] | TAGS
#safetensors #translation #en #vi #license-apache-2.0 #region-us
|
Finetune mT5-large with LoRA for English-Vietnamese translation
# How to use it
# Inference
## Translate sentences
For translate English to Vietnamese
For translate Vietnamese to English
# Hyperparameters Training
- Total time: ~168 hours
- Total dataset: ~5.6M samples
- Curve loss
!image/png
| [
"# How to use it",
"# Inference",
"## Translate sentences\nFor translate English to Vietnamese\n\n\nFor translate Vietnamese to English",
"# Hyperparameters Training\n\n- Total time: ~168 hours\n- Total dataset: ~5.6M samples\n\n- Curve loss\n\n!image/png"
] | [
"TAGS\n#safetensors #translation #en #vi #license-apache-2.0 #region-us \n",
"# How to use it",
"# Inference",
"## Translate sentences\nFor translate English to Vietnamese\n\n\nFor translate Vietnamese to English",
"# Hyperparameters Training\n\n- Total time: ~168 hours\n- Total dataset: ~5.6M samples\n\n- Curve loss\n\n!image/png"
] | [
26,
5,
4,
19,
31
] | [
"passage: TAGS\n#safetensors #translation #en #vi #license-apache-2.0 #region-us \n# How to use it# Inference## Translate sentences\nFor translate English to Vietnamese\n\n\nFor translate Vietnamese to English# Hyperparameters Training\n\n- Total time: ~168 hours\n- Total dataset: ~5.6M samples\n\n- Curve loss\n\n!image/png"
] | [
-0.018854709342122078,
0.04648366570472717,
-0.001187940244562924,
-0.0018019169801846147,
0.10133630782365799,
-0.06340909749269485,
0.077583447098732,
0.03909647837281227,
-0.0904584676027298,
-0.009416521526873112,
0.12917354702949524,
0.048087310045957565,
0.018044540658593178,
0.05971146747469902,
-0.10245204716920853,
-0.27248018980026245,
0.09378593415021896,
-0.020206009969115257,
-0.03823232650756836,
0.0960502102971077,
0.09517145156860352,
-0.08160810172557831,
0.11211847513914108,
-0.012177539989352226,
-0.040680937469005585,
0.011387016624212265,
-0.03255016356706619,
-0.07963811606168747,
0.12856800854206085,
-0.03517772629857063,
0.039553504437208176,
0.032609499990940094,
0.06497327983379364,
-0.15780194103717804,
0.006878910586237907,
0.05259712412953377,
-0.0071561080403625965,
-0.0015844153240323067,
-0.055859751999378204,
0.1163606345653534,
0.22701820731163025,
-0.17323176562786102,
0.038365814834833145,
0.03603469207882881,
-0.1432865411043167,
-0.10883951932191849,
-0.09238473325967789,
-0.16829994320869446,
0.2631997764110565,
0.10031168162822723,
-0.04002037271857262,
0.1926361471414566,
-0.1933814138174057,
0.0051916553638875484,
0.1835390031337738,
-0.39383214712142944,
-0.041760288178920746,
0.08233583718538284,
0.10807180404663086,
0.023483259603381157,
-0.1452334076166153,
0.09059961885213852,
0.09471607953310013,
0.020585011690855026,
-0.042335618287324905,
-0.1424141824245453,
-0.14682669937610626,
0.03393644839525223,
-0.17337378859519958,
0.08465250581502914,
0.11440558731555939,
0.06629638373851776,
-0.030195996165275574,
-0.039758358150720596,
-0.030039768666028976,
-0.026400376111268997,
-0.08500349521636963,
-0.06778164952993393,
0.08793288469314575,
0.004501132294535637,
0.09496139734983444,
-0.06389268487691879,
-0.09080902487039566,
-0.0746183916926384,
-0.11344853788614273,
-0.21000359952449799,
0.02307189628481865,
0.0055728210136294365,
-0.0027106713969260454,
0.010806821286678314,
-0.1795799732208252,
0.002805243246257305,
0.0019591874442994595,
-0.03889813274145126,
-0.01774626038968563,
0.01696416363120079,
0.03302234411239624,
-0.09276478737592697,
0.08589338511228561,
0.014401059597730637,
-0.14117370545864105,
0.02642136625945568,
-0.10607895255088806,
0.06069880723953247,
-0.034764789044857025,
0.1469530314207077,
-0.09905844926834106,
-0.06488319486379623,
-0.026505930349230766,
0.010189059190452099,
-0.0022567720152437687,
0.06951432675123215,
-0.09115487337112427,
-0.0175409447401762,
-0.030419906601309776,
0.0818324163556099,
-0.09217972308397293,
-0.006740711163729429,
-0.0327768512070179,
0.03009292297065258,
0.14386415481567383,
-0.01610591448843479,
-0.03542757034301758,
-0.045945923775434494,
-0.05776150897145271,
0.08233918994665146,
0.06569768488407135,
0.07279130816459656,
-0.03267249837517738,
-0.03208565339446068,
0.01217783521860838,
0.0321861132979393,
-0.06847760826349258,
-0.0908350870013237,
0.019801603630185127,
0.03392261266708374,
-0.04050842672586441,
-0.02700052782893181,
-0.10168521851301193,
-0.016190938651561737,
0.046328332275152206,
-0.0703708603978157,
0.05125671997666359,
-0.16132880747318268,
-0.06579558551311493,
-0.024724973365664482,
-0.010070589371025562,
0.05232676491141319,
-0.03452667221426964,
0.0280538871884346,
-0.038342852145433426,
0.13319391012191772,
-0.09522581100463867,
0.022533666342496872,
-0.10922746360301971,
0.026968609541654587,
-0.20827274024486542,
0.03432997316122055,
-0.03626836836338043,
0.06736479699611664,
-0.09716108441352844,
-0.07014906406402588,
0.016821006312966347,
0.04643804207444191,
-0.0006499747396446764,
0.14884492754936218,
-0.2557625472545624,
0.03630489856004715,
0.1821901947259903,
-0.03880791366100311,
-0.06774648278951645,
0.17490042746067047,
0.0006805872544646263,
0.10315264761447906,
0.0750613734126091,
0.16130654513835907,
-0.17574036121368408,
-0.13836245238780975,
0.14742200076580048,
0.056465912610292435,
-0.04779209569096565,
-0.062218982726335526,
0.05584852397441864,
-0.009081610478460789,
0.02621687762439251,
-0.0024578659795224667,
-0.04986990988254547,
0.029058504849672318,
-0.09072020649909973,
-0.07234600186347961,
0.038162898272275925,
-0.10831455886363983,
0.03230470418930054,
0.021391967311501503,
0.06224677711725235,
-0.11112416535615921,
-0.012377696111798286,
0.014935057610273361,
0.06351444125175476,
0.0316314734518528,
0.01372818648815155,
-0.08222448825836182,
-0.10822584480047226,
0.07202895730733871,
0.024439753964543343,
-0.004869547206908464,
0.07997775822877884,
-0.014959689229726791,
0.16711659729480743,
0.022326231002807617,
0.06864304840564728,
0.024906227365136147,
0.04024427384138107,
-0.03500578552484512,
0.041374579071998596,
0.10326343029737473,
0.029142584651708603,
-0.11526311188936234,
-0.10569063574075699,
0.13979284465312958,
-0.018379539251327515,
-0.003680879482999444,
-0.2536093294620514,
-0.012558415532112122,
-0.05029074475169182,
-0.04447048157453537,
0.01755487360060215,
0.08355167508125305,
-0.02038535103201866,
0.0032725513447076082,
-0.0208822563290596,
-0.038923006504774094,
0.05947468429803848,
-0.06486315280199051,
-0.022617371752858162,
0.19474339485168457,
-0.07243566960096359,
-0.0025105662643909454,
0.1355544775724411,
-0.1581779420375824,
-0.04930642619729042,
-0.024922961369156837,
0.042165789753198624,
-0.01970578357577324,
-0.11159021407365799,
0.0692613497376442,
0.025373201817274094,
0.07351300120353699,
0.1384073793888092,
-0.05857338756322861,
0.0665828287601471,
-0.0021213411819189787,
-0.12113473564386368,
0.04300161823630333,
0.11569737643003464,
0.04016857594251633,
-0.2900646924972534,
0.07986743003129959,
0.20876145362854004,
0.04194732382893562,
0.04263782873749733,
-0.03771600127220154,
-0.0014004040276631713,
-0.0031780540011823177,
0.1555459350347519,
0.002954769181087613,
0.08586669713258743,
-0.06294324994087219,
0.009928659535944462,
0.022475548088550568,
0.07451718300580978,
-0.05026697739958763,
-0.1928262859582901,
-0.09039930999279022,
-0.06890102475881577,
-0.07839622348546982,
0.002004990354180336,
0.040350571274757385,
-0.005364733282476664,
0.09365770220756531,
0.013552376069128513,
-0.09633227437734604,
-0.021814918145537376,
-0.009778600186109543,
-0.028515558689832687,
0.1941271722316742,
-0.1366504430770874,
-0.28054070472717285,
0.06123819202184677,
-0.03217272832989693,
0.06063292175531387,
-0.037359051406383514,
0.04254436120390892,
-0.2499232441186905,
-0.07943594455718994,
-0.02691677398979664,
0.014188973233103752,
0.046020254492759705,
-0.06003004312515259,
-0.03587250038981438,
0.0617499016225338,
-0.06442999839782715,
-0.03553122282028198,
-0.017195791006088257,
-0.10990767925977707,
-0.14387474954128265,
0.021221844479441643,
-0.13982051610946655,
-0.039601314812898636,
0.08224781602621078,
-0.027942543849349022,
-0.03061913326382637,
-0.01933952420949936,
0.13777834177017212,
-0.07391302287578583,
0.05150998383760452,
0.1392015814781189,
0.09752193838357925,
0.025608690455555916,
0.10812154412269592,
0.0006192108266986907,
-0.14198088645935059,
0.07729155570268631,
0.1212461069226265,
-0.041852861642837524,
-0.20701107382774353,
-0.08506911247968674,
-0.053466495126485825,
0.07250968366861343,
-0.002179516013711691,
0.016903406009078026,
-0.04766374081373215,
-0.0018358358647674322,
0.07274577021598816,
0.10239627957344055,
0.06905850023031235,
-0.04164661839604378,
0.11579475551843643,
-0.07329105585813522,
0.040475476533174515,
-0.08433259278535843,
-0.0370592400431633,
0.0924239233136177,
0.00029923286638222635,
0.14388170838356018,
0.10253848135471344,
-0.049720436334609985,
0.10243625938892365,
0.15444783866405487,
-0.027762433513998985,
0.03903116285800934,
-0.08965850621461868,
-0.08243144303560257,
-0.1081659346818924,
-0.029130754992365837,
0.049786582589149475,
0.04210406914353371,
-0.0058867973275482655,
-0.0885792076587677,
-0.039763059467077255,
0.12429066002368927,
0.11740729957818985,
0.19284148514270782,
0.15210846066474915,
-0.07182145863771439,
0.0587613619863987,
0.014360284432768822,
-0.009727112017571926,
-0.04912148416042328,
0.16904737055301666,
0.11165408790111542,
-0.11570346355438232,
0.09061092883348465,
-0.025391345843672752,
0.11585985124111176,
-0.03841695934534073,
0.052156783640384674,
-0.10222925245761871,
-0.061008427292108536,
0.016881516203284264,
0.051859382539987564,
-0.30131688714027405,
0.27257853746414185,
0.030213985592126846,
0.018200762569904327,
-0.035228513181209564,
-0.10360735654830933,
0.06766809523105621,
0.15034008026123047,
0.137989804148674,
0.04112936928868294,
-0.0177602656185627,
-0.03724193572998047,
-0.07880489528179169,
-0.02130122482776642,
0.02946249581873417,
0.15013329684734344,
-0.0719352513551712,
-0.009537197649478912,
-0.002792967716231942,
-0.0031968506518751383,
-0.05798279866576195,
-0.0831357017159462,
-0.024599779397249222,
-0.09984494000673294,
0.06537513434886932,
0.006491280160844326,
-0.006725124549120665,
0.028910266235470772,
-0.04844887554645538,
0.1327955722808838,
0.03226958215236664,
0.04104015603661537,
-0.04153023287653923,
0.04480709135532379,
0.1157665029168129,
-0.013370206579566002,
0.06917701661586761,
-0.025627154856920242,
-0.11873528361320496,
0.00044370393152348697,
0.04976099729537964,
0.2089577168226242,
-0.06842891126871109,
-0.04918638989329338,
-0.051734939217567444,
0.038376741111278534,
0.01842900551855564,
0.025912854820489883,
0.030045093968510628,
-0.019547000527381897,
-0.010952834039926529,
-0.11257919669151306,
-0.07030147314071655,
-0.11137912422418594,
0.03822234645485878,
0.03296356648206711,
-0.1288890242576599,
-0.1108684241771698,
-0.041355714201927185,
-0.1277891844511032,
0.18719084560871124,
0.28482064604759216,
-0.021818645298480988,
0.08233064413070679,
0.18744055926799774,
-0.0260821133852005,
-0.23742881417274475,
0.03299105539917946,
-0.051592979580163956,
0.01033782958984375,
0.11176601052284241,
-0.03970691189169884,
0.008316866122186184,
0.1329020857810974,
-0.041549816727638245,
0.013675235211849213,
-0.264439195394516,
-0.10691695660352707,
0.14749309420585632,
0.042214374989271164,
0.26284149289131165,
-0.1342742145061493,
0.02871663123369217,
-0.05837812274694443,
-0.048262033611536026,
0.1474175602197647,
-0.17651130259037018,
0.034914497286081314,
0.08358056098222733,
0.047011200338602066,
0.0018310699379071593,
-0.007199589628726244,
0.14541052281856537,
-0.09632605314254761,
0.10429146140813828,
-0.05955225229263306,
-0.11919092386960983,
-0.036291979253292084,
0.0018695325125008821,
0.10308133065700531,
-0.05782852694392204,
0.020039992406964302,
0.032820869237184525,
0.027623463422060013,
-0.05041823536157608,
-0.06313618272542953,
-0.010466405190527439,
-0.021957235410809517,
-0.07538213580846786,
0.09520852565765381,
-0.020967306569218636,
0.04146776348352432,
0.1735202670097351,
-0.014413607306778431,
-0.06539589911699295,
0.03156367316842079,
0.05154890567064285,
-0.03150903433561325,
0.1940687745809555,
-0.01672833040356636,
0.029175885021686554,
0.08370202034711838,
-0.22241275012493134,
0.023180771619081497,
0.057940661907196045,
-0.046821922063827515,
0.13231462240219116,
-0.04026198014616966,
0.004152866080403328,
0.07402712851762772,
0.057044968008995056,
-0.03558652848005295,
-0.14193841814994812,
0.05313714221119881,
0.08724421262741089,
0.1263834685087204,
0.05325102433562279,
0.025473907589912415,
-0.04967724531888962,
-0.011190379038453102,
-0.013520138338208199,
0.0965939611196518,
-0.11814875900745392,
0.1448645293712616,
-0.0044351136311888695,
0.03097483143210411,
-0.07839594036340714,
0.11103779822587967,
0.09997805207967758,
-0.11336202919483185,
0.02276732213795185,
0.149840846657753,
-0.10165706276893616,
-0.04046628251671791,
0.1653091162443161,
0.1097436398267746,
-0.05530562996864319,
-0.05947147309780121,
-0.01124324556440115,
-0.0781228095293045,
-0.023672837764024734,
0.12610764801502228,
0.051319099962711334,
-0.015201957896351814,
0.041660718619823456,
-0.047367095947265625,
-0.041857536882162094,
0.00554088456556201,
0.04291559383273125,
-0.027002766728401184,
0.014918411150574684,
-0.06932847201824188,
0.11484130471944809,
0.13663436472415924,
-0.040844958275556564,
0.020433876663446426,
-0.14247851073741913,
0.06556851416826248,
-0.11782924830913544,
-0.006161399185657501,
-0.017482265830039978,
-0.03221859037876129,
-0.01886175572872162,
-0.044891685247421265,
-0.06321153044700623,
-0.025019900873303413,
-0.11612994223833084,
0.030091769993305206,
-0.0031334897503256798,
0.04908695071935654,
-0.07287372648715973,
-0.09469287842512131,
0.08805625885725021,
0.03260737657546997,
0.026957569643855095,
-0.06585536897182465,
-0.025935936719179153,
0.09651397168636322,
-0.19167111814022064,
-0.048894427716732025,
0.11135775595903397,
0.05056530982255936,
0.10479655861854553,
0.050618670880794525,
-0.03850403055548668,
0.04946642369031906,
0.18908868730068207,
-0.017350124195218086,
0.08255652338266373,
-0.06447389721870422,
-0.12390081584453583,
-0.15628087520599365,
-0.1344572901725769,
0.022391311824321747,
0.0026320805773139,
-0.004632134456187487,
-0.03736941143870354,
0.10126454383134842,
-0.08322097361087799,
0.024491876363754272,
-0.03398649021983147,
-0.035081226378679276,
-0.07420793175697327,
-0.11187011003494263,
-0.036718547344207764,
-0.07310422509908676,
0.043052174150943756,
-0.051619984209537506,
0.15872833132743835,
0.0648290142416954,
0.1068916991353035,
0.0323965921998024,
-0.060188569128513336,
-0.1026037409901619,
0.021906903013586998,
0.12335646152496338,
0.06873972713947296,
-0.002559703541919589,
-0.1412392556667328,
0.08242987841367722,
0.018875839188694954,
0.021360065788030624,
0.08547475934028625,
0.13784724473953247,
0.04134908318519592,
0.11042052507400513,
-0.07813871651887894,
0.013713933527469635,
-0.029942801222205162,
0.13233758509159088,
-0.07577560842037201,
0.04716195538640022,
-0.06140197813510895,
0.10087123513221741,
0.14635410904884338,
-0.160316601395607,
0.07456360757350922,
-0.040849100798368454,
-0.06890816986560822,
-0.15805737674236298,
-0.021762151271104813,
-0.06296852231025696,
-0.2003057897090912,
-0.020524846389889717,
-0.020498162135481834,
0.11537028104066849,
0.06514350324869156,
0.16558827459812164,
-0.01439540646970272,
0.17640651762485504,
-0.09182710945606232,
-0.0403030700981617,
0.10994675755500793,
-0.00996874738484621,
0.04505027085542679,
-0.03313769772648811,
-0.004576151259243488,
0.040073905140161514,
-0.07132738083600998,
-0.00004555533450911753,
-0.009595246985554695,
-0.09472959488630295,
0.08290839940309525,
-0.03161881864070892,
-0.05139395222067833,
-0.012195432558655739,
0.02229711227118969,
0.0431261844933033,
0.09965251386165619,
0.09110171347856522,
0.0008972134673967957,
0.03825187310576439,
0.1696159988641739,
-0.05322735756635666,
-0.20851224660873413,
-0.12293823808431625,
0.23343653976917267,
0.009134521707892418,
0.06317983567714691,
-0.11318151652812958,
0.008156334049999714,
-0.0649232342839241,
0.27146127820014954,
0.10717649757862091,
-0.19682574272155762,
-0.08081234991550446,
-0.08638617396354675,
0.018265295773744583,
-0.07012990117073059,
0.09498134255409241,
0.10414329916238785,
0.15236273407936096,
-0.06956905871629715,
-0.11651330441236496,
-0.06961539387702942,
-0.04360228776931763,
-0.1502353847026825,
-0.03434385359287262,
0.11150668561458588,
0.0372641496360302,
0.007502933498471975,
0.15702307224273682,
0.0032468948047608137,
-0.04435136169195175,
-0.01925494708120823,
-0.08111487329006195,
-0.0955636203289032,
0.01875142753124237,
0.12386379390954971,
0.10656464099884033,
-0.01679341495037079,
-0.04251528158783913,
0.003736909246072173,
-0.03768129274249077,
-0.01515372097492218,
-0.003427404211834073,
-0.15350882709026337,
0.05015597864985466,
0.006577587220817804,
0.16203178465366364,
-0.005967925768345594,
0.026944084092974663,
0.055038053542375565,
-0.01440117321908474,
0.05213656276464462,
0.17896056175231934,
-0.009638361632823944,
-0.00454431539401412,
0.02588718757033348,
-0.04528478905558586,
-0.028109487146139145,
0.06294695287942886,
0.06118133291602135,
-0.06143670156598091,
0.08951634168624878,
-0.013488235883414745,
-0.08915087580680847,
-0.053045082837343216,
0.1191224604845047,
-0.07644791901111603,
0.0739174634218216,
0.07418970763683319,
0.013542285189032555,
0.11887470632791519,
-0.07554803788661957,
0.061413463205099106,
-0.03263699635863304,
-0.054895125329494476,
-0.07691919803619385,
-0.08487772196531296,
0.048925403505563736,
0.15646983683109283,
0.037179168313741684,
-0.15726208686828613,
0.023299768567085266,
-0.07527489215135574,
0.037523966282606125,
-0.07870914787054062,
0.07315753400325775,
0.10160929709672928,
-0.0011060814140364528,
-0.0366525799036026,
-0.29329177737236023,
-0.021167775616049767,
0.1207030862569809,
-0.04108953848481178,
-0.09119226783514023
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# Fine-Tuned_Model3
This model is a fine-tuned version of [google/vit-base-patch16-224](https://huggingface.co/google/vit-base-patch16-224) on the imagefolder dataset.
It achieves the following results on the evaluation set:
- Loss: 0.7362
- Accuracy: 0.608
- F1: 0.5096
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 32
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 30
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
| 3.2255 | 5.0 | 20 | 1.9574 | 0.512 | 0.3083 |
| 1.3773 | 10.0 | 40 | 0.8854 | 0.584 | 0.4617 |
| 0.869 | 15.0 | 60 | 0.7880 | 0.608 | 0.4795 |
| 0.7966 | 20.0 | 80 | 0.7732 | 0.6 | 0.4846 |
| 0.8458 | 25.0 | 100 | 0.7795 | 0.576 | 0.4112 |
| 0.8135 | 30.0 | 120 | 0.7362 | 0.608 | 0.5096 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.17.0
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["imagefolder"], "metrics": ["accuracy", "f1"], "base_model": "google/vit-base-patch16-224", "model-index": [{"name": "Fine-Tuned_Model3", "results": [{"task": {"type": "image-classification", "name": "Image Classification"}, "dataset": {"name": "imagefolder", "type": "imagefolder", "config": "default", "split": "train", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.608, "name": "Accuracy"}, {"type": "f1", "value": 0.5096170704866357, "name": "F1"}]}]}]} | image-classification | arpanl/Fine-Tuned_Model3 | [
"transformers",
"tensorboard",
"safetensors",
"vit",
"image-classification",
"generated_from_trainer",
"dataset:imagefolder",
"base_model:google/vit-base-patch16-224",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T06:31:34+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| Fine-Tuned\_Model3
==================
This model is a fine-tuned version of google/vit-base-patch16-224 on the imagefolder dataset.
It achieves the following results on the evaluation set:
* Loss: 0.7362
* Accuracy: 0.608
* F1: 0.5096
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 32
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 30
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.1.0+cu121
* Datasets 2.17.0
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 30",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 30",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
83,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 30### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
-0.12446990609169006,
0.129507377743721,
-0.0019213032210245728,
0.12662486732006073,
0.15288668870925903,
0.01589343510568142,
0.1468791514635086,
0.12362194061279297,
-0.05069790408015251,
0.05563133955001831,
0.14230184257030487,
0.1206381693482399,
0.027687858790159225,
0.15198321640491486,
-0.04559214413166046,
-0.2119566798210144,
0.019791072234511375,
0.04327630251646042,
-0.046364761888980865,
0.12765830755233765,
0.0900936871767044,
-0.13248717784881592,
0.11933479458093643,
0.022331325337290764,
-0.18325398862361908,
-0.003964535892009735,
0.03202195465564728,
-0.04332367330789566,
0.1324174851179123,
0.03552456572651863,
0.0978735163807869,
0.02175331301987171,
0.06453637778759003,
-0.15601572394371033,
0.012574677355587482,
0.07063796371221542,
-0.014266365207731724,
0.08765581995248795,
0.05266048386693001,
0.011137446388602257,
0.00881422683596611,
-0.08576571196317673,
0.03435087203979492,
0.022380733862519264,
-0.12149041146039963,
-0.23536978662014008,
-0.07614996284246445,
0.054249443113803864,
0.08459824323654175,
0.08706224709749222,
-0.007573842536658049,
0.14740124344825745,
-0.017424458637833595,
0.09631671756505966,
0.206583634018898,
-0.28068238496780396,
-0.07068989425897598,
0.039496712386608124,
0.016469191759824753,
0.08329284191131592,
-0.10345350950956345,
0.00469924695789814,
0.050439752638339996,
0.015737216919660568,
0.14078176021575928,
-0.008229064755141735,
-0.007895412854850292,
-0.010511058382689953,
-0.12304780632257462,
-0.055102817714214325,
0.19538910686969757,
0.09934432804584503,
-0.04980069026350975,
-0.06993235647678375,
-0.07336318492889404,
-0.1254146695137024,
-0.038552410900592804,
-0.00030480953864753246,
0.05138567090034485,
-0.02506234683096409,
-0.06470987200737,
-0.03525002300739288,
-0.11432155966758728,
-0.073687344789505,
-0.013013169169425964,
0.10386993736028671,
0.03894637152552605,
0.014372606761753559,
-0.017258964478969574,
0.0980173647403717,
-0.03189849853515625,
-0.14595608413219452,
0.008517615497112274,
0.02030741050839424,
0.01381322555243969,
-0.034323643893003464,
-0.03046242706477642,
-0.12448140978813171,
0.01186468917876482,
0.10573449730873108,
-0.042053643614053726,
0.044896434992551804,
0.000641263322904706,
0.0510115884244442,
-0.10696766525506973,
0.20064343512058258,
-0.06142492592334747,
-0.0035850610584020615,
0.033174172043800354,
0.11328618973493576,
0.058061230927705765,
-0.001012232038192451,
-0.11942891031503677,
0.01391695998609066,
0.11466732621192932,
-0.00019504253577906638,
-0.031428262591362,
0.07601800560951233,
-0.06949598342180252,
-0.03325918689370155,
0.08422911167144775,
-0.07142430543899536,
0.025134136900305748,
-0.002057180507108569,
-0.05462309718132019,
-0.06830760836601257,
0.04482667148113251,
0.008542426861822605,
-0.0017315612640231848,
0.05402912572026253,
-0.10618029534816742,
0.011458097025752068,
-0.06706549972295761,
-0.11108890920877457,
0.01021519023925066,
-0.09215980023145676,
0.02614584192633629,
-0.12654246389865875,
-0.1694798320531845,
-0.009430267848074436,
0.06535669416189194,
-0.025222796946763992,
-0.051418453454971313,
-0.03479591757059097,
-0.06646005809307098,
0.02205553837120533,
0.004212512169033289,
0.05073332414031029,
-0.06679338961839676,
0.09754706919193268,
0.04310159757733345,
0.06040900573134422,
-0.03180018439888954,
0.043122101575136185,
-0.09991294145584106,
0.05439642816781998,
-0.17272992432117462,
0.028111353516578674,
-0.044786155223846436,
0.09168227761983871,
-0.11143471300601959,
-0.0822773203253746,
-0.0031747871544212103,
-0.02262185700237751,
0.05857881158590317,
0.11359420418739319,
-0.13071398437023163,
-0.05784983932971954,
0.14921729266643524,
-0.094663105905056,
-0.1683247983455658,
0.12724772095680237,
-0.03288022428750992,
0.030696002766489983,
0.05650334432721138,
0.18203386664390564,
0.0815746933221817,
-0.11364946514368057,
-0.0013098404742777348,
-0.020876795053482056,
0.04385692998766899,
-0.05055753141641617,
0.09757508337497711,
0.008952989242970943,
-0.026227569207549095,
0.010817502625286579,
-0.10790888965129852,
0.0909220278263092,
-0.06929358839988708,
-0.08941235393285751,
-0.053489625453948975,
-0.09734133630990982,
0.05272938683629036,
0.06576033681631088,
0.057918012142181396,
-0.09509236365556717,
-0.09771917015314102,
0.02730569615960121,
0.08198035508394241,
-0.08161080628633499,
0.008383539505302906,
-0.07651208341121674,
0.1046188622713089,
-0.11888666450977325,
-0.009955791756510735,
-0.12509669363498688,
-0.04650842025876045,
0.03723626211285591,
-0.0550137422978878,
-0.01033865474164486,
-0.03219621628522873,
0.07792292535305023,
0.07269459962844849,
-0.07223165035247803,
-0.08478636294603348,
-0.0500405877828598,
0.003279293654486537,
-0.10091383010149002,
-0.1840447336435318,
-0.032992228865623474,
-0.025671295821666718,
0.15195678174495697,
-0.22839897871017456,
0.03862545266747475,
0.02600795403122902,
0.09010360389947891,
0.05914267897605896,
-0.02793765254318714,
-0.004861160181462765,
0.019440004602074623,
-0.040724411606788635,
-0.0844571441411972,
0.06266772001981735,
0.01867809146642685,
-0.09044285118579865,
-0.0038942787796258926,
-0.10219844430685043,
0.18507808446884155,
0.13238219916820526,
-0.0630134865641594,
-0.05629626661539078,
-0.0010833331616595387,
-0.03822780400514603,
-0.04075511172413826,
-0.03962963819503784,
-0.01500095147639513,
0.0760575383901596,
0.0013317709090188146,
0.1578223705291748,
-0.0958181619644165,
-0.01641342230141163,
0.04528525471687317,
-0.03091397136449814,
-0.03372574970126152,
0.09606140851974487,
0.07310453802347183,
-0.15766842663288116,
0.1508675366640091,
0.18705016374588013,
-0.06800266355276108,
0.12927453219890594,
-0.03519827127456665,
-0.06193378567695618,
-0.027951795607805252,
0.028275849297642708,
0.023057324811816216,
0.13836058974266052,
-0.1286085844039917,
-0.005343906115740538,
0.021844295784831047,
0.0001222498103743419,
-0.008412730880081654,
-0.21137556433677673,
-0.027237627655267715,
0.045783787965774536,
-0.04861275479197502,
0.011211560107767582,
-0.007813865318894386,
-0.02900841273367405,
0.07876042276620865,
0.015346208587288857,
-0.05289016664028168,
0.04860050603747368,
0.0022946633398532867,
-0.07488791644573212,
0.19678257405757904,
-0.07274562120437622,
-0.21142242848873138,
-0.14880000054836273,
-0.033020202070474625,
-0.07961511611938477,
0.03111588954925537,
0.05967757850885391,
-0.08676744252443314,
-0.0581037737429142,
-0.106836698949337,
-0.008947349153459072,
0.0355263352394104,
0.02779134176671505,
0.01873854361474514,
-0.008093610405921936,
0.12142468243837357,
-0.08986376225948334,
-0.005233029369264841,
-0.0042991419322788715,
-0.020306764170527458,
0.0410984605550766,
0.012866339646279812,
0.12324520200490952,
0.09403283149003983,
-0.012800537049770355,
0.018635917454957962,
-0.010816083289682865,
0.26125264167785645,
-0.08042465895414352,
-0.007046983111649752,
0.1633998453617096,
0.01185317151248455,
0.058411166071891785,
0.13183188438415527,
0.04938562214374542,
-0.09351592510938644,
0.01122781541198492,
0.023804103955626488,
-0.02285439521074295,
-0.17330783605575562,
-0.019812459126114845,
-0.043037448078393936,
-0.008120140060782433,
0.1400763988494873,
0.05220182612538338,
0.049268439412117004,
0.0944717526435852,
0.00023736615548841655,
0.0840202271938324,
-0.003454736666753888,
0.08085426688194275,
0.10280236601829529,
0.04264504089951515,
0.10607506334781647,
-0.03714367374777794,
-0.04065675660967827,
0.0323229543864727,
0.023626338690519333,
0.19703885912895203,
0.007992485538125038,
0.17255337536334991,
0.038481660187244415,
0.18260005116462708,
0.000506893964484334,
0.058484889566898346,
-0.019502371549606323,
-0.03603155165910721,
-0.009882834739983082,
-0.05329915136098862,
-0.02116975374519825,
0.044347215443849564,
-0.04437428340315819,
0.07293558120727539,
-0.08570196479558945,
0.019244100898504257,
0.06502804905176163,
0.2544978857040405,
0.05796429514884949,
-0.3784410357475281,
-0.10502169281244278,
0.007713588885962963,
-0.006771658547222614,
-0.06160619109869003,
0.0016164787812158465,
0.13813969492912292,
-0.055677350610494614,
0.0465666688978672,
-0.10224595665931702,
0.08082053810358047,
-0.04749390110373497,
0.028276877477765083,
0.08690560609102249,
0.06205376237630844,
0.005081302020698786,
0.06429342180490494,
-0.2263994663953781,
0.25784561038017273,
0.016889682039618492,
0.05690069496631622,
-0.04421854019165039,
0.0034482451155781746,
0.04437027499079704,
0.12186136841773987,
0.1060045063495636,
0.000847899354994297,
0.033496323972940445,
-0.1936575323343277,
-0.08376704901456833,
0.010286716744303703,
0.05374456197023392,
-0.06304527074098587,
0.08527480065822601,
-0.038238275796175,
-0.022585542872548103,
0.057192448526620865,
0.033240146934986115,
-0.08222229778766632,
-0.08757622539997101,
-0.004160361364483833,
0.04772787168622017,
0.021192006766796112,
-0.09336397051811218,
-0.09799030423164368,
-0.11113277822732925,
0.1269754022359848,
-0.011668638326227665,
-0.033801570534706116,
-0.1135840192437172,
0.08733093738555908,
0.05165061354637146,
-0.08948570489883423,
0.06467155367136002,
-0.02029864490032196,
0.12401539087295532,
0.03456760197877884,
-0.05384058505296707,
0.11517540365457535,
-0.06140716373920441,
-0.16430450975894928,
-0.05509275197982788,
0.10170482099056244,
-0.013251964934170246,
0.023595239967107773,
-0.0033504096791148186,
0.028041301295161247,
-0.0338953398168087,
-0.05428002402186394,
0.06113075464963913,
-0.017650453373789787,
0.05152374505996704,
0.010361327789723873,
-0.01325614470988512,
0.007273997645825148,
-0.06311619281768799,
-0.03556768596172333,
0.13043126463890076,
0.25850915908813477,
-0.08573324233293533,
-0.02106967568397522,
0.028896871954202652,
-0.039927560836076736,
-0.19538260996341705,
0.04434452950954437,
0.04759104922413826,
0.008583891205489635,
0.05656716600060463,
-0.14634409546852112,
0.08279015123844147,
0.08018620312213898,
-0.03703535348176956,
0.10063766688108444,
-0.26461461186408997,
-0.1304944008588791,
0.09392960369586945,
0.18875017762184143,
0.08307874202728271,
-0.1472942978143692,
-0.04674634337425232,
-0.025662899017333984,
-0.12584297358989716,
0.09403228759765625,
-0.0761321485042572,
0.10556157678365707,
-0.01924440637230873,
0.013949833810329437,
0.010041629895567894,
-0.059322331100702286,
0.1434849500656128,
-0.03698498383164406,
0.10756490379571915,
-0.05370619148015976,
-0.024013500660657883,
0.07180597633123398,
-0.06922794133424759,
0.054013606160879135,
-0.10025609284639359,
0.05980385094881058,
-0.06857746094465256,
-0.017378855496644974,
-0.07251343131065369,
0.03288408741354942,
-0.020364202558994293,
-0.03831639885902405,
-0.03703080490231514,
0.03854287415742874,
0.04615829139947891,
0.0011842948151752353,
0.21100644767284393,
0.041878774762153625,
0.09202823042869568,
0.12658526003360748,
0.044137634336948395,
-0.06888502091169357,
-0.09343378245830536,
-0.03976429998874664,
-0.030743993818759918,
0.08546317368745804,
-0.16806606948375702,
0.048236358910799026,
0.0936535894870758,
0.011559519916772842,
0.15934884548187256,
0.04829983785748482,
-0.027193034067749977,
0.01932169497013092,
0.07470571249723434,
-0.16227149963378906,
-0.15659014880657196,
-0.046637214720249176,
-0.028956718742847443,
-0.11320284754037857,
0.05161258578300476,
0.12511812150478363,
-0.08010450005531311,
0.00545159075409174,
-0.007573009468615055,
0.018892483785748482,
-0.014883878640830517,
0.1584310233592987,
0.07666274160146713,
0.040405165404081345,
-0.08410751074552536,
0.08901993185281754,
0.06090889871120453,
-0.10317815095186234,
-0.0008137666154652834,
0.013328644447028637,
-0.09995357692241669,
-0.04933781921863556,
0.042444318532943726,
0.1493636518716812,
-0.02106397971510887,
-0.05005848780274391,
-0.15005941689014435,
-0.09014321863651276,
0.041332729160785675,
0.09723043441772461,
0.09567543119192123,
0.017485329881310463,
-0.015897605568170547,
-0.0011914760107174516,
-0.09248141199350357,
0.12587836384773254,
0.03339752182364464,
0.09657955169677734,
-0.22284865379333496,
0.05119882896542549,
0.014243138954043388,
0.034060291945934296,
-0.017996611073613167,
0.03297320380806923,
-0.09376145154237747,
-0.01143172848969698,
-0.08976585417985916,
0.021460596472024918,
-0.026148958131670952,
0.011738424189388752,
-0.0069922697730362415,
-0.07742037624120712,
-0.06139185279607773,
0.03432033583521843,
-0.10138248652219772,
-0.044144243001937866,
0.04675999656319618,
0.06877566128969193,
-0.10228381305932999,
-0.04146941751241684,
0.03799721226096153,
-0.07735386490821838,
0.083299919962883,
0.01764766126871109,
0.009800580330193043,
0.030300026759505272,
-0.12282540649175644,
0.019968179985880852,
0.07454811781644821,
0.01040982361882925,
0.027345696464180946,
-0.10429815202951431,
0.0017888537840917706,
-0.008533187210559845,
-0.002059935126453638,
-0.006580295506864786,
0.11464697122573853,
-0.12921856343746185,
-0.02460494451224804,
-0.02657341957092285,
-0.03888733312487602,
-0.05962607264518738,
0.052827466279268265,
0.07758574187755585,
0.00510289566591382,
0.21243546903133392,
-0.08224163949489594,
0.003187467809766531,
-0.22441954910755157,
0.007509519811719656,
-0.006294675171375275,
-0.13191013038158417,
-0.13345113396644592,
-0.039739374071359634,
0.04788896441459656,
-0.07383419573307037,
0.0923924595117569,
0.017537767067551613,
0.025808246806263924,
0.03521474450826645,
0.02044989913702011,
0.0330023467540741,
0.021320119500160217,
0.19295980036258698,
-0.00830954872071743,
-0.02221623808145523,
0.06051652505993843,
0.022921210154891014,
0.11702732741832733,
0.1005043089389801,
0.09298844635486603,
0.15990161895751953,
-0.043495096266269684,
0.09779485315084457,
0.051023293286561966,
-0.025866391137242317,
-0.17335215210914612,
0.060821689665317535,
-0.0691339522600174,
0.1375378668308258,
-0.014550500549376011,
0.18704253435134888,
0.11293499916791916,
-0.15813332796096802,
0.017961613833904266,
-0.04506506025791168,
-0.06884658336639404,
-0.06136027351021767,
-0.13302385807037354,
-0.10690005123615265,
-0.17642317712306976,
0.013303551822900772,
-0.09066315740346909,
-0.003675914602354169,
0.0851161926984787,
-0.011685999110341072,
-0.03417177125811577,
0.21090342104434967,
0.04004770517349243,
-0.0004229727783240378,
0.06182803213596344,
0.0021484855096787214,
-0.06451139599084854,
-0.055819593369960785,
-0.08233519643545151,
0.03809230402112007,
-0.013416395522654057,
0.024982428178191185,
-0.04498783499002457,
-0.018892770633101463,
0.04934072867035866,
0.008632884360849857,
-0.11823468655347824,
0.01086986344307661,
0.012503426522016525,
0.011200704611837864,
0.010433821938931942,
0.002900701714679599,
0.027228711172938347,
-0.007801559753715992,
0.20662690699100494,
-0.05804495885968208,
-0.0058817476965487,
-0.11095121502876282,
0.12471267580986023,
0.010880470275878906,
-0.01461742538958788,
0.024950774386525154,
-0.0944819450378418,
0.03753110021352768,
0.21701210737228394,
0.14632250368595123,
-0.05925101786851883,
-0.0011222191387787461,
-0.004077531863003969,
-0.01660848595201969,
-0.03317364677786827,
0.08015899360179901,
0.08734561502933502,
-0.037994448095560074,
-0.06637781858444214,
-0.012096582911908627,
-0.054145727306604385,
-0.017340045422315598,
-0.02449187822639942,
0.04226125404238701,
0.022890213876962662,
0.020564353093504906,
-0.07593516260385513,
0.057063180953264236,
0.023980433121323586,
-0.08439626544713974,
0.0900256484746933,
-0.1948387622833252,
-0.13959269225597382,
-0.038514114916324615,
0.11448965221643448,
-0.005709950812160969,
0.037430766969919205,
-0.032104332000017166,
0.018337735906243324,
0.05775485932826996,
-0.020015837624669075,
-0.07753542065620422,
-0.09810028970241547,
0.059192415326833725,
-0.11370787769556046,
0.25859442353248596,
-0.037749920040369034,
0.0065338267013430595,
0.11144718527793884,
0.021582020446658134,
-0.11178795993328094,
0.06574095040559769,
0.030086811631917953,
-0.045323073863983154,
0.03228248283267021,
0.0987934023141861,
-0.023412466049194336,
0.11753619462251663,
0.03236446529626846,
-0.10024085640907288,
-0.007582260761409998,
-0.03646332398056984,
-0.05270697921514511,
-0.05449246987700462,
-0.030874567106366158,
-0.06675302982330322,
0.1329830437898636,
0.16903409361839294,
-0.042398497462272644,
-0.030223188921809196,
-0.0629587396979332,
0.03204810619354248,
0.10071404278278351,
0.019370686262845993,
-0.018434368073940277,
-0.21717782318592072,
0.02154693752527237,
0.016397010535001755,
0.0035492663737386465,
-0.2306765764951706,
-0.10576009005308151,
-0.016479775309562683,
-0.05362142249941826,
-0.07445240020751953,
0.08755197376012802,
0.12134641408920288,
0.04460898041725159,
-0.06246678903698921,
-0.048022013157606125,
-0.06849819421768188,
0.15567147731781006,
-0.12039714306592941,
-0.10227351635694504
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bert-base-uncased-finetuned-amazon_reviews_multi
This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2149
- Accuracy: 0.9402
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.215 | 1.0 | 1250 | 0.1709 | 0.9352 |
| 0.136 | 2.0 | 2500 | 0.2149 | 0.9402 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "bert-base-uncased", "model-index": [{"name": "bert-base-uncased-finetuned-amazon_reviews_multi", "results": []}]} | text-classification | JoelVIU/bert-base-uncased-finetuned-amazon_reviews_multi | [
"transformers",
"tensorboard",
"safetensors",
"bert",
"text-classification",
"generated_from_trainer",
"base_model:bert-base-uncased",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T06:32:09+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #bert #text-classification #generated_from_trainer #base_model-bert-base-uncased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| bert-base-uncased-finetuned-amazon\_reviews\_multi
==================================================
This model is a fine-tuned version of bert-base-uncased on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.2149
* Accuracy: 0.9402
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 2
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #bert #text-classification #generated_from_trainer #base_model-bert-base-uncased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
68,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #bert #text-classification #generated_from_trainer #base_model-bert-base-uncased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.08821936696767807,
0.09809461236000061,
-0.0021079694852232933,
0.10780711472034454,
0.14591750502586365,
0.025153733789920807,
0.15749092400074005,
0.11449519544839859,
-0.0660737082362175,
0.04462523013353348,
0.12498024106025696,
0.13014787435531616,
0.010781425051391125,
0.11824779957532883,
-0.0688638985157013,
-0.22624698281288147,
0.006858686916530132,
0.03052506409585476,
-0.06151825934648514,
0.11174045503139496,
0.09091722220182419,
-0.1227058470249176,
0.08985479176044464,
-0.007647586986422539,
-0.1784372478723526,
0.010598988272249699,
0.01815725490450859,
-0.050824832171201706,
0.13288240134716034,
0.03631322458386421,
0.1355244517326355,
0.020718514919281006,
0.0924074575304985,
-0.20263484120368958,
0.00901892688125372,
0.05794676020741463,
-0.008503382094204426,
0.08363906294107437,
0.033494532108306885,
0.013221297413110733,
0.08115286380052567,
-0.07914108037948608,
0.06384821981191635,
0.02069321647286415,
-0.11710357666015625,
-0.20381955802440643,
-0.07931444048881531,
0.040992893278598785,
0.08815018832683563,
0.07689282298088074,
-0.00947615597397089,
0.12117790430784225,
-0.060052573680877686,
0.08771209418773651,
0.2118978351354599,
-0.31896862387657166,
-0.06338334828615189,
0.052760761231184006,
0.030388271436095238,
0.07767406105995178,
-0.10548879206180573,
-0.02290228009223938,
0.06965646147727966,
0.024832775816321373,
0.12458054721355438,
-0.02536807395517826,
-0.07750265300273895,
0.010285105556249619,
-0.14765234291553497,
-0.015367607586085796,
0.1473287045955658,
0.04801209643483162,
-0.04157990217208862,
-0.04660116881132126,
-0.06590453535318375,
-0.14732636511325836,
-0.03808266296982765,
-0.01969512365758419,
0.05130979046225548,
-0.022161735221743584,
-0.06162091717123985,
-0.01417057029902935,
-0.1044955626130104,
-0.08125218749046326,
-0.06277438253164291,
0.14707915484905243,
0.038849711418151855,
0.003335579764097929,
-0.013267896138131618,
0.10006021708250046,
-0.04051513969898224,
-0.12962062656879425,
0.018970834091305733,
0.026263734325766563,
0.008208484388887882,
-0.056395430117845535,
-0.05395534634590149,
-0.06931990385055542,
0.023234574124217033,
0.15193040668964386,
-0.04701982066035271,
0.04696967080235481,
0.005437387619167566,
0.05059683695435524,
-0.09862755239009857,
0.1663200408220291,
-0.039466504007577896,
-0.020560111850500107,
0.016065720468759537,
0.07381491363048553,
0.0421258881688118,
-0.007070035673677921,
-0.12953227758407593,
0.02905821055173874,
0.10247618705034256,
0.01632522977888584,
-0.07143504172563553,
0.07903409004211426,
-0.04946732893586159,
-0.0028088544495403767,
0.020341647788882256,
-0.08799847960472107,
0.03240729868412018,
0.006001811008900404,
-0.05230124294757843,
-0.06255311518907547,
0.032950304448604584,
0.01947365142405033,
0.00715111568570137,
0.11068718135356903,
-0.08985814452171326,
0.008104035630822182,
-0.08973299711942673,
-0.12117727100849152,
0.02451775223016739,
-0.08141767978668213,
0.02375674992799759,
-0.10816492885351181,
-0.16751441359519958,
-0.0054976604878902435,
0.059343986213207245,
-0.02870982140302658,
-0.030264416709542274,
-0.056223031133413315,
-0.07265924662351608,
0.017908671870827675,
-0.018643934279680252,
0.08175715059041977,
-0.06377504765987396,
0.09462112188339233,
0.03777455911040306,
0.06722621619701385,
-0.051700737327337265,
0.03988759219646454,
-0.09917894005775452,
0.026675347238779068,
-0.17606721818447113,
0.015011966228485107,
-0.07579712569713593,
0.06145606189966202,
-0.07972010970115662,
-0.07340698689222336,
-0.0012638693442568183,
0.011607491411268711,
0.07106723636388779,
0.08860042691230774,
-0.16015872359275818,
-0.0676279067993164,
0.17137467861175537,
-0.09353213757276535,
-0.14034752547740936,
0.12733088433742523,
-0.0564877949655056,
0.0584409236907959,
0.05778200179338455,
0.1792062222957611,
0.05890676751732826,
-0.09646939486265182,
-0.0020821434445679188,
0.004002843052148819,
0.062201693654060364,
-0.04048023745417595,
0.06361686438322067,
0.0014583896845579147,
-0.004227496217936277,
0.017525121569633484,
-0.047704458236694336,
0.0495329350233078,
-0.0814017578959465,
-0.08679412305355072,
-0.04215218126773834,
-0.10053694248199463,
0.05435798317193985,
0.053404323756694794,
0.06919069588184357,
-0.1065114215016365,
-0.09243448823690414,
0.07872001826763153,
0.07277508825063705,
-0.07506780326366425,
0.020097365602850914,
-0.06927628815174103,
0.08442260324954987,
-0.05864223837852478,
-0.01632649451494217,
-0.15273237228393555,
-0.05328790098428726,
0.018840443342924118,
-0.006181970238685608,
0.01950100064277649,
0.007178320549428463,
0.07060573995113373,
0.07798291742801666,
-0.06923655420541763,
-0.021184511482715607,
-0.019107306376099586,
0.016681533306837082,
-0.12675397098064423,
-0.2060975432395935,
-0.01819251850247383,
-0.03333589434623718,
0.1287437081336975,
-0.23311661183834076,
0.05282003432512283,
0.0001520049263490364,
0.08656046539545059,
0.034034524112939835,
-0.0067826793529093266,
-0.048736535012722015,
0.0719003900885582,
-0.04602237790822983,
-0.056110698729753494,
0.0604894682765007,
0.009277703240513802,
-0.09090255200862885,
-0.040636345744132996,
-0.13938844203948975,
0.18436960875988007,
0.1376413255929947,
-0.0935848131775856,
-0.08183128386735916,
-0.00922318920493126,
-0.04237379506230354,
-0.030043015256524086,
-0.047588691115379333,
0.002742685843259096,
0.1420900672674179,
-0.015648555010557175,
0.1526661217212677,
-0.08264487981796265,
-0.03821792080998421,
0.02428801730275154,
-0.04561246186494827,
0.01032283902168274,
0.10439632087945938,
0.12213639169931412,
-0.11117558926343918,
0.1548563539981842,
0.17507940530776978,
-0.1007125973701477,
0.13650986552238464,
-0.04207636043429375,
-0.058749400079250336,
-0.02093697339296341,
0.0022936661262065172,
0.008908234536647797,
0.11307938396930695,
-0.13880480825901031,
-0.0007407191442325711,
0.006852506659924984,
0.013699830509722233,
0.01889962889254093,
-0.22065246105194092,
-0.027393585070967674,
0.035473473370075226,
-0.048603884875774384,
0.0019909930415451527,
-0.023336075246334076,
-0.01053670048713684,
0.09874564409255981,
-0.0015580495819449425,
-0.0863431766629219,
0.044455211609601974,
-0.0026169675402343273,
-0.08687146753072739,
0.21001768112182617,
-0.08401673287153244,
-0.12565883994102478,
-0.13848590850830078,
-0.07322285324335098,
-0.0410422682762146,
0.02396087907254696,
0.0656314268708229,
-0.07566198706626892,
-0.04255380854010582,
-0.10077830404043198,
0.004386895336210728,
0.03372437134385109,
0.03174688294529915,
0.019380351528525352,
0.003436225699260831,
0.07925098389387131,
-0.10626373440027237,
-0.010212716646492481,
-0.048599425703287125,
-0.0659560039639473,
0.028214843943715096,
0.028867779299616814,
0.11316850036382675,
0.14832165837287903,
-0.027263004332780838,
-0.004089855123311281,
-0.03220106288790703,
0.21811847388744354,
-0.05702028051018715,
-0.02160682901740074,
0.1178317591547966,
-0.03466589376330376,
0.04742473363876343,
0.13749682903289795,
0.06987863034009933,
-0.09346596151590347,
0.01833818294107914,
0.041005831211805344,
-0.02792939729988575,
-0.22154657542705536,
-0.03615972399711609,
-0.037909623235464096,
0.009910139255225658,
0.09840597957372665,
0.03510536253452301,
0.03012215532362461,
0.06535103917121887,
0.02980688214302063,
0.08439723402261734,
-0.014464019797742367,
0.07422810047864914,
0.12032734602689743,
0.03816685080528259,
0.12478221952915192,
-0.045660778880119324,
-0.05562494695186615,
0.03721786290407181,
0.0018160628387704492,
0.20653513073921204,
0.027421025559306145,
0.13816708326339722,
0.05816882103681564,
0.1511758267879486,
0.001512958318926394,
0.06847235560417175,
-0.015352229587733746,
-0.04593703895807266,
-0.011620619334280491,
-0.05018359050154686,
-0.025342168286442757,
0.04311951994895935,
-0.09436359256505966,
0.05500585585832596,
-0.10889514535665512,
0.009087524376809597,
0.059797246009111404,
0.23491083085536957,
0.04834276810288429,
-0.3150748014450073,
-0.0906156599521637,
0.02405450865626335,
-0.030819697305560112,
-0.0212396327406168,
0.03358348831534386,
0.12604863941669464,
-0.050491638481616974,
0.028470691293478012,
-0.07000349462032318,
0.08264011889696121,
-0.03858411684632301,
0.04513039067387581,
0.06958530843257904,
0.08507660776376724,
-0.007698400411754847,
0.07291840016841888,
-0.26702943444252014,
0.277108758687973,
0.01321440190076828,
0.06893313676118851,
-0.05181502550840378,
0.001937482156790793,
0.03380262851715088,
0.08721717447042465,
0.07341396808624268,
-0.021788926795125008,
-0.05907578393816948,
-0.1952105611562729,
-0.05565241351723671,
0.027072735130786896,
0.09348703920841217,
-0.031210867688059807,
0.0972396582365036,
-0.03514719754457474,
0.0026737770531326532,
0.0897209420800209,
-0.00960580538958311,
-0.07770346105098724,
-0.0981016755104065,
-0.014588744379580021,
0.03486288711428642,
-0.029296021908521652,
-0.07988514006137848,
-0.10492546856403351,
-0.13639704883098602,
0.1643073409795761,
-0.061190150678157806,
-0.021957024931907654,
-0.09562335908412933,
0.06545883417129517,
0.05191923305392265,
-0.07664451748132706,
0.05090055242180824,
0.009187201969325542,
0.08947570621967316,
0.01897364854812622,
-0.06215168535709381,
0.12776166200637817,
-0.07512927800416946,
-0.17320257425308228,
-0.07920318841934204,
0.0958791971206665,
0.021648332476615906,
0.04671391844749451,
-0.0022440776228904724,
0.0047818077728152275,
-0.012367071583867073,
-0.0797639712691307,
0.022681469097733498,
0.0007458238978870213,
0.06235652416944504,
0.011078170500695705,
-0.08222783356904984,
-0.0027863921131938696,
-0.05041421204805374,
-0.03406648337841034,
0.16449354588985443,
0.27297326922416687,
-0.09236051887273788,
0.004278761800378561,
0.06391842663288116,
-0.07241642475128174,
-0.21205702424049377,
0.03576347976922989,
0.036130961030721664,
0.0014419876970350742,
0.04028692469000816,
-0.1491943895816803,
0.12007444351911545,
0.10506334155797958,
-0.027964286506175995,
0.10542469471693039,
-0.277768075466156,
-0.135947123169899,
0.13596035540103912,
0.1506761908531189,
0.11337091028690338,
-0.15707312524318695,
-0.035385772585868835,
-0.03348167613148689,
-0.11186333745718002,
0.11375164985656738,
-0.12434922903776169,
0.11344508081674576,
-0.008808795362710953,
0.05902069807052612,
0.002432279521599412,
-0.05712321773171425,
0.12503811717033386,
-0.0060218218713998795,
0.10940933227539062,
-0.06171596050262451,
-0.030515005812048912,
0.04153886064887047,
-0.05239465460181236,
0.02030738815665245,
-0.10900434851646423,
0.03130079805850983,
-0.055156148970127106,
-0.03105039894580841,
-0.042748864740133286,
0.03672593832015991,
-0.03890617564320564,
-0.06554461270570755,
-0.039039477705955505,
0.02599671296775341,
0.036940596997737885,
-0.013298436999320984,
0.15004564821720123,
0.01113475114107132,
0.156352236866951,
0.14374634623527527,
0.08067628741264343,
-0.06797241419553757,
-0.022156821563839912,
-0.0022511091083288193,
-0.03805345296859741,
0.0685640200972557,
-0.15535643696784973,
0.03870537877082825,
0.12162525206804276,
0.011402474716305733,
0.15100182592868805,
0.07768996059894562,
-0.030403388664126396,
0.009675546549260616,
0.06837187707424164,
-0.15827052295207977,
-0.09366682916879654,
-0.0027398259844630957,
-0.031798020005226135,
-0.12072677910327911,
0.0667022168636322,
0.11465099453926086,
-0.07439368218183517,
0.007812743075191975,
-0.0037630703300237656,
0.011191878467798233,
-0.0434817373752594,
0.17357321083545685,
0.06630565971136093,
0.04696260765194893,
-0.07630450278520584,
0.07893523573875427,
0.046070970594882965,
-0.06760977953672409,
0.01531414408236742,
0.040935274213552475,
-0.0810401439666748,
-0.04651634022593498,
0.054560787975788116,
0.19822558760643005,
-0.032528553158044815,
-0.05474582687020302,
-0.14191384613513947,
-0.11598224192857742,
0.05839838087558746,
0.19776201248168945,
0.10127414762973785,
0.011168216355144978,
-0.03768888860940933,
0.023614218458533287,
-0.1132802739739418,
0.11153186857700348,
0.025900235399603844,
0.0841323509812355,
-0.15038317441940308,
0.11795246601104736,
0.0006963465129956603,
0.005906894337385893,
-0.026831485331058502,
0.04983964189887047,
-0.12790924310684204,
-0.008869478479027748,
-0.13094346225261688,
-0.007682700641453266,
-0.023757021874189377,
0.008830207400023937,
0.009432445280253887,
-0.05695900321006775,
-0.062312815338373184,
0.010071545839309692,
-0.10038956999778748,
-0.015987316146492958,
0.04034077748656273,
0.061626747250556946,
-0.12712348997592926,
-0.036727406084537506,
0.027221061289310455,
-0.0669950395822525,
0.06522275507450104,
0.0207328163087368,
0.02630869671702385,
0.05736708268523216,
-0.19031579792499542,
0.03169067203998566,
0.06822062283754349,
0.014125912450253963,
0.047573819756507874,
-0.08540370315313339,
-0.014358744025230408,
-0.006878476589918137,
0.045155227184295654,
0.021695246919989586,
0.08769465237855911,
-0.1275889277458191,
0.007107523735612631,
-0.02790873311460018,
-0.067097969353199,
-0.049892351031303406,
0.022061320021748543,
0.09071356803178787,
-0.0030249753035604954,
0.19711558520793915,
-0.0966540053486824,
0.016944056376814842,
-0.20338411629199982,
0.010568229481577873,
0.0025422577746212482,
-0.10719885677099228,
-0.11673707515001297,
-0.05981363728642464,
0.04378724843263626,
-0.060750920325517654,
0.15505515038967133,
0.014402170665562153,
0.017496244981884956,
0.03561726585030556,
-0.04065112769603729,
0.032598309218883514,
0.028411362320184708,
0.22462472319602966,
0.03305928781628609,
-0.03668489679694176,
0.010659845545887947,
0.033161722123622894,
0.10899409651756287,
0.06846455484628677,
0.17073896527290344,
0.1537363976240158,
-0.05340225249528885,
0.10154235363006592,
0.054445311427116394,
-0.06129519268870354,
-0.1375381052494049,
0.06657954305410385,
-0.04223598539829254,
0.0997837707400322,
-0.02546972595155239,
0.21023203432559967,
0.09028708934783936,
-0.1586644947528839,
0.013194099068641663,
-0.054783646017313004,
-0.08286139369010925,
-0.11673448234796524,
-0.04996589198708534,
-0.09627753496170044,
-0.1543273776769638,
0.004732871428132057,
-0.11352668702602386,
0.004941120743751526,
0.09254641085863113,
0.004893855191767216,
-0.012210378423333168,
0.16823574900627136,
-0.00123738928232342,
0.039961203932762146,
0.05205385014414787,
0.0055531286634504795,
-0.037593334913253784,
-0.10542110353708267,
-0.0880708172917366,
-0.0009550141403451562,
-0.016612067818641663,
0.01844264753162861,
-0.04991568252444267,
-0.023858286440372467,
0.04216134920716286,
-0.0032363790087401867,
-0.09834069013595581,
0.009999154135584831,
0.022287188097834587,
0.04942942410707474,
0.04441635683178902,
0.0010855391155928373,
0.012013304978609085,
0.0012162273051217198,
0.20792999863624573,
-0.07758822292089462,
-0.06202046573162079,
-0.10180255025625229,
0.21907423436641693,
0.026211557909846306,
0.020504504442214966,
0.01029377244412899,
-0.08813335746526718,
0.022410109639167786,
0.2310464084148407,
0.1872996985912323,
-0.07772710919380188,
0.0023262270260602236,
0.004514951258897781,
-0.011113518849015236,
-0.03660454601049423,
0.09368275105953217,
0.11731799691915512,
0.02766517736017704,
-0.07622034847736359,
-0.05645941197872162,
-0.03087715432047844,
-0.004379736725240946,
-0.03872079774737358,
0.055310629308223724,
0.03784023970365524,
0.010000905022025108,
-0.05204227194190025,
0.04854296147823334,
-0.03591065853834152,
-0.10960590094327927,
0.060292523354291916,
-0.19849641621112823,
-0.14837056398391724,
-0.009391087107360363,
0.12218066304922104,
-0.018638990819454193,
0.048747189342975616,
-0.03068026527762413,
-0.0071801720187067986,
0.0764058455824852,
-0.02678844891488552,
-0.07332267612218857,
-0.0728640928864479,
0.056646909564733505,
-0.08879389613866806,
0.23830202221870422,
-0.037625864148139954,
0.055200982838869095,
0.12921395897865295,
0.0461752712726593,
-0.06570279598236084,
0.07947782427072525,
0.04697827622294426,
-0.08085930347442627,
0.024108942598104477,
0.07074571400880814,
-0.04554462060332298,
0.12392577528953552,
0.052280426025390625,
-0.14050982892513275,
0.020552009344100952,
-0.056154798716306686,
-0.09566183388233185,
-0.053066007792949677,
-0.03543373569846153,
-0.06281024217605591,
0.13681168854236603,
0.19293922185897827,
-0.031091608107089996,
0.0010587273864075541,
-0.04643556475639343,
0.029022058472037315,
0.06430462747812271,
0.03566000983119011,
-0.033347636461257935,
-0.23287993669509888,
0.03059813193976879,
0.07467631250619888,
-0.006145186722278595,
-0.2812756597995758,
-0.08907955139875412,
0.001761922030709684,
-0.04736654832959175,
-0.10096511244773865,
0.07462041079998016,
0.1224583312869072,
0.053849443793296814,
-0.06563583761453629,
-0.10593988001346588,
-0.0717066079378128,
0.14902380108833313,
-0.1317976415157318,
-0.10141246765851974
] |
null | null | transformers | # ITT-AF/ITT-42dot_LLM-SFT-1.3B-v2.0
This model is a fine-tuned version of [42dot/42dot_LLM-SFT-1.3B](https://huggingface.co/42dot/42dot_LLM-SFT-1.3B) on an custom dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 24
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 96
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 1.0
- mixed_precision_training: Native AMP
### Training results
### Framework versions
- Transformers 4.36.2
- Pytorch 2.1.2+cu121
- Datasets 2.0.0
- Tokenizers 0.15.0 | {"license": "cc-by-nc-4.0"} | text-generation | ITT-AF/ITT-42dot_LLM-SFT-1.3B-v2.0 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T06:34:04+00:00 | [] | [] | TAGS
#transformers #safetensors #llama #text-generation #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # ITT-AF/ITT-42dot_LLM-SFT-1.3B-v2.0
This model is a fine-tuned version of 42dot/42dot_LLM-SFT-1.3B on an custom dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 24
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 96
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 1.0
- mixed_precision_training: Native AMP
### Training results
### Framework versions
- Transformers 4.36.2
- Pytorch 2.1.2+cu121
- Datasets 2.0.0
- Tokenizers 0.15.0 | [
"# ITT-AF/ITT-42dot_LLM-SFT-1.3B-v2.0\n\nThis model is a fine-tuned version of 42dot/42dot_LLM-SFT-1.3B on an custom dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 24\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 96\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1.0\n- mixed_precision_training: Native AMP",
"### Training results",
"### Framework versions\n\n- Transformers 4.36.2\n- Pytorch 2.1.2+cu121\n- Datasets 2.0.0\n- Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# ITT-AF/ITT-42dot_LLM-SFT-1.3B-v2.0\n\nThis model is a fine-tuned version of 42dot/42dot_LLM-SFT-1.3B on an custom dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 24\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 96\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1.0\n- mixed_precision_training: Native AMP",
"### Training results",
"### Framework versions\n\n- Transformers 4.36.2\n- Pytorch 2.1.2+cu121\n- Datasets 2.0.0\n- Tokenizers 0.15.0"
] | [
58,
51,
6,
12,
8,
3,
126,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# ITT-AF/ITT-42dot_LLM-SFT-1.3B-v2.0\n\nThis model is a fine-tuned version of 42dot/42dot_LLM-SFT-1.3B on an custom dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 24\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 96\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.36.2\n- Pytorch 2.1.2+cu121\n- Datasets 2.0.0\n- Tokenizers 0.15.0"
] | [
-0.12907032668590546,
0.11368664354085922,
-0.0007358461734838784,
0.07209643721580505,
0.11815119534730911,
0.022628897801041603,
0.09780332446098328,
0.13913516700267792,
-0.10154646635055542,
0.05422145128250122,
0.06743334978818893,
0.01924017071723938,
0.06429591029882431,
0.1335143744945526,
-0.029287287965416908,
-0.22299472987651825,
-0.00898036640137434,
0.013662722893059254,
-0.08527123183012009,
0.11087022721767426,
0.1161140501499176,
-0.06912861764431,
0.05173397436738014,
0.009454395622015,
-0.14025944471359253,
0.024514921009540558,
-0.03460788354277611,
-0.05469023063778877,
0.09284454584121704,
0.011471184901893139,
0.0965648964047432,
-0.007157579530030489,
0.13316814601421356,
-0.20064029097557068,
0.0022397488355636597,
0.083829365670681,
0.03428516909480095,
0.09689412266016006,
0.052876658737659454,
0.008172937668859959,
0.0823797807097435,
-0.13826528191566467,
0.07959207892417908,
0.013049245811998844,
-0.07310262322425842,
-0.18336065113544464,
-0.11646417528390884,
0.06827830523252487,
0.11561716347932816,
0.08680810779333115,
0.012938217259943485,
0.1564183533191681,
-0.055930349975824356,
0.06339891999959946,
0.1071944609284401,
-0.2512159049510956,
-0.07600811868906021,
0.06117215380072594,
0.04315103963017464,
0.08037770539522171,
-0.10796728730201721,
0.004831472411751747,
0.06900867074728012,
0.02981896512210369,
0.09781323373317719,
-0.0009205675451084971,
-0.038526542484760284,
-0.019745392724871635,
-0.13585391640663147,
-0.04884503036737442,
0.19209840893745422,
0.0748993530869484,
-0.0734449177980423,
-0.09917287528514862,
-0.025594353675842285,
-0.11872808635234833,
-0.032564010471105576,
-0.019154109060764313,
0.03951302170753479,
-0.038239557296037674,
-0.0595049187541008,
-0.054750192910432816,
-0.10472416132688522,
-0.09943345189094543,
0.015748487785458565,
0.13551799952983856,
0.02031705714762211,
0.00024768029106780887,
-0.025317436084151268,
0.10899856686592102,
-0.0015124890487641096,
-0.11940012127161026,
-0.039209842681884766,
0.00823961105197668,
-0.10030034929513931,
-0.08304551988840103,
-0.02905604988336563,
-0.052833158522844315,
-0.02719925343990326,
0.1364859938621521,
-0.027987267822027206,
0.03664012253284454,
0.014747700653970242,
0.003661748254671693,
-0.023160045966506004,
0.1261483132839203,
-0.06114886701107025,
-0.02570309117436409,
-0.010454901494085789,
0.12392277270555496,
0.021781379356980324,
-0.025427762418985367,
-0.07458508014678955,
-0.05188830569386482,
0.10494745522737503,
0.053985845297575,
-0.025738133117556572,
0.029078440740704536,
-0.03162715584039688,
-0.032824937254190445,
0.04034445807337761,
-0.13110913336277008,
0.05339319631457329,
-0.028043311089277267,
-0.0718255415558815,
-0.051315851509571075,
0.010653011500835419,
0.02427557110786438,
-0.0035659989807754755,
0.07961928844451904,
-0.07713770121335983,
-0.007271230686455965,
-0.09142686426639557,
-0.06710109859704971,
0.022826774045825005,
-0.002279252279549837,
-0.008853771723806858,
-0.08460113406181335,
-0.2327534556388855,
-0.040439873933792114,
0.04538010433316231,
-0.07717941701412201,
-0.030769936740398407,
-0.010474956594407558,
-0.06794483959674835,
0.027702851220965385,
-0.022521233186125755,
0.17641222476959229,
-0.05238024517893791,
0.07069104164838791,
0.033892907202243805,
0.018929488956928253,
0.004258052911609411,
0.029746156185865402,
-0.06037338078022003,
0.03605855628848076,
-0.08236918598413467,
0.08279943466186523,
-0.08701350539922714,
0.04929889738559723,
-0.12447290122509003,
-0.09609237313270569,
-0.04295440763235092,
-0.014235761016607285,
0.09093017131090164,
0.14321313798427582,
-0.1790246218442917,
-0.01828484982252121,
0.13990329205989838,
-0.07007226347923279,
-0.11294376105070114,
0.10049278289079666,
-0.031268347054719925,
0.07617131620645523,
0.06039587780833244,
0.14351218938827515,
0.14235259592533112,
-0.10738246887922287,
-0.036442942917346954,
0.03158717229962349,
0.07667562365531921,
-0.02872280590236187,
0.09594276547431946,
-0.009683515876531601,
-0.027548756450414658,
-0.011868540197610855,
-0.09656856954097748,
-0.0029711138922721148,
-0.09047039598226547,
-0.08661479502916336,
-0.040755949914455414,
-0.08644392341375351,
0.040366314351558685,
0.023508885875344276,
0.05263855308294296,
-0.06346354633569717,
-0.11947063356637955,
0.07731741666793823,
0.1444457471370697,
-0.05437947064638138,
0.012375127524137497,
-0.08550877869129181,
0.03751359134912491,
-0.07070936262607574,
-0.023499149829149246,
-0.2007453590631485,
-0.06843187659978867,
0.044216565787792206,
-0.02504572831094265,
0.0009283955441787839,
-0.03334317356348038,
0.08500771969556808,
0.056809645146131516,
-0.05121941864490509,
-0.03878415375947952,
-0.10541819781064987,
-0.007251556497067213,
-0.10171668231487274,
-0.15798833966255188,
-0.09015527367591858,
-0.03861037269234657,
0.21523869037628174,
-0.26327934861183167,
0.0012493862304836512,
-0.005041321739554405,
0.1272474080324173,
0.018224656581878662,
-0.053995732218027115,
0.0004428578249644488,
0.029098622500896454,
-0.014411699958145618,
-0.10247421264648438,
0.025756960734725,
0.027187444269657135,
-0.11310712993144989,
-0.0357794314622879,
-0.09507209807634354,
0.07556284219026566,
0.08412160724401474,
0.0939275398850441,
-0.10456056147813797,
-0.07769305258989334,
-0.04875840246677399,
-0.03975353389978409,
-0.04510870948433876,
-0.04237140342593193,
0.16324473917484283,
0.02229093387722969,
0.12971432507038116,
-0.06958691775798798,
-0.073684923350811,
0.0000789595942478627,
0.012218241579830647,
-0.012363945133984089,
0.05123402550816536,
0.011698082089424133,
-0.12368269264698029,
0.05962176248431206,
0.09604653716087341,
-0.05173959210515022,
0.13563737273216248,
-0.0570351667702198,
-0.10806157439947128,
-0.03124435618519783,
0.01801581121981144,
0.011813808232545853,
0.1272173970937729,
-0.06926585733890533,
0.010012108832597733,
0.023461632430553436,
0.0224944818764925,
0.025851473212242126,
-0.18119104206562042,
-0.005994922947138548,
0.03988804295659065,
-0.03595322370529175,
-0.03873860463500023,
-0.046714410185813904,
-0.01426349300891161,
0.08077200502157211,
0.011341703124344349,
-0.022618450224399567,
0.026041768491268158,
-0.022034885361790657,
-0.08342105895280838,
0.18525923788547516,
-0.09048177301883698,
-0.14598339796066284,
-0.14499138295650482,
0.05578508600592613,
-0.013887171633541584,
-0.024076083675026894,
-0.02370608039200306,
-0.056929245591163635,
-0.05154009908437729,
-0.107301726937294,
-0.023932909592986107,
-0.05464617535471916,
0.00012303490075282753,
0.08562976866960526,
0.0018137013539671898,
0.08618173748254776,
-0.13342402875423431,
0.030330287292599678,
-0.003647733246907592,
-0.06834118068218231,
0.012845962308347225,
0.0261679720133543,
0.10052786767482758,
0.11357526481151581,
-0.011307613924145699,
0.031471312046051025,
-0.0307698342949152,
0.1832638829946518,
-0.08859167993068695,
-0.013803408481180668,
0.11689180135726929,
0.0157183725386858,
0.040329813957214355,
0.10187216103076935,
0.0383010059595108,
-0.0836477056145668,
0.034093089401721954,
0.028943298384547234,
-0.014787566848099232,
-0.24978385865688324,
-0.027571359649300575,
-0.03877360001206398,
-0.07980306446552277,
0.11520399153232574,
0.04658297449350357,
0.025146694853901863,
0.05039602890610695,
-0.004456785973161459,
0.049125321209430695,
-0.027456071227788925,
0.0780283659696579,
0.06798861920833588,
0.041187409311532974,
0.0831640437245369,
-0.04509805515408516,
-0.026736902073025703,
0.07386337220668793,
0.0198383592069149,
0.2558669447898865,
-0.031242795288562775,
0.1251520961523056,
0.008141981437802315,
0.1300758421421051,
-0.021566692739725113,
0.05930398032069206,
0.021973008289933205,
-0.01141446828842163,
0.011224446818232536,
-0.06594789773225784,
-0.03298722207546234,
0.03267153725028038,
-0.003527045249938965,
0.03462207689881325,
-0.08627009391784668,
0.012848951853811741,
-0.001442659180611372,
0.2678047716617584,
0.07117913663387299,
-0.3367890417575836,
-0.06871271133422852,
0.010345395654439926,
-0.011840330436825752,
-0.08000374585390091,
-0.018028151243925095,
0.1350121647119522,
-0.11967809498310089,
0.07616332918405533,
-0.06454432755708694,
0.08635866641998291,
-0.027651315554976463,
-0.029844801872968674,
0.06994068622589111,
0.10861068964004517,
-0.017521457746624947,
0.07286588847637177,
-0.19771498441696167,
0.23074455559253693,
0.02877776511013508,
0.13724999129772186,
-0.04757675528526306,
0.037631601095199585,
0.01868179813027382,
0.08282171934843063,
0.09444872289896011,
-0.0022224776912480593,
-0.04360134154558182,
-0.1833031326532364,
-0.09217920154333115,
0.029054177924990654,
0.10130821168422699,
-0.04033614322543144,
0.1041504368185997,
-0.049286700785160065,
0.0007521102670580149,
0.05379354581236839,
-0.05033307895064354,
-0.17641699314117432,
-0.12281571328639984,
0.033144090324640274,
0.020079970359802246,
0.031018298119306564,
-0.10122013837099075,
-0.11670190095901489,
0.006891786586493254,
0.17457467317581177,
0.030737044289708138,
-0.0261640977114439,
-0.14780856668949127,
0.07999003678560257,
0.1396416872739792,
-0.04251416400074959,
0.02604125440120697,
0.03531067818403244,
0.1555134356021881,
0.025555195286870003,
-0.05488051474094391,
0.05228725075721741,
-0.07660114765167236,
-0.12745767831802368,
-0.03590884059667587,
0.14542588591575623,
0.024013103917241096,
0.056352339684963226,
0.03639779984951019,
0.013069641776382923,
0.014371693134307861,
-0.08875103294849396,
0.050298869609832764,
0.03037038818001747,
0.09651545435190201,
0.06183410435914993,
-0.04570910334587097,
0.04371226578950882,
-0.030103938654065132,
-0.031169729307293892,
0.13794216513633728,
0.24861593544483185,
-0.07256493717432022,
-0.005156212020665407,
0.06179053336381912,
-0.08643132448196411,
-0.17352688312530518,
0.07640427350997925,
0.13369937241077423,
0.02362404391169548,
0.031923193484544754,
-0.1853746473789215,
0.052191343158483505,
0.09937622398138046,
-0.028931913897395134,
0.08079709857702255,
-0.30861490964889526,
-0.12820659577846527,
0.07850942760705948,
0.09020882844924927,
-0.04397011920809746,
-0.13432709872722626,
-0.06198263540863991,
-0.03452843055129051,
-0.12403017282485962,
0.13361184298992157,
-0.06899292767047882,
0.10241467505693436,
0.012677695602178574,
0.08006899058818817,
0.027858739718794823,
-0.03706744313240051,
0.1758410781621933,
0.034357909113168716,
0.09081613272428513,
-0.028718795627355576,
0.07286886125802994,
0.10435187816619873,
-0.0787908062338829,
0.062091339379549026,
-0.03009761869907379,
0.051424019038677216,
-0.15960213541984558,
-0.01797821931540966,
-0.04390939697623253,
0.0657304972410202,
-0.04492413252592087,
-0.052453819662332535,
-0.016623765230178833,
0.07457514107227325,
0.01291213184595108,
-0.04349130764603615,
0.06850970536470413,
0.0333622545003891,
0.11797220259904861,
0.0668814554810524,
0.10629455745220184,
-0.01380646601319313,
-0.06336812674999237,
-0.0016197741497308016,
-0.021726584061980247,
0.06580933183431625,
-0.12030097842216492,
0.0061956848949193954,
0.10934728384017944,
0.07388994097709656,
0.15499554574489594,
0.02018321119248867,
-0.05385725945234299,
-0.004334849305450916,
0.05592944100499153,
-0.06385332345962524,
-0.11889950186014175,
-0.02664361521601677,
0.019154783338308334,
-0.17474959790706635,
0.0030747384298592806,
0.09514658153057098,
-0.08274930715560913,
-0.0046265278942883015,
-0.011288516223430634,
0.012268264777958393,
-0.030860647559165955,
0.20229144394397736,
0.04203501343727112,
0.06875265389680862,
-0.07328785955905914,
0.09646286815404892,
0.061311136931180954,
-0.07004125416278839,
0.03902359679341316,
0.013956858776509762,
-0.06206627935171127,
-0.022293228656053543,
0.027487168088555336,
0.10740357637405396,
-0.02653377503156662,
-0.05288923159241676,
-0.0533769316971302,
-0.08151829242706299,
0.03458402305841446,
0.03125244379043579,
0.042204294353723526,
-0.0032695711124688387,
-0.022967038676142693,
0.012595048174262047,
-0.14067919552326202,
0.08509648591279984,
0.044759731739759445,
0.09483758360147476,
-0.16092148423194885,
0.12451675534248352,
0.005303107667714357,
0.019556816667318344,
-0.013338902033865452,
0.002705499529838562,
-0.05183115974068642,
-0.008193697780370712,
-0.09322325885295868,
0.007829663343727589,
-0.02678280510008335,
0.006975941359996796,
-0.014301533810794353,
-0.04667524993419647,
-0.06502622365951538,
0.03855627775192261,
-0.06486326456069946,
-0.07606398314237595,
0.02312994934618473,
0.039368510246276855,
-0.11323995143175125,
-0.02246136963367462,
0.0245493296533823,
-0.10725300759077072,
0.06270013749599457,
0.04163534939289093,
0.04213014245033264,
0.03820650652050972,
-0.07525552809238434,
-0.0036859489046037197,
0.0405416376888752,
0.024150138720870018,
0.0548473559319973,
-0.12782247364521027,
-0.008471515960991383,
-0.003737468272447586,
0.0397469587624073,
0.018425723537802696,
0.04047122597694397,
-0.1189512237906456,
-0.05207546055316925,
-0.07718446850776672,
-0.016047338023781776,
-0.05336654186248779,
0.0694381520152092,
0.10234856605529785,
0.05804497003555298,
0.17772482335567474,
-0.06718320399522781,
0.025816982612013817,
-0.22423022985458374,
-0.03453930467367172,
-0.017894424498081207,
-0.03026879020035267,
-0.0800640732049942,
-0.03676594793796539,
0.08496589213609695,
-0.05614272505044937,
0.10826978832483292,
-0.008871067315340042,
0.10562022775411606,
0.041193000972270966,
-0.045864153653383255,
0.0047456067986786366,
0.0015818910906091332,
0.22000117599964142,
0.08373300731182098,
0.0026570912450551987,
0.10025322437286377,
-0.0028978039044886827,
0.0628473237156868,
0.030972402542829514,
0.1508224606513977,
0.1339421272277832,
-0.0342637337744236,
0.07832009345293045,
0.059063076972961426,
-0.11560754477977753,
-0.14019572734832764,
0.06423163414001465,
-0.01192059088498354,
0.11140751093626022,
-0.04705812409520149,
0.1251063197851181,
0.10648275911808014,
-0.17526353895664215,
0.03277301788330078,
-0.06252312660217285,
-0.11408250778913498,
-0.0858580619096756,
-0.05587884038686752,
-0.06792133301496506,
-0.13010291755199432,
0.003655659966170788,
-0.1232248917222023,
0.02796248160302639,
0.09879274666309357,
0.008880894631147385,
0.004750377032905817,
0.14839184284210205,
-0.030134864151477814,
-0.002132342429831624,
0.09247223287820816,
-0.018566997721791267,
0.017587218433618546,
-0.07398013770580292,
-0.0829525887966156,
0.06751315295696259,
0.02909203991293907,
0.09131300449371338,
-0.052261266857385635,
-0.012198654934763908,
0.058775246143341064,
0.01755395531654358,
-0.09132993221282959,
0.017927156761288643,
0.020925411954522133,
0.05325890704989433,
0.025684570893645287,
0.06596975028514862,
0.003881593933328986,
-0.03327900171279907,
0.28578928112983704,
-0.06372565776109695,
-0.059308841824531555,
-0.14472873508930206,
0.1451243907213211,
0.002614762634038925,
-0.01775680109858513,
0.0688379555940628,
-0.13154934346675873,
0.007685867138206959,
0.10405534505844116,
0.10263852775096893,
-0.06398955732584,
-0.02083495445549488,
-0.011644892394542694,
-0.0204011257737875,
-0.04353909194469452,
0.14567546546459198,
0.08138569444417953,
-0.0017720822943374515,
-0.052004918456077576,
-0.007017963100224733,
-0.02660863846540451,
-0.053907476365566254,
-0.046004682779312134,
0.079160675406456,
0.017324669286608696,
0.017803380265831947,
-0.031046465039253235,
0.10602836310863495,
0.04558016359806061,
-0.19680848717689514,
-0.010583587922155857,
-0.13329792022705078,
-0.18864667415618896,
-0.03398754447698593,
0.08053804934024811,
-0.0100252078846097,
0.04948849603533745,
-0.02458246797323227,
0.0014843875542283058,
0.08644615113735199,
-0.011792301200330257,
-0.021764516830444336,
-0.13736321032047272,
0.10635929554700851,
-0.06245410069823265,
0.2343938797712326,
-0.0037044889759272337,
0.07274027168750763,
0.09379015862941742,
0.0013915476156398654,
-0.16661441326141357,
0.02023395523428917,
0.06793366372585297,
-0.031498685479164124,
0.06761329621076584,
0.1533893197774887,
-0.04380779340863228,
0.08409539610147476,
0.0314340703189373,
-0.11895193159580231,
-0.02059050276875496,
-0.06769350171089172,
-0.01808979921042919,
-0.09967164695262909,
-0.01928478479385376,
-0.05590595304965973,
0.16307246685028076,
0.18637798726558685,
-0.06855539977550507,
0.012415215373039246,
-0.08032651990652084,
0.019721828401088715,
0.04513309895992279,
0.10688989609479904,
-0.032785575836896896,
-0.20558848977088928,
0.0020244906190782785,
0.022230930626392365,
0.028452331200242043,
-0.2542434334754944,
-0.07738206535577774,
0.043260060250759125,
-0.06667350977659225,
-0.031412918120622635,
0.1139899492263794,
0.09829536080360413,
0.03918387368321419,
-0.02717667631804943,
-0.09165994822978973,
-0.05503670871257782,
0.1312483549118042,
-0.15871082246303558,
-0.04226754605770111
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# outputs
This model was trained from scratch on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 1.4939
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 1
- eval_batch_size: 1
- seed: 42
- gradient_accumulation_steps: 16
- total_train_batch_size: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 5
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| 2.0181 | 0.24 | 4 | 1.9684 |
| 2.0616 | 0.47 | 8 | 1.8863 |
| 1.8467 | 0.71 | 12 | 1.8116 |
| 1.707 | 0.94 | 16 | 1.7309 |
| 1.7886 | 1.18 | 20 | 1.6529 |
| 1.6539 | 1.41 | 24 | 1.5884 |
| 1.5149 | 1.65 | 28 | 1.5568 |
| 1.4526 | 1.88 | 32 | 1.5390 |
| 1.5335 | 2.12 | 36 | 1.5283 |
| 1.5668 | 2.35 | 40 | 1.5211 |
| 1.3914 | 2.59 | 44 | 1.5158 |
| 1.5769 | 2.82 | 48 | 1.5113 |
| 1.3794 | 3.06 | 52 | 1.5075 |
| 1.5274 | 3.29 | 56 | 1.5043 |
| 1.5247 | 3.53 | 60 | 1.5016 |
| 1.4291 | 3.76 | 64 | 1.4993 |
| 1.4233 | 4.0 | 68 | 1.4974 |
| 1.4353 | 4.24 | 72 | 1.4960 |
| 1.6016 | 4.47 | 76 | 1.4949 |
| 1.4416 | 4.71 | 80 | 1.4942 |
| 1.4654 | 4.94 | 84 | 1.4939 |
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.2.0+cu118
- Datasets 2.17.0
- Tokenizers 0.15.1 | {"library_name": "peft", "tags": ["trl", "sft", "generated_from_trainer"], "base_model": "LLama_weights/tmp", "model-index": [{"name": "outputs", "results": []}]} | null | Basha738/outputs | [
"peft",
"tensorboard",
"safetensors",
"trl",
"sft",
"generated_from_trainer",
"base_model:LLama_weights/tmp",
"region:us"
] | 2024-02-08T06:34:14+00:00 | [] | [] | TAGS
#peft #tensorboard #safetensors #trl #sft #generated_from_trainer #base_model-LLama_weights/tmp #region-us
| outputs
=======
This model was trained from scratch on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 1.4939
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 1
* eval\_batch\_size: 1
* seed: 42
* gradient\_accumulation\_steps: 16
* total\_train\_batch\_size: 16
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 5
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* PEFT 0.8.2
* Transformers 4.37.2
* Pytorch 2.2.0+cu118
* Datasets 2.17.0
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 1\n* seed: 42\n* gradient\\_accumulation\\_steps: 16\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.2\n* Pytorch 2.2.0+cu118\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
"TAGS\n#peft #tensorboard #safetensors #trl #sft #generated_from_trainer #base_model-LLama_weights/tmp #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 1\n* seed: 42\n* gradient\\_accumulation\\_steps: 16\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.2\n* Pytorch 2.2.0+cu118\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
44,
141,
4,
39
] | [
"passage: TAGS\n#peft #tensorboard #safetensors #trl #sft #generated_from_trainer #base_model-LLama_weights/tmp #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 1\n* seed: 42\n* gradient\\_accumulation\\_steps: 16\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.2\n* Pytorch 2.2.0+cu118\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
-0.11290506273508072,
0.06460196524858475,
-0.00298514892347157,
0.09232161194086075,
0.140549898147583,
0.006438891869038343,
0.10556130856275558,
0.13328345119953156,
-0.12263218313455582,
0.0866960659623146,
0.11085882037878036,
0.07810774445533752,
0.03825432434678078,
0.1735166311264038,
-0.049536313861608505,
-0.25728103518486023,
0.001912323641590774,
0.0014358187327161431,
-0.11169923096895218,
0.1187455877661705,
0.0839153304696083,
-0.13267254829406738,
0.06561760604381561,
-0.01634341850876808,
-0.172222301363945,
0.030881626531481743,
0.013556918129324913,
-0.03173806890845299,
0.09437551349401474,
0.02559667080640793,
0.12921424210071564,
0.02083035372197628,
0.10131790488958359,
-0.19414421916007996,
0.008624640293419361,
0.08978481590747833,
0.013116450048983097,
0.08517049998044968,
0.08675312995910645,
-0.017000991851091385,
0.10458333045244217,
-0.10473066568374634,
0.06938993185758591,
0.01592247188091278,
-0.15232276916503906,
-0.3011346161365509,
-0.11154443025588989,
0.037313543260097504,
0.10928118228912354,
0.062032271176576614,
-0.01867496594786644,
0.1554574817419052,
-0.07755028456449509,
0.08459580689668655,
0.2886311411857605,
-0.26239973306655884,
-0.08648404479026794,
0.023193761706352234,
0.004873073659837246,
0.08413415402173996,
-0.11160267889499664,
-0.028660479933023453,
0.06270202994346619,
0.034593988209962845,
0.12188728898763657,
0.0007762771565467119,
-0.037964075803756714,
0.028520463034510612,
-0.1567889153957367,
-0.01191004365682602,
0.0951988622546196,
0.039010047912597656,
-0.028578585013747215,
-0.03912753984332085,
-0.062342461198568344,
-0.19488736987113953,
-0.050384800881147385,
-0.01222022995352745,
0.04925014078617096,
-0.05543976649641991,
-0.07578622549772263,
0.02548922598361969,
-0.06906374543905258,
-0.08837428689002991,
-0.011548708193004131,
0.1678289920091629,
0.06131717562675476,
0.012025491334497929,
-0.024359596893191338,
0.10512246191501617,
-0.05520498752593994,
-0.15379711985588074,
0.011640029959380627,
0.021586917340755463,
-0.05701662227511406,
-0.05672897398471832,
-0.03479841351509094,
-0.07668569684028625,
-0.002157914685085416,
0.14400506019592285,
-0.12285269796848297,
0.07973133772611618,
0.001092437538318336,
0.022536007687449455,
-0.0988362655043602,
0.1376703828573227,
-0.07256323099136353,
0.017592011019587517,
-0.013759970664978027,
0.08779821544885635,
0.0179781224578619,
-0.0017554620280861855,
-0.06150006502866745,
0.04672172665596008,
0.0726100504398346,
0.03315840661525726,
-0.06675317138433456,
0.019148577004671097,
-0.04313082993030548,
-0.00430607283487916,
0.052269063889980316,
-0.11163202673196793,
0.05911750718951225,
0.005467447452247143,
-0.07835479080677032,
-0.014360765926539898,
0.008514639921486378,
0.003135654143989086,
0.0032681680750101805,
0.15359434485435486,
-0.09317648410797119,
0.04054117947816849,
-0.09803380072116852,
-0.1326022893190384,
0.021455729380249977,
-0.04578090459108353,
-0.0028586210682988167,
-0.06116260960698128,
-0.13502153754234314,
-0.03615128621459007,
0.04430929943919182,
-0.06914526224136353,
-0.012828337028622627,
-0.034739866852760315,
-0.08759164065122604,
0.022213369607925415,
-0.0021702630911022425,
0.12316832691431046,
-0.06415948271751404,
0.10807406902313232,
0.03250522166490555,
0.07171051949262619,
0.002030361909419298,
0.033478450030088425,
-0.08224611729383469,
0.04564979299902916,
-0.2731154263019562,
0.033373452723026276,
-0.08089694380760193,
0.07305952906608582,
-0.10311008244752884,
-0.10297518968582153,
-0.02960197441279888,
-0.019174637272953987,
0.11528407037258148,
0.12723693251609802,
-0.1980525255203247,
-0.06904072314500809,
0.20626051723957062,
-0.10706990957260132,
-0.10402812063694,
0.1252063363790512,
-0.04053881764411926,
-0.015292622148990631,
0.058517392724752426,
0.19402490556240082,
0.06424079090356827,
-0.11583998054265976,
0.009347924962639809,
-0.06556398421525955,
0.09657607972621918,
-0.002322761807590723,
0.0662676990032196,
-0.018463360145688057,
0.020426074042916298,
0.0005631027743220329,
-0.027356727048754692,
0.04872197285294533,
-0.10706673562526703,
-0.069828562438488,
-0.03794652223587036,
-0.08620468527078629,
0.04126492887735367,
0.07022084295749664,
0.050312433391809464,
-0.11991133540868759,
-0.09564412385225296,
0.08849850296974182,
0.08798563480377197,
-0.05545271560549736,
0.03239923343062401,
-0.04826729744672775,
0.0811786949634552,
-0.06872481107711792,
-0.025050001218914986,
-0.17894989252090454,
-0.0564165934920311,
0.017581766471266747,
-0.004967991728335619,
0.004494934342801571,
-0.049152396619319916,
0.09932416677474976,
0.08395224809646606,
-0.08913936465978622,
-0.02104528248310089,
-0.04867066070437431,
-0.00601879321038723,
-0.1237930878996849,
-0.24663694202899933,
-0.03895322233438492,
-0.032304342836141586,
0.0781235545873642,
-0.2391575574874878,
0.02860972285270691,
0.03694171458482742,
0.1384200006723404,
0.053043000400066376,
-0.053233057260513306,
-0.014302159659564495,
0.08339396119117737,
-0.00972797255963087,
-0.07914447039365768,
0.024282997474074364,
-0.017986979335546494,
-0.06882745027542114,
-0.031444381922483444,
-0.13089603185653687,
0.18104447424411774,
0.11097810417413712,
0.010961534455418587,
-0.12248881161212921,
-0.0256077591329813,
-0.06698460876941681,
-0.03171389177441597,
-0.07566588371992111,
0.03308965638279915,
0.11309681832790375,
0.011065859347581863,
0.12779508531093597,
-0.09845238924026489,
-0.06672456860542297,
0.03299589827656746,
-0.025560179725289345,
0.026454277336597443,
0.1373223513364792,
0.08253542333841324,
-0.08817921578884125,
0.1326266974210739,
0.13909506797790527,
-0.07019409537315369,
0.1311260163784027,
-0.07369660586118698,
-0.10227654129266739,
-0.037008851766586304,
0.030869901180267334,
0.03622043877840042,
0.15283304452896118,
-0.03599078953266144,
0.012613681145012379,
0.00031094643054530025,
0.026414567604660988,
0.008716987445950508,
-0.20797084271907806,
-0.04035630077123642,
0.03629794716835022,
-0.05398624762892723,
-0.02255184017121792,
-0.03195599094033241,
0.02107502892613411,
0.1244412288069725,
-0.01113219652324915,
-0.03409126028418541,
-0.011319328099489212,
0.008053319528698921,
-0.0793972983956337,
0.20458586513996124,
-0.09338462352752686,
-0.07289989292621613,
-0.08943769335746765,
-0.02278812602162361,
-0.0331311896443367,
-0.01731244847178459,
0.041155096143484116,
-0.12356521189212799,
-0.03384126350283623,
-0.09231887757778168,
0.000029580036425613798,
0.0048199486918747425,
0.03688293695449829,
-0.006401766557246447,
-0.0026615895330905914,
0.09486372768878937,
-0.09715630859136581,
0.012046672403812408,
-0.04192240908741951,
-0.020639413967728615,
0.0321805477142334,
0.0464758463203907,
0.11066808551549911,
0.1598191112279892,
-0.021947680041193962,
0.03251674398779869,
-0.03216204792261124,
0.20993417501449585,
-0.07874231785535812,
-0.015337791293859482,
0.08023311197757721,
-0.005344390869140625,
0.06963658332824707,
0.12173867225646973,
0.07515520602464676,
-0.10678015649318695,
0.014649930410087109,
0.059989720582962036,
-0.049424976110458374,
-0.22295206785202026,
-0.0271456316113472,
-0.04004696011543274,
-0.03547394275665283,
0.09920051693916321,
0.04101207107305527,
-0.00900406762957573,
0.03220873326063156,
0.0006191668217070401,
0.026696711778640747,
-0.02288012206554413,
0.0707014724612236,
0.021472519263625145,
0.05068127065896988,
0.11560093611478806,
-0.053057264536619186,
-0.047256793826818466,
0.04554130882024765,
-0.037236813455820084,
0.23548227548599243,
-0.022413579747080803,
0.08037377893924713,
0.04881690442562103,
0.1719128042459488,
-0.002662915736436844,
0.0759902372956276,
-0.001655665459111333,
-0.05757004767656326,
0.011383784003555775,
-0.07076577842235565,
0.01491887029260397,
0.03283778578042984,
-0.046270500868558884,
0.05383997783064842,
-0.11537278443574905,
-0.020786553621292114,
0.06241963803768158,
0.2546767294406891,
0.06593994051218033,
-0.3169846832752228,
-0.05995349958539009,
0.01555679552257061,
-0.015041936188936234,
-0.017608527094125748,
0.007317651994526386,
0.14302316308021545,
-0.046496905386447906,
0.08563093096017838,
-0.07161901146173477,
0.08994103223085403,
-0.00042687461245805025,
0.0027081237640231848,
0.0716666579246521,
0.1140548437833786,
-0.039640381932258606,
0.03939094394445419,
-0.2520282566547394,
0.2967221438884735,
0.03048183023929596,
0.07652382552623749,
-0.030877260491251945,
0.0016808334039524198,
0.025653934106230736,
0.04594777897000313,
0.061741288751363754,
-0.00930564384907484,
-0.11673818528652191,
-0.2089191973209381,
-0.08272343873977661,
0.01642279140651226,
0.12870162725448608,
-0.022346679121255875,
0.11802680045366287,
0.0006967115914449096,
-0.008508453145623207,
0.05805543065071106,
-0.0534370057284832,
-0.0904364213347435,
-0.05701373517513275,
-0.01357004139572382,
-0.011093638837337494,
-0.06497831642627716,
-0.08453711867332458,
-0.10571279376745224,
-0.091163270175457,
0.1368231326341629,
-0.0025968796107918024,
-0.03667101263999939,
-0.14575280249118805,
0.0917169377207756,
0.12397699058055878,
-0.07342083752155304,
0.02880208007991314,
0.03209679573774338,
0.0836835727095604,
0.024288494139909744,
-0.04800887778401375,
0.1306043416261673,
-0.05236181244254112,
-0.19409379363059998,
-0.05997471883893013,
0.125332772731781,
0.0537450909614563,
0.057026609778404236,
-0.03733891248703003,
0.043201595544815063,
0.03242202475667,
-0.10036558657884598,
0.04368045553565025,
-0.006258795037865639,
0.051810916513204575,
0.03900904208421707,
-0.029429860413074493,
0.05384121462702751,
-0.05547125265002251,
-0.024247949942946434,
0.09757070243358612,
0.30713388323783875,
-0.09246158599853516,
0.02302415482699871,
0.01803177036345005,
-0.05589134246110916,
-0.16844616830348969,
0.07250090688467026,
0.07403850555419922,
0.00007121316593838856,
0.06447302550077438,
-0.1532856822013855,
0.08237957209348679,
0.12159474939107895,
-0.016970446333289146,
0.11751298606395721,
-0.31421801447868347,
-0.13456803560256958,
0.07623577117919922,
0.1375732272863388,
0.02648993767797947,
-0.1819537878036499,
-0.04411022365093231,
-0.0007463939837180078,
-0.07338874042034149,
0.06715299189090729,
-0.13849402964115143,
0.08474499732255936,
-0.017428934574127197,
0.028570506721735,
0.022401057183742523,
-0.056830115616321564,
0.13883619010448456,
-0.01761285588145256,
0.13092924654483795,
-0.038837794214487076,
0.019839275628328323,
0.05104796215891838,
-0.06451943516731262,
0.024849671870470047,
-0.06070045381784439,
0.03077416680753231,
-0.09907319396734238,
-0.007275763433426619,
-0.07961814105510712,
0.005366044584661722,
-0.040979932993650436,
-0.03971587494015694,
-0.062268346548080444,
0.03453080728650093,
0.041356973350048065,
-0.027857540175318718,
0.15343202650547028,
0.0018958962755277753,
0.21708478033542633,
0.11941628903150558,
0.05195153132081032,
-0.048063308000564575,
-0.04892018809914589,
0.02981312945485115,
-0.018390342593193054,
0.047237057238817215,
-0.19624871015548706,
0.03665899112820625,
0.1431306153535843,
0.043028734624385834,
0.11560411006212234,
0.06482008099555969,
-0.07688590884208679,
0.002473554341122508,
0.07983693480491638,
-0.14403310418128967,
-0.1080956757068634,
0.0016116704791784286,
-0.005946123972535133,
-0.12614071369171143,
0.039009224623441696,
0.08655446767807007,
-0.07386734336614609,
-0.009516283869743347,
-0.000539998582098633,
0.02837659791111946,
-0.04396044835448265,
0.24431826174259186,
0.06578297168016434,
0.062490321695804596,
-0.10500859469175339,
0.08133156597614288,
0.056457679718732834,
-0.08774584531784058,
0.0222002025693655,
0.12073450535535812,
-0.05748656764626503,
-0.004160458222031593,
0.10900475829839706,
0.17115886509418488,
-0.0017888626316562295,
-0.03392334282398224,
-0.15728257596492767,
-0.12267188727855682,
0.08006155490875244,
0.20263493061065674,
0.06507062911987305,
0.004891386721283197,
0.005095465108752251,
0.01805216073989868,
-0.12804298102855682,
0.1017155721783638,
0.07821157574653625,
0.09072283655405045,
-0.14044679701328278,
0.1643872857093811,
-0.002138710580766201,
0.00493514072149992,
-0.017586685717105865,
0.053020283579826355,
-0.13607168197631836,
0.013094307854771614,
-0.10217668861150742,
-0.033364929258823395,
-0.025457145646214485,
-0.009126953780651093,
-0.010443801060318947,
-0.0647527202963829,
-0.0615047924220562,
0.016951296478509903,
-0.12408331036567688,
-0.026495425030589104,
0.00992667581886053,
0.018728915601968765,
-0.13705602288246155,
-0.03130827099084854,
0.024569176137447357,
-0.08043237775564194,
0.04971056804060936,
0.033098798245191574,
0.04186556860804558,
0.055111292749643326,
-0.11628390848636627,
0.01992052048444748,
0.059687189757823944,
-0.024622242897748947,
0.0664217621088028,
-0.119950070977211,
-0.017217881977558136,
-0.03532521054148674,
0.056819360703229904,
0.024450786411762238,
0.028891796246170998,
-0.1322224885225296,
0.0015012248186394572,
-0.03851011395454407,
-0.07696910202503204,
-0.029139427468180656,
0.03307182341814041,
0.08763228356838226,
0.0275738388299942,
0.1352778524160385,
-0.0931442454457283,
0.03747105970978737,
-0.2466007024049759,
-0.021563217043876648,
-0.011686463840305805,
-0.08903823047876358,
-0.0767611414194107,
0.0031213993206620216,
0.08775077015161514,
-0.05570020526647568,
0.10699157416820526,
0.00748205091804266,
0.04292507842183113,
0.05944119393825531,
-0.06679830700159073,
0.013653844594955444,
0.03917795419692993,
0.16145379841327667,
0.02010149322450161,
-0.0511305071413517,
0.046380799263715744,
0.03588295355439186,
0.08613205701112747,
0.09740976989269257,
0.21273133158683777,
0.13955342769622803,
0.0023316144943237305,
0.0713200494647026,
0.03093457780778408,
-0.09111643582582474,
-0.11717361956834793,
0.07740205526351929,
-0.040455471724271774,
0.08950019627809525,
-0.01733090728521347,
0.1673208475112915,
0.1316327005624771,
-0.18945956230163574,
0.0465984083712101,
-0.04181760549545288,
-0.09121319651603699,
-0.12727147340774536,
-0.024501830339431763,
-0.08620429039001465,
-0.1809587925672531,
-0.002645249245688319,
-0.12076738476753235,
0.05168147012591362,
0.10187946259975433,
0.023349743336439133,
0.04027792811393738,
0.18108662962913513,
0.06085219606757164,
0.04634450003504753,
0.062227170914411545,
0.02745894528925419,
-0.018055854365229607,
-0.02227446250617504,
-0.07090891897678375,
0.03673015907406807,
-0.056484125554561615,
0.028448564931750298,
-0.04238511621952057,
-0.05823836103081703,
0.057121217250823975,
0.01735701598227024,
-0.1099637970328331,
0.025815453380346298,
0.017048414796590805,
0.06289652734994888,
0.04848766326904297,
0.024262147024273872,
0.011186785995960236,
-0.0423126220703125,
0.22652488946914673,
-0.09195475280284882,
-0.02924456261098385,
-0.12183096259832382,
0.2912781536579132,
0.04398028925061226,
-0.0021368442103266716,
0.01834159903228283,
-0.10012167692184448,
0.01938856765627861,
0.14746201038360596,
0.14282764494419098,
-0.038876116275787354,
-0.0009734516497701406,
-0.02641235664486885,
-0.016465039923787117,
-0.03493029996752739,
0.09201151877641678,
0.10577081888914108,
0.027827756479382515,
-0.08431747555732727,
-0.03728506714105606,
-0.058489639312028885,
-0.03208160027861595,
-0.0410899855196476,
0.04798145219683647,
0.054042521864175797,
0.01450357399880886,
-0.050475362688302994,
0.09438899159431458,
-0.023097656667232513,
-0.09653474390506744,
0.07755745947360992,
-0.2148134857416153,
-0.17480862140655518,
-0.019960373640060425,
0.05237439647316933,
-0.0016050407430157065,
0.054975420236587524,
-0.024733319878578186,
-0.016592208296060562,
0.06650152802467346,
-0.015200785361230373,
-0.06340333074331284,
-0.15721234679222107,
0.08221627026796341,
-0.11796945333480835,
0.20171141624450684,
-0.029115376994013786,
0.023711219429969788,
0.11357399821281433,
0.01862833835184574,
-0.09425950795412064,
0.046146560460329056,
0.05699152871966362,
-0.10709790140390396,
-0.01670907437801361,
0.15271331369876862,
-0.0451015830039978,
0.08691317588090897,
0.04238499328494072,
-0.12528614699840546,
0.007545332424342632,
-0.06508248299360275,
-0.04833412170410156,
-0.046403270214796066,
-0.022628987208008766,
-0.05736711621284485,
0.12466314435005188,
0.22994230687618256,
-0.036486998200416565,
0.041774339973926544,
-0.05670811980962753,
0.02725181169807911,
0.056182507425546646,
0.07677139341831207,
-0.01921169087290764,
-0.2582750916481018,
0.03551293909549713,
0.08215303719043732,
-0.01348846685141325,
-0.2492384910583496,
-0.0786878690123558,
0.02649838477373123,
-0.05961368978023529,
-0.08227917551994324,
0.11293765902519226,
0.08500345796346664,
0.06861256808042526,
-0.05584400147199631,
-0.14403970539569855,
-0.05150655657052994,
0.1672896444797516,
-0.1322261095046997,
-0.05897364020347595
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# gpt2-cybersac
This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 7.3745
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| 7.8257 | 1.0 | 1004 | 7.7518 |
| 7.4738 | 2.0 | 2008 | 7.4700 |
| 7.3193 | 3.0 | 3012 | 7.3745 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1
- Datasets 2.16.1
- Tokenizers 0.15.0
| {"license": "mit", "tags": ["generated_from_trainer"], "base_model": "gpt2", "model-index": [{"name": "gpt2-cybersac", "results": []}]} | text-generation | thomaslwang/gpt2-cybersac | [
"transformers",
"tensorboard",
"safetensors",
"gpt2",
"text-generation",
"generated_from_trainer",
"base_model:gpt2",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T06:34:29+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #gpt2 #text-generation #generated_from_trainer #base_model-gpt2 #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| gpt2-cybersac
=============
This model is a fine-tuned version of gpt2 on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 7.3745
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 64
* eval\_batch\_size: 64
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3.0
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.1
* Datasets 2.16.1
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1\n* Datasets 2.16.1\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #gpt2 #text-generation #generated_from_trainer #base_model-gpt2 #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1\n* Datasets 2.16.1\n* Tokenizers 0.15.0"
] | [
72,
98,
4,
30
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #gpt2 #text-generation #generated_from_trainer #base_model-gpt2 #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1\n* Datasets 2.16.1\n* Tokenizers 0.15.0"
] | [
-0.08091400563716888,
0.06340446323156357,
-0.002599051920697093,
0.10275588929653168,
0.1407833993434906,
0.016772793605923653,
0.16601140797138214,
0.12676945328712463,
-0.07275889068841934,
0.04694214463233948,
0.14434906840324402,
0.13430985808372498,
0.025535672903060913,
0.1378825157880783,
-0.06587649136781693,
-0.2366633415222168,
0.004112435970455408,
0.03468151018023491,
-0.056015338748693466,
0.12344570457935333,
0.0942722037434578,
-0.12256831675767899,
0.10176480561494827,
-0.010212235152721405,
-0.18491186201572418,
0.009247458539903164,
0.02644522674381733,
-0.056985631585121155,
0.13922952115535736,
0.03853742033243179,
0.10887493193149567,
0.028151921927928925,
0.07709625363349915,
-0.18276555836200714,
0.012299166060984135,
0.060244958847761154,
-0.003731349715963006,
0.08404011279344559,
0.050451863557100296,
0.0036992488894611597,
0.11758435517549515,
-0.0853501558303833,
0.0512097030878067,
0.013232932426035404,
-0.1363278329372406,
-0.21413549780845642,
-0.07524330168962479,
0.02249818854033947,
0.09785878658294678,
0.08157110959291458,
-0.019075196236371994,
0.13056278228759766,
-0.026868702843785286,
0.10178877413272858,
0.22910134494304657,
-0.32824718952178955,
-0.06454261392354965,
0.06684266030788422,
0.05276595801115036,
0.09316476434469223,
-0.09922446310520172,
0.010429139249026775,
0.0575791597366333,
0.030450697988271713,
0.1371288001537323,
-0.03329649567604065,
-0.012311812490224838,
0.009392005391418934,
-0.14323405921459198,
-0.03084300272166729,
0.15785813331604004,
0.03542467579245567,
-0.043460946530103683,
-0.06259655952453613,
-0.07177482545375824,
-0.13164885342121124,
-0.039068907499313354,
-0.02649591490626335,
0.044111430644989014,
-0.014986119233071804,
-0.09214209765195847,
-0.05326449126005173,
-0.11323512345552444,
-0.08330334722995758,
-0.05360211059451103,
0.14896191656589508,
0.03334397077560425,
0.009193358942866325,
-0.028027717024087906,
0.11258507519960403,
-0.053091730922460556,
-0.13408087193965912,
0.0007999468361958861,
0.02582278847694397,
0.018452011048793793,
-0.05046502500772476,
-0.05382312089204788,
-0.10542117804288864,
0.020940488204360008,
0.14042755961418152,
-0.07371629029512405,
0.058657415211200714,
0.005315168295055628,
0.04666788503527641,
-0.10076344758272171,
0.16990593075752258,
-0.033228032290935516,
-0.0235745832324028,
0.020432621240615845,
0.05986835062503815,
0.053369853645563126,
-0.018941812217235565,
-0.13711313903331757,
0.021615315228700638,
0.1153203547000885,
0.013912927359342575,
-0.06544116139411926,
0.08496321737766266,
-0.04184706136584282,
-0.006920883897691965,
0.01759539544582367,
-0.08405010402202606,
0.028722601011395454,
-0.010613180696964264,
-0.05612623319029808,
-0.06822134554386139,
0.02299819327890873,
0.018616197630763054,
-0.003066769102588296,
0.10458296537399292,
-0.08337478339672089,
0.01575058326125145,
-0.08297555148601532,
-0.12182029336690903,
0.01124613732099533,
-0.06990481168031693,
0.018237151205539703,
-0.10572066158056259,
-0.2008323073387146,
-0.004163122270256281,
0.05384490266442299,
-0.041433680802583694,
-0.023077601566910744,
-0.059138741344213486,
-0.07653933018445969,
0.01947648636996746,
-0.021606633439660072,
0.09909743070602417,
-0.061391353607177734,
0.10179141908884048,
0.050854988396167755,
0.07272703945636749,
-0.05859336256980896,
0.03085911087691784,
-0.09668135643005371,
0.03270908072590828,
-0.18189752101898193,
0.040474582463502884,
-0.03947700187563896,
0.0533490926027298,
-0.08274337649345398,
-0.0813772976398468,
-0.01723048835992813,
0.005659385118633509,
0.07488030940294266,
0.10148100554943085,
-0.15154419839382172,
-0.07387599349021912,
0.19162775576114655,
-0.0893179401755333,
-0.14494143426418304,
0.1399146318435669,
-0.05931026488542557,
0.055473580956459045,
0.07694455236196518,
0.2062188982963562,
0.0380173958837986,
-0.10113395750522614,
0.010195650160312653,
-0.0013601253740489483,
0.035051729530096054,
-0.03324255347251892,
0.07558199763298035,
0.0032906930427998304,
0.01441373024135828,
0.012333129532635212,
-0.042129866778850555,
0.04623395577073097,
-0.0842374786734581,
-0.07744187861680984,
-0.04253050312399864,
-0.09136302769184113,
0.0477861650288105,
0.04380473867058754,
0.07877799868583679,
-0.12509357929229736,
-0.09726506471633911,
0.052594687789678574,
0.05613085255026817,
-0.08242252469062805,
0.026217050850391388,
-0.07493878155946732,
0.08968625217676163,
-0.07906467467546463,
-0.008334527723491192,
-0.14054322242736816,
-0.05415116995573044,
0.013955122791230679,
0.019881315529346466,
0.024602241814136505,
0.009003161452710629,
0.08743946254253387,
0.08617763966321945,
-0.07157931476831436,
-0.017745422199368477,
0.0014431450981646776,
0.00020000735821668059,
-0.13264372944831848,
-0.18231020867824554,
-0.0075536081567406654,
-0.033360276371240616,
0.125146746635437,
-0.2314036786556244,
0.054108526557683945,
0.01819462887942791,
0.08113635331392288,
0.0309243593364954,
-0.022921502590179443,
-0.04035000130534172,
0.04745428264141083,
-0.051067106425762177,
-0.07023173570632935,
0.06448502093553543,
0.012200121767818928,
-0.1057373657822609,
-0.02716279961168766,
-0.17838342487812042,
0.19296656548976898,
0.1357201635837555,
-0.09378200024366379,
-0.08446534723043442,
-0.005480239633470774,
-0.03969299793243408,
-0.024857550859451294,
-0.04501011595129967,
-0.011979364790022373,
0.15388692915439606,
-0.013647104613482952,
0.15950262546539307,
-0.0904747024178505,
-0.04919872805476189,
0.029131488874554634,
-0.05591711029410362,
0.01141535583883524,
0.10523060709238052,
0.07319239526987076,
-0.11981856822967529,
0.14767573773860931,
0.16411401331424713,
-0.07206699997186661,
0.15976083278656006,
-0.02667459286749363,
-0.05388808995485306,
-0.03303477540612221,
0.022442348301410675,
0.014058270491659641,
0.098655104637146,
-0.13642649352550507,
-0.0069587924517691135,
0.008422448299825191,
0.019170934334397316,
0.020106296986341476,
-0.22532354295253754,
-0.033153872936964035,
0.04184400290250778,
-0.06297475844621658,
0.017039086669683456,
-0.00882650911808014,
-0.016090432181954384,
0.10229313373565674,
0.0057137105613946915,
-0.06894993036985397,
0.04935843124985695,
0.000029096596335875802,
-0.08455052226781845,
0.21122214198112488,
-0.06539187580347061,
-0.1615854948759079,
-0.13160176575183868,
-0.078531414270401,
-0.05407126992940903,
0.04051348939538002,
0.07054003328084946,
-0.07692641019821167,
-0.03815172612667084,
-0.11086869239807129,
0.019602108746767044,
0.0009491475648246706,
0.026621660217642784,
0.014506226405501366,
-0.013982496224343777,
0.06327400356531143,
-0.10354645550251007,
-0.019236411899328232,
-0.04526001214981079,
-0.06945546716451645,
0.045560430735349655,
0.02322208322584629,
0.10577081143856049,
0.13992182910442352,
-0.03042631968855858,
0.012946143746376038,
-0.0406620018184185,
0.23973706364631653,
-0.07750748097896576,
-0.012402473017573357,
0.13119910657405853,
-0.012906914576888084,
0.05553171783685684,
0.11734520643949509,
0.05700760334730148,
-0.10814371705055237,
0.019705453887581825,
0.026955805718898773,
-0.05287583917379379,
-0.20401838421821594,
-0.021802347153425217,
-0.0400354377925396,
0.013214768841862679,
0.09152472019195557,
0.03968290612101555,
0.03824673593044281,
0.07370925694704056,
0.019041478633880615,
0.08709486573934555,
0.0031079521868377924,
0.08584040403366089,
0.11286410689353943,
0.04140525683760643,
0.13666769862174988,
-0.04625602066516876,
-0.07132212072610855,
0.0422108992934227,
0.0017814768943935633,
0.20574480295181274,
0.021570948883891106,
0.13850514590740204,
0.04385301470756531,
0.1380602866411209,
0.0033165626227855682,
0.06289185583591461,
-0.01929328218102455,
-0.04700881615281105,
-0.01939200982451439,
-0.0483739860355854,
-0.02064760960638523,
0.04099232330918312,
-0.09313490986824036,
0.029928671196103096,
-0.09599879384040833,
0.004601954482495785,
0.05554357171058655,
0.20870080590248108,
0.05194360762834549,
-0.34087076783180237,
-0.08821044117212296,
0.03492075577378273,
-0.022452672943472862,
-0.029915453866124153,
0.025021953508257866,
0.13701944053173065,
-0.0615445114672184,
0.053524378687143326,
-0.09049063920974731,
0.07939238846302032,
-0.04465075209736824,
0.053210049867630005,
0.04892425984144211,
0.08353881537914276,
-0.020251382142305374,
0.07356101274490356,
-0.29053568840026855,
0.2697785794734955,
0.01882205717265606,
0.07932702451944351,
-0.06089385971426964,
0.0051641399040818214,
0.022253990173339844,
0.07268044352531433,
0.08641055971384048,
-0.02010398916900158,
-0.08113081753253937,
-0.19121773540973663,
-0.04930826276540756,
0.02943001687526703,
0.1161140576004982,
-0.03796834498643875,
0.10967332124710083,
-0.03304428234696388,
0.007005365565419197,
0.08067784458398819,
-0.016722310334444046,
-0.06621746718883514,
-0.09529609978199005,
-0.0006363068823702633,
0.03875508904457092,
-0.024586649611592293,
-0.07442598789930344,
-0.08988720923662186,
-0.13018611073493958,
0.17832277715206146,
-0.04489891603589058,
-0.04551846906542778,
-0.10403378307819366,
0.07371759414672852,
0.04395780712366104,
-0.08166977763175964,
0.033509183675050735,
0.011112118139863014,
0.08632324635982513,
0.022768590599298477,
-0.06453710794448853,
0.13124586641788483,
-0.0641680434346199,
-0.1737785041332245,
-0.060078080743551254,
0.11717826128005981,
0.012722667306661606,
0.042676907032728195,
0.00451699597761035,
0.0192408449947834,
-0.029001981019973755,
-0.08928052335977554,
0.048121530562639236,
-0.025708258152008057,
0.05259412154555321,
0.00506878225132823,
-0.0341353639960289,
0.020047705620527267,
-0.05147966369986534,
-0.03728051483631134,
0.16770552098751068,
0.28956282138824463,
-0.07868754863739014,
0.017598986625671387,
0.047279778867959976,
-0.07299679517745972,
-0.2021401822566986,
0.03202733397483826,
0.018377635627985,
-0.0005875263595953584,
0.048161204904317856,
-0.1526859551668167,
0.07929222285747528,
0.10012419521808624,
-0.028206318616867065,
0.1237388327717781,
-0.30914345383644104,
-0.1437864899635315,
0.10300253331661224,
0.15812243521213531,
0.15165606141090393,
-0.17210747301578522,
-0.04588634520769119,
-0.034526221454143524,
-0.0941685140132904,
0.10903310030698776,
-0.13019445538520813,
0.12891581654548645,
-0.013650509528815746,
0.07922127097845078,
0.00916418805718422,
-0.060751307755708694,
0.12514935433864594,
-0.019965332001447678,
0.11051701009273529,
-0.0765012800693512,
0.002450502011924982,
0.0725461095571518,
-0.05369163304567337,
0.03098166547715664,
-0.12445300072431564,
0.029125938192009926,
-0.053875599056482315,
-0.04007153958082199,
-0.04944121837615967,
0.038741376250982285,
-0.026456978172063828,
-0.07216384261846542,
-0.04888858646154404,
0.012330482713878155,
0.029144037514925003,
-0.01248947810381651,
0.16250672936439514,
0.00484208669513464,
0.15964072942733765,
0.14238843321800232,
0.08150551468133926,
-0.07568425685167313,
-0.01647135801613331,
-0.001437776256352663,
-0.0292658768594265,
0.05247194692492485,
-0.15548837184906006,
0.03240489587187767,
0.11823463439941406,
0.0073617021553218365,
0.1480601280927658,
0.08349521458148956,
-0.04276079311966896,
0.03428313881158829,
0.06681445986032486,
-0.17988061904907227,
-0.13820835947990417,
-0.016493823379278183,
-0.04859914630651474,
-0.09489525109529495,
0.08001389354467392,
0.13628728687763214,
-0.06776772439479828,
0.008753927424550056,
-0.01318502239882946,
0.017499113455414772,
-0.03842722997069359,
0.18904678523540497,
0.04654274135828018,
0.04229209944605827,
-0.0830148309469223,
0.07019301503896713,
0.03591267392039299,
-0.05523662269115448,
0.03172716498374939,
0.0589885376393795,
-0.07993384450674057,
-0.04189165681600571,
0.03496444970369339,
0.1994428038597107,
-0.05238231271505356,
-0.045115772634744644,
-0.14920170605182648,
-0.11410603672266006,
0.05178448185324669,
0.15958350896835327,
0.08655787259340286,
0.013035820797085762,
-0.03720620647072792,
0.02389785461127758,
-0.12145042419433594,
0.11138385534286499,
0.03771813586354256,
0.08800232410430908,
-0.14564000070095062,
0.13915444910526276,
-0.005954672582447529,
0.01316527184098959,
-0.0356389544904232,
0.038822516798973083,
-0.11744231730699539,
-0.003239176468923688,
-0.12678518891334534,
-0.01593630574643612,
-0.025244932621717453,
-0.007368056569248438,
-0.002958904718980193,
-0.05254899710416794,
-0.06474758684635162,
0.013285523280501366,
-0.10261011868715286,
-0.018453974276781082,
0.027162794023752213,
0.03741775080561638,
-0.1227014884352684,
-0.0414482057094574,
0.013640823774039745,
-0.06275799870491028,
0.0789996087551117,
0.0263031255453825,
0.014987649396061897,
0.05298786610364914,
-0.18628272414207458,
0.04001529514789581,
0.06229099631309509,
0.0020612962543964386,
0.034877173602581024,
-0.07339136302471161,
-0.013342960737645626,
-0.0006379900150932372,
0.057078246027231216,
0.031406428664922714,
0.07425731420516968,
-0.12359732389450073,
0.0195577684789896,
-0.03640333563089371,
-0.06534764915704727,
-0.05586053803563118,
0.04011442884802818,
0.059655096381902695,
0.01312791183590889,
0.18890707194805145,
-0.10528378933668137,
0.01663617417216301,
-0.20934580266475677,
0.00918753631412983,
0.016631945967674255,
-0.12923939526081085,
-0.1054343655705452,
-0.05789656937122345,
0.05982046574354172,
-0.05544830113649368,
0.15751561522483826,
0.02828788571059704,
0.01973036117851734,
0.034311916679143906,
-0.024976462125778198,
0.06179633364081383,
0.010010124184191227,
0.2291855812072754,
0.02962094359099865,
-0.05113435536623001,
0.01772666908800602,
0.048664264380931854,
0.1223754808306694,
0.06866933405399323,
0.18391013145446777,
0.13662202656269073,
-0.05766243487596512,
0.10599707067012787,
0.036199577152729034,
-0.043137382715940475,
-0.1599058359861374,
0.05081792548298836,
-0.04121796414256096,
0.10034683346748352,
-0.023869875818490982,
0.1977236419916153,
0.11962677538394928,
-0.15612557530403137,
0.009794489480555058,
-0.04823572561144829,
-0.08233193308115005,
-0.12183436006307602,
-0.08768434077501297,
-0.10271549969911575,
-0.1546679139137268,
0.008820489048957825,
-0.11606734246015549,
0.03290429711341858,
0.10044662654399872,
0.014977130107581615,
-0.02148994244635105,
0.1801089644432068,
0.014702030457556248,
0.023824742063879967,
0.050746284425258636,
-0.004243592265993357,
-0.027112657204270363,
-0.06922702491283417,
-0.08703900128602982,
0.002763173310086131,
-0.01838071644306183,
0.03417285904288292,
-0.040651679039001465,
-0.021839439868927002,
0.0322086475789547,
-0.018251417204737663,
-0.10255412757396698,
-0.0009264588006772101,
0.047571126371622086,
0.056253451853990555,
0.029862938448786736,
0.004860419314354658,
-0.004315641243010759,
-0.005365476012229919,
0.22685998678207397,
-0.07748401165008545,
-0.057899702340364456,
-0.0761021301150322,
0.2287118285894394,
0.023315725848078728,
0.015978358685970306,
0.004355898592621088,
-0.08761124312877655,
0.03146810829639435,
0.235707089304924,
0.19536040723323822,
-0.07830829918384552,
-0.0011383204255253077,
-0.017347605898976326,
-0.006199760362505913,
-0.013315094634890556,
0.10337003320455551,
0.09928677976131439,
0.04357524961233139,
-0.07909289747476578,
-0.03997420519590378,
-0.02459099516272545,
0.0012553459964692593,
-0.05060093104839325,
0.060699135065078735,
0.03885997459292412,
0.02077113464474678,
-0.04476838558912277,
0.061927929520606995,
-0.04071327671408653,
-0.10473911464214325,
0.028345279395580292,
-0.1943790316581726,
-0.13514794409275055,
0.006253289990127087,
0.11171767115592957,
-0.02128138393163681,
0.06614045798778534,
-0.03313097357749939,
0.005728274118155241,
0.035708777606487274,
-0.02004215307533741,
-0.0894695296883583,
-0.048917192965745926,
0.06269532442092896,
-0.11831299960613251,
0.22782956063747406,
-0.04042400047183037,
0.04908987507224083,
0.13110484182834625,
0.025307446718215942,
-0.07418505847454071,
0.0927954912185669,
0.03690613806247711,
-0.06332389265298843,
0.037183746695518494,
0.1030956581234932,
-0.036253221333026886,
0.089777372777462,
0.06007884442806244,
-0.1320602297782898,
0.008148632943630219,
-0.04132947325706482,
-0.0721394270658493,
-0.05276380106806755,
-0.052168406546115875,
-0.06591317057609558,
0.13422952592372894,
0.18368706107139587,
-0.030025670304894447,
0.014041696675121784,
-0.05472555384039879,
0.023296363651752472,
0.0726207047700882,
0.04649629816412926,
-0.030906159430742264,
-0.25225114822387695,
0.00943638663738966,
0.10500524938106537,
-0.01673498935997486,
-0.3103528618812561,
-0.08339107781648636,
-0.01542731188237667,
-0.03225192427635193,
-0.0993657335639,
0.08495522290468216,
0.12959185242652893,
0.05044623836874962,
-0.06040604040026665,
-0.09057044237852097,
-0.07316917926073074,
0.1626184582710266,
-0.1320618987083435,
-0.09531278163194656
] |
null | null | transformers | # SeverusWestLake-7B-DPO - AWQ
- Model creator: [s3nh](https://huggingface.co/s3nh)
- Original model: [SeverusWestLake-7B-DPO](https://huggingface.co/s3nh/SeverusWestLake-7B-DPO)

| {"license": "mit", "library_name": "transformers", "tags": ["mergekit", "merge"], "model_name": "SeverusWestLake-7B-DPO", "base_model": ["FelixChao/Sectumsempra-7B-DPO", "cognitivecomputations/WestLake-7B-v2-laser"], "model_creator": "s3nh", "model_type": "mistral", "pipeline_tag": "text-generation", "prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant\n", "quantized_by": "Suparious"} | text-generation | solidrust/SeverusWestLake-7B-DPO-AWQ | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"mergekit",
"merge",
"base_model:FelixChao/Sectumsempra-7B-DPO",
"base_model:cognitivecomputations/WestLake-7B-v2-laser",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"4-bit",
"region:us"
] | 2024-02-08T06:34:29+00:00 | [] | [] | TAGS
#transformers #safetensors #mistral #text-generation #mergekit #merge #base_model-FelixChao/Sectumsempra-7B-DPO #base_model-cognitivecomputations/WestLake-7B-v2-laser #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us
| # SeverusWestLake-7B-DPO - AWQ
- Model creator: s3nh
- Original model: SeverusWestLake-7B-DPO
!image
| [
"# SeverusWestLake-7B-DPO - AWQ\n\n- Model creator: s3nh\n- Original model: SeverusWestLake-7B-DPO\n\n!image"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #mergekit #merge #base_model-FelixChao/Sectumsempra-7B-DPO #base_model-cognitivecomputations/WestLake-7B-v2-laser #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n",
"# SeverusWestLake-7B-DPO - AWQ\n\n- Model creator: s3nh\n- Original model: SeverusWestLake-7B-DPO\n\n!image"
] | [
102,
37
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #mergekit #merge #base_model-FelixChao/Sectumsempra-7B-DPO #base_model-cognitivecomputations/WestLake-7B-v2-laser #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n# SeverusWestLake-7B-DPO - AWQ\n\n- Model creator: s3nh\n- Original model: SeverusWestLake-7B-DPO\n\n!image"
] | [
-0.10562489181756973,
0.03642871975898743,
-0.004760781768709421,
0.013056431896984577,
0.06663231551647186,
0.0017608021153137088,
0.19524694979190826,
0.04996975511312485,
0.05040602758526802,
0.008315237239003181,
0.09054649621248245,
0.12798632681369781,
0.04304938390851021,
0.1346019208431244,
-0.03943272680044174,
-0.1934709995985031,
0.09883400797843933,
0.03528236225247383,
-0.04042140766978264,
0.04037925973534584,
0.1017281785607338,
-0.08417224138975143,
0.09895431995391846,
-0.04426364228129387,
-0.06062827259302139,
0.028398480266332626,
-0.03144006058573723,
-0.05671035125851631,
0.08274386823177338,
0.06676328927278519,
0.023203428834676743,
0.12116314470767975,
0.021126670762896538,
-0.12142153084278107,
0.03232099860906601,
-0.003928014077246189,
-0.037319671362638474,
0.047382477670907974,
0.04667254909873009,
0.003260148223489523,
-0.004795733839273453,
-0.06785459816455841,
0.002215853426605463,
0.03567802160978317,
-0.07379835844039917,
0.0013836189173161983,
-0.06061201170086861,
0.1505347341299057,
0.10392116010189056,
0.03227146342396736,
0.009132759645581245,
0.0505770705640316,
0.015935171395540237,
0.098921999335289,
0.1659690886735916,
-0.22524333000183105,
-0.0316452719271183,
0.09738819301128387,
0.050233252346515656,
0.021870464086532593,
0.005504821427166462,
0.03384092077612877,
0.04763321205973625,
-0.03224701061844826,
-0.025064943358302116,
-0.0710906833410263,
0.13747777044773102,
-0.021682048216462135,
-0.11116176843643188,
0.02099008485674858,
0.2583809792995453,
0.041849534958601,
-0.040881894528865814,
-0.03865576535463333,
-0.1066713035106659,
0.005342263262718916,
-0.04747698828577995,
-0.04659854993224144,
0.026118531823158264,
-0.0019571399316191673,
0.045865852385759354,
-0.06252412497997284,
-0.06615468114614487,
-0.06347725540399551,
-0.09121078997850418,
0.18200267851352692,
0.0010266720782965422,
0.024783864617347717,
-0.0630471259355545,
0.08639399707317352,
-0.11278451234102249,
-0.1125965267419815,
-0.019937383010983467,
-0.08001776784658432,
0.008090447634458542,
0.035068877041339874,
-0.040703389793634415,
-0.10459761321544647,
0.0830271914601326,
0.21511642634868622,
0.0637802705168724,
0.020054105669260025,
0.05078829452395439,
0.06008925661444664,
-0.015263015404343605,
0.06769147515296936,
-0.09730686992406845,
-0.0904439315199852,
0.05441915616393089,
0.021260831505060196,
0.0897178053855896,
-0.04624592885375023,
-0.10316678136587143,
-0.04876410588622093,
0.055340275168418884,
-0.010986237786710262,
0.044690027832984924,
0.09711207449436188,
-0.016727037727832794,
-0.0391850546002388,
0.14742636680603027,
-0.06509879231452942,
-0.04740963876247406,
0.011258367449045181,
0.023236064240336418,
0.10632187873125076,
0.08309818804264069,
0.03453611955046654,
-0.01148039661347866,
0.026731593534350395,
-0.05557338893413544,
-0.052522871643304825,
-0.06102938577532768,
-0.04132896289229393,
0.03745565190911293,
-0.0716145932674408,
0.027068130671977997,
-0.1697976291179657,
-0.27913960814476013,
-0.003821803955361247,
-0.002833613660186529,
-0.016821594908833504,
-0.03475407138466835,
-0.050197429955005646,
-0.015088098123669624,
-0.006908611860126257,
-0.039335813373327255,
-0.03816302493214607,
-0.00925533752888441,
0.035917337983846664,
0.058498937636613846,
0.04475507512688637,
-0.24667969346046448,
0.016209473833441734,
-0.07809681445360184,
0.07224780321121216,
-0.13565681874752045,
0.03810863569378853,
-0.043319910764694214,
0.15365058183670044,
-0.007838616147637367,
0.0035837003961205482,
-0.06642342358827591,
0.04165942966938019,
0.04321253299713135,
0.1965263932943344,
-0.24913763999938965,
-0.05004493519663811,
0.08409859240055084,
-0.16083914041519165,
-0.19328196346759796,
0.08310375362634659,
0.0142278503626585,
0.04495907947421074,
0.10602093487977982,
0.2230421006679535,
0.08911920338869095,
-0.03538556024432182,
-0.012574932537972927,
0.008689066395163536,
-0.021608103066682816,
-0.09149426966905594,
0.07351802289485931,
0.029284760355949402,
-0.09456486999988556,
0.08078570663928986,
-0.06865967065095901,
0.00904786679893732,
-0.036744870245456696,
-0.048319704830646515,
-0.06568220257759094,
-0.07102001458406448,
0.0325913168489933,
-0.028034083545207977,
0.034268561750650406,
-0.08587602525949478,
-0.0003464943147264421,
0.12232412397861481,
0.07445377856492996,
-0.05591630935668945,
-0.009895606897771358,
-0.07902264595031738,
0.09299182146787643,
-0.0233126450330019,
0.06839622557163239,
-0.08900705724954605,
-0.05240966007113457,
-0.029771558940410614,
0.04788093641400337,
0.025667700916528702,
0.028376026079058647,
0.09131340682506561,
0.02680952288210392,
-0.007574755698442459,
-0.0736316442489624,
0.03960419446229935,
0.053920190781354904,
-0.03420836105942726,
-0.10869893431663513,
-0.008933345787227154,
-0.0445355586707592,
0.10814502835273743,
-0.06147898733615875,
0.08765339851379395,
0.026635337620973587,
0.1297958642244339,
0.0046895998530089855,
0.03190970420837402,
-0.025100288912653923,
0.0265598576515913,
-0.00460833078250289,
0.008741522207856178,
0.0788016989827156,
0.022342434152960777,
-0.13491329550743103,
0.09103094786405563,
-0.13346323370933533,
0.1516171246767044,
0.14564698934555054,
-0.08648157864809036,
0.027653753757476807,
-0.057000428438186646,
-0.017468834295868874,
-0.019108546897768974,
0.07890258729457855,
-0.13877743482589722,
-0.027602186426520348,
-0.006834762170910835,
0.08024092018604279,
-0.05955297127366066,
-0.005017935764044523,
0.003388330340385437,
-0.04247249290347099,
-0.05324005335569382,
0.10184851288795471,
0.05473201721906662,
-0.07261118292808533,
0.13409635424613953,
0.20412883162498474,
-0.013446791097521782,
0.12646126747131348,
-0.04460052773356438,
-0.0014522492419928312,
-0.03816164284944534,
-0.020009445026516914,
-0.02137279510498047,
0.08179881423711777,
-0.1630268096923828,
0.01751924306154251,
0.0470045767724514,
-0.011983836069703102,
0.06220005825161934,
-0.10581555217504501,
-0.023210866376757622,
0.004364258609712124,
-0.005766093265265226,
0.02773655392229557,
0.05236024037003517,
-0.019263654947280884,
0.08496736735105515,
-0.03065665438771248,
-0.06358969211578369,
0.06732945144176483,
-0.011463158763945103,
-0.08596393465995789,
0.13751524686813354,
-0.11103136837482452,
-0.18081502616405487,
-0.16712398827075958,
-0.0027328755240887403,
-0.13000614941120148,
0.041651371866464615,
0.06712692975997925,
-0.004649811889976263,
-0.02706201933324337,
-0.10119559615850449,
0.025100957602262497,
0.02470719814300537,
0.03884733095765114,
0.048308517783880234,
0.018482422456145287,
-0.0004338532453402877,
-0.07994646579027176,
-0.05238897353410721,
-0.016251850873231888,
0.009238026104867458,
0.0864136815071106,
-0.05023667961359024,
0.1320735663175583,
0.10757839679718018,
0.029998470097780228,
-0.004328992683440447,
-0.010078094899654388,
0.20026656985282898,
-0.0251818485558033,
0.02168155461549759,
0.22240331768989563,
-0.05836157500743866,
0.020628517493605614,
0.16712415218353271,
0.010491739958524704,
-0.07174950838088989,
0.026649927720427513,
-0.04784445837140083,
-0.04738346487283707,
-0.1615462601184845,
-0.05911945179104805,
-0.0644928440451622,
0.15266044437885284,
-0.01106052752584219,
0.04441892355680466,
0.018928758800029755,
0.07691600173711777,
-0.03623674437403679,
0.016382703557610512,
0.06329519301652908,
0.056693825870752335,
0.1887945532798767,
-0.053211111575365067,
0.0988832339644432,
-0.07115130126476288,
-0.08344012498855591,
0.06807060539722443,
0.10123275965452194,
0.08165156841278076,
0.09782152622938156,
0.09456153213977814,
0.04844824597239494,
0.0018679662607610226,
0.04534853249788284,
0.11663892865180969,
0.022441847249865532,
-0.025299072265625,
-0.049507688730955124,
-0.08303321897983551,
-0.03347301110625267,
0.04562884941697121,
-0.1159895807504654,
0.029557041823863983,
-0.0934181809425354,
0.007651497609913349,
0.0924600288271904,
0.1589299887418747,
0.007012668065726757,
-0.19508305191993713,
-0.062372032552957535,
0.1085396483540535,
0.04697912186384201,
-0.012889044359326363,
0.007807722315192223,
0.01115560345351696,
-0.01242709532380104,
0.1079925075173378,
0.01310005597770214,
0.07787007838487625,
-0.006311329081654549,
0.06019208952784538,
-0.07052561640739441,
0.0013036574237048626,
0.03315586969256401,
0.05698173865675926,
-0.27117952704429626,
0.1434236764907837,
0.006242765579372644,
0.011193514801561832,
0.0051813372410833836,
0.007197816856205463,
-0.03191769868135452,
0.20889024436473846,
0.10040923207998276,
-0.010065441951155663,
-0.0250406414270401,
-0.028321973979473114,
-0.06652429699897766,
0.029765252023935318,
0.04382074251770973,
-0.06551490724086761,
0.0476163811981678,
-0.005867300555109978,
-0.0037227857392281294,
0.004383449908345938,
0.09244056046009064,
-0.14678217470645905,
-0.11029183119535446,
0.07900767028331757,
-0.012485899031162262,
0.10178448259830475,
-0.09952179342508316,
-0.035765718668699265,
-0.16754856705665588,
0.16006016731262207,
-0.10398310422897339,
-0.05932971462607384,
-0.10838135331869125,
-0.08149999380111694,
0.021314134821295738,
-0.057229310274124146,
0.02615741826593876,
0.00017724561621434987,
0.06492356210947037,
-0.06190083175897598,
-0.17401419579982758,
0.08803998678922653,
-0.10735628008842468,
-0.11117969453334808,
-0.023704620078206062,
0.11986052244901657,
-0.057013604789972305,
-0.004291614517569542,
0.00908273458480835,
-0.030539385974407196,
-0.05109125003218651,
-0.09323035925626755,
0.027735797688364983,
0.14241643249988556,
-0.06361537426710129,
0.008679850958287716,
-0.03651668131351471,
-0.10903482139110565,
0.004086615517735481,
-0.014972371980547905,
0.11302291601896286,
0.26002880930900574,
-0.01110454648733139,
0.035779181867837906,
0.15156950056552887,
-0.06140056625008583,
-0.26861199736595154,
-0.07430744916200638,
-0.07402439415454865,
-0.0033498078119009733,
-0.006689721718430519,
-0.03423558920621872,
0.068255215883255,
0.12077422440052032,
-0.020073959603905678,
0.0018089988734573126,
-0.2614000737667084,
-0.11828958988189697,
0.13405399024486542,
0.08715468645095825,
0.2513774037361145,
-0.19069650769233704,
-0.07557231932878494,
-0.14182782173156738,
-0.20966272056102753,
0.1469632387161255,
-0.1421690434217453,
0.05172359198331833,
-0.02756606414914131,
0.019715577363967896,
0.0301392562687397,
-0.04710148274898529,
0.13595029711723328,
-0.008246623910963535,
0.07520484179258347,
-0.09688775986433029,
0.00989481434226036,
0.06519176065921783,
-0.0405818372964859,
0.08307516574859619,
-0.049798473715782166,
0.04122369363903999,
-0.03388402611017227,
-0.041952814906835556,
-0.01740163005888462,
0.06082584708929062,
-0.0345677025616169,
-0.06663966923952103,
-0.005163605324923992,
0.01657855324447155,
-0.05097274109721184,
0.01684807799756527,
0.14802499115467072,
-0.03398580104112625,
-0.01198986079543829,
0.14071321487426758,
0.08820106834173203,
-0.1402696818113327,
0.031053932383656502,
-0.024984726682305336,
-0.0598808154463768,
0.06658963859081268,
-0.024034680798649788,
0.02671782299876213,
0.11823156476020813,
-0.010535924695432186,
0.10363085567951202,
0.06256964802742004,
-0.026882855221629143,
-0.024485234171152115,
0.09224049746990204,
-0.22276516258716583,
-0.1956985592842102,
-0.06347592175006866,
0.04126576706767082,
-0.000952117086853832,
0.1789514273405075,
0.1685580462217331,
-0.04962676763534546,
0.008535650558769703,
-0.006024355534464121,
0.01054761279374361,
-0.06538569927215576,
0.08555150777101517,
0.06428435444831848,
0.02941901795566082,
-0.07739179581403732,
0.08204388618469238,
-0.018887242302298546,
-0.07136434316635132,
-0.021651482209563255,
0.046495307236909866,
-0.0949743464589119,
-0.10446111112833023,
-0.0938258171081543,
0.14246036112308502,
-0.081122025847435,
-0.07036575675010681,
-0.1368720531463623,
-0.16565188765525818,
0.024887189269065857,
0.12588825821876526,
0.06793753802776337,
0.02697572112083435,
0.04463374242186546,
0.0014737553428858519,
0.02262645587325096,
0.06510761380195618,
-0.021410757675766945,
0.10239685326814651,
-0.12194766849279404,
0.01678403466939926,
-0.05482497066259384,
0.059585344046354294,
-0.049360476434230804,
0.01820957660675049,
-0.16339576244354248,
-0.008141102269291878,
-0.11648555845022202,
-0.03586370870471001,
-0.10230779647827148,
-0.023024871945381165,
0.0032564017456024885,
-0.01420301292091608,
-0.01774682104587555,
0.0013963980600237846,
-0.03828893601894379,
-0.008525235578417778,
0.0018213833682239056,
0.05282388627529144,
-0.05707675591111183,
-0.023833902552723885,
0.039764005690813065,
-0.07018796354532242,
0.07494021207094193,
0.055923379957675934,
-0.024103563278913498,
0.043184444308280945,
-0.2027996927499771,
-0.006322725676000118,
0.06745176017284393,
0.03977154195308685,
-0.0030429877806454897,
0.0008392808958888054,
-0.035204362124204636,
0.0724310353398323,
-0.05440386012196541,
-0.0031103738583624363,
0.0665121003985405,
-0.08796408772468567,
-0.020659660920500755,
0.008495019748806953,
-0.07259443402290344,
-0.05421734228730202,
-0.04859442263841629,
0.08071230351924896,
0.08622920513153076,
0.12016674131155014,
-0.058868128806352615,
0.01891157403588295,
-0.13504642248153687,
0.01914951764047146,
0.04523076117038727,
-0.10530801862478256,
-0.0367421880364418,
-0.08441215753555298,
0.021110478788614273,
0.008645561523735523,
0.16679280996322632,
-0.07334452867507935,
-0.05440414696931839,
0.020703241229057312,
-0.04904770106077194,
0.04320447891950607,
0.07100527733564377,
0.3274634778499603,
0.05534818395972252,
0.04444640129804611,
-0.08235176652669907,
0.06514230370521545,
0.056444860994815826,
0.037661563605070114,
0.0669923946261406,
0.09578892588615417,
-0.11294138431549072,
0.1351012885570526,
0.12825407087802887,
0.08342301100492477,
-0.007573386188596487,
-0.029896285384893417,
-0.008356320671737194,
0.014672599732875824,
-0.06436090171337128,
0.16587060689926147,
0.16145439445972443,
-0.052876491099596024,
0.007789445109665394,
0.007656508591026068,
-0.04454630985856056,
-0.09233305603265762,
-0.12407036125659943,
-0.10892115533351898,
-0.14647193253040314,
-0.016784675419330597,
-0.10536077618598938,
-0.046264395117759705,
0.015090398490428925,
0.01753421686589718,
-0.061465878039598465,
0.10042089968919754,
0.009164905175566673,
-0.005470203701406717,
0.030252642929553986,
-0.020597251132130623,
-0.06910496950149536,
-0.05549890175461769,
-0.07756632566452026,
0.009941806085407734,
0.04205501824617386,
-0.04959389194846153,
0.020282402634620667,
-0.001315105240792036,
0.007725934032350779,
-0.016490885987877846,
-0.06836531311273575,
-0.026508739218115807,
0.011656505987048149,
0.01665467768907547,
0.11089292168617249,
0.02412423864006996,
-0.04706769064068794,
0.02950419671833515,
0.1311226487159729,
-0.02742348425090313,
-0.12243623286485672,
-0.06449130922555923,
0.0333336666226387,
0.019337039440870285,
0.09473215043544769,
-0.0040061939507722855,
-0.09270146489143372,
0.011811478063464165,
0.19295050203800201,
0.21739105880260468,
-0.07345376908779144,
0.03877090662717819,
-0.025625936686992645,
0.026269476860761642,
0.025013502687215805,
0.06523866951465607,
0.059925712645053864,
0.12360144406557083,
0.008973703719675541,
0.046952106058597565,
-0.02556004747748375,
-0.025963665917515755,
-0.07504455745220184,
0.06325247883796692,
0.034165043383836746,
-0.0171594750136137,
-0.028127243742346764,
0.09566503018140793,
-0.09901393204927444,
-0.02223069965839386,
-0.04411740228533745,
-0.10228735953569412,
-0.03912002593278885,
-0.04530197009444237,
0.09260525554418564,
-0.03346743807196617,
0.02576509490609169,
-0.0605432465672493,
-0.013557014055550098,
0.13453741371631622,
0.018304714933037758,
-0.09841043502092361,
-0.030827302485704422,
0.02099486067891121,
0.03142525628209114,
0.06700054556131363,
-0.03289197385311127,
0.0244978629052639,
0.08129042387008667,
0.03793754801154137,
-0.11629147082567215,
0.12729012966156006,
-0.013188311830163002,
-0.05506793409585953,
0.011962398886680603,
0.010021657682955265,
-0.01532448548823595,
0.049294695258140564,
0.035743892192840576,
-0.13131074607372284,
-0.012302319519221783,
0.10006105899810791,
-0.06793475896120071,
-0.06952448934316635,
0.10650249570608139,
-0.10043103247880936,
0.09412945806980133,
0.05127296224236488,
-0.04808717966079712,
0.006181302014738321,
-0.057732630521059036,
0.04870356619358063,
0.012458233162760735,
-0.029664693400263786,
0.013001251965761185,
-0.1501396745443344,
-0.019661404192447662,
0.07206308096647263,
0.017136406153440475,
-0.20322908461093903,
-0.0671936422586441,
-0.1592550128698349,
0.025263873860239983,
-0.06573198735713959,
0.04008079692721367,
0.16597601771354675,
-0.019428525120019913,
-0.03891589120030403,
-0.06216590106487274,
-0.020329948514699936,
0.09549165517091751,
-0.07642897218465805,
-0.1171891838312149
] |
null | null | diffusers |
# LoRA text2image fine-tuning - nightdude/kanji-lora
These are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were fine-tuned on the nightdude/sakana-kanji dataset. You can find some example images in the following.




| {"license": "creativeml-openrail-m", "tags": ["stable-diffusion", "stable-diffusion-diffusers", "text-to-image", "diffusers", "lora"], "base_model": "CompVis/stable-diffusion-v1-4", "inference": true} | text-to-image | nightdude/kanji-lora | [
"diffusers",
"stable-diffusion",
"stable-diffusion-diffusers",
"text-to-image",
"lora",
"base_model:CompVis/stable-diffusion-v1-4",
"license:creativeml-openrail-m",
"region:us"
] | 2024-02-08T06:35:22+00:00 | [] | [] | TAGS
#diffusers #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us
|
# LoRA text2image fine-tuning - nightdude/kanji-lora
These are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were fine-tuned on the nightdude/sakana-kanji dataset. You can find some example images in the following.
!img_0
!img_1
!img_2
!img_3
| [
"# LoRA text2image fine-tuning - nightdude/kanji-lora\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were fine-tuned on the nightdude/sakana-kanji dataset. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3"
] | [
"TAGS\n#diffusers #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us \n",
"# LoRA text2image fine-tuning - nightdude/kanji-lora\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were fine-tuned on the nightdude/sakana-kanji dataset. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3"
] | [
66,
93
] | [
"passage: TAGS\n#diffusers #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us \n# LoRA text2image fine-tuning - nightdude/kanji-lora\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were fine-tuned on the nightdude/sakana-kanji dataset. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3"
] | [
-0.05316605418920517,
-0.127903550863266,
-0.0014596331166103482,
0.012051735073328018,
0.06352008134126663,
0.028487004339694977,
0.22753800451755524,
0.11040426790714264,
-0.10088705271482468,
0.06337901949882507,
0.027909278869628906,
0.002505481941625476,
0.06274986267089844,
0.10665522515773773,
0.02219214104115963,
-0.25689926743507385,
0.013761523179709911,
-0.012637909501791,
-0.04797176271677017,
0.07006634026765823,
0.09960303455591202,
-0.0446910634636879,
0.11315546184778214,
-0.017514243721961975,
-0.12986432015895844,
-0.01722165197134018,
-0.05184317007660866,
-0.11490947008132935,
0.0636788159608841,
0.04124892130494118,
-0.02885696291923523,
0.11545316129922867,
0.06263229250907898,
-0.12135536968708038,
0.026840554550290108,
-0.024662893265485764,
-0.023203393444418907,
0.05722029507160187,
-0.04294488951563835,
0.04095836728811264,
0.12943074107170105,
-0.029657498002052307,
-0.050092633813619614,
0.013531525619328022,
-0.10335951298475266,
-0.1039905995130539,
-0.03735991194844246,
-0.012560161761939526,
0.08286122977733612,
-0.012531736865639687,
0.0007858013268560171,
-0.000047684083256172016,
-0.09818878769874573,
0.018530936911702156,
0.2916872799396515,
-0.20904597640037537,
-0.018806561827659607,
0.19374100863933563,
0.03971274942159653,
0.12113044410943985,
-0.05872088298201561,
0.06460727751255035,
0.06956467032432556,
-0.09306206554174423,
0.07068105041980743,
-0.051036350429058075,
-0.009665531106293201,
-0.04943278431892395,
-0.0743434876203537,
0.10529514402151108,
0.24162322282791138,
-0.004758676514029503,
-0.04837894067168236,
-0.14809605479240417,
-0.0019732911605387926,
0.05868389829993248,
-0.07537117600440979,
0.03453387692570686,
-0.002576821018010378,
-0.025951707735657692,
0.06584160029888153,
-0.0007057348848320544,
-0.07742289453744888,
-0.019612688571214676,
-0.03179695084691048,
0.1754598617553711,
0.017321595922112465,
0.024600371718406677,
0.00175651372410357,
0.08181484043598175,
-0.18613775074481964,
-0.16470219194889069,
-0.008985022082924843,
-0.049351535737514496,
0.09713300317525864,
0.07406196743249893,
0.03982330858707428,
-0.08888881653547287,
0.06158270686864853,
0.08627738058567047,
0.1993916928768158,
-0.00874504167586565,
0.009881600737571716,
0.10657142102718353,
-0.005743378307670355,
0.07934422045946121,
-0.07330088317394257,
-0.08278767764568329,
0.0699186697602272,
0.014840828254818916,
0.1087040826678276,
-0.05141036957502365,
-0.08855868875980377,
-0.09307639300823212,
-0.0940941721200943,
0.06782763451337814,
-0.059730466455221176,
0.04684138298034668,
-0.0678514689207077,
-0.028021089732646942,
0.16484908759593964,
-0.029345672577619553,
0.001210891641676426,
-0.037327177822589874,
-0.03124343231320381,
0.1681218445301056,
0.07948954403400421,
0.03015478141605854,
0.05436854064464569,
0.035737525671720505,
-0.0262320414185524,
0.0264323391020298,
-0.0043416013941168785,
-0.07352223992347717,
0.0349850058555603,
-0.1618218570947647,
0.009743329137563705,
-0.14050598442554474,
-0.03097139298915863,
0.055042821913957596,
0.009418906643986702,
-0.04509783908724785,
0.06534117460250854,
-0.05056910589337349,
-0.031558915972709656,
-0.004700012505054474,
0.00017987992032431066,
0.027729736641049385,
-0.06130823493003845,
0.045873451977968216,
0.05454450845718384,
0.163425475358963,
-0.06538940221071243,
0.00585463922470808,
-0.07785490155220032,
0.020235443487763405,
-0.11552761495113373,
-0.04623937979340553,
-0.005044615361839533,
0.02672012895345688,
-0.07524172961711884,
-0.013808450661599636,
0.004508591257035732,
0.011195399798452854,
0.014857448637485504,
0.18547241389751434,
-0.24701443314552307,
-0.04269848391413689,
0.06817227602005005,
-0.1423201709985733,
-0.0610196590423584,
0.06728173792362213,
0.04323620721697807,
0.16404664516448975,
0.05118977278470993,
0.12716349959373474,
0.048421867191791534,
-0.20072847604751587,
0.07444838434457779,
-0.05091838538646698,
-0.051575906574726105,
-0.06314167380332947,
0.09705068171024323,
0.08944620192050934,
0.0017718252493068576,
0.06992553919553757,
-0.053283609449863434,
0.05756461247801781,
-0.06800998747348785,
-0.04061346501111984,
0.02698059007525444,
-0.007893114350736141,
0.018311556428670883,
-0.013219902291893959,
0.03901410102844238,
-0.047126173973083496,
-0.017191363498568535,
-0.07408876717090607,
0.1101103201508522,
-0.026686541736125946,
0.05988776311278343,
-0.09808244556188583,
0.19080618023872375,
-0.12606410682201385,
0.002442219527438283,
-0.011951561085879803,
0.12684351205825806,
0.041869986802339554,
0.1753225028514862,
0.04281647130846977,
-0.019122153520584106,
0.07971441000699997,
0.03349977359175682,
-0.049009546637535095,
0.009020759724080563,
0.06117691844701767,
-0.023385243490338326,
0.00011297173477942124,
-0.2004636526107788,
0.04795892536640167,
-0.07859480381011963,
0.02882969379425049,
-0.1283259242773056,
-0.0300186425447464,
0.040915101766586304,
0.10546600073575974,
0.10052531212568283,
0.007574562449008226,
0.06368027627468109,
-0.002210465958341956,
-0.07424133270978928,
0.012039086781442165,
-0.0008607096970081329,
-0.029095502570271492,
-0.12933003902435303,
0.25363272428512573,
-0.08947614580392838,
0.08045628666877747,
0.10065368562936783,
-0.004281025379896164,
-0.026495205238461494,
-0.04539451375603676,
-0.0027119657024741173,
-0.007058974355459213,
-0.0024359552189707756,
0.016215968877077103,
-0.05371718108654022,
-0.049555376172065735,
0.08101287484169006,
-0.024691695347428322,
0.04919520765542984,
0.0742623582482338,
-0.07040423899888992,
-0.09408298134803772,
0.05635442957282066,
0.037011370062828064,
0.11228903383016586,
0.0033086384646594524,
0.1266976147890091,
-0.018211161717772484,
0.12718342244625092,
-0.027280591428279877,
-0.08308134227991104,
-0.061972856521606445,
0.021316491067409515,
0.06479568034410477,
0.17570006847381592,
-0.005276735406368971,
-0.019833166152238846,
0.02548721246421337,
-0.015825491398572922,
0.028116077184677124,
-0.07817596942186356,
-0.06330884993076324,
0.027032770216464996,
-0.043293315917253494,
0.05772053077816963,
0.11115481704473495,
-0.09822972118854523,
0.1429016888141632,
-0.06806453317403793,
-0.047744281589984894,
-0.05385711416602135,
-0.022755062207579613,
-0.037622589617967606,
0.05854762718081474,
-0.07862741500139236,
-0.17567504942417145,
-0.12745407223701477,
0.035933755338191986,
0.011072423309087753,
-0.03210132569074631,
0.06662321090698242,
-0.11077608168125153,
-0.08792899549007416,
-0.10450851172208786,
0.12142670154571533,
0.10217829048633575,
0.025484181940555573,
-0.016080573201179504,
-0.002496195025742054,
-0.06099449843168259,
-0.07557518780231476,
0.005083631258457899,
-0.02463565766811371,
0.08566413819789886,
0.07605347037315369,
-0.05306507274508476,
0.14513051509857178,
0.10483936965465546,
0.023223217576742172,
-0.004202928394079208,
0.05164070799946785,
0.10211194306612015,
-0.04357186704874039,
0.08460631221532822,
0.15828858315944672,
0.0536244735121727,
0.008267025463283062,
0.09254620969295502,
0.02321290783584118,
-0.03172760829329491,
0.0037564942613244057,
-0.019812200218439102,
-0.12156376242637634,
-0.028742805123329163,
-0.02107398398220539,
-0.08646083623170853,
0.015551966615021229,
-0.00882970541715622,
-0.0032971263863146305,
0.018861910328269005,
0.09459029883146286,
-0.0012481248704716563,
-0.018735799938440323,
0.10565382987260818,
0.0921517089009285,
0.010241275653243065,
-0.0354035310447216,
-0.003041032236069441,
-0.060131967067718506,
-0.08098334819078445,
0.13523654639720917,
-0.03222087398171425,
0.10371677577495575,
-0.06200703978538513,
0.1319812536239624,
0.061429087072610855,
0.06290572881698608,
0.08844173699617386,
0.0445222370326519,
-0.08302218466997147,
-0.03291136771440506,
-0.051013220101594925,
-0.13222159445285797,
0.07223405689001083,
0.06543386727571487,
0.0398687943816185,
-0.036307305097579956,
0.009015143848955631,
0.006421645171940327,
0.0046271588653326035,
0.044911205768585205,
0.15008652210235596,
-0.2276601791381836,
0.07599198818206787,
0.04896153509616852,
0.09491182118654251,
-0.017453128471970558,
0.018274610862135887,
0.09450355172157288,
-0.025753924623131752,
0.060751404613256454,
-0.05063508078455925,
0.05270068719983101,
-0.0018095027189701796,
-0.08546049147844315,
-0.14661943912506104,
0.06444419920444489,
-0.03887602314352989,
0.0017123802099376917,
-0.23383614420890808,
0.06739386171102524,
0.005172302946448326,
0.021862996742129326,
-0.07829470187425613,
-0.0472603514790535,
0.013106009922921658,
0.02207665704190731,
0.06330303847789764,
0.00597167294472456,
0.12665551900863647,
-0.07215071469545364,
-0.19325245916843414,
-0.004330793395638466,
0.05956050753593445,
-0.04443925619125366,
0.029205424711108208,
0.032828282564878464,
-0.02973417192697525,
0.015554440207779408,
-0.025215666741132736,
-0.14072445034980774,
-0.05632443726062775,
0.05575364828109741,
0.12063534557819366,
-0.02364966832101345,
-0.023255841806530952,
-0.08799933642148972,
-0.04640081524848938,
0.07594238221645355,
0.01033175177872181,
-0.06247291713953018,
-0.08054860681295395,
-0.12562203407287598,
0.12974996864795685,
-0.008264843374490738,
-0.009598342701792717,
-0.02940271608531475,
0.09090835601091385,
-0.03213926777243614,
-0.13424989581108093,
0.051385827362537384,
-0.02808181755244732,
-0.12927722930908203,
-0.08185526728630066,
0.12229179590940475,
-0.0022087451070547104,
-0.016796069219708443,
-0.002621019957587123,
0.021186817437410355,
0.05726999789476395,
-0.048974744975566864,
0.01719195768237114,
0.08952439576387405,
-0.08342071622610092,
0.06568646430969238,
-0.023772669956088066,
0.01587597094476223,
-0.03332950919866562,
-0.08031415939331055,
0.05246521532535553,
0.269439697265625,
-0.052479613572359085,
0.13772310316562653,
0.10636789351701736,
-0.03267383575439453,
-0.1452694535255432,
-0.08250151574611664,
-0.0834013968706131,
0.04305507615208626,
0.04523377865552902,
-0.032969679683446884,
0.0456356480717659,
0.06600618362426758,
-0.019875748082995415,
0.21033728122711182,
-0.43959301710128784,
-0.12678298354148865,
-0.023701420053839684,
0.15437301993370056,
0.2510429322719574,
-0.1408194750547409,
-0.029388798400759697,
-0.03320247307419777,
-0.15263383090496063,
-0.0337204746901989,
0.04731246829032898,
0.09997788816690445,
-0.09566665440797806,
0.0037210655864328146,
0.03921456262469292,
-0.018757840618491173,
0.1681617945432663,
-0.004719150252640247,
0.018845219165086746,
-0.15377314388751984,
-0.09926818311214447,
0.1661268174648285,
-0.04810355603694916,
0.05163397639989853,
-0.1591326892375946,
-0.008961893618106842,
-0.12726442515850067,
-0.002030493225902319,
0.027204236015677452,
-0.0181445125490427,
-0.039341818541288376,
-0.05617592856287956,
-0.10681931674480438,
0.09971191734075546,
0.005342730786651373,
-0.03036360628902912,
0.18910960853099823,
0.02227167785167694,
-0.04211093857884407,
0.0827564001083374,
-0.03646513819694519,
0.09080329537391663,
0.010720187798142433,
-0.03830898180603981,
-0.005221775732934475,
0.11757280677556992,
-0.10408825427293777,
0.02981586754322052,
0.11198867857456207,
0.07033643871545792,
0.09180848300457001,
-0.018582947552204132,
-0.10674899071455002,
0.11741367727518082,
0.18833675980567932,
-0.07138188183307648,
-0.008533500134944916,
-0.006664377171546221,
-0.13422012329101562,
0.10866602510213852,
0.056579895317554474,
0.10868308693170547,
-0.1007702499628067,
0.0008723256760276854,
0.013846689835190773,
0.10799336433410645,
-0.05854148045182228,
0.12639136612415314,
0.026495836675167084,
-0.0013880492188036442,
-0.0765763372182846,
0.0863742083311081,
-0.016732076182961464,
0.062219198793172836,
-0.05758964642882347,
0.05807062238454819,
-0.06678615510463715,
-0.019129756838083267,
0.1474532186985016,
0.05461056903004646,
-0.12792401015758514,
0.002604002133011818,
-0.12110105156898499,
-0.11228827387094498,
-0.0061314962804317474,
0.05838686227798462,
0.038354963064193726,
0.04272746667265892,
0.019990872591733932,
-0.01874598301947117,
-0.08609657734632492,
0.05469664931297302,
0.12099602073431015,
0.09660797566175461,
-0.14369256794452667,
-0.0675186961889267,
-0.017284706234931946,
0.0004414654104039073,
-0.06810791790485382,
-0.08168360590934753,
-0.08317779749631882,
-0.004440590273588896,
-0.13653071224689484,
0.0768006443977356,
-0.14276249706745148,
-0.0359182246029377,
-0.0674041360616684,
-0.025571437552571297,
-0.030061645433306694,
-0.005678053945302963,
-0.00930071622133255,
-0.003053357359021902,
-0.0503145232796669,
0.06914487481117249,
-0.02291635423898697,
-0.0949244350194931,
-0.009854437783360481,
-0.10541996359825134,
0.044303301721811295,
0.021225502714514732,
-0.011576044373214245,
-0.012164393439888954,
-0.08110478520393372,
0.007683637086302042,
0.12247655540704727,
-0.0044550164602696896,
0.004310997202992439,
-0.08663881570100784,
0.014171433635056019,
-0.012187562882900238,
0.01770123466849327,
-0.02090589888393879,
0.08149388432502747,
-0.1059565395116806,
0.008939662016928196,
-0.19165928661823273,
-0.01752258464694023,
-0.05659713223576546,
0.0668359324336052,
0.09492938220500946,
0.08712279051542282,
0.1186988428235054,
-0.07662304490804672,
0.042496293783187866,
-0.03883367404341698,
-0.0030498774722218513,
-0.003714535851031542,
-0.08650236576795578,
0.029291441664099693,
-0.00968005321919918,
-0.029606590047478676,
-0.043919339776039124,
0.10947326570749283,
0.00016938141197897494,
-0.19417084753513336,
-0.024734513834118843,
-0.08517339080572128,
-0.007566153537482023,
0.03563372790813446,
0.2331077754497528,
0.027414515614509583,
0.0341373085975647,
-0.07821627706289291,
0.07635896652936935,
0.08925387263298035,
0.023148158565163612,
-0.004494918044656515,
0.11063958704471588,
-0.006860257592052221,
0.11179321259260178,
0.05068599432706833,
-0.05214989557862282,
0.0253159087151289,
0.07111204415559769,
-0.09572605043649673,
0.0752120316028595,
-0.015233531594276428,
0.004548563156276941,
0.22067204117774963,
-0.08050992339849472,
-0.02733268402516842,
0.03416227921843529,
-0.028179040178656578,
0.00938814040273428,
-0.21999302506446838,
-0.094418965280056,
-0.13137775659561157,
0.04006708785891533,
-0.07425019890069962,
0.03360055759549141,
0.042850758880376816,
0.04771755263209343,
0.0452430360019207,
0.057505082339048386,
0.04667067900300026,
-0.11084263771772385,
0.15336041152477264,
-0.014656337909400463,
-0.10854493081569672,
0.03365713357925415,
0.01590188965201378,
0.031643085181713104,
-0.02656496688723564,
-0.027418650686740875,
0.09304425865411758,
0.058539390563964844,
-0.0015064169419929385,
-0.02471732534468174,
-0.10258214920759201,
-0.015285752713680267,
-0.02813088148832321,
0.1190285012125969,
0.17106108367443085,
0.0835329219698906,
-0.047466352581977844,
0.002144949045032263,
0.1360931694507599,
-0.014913287945091724,
-0.015403411351144314,
-0.08889426290988922,
-0.04135563597083092,
-0.06784151494503021,
0.004280258901417255,
-0.05293114110827446,
-0.08144673705101013,
-0.005104356911033392,
0.18217240273952484,
0.202095165848732,
0.003329323371872306,
-0.019201504066586494,
-0.08982852846384048,
-0.0008981871651485562,
-0.0013993819011375308,
0.09063021838665009,
0.0158116165548563,
0.15998777747154236,
-0.035199254751205444,
-0.03547457978129387,
-0.13117602467536926,
-0.017833733931183815,
-0.07165737450122833,
-0.009780749678611755,
0.03257148340344429,
-0.08018531650304794,
-0.07651802152395248,
0.07585927098989487,
-0.0425148606300354,
-0.09095612168312073,
0.12416412681341171,
-0.1453201025724411,
-0.06139621138572693,
-0.08595316112041473,
-0.022311247885227203,
0.044196877628564835,
0.01518932543694973,
-0.08667215704917908,
-0.023036019876599312,
-0.03925304114818573,
0.035314176231622696,
-0.06967940926551819,
-0.04940330237150192,
-0.03139331936836243,
-0.04727796092629433,
0.14133231341838837,
-0.018314452841877937,
0.0386495478451252,
0.009116467088460922,
0.03853246942162514,
-0.09119532257318497,
0.07469694316387177,
-0.03096703626215458,
-0.04467487335205078,
-0.0700811967253685,
0.11987844109535217,
-0.011071907356381416,
0.06602595001459122,
0.05932093411684036,
-0.017422761768102646,
0.028875049203634262,
0.10220184922218323,
-0.016512775793671608,
-0.07118202745914459,
-0.011684701777994633,
-0.0903404951095581,
0.11861360818147659,
0.03132754936814308,
-0.04470665007829666,
-0.058834612369537354,
-0.004038104321807623,
0.046057622879743576,
0.07174194604158401,
-0.026025250554084778,
0.031444333493709564,
-0.08100984245538712,
-0.08580675721168518,
0.002668118802830577,
0.039191897958517075,
-0.18107542395591736,
-0.07747145742177963,
-0.15077435970306396,
-0.027884425595402718,
-0.011244392022490501,
0.10728257894515991,
0.20540277659893036,
0.013975118286907673,
0.009771726094186306,
-0.316349059343338,
0.04709215089678764,
0.0952526107430458,
-0.13927878439426422,
-0.08627916872501373
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.8.2 | {"library_name": "peft", "base_model": "meta-llama/Llama-2-7b-hf"} | null | ManuThakur/Llama2Trained | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:meta-llama/Llama-2-7b-hf",
"region:us"
] | 2024-02-08T06:37:31+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #safetensors #arxiv-1910.09700 #base_model-meta-llama/Llama-2-7b-hf #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.8.2 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
"TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-meta-llama/Llama-2-7b-hf #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
41,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-meta-llama/Llama-2-7b-hf #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2"
] | [
-0.13237035274505615,
0.20393601059913635,
-0.002511046128347516,
0.02874687872827053,
0.07912357151508331,
0.019634027034044266,
0.04979075863957405,
0.13531364500522614,
0.020043307915329933,
0.10580451786518097,
0.0737132653594017,
0.11718367785215378,
0.10926163196563721,
0.20654499530792236,
0.003890186781063676,
-0.15440793335437775,
0.024214256554841995,
-0.08298544585704803,
0.006814117077738047,
0.1290476769208908,
0.14319083094596863,
-0.10468140989542007,
0.0831538662314415,
-0.014203370548784733,
0.0008161105797626078,
-0.03246506303548813,
-0.06674343347549438,
-0.015596466138958931,
0.04917285591363907,
0.02522817626595497,
0.05882670730352402,
-0.010089844465255737,
0.0929119735956192,
-0.26152917742729187,
0.018749000504612923,
0.04154228791594505,
0.009074261412024498,
0.08363344520330429,
0.0979103073477745,
-0.04074648395180702,
0.12078511714935303,
-0.024994686245918274,
0.13832204043865204,
0.09345067292451859,
-0.08226727694272995,
-0.233157217502594,
-0.06684722006320953,
0.07271547615528107,
0.18968668580055237,
0.08927863836288452,
-0.044125091284513474,
0.14097759127616882,
-0.07517150044441223,
0.02484818734228611,
0.04656748101115227,
-0.09290260076522827,
-0.06676048040390015,
0.0702265128493309,
0.13261590898036957,
0.0625041052699089,
-0.12113244831562042,
-0.03750992938876152,
0.03344248607754707,
0.044793009757995605,
0.06062353774905205,
0.005180627107620239,
0.16268815100193024,
0.034240271896123886,
-0.14592847228050232,
-0.05353321507573128,
0.14678435027599335,
0.01157673355191946,
-0.04636283218860626,
-0.21997328102588654,
-0.0027822081465274096,
-0.09489403665065765,
-0.022923149168491364,
-0.05228540673851967,
0.03324316069483757,
0.00603833794593811,
0.1196645051240921,
-0.042089227586984634,
-0.09635167568922043,
-0.029711460694670677,
0.0996040627360344,
0.05452839657664299,
0.02769845724105835,
-0.02099502831697464,
0.010653719305992126,
0.1290775090456009,
0.08296726644039154,
-0.1341402530670166,
-0.07021861523389816,
-0.0753326416015625,
-0.04316629841923714,
-0.03228989988565445,
0.03893959894776344,
0.019871119409799576,
0.07120058685541153,
0.2619621157646179,
-0.022196462377905846,
0.06401924788951874,
0.061033982783555984,
0.01709051802754402,
0.04062429443001747,
0.10795178264379501,
-0.03382651507854462,
-0.15705206990242004,
-0.007360270246863365,
0.10362072288990021,
-0.004135396331548691,
-0.02802850492298603,
-0.045986633747816086,
0.03152812272310257,
0.044165465980768204,
0.11501371115446091,
0.11203816533088684,
-0.019931387156248093,
-0.07717939466238022,
-0.05966082587838173,
0.19364216923713684,
-0.16149258613586426,
0.038572292774915695,
0.02467195875942707,
-0.006866174750030041,
-0.06484853476285934,
0.0073310090228915215,
0.016164373606443405,
-0.027354510501027107,
0.0603426918387413,
-0.0646006166934967,
-0.04179375246167183,
-0.1283673793077469,
-0.02387934736907482,
0.032629046589136124,
0.0170845165848732,
-0.0421639084815979,
-0.046661876142024994,
-0.08786044269800186,
-0.11000633984804153,
0.10926247388124466,
-0.05313732475042343,
-0.052913907915353775,
-0.02804330736398697,
-0.08941388875246048,
0.022293368354439735,
0.027490468695759773,
0.0755976140499115,
-0.02891632728278637,
0.052480049431324005,
0.003703000722452998,
0.059941843152046204,
0.0814133733510971,
0.027145687490701675,
-0.08097686618566513,
0.06685694307088852,
-0.19895170629024506,
0.07886288315057755,
-0.08557034283876419,
0.035526763647794724,
-0.16191443800926208,
-0.008882720954716206,
0.015485688112676144,
0.028551144525408745,
0.0418417863547802,
0.16628479957580566,
-0.21890771389007568,
-0.021091977134346962,
0.15901808440685272,
-0.10847076028585434,
-0.1374696046113968,
0.0436418242752552,
-0.04286689683794975,
0.18280568718910217,
0.028055870905518532,
0.010343263857066631,
0.09726855903863907,
-0.16840705275535583,
-0.02907063439488411,
-0.021288467571139336,
0.0036895605735480785,
0.07365763932466507,
0.09041544795036316,
-0.09089618921279907,
-0.0016403654590249062,
0.012144356034696102,
-0.06943254172801971,
-0.015110267326235771,
-0.04118245840072632,
-0.10628213733434677,
0.002018203027546406,
-0.09110194444656372,
0.023759065195918083,
0.0035124430432915688,
-0.09477277845144272,
-0.008542876690626144,
-0.1573835164308548,
-0.0652049109339714,
0.09409166127443314,
0.0002530320198275149,
-0.024702679365873337,
-0.10900412499904633,
0.06465248018503189,
-0.03883763402700424,
-0.026517964899539948,
-0.14125961065292358,
-0.023071611300110817,
0.01673055998980999,
-0.14134323596954346,
-0.01001854706555605,
-0.12183605134487152,
0.06567396223545074,
0.005137317348271608,
-0.0481104739010334,
-0.04708600044250488,
-0.004086394794285297,
0.0014921361580491066,
-0.05505292862653732,
-0.23444515466690063,
-0.028233496472239494,
-0.05085372179746628,
0.16539393365383148,
-0.2289838343858719,
0.044271692633628845,
0.014694449491798878,
0.11615854501724243,
-0.0018446118338033557,
-0.0661761611700058,
0.022094158455729485,
-0.07084274291992188,
-0.025033291429281235,
-0.07177132368087769,
-0.0071777342818677425,
0.00008959023398347199,
-0.029647991061210632,
0.015313859097659588,
-0.10952108353376389,
-0.053884293884038925,
0.100620798766613,
0.060472261160612106,
-0.14894865453243256,
0.008543584495782852,
-0.03779032453894615,
-0.06071627512574196,
-0.07427168637514114,
-0.0695083886384964,
0.0856412947177887,
0.052977994084358215,
0.03996400535106659,
-0.0812206119298935,
-0.07201940566301346,
0.005019875708967447,
-0.02742239646613598,
-0.005877636838704348,
0.11996077746152878,
0.07278608530759811,
-0.10015858709812164,
0.0890948474407196,
0.07567999511957169,
0.012905389070510864,
0.07863839715719223,
-0.028960783034563065,
-0.10615462064743042,
-0.03149069845676422,
0.05891314521431923,
0.0075002689845860004,
0.18196412920951843,
-0.07219336181879044,
0.05777830258011818,
0.046155888587236404,
-0.046635568141937256,
0.05089704319834709,
-0.09103982150554657,
0.0068960352800786495,
0.00045980032882653177,
-0.017081741243600845,
0.029599705711007118,
-0.020320137962698936,
0.006365274079144001,
0.07632698118686676,
0.05559656023979187,
0.02392573468387127,
0.023359429091215134,
-0.037590380758047104,
-0.1454712599515915,
0.18398217856884003,
-0.09283597022294998,
-0.235765740275383,
-0.15705986320972443,
0.0616452731192112,
0.049257904291152954,
-0.015689486637711525,
0.02697811834514141,
-0.055544715374708176,
-0.10059839487075806,
-0.08630408346652985,
-0.001965506933629513,
0.033574361354112625,
-0.05912783369421959,
-0.07473962754011154,
0.045523062348365784,
0.04523130878806114,
-0.11779510229825974,
0.02612960711121559,
0.06724361330270767,
-0.01014306303113699,
0.002122951438650489,
0.05421233922243118,
0.09625556319952011,
0.1871589571237564,
-0.0047584883868694305,
0.006493487861007452,
0.06463784724473953,
0.27302834391593933,
-0.16097134351730347,
0.10603976994752884,
0.1468280404806137,
-0.06509615480899811,
0.06928659975528717,
0.1811111718416214,
0.024897225201129913,
-0.0959320068359375,
0.024916043505072594,
0.02835996262729168,
-0.01960386149585247,
-0.2740720212459564,
-0.0512622706592083,
-0.015117009170353413,
-0.08622704446315765,
0.07128944247961044,
0.08718991279602051,
0.07891540229320526,
0.03938929736614227,
-0.05623466521501541,
-0.11011259257793427,
0.02521095983684063,
0.10682129859924316,
-0.01211885642260313,
0.003295447211712599,
0.08167944848537445,
-0.04613311216235161,
0.007927946746349335,
0.08699803054332733,
-0.01990879327058792,
0.1374768167734146,
0.04775961861014366,
0.09206060320138931,
0.08603846281766891,
0.10468525439500809,
-0.011216369457542896,
0.031460702419281006,
0.01713097095489502,
0.023083847016096115,
0.025577327236533165,
-0.0892123356461525,
0.00939508993178606,
0.11217135936021805,
0.02443520911037922,
0.02237142249941826,
0.016059260815382004,
-0.042084116488695145,
0.035355109721422195,
0.19778503477573395,
0.02863113395869732,
-0.21936152875423431,
-0.08315163850784302,
0.04950554668903351,
-0.07752750813961029,
-0.15846198797225952,
-0.0069001950323581696,
0.02585102617740631,
-0.16377925872802734,
0.015679948031902313,
-0.04114160314202309,
0.10047675669193268,
-0.07824478298425674,
-0.04026156663894653,
0.11029542237520218,
0.047400183975696564,
-0.01943347603082657,
0.05451195687055588,
-0.19536079466342926,
0.10843666642904282,
0.02992161363363266,
0.07536879926919937,
-0.08786998689174652,
0.09398660063743591,
0.006047630682587624,
-0.019160762429237366,
0.16931316256523132,
-0.0001144029592978768,
-0.049934081733226776,
-0.08560120314359665,
-0.09227954596281052,
0.0015766898868605494,
0.07818529009819031,
-0.12631447613239288,
0.0825691819190979,
-0.03569265082478523,
-0.024482207372784615,
-0.008127174340188503,
-0.08541606366634369,
-0.1325976550579071,
-0.14982733130455017,
0.05399367958307266,
-0.0976201519370079,
0.02554609440267086,
-0.08825770765542984,
-0.05347679927945137,
0.016768373548984528,
0.18224331736564636,
-0.21447692811489105,
-0.10864878445863724,
-0.14267513155937195,
-0.11213549226522446,
0.16079570353031158,
-0.042837124317884445,
0.08159231394529343,
0.00010400224709883332,
0.15704618394374847,
0.01110734511166811,
-0.015090357512235641,
0.08682332187891006,
-0.09437134861946106,
-0.19026298820972443,
-0.04887847229838371,
0.16311104595661163,
0.1444961428642273,
0.029530119150877,
-0.005065699107944965,
0.02549002133309841,
-0.06952440738677979,
-0.11216824501752853,
0.02609189972281456,
0.16361786425113678,
0.07300680130720139,
-0.012950204312801361,
-0.025871867313981056,
-0.0997539535164833,
-0.05963310971856117,
-0.04339827224612236,
-0.00898770522326231,
0.20425592362880707,
-0.06497634947299957,
0.14582973718643188,
0.10464579612016678,
-0.05606960505247116,
-0.21339629590511322,
0.03492094576358795,
0.04277806729078293,
0.026418045163154602,
0.04313372075557709,
-0.18166027963161469,
0.09741673618555069,
-0.014149999246001244,
-0.08650295436382294,
0.17498920857906342,
-0.17328102886676788,
-0.13439859449863434,
0.1159968227148056,
0.025544147938489914,
-0.21331895887851715,
-0.13972461223602295,
-0.10190334171056747,
-0.0198976993560791,
-0.126362144947052,
0.036111894994974136,
-0.0036879852414131165,
0.00850605871528387,
0.012948633171617985,
0.018173353746533394,
0.039593230932950974,
-0.05594787001609802,
0.21268853545188904,
-0.03937339782714844,
0.000047609177272534,
-0.050931964069604874,
-0.06770505011081696,
0.023772839456796646,
-0.0565045028924942,
0.12416863441467285,
-0.01210821233689785,
0.039195943623781204,
-0.17265570163726807,
-0.04285977780818939,
-0.058010976761579514,
0.03728554770350456,
-0.09242235124111176,
-0.0793662965297699,
-0.04483490809798241,
0.09155189245939255,
0.09041202813386917,
-0.018728721886873245,
0.0019666242878884077,
-0.09585212171077728,
0.07403325289487839,
0.20964933931827545,
0.20306745171546936,
0.0681707113981247,
-0.05247919633984566,
0.02836998738348484,
-0.03519117832183838,
0.04444263130426407,
-0.2148476094007492,
0.0430048331618309,
0.0631239265203476,
0.024400800466537476,
0.06267635524272919,
-0.01054441649466753,
-0.1590016484260559,
-0.07973737269639969,
0.08659059554338455,
-0.0608268640935421,
-0.16209019720554352,
-0.03262902423739433,
0.02129248157143593,
-0.2115628719329834,
-0.04105594381690025,
0.03599734604358673,
-0.014814808964729309,
-0.03840542584657669,
0.021407432854175568,
0.07970889657735825,
-0.028947602957487106,
0.1049608662724495,
0.09329938143491745,
0.09604475647211075,
-0.09774979948997498,
0.05453461781144142,
0.07179035246372223,
-0.031663764268159866,
0.03226640820503235,
0.1210775151848793,
-0.04315068572759628,
-0.046701591461896896,
0.08053972572088242,
0.11871292442083359,
-0.00035442441003397107,
-0.06335891038179398,
-0.0028557574842125177,
-0.0440225712954998,
0.054060470312833786,
0.10412941128015518,
0.036388467997312546,
0.0012024412862956524,
0.07687212526798248,
0.028011957183480263,
-0.09147296100854874,
0.12449978291988373,
0.06066809967160225,
0.02483541890978813,
-0.05523430183529854,
-0.038621995598077774,
-0.015819178894162178,
-0.0028008304070681334,
-0.01961326226592064,
-0.0014547118917107582,
-0.08309019356966019,
0.0061004795134067535,
-0.13227513432502747,
0.022323906421661377,
-0.07725922018289566,
0.00378548726439476,
0.036021001636981964,
-0.046576302498579025,
0.0013563713291659951,
-0.0008801636286079884,
-0.07430332899093628,
-0.05454954877495766,
-0.01629588007926941,
0.07790114730596542,
-0.13923588395118713,
0.03906119614839554,
0.07606222480535507,
-0.10726266354322433,
0.06878530234098434,
-0.007731399964541197,
0.008601504378020763,
0.0010856596054509282,
-0.13779860734939575,
0.05484551563858986,
-0.028775036334991455,
-0.006356567144393921,
0.005071246065199375,
-0.19585701823234558,
-0.00865773856639862,
-0.03182972967624664,
-0.0634872317314148,
0.019731810316443443,
-0.001073729363270104,
-0.11955288797616959,
0.1077868640422821,
0.004837313666939735,
-0.05712589994072914,
-0.0236744936555624,
0.042738161981105804,
0.0863419771194458,
-0.0053856209851801395,
0.12532570958137512,
-0.0293873380869627,
0.07612910121679306,
-0.17633569240570068,
-0.010070881806313992,
-0.015794692561030388,
0.05993741378188133,
-0.019834399223327637,
-0.03712667524814606,
0.06236843764781952,
-0.027145320549607277,
0.17265751957893372,
-0.004146610386669636,
0.07253459841012955,
0.0493277981877327,
0.008650471456348896,
0.04884583130478859,
0.07257263362407684,
0.06367837637662888,
-0.017801770940423012,
0.00016894470900297165,
0.04386947304010391,
-0.002970502246171236,
-0.051965516060590744,
-0.15762734413146973,
0.06277678161859512,
0.17842786014080048,
0.056998081505298615,
0.030175408348441124,
0.012138530611991882,
-0.12049488723278046,
-0.07329574972391129,
0.10845038294792175,
-0.021686408668756485,
-0.031095284968614578,
-0.06442723423242569,
0.21323516964912415,
0.1388614922761917,
-0.19825653731822968,
0.0702671930193901,
-0.06280558556318283,
-0.04658647999167442,
-0.14314492046833038,
-0.17366671562194824,
-0.059809304773807526,
-0.0547034814953804,
-0.026051264256238937,
-0.054700352251529694,
0.04570859298110008,
0.047346316277980804,
-0.0016739139100536704,
-0.02772514894604683,
0.1126171201467514,
0.02765420638024807,
-0.032165806740522385,
0.04451003298163414,
0.05619681254029274,
0.03682970255613327,
-0.09137814491987228,
0.007322985213249922,
0.0029695341363549232,
0.014342821203172207,
0.06777288764715195,
0.01613135077059269,
-0.06992621719837189,
0.02725713886320591,
-0.020467489957809448,
-0.12120343744754791,
0.042514219880104065,
-0.005491400603204966,
-0.02191038616001606,
0.14766326546669006,
0.039597559720277786,
0.008086306042969227,
-0.014769108034670353,
0.22978916764259338,
-0.079631008207798,
-0.08263124525547028,
-0.1393512636423111,
0.07894771546125412,
-0.07535439729690552,
0.020168637856841087,
0.02652786672115326,
-0.12502749264240265,
0.017455779016017914,
0.17437158524990082,
0.11967697739601135,
-0.01862110011279583,
0.005760727450251579,
0.04387581720948219,
0.003006097162142396,
-0.04732988774776459,
0.01692454144358635,
0.05290905013680458,
0.19558346271514893,
-0.0746847614645958,
0.054245725274086,
-0.01774757355451584,
-0.08059251308441162,
-0.020728278905153275,
0.09288354963064194,
-0.009933017194271088,
-0.004748775623738766,
-0.06074956804513931,
0.149005725979805,
-0.0759778842329979,
-0.20890262722969055,
0.06107410788536072,
-0.057474348694086075,
-0.13986754417419434,
-0.043588198721408844,
0.03270360454916954,
-0.02818191610276699,
-0.0004342520551290363,
0.05878293514251709,
-0.041880737990140915,
0.1787300854921341,
0.02775873802602291,
-0.04535049945116043,
-0.08805633336305618,
0.060195520520210266,
-0.15322564542293549,
0.28409940004348755,
0.02300625666975975,
0.06475372612476349,
0.11462150514125824,
-0.023716775700449944,
-0.14765876531600952,
0.016111766919493675,
0.11251717060804367,
-0.07146475464105606,
0.06923303008079529,
0.16616879403591156,
0.00888645276427269,
0.12871026992797852,
0.06517354398965836,
-0.04169101640582085,
0.03372213616967201,
-0.08477409183979034,
-0.04430316761136055,
-0.1301726996898651,
0.07585147768259048,
-0.09351208806037903,
0.15738072991371155,
0.11715016514062881,
-0.07169844210147858,
0.010452828370034695,
-0.02282477170228958,
0.09099912643432617,
0.012017005123198032,
0.10486294329166412,
0.01101954746991396,
-0.19380232691764832,
0.04388235881924629,
0.012521770782768726,
0.09230010956525803,
-0.21009819209575653,
-0.05027567222714424,
0.04558335989713669,
-0.022896859794855118,
-0.06855283677577972,
0.11809497326612473,
0.03357189893722534,
0.028112467378377914,
-0.037041857838630676,
-0.032784342765808105,
0.007307000923901796,
0.151776984333992,
-0.11639050394296646,
-0.019398227334022522
] |
null | null | null |
# **Reinforce** Agent playing **Pixelcopter-PLE-v0**
This is a trained model of a **Reinforce** agent playing **Pixelcopter-PLE-v0** .
To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: https://huggingface.co/deep-rl-course/unit4/introduction
| {"tags": ["Pixelcopter-PLE-v0", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class"], "model-index": [{"name": "Pixelcopter-PLE-v0", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "Pixelcopter-PLE-v0", "type": "Pixelcopter-PLE-v0"}, "metrics": [{"type": "mean_reward", "value": "25.70 +/- 13.80", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | turgutburak01/Pixelcopter-PLE-v0 | [
"Pixelcopter-PLE-v0",
"reinforce",
"reinforcement-learning",
"custom-implementation",
"deep-rl-class",
"model-index",
"region:us"
] | 2024-02-08T06:38:29+00:00 | [] | [] | TAGS
#Pixelcopter-PLE-v0 #reinforce #reinforcement-learning #custom-implementation #deep-rl-class #model-index #region-us
|
# Reinforce Agent playing Pixelcopter-PLE-v0
This is a trained model of a Reinforce agent playing Pixelcopter-PLE-v0 .
To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL
| [
"# Reinforce Agent playing Pixelcopter-PLE-v0\n This is a trained model of a Reinforce agent playing Pixelcopter-PLE-v0 .\n To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL"
] | [
"TAGS\n#Pixelcopter-PLE-v0 #reinforce #reinforcement-learning #custom-implementation #deep-rl-class #model-index #region-us \n",
"# Reinforce Agent playing Pixelcopter-PLE-v0\n This is a trained model of a Reinforce agent playing Pixelcopter-PLE-v0 .\n To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL"
] | [
41,
58
] | [
"passage: TAGS\n#Pixelcopter-PLE-v0 #reinforce #reinforcement-learning #custom-implementation #deep-rl-class #model-index #region-us \n# Reinforce Agent playing Pixelcopter-PLE-v0\n This is a trained model of a Reinforce agent playing Pixelcopter-PLE-v0 .\n To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL"
] | [
0.0073175891302526,
-0.2259262204170227,
-0.0017347558168694377,
0.05054566636681557,
0.0658537745475769,
-0.055378563702106476,
0.1412602812051773,
0.05916554853320122,
-0.04990595206618309,
0.059261854737997055,
0.14166708290576935,
0.03996060788631439,
0.022112762555480003,
0.1513713151216507,
0.09764605015516281,
-0.2469022423028946,
0.07438477873802185,
0.01641594059765339,
0.008152224123477936,
0.09583204984664917,
0.060265738517045975,
-0.1405058205127716,
0.037032704800367355,
-0.01332044042646885,
-0.13650871813297272,
0.0010478810872882605,
-0.021802188828587532,
-0.03625129908323288,
0.15681709349155426,
0.006844013463705778,
0.09602472931146622,
-0.001560068572871387,
0.06475798785686493,
-0.12438877671957016,
0.05466329678893089,
0.06455880403518677,
-0.06293967366218567,
0.058029334992170334,
-0.057374246418476105,
0.11959903687238693,
0.04641333222389221,
-0.01578129455447197,
0.054811324924230576,
0.010941818356513977,
-0.14131468534469604,
-0.006710252724587917,
0.007013716734945774,
0.15098218619823456,
0.1339312642812729,
0.01409265398979187,
-0.0014771400019526482,
0.1363491266965866,
-0.16774429380893707,
0.045684073120355606,
0.061802688986063004,
-0.2633039951324463,
-0.04168876260519028,
0.12259352207183838,
0.08951573073863983,
0.06848238408565521,
-0.060910262167453766,
0.07636868953704834,
0.049813780933618546,
0.013985024765133858,
0.023094501346349716,
-0.042509064078330994,
-0.040479615330696106,
0.02289252169430256,
-0.0921095609664917,
-0.05999262258410454,
0.11517233401536942,
-0.006806366611272097,
0.03735918551683426,
-0.12476086616516113,
-0.015330453403294086,
-0.07314357161521912,
-0.05917041376233101,
-0.082573801279068,
0.07563583552837372,
0.030191516503691673,
-0.048283837735652924,
-0.08895846456289291,
-0.056533291935920715,
-0.11489585787057877,
-0.023082571104168892,
-0.07226225733757019,
0.005096882116049528,
-0.03157244250178337,
-0.035645097494125366,
0.09446526318788528,
-0.0021088174544274807,
-0.015028090216219425,
-0.03452150896191597,
-0.05930153280496597,
-0.04213470220565796,
-0.02359505370259285,
-0.03510070592164993,
-0.059062156826257706,
0.054655663669109344,
0.0680202916264534,
0.04938843473792076,
0.09133565425872803,
-0.0467856265604496,
0.1667373925447464,
-0.03256719931960106,
0.08078566938638687,
-0.011897698976099491,
0.2012830525636673,
0.11370102316141129,
0.12129533290863037,
0.06716908514499664,
-0.05294690653681755,
-0.16726544499397278,
0.039163749665021896,
0.12641896307468414,
0.07664673775434494,
-0.032492902129888535,
0.018162984400987625,
-0.12440363317728043,
0.05439428985118866,
-0.14826108515262604,
-0.06745084375143051,
0.024251462891697884,
0.01822635903954506,
-0.060682263225317,
0.03656952083110809,
-0.0028792342636734247,
0.003339326474815607,
0.004654870834201574,
-0.16432709991931915,
-0.05568019300699234,
0.028964387252926826,
-0.15712425112724304,
-0.06656725704669952,
0.06277995556592941,
-0.10113482922315598,
-0.012132617644965649,
-0.16982388496398926,
-0.16305199265480042,
-0.03628521412611008,
0.017857929691672325,
-0.040613796561956406,
-0.056917786598205566,
-0.14010562002658844,
-0.019415250048041344,
-0.045320261269807816,
-0.004312154371291399,
0.044072363525629044,
0.0020940210670232773,
0.04635847359895706,
0.0066573889926075935,
0.09289347380399704,
0.010714372619986534,
-0.0014722738415002823,
-0.04595406726002693,
0.0909833237528801,
-0.30731555819511414,
0.07525643706321716,
-0.08645553886890411,
0.05539081245660782,
-0.057316381484270096,
-0.0926317572593689,
-0.007509906310588121,
0.06277763843536377,
0.060464419424533844,
0.20788121223449707,
-0.2800109386444092,
-0.07025618106126785,
0.13655538856983185,
-0.09533236175775528,
-0.13146020472049713,
0.0513952374458313,
-0.050213608890771866,
0.07593657076358795,
0.027370907366275787,
0.140700101852417,
-0.028026295825839043,
-0.15554022789001465,
0.06281048059463501,
0.04586128890514374,
-0.11356306821107864,
0.019295670092105865,
0.03597676753997803,
0.06723599135875702,
0.05744141340255737,
-0.036986757069826126,
-0.04105675220489502,
0.08096802979707718,
-0.07076814025640488,
-0.037564266473054886,
0.04588831216096878,
-0.0579565204679966,
0.1630958467721939,
0.033971156924963,
0.09856503456830978,
-0.04149768501520157,
-0.07435470074415207,
-0.005698562134057283,
0.038746561855077744,
-0.08962973952293396,
0.025353478267788887,
-0.18320298194885254,
0.2423991560935974,
-0.02621818706393242,
0.027546977624297142,
-0.16845986247062683,
-0.0588528998196125,
0.011087946593761444,
0.21568740904331207,
0.030399197712540627,
0.12989304959774017,
0.07485637813806534,
-0.01250512059777975,
0.014156299643218517,
-0.06183977797627449,
-0.1972363442182541,
-0.03247830644249916,
0.008314179256558418,
-0.058311350643634796,
-0.04934588819742203,
-0.0900716632604599,
0.10427892208099365,
-0.19334633648395538,
-0.005319371819496155,
0.08282599598169327,
0.023504555225372314,
0.03946567326784134,
0.0035407328978180885,
-0.03634254261851311,
0.055148303508758545,
0.02030518464744091,
-0.08980578929185867,
0.14668866991996765,
0.0035520538222044706,
-0.03514726087450981,
-0.03927676007151604,
-0.03267495706677437,
0.05703731253743172,
0.08045367896556854,
-0.18214593827724457,
-0.0733821839094162,
-0.0838410034775734,
-0.02458474040031433,
0.050523869693279266,
0.036679428070783615,
0.02738112211227417,
0.44813573360443115,
0.057562243193387985,
0.09003535658121109,
-0.08811535686254501,
0.039806611835956573,
0.012785476632416248,
-0.031281858682632446,
0.013625281862914562,
0.04725322127342224,
0.11279468983411789,
0.028284218162298203,
0.01669839769601822,
0.03680038824677467,
0.01938779093325138,
0.08824212104082108,
-0.10939645022153854,
-0.003965397831052542,
0.002614045049995184,
0.038018375635147095,
0.03672022372484207,
0.07190682739019394,
0.015936892479658127,
-0.09583546966314316,
-0.030848123133182526,
-0.11166880279779434,
0.015594755299389362,
-0.20979784429073334,
-0.025905707851052284,
-0.029619399458169937,
0.0003502996696624905,
0.09109684824943542,
0.04222718998789787,
-0.04444896802306175,
0.035467714071273804,
0.03947039321064949,
-0.0861397460103035,
0.0594942644238472,
-0.014317752793431282,
-0.07008631527423859,
0.13023322820663452,
-0.1002996563911438,
-0.3153233230113983,
-0.08797995746135712,
0.05698639526963234,
0.05295826122164726,
0.06816939264535904,
-0.05876303091645241,
-0.09240786731243134,
0.03294730558991432,
-0.06836386770009995,
-0.0017794050509110093,
0.0037346978206187487,
-0.051060982048511505,
0.07253886014223099,
0.08541567623615265,
-0.014505518600344658,
-0.08911184966564178,
-0.006620637606829405,
-0.041561197489500046,
-0.124965138733387,
0.044060997664928436,
-0.03760828450322151,
0.00007921225915197283,
0.18620672821998596,
0.03724536672234535,
0.06256633251905441,
-0.06291008740663528,
0.07596296072006226,
-0.09150096774101257,
0.0004740063741337508,
0.18428465723991394,
-0.015377625823020935,
-0.004100616089999676,
-0.03996327146887779,
-0.0259257685393095,
-0.10829219967126846,
0.053985193371772766,
-0.07330703735351562,
-0.07349077612161636,
-0.0023273853585124016,
-0.07770214974880219,
-0.0351552739739418,
0.0012160884216427803,
0.07817990332841873,
0.029699061065912247,
-0.09635239094495773,
0.04920589178800583,
0.1298678070306778,
0.0931883230805397,
0.03626195341348648,
0.023981640115380287,
0.13739009201526642,
-0.11230582743883133,
0.019063033163547516,
-0.05148853361606598,
-0.1041760966181755,
-0.042787205427885056,
-0.0714287981390953,
0.07368279993534088,
0.06034531816840172,
-0.09970010071992874,
0.05144011229276657,
0.041872985661029816,
0.0883496031165123,
0.1373600959777832,
-0.04213863983750343,
-0.11244629323482513,
-0.041393622756004333,
-0.022004956379532814,
-0.1777329444885254,
0.0341336652636528,
0.22155584394931793,
0.0073304991237819195,
-0.10497386753559113,
0.07876885682344437,
-0.005956185050308704,
0.11527370661497116,
0.031222699210047722,
-0.278682678937912,
0.016931315883994102,
0.00203216471709311,
0.042359162122011185,
-0.047676295042037964,
0.10937416553497314,
0.11747439950704575,
-0.14421136677265167,
-0.06650938838720322,
-0.03273930773139,
0.044137366116046906,
-0.15618287026882172,
0.036923591047525406,
-0.12602220475673676,
0.06240779533982277,
0.050940994173288345,
0.05090156942605972,
-0.2197665423154831,
0.06881614029407501,
-0.0274215005338192,
0.06763827055692673,
-0.062248338013887405,
-0.01823522336781025,
0.04473711550235748,
0.025079863145947456,
0.14955177903175354,
-0.014347962103784084,
0.14454017579555511,
-0.09031219780445099,
-0.11753576993942261,
0.0027052261866629124,
0.08532248437404633,
0.013173088431358337,
0.013580933213233948,
0.0026939227245748043,
0.041669201105833054,
-0.02811569906771183,
0.17063532769680023,
-0.08147624880075455,
-0.022407781332731247,
-0.06592555344104767,
-0.018158966675400734,
0.2039334923028946,
-0.12064731866121292,
-0.10121093690395355,
-0.11619500070810318,
0.08663272857666016,
-0.04296411573886871,
0.08175522089004517,
-0.020344657823443413,
0.049704354256391525,
-0.02509051002562046,
0.007178863976150751,
0.09594997018575668,
0.01950966566801071,
0.08983828872442245,
-0.09791163355112076,
-0.019585272297263145,
0.13838915526866913,
-0.037155888974666595,
-0.036971647292375565,
-0.019425252452492714,
0.11054370552301407,
-0.0358734093606472,
0.08033111691474915,
0.03929615020751953,
0.03664831817150116,
0.03428546339273453,
-0.039165496826171875,
0.10309428721666336,
0.10041618347167969,
-0.06291446089744568,
0.03864621743559837,
-0.07954532653093338,
0.26597461104393005,
0.040773067623376846,
0.07301845401525497,
0.28390514850616455,
0.19391325116157532,
-0.03036464750766754,
0.10683353990316391,
-0.017607249319553375,
-0.024403288960456848,
-0.2950931787490845,
0.0006976581644266844,
0.027765681967139244,
0.11812873929738998,
0.01744898222386837,
-0.20587195456027985,
-0.1211688369512558,
-0.03560304269194603,
-0.007791717536747456,
0.0310499370098114,
-0.2441052496433258,
-0.06442268192768097,
0.06107868626713753,
0.13779635727405548,
0.15878525376319885,
-0.05917542055249214,
-0.007856467738747597,
0.029358724132180214,
0.07593556493520737,
0.017292039468884468,
-0.11598441749811172,
0.11550791561603546,
0.025637371465563774,
-0.05708931386470795,
0.0267958827316761,
-0.044003549963235855,
0.04214555397629738,
-0.17736166715621948,
0.10933554917573929,
-0.05924695357680321,
-0.08421005308628082,
0.07140472531318665,
-0.02217724733054638,
-0.048552993685007095,
0.0789642184972763,
0.020652711391448975,
-0.13173207640647888,
0.038154006004333496,
0.005618774797767401,
0.04346654564142227,
-0.004941361024975777,
-0.019811764359474182,
-0.029163256287574768,
0.07706235349178314,
-0.03806605935096741,
0.09605937451124191,
0.19590972363948822,
-0.0573095865547657,
0.03974950686097145,
0.085201695561409,
0.09593135863542557,
-0.05523005872964859,
-0.0809539332985878,
-0.03812742978334427,
-0.005277194548398256,
0.0674438327550888,
-0.08598461747169495,
-0.019085103645920753,
0.07938229292631149,
0.015313901007175446,
0.14910826086997986,
0.14389736950397491,
-0.08835655450820923,
0.11321785300970078,
0.10694554448127747,
-0.11366690695285797,
-0.08583837002515793,
-0.02963297814130783,
0.0009990704711526632,
0.04910186678171158,
-0.048617590218782425,
0.05932905897498131,
-0.1035301461815834,
0.012819357216358185,
0.03532040864229202,
0.0038119733799248934,
-0.09975302964448929,
0.009764863178133965,
0.08645275235176086,
0.06119582802057266,
-0.0567571222782135,
0.09250631928443909,
-0.0019178141374140978,
-0.10868195444345474,
0.07241881638765335,
0.009918469935655594,
-0.021528873592615128,
-0.06352251768112183,
0.03211374953389168,
0.2370220273733139,
0.13945111632347107,
-0.04336636886000633,
-0.12396618723869324,
-0.15508891642093658,
0.037849195301532745,
0.024356422945857048,
0.051251959055662155,
0.0062240250408649445,
-0.06906022876501083,
0.01234503649175167,
-0.04392383247613907,
0.005266309250146151,
-0.05930564925074577,
-0.047703344374895096,
-0.12081446498632431,
0.1154373437166214,
0.053290288895368576,
0.11705748736858368,
-0.0842847004532814,
-0.07057584822177887,
-0.1921386867761612,
0.09190598875284195,
0.041707299649715424,
-0.05532265454530716,
0.06002674251794815,
-0.030134430155158043,
0.017344338819384575,
0.11256659775972366,
-0.051967836916446686,
0.008543911390006542,
-0.09269233793020248,
0.03236149623990059,
0.03133073076605797,
0.04903566092252731,
-0.004612727556377649,
-0.017903391271829605,
0.04399999976158142,
-0.05730267986655235,
0.07619527727365494,
-0.07757602632045746,
-0.033709146082401276,
0.0645759105682373,
-0.16051416099071503,
-0.054324716329574585,
0.08708633482456207,
0.013749903067946434,
0.02590017393231392,
-0.05825240537524223,
0.019142305478453636,
-0.05566488951444626,
-0.04483235627412796,
0.01169554702937603,
-0.05552767962217331,
-0.011517677456140518,
0.05293213203549385,
-0.05287189036607742,
-0.040493328124284744,
-0.06794002652168274,
0.061874233186244965,
-0.07247710227966309,
0.09816460311412811,
0.031187955290079117,
-0.10892423242330551,
0.07648903876543045,
-0.037552736699581146,
-0.0049397205002605915,
-0.009439278393983841,
0.039307788014411926,
0.15598824620246887,
-0.1606634259223938,
0.05345672369003296,
-0.0484454482793808,
0.13272921741008759,
0.046888746321201324,
-0.04458791762590408,
-0.020207170397043228,
0.02469455823302269,
-0.05549024045467377,
0.06932897865772247,
0.15877580642700195,
0.09880131483078003,
0.02571805939078331,
0.008134597912430763,
0.10187267512083054,
0.1060529574751854,
0.08136752992868423,
0.08394161611795425,
-0.03428563475608826,
-0.11287897825241089,
0.14338994026184082,
0.09748584777116776,
0.024613093584775925,
0.21077860891819,
0.17944025993347168,
0.03125298395752907,
0.03018142655491829,
-0.06512103229761124,
0.17325744032859802,
0.061261482536792755,
-0.08229418843984604,
0.014424329623579979,
0.03221147879958153,
-0.049809664487838745,
-0.047004032880067825,
-0.09757380187511444,
-0.029556652531027794,
-0.24085633456707,
0.10851483792066574,
-0.057250600308179855,
-0.09750643372535706,
0.022772664204239845,
0.02990041859447956,
-0.018839845433831215,
0.11280566453933716,
-0.07735858112573624,
0.012980576604604721,
0.18577688932418823,
-0.03825045004487038,
-0.022322099655866623,
-0.1633504331111908,
-0.11154003441333771,
-0.014046176336705685,
-0.11750495433807373,
0.025494296103715897,
0.06305963546037674,
0.01117965579032898,
0.04399528726935387,
0.028923438861966133,
-0.020834028720855713,
0.019218796864151955,
-0.05903913825750351,
-0.042673509567976,
-0.01891910657286644,
0.02202831581234932,
-0.09593231230974197,
-0.03627033904194832,
0.12151803076267242,
-0.03246605768799782,
-0.08207374066114426,
-0.006544890813529491,
0.07848484069108963,
-0.042620159685611725,
0.09450104832649231,
-0.07687012106180191,
-0.03479038178920746,
-0.06794454902410507,
0.268902063369751,
0.09388194978237152,
-0.20183001458644867,
0.03341769427061081,
-0.030470456928014755,
0.026735708117485046,
-0.09215684235095978,
0.16250114142894745,
0.0899243950843811,
0.049168527126312256,
-0.12686687707901,
-0.003401300171390176,
-0.09992645680904388,
-0.0028723697178065777,
-0.12552696466445923,
-0.14725084602832794,
0.12093491852283478,
-0.003848524997010827,
-0.06547791510820389,
0.02844911813735962,
-0.15909899771213531,
0.06585367769002914,
0.0978507474064827,
-0.1514272391796112,
-0.038227714598178864,
-0.06086801365017891,
0.06072385236620903,
0.026465637609362602,
0.13005392253398895,
-0.05080926790833473,
0.012067130766808987,
-0.0656723901629448,
-0.011309894733130932,
-0.0000654291216051206,
-0.017478201538324356,
0.001532604917883873,
-0.09828947484493256,
0.05038110539317131,
-0.0835796371102333,
0.12184429168701172,
0.05709611251950264,
0.005326167680323124,
0.008464806713163853,
0.0648408755660057,
-0.02414623089134693,
-0.10202058404684067,
-0.01877439208328724,
0.033475372940301895,
0.03998998552560806,
0.010373802855610847,
0.034506846219301224,
0.0006507808575406671,
0.07714920490980148,
-0.011413984932005405,
-0.027285432443022728,
-0.058209117501974106,
0.03936338797211647,
-0.10441672056913376,
0.10461361706256866,
0.0013552121818065643,
-0.02240127883851528,
-0.010913821868598461,
-0.05532446503639221,
0.045815300196409225,
0.04572062939405441,
0.029743505641818047,
-0.05261747166514397,
-0.09262793511152267,
-0.021781492978334427,
0.023900283500552177,
-0.11539579927921295,
-0.18497975170612335,
-0.0664035826921463,
-0.15038692951202393,
-0.01633414439857006,
-0.0620744526386261,
0.08902198076248169,
0.13558129966259003,
0.030392181128263474,
-0.04822919890284538,
-0.12171997129917145,
0.025026977062225342,
0.13544774055480957,
-0.03851630911231041,
-0.07532322406768799
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# smolm-autoreg-bpe-counterfactual-babylm-random_removal-seed_211-3e-4
This model was trained from scratch on the kanishka/counterfactual-babylm-random_removal dataset.
It achieves the following results on the evaluation set:
- Loss: 3.4079
- Accuracy: 0.4097
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0003
- train_batch_size: 32
- eval_batch_size: 64
- seed: 211
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 32000
- num_epochs: 20.0
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:------:|:---------------:|:--------:|
| 3.7403 | 1.0 | 18586 | 3.8964 | 0.3473 |
| 3.4414 | 2.0 | 37172 | 3.6106 | 0.3762 |
| 3.2952 | 3.0 | 55758 | 3.5047 | 0.3879 |
| 3.2087 | 4.0 | 74344 | 3.4663 | 0.3950 |
| 3.149 | 5.0 | 92930 | 3.4383 | 0.3987 |
| 3.101 | 6.0 | 111516 | 3.3864 | 0.4021 |
| 3.0614 | 7.0 | 130102 | 3.3728 | 0.4053 |
| 3.0311 | 8.0 | 148688 | 3.3712 | 0.4057 |
| 2.9996 | 9.0 | 167274 | 3.3636 | 0.4071 |
| 2.9768 | 10.0 | 185860 | 3.3474 | 0.4088 |
| 2.9515 | 11.0 | 204446 | 3.3726 | 0.4089 |
| 2.9309 | 12.0 | 223032 | 3.3788 | 0.4076 |
| 2.9078 | 13.0 | 241618 | 3.3546 | 0.4109 |
| 2.8874 | 14.0 | 260204 | 3.3762 | 0.4093 |
| 2.8664 | 15.0 | 278790 | 3.3832 | 0.4096 |
| 2.8486 | 16.0 | 297376 | 3.3725 | 0.4112 |
| 2.827 | 17.0 | 315962 | 3.3913 | 0.4099 |
| 2.8082 | 18.0 | 334548 | 3.3945 | 0.4100 |
| 2.7888 | 19.0 | 353134 | 3.4077 | 0.4095 |
| 2.7762 | 20.0 | 371720 | 3.4079 | 0.4097 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"tags": ["generated_from_trainer"], "datasets": ["kanishka/counterfactual-babylm-random_removal"], "metrics": ["accuracy"], "model-index": [{"name": "smolm-autoreg-bpe-counterfactual-babylm-random_removal-seed_211-3e-4", "results": [{"task": {"type": "text-generation", "name": "Causal Language Modeling"}, "dataset": {"name": "kanishka/counterfactual-babylm-random_removal", "type": "kanishka/counterfactual-babylm-random_removal"}, "metrics": [{"type": "accuracy", "value": 0.40973803931895864, "name": "Accuracy"}]}]}]} | text-generation | kanishka/smolm-autoreg-bpe-counterfactual-babylm-random_removal-seed_211-3e-4 | [
"transformers",
"tensorboard",
"safetensors",
"opt",
"text-generation",
"generated_from_trainer",
"dataset:kanishka/counterfactual-babylm-random_removal",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T06:38:30+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #opt #text-generation #generated_from_trainer #dataset-kanishka/counterfactual-babylm-random_removal #model-index #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| smolm-autoreg-bpe-counterfactual-babylm-random\_removal-seed\_211-3e-4
======================================================================
This model was trained from scratch on the kanishka/counterfactual-babylm-random\_removal dataset.
It achieves the following results on the evaluation set:
* Loss: 3.4079
* Accuracy: 0.4097
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0003
* train\_batch\_size: 32
* eval\_batch\_size: 64
* seed: 211
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 32000
* num\_epochs: 20.0
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 64\n* seed: 211\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 32000\n* num\\_epochs: 20.0\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #opt #text-generation #generated_from_trainer #dataset-kanishka/counterfactual-babylm-random_removal #model-index #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 64\n* seed: 211\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 32000\n* num\\_epochs: 20.0\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
82,
132,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #opt #text-generation #generated_from_trainer #dataset-kanishka/counterfactual-babylm-random_removal #model-index #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 64\n* seed: 211\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 32000\n* num\\_epochs: 20.0\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.13828331232070923,
0.12733496725559235,
-0.0014970721676945686,
0.06526648253202438,
0.12099514156579971,
0.011192343197762966,
0.15518604218959808,
0.13008545339107513,
-0.0976354107260704,
0.07581731677055359,
0.1422099471092224,
0.06627632677555084,
0.049948450177907944,
0.1410091519355774,
-0.06970104575157166,
-0.2773881256580353,
0.02406955696642399,
0.04192548990249634,
-0.10696268081665039,
0.12756912410259247,
0.09499906748533249,
-0.11121256649494171,
0.07544072717428207,
0.05080099403858185,
-0.15645670890808105,
-0.014275025576353073,
0.00547034852206707,
-0.07195966690778732,
0.09947826713323593,
0.04010133445262909,
0.12649530172348022,
0.03760673105716705,
0.06629407405853271,
-0.16314256191253662,
0.01495745312422514,
0.05529529228806496,
0.021927395835518837,
0.10441084206104279,
0.06995617598295212,
-0.0239107683300972,
0.12134572118520737,
-0.05540251359343529,
0.06445173919200897,
0.050621479749679565,
-0.11698070913553238,
-0.23295621573925018,
-0.08315745741128922,
0.0805797129869461,
0.08948302268981934,
0.07393907755613327,
-0.01854333095252514,
0.14988568425178528,
-0.022373778745532036,
0.0933433249592781,
0.2289993315935135,
-0.24891667068004608,
-0.09886886924505234,
-0.01844041980803013,
0.05972811207175255,
0.037418901920318604,
-0.11077069491147995,
-0.011517844162881374,
0.07251065969467163,
0.020313221961259842,
0.0987354964017868,
0.005354058928787708,
0.04168827086687088,
-0.02297770231962204,
-0.1410687118768692,
-0.062307532876729965,
0.16026142239570618,
0.07493442296981812,
-0.06154869124293327,
-0.0806681364774704,
-0.052804574370384216,
-0.19080311059951782,
-0.04624776542186737,
0.010787829756736755,
0.010056284256279469,
-0.06205747649073601,
-0.12575341761112213,
-0.030106142163276672,
-0.10112538188695908,
-0.10203337669372559,
0.006581378635019064,
0.21217040717601776,
0.04780062660574913,
-0.004533547442406416,
-0.035073865205049515,
0.12593470513820648,
0.05358486622571945,
-0.1623779535293579,
-0.02705756016075611,
0.020374860614538193,
-0.033944081515073776,
-0.038329459726810455,
-0.05705836042761803,
-0.05867364630103111,
0.0026014987379312515,
0.16453325748443604,
-0.0807693675160408,
0.04811650887131691,
0.003030303865671158,
0.026300916448235512,
-0.08759841322898865,
0.16737785935401917,
-0.04226807504892349,
0.007514753378927708,
0.00005980748755973764,
0.1325409710407257,
0.020508309826254845,
-0.008326912298798561,
-0.06127656251192093,
0.019005510956048965,
0.1257636845111847,
0.041471805423498154,
-0.03004913590848446,
0.044380415230989456,
-0.051141947507858276,
-0.03142041340470314,
0.07790873199701309,
-0.07588892430067062,
-0.0010631268378347158,
0.01401794608682394,
-0.08922354876995087,
-0.03907289355993271,
0.018380600959062576,
0.020989766344428062,
0.011216492392122746,
0.08965282887220383,
-0.11381793022155762,
-0.019360756501555443,
-0.09023331850767136,
-0.0936964824795723,
0.007379692979156971,
-0.03119516745209694,
0.005207822658121586,
-0.1027563139796257,
-0.173553928732872,
-0.03076556883752346,
0.024759162217378616,
-0.026754610240459442,
-0.05887879058718681,
-0.04958483576774597,
-0.09443068504333496,
0.039431508630514145,
-0.010965187102556229,
0.11909367889165878,
-0.04692449793219566,
0.11527878791093826,
0.07127980887889862,
0.04357341676950455,
0.013443538919091225,
0.029569966718554497,
-0.08616137504577637,
0.06724642217159271,
-0.11951404809951782,
0.06197826564311981,
-0.06284110248088837,
0.05647983402013779,
-0.10102048516273499,
-0.12137114256620407,
-0.009512930177152157,
-0.0023014647886157036,
0.08995790779590607,
0.12057054787874222,
-0.10931567847728729,
-0.0938439890742302,
0.16501866281032562,
-0.10290740430355072,
-0.1718919426202774,
0.1100853830575943,
-0.021273987367749214,
0.03501734510064125,
0.05314917862415314,
0.1356443464756012,
0.08931387960910797,
-0.07246388494968414,
-0.03916546702384949,
-0.06524760276079178,
0.09959613531827927,
-0.011672898195683956,
0.10158680379390717,
0.008649186231195927,
-0.0310148186981678,
0.00912504643201828,
-0.06229262799024582,
0.05355421081185341,
-0.11226401478052139,
-0.08987797796726227,
-0.03164701163768768,
-0.11536629498004913,
0.050693344324827194,
0.05576947331428528,
0.07623574882745743,
-0.09592318534851074,
-0.11242520809173584,
0.06809157133102417,
0.12804627418518066,
-0.07617606967687607,
0.013379520736634731,
-0.08731893450021744,
0.06822957843542099,
-0.0715845450758934,
-0.025548217818140984,
-0.16803592443466187,
-0.08991148322820663,
0.026668068021535873,
-0.03616034612059593,
-0.011565865948796272,
-0.07326751202344894,
0.09065455198287964,
0.08395706862211227,
-0.06202009692788124,
-0.08516324311494827,
-0.06759846955537796,
-0.016025710850954056,
-0.0938856303691864,
-0.17774598300457,
-0.07862400263547897,
-0.02309064194560051,
0.1826385110616684,
-0.23216570913791656,
0.050230614840984344,
-0.0018077751155942678,
0.14860950410366058,
0.05772308632731438,
-0.050287749618291855,
-0.002250443212687969,
0.04973233863711357,
-0.04401249438524246,
-0.07100662589073181,
0.0384768545627594,
0.019595270976424217,
-0.1219211295247078,
0.0027899681590497494,
-0.13383224606513977,
0.13716860115528107,
0.1033513993024826,
0.016874661669135094,
-0.0951170027256012,
-0.015551666729152203,
-0.08844681084156036,
-0.042832937091588974,
-0.026831310242414474,
-0.02419406920671463,
0.10133962333202362,
0.036851897835731506,
0.1543198823928833,
-0.09702233225107193,
-0.057587359100580215,
0.03492416813969612,
-0.017261521890759468,
-0.020148906856775284,
0.11329679191112518,
0.03246832638978958,
-0.11001550406217575,
0.11533834785223007,
0.07643134891986847,
-0.08628154546022415,
0.1703885942697525,
-0.0639893114566803,
-0.09373442828655243,
-0.02997000142931938,
0.02447156421840191,
0.050322018563747406,
0.11823777854442596,
-0.08423028141260147,
0.007930605672299862,
0.02721938118338585,
-0.0003317372174933553,
0.024207331240177155,
-0.20124293863773346,
-0.017199581488966942,
0.039461273699998856,
-0.03228778392076492,
-0.028956664726138115,
0.011799895204603672,
-0.002005064394325018,
0.09587559103965759,
-0.00000845272825245047,
-0.02245396003127098,
0.016526104882359505,
0.0013633720809593797,
-0.08107146620750427,
0.21997568011283875,
-0.055775973945856094,
-0.14001241326332092,
-0.17351411283016205,
0.004243055824190378,
-0.07807964831590652,
0.003911362029612064,
0.03325200080871582,
-0.07851459830999374,
-0.024060193449258804,
-0.08890394866466522,
0.011644678190350533,
-0.013315810821950436,
0.031749606132507324,
0.015994485467672348,
0.0005040064570493996,
0.09071327745914459,
-0.1144135594367981,
0.01629844307899475,
-0.01600676402449608,
-0.030244167894124985,
0.03622180595993996,
0.015050729736685753,
0.10322212427854538,
0.12935630977153778,
0.0053581357933580875,
0.018958738073706627,
-0.03214399889111519,
0.2079959213733673,
-0.0896613672375679,
-0.0412311777472496,
0.13261264562606812,
-0.006937046535313129,
0.05020540952682495,
0.08223790675401688,
0.04941047355532646,
-0.09137707948684692,
0.040499184280633926,
0.053428925573825836,
-0.02477254718542099,
-0.22966697812080383,
-0.014466064982116222,
-0.037154391407966614,
0.00019533972954377532,
0.1193663701415062,
0.03162187337875366,
0.0495726615190506,
0.08334945142269135,
-0.04737550765275955,
0.0349954254925251,
-0.037370145320892334,
0.10031713545322418,
0.06755580008029938,
0.04910427704453468,
0.13277755677700043,
-0.02998168393969536,
-0.05541175603866577,
0.027011264115571976,
-0.032495588064193726,
0.2195587158203125,
-0.02995661087334156,
0.18067310750484467,
0.047168243676424026,
0.14007185399532318,
0.016050662845373154,
0.08389624953269958,
0.01799076795578003,
-0.03706015273928642,
0.02989470772445202,
-0.057320863008499146,
-0.03691871464252472,
0.04362224414944649,
-0.01245405524969101,
0.09492485970258713,
-0.13988076150417328,
0.00005200646774028428,
0.01642216555774212,
0.29240259528160095,
0.06280721724033356,
-0.37348198890686035,
-0.1359163373708725,
0.011358136311173439,
-0.05180569365620613,
-0.07393897324800491,
0.009395508095622063,
0.11055318266153336,
-0.10410688817501068,
0.05673694610595703,
-0.10562240332365036,
0.09824293106794357,
-0.04159197956323624,
-0.005129422061145306,
0.044979266822338104,
0.08447443693876266,
-0.03736751154065132,
0.05530242249369621,
-0.22727438807487488,
0.2781114876270294,
0.0016691609052941203,
0.09007322788238525,
-0.04137858748435974,
0.01098642684519291,
0.04906526952981949,
0.025265783071517944,
0.05333181470632553,
-0.018330717459321022,
-0.065323606133461,
-0.20490263402462006,
-0.08558209985494614,
0.0240324754267931,
0.10550113022327423,
-0.07359679043292999,
0.14371605217456818,
-0.02712504379451275,
-0.0017186065670102835,
0.0577511265873909,
-0.06835372000932693,
-0.12108632922172546,
-0.07360174506902695,
0.02627510204911232,
0.03613661229610443,
0.06288732588291168,
-0.11794491112232208,
-0.12966492772102356,
-0.02378193847835064,
0.1431090384721756,
-0.051167819648981094,
-0.06307989358901978,
-0.1362427920103073,
0.08339176326990128,
0.14517861604690552,
-0.08166822791099548,
0.030780108645558357,
-0.0036882886197417974,
0.1697758138179779,
0.014892793260514736,
-0.02676372230052948,
0.07197181135416031,
-0.07218421995639801,
-0.2427264153957367,
-0.03560938686132431,
0.15567384660243988,
0.04211609065532684,
0.04875055328011513,
-0.018923001363873482,
0.02346823178231716,
-0.026260625571012497,
-0.07526883482933044,
0.04765823110938072,
-0.01971689611673355,
0.03596445545554161,
0.023695316165685654,
-0.008729018270969391,
0.040972184389829636,
-0.06597869098186493,
-0.027986157685518265,
0.1286509782075882,
0.32001441717147827,
-0.05236200988292694,
-0.029646478593349457,
0.03512437641620636,
-0.034410443156957626,
-0.12992815673351288,
0.04185003414750099,
0.12845510244369507,
0.028854666277766228,
0.005055509507656097,
-0.1831628978252411,
0.049474772065877914,
0.08336116373538971,
-0.040864795446395874,
0.13138003647327423,
-0.26553115248680115,
-0.14507004618644714,
0.10374844819307327,
0.14797475934028625,
0.046282313764095306,
-0.1592034548521042,
-0.06990108639001846,
-0.018568433821201324,
-0.12237096577882767,
0.13719117641448975,
-0.029375357553362846,
0.11517408490180969,
-0.023753609508275986,
0.07870850712060928,
0.024271495640277863,
-0.05761745199561119,
0.15883305668830872,
-0.0328432060778141,
0.07098331302404404,
-0.01089303195476532,
0.0018197926692664623,
0.09145504236221313,
-0.07879146933555603,
0.01775171235203743,
-0.0766262412071228,
0.04823476821184158,
-0.1294201761484146,
-0.037572380155324936,
-0.09172788262367249,
0.05927581712603569,
-0.04762841761112213,
-0.04027358070015907,
-0.02415415458381176,
0.04178258776664734,
0.028295211493968964,
-0.0036697990726679564,
0.17932674288749695,
-0.00023094408970791847,
0.17957769334316254,
0.11949270963668823,
0.0878850668668747,
-0.014815153554081917,
-0.03229489549994469,
-0.02136785164475441,
-0.02156822569668293,
0.07357605546712875,
-0.13355262577533722,
0.019337253645062447,
0.12558738887310028,
0.05476941540837288,
0.13712312281131744,
0.06160370260477066,
-0.06933058798313141,
0.033891063183546066,
0.06968051195144653,
-0.11780773103237152,
-0.0936528667807579,
-0.056512050330638885,
0.03149329498410225,
-0.18088950216770172,
0.04647020995616913,
0.12295274436473846,
-0.07793094217777252,
-0.011861016042530537,
-0.004524874035269022,
-0.01117013767361641,
-0.019750181585550308,
0.20733381807804108,
0.06586162000894547,
0.08379475772380829,
-0.08646851032972336,
0.06558186560869217,
0.021812256425619125,
-0.11030878126621246,
0.023865798488259315,
0.04936007410287857,
-0.05749187991023064,
-0.014781555160880089,
0.0005785655812360346,
0.12587060034275055,
-0.07929754257202148,
-0.05200551077723503,
-0.17177565395832062,
-0.10780590027570724,
0.06900189071893692,
0.1475873440504074,
0.05470966547727585,
0.03636850044131279,
-0.016868842765688896,
0.03578852862119675,
-0.13349977135658264,
0.12241671234369278,
0.08476465940475464,
0.10490836203098297,
-0.15298902988433838,
0.15946781635284424,
-0.006822614464908838,
0.015584348700940609,
-0.010995866730809212,
0.008181510493159294,
-0.09653130918741226,
0.0024409135803580284,
-0.14236994087696075,
-0.017069997265934944,
-0.054225094616413116,
-0.009034126996994019,
-0.0032494112383574247,
-0.06163503974676132,
-0.08861000090837479,
0.018178947269916534,
-0.1117064356803894,
-0.04948549345135689,
0.03292545676231384,
0.04586542770266533,
-0.12091659009456635,
-0.008320033550262451,
0.048518575727939606,
-0.1064734011888504,
0.06917663663625717,
0.05293084681034088,
0.033411409705877304,
0.0426546074450016,
-0.04307950660586357,
0.03539267182350159,
0.024083569645881653,
-0.012537005357444286,
0.028315329924225807,
-0.11590680480003357,
0.012205487117171288,
-0.02423928491771221,
0.03209680691361427,
0.012699571438133717,
0.04706913232803345,
-0.13172054290771484,
-0.03576769307255745,
0.0015901199076324701,
-0.013837412931025028,
-0.06591834872961044,
0.03177269548177719,
0.046649087220430374,
0.02769836224615574,
0.18700291216373444,
-0.0676126778125763,
0.01887224242091179,
-0.23168158531188965,
0.007370565086603165,
-0.014523589052259922,
-0.11479087173938751,
-0.0660618245601654,
-0.021793484687805176,
0.07237889617681503,
-0.06157481297850609,
0.07665116339921951,
-0.06324446946382523,
0.052296318113803864,
0.03698478266596794,
-0.03479842096567154,
0.058011867105960846,
0.03389498591423035,
0.2504591941833496,
0.04987931251525879,
-0.024393288418650627,
0.06556544452905655,
0.014830762520432472,
0.07125740498304367,
0.06580294668674469,
0.18105866014957428,
0.11377294361591339,
-0.04996419697999954,
0.10516413301229477,
0.03947782889008522,
-0.04283728823065758,
-0.1377168744802475,
0.023714745417237282,
-0.024054143577814102,
0.09488610923290253,
0.011313480325043201,
0.17144161462783813,
0.1413525640964508,
-0.16662956774234772,
0.0137409009039402,
-0.02195855788886547,
-0.0686124861240387,
-0.0918222963809967,
-0.06274725496768951,
-0.08528800308704376,
-0.14610549807548523,
0.023757541552186012,
-0.13037467002868652,
0.006260162219405174,
0.0725981816649437,
0.015339981764554977,
-0.0007646625745110214,
0.18092384934425354,
0.0453953891992569,
0.014380350708961487,
0.09387350082397461,
0.004460064694285393,
-0.022139806300401688,
-0.038739435374736786,
-0.12033244222402573,
0.05043955147266388,
-0.025936627760529518,
0.043228283524513245,
-0.06752069294452667,
-0.0783785879611969,
0.06315438449382782,
0.01932225562632084,
-0.12817993760108948,
0.017673656344413757,
0.020120395347476006,
0.07438014447689056,
0.017809845507144928,
0.013508439995348454,
0.027714287862181664,
-0.022184375673532486,
0.24119894206523895,
-0.08006514608860016,
-0.011389423161745071,
-0.14514118432998657,
0.2353755235671997,
0.005710735451430082,
-0.04053034633398056,
0.04396858066320419,
-0.08909210562705994,
-0.0007388952071778476,
0.17694799602031708,
0.10767456889152527,
-0.02020389214158058,
-0.02032880298793316,
-0.0005087242461740971,
-0.01577155478298664,
-0.04971316084265709,
0.1001807451248169,
0.06985291093587875,
0.024754399433732033,
-0.06673777103424072,
-0.05130680277943611,
-0.049736108630895615,
-0.04134434089064598,
0.006432961206883192,
0.08622868359088898,
0.027725830674171448,
-0.012112954631447792,
-0.044430866837501526,
0.08335269242525101,
-0.06763676553964615,
-0.09626930207014084,
0.038324207067489624,
-0.1673223078250885,
-0.1799706220626831,
-0.039033349603414536,
0.048460427671670914,
0.00873936340212822,
0.06158570572733879,
-0.012711843475699425,
-0.011361280456185341,
0.08590896427631378,
-0.018981939181685448,
-0.042251892387866974,
-0.11751073598861694,
0.08006078004837036,
-0.07256346195936203,
0.22471661865711212,
-0.04149404540657997,
0.022659320384263992,
0.1273125261068344,
0.015903832390904427,
-0.10858671367168427,
0.04455040022730827,
0.07077954709529877,
-0.1065700426697731,
0.045792270451784134,
0.20109957456588745,
-0.025017939507961273,
0.11087685078382492,
0.03672732785344124,
-0.15127968788146973,
-0.004227418452501297,
-0.07031894475221634,
-0.04063752666115761,
-0.07814078778028488,
-0.014586208388209343,
-0.040016356855630875,
0.12675011157989502,
0.2203827053308487,
-0.07104555517435074,
-0.01710587926208973,
-0.05049217492341995,
0.04529941827058792,
0.09317369014024734,
0.07784216105937958,
-0.021453870460391045,
-0.30284395813941956,
0.003840648103505373,
0.048030491918325424,
-0.018395770341157913,
-0.3265220522880554,
-0.07242552936077118,
-0.0009786451701074839,
-0.06776434183120728,
-0.041912563145160675,
0.09452223032712936,
0.07570795714855194,
0.0433032289147377,
-0.04218147695064545,
-0.03344607725739479,
-0.07289204746484756,
0.1757867932319641,
-0.15828688442707062,
-0.0807192474603653
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | anish005/mistral-reddit | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"4-bit",
"region:us"
] | 2024-02-08T06:40:22+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #mistral #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
59,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.04788382723927498,
0.16171279549598694,
-0.005352917592972517,
0.02136841043829918,
0.09686184674501419,
0.015111604705452919,
0.07137951999902725,
0.10955451428890228,
-0.020038405433297157,
0.11244286596775055,
0.03330032527446747,
0.09441220015287399,
0.11357662081718445,
0.14772894978523254,
-0.003575492650270462,
-0.232261523604393,
0.05006932094693184,
-0.1246371790766716,
-0.03666049614548683,
0.11627218872308731,
0.15057805180549622,
-0.10109459608793259,
0.0756460502743721,
-0.030913641676306725,
-0.009812407195568085,
-0.033390406519174576,
-0.05693698301911354,
-0.04944330081343651,
0.05103539675474167,
0.07355327904224396,
0.06860782206058502,
0.004711335990577936,
0.09396199136972427,
-0.2655787467956543,
0.020426444709300995,
0.07093948125839233,
-0.0019974696915596724,
0.07591848820447922,
0.05331665277481079,
-0.07516877353191376,
0.09268398582935333,
-0.050851862877607346,
0.14750060439109802,
0.07999978214502335,
-0.09178632497787476,
-0.1916678249835968,
-0.08780492842197418,
0.1011761948466301,
0.18467943370342255,
0.04421152547001839,
-0.023150887340307236,
0.10070723295211792,
-0.08664919435977936,
0.011689732782542706,
0.05446745082736015,
-0.06747105717658997,
-0.052418507635593414,
0.06491605937480927,
0.0793420672416687,
0.0767722949385643,
-0.12429667264223099,
-0.02174406498670578,
0.008611103519797325,
0.00887030828744173,
0.0814518854022026,
0.02427673526108265,
0.15523891150951385,
0.04025983437895775,
-0.12765221297740936,
-0.04938573017716408,
0.1069699227809906,
0.04103906825184822,
-0.04726257547736168,
-0.25091874599456787,
-0.02940361201763153,
-0.025307154282927513,
-0.0306085217744112,
-0.03984487056732178,
0.04118539020419121,
-0.006947068031877279,
0.08044755458831787,
-0.00699204858392477,
-0.07604070007801056,
-0.03760726749897003,
0.06074252352118492,
0.061373963952064514,
0.026066971942782402,
-0.012132911942899227,
0.010845172218978405,
0.11657512187957764,
0.10491005331277847,
-0.12471938878297806,
-0.05219554528594017,
-0.06468912214040756,
-0.07947935909032822,
-0.043697141110897064,
0.03412042558193207,
0.041996799409389496,
0.0503966324031353,
0.24876368045806885,
0.013237204402685165,
0.05510834604501724,
0.03997663035988808,
0.009734313935041428,
0.06435027718544006,
0.11203338205814362,
-0.06008143350481987,
-0.09627178311347961,
-0.027062542736530304,
0.09033986181020737,
0.010054918937385082,
-0.04071144387125969,
-0.05739009380340576,
0.0623612105846405,
0.0185408778488636,
0.11882077902555466,
0.08993595838546753,
0.0032226061448454857,
-0.07099147886037827,
-0.06380297243595123,
0.1967260092496872,
-0.16188356280326843,
0.047058627009391785,
0.0354633666574955,
-0.038969624787569046,
-0.0023125922307372093,
0.007358122151345015,
0.025436079129576683,
-0.01989656314253807,
0.09119832515716553,
-0.05659566447138786,
-0.040517594665288925,
-0.10913336277008057,
-0.03569170460104942,
0.032513219863176346,
0.010858619585633278,
-0.03217906504869461,
-0.030786113813519478,
-0.08418484032154083,
-0.06755144894123077,
0.09449765086174011,
-0.07364831864833832,
-0.053648900240659714,
-0.01785014383494854,
-0.07429704070091248,
0.025341181084513664,
0.02034521847963333,
0.07678651064634323,
-0.019926751032471657,
0.04245077818632126,
-0.05644877254962921,
0.059700917452573776,
0.10742949694395065,
0.033078111708164215,
-0.05497613549232483,
0.06192035973072052,
-0.24182282388210297,
0.10019542276859283,
-0.06878995895385742,
0.05493709444999695,
-0.1513376086950302,
-0.0262905303388834,
0.049252137541770935,
0.0076136840507388115,
-0.010768214240670204,
0.13709646463394165,
-0.21793904900550842,
-0.028675777837634087,
0.16038778424263,
-0.09573561698198318,
-0.07581635564565659,
0.061356328427791595,
-0.053495727479457855,
0.10603067278862,
0.041206974536180496,
-0.0253668911755085,
0.06327881664037704,
-0.1329721212387085,
0.0035808775573968887,
-0.045546747744083405,
-0.018045976758003235,
0.16059941053390503,
0.07648856192827225,
-0.06927888095378876,
0.07070852816104889,
0.024078376591205597,
-0.026113634929060936,
-0.046159777790308,
-0.018227294087409973,
-0.1095207929611206,
0.010817212983965874,
-0.060132887214422226,
0.02282119169831276,
-0.025257518514990807,
-0.09224134683609009,
-0.028882192447781563,
-0.17473143339157104,
-0.01543671078979969,
0.0841502845287323,
-0.008552714250981808,
-0.019770942628383636,
-0.11773128807544708,
0.014404429122805595,
0.038445498794317245,
0.0027449382469058037,
-0.13180501759052277,
-0.050576433539390564,
0.027280667796730995,
-0.1619797945022583,
0.03360215947031975,
-0.051585033535957336,
0.05001852661371231,
0.0318891666829586,
-0.03249809890985489,
-0.028068941086530685,
0.022396177053451538,
0.005391599610447884,
-0.013861955143511295,
-0.24684561789035797,
-0.02524387463927269,
-0.022996457293629646,
0.16599495708942413,
-0.21521428227424622,
0.03807735815644264,
0.0733959823846817,
0.1517697423696518,
0.009051989763975143,
-0.03698690980672836,
0.0017559787956997752,
-0.07388318330049515,
-0.03124045953154564,
-0.05945143476128578,
-0.007037308998405933,
-0.03604515269398689,
-0.05787331610918045,
0.047536347061395645,
-0.16902238130569458,
-0.02949199639260769,
0.10039274394512177,
0.06592373549938202,
-0.13678009808063507,
-0.022425442934036255,
-0.034927621483802795,
-0.04287220165133476,
-0.05416050925850868,
-0.05881212651729584,
0.10532104223966599,
0.05762924998998642,
0.04464157670736313,
-0.0650712326169014,
-0.07518194615840912,
0.00088359450455755,
-0.020143218338489532,
-0.023713968694210052,
0.09234753251075745,
0.07105407863855362,
-0.126350998878479,
0.09208080917596817,
0.10551702976226807,
0.08511383831501007,
0.09815585613250732,
-0.024149423465132713,
-0.08191148191690445,
-0.050659939646720886,
0.02379353903234005,
0.01600629836320877,
0.13259312510490417,
-0.010838953778147697,
0.05292908474802971,
0.04124988242983818,
-0.013232617639005184,
0.009245194494724274,
-0.0925365537405014,
0.03198198601603508,
0.03315291553735733,
-0.018429066985845566,
0.039537809789180756,
-0.03881950303912163,
0.020082874223589897,
0.08976095914840698,
0.047349054366350174,
0.039120472967624664,
0.014505230821669102,
-0.046636730432510376,
-0.11192648112773895,
0.16611367464065552,
-0.12793833017349243,
-0.23291675746440887,
-0.14571763575077057,
0.003718912834301591,
0.03641049191355705,
-0.010390745475888252,
0.002204331336542964,
-0.06504169851541519,
-0.11800546944141388,
-0.09107901155948639,
0.010856508277356625,
0.049631036818027496,
-0.08566083759069443,
-0.05643118917942047,
0.05523066222667694,
0.039479292929172516,
-0.14542964100837708,
0.01921185478568077,
0.04928894340991974,
-0.09167466312646866,
-0.008233107626438141,
0.08074086904525757,
0.06674882769584656,
0.18043169379234314,
0.013242475688457489,
-0.022343328222632408,
0.032658565789461136,
0.21998950839042664,
-0.1353374421596527,
0.1128942146897316,
0.14020681381225586,
-0.09332811832427979,
0.08355985581874847,
0.20060832798480988,
0.04187845438718796,
-0.10058243572711945,
0.03296395763754845,
0.017997587099671364,
-0.030420765280723572,
-0.24256370961666107,
-0.07092253863811493,
-0.00026266687200404704,
-0.0599735751748085,
0.07366035133600235,
0.08954169601202011,
0.09123681485652924,
0.01494339108467102,
-0.0955287516117096,
-0.080891452729702,
0.056770894676446915,
0.10385555773973465,
0.019311824813485146,
-0.012641520239412785,
0.09103459119796753,
-0.03278684988617897,
0.016931859776377678,
0.0904497355222702,
0.0008944828878156841,
0.17511117458343506,
0.058405566960573196,
0.18374158442020416,
0.0765325129032135,
0.07154922187328339,
0.015365427359938622,
0.009895091876387596,
0.017764348536729813,
0.02660132572054863,
-0.0053646075539290905,
-0.08453443646430969,
-0.014433449134230614,
0.11945675313472748,
0.07353336364030838,
0.017197363078594208,
0.016192223876714706,
-0.04000629484653473,
0.08344162255525589,
0.17407093942165375,
-0.003780076280236244,
-0.18052507936954498,
-0.06431038677692413,
0.08350689709186554,
-0.09346359968185425,
-0.10017222911119461,
-0.02494942955672741,
0.030767329037189484,
-0.17044265568256378,
0.0249007735401392,
-0.016930779442191124,
0.11206945031881332,
-0.13528640568256378,
-0.019095007330179214,
0.06340263038873672,
0.07177523523569107,
-0.0006523873889818788,
0.058229442685842514,
-0.16294988989830017,
0.10450614243745804,
0.012098570354282856,
0.06693841516971588,
-0.09612328559160233,
0.09953869134187698,
-0.005955029278993607,
-0.010155374184250832,
0.1313311606645584,
0.009115277789533138,
-0.07581817358732224,
-0.07944932579994202,
-0.09122282266616821,
-0.009041238576173782,
0.126266211271286,
-0.14647246897220612,
0.08482405543327332,
-0.03597019985318184,
-0.0416097566485405,
0.002930275397375226,
-0.10596253722906113,
-0.12220548838376999,
-0.18631164729595184,
0.055513981729745865,
-0.13507777452468872,
0.03854088857769966,
-0.10657316446304321,
-0.035541050136089325,
-0.030116569250822067,
0.18516884744167328,
-0.22976601123809814,
-0.06906338781118393,
-0.15047605335712433,
-0.09873856604099274,
0.14586862921714783,
-0.050321947783231735,
0.08481817692518234,
-0.00589280528947711,
0.1804574877023697,
0.02166794426739216,
-0.021489109843969345,
0.09810362011194229,
-0.09247367084026337,
-0.19692669808864594,
-0.08017813414335251,
0.15722282230854034,
0.13640479743480682,
0.036161039024591446,
-0.003470085794106126,
0.038310710340738297,
-0.019128555431962013,
-0.12300188839435577,
0.021808674558997154,
0.17748361825942993,
0.06226111575961113,
0.02378440462052822,
-0.025610120967030525,
-0.11692396551370621,
-0.06900777667760849,
-0.03363456577062607,
0.030739158391952515,
0.1859661191701889,
-0.07158373296260834,
0.18602654337882996,
0.14774003624916077,
-0.058341678231954575,
-0.19670341908931732,
0.009590700268745422,
0.0356709361076355,
0.0062993373721838,
0.03402268886566162,
-0.20171645283699036,
0.08260589838027954,
-0.0000703737823641859,
-0.05092230439186096,
0.12990811467170715,
-0.1724688857793808,
-0.15031461417675018,
0.07340911030769348,
0.036582015454769135,
-0.191009521484375,
-0.11979404836893082,
-0.08877003937959671,
-0.05305791646242142,
-0.18255825340747833,
0.10235996544361115,
0.03505839407444,
0.007234846707433462,
0.033568330109119415,
0.030296791344881058,
0.016846131533384323,
-0.03902881219983101,
0.19317437708377838,
-0.025881100445985794,
0.03175598382949829,
-0.08486942201852798,
-0.0721178725361824,
0.04698624834418297,
-0.05440608412027359,
0.07560842484235764,
-0.02850610576570034,
0.010811456479132175,
-0.10112031549215317,
-0.04238447546958923,
-0.02994711697101593,
0.014171373099088669,
-0.09643256664276123,
-0.0892103835940361,
-0.04899745434522629,
0.09385206550359726,
0.09383191168308258,
-0.03679990395903587,
-0.033308759331703186,
-0.0708332359790802,
0.04319954290986061,
0.1834612935781479,
0.1771630197763443,
0.04282272979617119,
-0.07718019932508469,
-0.004353965632617474,
-0.012391943484544754,
0.04512987285852432,
-0.216888889670372,
0.0646008849143982,
0.04998873919248581,
0.017488451674580574,
0.119838647544384,
-0.02023271657526493,
-0.15518377721309662,
-0.06958208978176117,
0.06293158233165741,
-0.05947147309780121,
-0.19729353487491608,
0.005153949372470379,
0.05639190226793289,
-0.16896353662014008,
-0.04793788120150566,
0.04407742992043495,
-0.004272493068128824,
-0.04013913497328758,
0.019694777205586433,
0.08993566036224365,
0.003983452916145325,
0.06979537010192871,
0.057179566472768784,
0.08297405391931534,
-0.10303977131843567,
0.07298099994659424,
0.08502772450447083,
-0.07904176414012909,
0.02611508034169674,
0.09225213527679443,
-0.05959508195519447,
-0.03061521053314209,
0.024558385834097862,
0.08217264711856842,
0.011403042823076248,
-0.04143837094306946,
0.011890463531017303,
-0.10493540018796921,
0.061552610248327255,
0.0872589722275734,
0.033055905252695084,
0.014960144646465778,
0.0323425829410553,
0.04615075886249542,
-0.06838563084602356,
0.12262481451034546,
0.028603000566363335,
0.01619773730635643,
-0.039672788232564926,
-0.04883408173918724,
0.023656455799937248,
-0.03148266673088074,
-0.006783190183341503,
-0.034929223358631134,
-0.07470342516899109,
-0.017337948083877563,
-0.16813358664512634,
-0.015706919133663177,
-0.04851626604795456,
0.01141374558210373,
0.030727919191122055,
-0.039717670530080795,
0.008398020640015602,
0.007660517003387213,
-0.0750393494963646,
-0.06366884708404541,
-0.022168075665831566,
0.09360232949256897,
-0.16274383664131165,
0.0231650248169899,
0.08795329183340073,
-0.12010334432125092,
0.093709796667099,
0.017991894856095314,
-0.005580618511885405,
0.030415862798690796,
-0.15203481912612915,
0.03863019123673439,
-0.030480829998850822,
0.014001374132931232,
0.0430658757686615,
-0.2246030569076538,
-0.00014216898125596344,
-0.03392428159713745,
-0.06211007013916969,
-0.008089113049209118,
-0.03614491969347,
-0.11279971152544022,
0.10460628569126129,
0.00755698699504137,
-0.09058507531881332,
-0.031222015619277954,
0.03175928816199303,
0.08461960405111313,
-0.023131132125854492,
0.15904083847999573,
-0.003058732021600008,
0.07367146760225296,
-0.16738978028297424,
-0.019550222903490067,
-0.009911867789924145,
0.019858263432979584,
-0.021036015823483467,
-0.013049607165157795,
0.039891984313726425,
-0.023009097203612328,
0.1832437962293625,
-0.02614782750606537,
0.02115444466471672,
0.06662114709615707,
0.031309690326452255,
-0.027116473764181137,
0.10507345199584961,
0.05415768921375275,
0.02187212184071541,
0.019098330289125443,
0.0009401091956533492,
-0.04275880753993988,
-0.026286903768777847,
-0.20222914218902588,
0.06478860974311829,
0.14196400344371796,
0.09015882015228271,
-0.019612208008766174,
0.082443006336689,
-0.09847161918878555,
-0.11266232281923294,
0.12008036673069,
-0.05389230325818062,
-0.005624994169920683,
-0.06746947765350342,
0.1300724893808365,
0.1476544737815857,
-0.19186675548553467,
0.07097877562046051,
-0.06951600313186646,
-0.049371387809515,
-0.11596925556659698,
-0.19563089311122894,
-0.0579688623547554,
-0.05182981118559837,
-0.01601085439324379,
-0.04734842851758003,
0.07465895265340805,
0.05611773207783699,
0.007687699515372515,
-0.0008743742946535349,
0.06143457442522049,
-0.025334985926747322,
-0.00020855919865425676,
0.026871640235185623,
0.06543756276369095,
0.01277367677539587,
-0.028740158304572105,
0.017883067950606346,
-0.009029596112668514,
0.04200661554932594,
0.06335420906543732,
0.04628584161400795,
-0.029481444507837296,
0.015326657332479954,
-0.04012451693415642,
-0.10723566263914108,
0.0419759564101696,
-0.027193402871489525,
-0.08225540816783905,
0.14785805344581604,
0.02399633452296257,
0.009130166843533516,
-0.019807780161499977,
0.24066896736621857,
-0.07349257171154022,
-0.09820521622896194,
-0.1490466147661209,
0.10572494566440582,
-0.04344555735588074,
0.06301548331975937,
0.046267855912446976,
-0.10314686596393585,
0.017511071637272835,
0.12181373685598373,
0.1645330935716629,
-0.0424080528318882,
0.020746879279613495,
0.0270154420286417,
0.0042765699326992035,
-0.03603396192193031,
0.05116730183362961,
0.06923956423997879,
0.15690730512142181,
-0.04900969937443733,
0.09728439897298813,
-0.0028536769095808268,
-0.09575258195400238,
-0.03706898167729378,
0.11544451117515564,
-0.015991076827049255,
0.016934793442487717,
-0.05673288553953171,
0.11960674822330475,
-0.06204056739807129,
-0.23141460120677948,
0.059825554490089417,
-0.0669918954372406,
-0.136424258351326,
-0.021505871787667274,
0.08302503079175949,
-0.012656555511057377,
0.027061283588409424,
0.07233231514692307,
-0.07476846873760223,
0.1983826905488968,
0.036584943532943726,
-0.05408371239900589,
-0.05269603058695793,
0.08472228795289993,
-0.10221095383167267,
0.2712222933769226,
0.01703515276312828,
0.05113283172249794,
0.10301852226257324,
-0.012269208207726479,
-0.13325491547584534,
0.021790657192468643,
0.09555859118700027,
-0.09376159310340881,
0.04151248186826706,
0.19850991666316986,
0.00041501864325255156,
0.12115148454904556,
0.08070485293865204,
-0.07617492228746414,
0.04874774441123009,
-0.09502684324979782,
-0.07229389250278473,
-0.08947079628705978,
0.09731078147888184,
-0.07675671577453613,
0.14196333289146423,
0.13027575612068176,
-0.05282389745116234,
0.009622754529118538,
-0.02872757986187935,
0.046359382569789886,
0.0037509698886424303,
0.1005694642663002,
0.008517593145370483,
-0.18598082661628723,
0.021583275869488716,
0.013244304805994034,
0.10594569146633148,
-0.1648784875869751,
-0.09909144788980484,
0.03998237103223801,
0.0028793413657695055,
-0.06011311709880829,
0.12933798134326935,
0.060960717499256134,
0.04498228803277016,
-0.0423818863928318,
-0.02333086170256138,
-0.009747753851115704,
0.13612231612205505,
-0.10099201649427414,
0.0027629989199340343
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | Jaerim/bloom-7b1-lora-tagger_3 | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-08T06:49:56+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
31,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06646376848220825,
0.2168014943599701,
-0.00225935154594481,
0.023818302899599075,
0.1271018385887146,
-0.001635765191167593,
0.04218708351254463,
0.13324736058712006,
-0.020175931975245476,
0.11144465953111649,
0.046588581055402756,
0.09377603232860565,
0.09928803145885468,
0.18404334783554077,
0.04859916493296623,
-0.2059975117444992,
0.007056170143187046,
-0.09090408682823181,
0.014076028019189835,
0.1116579994559288,
0.13719257712364197,
-0.10291384905576706,
0.08272874355316162,
-0.04045208916068077,
-0.02019004337489605,
0.00012576708104461432,
-0.09259183704853058,
-0.07032395154237747,
0.06885425746440887,
0.06264153122901917,
0.051234472543001175,
0.001456156256608665,
0.09140396863222122,
-0.2864592671394348,
0.017265573143959045,
0.08406311273574829,
0.0027674848679453135,
0.06290827691555023,
0.07236549258232117,
-0.07389893382787704,
0.11328595131635666,
-0.08021481335163116,
0.13019037246704102,
0.08625296503305435,
-0.062064990401268005,
-0.23071379959583282,
-0.07525765895843506,
0.0963398814201355,
0.12251301854848862,
0.06215599179267883,
-0.022921854630112648,
0.15455181896686554,
-0.06248689442873001,
0.012971068732440472,
0.1294165402650833,
-0.11526761949062347,
-0.05572471022605896,
0.061741601675748825,
0.11775490641593933,
0.10740239918231964,
-0.14110268652439117,
-0.0017287094378843904,
0.04900608956813812,
0.029121357947587967,
0.08589313924312592,
0.022661056369543076,
0.12003941088914871,
0.04652795568108559,
-0.13695219159126282,
-0.04037507623434067,
0.12011898308992386,
0.038862764835357666,
-0.06446044892072678,
-0.2168138176202774,
-0.006778308190405369,
-0.0601806715130806,
-0.014732478186488152,
-0.07019448280334473,
0.039128515869379044,
-0.02470310963690281,
0.07317749410867691,
-0.04465159401297569,
-0.1063927412033081,
-0.0421026237308979,
0.0892222449183464,
0.07748593389987946,
0.011527054943144321,
-0.02519804798066616,
0.04627908393740654,
0.13455867767333984,
0.05402068421244621,
-0.10399353504180908,
-0.07017925381660461,
-0.06942764669656754,
-0.09420394152402878,
-0.04035796597599983,
0.056760527193546295,
0.031942449510097504,
0.02665667235851288,
0.22703726589679718,
0.016653569415211678,
0.04155244305729866,
0.0224777739495039,
0.01032855175435543,
0.043662428855895996,
0.0955500528216362,
-0.05303520709276199,
-0.15660029649734497,
-0.04072032496333122,
0.09077946096658707,
-0.0027527001220732927,
-0.036689214408397675,
-0.03966725245118141,
0.03849169611930847,
0.06843466311693192,
0.13122352957725525,
0.07552056759595871,
-0.017929591238498688,
-0.04813180863857269,
-0.030096933245658875,
0.23523783683776855,
-0.1493375599384308,
0.04426715523004532,
-0.02271856553852558,
-0.01804111897945404,
-0.03908449783921242,
0.03597262129187584,
0.022118929773569107,
-0.000004518366949923802,
0.09706240892410278,
-0.058981191366910934,
-0.05378659814596176,
-0.10168042778968811,
-0.03272576630115509,
0.04088849574327469,
-0.013975566253066063,
-0.010589460842311382,
-0.09025166928768158,
-0.09490354359149933,
-0.04766594246029854,
0.05537205561995506,
-0.05123869329690933,
-0.03770573064684868,
0.009465423412621021,
-0.08151785284280777,
-0.005444355774670839,
-0.005417742300778627,
0.10699385404586792,
-0.03222226724028587,
0.04445803165435791,
-0.027600755915045738,
0.05225523188710213,
0.09919606149196625,
0.031576547771692276,
-0.0773419588804245,
0.0561848059296608,
-0.22559374570846558,
0.07503069192171097,
-0.11481974273920059,
0.04335082694888115,
-0.1704932004213333,
-0.042439818382263184,
0.005444696638733149,
0.0139949731528759,
0.013206101022660732,
0.12720820307731628,
-0.19255615770816803,
-0.01654396951198578,
0.13260798156261444,
-0.09212633967399597,
-0.118110790848732,
0.07884611934423447,
-0.029701577499508858,
0.1624738723039627,
0.04682036489248276,
-0.027025915682315826,
0.09224298596382141,
-0.16434773802757263,
-0.07092688232660294,
-0.00949116237461567,
-0.01727987825870514,
0.12109188735485077,
0.07512219995260239,
-0.05991523340344429,
0.046571120619773865,
0.02832140028476715,
-0.038078423589468,
-0.04424772411584854,
-0.050857074558734894,
-0.10884185880422592,
-0.01070026308298111,
-0.08987759798765182,
0.04065500199794769,
-0.01250192429870367,
-0.07916021347045898,
-0.029885273426771164,
-0.18612512946128845,
-0.0030564051121473312,
0.10038342326879501,
0.0035033065360039473,
-0.005652366206049919,
-0.08666291832923889,
0.026358824223279953,
-0.03112892620265484,
-0.008404186926782131,
-0.16764774918556213,
-0.04399421438574791,
0.046902090311050415,
-0.16094985604286194,
0.020117372274398804,
-0.06413903087377548,
0.06334125250577927,
0.03641495108604431,
-0.05590536445379257,
-0.0248766727745533,
-0.01730942726135254,
0.011945613659918308,
-0.05083848536014557,
-0.18994836509227753,
-0.056277405470609665,
-0.037882111966609955,
0.149809330701828,
-0.25956398248672485,
0.032966937869787216,
0.051140617579221725,
0.14649195969104767,
0.00406361510977149,
-0.05115427449345589,
0.01429014839231968,
-0.05360214412212372,
-0.054652128368616104,
-0.06746816635131836,
-0.006135428790003061,
-0.027576493099331856,
-0.05147203803062439,
0.019243421033024788,
-0.1755700707435608,
-0.021410830318927765,
0.09424154460430145,
0.12876708805561066,
-0.1486445665359497,
-0.018640631809830666,
-0.048725154250860214,
-0.06339836865663528,
-0.0715010017156601,
-0.07038594037294388,
0.10712739825248718,
0.0513901449739933,
0.04796046018600464,
-0.07435787469148636,
-0.07092321664094925,
0.02726263552904129,
0.006906150374561548,
-0.03382374346256256,
0.08727246522903442,
0.05199531093239784,
-0.09209315478801727,
0.0756213590502739,
0.1092359870672226,
0.07177663594484329,
0.09363535046577454,
0.01574566215276718,
-0.11756632477045059,
-0.028492970392107964,
0.036266472190618515,
0.02740776725113392,
0.1465986967086792,
-0.05952361226081848,
0.04016614332795143,
0.04494241625070572,
-0.04170418903231621,
0.022319864481687546,
-0.08787637203931808,
0.024075502529740334,
0.025203049182891846,
-0.0034381982404738665,
0.06284574419260025,
-0.02525499276816845,
-0.0050758360885083675,
0.07016654312610626,
0.047779910266399384,
0.04621000960469246,
0.009655474685132504,
-0.01720241829752922,
-0.1047825813293457,
0.16950392723083496,
-0.0951867327094078,
-0.269941508769989,
-0.17632324993610382,
0.026197833940386772,
0.04035249724984169,
-0.022378476336598396,
0.031619444489479065,
-0.07056326419115067,
-0.10630585998296738,
-0.1060405746102333,
-0.002429972169920802,
0.01714223250746727,
-0.06364088505506516,
-0.0741225928068161,
0.07348573952913284,
0.04382912442088127,
-0.14902326464653015,
0.038552410900592804,
0.055694397538900375,
-0.057955220341682434,
-0.0233661737293005,
0.09118817001581192,
0.12397737801074982,
0.14583967626094818,
-0.021366750821471214,
-0.028626007959246635,
0.029004426673054695,
0.19620531797409058,
-0.13469526171684265,
0.10371150821447372,
0.13814030587673187,
-0.04545360431075096,
0.08360563963651657,
0.1560150384902954,
0.029186224564909935,
-0.08317049592733383,
0.05044832453131676,
0.04082648828625679,
-0.043159641325473785,
-0.2666129767894745,
-0.0534592866897583,
0.012832709588110447,
-0.06255637854337692,
0.09786593168973923,
0.10183793306350708,
0.11542957276105881,
0.034910861402750015,
-0.07166364789009094,
-0.043925940990448,
-0.0058974819257855415,
0.11737963557243347,
-0.05490213260054588,
-0.012639665976166725,
0.07686592638492584,
-0.05086168646812439,
0.005355054512619972,
0.10266812145709991,
0.02973790094256401,
0.17442677915096283,
0.020399179309606552,
0.11231429129838943,
0.06195578724145889,
0.08633565157651901,
0.0007386076031252742,
0.02951662428677082,
0.05147615820169449,
0.017203815281391144,
-0.002300140680745244,
-0.10421168059110641,
-0.006156572140753269,
0.1449710875749588,
0.028103826567530632,
0.029669636860489845,
-0.0018948549404740334,
-0.005003341939300299,
0.05121048167347908,
0.1746254414319992,
-0.011592294089496136,
-0.22072425484657288,
-0.0845772922039032,
0.06936841458082199,
-0.06218599155545235,
-0.12968985736370087,
-0.026130788028240204,
0.045467354357242584,
-0.17519839107990265,
0.026703642681241035,
-0.027433741837739944,
0.0919293761253357,
-0.09345759451389313,
-0.02221956104040146,
0.03687324374914169,
0.084866963326931,
-0.014529162086546421,
0.08703910559415817,
-0.14498743414878845,
0.11886418610811234,
0.02978132851421833,
0.09024628251791,
-0.11081171780824661,
0.07909037172794342,
-0.007550720125436783,
0.009180475026369095,
0.19379350543022156,
-0.011335089802742004,
-0.03514958545565605,
-0.08774717897176743,
-0.11210042238235474,
-0.013537433929741383,
0.12687496840953827,
-0.1243172138929367,
0.08773399889469147,
-0.015198243781924248,
-0.044079482555389404,
0.00937260314822197,
-0.12100647389888763,
-0.17273177206516266,
-0.19628387689590454,
0.05585884302854538,
-0.09575839340686798,
0.025643249973654747,
-0.11914430558681488,
-0.07089093327522278,
-0.02952558360993862,
0.241120383143425,
-0.1745356321334839,
-0.06510113179683685,
-0.1468164622783661,
-0.046294767409563065,
0.1662203073501587,
-0.04437198117375374,
0.0718095526099205,
-0.0208172257989645,
0.20345525443553925,
0.005988610442727804,
-0.004939318168908358,
0.06724198162555695,
-0.08892562240362167,
-0.16873881220817566,
-0.06771010160446167,
0.1510489284992218,
0.11680185794830322,
0.04907919466495514,
-0.002248800592496991,
0.0011772146681323647,
-0.016943959519267082,
-0.1137804463505745,
-0.0033210667315870523,
0.16037839651107788,
0.03878779336810112,
0.025986969470977783,
-0.05243593826889992,
-0.08797456324100494,
-0.06899320334196091,
-0.06853509694337845,
0.06221301481127739,
0.19590823352336884,
-0.10376439243555069,
0.1700313836336136,
0.147536963224411,
-0.07305635511875153,
-0.23175598680973053,
0.035342130810022354,
0.04983805492520332,
0.0014306638622656465,
0.04886869341135025,
-0.18252557516098022,
0.10521943867206573,
0.019543392583727837,
-0.05505957826972008,
0.13485197722911835,
-0.1557481735944748,
-0.1552847921848297,
0.0722852572798729,
0.03904085233807564,
-0.22423844039440155,
-0.1354004591703415,
-0.09622503817081451,
-0.05825018882751465,
-0.14065024256706238,
0.06054598465561867,
-0.002136280992999673,
0.015948504209518433,
0.03500790148973465,
-0.0015643214574083686,
0.027123261243104935,
-0.058935679495334625,
0.18609118461608887,
-0.004065449349582195,
0.020676052197813988,
-0.060264769941568375,
-0.0478842556476593,
0.09839435666799545,
-0.06130504235625267,
0.12208222597837448,
0.004057085141539574,
0.01594383642077446,
-0.10362856835126877,
-0.048314861953258514,
-0.04328322783112526,
0.05154227837920189,
-0.07548051327466965,
-0.10070807486772537,
-0.043625857681035995,
0.08841723203659058,
0.07005169242620468,
-0.03383097052574158,
0.00549331633374095,
-0.07189501076936722,
0.10019614547491074,
0.17795267701148987,
0.17573626339435577,
0.009926567785441875,
-0.07241068035364151,
0.01677953451871872,
-0.04142116755247116,
0.044231921434402466,
-0.2513144314289093,
0.03756171092391014,
0.06098250672221184,
0.029438555240631104,
0.09217222779989243,
-0.020435843616724014,
-0.1820858269929886,
-0.04050002992153168,
0.08094815909862518,
-0.05452597141265869,
-0.22617179155349731,
-0.019085140898823738,
0.0954197570681572,
-0.2020406424999237,
-0.007372708059847355,
0.03995226323604584,
-0.048725228756666183,
-0.023169852793216705,
0.00010950004070764408,
0.06317184865474701,
0.002471912419423461,
0.09773622453212738,
0.0735151618719101,
0.09715340286493301,
-0.08337292820215225,
0.10562895983457565,
0.10150538384914398,
-0.09572599828243256,
0.03605884686112404,
0.06754924356937408,
-0.05300498008728027,
-0.043293699622154236,
0.03665391728281975,
0.033023297786712646,
0.005234600510448217,
-0.060321882367134094,
0.013913018628954887,
-0.036497246474027634,
0.044923391193151474,
0.08326134830713272,
0.03754979372024536,
-0.013354414142668247,
0.06462216377258301,
0.03401726484298706,
-0.10898099094629288,
0.10366570204496384,
0.01731540448963642,
0.04105307161808014,
-0.08384523540735245,
-0.019968897104263306,
0.035425446927547455,
0.030576206743717194,
-0.01765924133360386,
-0.02306121215224266,
-0.02860277332365513,
-0.01614218018949032,
-0.14299540221691132,
-0.023106401786208153,
-0.07243485748767853,
0.006181265693157911,
0.014656842686235905,
-0.031884219497442245,
-0.011233693920075893,
0.02475680410861969,
-0.06979699432849884,
-0.07426341623067856,
-0.006949664559215307,
0.09833318740129471,
-0.15115703642368317,
0.008848577737808228,
0.06907843053340912,
-0.11088496446609497,
0.08190931379795074,
-0.008411259390413761,
0.016245156526565552,
0.022527478635311127,
-0.15448406338691711,
0.05601610988378525,
0.0008648968650959432,
0.01916889287531376,
0.025886621326208115,
-0.16471809148788452,
0.004104440100491047,
-0.04661374166607857,
-0.02149827405810356,
-0.00004464812809601426,
-0.02647159807384014,
-0.12325995415449142,
0.06858719140291214,
-0.015622655861079693,
-0.035931166261434555,
-0.02701525390148163,
0.0539589487016201,
0.07888586074113846,
-0.027474910020828247,
0.10445091128349304,
-0.008690856397151947,
0.04941811040043831,
-0.16801609098911285,
-0.02470702864229679,
-0.04982255399227142,
0.019377702847123146,
0.009884213097393513,
-0.007693959400057793,
0.04183054715394974,
-0.00976533442735672,
0.21883612871170044,
-0.05075952783226967,
0.1607085019350052,
0.05847611650824547,
-0.017352959141135216,
-0.0007513365126214921,
0.06180921941995621,
0.05997028574347496,
0.04658793285489082,
0.009480604901909828,
0.023740366101264954,
-0.022450892254710197,
-0.006695089396089315,
-0.15932634472846985,
0.01890849508345127,
0.14999441802501678,
0.06301083415746689,
0.024745315313339233,
0.05866100639104843,
-0.12775006890296936,
-0.12135478109121323,
0.09311001747846603,
-0.026755332946777344,
0.00928465835750103,
-0.08245618641376495,
0.1358020007610321,
0.14980104565620422,
-0.14000412821769714,
0.05256148427724838,
-0.06134212389588356,
-0.05217423290014267,
-0.10388828068971634,
-0.12032219022512436,
-0.05887215584516525,
-0.053666237741708755,
0.002330566756427288,
-0.03760887682437897,
0.054546963423490524,
0.03344334661960602,
-0.009351172484457493,
-0.00022941511997487396,
0.13597318530082703,
-0.019751882180571556,
-0.0028988157864660025,
0.048313532024621964,
0.03693558648228645,
0.02373051457107067,
-0.05275435373187065,
0.02940409444272518,
0.02539868652820587,
0.032232340425252914,
0.06546790152788162,
0.033412106335163116,
-0.047448933124542236,
0.03804153576493263,
-0.0025254099164158106,
-0.11207924783229828,
0.019641218706965446,
-0.00460948096588254,
-0.0742158442735672,
0.1268945336341858,
0.0407399944961071,
0.010224059224128723,
-0.03741471841931343,
0.24361543357372284,
-0.06653323769569397,
-0.06378097087144852,
-0.13251738250255585,
0.10491154342889786,
-0.0027236645109951496,
0.06476365029811859,
0.023412218317389488,
-0.1284150779247284,
0.005243356805294752,
0.13858191668987274,
0.12181595712900162,
0.0045748427510261536,
0.009228081442415714,
0.0518609918653965,
0.0025186820421367884,
-0.06998204439878464,
0.054019294679164886,
0.06992026418447495,
0.12919506430625916,
-0.07847554981708527,
0.07680778950452805,
0.0006860480643808842,
-0.08370215445756912,
-0.02947772853076458,
0.11312682181596756,
-0.0409729965031147,
0.03491825982928276,
-0.047444481402635574,
0.10916327685117722,
-0.05787910893559456,
-0.29412412643432617,
0.02350960113108158,
-0.09588567912578583,
-0.15202060341835022,
-0.018367812037467957,
0.05944539234042168,
-0.02624768204987049,
0.018029648810625076,
0.06971040368080139,
-0.06011629104614258,
0.20098382234573364,
0.0335683599114418,
-0.07864278554916382,
-0.0664360448718071,
0.04837050288915634,
-0.06564252078533173,
0.2949807047843933,
0.008418165147304535,
0.02863333560526371,
0.10770907253026962,
-0.03253700211644173,
-0.18271861970424652,
0.010723991319537163,
0.1133992001414299,
-0.08056149631738663,
0.08200647681951523,
0.19000613689422607,
-0.012578671798110008,
0.1209007054567337,
0.05294662341475487,
-0.047376248985528946,
0.04217283055186272,
-0.03389401361346245,
-0.051268599927425385,
-0.10752558708190918,
0.058453381061553955,
-0.05909625440835953,
0.15447644889354706,
0.10152646154165268,
-0.05671518296003342,
-0.004550917539745569,
-0.05555408447980881,
0.04875178262591362,
0.01804669201374054,
0.12263146042823792,
0.02951994352042675,
-0.1865430772304535,
0.032826557755470276,
-0.01144319772720337,
0.10186848044395447,
-0.25588861107826233,
-0.08421015739440918,
0.08833149075508118,
-0.011924264021217823,
-0.05105875805020332,
0.10560628771781921,
0.057650718837976456,
0.04243382066488266,
-0.043439045548439026,
-0.10480839014053345,
-0.02186836116015911,
0.14663739502429962,
-0.1469624787569046,
-0.025013303384184837
] |
null | null | transformers | # merged
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
## Merge Details
### Merge Method
This model was merged using the passthrough merge method.
### Models Merged
The following models were included in the merge:
* [ChaiML/season_4_top_solution](https://huggingface.co/ChaiML/season_4_top_solution)
### Configuration
The following YAML configuration was used to produce this model:
```yaml
dtype: bfloat16
merge_method: passthrough
slices:
- sources:
- layer_range: [0, 30]
model:
model:
path: ChaiML/season_4_top_solution
- sources:
- layer_range: [10, 40]
model:
model:
path: ChaiML/season_4_top_solution
```
| {"library_name": "transformers", "tags": ["mergekit", "merge"], "base_model": ["ChaiML/season_4_top_solution"]} | text-generation | gotchu/s8-knarf | [
"transformers",
"safetensors",
"llama",
"text-generation",
"mergekit",
"merge",
"base_model:ChaiML/season_4_top_solution",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T06:50:26+00:00 | [] | [] | TAGS
#transformers #safetensors #llama #text-generation #mergekit #merge #base_model-ChaiML/season_4_top_solution #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # merged
This is a merge of pre-trained language models created using mergekit.
## Merge Details
### Merge Method
This model was merged using the passthrough merge method.
### Models Merged
The following models were included in the merge:
* ChaiML/season_4_top_solution
### Configuration
The following YAML configuration was used to produce this model:
| [
"# merged\n\nThis is a merge of pre-trained language models created using mergekit.",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the passthrough merge method.",
"### Models Merged\n\nThe following models were included in the merge:\n* ChaiML/season_4_top_solution",
"### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #mergekit #merge #base_model-ChaiML/season_4_top_solution #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# merged\n\nThis is a merge of pre-trained language models created using mergekit.",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the passthrough merge method.",
"### Models Merged\n\nThe following models were included in the merge:\n* ChaiML/season_4_top_solution",
"### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
72,
19,
4,
17,
29,
17
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #mergekit #merge #base_model-ChaiML/season_4_top_solution #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# merged\n\nThis is a merge of pre-trained language models created using mergekit.## Merge Details### Merge Method\n\nThis model was merged using the passthrough merge method.### Models Merged\n\nThe following models were included in the merge:\n* ChaiML/season_4_top_solution### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
-0.06761243939399719,
-0.142740860581398,
-0.0021057273261249065,
-0.0336528979241848,
0.12789930403232574,
0.030466478317975998,
0.19344773888587952,
0.023075919598340988,
0.0036073438823223114,
0.012661099433898926,
0.014597713947296143,
0.10411302745342255,
0.06168333813548088,
0.19056786596775055,
-0.044030118733644485,
-0.13674531877040863,
0.059748757630586624,
-0.04980362206697464,
-0.19018331170082092,
0.10590068995952606,
0.08144126832485199,
-0.07756754010915756,
0.11693236976861954,
0.028768450021743774,
-0.18758323788642883,
0.0633242279291153,
-0.022208664566278458,
0.011119463481009007,
0.1024223119020462,
0.11638275533914566,
0.12122946977615356,
0.02650836482644081,
-0.046620771288871765,
-0.17818763852119446,
0.06432143598794937,
0.016852712258696556,
0.005874840077012777,
0.009418129920959473,
0.08086742460727692,
-0.008031228557229042,
0.09247303754091263,
-0.08665934950113297,
-0.017861846834421158,
0.06608514487743378,
-0.10732768476009369,
0.01296653039753437,
-0.09562461078166962,
0.056258250027894974,
0.2109098881483078,
0.0046988497488200665,
-0.03520035743713379,
-0.0054289549589157104,
0.012860098853707314,
0.0783834233880043,
-0.014856180176138878,
-0.3449316620826721,
0.014723585918545723,
0.1201985701918602,
0.07650275528430939,
-0.11575498431921005,
0.05044294521212578,
0.0729188397526741,
0.06592381745576859,
-0.04825850948691368,
0.02710762992501259,
-0.041316140443086624,
0.13085517287254333,
-0.07168598473072052,
-0.12346227467060089,
-0.03258352354168892,
0.14084061980247498,
-0.00206738174892962,
-0.023146461695432663,
-0.09683994948863983,
-0.141667440533638,
0.05601752921938896,
0.009637963026762009,
0.001052699750289321,
0.006857884582132101,
0.02444642409682274,
0.07215877622365952,
-0.0688626766204834,
-0.08502517640590668,
-0.023212704807519913,
-0.1497822105884552,
0.2573879063129425,
0.07311135530471802,
0.04675714671611786,
-0.09827720373868942,
0.06079927086830139,
-0.041195161640644073,
-0.10756353288888931,
0.030788948759436607,
-0.038542792201042175,
-0.06524945795536041,
-0.008176298812031746,
-0.10228065401315689,
-0.1750500202178955,
0.1318642497062683,
0.1277630478143692,
-0.09826791286468506,
0.008571702055633068,
0.08641398698091507,
0.06594937294721603,
0.07969645410776138,
0.009755263105034828,
-0.15225224196910858,
-0.09600617736577988,
0.054930608719587326,
-0.0039838082157075405,
0.13973338901996613,
-0.0022199638187885284,
-0.12310735136270523,
-0.020914597436785698,
-0.0332491509616375,
0.028772776946425438,
0.0366705060005188,
0.10229307413101196,
-0.05267379805445671,
-0.06506408751010895,
0.08917753398418427,
-0.10383091866970062,
0.013360390439629555,
-0.016830896958708763,
0.008006000891327858,
-0.03596397861838341,
0.10953094810247421,
0.06989683210849762,
0.0053845481015741825,
0.07994289696216583,
-0.013317296281456947,
-0.01472637802362442,
-0.08653365820646286,
-0.08144281059503555,
-0.005992460064589977,
-0.03566666319966316,
-0.0033453740179538727,
-0.057958684861660004,
-0.271245539188385,
-0.02618265524506569,
0.029131922870874405,
-0.022833138704299927,
-0.027537772431969643,
-0.037800226360559464,
0.023310353979468346,
-0.027255704626441002,
-0.02517630159854889,
-0.03422527760267258,
-0.025097232311964035,
-0.028933456167578697,
-0.03150118142366409,
0.05701170116662979,
-0.11062639951705933,
0.029796073213219643,
-0.06938319653272629,
0.13292072713375092,
-0.12949806451797485,
0.13776962459087372,
0.0443321168422699,
0.0916093960404396,
-0.06929440051317215,
0.025586377829313278,
-0.030892658978700638,
0.06893490254878998,
0.05481860041618347,
0.1823584884405136,
-0.10846046358346939,
-0.07271716743707657,
0.15949048101902008,
-0.16194495558738708,
-0.14311280846595764,
0.0726533904671669,
-0.004366873297840357,
0.05662059038877487,
0.07814417034387589,
0.18720212578773499,
0.07525888830423355,
-0.03277478367090225,
-0.05036585405468941,
-0.009517140686511993,
-0.039272211492061615,
-0.05334141105413437,
0.0563155822455883,
0.024806629866361618,
-0.14462903141975403,
0.023072276264429092,
0.059704285115003586,
0.16303154826164246,
-0.05326031893491745,
-0.049524083733558655,
-0.06960340589284897,
-0.06700950860977173,
0.057537615299224854,
-0.044385876506567,
0.025318970903754234,
-0.0643860399723053,
0.03556094318628311,
0.13640868663787842,
0.08725285530090332,
-0.046465106308460236,
0.01420536171644926,
-0.03876478224992752,
0.10733090341091156,
-0.1452765017747879,
0.045519061386585236,
-0.06656118482351303,
-0.09869805723428726,
-0.04204355180263519,
0.06973887234926224,
0.04572577029466629,
0.012604007497429848,
0.06578224897384644,
0.03964705020189285,
-0.040968459099531174,
-0.06429032981395721,
0.13914085924625397,
0.020425375550985336,
-0.027968842536211014,
-0.20196083188056946,
-0.06715484708547592,
-0.07028839737176895,
0.24383363127708435,
-0.03406777232885361,
0.07368536293506622,
-0.05114021152257919,
0.23225176334381104,
-0.028776589781045914,
0.04994559660553932,
0.10394024848937988,
0.040410444140434265,
-0.05051172897219658,
0.018444271758198738,
0.07654319703578949,
0.0400288887321949,
-0.17757058143615723,
0.1850624978542328,
-0.1781235933303833,
0.04605155438184738,
0.10776340961456299,
-0.08346717804670334,
0.00951626431196928,
-0.09377598762512207,
-0.022987212985754013,
-0.05581212043762207,
0.01832377351820469,
-0.06438495218753815,
0.10277815908193588,
0.01924196071922779,
0.15039163827896118,
-0.02694021165370941,
0.01695503480732441,
-0.014001558534801006,
-0.08239208161830902,
-0.038653772324323654,
0.044368792325258255,
-0.019945811480283737,
-0.15262994170188904,
0.12323000282049179,
0.19546274840831757,
0.07343040406703949,
0.15362021327018738,
-0.016887230798602104,
0.01932809129357338,
-0.06024467945098877,
0.03529457747936249,
-0.028986280784010887,
-0.031264834105968475,
-0.01683366671204567,
0.021124469116330147,
0.04431889206171036,
-0.03956327214837074,
0.06644658744335175,
-0.13374027609825134,
0.028262630105018616,
0.08166808634996414,
-0.011267594993114471,
0.14933450520038605,
0.09843070805072784,
0.00950612872838974,
0.038807086646556854,
0.003943809773772955,
0.027744680643081665,
0.02689998783171177,
-0.017706584185361862,
-0.09488198161125183,
0.17683278024196625,
-0.12638279795646667,
-0.23772498965263367,
-0.22013714909553528,
-0.09448189288377762,
-0.17215462028980255,
0.013936917297542095,
0.05876852199435234,
-0.0841251015663147,
-0.05395792797207832,
-0.08186861127614975,
0.16842304170131683,
0.059631992131471634,
0.009723966009914875,
-0.05578845739364624,
-0.050884537398815155,
0.033570967614650726,
-0.07998786121606827,
-0.010360369458794594,
-0.004890142939984798,
-0.06671608239412308,
0.05898052453994751,
-0.026938196271657944,
0.11042825132608414,
0.16402146220207214,
-0.031834084540605545,
-0.0035832375288009644,
-0.010256655514240265,
0.1681322604417801,
-0.027508359402418137,
0.08150281757116318,
0.20588818192481995,
-0.08720578253269196,
0.05751550942659378,
0.2586362957954407,
0.011112739332020283,
-0.029717013239860535,
0.014528011903166771,
-0.024693693965673447,
-0.1122337132692337,
-0.1767898052930832,
-0.1375913918018341,
-0.08863283693790436,
-0.01631048321723938,
0.002442058641463518,
0.04343004524707794,
0.05085621029138565,
0.0784367248415947,
-0.07753497362136841,
-0.0031092159915715456,
-0.010174566879868507,
0.0299350805580616,
0.22746887803077698,
-0.025386318564414978,
0.09775153547525406,
-0.08002372831106186,
-0.043291278183460236,
0.05312647297978401,
0.07351324707269669,
0.10213957726955414,
0.05596422031521797,
0.04159430041909218,
0.14606700837612152,
-0.0022606297861784697,
0.07073165476322174,
0.06382003426551819,
-0.05577288195490837,
0.025764282792806625,
-0.0038561944384127855,
-0.1005842313170433,
-0.011826125904917717,
0.07089720666408539,
-0.1217406839132309,
0.06988625228404999,
-0.06335903704166412,
0.039306558668613434,
0.0880490392446518,
0.15907469391822815,
0.12060447782278061,
-0.22892692685127258,
-0.05880552530288696,
0.07376863062381744,
0.02609846368432045,
-0.02189251407980919,
-0.03539332002401352,
0.0021473087836056948,
-0.0179787315428257,
0.22001147270202637,
-0.01720855012536049,
0.10426751524209976,
-0.007758045103400946,
0.012207882478833199,
-0.026031818240880966,
0.10399188101291656,
0.004978592973202467,
0.029138419777154922,
-0.09474785625934601,
0.17272022366523743,
0.04399225860834122,
-0.011663142591714859,
0.041755445301532745,
0.03750303387641907,
0.03770187497138977,
0.2549665570259094,
0.011117071844637394,
0.01804533414542675,
-0.009202671237289906,
-0.004687131382524967,
-0.13414977490901947,
-0.0013099790085107088,
-0.02540789544582367,
-0.03255625069141388,
0.10193224996328354,
-0.05060514062643051,
-0.033925145864486694,
-0.002806835575029254,
0.08725686371326447,
-0.07501976937055588,
-0.1219387874007225,
0.031181370839476585,
0.12251991033554077,
0.05583205819129944,
-0.0605497770011425,
-0.02535538375377655,
-0.12135466188192368,
0.2712673544883728,
0.06488891690969467,
-0.1215224415063858,
-0.09424779564142227,
-0.0447106808423996,
0.07341805845499039,
-0.051653604954481125,
0.0888170376420021,
-0.04155900701880455,
0.03642981871962547,
-0.0475919246673584,
-0.19892849028110504,
0.07323477417230606,
-0.0771809071302414,
-0.01965983211994171,
-0.010770364664494991,
0.1277683526277542,
-0.08497081696987152,
-0.014405734837055206,
0.020907841622829437,
0.05056378245353699,
-0.0769835114479065,
-0.0420902855694294,
-0.014895680360496044,
0.2572650909423828,
0.04501621425151825,
0.16752156615257263,
-0.03894157335162163,
-0.17983321845531464,
0.03282513469457626,
-0.04156435281038284,
0.18426765501499176,
0.19972673058509827,
-0.03907342627644539,
0.09379161894321442,
0.11674205213785172,
-0.08896134048700333,
-0.2512457072734833,
-0.08813782036304474,
-0.06251047551631927,
0.10019564628601074,
-0.04442518576979637,
-0.04931266978383064,
0.06652777642011642,
0.05739818140864372,
-0.018417339771986008,
0.007688471116125584,
-0.2530641257762909,
-0.19943773746490479,
0.050413258373737335,
0.08716162294149399,
0.3564630150794983,
-0.13357071578502655,
-0.08286605030298233,
-0.0913674384355545,
-0.07182539254426956,
-0.02433832176029682,
-0.23710303008556366,
0.07296675443649292,
-0.01497607957571745,
0.023885637521743774,
0.023521788418293,
-0.04705352708697319,
0.1286763846874237,
-0.03060179017484188,
0.039888180792331696,
-0.08730636537075043,
0.06061297655105591,
0.08142626285552979,
-0.07290643453598022,
0.0781710296869278,
-0.16167199611663818,
0.022509917616844177,
-0.07270606607198715,
-0.037759117782115936,
0.004460237920284271,
0.06019562482833862,
-0.01896909810602665,
-0.039333827793598175,
-0.07506684958934784,
-0.03525327891111374,
0.034570492804050446,
-0.003396532731130719,
0.08948234468698502,
-0.050653476268053055,
0.12154112011194229,
0.24241653084754944,
0.10891552269458771,
-0.07644075155258179,
0.009539763443171978,
0.02196701243519783,
-0.06709066778421402,
0.06087623909115791,
-0.12902683019638062,
0.010386835783720016,
0.09406737983226776,
-0.014129115268588066,
0.1018001064658165,
0.010240326635539532,
-0.006936208810657263,
0.027876120060682297,
0.13178063929080963,
-0.18139834702014923,
-0.299674928188324,
-0.025564387440681458,
-0.011618100106716156,
0.00036007369635626674,
0.11764174699783325,
0.17512209713459015,
-0.04381026700139046,
0.002155440393835306,
0.008245982229709625,
0.011615200899541378,
-0.0813247337937355,
0.07991079241037369,
-0.01930108293890953,
0.026421019807457924,
-0.09614701569080353,
0.051796574145555496,
0.04179976135492325,
-0.07532162964344025,
0.007184797432273626,
0.024681994691491127,
-0.10251788794994354,
-0.08519679307937622,
-0.06658731400966644,
0.2015969455242157,
-0.09667611122131348,
-0.10231037437915802,
-0.12384580075740814,
-0.16613684594631195,
0.007053451146930456,
0.11921576410531998,
0.060624707490205765,
0.04550952464342117,
0.017797190696001053,
-0.07459421455860138,
-0.04588505998253822,
0.09131266921758652,
0.06249493360519409,
0.0733010545372963,
-0.17191986739635468,
0.08149322122335434,
0.004759191535413265,
0.08234266191720963,
-0.06339104473590851,
-0.02612956427037716,
-0.07757510989904404,
-0.012902823276817799,
-0.1980748325586319,
-0.006069868803024292,
-0.1460813581943512,
-0.04606645926833153,
-0.004396392498165369,
-0.01636114902794361,
-0.02323833666741848,
0.019208531826734543,
-0.02708418294787407,
-0.0012443347368389368,
-0.05496792122721672,
0.013714298605918884,
-0.06842787563800812,
-0.05364836007356644,
-0.0022490352857857943,
-0.06056087091565132,
0.042693331837654114,
0.015524575486779213,
-0.061981506645679474,
-0.05145541951060295,
-0.06520537286996841,
-0.06115267425775528,
0.0897492840886116,
-0.018495285883545876,
0.025810519233345985,
-0.08377821743488312,
-0.030124735087156296,
0.03374386951327324,
-0.05202857404947281,
-0.040095411241054535,
0.04990821331739426,
-0.02376713976264,
0.04472888633608818,
-0.019259551540017128,
0.016708742827177048,
-0.03366124629974365,
-0.017840970307588577,
0.05836951732635498,
0.08267754316329956,
0.12591677904129028,
-0.07626919448375702,
0.019678810611367226,
-0.14620938897132874,
-0.003801557468250394,
-0.022412460297346115,
-0.13240733742713928,
-0.09622689336538315,
-0.1429111510515213,
-0.019138600677251816,
-0.006895937025547028,
0.21275834739208221,
0.040699854493141174,
-0.08810726553201675,
0.01870187558233738,
0.008015486411750317,
0.08880440145730972,
0.084284707903862,
0.25351566076278687,
-0.01649065688252449,
0.042106419801712036,
-0.09462942183017731,
0.05316777899861336,
0.034742217510938644,
-0.042735505849123,
0.05176347494125366,
0.05104900151491165,
-0.021304139867424965,
0.0743960365653038,
0.06608758121728897,
0.01381915993988514,
-0.07170440256595612,
-0.22063863277435303,
-0.12110215425491333,
0.06644771248102188,
-0.017888227477669716,
0.13151594996452332,
0.15788264572620392,
-0.15034466981887817,
0.059700857847929,
0.04959053173661232,
-0.010889211669564247,
-0.10899118334054947,
-0.01584029570221901,
-0.10400167852640152,
-0.15357735753059387,
-0.039535362273454666,
-0.0678909569978714,
-0.04032038897275925,
-0.000593578617554158,
0.0027094874531030655,
0.019839758053421974,
0.22628509998321533,
0.027426695451140404,
-0.0014848889550194144,
0.01671954058110714,
-0.024151548743247986,
0.01304530818015337,
-0.009322179481387138,
-0.02554689720273018,
0.03698301315307617,
-0.03328549116849899,
-0.023844216018915176,
0.05658023804426193,
0.0695887953042984,
0.08538182824850082,
-0.058291055262088776,
-0.07799072563648224,
-0.021572468802332878,
0.06253848224878311,
0.11047624051570892,
-0.08141741156578064,
0.03717828541994095,
-0.06988748162984848,
-0.0015213459264487028,
0.03903829678893089,
-0.048676930367946625,
-0.08708367496728897,
-0.06349599361419678,
0.20061810314655304,
-0.038171492516994476,
0.023808153346180916,
0.026119466871023178,
-0.07290620356798172,
0.008890477009117603,
0.14555859565734863,
0.3476392328739166,
-0.009537232108414173,
0.006633502431213856,
-0.08356525003910065,
0.030388658866286278,
0.028828680515289307,
0.10942403972148895,
-0.006335549056529999,
0.14645889401435852,
-0.04113755375146866,
0.04274637624621391,
-0.0386643260717392,
-0.10630834102630615,
-0.08048927038908005,
0.02366589568555355,
-0.031917013227939606,
-0.049290288239717484,
0.03684314340353012,
0.08729659020900726,
-0.07361927628517151,
-0.021330857649445534,
0.05889030545949936,
-0.14441229403018951,
-0.07714033871889114,
-0.09263473749160767,
0.15845398604869843,
0.013787987641990185,
0.04246135056018829,
-0.07621587812900543,
0.0012427198234945536,
0.11419670283794403,
0.010086767375469208,
-0.10871510207653046,
-0.03021477721631527,
0.06009557098150253,
-0.035500265657901764,
-0.02685130201280117,
-0.002782420488074422,
-0.004375019110739231,
0.08248859643936157,
-0.024608954787254333,
-0.1076752096414566,
-0.0018262348603457212,
-0.01229084562510252,
-0.028409749269485474,
0.028874587267637253,
-0.01460476778447628,
-0.018100472167134285,
-0.04796426370739937,
0.004861354827880859,
-0.16603444516658783,
0.029781010001897812,
-0.03325110673904419,
-0.08763759583234787,
-0.06435532122850418,
0.05347248166799545,
-0.030936896800994873,
0.118751659989357,
0.13791416585445404,
-0.04267067462205887,
0.025315556675195694,
-0.011323494836688042,
0.027849774807691574,
0.05208253860473633,
0.10146196186542511,
0.006664627697318792,
-0.15501435101032257,
0.007474126759916544,
0.056350767612457275,
0.06767645478248596,
-0.3091873824596405,
-0.09323783218860626,
-0.0964648574590683,
0.003940503112971783,
-0.08244174718856812,
0.13707563281059265,
0.20796778798103333,
0.026194443926215172,
-0.01465279795229435,
-0.15278410911560059,
0.010987953282892704,
0.07891152054071426,
-0.05430950969457626,
-0.10523738712072372
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# image_classification
This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the imagefolder dataset.
It achieves the following results on the evaluation set:
- Loss: 1.1599
- Accuracy: 0.5813
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 13
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| No log | 1.0 | 40 | 1.8887 | 0.35 |
| No log | 2.0 | 80 | 1.5494 | 0.425 |
| No log | 3.0 | 120 | 1.4015 | 0.5188 |
| No log | 4.0 | 160 | 1.2919 | 0.55 |
| No log | 5.0 | 200 | 1.2205 | 0.5813 |
| No log | 6.0 | 240 | 1.2246 | 0.575 |
| No log | 7.0 | 280 | 1.2053 | 0.5312 |
| No log | 8.0 | 320 | 1.1487 | 0.5687 |
| No log | 9.0 | 360 | 1.1727 | 0.5437 |
| No log | 10.0 | 400 | 1.1459 | 0.55 |
| No log | 11.0 | 440 | 1.1313 | 0.5813 |
| No log | 12.0 | 480 | 1.0990 | 0.6062 |
| 1.1138 | 13.0 | 520 | 1.1020 | 0.6188 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["imagefolder"], "metrics": ["accuracy"], "base_model": "google/vit-base-patch16-224-in21k", "model-index": [{"name": "image_classification", "results": [{"task": {"type": "image-classification", "name": "Image Classification"}, "dataset": {"name": "imagefolder", "type": "imagefolder", "config": "default", "split": "train", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.58125, "name": "Accuracy"}]}]}]} | image-classification | yangswei/image_classification | [
"transformers",
"tensorboard",
"safetensors",
"vit",
"image-classification",
"generated_from_trainer",
"dataset:imagefolder",
"base_model:google/vit-base-patch16-224-in21k",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T06:51:26+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| image\_classification
=====================
This model is a fine-tuned version of google/vit-base-patch16-224-in21k on the imagefolder dataset.
It achieves the following results on the evaluation set:
* Loss: 1.1599
* Accuracy: 0.5813
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 13
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 13",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 13",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
86,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 13### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.11764444410800934,
0.13512200117111206,
-0.002589508658275008,
0.12045709043741226,
0.14152748882770538,
0.004859213251620531,
0.13963647186756134,
0.1385362297296524,
-0.07143764942884445,
0.08122877776622772,
0.14856930077075958,
0.13030961155891418,
0.029905881732702255,
0.18729659914970398,
-0.04841630160808563,
-0.22331304848194122,
0.026304438710212708,
0.04659973457455635,
-0.04857369884848595,
0.12089377641677856,
0.08843084424734116,
-0.1368962675333023,
0.11702775210142136,
0.025227433070540428,
-0.20190371572971344,
-0.007630900479853153,
0.029175853356719017,
-0.05752947926521301,
0.11698196828365326,
0.03700752183794975,
0.0912281796336174,
0.027913378551602364,
0.05264570936560631,
-0.14976441860198975,
0.010312250815331936,
0.07499676942825317,
-0.008718756958842278,
0.09642095863819122,
0.05588778108358383,
0.01283137034624815,
0.012070349417626858,
-0.09279842674732208,
0.040429770946502686,
0.026327237486839294,
-0.11253516376018524,
-0.23308773338794708,
-0.08563393354415894,
0.057643093168735504,
0.07787211239337921,
0.07102856040000916,
-0.003061442170292139,
0.14272943139076233,
-0.006720641162246466,
0.09734538942575455,
0.22933481633663177,
-0.27243363857269287,
-0.07755240797996521,
0.040147412568330765,
0.01753450557589531,
0.07819264382123947,
-0.1013517901301384,
0.011260110884904861,
0.05913977697491646,
0.015300070866942406,
0.15208584070205688,
-0.004673824179917574,
-0.013608341105282307,
-0.02613564394414425,
-0.12513789534568787,
-0.06483881175518036,
0.19181954860687256,
0.0891055092215538,
-0.04710124433040619,
-0.08061584830284119,
-0.0809321179986,
-0.1400146186351776,
-0.046531613916158676,
-0.011150557547807693,
0.057746175676584244,
-0.03411676362156868,
-0.06340470910072327,
-0.03488915413618088,
-0.09833012521266937,
-0.07019275426864624,
-0.015052413567900658,
0.09754949808120728,
0.054355017840862274,
0.012958547100424767,
-0.019707130268216133,
0.08153021335601807,
-0.040454570204019547,
-0.14453008770942688,
-0.00784135702997446,
0.01813052035868168,
0.025137145072221756,
-0.031004641205072403,
-0.025600699707865715,
-0.1118408665060997,
0.02071315236389637,
0.11035110801458359,
-0.0710933730006218,
0.05702487751841545,
-0.019823284819722176,
0.05147066339850426,
-0.11218857020139694,
0.19005782902240753,
-0.049158353358507156,
0.019304519519209862,
0.039781201630830765,
0.10269298404455185,
0.050096914172172546,
-0.002129604807123542,
-0.10551011562347412,
0.018774231895804405,
0.12039201706647873,
0.005552207585424185,
-0.03542640060186386,
0.08170582354068756,
-0.06239338219165802,
-0.029671501368284225,
0.0782092809677124,
-0.08725659549236298,
0.025570768862962723,
-0.006866145879030228,
-0.053714364767074585,
-0.05802810564637184,
0.04711175337433815,
-0.011755510233342648,
-0.01319239940494299,
0.042459990829229355,
-0.10256795585155487,
0.011821204796433449,
-0.0674973651766777,
-0.10702585428953171,
0.013958062045276165,
-0.11495786160230637,
0.015685664489865303,
-0.12262342125177383,
-0.1391865760087967,
-0.013133692555129528,
0.060412224382162094,
-0.0299366507679224,
-0.05301636457443237,
-0.041731175035238266,
-0.08092506974935532,
0.026501361280679703,
0.004902180749922991,
0.04675145819783211,
-0.057117000222206116,
0.08835763484239578,
0.04579019919037819,
0.07608752697706223,
-0.018667640164494514,
0.04670323058962822,
-0.08841915428638458,
0.05831769481301308,
-0.20483596622943878,
0.037254828959703445,
-0.059762995690107346,
0.0871657133102417,
-0.12046878784894943,
-0.08726757019758224,
0.00030698769842274487,
-0.020322751253843307,
0.0645289272069931,
0.10858000069856644,
-0.13768890500068665,
-0.05775922164320946,
0.17217335104942322,
-0.10212460160255432,
-0.1556587964296341,
0.11335211247205734,
-0.030357878655195236,
0.029488753527402878,
0.054946426302194595,
0.19757983088493347,
0.08099635690450668,
-0.10896632075309753,
-0.007540930062532425,
-0.03140242025256157,
0.03474791720509529,
-0.0542144849896431,
0.07627690583467484,
-0.0008913844940252602,
-0.011290287598967552,
0.0229901522397995,
-0.09551648050546646,
0.0630020871758461,
-0.07357678562402725,
-0.08483672142028809,
-0.06562598794698715,
-0.0882251113653183,
0.04349168390035629,
0.059430111199617386,
0.06544730067253113,
-0.10060568898916245,
-0.09089488536119461,
0.027707798406481743,
0.08103296905755997,
-0.09358508884906769,
0.017559798434376717,
-0.08080413937568665,
0.1117594838142395,
-0.10985339432954788,
0.0007958461064845324,
-0.13462485373020172,
-0.03258748725056648,
0.04881887510418892,
-0.06264283508062363,
-0.009185576811432838,
-0.037363965064287186,
0.0733700767159462,
0.06154854968190193,
-0.06359858810901642,
-0.07255588471889496,
-0.04064885899424553,
-0.0033952370285987854,
-0.09865549206733704,
-0.1921132504940033,
-0.024556921795010567,
-0.027382701635360718,
0.10429539531469345,
-0.21726956963539124,
0.04201642796397209,
0.05168306082487106,
0.10042411834001541,
0.058460962027311325,
-0.03222675621509552,
0.0036715504247695208,
0.017343834042549133,
-0.039991382509469986,
-0.0893605425953865,
0.062201112508773804,
0.014153112657368183,
-0.06716355681419373,
0.006629870273172855,
-0.09946059435606003,
0.17481762170791626,
0.1300860047340393,
-0.032943420112133026,
-0.0632399469614029,
-0.00400611013174057,
-0.044182319194078445,
-0.035492442548274994,
-0.03642500936985016,
0.008564289659261703,
0.08287134021520615,
-0.009528233669698238,
0.16156966984272003,
-0.1051715612411499,
-0.02630491927266121,
0.058716658502817154,
-0.029900044202804565,
-0.03812507539987564,
0.0896214172244072,
0.07011791318655014,
-0.1343931257724762,
0.14796310663223267,
0.16474199295043945,
-0.06841951608657837,
0.12494634836912155,
-0.04831704497337341,
-0.06263326853513718,
-0.02280440926551819,
0.041197869926691055,
0.03274824842810631,
0.1280038058757782,
-0.11916659027338028,
-0.012296305038034916,
0.023531144484877586,
0.003234709380194545,
-0.007457838859409094,
-0.20078811049461365,
-0.00925714336335659,
0.0388016514480114,
-0.06060701608657837,
0.025046344846487045,
-0.005864563398063183,
-0.021288294345140457,
0.08464224636554718,
0.008177232928574085,
-0.04257797449827194,
0.046960532665252686,
0.011051027104258537,
-0.0703202560544014,
0.19552773237228394,
-0.08368364721536636,
-0.21444712579250336,
-0.1324203461408615,
-0.022895105183124542,
-0.07981423288583755,
0.02140718139708042,
0.058641888201236725,
-0.09433171898126602,
-0.0565846785902977,
-0.10398397594690323,
-0.013813691213726997,
0.030053643509745598,
0.039813462644815445,
0.04285653308033943,
-0.0030209491960704327,
0.13106122612953186,
-0.09951378405094147,
-0.007041654083877802,
-0.010529419407248497,
-0.024749042466282845,
0.04835669323801994,
0.01947885937988758,
0.12096858024597168,
0.08704964071512222,
-0.027948306873440742,
0.033293843269348145,
-0.021374141797423363,
0.24313397705554962,
-0.07297474890947342,
-0.0023365309461951256,
0.14955854415893555,
0.01986546628177166,
0.0682375431060791,
0.1311814934015274,
0.038684118539094925,
-0.10270018130540848,
0.008326326496899128,
0.02277129888534546,
-0.02626965194940567,
-0.1892206221818924,
-0.017310189083218575,
-0.039190880954265594,
-0.0033366098068654537,
0.15156055986881256,
0.05579404532909393,
0.05959703400731087,
0.09367187321186066,
0.000661893398500979,
0.09127334505319595,
-0.0055085644125938416,
0.08726560324430466,
0.11121205240488052,
0.04689955711364746,
0.10906160622835159,
-0.043784789741039276,
-0.02814451977610588,
0.03241283446550369,
0.017767248675227165,
0.2261905074119568,
0.00018619047477841377,
0.17437347769737244,
0.04733264073729515,
0.18962348997592926,
0.01753518544137478,
0.05614832416176796,
-0.021440016105771065,
-0.026949504390358925,
-0.009223824366927147,
-0.05370432883501053,
-0.020865973085165024,
0.03586485981941223,
-0.05119192972779274,
0.06472262740135193,
-0.09288249909877777,
0.04214414209127426,
0.06292396783828735,
0.2632756531238556,
0.03747488185763359,
-0.3779374659061432,
-0.09564990550279617,
-0.006072385236620903,
-0.014438515529036522,
-0.06247309222817421,
0.004186393227428198,
0.14460279047489166,
-0.060622889548540115,
0.059474509209394455,
-0.10481241345405579,
0.08202364295721054,
-0.049894049763679504,
0.02150817960500717,
0.07837644219398499,
0.08868986368179321,
0.008372707292437553,
0.05654231831431389,
-0.24768905341625214,
0.25872477889060974,
0.01516481302678585,
0.06299551576375961,
-0.04647960513830185,
0.013162878341972828,
0.03546437993645668,
0.10373103618621826,
0.1106199100613594,
-0.005229472182691097,
-0.016594525426626205,
-0.17520642280578613,
-0.08914074301719666,
0.00697906082496047,
0.07169909030199051,
-0.046438198536634445,
0.08142077922821045,
-0.03018474578857422,
-0.022935478016734123,
0.05089205130934715,
-0.004208697006106377,
-0.0891035869717598,
-0.09378670901060104,
-0.00762608926743269,
0.042228102684020996,
0.013459181413054466,
-0.09488574415445328,
-0.09803754091262817,
-0.10428707301616669,
0.13289755582809448,
-0.01745361089706421,
-0.041044894605875015,
-0.11994276940822601,
0.0865074023604393,
0.05955522134900093,
-0.09321940690279007,
0.08078111708164215,
-0.02865653671324253,
0.1357981264591217,
0.029795458540320396,
-0.06278637051582336,
0.11005957424640656,
-0.06004214659333229,
-0.17319291830062866,
-0.04652365297079086,
0.10509270429611206,
-0.019316820427775383,
0.025458350777626038,
-0.00048521097050979733,
0.02748608961701393,
-0.011054742150008678,
-0.06002514809370041,
0.0585169643163681,
0.012034153565764427,
0.057546231895685196,
-0.01444156002253294,
-0.020859505981206894,
0.006801279727369547,
-0.06444567441940308,
-0.0297895148396492,
0.13443174958229065,
0.24594737589359283,
-0.09737911075353622,
0.004677808377891779,
0.019995464012026787,
-0.0515986904501915,
-0.19615793228149414,
0.046961456537246704,
0.06626344472169876,
0.0011840381193906069,
0.03195241093635559,
-0.1555091142654419,
0.07275909930467606,
0.08196834474802017,
-0.030794763937592506,
0.09419076144695282,
-0.2662884294986725,
-0.13322611153125763,
0.08016213029623032,
0.18357788026332855,
0.06874960660934448,
-0.1449221819639206,
-0.055395353585481644,
-0.010100364685058594,
-0.09369542449712753,
0.09211120009422302,
-0.057994063943624496,
0.10525432229042053,
-0.028548715636134148,
0.004616732709109783,
0.005854741670191288,
-0.057884424924850464,
0.12890754640102386,
-0.03200128674507141,
0.10716796666383743,
-0.057005077600479126,
-0.010405335575342178,
0.07675597816705704,
-0.07650814205408096,
0.0638655349612236,
-0.08939897269010544,
0.06398385763168335,
-0.06266579777002335,
-0.016289029270410538,
-0.07105870544910431,
0.032188836485147476,
-0.018952373415231705,
-0.026158887892961502,
-0.05123016610741615,
0.024189038202166557,
0.05222386494278908,
-0.0013514412567019463,
0.1991061568260193,
0.050031546503305435,
0.08935361355543137,
0.13871558010578156,
0.04288501664996147,
-0.07489487528800964,
-0.10000263899564743,
-0.02747277542948723,
-0.027515830472111702,
0.08667431026697159,
-0.1858905553817749,
0.049684587866067886,
0.09706258028745651,
0.010297784581780434,
0.14298567175865173,
0.04670065641403198,
-0.033227287232875824,
0.019350646063685417,
0.07108762115240097,
-0.15481282770633698,
-0.1636543571949005,
-0.03231717273592949,
-0.019337838515639305,
-0.11691375076770782,
0.06289277970790863,
0.11156196892261505,
-0.08573035895824432,
0.0031364557798951864,
-0.008033327758312225,
0.015850963070988655,
-0.001984128262847662,
0.16285808384418488,
0.0811963602900505,
0.04528191313147545,
-0.09190085530281067,
0.09902695566415787,
0.05369001626968384,
-0.1053948774933815,
0.022020861506462097,
0.026600094512104988,
-0.10377729684114456,
-0.03789014369249344,
0.06682208925485611,
0.14378942549228668,
0.00033475851523689926,
-0.05008783936500549,
-0.14449909329414368,
-0.09342958778142929,
0.058374036103487015,
0.12437126785516739,
0.09284667670726776,
0.015723219141364098,
-0.011195636354386806,
0.0006879014545120299,
-0.1046130433678627,
0.1197706088423729,
0.032290246337652206,
0.09787023067474365,
-0.21870984137058258,
0.05767025798559189,
0.018016472458839417,
0.032032255083322525,
-0.01947706565260887,
0.029871582984924316,
-0.09776318073272705,
-0.016205523163080215,
-0.0601126067340374,
0.0421827994287014,
-0.036765385419130325,
0.005503166001290083,
-0.006183840334415436,
-0.06899906694889069,
-0.06144210696220398,
0.04135039448738098,
-0.1007462590932846,
-0.04604232311248779,
0.03564571589231491,
0.06997189670801163,
-0.10228724032640457,
-0.029401684179902077,
0.025855081155896187,
-0.08015374094247818,
0.08018931746482849,
0.013313505798578262,
0.0003344202705193311,
0.023971425369381905,
-0.10144920647144318,
0.01170396152883768,
0.0844736322760582,
0.0023789203260093927,
0.029592592269182205,
-0.10345650464296341,
0.006262886803597212,
-0.0018232337897643447,
0.002120407996699214,
-0.008592193014919758,
0.1037750393152237,
-0.1340494155883789,
-0.025066139176487923,
-0.037659477442502975,
-0.0332927368581295,
-0.05909410119056702,
0.06091965734958649,
0.08533547818660736,
-0.0029338907916098833,
0.20013189315795898,
-0.08626281470060349,
0.0012225453974679112,
-0.22378350794315338,
0.004308170638978481,
-0.004870743956416845,
-0.1362777054309845,
-0.12583312392234802,
-0.028006596490740776,
0.052500441670417786,
-0.07272263616323471,
0.09633250534534454,
0.015528041869401932,
0.005463740788400173,
0.03590396046638489,
0.003014509566128254,
-0.001735204947181046,
0.02674783580005169,
0.18505123257637024,
-0.007147321943193674,
-0.021157708019018173,
0.07146983593702316,
0.01909119077026844,
0.11649779230356216,
0.0840657502412796,
0.10088535398244858,
0.16292303800582886,
-0.04234790802001953,
0.10513481497764587,
0.04933501407504082,
-0.022084658965468407,
-0.1734275370836258,
0.10123051702976227,
-0.07545798271894455,
0.14333204925060272,
-0.013059891760349274,
0.16278716921806335,
0.12182081490755081,
-0.1579466015100479,
0.028677551075816154,
-0.02910970337688923,
-0.07299938797950745,
-0.0713169202208519,
-0.14282052218914032,
-0.11673738062381744,
-0.18597784638404846,
0.014860942959785461,
-0.09846791625022888,
0.006735004484653473,
0.07586970925331116,
-0.009442432783544064,
-0.022980276495218277,
0.2072959542274475,
0.05031755566596985,
-0.0018605367513373494,
0.07015055418014526,
0.000791032100096345,
-0.06759081780910492,
-0.05971204489469528,
-0.08503936976194382,
0.038723040372133255,
-0.008470187894999981,
0.03299541398882866,
-0.030025074258446693,
-0.006795855239033699,
0.049253251403570175,
-0.0005805885302834213,
-0.11057381331920624,
0.017208252102136612,
0.016095079481601715,
0.010045529343187809,
0.003046760568395257,
0.0037861166056245565,
0.0076500470750033855,
-0.00897123571485281,
0.18375752866268158,
-0.054615505039691925,
-0.0060697584412992,
-0.11899099498987198,
0.12503525614738464,
0.02825990691781044,
-0.015883982181549072,
0.028842493891716003,
-0.07998210936784744,
0.02650350145995617,
0.21896962821483612,
0.14496728777885437,
-0.018135562539100647,
-0.0015389241743832827,
-0.006991059985011816,
-0.01981694996356964,
-0.02901512198150158,
0.0935748741030693,
0.09348926693201065,
-0.043009962886571884,
-0.05549543350934982,
-0.025126053020358086,
-0.04698945954442024,
-0.016143640503287315,
-0.03650468960404396,
0.0376470722258091,
0.01750567927956581,
0.014662797562777996,
-0.06436789035797119,
0.04223893582820892,
0.01847166009247303,
-0.07113940268754959,
0.08869649469852448,
-0.1957550197839737,
-0.13880327343940735,
-0.033047061413526535,
0.10037659108638763,
-0.005442218855023384,
0.02980172634124756,
-0.021247144788503647,
0.015624051913619041,
0.06931367516517639,
-0.02371610701084137,
-0.08351006358861923,
-0.09853099286556244,
0.050749458372592926,
-0.1301218867301941,
0.24626897275447845,
-0.03196502476930618,
0.0067982180044054985,
0.11156277358531952,
0.0185833927243948,
-0.1193198636174202,
0.05168553441762924,
0.02622678317129612,
-0.037482794374227524,
0.02558443695306778,
0.11117885261774063,
-0.020647253841161728,
0.10736607760190964,
0.03427661955356598,
-0.08952891081571579,
-0.018075330182909966,
-0.053039148449897766,
-0.04048994183540344,
-0.056705474853515625,
-0.02405516803264618,
-0.06871236115694046,
0.12355363368988037,
0.1737527847290039,
-0.04224829003214836,
-0.026600105687975883,
-0.0633816346526146,
0.037416111677885056,
0.09020824730396271,
0.017519410699605942,
-0.013283582404255867,
-0.22859162092208862,
0.016995728015899658,
0.022902945056557655,
-0.003902466269209981,
-0.21472617983818054,
-0.10973909497261047,
-0.017607787624001503,
-0.052848756313323975,
-0.08901337534189224,
0.08712980896234512,
0.11765968054533005,
0.05002998560667038,
-0.060335226356983185,
-0.045158326625823975,
-0.06922067701816559,
0.16004028916358948,
-0.12357012927532196,
-0.08564318716526031
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# TrOCR_0208-2
This model is a fine-tuned version of [microsoft/trocr-base-stage1](https://huggingface.co/microsoft/trocr-base-stage1) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 1.2584
- Cer: 0.1211
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 20
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Cer |
|:-------------:|:-----:|:----:|:---------------:|:------:|
| 1.3873 | 1.71 | 500 | 1.6813 | 0.2361 |
| 0.8298 | 3.42 | 1000 | 1.7390 | 0.2441 |
| 0.5587 | 5.14 | 1500 | 1.5896 | 0.2090 |
| 0.376 | 6.85 | 2000 | 1.4717 | 0.1775 |
| 0.2847 | 8.56 | 2500 | 1.5528 | 0.1928 |
| 0.2376 | 10.27 | 3000 | 1.4412 | 0.1727 |
| 0.2101 | 11.99 | 3500 | 1.3770 | 0.1592 |
| 0.2551 | 13.7 | 4000 | 1.4311 | 0.1564 |
| 0.226 | 15.41 | 4500 | 1.2536 | 0.1337 |
| 0.1365 | 17.12 | 5000 | 1.2753 | 0.1272 |
| 0.14 | 18.84 | 5500 | 1.2584 | 0.1211 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.13.0
- Tokenizers 0.15.0
| {"tags": ["generated_from_trainer"], "base_model": "microsoft/trocr-base-stage1", "model-index": [{"name": "TrOCR_0208-2", "results": []}]} | null | yoon1000/TrOCR_0208-2 | [
"transformers",
"tensorboard",
"safetensors",
"vision-encoder-decoder",
"generated_from_trainer",
"base_model:microsoft/trocr-base-stage1",
"endpoints_compatible",
"region:us"
] | 2024-02-08T06:51:30+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #vision-encoder-decoder #generated_from_trainer #base_model-microsoft/trocr-base-stage1 #endpoints_compatible #region-us
| TrOCR\_0208-2
=============
This model is a fine-tuned version of microsoft/trocr-base-stage1 on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 1.2584
* Cer: 0.1211
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 20
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.1+cu121
* Datasets 2.13.0
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 20\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.13.0\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #vision-encoder-decoder #generated_from_trainer #base_model-microsoft/trocr-base-stage1 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 20\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.13.0\n* Tokenizers 0.15.0"
] | [
56,
113,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #vision-encoder-decoder #generated_from_trainer #base_model-microsoft/trocr-base-stage1 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 20\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.13.0\n* Tokenizers 0.15.0"
] | [
-0.10136249661445618,
0.04104245454072952,
-0.0011656505521386862,
0.09182430803775787,
0.15523108839988708,
0.011195376515388489,
0.12403081357479095,
0.0929989293217659,
-0.10167673230171204,
0.05386224016547203,
0.12086964398622513,
0.11242122948169708,
0.021966222673654556,
0.14176194369792938,
-0.04681845381855965,
-0.2618558406829834,
0.016962645575404167,
0.04084288328886032,
-0.06808821856975555,
0.11732184141874313,
0.07198160141706467,
-0.14962396025657654,
0.07333856076002121,
-0.014492969959974289,
-0.2131451517343521,
0.017697252333164215,
0.01622425764799118,
-0.033688995987176895,
0.13350850343704224,
0.04213126003742218,
0.14364778995513916,
0.01428316067904234,
0.09481186419725418,
-0.20330406725406647,
0.0187858734279871,
0.08493708819150925,
0.005380341783165932,
0.07420040667057037,
0.06888136267662048,
0.017412636429071426,
0.0868072435259819,
-0.1059679165482521,
0.07298089563846588,
0.018254922702908516,
-0.1415831744670868,
-0.23535211384296417,
-0.06995725631713867,
0.007075622212141752,
0.08619657158851624,
0.09400925040245056,
-0.01593143120408058,
0.15670141577720642,
-0.04323451966047287,
0.11058937758207321,
0.2060774713754654,
-0.26649338006973267,
-0.08345858752727509,
0.025955703109502792,
0.055123791098594666,
0.10474283993244171,
-0.12270311266183853,
-0.0024111061356961727,
0.055388953536748886,
0.035027407109737396,
0.13808272778987885,
-0.02918299101293087,
-0.06685016304254532,
-0.004541054368019104,
-0.15999744832515717,
-0.028177879750728607,
0.08469926565885544,
0.05885259807109833,
-0.024581339210271835,
-0.030846212059259415,
-0.0847574770450592,
-0.183625727891922,
-0.05241461470723152,
-0.025770800188183784,
0.048067715018987656,
-0.027978530153632164,
-0.10676618665456772,
-0.03882588818669319,
-0.10741669684648514,
-0.07838144898414612,
-0.05383015424013138,
0.1422288715839386,
0.03381556645035744,
0.004976069554686546,
-0.018406067043542862,
0.10408798605203629,
-0.01105285994708538,
-0.13245616853237152,
0.010970397852361202,
0.02275346778333187,
-0.06640985608100891,
-0.05078279599547386,
-0.06178462132811546,
-0.07273072004318237,
-0.010900220833718777,
0.1026000902056694,
-0.04110374301671982,
0.07678427547216415,
-0.007356527727097273,
0.052450474351644516,
-0.12089550495147705,
0.1906966269016266,
-0.08478749543428421,
0.0023216085974127054,
-0.0008683409541845322,
0.06443305313587189,
0.021755244582891464,
-0.005446861032396555,
-0.09325416386127472,
0.024357913061976433,
0.1215076744556427,
-0.00032986601581797004,
-0.0644494965672493,
0.07465378940105438,
-0.05026518926024437,
-0.002177664777263999,
-0.019256481900811195,
-0.08236194401979446,
0.042114511132240295,
-0.006902105174958706,
-0.050736792385578156,
-0.030871769413352013,
0.036631401628255844,
0.018806176260113716,
0.0034054615534842014,
0.07832222431898117,
-0.08750931173563004,
0.046575989574193954,
-0.11352665722370148,
-0.1378965526819229,
0.012788736261427402,
-0.0637294203042984,
0.016116876155138016,
-0.09893019497394562,
-0.1301303505897522,
-0.022389549762010574,
0.058929041028022766,
-0.024961771443486214,
0.019727109000086784,
-0.049306921660900116,
-0.07854661345481873,
0.010990066453814507,
-0.01748831383883953,
0.1011984795331955,
-0.054077547043561935,
0.10789749026298523,
0.05869530886411667,
0.0691501647233963,
-0.06506510823965073,
0.03318015858530998,
-0.0863814428448677,
0.038309283554553986,
-0.22103500366210938,
0.058488525450229645,
-0.055566005408763885,
0.078011155128479,
-0.08555514365434647,
-0.09283559769392014,
-0.018384216353297234,
0.013041684404015541,
0.10380948334932327,
0.08424142748117447,
-0.2042497843503952,
-0.0487879142165184,
0.1641438603401184,
-0.09297596663236618,
-0.11773467808961868,
0.11711788177490234,
-0.05313583090901375,
0.030258378013968468,
0.06523099541664124,
0.19101786613464355,
0.07151895016431808,
-0.10786954313516617,
0.032178547233343124,
-0.03859509527683258,
0.0338779054582119,
-0.03020070120692253,
0.05454817786812782,
0.02060241624712944,
0.0265506599098444,
0.006996284704655409,
-0.03988800197839737,
0.07287207245826721,
-0.10289320349693298,
-0.07468308508396149,
-0.045501723885536194,
-0.08548600226640701,
0.018412111327052116,
0.07089533656835556,
0.050693437457084656,
-0.11234115064144135,
-0.08295869827270508,
0.0941343605518341,
0.0646909773349762,
-0.09645799547433853,
0.037253886461257935,
-0.08330877870321274,
0.021510422229766846,
-0.09852882474660873,
-0.02672003023326397,
-0.18059547245502472,
-0.05848383903503418,
-0.0032603899016976357,
0.0002773697779048234,
0.015156576409935951,
0.02312060073018074,
0.08610009402036667,
0.07596740126609802,
-0.05796017497777939,
-0.05293852463364601,
-0.0259809922426939,
0.011939176358282566,
-0.1286594569683075,
-0.2028454691171646,
-0.03423686698079109,
-0.03207238018512726,
0.10399220883846283,
-0.2327861785888672,
0.021928709000349045,
0.02757425233721733,
0.10769031941890717,
0.049170129001140594,
-0.02919216826558113,
-0.027778740972280502,
0.08726321160793304,
-0.022347353398799896,
-0.0854855626821518,
0.056858718395233154,
0.00021090151858516037,
-0.09355334937572479,
-0.029723389074206352,
-0.16705724596977234,
0.17462614178657532,
0.12759019434452057,
-0.10856198519468307,
-0.0864260271191597,
0.005113325547426939,
-0.04779781401157379,
-0.04090845212340355,
-0.027647849172353745,
0.0039734807796776295,
0.16823381185531616,
-0.005211552605032921,
0.1309678703546524,
-0.06625892221927643,
-0.03226790577173233,
0.026831064373254776,
-0.03274940326809883,
0.0034818495623767376,
0.10325685143470764,
0.06224943697452545,
-0.13052484393119812,
0.12178691476583481,
0.13003526628017426,
-0.08112829178571701,
0.13862481713294983,
-0.03621979430317879,
-0.07957478612661362,
-0.022887621074914932,
0.01055555697530508,
0.021257350221276283,
0.15305735170841217,
-0.12222559750080109,
-0.0025697508826851845,
-0.010252883657813072,
0.014496956020593643,
0.024043487384915352,
-0.24203525483608246,
-0.024763213470578194,
0.035998497158288956,
-0.04057192802429199,
0.020959168672561646,
-0.038467343896627426,
-0.0015834481455385685,
0.09213308990001678,
-0.0024858431424945593,
-0.06705956906080246,
0.025140604004263878,
-0.00882559921592474,
-0.07712706923484802,
0.20236656069755554,
-0.07685903459787369,
-0.1616036295890808,
-0.11464221775531769,
-0.040862761437892914,
-0.018601831048727036,
0.018747376278042793,
0.051995985209941864,
-0.07457833737134933,
-0.03831063210964203,
-0.10004913806915283,
0.0003752573684323579,
0.017992885783314705,
0.0305629875510931,
0.008061020635068417,
-0.009599046781659126,
0.10755575448274612,
-0.09234176576137543,
0.007203348912298679,
-0.04360175505280495,
-0.05002470687031746,
0.0483013354241848,
0.03738832101225853,
0.1488855481147766,
0.15352371335029602,
-0.02916104719042778,
0.010140052996575832,
-0.02928842417895794,
0.21612174808979034,
-0.0976305678486824,
-0.02340233139693737,
0.10344256460666656,
-0.023832708597183228,
0.042424995452165604,
0.12667927145957947,
0.051440734416246414,
-0.11198431998491287,
0.03150564804673195,
0.043209195137023926,
-0.03208896890282631,
-0.17221787571907043,
-0.019949914887547493,
-0.04803793132305145,
-0.048327989876270294,
0.08206740766763687,
0.02802538499236107,
-0.0008942246786318719,
0.05735548585653305,
0.031092900782823563,
0.04448022320866585,
-0.0253420639783144,
0.07322520762681961,
0.10398966819047928,
0.030310748144984245,
0.11700507998466492,
-0.05444059893488884,
-0.0687091276049614,
0.020068857818841934,
-0.00820763036608696,
0.21924081444740295,
0.012057531625032425,
0.12619730830192566,
0.04767535626888275,
0.16589564085006714,
0.015323871746659279,
0.06331706792116165,
0.011292099952697754,
-0.05700714886188507,
-0.0022748326882719994,
-0.04259679839015007,
-0.03355953097343445,
0.02060925029218197,
-0.04896857216954231,
0.04783744364976883,
-0.09149150550365448,
-0.0035083310212939978,
0.057057008147239685,
0.25319212675094604,
0.061079833656549454,
-0.35403233766555786,
-0.08088039606809616,
0.005633076652884483,
-0.02010050043463707,
-0.04128461703658104,
0.003323439508676529,
0.15870147943496704,
-0.05341213196516037,
0.03566182777285576,
-0.09629230201244354,
0.08491326123476028,
-0.055586379021406174,
0.03753342106938362,
0.05503574386239052,
0.1058427020907402,
-0.007805925328284502,
0.040212057530879974,
-0.2861754298210144,
0.2845037281513214,
0.014344538561999798,
0.08763706684112549,
-0.03299481421709061,
-0.022911736741662025,
0.027014991268515587,
0.07216005027294159,
0.05327242985367775,
-0.022186722606420517,
-0.07040370255708694,
-0.22715462744235992,
-0.05086050555109978,
0.040153082460165024,
0.13452422618865967,
-0.0015674817841500044,
0.12030500918626785,
-0.024376217275857925,
0.00567314587533474,
0.08461311459541321,
-0.04441603645682335,
-0.09788387268781662,
-0.08159653842449188,
-0.016890985891222954,
0.022945912554860115,
0.013424293138086796,
-0.0880521610379219,
-0.10124465078115463,
-0.08602042496204376,
0.14252182841300964,
-0.02815914899110794,
0.008762942627072334,
-0.1308450549840927,
0.10583186894655228,
0.08163801580667496,
-0.06836647540330887,
0.045815449208021164,
0.024554265663027763,
0.09358450770378113,
0.043813545256853104,
-0.05998286232352257,
0.145164355635643,
-0.050931770354509354,
-0.14460042119026184,
-0.058110203593969345,
0.0847473070025444,
0.0305893886834383,
0.04094202071428299,
-0.009422151371836662,
0.010137919336557388,
0.008708136156201363,
-0.07312419265508652,
0.0356016606092453,
-0.008337892591953278,
0.03092416375875473,
0.030714698135852814,
-0.026061397045850754,
0.027146359905600548,
-0.05183299630880356,
-0.016289122402668,
0.15011300146579742,
0.2442316859960556,
-0.07761524617671967,
-0.03546185791492462,
0.04336864501237869,
-0.062012091279029846,
-0.2015238106250763,
0.10689664632081985,
0.044990845024585724,
0.015022565610706806,
0.048871710896492004,
-0.12757709622383118,
0.08197169005870819,
0.09617694467306137,
-0.013231167569756508,
0.10725422948598862,
-0.3064803183078766,
-0.13114866614341736,
0.0981326624751091,
0.19891807436943054,
0.05827941372990608,
-0.16270388662815094,
-0.020468097180128098,
-0.023075593635439873,
-0.10415970534086227,
0.061288535594940186,
-0.10666537284851074,
0.13160675764083862,
0.0042405338026583195,
0.03755100816488266,
0.006300531327724457,
-0.06554234772920609,
0.1241217702627182,
-0.0443265475332737,
0.13665276765823364,
-0.05853600800037384,
0.022086672484874725,
0.08129802346229553,
-0.05611656233668327,
-0.002854971680790186,
-0.026429153978824615,
0.03348907083272934,
-0.04392799362540245,
-0.029404548928141594,
-0.061555974185466766,
0.01081850565969944,
-0.01921248435974121,
-0.05458410084247589,
-0.037442877888679504,
0.040867988020181656,
0.03804808109998703,
-0.014440969564020634,
0.14722305536270142,
-0.026917658746242523,
0.1166350394487381,
0.10350486636161804,
0.07221060991287231,
-0.056991517543792725,
-0.05593037232756615,
0.003082921262830496,
-0.014969908632338047,
0.05823058262467384,
-0.1295759081840515,
0.034993819892406464,
0.143244668841362,
0.008164905942976475,
0.14594751596450806,
0.07432552427053452,
-0.047023698687553406,
0.048125799745321274,
0.06297194212675095,
-0.13696186244487762,
-0.1631390005350113,
0.007374597247689962,
-0.02846340276300907,
-0.09081101417541504,
0.06795395165681839,
0.11203105747699738,
-0.0805635079741478,
0.0169574823230505,
-0.013363167643547058,
-0.010826402343809605,
-0.053499020636081696,
0.20047491788864136,
0.05162256583571434,
0.04581476002931595,
-0.08549986779689789,
0.08155923336744308,
0.03258148208260536,
-0.12292671203613281,
0.012089408002793789,
0.06956841051578522,
-0.07174190133810043,
-0.03185640275478363,
0.09035427868366241,
0.20725448429584503,
-0.03746733441948891,
-0.04807793349027634,
-0.14325660467147827,
-0.11562685668468475,
0.05478733777999878,
0.2275850474834442,
0.07826269418001175,
0.018408047035336494,
-0.03381740674376488,
0.03537794202566147,
-0.12435929477214813,
0.07932272553443909,
0.03770646080374718,
0.08999034762382507,
-0.16124853491783142,
0.16730263829231262,
0.007117726374417543,
0.014362613670527935,
-0.03709263354539871,
0.028066791594028473,
-0.1206761971116066,
0.016363387927412987,
-0.15098074078559875,
-0.01754303090274334,
-0.015512308105826378,
0.0024879449047148228,
0.006529727950692177,
-0.06880830973386765,
-0.0719861313700676,
0.017408179119229317,
-0.10570078343153,
-0.034286387264728546,
0.04521540552377701,
0.01840723305940628,
-0.11162008345127106,
-0.043288469314575195,
0.018534008413553238,
-0.06341665983200073,
0.05524645373225212,
0.020632734522223473,
0.01263673510402441,
0.03864631429314613,
-0.18620453774929047,
0.0115232327952981,
0.08573459088802338,
-0.0160340778529644,
0.05074542388319969,
-0.057115357369184494,
-0.02101861499249935,
-0.007388032041490078,
0.07920742779970169,
0.008385022170841694,
0.08794784545898438,
-0.11599715054035187,
-0.019539974629878998,
-0.06089319288730621,
-0.054196860641241074,
-0.05633436143398285,
0.0368548221886158,
0.06547028571367264,
0.037025436758995056,
0.16666828095912933,
-0.10440348088741302,
0.022368036210536957,
-0.2181624472141266,
-0.006726088933646679,
-0.001849491149187088,
-0.10865401476621628,
-0.05519206076860428,
-0.056682027876377106,
0.07793004810810089,
-0.05789424851536751,
0.11441142857074738,
-0.009217704646289349,
0.07133503258228302,
0.03757898136973381,
-0.0679977610707283,
0.024263393133878708,
0.04630620777606964,
0.22414670884609222,
0.01581418700516224,
-0.03895511105656624,
0.06910424679517746,
0.07165698707103729,
0.11379416286945343,
0.13287749886512756,
0.16546481847763062,
0.17343631386756897,
-0.04577787220478058,
0.12038253992795944,
0.04305446892976761,
-0.038916878402233124,
-0.11352583020925522,
0.06645499914884567,
-0.05651411786675453,
0.09194548428058624,
-0.03441952168941498,
0.17782796919345856,
0.11626186966896057,
-0.15708371996879578,
0.023406269028782845,
-0.04406638443470001,
-0.0955117866396904,
-0.06940248608589172,
-0.04289810359477997,
-0.10104671120643616,
-0.15884040296077728,
0.0032463413663208485,
-0.10454633831977844,
0.018964160233736038,
0.12556084990501404,
0.017234576866030693,
-0.008339956402778625,
0.22514908015727997,
0.06266986578702927,
0.02902568317949772,
0.06563392281532288,
0.011773304082453251,
-0.017846938222646713,
-0.06780864298343658,
-0.08388052880764008,
0.04016866162419319,
-0.031046463176608086,
0.03421241417527199,
-0.043027542531490326,
-0.05191643536090851,
0.056177977472543716,
-0.008226368576288223,
-0.11385847628116608,
0.018108466640114784,
0.040741462260484695,
0.045343875885009766,
0.027674641460180283,
0.02492421492934227,
-0.011797336861491203,
-0.02869882434606552,
0.22759738564491272,
-0.08757159113883972,
-0.06007527559995651,
-0.10242222994565964,
0.2250511646270752,
0.011433582752943039,
0.02114136703312397,
-0.0042500849813222885,
-0.09235744178295135,
0.03553733602166176,
0.21074554324150085,
0.13893893361091614,
-0.1003798171877861,
-0.004784749820828438,
-0.018759731203317642,
-0.011990813538432121,
-0.054362230002880096,
0.10655473172664642,
0.11901441961526871,
0.0022242132108658552,
-0.09825449436903,
-0.06839735805988312,
-0.037434887140989304,
-0.019428875297307968,
-0.028491947799921036,
0.004324212204664946,
0.054087307304143906,
0.03618580847978592,
-0.07029379904270172,
0.0732407495379448,
-0.029360134154558182,
-0.1000271737575531,
0.0915641337633133,
-0.18185828626155853,
-0.14060834050178528,
-0.020217182114720345,
0.1203862726688385,
-0.020664462819695473,
0.04655981436371803,
-0.04350212961435318,
0.008343071676790714,
0.03563931956887245,
-0.01630849391222,
-0.05906345695257187,
-0.09133142977952957,
0.045043110847473145,
-0.12832680344581604,
0.1976548731327057,
-0.05023111402988434,
0.017719581723213196,
0.13495637476444244,
0.050457485020160675,
-0.08150410652160645,
0.09464934468269348,
0.023110169917345047,
-0.09962882101535797,
0.023426640778779984,
0.1268542855978012,
-0.03781186789274216,
0.10256791114807129,
0.04860052093863487,
-0.1478590965270996,
0.036051083356142044,
-0.10695310682058334,
-0.03912649676203728,
-0.03060220368206501,
-0.06105760857462883,
-0.062053047120571136,
0.1347409039735794,
0.1851874440908432,
-0.023166963830590248,
0.036340419203042984,
-0.06367196142673492,
0.010864805430173874,
0.05691659078001976,
0.08258628845214844,
-0.03850019350647926,
-0.2471030056476593,
0.02357967011630535,
0.06132473796606064,
-0.037201859056949615,
-0.2526988387107849,
-0.10762322694063187,
0.007633021101355553,
-0.06390713900327682,
-0.06389429420232773,
0.10267697274684906,
0.10433375090360641,
0.06602808088064194,
-0.0641484409570694,
-0.1470121592283249,
-0.05429985374212265,
0.17953786253929138,
-0.136148601770401,
-0.08687085658311844
] |
null | null | transformers | # MiquMaid-v2-70B 2.4bpw
## Description
Exllama quant of [NeverSleep/MiquMaid-v2-70B](https://huggingface.co/NeverSleep/MiquMaid-v2-70B)
## Other quants:
EXL2: [4bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-4bpw-exl2), [3.5bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-3.5bpw-exl2), [3bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-3bpw-exl2), [2.4bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-2.4bpw-exl2), [2.3bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-2.3bpw-exl2)
2.4bpw is probably the most you can fit in a 24gb card
GGUF:
[2bit Imatrix GGUF](https://huggingface.co/Kooten/MiquMaid-v2-70B-Imatrix-GGUF)
## Prompt format: Alpaca
```
### Instruction:
{system prompt}
### Input:
{input}
### Response:
{reply}
```
## Contact
Kooten on discord
[ko-fi.com/kooten](https://ko-fi.com/kooten) | {"license": "cc-by-nc-4.0", "tags": ["not-for-all-audiences", "nsfw"]} | text-generation | Kooten/MiquMaid-v2-70B-2.4bpw-exl2 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"not-for-all-audiences",
"nsfw",
"conversational",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T06:55:07+00:00 | [] | [] | TAGS
#transformers #safetensors #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # MiquMaid-v2-70B 2.4bpw
## Description
Exllama quant of NeverSleep/MiquMaid-v2-70B
## Other quants:
EXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw
2.4bpw is probably the most you can fit in a 24gb card
GGUF:
2bit Imatrix GGUF
## Prompt format: Alpaca
## Contact
Kooten on discord
URL | [
"# MiquMaid-v2-70B 2.4bpw",
"## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B",
"## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF",
"## Prompt format: Alpaca",
"## Contact\nKooten on discord\n\nURL"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# MiquMaid-v2-70B 2.4bpw",
"## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B",
"## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF",
"## Prompt format: Alpaca",
"## Contact\nKooten on discord\n\nURL"
] | [
75,
14,
21,
60,
8,
7
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# MiquMaid-v2-70B 2.4bpw## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF## Prompt format: Alpaca## Contact\nKooten on discord\n\nURL"
] | [
-0.05536721646785736,
0.041095953434705734,
-0.005616066977381706,
0.08649308234453201,
0.025252895429730415,
0.023109616711735725,
0.14957532286643982,
0.12387951463460922,
0.002293301746249199,
0.011250326409935951,
0.04677974805235863,
0.11005803942680359,
0.04742090031504631,
0.16318507492542267,
-0.05459979176521301,
-0.09710223227739334,
0.028902357444167137,
0.03199230134487152,
0.07678759098052979,
0.11509665846824646,
0.0652657151222229,
-0.031223038211464882,
0.055522602051496506,
-0.0451180599629879,
-0.10840820521116257,
-0.009847505949437618,
-0.025824926793575287,
-0.06644352525472641,
0.04780501872301102,
0.08139406144618988,
0.04041016846895218,
0.10467026382684708,
-0.032140057533979416,
-0.1826706975698471,
0.027486702427268028,
0.03882202133536339,
-0.060944296419620514,
0.027478037402033806,
0.029945330694317818,
-0.08381341397762299,
0.005503260530531406,
0.005256712902337313,
-0.06775756925344467,
0.0728074386715889,
-0.09106133133172989,
-0.07724066078662872,
-0.04898356273770332,
0.11664412170648575,
0.0732918456196785,
0.02732684835791588,
-0.0029023545794188976,
0.09513900429010391,
-0.010654107667505741,
0.08181695640087128,
0.2301764190196991,
-0.25752493739128113,
-0.019602181389927864,
0.08651480078697205,
0.02727694623172283,
0.09409356862306595,
-0.07355023175477982,
0.031874947249889374,
0.024884207174181938,
-0.028935538604855537,
0.08012184500694275,
-0.020935194566845894,
0.060911208391189575,
-0.03262222558259964,
-0.08878209441900253,
0.009033609181642532,
0.14562898874282837,
0.03855307027697563,
-0.06249652057886124,
-0.04950924590229988,
-0.08129836618900299,
-0.10791239887475967,
-0.0494922436773777,
0.044517528265714645,
0.010665844194591045,
-0.011605054140090942,
-0.04463577643036842,
-0.001669973018579185,
-0.1074347123503685,
-0.04725448787212372,
-0.12969157099723816,
0.17361536622047424,
0.007061317563056946,
0.02972109243273735,
0.009701378643512726,
0.034560903906822205,
-0.08602398633956909,
-0.08700334280729294,
-0.08839686959981918,
-0.05994725972414017,
0.06753434985876083,
0.0010367853101342916,
0.006114479154348373,
0.08255401253700256,
0.08921174705028534,
0.14502617716789246,
-0.06975319981575012,
0.08816999942064285,
0.0072898464277386665,
0.0645846351981163,
-0.030989572405815125,
-0.11975692957639694,
-0.01658794842660427,
-0.05386064946651459,
0.05345587804913521,
0.03026391752064228,
0.08435775339603424,
-0.022466639056801796,
-0.10252059251070023,
0.0038177690003067255,
-0.0011503593996167183,
0.04030900448560715,
0.018500329926609993,
0.05614693462848663,
-0.04967248812317848,
-0.024317726492881775,
0.18618515133857727,
-0.06426923722028732,
0.020163534209132195,
0.058442723006010056,
0.027914810925722122,
0.04695424810051918,
0.06677166372537613,
-0.010623931884765625,
-0.02346028946340084,
0.05120179057121277,
-0.030490946024656296,
0.016006628051400185,
-0.01736266165971756,
-0.05878130719065666,
0.08366920799016953,
-0.04083693027496338,
-0.021365264430642128,
-0.18543629348278046,
-0.150686576962471,
0.08882453292608261,
-0.0062015000730752945,
-0.045240797102451324,
0.01620103046298027,
0.0071777761913836,
-0.032353565096855164,
-0.00480873417109251,
-0.0516539104282856,
0.017305167391896248,
-0.07674531638622284,
0.0873623862862587,
0.08716695010662079,
0.13274893164634705,
-0.11471287161111832,
-0.017906714230775833,
-0.04576783999800682,
0.0479239821434021,
-0.08411931991577148,
-0.0023155142553150654,
-0.12088300287723541,
-0.029959965497255325,
-0.06598440557718277,
-0.02909747324883938,
-0.050441011786460876,
0.04511846974492073,
-0.009161867201328278,
0.09390594810247421,
-0.07042858004570007,
-0.043072931468486786,
0.263116717338562,
-0.14608481526374817,
-0.10752486437559128,
0.11411112546920776,
0.05302312970161438,
-0.06271703541278839,
0.04696505144238472,
0.08402395248413086,
0.1350938379764557,
-0.11118502169847488,
-0.10385719686746597,
0.06978544592857361,
-0.007902256213128567,
-0.09109405428171158,
0.08950591832399368,
0.079718679189682,
0.009787121787667274,
0.028866970911622047,
0.13447947800159454,
0.025212706997990608,
0.009671151638031006,
-0.05845178663730621,
-0.06501512229442596,
-0.08639383316040039,
0.08379844576120377,
-0.016913505271077156,
-0.03964821249246597,
-0.124032162129879,
-0.09659655392169952,
0.013912136666476727,
0.07083723694086075,
-0.01834043115377426,
-0.016896579414606094,
-0.1400366872549057,
0.11175001412630081,
-0.11024748533964157,
0.028532136231660843,
-0.04844516143202782,
-0.02834445796906948,
-0.0352526493370533,
-0.02491305209696293,
0.049441978335380554,
0.06916160136461258,
0.0932871550321579,
0.07237327843904495,
-0.0990065410733223,
0.03370900824666023,
0.12256928533315659,
0.025604678317904472,
-0.005074118729680777,
-0.08503341674804688,
0.05543145537376404,
-0.024100495502352715,
0.14415985345840454,
-0.09427737444639206,
0.046571604907512665,
0.15622691810131073,
0.0870295837521553,
0.004510888829827309,
-0.043034330010414124,
0.06763316690921783,
0.01287285890430212,
-0.046841226518154144,
0.0033696831669658422,
0.07389360666275024,
-0.009126827120780945,
-0.009763652458786964,
0.11912830173969269,
-0.2580769658088684,
0.18420366942882538,
0.18209481239318848,
-0.024330563843250275,
-0.030786460265517235,
-0.053832389414310455,
0.01447885762900114,
-0.019493546336889267,
-0.019312487915158272,
-0.06303591281175613,
0.024796675890684128,
-0.023179780691862106,
0.13348381221294403,
-0.09997757524251938,
-0.048480063676834106,
0.05795455724000931,
-0.05554281175136566,
-0.008190243504941463,
0.04361358657479286,
0.07608220726251602,
-0.12148001044988632,
0.14940406382083893,
0.17931252717971802,
-0.07654894143342972,
0.12384331971406937,
0.025429682806134224,
-0.010422869585454464,
-0.034722715616226196,
0.04524153843522072,
0.03680000081658363,
0.07654599100351334,
-0.07120171189308167,
0.056843798607587814,
0.05700297653675079,
-0.014500269666314125,
0.031024441123008728,
-0.13973551988601685,
-0.032899536192417145,
0.04001512750983238,
-0.09117165207862854,
-0.06925716251134872,
0.002085962099954486,
-0.0112796351313591,
0.09668377786874771,
-0.05187634378671646,
0.01560357678681612,
0.02631830796599388,
0.0460829958319664,
-0.08109521120786667,
0.14490394294261932,
-0.09616557508707047,
-0.3102935254573822,
-0.08526556938886642,
-0.1270912140607834,
-0.05667433887720108,
0.02313552424311638,
0.0922485813498497,
-0.05256413668394089,
-0.04397611692547798,
-0.10422392189502716,
-0.064815454185009,
-0.005433007143437862,
-0.0030826725997030735,
0.01240744348615408,
-0.008915485814213753,
0.07399342209100723,
-0.08883367478847504,
-0.007996860891580582,
0.08231991529464722,
-0.048027679324150085,
0.16435542702674866,
-0.00881101656705141,
0.0906619057059288,
0.11754753440618515,
-0.0058190175332129,
-0.025890018790960312,
-0.03404988721013069,
0.12230055034160614,
-0.05423176288604736,
0.0675327330827713,
0.1270984560251236,
-0.013793401420116425,
0.06800294667482376,
0.08324204385280609,
0.031696733087301254,
-0.06287404894828796,
-0.05105035379528999,
0.01754540205001831,
-0.11981865018606186,
-0.13460318744182587,
-0.03978158161044121,
-0.042137742042541504,
0.15678246319293976,
0.04483442381024361,
0.0621379017829895,
0.045193325728178024,
0.12762127816677094,
-0.06624346971511841,
-0.014091335237026215,
0.06684915721416473,
0.08357810229063034,
0.12101621925830841,
0.024898014962673187,
0.15765763819217682,
-0.08555121719837189,
-0.015439176931977272,
0.1189781054854393,
0.11187269538640976,
0.027878981083631516,
-0.007739008404314518,
0.10362441092729568,
0.07023274898529053,
0.02216573804616928,
0.11389419436454773,
0.07293906062841415,
-0.0026692405808717012,
-0.04561365023255348,
-0.03359517827630043,
-0.06651657074689865,
0.00010636813385644928,
0.026901857927441597,
-0.003270574379712343,
-0.04068699851632118,
0.02800673432648182,
-0.0353844091296196,
0.0808243453502655,
0.026201311498880386,
0.082752525806427,
-0.21919570863246918,
-0.03887892886996269,
0.07809318602085114,
-0.00763695826753974,
-0.03921186551451683,
0.027920112013816833,
0.08535977452993393,
0.03226185590028763,
0.16677822172641754,
-0.018561331555247307,
0.07018299400806427,
-0.07795344293117523,
0.022597795352339745,
-0.035807445645332336,
0.06352697312831879,
-0.009894194081425667,
0.10098353773355484,
-0.2687614858150482,
0.05366746708750725,
0.039292577654123306,
0.013427561149001122,
-0.05176352337002754,
-0.025425001978874207,
0.012088422663509846,
-0.006062290631234646,
0.08725295960903168,
0.011655528098344803,
0.0701567679643631,
-0.13967463374137878,
-0.08597035706043243,
0.012713338248431683,
0.07530714571475983,
0.050019536167383194,
0.05061069130897522,
0.035620611160993576,
-0.05331692099571228,
-0.0025089101400226355,
-0.04147062078118324,
-0.04639279097318649,
-0.11067239195108414,
0.09835325926542282,
0.15504474937915802,
-0.03223725035786629,
-0.02371024526655674,
-0.02494705468416214,
-0.20577798783779144,
0.21850115060806274,
-0.1264043152332306,
-0.08909271657466888,
-0.04957837611436844,
-0.0830666720867157,
-0.0006954535492695868,
-0.02092759683728218,
0.03791984170675278,
-0.039343129843473434,
0.11668367683887482,
-0.03669549897313118,
-0.06828226894140244,
0.06385896354913712,
-0.12090959399938583,
-0.12632153928279877,
-0.03309705853462219,
0.14715027809143066,
-0.09887982904911041,
0.026402510702610016,
0.03144315257668495,
-0.008784758858382702,
-0.012836925685405731,
-0.1301671415567398,
0.0921098068356514,
0.022546909749507904,
-0.02057018131017685,
0.01844470575451851,
0.008513814769685268,
-0.04568218067288399,
-0.0036149458028376102,
-0.04835133254528046,
0.11729985475540161,
0.3371688425540924,
-0.03398217633366585,
0.10424253344535828,
0.11825858801603317,
-0.03994102030992508,
-0.23225033283233643,
-0.15137062966823578,
-0.08843457698822021,
-0.07970812916755676,
-0.026860609650611877,
-0.07846304774284363,
0.05339307710528374,
0.12656594812870026,
-0.04079795628786087,
0.20145714282989502,
-0.13187210261821747,
-0.11129818856716156,
0.018926244229078293,
0.10885350406169891,
0.2575282156467438,
-0.1771252155303955,
-0.06262843310832977,
-0.07589274644851685,
-0.17670992016792297,
0.18982526659965515,
-0.14434674382209778,
0.10623525083065033,
-0.02410399727523327,
0.05273497849702835,
-0.015451711602509022,
-0.07957937568426132,
0.11658160388469696,
-0.06524287909269333,
0.00028902519261464477,
-0.1274566650390625,
0.06916455179452896,
0.040632762014865875,
-0.019056769087910652,
0.01961738057434559,
-0.2317436933517456,
-0.010109981521964073,
-0.06661655008792877,
-0.055344197899103165,
-0.038204848766326904,
-0.0002934721123892814,
-0.04394688829779625,
-0.038108572363853455,
-0.04765348881483078,
-0.0105156684294343,
0.012947780080139637,
-0.03459939733147621,
-0.02979549579322338,
-0.06989263743162155,
0.09145428240299225,
0.30834758281707764,
0.09818128496408463,
-0.07260797917842865,
-0.06369854509830475,
-0.06322066485881805,
-0.09720383584499359,
0.008099596947431564,
-0.15025833249092102,
0.060442034155130386,
0.03833993524312973,
0.005611020606011152,
0.06505591422319412,
0.04103529825806618,
-0.03566521033644676,
0.026385359466075897,
0.15117418766021729,
-0.1420539915561676,
-0.025383643805980682,
0.005744248628616333,
0.10724203288555145,
0.006713702343404293,
0.016555843874812126,
0.14005059003829956,
-0.017725612968206406,
0.01692129112780094,
0.047448236495256424,
0.024330440908670425,
-0.0012325564166530967,
0.1179145872592926,
0.03661298751831055,
0.07786592841148376,
-0.11317276954650879,
0.06466946750879288,
-0.035433974117040634,
-0.042618557810783386,
-0.0035752446856349707,
0.0669107660651207,
-0.0884891152381897,
-0.1281280368566513,
0.006051603239029646,
-0.0466439351439476,
-0.05437345802783966,
-0.06234839931130409,
-0.036832451820373535,
-0.13813082873821259,
0.028203578665852547,
0.1642111986875534,
0.01278737187385559,
-0.02298760786652565,
0.08439895510673523,
0.03454098850488663,
-0.0774833932518959,
0.08501076698303223,
-0.018527282401919365,
0.09303252398967743,
-0.12543988227844238,
0.0490594319999218,
-0.0017908571753650904,
-0.00802757777273655,
-0.04654036462306976,
0.035804618149995804,
-0.10524250566959381,
-0.03469131141901016,
-0.14994244277477264,
0.10340278595685959,
-0.10387622565031052,
0.0026456303894519806,
-0.016384216025471687,
0.011940308846533298,
-0.02117007039487362,
-0.04224124923348427,
-0.06891471892595291,
0.010808376595377922,
0.02026774175465107,
-0.02000625617802143,
-0.10386817902326584,
-0.027497103437781334,
0.026054084300994873,
-0.030456237494945526,
0.05619943141937256,
0.02605430781841278,
-0.053291045129299164,
-0.007232055068016052,
-0.19211269915103912,
-0.01876760832965374,
0.08998364955186844,
0.05446423962712288,
-0.037845030426979065,
0.008032258599996567,
0.044298309832811356,
0.051243387162685394,
0.026726284995675087,
0.03709058463573456,
0.09336244314908981,
-0.0683683529496193,
0.02638864517211914,
-0.07996077835559845,
-0.06395740061998367,
-0.03317223861813545,
-0.013171014375984669,
0.13708969950675964,
-0.06854964047670364,
0.1150956004858017,
-0.0704563558101654,
-0.010517766699194908,
-0.13622696697711945,
0.02516622468829155,
0.03330613300204277,
-0.1701171100139618,
-0.04276920109987259,
-0.03287394717335701,
0.022367076948285103,
-0.0348806232213974,
0.17920146882534027,
-0.08072645217180252,
-0.1335446983575821,
0.030534835532307625,
-0.08866804093122482,
-0.029492376372218132,
-0.008332783356308937,
0.22265055775642395,
0.06294985115528107,
-0.006019707769155502,
-0.11789993941783905,
0.016651177778840065,
0.07495841383934021,
-0.1254684031009674,
0.08059800416231155,
0.10918678343296051,
-0.08126256614923477,
0.05238568037748337,
0.05360392853617668,
-0.06261157244443893,
-0.06089036166667938,
0.011257740668952465,
-0.1587831974029541,
0.07830321043729782,
-0.0011493070051074028,
0.07738592475652695,
0.11882204562425613,
-0.03659709542989731,
0.0018255503382533789,
-0.05900957062840462,
-0.023674357682466507,
-0.11977118998765945,
-0.05724676698446274,
-0.09749093651771545,
-0.08958438783884048,
0.01630418561398983,
-0.06845877319574356,
0.011070778593420982,
0.05486046522855759,
0.05751694366335869,
0.021000977605581284,
0.14155134558677673,
-0.020942477509379387,
0.0036805416457355022,
-0.010742001235485077,
-0.024185391142964363,
-0.031032098457217216,
0.06429726630449295,
-0.05598151311278343,
0.06568493694067001,
-0.02417454868555069,
0.03675103187561035,
0.03147448971867561,
0.039646562188863754,
0.015576764941215515,
-0.06811962276697159,
-0.09642806649208069,
-0.041563015431165695,
0.035806577652692795,
0.06705813109874725,
0.12257827818393707,
-0.026944924145936966,
0.00045283016515895724,
0.019346704706549644,
0.08214554935693741,
-0.0193247739225626,
-0.10565540194511414,
-0.015138130635023117,
0.15201792120933533,
-0.08618732541799545,
0.06145408749580383,
-0.01939547061920166,
-0.027003366500139236,
0.023551374673843384,
0.2431560903787613,
0.16405384242534637,
-0.03568430617451668,
0.05659674108028412,
0.015377906151115894,
0.01920650340616703,
0.06692832708358765,
0.08893518894910812,
0.06200776249170303,
0.17942148447036743,
-0.027044229209423065,
-0.0025605312548577785,
0.04211198911070824,
-0.003506310749799013,
-0.07788021862506866,
0.06274110823869705,
-0.024094613268971443,
0.013935434632003307,
-0.06922365725040436,
0.0004747009661514312,
-0.07316520810127258,
0.0364312119781971,
0.0036987541243433952,
-0.10834040492773056,
-0.05637107044458389,
0.0036430535838007927,
0.08282823860645294,
0.03449172526597977,
-0.018016962334513664,
-0.021460313349962234,
-0.015466035343706608,
-0.017708798870444298,
-0.015394114889204502,
-0.16913895308971405,
0.06477700173854828,
-0.03531317412853241,
-0.00013111496809870005,
0.060975782573223114,
-0.03543916344642639,
0.13325513899326324,
0.10884106904268265,
-0.026040516793727875,
-0.060108862817287445,
0.17411290109157562,
0.006369365844875574,
-0.17572316527366638,
0.041393380612134933,
0.12910199165344238,
-0.043854229152202606,
0.02587788552045822,
0.09135317802429199,
-0.05597565323114395,
-0.046600356698036194,
0.09534341841936111,
-0.06995607912540436,
-0.011892503127455711,
0.03803428262472153,
-0.03523196652531624,
0.07724712789058685,
0.05606958642601967,
-0.03979726508259773,
-0.044405288994312286,
-0.015402681194245815,
0.07267772406339645,
-0.03761781007051468,
-0.056407686322927475,
-0.01667267270386219,
-0.18295541405677795,
-0.04121660068631172,
0.07302749156951904,
0.13005025684833527,
-0.3128089904785156,
-0.012653334997594357,
-0.08699717372655869,
0.0002521966816857457,
-0.1468711495399475,
0.05406820401549339,
0.18560557067394257,
0.0314750075340271,
-0.06653007864952087,
-0.11481979489326477,
-0.015697697177529335,
0.09590569883584976,
-0.09372441470623016,
-0.08272730559110641
] |
null | null | null |
见[nenekochan/Yi-6B-yoruno](https://huggingface.co/nenekochan/Yi-6B-yoruno)
| {"language": ["zh"], "license": "cc-by-nc-4.0", "tags": ["not-for-all-audiences"], "base_model": "nenekochan/Yi-6B-yoruno", "inference": false} | null | nenekochan/Yi-6B-yoruno-GGUF | [
"gguf",
"not-for-all-audiences",
"zh",
"base_model:nenekochan/Yi-6B-yoruno",
"license:cc-by-nc-4.0",
"region:us"
] | 2024-02-08T06:56:27+00:00 | [] | [
"zh"
] | TAGS
#gguf #not-for-all-audiences #zh #base_model-nenekochan/Yi-6B-yoruno #license-cc-by-nc-4.0 #region-us
|
见nenekochan/Yi-6B-yoruno
| [] | [
"TAGS\n#gguf #not-for-all-audiences #zh #base_model-nenekochan/Yi-6B-yoruno #license-cc-by-nc-4.0 #region-us \n"
] | [
47
] | [
"passage: TAGS\n#gguf #not-for-all-audiences #zh #base_model-nenekochan/Yi-6B-yoruno #license-cc-by-nc-4.0 #region-us \n"
] | [
-0.026661058887839317,
0.0905449315905571,
-0.005499874707311392,
0.038610197603702545,
-0.005202499218285084,
0.07174187898635864,
0.22486838698387146,
0.05639887973666191,
0.1400977075099945,
0.003241452621296048,
0.0909249559044838,
-0.006399211473762989,
0.019856425002217293,
-0.030080070719122887,
-0.009976125322282314,
-0.12685169279575348,
0.04288376867771149,
0.07796989381313324,
0.09845148026943207,
0.0309718307107687,
0.03857336938381195,
-0.019862109795212746,
0.01401637215167284,
0.004860697314143181,
-0.10469396412372589,
0.0018838506657630205,
0.0067830015905201435,
-0.04017762467265129,
0.037337180227041245,
0.03407203406095505,
0.043828085064888,
0.12185635417699814,
-0.031374696642160416,
-0.11537811160087585,
0.0291490126401186,
-0.053157951682806015,
-0.1976330727338791,
0.03612598031759262,
0.02017715573310852,
0.01826580800116062,
0.23006747663021088,
0.0911717638373375,
-0.16486158967018127,
0.0654609277844429,
-0.2088834047317505,
-0.10143602639436722,
-0.07975738495588303,
0.18595880270004272,
0.01359429582953453,
0.041089512407779694,
-0.00667526014149189,
0.07151222974061966,
-0.17636239528656006,
-0.05103171616792679,
0.10488519817590714,
-0.29815059900283813,
-0.00728545431047678,
0.21301932632923126,
-0.01748448796570301,
0.044581614434719086,
-0.11725503206253052,
0.09835321456193924,
0.08063114434480667,
-0.06056372448801994,
-0.15304574370384216,
-0.040969640016555786,
0.2108604609966278,
0.049139171838760376,
-0.0383366197347641,
0.008419893682003021,
0.21047241985797882,
0.11122854799032211,
-0.009157773107290268,
0.09481319785118103,
-0.01377552654594183,
0.02803863026201725,
-0.06482034176588058,
0.05682259425520897,
0.012581110000610352,
0.12731367349624634,
0.0052676997147500515,
-0.07667410373687744,
-0.1225057989358902,
-0.07351648062467575,
-0.11436698585748672,
0.14758571982383728,
-0.009536621160805225,
0.06470245122909546,
-0.07685625553131104,
0.024715278297662735,
-0.24723686277866364,
-0.0869554802775383,
-0.08101911842823029,
-0.05918420851230621,
0.08827326446771622,
-0.0007937372429296374,
0.022842541337013245,
0.17095695436000824,
0.1741706281900406,
0.13926617801189423,
-0.04476363584399223,
0.019486404955387115,
0.008808927610516548,
0.135541170835495,
0.0016773805255070329,
-0.09378059208393097,
-0.03658295050263405,
0.1345628798007965,
0.0695289745926857,
-0.036174606531858444,
0.07106468081474304,
0.0021805341821163893,
-0.13379448652267456,
-0.05267338827252388,
-0.13005603849887848,
0.05568287894129753,
0.017369281500577927,
-0.021262124180793762,
-0.01999037340283394,
0.026293544098734856,
0.23151369392871857,
0.10380128026008606,
-0.06745492666959763,
0.049029961228370667,
-0.005267901811748743,
-0.02891944721341133,
-0.1149829775094986,
0.08904310315847397,
0.1251569241285324,
0.043013863265514374,
-0.15220907330513,
-0.01408824510872364,
0.06551570445299149,
0.11560161411762238,
0.11206263303756714,
-0.04929257929325104,
0.10311944037675858,
-0.12166562676429749,
-0.08785492181777954,
0.04085040092468262,
-0.03446623682975769,
-0.03692832589149475,
0.004453662782907486,
0.036295246332883835,
0.02493511699140072,
-0.04771417751908302,
-0.04774636775255203,
-0.06232358515262604,
-0.10095665603876114,
0.10277644544839859,
0.04902392625808716,
-0.016546420753002167,
-0.11685749888420105,
-0.03908678516745567,
-0.0821014791727066,
0.07636188715696335,
0.045898810029029846,
-0.04174934700131416,
-0.1211678758263588,
0.05407613888382912,
-0.0023194255772978067,
0.00018976219871547073,
-0.0975629985332489,
-0.0018189763650298119,
-0.0306499432772398,
0.20159748196601868,
-0.027129871770739555,
-0.04040931537747383,
0.09808436781167984,
-0.11094538122415543,
-0.07970987260341644,
0.06117856875061989,
0.042268041521310806,
-0.05169067159295082,
0.03676401078701019,
0.3068903088569641,
-0.008077288046479225,
-0.15381264686584473,
-0.06724463403224945,
0.10840731859207153,
-0.0808788537979126,
-0.08445753157138824,
0.121327705681324,
-0.03874512389302254,
-0.15615206956863403,
0.02419568970799446,
-0.11994924396276474,
0.05941889435052872,
-0.04036537930369377,
-0.09100130200386047,
0.016510823741555214,
-0.0679360032081604,
0.127387136220932,
0.032543379813432693,
0.08602853864431381,
-0.025932367891073227,
-0.02973187156021595,
-0.12881195545196533,
0.12668156623840332,
0.09684424102306366,
-0.015105435624718666,
-0.11280577629804611,
0.13635538518428802,
-0.06436046957969666,
-0.013483908027410507,
0.04791618511080742,
0.027664965018630028,
-0.0220036581158638,
0.008636685088276863,
0.13300952315330505,
0.1241435632109642,
0.022716602310538292,
-0.04807642474770546,
-0.08935657143592834,
-0.009193171747028828,
-0.031091133132576942,
0.0349084697663784,
0.06612925976514816,
-0.11635849624872208,
0.07438730448484421,
0.01954551227390766,
0.09960521757602692,
-0.15265604853630066,
-0.034250251948833466,
0.16349266469478607,
0.00574432173743844,
-0.11062336713075638,
0.032848335802555084,
0.06740915775299072,
-0.02150426246225834,
0.04194719344377518,
0.0290321484208107,
0.11630403995513916,
0.06008199229836464,
-0.12810783088207245,
0.07241366803646088,
-0.017202267423272133,
0.10643521696329117,
0.12067938596010208,
-0.017499348148703575,
0.02595360390841961,
-0.03280487656593323,
-0.020982617512345314,
0.001649766112677753,
0.10040029883384705,
0.11546909809112549,
0.0507226325571537,
-0.09183809161186218,
0.012097843922674656,
-0.05983062833547592,
0.04397342726588249,
0.051516350358724594,
-0.10037573426961899,
-0.05465434864163399,
0.06144353374838829,
0.17306728661060333,
-0.19044536352157593,
0.17060449719429016,
0.19386176764965057,
-0.013772871345281601,
0.2123526632785797,
-0.04967089742422104,
0.027834312990307808,
-0.09431105107069016,
0.028186162933707237,
-0.03596002981066704,
0.23033596575260162,
-0.14952296018600464,
0.04859757423400879,
0.010916237719357014,
-0.020447324961423874,
0.09724763035774231,
-0.14625166356563568,
-0.14984917640686035,
-0.06238069757819176,
-0.079392209649086,
-0.12635470926761627,
0.09126511961221695,
-0.15545563399791718,
0.03032379038631916,
0.015828024595975876,
-0.026016544550657272,
0.12650802731513977,
0.011236662045121193,
-0.08762532472610474,
0.059003084897994995,
-0.08304446935653687,
-0.004373059142380953,
-0.027511827647686005,
-0.031235219910740852,
-0.12375036627054214,
0.04545636102557182,
0.057923056185245514,
-0.11878606677055359,
0.017867466434836388,
0.05466768145561218,
-0.13278046250343323,
-0.0686059296131134,
-0.052105240523815155,
0.022525770589709282,
0.024896832183003426,
-0.02848203480243683,
-0.07032331079244614,
-0.028909048065543175,
-0.11178658902645111,
-0.11967265605926514,
0.06995145231485367,
-0.08633268624544144,
0.11441193521022797,
0.09629704803228378,
0.12149744480848312,
0.03774239122867584,
-0.02796490676701069,
0.16805781424045563,
-0.1321011185646057,
-0.029595842584967613,
0.13962602615356445,
0.08647768199443817,
0.025839392095804214,
0.13340508937835693,
0.09326566755771637,
-0.04488053545355797,
-0.056705981492996216,
-0.0027672876603901386,
-0.1114281490445137,
-0.1256190538406372,
-0.05204794555902481,
-0.09271349012851715,
0.06291759014129639,
-0.011651215143501759,
0.08243006467819214,
0.23220595717430115,
0.11066873371601105,
-0.05443481355905533,
-0.05547109246253967,
-0.03825024515390396,
0.02208593487739563,
0.1373511254787445,
0.028532588854432106,
-0.035756878554821014,
-0.07845369726419449,
0.04840540513396263,
0.15764763951301575,
0.08804772049188614,
0.06232789158821106,
0.12820154428482056,
0.1149456799030304,
0.16313503682613373,
0.21821603178977966,
0.16214555501937866,
-0.09756730496883392,
0.013755854219198227,
-0.06864053010940552,
-0.013830685056746006,
-0.05591214448213577,
0.05375000834465027,
0.01960100419819355,
0.01741364225745201,
-0.169950932264328,
0.046584077179431915,
-0.24007022380828857,
0.0654020607471466,
-0.04400356113910675,
0.057500410825014114,
0.04626063257455826,
0.016395878046751022,
0.04473140090703964,
0.14239485561847687,
0.02084677293896675,
0.05999084934592247,
-0.038357749581336975,
-0.06758798658847809,
0.09334547072649002,
0.039611104875802994,
0.03767623007297516,
0.05430474877357483,
0.026433797553181648,
-0.11891042441129684,
-0.04585516080260277,
-0.011896715499460697,
0.09501640498638153,
-0.2166971117258072,
0.1964467316865921,
0.03969038650393486,
0.03326286002993584,
-0.034602005034685135,
-0.12079410254955292,
0.06363777071237564,
0.1425143927335739,
0.169012188911438,
0.07088860124349594,
-0.07989274710416794,
-0.02739618346095085,
-0.03849596530199051,
0.053144946694374084,
0.014710243791341782,
-0.06289221346378326,
-0.11391530930995941,
0.046084072440862656,
0.04388737678527832,
0.019355474039912224,
0.09243408590555191,
-0.20794092118740082,
-0.03113741986453533,
0.0893375426530838,
0.05467235669493675,
0.008245639503002167,
-0.030395207926630974,
0.0160411074757576,
-0.05996938422322273,
0.05376231297850609,
-0.09749650210142136,
-0.027427755296230316,
-0.05265957862138748,
-0.0744168683886528,
0.041309986263513565,
-0.059480708092451096,
-0.023321695625782013,
-0.04928077757358551,
-0.0677403137087822,
-0.09655467420816422,
-0.1565091907978058,
0.06070232018828392,
-0.04980230703949928,
-0.10403314977884293,
0.01008553709834814,
0.1310943365097046,
0.02337733469903469,
0.058295540511608124,
0.029515203088521957,
-0.011434623971581459,
0.022450122982263565,
-0.16824489831924438,
0.07772582024335861,
-0.17185847461223602,
-0.0335092730820179,
-0.0569579191505909,
0.06528645008802414,
-0.002031188691034913,
-0.043976906687021255,
-0.10550546646118164,
0.06156313046813011,
0.4705933630466461,
-0.016978669911623,
0.12363167107105255,
0.3306851089000702,
-0.04655063524842262,
-0.12314625084400177,
-0.13786643743515015,
-0.17917361855506897,
-0.11145101487636566,
-0.05313219875097275,
-0.171022430062294,
-0.026883797720074654,
0.18239498138427734,
-0.11136692017316818,
0.2559732496738434,
-0.17132392525672913,
-0.05928957089781761,
0.08912196010351181,
0.008909349329769611,
0.49022334814071655,
-0.13220027089118958,
-0.10490802675485611,
0.04769153147935867,
-0.11403864622116089,
0.11345809698104858,
-0.046922408044338226,
0.09498821198940277,
-0.03156019374728203,
-0.04102982208132744,
-0.0032049762085080147,
0.0020673677790910006,
0.172369122505188,
0.021981066092848778,
0.04972310736775398,
-0.07245832681655884,
-0.1673697680234909,
0.17949925363063812,
0.023861346766352654,
-0.0814521536231041,
-0.08115624636411667,
0.003902826923877001,
-0.027669476345181465,
-0.01197951752692461,
-0.09271810203790665,
0.09678293019533157,
-0.016154086217284203,
-0.12596066296100616,
-0.15380346775054932,
0.06716643273830414,
-0.08688230812549591,
0.011025900952517986,
0.19363687932491302,
-0.09717557579278946,
0.051650308072566986,
0.02240489423274994,
-0.03788372874259949,
-0.09308378398418427,
-0.004756526090204716,
-0.08850155025720596,
-0.0733095034956932,
0.08655881136655807,
-0.17940057814121246,
-0.02230468951165676,
0.04812780022621155,
0.06519358605146408,
0.038997583091259,
0.03653611242771149,
-0.11753356456756592,
0.11634545028209686,
0.13775412738323212,
-0.14906199276447296,
-0.0697489082813263,
-0.07361546158790588,
-0.05249553173780441,
0.15283417701721191,
0.042507193982601166,
0.02888251468539238,
0.002566701965406537,
-0.013704271987080574,
0.020748751237988472,
-0.01742234081029892,
-0.13535860180854797,
-0.06776537001132965,
0.11289840936660767,
-0.03526081144809723,
-0.1042080670595169,
0.062077172100543976,
0.03583935275673866,
0.12504692375659943,
-0.05674272030591965,
0.027771878987550735,
-0.06925780326128006,
-0.09621066600084305,
-0.19877071678638458,
0.06293253600597382,
-0.1992800235748291,
-0.05842785909771919,
-0.04865417256951332,
-0.04270896315574646,
-0.04809798300266266,
0.09285251051187515,
0.027214838191866875,
0.07145673036575317,
0.04635203257203102,
0.016376622021198273,
-0.00255291978828609,
-0.10042405873537064,
-0.08205853402614594,
0.03508797660470009,
-0.10745305567979813,
-0.1421753168106079,
0.04155559092760086,
0.1120680421590805,
-0.07922976464033127,
-0.05370207875967026,
-0.16149526834487915,
0.024609509855508804,
-0.09296352416276932,
0.014487884007394314,
-0.16262176632881165,
-0.004029275383800268,
-0.014904518611729145,
-0.055069223046302795,
-0.07104907929897308,
0.010246001183986664,
-0.07980012893676758,
0.056054264307022095,
0.08422324061393738,
0.06759481132030487,
-0.03402814269065857,
0.018117176368832588,
0.05755464360117912,
0.03127814456820488,
0.08387678861618042,
0.10768619179725647,
0.0676952600479126,
0.14085379242897034,
-0.25984129309654236,
0.03416125476360321,
0.013617333956062794,
-0.02823750488460064,
-0.04540951922535896,
0.08438307046890259,
0.009164768271148205,
0.00817177351564169,
-0.02080639824271202,
0.03848422318696976,
-0.08212728798389435,
-0.11221771687269211,
-0.08579632639884949,
-0.01051285583525896,
-0.0467349998652935,
-0.013190685771405697,
-0.02707301452755928,
0.1618724912405014,
0.04405052214860916,
0.08097276836633682,
0.0011022419203072786,
-0.008922652341425419,
-0.012481249868869781,
-0.02148864045739174,
-0.024189958348870277,
-0.10461746156215668,
0.008861014619469643,
-0.06532116234302521,
-0.07644105702638626,
0.01603027805685997,
0.21264855563640594,
-0.06739240884780884,
-0.17590545117855072,
-0.012474643997848034,
0.11770553141832352,
0.06626995652914047,
-0.03730320557951927,
0.3892042636871338,
0.049407463520765305,
0.0005271050613373518,
-0.11066264659166336,
0.10120741277933121,
-0.06961280852556229,
-0.1889486461877823,
-0.015234533697366714,
0.011389806866645813,
-0.07948076725006104,
-0.042692795395851135,
0.07014229148626328,
-0.08349527418613434,
0.05924322456121445,
0.1026693806052208,
-0.015514873899519444,
-0.08010166138410568,
-0.0208444707095623,
-0.013794581405818462,
0.24483752250671387,
-0.034877121448516846,
-0.009027530439198017,
0.00952915195375681,
-0.02107326127588749,
-0.07008058577775955,
-0.21871913969516754,
-0.03560461848974228,
-0.1679067760705948,
0.12409016489982605,
-0.02875329926609993,
0.00008071696356637403,
0.15271569788455963,
0.007305186707526445,
-0.10448218137025833,
-0.008546660654246807,
0.037788838148117065,
-0.0758458748459816,
0.04650134593248367,
-0.028856169432401657,
-0.07444648444652557,
-0.12792372703552246,
-0.05947110056877136,
0.03632860258221626,
-0.12901003658771515,
-0.004244013223797083,
-0.025514381006360054,
0.03348580375313759,
-0.016220998018980026,
-0.07954675704240799,
-0.04755159467458725,
-0.020310834050178528,
0.06349015980958939,
0.10382160544395447,
0.1884600669145584,
-0.03201891854405403,
-0.008435053750872612,
0.04435797408223152,
0.10507737100124359,
0.07524392008781433,
-0.03841296583414078,
-0.03124869614839554,
0.04405738785862923,
-0.10687430948019028,
0.02269514836370945,
-0.027721194550395012,
-0.03383364528417587,
0.14649538695812225,
0.19471639394760132,
0.165367990732193,
-0.060858264565467834,
0.04772879555821419,
-0.05943099781870842,
0.014569760300219059,
0.03339407965540886,
0.06683824211359024,
-0.04463199898600578,
0.13030950725078583,
-0.11571606248617172,
-0.00669187493622303,
-0.03889603540301323,
0.06743880361318588,
-0.05494688078761101,
0.0899902954697609,
0.07970428466796875,
-0.05303238332271576,
-0.10135333985090256,
0.10370031744241714,
-0.18674728274345398,
0.10043583065271378,
0.10130999237298965,
-0.017164485529065132,
0.037375930696725845,
-0.01628248207271099,
-0.06276543438434601,
0.09001078456640244,
0.05997305363416672,
-0.10949888080358505,
-0.05130322277545929,
-0.16504010558128357,
0.05046640336513519,
-0.20626333355903625,
-0.08936043083667755,
0.08364196866750717,
0.1890195608139038,
0.23074480891227722,
-0.04620203375816345,
0.058419369161129,
0.06126277521252632,
0.02320392243564129,
-0.07488510757684708,
0.184667706489563,
0.03501497954130173,
-0.12491831928491592,
-0.11928005516529083,
-0.11068461835384369,
0.03739981725811958,
-0.05872778967022896,
0.04313452169299126,
-0.009478509426116943,
0.03890237212181091,
0.1540495753288269,
-0.03208247944712639,
-0.028794974088668823,
0.11491173505783081,
-0.12021491676568985,
0.0969771146774292,
-0.11044061183929443,
-0.03414588049054146,
-0.15233711898326874,
-0.0438668318092823,
0.07037526369094849,
0.08014161884784698,
-0.06063784658908844,
-0.014854948967695236,
0.02786206267774105,
0.030347907915711403,
0.18394646048545837,
-0.023319557309150696,
-0.07684189826250076,
0.00176699785515666,
-0.08121786266565323,
0.09851302951574326,
-0.055586718022823334,
0.018611406907439232,
0.09238631278276443,
-0.042573899030685425,
0.004694420378655195,
-0.27044641971588135,
-0.021039016544818878,
-0.06437614560127258,
-0.04325375333428383,
-0.0862412080168724
] |
null | null | transformers | # MiquMaid-v2-70B-DPO 2.4bpw
## Description
Exllama quant of [NeverSleep/MiquMaid-v2-70B-DPO](https://huggingface.co/NeverSleep/MiquMaid-v2-70B-DPO)
## Other quants:
EXL2: [4bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-4bpw-exl2), [3.5bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-3.5bpw-exl2), [3bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-3bpw-exl2), [2.4bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-2.4bpw-exl2), [2.3bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-2.3bpw-exl2)
2.4bpw is probably the most you can fit in a 24gb card
GGUF:
[2bit Imatrix GGUF](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-Imatrix-GGUF)
## Prompt format: Alpaca
```
### Instruction:
{system prompt}
### Input:
{input}
### Response:
{reply}
```
## Contact
Kooten on discord
[ko-fi.com/kooten](https://ko-fi.com/kooten) | {"license": "cc-by-nc-4.0", "tags": ["not-for-all-audiences", "nsfw"]} | text-generation | Kooten/MiquMaid-v2-70B-DPO-2.4bpw-exl2 | [
"transformers",
"pytorch",
"llama",
"text-generation",
"not-for-all-audiences",
"nsfw",
"conversational",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T06:56:41+00:00 | [] | [] | TAGS
#transformers #pytorch #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # MiquMaid-v2-70B-DPO 2.4bpw
## Description
Exllama quant of NeverSleep/MiquMaid-v2-70B-DPO
## Other quants:
EXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw
2.4bpw is probably the most you can fit in a 24gb card
GGUF:
2bit Imatrix GGUF
## Prompt format: Alpaca
## Contact
Kooten on discord
URL | [
"# MiquMaid-v2-70B-DPO 2.4bpw",
"## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B-DPO",
"## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF",
"## Prompt format: Alpaca",
"## Contact\nKooten on discord\n\nURL"
] | [
"TAGS\n#transformers #pytorch #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# MiquMaid-v2-70B-DPO 2.4bpw",
"## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B-DPO",
"## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF",
"## Prompt format: Alpaca",
"## Contact\nKooten on discord\n\nURL"
] | [
74,
17,
24,
60,
8,
7
] | [
"passage: TAGS\n#transformers #pytorch #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# MiquMaid-v2-70B-DPO 2.4bpw## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B-DPO## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF## Prompt format: Alpaca## Contact\nKooten on discord\n\nURL"
] | [
-0.04528329521417618,
0.07163137197494507,
-0.006452422589063644,
0.0898926854133606,
0.0462457500398159,
0.02721884846687317,
0.12119237333536148,
0.14412756264209747,
0.015651941299438477,
0.007050621323287487,
0.06058143824338913,
0.127939373254776,
0.06306967884302139,
0.15848994255065918,
-0.0401102751493454,
-0.15914487838745117,
0.015428673475980759,
0.0634930282831192,
0.09445460885763168,
0.12359726428985596,
0.06823231279850006,
-0.024871399626135826,
0.04905260354280472,
-0.030898669734597206,
-0.13080698251724243,
-0.00382184237241745,
-0.05011551454663277,
-0.06484078615903854,
0.04989341273903847,
0.09353107959032059,
0.030266525223851204,
0.08950532227754593,
-0.03287433460354805,
-0.17654378712177277,
0.022447781637310982,
0.0327940508723259,
-0.05216139554977417,
0.03640494868159294,
0.04565524682402611,
-0.04865781590342522,
0.038196634501218796,
0.028419090434908867,
-0.07467969506978989,
0.07283339649438858,
-0.10249444842338562,
-0.0900149717926979,
-0.05357963964343071,
0.1409021019935608,
0.0677829384803772,
0.03989473730325699,
-0.007512970361858606,
0.11159121245145798,
-0.037424225360155106,
0.07373037189245224,
0.2272694855928421,
-0.2451624870300293,
-0.021859019994735718,
0.08042746037244797,
0.039312053471803665,
0.05450497567653656,
-0.06577441841363907,
0.004759800620377064,
0.01180858165025711,
-0.005850449204444885,
0.02050834521651268,
-0.03366920351982117,
0.04863416776061058,
-0.023294277489185333,
-0.09160565584897995,
-0.010670259594917297,
0.15132534503936768,
0.06005792319774628,
-0.04678693413734436,
-0.03198288008570671,
-0.07936102151870728,
-0.12744030356407166,
-0.039146389812231064,
0.0331723652780056,
-0.0027068022172898054,
-0.014572324231266975,
-0.044930342584848404,
-0.0006742742261849344,
-0.10563389211893082,
-0.04218386113643646,
-0.14066822826862335,
0.1635373830795288,
0.0012726489221677184,
0.027177471667528152,
0.003951788414269686,
0.047059111297130585,
-0.0887766182422638,
-0.0860796868801117,
-0.08405060321092606,
-0.08869814872741699,
0.039454806596040726,
-0.0011910616885870695,
-0.009989743120968342,
0.10326240956783295,
0.0789450854063034,
0.17500923573970795,
-0.027900468558073044,
0.07528751343488693,
0.029413258656859398,
0.06505939364433289,
0.0025558907072991133,
-0.07921513170003891,
-0.03366680443286896,
-0.041064899414777756,
0.03164006397128105,
0.013581397943198681,
0.04560493677854538,
-0.02283940464258194,
-0.1131848394870758,
-0.021274961531162262,
-0.008426075801253319,
0.02094775065779686,
0.04113501310348511,
0.05120369791984558,
-0.06194581091403961,
-0.06193489953875542,
0.184137761592865,
-0.04908068850636482,
0.02315993420779705,
0.059230584651231766,
-0.0043141767382621765,
0.10196304321289062,
0.04789521172642708,
-0.0029364656656980515,
-0.0574350506067276,
0.05455520376563072,
-0.03349059820175171,
0.02752000465989113,
-0.023661671206355095,
-0.04995246231555939,
0.09059232473373413,
-0.062404338270425797,
-0.014270851388573647,
-0.16976916790008545,
-0.13448341190814972,
0.07022368907928467,
-0.003031268250197172,
-0.05884183570742607,
-0.039791665971279144,
0.003048209473490715,
-0.01874348893761635,
-0.007841820828616619,
-0.06032755225896835,
0.03461231663823128,
-0.069976307451725,
0.08790339529514313,
0.07529346644878387,
0.124673031270504,
-0.11897435039281845,
0.0012562461197376251,
-0.03872247040271759,
0.02734135463833809,
-0.07426171004772186,
0.017842212691903114,
-0.11437317728996277,
-0.029125668108463287,
-0.08240397274494171,
-0.049472711980342865,
-0.028614388778805733,
0.03774334862828255,
-0.0043097627349197865,
0.10732276737689972,
-0.058063577860593796,
-0.058130908757448196,
0.24316895008087158,
-0.09083952754735947,
-0.06674516201019287,
0.1272088587284088,
0.05489056557416916,
-0.08210477977991104,
0.04769393801689148,
0.07109421491622925,
0.1147196888923645,
-0.0989437848329544,
-0.07753217965364456,
0.08143980801105499,
-0.010307874530553818,
-0.08024390786886215,
0.09421651810407639,
0.05149117857217789,
0.03741482272744179,
0.032477885484695435,
0.1582486629486084,
0.03270679712295532,
-0.002578867832198739,
-0.06835747510194778,
-0.0513540580868721,
-0.06125880405306816,
0.07678072154521942,
0.00765292439609766,
-0.0288084764033556,
-0.0951407328248024,
-0.11145547777414322,
-0.0006623206427320838,
0.0766461044549942,
-0.003639130387455225,
-0.007251658942550421,
-0.12011557817459106,
0.09802862256765366,
-0.08426712453365326,
0.03035138174891472,
-0.07477986067533493,
0.0021347226575016975,
-0.03033488616347313,
-0.00821926724165678,
0.08062634617090225,
0.10364771634340286,
0.07265113294124603,
0.03516825661063194,
-0.07812323421239853,
0.03522464632987976,
0.09419062733650208,
0.0034943101927638054,
-0.005134623497724533,
-0.06783166527748108,
0.027492238208651543,
-0.01253916509449482,
0.12063980847597122,
-0.07953637093305588,
0.025750774890184402,
0.13455478847026825,
0.10318318754434586,
-0.025927789509296417,
-0.034950561821460724,
0.06508994847536087,
0.03175840899348259,
-0.042625267058610916,
0.016545845195651054,
0.0872839167714119,
-0.018271617591381073,
-0.01626809500157833,
0.09280532598495483,
-0.21007531881332397,
0.11604064702987671,
0.17388203740119934,
-0.06291733682155609,
-0.02991160936653614,
-0.06131114065647125,
-0.009001716040074825,
-0.026116440072655678,
0.0108818793669343,
-0.0558149516582489,
0.07376930117607117,
-0.014966060407459736,
0.12646397948265076,
-0.08400431275367737,
-0.0533720999956131,
0.05316966772079468,
-0.045308489352464676,
0.0024740127846598625,
0.06781847029924393,
0.10850757360458374,
-0.11409509927034378,
0.15815870463848114,
0.14045053720474243,
-0.08816653490066528,
0.13273866474628448,
0.026954639703035355,
-0.030716726556420326,
-0.033694881945848465,
0.015496689826250076,
0.008249777369201183,
0.057269901037216187,
-0.11634588986635208,
0.05585157126188278,
0.08057928830385208,
-0.011945321224629879,
0.025167811661958694,
-0.13197088241577148,
-0.04010554403066635,
0.020707838237285614,
-0.0700286328792572,
-0.0838564783334732,
-0.003250167006626725,
0.005895698443055153,
0.1021244004368782,
-0.015017818659543991,
-0.0017467769794166088,
0.03929918631911278,
0.042918767780065536,
-0.07136955112218857,
0.1400219053030014,
-0.09697015583515167,
-0.3062552213668823,
-0.08039301633834839,
-0.14406630396842957,
-0.08165093511343002,
0.01735171489417553,
0.09327059239149094,
-0.0687723383307457,
-0.030832545831799507,
-0.04780550301074982,
-0.045138776302337646,
-0.0448143295943737,
-0.010994265787303448,
0.01062546856701374,
-0.01665470562875271,
0.057832423597574234,
-0.07968410104513168,
-0.020553812384605408,
0.05868648365139961,
-0.041562534868717194,
0.16663700342178345,
-0.023025739938020706,
0.0900566503405571,
0.13824474811553955,
0.017190564423799515,
-0.016270335763692856,
-0.03229590505361557,
0.15724404156208038,
-0.06636595726013184,
0.050204142928123474,
0.14103077352046967,
-0.010723590850830078,
0.07745099067687988,
0.09187301248311996,
0.04077059403061867,
-0.05148950219154358,
-0.05275334417819977,
0.030395707115530968,
-0.09769417345523834,
-0.1466764658689499,
-0.03707330301403999,
-0.06356684863567352,
0.1448662281036377,
0.05608278885483742,
0.06016278639435768,
0.06789299100637436,
0.11671984940767288,
-0.0765274167060852,
0.00217574299313128,
0.03049309365451336,
0.09042918682098389,
0.16819067299365997,
0.021895350888371468,
0.15410727262496948,
-0.06035047769546509,
-0.03689112886786461,
0.12719614803791046,
0.1320013701915741,
0.04616347327828407,
0.007890940643846989,
0.1078610047698021,
0.056643176823854446,
0.01637193001806736,
0.09680477529764175,
0.0455782525241375,
0.013320399448275566,
-0.03150862455368042,
-0.04039545729756355,
-0.0589636005461216,
-0.00733917485922575,
0.043358057737350464,
0.07051298767328262,
-0.07730213552713394,
0.030801959335803986,
-0.08793442696332932,
0.06702514737844467,
0.06651318818330765,
0.09082905203104019,
-0.20864856243133545,
-0.02963423915207386,
0.049608927220106125,
-0.027823233976960182,
-0.041616056114435196,
0.019097687676548958,
0.0048203072510659695,
-0.02216091938316822,
0.16462264955043793,
-0.009996399283409119,
0.09373197704553604,
-0.07609635591506958,
0.02394089102745056,
-0.0336696058511734,
0.05475431680679321,
0.0032541267573833466,
0.11388926953077316,
-0.2912919521331787,
0.07416396588087082,
0.023151226341724396,
0.0008483300916850567,
-0.0707821398973465,
-0.008730041794478893,
0.013543725945055485,
-0.03793439641594887,
0.0890703946352005,
0.012301306240260601,
0.11272485554218292,
-0.13513131439685822,
-0.06880045682191849,
0.01789858750998974,
0.05891025438904762,
0.03406525403261185,
0.038898441940546036,
0.049576397985219955,
-0.04131985828280449,
-0.011378156021237373,
-0.004000697750598192,
-0.037983596324920654,
-0.09390521049499512,
0.12354180961847305,
0.11992531269788742,
-0.020762450993061066,
0.013536415062844753,
-0.032947760075330734,
-0.18882542848587036,
0.22462955117225647,
-0.11887657642364502,
-0.09915962815284729,
-0.06057421490550041,
-0.057684849947690964,
0.020852483808994293,
-0.03707578778266907,
0.019683770835399628,
-0.05346710607409477,
0.09027983248233795,
-0.0326990969479084,
-0.0836855098605156,
0.08107277005910873,
-0.10390911996364594,
-0.13462325930595398,
-0.03570057079195976,
0.15627582371234894,
-0.0713052973151207,
0.03943798691034317,
0.019598297774791718,
-0.01245953980833292,
-0.044795695692300797,
-0.12616638839244843,
0.10435653477907181,
0.006767628248780966,
-0.0030304021202027798,
0.01629154570400715,
0.017540719360113144,
-0.011568732559680939,
-0.038675326853990555,
-0.03445075824856758,
0.14126884937286377,
0.3582994043827057,
-0.03493929281830788,
0.11785607039928436,
0.11852559447288513,
-0.04013323411345482,
-0.21210768818855286,
-0.16463030874729156,
-0.03908853977918625,
-0.08053785562515259,
-0.05867880955338478,
-0.138952374458313,
0.04111839458346367,
0.13380493223667145,
-0.03485734388232231,
0.1909855753183365,
-0.14483244717121124,
-0.09902530163526535,
0.04036460071802139,
0.053661808371543884,
0.24369053542613983,
-0.15806622803211212,
-0.060376424342393875,
-0.06081567704677582,
-0.2034968137741089,
0.22196625173091888,
-0.038331735879182816,
0.12462195008993149,
-0.03901337459683418,
0.09878508001565933,
-0.0021295088808983564,
-0.05908086523413658,
0.12516170740127563,
-0.023898525163531303,
-0.02979055605828762,
-0.1173824891448021,
0.04384712129831314,
0.006777587346732616,
-0.007371976040303707,
0.010031231679022312,
-0.17720730602741241,
-0.02875439077615738,
-0.13824231922626495,
-0.059282924979925156,
-0.07689593732357025,
0.017675895243883133,
-0.032586731016635895,
-0.05288975313305855,
-0.053078535944223404,
-0.002714537549763918,
0.01391583401709795,
-0.022026503458619118,
-0.013614046387374401,
-0.07199207693338394,
0.0757441595196724,
0.3028547167778015,
0.09156999737024307,
-0.07679930329322815,
-0.09435424208641052,
-0.06671026349067688,
-0.0948554277420044,
0.021586278453469276,
-0.1196807399392128,
0.03649064153432846,
0.0789615735411644,
0.008781321346759796,
0.06257959455251694,
0.04451083019375801,
-0.059990499168634415,
0.020105648785829544,
0.12036018073558807,
-0.13913343846797943,
0.0031466607470065355,
-0.012797651812434196,
0.06317488104104996,
0.0004895043675787747,
-0.011513053439557552,
0.13592085242271423,
-0.00818924605846405,
-0.013584510423243046,
0.058785196393728256,
0.018331237137317657,
-0.020136382430791855,
0.11410196870565414,
0.03432319685816765,
0.07384408265352249,
-0.13042877614498138,
0.06291990727186203,
-0.010775908827781677,
-0.05863502621650696,
-0.010105149820446968,
0.10417993366718292,
-0.06937842071056366,
-0.12815244495868683,
-0.017546208575367928,
-0.08106564730405807,
-0.09955655038356781,
-0.039303962141275406,
-0.02379108965396881,
-0.11868651211261749,
0.06567592173814774,
0.11985321342945099,
0.0214877687394619,
-0.023858657106757164,
0.06417647749185562,
0.04118369147181511,
-0.06942077726125717,
0.05954727157950401,
-0.007935027591884136,
0.0827648788690567,
-0.10057100653648376,
0.05627457797527313,
-0.007071836851537228,
0.045663751661777496,
-0.04201728105545044,
0.011263292282819748,
-0.12162399291992188,
-0.029296578839421272,
-0.14019446074962616,
0.05640781670808792,
-0.11359187215566635,
-0.004185907077044249,
-0.019457103684544563,
-0.0014538337709382176,
-0.02862389385700226,
-0.040162645280361176,
-0.09108053147792816,
0.005106726195663214,
0.0204436257481575,
-0.006414234172552824,
-0.09440050274133682,
-0.027568846940994263,
0.029350578784942627,
-0.02459637075662613,
0.07276730239391327,
0.05146758258342743,
-0.05297821760177612,
-0.019372396171092987,
-0.1498912125825882,
-0.033467549830675125,
0.07894154638051987,
0.07147682458162308,
-0.024189600721001625,
0.024439599364995956,
0.04482119530439377,
0.04017573967576027,
0.03610227257013321,
0.029948171228170395,
0.07726418226957321,
-0.08972108364105225,
-0.016648413613438606,
-0.07050257176160812,
-0.09997420758008957,
-0.052113085985183716,
-0.009810904040932655,
0.12614473700523376,
-0.03147149085998535,
0.10636845231056213,
-0.048485904932022095,
0.027273094281554222,
-0.1380540132522583,
0.020464254543185234,
0.029605790972709656,
-0.1483004093170166,
-0.006754735950380564,
-0.05732715129852295,
0.03101530484855175,
-0.028723331168293953,
0.16346248984336853,
-0.07850834727287292,
-0.10358870774507523,
0.028283532708883286,
-0.07088905572891235,
-0.03706059232354164,
-0.01753208227455616,
0.19589252769947052,
0.07834082096815109,
-0.012333392165601254,
-0.09437050670385361,
0.027053268626332283,
0.05555791035294533,
-0.07573895156383514,
0.08729879558086395,
0.08243527263402939,
-0.027995288372039795,
0.043542779982089996,
0.06570974737405777,
-0.0732206478714943,
-0.08390334248542786,
0.006977423094213009,
-0.12231682240962982,
0.10399163514375687,
-0.019946131855249405,
0.08957935869693756,
0.08756081014871597,
-0.042119238525629044,
0.018063200637698174,
-0.06798744946718216,
-0.034181322902441025,
-0.11061698198318481,
-0.08541616797447205,
-0.08318426460027695,
-0.09562535583972931,
0.029740316793322563,
-0.07634840905666351,
0.03253920003771782,
0.059507858008146286,
0.0473824068903923,
-0.015896154567599297,
0.08763337135314941,
-0.016738703474402428,
-0.03312166407704353,
0.010083924978971481,
-0.02478504739701748,
-0.020216811448335648,
0.0317029245197773,
-0.03501057252287865,
0.06746170669794083,
-0.012812582775950432,
0.045526403933763504,
0.025530897080898285,
0.019834039732813835,
-0.0032591265626251698,
-0.06828749179840088,
-0.09333505481481552,
-0.03524521738290787,
0.035732828080654144,
0.07947167754173279,
0.15581165254116058,
-0.03362594172358513,
0.015329455956816673,
0.016099698841571808,
0.07745765149593353,
-0.027735436335206032,
-0.08362320065498352,
-0.03638264909386635,
0.19769547879695892,
-0.0861925333738327,
0.05161915719509125,
-0.013659772463142872,
-0.022225577384233475,
-0.009700963273644447,
0.24597641825675964,
0.15538020431995392,
-0.03997454047203064,
0.0310515109449625,
0.045467495918273926,
0.02256855182349682,
0.07640795409679413,
0.0907912328839302,
0.07969585806131363,
0.1816909909248352,
-0.05036146938800812,
0.005301611963659525,
0.028939485549926758,
0.0008707839879207313,
-0.08083228766918182,
0.061098821461200714,
0.01687912829220295,
-0.018574412912130356,
-0.0639202743768692,
0.02103685401380062,
-0.12717978656291962,
0.06778627634048462,
-0.003444356843829155,
-0.11873090267181396,
-0.07583507150411606,
0.006522797048091888,
0.052008599042892456,
0.04581400007009506,
0.012317890301346779,
-0.03108825534582138,
-0.01044714916497469,
0.021397674456238747,
-0.019776789471507072,
-0.17685963213443756,
0.05629056692123413,
0.005051653832197189,
-0.0023562079295516014,
0.029236188158392906,
-0.039799533784389496,
0.15717796981334686,
0.10553140193223953,
0.008699931204319,
-0.05069193243980408,
0.13004064559936523,
0.02007257007062435,
-0.16977433860301971,
0.03929569572210312,
0.1285264641046524,
-0.03408025577664375,
-0.0004818830930162221,
0.09348344057798386,
-0.06363322585821152,
-0.05150986835360527,
0.10830113291740417,
-0.029827432706952095,
-0.017348160967230797,
0.06194726377725601,
-0.0541544184088707,
0.07221902906894684,
0.06545151770114899,
-0.05014830082654953,
-0.07010024785995483,
-0.04440038278698921,
0.05321282148361206,
-0.024059981107711792,
-0.055699631571769714,
-0.05668584257364273,
-0.16982164978981018,
-0.05214448273181915,
0.05430714786052704,
0.1254342645406723,
-0.24346593022346497,
-0.006135833915323019,
-0.08637513965368271,
0.00045734900049865246,
-0.12948040664196014,
0.06127529591321945,
0.09904332458972931,
0.01079505030065775,
-0.05136748030781746,
-0.10356951504945755,
-0.01981150545179844,
0.07917719334363937,
-0.12312391400337219,
-0.07185251265764236
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# 1000_STEPS_5e7
This model is a fine-tuned version of [meta-llama/Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b-hf) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.6624
- Rewards/chosen: -0.0929
- Rewards/rejected: -0.1682
- Rewards/accuracies: 0.5451
- Rewards/margins: 0.0754
- Logps/rejected: -16.8218
- Logps/chosen: -15.0454
- Logits/rejected: -0.1317
- Logits/chosen: -0.1315
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-07
- train_batch_size: 4
- eval_batch_size: 1
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 8
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_steps: 100
- training_steps: 1200
### Training results
| Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
|:-------------:|:-----:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
| 0.6932 | 0.1 | 50 | 0.6927 | 0.0015 | 0.0005 | 0.4242 | 0.0010 | -15.1347 | -14.1020 | -0.0215 | -0.0215 |
| 0.6901 | 0.2 | 100 | 0.6901 | -0.0121 | -0.0185 | 0.4835 | 0.0063 | -15.3239 | -14.2383 | -0.0268 | -0.0268 |
| 0.6837 | 0.29 | 150 | 0.6838 | -0.0153 | -0.0352 | 0.5209 | 0.0199 | -15.4913 | -14.2697 | -0.0332 | -0.0331 |
| 0.6775 | 0.39 | 200 | 0.6808 | -0.0261 | -0.0529 | 0.5363 | 0.0268 | -15.6684 | -14.3783 | -0.0453 | -0.0451 |
| 0.6761 | 0.49 | 250 | 0.6779 | -0.0555 | -0.0896 | 0.5297 | 0.0340 | -16.0350 | -14.6723 | -0.0533 | -0.0531 |
| 0.6692 | 0.59 | 300 | 0.6771 | -0.0812 | -0.1192 | 0.5121 | 0.0380 | -16.3311 | -14.9285 | -0.0659 | -0.0657 |
| 0.683 | 0.68 | 350 | 0.6739 | -0.0352 | -0.0789 | 0.5385 | 0.0437 | -15.9286 | -14.4687 | -0.0583 | -0.0581 |
| 0.677 | 0.78 | 400 | 0.6733 | -0.0430 | -0.0895 | 0.5451 | 0.0464 | -16.0340 | -14.5472 | -0.0635 | -0.0633 |
| 0.6665 | 0.88 | 450 | 0.6692 | -0.0347 | -0.0904 | 0.5516 | 0.0557 | -16.0436 | -14.4638 | -0.0724 | -0.0721 |
| 0.6559 | 0.98 | 500 | 0.6668 | -0.0374 | -0.0989 | 0.5516 | 0.0615 | -16.1283 | -14.4907 | -0.0752 | -0.0750 |
| 0.6406 | 1.07 | 550 | 0.6665 | -0.0482 | -0.1114 | 0.5582 | 0.0632 | -16.2528 | -14.5988 | -0.0983 | -0.0981 |
| 0.6301 | 1.17 | 600 | 0.6656 | -0.0655 | -0.1316 | 0.5495 | 0.0661 | -16.4553 | -14.7718 | -0.1067 | -0.1065 |
| 0.6206 | 1.27 | 650 | 0.6648 | -0.0581 | -0.1265 | 0.5407 | 0.0684 | -16.4041 | -14.6977 | -0.1163 | -0.1161 |
| 0.6015 | 1.37 | 700 | 0.6641 | -0.0734 | -0.1439 | 0.5495 | 0.0706 | -16.5788 | -14.8504 | -0.1219 | -0.1217 |
| 0.6299 | 1.46 | 750 | 0.6637 | -0.0883 | -0.1601 | 0.5429 | 0.0719 | -16.7407 | -14.9994 | -0.1233 | -0.1231 |
| 0.6031 | 1.56 | 800 | 0.6630 | -0.0881 | -0.1617 | 0.5407 | 0.0736 | -16.7566 | -14.9977 | -0.1270 | -0.1267 |
| 0.6474 | 1.66 | 850 | 0.6633 | -0.0908 | -0.1640 | 0.5451 | 0.0733 | -16.7795 | -15.0245 | -0.1278 | -0.1276 |
| 0.6229 | 1.76 | 900 | 0.6630 | -0.0923 | -0.1664 | 0.5473 | 0.0740 | -16.8031 | -15.0403 | -0.1290 | -0.1288 |
| 0.6085 | 1.86 | 950 | 0.6628 | -0.0924 | -0.1669 | 0.5429 | 0.0745 | -16.8081 | -15.0405 | -0.1305 | -0.1303 |
| 0.6062 | 1.95 | 1000 | 0.6629 | -0.0932 | -0.1676 | 0.5363 | 0.0744 | -16.8155 | -15.0492 | -0.1302 | -0.1300 |
| 0.6236 | 2.05 | 1050 | 0.6628 | -0.0937 | -0.1684 | 0.5451 | 0.0747 | -16.8231 | -15.0539 | -0.1314 | -0.1311 |
| 0.6217 | 2.15 | 1100 | 0.6629 | -0.0932 | -0.1676 | 0.5451 | 0.0744 | -16.8150 | -15.0489 | -0.1308 | -0.1305 |
| 0.6255 | 2.25 | 1150 | 0.6624 | -0.0927 | -0.1682 | 0.5495 | 0.0755 | -16.8214 | -15.0442 | -0.1315 | -0.1313 |
| 0.598 | 2.34 | 1200 | 0.6624 | -0.0929 | -0.1682 | 0.5451 | 0.0754 | -16.8218 | -15.0454 | -0.1317 | -0.1315 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.0.0+cu117
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"tags": ["trl", "dpo", "generated_from_trainer"], "base_model": "meta-llama/Llama-2-7b-hf", "model-index": [{"name": "1000_STEPS_5e7", "results": []}]} | text-generation | tsavage68/1200STEPS_5e7_0.1beta_DPO_zeroshot | [
"transformers",
"safetensors",
"llama",
"text-generation",
"trl",
"dpo",
"generated_from_trainer",
"base_model:meta-llama/Llama-2-7b-hf",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T06:57:08+00:00 | [] | [] | TAGS
#transformers #safetensors #llama #text-generation #trl #dpo #generated_from_trainer #base_model-meta-llama/Llama-2-7b-hf #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| 1000\_STEPS\_5e7
================
This model is a fine-tuned version of meta-llama/Llama-2-7b-hf on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 0.6624
* Rewards/chosen: -0.0929
* Rewards/rejected: -0.1682
* Rewards/accuracies: 0.5451
* Rewards/margins: 0.0754
* Logps/rejected: -16.8218
* Logps/chosen: -15.0454
* Logits/rejected: -0.1317
* Logits/chosen: -0.1315
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-07
* train\_batch\_size: 4
* eval\_batch\_size: 1
* seed: 42
* gradient\_accumulation\_steps: 2
* total\_train\_batch\_size: 8
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: cosine
* lr\_scheduler\_warmup\_steps: 100
* training\_steps: 1200
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.0.0+cu117
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-07\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 1\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 8\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 100\n* training\\_steps: 1200",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.0.0+cu117\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #trl #dpo #generated_from_trainer #base_model-meta-llama/Llama-2-7b-hf #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-07\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 1\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 8\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 100\n* training\\_steps: 1200",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.0.0+cu117\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
78,
145,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #trl #dpo #generated_from_trainer #base_model-meta-llama/Llama-2-7b-hf #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-07\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 1\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 8\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 100\n* training\\_steps: 1200### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.0.0+cu117\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.13937005400657654,
0.09209383279085159,
-0.0020291083492338657,
0.07624784857034683,
0.14197181165218353,
0.01198587566614151,
0.10289172828197479,
0.1343139410018921,
-0.10743379592895508,
0.08814225345849991,
0.1353447288274765,
0.11547805368900299,
0.05357689782977104,
0.18019822239875793,
-0.0353400744497776,
-0.3140809237957001,
0.005758704151958227,
-0.015572793781757355,
-0.16373947262763977,
0.12805800139904022,
0.0929272249341011,
-0.12013629078865051,
0.05052344873547554,
-0.03451646864414215,
-0.12046653032302856,
-0.03495026007294655,
-0.022787991911172867,
-0.04159916564822197,
0.13627614080905914,
0.005083560943603516,
0.11143279075622559,
0.05547409877181053,
0.09781313687562943,
-0.2242133766412735,
0.011788834817707539,
0.061056461185216904,
0.043318115174770355,
0.0841531828045845,
0.06671314686536789,
-0.02993115410208702,
0.0844661071896553,
-0.09975263476371765,
0.07061567157506943,
0.0357455350458622,
-0.1216992735862732,
-0.22973686456680298,
-0.10280755162239075,
0.05498727783560753,
0.15296289324760437,
0.08189421147108078,
-0.02215447835624218,
0.0714239627122879,
-0.08031080663204193,
0.08298638463020325,
0.24331200122833252,
-0.2672750949859619,
-0.08758027851581573,
0.06317945569753647,
0.06356534361839294,
0.06975310295820236,
-0.12509465217590332,
-0.008038238622248173,
0.04130522161722183,
0.012570145539939404,
0.12288789451122284,
0.005998064298182726,
0.07066870480775833,
0.006162129808217287,
-0.14932526648044586,
-0.036003585904836655,
0.10852663218975067,
0.0779898464679718,
-0.03880893811583519,
-0.082615926861763,
-0.03851707652211189,
-0.22931471467018127,
-0.040050607174634933,
-0.018948938697576523,
0.03094690665602684,
-0.05282660201191902,
-0.09719119220972061,
0.0073349690064787865,
-0.07258696109056473,
-0.10739200562238693,
0.04923811927437782,
0.14694339036941528,
0.03847178816795349,
-0.04582024738192558,
0.026293538510799408,
0.16388416290283203,
0.05575529858469963,
-0.1538512259721756,
-0.005005211569368839,
0.015072371810674667,
-0.07688244432210922,
-0.0431162565946579,
-0.03002716414630413,
0.007287744898349047,
0.008019793778657913,
0.14921027421951294,
-0.04924026504158974,
0.04289380833506584,
0.05009779706597328,
0.02778293937444687,
-0.11386648565530777,
0.14739292860031128,
-0.08114276826381683,
-0.0909755602478981,
-0.0319039523601532,
0.14586208760738373,
-0.0044690570794045925,
-0.00756392627954483,
-0.08274732530117035,
0.010149374604225159,
0.11555486917495728,
0.06983677297830582,
-0.02803565189242363,
0.04032033681869507,
-0.07085830718278885,
-0.015243512578308582,
0.030813755467534065,
-0.09785155951976776,
0.015414385125041008,
-0.0005646378849633038,
-0.08466047793626785,
-0.053950753062963486,
0.004682897590100765,
0.017290910705924034,
0.01906920224428177,
0.13481786847114563,
-0.08095134049654007,
-0.024960359558463097,
-0.09990638494491577,
-0.09697283804416656,
0.006953186821192503,
-0.07952236384153366,
-0.008293886668980122,
-0.07879337668418884,
-0.15913300216197968,
-0.06218709424138069,
0.05077872425317764,
-0.05981330946087837,
-0.0680171474814415,
-0.08528351038694382,
-0.103064626455307,
0.02880183421075344,
-0.008317697793245316,
0.1476806402206421,
-0.048565689474344254,
0.13458295166492462,
0.012350265868008137,
0.07657577097415924,
0.059316545724868774,
0.0442744679749012,
-0.05193745344877243,
0.06983493268489838,
-0.2220771312713623,
0.06918685883283615,
-0.06884198635816574,
0.09504179656505585,
-0.12480060011148453,
-0.09500744193792343,
-0.031359538435935974,
-0.009828821755945683,
0.09488866478204727,
0.16039492189884186,
-0.17588116228580475,
-0.07014579325914383,
0.20278577506542206,
-0.06561114639043808,
-0.13311977684497833,
0.10945416241884232,
-0.03173181042075157,
0.044766563922166824,
0.03981299698352814,
0.1405709683895111,
0.10052653402090073,
-0.08001033216714859,
0.01085414458066225,
-0.04330103099346161,
0.09650623053312302,
0.022852715104818344,
0.09992334246635437,
-0.033106546849012375,
0.01584435999393463,
-0.004744445905089378,
-0.06718537956476212,
0.04970719292759895,
-0.10317332297563553,
-0.08355392515659332,
-0.002291724318638444,
-0.10254372656345367,
0.06270015239715576,
0.04279943183064461,
0.046885162591934204,
-0.08839426189661026,
-0.11335418373346329,
0.011559919454157352,
0.11090286821126938,
-0.07462358474731445,
0.016285276040434837,
-0.04045813903212547,
0.06906090676784515,
-0.03206987306475639,
-0.0030654093716293573,
-0.14901359379291534,
-0.060116857290267944,
0.02106832154095173,
0.025876814499497414,
-0.01924283243715763,
-0.028550148010253906,
0.08145859837532043,
0.06831791251897812,
-0.0835299864411354,
-0.08112479746341705,
-0.055597469210624695,
-0.004856047686189413,
-0.10897368937730789,
-0.2395460158586502,
-0.05727856233716011,
-0.03341536968946457,
0.21451909840106964,
-0.26122984290122986,
0.04973379522562027,
0.006761663593351841,
0.12160058319568634,
0.036873891949653625,
-0.034296754747629166,
-0.005329058971256018,
0.05403830111026764,
-0.03268282860517502,
-0.08943790197372437,
0.042798250913619995,
-0.009486480616033077,
-0.13127243518829346,
-0.011755327694118023,
-0.12849320471286774,
0.1203506737947464,
0.09667880833148956,
0.009051306173205376,
-0.1391485333442688,
-0.08937125653028488,
-0.07038138806819916,
-0.04760390520095825,
-0.02784503623843193,
-0.009034383110702038,
0.09292100369930267,
0.03746894374489784,
0.1241302490234375,
-0.07919151335954666,
-0.0562298558652401,
0.028775926679372787,
-0.005566742271184921,
0.019217541441321373,
0.14702419936656952,
0.047420866787433624,
-0.060953933745622635,
0.12402421236038208,
0.1269388645887375,
-0.04331333190202713,
0.14464348554611206,
-0.044670525938272476,
-0.09393966943025589,
-0.026243852451443672,
0.06039916351437569,
0.04576009884476662,
0.13352522253990173,
-0.07957086712121964,
-0.00770802004262805,
0.0022338347043842077,
0.028537893667817116,
0.0034128434490412474,
-0.2028670758008957,
-0.04771508648991585,
0.04881114140152931,
-0.05470865219831467,
0.0017386714462190866,
-0.025611523538827896,
-0.017861977219581604,
0.10486191511154175,
0.04341351240873337,
-0.056681565940380096,
0.012508542276918888,
-0.006608553696423769,
-0.07912693917751312,
0.23010538518428802,
-0.08531400561332703,
-0.11891120672225952,
-0.10957993566989899,
0.00955918151885271,
-0.010547696612775326,
0.015960050746798515,
0.029236318543553352,
-0.10196240246295929,
0.011777526699006557,
-0.06989513337612152,
0.018529240041971207,
-0.02135089784860611,
0.03950955346226692,
-0.026560094207525253,
0.022908106446266174,
0.05335814878344536,
-0.08270629495382309,
0.021681124344468117,
-0.01601371541619301,
-0.053150665014982224,
0.04454561322927475,
0.002915446413680911,
0.111930251121521,
0.1844976544380188,
0.017929870635271072,
0.017207428812980652,
-0.04843877628445625,
0.14349894225597382,
-0.13218067586421967,
0.003842497942969203,
0.0991789922118187,
0.021432599052786827,
0.0572093240916729,
0.16269297897815704,
0.037435419857501984,
-0.09844201803207397,
0.056685321033000946,
0.04118185490369797,
-0.024711372330784798,
-0.21740354597568512,
-0.005786220543086529,
-0.038540396839380264,
0.030650535598397255,
0.10318521410226822,
0.03022126853466034,
0.02865363284945488,
0.061660949140787125,
-0.022487595677375793,
-0.0010524694807827473,
0.010043071582913399,
0.0764811784029007,
-0.006863452028483152,
0.030851589515805244,
0.12043106555938721,
-0.014866175130009651,
-0.04072002321481705,
0.00881493091583252,
0.020328423008322716,
0.2335279881954193,
-0.030388426035642624,
0.11797907203435898,
0.04637526720762253,
0.15047746896743774,
-0.01056051068007946,
0.08285268396139145,
0.026842154562473297,
-0.04669279605150223,
0.004601517226547003,
-0.05843791738152504,
-0.02663506753742695,
0.056498970836400986,
0.01029857899993658,
0.06194264814257622,
-0.1462256908416748,
0.039592571556568146,
0.04102298617362976,
0.3222641050815582,
0.09464926272630692,
-0.31987690925598145,
-0.0960923582315445,
0.011048195883631706,
-0.04333033412694931,
-0.0388074591755867,
0.015209050849080086,
0.12290875613689423,
-0.11192797869443893,
0.04038098081946373,
-0.08408841490745544,
0.07311125099658966,
-0.05531281977891922,
0.0008828259888105094,
0.06136415898799896,
0.07097688317298889,
-0.032423555850982666,
0.061034802347421646,
-0.2688913941383362,
0.31325727701187134,
-0.005246073938906193,
0.06693460792303085,
-0.033683158457279205,
0.01103098876774311,
0.03642265498638153,
0.044865284115076065,
0.11345254629850388,
-0.006750098429620266,
-0.016546467319130898,
-0.19616566598415375,
-0.09673366695642471,
0.0013773217797279358,
0.14273542165756226,
-0.1359352022409439,
0.1344047337770462,
-0.020434102043509483,
-0.022626329213380814,
0.049730829894542694,
-0.05435692146420479,
-0.07615739107131958,
-0.07693483680486679,
0.016174191609025,
-0.04851147532463074,
0.09301113337278366,
-0.11193064600229263,
-0.10280857235193253,
-0.05614318326115608,
0.16082169115543365,
-0.08610668778419495,
-0.023284196853637695,
-0.14482073485851288,
0.08164215832948685,
0.11638359725475311,
-0.07414192706346512,
0.0487777516245842,
0.012150624766945839,
0.09813187271356583,
0.006800789386034012,
0.01985934004187584,
0.12790992856025696,
-0.07529273629188538,
-0.24359376728534698,
-0.07340109348297119,
0.1643402874469757,
0.04767289385199547,
0.06165962293744087,
-0.020337877795100212,
0.012956966646015644,
0.0092060761526227,
-0.08144957572221756,
0.054815590381622314,
0.004962868522852659,
0.06702413409948349,
0.04342614859342575,
-0.047975022345781326,
0.07514025270938873,
-0.07194451987743378,
-0.06208045035600662,
0.1304566115140915,
0.3341014087200165,
-0.09605240821838379,
0.011811847798526287,
0.04898189380764961,
-0.03425712510943413,
-0.18282222747802734,
0.05107518285512924,
0.11070577800273895,
0.03608643263578415,
0.008653547614812851,
-0.1824527233839035,
0.0459582544863224,
0.10539684444665909,
-0.033473435789346695,
0.1082499697804451,
-0.308626651763916,
-0.13372308015823364,
0.07228982448577881,
0.12110522389411926,
-0.0061678639613091946,
-0.17537540197372437,
-0.05783936381340027,
-0.013882112689316273,
-0.08036921173334122,
0.04554173722863197,
-0.05054996535181999,
0.11889444291591644,
-0.009596078656613827,
0.016744118183851242,
0.02771838940680027,
-0.06388910114765167,
0.1408374309539795,
0.00148108194116503,
0.07918327301740646,
-0.025209898129105568,
-0.0061545222997665405,
0.02446240745484829,
-0.09568937122821808,
0.0016326154582202435,
-0.06597421318292618,
0.0360775850713253,
-0.10459030419588089,
-0.025860249996185303,
-0.09399683773517609,
0.03571382537484169,
-0.06308530271053314,
-0.07563599199056625,
-0.020689038559794426,
0.06913311779499054,
0.05954167619347572,
-0.0007040438940748572,
0.11463450640439987,
-0.04473469778895378,
0.1808130443096161,
0.07638783752918243,
0.1001967191696167,
-0.0038610550109297037,
-0.027296148240566254,
0.006767478305846453,
-0.018628787249326706,
0.04606995731592178,
-0.16271570324897766,
0.010049713775515556,
0.14161530137062073,
0.04779312014579773,
0.13814783096313477,
0.07453161478042603,
-0.054145026952028275,
-0.003070964477956295,
0.08509357273578644,
-0.10699799656867981,
-0.09981728345155716,
-0.016803283244371414,
-0.004288865718990564,
-0.15469825267791748,
0.05846714600920677,
0.1023048609495163,
-0.05937797576189041,
-0.00496553722769022,
0.0029440487269312143,
0.015171940438449383,
-0.03800489008426666,
0.21698476374149323,
0.06434078514575958,
0.1071673035621643,
-0.07024383544921875,
0.0764748752117157,
0.03904002904891968,
-0.13006560504436493,
0.002746624406427145,
0.09261035919189453,
-0.09180807322263718,
-0.017514148727059364,
0.035619962960481644,
0.06920240819454193,
-0.010389581322669983,
-0.014269297011196613,
-0.13615691661834717,
-0.12403371930122375,
0.0624946728348732,
0.13329921662807465,
0.04687168076634407,
0.039114031940698624,
-0.007045170292258263,
0.04627292603254318,
-0.13724574446678162,
0.1180795207619667,
0.07167172431945801,
0.09823071956634521,
-0.15469194948673248,
0.17069844901561737,
-0.015785278752446175,
0.018625833094120026,
-0.006836122367531061,
0.02722197212278843,
-0.12116333842277527,
0.010288044810295105,
-0.08774534612894058,
-0.0743904560804367,
-0.05015110969543457,
-0.024260349571704865,
-0.012647783383727074,
-0.03758352994918823,
-0.015757760033011436,
-0.007113379426300526,
-0.1103193610906601,
-0.05298372730612755,
-0.0058640409260988235,
0.04197394847869873,
-0.09265466034412384,
-0.03410939872264862,
0.028666824102401733,
-0.11579620093107224,
0.09790012985467911,
0.02254354953765869,
0.055717818439006805,
0.007217543199658394,
-0.08719049394130707,
0.04829676076769829,
0.02535640448331833,
-0.03499305248260498,
0.035675350576639175,
-0.13096533715724945,
-0.01965995691716671,
-0.06982916593551636,
0.02397266775369644,
0.0203285813331604,
0.013294990174472332,
-0.13939903676509857,
0.015977909788489342,
-0.037332408130168915,
-0.04366481304168701,
-0.06814129650592804,
0.04839589446783066,
0.04513385891914368,
0.009635450318455696,
0.13742737472057343,
-0.07355690002441406,
0.05345592647790909,
-0.21947528421878815,
-0.020100027322769165,
-0.02093590423464775,
-0.07430737465620041,
-0.06056556478142738,
-0.031020740047097206,
0.09030469506978989,
-0.06114991009235382,
0.047635722905397415,
-0.05746043846011162,
0.06682378798723221,
0.02913726679980755,
-0.11421322077512741,
0.08171005547046661,
0.055536553263664246,
0.17864370346069336,
0.058314621448516846,
-0.03743729367852211,
0.05648672953248024,
0.05342580005526543,
0.07373485714197159,
0.08592306077480316,
0.19374683499336243,
0.14105406403541565,
0.007968307472765446,
0.09534938633441925,
0.03102513588964939,
-0.1110210195183754,
-0.1541706621646881,
0.07314055413007736,
-0.03505276143550873,
0.08264201879501343,
-0.03079644776880741,
0.18582677841186523,
0.13449834287166595,
-0.19853556156158447,
0.01753038913011551,
-0.041448671370744705,
-0.09444457292556763,
-0.09082191437482834,
-0.0404319129884243,
-0.06626821309328079,
-0.17376352846622467,
0.0028563092928379774,
-0.10460999608039856,
0.00831969827413559,
0.09920202195644379,
0.026967335492372513,
0.031220313161611557,
0.1847940981388092,
0.05994366481900215,
0.03177899122238159,
0.10401062667369843,
0.02888602204620838,
0.01401999220252037,
-0.04005848616361618,
-0.1134238988161087,
0.008570685051381588,
-0.05836963653564453,
0.029734386131167412,
-0.08070791512727737,
-0.09870156645774841,
0.055001188069581985,
0.040454138070344925,
-0.10748230665922165,
0.023134827613830566,
0.015767911449074745,
0.0649607852101326,
0.07898681610822678,
0.013658653013408184,
-0.012024866417050362,
-0.02646794356405735,
0.2742142975330353,
-0.1111980527639389,
-0.056831490248441696,
-0.11641282588243484,
0.27378931641578674,
0.024126814678311348,
0.004807177931070328,
0.00666216854006052,
-0.10406158864498138,
0.021564120426774025,
0.16947075724601746,
0.16913212835788727,
-0.05730978772044182,
-0.010973780415952206,
0.023305807262659073,
-0.016855154186487198,
-0.043159641325473785,
0.08007316291332245,
0.11910310387611389,
0.037991978228092194,
-0.07807090878486633,
-0.028206054121255875,
-0.02099498361349106,
-0.059245746582746506,
-0.02706841006875038,
0.07366928458213806,
0.04228077828884125,
0.021213842555880547,
-0.036017827689647675,
0.11586713790893555,
-0.04159124195575714,
-0.13957782089710236,
0.06374640762805939,
-0.19778349995613098,
-0.1784490942955017,
-0.060258377343416214,
0.019480273127555847,
-0.0050139655359089375,
0.07331284880638123,
-0.003197163576260209,
-0.02950623258948326,
0.08898305147886276,
-0.002756356028839946,
-0.020961115136742592,
-0.11304021626710892,
0.06816526502370834,
-0.060707367956638336,
0.19851577281951904,
-0.06321528553962708,
-0.018349429592490196,
0.13989515602588654,
0.02769191563129425,
-0.091034434735775,
0.040359608829021454,
0.09435314685106277,
-0.08921998739242554,
0.04423314332962036,
0.16402000188827515,
-0.03744226694107056,
0.11586513370275497,
0.047006718814373016,
-0.15157300233840942,
0.0275814738124609,
-0.08123806118965149,
-0.0596969872713089,
-0.09166955947875977,
0.004653610289096832,
-0.022848570719361305,
0.14669978618621826,
0.23772218823432922,
-0.06787397712469101,
0.025132369250059128,
-0.05768513306975365,
0.004722267854958773,
0.06028497964143753,
0.10006479918956757,
-0.030735470354557037,
-0.26279494166374207,
0.01312173530459404,
0.04721770063042641,
-0.000679207150824368,
-0.2755245566368103,
-0.09116538614034653,
0.03274394944310188,
-0.054978031665086746,
-0.077473483979702,
0.10464753955602646,
0.05649590119719505,
0.0574377179145813,
-0.047777459025382996,
-0.09687840938568115,
-0.0709979459643364,
0.198967844247818,
-0.17839039862155914,
-0.07672049105167389
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# audio_classification
This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the minds14 dataset.
It achieves the following results on the evaluation set:
- Loss: 2.6585
- Accuracy: 0.0973
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 3e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| No log | 1.0 | 29 | 2.6509 | 0.0796 |
| No log | 2.0 | 58 | 2.6585 | 0.0973 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["minds14"], "metrics": ["accuracy"], "base_model": "facebook/wav2vec2-base", "model-index": [{"name": "audio_classification", "results": [{"task": {"type": "audio-classification", "name": "Audio Classification"}, "dataset": {"name": "minds14", "type": "minds14", "config": "en-US", "split": "train", "args": "en-US"}, "metrics": [{"type": "accuracy", "value": 0.09734513274336283, "name": "Accuracy"}]}]}]} | audio-classification | rendy-k/audio_classification | [
"transformers",
"tensorboard",
"safetensors",
"wav2vec2",
"audio-classification",
"generated_from_trainer",
"dataset:minds14",
"base_model:facebook/wav2vec2-base",
"license:apache-2.0",
"model-index",
"endpoints_compatible",
"region:us"
] | 2024-02-08T06:57:41+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #wav2vec2 #audio-classification #generated_from_trainer #dataset-minds14 #base_model-facebook/wav2vec2-base #license-apache-2.0 #model-index #endpoints_compatible #region-us
| audio\_classification
=====================
This model is a fine-tuned version of facebook/wav2vec2-base on the minds14 dataset.
It achieves the following results on the evaluation set:
* Loss: 2.6585
* Accuracy: 0.0973
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 3e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 2
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #wav2vec2 #audio-classification #generated_from_trainer #dataset-minds14 #base_model-facebook/wav2vec2-base #license-apache-2.0 #model-index #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
77,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #wav2vec2 #audio-classification #generated_from_trainer #dataset-minds14 #base_model-facebook/wav2vec2-base #license-apache-2.0 #model-index #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.12484735250473022,
0.11707527190446854,
-0.0012715125922113657,
0.09490085393190384,
0.1295379251241684,
-0.0024933998938649893,
0.1438623070716858,
0.09460745751857758,
-0.0638435110449791,
0.04617027938365936,
0.11434158682823181,
0.08452221006155014,
0.03081836923956871,
0.1387488842010498,
-0.05481557175517082,
-0.1849554181098938,
0.02917948178946972,
0.02513938769698143,
-0.06582895666360855,
0.12042973190546036,
0.08179737627506256,
-0.11497028172016144,
0.08804105967283249,
0.0018190565751865506,
-0.16075395047664642,
0.003118857042863965,
0.023724820464849472,
-0.05280587077140808,
0.1175159439444542,
0.03549659997224808,
0.09389273077249527,
0.025608401745557785,
0.07190589606761932,
-0.1921335756778717,
0.019052842631936073,
0.07904773950576782,
-0.00407566549256444,
0.07974561303853989,
0.058502305299043655,
0.007259179838001728,
0.018399327993392944,
-0.07569289952516556,
0.035539235919713974,
0.03834354504942894,
-0.1341593712568283,
-0.2646881937980652,
-0.08395806699991226,
0.04601430147886276,
0.1024370789527893,
0.09555376321077347,
-0.02703631855547428,
0.12205881625413895,
-0.009372501634061337,
0.09276803582906723,
0.205942764878273,
-0.27436262369155884,
-0.06252714991569519,
0.04778366908431053,
0.04727894067764282,
0.1020747646689415,
-0.09944163262844086,
0.007437779568135738,
0.0695761889219284,
0.031592562794685364,
0.14710216224193573,
-0.02991085685789585,
-0.01975225657224655,
-0.014613354578614235,
-0.13405297696590424,
-0.03642401099205017,
0.19762279093265533,
0.10011976957321167,
-0.05720123276114464,
-0.051279742270708084,
-0.061910197138786316,
-0.1319425106048584,
-0.03821627423167229,
-0.009740347974002361,
0.044274620711803436,
-0.013130935840308666,
-0.07289864867925644,
-0.01139412634074688,
-0.09429187327623367,
-0.0888318419456482,
-0.00989790540188551,
0.10066209733486176,
0.003220942337065935,
0.007559556979686022,
-0.021766291931271553,
0.07702869921922684,
-0.031755588948726654,
-0.15359890460968018,
0.022626403719186783,
0.027127036824822426,
-0.03300001099705696,
-0.03617917373776436,
-0.037240900099277496,
-0.0754561647772789,
0.017647920176386833,
0.09845907986164093,
-0.05503833293914795,
0.055284708738327026,
-0.003626354970037937,
0.05462656170129776,
-0.0799015685915947,
0.12907525897026062,
-0.06921348720788956,
-0.07897325605154037,
0.013140331953763962,
0.09645111858844757,
0.048897404223680496,
-0.017821792513132095,
-0.12125254422426224,
0.03244252875447273,
0.1114007830619812,
-0.00006155335722723976,
-0.041490569710731506,
0.03951474279165268,
-0.07912972569465637,
-0.026407968252897263,
0.03793230280280113,
-0.06794566661119461,
0.015301770530641079,
0.010295256972312927,
-0.025250719860196114,
-0.056547824293375015,
0.014124181121587753,
0.04523270204663277,
0.027465760707855225,
0.07481551170349121,
-0.08675540238618851,
-0.01618194952607155,
-0.0589228980243206,
-0.09748534858226776,
0.03343575820326805,
-0.0732872262597084,
0.041117675602436066,
-0.09739820659160614,
-0.17667700350284576,
-0.011043649166822433,
0.08003029972314835,
-0.015598388388752937,
-0.0598619319498539,
-0.043604373931884766,
-0.08084779977798462,
0.018383175134658813,
-0.030779516324400902,
0.1052766814827919,
-0.08090746402740479,
0.09274644404649734,
0.03365014120936394,
0.062382910400629044,
-0.05182626470923424,
0.05034879967570305,
-0.07644517719745636,
0.05852871388196945,
-0.18525394797325134,
0.027509111911058426,
-0.06872102618217468,
0.06446599215269089,
-0.10879678279161453,
-0.06553438305854797,
-0.016143139451742172,
0.0068389554508030415,
0.06600474566221237,
0.07883749902248383,
-0.19277918338775635,
-0.08258301764726639,
0.1488567441701889,
-0.08845541626214981,
-0.16625969111919403,
0.1471027433872223,
-0.03511517867445946,
0.0432281456887722,
0.06875938922166824,
0.22382284700870514,
0.08816584199666977,
-0.11347917467355728,
-0.0008893265039660037,
-0.01523788832128048,
0.0693274587392807,
-0.05534506216645241,
0.10489212721586227,
0.0008448170265182853,
-0.00048607157077640295,
-0.0058913519605994225,
-0.03744779899716377,
0.06416469067335129,
-0.06319485604763031,
-0.08420638740062714,
-0.028099732473492622,
-0.1235804483294487,
0.03260624781250954,
0.046210113912820816,
0.021205704659223557,
-0.10174384713172913,
-0.08212054520845413,
0.04769457131624222,
0.09279783070087433,
-0.06138506159186363,
0.010023345239460468,
-0.07845748215913773,
0.09105327725410461,
-0.12092927098274231,
-0.028774816542863846,
-0.14682647585868835,
-0.002518940716981888,
-0.0005665787612088025,
-0.022082611918449402,
0.0016547825653105974,
0.03306004777550697,
0.06786010414361954,
0.06554245203733444,
-0.06926064193248749,
-0.06722510606050491,
-0.03157452493906021,
0.0325588658452034,
-0.08952900767326355,
-0.21256743371486664,
-0.02479131706058979,
-0.039062242954969406,
0.14649680256843567,
-0.21988530457019806,
0.020786283537745476,
0.024400213733315468,
0.08181865513324738,
0.06039430946111679,
-0.0338049940764904,
-0.000944136583711952,
0.05677226185798645,
-0.009739921428263187,
-0.0669734999537468,
0.05142473801970482,
0.026172857731580734,
-0.09545303881168365,
-0.02447533793747425,
-0.13993960618972778,
0.22258631885051727,
0.12710615992546082,
-0.0462193563580513,
-0.04322937875986099,
0.03181549906730652,
-0.03541313856840134,
-0.04016917943954468,
-0.05022453889250755,
-0.02244318276643753,
0.13265854120254517,
0.012783650308847427,
0.13469763100147247,
-0.10240486264228821,
-0.023211240768432617,
0.05787345767021179,
-0.039361484348773956,
-0.0025428859516978264,
0.1014920100569725,
0.04883628338575363,
-0.13585348427295685,
0.16097693145275116,
0.20705869793891907,
-0.05515507981181145,
0.16508418321609497,
-0.06004953756928444,
-0.0730825886130333,
-0.03667179495096207,
-0.012445389293134212,
0.0015894024400040507,
0.16098496317863464,
-0.1252424269914627,
0.019908098503947258,
0.013408299535512924,
0.040232185274362564,
-0.0001725157635519281,
-0.19685178995132446,
-0.03483724966645241,
0.050469622015953064,
-0.05644255131483078,
-0.05173502862453461,
-0.010141673497855663,
-0.03022344596683979,
0.07272329926490784,
0.0005631073727272451,
-0.07423438876867294,
0.045044079422950745,
-0.013512113131582737,
-0.08976589888334274,
0.16876469552516937,
-0.07856021076440811,
-0.16728734970092773,
-0.14647726714611053,
-0.06213754042983055,
-0.05202651768922806,
0.028412001207470894,
0.07013647258281708,
-0.0752338096499443,
-0.04063744843006134,
-0.09246137738227844,
0.010201902128756046,
0.021326260641217232,
0.01870453730225563,
0.06704574823379517,
0.0022263340651988983,
0.1251649558544159,
-0.08597058802843094,
0.003820964600890875,
-0.0018213981529697776,
-0.001831083674915135,
-0.0013934893067926168,
0.02751181274652481,
0.10970520228147507,
0.12966112792491913,
0.02533911168575287,
0.0024728430435061455,
-0.013426016084849834,
0.27761927247047424,
-0.10319187492132187,
0.0008125572348944843,
0.15034763514995575,
-0.04172156751155853,
0.03764481842517853,
0.15155655145645142,
0.05794214829802513,
-0.10385149717330933,
0.006799089722335339,
0.04175226390361786,
-0.03687591850757599,
-0.20565713942050934,
-0.026991549879312515,
-0.04552484676241875,
0.0066283089108765125,
0.06641057133674622,
0.022173607721924782,
0.03343679755926132,
0.0614020898938179,
0.019616466015577316,
0.06276588886976242,
-0.01085413247346878,
0.04751039296388626,
0.05876016616821289,
0.035002924501895905,
0.11127398163080215,
-0.051250722259283066,
-0.014879907481372356,
0.04308967664837837,
0.03139646723866463,
0.18478956818580627,
0.05377580225467682,
0.16205966472625732,
0.06813372671604156,
0.13487643003463745,
0.00454677501693368,
0.05884324759244919,
-0.016628297045826912,
-0.04675677418708801,
-0.017596621066331863,
-0.048749472945928574,
-0.02677755057811737,
0.029429595917463303,
-0.05245594307780266,
0.06521552801132202,
-0.10054267197847366,
0.013258371502161026,
0.04359909147024155,
0.26077282428741455,
0.06659664958715439,
-0.3255167007446289,
-0.10322301834821701,
0.035270534455776215,
-0.02852296456694603,
-0.04336320981383324,
0.034417930990457535,
0.16911642253398895,
-0.0397203154861927,
0.05102132260799408,
-0.058852292597293854,
0.09496147930622101,
-0.04135970398783684,
0.04766126349568367,
0.06884796172380447,
0.06621910631656647,
-0.01340487040579319,
0.059502504765987396,
-0.23460602760314941,
0.2854408621788025,
0.034133430570364,
0.08380155265331268,
-0.03354845568537712,
-0.010915695689618587,
0.02568569965660572,
0.08309552073478699,
0.11150070279836655,
0.001090604462660849,
-0.0736638605594635,
-0.1838684231042862,
-0.07027561962604523,
0.019981762394309044,
0.06437736749649048,
-0.035249706357717514,
0.0784880667924881,
-0.03159720078110695,
-0.006033327430486679,
0.07916506379842758,
0.022756481543183327,
-0.08963575959205627,
-0.07696188241243362,
-0.018466204404830933,
0.08285519480705261,
0.007195105776190758,
-0.08891604840755463,
-0.08861405402421951,
-0.1379554271697998,
0.10051674395799637,
-0.0332503616809845,
-0.015676887705922127,
-0.10115478187799454,
0.033405277878046036,
0.062344830483198166,
-0.06183331832289696,
0.04374389722943306,
0.023014549165964127,
0.1216469258069992,
0.018999313935637474,
-0.045308008790016174,
0.12198583036661148,
-0.09019625186920166,
-0.1661914587020874,
-0.06569938361644745,
0.1457890272140503,
0.032617539167404175,
0.05129947513341904,
0.0012325544375926256,
0.015208469703793526,
0.0009087700163945556,
-0.05820893868803978,
0.06816229969263077,
-0.016389193013310432,
0.03974852338433266,
-0.01536358892917633,
0.011279935948550701,
-0.008950544521212578,
-0.04820192605257034,
-0.05133151262998581,
0.13067041337490082,
0.29569530487060547,
-0.0637117251753807,
0.026017697528004646,
0.09549218416213989,
-0.050317153334617615,
-0.17834267020225525,
0.06027880311012268,
0.043513957411050797,
0.014818063005805016,
0.07812421023845673,
-0.15452256798744202,
0.10679788142442703,
0.069767527282238,
-0.034738022834062576,
0.0921747088432312,
-0.26526373624801636,
-0.11656726151704788,
0.11022122204303741,
0.140851691365242,
0.11417040973901749,
-0.13953030109405518,
-0.0433858297765255,
-0.057819664478302,
-0.156254842877388,
0.09911740571260452,
-0.18820065259933472,
0.10423354804515839,
0.020579587668180466,
0.05389147996902466,
0.00219038431532681,
-0.04657748341560364,
0.13689671456813812,
0.0004064989334437996,
0.10695254802703857,
-0.0446910485625267,
0.03294522315263748,
0.08122421056032181,
-0.055926937609910965,
0.0015719548100605607,
-0.10930556803941727,
0.02850429341197014,
-0.06743570417165756,
-0.029403697699308395,
-0.06036437675356865,
0.01301483903080225,
-0.0359654575586319,
-0.05660069361329079,
-0.021320605650544167,
0.04636826366186142,
0.05269988253712654,
-0.017357245087623596,
0.14883124828338623,
0.003268853295594454,
0.11785639822483063,
0.14353781938552856,
0.09774080663919449,
-0.08162819594144821,
-0.06907051801681519,
-0.011872506700456142,
-0.05213188752532005,
0.07252702862024307,
-0.1549125760793686,
0.05782744660973549,
0.09085146337747574,
0.008258596993982792,
0.14164607226848602,
0.049779485911130905,
-0.06337147206068039,
0.009526395238935947,
0.06608428806066513,
-0.15865810215473175,
-0.15278086066246033,
-0.031026378273963928,
-0.006137753836810589,
-0.11916018277406693,
0.03826194256544113,
0.1303533911705017,
-0.05731617286801338,
0.016277261078357697,
0.0012007573386654258,
0.01783422939479351,
-0.07670629769563675,
0.18039223551750183,
0.06251095235347748,
0.06317508220672607,
-0.08993331342935562,
0.09236009418964386,
0.032016925513744354,
-0.10235415399074554,
0.021910546347498894,
0.0005689472309313715,
-0.08395929634571075,
-0.04286607354879379,
0.0411781445145607,
0.18974202871322632,
0.013595358468592167,
-0.08007659018039703,
-0.12410491704940796,
-0.11455550044775009,
0.04093264043331146,
0.17032262682914734,
0.0815248191356659,
0.008266513235867023,
-0.022015277296304703,
0.006581773050129414,
-0.09438001364469528,
0.12403330206871033,
0.0372127965092659,
0.07360292226076126,
-0.213899627327919,
0.06596450507640839,
-0.003229133551940322,
0.010275183245539665,
-0.024919399991631508,
0.04100212827324867,
-0.10334865748882294,
0.017918085679411888,
-0.14876747131347656,
-0.007795917801558971,
-0.0323900543153286,
0.007560146041214466,
-0.00898180715739727,
-0.06574473530054092,
-0.07628552615642548,
0.02508343942463398,
-0.09952674061059952,
-0.017984846606850624,
0.04932618886232376,
0.06288067251443863,
-0.11477616429328918,
-0.03284217417240143,
0.03663058951497078,
-0.07372353225946426,
0.06339894980192184,
0.01346275769174099,
0.013608230277895927,
0.037806469947099686,
-0.1639920324087143,
0.01502749603241682,
0.06288458406925201,
0.010295753367245197,
0.04831678047776222,
-0.13824237883090973,
-0.02626296691596508,
-0.007877803407609463,
0.03479447215795517,
0.008508067578077316,
0.09564485400915146,
-0.09917541593313217,
-0.017046796157956123,
-0.018035301938652992,
-0.03693811222910881,
-0.04985484480857849,
0.018655534833669662,
0.1235344186425209,
0.024795975536108017,
0.2192874401807785,
-0.07434778660535812,
0.011780946515500546,
-0.211531862616539,
0.018356330692768097,
-0.0026497526559978724,
-0.12010350078344345,
-0.09246644377708435,
-0.04940519481897354,
0.045868437737226486,
-0.07525834441184998,
0.11899318546056747,
-0.018964799121022224,
0.03589117154479027,
0.032071541994810104,
-0.05944383889436722,
0.022031603381037712,
0.025914842262864113,
0.2259315848350525,
0.010815308429300785,
-0.026509493589401245,
0.03591792285442352,
0.017209835350513458,
0.10884202271699905,
0.13762271404266357,
0.13133157789707184,
0.186716690659523,
-0.027895187959074974,
0.08618831634521484,
0.05331837013363838,
-0.04873025417327881,
-0.16269566118717194,
0.06499011069536209,
-0.06238187849521637,
0.11428090929985046,
-0.0015092840185388923,
0.19320803880691528,
0.09910731762647629,
-0.1741790622472763,
0.030220966786146164,
-0.07081635296344757,
-0.06759047508239746,
-0.0928296446800232,
-0.09562905877828598,
-0.09165035933256149,
-0.15688104927539825,
-0.0019510629354044795,
-0.11087345331907272,
0.015167728997766972,
0.10290968418121338,
-0.001813331269659102,
-0.014517945237457752,
0.1944192498922348,
-0.002118578879162669,
0.0253287460654974,
0.05889308452606201,
0.0005119132110849023,
-0.07258614152669907,
-0.06754174828529358,
-0.07602675259113312,
0.053963396698236465,
-0.02318691276013851,
0.021754726767539978,
-0.06135357543826103,
-0.05666846036911011,
0.05266382545232773,
0.0026063406839966774,
-0.11915493756532669,
0.007270761299878359,
0.010323849506676197,
0.0387762226164341,
0.05496960133314133,
0.021868126466870308,
0.017962845042347908,
0.011001713573932648,
0.21362406015396118,
-0.0764354020357132,
-0.03644097223877907,
-0.12950436770915985,
0.12980718910694122,
-0.027648335322737694,
0.0012812039349228144,
0.0059320623986423016,
-0.09054621309041977,
0.03664901480078697,
0.21576432883739471,
0.16272079944610596,
-0.09030483663082123,
-0.0007937264163047075,
-0.025579214096069336,
-0.007305496837943792,
-0.06483174860477448,
0.07962843030691147,
0.10563119500875473,
-0.012459482066333294,
-0.07128473371267319,
-0.03357226401567459,
-0.06798898428678513,
-0.03237640857696533,
0.0021927962079644203,
0.07202836871147156,
0.02719135954976082,
0.002268813084810972,
-0.04065809026360512,
0.09345653653144836,
-0.05323449522256851,
-0.137278750538826,
0.011092192493379116,
-0.17542901635169983,
-0.15018457174301147,
-0.042398739606142044,
0.11037079989910126,
0.01533514354377985,
0.030683349817991257,
-0.03708893060684204,
0.01237449236214161,
0.07097878307104111,
-0.01550906989723444,
-0.04278058186173439,
-0.0935019850730896,
0.05803387239575386,
-0.10262289643287659,
0.216199591755867,
-0.04581008106470108,
0.034735117107629776,
0.10988343507051468,
0.046289097517728806,
-0.08722569048404694,
0.08429776877164841,
0.045406464487314224,
-0.06800892949104309,
0.0032570320181548595,
0.1320948451757431,
-0.04487133026123047,
0.1497625708580017,
0.049482159316539764,
-0.1277311146259308,
-0.008988561108708382,
-0.050278499722480774,
-0.06545261293649673,
-0.040058434009552,
-0.04385743662714958,
-0.063344307243824,
0.13933296501636505,
0.13457724452018738,
-0.055025432258844376,
-0.013290191069245338,
-0.05028654634952545,
0.04285867512226105,
0.08577297627925873,
0.030888661742210388,
-0.021984931081533432,
-0.24559873342514038,
0.026708705350756645,
0.04546472057700157,
-0.00922396406531334,
-0.27034279704093933,
-0.08730417490005493,
-0.012363002635538578,
-0.053453925997018814,
-0.06746400892734528,
0.07264126092195511,
0.1021534875035286,
0.04672139510512352,
-0.07088383287191391,
-0.03489166125655174,
-0.06713555008172989,
0.15753528475761414,
-0.1219107136130333,
-0.10119491070508957
] |
null | null | transformers | # MiquMaid-v2-70B-DPO 3.5bpw
## Description
Exllama quant of [NeverSleep/MiquMaid-v2-70B-DPO](https://huggingface.co/NeverSleep/MiquMaid-v2-70B-DPO)
## Other quants:
EXL2: [4bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-4bpw-exl2), [3.5bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-3.5bpw-exl2), [3bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-3bpw-exl2), [2.4bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-2.4bpw-exl2), [2.3bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-2.3bpw-exl2)
2.4bpw is probably the most you can fit in a 24gb card
GGUF:
[2bit Imatrix GGUF](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-Imatrix-GGUF)
## Prompt format: Alpaca
```
### Instruction:
{system prompt}
### Input:
{input}
### Response:
{reply}
```
## Contact
Kooten on discord
[ko-fi.com/kooten](https://ko-fi.com/kooten) | {"license": "cc-by-nc-4.0", "tags": ["not-for-all-audiences", "nsfw"]} | text-generation | Kooten/MiquMaid-v2-70B-DPO-3.5bpw-exl2 | [
"transformers",
"pytorch",
"llama",
"text-generation",
"not-for-all-audiences",
"nsfw",
"conversational",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T06:59:42+00:00 | [] | [] | TAGS
#transformers #pytorch #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # MiquMaid-v2-70B-DPO 3.5bpw
## Description
Exllama quant of NeverSleep/MiquMaid-v2-70B-DPO
## Other quants:
EXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw
2.4bpw is probably the most you can fit in a 24gb card
GGUF:
2bit Imatrix GGUF
## Prompt format: Alpaca
## Contact
Kooten on discord
URL | [
"# MiquMaid-v2-70B-DPO 3.5bpw",
"## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B-DPO",
"## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF",
"## Prompt format: Alpaca",
"## Contact\nKooten on discord\n\nURL"
] | [
"TAGS\n#transformers #pytorch #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# MiquMaid-v2-70B-DPO 3.5bpw",
"## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B-DPO",
"## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF",
"## Prompt format: Alpaca",
"## Contact\nKooten on discord\n\nURL"
] | [
74,
17,
24,
60,
8,
7
] | [
"passage: TAGS\n#transformers #pytorch #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# MiquMaid-v2-70B-DPO 3.5bpw## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B-DPO## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF## Prompt format: Alpaca## Contact\nKooten on discord\n\nURL"
] | [
-0.04582490026950836,
0.0684068575501442,
-0.006411169655621052,
0.08963239938020706,
0.047497086226940155,
0.026873638853430748,
0.12043806165456772,
0.143168643116951,
0.013703133910894394,
0.006306761410087347,
0.06044897064566612,
0.12961165606975555,
0.06286982446908951,
0.16042953729629517,
-0.04037059098482132,
-0.15921349823474884,
0.015951594337821007,
0.06383323669433594,
0.09659018367528915,
0.12402684986591339,
0.06875807791948318,
-0.025508197024464607,
0.048755645751953125,
-0.031528495252132416,
-0.12988236546516418,
-0.003468544688075781,
-0.05160415545105934,
-0.0651833713054657,
0.04912799224257469,
0.09295079112052917,
0.030017156153917313,
0.08990386873483658,
-0.034092508256435394,
-0.17658768594264984,
0.022853920236229897,
0.0326879657804966,
-0.051645342260599136,
0.035571277141571045,
0.04587722197175026,
-0.04887495934963226,
0.038535572588443756,
0.02824425883591175,
-0.07422705739736557,
0.07308534532785416,
-0.10205641388893127,
-0.08794225752353668,
-0.053595904260873795,
0.14069323241710663,
0.06790199875831604,
0.04086223244667053,
-0.007748959120362997,
0.112639881670475,
-0.03812000900506973,
0.07493887841701508,
0.22947010397911072,
-0.24385492503643036,
-0.022439002990722656,
0.08317763358354568,
0.039192330092191696,
0.05493713170289993,
-0.06554973125457764,
0.005218737758696079,
0.011566444300115108,
-0.005284345708787441,
0.021744443103671074,
-0.03344579041004181,
0.0493793822824955,
-0.023433683440089226,
-0.09265179932117462,
-0.0105742821469903,
0.15324871242046356,
0.06043320149183273,
-0.04642539098858833,
-0.03394947946071625,
-0.07827083766460419,
-0.128587007522583,
-0.03959805518388748,
0.03271324560046196,
-0.002936763223260641,
-0.01560176257044077,
-0.04460400342941284,
0.00013509001291822642,
-0.10542383790016174,
-0.04201354458928108,
-0.14006704092025757,
0.16337531805038452,
0.0014881024835631251,
0.02625935897231102,
0.004008527845144272,
0.047091949731111526,
-0.08856236189603806,
-0.08503749966621399,
-0.08470926433801651,
-0.0897606834769249,
0.03980621322989464,
-0.0021973031107336283,
-0.009887533262372017,
0.10075703263282776,
0.07867159694433212,
0.1741325706243515,
-0.027983633801341057,
0.07548202574253082,
0.03069010190665722,
0.06500085443258286,
0.0031937523745000362,
-0.07913748919963837,
-0.03440814092755318,
-0.04152001440525055,
0.03154759109020233,
0.014269739389419556,
0.045309558510780334,
-0.022522415965795517,
-0.11244446039199829,
-0.02087460458278656,
-0.008135161362588406,
0.02145155519247055,
0.040778160095214844,
0.0524558387696743,
-0.062202244997024536,
-0.06274721771478653,
0.18581002950668335,
-0.04827360808849335,
0.021300364285707474,
0.05829671397805214,
-0.004439160227775574,
0.10424323379993439,
0.047731515020132065,
-0.0035930145531892776,
-0.057569172233343124,
0.05589073896408081,
-0.033560652285814285,
0.0272708460688591,
-0.02400396578013897,
-0.050247952342033386,
0.09007219225168228,
-0.06337643414735794,
-0.014024478383362293,
-0.1705516129732132,
-0.1355183869600296,
0.06971427798271179,
-0.004046582616865635,
-0.058166228234767914,
-0.03931889310479164,
0.0022773866076022387,
-0.018291745334863663,
-0.00838204100728035,
-0.060626156628131866,
0.034805603325366974,
-0.07005590945482254,
0.08743729442358017,
0.07600720226764679,
0.12460678815841675,
-0.11840873956680298,
0.0019584933761507273,
-0.03880602866411209,
0.026164934039115906,
-0.07519033551216125,
0.01780175045132637,
-0.11506011337041855,
-0.027500716969370842,
-0.08192585408687592,
-0.05028565600514412,
-0.02886795438826084,
0.03794778883457184,
-0.004283525515347719,
0.10768861323595047,
-0.05667106434702873,
-0.057809725403785706,
0.2422356903553009,
-0.09024515748023987,
-0.06726867705583572,
0.1266731470823288,
0.05558012053370476,
-0.08360251784324646,
0.048102956265211105,
0.0708654448390007,
0.11421101540327072,
-0.09976281970739365,
-0.07826036959886551,
0.08152056485414505,
-0.010199741460382938,
-0.08231049031019211,
0.09392035007476807,
0.05199035629630089,
0.03676558658480644,
0.03334128111600876,
0.1575973927974701,
0.03342241048812866,
-0.0028626136481761932,
-0.06767885386943817,
-0.05103132873773575,
-0.061980921775102615,
0.07522516697645187,
0.007884600199759007,
-0.028491560369729996,
-0.09562642872333527,
-0.11036937683820724,
-0.0017878146609291434,
0.07671914249658585,
-0.003546590218320489,
-0.007503082510083914,
-0.1196042150259018,
0.09845466911792755,
-0.08503146469593048,
0.03068271465599537,
-0.07490643858909607,
0.0019682422280311584,
-0.030514175072312355,
-0.006882324814796448,
0.08002013713121414,
0.10305605828762054,
0.07230190932750702,
0.03384450823068619,
-0.07829535007476807,
0.03472060710191727,
0.0948038175702095,
0.0033809212036430836,
-0.004210072103887796,
-0.06718618422746658,
0.027233513072133064,
-0.012968409806489944,
0.1175948902964592,
-0.07787806540727615,
0.025900082662701607,
0.13329587876796722,
0.10208495706319809,
-0.026521610096096992,
-0.034858588129282,
0.06685624271631241,
0.032981112599372864,
-0.04253131151199341,
0.016709787771105766,
0.08756591379642487,
-0.018490871414542198,
-0.015111759305000305,
0.09264140576124191,
-0.21026116609573364,
0.11715582758188248,
0.17362883687019348,
-0.06294393539428711,
-0.030599232763051987,
-0.06301114708185196,
-0.008627192117273808,
-0.02563450112938881,
0.011065253056585789,
-0.055589038878679276,
0.07276608794927597,
-0.014822155237197876,
0.12653279304504395,
-0.08420581370592117,
-0.053510136902332306,
0.0533212274312973,
-0.04417143762111664,
0.0018029269995167851,
0.0679749995470047,
0.10939477384090424,
-0.11373770236968994,
0.15822981297969818,
0.14058485627174377,
-0.08779773116111755,
0.13361763954162598,
0.027351167052984238,
-0.03043307363986969,
-0.032340843230485916,
0.015464241616427898,
0.00789677258580923,
0.056895770132541656,
-0.11704664677381516,
0.05522652715444565,
0.0800691768527031,
-0.011647751554846764,
0.025949109345674515,
-0.1300257295370102,
-0.04105745255947113,
0.021214362233877182,
-0.0695265457034111,
-0.08277267217636108,
-0.0016601737588644028,
0.005137088242918253,
0.101523257791996,
-0.015549113042652607,
-0.001983784371986985,
0.03867767006158829,
0.04322611540555954,
-0.07111696898937225,
0.14121945202350616,
-0.09647230803966522,
-0.30733025074005127,
-0.07857482135295868,
-0.14495520293712616,
-0.08292830735445023,
0.016102999448776245,
0.09353078901767731,
-0.06979130953550339,
-0.03137896582484245,
-0.048107802867889404,
-0.045265909284353256,
-0.04490061104297638,
-0.011240028776228428,
0.010262654162943363,
-0.01788639836013317,
0.056656625121831894,
-0.08015856146812439,
-0.020541833713650703,
0.05825645849108696,
-0.04098339378833771,
0.1668710857629776,
-0.024308543652296066,
0.09128880500793457,
0.13803304731845856,
0.017689021304249763,
-0.01584150642156601,
-0.03321511298418045,
0.15794293582439423,
-0.06563503295183182,
0.04938054829835892,
0.14132972061634064,
-0.008156443014740944,
0.07705025374889374,
0.09128060191869736,
0.04067555442452431,
-0.05199521780014038,
-0.05198335647583008,
0.030381476506590843,
-0.0981254130601883,
-0.14651194214820862,
-0.037517156451940536,
-0.06446903944015503,
0.14517949521541595,
0.05526777356863022,
0.060085415840148926,
0.06875264644622803,
0.11728721857070923,
-0.07688433676958084,
0.0008447429863736033,
0.03042089380323887,
0.09071183204650879,
0.16828912496566772,
0.021669529378414154,
0.15510448813438416,
-0.06039077416062355,
-0.03671281412243843,
0.12675373256206512,
0.13252462446689606,
0.04581773653626442,
0.0076622795313596725,
0.1067720577120781,
0.057230591773986816,
0.015584813430905342,
0.09663423895835876,
0.04717424511909485,
0.012874344363808632,
-0.032283734530210495,
-0.04044756665825844,
-0.05916476994752884,
-0.007050851359963417,
0.04156402125954628,
0.07063524425029755,
-0.07718320190906525,
0.030122652649879456,
-0.08814846724271774,
0.06716170161962509,
0.06671981513500214,
0.08936896175146103,
-0.2088191658258438,
-0.028593817725777626,
0.04828054830431938,
-0.027664581313729286,
-0.04099032282829285,
0.019330671057105064,
0.0027588882949203253,
-0.022044865414500237,
0.16534073650836945,
-0.009597979485988617,
0.09387727826833725,
-0.07612869143486023,
0.02398321032524109,
-0.03314380720257759,
0.05616947263479233,
0.0026691912207752466,
0.11348195374011993,
-0.29423680901527405,
0.07459088414907455,
0.022921649739146233,
0.0013921256177127361,
-0.070274218916893,
-0.009420517832040787,
0.013431424275040627,
-0.037070389837026596,
0.08954539149999619,
0.011687802150845528,
0.11529877036809921,
-0.13591837882995605,
-0.06850113719701767,
0.017961157485842705,
0.05855780839920044,
0.03528587892651558,
0.03812882676720619,
0.050028253346681595,
-0.04115259274840355,
-0.011492578312754631,
-0.0028778850100934505,
-0.039288230240345,
-0.09423855692148209,
0.12312605232000351,
0.12090452015399933,
-0.021107815206050873,
0.014487851411104202,
-0.033293504267930984,
-0.19208550453186035,
0.22321012616157532,
-0.1170232743024826,
-0.09947218745946884,
-0.06058977171778679,
-0.05823701620101929,
0.020056450739502907,
-0.036763016134500504,
0.019995251670479774,
-0.054087672382593155,
0.08842375129461288,
-0.03354021534323692,
-0.08400633186101913,
0.08057116717100143,
-0.10512647032737732,
-0.13273866474628448,
-0.03524647280573845,
0.15523095428943634,
-0.07117729634046555,
0.03926239162683487,
0.01936325989663601,
-0.013390879146754742,
-0.04495241492986679,
-0.1269797384738922,
0.10327374935150146,
0.006790191400796175,
-0.0036693215370178223,
0.017119446769356728,
0.017938394099473953,
-0.012902448885142803,
-0.03939798101782799,
-0.034646034240722656,
0.14153218269348145,
0.35773348808288574,
-0.03405236825346947,
0.11649525910615921,
0.11892581731081009,
-0.039593406021595,
-0.21185939013957977,
-0.16374672949314117,
-0.03986865654587746,
-0.08046954870223999,
-0.05837932229042053,
-0.13859760761260986,
0.0410170704126358,
0.13381896913051605,
-0.03437136486172676,
0.18987716734409332,
-0.14507335424423218,
-0.09870345145463943,
0.04097047075629234,
0.0551617406308651,
0.243429496884346,
-0.1582370549440384,
-0.05891342833638191,
-0.05965219438076019,
-0.20286250114440918,
0.22145499289035797,
-0.03591753542423248,
0.1256483644247055,
-0.03981955349445343,
0.09790444374084473,
-0.002509930869564414,
-0.059963829815387726,
0.12346718460321426,
-0.02461555041372776,
-0.028667019680142403,
-0.11712814122438431,
0.045186419039964676,
0.0065001375041902065,
-0.006960918661206961,
0.011699987575411797,
-0.17834196984767914,
-0.028735652565956116,
-0.13756883144378662,
-0.060157280415296555,
-0.07768747210502625,
0.018088795244693756,
-0.03221731260418892,
-0.053302861750125885,
-0.05346861109137535,
-0.002044514287263155,
0.013892869465053082,
-0.022210154682397842,
-0.013020721264183521,
-0.07299872487783432,
0.07618267089128494,
0.30049827694892883,
0.09260325878858566,
-0.07857368886470795,
-0.09534958004951477,
-0.06770896911621094,
-0.0947115570306778,
0.021391406655311584,
-0.11831045895814896,
0.03503655642271042,
0.07924559712409973,
0.008232727646827698,
0.062048621475696564,
0.044595230370759964,
-0.059853482991456985,
0.019759217277169228,
0.12042315304279327,
-0.13908174633979797,
0.004425274673849344,
-0.013285956345498562,
0.06379378587007523,
0.0004541476664599031,
-0.010039888322353363,
0.13431501388549805,
-0.008667916990816593,
-0.013672774657607079,
0.058194033801555634,
0.017729058861732483,
-0.018992755562067032,
0.11546573042869568,
0.03496812656521797,
0.07397865504026413,
-0.13057760894298553,
0.06279075145721436,
-0.011849385686218739,
-0.0590694397687912,
-0.010049685835838318,
0.10481002926826477,
-0.07038460671901703,
-0.12835773825645447,
-0.017328836023807526,
-0.08210008591413498,
-0.10034860670566559,
-0.03863706812262535,
-0.024345340207219124,
-0.11874531209468842,
0.06606581807136536,
0.12094057351350784,
0.02172795496881008,
-0.024814993143081665,
0.06406383961439133,
0.04073449969291687,
-0.06960268318653107,
0.059185344725847244,
-0.00878099724650383,
0.08285075426101685,
-0.10113184154033661,
0.0546618290245533,
-0.006787142250686884,
0.04524298384785652,
-0.04187369719147682,
0.011183404363691807,
-0.12183726578950882,
-0.028870420530438423,
-0.13848167657852173,
0.05508766323328018,
-0.11466217786073685,
-0.004002262372523546,
-0.02024228312075138,
-0.0004070240247529,
-0.0284191332757473,
-0.039480697363615036,
-0.09090229123830795,
0.005418331827968359,
0.021178821101784706,
-0.006525960750877857,
-0.093484066426754,
-0.026871223002672195,
0.027927404269576073,
-0.023714831098914146,
0.07242493331432343,
0.05014222487807274,
-0.05375358834862709,
-0.019420063123106956,
-0.15223382413387299,
-0.03314598649740219,
0.07887186855077744,
0.07142087817192078,
-0.0245599914342165,
0.025776753202080727,
0.04539560154080391,
0.04111924767494202,
0.03631201013922691,
0.030261505395174026,
0.07549787312746048,
-0.09010141342878342,
-0.016714155673980713,
-0.06997295469045639,
-0.09816251695156097,
-0.05176657810807228,
-0.009239599108695984,
0.12638480961322784,
-0.03132988139986992,
0.1059076264500618,
-0.0489514134824276,
0.027195315808057785,
-0.13752956688404083,
0.020726293325424194,
0.029659492895007133,
-0.1493890881538391,
-0.008127909153699875,
-0.05683813989162445,
0.03150832653045654,
-0.028273679316043854,
0.1645122617483139,
-0.07818209379911423,
-0.10317167639732361,
0.02819615975022316,
-0.06963564455509186,
-0.03598532825708389,
-0.018087593838572502,
0.19610115885734558,
0.07746971398591995,
-0.012387366034090519,
-0.09268632531166077,
0.028744330629706383,
0.05596093833446503,
-0.07497544586658478,
0.08811493217945099,
0.08290284872055054,
-0.0293477401137352,
0.04391024634242058,
0.064759261906147,
-0.07300133258104324,
-0.08227050304412842,
0.004626152105629444,
-0.1207580491900444,
0.10451983660459518,
-0.02028585970401764,
0.08787745982408524,
0.08778759092092514,
-0.041231583803892136,
0.01808958314359188,
-0.06765864044427872,
-0.03377069532871246,
-0.11065781861543655,
-0.08681938797235489,
-0.08246675133705139,
-0.09624631702899933,
0.03003651276230812,
-0.07718263566493988,
0.03152571618556976,
0.05982353538274765,
0.04855962097644806,
-0.016038045287132263,
0.08917171508073807,
-0.01532548014074564,
-0.03290318325161934,
0.011603114195168018,
-0.025753306224942207,
-0.02083616703748703,
0.030904343351721764,
-0.033984798938035965,
0.06716516613960266,
-0.014273342676460743,
0.045488156378269196,
0.025014327839016914,
0.018197204917669296,
-0.0030380445532500744,
-0.06790199875831604,
-0.09326570481061935,
-0.03524523228406906,
0.035950079560279846,
0.07916548103094101,
0.15833036601543427,
-0.03399674966931343,
0.014815525151789188,
0.01596844010055065,
0.07704873383045197,
-0.026769990101456642,
-0.08346682041883469,
-0.036198049783706665,
0.1967553347349167,
-0.08492948859930038,
0.051365457475185394,
-0.014073158614337444,
-0.021390628069639206,
-0.009151048958301544,
0.24680779874324799,
0.1558638960123062,
-0.03880663588643074,
0.03142015263438225,
0.0449141189455986,
0.022672144696116447,
0.07627435773611069,
0.09049862623214722,
0.080573670566082,
0.18151302635669708,
-0.050170548260211945,
0.0064696972258389,
0.02926088310778141,
0.0005708857788704336,
-0.07937124371528625,
0.058487117290496826,
0.01898510940372944,
-0.018463321030139923,
-0.06325871497392654,
0.021527735516428947,
-0.12836682796478271,
0.07136579602956772,
-0.005507428664714098,
-0.11806054413318634,
-0.07548294961452484,
0.005051296204328537,
0.050737399607896805,
0.04634450748562813,
0.011956246569752693,
-0.030381200835108757,
-0.009805100969970226,
0.02013942413032055,
-0.019740281626582146,
-0.17666365206241608,
0.05412799492478371,
0.005300190765410662,
-0.0005348154809325933,
0.028260236606001854,
-0.04000202566385269,
0.15599164366722107,
0.10570362210273743,
0.008327197283506393,
-0.05073821544647217,
0.1309603899717331,
0.019429882988333702,
-0.16900981962680817,
0.03996788710355759,
0.12783561646938324,
-0.03354188799858093,
0.0007112044258974493,
0.09201359003782272,
-0.06345468759536743,
-0.050123605877161026,
0.10716459155082703,
-0.029041798785328865,
-0.017352398484945297,
0.06418813765048981,
-0.05377950146794319,
0.07307108491659164,
0.06561677902936935,
-0.049810558557510376,
-0.07091116905212402,
-0.04481380060315132,
0.053485672920942307,
-0.0244421549141407,
-0.055541422218084335,
-0.057107239961624146,
-0.1694939285516739,
-0.052250444889068604,
0.0535535030066967,
0.12531651556491852,
-0.24392783641815186,
-0.005307360086590052,
-0.08672872185707092,
0.00007102619565557688,
-0.12976598739624023,
0.06058349460363388,
0.09938614070415497,
0.01175694353878498,
-0.051241058856248856,
-0.10182222723960876,
-0.020083986222743988,
0.0793035551905632,
-0.12403896450996399,
-0.07143314182758331
] |
null | null | null |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# outputs
This model is a fine-tuned version of [tiiuae/falcon-7b-instruct](https://huggingface.co/tiiuae/falcon-7b-instruct) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 4
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 100
- training_steps: 200
- mixed_precision_training: Native AMP
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "tiiuae/falcon-7b-instruct", "model-index": [{"name": "outputs", "results": []}]} | null | madhiarasan/outputs | [
"tensorboard",
"safetensors",
"generated_from_trainer",
"base_model:tiiuae/falcon-7b-instruct",
"license:apache-2.0",
"region:us"
] | 2024-02-08T07:00:02+00:00 | [] | [] | TAGS
#tensorboard #safetensors #generated_from_trainer #base_model-tiiuae/falcon-7b-instruct #license-apache-2.0 #region-us
|
# outputs
This model is a fine-tuned version of tiiuae/falcon-7b-instruct on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 4
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 100
- training_steps: 200
- mixed_precision_training: Native AMP
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| [
"# outputs\n\nThis model is a fine-tuned version of tiiuae/falcon-7b-instruct on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 4\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 16\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 100\n- training_steps: 200\n- mixed_precision_training: Native AMP",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#tensorboard #safetensors #generated_from_trainer #base_model-tiiuae/falcon-7b-instruct #license-apache-2.0 #region-us \n",
"# outputs\n\nThis model is a fine-tuned version of tiiuae/falcon-7b-instruct on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 4\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 16\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 100\n- training_steps: 200\n- mixed_precision_training: Native AMP",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
46,
33,
6,
12,
8,
3,
139,
4,
33
] | [
"passage: TAGS\n#tensorboard #safetensors #generated_from_trainer #base_model-tiiuae/falcon-7b-instruct #license-apache-2.0 #region-us \n# outputs\n\nThis model is a fine-tuned version of tiiuae/falcon-7b-instruct on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 4\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 16\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 100\n- training_steps: 200\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.12358558177947998,
0.1481574922800064,
-0.0028361729346215725,
0.0815410315990448,
0.11599026620388031,
0.03526788204908371,
0.09433604031801224,
0.14443708956241608,
-0.017918305471539497,
0.09506421536207199,
0.0841808095574379,
0.018607012927532196,
0.06941516697406769,
0.17237073183059692,
-0.014405608177185059,
-0.22827079892158508,
0.014832638204097748,
-0.05790885165333748,
-0.07396478950977325,
0.10119909048080444,
0.09843865036964417,
-0.08179072290658951,
0.05726584419608116,
-0.0012898591812700033,
-0.11184774339199066,
-0.007052178028970957,
-0.04024030268192291,
-0.05459512025117874,
0.09416410326957703,
-0.009748686105012894,
0.06600181013345718,
0.012576139532029629,
0.1214815154671669,
-0.21351757645606995,
-0.0005664888303726912,
0.08088859915733337,
0.055312782526016235,
0.10003315657377243,
0.07087728381156921,
0.003466640366241336,
0.04550972953438759,
-0.14204424619674683,
0.10072901099920273,
0.028541594743728638,
-0.0750749483704567,
-0.1451718509197235,
-0.10262355208396912,
0.06797538697719574,
0.09690530598163605,
0.07209686934947968,
0.013543505221605301,
0.1028401181101799,
-0.0785364955663681,
0.0554509162902832,
0.17967963218688965,
-0.26331818103790283,
-0.05052526667714119,
0.027292072772979736,
0.061510078608989716,
0.070132777094841,
-0.11555181443691254,
-0.024246059358119965,
0.036184195429086685,
0.009934795089066029,
0.10524017363786697,
0.015776898711919785,
-0.04255658760666847,
0.004081118851900101,
-0.12209203839302063,
-0.033388007432222366,
0.12078936398029327,
0.05402424931526184,
-0.05993184447288513,
-0.15286573767662048,
-0.03786647692322731,
-0.13274578750133514,
-0.015943801030516624,
0.001909784972667694,
0.020176541060209274,
-0.043845050036907196,
-0.03540018945932388,
-0.04318471997976303,
-0.05555637925863266,
-0.07589996606111526,
0.03487063944339752,
0.11400501430034637,
0.0431194044649601,
0.016622483730316162,
0.01221088133752346,
0.1162349209189415,
0.009213458746671677,
-0.11395590752363205,
-0.002953333081677556,
-0.018957363441586494,
-0.13478310406208038,
-0.0373060442507267,
-0.013194169849157333,
0.018329720944166183,
0.005863973405212164,
0.145616814494133,
-0.023060033097863197,
0.07239240407943726,
0.034620314836502075,
0.0024239502381533384,
-0.012112504802644253,
0.14009793102741241,
-0.060618139803409576,
-0.043154262006282806,
-0.008756178431212902,
0.13610006868839264,
0.012586467899382114,
-0.023488959297537804,
-0.07506869733333588,
-0.030220404267311096,
0.0705891102552414,
0.09268010407686234,
-0.027664827182888985,
0.011076501570641994,
-0.05922393500804901,
-0.022934984415769577,
0.07444655895233154,
-0.13281409442424774,
0.048037104308605194,
0.004442840348929167,
-0.08454302698373795,
-0.047514740377664566,
0.004729656036943197,
-0.000057537043176125735,
-0.0355624258518219,
0.060824811458587646,
-0.07430149614810944,
-0.032356295734643936,
-0.06727854162454605,
-0.04371747002005577,
0.0075277406722307205,
-0.10494373738765717,
-0.008678685873746872,
-0.0470530167222023,
-0.17864103615283966,
-0.03063208796083927,
0.04705839976668358,
-0.09435819834470749,
-0.06901001930236816,
-0.022228429093956947,
-0.06049812212586403,
0.03243926540017128,
-0.011672437191009521,
0.15035605430603027,
-0.051640354096889496,
0.070429727435112,
-0.026154058054089546,
0.037731923162937164,
0.06849723309278488,
0.043111976236104965,
-0.0669761374592781,
0.06070813536643982,
-0.12379426509141922,
0.08559264987707138,
-0.09805528819561005,
0.02044680528342724,
-0.14020229876041412,
-0.06879456341266632,
0.026464873924851418,
-0.0295029915869236,
0.07444620132446289,
0.1398102343082428,
-0.16942821443080902,
0.004735699389129877,
0.1289624720811844,
-0.05522622540593147,
-0.07288940995931625,
0.09292426705360413,
-0.030776113271713257,
0.004181841388344765,
0.027932120487093925,
0.15319201350212097,
0.10268940031528473,
-0.15947121381759644,
0.013113327324390411,
0.03742020204663277,
0.06774044781923294,
0.012072434648871422,
0.07484222203493118,
-0.017351100221276283,
0.009466231800615788,
0.012223551981151104,
-0.0511556938290596,
0.008070237003266811,
-0.06477570533752441,
-0.07656771689653397,
-0.047458574175834656,
-0.092702217400074,
0.04104682058095932,
-0.004194187931716442,
0.0034931490663439035,
-0.05410807952284813,
-0.129741370677948,
0.0060617076233029366,
0.14817020297050476,
-0.05316200479865074,
-0.0001105849732994102,
-0.07493560016155243,
0.0404120534658432,
-0.019876670092344284,
-0.027576494961977005,
-0.1715870350599289,
-0.11546476185321808,
0.05948691815137863,
-0.08126591145992279,
0.01255062222480774,
-0.019204629585146904,
0.055183157324790955,
0.06688185781240463,
-0.029984764754772186,
-0.04080260172486305,
-0.05873556807637215,
-0.011226397939026356,
-0.0897429808974266,
-0.18673260509967804,
-0.07363313436508179,
-0.04216018691658974,
0.19558119773864746,
-0.22367551922798157,
0.0013061914360150695,
-0.028231633827090263,
0.12208143621683121,
0.033520136028528214,
-0.06236379221081734,
0.03388630971312523,
0.011467482894659042,
0.008869335986673832,
-0.1198221743106842,
0.024184079840779305,
0.008460861630737782,
-0.11560416221618652,
-0.0026686200872063637,
-0.13105539977550507,
-0.0028233928605914116,
0.05086659640073776,
0.1768837720155716,
-0.10772453993558884,
-0.08525989949703217,
-0.06702983379364014,
-0.05343690142035484,
-0.060553621500730515,
0.00708638085052371,
0.1644420027732849,
0.04706478863954544,
0.10774825513362885,
-0.07230013608932495,
-0.08292564004659653,
0.012406463734805584,
0.01567353494465351,
-0.01873977854847908,
0.09167446196079254,
0.05644431337714195,
-0.08850088715553284,
0.06659470498561859,
0.08856799453496933,
-0.019206784665584564,
0.10963776707649231,
-0.05481959506869316,
-0.10917539149522781,
-0.02118157595396042,
0.04655403271317482,
0.005016314331442118,
0.15099908411502838,
-0.042935729026794434,
0.026138849556446075,
0.04656393826007843,
0.010134049691259861,
0.02342759631574154,
-0.15778318047523499,
-0.015585117042064667,
0.015029528178274632,
-0.04903750494122505,
-0.03327646479010582,
-0.019733011722564697,
0.028498755767941475,
0.07448666542768478,
0.033201806247234344,
-0.02502969279885292,
0.019794417545199394,
-0.016222557052969933,
-0.06715324521064758,
0.16855794191360474,
-0.11265269666910172,
-0.1396743655204773,
-0.12072620540857315,
0.04965854808688164,
-0.020444409921765327,
-0.03947575017809868,
-0.023886913433670998,
-0.056354787200689316,
-0.061924226582050323,
-0.11086885631084442,
-0.042455416172742844,
-0.030602263286709785,
-0.00591091625392437,
0.03025231882929802,
0.02120542898774147,
0.09037787467241287,
-0.11368139833211899,
0.02445182017982006,
-0.0023325402289628983,
-0.0582507885992527,
0.0011583002051338553,
0.06418745219707489,
0.0747973769903183,
0.11873253434896469,
0.006912962533533573,
0.014154423959553242,
-0.03046380914747715,
0.21290244162082672,
-0.09190760552883148,
0.023487374186515808,
0.10996916890144348,
-0.0016235212096944451,
0.06445260345935822,
0.1429356336593628,
0.021684234961867332,
-0.08775200694799423,
0.029192568734288216,
0.05107254162430763,
-0.013410251587629318,
-0.25446954369544983,
-0.0326092354953289,
-0.0304682906717062,
-0.04083666577935219,
0.12139714509248734,
0.05769159272313118,
-0.035501811653375626,
0.044550903141498566,
-0.024300143122673035,
0.007943343371152878,
-0.012765287421643734,
0.07137361913919449,
0.015227784402668476,
0.03993895277380943,
0.07020529359579086,
-0.011370951309800148,
-0.0028936059679836035,
0.0664815753698349,
0.06672744452953339,
0.2609007954597473,
-0.04965199530124664,
0.1607070118188858,
-0.0011343079386278987,
0.16688115894794464,
-0.021131744608283043,
0.05633149668574333,
0.004059026017785072,
0.005264569539576769,
-0.0018031193176284432,
-0.05882655829191208,
-0.01712816208600998,
0.035304099321365356,
0.010938317514955997,
0.027050955221056938,
-0.07945220917463303,
0.048892490565776825,
0.0007463894435204566,
0.27247148752212524,
0.060836270451545715,
-0.29007434844970703,
-0.06896112114191055,
0.017573434859514236,
-0.03894722834229469,
-0.09804033488035202,
0.01418171264231205,
0.13960979878902435,
-0.13316774368286133,
0.04490472748875618,
-0.07038997858762741,
0.07754424214363098,
-0.056254930794239044,
-0.02587171457707882,
0.0697610154747963,
0.14264680445194244,
-0.009934663772583008,
0.09174997359514236,
-0.14839290082454681,
0.20448240637779236,
0.01427463348954916,
0.08358744531869888,
-0.06950752437114716,
0.042981307953596115,
-0.005475680809468031,
0.030833333730697632,
0.1302979737520218,
-0.0019229318713769317,
-0.05142099782824516,
-0.16181492805480957,
-0.14378900825977325,
0.016765208914875984,
0.11805252730846405,
-0.05161800608038902,
0.0771837830543518,
-0.055724095553159714,
-0.009969659149646759,
0.026013977825641632,
-0.11322801560163498,
-0.13210880756378174,
-0.15653283894062042,
0.027400068938732147,
-0.006551031023263931,
-0.01989026553928852,
-0.0905541330575943,
-0.08442658185958862,
0.0041048419661819935,
0.15043668448925018,
-0.01223783940076828,
-0.056903865188360214,
-0.1554747074842453,
0.028177136555314064,
0.1930294781923294,
-0.04393091797828674,
0.034391649067401886,
0.000039075897802831605,
0.15215691924095154,
0.024155547842383385,
-0.0570361465215683,
0.06858097016811371,
-0.07363882660865784,
-0.18311795592308044,
-0.06750483065843582,
0.15215086936950684,
0.04920034483075142,
0.0426626056432724,
-0.009929545223712921,
0.012657309882342815,
0.00008890373283065856,
-0.08487104624509811,
0.03638375177979469,
0.05580689013004303,
0.03869796171784401,
0.05209130048751831,
-0.056295234709978104,
0.0551835261285305,
-0.035655297338962555,
-0.029699454084038734,
0.10301230847835541,
0.23647791147232056,
-0.09355658292770386,
0.11653342097997665,
0.07471121847629547,
-0.0424531064927578,
-0.1586245596408844,
0.023701516911387444,
0.11574563384056091,
0.04060142859816551,
0.04401077330112457,
-0.17795869708061218,
0.07954870164394379,
0.09909331798553467,
-0.03259889408946037,
0.08472497016191483,
-0.3163225054740906,
-0.12099102884531021,
0.06795087456703186,
0.07606247067451477,
-0.03297276049852371,
-0.11318887770175934,
-0.0521085225045681,
-0.018739929422736168,
-0.064348503947258,
0.09029003232717514,
-0.049707625061273575,
0.10925047099590302,
-0.009217713959515095,
0.08941475301980972,
0.0400041863322258,
-0.04293833673000336,
0.1498217135667801,
0.01605798862874508,
0.06895909458398819,
-0.038252539932727814,
0.033640772104263306,
0.03592507913708687,
-0.08387730270624161,
0.05692538246512413,
-0.07331225275993347,
0.07603804767131805,
-0.14048299193382263,
-0.009316932410001755,
-0.054732102900743484,
0.052570488303899765,
-0.0506269671022892,
-0.056068070232868195,
-0.034913696348667145,
0.07658412307500839,
0.05357597395777702,
-0.03849700465798378,
0.042199134826660156,
0.03528190031647682,
0.05380832403898239,
0.11607932299375534,
0.0724262148141861,
0.008618474937975407,
-0.10809160023927689,
0.0061385962180793285,
-0.014648283831775188,
0.05019594356417656,
-0.1192307397723198,
0.017728079110383987,
0.10547086596488953,
0.0436977781355381,
0.11635709553956985,
0.012182096019387245,
-0.09012223780155182,
-0.01153769250959158,
0.036502905189991,
-0.10277045518159866,
-0.1427367776632309,
0.00432777963578701,
0.019205693155527115,
-0.15421240031719208,
-0.0073091513477265835,
0.09557913988828659,
-0.03962130844593048,
-0.023291336372494698,
-0.011367700062692165,
0.041102003306150436,
0.004109154921025038,
0.17352059483528137,
0.04588991776108742,
0.08071355521678925,
-0.07991132140159607,
0.1383599489927292,
0.08162000775337219,
-0.08194100111722946,
0.047263212502002716,
0.06851954758167267,
-0.09057359397411346,
-0.0060852388851344585,
0.06290483474731445,
0.07853981852531433,
0.017032533884048462,
-0.031107226386666298,
-0.04332719370722771,
-0.09429571032524109,
0.048611778765916824,
0.040656767785549164,
0.015033050440251827,
-0.0052826483733952045,
-0.01840614527463913,
0.011187728494405746,
-0.13599807024002075,
0.10612490028142929,
0.061311669647693634,
0.06439578533172607,
-0.14370249211788177,
0.058279942721128464,
-0.0018178238533437252,
0.03754038363695145,
-0.00794994831085205,
0.02355380542576313,
-0.08336935192346573,
-0.023624321445822716,
-0.09899020195007324,
0.016321200877428055,
-0.030280176550149918,
0.016053413972258568,
-0.014901114627718925,
-0.037829816341400146,
-0.015005718916654587,
0.03219263628125191,
-0.06555993854999542,
-0.06777353584766388,
-0.00822826661169529,
0.05545492842793465,
-0.12554305791854858,
-0.010438020341098309,
0.0312928631901741,
-0.11582429707050323,
0.10570847243070602,
0.041192445904016495,
0.034001726657152176,
0.006424612365663052,
-0.0940796434879303,
0.005364434327930212,
0.0021692789159715176,
0.020786598324775696,
0.05940202251076698,
-0.12669648230075836,
-0.020692674443125725,
-0.05315324291586876,
-0.001945883734151721,
-0.003922124858945608,
0.023393958806991577,
-0.13678069412708282,
-0.04355524852871895,
-0.03442845866084099,
-0.04268251359462738,
-0.0683896467089653,
0.052677154541015625,
0.09691975265741348,
0.023325597867369652,
0.13519693911075592,
-0.051583681255578995,
0.04628133028745651,
-0.18252022564411163,
-0.0337858609855175,
-0.010061467066407204,
0.0028862780891358852,
-0.06294120103120804,
-0.004473901353776455,
0.08832681179046631,
-0.05212737247347832,
0.12100556492805481,
-0.035642918199300766,
0.07996934652328491,
0.016089480370283127,
-0.043210189789533615,
0.00011578006524359807,
0.02890698052942753,
0.1463354080915451,
0.054258860647678375,
-0.007850643247365952,
0.0864766389131546,
-0.04405389353632927,
0.03593166545033455,
0.06030290573835373,
0.14681167900562286,
0.13524256646633148,
0.022784283384680748,
0.053293220698833466,
0.06762415170669556,
-0.14073123037815094,
-0.1349247246980667,
0.13744531571865082,
-0.0365564338862896,
0.1040985956788063,
-0.052680812776088715,
0.1981840282678604,
0.08690586686134338,
-0.1868092268705368,
0.05063747987151146,
-0.047874629497528076,
-0.1173030361533165,
-0.09733117371797562,
-0.08457966893911362,
-0.061227839440107346,
-0.11033327132463455,
0.02185039035975933,
-0.10013851523399353,
0.04351188987493515,
0.07880620658397675,
0.015247967094182968,
0.02620142139494419,
0.12809878587722778,
-0.0023224775213748217,
0.007767773699015379,
0.07347695529460907,
0.05207260698080063,
0.005688159726560116,
-0.015419145114719868,
-0.07190646231174469,
0.04868297278881073,
-0.0014463787665590644,
0.07890766859054565,
-0.0431414395570755,
0.01681865006685257,
0.058084700256586075,
0.001487002125941217,
-0.06829819083213806,
0.021279141306877136,
-0.004817333072423935,
0.025136673822999,
0.041183952242136,
0.05946854501962662,
0.015008723363280296,
-0.0592740997672081,
0.29855892062187195,
-0.07416115701198578,
-0.06012321263551712,
-0.13982300460338593,
0.16498857736587524,
-0.008786074817180634,
-0.01972462609410286,
0.06105346232652664,
-0.11355874687433243,
-0.02874814346432686,
0.14444151520729065,
0.13918420672416687,
-0.08235939592123032,
-0.029574979096651077,
-0.020351162180304527,
-0.019544359296560287,
-0.06979760527610779,
0.13289956748485565,
0.08770646154880524,
0.027414938434958458,
-0.05720537155866623,
0.014561216346919537,
-0.0026121768169105053,
-0.04915403202176094,
-0.0845073014497757,
0.1068047285079956,
-0.004489647690206766,
0.010538444854319096,
-0.04456951841711998,
0.07113850116729736,
-0.0008219785522669554,
-0.20037177205085754,
0.016977911815047264,
-0.12694290280342102,
-0.20016269385814667,
-0.04063278064131737,
0.05543818324804306,
-0.013434997759759426,
0.05102825537323952,
-0.0015940777957439423,
-0.0008149670902639627,
0.12486368417739868,
-0.015609174966812134,
-0.03545529395341873,
-0.08686000108718872,
0.06981708109378815,
-0.03871632739901543,
0.2383890599012375,
0.011614104732871056,
0.04991347715258598,
0.09390527755022049,
0.009923743084073067,
-0.1776631772518158,
0.013855038210749626,
0.09480544924736023,
-0.058095257729291916,
0.045780543237924576,
0.17796528339385986,
-0.03809177130460739,
0.08341337740421295,
0.04187745600938797,
-0.07860460132360458,
-0.02146957255899906,
-0.06972470134496689,
-0.011888261884450912,
-0.08218063414096832,
-0.004017912317067385,
-0.045759785920381546,
0.17454157769680023,
0.20100513100624084,
-0.05501183494925499,
-0.0041144778952002525,
-0.06412743031978607,
0.01704859733581543,
0.0372953936457634,
0.08426505327224731,
-0.0006031871307641268,
-0.17394527792930603,
0.017260372638702393,
0.05416686460375786,
0.0430595837533474,
-0.21094602346420288,
-0.07697997987270355,
0.02084173634648323,
-0.05211976170539856,
-0.051265109330415726,
0.12877556681632996,
0.04938989505171776,
0.026807520538568497,
-0.030617740005254745,
-0.08471183478832245,
-0.03801259770989418,
0.128890261054039,
-0.1534092128276825,
-0.03204086422920227
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# phi-1_5-query_classify
This model is a fine-tuned version of [microsoft/phi-1_5](https://huggingface.co/microsoft/phi-1_5) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 4
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- training_steps: 1000
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.38.0.dev0
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1 | {"license": "mit", "library_name": "peft", "tags": ["generated_from_trainer"], "base_model": "microsoft/phi-1_5", "model-index": [{"name": "phi-1_5-query_classify", "results": []}]} | null | akashAD/phi-1_5-query_classify | [
"peft",
"tensorboard",
"safetensors",
"generated_from_trainer",
"base_model:microsoft/phi-1_5",
"license:mit",
"region:us"
] | 2024-02-08T07:00:07+00:00 | [] | [] | TAGS
#peft #tensorboard #safetensors #generated_from_trainer #base_model-microsoft/phi-1_5 #license-mit #region-us
|
# phi-1_5-query_classify
This model is a fine-tuned version of microsoft/phi-1_5 on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 4
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- training_steps: 1000
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.38.0.dev0
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1 | [
"# phi-1_5-query_classify\n\nThis model is a fine-tuned version of microsoft/phi-1_5 on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 4\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- training_steps: 1000",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#peft #tensorboard #safetensors #generated_from_trainer #base_model-microsoft/phi-1_5 #license-mit #region-us \n",
"# phi-1_5-query_classify\n\nThis model is a fine-tuned version of microsoft/phi-1_5 on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 4\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- training_steps: 1000",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
41,
35,
6,
12,
8,
3,
89,
4,
44
] | [
"passage: TAGS\n#peft #tensorboard #safetensors #generated_from_trainer #base_model-microsoft/phi-1_5 #license-mit #region-us \n# phi-1_5-query_classify\n\nThis model is a fine-tuned version of microsoft/phi-1_5 on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 4\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- training_steps: 1000### Training results### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.1615510731935501,
0.11363256722688675,
-0.0008821933879517019,
0.09566719830036163,
0.14523105323314667,
-0.009246927686035633,
0.09629359096288681,
0.10994727909564972,
-0.09863488376140594,
0.10374224185943604,
0.08602546155452728,
0.047842178493738174,
0.06764935702085495,
0.21236860752105713,
-0.01882587932050228,
-0.21801048517227173,
0.041186824440956116,
0.0023328259121626616,
0.022737465798854828,
0.11407031863927841,
0.08292990922927856,
-0.11576952785253525,
0.06781783699989319,
0.010305371135473251,
-0.14427369832992554,
-0.029690464958548546,
-0.008939347229897976,
-0.06024322658777237,
0.09249363839626312,
-0.016175543889403343,
0.1195589080452919,
0.010082188062369823,
0.11312621831893921,
-0.14818985760211945,
0.010158071294426918,
0.09404871612787247,
0.04581715539097786,
0.10752422362565994,
0.07226806879043579,
0.03403082862496376,
0.01628980040550232,
-0.10082179307937622,
0.07857857644557953,
0.03180054575204849,
-0.1044699028134346,
-0.20915710926055908,
-0.11073239147663116,
0.11415138840675354,
0.08983201533555984,
0.11181356757879257,
-0.0001541662059025839,
0.1870352178812027,
-0.04738970100879669,
0.06370321661233902,
0.2072947919368744,
-0.26918208599090576,
-0.064527228474617,
0.05124631151556969,
0.055613718926906586,
0.08523526787757874,
-0.08995269238948822,
-0.04903160780668259,
0.055010102689266205,
0.04855573922395706,
0.0960056260228157,
0.0013378263683989644,
-0.01863234117627144,
-0.06440120190382004,
-0.15397736430168152,
-0.057062067091464996,
0.13199865818023682,
0.036764226853847504,
-0.07178449630737305,
-0.0744800716638565,
-0.07432912290096283,
-0.16087166965007782,
-0.02324642241001129,
-0.0242958702147007,
0.017980609089136124,
-0.049501266330480576,
-0.03978309780359268,
-0.06041004881262779,
-0.09417301416397095,
-0.07639850676059723,
-0.0061302389949560165,
0.09829557687044144,
0.05870024114847183,
0.04022211208939552,
-0.014617553912103176,
0.1314803510904312,
-0.026309825479984283,
-0.09113579988479614,
-0.012883923947811127,
-0.023532738909125328,
-0.014390443451702595,
-0.0477379709482193,
-0.030639413744211197,
0.00852486677467823,
-0.004388263914734125,
0.09778818488121033,
-0.17034882307052612,
0.029746247455477715,
0.022944526746869087,
0.04806710034608841,
-0.04754291847348213,
0.11984865367412567,
-0.07823280990123749,
0.05893344059586525,
0.03707012161612511,
0.12488365918397903,
0.04330800473690033,
-0.001244239159859717,
-0.10485855489969254,
-0.025148462504148483,
0.14101353287696838,
0.06306739151477814,
-0.021808937191963196,
0.016384325921535492,
-0.054094281047582626,
-0.03516293317079544,
0.11607778072357178,
-0.1027805283665657,
0.03696194291114807,
0.0052397288382053375,
-0.07307971268892288,
-0.06854887306690216,
0.04777882248163223,
-0.0018462236039340496,
-0.07435869425535202,
0.04712766036391258,
-0.09080000221729279,
0.019557751715183258,
-0.09325828403234482,
-0.07823291420936584,
0.02208009362220764,
-0.10711617767810822,
-0.02173038199543953,
-0.1101226732134819,
-0.1910005360841751,
-0.04089648276567459,
-0.010817953385412693,
-0.04954845830798149,
-0.0540560819208622,
0.00917891040444374,
-0.09484127163887024,
-0.021419567987322807,
-0.044033296406269073,
0.105988048017025,
-0.05194670706987381,
0.10979009419679642,
0.004109618254005909,
0.01865437626838684,
-0.046375177800655365,
0.044667571783065796,
-0.0798732191324234,
0.03724447637796402,
-0.19315508008003235,
0.048726778477430344,
-0.08094339072704315,
0.022334640845656395,
-0.11193954944610596,
-0.12047360837459564,
-0.005396990571171045,
-0.03824881091713905,
0.09752203524112701,
0.108735091984272,
-0.13366441428661346,
-0.002109564607962966,
0.1567864865064621,
-0.10617217421531677,
-0.07415512949228287,
0.08675480633974075,
-0.003981044515967369,
0.010300121270120144,
0.047614529728889465,
0.17368826270103455,
0.0716778039932251,
-0.16457197070121765,
0.012690851464867592,
0.030905814841389656,
0.044245507568120956,
-0.09636881947517395,
0.07396493852138519,
-0.04233919829130173,
0.04339425638318062,
0.021382208913564682,
-0.06531542539596558,
0.017559465020895004,
-0.09114692360162735,
-0.06378578394651413,
-0.05407644435763359,
-0.09549058973789215,
0.0401589497923851,
0.0541083849966526,
0.03728621453046799,
-0.05536897853016853,
-0.08486833423376083,
0.09503006190061569,
0.12845651805400848,
-0.027849363163113594,
0.019795700907707214,
-0.08129817992448807,
0.12432229518890381,
-0.07046107202768326,
-0.03753142058849335,
-0.23061347007751465,
-0.08426029235124588,
0.04200529307126999,
-0.04738547280430794,
-0.013602808117866516,
-0.009190764278173447,
0.04069629684090614,
0.1015520840883255,
-0.02942429482936859,
-0.041424863040447235,
-0.10055655986070633,
-0.0021177444141358137,
-0.1605559140443802,
-0.16074037551879883,
-0.06596264988183975,
-0.01797088049352169,
0.15017084777355194,
-0.19009824097156525,
0.03799627721309662,
0.006734008435159922,
0.1336570829153061,
0.020699087530374527,
-0.05611489340662956,
0.017847085371613503,
0.08723629266023636,
0.01620316132903099,
-0.09545772522687912,
0.07461771368980408,
0.031004484742879868,
-0.04912382364273071,
-0.05646860599517822,
-0.15929904580116272,
0.07947196066379547,
0.10403513163328171,
0.06839440762996674,
-0.09127151221036911,
-0.024683741852641106,
-0.08616263419389725,
-0.03141503408551216,
-0.07989664375782013,
0.010740529745817184,
0.14881564676761627,
0.022939909249544144,
0.14742308855056763,
-0.09817824512720108,
-0.04797625541687012,
0.030341897159814835,
-0.012690600007772446,
0.01975492760539055,
0.08830220252275467,
0.04190545901656151,
-0.07280761748552322,
0.11957040429115295,
0.08653741329908371,
-0.051993418484926224,
0.07927294820547104,
-0.0712875947356224,
-0.10190853476524353,
-0.007392391562461853,
0.039974670857191086,
0.0017776759341359138,
0.17015910148620605,
-0.02479030378162861,
0.02241632342338562,
0.04324064031243324,
0.012933259829878807,
0.03342122212052345,
-0.21576376259326935,
-0.006305558141320944,
-0.014933092519640923,
-0.03578953072428703,
-0.02245386317372322,
-0.026555480435490608,
0.0381305068731308,
0.09357915818691254,
0.02549016661942005,
-0.03024127520620823,
0.026083260774612427,
0.01069187093526125,
-0.09073285758495331,
0.1926453411579132,
-0.09902931749820709,
-0.1687110811471939,
-0.12159693986177444,
0.11586078256368637,
-0.043545059859752655,
-0.016239013522863388,
0.015677038580179214,
-0.042848456650972366,
-0.03055519610643387,
-0.10956355929374695,
-0.04576503112912178,
-0.018581928685307503,
-0.010043608956038952,
-0.0024286750704050064,
0.014030049555003643,
0.12547554075717926,
-0.10003779828548431,
-0.0022083022631704807,
-0.03565163165330887,
-0.06093978136777878,
0.01816447637975216,
0.05404557287693024,
0.09101561456918716,
0.11808005720376968,
-0.024054400622844696,
0.012340213172137737,
-0.04672403261065483,
0.23165149986743927,
-0.08432191610336304,
-0.016278646886348724,
0.18360431492328644,
0.009222259745001793,
0.06209656968712807,
0.08928640931844711,
0.019435761496424675,
-0.07514718174934387,
0.008302946574985981,
0.029122499749064445,
-0.025658734142780304,
-0.2271510660648346,
-0.06310449540615082,
-0.01634078286588192,
0.0016868208767846227,
0.08716677129268646,
0.05056484788656235,
0.04716434329748154,
0.056064460426568985,
-0.037278912961483,
0.032389115542173386,
-0.03573337942361832,
0.08182288706302643,
0.07032747566699982,
0.011175793595612049,
0.10142554342746735,
-0.03173552826046944,
-0.00044680607970803976,
0.06147635728120804,
0.01910489611327648,
0.25089213252067566,
-0.04112295061349869,
0.06305665522813797,
0.04633601754903793,
0.2070508897304535,
0.003981775138527155,
0.04819915443658829,
0.030254976823925972,
0.008961541578173637,
0.019021963700652122,
-0.06058673933148384,
-0.04779527336359024,
0.004093740601092577,
-0.03974573686718941,
0.06206119805574417,
-0.12248626351356506,
0.017994241788983345,
-0.00075267389183864,
0.2976453900337219,
0.04754519462585449,
-0.35053595900535583,
-0.11517967283725739,
-0.021126996725797653,
-0.00018909023492597044,
-0.08305641263723373,
-0.005077267996966839,
0.13925495743751526,
-0.13883423805236816,
0.01301402784883976,
-0.06346399337053299,
0.09131178259849548,
-0.06306443363428116,
0.00946498941630125,
0.04146762937307358,
0.11873416602611542,
0.006974691990762949,
0.06418798863887787,
-0.1906488984823227,
0.22778914868831635,
0.022807221859693527,
0.1077699288725853,
-0.03361763060092926,
0.004431452136486769,
-0.000633072282653302,
0.12374240159988403,
0.10522547364234924,
-0.004542261827737093,
0.037338241934776306,
-0.21719656884670258,
-0.15264861285686493,
0.0005712254787795246,
0.0683022141456604,
0.000547126226592809,
0.06023719161748886,
-0.0186209324747324,
0.03727215901017189,
0.015204302035272121,
-0.060714587569236755,
-0.15704259276390076,
-0.05302274227142334,
0.011186208575963974,
0.002439098432660103,
-0.012858755886554718,
-0.10389739274978638,
-0.11024636775255203,
-0.020901434123516083,
0.1093829944729805,
-0.050744205713272095,
-0.08710804581642151,
-0.14045235514640808,
0.10509366542100906,
0.09534881263971329,
-0.05929609388113022,
0.013291758485138416,
-0.016787853091955185,
0.12198130041360855,
0.011709059588611126,
-0.05486679449677467,
0.07054220885038376,
-0.06178928539156914,
-0.20986968278884888,
-0.04285350814461708,
0.1265704333782196,
0.03879835456609726,
0.046434786170721054,
-0.008809187449514866,
-0.0022455754224210978,
-0.0031289176549762487,
-0.08872640877962112,
0.02654268592596054,
0.08170361071825027,
0.12270403653383255,
0.01573089137673378,
-0.05156216770410538,
0.0483645536005497,
-0.04142460972070694,
-0.005245942156761885,
0.10274682939052582,
0.19851958751678467,
-0.09618111699819565,
0.036212362349033356,
0.08103064447641373,
-0.04685748368501663,
-0.21040530502796173,
0.05635407939553261,
0.11579690873622894,
0.016309961676597595,
0.005046521779149771,
-0.1495097279548645,
0.06481044739484787,
0.1488969326019287,
-0.05613227188587189,
0.08725956827402115,
-0.2992357015609741,
-0.1308871954679489,
0.09859149158000946,
0.14030984044075012,
0.03888024017214775,
-0.1594155877828598,
-0.052176084369421005,
-0.00788675807416439,
-0.0753002017736435,
0.05863131955265999,
-0.17554253339767456,
0.09780001640319824,
-0.013046267442405224,
0.06824071705341339,
0.011409636586904526,
-0.04683555290102959,
0.13589100539684296,
-0.010466677136719227,
0.0864168331027031,
-0.05052954703569412,
0.04541885107755661,
0.08043400943279266,
-0.0707692801952362,
0.06769099086523056,
0.0411183200776577,
0.07891993224620819,
-0.10490141063928604,
-0.007698534056544304,
-0.09082885086536407,
0.06404935568571091,
-0.04968523979187012,
-0.05508438125252724,
-0.03074459731578827,
0.06325669586658478,
-0.00924626924097538,
-0.026637999340891838,
0.06835661828517914,
0.02492956817150116,
0.12592649459838867,
0.1611839234828949,
0.07931139320135117,
-0.014438281767070293,
-0.11984609812498093,
0.005094519350677729,
-0.04546802490949631,
0.07372138649225235,
-0.11317617446184158,
-0.01110553927719593,
0.11470110714435577,
0.015938803553581238,
0.09877452254295349,
0.03414906933903694,
-0.08044088631868362,
0.00896474439650774,
0.027461206540465355,
-0.12874345481395721,
-0.18792876601219177,
0.009087944403290749,
0.031617842614650726,
-0.12023262679576874,
0.06539005786180496,
0.10747959464788437,
-0.08844434469938278,
-0.028439665213227272,
-0.015652451664209366,
0.009576797485351562,
-0.02119367942214012,
0.1771925687789917,
0.08145471662282944,
0.07463761419057846,
-0.09506414830684662,
0.12609811127185822,
0.08535981923341751,
-0.06632152199745178,
0.029216343536973,
0.046101827174425125,
-0.1313232034444809,
-0.03486097976565361,
0.05833332613110542,
0.11414807289838791,
-0.02022107131779194,
-0.08050661534070969,
-0.09012968093156815,
-0.10140407085418701,
0.05016431212425232,
0.10738161206245422,
0.052804138511419296,
-0.007172450888901949,
0.008750272914767265,
0.017353378236293793,
-0.13973775506019592,
0.0849425345659256,
-0.010102865286171436,
0.06932902336120605,
-0.17686931788921356,
0.14542818069458008,
0.021054638549685478,
0.06638084352016449,
-0.021292787045240402,
0.030919993296265602,
-0.08725520223379135,
-0.00111687695607543,
-0.12009425461292267,
0.00822935439646244,
-0.032560184597969055,
-0.017206232994794846,
-0.014921717345714569,
-0.050164975225925446,
-0.007481559179723263,
0.07334887981414795,
-0.06654161214828491,
-0.05505867302417755,
0.015066766180098057,
0.033745940774679184,
-0.11397527903318405,
-0.000781437149271369,
-0.01410055160522461,
-0.062264300882816315,
0.08974950760602951,
0.04296676069498062,
0.03683382645249367,
0.0013319944264367223,
-0.028716864064335823,
0.026118706911802292,
0.04951326549053192,
0.014912990853190422,
0.06463097780942917,
-0.04808613285422325,
-0.02199670672416687,
-0.014744204469025135,
0.0168733112514019,
0.02233145758509636,
0.09464428573846817,
-0.13764777779579163,
-0.018832847476005554,
-0.01004033163189888,
-0.005249597132205963,
-0.07518725842237473,
0.038798294961452484,
0.08361013978719711,
0.05115634575486183,
0.12958042323589325,
-0.08132157474756241,
0.036930061876773834,
-0.20067012310028076,
-0.02315410226583481,
-0.016909243538975716,
-0.02688807062804699,
-0.06749984622001648,
-0.01113987062126398,
0.09936395287513733,
-0.029095739126205444,
0.08321962505578995,
0.013816727325320244,
0.12140664458274841,
0.03106420487165451,
-0.038233838975429535,
0.005760107655078173,
0.024184463545680046,
0.12538693845272064,
0.04300867021083832,
-0.014102362096309662,
0.11517219245433807,
0.011817693710327148,
0.05670884996652603,
0.07309071719646454,
0.19184191524982452,
0.15901458263397217,
0.015344708226621151,
0.06959794461727142,
0.0701909288764,
-0.07345455139875412,
-0.15540438890457153,
0.012622962705790997,
-0.020936405286192894,
0.08654669672250748,
-0.06285829842090607,
0.1374797224998474,
0.1007467657327652,
-0.14978913962841034,
0.026577899232506752,
-0.046779025346040726,
-0.09766501933336258,
-0.09736386686563492,
-0.03353288024663925,
-0.05217091739177704,
-0.12005380541086197,
-0.01032733079046011,
-0.10997258126735687,
-0.038592129945755005,
0.1207774206995964,
-0.0045754676684737206,
-0.02163795940577984,
0.15770666301250458,
0.02251291647553444,
0.00606551906093955,
0.05323489010334015,
0.024039853364229202,
0.025472069159150124,
-0.08406609296798706,
-0.08301243185997009,
0.07343081384897232,
-0.0026960589457303286,
0.09962733834981918,
-0.031219955533742905,
-0.015226905234158039,
0.052688200026750565,
0.002039551269263029,
-0.07358928769826889,
0.030879413709044456,
0.025651130825281143,
0.01777307502925396,
0.10748991370201111,
0.04048360511660576,
0.008870276622474194,
-0.04435451328754425,
0.23898553848266602,
-0.058362703770399094,
-0.062405455857515335,
-0.11417655646800995,
0.2580143213272095,
0.013482117094099522,
-0.045909248292446136,
0.06156878173351288,
-0.13521628081798553,
-0.00800992175936699,
0.17069998383522034,
0.13403360545635223,
-0.07218077033758163,
-0.01326330378651619,
-0.0029177481774240732,
-0.025878477841615677,
-0.06986602395772934,
0.12870736420154572,
0.12468475103378296,
0.05160382390022278,
-0.08409570157527924,
0.017410404980182648,
-0.013260312378406525,
-0.02149474062025547,
-0.11457313597202301,
0.021151913329958916,
0.010623909533023834,
0.02397264540195465,
-0.07144346088171005,
0.06536101549863815,
-0.025763390585780144,
-0.12210509926080704,
0.06485305726528168,
-0.13173767924308777,
-0.17928145825862885,
-0.01133851706981659,
0.05354122072458267,
-0.018771955743432045,
0.05147720128297806,
-0.030713766813278198,
0.0433182492852211,
0.11402274668216705,
-0.04110591858625412,
-0.03685970976948738,
-0.08907687664031982,
0.06893018633127213,
-0.0665394589304924,
0.20785613358020782,
-0.02255823090672493,
0.0834662914276123,
0.09902120381593704,
0.013156370259821415,
-0.16448478400707245,
0.05939431115984917,
0.04047178104519844,
-0.04129984602332115,
0.014381668530404568,
0.10898681730031967,
-0.008520128205418587,
0.014850101433694363,
0.021595776081085205,
-0.14718911051750183,
-0.04999638721346855,
-0.04963771626353264,
-0.008694134652614594,
-0.06697427481412888,
-0.0072712074033916,
-0.05717737600207329,
0.1397949457168579,
0.1529671847820282,
-0.057593487203121185,
-0.010253280401229858,
-0.07054571062326431,
0.004453916568309069,
0.029287850484251976,
0.017066407948732376,
0.001994677819311619,
-0.19359830021858215,
0.039407338947057724,
0.02136397175490856,
0.011910008266568184,
-0.2329895794391632,
-0.04330424219369888,
0.02743825502693653,
-0.0759686604142189,
-0.08510231226682663,
0.08270476758480072,
0.009668268263339996,
0.03824519366025925,
-0.04638076201081276,
-0.0743911936879158,
-0.052453167736530304,
0.13951753079891205,
-0.1207919642329216,
-0.0763198509812355
] |
null | null | transformers | # MiquMaid-v2-70B-DPO 3bpw
## Description
Exllama quant of [NeverSleep/MiquMaid-v2-70B-DPO](https://huggingface.co/NeverSleep/MiquMaid-v2-70B-DPO)
## Other quants:
EXL2: [4bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-4bpw-exl2), [3.5bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-3.5bpw-exl2), [3bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-3bpw-exl2), [2.4bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-2.4bpw-exl2), [2.3bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-2.3bpw-exl2)
2.4bpw is probably the most you can fit in a 24gb card
GGUF:
[2bit Imatrix GGUF](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-Imatrix-GGUF)
## Prompt format: Alpaca
```
### Instruction:
{system prompt}
### Input:
{input}
### Response:
{reply}
```
## Contact
Kooten on discord
[ko-fi.com/kooten](https://ko-fi.com/kooten) | {"license": "cc-by-nc-4.0", "tags": ["not-for-all-audiences", "nsfw"]} | text-generation | Kooten/MiquMaid-v2-70B-DPO-3bpw-exl2 | [
"transformers",
"pytorch",
"llama",
"text-generation",
"not-for-all-audiences",
"nsfw",
"conversational",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T07:00:23+00:00 | [] | [] | TAGS
#transformers #pytorch #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # MiquMaid-v2-70B-DPO 3bpw
## Description
Exllama quant of NeverSleep/MiquMaid-v2-70B-DPO
## Other quants:
EXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw
2.4bpw is probably the most you can fit in a 24gb card
GGUF:
2bit Imatrix GGUF
## Prompt format: Alpaca
## Contact
Kooten on discord
URL | [
"# MiquMaid-v2-70B-DPO 3bpw",
"## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B-DPO",
"## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF",
"## Prompt format: Alpaca",
"## Contact\nKooten on discord\n\nURL"
] | [
"TAGS\n#transformers #pytorch #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# MiquMaid-v2-70B-DPO 3bpw",
"## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B-DPO",
"## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF",
"## Prompt format: Alpaca",
"## Contact\nKooten on discord\n\nURL"
] | [
74,
17,
24,
60,
8,
7
] | [
"passage: TAGS\n#transformers #pytorch #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# MiquMaid-v2-70B-DPO 3bpw## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B-DPO## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF## Prompt format: Alpaca## Contact\nKooten on discord\n\nURL"
] | [
-0.04559050500392914,
0.06824549287557602,
-0.006388335023075342,
0.08988054096698761,
0.047118205577135086,
0.026196110993623734,
0.12174732238054276,
0.14308159053325653,
0.013500167988240719,
0.005482081323862076,
0.060952965170145035,
0.12988772988319397,
0.06264078617095947,
0.1582416445016861,
-0.039360448718070984,
-0.1602913737297058,
0.015588885173201561,
0.06388508528470993,
0.09820490330457687,
0.12457893043756485,
0.06860631704330444,
-0.02580440603196621,
0.0493277870118618,
-0.031243644654750824,
-0.1293928623199463,
-0.0034809575881808996,
-0.05143994465470314,
-0.06516003608703613,
0.0503556989133358,
0.09295450150966644,
0.03012908436357975,
0.08965755254030228,
-0.03372069448232651,
-0.17687362432479858,
0.023328598588705063,
0.030981646850705147,
-0.05234508216381073,
0.035862114280462265,
0.04483089596033096,
-0.04821387678384781,
0.039924897253513336,
0.02784641645848751,
-0.07442740350961685,
0.07232176512479782,
-0.1031886413693428,
-0.08600997179746628,
-0.053677964955568314,
0.14217513799667358,
0.06799492985010147,
0.04177560657262802,
-0.0080355703830719,
0.11371451616287231,
-0.03886375203728676,
0.07449328899383545,
0.23002539575099945,
-0.24467039108276367,
-0.022382307797670364,
0.08232863247394562,
0.03878180310130119,
0.05440903827548027,
-0.06556332856416702,
0.00562680047005415,
0.011430315673351288,
-0.00529492786154151,
0.020736103877425194,
-0.03392624482512474,
0.04846573248505592,
-0.022681159898638725,
-0.09289530664682388,
-0.011070606298744678,
0.15238937735557556,
0.0599236823618412,
-0.04595012962818146,
-0.034343376755714417,
-0.07798583805561066,
-0.12682661414146423,
-0.03962298482656479,
0.0323217511177063,
-0.0033694463782012463,
-0.015411986038088799,
-0.04577292874455452,
-0.0010069870622828603,
-0.10656531900167465,
-0.041472647339105606,
-0.14029622077941895,
0.16278153657913208,
0.0015777143416926265,
0.025784911587834358,
0.0032211646903306246,
0.047401487827301025,
-0.08662254363298416,
-0.08483143150806427,
-0.08331389725208282,
-0.08938184380531311,
0.04092119634151459,
-0.0020182859152555466,
-0.00946969073265791,
0.10156124830245972,
0.07845783978700638,
0.17323482036590576,
-0.029112711548805237,
0.07396457344293594,
0.03086676262319088,
0.0660175010561943,
0.003545823972672224,
-0.07862488925457001,
-0.0352088138461113,
-0.04034828394651413,
0.03161429986357689,
0.013692802749574184,
0.04599880799651146,
-0.023036645725369453,
-0.11379383504390717,
-0.020520852878689766,
-0.00882832519710064,
0.022347761318087578,
0.041305143386125565,
0.05248548462986946,
-0.06235513463616371,
-0.06236722692847252,
0.18690775334835052,
-0.04765309393405914,
0.021485688164830208,
0.05807637795805931,
-0.004587532486766577,
0.10442305356264114,
0.04748380184173584,
-0.0027080911677330732,
-0.05782336741685867,
0.05563892051577568,
-0.033768460154533386,
0.028184637427330017,
-0.023827800527215004,
-0.049107540398836136,
0.09056676179170609,
-0.061263177543878555,
-0.014561742544174194,
-0.17010124027729034,
-0.13517487049102783,
0.06973911821842194,
-0.004629223607480526,
-0.05731154978275299,
-0.03976315259933472,
0.0013198364758864045,
-0.017812859266996384,
-0.008803943172097206,
-0.060639917850494385,
0.03320101648569107,
-0.0697341039776802,
0.08761176466941833,
0.07492191344499588,
0.12439697235822678,
-0.11987403780221939,
0.0020946913864463568,
-0.03980548307299614,
0.02622750960290432,
-0.07431412488222122,
0.017303382977843285,
-0.11355297267436981,
-0.02734231948852539,
-0.08070507645606995,
-0.050626639276742935,
-0.028395257890224457,
0.0384494848549366,
-0.00471697561442852,
0.10912048816680908,
-0.05709315463900566,
-0.058587539941072464,
0.2430669069290161,
-0.09019219875335693,
-0.06799827516078949,
0.12773531675338745,
0.055166278034448624,
-0.08308716118335724,
0.04933766648173332,
0.07230602949857712,
0.11135275661945343,
-0.09970803558826447,
-0.07809452712535858,
0.08191464096307755,
-0.01174345426261425,
-0.08165382593870163,
0.09373214095830917,
0.051825933158397675,
0.0375211276113987,
0.032556917518377304,
0.15808434784412384,
0.033341288566589355,
-0.0034550477284938097,
-0.06827958673238754,
-0.04970458149909973,
-0.0603308379650116,
0.07571684569120407,
0.008153258822858334,
-0.028269579634070396,
-0.09536049515008926,
-0.10992567241191864,
-0.0015068466309458017,
0.07619946449995041,
-0.0034522423520684242,
-0.006539048161357641,
-0.11979123204946518,
0.09839706122875214,
-0.0834730789065361,
0.030019469559192657,
-0.07405144721269608,
0.0016904993681237102,
-0.031382374465465546,
-0.004861239343881607,
0.08118892461061478,
0.10350069403648376,
0.07213311642408371,
0.03370225429534912,
-0.07826340943574905,
0.03419800102710724,
0.09464447945356369,
0.0025050370022654533,
-0.004523979499936104,
-0.06735322624444962,
0.027854297310113907,
-0.013053334318101406,
0.11793442070484161,
-0.07766890525817871,
0.025512127205729485,
0.13318991661071777,
0.10125917196273804,
-0.026402244344353676,
-0.03440733626484871,
0.06643553078174591,
0.03326971083879471,
-0.04181911051273346,
0.017242267727851868,
0.08813535422086716,
-0.01820305734872818,
-0.017206387594342232,
0.09350842982530594,
-0.2104381024837494,
0.1154768094420433,
0.17357490956783295,
-0.06478982418775558,
-0.029969485476613045,
-0.0642695501446724,
-0.009316673502326012,
-0.024859435856342316,
0.011393238790333271,
-0.05541222169995308,
0.07519698143005371,
-0.01469894964247942,
0.12602977454662323,
-0.08437944948673248,
-0.05386049672961235,
0.05162588879466057,
-0.04496888071298599,
0.001616297522559762,
0.06912883371114731,
0.10974669456481934,
-0.11460991203784943,
0.15835703909397125,
0.14147047698497772,
-0.08557333797216415,
0.13432717323303223,
0.028006965294480324,
-0.030153216794133186,
-0.03276491537690163,
0.014813254587352276,
0.007349101360887289,
0.05653955414891243,
-0.1174163892865181,
0.05494949221611023,
0.07980010658502579,
-0.011529658921062946,
0.026350252330303192,
-0.13102243840694427,
-0.04243847355246544,
0.020339323207736015,
-0.06971977651119232,
-0.08307735621929169,
-0.0013011920964345336,
0.004965013824403286,
0.10173270851373672,
-0.014654146507382393,
-0.002229053061455488,
0.039108987897634506,
0.04271165654063225,
-0.07181143760681152,
0.14043442904949188,
-0.09632992744445801,
-0.3080297112464905,
-0.08001209795475006,
-0.1446453034877777,
-0.08258570730686188,
0.016170289367437363,
0.09371967613697052,
-0.0694844201207161,
-0.030376099050045013,
-0.04833959415555,
-0.04460674524307251,
-0.044897016137838364,
-0.01229635626077652,
0.009725930169224739,
-0.017906980589032173,
0.05579096078872681,
-0.0803367868065834,
-0.021338824182748795,
0.056904230266809464,
-0.04210411384701729,
0.16783471405506134,
-0.025261471047997475,
0.09174906462430954,
0.1381477266550064,
0.018445413559675217,
-0.015577751211822033,
-0.03357023373246193,
0.15828078985214233,
-0.0662425234913826,
0.049144499003887177,
0.14174580574035645,
-0.008761610835790634,
0.0774737074971199,
0.09139982610940933,
0.0412340871989727,
-0.05241844803094864,
-0.051705509424209595,
0.03013664484024048,
-0.09790278226137161,
-0.14722688496112823,
-0.038155216723680496,
-0.06405670195817947,
0.14546048641204834,
0.054326366633176804,
0.059863053262233734,
0.06916607916355133,
0.11644928902387619,
-0.07739847898483276,
0.0000853223173180595,
0.02965034544467926,
0.09024147689342499,
0.16870440542697906,
0.02035597898066044,
0.1539788544178009,
-0.05985434353351593,
-0.038333140313625336,
0.12704986333847046,
0.13196031749248505,
0.044590599834918976,
0.008188918232917786,
0.10701863467693329,
0.05819878354668617,
0.016283418983221054,
0.09577569365501404,
0.046257805079221725,
0.01322256214916706,
-0.03285762667655945,
-0.040937863290309906,
-0.0596817247569561,
-0.00685481633991003,
0.04133853688836098,
0.0718691423535347,
-0.07713006436824799,
0.030017031356692314,
-0.08765096217393875,
0.06776813417673111,
0.06612499058246613,
0.08896058052778244,
-0.20802456140518188,
-0.02863362431526184,
0.04846455901861191,
-0.027438122779130936,
-0.040306687355041504,
0.019458632916212082,
0.0030474269296973944,
-0.02399148792028427,
0.16416162252426147,
-0.008778770454227924,
0.09412536025047302,
-0.07747261226177216,
0.02462742291390896,
-0.03554907068610191,
0.05414997786283493,
0.002567762741819024,
0.11414836347103119,
-0.2933025062084198,
0.07641066610813141,
0.022885628044605255,
0.0011850126320496202,
-0.07005676627159119,
-0.010289476253092289,
0.014078100211918354,
-0.03490101546049118,
0.09074927121400833,
0.012261776253581047,
0.1167975664138794,
-0.13575586676597595,
-0.06840462982654572,
0.01806970126926899,
0.058582909405231476,
0.035087406635284424,
0.03774651885032654,
0.04974000155925751,
-0.040424756705760956,
-0.01195287425071001,
-0.001626082113943994,
-0.039080116897821426,
-0.09414488822221756,
0.12276504188776016,
0.12031853944063187,
-0.020588746294379234,
0.014483679085969925,
-0.033752672374248505,
-0.1926722228527069,
0.2231282889842987,
-0.11499723792076111,
-0.09887710958719254,
-0.06012018769979477,
-0.058868784457445145,
0.019404347985982895,
-0.03817281499505043,
0.018160570412874222,
-0.05358947440981865,
0.08826334029436111,
-0.03447321802377701,
-0.0834827795624733,
0.0808490514755249,
-0.10449660569429398,
-0.1332552582025528,
-0.03482488915324211,
0.156071737408638,
-0.06973135471343994,
0.03984996676445007,
0.01986514776945114,
-0.013383783400058746,
-0.04544590786099434,
-0.12730464339256287,
0.1041497141122818,
0.006427581887692213,
-0.0032592301722615957,
0.016218356788158417,
0.016938192769885063,
-0.011493629775941372,
-0.04037332534790039,
-0.0346212275326252,
0.1437455713748932,
0.35855334997177124,
-0.03362540900707245,
0.11728550493717194,
0.12044008076190948,
-0.03951504826545715,
-0.21306785941123962,
-0.16468219459056854,
-0.04167376086115837,
-0.08058619499206543,
-0.05792120844125748,
-0.1387893706560135,
0.04041588306427002,
0.13345852494239807,
-0.03450454771518707,
0.1876789778470993,
-0.14621035754680634,
-0.09879523515701294,
0.04056508094072342,
0.055002957582473755,
0.24272598326206207,
-0.15756253898143768,
-0.05882555618882179,
-0.05980026721954346,
-0.20135696232318878,
0.22079505026340485,
-0.035689570009708405,
0.12574942409992218,
-0.0400204062461853,
0.10001415759325027,
-0.0022321161814033985,
-0.05996369197964668,
0.12321913242340088,
-0.024160493165254593,
-0.02845074236392975,
-0.11750750988721848,
0.042883746325969696,
0.006900678854435682,
-0.006519833113998175,
0.011311694979667664,
-0.1776265949010849,
-0.02927294373512268,
-0.13826976716518402,
-0.05954210087656975,
-0.07813160121440887,
0.01862361654639244,
-0.03233624994754791,
-0.054189532995224,
-0.05368826538324356,
-0.001077271532267332,
0.013663958758115768,
-0.02109837345778942,
-0.011807403527200222,
-0.07428678125143051,
0.07593999058008194,
0.298345148563385,
0.09377123415470123,
-0.08029107749462128,
-0.09377069026231766,
-0.06808627396821976,
-0.09443455189466476,
0.021602652966976166,
-0.11691045016050339,
0.03488701581954956,
0.07946551591157913,
0.008274652995169163,
0.06306034326553345,
0.04464176297187805,
-0.05994390323758125,
0.0206358190625906,
0.12071118503808975,
-0.13936537504196167,
0.0035366136580705643,
-0.013754235580563545,
0.06328856199979782,
0.001124756527133286,
-0.009288421832025051,
0.13467656075954437,
-0.008659417741000652,
-0.014318323694169521,
0.05789612606167793,
0.01790391467511654,
-0.019867895171046257,
0.11238133907318115,
0.03601672500371933,
0.07288673520088196,
-0.13067403435707092,
0.06225140392780304,
-0.011244507506489754,
-0.05880418419837952,
-0.010518052615225315,
0.10577401518821716,
-0.07008152455091476,
-0.12933845818042755,
-0.017984597012400627,
-0.08029487729072571,
-0.10140030831098557,
-0.038835495710372925,
-0.02466098964214325,
-0.11732161045074463,
0.06578025966882706,
0.12024202942848206,
0.021930720657110214,
-0.02445288375020027,
0.06330573558807373,
0.040455903857946396,
-0.06877273321151733,
0.058907341212034225,
-0.009436211548745632,
0.08171407133340836,
-0.10011324286460876,
0.05593390762805939,
-0.007000438868999481,
0.04656244441866875,
-0.042096808552742004,
0.010454392991960049,
-0.12199129164218903,
-0.028482412919402122,
-0.13777586817741394,
0.05355900153517723,
-0.11562291532754898,
-0.004411663860082626,
-0.019834626466035843,
-0.0001588420127518475,
-0.028319230303168297,
-0.0401291660964489,
-0.0909336656332016,
0.006102205719798803,
0.02165692113339901,
-0.005732030142098665,
-0.0947655662894249,
-0.02712368778884411,
0.0288098081946373,
-0.022916939109563828,
0.07339967042207718,
0.05027583986520767,
-0.053853265941143036,
-0.018906990066170692,
-0.15287376940250397,
-0.032895758748054504,
0.07891236990690231,
0.07177227735519409,
-0.024425838142633438,
0.026782222092151642,
0.04453924298286438,
0.04097636044025421,
0.036139730364084244,
0.031025642529129982,
0.07564596086740494,
-0.09016066044569016,
-0.015179342590272427,
-0.06932400166988373,
-0.09984796494245529,
-0.05200488120317459,
-0.008859457448124886,
0.124767005443573,
-0.031192095950245857,
0.10604539513587952,
-0.04947243258357048,
0.028001660481095314,
-0.13656450808048248,
0.020905056968331337,
0.030289633199572563,
-0.14875876903533936,
-0.00879506766796112,
-0.05866975337266922,
0.030798863619565964,
-0.028137467801570892,
0.16427136957645416,
-0.07779567688703537,
-0.10442385822534561,
0.028044816106557846,
-0.06883969902992249,
-0.035729724913835526,
-0.01920166239142418,
0.195208340883255,
0.0768575593829155,
-0.013034977950155735,
-0.09336164593696594,
0.028258873149752617,
0.05470680445432663,
-0.07468397915363312,
0.08837109059095383,
0.08206506818532944,
-0.02802850492298603,
0.043480150401592255,
0.06569825112819672,
-0.07119356840848923,
-0.08243247121572495,
0.0018002165015786886,
-0.1200653612613678,
0.10302755236625671,
-0.019865816459059715,
0.08865819871425629,
0.08807998895645142,
-0.04138169810175896,
0.018139047548174858,
-0.06729256361722946,
-0.03391994163393974,
-0.11012914776802063,
-0.08731773495674133,
-0.08211935311555862,
-0.09510087221860886,
0.03055810183286667,
-0.07728131860494614,
0.031176894903182983,
0.05964452028274536,
0.04882122948765755,
-0.017566930502653122,
0.08921512216329575,
-0.015500808134675026,
-0.03333238884806633,
0.011263920925557613,
-0.02561451494693756,
-0.0193928312510252,
0.02882024459540844,
-0.03337178751826286,
0.0667535737156868,
-0.014413497410714626,
0.045600369572639465,
0.025282632559537888,
0.01835131086409092,
-0.004197501577436924,
-0.06884007900953293,
-0.09311959892511368,
-0.03553645685315132,
0.03599797189235687,
0.08031683415174484,
0.15826885402202606,
-0.03346002846956253,
0.015185556374490261,
0.016063109040260315,
0.07714559882879257,
-0.0270018819719553,
-0.08423137664794922,
-0.03575523570179939,
0.19895073771476746,
-0.0857381746172905,
0.05178721249103546,
-0.01507615763694048,
-0.020429572090506554,
-0.009043817408382893,
0.24779543280601501,
0.1580526977777481,
-0.04087318852543831,
0.03172179311513901,
0.04432220011949539,
0.02349538914859295,
0.07619480043649673,
0.0899079442024231,
0.08030176907777786,
0.18204829096794128,
-0.05086443945765495,
0.005983550567179918,
0.027447452768683434,
0.0008100030245259404,
-0.08039986342191696,
0.05833578109741211,
0.019409039989113808,
-0.01897098682820797,
-0.06388989090919495,
0.02247668243944645,
-0.1295699179172516,
0.07127043604850769,
-0.006222025491297245,
-0.11824635416269302,
-0.07495326548814774,
0.005348408594727516,
0.05171086639165878,
0.04725990444421768,
0.013414736837148666,
-0.029927417635917664,
-0.010024065151810646,
0.020464519038796425,
-0.0186306145042181,
-0.1773025095462799,
0.056071050465106964,
0.006775796413421631,
-0.0007009959081187844,
0.028254181146621704,
-0.04012545943260193,
0.1563853919506073,
0.10534163564443588,
0.008349584415555,
-0.049141526222229004,
0.13158491253852844,
0.020128680393099785,
-0.16747167706489563,
0.0387866348028183,
0.12641917169094086,
-0.03217533975839615,
-0.0007775741396471858,
0.09280615299940109,
-0.0632646307349205,
-0.04857153818011284,
0.10819081217050552,
-0.02934306487441063,
-0.01738853193819523,
0.06483396142721176,
-0.053873028606176376,
0.07305310666561127,
0.06559193879365921,
-0.049667470157146454,
-0.07109924405813217,
-0.044763535261154175,
0.05215161293745041,
-0.024036968126893044,
-0.057324182242155075,
-0.057252854108810425,
-0.16896845400333405,
-0.05282648652791977,
0.05202891305088997,
0.12504971027374268,
-0.24422687292099,
-0.005599854048341513,
-0.08731099963188171,
0.00023018023057375103,
-0.1298786997795105,
0.060332342982292175,
0.09848855435848236,
0.011204843409359455,
-0.05002040043473244,
-0.10026684403419495,
-0.019864913076162338,
0.07836350053548813,
-0.12341929972171783,
-0.07085295021533966
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# tamil-llama-7b-instruct-quantized-ASR-output-fine-tuning
This model is a fine-tuned version of [abhinand/tamil-llama-7b-instruct-v0.1](https://huggingface.co/abhinand/tamil-llama-7b-instruct-v0.1) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 5.3527
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 1
- eval_batch_size: 1
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 4
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: constant
- lr_scheduler_warmup_ratio: 0.03
- training_steps: 1500
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| 2.4659 | 2.58 | 500 | 4.2383 |
| 1.9248 | 5.17 | 1000 | 4.6944 |
| 1.4112 | 7.75 | 1500 | 5.3527 |
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.0
- Pytorch 2.1.2
- Datasets 2.16.1
- Tokenizers 0.15.1 | {"language": ["ta"], "license": "llama2", "library_name": "peft", "tags": ["trl", "sft", "Tamil-ASR, ASR-fine-tuning, Tamil-llama", "generated_from_trainer"], "base_model": "abhinand/tamil-llama-7b-instruct-v0.1", "model-index": [{"name": "tamil-llama-7b-instruct-quantized-ASR-output-fine-tuning", "results": []}]} | null | sujith013/tamil-llama-7b-instruct-quantized-ASR-output-fine-tuning | [
"peft",
"tensorboard",
"safetensors",
"trl",
"sft",
"Tamil-ASR, ASR-fine-tuning, Tamil-llama",
"generated_from_trainer",
"ta",
"base_model:abhinand/tamil-llama-7b-instruct-v0.1",
"license:llama2",
"region:us"
] | 2024-02-08T07:03:25+00:00 | [] | [
"ta"
] | TAGS
#peft #tensorboard #safetensors #trl #sft #Tamil-ASR, ASR-fine-tuning, Tamil-llama #generated_from_trainer #ta #base_model-abhinand/tamil-llama-7b-instruct-v0.1 #license-llama2 #region-us
| tamil-llama-7b-instruct-quantized-ASR-output-fine-tuning
========================================================
This model is a fine-tuned version of abhinand/tamil-llama-7b-instruct-v0.1 on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 5.3527
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0002
* train\_batch\_size: 1
* eval\_batch\_size: 1
* seed: 42
* gradient\_accumulation\_steps: 4
* total\_train\_batch\_size: 4
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: constant
* lr\_scheduler\_warmup\_ratio: 0.03
* training\_steps: 1500
### Training results
### Framework versions
* PEFT 0.8.2
* Transformers 4.37.0
* Pytorch 2.1.2
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 1\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: constant\n* lr\\_scheduler\\_warmup\\_ratio: 0.03\n* training\\_steps: 1500",
"### Training results",
"### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#peft #tensorboard #safetensors #trl #sft #Tamil-ASR, ASR-fine-tuning, Tamil-llama #generated_from_trainer #ta #base_model-abhinand/tamil-llama-7b-instruct-v0.1 #license-llama2 #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 1\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: constant\n* lr\\_scheduler\\_warmup\\_ratio: 0.03\n* training\\_steps: 1500",
"### Training results",
"### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
79,
143,
4,
36
] | [
"passage: TAGS\n#peft #tensorboard #safetensors #trl #sft #Tamil-ASR, ASR-fine-tuning, Tamil-llama #generated_from_trainer #ta #base_model-abhinand/tamil-llama-7b-instruct-v0.1 #license-llama2 #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 1\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: constant\n* lr\\_scheduler\\_warmup\\_ratio: 0.03\n* training\\_steps: 1500### Training results### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.14538884162902832,
0.03815491870045662,
-0.0023391630966216326,
0.1066264882683754,
0.14292113482952118,
-0.0018487201305106282,
0.0632292702794075,
0.15247978270053864,
-0.08284144103527069,
0.0774601474404335,
0.12172408401966095,
0.08873046189546585,
0.07395186275243759,
0.1571068912744522,
-0.006433531641960144,
-0.3380865454673767,
-0.028495103120803833,
-0.017069755122065544,
-0.18127678334712982,
0.12056158483028412,
0.10627390444278717,
-0.10261797159910202,
0.057704370468854904,
-0.0182122103869915,
-0.06672323495149612,
0.005819692276418209,
-0.023220280185341835,
-0.046662647277116776,
0.08885175734758377,
-0.010047557763755322,
0.0800686776638031,
0.014891048893332481,
0.10050255060195923,
-0.22431419789791107,
0.017590392380952835,
0.027140647172927856,
0.03700204938650131,
0.051400721073150635,
0.08156297355890274,
-0.006554684601724148,
0.13292068243026733,
-0.058476924896240234,
0.08168382197618484,
0.036772336810827255,
-0.12431833893060684,
-0.24350707232952118,
-0.09128941595554352,
0.08838705718517303,
0.12815828621387482,
0.08074752241373062,
-0.03602245822548866,
0.07690918445587158,
-0.07675259560346603,
0.09695141017436981,
0.19739143550395966,
-0.250245600938797,
-0.09489277005195618,
0.08903240412473679,
0.05730539560317993,
0.12686100602149963,
-0.1362653374671936,
-0.0640350952744484,
0.04432615637779236,
0.037319377064704895,
0.12173032760620117,
-0.007823444902896881,
0.07863558083772659,
0.02783290110528469,
-0.13087275624275208,
-0.013634220696985722,
0.10955806821584702,
0.048246100544929504,
-0.02451348304748535,
-0.07277281582355499,
-0.020820382982492447,
-0.23973754048347473,
-0.04796963930130005,
0.02135368436574936,
0.009688504040241241,
-0.02917211875319481,
-0.007700631860643625,
0.02928175963461399,
-0.0830841138958931,
-0.07742225378751755,
0.1222001388669014,
0.05680252984166145,
0.04988747835159302,
-0.03031175397336483,
-0.02880791202187538,
0.11768561601638794,
0.02331913821399212,
-0.1576397567987442,
-0.003080237191170454,
0.02652609534561634,
-0.061038270592689514,
-0.026145612820982933,
0.013700490817427635,
-0.02232077717781067,
-0.013355908915400505,
0.11913306266069412,
-0.08877414464950562,
0.07144657522439957,
0.05929406359791756,
0.020470857620239258,
-0.11966940015554428,
0.14850465953350067,
-0.09086354076862335,
-0.01118966843932867,
-0.04773103445768356,
0.17088472843170166,
0.011940578930079937,
-0.00547296367585659,
-0.04962761327624321,
0.03967176005244255,
0.080105721950531,
0.058362219482660294,
-0.05944356694817543,
0.020095989108085632,
-0.07119113206863403,
-0.01790386624634266,
0.05304252356290817,
-0.1312156766653061,
0.02675325609743595,
-0.008566241711378098,
-0.06959444284439087,
-0.09079950302839279,
0.00031635977211408317,
0.009582556784152985,
0.014350906014442444,
0.13927777111530304,
-0.10741543769836426,
0.023051172494888306,
-0.11123701184988022,
-0.07689478993415833,
0.0029546136502176523,
-0.11653633415699005,
-0.018565775826573372,
-0.05562802031636238,
-0.15789081156253815,
-0.037530381232500076,
0.0481862798333168,
-0.09077530354261398,
-0.05572189763188362,
-0.047658421099185944,
-0.1024378314614296,
0.021340996026992798,
-0.02114267274737358,
0.13884450495243073,
-0.05470064654946327,
0.12309828400611877,
0.024139726534485817,
0.08532440662384033,
0.08781521767377853,
0.05713445693254471,
-0.05984813719987869,
0.07777214795351028,
-0.15987224876880646,
0.04611065611243248,
-0.1064477413892746,
0.09196612983942032,
-0.11221221834421158,
-0.09876398742198944,
-0.054121196269989014,
-0.010442198254168034,
0.09471902251243591,
0.14414319396018982,
-0.1804402619600296,
-0.07664395868778229,
0.20816312730312347,
-0.09507216513156891,
-0.1545325219631195,
0.09920325130224228,
-0.002083889441564679,
0.07562775909900665,
0.0018968769581988454,
0.15153904259204865,
0.08661868423223495,
-0.0678742453455925,
-0.024290114641189575,
-0.05460745841264725,
0.11447519063949585,
0.03347041830420494,
0.09281818568706512,
-0.03429674729704857,
0.07631873339414597,
0.0068761082366108894,
-0.0421423614025116,
0.033642932772636414,
-0.09662659466266632,
-0.09999686479568481,
-0.02570200338959694,
-0.1030510887503624,
0.02934245392680168,
0.072383813560009,
0.05951627716422081,
-0.05850435793399811,
-0.08189820498228073,
0.0028445892967283726,
0.09896617382764816,
-0.05601334571838379,
0.02377723529934883,
-0.0755554586648941,
0.0834670290350914,
-0.06552963703870773,
-0.003937825094908476,
-0.1548876166343689,
-0.01194615289568901,
0.03914898261427879,
-0.0062027303501963615,
-0.016020668670535088,
-0.055640604346990585,
0.10523461550474167,
0.0405300073325634,
-0.10564029961824417,
-0.06299154460430145,
-0.051278166472911835,
-0.010373124852776527,
-0.10446170717477798,
-0.2358931452035904,
-0.05975588038563728,
-0.035372648388147354,
0.1809716671705246,
-0.28075680136680603,
0.06027001887559891,
0.06891357898712158,
0.14229728281497955,
0.06045154109597206,
-0.06662449240684509,
0.04021751880645752,
0.0658293217420578,
-0.002197820693254471,
-0.1098708063364029,
0.03758411854505539,
-0.023097669705748558,
-0.1076294556260109,
-0.03200814500451088,
-0.118193618953228,
0.12997201085090637,
0.0816422700881958,
0.03485756739974022,
-0.1134534552693367,
-0.024739425629377365,
-0.07239015400409698,
-0.05106345936655998,
-0.03879418224096298,
0.045635927468538284,
0.04947507381439209,
0.050343889743089676,
0.15551690757274628,
-0.06553949415683746,
-0.03832376003265381,
0.05435733124613762,
-0.033924926072359085,
0.02298569679260254,
0.1657916009426117,
0.08153359591960907,
-0.021906370297074318,
0.1351305991411209,
0.08461126685142517,
-0.05257609486579895,
0.15888360142707825,
-0.05212665721774101,
-0.10685861110687256,
-0.01679573394358158,
0.05902370065450668,
0.0507676936686039,
0.1360044628381729,
-0.09727557748556137,
0.005630797706544399,
-0.0011261184699833393,
0.012172786518931389,
-0.0046439687721431255,
-0.18958251178264618,
-0.0693218782544136,
0.03362146019935608,
-0.044782985001802444,
-0.038084130734205246,
0.028444156050682068,
-0.047469768673181534,
0.0865308940410614,
-0.0005017759394831955,
-0.016920069232583046,
0.006289852317422628,
0.011964280158281326,
-0.10775739699602127,
0.19934645295143127,
-0.07797666639089584,
-0.15346252918243408,
-0.15219159424304962,
0.038671400398015976,
0.0012069862568750978,
-0.013783479109406471,
0.02069627307355404,
-0.17479640245437622,
-0.015001426450908184,
-0.08339489996433258,
-0.007177543360739946,
0.005245237611234188,
0.01998252049088478,
-0.05582364276051521,
0.01831389218568802,
0.09318462759256363,
-0.08713055402040482,
0.03191085159778595,
0.010341515764594078,
-0.06258810311555862,
0.06552465260028839,
0.025119785219430923,
0.08325889706611633,
0.15423287451267242,
0.009513840079307556,
0.027513591572642326,
-0.013263650238513947,
0.18105894327163696,
-0.12853151559829712,
0.04296664148569107,
0.09538239985704422,
0.019308147951960564,
0.061335887759923935,
0.16800837218761444,
0.046839285641908646,
-0.08924148976802826,
0.03718721866607666,
0.05765461176633835,
-0.026866108179092407,
-0.23866057395935059,
0.004715210758149624,
-0.0474172979593277,
0.0238513071089983,
0.11909664422273636,
0.06414400786161423,
0.026411501690745354,
0.06779766827821732,
-0.011577319353818893,
0.0070640211924910545,
-0.004592880140990019,
0.07067549228668213,
-0.017521554604172707,
0.02420021779835224,
0.1160082146525383,
-0.0458068922162056,
-0.0007049575797282159,
0.023299366235733032,
0.03273576498031616,
0.25419458746910095,
0.00362519477494061,
0.08333255350589752,
0.057398539036512375,
0.19466589391231537,
-0.008881560526788235,
0.04539771005511284,
-0.028006508946418762,
-0.06654036045074463,
0.03358493000268936,
-0.051225073635578156,
-0.020171955227851868,
0.08121699094772339,
0.004656986799091101,
0.05304114520549774,
-0.08432251214981079,
0.04308927804231644,
0.03967467322945595,
0.3120890259742737,
0.04902644082903862,
-0.27102580666542053,
-0.10682554543018341,
0.01983317732810974,
-0.06132129579782486,
-0.03797874227166176,
0.004214801359921694,
0.13067571818828583,
-0.06926696747541428,
0.06192809343338013,
-0.11314308643341064,
0.0674368292093277,
-0.011068087071180344,
-0.020720379427075386,
0.07633402198553085,
0.03110693022608757,
-0.017406504601240158,
0.05623488873243332,
-0.2703779637813568,
0.3364476263523102,
0.014884291216731071,
0.08656112849712372,
0.005535244941711426,
0.0077432673424482346,
0.029123730957508087,
0.04429245740175247,
0.0899856686592102,
-0.0034695309586822987,
-0.00614909129217267,
-0.24200113117694855,
-0.057669781148433685,
0.005270098801702261,
0.13830940425395966,
-0.07221890985965729,
0.12567639350891113,
-0.03437730297446251,
0.0021439336705952883,
0.03253287076950073,
-0.10069523751735687,
-0.09545782953500748,
-0.06255657225847244,
0.050907377153635025,
-0.03275526314973831,
0.014304042793810368,
-0.10848531126976013,
-0.11305351555347443,
-0.077090322971344,
0.07747293263673782,
-0.14664119482040405,
-0.036720842123031616,
-0.1136305183172226,
0.11307478696107864,
0.14069002866744995,
-0.07532883435487747,
0.04675619676709175,
0.014628995209932327,
0.07884861528873444,
0.018270859494805336,
0.07749413698911667,
0.09084620326757431,
-0.04823708534240723,
-0.2340579777956009,
-0.021404296159744263,
0.1321275532245636,
0.051875293254852295,
0.04432002827525139,
-0.0438510961830616,
0.050359826534986496,
0.030178578570485115,
-0.08057545870542526,
0.0852937325835228,
0.016876671463251114,
-0.011757823638617992,
0.025978801771998405,
-0.08374829590320587,
0.061230361461639404,
-0.05757927894592285,
-0.11924031376838684,
0.07126925885677338,
0.320117324590683,
-0.08555746078491211,
0.04420117661356926,
0.026371300220489502,
-0.06817469745874405,
-0.15426456928253174,
0.051595207303762436,
0.09915326535701752,
0.03746489807963371,
0.07374611496925354,
-0.18894249200820923,
0.04040313884615898,
0.09769769012928009,
-0.038312364369630814,
0.07857085764408112,
-0.2824353873729706,
-0.13889609277248383,
0.041240423917770386,
0.11953651160001755,
-0.07184689491987228,
-0.20043767988681793,
-0.06833179295063019,
0.022675054147839546,
-0.09485575556755066,
-0.039043985307216644,
-0.008129006251692772,
0.10757715255022049,
-0.04399024322628975,
0.03283964842557907,
0.013965205289423466,
-0.06969301402568817,
0.1545865386724472,
-0.01611473597586155,
0.05480028688907623,
-0.024907024577260017,
-0.014866518788039684,
0.04136800765991211,
-0.05524713918566704,
0.06332015991210938,
-0.05174974352121353,
0.008567875251173973,
-0.12750136852264404,
-0.012928582727909088,
-0.08963624387979507,
0.010091031901538372,
-0.053125567734241486,
-0.038113102316856384,
-0.028350869193673134,
0.09589722007513046,
0.03997645527124405,
-0.003440259024500847,
0.11412344127893448,
-0.06717897951602936,
0.1767297238111496,
0.08385065197944641,
0.11419402807950974,
-0.04842892289161682,
-0.05228598415851593,
-0.01629006490111351,
0.0036435655783861876,
0.036007367074489594,
-0.21101492643356323,
0.021468203514814377,
0.11608782410621643,
0.0453629270195961,
0.16590958833694458,
0.058598592877388,
-0.04401533305644989,
0.021820735186338425,
0.12220368534326553,
-0.032624952495098114,
-0.15887299180030823,
0.008459966629743576,
0.01885583996772766,
-0.17522816359996796,
0.018765568733215332,
0.07899098843336105,
-0.08713177591562271,
-0.008968709036707878,
-0.00625983253121376,
0.05199718475341797,
-0.05893012881278992,
0.21560828387737274,
0.08364265412092209,
0.09213277697563171,
-0.05635945498943329,
0.08991599828004837,
0.04953937977552414,
-0.1458457112312317,
0.008541833609342575,
0.13107939064502716,
-0.08717061579227448,
-0.031829189509153366,
0.004229782149195671,
0.04718063771724701,
0.008057521656155586,
0.0038043991662561893,
-0.14723782241344452,
-0.07961481809616089,
0.06925826519727707,
0.10950689762830734,
0.03191692382097244,
0.014007987454533577,
-0.03873458877205849,
0.037885285913944244,
-0.1188354343175888,
0.12478865683078766,
0.10508249700069427,
0.08752304315567017,
-0.17155487835407257,
0.12447579950094223,
0.020815592259168625,
-0.007078999187797308,
0.007104780059307814,
0.02065967582166195,
-0.08786805719137192,
0.01605294458568096,
-0.04301542043685913,
-0.05998352915048599,
-0.05136711150407791,
-0.00900895893573761,
-0.019056163728237152,
-0.06024527549743652,
-0.043587248772382736,
0.010153437964618206,
-0.11136931926012039,
-0.04297082498669624,
-0.0203604344278574,
0.08372705429792404,
-0.10154446959495544,
-0.03418572247028351,
0.05788695439696312,
-0.07766702771186829,
0.05408621206879616,
-0.007571117952466011,
-0.011536945588886738,
0.02532205916941166,
-0.0636448860168457,
0.05542019009590149,
0.014260031282901764,
-0.02445991151034832,
0.0417068675160408,
-0.14077946543693542,
-0.006445868406444788,
-0.054896026849746704,
0.027658412232995033,
0.0475601889193058,
0.05504245683550835,
-0.14572151005268097,
0.023384900763630867,
-0.08201289176940918,
-0.05634979531168938,
-0.07203985005617142,
0.09443848580121994,
0.04452729597687721,
0.02440873719751835,
0.15257984399795532,
-0.09106644243001938,
0.03658682852983475,
-0.23185397684574127,
-0.007130997721105814,
-0.024269428104162216,
-0.06842120736837387,
-0.0739908218383789,
-0.007509114686399698,
0.07793217152357101,
-0.08673068135976791,
0.027694808319211006,
-0.024562008678913116,
0.03286067023873329,
0.05748386308550835,
-0.13837146759033203,
0.052288178354501724,
0.026701029390096664,
0.17287513613700867,
0.06912947446107864,
-0.026962896808981895,
0.05531030148267746,
0.0370413139462471,
0.012197296135127544,
0.11829415708780289,
0.11951232701539993,
0.129713773727417,
-0.05275316536426544,
0.06956546753644943,
0.023445267230272293,
-0.1346282958984375,
-0.14011336863040924,
0.11599337309598923,
-0.06735454499721527,
0.09612362086772919,
-0.027837667614221573,
0.17665213346481323,
0.22288598120212555,
-0.19617116451263428,
0.05413440987467766,
-0.06189922243356705,
-0.11422734707593918,
-0.08304262161254883,
-0.06275875866413116,
-0.08041398972272873,
-0.18077926337718964,
0.037104200571775436,
-0.09803329408168793,
0.027870675548911095,
0.11489413678646088,
0.03911091387271881,
0.05372143164277077,
0.21941450238227844,
0.044073160737752914,
0.06589208543300629,
0.0754489153623581,
0.027494490146636963,
0.0024235162418335676,
-0.06889920681715012,
-0.06160391494631767,
-0.017846539616584778,
-0.09140440821647644,
0.014648144133388996,
-0.06584921479225159,
-0.11034766584634781,
0.030117304995656013,
0.031655892729759216,
-0.09246152639389038,
0.03187796473503113,
0.023124946281313896,
0.03555810824036598,
0.08905348926782608,
0.013400869444012642,
-0.0020808111876249313,
-0.039300572127103806,
0.25009170174598694,
-0.0790150836110115,
-0.03361611068248749,
-0.14362332224845886,
0.240578755736351,
0.05021921917796135,
-0.011325502768158913,
0.007174832280725241,
-0.09594657272100449,
0.022923266515135765,
0.1809404194355011,
0.13230882585048676,
-0.01588062196969986,
-0.02097439020872116,
0.009179939515888691,
-0.004099156707525253,
-0.07458710670471191,
0.09167921543121338,
0.1124144196510315,
0.02971186861395836,
-0.10244208574295044,
-0.02129688858985901,
-0.07101336121559143,
-0.0745217427611351,
-0.040969960391521454,
0.05981442332267761,
0.04651087895035744,
0.04637489467859268,
-0.06670969724655151,
0.12916335463523865,
-0.015528888441622257,
-0.10632835328578949,
0.060949310660362244,
-0.18283464014530182,
-0.16293556988239288,
-0.05589704588055611,
-0.000817062275018543,
-0.01759485714137554,
0.05973074585199356,
0.006947912275791168,
-0.02517554722726345,
0.06489314883947372,
0.030936241149902344,
-0.009449480101466179,
-0.14118319749832153,
0.11498155444860458,
-0.09792304039001465,
0.17600856721401215,
-0.06516977399587631,
-0.007983983494341373,
0.12572267651557922,
0.03196893259882927,
-0.07433441281318665,
0.018517641350626945,
0.08919911831617355,
-0.025837747380137444,
-0.02404463104903698,
0.18485158681869507,
-0.0355788879096508,
0.10918714851140976,
0.03480175882577896,
-0.08730145543813705,
0.02642650343477726,
-0.12185556441545486,
-0.045732833445072174,
-0.07763807475566864,
0.029655568301677704,
-0.014617478474974632,
0.1343802958726883,
0.24862593412399292,
-0.06444410234689713,
-0.009533854201436043,
-0.08114730566740036,
0.024102244526147842,
0.038403887301683426,
0.07252015173435211,
-0.035257935523986816,
-0.25346410274505615,
0.030062396079301834,
0.002338557969778776,
-0.004627244081348181,
-0.22943635284900665,
-0.1117553785443306,
0.07160911709070206,
-0.05851729214191437,
-0.07232312113046646,
0.12891709804534912,
0.06437033414840698,
0.07626499980688095,
-0.05328083038330078,
-0.1492500603199005,
-0.038015175610780716,
0.18256472051143646,
-0.16951987147331238,
-0.06452617794275284
] |
null | null | transformers | # Kunocchini-7b - AWQ
- Model creator: [Test157t](https://huggingface.co/Test157t)
- Original model: [Kunocchini-7b](https://huggingface.co/Test157t/Kunocchini-7b)

| {"language": ["en"], "license": "apache-2.0", "library_name": "transformers", "tags": ["mergekit", "merge", "alpaca", "mistral"], "model_name": "Kunocchini-7b", "base_model": ["SanjiWatsuki/Kunoichi-DPO-v2-7B", "Epiculous/Fett-uccine-7B"], "model_creator": "Test157t", "model_type": "mistral", "pipeline_tag": "text-generation", "prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant\n", "quantized_by": "Suparious"} | text-generation | solidrust/Kunocchini-7b-AWQ | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"mergekit",
"merge",
"alpaca",
"en",
"base_model:SanjiWatsuki/Kunoichi-DPO-v2-7B",
"base_model:Epiculous/Fett-uccine-7B",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"4-bit",
"region:us"
] | 2024-02-08T07:03:47+00:00 | [] | [
"en"
] | TAGS
#transformers #safetensors #mistral #text-generation #mergekit #merge #alpaca #en #base_model-SanjiWatsuki/Kunoichi-DPO-v2-7B #base_model-Epiculous/Fett-uccine-7B #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us
| # Kunocchini-7b - AWQ
- Model creator: Test157t
- Original model: Kunocchini-7b
!image/jpeg
| [
"# Kunocchini-7b - AWQ\n\n- Model creator: Test157t\n- Original model: Kunocchini-7b\n\n!image/jpeg"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #mergekit #merge #alpaca #en #base_model-SanjiWatsuki/Kunoichi-DPO-v2-7B #base_model-Epiculous/Fett-uccine-7B #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n",
"# Kunocchini-7b - AWQ\n\n- Model creator: Test157t\n- Original model: Kunocchini-7b\n\n!image/jpeg"
] | [
109,
31
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #mergekit #merge #alpaca #en #base_model-SanjiWatsuki/Kunoichi-DPO-v2-7B #base_model-Epiculous/Fett-uccine-7B #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n# Kunocchini-7b - AWQ\n\n- Model creator: Test157t\n- Original model: Kunocchini-7b\n\n!image/jpeg"
] | [
-0.10694345831871033,
0.08837515860795975,
-0.002655352232977748,
0.059400055557489395,
0.033478762954473495,
0.017394866794347763,
0.21687333285808563,
0.12222938984632492,
0.0540500171482563,
0.0471656434237957,
0.11509285867214203,
0.07713142037391663,
0.07050606608390808,
0.1121034324169159,
-0.03851211071014404,
-0.15266887843608856,
0.1048201471567154,
0.06350798904895782,
-0.044042132794857025,
0.0823831632733345,
0.09367155283689499,
-0.0626320168375969,
0.11568769812583923,
-0.0024239264894276857,
-0.11870773881673813,
0.039773810654878616,
-0.03710804879665375,
-0.05385814979672432,
0.06824059039354324,
0.06649096310138702,
0.09257017821073532,
0.11564501374959946,
0.027261916548013687,
-0.12004461884498596,
0.04364005848765373,
-0.002356848679482937,
-0.02274712175130844,
0.019857844337821007,
0.06577545404434204,
0.0011486461153253913,
0.009340407326817513,
0.011233238503336906,
-0.061544086784124374,
0.025625841692090034,
-0.07865601032972336,
0.008487682789564133,
-0.051129039376974106,
0.19617372751235962,
0.057175058871507645,
0.08595076203346252,
0.0037178248167037964,
0.06286215037107468,
0.011900614947080612,
0.048817917704582214,
0.14222659170627594,
-0.16257792711257935,
-0.08505608141422272,
0.0713549256324768,
0.0516229122877121,
0.01979435235261917,
-0.011344868689775467,
0.030415836721658707,
0.09335020184516907,
-0.054693035781383514,
-0.07514118403196335,
-0.08627799153327942,
0.16299527883529663,
-0.039882298558950424,
-0.07403481006622314,
0.030967755243182182,
0.19231505692005157,
0.04255615174770355,
-0.05289415642619133,
-0.006824058946222067,
-0.0635426789522171,
0.07697898149490356,
-0.034487947821617126,
-0.04852961003780365,
0.011217574588954449,
0.02214064449071884,
0.08065448701381683,
-0.05101887509226799,
-0.09792247414588928,
-0.038427047431468964,
-0.0721009224653244,
0.18458051979541779,
0.03247484937310219,
0.0061336699873209,
-0.013434451073408127,
0.04784552752971649,
-0.06169181317090988,
-0.16760031878948212,
-0.039392367005348206,
-0.07163409143686295,
0.13856373727321625,
0.014333552680909634,
-0.014816848561167717,
-0.0990481749176979,
0.16363756358623505,
0.16250163316726685,
-0.038292210549116135,
0.031009873375296593,
0.04837137088179588,
0.04887644946575165,
-0.018222618848085403,
0.028324313461780548,
-0.06454761326313019,
-0.08411970734596252,
0.04107106477022171,
0.13393817842006683,
0.0492703914642334,
-0.01721910387277603,
-0.10630892962217331,
-0.0535694882273674,
0.0028695606160908937,
-0.009167959913611412,
0.0988791435956955,
0.03080226294696331,
-0.052990954369306564,
-0.048637501895427704,
0.14069803059101105,
-0.06384105980396271,
-0.0359555222094059,
-0.008610769174993038,
-0.01065288856625557,
0.11175324022769928,
-0.009482675231993198,
0.05079144239425659,
-0.018529411405324936,
0.00632982375100255,
-0.06539323925971985,
-0.04506229609251022,
-0.0428769588470459,
0.013101059943437576,
0.05717470869421959,
-0.14971373975276947,
0.03293254226446152,
-0.15330642461776733,
-0.10736873745918274,
-0.014002295210957527,
0.053340714424848557,
-0.040013328194618225,
-0.03224565088748932,
-0.018533583730459213,
0.02029179222881794,
-0.0007972113089635968,
0.009320612996816635,
-0.09661341458559036,
-0.043732985854148865,
0.001506506116129458,
0.039688270539045334,
0.051925648003816605,
-0.16698268055915833,
0.023512857034802437,
-0.11060357093811035,
0.08495296537876129,
-0.1132861077785492,
-0.02759663388133049,
-0.041971657425165176,
0.1046627014875412,
-0.08091425150632858,
0.008859742432832718,
-0.06904879212379456,
0.010073421522974968,
0.056515175849199295,
0.22229835391044617,
-0.07601681351661682,
-0.041429005563259125,
0.09488344192504883,
-0.1292862743139267,
-0.20886412262916565,
0.10518988966941833,
0.020802132785320282,
0.08350767195224762,
0.044537436217069626,
0.19053484499454498,
0.056610651314258575,
-0.057960424572229385,
-0.028492428362369537,
0.0360339991748333,
0.046463917940855026,
-0.07076435536146164,
0.11046606302261353,
-0.011695461347699165,
-0.08593232929706573,
0.03942469507455826,
-0.08455090969800949,
-0.006268287543207407,
-0.02513730153441429,
-0.0829513892531395,
-0.05248658359050751,
-0.09015276283025742,
-0.0008608024218119681,
0.0013817056315019727,
0.0393267385661602,
-0.03050236590206623,
-0.042979028075933456,
-0.00026446414995007217,
0.09140709042549133,
-0.052438460290431976,
-0.0059524220414459705,
-0.06582346558570862,
0.24359437823295593,
-0.04848488047719002,
0.012980757281184196,
-0.02565019577741623,
0.02357546240091324,
0.019111711531877518,
-0.011692154221236706,
0.00924463476985693,
-0.05711241066455841,
0.03775882348418236,
0.012964693829417229,
-0.03897593542933464,
-0.06124011054635048,
0.031207680702209473,
0.027290623635053635,
0.011015775613486767,
-0.18574154376983643,
0.04047476127743721,
-0.04775339737534523,
0.17545770108699799,
-0.0629919096827507,
0.03451739251613617,
0.01173896063119173,
0.13627147674560547,
-0.03133765235543251,
0.043443650007247925,
-0.0027429668698459864,
-0.011928299441933632,
-0.04430466517806053,
0.007742779795080423,
0.09497663378715515,
0.006300770677626133,
-0.1187305822968483,
0.10185106098651886,
-0.03663884103298187,
0.10164443403482437,
0.13240964710712433,
-0.040535055100917816,
0.010055337101221085,
-0.09309867769479752,
0.01567198894917965,
-0.019946519285440445,
0.06666947156190872,
0.02379675954580307,
-0.003723129630088806,
-0.01850486733019352,
0.08024061471223831,
-0.14345481991767883,
0.054664939641952515,
0.03937987983226776,
-0.06671424210071564,
-0.06929043680429459,
0.07539154589176178,
0.07845601439476013,
-0.1897772252559662,
0.13432934880256653,
0.2865854799747467,
-0.04135442525148392,
0.17093366384506226,
-0.03616863489151001,
-0.0244076456874609,
-0.01347016915678978,
0.09132616221904755,
-0.003289590822532773,
0.09265884011983871,
-0.14608901739120483,
0.054026953876018524,
0.008354727178812027,
-0.03821401298046112,
0.04049594700336456,
-0.10357672721147537,
-0.026389528065919876,
-0.003922414034605026,
-0.02240508422255516,
0.07220017910003662,
0.03825140371918678,
-0.0658283531665802,
0.060931071639060974,
0.03525412827730179,
-0.11242973804473877,
0.04632286727428436,
0.030119312927126884,
-0.03534865379333496,
0.19479838013648987,
-0.07722045481204987,
-0.18762138485908508,
-0.1368200182914734,
-0.025191236287355423,
-0.11515876650810242,
0.017819801345467567,
0.060536980628967285,
-0.029559185728430748,
-0.00029138641548343003,
-0.1069164127111435,
-0.027588175609707832,
0.030048642307519913,
-0.019967250525951385,
0.01603987254202366,
0.011082153767347336,
-0.01116324495524168,
-0.05786465108394623,
-0.03008054383099079,
0.037756212055683136,
0.002605903660878539,
0.07843668758869171,
-0.12307249009609222,
0.147603839635849,
0.08821051567792892,
0.0213176216930151,
-0.011362455785274506,
0.013997248373925686,
0.15636372566223145,
-0.07387209683656693,
0.07408527284860611,
0.20436304807662964,
-0.00759287690743804,
0.06565940380096436,
0.21329297125339508,
-0.009305387735366821,
-0.03531460464000702,
-0.0017357835313305259,
-0.03516113758087158,
-0.008200585842132568,
-0.1597650647163391,
-0.006986959837377071,
-0.05513325333595276,
0.10939420759677887,
0.04431111738085747,
0.019261041656136513,
0.047614846378564835,
0.1649271696805954,
-0.05795620381832123,
-0.02697593905031681,
0.010827505961060524,
0.07655475288629532,
0.11250197887420654,
-0.018582701683044434,
0.10208349674940109,
-0.061630502343177795,
-0.06884650141000748,
0.09752849489450455,
0.0933661162853241,
0.05800405517220497,
0.03121265396475792,
0.06967899948358536,
0.06925492733716965,
0.16145013272762299,
0.07674716413021088,
0.0308830663561821,
-0.017675040289759636,
-0.03440592437982559,
-0.035650696605443954,
-0.09370457381010056,
0.03274073079228401,
0.06229792535305023,
-0.13268665969371796,
-0.0066010430455207825,
-0.04042026400566101,
0.03623558580875397,
0.03204241022467613,
0.20505695044994354,
0.06638790667057037,
-0.23499764502048492,
-0.020858369767665863,
0.03132517635822296,
0.07574925571680069,
-0.07476027309894562,
-0.006561460439115763,
-0.007665209472179413,
-0.04492152854800224,
0.11763351410627365,
-0.018584774807095528,
0.09634773433208466,
0.06219075247645378,
0.020316531881690025,
-0.03422413766384125,
0.017391392961144447,
-0.039626434445381165,
0.053229741752147675,
-0.16038604080677032,
0.19878152012825012,
0.007358587346971035,
-0.007833519950509071,
-0.009755454957485199,
0.01473112590610981,
0.053025029599666595,
0.24784527719020844,
0.11616979539394379,
-0.006603278685361147,
0.014688167721033096,
-0.06859263777732849,
-0.09265754371881485,
0.03367418423295021,
0.06839963793754578,
0.002369755646213889,
0.03875814750790596,
-0.01273959781974554,
-0.006167527288198471,
-0.015966765582561493,
0.14745362102985382,
-0.18196316063404083,
-0.1049974262714386,
0.08230815827846527,
0.026112448424100876,
0.05988689139485359,
-0.07782554626464844,
-0.05006624758243561,
-0.04588116332888603,
0.11713286489248276,
-0.010188696905970573,
-0.03939315304160118,
-0.1000961884856224,
-0.07009757310152054,
0.07778429239988327,
-0.08506301790475845,
0.04486299306154251,
-0.12186544388532639,
0.018026789650321007,
-0.06107460707426071,
-0.09575239568948746,
0.07200507819652557,
-0.08732561022043228,
-0.10435625910758972,
-0.022862423211336136,
0.030361570417881012,
-0.04863093048334122,
-0.01380886510014534,
0.05203234776854515,
-0.014596770517528057,
-0.07646326720714569,
-0.04660730063915253,
-0.054169707000255585,
-0.0006603325600735843,
-0.01973443292081356,
0.0430070236325264,
-0.08160298317670822,
-0.16673357784748077,
-0.04150564968585968,
-0.07398243993520737,
0.04348742961883545,
0.2752189636230469,
0.00349480751901865,
0.000834732607472688,
0.1888323426246643,
-0.07606197893619537,
-0.23823608458042145,
-0.05253651365637779,
-0.07108201086521149,
0.009163605980575085,
-0.056927841156721115,
-0.07297544181346893,
0.11119537800550461,
0.11533138155937195,
-0.03545061871409416,
0.05683527886867523,
-0.1969749480485916,
-0.08713170886039734,
0.04520638287067413,
0.09674937278032303,
0.2564798891544342,
-0.16635507345199585,
-0.045188091695308685,
-0.10162651538848877,
-0.1817694753408432,
-0.018172770738601685,
-0.10304684191942215,
0.05851690098643303,
-0.06800291687250137,
-0.08807330578565598,
0.005015425384044647,
-0.025050630792975426,
0.14653830230236053,
-0.0673719272017479,
0.027233703061938286,
-0.10804368555545807,
0.002684718696400523,
0.040932267904281616,
-0.016572941094636917,
0.03939748927950859,
-0.17123627662658691,
0.02636622078716755,
-0.026859426870942116,
-0.04087754711508751,
-0.019613614305853844,
0.08321937918663025,
-0.06307190656661987,
-0.06203572452068329,
-0.0373242050409317,
0.08310779184103012,
-0.026611115783452988,
0.03673458844423294,
0.14064115285873413,
-0.04816897213459015,
0.07356615364551544,
0.18389791250228882,
0.043272245675325394,
-0.14060698449611664,
0.11373649537563324,
-0.0450933538377285,
-0.061320219188928604,
0.044262923300266266,
-0.09993881732225418,
-0.0030682901851832867,
0.09035871177911758,
-0.021869348362088203,
0.05812738463282585,
0.01391973253339529,
-0.04144345223903656,
-0.015737386420369148,
0.1310146600008011,
-0.1139516532421112,
-0.09141650795936584,
-0.07421579957008362,
0.007453897036612034,
-0.046899113804101944,
0.07088746875524521,
0.16702629625797272,
-0.04357350245118141,
-0.015614019706845284,
-0.015955301001667976,
-0.00973939523100853,
-0.0777682214975357,
0.06450272351503372,
0.07436521351337433,
-0.010856782086193562,
-0.020933859050273895,
0.10369417071342468,
0.015153088606894016,
-0.10424739867448807,
-0.038832128047943115,
0.06202185899019241,
-0.09908866882324219,
-0.1192922294139862,
0.0542784109711647,
0.1235181912779808,
-0.17879176139831543,
-0.07754045724868774,
-0.18311546742916107,
-0.11218902468681335,
0.018620071932673454,
0.1291436403989792,
0.09867703169584274,
-0.03735554218292236,
0.09556981921195984,
-0.06019776314496994,
-0.008470225147902966,
0.06874653697013855,
0.05431973189115524,
0.08463814854621887,
-0.08902809768915176,
-0.1117807999253273,
-0.02519916743040085,
0.026101967319846153,
-0.03505094349384308,
0.016412673518061638,
-0.1498679518699646,
-0.02457062341272831,
-0.18844681978225708,
0.03864533081650734,
-0.10221990197896957,
-0.03462034836411476,
-0.04952198639512062,
-0.02356228604912758,
-0.031514719128608704,
0.005504976958036423,
-0.04321534186601639,
-0.016968494281172752,
-0.022191008552908897,
0.0775766670703888,
-0.059164080768823624,
0.0032000462524592876,
0.07703354954719543,
-0.0453357920050621,
0.05293408781290054,
-0.016191862523555756,
-0.027129940688610077,
-0.004312545992434025,
-0.18473151326179504,
0.044226035475730896,
-0.00761444540694356,
0.037790536880493164,
-0.029579924419522285,
-0.050051867961883545,
0.0061639281921088696,
0.0856141671538353,
-0.011394818313419819,
0.015613923780620098,
0.10398081690073013,
-0.06350281834602356,
-0.04419635608792305,
-0.05447117239236832,
-0.10204600542783737,
-0.042311713099479675,
0.04240524768829346,
0.12283902615308762,
0.018809624016284943,
0.10905244946479797,
-0.045724958181381226,
0.02770077995955944,
-0.10159250348806381,
0.014392399229109287,
0.031665053218603134,
-0.12462112307548523,
-0.08425511419773102,
-0.08890847116708755,
-0.012739705853164196,
0.0137145034968853,
0.07739042490720749,
-0.05167568475008011,
-0.061862289905548096,
0.029717842116951942,
-0.08917322754859924,
0.13162755966186523,
0.03857976943254471,
0.3168564736843109,
0.05862647294998169,
0.025307245552539825,
-0.0719049870967865,
0.029027406126260757,
0.02436830662190914,
0.10360740125179291,
-0.00006599030893994495,
0.2621728181838989,
-0.02071145921945572,
0.11158876866102219,
0.10817354172468185,
0.08236848562955856,
0.06366448104381561,
-0.020268630236387253,
-0.026007741689682007,
0.025553220883011818,
-0.002916125813499093,
0.1324785202741623,
0.18368631601333618,
-0.10366468876600266,
-0.021171243861317635,
-0.030717261135578156,
-0.04995088279247284,
-0.07840476185083389,
-0.1494816392660141,
-0.11120893061161041,
-0.16940078139305115,
0.014354556798934937,
-0.0757032185792923,
-0.022322416305541992,
0.01374739222228527,
0.014571310952305794,
-0.056952398270368576,
0.16461756825447083,
0.020875990390777588,
-0.060756079852581024,
0.10373833775520325,
0.0030190146062523127,
-0.038141194730997086,
0.020127514377236366,
-0.09959296882152557,
-0.0221351720392704,
-0.020571015775203705,
-0.04912697896361351,
-0.002162807621061802,
0.008001435548067093,
0.039726391434669495,
-0.018349919468164444,
-0.08550199866294861,
-0.00014743457722943276,
-0.016814596951007843,
0.06082843244075775,
0.09173224121332169,
-0.004165548365563154,
0.010006936267018318,
0.03233780339360237,
0.11094147711992264,
0.024360336363315582,
-0.12663418054580688,
-0.06276445835828781,
0.00338231073692441,
-0.043097563087940216,
0.0273085068911314,
-0.016246473416686058,
-0.05048005282878876,
0.04917418211698532,
0.20901130139827728,
0.1970893144607544,
-0.09179440140724182,
0.003718996187672019,
-0.01545010507106781,
0.0183198731392622,
0.017027320340275764,
0.05909667909145355,
0.022815803065896034,
0.007281818427145481,
-0.039136696606874466,
0.017343290150165558,
-0.04886132478713989,
-0.011537056416273117,
-0.1340760886669159,
-0.014924183487892151,
0.079200379550457,
-0.06342765688896179,
-0.057148322463035583,
0.1050109714269638,
-0.061642035841941833,
0.04053197801113129,
0.05288929119706154,
-0.104364775121212,
-0.08471541106700897,
-0.08964294195175171,
0.07927091419696808,
-0.0021636676974594593,
0.026074375957250595,
-0.06462966650724411,
0.013482565991580486,
0.020332051441073418,
0.01100755762308836,
-0.0864366739988327,
-0.107904352247715,
0.020057957619428635,
-0.032615870237350464,
0.10123427212238312,
0.005694798193871975,
0.02958795800805092,
0.12015337496995926,
0.028509385883808136,
-0.09214106947183609,
0.06970694661140442,
0.03455035388469696,
-0.015826784074306488,
0.009137529879808426,
-0.05964171141386032,
-0.0019439575262367725,
0.053891871124506,
0.05522589385509491,
-0.0055726137943565845,
0.041366636753082275,
0.05460454151034355,
-0.032426897436380386,
-0.10693240165710449,
0.10044541954994202,
-0.14932651817798615,
0.14032982289791107,
0.09777020663022995,
-0.03898324444890022,
-0.05369940772652626,
-0.018632199615240097,
0.0773223340511322,
0.07755447179079056,
-0.024158187210559845,
-0.012860003858804703,
-0.13829514384269714,
-0.04089939221739769,
-0.03939628228545189,
0.044286470860242844,
-0.2607298493385315,
-0.07150134444236755,
-0.15354925394058228,
0.0156716275960207,
-0.08225303143262863,
0.0704742893576622,
0.15341144800186157,
0.002049789298325777,
-0.004756886046379805,
-0.17670343816280365,
-0.03834297135472298,
0.06192617863416672,
-0.030890600755810738,
-0.11186923086643219
] |
null | null | transformers |
dreamsim! now in quasi-transformers quasi-diffusers form.
this probably won't work for you! but if it works for what i'm experimenting with, i'll try to get it upstreamed. | {"license": "mit", "library_name": "transformers", "datasets": ["PerceptionEval/DreamSim"]} | null | neggles/dreamsim | [
"transformers",
"dataset:PerceptionEval/DreamSim",
"license:mit",
"endpoints_compatible",
"region:us"
] | 2024-02-08T07:04:13+00:00 | [] | [] | TAGS
#transformers #dataset-PerceptionEval/DreamSim #license-mit #endpoints_compatible #region-us
|
dreamsim! now in quasi-transformers quasi-diffusers form.
this probably won't work for you! but if it works for what i'm experimenting with, i'll try to get it upstreamed. | [] | [
"TAGS\n#transformers #dataset-PerceptionEval/DreamSim #license-mit #endpoints_compatible #region-us \n"
] | [
33
] | [
"passage: TAGS\n#transformers #dataset-PerceptionEval/DreamSim #license-mit #endpoints_compatible #region-us \n"
] | [
-0.09553434699773788,
0.20660559833049774,
-0.005656282417476177,
-0.013924038037657738,
0.13537795841693878,
0.04216877371072769,
0.11737984418869019,
0.04683246091008186,
0.0806841105222702,
-0.04113897681236267,
0.13893717527389526,
0.1645231693983078,
0.0067439922131598,
0.07789706438779831,
-0.05525866150856018,
-0.22707155346870422,
0.11177395284175873,
0.0680987685918808,
-0.09475668519735336,
0.05864325165748596,
0.08907055854797363,
-0.08472836017608643,
0.1331523209810257,
-0.034887995570898056,
-0.16163477301597595,
-0.020861869677901268,
0.004881520755589008,
-0.07683810591697693,
0.08707471191883087,
0.06699976325035095,
0.11661309003829956,
0.04708123952150345,
-0.027788333594799042,
-0.21289552748203278,
0.0004023966030217707,
-0.0054352907463908195,
-0.060670774430036545,
0.01554957590997219,
0.057104483246803284,
-0.007248400244861841,
0.031283460557460785,
0.01616496406495571,
0.023076295852661133,
0.07154383510351181,
-0.1306203156709671,
-0.07978689670562744,
-0.0701371356844902,
0.03949403017759323,
0.04499528184533119,
0.0637676939368248,
0.03414011374115944,
0.09762134402990341,
-0.06180764362215996,
0.09000307321548462,
0.08333111554384232,
-0.23369216918945312,
0.022468024864792824,
0.18011319637298584,
-0.01524700690060854,
-0.07631789892911911,
-0.009051002562046051,
0.08304901421070099,
0.05619519576430321,
0.06201355531811714,
0.003979732748121023,
-0.05543626844882965,
-0.0994584858417511,
0.08223167806863785,
-0.028284912928938866,
-0.09199236333370209,
0.2851931154727936,
-0.0014827307313680649,
0.022476566955447197,
-0.05384470149874687,
-0.082785964012146,
-0.07607247680425644,
-0.03203681483864784,
0.005635999608784914,
0.044800564646720886,
0.11108177900314331,
-0.05350381135940552,
0.029902512207627296,
-0.12947943806648254,
0.05167507007718086,
-0.21972733736038208,
0.052030544728040695,
-0.03029412217438221,
0.12235627323389053,
-0.16137051582336426,
0.01776217296719551,
-0.006802916526794434,
-0.07415314763784409,
-0.07600042223930359,
-0.12902721762657166,
0.00957051943987608,
-0.0034834975376725197,
-0.04221267253160477,
0.050961337983608246,
0.08292660862207413,
0.08092706650495529,
0.12537901103496552,
-0.01247633621096611,
0.05614322051405907,
0.08141698688268661,
0.07599926739931107,
0.07944483309984207,
-0.006299377419054508,
-0.052733466029167175,
-0.01265911478549242,
-0.23292873799800873,
0.03910542652010918,
-0.024027051404118538,
-0.10516233742237091,
-0.009287286549806595,
-0.0834418535232544,
0.09032749384641647,
-0.008855435997247696,
0.04015679284930229,
-0.08317288011312485,
0.0017776215681806207,
0.04537881538271904,
0.002340463688597083,
-0.01756098121404648,
-0.06713543832302094,
-0.009103861637413502,
0.09609626978635788,
0.04421694949269295,
0.00591112207621336,
0.014718413352966309,
0.04514341056346893,
-0.08600199967622757,
-0.05108005180954933,
-0.03210766986012459,
-0.02124641090631485,
0.10393243283033371,
-0.13116392493247986,
0.1410677582025528,
-0.1394238919019699,
-0.19902965426445007,
0.017823245376348495,
0.12455333024263382,
-0.020740581676363945,
-0.01511826179921627,
-0.013556652702391148,
-0.10012075304985046,
-0.014941747300326824,
-0.08447451144456863,
-0.0977126806974411,
-0.046439215540885925,
0.017550820484757423,
-0.05606155842542648,
0.06286902725696564,
-0.17541900277137756,
0.05847705528140068,
-0.15540413558483124,
0.04732932522892952,
0.019617605954408646,
0.09161300212144852,
-0.04950138181447983,
0.24770532548427582,
-0.026740198954939842,
-0.0102377999573946,
0.036229152232408524,
0.047273557633161545,
-0.07993825525045395,
0.1494024693965912,
-0.1937168538570404,
-0.025129791349172592,
0.15651078522205353,
-0.14032362401485443,
-0.31840604543685913,
0.02360256016254425,
-0.024805301800370216,
0.11878085881471634,
0.10359925031661987,
0.17438538372516632,
0.04516223445534706,
-0.11881329864263535,
-0.03840607404708862,
0.11432869732379913,
-0.1321781575679779,
-0.22938916087150574,
0.027864381670951843,
0.03524081036448479,
0.006582463160157204,
0.07066554576158524,
0.011369592510163784,
0.1279127597808838,
-0.05281829833984375,
-0.06777223199605942,
-0.07718216627836227,
-0.0639619380235672,
-0.058362238109111786,
0.04459923505783081,
0.03967861831188202,
-0.08086325973272324,
0.07230120897293091,
0.0669206753373146,
-0.0037967534735798836,
0.027633454650640488,
0.006537873297929764,
-0.08237159252166748,
-0.049532853066921234,
-0.07218358665704727,
-0.0160194281488657,
-0.118702732026577,
-0.12026922404766083,
-0.035863686352968216,
0.050116218626499176,
-0.010598307475447655,
0.21554605662822723,
0.05534842982888222,
-0.03190666809678078,
0.00981411337852478,
-0.012933679856359959,
0.14138276875019073,
0.09336616098880768,
-0.023943915963172913,
-0.11182009428739548,
0.02109072171151638,
-0.056710004806518555,
-0.13444426655769348,
-0.013912859372794628,
-0.011419269256293774,
0.11561614274978638,
0.17873866856098175,
0.02207961492240429,
0.026181381195783615,
-0.036211445927619934,
0.005528942681849003,
-0.039418775588274,
-0.03551527485251427,
0.09410951286554337,
-0.008728652261197567,
-0.06997590512037277,
0.24166417121887207,
-0.10108049213886261,
0.31352123618125916,
0.1998690813779831,
-0.16316138207912445,
0.07624776661396027,
0.025988852605223656,
-0.01703352853655815,
0.018410250544548035,
0.0487106516957283,
0.05597511678934097,
0.01732088439166546,
-0.0008313836879096925,
0.0993274450302124,
-0.005919596645981073,
0.03622612729668617,
0.02424526773393154,
-0.05915059894323349,
-0.1293046623468399,
0.017092861235141754,
0.04035578668117523,
-0.21883074939250946,
0.22792735695838928,
0.21350207924842834,
0.15809865295886993,
0.11994516849517822,
-0.09193538874387741,
0.02201629802584648,
-0.03662525862455368,
-0.056648094207048416,
-0.08934912085533142,
0.15332482755184174,
-0.23860833048820496,
-0.007627315819263458,
0.07947767525911331,
0.02748788334429264,
0.08188427239656448,
-0.16462519764900208,
-0.08397965133190155,
0.05506080016493797,
0.005129815079271793,
-0.02544054202735424,
0.09869993478059769,
-0.039929043501615524,
0.0074763428419828415,
-0.0476657897233963,
-0.04157363623380661,
0.1749042272567749,
-0.013554764911532402,
-0.086812324821949,
0.12715479731559753,
-0.17455539107322693,
-0.25827470421791077,
-0.14483052492141724,
-0.15520676970481873,
-0.009532546624541283,
0.03731883689761162,
0.10128939896821976,
-0.03392410650849342,
-0.09086950123310089,
0.016852807253599167,
0.0348728783428669,
-0.11428970843553543,
0.011410009115934372,
0.01412571407854557,
0.08924486488103867,
-0.09950073808431625,
-0.08586273342370987,
-0.043724652379751205,
0.006384312640875578,
0.0644349604845047,
0.09423090517520905,
-0.12671183049678802,
0.12181013822555542,
0.1150960773229599,
0.02996620535850525,
0.03441304340958595,
-0.022271761670708656,
0.3300088047981262,
-0.06968033313751221,
-0.041210126131772995,
0.1673230081796646,
0.05923861265182495,
0.03393181785941124,
0.28512686491012573,
0.04592311382293701,
-0.11064114421606064,
-0.006611230317503214,
-0.09147709608078003,
-0.12858152389526367,
-0.2057800590991974,
-0.12616239488124847,
-0.10837600380182266,
0.007860014215111732,
0.014218566007912159,
0.06536629050970078,
0.08715212345123291,
0.15127751231193542,
0.04385295882821083,
-0.001686671283096075,
-0.07435563951730728,
0.06439968198537827,
0.2443879097700119,
-0.08145128935575485,
0.034376274794340134,
-0.15137791633605957,
-0.05965716019272804,
0.09347403049468994,
0.12424714863300323,
0.2310456782579422,
0.13134074211120605,
0.0553186759352684,
0.11654740571975708,
0.1227150708436966,
0.10477849841117859,
0.08278580754995346,
0.09767355024814606,
-0.0409814827144146,
-0.0350472554564476,
-0.0013125715777277946,
-0.10817883163690567,
0.11249913275241852,
0.018107717856764793,
-0.11769511550664902,
0.0070450264029204845,
-0.1043064296245575,
0.0449703112244606,
0.11203697323799133,
0.12761034071445465,
-0.24540628492832184,
-0.04817618802189827,
0.12277445942163467,
0.04725099727511406,
-0.06485173851251602,
0.0784311294555664,
0.10058284550905228,
-0.06378951668739319,
0.10478752851486206,
-0.04833066090941429,
0.08915577083826065,
0.009952683933079243,
0.038008540868759155,
-0.021575454622507095,
-0.12952812016010284,
0.08205513656139374,
0.08708885312080383,
-0.2912271320819855,
0.257068008184433,
-0.013913744129240513,
0.026260558515787125,
-0.07687260210514069,
-0.008632968179881573,
-0.02166871353983879,
0.266574889421463,
0.13170628249645233,
0.05148623138666153,
-0.2108079493045807,
-0.09565584361553192,
0.08658729493618011,
0.028165195137262344,
0.06147608906030655,
0.04226473346352577,
-0.0348474383354187,
-0.0011810768628492951,
0.0066399830393493176,
0.00572485476732254,
0.08654622733592987,
-0.017437972128391266,
-0.17166611552238464,
0.013921558856964111,
0.11509037017822266,
0.14049725234508514,
-0.03226577118039131,
0.028414899483323097,
-0.07145722955465317,
0.07786168903112411,
-0.07891088724136353,
-0.059552714228630066,
-0.11808145046234131,
-0.17970016598701477,
0.0928245261311531,
-0.03612440824508667,
0.08966708183288574,
-0.07408636808395386,
-0.005311200860887766,
-0.06351044028997421,
-0.19280792772769928,
0.11114899069070816,
-0.16284145414829254,
0.007827265188097954,
-0.09300285577774048,
0.05011343955993652,
-0.07643221318721771,
-0.004925127141177654,
0.02608652412891388,
0.04684680700302124,
-0.07928918302059174,
-0.10950395464897156,
0.022762250155210495,
0.11488012224435806,
-0.022936517372727394,
0.05789798125624657,
0.027557700872421265,
0.013104011304676533,
0.064963199198246,
-0.04644959047436714,
0.1396130472421646,
0.1597082018852234,
-0.08468138426542282,
0.1906338781118393,
0.1325991302728653,
-0.10725385695695877,
-0.33115053176879883,
-0.06292341649532318,
-0.18998047709465027,
-0.025839364156126976,
0.028059333562850952,
-0.1268288493156433,
0.1403195858001709,
-0.026790976524353027,
-0.057760320603847504,
0.06424132734537125,
-0.2740081250667572,
-0.08194388449192047,
0.12444309145212173,
0.06343525648117065,
0.38828858733177185,
-0.1096118837594986,
-0.07331959903240204,
-0.09394952654838562,
-0.2595784664154053,
0.22406154870986938,
0.06399662792682648,
0.07526455074548721,
0.01007335726171732,
0.12615616619586945,
-0.0008151205256581306,
-0.05844452604651451,
0.18669266998767853,
0.050061408430337906,
0.10270892083644867,
-0.09854111075401306,
0.024341700598597527,
0.2060420662164688,
-0.03246891871094704,
0.04771234095096588,
0.05794694647192955,
0.03275361657142639,
-0.089153952896595,
-0.00928431935608387,
-0.07180027663707733,
0.01652958244085312,
0.018974091857671738,
-0.08994245529174805,
-0.09135329723358154,
0.0200482290238142,
0.09401123225688934,
-0.027098197489976883,
0.2228863686323166,
0.028692156076431274,
-0.06373138725757599,
0.08314364403486252,
0.0035508086439222097,
-0.20287054777145386,
-0.12269794940948486,
-0.06548654288053513,
-0.07688810676336288,
0.054004162549972534,
-0.16719746589660645,
0.0019848693627864122,
0.1448289006948471,
-0.028078092262148857,
0.055935703217983246,
0.0787605345249176,
-0.033588748425245285,
0.0013004835927858949,
0.10252776741981506,
-0.0867636650800705,
-0.1759173423051834,
-0.009713626466691494,
0.022715633735060692,
0.1120329424738884,
0.09544051438570023,
0.1111547127366066,
-0.029010754078626633,
0.03450216352939606,
-0.047546226531267166,
0.041306816041469574,
-0.06970246881246567,
0.028277644887566566,
0.010237579233944416,
0.011821093037724495,
-0.1589604765176773,
0.1320519745349884,
-0.04254766181111336,
-0.26020267605781555,
-0.003697126405313611,
-0.04225005581974983,
-0.1602129489183426,
-0.10469192266464233,
-0.034669987857341766,
0.06916399300098419,
-0.21498771011829376,
-0.10927136987447739,
0.005444599315524101,
-0.1694122850894928,
0.04712466895580292,
0.15882664918899536,
0.09708710759878159,
0.15601037442684174,
-0.035269610583782196,
-0.08019888401031494,
-0.012369023635983467,
-0.03429005295038223,
-0.07111550867557526,
0.04300657659769058,
-0.18891113996505737,
-0.07125872373580933,
-0.03721155598759651,
0.10534825921058655,
-0.07665479928255081,
-0.050193846225738525,
-0.10536644607782364,
0.04514964669942856,
-0.16302037239074707,
0.004627868067473173,
-0.19037973880767822,
-0.01747826859354973,
0.05336320027709007,
-0.06097705289721489,
-0.0831555649638176,
0.009614247828722,
-0.12558239698410034,
0.05709642916917801,
0.045144643634557724,
0.057816389948129654,
-0.03798453137278557,
-0.061361417174339294,
0.05758316442370415,
-0.020429028198122978,
0.04902474582195282,
0.05982846021652222,
-0.10840511322021484,
0.06917490810155869,
-0.22129172086715698,
-0.12378169596195221,
0.1485210657119751,
-0.001393225509673357,
0.08897166699171066,
0.036264900118112564,
-0.057122863829135895,
0.13070887327194214,
0.001988231437280774,
0.05031447857618332,
-0.013835330493748188,
-0.12811145186424255,
-0.11723881959915161,
0.019172176718711853,
-0.07539200037717819,
-0.043730467557907104,
-0.04261047765612602,
0.196062833070755,
0.001733769429847598,
0.16449320316314697,
-0.01913974992930889,
0.07085579633712769,
-0.006853039842098951,
-0.013516467995941639,
-0.013735753484070301,
-0.13847889006137848,
-0.029179366305470467,
-0.05337823927402496,
-0.05021489039063454,
-0.011058440431952477,
0.38617783784866333,
0.03002072684466839,
-0.08768215030431747,
0.05172811076045036,
0.09594913572072983,
-0.017271367833018303,
0.016995880752801895,
0.371626079082489,
0.06104334071278572,
0.006903562229126692,
-0.08663302659988403,
0.1322750449180603,
0.017207294702529907,
0.008209030143916607,
0.007300173863768578,
0.10277962684631348,
0.000012686929039773531,
0.15004019439220428,
0.019100407138466835,
-0.003115343861281872,
-0.03060486912727356,
-0.12627281248569489,
0.014009338803589344,
0.09930379688739777,
-0.037517305463552475,
0.006063925102353096,
0.07088585197925568,
-0.0586521290242672,
0.07333777099847794,
-0.045752525329589844,
-0.03729790449142456,
-0.10237425565719604,
-0.14150729775428772,
-0.06999283283948898,
-0.11745747923851013,
0.037805404514074326,
-0.07599306851625443,
0.0232149139046669,
0.22992996871471405,
0.039218369871377945,
-0.0587102547287941,
0.03115350566804409,
-0.037428177893161774,
-0.018405551090836525,
-0.0033954449463635683,
0.0014880939852446318,
0.00909662526100874,
-0.02879379875957966,
0.025978874415159225,
-0.08512308448553085,
-0.015955699607729912,
-0.08434951305389404,
0.03663713485002518,
-0.001932366518303752,
0.0012527921935543418,
-0.08530207723379135,
-0.06605515629053116,
-0.08520438522100449,
0.0774679183959961,
-0.05294593423604965,
0.0685124322772026,
0.019473446533083916,
0.011585202999413013,
0.053791843354701996,
0.15853819251060486,
-0.08125075697898865,
-0.07458087801933289,
-0.08145801723003387,
0.08219250291585922,
0.009424944408237934,
0.11903568357229233,
0.019733494147658348,
-0.018532540649175644,
-0.10783728212118149,
0.18576739728450775,
0.2955370843410492,
-0.037707310169935226,
-0.008036476559937,
-0.04737429320812225,
0.02171042189002037,
-0.021725622937083244,
0.1194358617067337,
-0.004593086428940296,
0.24066424369812012,
-0.07922965288162231,
0.012893162667751312,
-0.01665167137980461,
-0.06059686094522476,
-0.05922771617770195,
0.011524485424160957,
0.00543881906196475,
-0.08515490591526031,
-0.06771377474069595,
0.14879558980464935,
-0.21452240645885468,
0.08462277054786682,
0.23660168051719666,
-0.14411254227161407,
-0.012266242876648903,
-0.03307072073221207,
0.1784890741109848,
0.061595432460308075,
0.08090005815029144,
-0.06600961834192276,
-0.04407176375389099,
0.19892354309558868,
0.026173755526542664,
-0.3480692207813263,
0.01810946874320507,
0.08439300954341888,
-0.038966674357652664,
0.11835446208715439,
-0.01899244263768196,
0.05195700004696846,
0.08241178095340729,
0.06267097592353821,
-0.11775439232587814,
0.04446861520409584,
0.03165091574192047,
-0.05130511149764061,
-0.03909491002559662,
-0.019980700686573982,
-0.02090802788734436,
-0.023104393854737282,
0.03620420768857002,
-0.2225172221660614,
0.019937511533498764,
0.0703384280204773,
-0.003044441342353821,
-0.05936223268508911,
-0.02315395511686802,
-0.053833264857530594,
0.05295420438051224,
-0.018841998651623726,
-0.014783680438995361,
-0.008802643045783043,
-0.005411248188465834,
0.031996797770261765,
0.05472295358777046,
-0.10822184383869171,
-0.010968109592795372,
0.025933044031262398,
-0.06049104779958725,
0.13010752201080322,
-0.015264809131622314,
-0.15262199938297272,
-0.03871942684054375,
-0.06835196167230606,
-0.002474856097251177,
-0.048538852483034134,
0.07375658303499222,
0.10781414806842804,
0.024991948157548904,
-0.026862282305955887,
-0.02278846688568592,
0.0341036319732666,
0.03558763861656189,
-0.010147236287593842,
-0.13334517180919647
] |
null | null | null |
# Lora of furina/フリーナ/芙宁娜 (Genshin Impact)
## What Is This?
This is the LoRA model of waifu furina/フリーナ/芙宁娜 (Genshin Impact).
## How Is It Trained?
* This model is trained with [HCP-Diffusion](https://github.com/7eu7d7/HCP-Diffusion).
* The [auto-training framework](https://github.com/deepghs/cyberharem) is maintained by [DeepGHS Team](https://huggingface.co/deepghs).
* The base model used for training is [deepghs/animefull-latest](https://huggingface.co/deepghs/animefull-latest).
* Dataset used for training is the `stage3-p480-800` in [CyberHarem/furina_genshin](https://huggingface.co/datasets/CyberHarem/furina_genshin), which contains 1295 images.
* Batch size is 4, resolution is 720x720, clustering into 5 buckets.
* Batch size for regularization dataset is 1, resolution is 720x720, clustering into 20 buckets.
* Trained for 10000 steps, 40 checkpoints were saved and evaluated.
* **Trigger word is `furina_genshin`.**
* Pruned core tags for this waifu are `blue_eyes, blue_hair, bangs, white_hair, ahoge, hair_between_eyes, long_hair, multicolored_hair, hat, bow, streaked_hair, very_long_hair, breasts`. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable.
## How to Use It?
### If You Are Using A1111 WebUI v1.7+
**Just use it like the classic LoRA**. The LoRA we provided are bundled with the embedding file.
### If You Are Using A1111 WebUI v1.6 or Lower
After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.
For example, if you want to use the model from step 5500, you need to download [`5500/furina_genshin.pt`](https://huggingface.co/CyberHarem/furina_genshin/resolve/main/5500/furina_genshin.pt) as the embedding and [`5500/furina_genshin.safetensors`](https://huggingface.co/CyberHarem/furina_genshin/resolve/main/5500/furina_genshin.safetensors) for loading Lora. By using both files together, you can generate images for the desired characters.
## Which Step Should I Use?
We selected 5 good steps for you to choose. The best one is step 5500.
1760 images (1.93 GiB) were generated for auto-testing.

The base model used for generating preview images is [Meina/MeinaMix_V11](https://huggingface.co/Meina/MeinaMix_V11).
Here are the preview of the recommended steps:
| Step | Epoch | CCIP | AI Corrupt | Bikini Plus | Score | Download | pattern_0_0 | pattern_0_1 | pattern_1_0 | pattern_1_1 | pattern_2 | pattern_3 | pattern_4_0 | pattern_4_1 | pattern_5 | portrait_0 | portrait_1 | portrait_2 | full_body_0 | full_body_1 | profile_0 | profile_1 | free_0 | free_1 | shorts | maid_0 | maid_1 | miko | yukata | suit | china | bikini_0 | bikini_1 | bikini_2 | sit | squat | kneel | jump | crossed_arms | angry | smile | cry | grin | n_lie_0 | n_lie_1 | n_stand_0 | n_stand_1 | n_stand_2 | n_sex_0 | n_sex_1 |
|-------:|--------:|:----------|:-------------|:--------------|:----------|:---------------------------------------------------------------------------------------------------|:-----------------------------------------------|:-----------------------------------------------|:-----------------------------------------------|:-----------------------------------------------|:-------------------------------------------|:-------------------------------------------|:-----------------------------------------------|:-----------------------------------------------|:-------------------------------------------|:---------------------------------------------|:---------------------------------------------|:---------------------------------------------|:-----------------------------------------------|:-----------------------------------------------|:-------------------------------------------|:-------------------------------------------|:-------------------------------------|:-------------------------------------|:-------------------------------------|:-------------------------------------|:-------------------------------------|:---------------------------------|:-------------------------------------|:---------------------------------|:-----------------------------------|:-----------------------------------------|:-----------------------------------------|:-----------------------------------------|:-------------------------------|:-----------------------------------|:-----------------------------------|:---------------------------------|:-------------------------------------------------|:-----------------------------------|:-----------------------------------|:-------------------------------|:---------------------------------|:---------------------------------------|:---------------------------------------|:-------------------------------------------|:-------------------------------------------|:-------------------------------------------|:---------------------------------------|:---------------------------------------|
| 5500 | 17 | **0.991** | **0.956** | 0.849 | **0.832** | [Download](https://huggingface.co/CyberHarem/furina_genshin/resolve/main/5500/furina_genshin.zip) |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |
| 3250 | 11 | 0.990 | 0.947 | **0.850** | 0.824 | [Download](https://huggingface.co/CyberHarem/furina_genshin/resolve/main/3250/furina_genshin.zip) |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |
| 750 | 3 | 0.988 | 0.948 | 0.850 | 0.814 | [Download](https://huggingface.co/CyberHarem/furina_genshin/resolve/main/750/furina_genshin.zip) |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |
| 10000 | 31 | 0.990 | 0.917 | 0.844 | 0.814 | [Download](https://huggingface.co/CyberHarem/furina_genshin/resolve/main/10000/furina_genshin.zip) |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |
| 4750 | 15 | 0.988 | 0.932 | 0.847 | 0.812 | [Download](https://huggingface.co/CyberHarem/furina_genshin/resolve/main/4750/furina_genshin.zip) |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |
## Anything Else?
Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:
1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.
2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.
3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.
4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.
5. Individuals who finds the generated image content offensive to their values.
## All Steps
We uploaded the files in all steps. you can check the images, metrics and download them in the following links:
* [Steps From 7750 to 10000](all/0.md)
* [Steps From 5250 to 7500](all/1.md)
* [Steps From 2750 to 5000](all/2.md)
* [Steps From 250 to 2500](all/3.md)
| {"license": "mit", "tags": ["art", "not-for-all-audiences"], "datasets": ["CyberHarem/furina_genshin"], "pipeline_tag": "text-to-image"} | text-to-image | CyberHarem/furina_genshin | [
"art",
"not-for-all-audiences",
"text-to-image",
"dataset:CyberHarem/furina_genshin",
"license:mit",
"region:us"
] | 2024-02-08T07:07:03+00:00 | [] | [] | TAGS
#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/furina_genshin #license-mit #region-us
| Lora of furina/フリーナ/芙宁娜 (Genshin Impact)
========================================
What Is This?
-------------
This is the LoRA model of waifu furina/フリーナ/芙宁娜 (Genshin Impact).
How Is It Trained?
------------------
* This model is trained with HCP-Diffusion.
* The auto-training framework is maintained by DeepGHS Team.
* The base model used for training is deepghs/animefull-latest.
* Dataset used for training is the 'stage3-p480-800' in CyberHarem/furina\_genshin, which contains 1295 images.
* Batch size is 4, resolution is 720x720, clustering into 5 buckets.
* Batch size for regularization dataset is 1, resolution is 720x720, clustering into 20 buckets.
* Trained for 10000 steps, 40 checkpoints were saved and evaluated.
* Trigger word is 'furina\_genshin'.
* Pruned core tags for this waifu are 'blue\_eyes, blue\_hair, bangs, white\_hair, ahoge, hair\_between\_eyes, long\_hair, multicolored\_hair, hat, bow, streaked\_hair, very\_long\_hair, breasts'. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable.
How to Use It?
--------------
### If You Are Using A1111 WebUI v1.7+
Just use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.
### If You Are Using A1111 WebUI v1.6 or Lower
After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.
For example, if you want to use the model from step 5500, you need to download '5500/furina\_genshin.pt' as the embedding and '5500/furina\_genshin.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.
Which Step Should I Use?
------------------------
We selected 5 good steps for you to choose. The best one is step 5500.
1760 images (1.93 GiB) were generated for auto-testing.
!Metrics Plot
The base model used for generating preview images is Meina/MeinaMix\_V11.
Here are the preview of the recommended steps:
Anything Else?
--------------
Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:
1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.
2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.
3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.
4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.
5. Individuals who finds the generated image content offensive to their values.
All Steps
---------
We uploaded the files in all steps. you can check the images, metrics and download them in the following links:
* Steps From 7750 to 10000
* Steps From 5250 to 7500
* Steps From 2750 to 5000
* Steps From 250 to 2500
| [
"### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.",
"### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 5500, you need to download '5500/furina\\_genshin.pt' as the embedding and '5500/furina\\_genshin.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 5500.\n\n\n1760 images (1.93 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 7750 to 10000\n* Steps From 5250 to 7500\n* Steps From 2750 to 5000\n* Steps From 250 to 2500"
] | [
"TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/furina_genshin #license-mit #region-us \n",
"### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.",
"### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 5500, you need to download '5500/furina\\_genshin.pt' as the embedding and '5500/furina\\_genshin.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 5500.\n\n\n1760 images (1.93 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 7750 to 10000\n* Steps From 5250 to 7500\n* Steps From 2750 to 5000\n* Steps From 250 to 2500"
] | [
43,
38,
470
] | [
"passage: TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/furina_genshin #license-mit #region-us \n### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file."
] | [
0.02063136361539364,
-0.001966506242752075,
-0.004067449830472469,
0.09308246523141861,
0.08742878586053848,
0.08389410376548767,
0.22559499740600586,
0.09443168342113495,
0.13008880615234375,
-0.0774877592921257,
0.08097697049379349,
0.03464508429169655,
-0.01268518716096878,
0.02718823403120041,
-0.028996845707297325,
-0.16388915479183197,
-0.06145775318145752,
-0.0068334839306771755,
-0.01739662140607834,
0.007999647408723831,
0.08961762487888336,
0.005961066111922264,
0.10754025727510452,
-0.049905985593795776,
-0.04860205948352814,
0.05010012909770012,
-0.014414798468351364,
-0.03705200180411339,
0.03168965131044388,
0.09478821605443954,
0.11836738884449005,
0.009569711983203888,
0.06619317084550858,
-0.1423998922109604,
0.06971034407615662,
-0.002357833320274949,
-0.10916425287723541,
0.006924204993993044,
0.017383994534611702,
-0.02986602857708931,
0.13058778643608093,
0.04687919467687607,
-0.11337550729513168,
0.04215989261865616,
-0.12762030959129333,
-0.019861841574311256,
-0.03850192949175835,
0.03476361930370331,
0.13068333268165588,
0.06923571974039078,
0.013178346678614616,
0.03264879435300827,
-0.046287134289741516,
0.09116657078266144,
0.12424224615097046,
-0.13271555304527283,
-0.07246861606836319,
0.11986097693443298,
0.008026670664548874,
0.1387220174074173,
-0.08769746869802475,
0.10494617372751236,
0.07406624406576157,
-0.05260733515024185,
-0.1365770846605301,
-0.08881741017103195,
-0.20223891735076904,
-0.009014341048896313,
0.0015769184101372957,
0.025755764916539192,
0.42187854647636414,
0.06601591408252716,
0.04813048243522644,
0.06276163458824158,
-0.07336971163749695,
-0.010833683423697948,
-0.09725566953420639,
0.12981081008911133,
0.034809429198503494,
0.09582022577524185,
-0.051902249455451965,
-0.12374011427164078,
-0.11080845445394516,
-0.08940393477678299,
-0.05972230061888695,
-0.01521717756986618,
0.024964189156889915,
0.12437113374471664,
-0.18894217908382416,
0.0014572375221177936,
-0.06347968429327011,
-0.11348038911819458,
0.022804727777838707,
-0.10295918583869934,
0.1665894091129303,
0.0693678930401802,
-0.0238169077783823,
0.012971444986760616,
0.25050103664398193,
0.13133619725704193,
0.16610629856586456,
0.044331107288599014,
-0.10273228585720062,
0.1287229359149933,
0.04901270568370819,
-0.11087936908006668,
-0.028953339904546738,
-0.08289206027984619,
0.14932696521282196,
-0.06303738057613373,
0.09871520102024078,
-0.06417860090732574,
-0.11889234185218811,
0.02309919334948063,
-0.09041929990053177,
0.06031400337815285,
0.017942005768418312,
0.010398016311228275,
-0.043848197907209396,
0.0430808924138546,
0.051002971827983856,
-0.031988419592380524,
-0.002136409282684326,
-0.012715289369225502,
-0.06599804013967514,
0.040221888571977615,
0.09550627321004868,
0.02981913834810257,
0.06396650522947311,
-0.011290979571640491,
-0.02782982401549816,
-0.004136357922106981,
-0.05715888366103172,
-0.002413844922557473,
0.038967858999967575,
0.036814019083976746,
0.09658931195735931,
-0.15665890276432037,
-0.0762500986456871,
-0.0033783954568207264,
0.05290999263525009,
-0.008997089229524136,
0.09532798081636429,
-0.004387429915368557,
0.04090951010584831,
0.0011850175214931369,
-0.019492506980895996,
0.047829773277044296,
-0.10661338269710541,
0.07806449383497238,
-0.020739251747727394,
0.09156773239374161,
-0.19920383393764496,
-0.004525835160166025,
-0.04986602067947388,
0.007037294562906027,
0.06052248179912567,
-0.008082094602286816,
-0.11187456548213959,
0.08623780310153961,
-0.008130813017487526,
0.06115899980068207,
-0.09177729487419128,
0.0502496212720871,
0.020937874913215637,
0.07699990272521973,
-0.0835246592760086,
0.017873352393507957,
0.13183830678462982,
-0.1455797255039215,
-0.13980817794799805,
0.09830587357282639,
-0.022137247025966644,
0.015101557597517967,
0.04213833436369896,
0.15873973071575165,
0.17587751150131226,
-0.15706795454025269,
-0.017545223236083984,
0.05708640441298485,
-0.02593400701880455,
-0.057899899780750275,
-0.016468141227960587,
0.12299005687236786,
0.0030552023090422153,
0.033622439950704575,
-0.03652983158826828,
0.12571637332439423,
-0.041716888546943665,
-0.08251343667507172,
-0.03564097732305527,
-0.08099868893623352,
-0.06503285467624664,
0.04280487075448036,
-0.00224079517647624,
-0.04713359847664833,
0.014160679653286934,
-0.13510490953922272,
0.16119831800460815,
0.002179378177970648,
0.01846638135612011,
-0.08342655748128891,
0.12315617501735687,
0.018073882907629013,
0.004933269694447517,
0.0008151165093295276,
-0.0498083233833313,
-0.0941024199128151,
0.24196133017539978,
0.07868633419275284,
0.1112714409828186,
0.06608846783638,
-0.044526491314172745,
-0.0845038890838623,
0.02795787714421749,
0.01067365426570177,
-0.03928709402680397,
0.03648597374558449,
-0.10932664573192596,
0.041138481348752975,
-0.009837052784860134,
0.03660855069756508,
-0.005042464006692171,
-0.02772827073931694,
0.07870744913816452,
0.026034805923700333,
-0.02377321943640709,
0.07987428456544876,
0.043158430606126785,
-0.025892596691846848,
-0.07300232350826263,
0.007425438147038221,
0.06708025187253952,
-0.01119326427578926,
-0.09397920966148376,
0.012368018738925457,
0.004909964744001627,
0.04088751599192619,
0.2042549103498459,
-0.2187018245458603,
0.027451166883111,
0.02093079872429371,
0.042336709797382355,
0.04668297618627548,
-0.02019302174448967,
-0.049273427575826645,
0.030118882656097412,
-0.04368637874722481,
0.06546510010957718,
-0.011831037700176239,
0.0805012434720993,
-0.014731942676007748,
-0.1304071694612503,
-0.004538203123956919,
-0.02734546735882759,
0.17388859391212463,
-0.17962563037872314,
0.06651227176189423,
0.13975150883197784,
-0.14110605418682098,
0.14695242047309875,
-0.00044412401621229947,
-0.008962443098425865,
0.01676851138472557,
0.03361928462982178,
0.000028987842597416602,
0.11696923524141312,
-0.11355716735124588,
-0.022553276270627975,
0.027900496497750282,
-0.08776425570249557,
0.026367945596575737,
-0.12332494556903839,
-0.1121269091963768,
-0.07734160125255585,
-0.05065983161330223,
0.008537701331079006,
0.012870395556092262,
-0.05814233049750328,
0.08682891726493835,
-0.09105132520198822,
-0.08119846135377884,
-0.023162249475717545,
-0.0786362737417221,
0.019975952804088593,
0.006689172703772783,
-0.053479745984077454,
-0.13175158202648163,
-0.12385038286447525,
-0.07919464260339737,
-0.1478252410888672,
0.00403963029384613,
0.06612614542245865,
-0.11267774552106857,
-0.05088772624731064,
0.0004236796230543405,
-0.03402626886963844,
0.07995785772800446,
-0.06796850264072418,
0.009387091733515263,
0.05665038153529167,
-0.026637494564056396,
-0.1625625342130661,
0.001088478253223002,
-0.05937911942601204,
-0.07688325643539429,
0.15614266693592072,
-0.13706815242767334,
0.18392089009284973,
-0.043740347027778625,
0.046790711581707,
0.057728543877601624,
0.021908000111579895,
0.13746477663516998,
-0.11343623697757721,
0.08702147006988525,
0.19252990186214447,
0.038182541728019714,
0.0640781968832016,
0.10837358236312866,
0.0880613625049591,
-0.10481308400630951,
0.03630243241786957,
0.08434353768825531,
-0.1109103262424469,
-0.07256586104631424,
-0.05229958891868591,
-0.1099151074886322,
-0.024935107678174973,
0.061658136546611786,
0.06479883939027786,
0.0827934667468071,
0.12580950558185577,
-0.07392393052577972,
-0.0025474056601524353,
0.11048933118581772,
0.03958955407142639,
0.0841023325920105,
0.011484530754387379,
0.06785490363836288,
-0.13026919960975647,
-0.047073088586330414,
0.16509617865085602,
0.20776155591011047,
0.2533659040927887,
0.01591913402080536,
0.07968820631504059,
0.1106654480099678,
0.08566122502088547,
0.11049725860357285,
0.042495809495449066,
0.011636887677013874,
0.02090311609208584,
-0.07100331038236618,
-0.05124332383275032,
0.03902282938361168,
0.006443806923925877,
-0.05668460577726364,
-0.14660263061523438,
0.112579345703125,
0.0021699408534914255,
0.07802408188581467,
0.12809564173221588,
0.042487241327762604,
-0.10836111754179001,
0.13876037299633026,
0.09402570128440857,
0.09853607416152954,
-0.06484895944595337,
0.14393050968647003,
0.07592985779047012,
-0.003979164641350508,
0.15035492181777954,
0.01979963853955269,
0.14670142531394958,
-0.04386351630091667,
-0.07464907318353653,
-0.05587434023618698,
-0.04214901477098465,
0.010229276493191719,
0.03139035403728485,
-0.229093998670578,
0.08421715348958969,
0.05458127707242966,
0.009223241358995438,
-0.02028404362499714,
-0.05337238311767578,
0.1770862489938736,
0.13968688249588013,
0.09026803076267242,
0.0193443913012743,
-0.04545171558856964,
-0.02893994003534317,
-0.07791946083307266,
0.047099579125642776,
0.027564015239477158,
0.06065551936626434,
-0.03879716992378235,
-0.10244633257389069,
-0.02326420694589615,
0.00268092704936862,
0.017566189169883728,
-0.08031732589006424,
-0.10990902036428452,
-0.047007717192173004,
0.2704622745513916,
-0.048073071986436844,
0.06046495586633682,
0.059633709490299225,
0.015667038038372993,
-0.017216037958860397,
0.04191748797893524,
-0.04390353709459305,
-0.015841694548726082,
0.0064361910335719585,
0.013722511939704418,
0.01712546870112419,
-0.0371035560965538,
-0.05864069238305092,
-0.010498398914933205,
-0.08823496103286743,
-0.10760574042797089,
0.003234458388760686,
-0.041002217680215836,
-0.008164915256202221,
-0.02012334205210209,
0.017551742494106293,
-0.08846620470285416,
-0.02126779779791832,
0.02203742414712906,
0.04007299989461899,
-0.0599510483443737,
-0.13435815274715424,
0.006987391971051693,
-0.014276482164859772,
-0.07424803078174591,
0.024669963866472244,
-0.12293563038110733,
-0.06697285920381546,
-0.06201926991343498,
-0.03660818189382553,
0.11193042248487473,
0.22884199023246765,
-0.03691650554537773,
-0.011099979281425476,
0.11540786176919937,
-0.08984501659870148,
-0.31959354877471924,
-0.15261200070381165,
-0.16035863757133484,
-0.09919676184654236,
0.024694330990314484,
-0.09483373165130615,
0.030329253524541855,
0.09791705012321472,
-0.039176929742097855,
0.22954382002353668,
-0.1885816752910614,
-0.09886320680379868,
0.057825151830911636,
0.10509101301431656,
0.295073926448822,
-0.24438947439193726,
0.015688559040427208,
-0.09514027833938599,
-0.01863790675997734,
0.03158051148056984,
-0.08156250417232513,
0.11927708983421326,
0.04415062069892883,
0.05047261714935303,
-0.007339892443269491,
-0.005602825898677111,
0.13098257780075073,
-0.08530011773109436,
0.13132159411907196,
-0.10705619305372238,
-0.0783129557967186,
0.21912530064582825,
-0.020011920481920242,
0.001358875771984458,
-0.20024162530899048,
-0.03315184265375137,
-0.011580201797187328,
0.033065564930438995,
-0.006128721870481968,
0.05239135026931763,
-0.004424794111400843,
-0.02967400848865509,
-0.14297528564929962,
-0.01648416742682457,
-0.03679249435663223,
0.06754843890666962,
0.2398601919412613,
-0.062269240617752075,
-0.05471041798591614,
0.0443902313709259,
-0.01884733885526657,
0.11554649472236633,
0.020689349621534348,
-0.04686175659298897,
-0.05332694947719574,
0.09994570910930634,
-0.1970922201871872,
0.063298799097538,
-0.0007646753219887614,
-0.005302996374666691,
0.024245373904705048,
0.014361375942826271,
0.02788952738046646,
0.12395945936441422,
0.16746513545513153,
-0.028170917183160782,
-0.06057864427566528,
-0.03299759700894356,
0.007086495868861675,
0.1355787217617035,
-0.013565478846430779,
0.12162567675113678,
0.0026302358601242304,
0.038937781006097794,
0.010940068401396275,
0.06004186347126961,
-0.08178666979074478,
-0.08808999508619308,
0.0857321247458458,
-0.050525739789009094,
-0.08792934566736221,
0.08056974411010742,
0.05224747955799103,
0.06118367239832878,
0.01555580087006092,
0.021797964349389076,
0.004262207541614771,
-0.12915000319480896,
-0.004560430534183979,
0.1983027309179306,
-0.08627885580062866,
-0.06620065867900848,
-0.07899788022041321,
0.005271814297884703,
-0.11483897268772125,
0.05675453320145607,
0.040884830057621,
-0.02170618064701557,
0.12375061959028244,
-0.027709605172276497,
-0.04943624883890152,
0.009418520145118237,
-0.07432913035154343,
0.037358153611421585,
-0.14392079412937164,
-0.193904846906662,
0.05115702003240585,
-0.02557704597711563,
-0.07233234494924545,
-0.08686404675245285,
-0.0829099491238594,
0.05820130184292793,
-0.14288409054279327,
0.1367107331752777,
-0.05836814269423485,
0.04989580437541008,
-0.039200667291879654,
-0.04741775244474411,
-0.10287265479564667,
-0.010419717989861965,
-0.04480063170194626,
-0.02733493223786354,
0.05740756914019585,
0.0008014675695449114,
-0.12176322937011719,
-0.11192494630813599,
0.05227495729923248,
-0.006997196935117245,
-0.0025248366873711348,
0.018799327313899994,
-0.07165538519620895,
0.011041710153222084,
-0.21851611137390137,
-0.042402904480695724,
0.08613522350788116,
0.040143340826034546,
-0.09862294048070908,
0.1395276039838791,
0.04555295407772064,
-0.024086082354187965,
0.03714672848582268,
0.006663348991423845,
0.14427267014980316,
-0.07031701505184174,
0.031240757554769516,
-0.13709498941898346,
-0.14074444770812988,
-0.030508683994412422,
0.03409608080983162,
0.2396184206008911,
0.08356516808271408,
0.11453827470541,
-0.05502690002322197,
0.023517245426774025,
-0.018902216106653214,
0.0785134956240654,
0.019728826358914375,
-0.0980018824338913,
-0.07533524185419083,
-0.16522440314292908,
-0.06266290694475174,
-0.0675647035241127,
0.17895959317684174,
0.04198586940765381,
-0.12073011696338654,
0.0031159818172454834,
0.0985165685415268,
-0.15714097023010254,
-0.00656132260337472,
0.19069577753543854,
-0.04210099205374718,
0.013936547562479973,
-0.15302519500255585,
0.039160437881946564,
0.08042767643928528,
-0.020854976028203964,
-0.013286413624882698,
0.10824128240346909,
-0.026333404704928398,
-0.010598661378026009,
0.022460894659161568,
-0.03699161857366562,
0.07603020966053009,
-0.051641128957271576,
0.06286115944385529,
0.007164948154240847,
-0.0573832169175148,
-0.10849395394325256,
0.1776425987482071,
-0.014335974119603634,
0.02029932476580143,
-0.058072734624147415,
-0.0023532300256192684,
-0.0951792448759079,
-0.11880478262901306,
-0.07794037461280823,
-0.1521134078502655,
0.06877679377794266,
-0.07246100902557373,
0.015048118308186531,
-0.012350399047136307,
0.02222038060426712,
-0.08599846065044403,
0.011182229965925217,
-0.17541271448135376,
-0.03295267000794411,
0.017741607502102852,
-0.010517670772969723,
-0.023495592176914215,
-0.03155849128961563,
-0.056692276149988174,
0.04037224501371384,
-0.06095006689429283,
-0.06283359229564667,
0.060239262878894806,
0.09093807637691498,
0.05764807388186455,
-0.14506325125694275,
-0.1099768802523613,
-0.06878990679979324,
0.02422149106860161,
0.0707649290561676,
0.17823907732963562,
0.025239599868655205,
-0.007869920693337917,
0.0366964228451252,
0.1629951298236847,
0.011288512498140335,
-0.047349777072668076,
-0.06740102171897888,
-0.1311212182044983,
-0.12847523391246796,
-0.00907034520059824,
-0.05531249940395355,
-0.033554062247276306,
0.012438775971531868,
0.2438674420118332,
0.19516633450984955,
-0.1331038624048233,
0.03445027396082878,
-0.0637759193778038,
0.03503032028675079,
-0.020044278353452682,
0.15414902567863464,
0.04633779078722,
0.12523101270198822,
-0.028399989008903503,
-0.04064793139696121,
-0.06190883368253708,
0.016257012262940407,
-0.10258867591619492,
0.020121585577726364,
-0.0017541892593726516,
-0.0667179748415947,
-0.07066783308982849,
0.09089042246341705,
-0.12925685942173004,
0.042638760060071945,
0.18011963367462158,
-0.14657679200172424,
-0.009904816746711731,
-0.03257455676794052,
0.05071938410401344,
0.11328626424074173,
0.017976753413677216,
-0.07774310559034348,
-0.0037855729460716248,
0.007672405336052179,
0.03154725581407547,
-0.17823360860347748,
-0.10118065774440765,
-0.013713724911212921,
-0.1150989979505539,
0.12640051543712616,
-0.006142603699117899,
0.0035269579384475946,
0.03176566958427429,
-0.07624257355928421,
-0.0069709960371255875,
0.16727912425994873,
0.01768195629119873,
-0.0536552369594574,
-0.030534887686371803,
-0.027902299538254738,
-0.10428634285926819,
0.08348623663187027,
0.07915928214788437,
0.0387050025165081,
-0.011799951083958149,
0.15638305246829987,
-0.022158749401569366,
-0.035794127732515335,
0.12567901611328125,
-0.17630867660045624,
0.09640127420425415,
0.013099360279738903,
-0.02371886558830738,
-0.08107023686170578,
-0.04923111945390701,
0.041165199130773544,
0.08354230970144272,
-0.16258007287979126,
-0.043006863445043564,
0.04147496819496155,
-0.11320789903402328,
0.07138777524232864,
0.04154369607567787,
-0.08664262294769287,
0.01941647194325924,
-0.12450280040502548,
0.0012744928244501352,
-0.10503152757883072,
0.03293466940522194,
0.2079509049654007,
-0.025966927409172058,
0.013933198526501656,
-0.1290113776922226,
0.06822633743286133,
-0.038864947855472565,
-0.04012955352663994,
-0.07743152230978012
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# t5-base-squad-model1
This model is a fine-tuned version of [t5-base](https://huggingface.co/t5-base) on the squad dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 16
- seed: 59
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["varun-v-rao/squad"], "base_model": "t5-base", "model-index": [{"name": "t5-base-squad-model1", "results": []}]} | question-answering | varun-v-rao/t5-base-squad-model1 | [
"transformers",
"tensorboard",
"safetensors",
"t5",
"question-answering",
"generated_from_trainer",
"dataset:varun-v-rao/squad",
"base_model:t5-base",
"license:apache-2.0",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T07:07:29+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #t5 #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-t5-base #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us
|
# t5-base-squad-model1
This model is a fine-tuned version of t5-base on the squad dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 16
- seed: 59
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| [
"# t5-base-squad-model1\n\nThis model is a fine-tuned version of t5-base on the squad dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 59\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #t5 #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-t5-base #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us \n",
"# t5-base-squad-model1\n\nThis model is a fine-tuned version of t5-base on the squad dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 59\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
80,
30,
6,
12,
8,
3,
90,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #t5 #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-t5-base #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us \n# t5-base-squad-model1\n\nThis model is a fine-tuned version of t5-base on the squad dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 59\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3### Training results### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
-0.08270397037267685,
0.10978949815034866,
-0.0016183212865144014,
0.1019439846277237,
0.13991354405879974,
0.021683230996131897,
0.12096034735441208,
0.1268412172794342,
-0.10741498321294785,
0.05212635174393654,
0.05748045817017555,
0.05884094536304474,
0.04485807195305824,
0.11291296780109406,
-0.025496840476989746,
-0.21726375818252563,
0.005962212570011616,
-0.01756042055785656,
-0.06675230711698532,
0.11591785401105881,
0.10403614491224289,
-0.09329898655414581,
0.08531630784273148,
-0.016263695433735847,
-0.13365304470062256,
0.057423606514930725,
-0.00858111772686243,
-0.04539952799677849,
0.11492392420768738,
0.029095904901623726,
0.0835646390914917,
0.002929957117885351,
0.14278170466423035,
-0.23471347987651825,
0.007328800391405821,
0.0978313460946083,
0.00753290019929409,
0.06830695271492004,
0.0369081050157547,
-0.001540431403554976,
0.101357102394104,
-0.1586064100265503,
0.09976335614919662,
0.03922553360462189,
-0.07651356607675552,
-0.13394975662231445,
-0.0783032700419426,
0.07967139780521393,
0.10521470755338669,
0.1063932329416275,
0.0034010824747383595,
0.16092228889465332,
-0.0743069276213646,
0.09304805099964142,
0.1969931274652481,
-0.28116998076438904,
-0.05905573070049286,
0.07634026557207108,
0.060443244874477386,
0.10192082822322845,
-0.10368462651968002,
-0.014866068959236145,
0.049450092017650604,
0.02311905100941658,
0.0890093445777893,
-0.021972134709358215,
-0.11549200862646103,
0.02463529258966446,
-0.14080767333507538,
-0.012588283978402615,
0.18394917249679565,
0.06514568626880646,
-0.03926021605730057,
-0.06673436611890793,
-0.07352251559495926,
-0.07061101496219635,
-0.012144179083406925,
-0.05598364397883415,
0.04677914083003998,
-0.04401232302188873,
-0.04906321316957474,
-0.05423237010836601,
-0.08681133389472961,
-0.08538970351219177,
-0.004320845473557711,
0.0698843002319336,
0.054872363805770874,
0.01323106698691845,
-0.046079959720373154,
0.1013493612408638,
-0.017883749678730965,
-0.1069478988647461,
-0.017087284475564957,
0.010696986690163612,
-0.08816476911306381,
-0.055159855633974075,
-0.013329064473509789,
-0.05002473667263985,
0.025778399780392647,
0.13774561882019043,
-0.06268839538097382,
0.04767030477523804,
0.004849668126553297,
-0.0012413270305842161,
-0.030371075496077538,
0.11504853516817093,
-0.07381954044103622,
-0.05169719457626343,
0.029347607865929604,
0.09343547374010086,
0.03446340188384056,
-0.00044669213821180165,
-0.08987743407487869,
-0.016939878463745117,
0.10580606758594513,
0.08255426585674286,
-0.02549278549849987,
0.05019707232713699,
-0.011162638664245605,
-0.022872216999530792,
0.02117479406297207,
-0.1391095072031021,
0.02424323558807373,
-0.019326478242874146,
-0.07212650775909424,
-0.07289918512105942,
0.04114800691604614,
0.002878230530768633,
-0.032932594418525696,
0.041147831827402115,
-0.08554042130708694,
-0.014834984205663204,
-0.08080341666936874,
-0.08393432199954987,
0.021469198167324066,
-0.04299323260784149,
-0.006106294225901365,
-0.083989717066288,
-0.22945544123649597,
-0.029902633279561996,
0.025594884529709816,
-0.04313715919852257,
-0.042475711554288864,
-0.04822495952248573,
-0.07429821789264679,
-0.011303371749818325,
-0.0007077789050526917,
0.11029709875583649,
-0.05685829743742943,
0.07930993288755417,
0.00266831461340189,
0.02088163234293461,
0.013802707195281982,
0.02634204737842083,
-0.1056603193283081,
0.020961152389645576,
-0.13438676297664642,
0.041785191744565964,
-0.0792146846652031,
0.04626385495066643,
-0.1212674230337143,
-0.09252946823835373,
0.00832956563681364,
-0.016463322564959526,
0.04474184289574623,
0.12422803789377213,
-0.1865200400352478,
-0.003972065169364214,
0.15236680209636688,
-0.09637623280286789,
-0.11777240037918091,
0.10911865532398224,
-0.04237880930304527,
0.03719297796487808,
0.07232605665922165,
0.16293679177761078,
0.10399848222732544,
-0.14069122076034546,
0.002156424568966031,
-0.001395785715430975,
0.03584422543644905,
0.004094459116458893,
0.06055222451686859,
-0.0025680342223495245,
0.0012371775228530169,
0.003403719048947096,
-0.07898734509944916,
0.00505789602175355,
-0.07801681011915207,
-0.08727431297302246,
-0.06522504985332489,
-0.10310408473014832,
0.025568991899490356,
0.04389858618378639,
0.01632246933877468,
-0.07829157263040543,
-0.11738795787096024,
0.10757963359355927,
0.12947741150856018,
-0.053300224244594574,
0.009324357844889164,
-0.08272211253643036,
0.04323078319430351,
-0.032936941832304,
-0.023578764870762825,
-0.17152416706085205,
-0.1397923231124878,
0.023340972140431404,
-0.06294207274913788,
0.05033320188522339,
0.02160671167075634,
0.06478575617074966,
0.0740148052573204,
-0.06660906225442886,
-0.03144877403974533,
-0.0688285157084465,
0.01591859944164753,
-0.08327122032642365,
-0.21423602104187012,
-0.036104682832956314,
-0.018590975552797318,
0.1304529458284378,
-0.25092509388923645,
0.033052414655685425,
-0.03721025213599205,
0.11829615384340286,
0.026798389852046967,
-0.03569832444190979,
0.00907360389828682,
0.030912162736058235,
-0.017808984965085983,
-0.0938597023487091,
0.029118690639734268,
-0.00863025151193142,
-0.07551880925893784,
-0.0629766434431076,
-0.1549314260482788,
0.08429936319589615,
0.07389993965625763,
0.051863208413124084,
-0.10095907747745514,
0.018215594813227654,
-0.059335123747587204,
-0.048093296587467194,
-0.08218024671077728,
-0.016766928136348724,
0.13248573243618011,
-0.005490233190357685,
0.11033552139997482,
-0.07114818692207336,
-0.06916575878858566,
-0.0027311164885759354,
-0.003004896454513073,
0.004217774607241154,
0.09150822460651398,
0.06822478026151657,
-0.13727182149887085,
0.09823555499315262,
0.08210371434688568,
-0.05388713255524635,
0.13980403542518616,
-0.06097397953271866,
-0.08045820891857147,
-0.027460526674985886,
0.034822482615709305,
-0.006792261730879545,
0.14842158555984497,
-0.10395403206348419,
0.013641477562487125,
0.01749161072075367,
0.0012959038140252233,
0.020588351413607597,
-0.17354494333267212,
-0.028226027265191078,
0.02296576090157032,
-0.06085857376456261,
-0.01906188391149044,
-0.02670346572995186,
0.02309659868478775,
0.101713627576828,
0.001998668070882559,
-0.029387380927801132,
0.019174624234437943,
-0.008636941201984882,
-0.10627946257591248,
0.1939750462770462,
-0.10576634109020233,
-0.1424643099308014,
-0.1130862832069397,
0.06864507496356964,
-0.06236479803919792,
-0.031230736523866653,
0.03108714334666729,
-0.08465274423360825,
-0.055374883115291595,
-0.12408352643251419,
-0.005515581462532282,
-0.01518088299781084,
-0.017500758171081543,
-0.00556944590061903,
0.03134453296661377,
0.09189599752426147,
-0.1453140527009964,
0.025195524096488953,
-0.009877954609692097,
-0.1158125102519989,
-0.019480668008327484,
0.047578804194927216,
0.12974537909030914,
0.1370648592710495,
-0.025706231594085693,
0.01249675266444683,
-0.037168942391872406,
0.21001587808132172,
-0.05953454598784447,
0.032676197588443756,
0.1271672248840332,
0.01952684298157692,
0.04473230615258217,
0.14821286499500275,
0.026148445904254913,
-0.08739562332630157,
0.046452976763248444,
0.08450587093830109,
-0.013936678878962994,
-0.2583655118942261,
-0.03170673921704292,
-0.024392059072852135,
-0.04757969081401825,
0.06300538033246994,
0.06924978643655777,
0.04742088541388512,
0.03811457008123398,
-0.012904776260256767,
0.02678171545267105,
0.002809701720252633,
0.07906465232372284,
0.08538364619016647,
0.03197823837399483,
0.08866287767887115,
-0.04914354905486107,
-0.03349849581718445,
0.06354352086782455,
0.018416384235024452,
0.2535839378833771,
-0.02262287214398384,
0.102730393409729,
0.049797043204307556,
0.1587177813053131,
-0.03031276725232601,
0.03457021340727806,
0.008566378615796566,
-0.0022766899783164263,
0.01694803312420845,
-0.067125603556633,
0.003601079573854804,
0.04097104072570801,
-0.057028427720069885,
0.06041640788316727,
-0.06764344871044159,
0.053489118814468384,
0.03508198261260986,
0.24357210099697113,
0.03136642649769783,
-0.28009381890296936,
-0.07964445650577545,
0.01453288272023201,
-0.034103091806173325,
-0.05041831359267235,
0.031949322670698166,
0.1376693695783615,
-0.09594277292490005,
0.047249507158994675,
-0.051506735384464264,
0.08064903318881989,
-0.015286548994481564,
-0.008020536042749882,
0.056508779525756836,
0.12061756104230881,
-0.022241730242967606,
0.10044458508491516,
-0.21770982444286346,
0.2247689813375473,
0.023489544168114662,
0.11542050540447235,
-0.048816513270139694,
0.020267652347683907,
0.006018833257257938,
0.11216709017753601,
0.12973012030124664,
-0.002896790625527501,
-0.04485699161887169,
-0.15841053426265717,
-0.06345140933990479,
0.05194912850856781,
0.09224347770214081,
-0.01975312829017639,
0.08666124194860458,
-0.057660143822431564,
-0.0054636187851428986,
0.06739567965269089,
-0.06212109327316284,
-0.19933490455150604,
-0.10687638074159622,
-0.004063331056386232,
0.0019411941757425666,
-0.039716243743896484,
-0.09650812298059464,
-0.08936414122581482,
-0.027682559564709663,
0.16268885135650635,
-0.0021825910080224276,
-0.05014677718281746,
-0.12692022323608398,
0.06658060848712921,
0.10824435204267502,
-0.06481681764125824,
0.01782870478928089,
0.021229304373264313,
0.10331633687019348,
0.05177329108119011,
-0.0895400270819664,
0.07137305289506912,
-0.0697069987654686,
-0.1605202555656433,
-0.0559382438659668,
0.1243269219994545,
0.055682968348264694,
0.043971773236989975,
0.003980136476457119,
0.004198232665657997,
0.01890973001718521,
-0.08820705115795135,
-0.0027502619195729494,
0.07145141065120697,
0.06586122512817383,
0.048920612782239914,
-0.09587733447551727,
-0.01822180673480034,
-0.0476885586977005,
-0.03397493064403534,
0.11739138513803482,
0.1916479915380478,
-0.08096904307603836,
0.06313718855381012,
0.08967926353216171,
-0.08320140838623047,
-0.17245343327522278,
0.06314561516046524,
0.049618907272815704,
0.0004228983598295599,
0.06921525299549103,
-0.15241101384162903,
0.12359195202589035,
0.10591449588537216,
-0.01864674501121044,
0.055850591510534286,
-0.3450494706630707,
-0.14523805677890778,
0.08917193859815598,
0.14581817388534546,
-0.004880319349467754,
-0.16517037153244019,
-0.030205026268959045,
-0.015641946345567703,
-0.13543398678302765,
0.10720646381378174,
-0.13716618716716766,
0.08308687806129456,
-0.0002543374721426517,
0.08844469487667084,
0.01943136751651764,
-0.036942049860954285,
0.11562381684780121,
0.025812705978751183,
0.09975375980138779,
-0.06794895976781845,
0.003959489054977894,
0.12091433256864548,
-0.06806983798742294,
0.07614173740148544,
-0.056802451610565186,
0.0896318331360817,
-0.1237623319029808,
-0.018200716003775597,
-0.06970779597759247,
0.06373855471611023,
-0.06008854880928993,
-0.07509596645832062,
-0.05749031901359558,
0.06691913306713104,
0.057131920009851456,
-0.040993016213178635,
0.0765731930732727,
0.01148291677236557,
0.12451516836881638,
0.09789495915174484,
0.11078441888093948,
0.0016616806387901306,
-0.07820961624383926,
0.0025479462929069996,
-0.02430100366473198,
0.0481165386736393,
-0.13329178094863892,
0.02818789891898632,
0.1133621335029602,
0.052714649587869644,
0.1353999376296997,
0.029170401394367218,
-0.05459792912006378,
-0.015723025426268578,
0.036785852164030075,
-0.12345438450574875,
-0.2071353942155838,
-0.010678372345864773,
-0.054111920297145844,
-0.14601737260818481,
0.04590873047709465,
0.09185963869094849,
-0.07557093352079391,
-0.008083623833954334,
-0.015344521030783653,
0.04169629141688347,
-0.02937237173318863,
0.17334598302841187,
0.07096750289201736,
0.06630886346101761,
-0.07071883231401443,
0.11995948851108551,
0.07953426986932755,
-0.0551060326397419,
0.04048733040690422,
0.08092840760946274,
-0.08946768939495087,
-0.03319374844431877,
0.08590128272771835,
0.1776527762413025,
-0.057137202471494675,
-0.05464024841785431,
-0.10038118064403534,
-0.08703447878360748,
0.031235933303833008,
0.1661563664674759,
0.0310638677328825,
-0.007457967381924391,
-0.008215305395424366,
0.02771291695535183,
-0.1353810578584671,
0.12659603357315063,
0.04022783413529396,
0.07639174908399582,
-0.16189706325531006,
0.09754686057567596,
0.014798431657254696,
0.03438350930809975,
-0.018976949155330658,
0.04399050027132034,
-0.1000407263636589,
-0.014596221968531609,
-0.15232466161251068,
-0.006541073322296143,
-0.02557818964123726,
0.019934769719839096,
-0.005989396013319492,
-0.062426213175058365,
-0.05194705352187157,
0.060481250286102295,
-0.05463829264044762,
-0.05079352855682373,
0.031713131815195084,
0.06115954369306564,
-0.16772982478141785,
-0.012188362888991833,
0.01529867947101593,
-0.07893054187297821,
0.07231387495994568,
0.02978474833071232,
0.014359768480062485,
0.04138531908392906,
-0.150034099817276,
0.020202819257974625,
0.016330083832144737,
0.03362901881337166,
0.058582425117492676,
-0.09995348006486893,
-0.00976712815463543,
-0.0214247889816761,
0.02954789251089096,
0.02123783342540264,
0.04184107109904289,
-0.11192131787538528,
0.0006899954751133919,
-0.06764167547225952,
-0.04533166065812111,
-0.058653391897678375,
0.03575195372104645,
0.098702572286129,
0.00845879316329956,
0.1580723077058792,
-0.09061697125434875,
0.033529579639434814,
-0.22269785404205322,
-0.032377690076828,
0.010005883872509003,
-0.023649748414754868,
-0.0786973237991333,
-0.024384407326579094,
0.06813996285200119,
-0.06853276491165161,
0.1101449504494667,
-0.003754019271582365,
0.10438745468854904,
0.04617021977901459,
-0.02890244871377945,
0.020103955641388893,
0.00639326311647892,
0.18636935949325562,
0.014788951724767685,
-0.020988548174500465,
0.07058356702327728,
-0.029966523870825768,
0.05592964589595795,
-0.003671477548778057,
0.1890474110841751,
0.15142738819122314,
-0.0549536794424057,
0.04789261892437935,
0.08432288467884064,
-0.10729055851697922,
-0.11702923476696014,
0.0904126763343811,
-0.018745344132184982,
0.10200098901987076,
-0.04953193664550781,
0.1784585565328598,
0.14992563426494598,
-0.15717890858650208,
0.03689837455749512,
-0.06138933077454567,
-0.11090080440044403,
-0.10725561529397964,
-0.06579603999853134,
-0.08483016490936279,
-0.13291648030281067,
0.01907222718000412,
-0.12970688939094543,
0.03199153020977974,
0.07857752591371536,
0.015199186280369759,
0.0026732233818620443,
0.17628635466098785,
0.007375357206910849,
0.028030358254909515,
0.027489330619573593,
0.019264770671725273,
-0.012708812952041626,
-0.04648001864552498,
-0.0402589812874794,
0.06225660443305969,
-0.008038251660764217,
0.04516153037548065,
-0.04485367238521576,
0.006186122074723244,
0.04836837947368622,
-0.02197950705885887,
-0.058847032487392426,
0.025347445160150528,
0.02066740393638611,
0.02805333398282528,
0.06159408763051033,
0.06817158311605453,
-0.006336200051009655,
-0.022544510662555695,
0.29188334941864014,
-0.06815799325704575,
-0.11347043514251709,
-0.12038113176822662,
0.1921592354774475,
0.019503196701407433,
-0.013594439253211021,
0.06811479479074478,
-0.13781039416790009,
0.013889409601688385,
0.17594769597053528,
0.15267042815685272,
-0.03905807062983513,
-0.012312867678701878,
-0.02849135734140873,
-0.01686166785657406,
-0.06223335117101669,
0.10687105357646942,
0.10565702617168427,
0.04639393836259842,
-0.04945675656199455,
-0.017798535525798798,
-0.014963868074119091,
-0.020064610987901688,
-0.07761382311582565,
0.0802193209528923,
0.032254185527563095,
0.006365085486322641,
-0.01750349998474121,
0.09390218555927277,
-0.010563245043158531,
-0.19355717301368713,
0.02556999772787094,
-0.1395868957042694,
-0.1561078578233719,
-0.036401793360710144,
0.11665621399879456,
-0.03183870017528534,
0.029079057276248932,
-0.02022346667945385,
0.0022163796238601208,
0.11161722987890244,
-0.004827193915843964,
-0.08616483956575394,
-0.11701919138431549,
0.07688329368829727,
-0.07678353786468506,
0.24880123138427734,
-0.008218103088438511,
0.05497114360332489,
0.11568435281515121,
-0.007289593573659658,
-0.14320261776447296,
0.07817695289850235,
0.06754224747419357,
-0.0603354349732399,
0.032734233886003494,
0.14082548022270203,
-0.03573085367679596,
0.11265586316585541,
0.035556890070438385,
-0.0941835567355156,
-0.010602409951388836,
-0.032419223338365555,
-0.025511151179671288,
-0.1005004271864891,
0.007912343367934227,
-0.07236669212579727,
0.15231598913669586,
0.17258307337760925,
-0.04353514313697815,
0.03431278094649315,
-0.09341748803853989,
0.02940981276333332,
0.06419367343187332,
0.07911121100187302,
0.009002942591905594,
-0.1962486207485199,
0.03598034381866455,
-0.0023051598109304905,
0.02752142772078514,
-0.2533350884914398,
-0.08534646779298782,
0.043560758233070374,
-0.04860760644078255,
-0.07114001363515854,
0.11649027466773987,
0.1269657462835312,
0.04598820209503174,
-0.037861306220293045,
-0.12946002185344696,
-0.04624840244650841,
0.16058096289634705,
-0.15557339787483215,
-0.06204187870025635
] |
null | null | setfit |
# SetFit with sentence-transformers/paraphrase-mpnet-base-v2
This is a [SetFit](https://github.com/huggingface/setfit) model that can be used for Text Classification. This SetFit model uses [sentence-transformers/paraphrase-mpnet-base-v2](https://huggingface.co/sentence-transformers/paraphrase-mpnet-base-v2) as the Sentence Transformer embedding model. A [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance is used for classification.
The model has been trained using an efficient few-shot learning technique that involves:
1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning.
2. Training a classification head with features from the fine-tuned Sentence Transformer.
## Model Details
### Model Description
- **Model Type:** SetFit
- **Sentence Transformer body:** [sentence-transformers/paraphrase-mpnet-base-v2](https://huggingface.co/sentence-transformers/paraphrase-mpnet-base-v2)
- **Classification head:** a [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance
- **Maximum Sequence Length:** 512 tokens
- **Number of Classes:** 30 classes
<!-- - **Training Dataset:** [Unknown](https://huggingface.co/datasets/unknown) -->
<!-- - **Language:** Unknown -->
<!-- - **License:** Unknown -->
### Model Sources
- **Repository:** [SetFit on GitHub](https://github.com/huggingface/setfit)
- **Paper:** [Efficient Few-Shot Learning Without Prompts](https://arxiv.org/abs/2209.11055)
- **Blogpost:** [SetFit: Efficient Few-Shot Learning Without Prompts](https://huggingface.co/blog/setfit)
### Model Labels
| Label | Examples |
|:----------------------------|:-----------------------------------------------------------------------------------------------------------------------------------------------------|
| ls | <ul><li>'List all files and directories'</li><li>'Show files in the current directory'</li><li>'Display contents of the current directory'</li></ul> |
| cd | <ul><li>'Change to the specified directory'</li><li>'Move to the home directory'</li><li>'Navigate to the specified directory path'</li></ul> |
| mkdir docs | <ul><li>"Create a new directory named 'docs'"</li></ul> |
| mkdir projects | <ul><li>"Make a directory named 'projects'"</li></ul> |
| mkdir data | <ul><li>"Create a folder called 'data'"</li></ul> |
| mkdir images | <ul><li>"Make a directory named 'images'"</li></ul> |
| mkdir scripts | <ul><li>"Create a new folder named 'scripts'"</li></ul> |
| rm example.txt | <ul><li>"Remove the file named 'example.txt'"</li></ul> |
| rm temp.txt | <ul><li>"Delete the file called 'temp.txt'"</li></ul> |
| rm file1 | <ul><li>"Remove the file named 'file1'"</li></ul> |
| rm file2 | <ul><li>"Delete the file named 'file2'"</li></ul> |
| rm backup.txt | <ul><li>"Remove the file named 'backup.txt'"</li></ul> |
| cp file1 /destination | <ul><li>'Copy file1 to directory /destination'</li></ul> |
| cp file2 /backup | <ul><li>'Duplicate file2 to directory /backup'</li></ul> |
| cp file3 /archive | <ul><li>'Copy file3 to folder /archive'</li></ul> |
| cp file4 /temp | <ul><li>'Duplicate file4 to folder /temp'</li></ul> |
| cp file5 /images | <ul><li>'Copy file5 to directory /images'</li></ul> |
| mv file2 /new_location | <ul><li>'Move file2 to directory /new_location'</li></ul> |
| mv file3 /backup | <ul><li>'Transfer file3 to directory /backup'</li></ul> |
| mv file4 /archive | <ul><li>'Move file4 to folder /archive'</li></ul> |
| mv file5 /temp | <ul><li>'Transfer file5 to folder /temp'</li></ul> |
| mv file6 /images | <ul><li>'Move file6 to directory /images'</li></ul> |
| cat README.md | <ul><li>"Display the contents of file 'README.md'"</li></ul> |
| cat notes.txt | <ul><li>"Show the content of file 'notes.txt'"</li></ul> |
| cat data.csv | <ul><li>"Print the contents of file 'data.csv'"</li></ul> |
| cat script.sh | <ul><li>"Display the content of file 'script.sh'"</li></ul> |
| cat config.ini | <ul><li>"Show the contents of file 'config.ini'"</li></ul> |
| grep 'pattern' data.txt | <ul><li>"Search for 'pattern' in file 'data.txt'"</li></ul> |
| grep 'word' text.txt | <ul><li>"Find occurrences of 'word' in file 'text.txt'"</li></ul> |
| grep 'keyword' document.txt | <ul><li>"Search for 'keyword' in file 'document.txt'"</li></ul> |
## Evaluation
### Metrics
| Label | Accuracy |
|:--------|:---------|
| **all** | 0.0 |
## Uses
### Direct Use for Inference
First install the SetFit library:
```bash
pip install setfit
```
Then you can load this model and run inference.
```python
from setfit import SetFitModel
# Download from the 🤗 Hub
model = SetFitModel.from_pretrained("souvenger/NLP2Linux")
# Run inference
preds = model("Install package 'vim' as superuser")
```
<!--
### Downstream Use
*List how someone could finetune this model on their own dataset.*
-->
<!--
### Out-of-Scope Use
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
-->
<!--
## Bias, Risks and Limitations
*What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
-->
<!--
### Recommendations
*What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
-->
## Training Details
### Training Set Metrics
| Training set | Min | Median | Max |
|:-------------|:----|:-------|:----|
| Word count | 5 | 5.6667 | 9 |
| Label | Training Sample Count |
|:----------------------------|:----------------------|
| cat README.md | 1 |
| cat config.ini | 1 |
| cat data.csv | 1 |
| cat notes.txt | 1 |
| cat script.sh | 1 |
| cd | 10 |
| cp file1 /destination | 1 |
| cp file2 /backup | 1 |
| cp file3 /archive | 1 |
| cp file4 /temp | 1 |
| cp file5 /images | 1 |
| grep 'keyword' document.txt | 1 |
| grep 'pattern' data.txt | 1 |
| grep 'word' text.txt | 1 |
| ls | 10 |
| mkdir data | 1 |
| mkdir docs | 1 |
| mkdir images | 1 |
| mkdir projects | 1 |
| mkdir scripts | 1 |
| mv file2 /new_location | 1 |
| mv file3 /backup | 1 |
| mv file4 /archive | 1 |
| mv file5 /temp | 1 |
| mv file6 /images | 1 |
| rm backup.txt | 1 |
| rm example.txt | 1 |
| rm file1 | 1 |
| rm file2 | 1 |
| rm temp.txt | 1 |
### Training Hyperparameters
- batch_size: (8, 8)
- num_epochs: (1, 1)
- max_steps: -1
- sampling_strategy: oversampling
- num_iterations: 20
- body_learning_rate: (2e-05, 2e-05)
- head_learning_rate: 2e-05
- loss: CosineSimilarityLoss
- distance_metric: cosine_distance
- margin: 0.25
- end_to_end: False
- use_amp: False
- warmup_proportion: 0.1
- seed: 42
- eval_max_steps: -1
- load_best_model_at_end: False
### Training Results
| Epoch | Step | Training Loss | Validation Loss |
|:------:|:----:|:-------------:|:---------------:|
| 0.0042 | 1 | 0.1215 | - |
| 0.2083 | 50 | 0.0232 | - |
| 0.4167 | 100 | 0.01 | - |
| 0.625 | 150 | 0.0044 | - |
| 0.8333 | 200 | 0.0025 | - |
### Framework Versions
- Python: 3.10.13
- SetFit: 1.0.3
- Sentence Transformers: 2.3.1
- Transformers: 4.37.0
- PyTorch: 2.1.2
- Datasets: 2.1.0
- Tokenizers: 0.15.1
## Citation
### BibTeX
```bibtex
@article{https://doi.org/10.48550/arxiv.2209.11055,
doi = {10.48550/ARXIV.2209.11055},
url = {https://arxiv.org/abs/2209.11055},
author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren},
keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
title = {Efficient Few-Shot Learning Without Prompts},
publisher = {arXiv},
year = {2022},
copyright = {Creative Commons Attribution 4.0 International}
}
```
<!--
## Glossary
*Clearly define terms in order to be accessible across audiences.*
-->
<!--
## Model Card Authors
*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
-->
<!--
## Model Card Contact
*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
--> | {"library_name": "setfit", "tags": ["setfit", "sentence-transformers", "text-classification", "generated_from_setfit_trainer"], "metrics": ["accuracy"], "widget": [{"text": "Upgrade all installed packages with superuser privileges"}, {"text": "Install package 'vim' as superuser"}, {"text": "Remove package 'firefox' with superuser privileges"}, {"text": "Change permissions of directory 'docs' to writable"}, {"text": "Update package lists using superuser privileges"}], "pipeline_tag": "text-classification", "inference": true, "base_model": "sentence-transformers/paraphrase-mpnet-base-v2", "model-index": [{"name": "SetFit with sentence-transformers/paraphrase-mpnet-base-v2", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "Unknown", "type": "unknown", "split": "test"}, "metrics": [{"type": "accuracy", "value": 0.0, "name": "Accuracy"}]}]}]} | text-classification | souvenger/NLP2Linux | [
"setfit",
"safetensors",
"mpnet",
"sentence-transformers",
"text-classification",
"generated_from_setfit_trainer",
"arxiv:2209.11055",
"base_model:sentence-transformers/paraphrase-mpnet-base-v2",
"model-index",
"region:us"
] | 2024-02-08T07:09:07+00:00 | [
"2209.11055"
] | [] | TAGS
#setfit #safetensors #mpnet #sentence-transformers #text-classification #generated_from_setfit_trainer #arxiv-2209.11055 #base_model-sentence-transformers/paraphrase-mpnet-base-v2 #model-index #region-us
| SetFit with sentence-transformers/paraphrase-mpnet-base-v2
==========================================================
This is a SetFit model that can be used for Text Classification. This SetFit model uses sentence-transformers/paraphrase-mpnet-base-v2 as the Sentence Transformer embedding model. A LogisticRegression instance is used for classification.
The model has been trained using an efficient few-shot learning technique that involves:
1. Fine-tuning a Sentence Transformer with contrastive learning.
2. Training a classification head with features from the fine-tuned Sentence Transformer.
Model Details
-------------
### Model Description
* Model Type: SetFit
* Sentence Transformer body: sentence-transformers/paraphrase-mpnet-base-v2
* Classification head: a LogisticRegression instance
* Maximum Sequence Length: 512 tokens
* Number of Classes: 30 classes
### Model Sources
* Repository: SetFit on GitHub
* Paper: Efficient Few-Shot Learning Without Prompts
* Blogpost: SetFit: Efficient Few-Shot Learning Without Prompts
### Model Labels
Evaluation
----------
### Metrics
Uses
----
### Direct Use for Inference
First install the SetFit library:
Then you can load this model and run inference.
Training Details
----------------
### Training Set Metrics
### Training Hyperparameters
* batch\_size: (8, 8)
* num\_epochs: (1, 1)
* max\_steps: -1
* sampling\_strategy: oversampling
* num\_iterations: 20
* body\_learning\_rate: (2e-05, 2e-05)
* head\_learning\_rate: 2e-05
* loss: CosineSimilarityLoss
* distance\_metric: cosine\_distance
* margin: 0.25
* end\_to\_end: False
* use\_amp: False
* warmup\_proportion: 0.1
* seed: 42
* eval\_max\_steps: -1
* load\_best\_model\_at\_end: False
### Training Results
### Framework Versions
* Python: 3.10.13
* SetFit: 1.0.3
* Sentence Transformers: 2.3.1
* Transformers: 4.37.0
* PyTorch: 2.1.2
* Datasets: 2.1.0
* Tokenizers: 0.15.1
### BibTeX
| [
"### Model Description\n\n\n* Model Type: SetFit\n* Sentence Transformer body: sentence-transformers/paraphrase-mpnet-base-v2\n* Classification head: a LogisticRegression instance\n* Maximum Sequence Length: 512 tokens\n* Number of Classes: 30 classes",
"### Model Sources\n\n\n* Repository: SetFit on GitHub\n* Paper: Efficient Few-Shot Learning Without Prompts\n* Blogpost: SetFit: Efficient Few-Shot Learning Without Prompts",
"### Model Labels\n\n\n\nEvaluation\n----------",
"### Metrics\n\n\n\nUses\n----",
"### Direct Use for Inference\n\n\nFirst install the SetFit library:\n\n\nThen you can load this model and run inference.\n\n\nTraining Details\n----------------",
"### Training Set Metrics",
"### Training Hyperparameters\n\n\n* batch\\_size: (8, 8)\n* num\\_epochs: (1, 1)\n* max\\_steps: -1\n* sampling\\_strategy: oversampling\n* num\\_iterations: 20\n* body\\_learning\\_rate: (2e-05, 2e-05)\n* head\\_learning\\_rate: 2e-05\n* loss: CosineSimilarityLoss\n* distance\\_metric: cosine\\_distance\n* margin: 0.25\n* end\\_to\\_end: False\n* use\\_amp: False\n* warmup\\_proportion: 0.1\n* seed: 42\n* eval\\_max\\_steps: -1\n* load\\_best\\_model\\_at\\_end: False",
"### Training Results",
"### Framework Versions\n\n\n* Python: 3.10.13\n* SetFit: 1.0.3\n* Sentence Transformers: 2.3.1\n* Transformers: 4.37.0\n* PyTorch: 2.1.2\n* Datasets: 2.1.0\n* Tokenizers: 0.15.1",
"### BibTeX"
] | [
"TAGS\n#setfit #safetensors #mpnet #sentence-transformers #text-classification #generated_from_setfit_trainer #arxiv-2209.11055 #base_model-sentence-transformers/paraphrase-mpnet-base-v2 #model-index #region-us \n",
"### Model Description\n\n\n* Model Type: SetFit\n* Sentence Transformer body: sentence-transformers/paraphrase-mpnet-base-v2\n* Classification head: a LogisticRegression instance\n* Maximum Sequence Length: 512 tokens\n* Number of Classes: 30 classes",
"### Model Sources\n\n\n* Repository: SetFit on GitHub\n* Paper: Efficient Few-Shot Learning Without Prompts\n* Blogpost: SetFit: Efficient Few-Shot Learning Without Prompts",
"### Model Labels\n\n\n\nEvaluation\n----------",
"### Metrics\n\n\n\nUses\n----",
"### Direct Use for Inference\n\n\nFirst install the SetFit library:\n\n\nThen you can load this model and run inference.\n\n\nTraining Details\n----------------",
"### Training Set Metrics",
"### Training Hyperparameters\n\n\n* batch\\_size: (8, 8)\n* num\\_epochs: (1, 1)\n* max\\_steps: -1\n* sampling\\_strategy: oversampling\n* num\\_iterations: 20\n* body\\_learning\\_rate: (2e-05, 2e-05)\n* head\\_learning\\_rate: 2e-05\n* loss: CosineSimilarityLoss\n* distance\\_metric: cosine\\_distance\n* margin: 0.25\n* end\\_to\\_end: False\n* use\\_amp: False\n* warmup\\_proportion: 0.1\n* seed: 42\n* eval\\_max\\_steps: -1\n* load\\_best\\_model\\_at\\_end: False",
"### Training Results",
"### Framework Versions\n\n\n* Python: 3.10.13\n* SetFit: 1.0.3\n* Sentence Transformers: 2.3.1\n* Transformers: 4.37.0\n* PyTorch: 2.1.2\n* Datasets: 2.1.0\n* Tokenizers: 0.15.1",
"### BibTeX"
] | [
72,
64,
52,
8,
8,
31,
7,
176,
4,
55,
6
] | [
"passage: TAGS\n#setfit #safetensors #mpnet #sentence-transformers #text-classification #generated_from_setfit_trainer #arxiv-2209.11055 #base_model-sentence-transformers/paraphrase-mpnet-base-v2 #model-index #region-us \n### Model Description\n\n\n* Model Type: SetFit\n* Sentence Transformer body: sentence-transformers/paraphrase-mpnet-base-v2\n* Classification head: a LogisticRegression instance\n* Maximum Sequence Length: 512 tokens\n* Number of Classes: 30 classes### Model Sources\n\n\n* Repository: SetFit on GitHub\n* Paper: Efficient Few-Shot Learning Without Prompts\n* Blogpost: SetFit: Efficient Few-Shot Learning Without Prompts### Model Labels\n\n\n\nEvaluation\n----------### Metrics\n\n\n\nUses\n----### Direct Use for Inference\n\n\nFirst install the SetFit library:\n\n\nThen you can load this model and run inference.\n\n\nTraining Details\n----------------### Training Set Metrics### Training Hyperparameters\n\n\n* batch\\_size: (8, 8)\n* num\\_epochs: (1, 1)\n* max\\_steps: -1\n* sampling\\_strategy: oversampling\n* num\\_iterations: 20\n* body\\_learning\\_rate: (2e-05, 2e-05)\n* head\\_learning\\_rate: 2e-05\n* loss: CosineSimilarityLoss\n* distance\\_metric: cosine\\_distance\n* margin: 0.25\n* end\\_to\\_end: False\n* use\\_amp: False\n* warmup\\_proportion: 0.1\n* seed: 42\n* eval\\_max\\_steps: -1\n* load\\_best\\_model\\_at\\_end: False### Training Results### Framework Versions\n\n\n* Python: 3.10.13\n* SetFit: 1.0.3\n* Sentence Transformers: 2.3.1\n* Transformers: 4.37.0\n* PyTorch: 2.1.2\n* Datasets: 2.1.0\n* Tokenizers: 0.15.1### BibTeX"
] | [
-0.08306419849395752,
0.15945403277873993,
-0.0060103889554739,
0.06521530449390411,
0.12377969920635223,
0.05624328926205635,
0.07240543514490128,
0.15473853051662445,
-0.010996981523931026,
0.151503786444664,
0.0534738190472126,
0.12560825049877167,
0.09557919949293137,
0.19554723799228668,
-0.005747056100517511,
-0.28780221939086914,
0.010044028051197529,
-0.10422865301370621,
-0.01802070438861847,
0.09685008972883224,
0.10467687249183655,
-0.07374989241361618,
0.046452946960926056,
-0.04534049332141876,
-0.03217892348766327,
-0.025561485439538956,
-0.04226548969745636,
-0.03346342220902443,
0.01680644229054451,
0.04012173414230347,
0.013068373315036297,
-0.0070953937247395515,
0.05249405279755592,
-0.31340962648391724,
0.016582032665610313,
0.09373284131288528,
-0.006040000822395086,
0.08359427005052567,
0.09700480103492737,
-0.11251979321241379,
0.11309881508350372,
-0.07442495971918106,
0.07468848675489426,
0.05376794934272766,
-0.15190957486629486,
-0.17388853430747986,
-0.06985221803188324,
0.09546411782503128,
0.16799215972423553,
0.0739935040473938,
-0.06249697506427765,
0.04668273404240608,
-0.03653983771800995,
0.07735436409711838,
0.1646042764186859,
-0.22632050514221191,
-0.07051943987607956,
0.04380214586853981,
0.04193171113729477,
0.0348486490547657,
-0.10372652113437653,
-0.04438900575041771,
0.0027497457340359688,
0.04062901809811592,
0.061883680522441864,
0.01585368625819683,
0.11175132542848587,
-0.008436216972768307,
-0.1409788429737091,
-0.05979929119348526,
0.06475699692964554,
0.04389418289065361,
-0.018240470439195633,
-0.17629781365394592,
-0.02437625452876091,
-0.12398166209459305,
-0.04938157647848129,
0.013852213509380817,
-0.0006693379837088287,
0.001975112594664097,
0.011022412218153477,
0.015816612169146538,
-0.03874383866786957,
-0.04341976344585419,
0.0350065603852272,
0.02774740941822529,
0.0395895391702652,
-0.044774480164051056,
0.03888775408267975,
0.09064748883247375,
0.011721078306436539,
-0.16997051239013672,
-0.036418892443180084,
-0.03830600157380104,
-0.10143724083900452,
-0.03999612107872963,
0.013647971674799919,
0.04237251356244087,
0.05926501005887985,
0.2471853792667389,
-0.062146078795194626,
0.10339917242527008,
-0.00596210639923811,
0.02937982603907585,
0.004063509404659271,
0.0535392090678215,
-0.08955313265323639,
-0.12359520047903061,
-0.07236062735319138,
0.0875568762421608,
-0.0315556675195694,
-0.01147793885320425,
0.014020343311131,
0.05911409854888916,
0.056143637746572495,
0.0785764828324318,
0.05932066962122917,
0.04744559898972511,
-0.07313376665115356,
-0.04478498548269272,
0.05172021687030792,
-0.15383419394493103,
0.04554343968629837,
0.06235058978199959,
-0.07459437847137451,
-0.06273087859153748,
0.06507502496242523,
0.0038165701553225517,
-0.0683789923787117,
0.08992967754602432,
-0.049132782965898514,
0.01028316468000412,
-0.07566419988870621,
-0.08976297080516815,
0.04196691885590553,
-0.01579234004020691,
-0.052101101726293564,
-0.04982293024659157,
-0.09925519675016403,
-0.11603118479251862,
0.08032917976379395,
-0.10680464655160904,
-0.04324553161859512,
-0.09204570204019547,
-0.08815772086381912,
0.049046941101551056,
0.007296509109437466,
0.07803770899772644,
-0.06334852427244186,
0.04492202028632164,
-0.015388239175081253,
0.07461511343717575,
0.16010387241840363,
0.04575097933411598,
-0.05256476625800133,
0.08181483298540115,
-0.19728635251522064,
0.12976710498332977,
-0.10695052891969681,
0.06183525174856186,
-0.17897383868694305,
-0.050654567778110504,
-0.009730033576488495,
0.016352640464901924,
0.07949844747781754,
0.13237574696540833,
-0.20547610521316528,
-0.05469035729765892,
0.22095517814159393,
-0.07271774113178253,
-0.11322180926799774,
0.06083234027028084,
-0.039287734776735306,
0.0917772725224495,
0.03638705611228943,
0.12273266166448593,
0.13974712789058685,
-0.06633183360099792,
-0.01226100791245699,
-0.09057974070310593,
0.026091011241078377,
0.16723662614822388,
0.05341293662786484,
-0.04828739911317825,
0.05176433175802231,
0.01026188489049673,
-0.03276251256465912,
0.009883805178105831,
-0.04663104936480522,
-0.08157182484865189,
0.020926060155034065,
-0.0470835380256176,
0.0027330326847732067,
0.03491302207112312,
-0.03301113471388817,
-0.04099274426698685,
-0.13282980024814606,
0.047748807817697525,
0.0636974424123764,
-0.061834774911403656,
0.005945851095020771,
-0.08879592269659042,
0.008830280043184757,
0.06776056438684464,
0.021823348477482796,
-0.18423479795455933,
-0.03297213464975357,
0.01207362487912178,
-0.05883998051285744,
0.05440043285489082,
-0.06445521861314774,
0.07709214091300964,
0.045919034630060196,
-0.03826301917433739,
-0.03525771200656891,
0.06693731993436813,
0.01319510955363512,
-0.05380294471979141,
-0.22686032950878143,
-0.0345936045050621,
-0.027081865817308426,
0.211131751537323,
-0.25634825229644775,
0.047987934201955795,
-0.010588591918349266,
0.15271857380867004,
0.0077651506289839745,
-0.053436484187841415,
0.02362142689526081,
-0.01208439189940691,
-0.020309127867221832,
-0.07655719667673111,
0.019541841000318527,
0.00033104370231740177,
-0.047923944890499115,
-0.006820505950599909,
-0.19882506132125854,
-0.09600753337144852,
0.09353160113096237,
0.02724834904074669,
-0.16371357440948486,
-0.10440602898597717,
-0.0372939258813858,
-0.048986539244651794,
-0.04330918937921524,
-0.045179493725299835,
0.12144399434328079,
0.05183636397123337,
0.07474332302808762,
-0.05036497861146927,
-0.06157534942030907,
-0.004273590631783009,
-0.008172059431672096,
-0.004947381094098091,
0.17370982468128204,
-0.013135990127921104,
-0.11582779884338379,
0.08965601772069931,
0.05604716017842293,
-0.023132972419261932,
0.06577911972999573,
-0.01570240966975689,
-0.06673341244459152,
-0.06981740146875381,
0.07344559580087662,
0.06264536827802658,
0.03473358601331711,
-0.09020035713911057,
0.03365153819322586,
0.0377485528588295,
0.007015651557594538,
-0.0009756272775121033,
-0.07940676808357239,
0.013453164137899876,
0.020951779559254646,
-0.03618774935603142,
0.0406404547393322,
-0.03354950249195099,
0.017493627965450287,
0.07968000322580338,
0.02481711469590664,
0.034631986171007156,
-0.030304262414574623,
-0.055337339639663696,
-0.12103348225355148,
0.18220335245132446,
-0.1155749186873436,
-0.2069687396287918,
-0.07852838188409805,
-0.030154511332511902,
0.0037427505012601614,
-0.02119745872914791,
0.024701083078980446,
-0.0478348545730114,
-0.05628778412938118,
-0.11610279977321625,
0.03208725526928902,
0.05067478120326996,
-0.04461086541414261,
-0.03642251342535019,
0.04681076854467392,
0.08809453248977661,
-0.09149449318647385,
0.016055811196565628,
0.01801031455397606,
-0.05186733230948448,
0.011288854293525219,
0.019296202808618546,
0.03951825946569443,
0.15780559182167053,
0.08088843524456024,
0.0438959077000618,
0.004349769093096256,
0.23239800333976746,
-0.09615170955657959,
0.05986018478870392,
0.09195516258478165,
-0.01302669383585453,
0.08301080018281937,
0.2327585220336914,
0.03948119655251503,
-0.08019038289785385,
0.0639924556016922,
0.07040547579526901,
-0.02573353983461857,
-0.21405096352100372,
-0.035900115966796875,
-0.03931707143783569,
0.03433271124958992,
0.17931315302848816,
0.05004905164241791,
0.046665120869874954,
0.03930859640240669,
-0.05856930837035179,
-0.06967372447252274,
0.1167617067694664,
0.09909315407276154,
0.0026481428649276495,
0.030205892398953438,
0.10117104649543762,
-0.010390233248472214,
0.016105275601148605,
0.02531011402606964,
-0.02792784757912159,
0.15404625236988068,
-0.032915301620960236,
0.11532606929540634,
0.08917258679866791,
0.11824629455804825,
-0.03942105919122696,
0.05133433640003204,
-0.016469968482851982,
0.04306197538971901,
0.026381537318229675,
-0.08721005916595459,
0.00693946797400713,
0.06349486857652664,
0.01628451608121395,
0.015479576773941517,
-0.040282826870679855,
0.008563816547393799,
0.10847361385822296,
0.16232235729694366,
0.08141685277223587,
-0.27324163913726807,
-0.03803710639476776,
0.04166974127292633,
-0.07599548995494843,
-0.06939451396465302,
-0.022969629615545273,
0.032133668661117554,
-0.125735804438591,
0.07991844415664673,
-0.05605626106262207,
0.09068676829338074,
-0.043126340955495834,
-0.0072276475839316845,
0.10141731798648834,
0.10998935997486115,
-0.009533685632050037,
0.04430420696735382,
-0.20722077786922455,
0.15484625101089478,
-0.00469476031139493,
0.06088852882385254,
-0.07911935448646545,
0.054222699254751205,
0.038117628544569016,
-0.09848226606845856,
0.14830540120601654,
-0.015938876196742058,
-0.13180525600910187,
-0.12686635553836823,
-0.0671084076166153,
-0.04773302376270294,
0.10181478410959244,
-0.15081390738487244,
0.1217225193977356,
-0.0076804738491773605,
-0.04826366901397705,
-0.009493744932115078,
-0.02839755266904831,
-0.11674796044826508,
-0.12506286799907684,
0.03201596438884735,
-0.06610722839832306,
0.06988195329904556,
-0.05162622779607773,
-0.029404066503047943,
-0.06143992021679878,
0.17773766815662384,
-0.21073201298713684,
-0.06104395166039467,
-0.12764297425746918,
0.08595159649848938,
0.16507716476917267,
-0.0834602564573288,
0.047895316034555435,
0.0010091315489262342,
0.13669908046722412,
0.015753712505102158,
-0.034811511635780334,
0.09913871437311172,
-0.07362084835767746,
-0.20104748010635376,
-0.03407454490661621,
0.18481042981147766,
0.0696035772562027,
0.04594269394874573,
0.0037655127234756947,
0.06517089158296585,
0.00320642814040184,
-0.10254216194152832,
0.014790632762014866,
0.06823193281888962,
0.053024593740701675,
0.046140704303979874,
-0.07088591903448105,
-0.0503225214779377,
-0.11804362386465073,
0.009831750765442848,
0.07321006804704666,
0.2288341075181961,
-0.08443636447191238,
0.08287707716226578,
0.03676040098071098,
-0.07117178291082382,
-0.18887974321842194,
-0.016092142090201378,
0.10160348564386368,
-0.01644817925989628,
0.052720680832862854,
-0.1751059740781784,
0.11265874654054642,
0.08470947295427322,
-0.0007759762811474502,
0.06569594144821167,
-0.30845877528190613,
-0.14853936433792114,
0.05217529088258743,
0.03252636641263962,
-0.21583305299282074,
-0.16703180968761444,
-0.0829220712184906,
-0.03797765076160431,
-0.07865903526544571,
0.13212984800338745,
-0.06712409853935242,
0.06985606998205185,
0.04315083846449852,
-0.010379813611507416,
0.044051434844732285,
-0.04339565336704254,
0.14671194553375244,
0.0288795568048954,
0.0314176119863987,
-0.09117384999990463,
-0.03391359746456146,
-0.025613605976104736,
-0.09084641188383102,
0.07025197893381119,
-0.06553500145673752,
0.01790647953748703,
-0.11217422783374786,
-0.029788456857204437,
-0.06721002608537674,
-0.013190696015954018,
-0.09471023082733154,
-0.01783859357237816,
-0.009300073608756065,
0.10187038034200668,
0.11657354235649109,
0.007128983736038208,
0.05103268474340439,
-0.07222916930913925,
0.1427888125181198,
0.1987677812576294,
0.09684336185455322,
0.06757017970085144,
-0.11320933699607849,
0.03561650216579437,
0.0161020178347826,
-0.01096824649721384,
-0.20106305181980133,
0.0753876119852066,
0.11114069074392319,
0.028895115479826927,
0.16412946581840515,
0.026443924754858017,
-0.12304510176181793,
-0.05290762335062027,
0.09477457404136658,
-0.07374955713748932,
-0.11348854005336761,
0.02350720576941967,
0.08651448786258698,
-0.18558284640312195,
-0.07394657284021378,
0.08631671965122223,
-0.013085944578051567,
-0.006923815235495567,
0.04920841380953789,
0.10835961997509003,
-0.018083317205309868,
0.17918437719345093,
0.015973443165421486,
0.07736477255821228,
-0.10091931372880936,
0.10300947725772858,
0.1067965030670166,
-0.03955155983567238,
0.025972610339522362,
0.20607581734657288,
-0.06095372512936592,
-0.06313516199588776,
-0.001575070433318615,
0.043169938027858734,
0.0439266711473465,
-0.0021548208314925432,
0.009435461834073067,
-0.09849920123815536,
0.05535241588950157,
0.08342595398426056,
0.012823466211557388,
0.02209969237446785,
0.029416806995868683,
0.013752326369285583,
-0.06456704437732697,
0.1302647441625595,
0.13785311579704285,
0.03726176172494888,
-0.037625931203365326,
0.12073757499456406,
0.002231118967756629,
-0.03228861093521118,
0.008378570899367332,
-0.003806233871728182,
-0.1428249031305313,
0.01734251342713833,
-0.04342477396130562,
0.0461735837161541,
-0.14496789872646332,
-0.013890773057937622,
0.024099385365843773,
0.0012850501807406545,
-0.016022847965359688,
-0.018020346760749817,
-0.07556674629449844,
-0.09920841455459595,
-0.034121256321668625,
0.07838473469018936,
-0.14210957288742065,
-0.03407648950815201,
0.048201389610767365,
-0.12146050482988358,
0.08427290618419647,
0.02498578280210495,
0.008095447905361652,
-0.0016997085185721517,
-0.09517102688550949,
0.002660144353285432,
-0.011552243493497372,
-0.009557360783219337,
0.02309376932680607,
-0.20965439081192017,
0.020347001031041145,
-0.08098479360342026,
-0.041779641062021255,
0.02315853163599968,
-0.01365124061703682,
-0.13910728693008423,
0.06350631266832352,
-0.023358739912509918,
-0.04377387464046478,
-0.07233728468418121,
0.026517832651734352,
0.07518287748098373,
-0.022745834663510323,
0.12545791268348694,
-0.06521070003509521,
0.08524662256240845,
-0.21612615883350372,
-0.011245641857385635,
0.004044529981911182,
-0.03601302579045296,
0.011295735836029053,
-0.019329670816659927,
0.10531461238861084,
-0.050317879766225815,
0.05983857065439224,
-0.020200081169605255,
0.0016427092486992478,
0.04571877792477608,
-0.057615216821432114,
-0.006810610182583332,
0.08344200998544693,
0.05464641749858856,
0.03421761840581894,
-0.037200964987277985,
-0.014547395519912243,
0.01579013653099537,
-0.0051322216168046,
-0.041133083403110504,
0.12029191851615906,
0.16834767162799835,
0.10292693227529526,
0.005582103040069342,
0.04039636626839638,
-0.13760198652744293,
-0.044428750872612,
0.19007496535778046,
-0.07018454372882843,
0.05176950618624687,
-0.06919210404157639,
0.07927727699279785,
0.096622034907341,
-0.2281593382358551,
0.07882778346538544,
-0.06587369740009308,
-0.09185878187417984,
-0.0763116329908371,
-0.18177524209022522,
-0.07163520157337189,
-0.0738641694188118,
-0.02340652421116829,
-0.11749430745840073,
0.03295306861400604,
0.062375862151384354,
0.02278575301170349,
0.03332945331931114,
0.08261419832706451,
-0.0200524739921093,
-0.028666384518146515,
0.06452487409114838,
0.02971368096768856,
0.033343516290187836,
-0.0018313609762117267,
-0.021476153284311295,
0.00527555076405406,
0.02191106230020523,
0.06787109375,
0.02076893113553524,
-0.0055960360914468765,
0.047261353582143784,
-0.032462768256664276,
-0.1103258952498436,
0.02922710031270981,
-0.02675922028720379,
-0.02223973348736763,
0.15750505030155182,
0.07411342114210129,
-0.011943555437028408,
-0.016339153051376343,
0.21544551849365234,
-0.06582079082727432,
-0.08187293261289597,
-0.1915305256843567,
0.19493629038333893,
0.024350393563508987,
0.017850477248430252,
-0.029414815828204155,
-0.11447067558765411,
-0.015790656208992004,
0.14322754740715027,
0.1704721748828888,
-0.03632432222366333,
0.008539223112165928,
0.05901405215263367,
0.004512257874011993,
-0.0030537534039467573,
0.015034159645438194,
0.10575024783611298,
0.11790728569030762,
-0.056560929864645004,
0.0737227275967598,
0.018605776131153107,
-0.09553301334381104,
-0.0799395963549614,
0.07499447464942932,
0.015246434137225151,
0.03320840373635292,
-0.011844598688185215,
0.15449543297290802,
-0.12850064039230347,
-0.17100195586681366,
0.03647247329354286,
-0.14350035786628723,
-0.1563165783882141,
-0.05859643220901489,
0.0019425860373303294,
0.04618639498949051,
0.06428535282611847,
0.046589016914367676,
-0.04455365985631943,
0.08860024064779282,
0.021286359056830406,
0.009240536950528622,
-0.080538809299469,
0.028106864541769028,
-0.0723872184753418,
0.2352437525987625,
-0.024636531248688698,
-0.0035197101533412933,
0.14423638582229614,
-0.01297989021986723,
-0.12277987599372864,
0.008051250129938126,
0.09828075766563416,
-0.115563303232193,
0.0776204988360405,
0.16017718613147736,
-0.03241688013076782,
0.0838785320520401,
0.09636791795492172,
-0.0981600359082222,
0.013276557438075542,
-0.04769507050514221,
-0.016470715403556824,
-0.08846442401409149,
0.04458135366439819,
-0.04019898548722267,
0.1259990781545639,
0.21414893865585327,
-0.07955477386713028,
0.003839150071144104,
-0.04020746052265167,
-0.00010871962149394676,
-0.026157265529036522,
0.10318376123905182,
-0.03518553823232651,
-0.23829440772533417,
0.06475360691547394,
0.025607338175177574,
0.09017840772867203,
-0.19532018899917603,
-0.08891493827104568,
0.06967543810606003,
-0.03544342890381813,
-0.07844912260770798,
0.14435110986232758,
0.09351013600826263,
0.024435147643089294,
-0.03824109584093094,
-0.15069608390331268,
-0.0049976264126598835,
0.21175169944763184,
-0.11435126513242722,
-0.05753830820322037
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | humung/polyglot-ko-12.8b-vlending-v0.5 | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-08T07:09:21+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
31,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06646376848220825,
0.2168014943599701,
-0.00225935154594481,
0.023818302899599075,
0.1271018385887146,
-0.001635765191167593,
0.04218708351254463,
0.13324736058712006,
-0.020175931975245476,
0.11144465953111649,
0.046588581055402756,
0.09377603232860565,
0.09928803145885468,
0.18404334783554077,
0.04859916493296623,
-0.2059975117444992,
0.007056170143187046,
-0.09090408682823181,
0.014076028019189835,
0.1116579994559288,
0.13719257712364197,
-0.10291384905576706,
0.08272874355316162,
-0.04045208916068077,
-0.02019004337489605,
0.00012576708104461432,
-0.09259183704853058,
-0.07032395154237747,
0.06885425746440887,
0.06264153122901917,
0.051234472543001175,
0.001456156256608665,
0.09140396863222122,
-0.2864592671394348,
0.017265573143959045,
0.08406311273574829,
0.0027674848679453135,
0.06290827691555023,
0.07236549258232117,
-0.07389893382787704,
0.11328595131635666,
-0.08021481335163116,
0.13019037246704102,
0.08625296503305435,
-0.062064990401268005,
-0.23071379959583282,
-0.07525765895843506,
0.0963398814201355,
0.12251301854848862,
0.06215599179267883,
-0.022921854630112648,
0.15455181896686554,
-0.06248689442873001,
0.012971068732440472,
0.1294165402650833,
-0.11526761949062347,
-0.05572471022605896,
0.061741601675748825,
0.11775490641593933,
0.10740239918231964,
-0.14110268652439117,
-0.0017287094378843904,
0.04900608956813812,
0.029121357947587967,
0.08589313924312592,
0.022661056369543076,
0.12003941088914871,
0.04652795568108559,
-0.13695219159126282,
-0.04037507623434067,
0.12011898308992386,
0.038862764835357666,
-0.06446044892072678,
-0.2168138176202774,
-0.006778308190405369,
-0.0601806715130806,
-0.014732478186488152,
-0.07019448280334473,
0.039128515869379044,
-0.02470310963690281,
0.07317749410867691,
-0.04465159401297569,
-0.1063927412033081,
-0.0421026237308979,
0.0892222449183464,
0.07748593389987946,
0.011527054943144321,
-0.02519804798066616,
0.04627908393740654,
0.13455867767333984,
0.05402068421244621,
-0.10399353504180908,
-0.07017925381660461,
-0.06942764669656754,
-0.09420394152402878,
-0.04035796597599983,
0.056760527193546295,
0.031942449510097504,
0.02665667235851288,
0.22703726589679718,
0.016653569415211678,
0.04155244305729866,
0.0224777739495039,
0.01032855175435543,
0.043662428855895996,
0.0955500528216362,
-0.05303520709276199,
-0.15660029649734497,
-0.04072032496333122,
0.09077946096658707,
-0.0027527001220732927,
-0.036689214408397675,
-0.03966725245118141,
0.03849169611930847,
0.06843466311693192,
0.13122352957725525,
0.07552056759595871,
-0.017929591238498688,
-0.04813180863857269,
-0.030096933245658875,
0.23523783683776855,
-0.1493375599384308,
0.04426715523004532,
-0.02271856553852558,
-0.01804111897945404,
-0.03908449783921242,
0.03597262129187584,
0.022118929773569107,
-0.000004518366949923802,
0.09706240892410278,
-0.058981191366910934,
-0.05378659814596176,
-0.10168042778968811,
-0.03272576630115509,
0.04088849574327469,
-0.013975566253066063,
-0.010589460842311382,
-0.09025166928768158,
-0.09490354359149933,
-0.04766594246029854,
0.05537205561995506,
-0.05123869329690933,
-0.03770573064684868,
0.009465423412621021,
-0.08151785284280777,
-0.005444355774670839,
-0.005417742300778627,
0.10699385404586792,
-0.03222226724028587,
0.04445803165435791,
-0.027600755915045738,
0.05225523188710213,
0.09919606149196625,
0.031576547771692276,
-0.0773419588804245,
0.0561848059296608,
-0.22559374570846558,
0.07503069192171097,
-0.11481974273920059,
0.04335082694888115,
-0.1704932004213333,
-0.042439818382263184,
0.005444696638733149,
0.0139949731528759,
0.013206101022660732,
0.12720820307731628,
-0.19255615770816803,
-0.01654396951198578,
0.13260798156261444,
-0.09212633967399597,
-0.118110790848732,
0.07884611934423447,
-0.029701577499508858,
0.1624738723039627,
0.04682036489248276,
-0.027025915682315826,
0.09224298596382141,
-0.16434773802757263,
-0.07092688232660294,
-0.00949116237461567,
-0.01727987825870514,
0.12109188735485077,
0.07512219995260239,
-0.05991523340344429,
0.046571120619773865,
0.02832140028476715,
-0.038078423589468,
-0.04424772411584854,
-0.050857074558734894,
-0.10884185880422592,
-0.01070026308298111,
-0.08987759798765182,
0.04065500199794769,
-0.01250192429870367,
-0.07916021347045898,
-0.029885273426771164,
-0.18612512946128845,
-0.0030564051121473312,
0.10038342326879501,
0.0035033065360039473,
-0.005652366206049919,
-0.08666291832923889,
0.026358824223279953,
-0.03112892620265484,
-0.008404186926782131,
-0.16764774918556213,
-0.04399421438574791,
0.046902090311050415,
-0.16094985604286194,
0.020117372274398804,
-0.06413903087377548,
0.06334125250577927,
0.03641495108604431,
-0.05590536445379257,
-0.0248766727745533,
-0.01730942726135254,
0.011945613659918308,
-0.05083848536014557,
-0.18994836509227753,
-0.056277405470609665,
-0.037882111966609955,
0.149809330701828,
-0.25956398248672485,
0.032966937869787216,
0.051140617579221725,
0.14649195969104767,
0.00406361510977149,
-0.05115427449345589,
0.01429014839231968,
-0.05360214412212372,
-0.054652128368616104,
-0.06746816635131836,
-0.006135428790003061,
-0.027576493099331856,
-0.05147203803062439,
0.019243421033024788,
-0.1755700707435608,
-0.021410830318927765,
0.09424154460430145,
0.12876708805561066,
-0.1486445665359497,
-0.018640631809830666,
-0.048725154250860214,
-0.06339836865663528,
-0.0715010017156601,
-0.07038594037294388,
0.10712739825248718,
0.0513901449739933,
0.04796046018600464,
-0.07435787469148636,
-0.07092321664094925,
0.02726263552904129,
0.006906150374561548,
-0.03382374346256256,
0.08727246522903442,
0.05199531093239784,
-0.09209315478801727,
0.0756213590502739,
0.1092359870672226,
0.07177663594484329,
0.09363535046577454,
0.01574566215276718,
-0.11756632477045059,
-0.028492970392107964,
0.036266472190618515,
0.02740776725113392,
0.1465986967086792,
-0.05952361226081848,
0.04016614332795143,
0.04494241625070572,
-0.04170418903231621,
0.022319864481687546,
-0.08787637203931808,
0.024075502529740334,
0.025203049182891846,
-0.0034381982404738665,
0.06284574419260025,
-0.02525499276816845,
-0.0050758360885083675,
0.07016654312610626,
0.047779910266399384,
0.04621000960469246,
0.009655474685132504,
-0.01720241829752922,
-0.1047825813293457,
0.16950392723083496,
-0.0951867327094078,
-0.269941508769989,
-0.17632324993610382,
0.026197833940386772,
0.04035249724984169,
-0.022378476336598396,
0.031619444489479065,
-0.07056326419115067,
-0.10630585998296738,
-0.1060405746102333,
-0.002429972169920802,
0.01714223250746727,
-0.06364088505506516,
-0.0741225928068161,
0.07348573952913284,
0.04382912442088127,
-0.14902326464653015,
0.038552410900592804,
0.055694397538900375,
-0.057955220341682434,
-0.0233661737293005,
0.09118817001581192,
0.12397737801074982,
0.14583967626094818,
-0.021366750821471214,
-0.028626007959246635,
0.029004426673054695,
0.19620531797409058,
-0.13469526171684265,
0.10371150821447372,
0.13814030587673187,
-0.04545360431075096,
0.08360563963651657,
0.1560150384902954,
0.029186224564909935,
-0.08317049592733383,
0.05044832453131676,
0.04082648828625679,
-0.043159641325473785,
-0.2666129767894745,
-0.0534592866897583,
0.012832709588110447,
-0.06255637854337692,
0.09786593168973923,
0.10183793306350708,
0.11542957276105881,
0.034910861402750015,
-0.07166364789009094,
-0.043925940990448,
-0.0058974819257855415,
0.11737963557243347,
-0.05490213260054588,
-0.012639665976166725,
0.07686592638492584,
-0.05086168646812439,
0.005355054512619972,
0.10266812145709991,
0.02973790094256401,
0.17442677915096283,
0.020399179309606552,
0.11231429129838943,
0.06195578724145889,
0.08633565157651901,
0.0007386076031252742,
0.02951662428677082,
0.05147615820169449,
0.017203815281391144,
-0.002300140680745244,
-0.10421168059110641,
-0.006156572140753269,
0.1449710875749588,
0.028103826567530632,
0.029669636860489845,
-0.0018948549404740334,
-0.005003341939300299,
0.05121048167347908,
0.1746254414319992,
-0.011592294089496136,
-0.22072425484657288,
-0.0845772922039032,
0.06936841458082199,
-0.06218599155545235,
-0.12968985736370087,
-0.026130788028240204,
0.045467354357242584,
-0.17519839107990265,
0.026703642681241035,
-0.027433741837739944,
0.0919293761253357,
-0.09345759451389313,
-0.02221956104040146,
0.03687324374914169,
0.084866963326931,
-0.014529162086546421,
0.08703910559415817,
-0.14498743414878845,
0.11886418610811234,
0.02978132851421833,
0.09024628251791,
-0.11081171780824661,
0.07909037172794342,
-0.007550720125436783,
0.009180475026369095,
0.19379350543022156,
-0.011335089802742004,
-0.03514958545565605,
-0.08774717897176743,
-0.11210042238235474,
-0.013537433929741383,
0.12687496840953827,
-0.1243172138929367,
0.08773399889469147,
-0.015198243781924248,
-0.044079482555389404,
0.00937260314822197,
-0.12100647389888763,
-0.17273177206516266,
-0.19628387689590454,
0.05585884302854538,
-0.09575839340686798,
0.025643249973654747,
-0.11914430558681488,
-0.07089093327522278,
-0.02952558360993862,
0.241120383143425,
-0.1745356321334839,
-0.06510113179683685,
-0.1468164622783661,
-0.046294767409563065,
0.1662203073501587,
-0.04437198117375374,
0.0718095526099205,
-0.0208172257989645,
0.20345525443553925,
0.005988610442727804,
-0.004939318168908358,
0.06724198162555695,
-0.08892562240362167,
-0.16873881220817566,
-0.06771010160446167,
0.1510489284992218,
0.11680185794830322,
0.04907919466495514,
-0.002248800592496991,
0.0011772146681323647,
-0.016943959519267082,
-0.1137804463505745,
-0.0033210667315870523,
0.16037839651107788,
0.03878779336810112,
0.025986969470977783,
-0.05243593826889992,
-0.08797456324100494,
-0.06899320334196091,
-0.06853509694337845,
0.06221301481127739,
0.19590823352336884,
-0.10376439243555069,
0.1700313836336136,
0.147536963224411,
-0.07305635511875153,
-0.23175598680973053,
0.035342130810022354,
0.04983805492520332,
0.0014306638622656465,
0.04886869341135025,
-0.18252557516098022,
0.10521943867206573,
0.019543392583727837,
-0.05505957826972008,
0.13485197722911835,
-0.1557481735944748,
-0.1552847921848297,
0.0722852572798729,
0.03904085233807564,
-0.22423844039440155,
-0.1354004591703415,
-0.09622503817081451,
-0.05825018882751465,
-0.14065024256706238,
0.06054598465561867,
-0.002136280992999673,
0.015948504209518433,
0.03500790148973465,
-0.0015643214574083686,
0.027123261243104935,
-0.058935679495334625,
0.18609118461608887,
-0.004065449349582195,
0.020676052197813988,
-0.060264769941568375,
-0.0478842556476593,
0.09839435666799545,
-0.06130504235625267,
0.12208222597837448,
0.004057085141539574,
0.01594383642077446,
-0.10362856835126877,
-0.048314861953258514,
-0.04328322783112526,
0.05154227837920189,
-0.07548051327466965,
-0.10070807486772537,
-0.043625857681035995,
0.08841723203659058,
0.07005169242620468,
-0.03383097052574158,
0.00549331633374095,
-0.07189501076936722,
0.10019614547491074,
0.17795267701148987,
0.17573626339435577,
0.009926567785441875,
-0.07241068035364151,
0.01677953451871872,
-0.04142116755247116,
0.044231921434402466,
-0.2513144314289093,
0.03756171092391014,
0.06098250672221184,
0.029438555240631104,
0.09217222779989243,
-0.020435843616724014,
-0.1820858269929886,
-0.04050002992153168,
0.08094815909862518,
-0.05452597141265869,
-0.22617179155349731,
-0.019085140898823738,
0.0954197570681572,
-0.2020406424999237,
-0.007372708059847355,
0.03995226323604584,
-0.048725228756666183,
-0.023169852793216705,
0.00010950004070764408,
0.06317184865474701,
0.002471912419423461,
0.09773622453212738,
0.0735151618719101,
0.09715340286493301,
-0.08337292820215225,
0.10562895983457565,
0.10150538384914398,
-0.09572599828243256,
0.03605884686112404,
0.06754924356937408,
-0.05300498008728027,
-0.043293699622154236,
0.03665391728281975,
0.033023297786712646,
0.005234600510448217,
-0.060321882367134094,
0.013913018628954887,
-0.036497246474027634,
0.044923391193151474,
0.08326134830713272,
0.03754979372024536,
-0.013354414142668247,
0.06462216377258301,
0.03401726484298706,
-0.10898099094629288,
0.10366570204496384,
0.01731540448963642,
0.04105307161808014,
-0.08384523540735245,
-0.019968897104263306,
0.035425446927547455,
0.030576206743717194,
-0.01765924133360386,
-0.02306121215224266,
-0.02860277332365513,
-0.01614218018949032,
-0.14299540221691132,
-0.023106401786208153,
-0.07243485748767853,
0.006181265693157911,
0.014656842686235905,
-0.031884219497442245,
-0.011233693920075893,
0.02475680410861969,
-0.06979699432849884,
-0.07426341623067856,
-0.006949664559215307,
0.09833318740129471,
-0.15115703642368317,
0.008848577737808228,
0.06907843053340912,
-0.11088496446609497,
0.08190931379795074,
-0.008411259390413761,
0.016245156526565552,
0.022527478635311127,
-0.15448406338691711,
0.05601610988378525,
0.0008648968650959432,
0.01916889287531376,
0.025886621326208115,
-0.16471809148788452,
0.004104440100491047,
-0.04661374166607857,
-0.02149827405810356,
-0.00004464812809601426,
-0.02647159807384014,
-0.12325995415449142,
0.06858719140291214,
-0.015622655861079693,
-0.035931166261434555,
-0.02701525390148163,
0.0539589487016201,
0.07888586074113846,
-0.027474910020828247,
0.10445091128349304,
-0.008690856397151947,
0.04941811040043831,
-0.16801609098911285,
-0.02470702864229679,
-0.04982255399227142,
0.019377702847123146,
0.009884213097393513,
-0.007693959400057793,
0.04183054715394974,
-0.00976533442735672,
0.21883612871170044,
-0.05075952783226967,
0.1607085019350052,
0.05847611650824547,
-0.017352959141135216,
-0.0007513365126214921,
0.06180921941995621,
0.05997028574347496,
0.04658793285489082,
0.009480604901909828,
0.023740366101264954,
-0.022450892254710197,
-0.006695089396089315,
-0.15932634472846985,
0.01890849508345127,
0.14999441802501678,
0.06301083415746689,
0.024745315313339233,
0.05866100639104843,
-0.12775006890296936,
-0.12135478109121323,
0.09311001747846603,
-0.026755332946777344,
0.00928465835750103,
-0.08245618641376495,
0.1358020007610321,
0.14980104565620422,
-0.14000412821769714,
0.05256148427724838,
-0.06134212389588356,
-0.05217423290014267,
-0.10388828068971634,
-0.12032219022512436,
-0.05887215584516525,
-0.053666237741708755,
0.002330566756427288,
-0.03760887682437897,
0.054546963423490524,
0.03344334661960602,
-0.009351172484457493,
-0.00022941511997487396,
0.13597318530082703,
-0.019751882180571556,
-0.0028988157864660025,
0.048313532024621964,
0.03693558648228645,
0.02373051457107067,
-0.05275435373187065,
0.02940409444272518,
0.02539868652820587,
0.032232340425252914,
0.06546790152788162,
0.033412106335163116,
-0.047448933124542236,
0.03804153576493263,
-0.0025254099164158106,
-0.11207924783229828,
0.019641218706965446,
-0.00460948096588254,
-0.0742158442735672,
0.1268945336341858,
0.0407399944961071,
0.010224059224128723,
-0.03741471841931343,
0.24361543357372284,
-0.06653323769569397,
-0.06378097087144852,
-0.13251738250255585,
0.10491154342889786,
-0.0027236645109951496,
0.06476365029811859,
0.023412218317389488,
-0.1284150779247284,
0.005243356805294752,
0.13858191668987274,
0.12181595712900162,
0.0045748427510261536,
0.009228081442415714,
0.0518609918653965,
0.0025186820421367884,
-0.06998204439878464,
0.054019294679164886,
0.06992026418447495,
0.12919506430625916,
-0.07847554981708527,
0.07680778950452805,
0.0006860480643808842,
-0.08370215445756912,
-0.02947772853076458,
0.11312682181596756,
-0.0409729965031147,
0.03491825982928276,
-0.047444481402635574,
0.10916327685117722,
-0.05787910893559456,
-0.29412412643432617,
0.02350960113108158,
-0.09588567912578583,
-0.15202060341835022,
-0.018367812037467957,
0.05944539234042168,
-0.02624768204987049,
0.018029648810625076,
0.06971040368080139,
-0.06011629104614258,
0.20098382234573364,
0.0335683599114418,
-0.07864278554916382,
-0.0664360448718071,
0.04837050288915634,
-0.06564252078533173,
0.2949807047843933,
0.008418165147304535,
0.02863333560526371,
0.10770907253026962,
-0.03253700211644173,
-0.18271861970424652,
0.010723991319537163,
0.1133992001414299,
-0.08056149631738663,
0.08200647681951523,
0.19000613689422607,
-0.012578671798110008,
0.1209007054567337,
0.05294662341475487,
-0.047376248985528946,
0.04217283055186272,
-0.03389401361346245,
-0.051268599927425385,
-0.10752558708190918,
0.058453381061553955,
-0.05909625440835953,
0.15447644889354706,
0.10152646154165268,
-0.05671518296003342,
-0.004550917539745569,
-0.05555408447980881,
0.04875178262591362,
0.01804669201374054,
0.12263146042823792,
0.02951994352042675,
-0.1865430772304535,
0.032826557755470276,
-0.01144319772720337,
0.10186848044395447,
-0.25588861107826233,
-0.08421015739440918,
0.08833149075508118,
-0.011924264021217823,
-0.05105875805020332,
0.10560628771781921,
0.057650718837976456,
0.04243382066488266,
-0.043439045548439026,
-0.10480839014053345,
-0.02186836116015911,
0.14663739502429962,
-0.1469624787569046,
-0.025013303384184837
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# image_classification
This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.1874
- Accuracy: 0.9517
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| No log | 1.0 | 52 | 0.2941 | 0.9227 |
| No log | 2.0 | 104 | 0.2064 | 0.9517 |
| No log | 3.0 | 156 | 0.2221 | 0.9372 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "google/vit-base-patch16-224-in21k", "model-index": [{"name": "image_classification", "results": []}]} | image-classification | rendy-k/image_classification | [
"transformers",
"tensorboard",
"safetensors",
"vit",
"image-classification",
"generated_from_trainer",
"base_model:google/vit-base-patch16-224-in21k",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T07:15:47+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| image\_classification
=====================
This model is a fine-tuned version of google/vit-base-patch16-224-in21k on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 0.1874
* Accuracy: 0.9517
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
75,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.10594063252210617,
0.08845723420381546,
-0.0017734578577801585,
0.1274452656507492,
0.16747550666332245,
0.015808578580617905,
0.14262361824512482,
0.10733547806739807,
-0.06078345328569412,
0.0580371730029583,
0.12698964774608612,
0.10785748809576035,
0.02874813973903656,
0.151635080575943,
-0.04963736981153488,
-0.2168433964252472,
0.011168776080012321,
0.03095514141023159,
-0.04811825230717659,
0.11507515609264374,
0.08195509016513824,
-0.13784722983837128,
0.11778523772954941,
0.0062084258534014225,
-0.1899637132883072,
0.0011305606458336115,
0.03588073328137398,
-0.04471570998430252,
0.13196900486946106,
0.044009678065776825,
0.1277812421321869,
0.022960685193538666,
0.09207423776388168,
-0.1532478630542755,
0.011612383648753166,
0.08348769694566727,
-0.014117751270532608,
0.08285434544086456,
0.058402545750141144,
0.0326942540705204,
0.026064394041895866,
-0.08764725178480148,
0.04083879664540291,
0.01767805777490139,
-0.1186753362417221,
-0.2093178927898407,
-0.07382980734109879,
0.024831226095557213,
0.10430760681629181,
0.08294609189033508,
-0.009821920655667782,
0.14484430849552155,
-0.03662417456507683,
0.0903729647397995,
0.18690207600593567,
-0.28518638014793396,
-0.07708466053009033,
0.04965297132730484,
0.015341293066740036,
0.0778593197464943,
-0.09941516816616058,
0.007553376257419586,
0.05620948597788811,
0.016047563403844833,
0.11943121999502182,
-0.006997319404035807,
-0.05903029069304466,
-0.01129148155450821,
-0.13252627849578857,
-0.021626612171530724,
0.14326338469982147,
0.08411920070648193,
-0.04569058120250702,
-0.03888527676463127,
-0.07429411262273788,
-0.13663624227046967,
-0.05464022979140282,
0.004361697472631931,
0.052237849682569504,
-0.033340103924274445,
-0.0722898617386818,
-0.022496037185192108,
-0.11196130514144897,
-0.07977744191884995,
-0.04314190521836281,
0.09406956285238266,
0.038687512278556824,
0.011658417992293835,
-0.017768751829862595,
0.09588415920734406,
-0.04710201174020767,
-0.12806668877601624,
0.021072065457701683,
0.0269866231828928,
0.0008201840100809932,
-0.048768457025289536,
-0.03483189642429352,
-0.10688547044992447,
0.007417092565447092,
0.09053625166416168,
-0.04224000871181488,
0.05215167626738548,
-0.016305992379784584,
0.053821057081222534,
-0.11995016038417816,
0.1883580982685089,
-0.06638734042644501,
0.0009583496139384806,
0.029502101242542267,
0.10603964328765869,
0.05804107338190079,
0.003583901794627309,
-0.13628411293029785,
0.01708259992301464,
0.1017795205116272,
0.003316165879368782,
-0.04297018423676491,
0.07947196811437607,
-0.05588509142398834,
-0.010638980194926262,
0.07064312696456909,
-0.061836808919906616,
0.03516871854662895,
-0.00412219762802124,
-0.055332526564598083,
-0.06269211322069168,
0.04983394965529442,
0.013041095808148384,
0.009082824923098087,
0.10001324862241745,
-0.10160189867019653,
0.0253615640103817,
-0.08185767382383347,
-0.11994560062885284,
0.015152815729379654,
-0.09471549093723297,
0.020720940083265305,
-0.13681672513484955,
-0.1461969017982483,
-0.001444731722585857,
0.06291249394416809,
-0.038304392248392105,
-0.02973189577460289,
-0.03913617879152298,
-0.08020113408565521,
0.01799837313592434,
0.0006584526272490621,
0.0821656733751297,
-0.0661734938621521,
0.10321972519159317,
0.04785827547311783,
0.07443667203187943,
-0.05070575699210167,
0.036447204649448395,
-0.10000156611204147,
0.04522474482655525,
-0.19629652798175812,
0.002088688313961029,
-0.0565408356487751,
0.08699357509613037,
-0.0924595445394516,
-0.07775197923183441,
-0.005169438663870096,
-0.009557240642607212,
0.0677090510725975,
0.09759485721588135,
-0.15999692678451538,
-0.04658062011003494,
0.1310330331325531,
-0.09334225207567215,
-0.14957498013973236,
0.13863016664981842,
-0.028138810768723488,
0.01637820340692997,
0.04944712296128273,
0.1753847897052765,
0.06340961903333664,
-0.13487571477890015,
0.006759908050298691,
-0.019961223006248474,
0.04032299295067787,
-0.05740085989236832,
0.07513551414012909,
0.011746427975594997,
-0.023193197324872017,
0.0064379009418189526,
-0.08509350568056107,
0.09090834856033325,
-0.07697758823633194,
-0.08229788392782211,
-0.04757151007652283,
-0.10267127305269241,
0.04310256987810135,
0.06607218831777573,
0.05421169474720955,
-0.10480470210313797,
-0.09761308133602142,
0.03853253647685051,
0.07421339303255081,
-0.06060029938817024,
0.008153332397341728,
-0.08005984127521515,
0.10691799223423004,
-0.10751590132713318,
-0.014107989147305489,
-0.11939078569412231,
-0.06173627823591232,
0.025136001408100128,
-0.03504244610667229,
0.0028062586206942797,
-0.02558048442006111,
0.07555704563856125,
0.08528058230876923,
-0.07549721002578735,
-0.053845204412937164,
-0.02559848316013813,
0.01271100528538227,
-0.11714420467615128,
-0.1917707324028015,
-0.022125398740172386,
-0.019889211282134056,
0.14292103052139282,
-0.23566347360610962,
0.02914389967918396,
0.010201849974691868,
0.09953378140926361,
0.058450888842344284,
-0.013380005955696106,
-0.02688777446746826,
0.04577014595270157,
-0.03906020522117615,
-0.08000088483095169,
0.06499172747135162,
0.004292610567063093,
-0.05704144388437271,
-0.01891751028597355,
-0.12425325810909271,
0.18591274321079254,
0.14623309671878815,
-0.08812517672777176,
-0.07352937757968903,
0.010089398361742496,
-0.0287910345941782,
-0.0299181267619133,
-0.05148031935095787,
-0.0008984978194348514,
0.08035637438297272,
-0.005576944909989834,
0.1575467586517334,
-0.08705244213342667,
-0.013954507187008858,
0.0534968227148056,
-0.03360870108008385,
-0.027110731229186058,
0.08727233111858368,
0.13704581558704376,
-0.14525610208511353,
0.14731110632419586,
0.18857088685035706,
-0.09011339396238327,
0.12267163395881653,
-0.025439612567424774,
-0.06161880120635033,
-0.017819473519921303,
0.024461131542921066,
0.020228836685419083,
0.1571156084537506,
-0.11419028788805008,
-0.0007276919786818326,
0.008710864000022411,
-0.000876626290846616,
-0.0063386945985257626,
-0.2266739010810852,
-0.03024623543024063,
0.03082679770886898,
-0.044318728148937225,
0.004085292108356953,
-0.023786848410964012,
-0.013926473446190357,
0.09497568756341934,
0.012842647731304169,
-0.07168250530958176,
0.03814786300063133,
0.007063243538141251,
-0.06903266906738281,
0.19299696385860443,
-0.07817833125591278,
-0.1931857317686081,
-0.14108125865459442,
-0.046841759234666824,
-0.05798037722706795,
0.03670164942741394,
0.05031266808509827,
-0.09719958156347275,
-0.06381234526634216,
-0.10664860904216766,
0.00432292465120554,
0.04653463885188103,
0.03128731995820999,
0.005729840602725744,
0.006119322031736374,
0.12169058620929718,
-0.07822448015213013,
0.0006773995701223612,
-0.02078738436102867,
-0.037768710404634476,
0.04007638618350029,
0.03893153741955757,
0.1266191303730011,
0.10742015391588211,
-0.03362791985273361,
0.001832211622968316,
-0.019215216860175133,
0.24182546138763428,
-0.07923873513936996,
0.006093005649745464,
0.14163434505462646,
0.0005705058574676514,
0.051968760788440704,
0.150914266705513,
0.052485473453998566,
-0.10298285633325577,
0.019937871024012566,
0.04030151292681694,
-0.022748257964849472,
-0.16480687260627747,
-0.02704290859401226,
-0.034403372555971146,
-0.004538108594715595,
0.12435605376958847,
0.04640733078122139,
0.022480741143226624,
0.08416466414928436,
0.009378817863762379,
0.09583129733800888,
-0.019076703116297722,
0.07079003751277924,
0.10876311361789703,
0.04278842732310295,
0.11632424592971802,
-0.046285733580589294,
-0.049393266439437866,
0.032223936170339584,
0.012065345421433449,
0.19861792027950287,
0.0134430518373847,
0.11721637845039368,
0.0386316291987896,
0.18132108449935913,
0.012911917641758919,
0.05080196633934975,
-0.012304572388529778,
-0.05268599092960358,
-0.01281944289803505,
-0.04991624131798744,
-0.004027975723147392,
0.036247722804546356,
-0.05296771973371506,
0.055381935089826584,
-0.08966561406850815,
0.025239281356334686,
0.0589987151324749,
0.21993063390254974,
0.06746707856655121,
-0.36730697751045227,
-0.08741390705108643,
0.010996378026902676,
-0.0037984238006174564,
-0.059285424649715424,
0.014943322166800499,
0.14641724526882172,
-0.04017495736479759,
0.03851805999875069,
-0.0973140075802803,
0.07802986353635788,
-0.036661870777606964,
0.03537258133292198,
0.10173043608665466,
0.08263641595840454,
-0.0025735299568623304,
0.0592440664768219,
-0.2392483502626419,
0.2720709443092346,
0.01929396577179432,
0.06654781848192215,
-0.03490668162703514,
-0.00345648848451674,
0.03696134686470032,
0.12209669500589371,
0.08584710210561752,
-0.0094683188945055,
-0.0045754374004900455,
-0.22137868404388428,
-0.06391218304634094,
0.015967044979333878,
0.08157192170619965,
-0.02708916924893856,
0.09251043200492859,
-0.04137831926345825,
-0.028427088633179665,
0.07352384179830551,
0.0008174997055903077,
-0.09613318741321564,
-0.08466104418039322,
-0.01599203050136566,
0.04362067952752113,
-0.007351035717874765,
-0.0824410617351532,
-0.10041864216327667,
-0.14058208465576172,
0.11360464990139008,
-0.0172681026160717,
-0.01975809782743454,
-0.10209551453590393,
0.08580399304628372,
0.05166292563080788,
-0.0786176323890686,
0.0732993558049202,
-0.015183492563664913,
0.10383445024490356,
0.03226521611213684,
-0.059149570763111115,
0.1335209459066391,
-0.07215198874473572,
-0.16677656769752502,
-0.07503088563680649,
0.08610699325799942,
-0.020154651254415512,
0.026842841878533363,
-0.008523069322109222,
0.02264067530632019,
-0.012523822486400604,
-0.05234616622328758,
0.054024502635002136,
-0.025204788893461227,
0.04756578430533409,
0.0014661047607660294,
-0.029808681458234787,
-0.002934676595032215,
-0.061034828424453735,
-0.03552309796214104,
0.11198589205741882,
0.27935031056404114,
-0.08749557286500931,
-0.024621648713946342,
0.03665051981806755,
-0.040608927607536316,
-0.21124114096164703,
0.061805158853530884,
0.03190341591835022,
-0.0007600279059261084,
0.04712197557091713,
-0.1321321576833725,
0.09354400634765625,
0.08692976087331772,
-0.03787137567996979,
0.12883013486862183,
-0.2562580704689026,
-0.13291695713996887,
0.09820496290922165,
0.19180895388126373,
0.08489222079515457,
-0.14767906069755554,
-0.034961946308612823,
-0.024681802839040756,
-0.10828366130590439,
0.10121457278728485,
-0.0812772884964943,
0.1019766703248024,
-0.010560473427176476,
0.03032168187201023,
0.007048462517559528,
-0.0584542490541935,
0.13420188426971436,
-0.05781706050038338,
0.114717036485672,
-0.07107487320899963,
-0.023039180785417557,
0.03284776583313942,
-0.05530534312129021,
0.031263161450624466,
-0.0827709510922432,
0.044913340359926224,
-0.03325982391834259,
-0.01232141349464655,
-0.06196151301264763,
0.03352374583482742,
-0.02384762652218342,
-0.03897285461425781,
-0.045314572751522064,
0.04078703373670578,
0.016635864973068237,
-0.010282821021974087,
0.18079203367233276,
0.0364624448120594,
0.10614590346813202,
0.11650462448596954,
0.045148443430662155,
-0.07142512500286102,
-0.04624462127685547,
-0.013295036740601063,
-0.038917455822229385,
0.08725384622812271,
-0.15835973620414734,
0.03720584511756897,
0.10083437711000443,
0.0004163091944064945,
0.1382332444190979,
0.05619203299283981,
-0.02652691677212715,
0.02193383313715458,
0.08557406067848206,
-0.15971975028514862,
-0.1336013376712799,
-0.02840835228562355,
-0.021013814955949783,
-0.09196456521749496,
0.0517796128988266,
0.11750270426273346,
-0.09257565438747406,
0.012205960229039192,
-0.0035705415066331625,
0.021229663863778114,
-0.01761971041560173,
0.16488003730773926,
0.0826575830578804,
0.03703774884343147,
-0.07939039915800095,
0.08933428674936295,
0.06586338579654694,
-0.11102009564638138,
-0.0026346591766923666,
0.03451653569936752,
-0.09860394150018692,
-0.0534595362842083,
0.05745626240968704,
0.16573460400104523,
-0.019246814772486687,
-0.04749401658773422,
-0.14345254004001617,
-0.10374907404184341,
0.04222876578569412,
0.15168900787830353,
0.0872710794210434,
-0.003922241274267435,
-0.00980400387197733,
0.014992525801062584,
-0.1134660467505455,
0.1248055249452591,
0.02179286628961563,
0.10713569074869156,
-0.2159658521413803,
0.08200620114803314,
0.010479098185896873,
0.027549833059310913,
-0.024375002831220627,
0.05661387741565704,
-0.10923871397972107,
-0.01041620783507824,
-0.10486246645450592,
0.010391434654593468,
-0.013766221702098846,
0.009569471701979637,
0.0034949928522109985,
-0.07372334599494934,
-0.06467819958925247,
0.019709721207618713,
-0.09937705099582672,
-0.03438059985637665,
0.05165758728981018,
0.052864719182252884,
-0.10256284475326538,
-0.044540345668792725,
0.027313802391290665,
-0.06613892316818237,
0.06556759029626846,
0.004576642531901598,
0.02401791699230671,
0.04253925755620003,
-0.15717516839504242,
0.010957987047731876,
0.0737895518541336,
0.014484528452157974,
0.04710585996508598,
-0.07805876433849335,
0.0005261367186903954,
-0.014755732379853725,
0.03827034309506416,
-0.0029256725683808327,
0.12521186470985413,
-0.1280849426984787,
-0.010214880108833313,
-0.030833197757601738,
-0.05189375951886177,
-0.048300981521606445,
0.04160019010305405,
0.07505190372467041,
-0.005057074595242739,
0.21092668175697327,
-0.09911557286977768,
0.0010294284438714385,
-0.21269825100898743,
0.005849240347743034,
-0.01635023020207882,
-0.12514054775238037,
-0.1293465793132782,
-0.04219939559698105,
0.054471779614686966,
-0.07021959125995636,
0.0897296592593193,
0.03978227078914642,
0.04136349633336067,
0.03909298777580261,
-0.005412998143583536,
0.01886283978819847,
0.030922068282961845,
0.18488085269927979,
0.003951858263462782,
-0.026957836002111435,
0.029606427997350693,
0.02895059622824192,
0.12494699656963348,
0.07432689517736435,
0.12052099406719208,
0.158183291554451,
-0.05486864596605301,
0.11155582219362259,
0.06236451491713524,
-0.05341944098472595,
-0.16064715385437012,
0.06599103659391403,
-0.08984194695949554,
0.1270543485879898,
-0.03161422535777092,
0.18494565784931183,
0.10318756848573685,
-0.14270152151584625,
0.0016963576199486852,
-0.07137976586818695,
-0.07215830683708191,
-0.07362056523561478,
-0.09578045457601547,
-0.10970981419086456,
-0.17377439141273499,
0.010629448108375072,
-0.08310059458017349,
0.005046343896538019,
0.10107998549938202,
-0.008042685687541962,
-0.027723506093025208,
0.21887221932411194,
0.03622277453541756,
0.01728392392396927,
0.05712415277957916,
0.012830018997192383,
-0.05240478739142418,
-0.05857173353433609,
-0.08773774653673172,
0.026066672056913376,
-0.025883937254548073,
0.014154992997646332,
-0.04907750338315964,
-0.03690871596336365,
0.053404875099658966,
0.02252829447388649,
-0.11582762002944946,
0.016229696571826935,
0.017182666808366776,
0.02486554905772209,
0.02892487868666649,
-0.0010271413484588265,
0.038169015198946,
-0.007680388633161783,
0.21005405485630035,
-0.06754778325557709,
-0.03170780465006828,
-0.09634329378604889,
0.17323388159275055,
0.003957875072956085,
0.005770971532911062,
0.007013963535428047,
-0.09652459621429443,
0.029683837667107582,
0.2346685230731964,
0.15382707118988037,
-0.08395355194807053,
-0.007586681749671698,
-0.007199040148407221,
-0.016148190945386887,
-0.0512663796544075,
0.1075073853135109,
0.10340624302625656,
-0.027314404025673866,
-0.07169105112552643,
-0.041397806257009506,
-0.051854558289051056,
-0.0008105550659820437,
-0.02988404594361782,
0.02631174400448799,
0.02849878929555416,
0.02699587307870388,
-0.07420384883880615,
0.059380121529102325,
0.011177068576216698,
-0.08895054459571838,
0.08721030503511429,
-0.19735345244407654,
-0.14562664926052094,
-0.027130136266350746,
0.14509205520153046,
-0.01613767445087433,
0.03442130610346794,
-0.05105980858206749,
0.014624156057834625,
0.05160563066601753,
-0.02533918246626854,
-0.06576118618249893,
-0.09990835189819336,
0.0421198233962059,
-0.11588603258132935,
0.26565900444984436,
-0.0392838791012764,
0.017511874437332153,
0.11695042252540588,
0.0266891960054636,
-0.10328477621078491,
0.08103692531585693,
0.034441716969013214,
-0.07164473086595535,
0.039233602583408356,
0.10328523814678192,
-0.033797211945056915,
0.12982456386089325,
0.03896244987845421,
-0.11112165451049805,
0.00999599788337946,
-0.057903848588466644,
-0.06060374900698662,
-0.049703359603881836,
-0.023396383970975876,
-0.07057115435600281,
0.13335832953453064,
0.16843360662460327,
-0.02842838503420353,
-0.014729839749634266,
-0.058793146163225174,
0.02356390655040741,
0.08464354276657104,
0.03628180921077728,
-0.02241545543074608,
-0.21631188690662384,
0.02464842051267624,
0.022245818749070168,
-0.001133415731601417,
-0.24539954960346222,
-0.07971504330635071,
-0.018693502992391586,
-0.05165579169988632,
-0.09345772862434387,
0.09261204302310944,
0.13297678530216217,
0.04697401076555252,
-0.0696195736527443,
-0.07840447127819061,
-0.06435853242874146,
0.15703271329402924,
-0.12127187103033066,
-0.10557734966278076
] |
null | null | keras |
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
| Hyperparameters | Value |
| :-- | :-- |
| name | Adam |
| weight_decay | None |
| clipnorm | None |
| global_clipnorm | None |
| clipvalue | None |
| use_ema | False |
| ema_momentum | 0.99 |
| ema_overwrite_frequency | None |
| jit_compile | True |
| is_legacy_optimizer | False |
| learning_rate | 0.0010000000474974513 |
| beta_1 | 0.9 |
| beta_2 | 0.999 |
| epsilon | 1e-07 |
| amsgrad | False |
| training_precision | float32 |
## Model Plot
<details>
<summary>View Model Plot</summary>

</details> | {"library_name": "keras", "tags": ["binary-classification", "keras", "tensorflow"]} | null | shuvom/churn-cl-v1 | [
"keras",
"binary-classification",
"tensorflow",
"region:us"
] | 2024-02-08T07:16:03+00:00 | [] | [] | TAGS
#keras #binary-classification #tensorflow #region-us
| Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
Model Plot
----------
View Model Plot
!Model Image
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n\nModel Plot\n----------\n\n\n\nView Model Plot\n!Model Image"
] | [
"TAGS\n#keras #binary-classification #tensorflow #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n\nModel Plot\n----------\n\n\n\nView Model Plot\n!Model Image"
] | [
19,
28
] | [
"passage: TAGS\n#keras #binary-classification #tensorflow #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n\nModel Plot\n----------\n\n\n\nView Model Plot\n!Model Image"
] | [
-0.020976699888706207,
0.04339887574315071,
-0.00347478361800313,
0.04186830297112465,
0.18364962935447693,
0.07889861613512039,
0.13544806838035583,
0.09911441057920456,
-0.017392907291650772,
-0.0008269893005490303,
0.09771728515625,
0.06912746280431747,
-0.003348276950418949,
0.044903144240379333,
-0.014624292030930519,
-0.308478444814682,
-0.04603499174118042,
-0.017174776643514633,
-0.15209613740444183,
0.12202247977256775,
0.06733229011297226,
-0.14481426775455475,
0.0443350113928318,
-0.013460823334753513,
-0.2763465642929077,
0.04463745653629303,
0.028024420142173767,
-0.0422818697988987,
0.12503308057785034,
-0.026350362226366997,
0.20722006261348724,
-0.004760674200952053,
0.08437477052211761,
-0.16042041778564453,
0.03908838331699371,
0.08308286964893341,
-0.01143887359648943,
0.07809092849493027,
0.00707343639805913,
-0.01039188727736473,
0.07481582462787628,
-0.06268767267465591,
0.07831122726202011,
-0.025203518569469452,
-0.18158157169818878,
-0.10515434294939041,
-0.050785940140485764,
-0.011201271787285805,
0.09499586373567581,
0.026654206216335297,
-0.016634438186883926,
0.14070181548595428,
-0.1487990766763687,
0.021589355543255806,
0.0457296259701252,
-0.17724932730197906,
-0.07402556389570236,
0.16294139623641968,
0.0018299245275557041,
0.061850786209106445,
-0.11651346832513809,
0.09811359643936157,
0.08193445205688477,
-0.007684560026973486,
0.04796018823981285,
-0.035622380673885345,
-0.10662500560283661,
0.06633047014474869,
-0.14900502562522888,
0.04771034047007561,
0.03018615022301674,
-0.012173064984381199,
0.038162749260663986,
0.02482299879193306,
-0.043473754078149796,
-0.007947084493935108,
-0.01818023808300495,
-0.058693189173936844,
0.07351483404636383,
0.031102977693080902,
0.03302711248397827,
-0.0840931236743927,
-0.11549296230077744,
-0.038068987429142,
-0.09488770365715027,
0.05232490226626396,
-0.01660342514514923,
0.045807357877492905,
-0.10450389236211777,
0.05981200933456421,
-0.16418369114398956,
-0.06989593803882599,
0.13556544482707977,
-0.0386301726102829,
-0.07728628814220428,
-0.09203220903873444,
-0.026737315580248833,
-0.16642218828201294,
0.03584803640842438,
-0.09381130337715149,
0.09123142808675766,
0.04803230240941048,
-0.05531369894742966,
0.07789035141468048,
0.035457342863082886,
0.09545939415693283,
-0.06812775135040283,
-0.04295938089489937,
-0.014034119434654713,
0.018557462841272354,
-0.0325775109231472,
-0.01618916355073452,
-0.1086934506893158,
0.01885971426963806,
0.007424151059240103,
0.03895605355501175,
-0.10487020015716553,
0.034095775336027145,
-0.0440635159611702,
-0.010542242787778378,
-0.09544094651937485,
-0.05181053653359413,
0.06706230342388153,
-0.022917738184332848,
-0.05668174475431442,
0.04466162994503975,
-0.023174000903964043,
0.05546892061829567,
0.02577051892876625,
0.11734239757061005,
-0.07816632091999054,
0.06499101966619492,
-0.13996416330337524,
-0.09226977825164795,
-0.033741407096385956,
-0.04684107005596161,
-0.015356622636318207,
-0.014791152440011501,
-0.1489420384168625,
-0.0038705002516508102,
0.10788162797689438,
-0.049265529960393906,
0.09018735587596893,
-0.03983752802014351,
-0.03470483794808388,
0.0020076700020581484,
0.042635273188352585,
0.136612206697464,
-0.016618426889181137,
0.048164840787649155,
-0.04953908547759056,
0.09570032358169556,
-0.03552189841866493,
0.02196585386991501,
-0.07745871692895889,
0.04825618118047714,
-0.30636611580848694,
0.0656021237373352,
-0.043245021253824234,
0.09010763466358185,
-0.053183481097221375,
-0.052786935120821,
-0.026303069666028023,
0.035041190683841705,
0.012160114012658596,
0.13841606676578522,
-0.37941688299179077,
-0.05009089782834053,
0.14461460709571838,
-0.09033933281898499,
-0.003149570431560278,
0.0383162759244442,
-0.14762362837791443,
0.0928216204047203,
0.0905904546380043,
0.3739500343799591,
0.047747902572155,
-0.08404933661222458,
0.11640554666519165,
-0.03342849761247635,
-0.06984854489564896,
0.06432339549064636,
-0.012607910670340061,
-0.01763041317462921,
-0.01508027222007513,
-0.03160706162452698,
-0.04414394497871399,
0.02590661123394966,
-0.10654427111148834,
-0.08837227523326874,
0.00212616054341197,
-0.06504666805267334,
0.1009315773844719,
0.03557851165533066,
0.06961978226900101,
-0.09130870550870895,
0.012674078345298767,
0.13291750848293304,
0.05043890327215195,
-0.02731998823583126,
-0.03933367878198624,
-0.04344717040657997,
-0.039580218493938446,
-0.03140132501721382,
-0.048966217786073685,
-0.11501944065093994,
-0.08534162491559982,
-0.014193050563335419,
0.09232660382986069,
0.04475674033164978,
0.09084991365671158,
0.09681137651205063,
0.03377064689993858,
-0.05189516022801399,
0.03938402980566025,
0.020420609042048454,
0.016952108591794968,
-0.12675179541110992,
-0.2219688594341278,
0.025368403643369675,
-0.058774128556251526,
0.08979596197605133,
-0.2883453071117401,
0.00013780857261735946,
-0.019871706143021584,
0.08759915828704834,
0.06307289749383926,
-0.013263409025967121,
-0.000777311681304127,
0.009137425571680069,
-0.019374115392565727,
-0.0591982863843441,
0.062317054718732834,
-0.07529879361391068,
-0.06003748998045921,
-0.007170591503381729,
-0.0732424333691597,
0.025422709062695503,
0.12205282598733902,
-0.26069608330726624,
-0.1186583936214447,
-0.04629600793123245,
-0.027865860611200333,
0.005748738069087267,
-0.036092422902584076,
0.01787232607603073,
0.13725852966308594,
-0.03517475724220276,
0.08439279347658157,
0.02653331868350506,
-0.007678791414946318,
-0.00878026895225048,
-0.07433411478996277,
0.014860281720757484,
0.02094302885234356,
0.1734134405851364,
-0.19146999716758728,
0.10499143600463867,
0.1693447083234787,
-0.03566587716341019,
0.2116820514202118,
0.00639694556593895,
-0.010248973965644836,
-0.053016673773527145,
-0.03725529834628105,
-0.014613013714551926,
0.15833888947963715,
-0.16284774243831635,
-0.015447394922375679,
-0.026719102635979652,
-0.0012141152983531356,
0.05072198063135147,
-0.2325066328048706,
-0.08792052417993546,
0.00453084846958518,
-0.021980801597237587,
0.03828292712569237,
-0.02668709307909012,
-0.028212247416377068,
0.10449839383363724,
-0.013543959707021713,
-0.03829748183488846,
0.024441681802272797,
-0.036821961402893066,
-0.05818618834018707,
0.19310565292835236,
-0.08690603077411652,
-0.06842384487390518,
0.004471332300454378,
-0.009418805129826069,
0.08793004602193832,
0.03527479246258736,
-0.015327684581279755,
-0.1799377202987671,
-0.0019897520542144775,
-0.001951946527697146,
-0.0022071751300245523,
-0.006614843849092722,
0.034449297934770584,
-0.04423307999968529,
0.046549372375011444,
-0.018437378108501434,
-0.07332085818052292,
-0.042768996208906174,
-0.10128790885210037,
-0.03169427067041397,
0.03185906261205673,
0.02057388424873352,
0.05606810748577118,
0.254011869430542,
-0.03248220682144165,
0.0029180175624787807,
-0.010219074785709381,
0.1672016829252243,
-0.12178336828947067,
0.031075486913323402,
0.017279667779803276,
-0.07632483541965485,
0.03425206243991852,
0.11721287667751312,
0.07793683558702469,
-0.11677294969558716,
0.004875375423580408,
0.06695292145013809,
-0.10172716528177261,
-0.19417309761047363,
-0.050135206431150436,
-0.050750572234392166,
-0.11142438650131226,
-0.0051450058817863464,
0.00975918397307396,
0.03860288858413696,
0.042679090052843094,
0.14391067624092102,
0.032789647579193115,
-0.04018842428922653,
-0.0417853482067585,
-0.015638243407011032,
-0.02327808365225792,
0.046219199895858765,
-0.04930045083165169,
-0.11686632037162781,
0.07239110767841339,
-0.013978724367916584,
0.3532918393611908,
0.10599947720766068,
-0.059478435665369034,
0.0742301493883133,
0.01578829064965248,
0.06492779403924942,
0.0025160687509924173,
-0.027243446558713913,
-0.08494585007429123,
-0.05417085811495781,
-0.04964509233832359,
0.034244511276483536,
0.029442446306347847,
0.050243280827999115,
-0.062304701656103134,
-0.032277509570121765,
-0.05688995122909546,
0.14198195934295654,
0.06257567554712296,
0.0868903398513794,
-0.21217410266399384,
0.05310498923063278,
0.058538008481264114,
0.01613224484026432,
0.009115851484239101,
-0.004318295046687126,
0.13067510724067688,
-0.0988268256187439,
0.050249263644218445,
-0.042022041976451874,
0.05241110920906067,
-0.06059795245528221,
0.05097708851099014,
0.04138396680355072,
0.04848507046699524,
-0.046527620404958725,
0.0330158956348896,
-0.19803082942962646,
0.31624677777290344,
0.025926286354660988,
0.05332726985216141,
-0.0655408501625061,
-0.08313293009996414,
0.09500136226415634,
0.09466878324747086,
0.07767107337713242,
0.029992643743753433,
-0.03532020375132561,
-0.0812198668718338,
-0.030454106628894806,
0.007095073349773884,
0.1553030163049698,
-0.04699008911848068,
0.022343436256051064,
0.02246665768325329,
0.023686900734901428,
0.07321157306432724,
0.03235579654574394,
-0.1211652085185051,
-0.01847461611032486,
-0.06065651774406433,
-0.04741901904344559,
-0.15208062529563904,
-0.02561967261135578,
-0.050407540053129196,
-0.01564851962029934,
0.15301428735256195,
0.08552143722772598,
0.011142979376018047,
-0.09722864627838135,
0.11820033937692642,
0.0438498854637146,
0.03344504162669182,
0.03153667226433754,
0.03950516879558563,
0.026272838935256004,
0.016009263694286346,
-0.09664905816316605,
0.14231836795806885,
-0.0390818752348423,
-0.040836770087480545,
-0.1433388888835907,
0.1245153397321701,
0.05353759229183197,
0.05445610359311104,
0.04337039217352867,
0.041381996124982834,
0.05550239607691765,
-0.09340876340866089,
0.14556215703487396,
-0.0014937870437279344,
0.023646749556064606,
0.08518507331609726,
-0.08920841664075851,
-0.04783446341753006,
-0.023236246779561043,
0.016793671995401382,
0.25183263421058655,
0.20838606357574463,
-0.079676054418087,
0.062474511563777924,
0.03181711211800575,
-0.12044250220060349,
-0.21922354400157928,
0.2400149703025818,
0.03860707953572273,
-0.009399968199431896,
0.12490477412939072,
-0.18940269947052002,
0.09990432858467102,
0.12883266806602478,
-0.005139338783919811,
0.14859427511692047,
-0.38455042243003845,
-0.08757661283016205,
-0.0016773889074102044,
0.11347933113574982,
0.15377214550971985,
-0.1635422557592392,
-0.05250401049852371,
-0.06165669485926628,
0.0015270146541297436,
0.19786043465137482,
-0.20038367807865143,
0.11275964975357056,
0.0453936830163002,
0.08607113361358643,
0.021976742893457413,
-0.08246215432882309,
0.19597110152244568,
0.019483361393213272,
0.13048632442951202,
-0.07329300791025162,
-0.21120484173297882,
0.17052416503429413,
0.0009632161818444729,
-0.067224882543087,
0.09275180846452713,
-0.011875002644956112,
-0.1519179344177246,
-0.007754131220281124,
-0.06069333106279373,
0.020869052037596703,
-0.025109391659498215,
-0.09977307915687561,
-0.04515446722507477,
0.07297327369451523,
0.01853739656507969,
0.0046765003353357315,
0.003218834288418293,
-0.01648571528494358,
0.09489856660366058,
0.04269939288496971,
0.06570170074701309,
0.13382939994335175,
-0.0435711033642292,
-0.0010021283524110913,
0.006048016715794802,
0.05505111813545227,
-0.22494417428970337,
0.02557753585278988,
0.1211625188589096,
0.028797760605812073,
0.1477666050195694,
0.09294634312391281,
-0.0184304378926754,
0.08042202889919281,
0.10244657844305038,
-0.17433670163154602,
-0.11051975190639496,
0.039682283997535706,
-0.10612735897302628,
0.049302540719509125,
0.04097043350338936,
0.06221867725253105,
-0.052825264632701874,
0.012633471749722958,
0.02239070273935795,
-0.004165896214544773,
-0.14706507325172424,
0.1875305324792862,
0.135096937417984,
0.017266999930143356,
-0.040082450956106186,
0.08211836963891983,
0.04778926074504852,
-0.02952166460454464,
0.03550301492214203,
0.08201020956039429,
-0.07028252631425858,
-0.05901465564966202,
0.09357257932424545,
0.3533781170845032,
-0.02833661623299122,
0.016645710915327072,
-0.01123801153153181,
-0.0504026934504509,
-0.00574389286339283,
0.12811724841594696,
0.10035686194896698,
0.0008415559423156083,
-0.05166179686784744,
0.04000268131494522,
-0.06671321392059326,
-0.006725423038005829,
-0.0229390487074852,
0.02411779947578907,
-0.07711803168058395,
0.1767762303352356,
0.009068545885384083,
0.0021380700636655092,
-0.09094895422458649,
-0.011676819063723087,
-0.163379967212677,
0.043968524783849716,
-0.08492697030305862,
-0.02962791733443737,
0.05804458633065224,
-0.05157458037137985,
0.0318167582154274,
-0.029445668682456017,
-0.061622489243745804,
-0.038298722356557846,
-0.12289047986268997,
0.01623133383691311,
0.02445175312459469,
-0.005285696592181921,
-0.09054873883724213,
-0.06716682761907578,
0.035849250853061676,
0.0016500905621796846,
0.07073979824781418,
0.02732304483652115,
0.0353386253118515,
0.05910876765847206,
-0.23399023711681366,
-0.05935274809598923,
0.10016173869371414,
-0.022410020232200623,
0.09459580481052399,
0.014593730680644512,
-0.027629053220152855,
-0.06642927974462509,
0.07776354998350143,
0.05570153519511223,
0.11795026808977127,
-0.08175278455018997,
-0.0006842946168035269,
-0.025104494765400887,
-0.15980882942676544,
-0.01261658500880003,
0.010057945735752583,
0.0477244071662426,
0.04081112891435623,
0.10754472017288208,
-0.05083534121513367,
0.06009272113442421,
-0.16282455623149872,
-0.03764565661549568,
-0.036345917731523514,
-0.049172475934028625,
0.039531320333480835,
-0.0006397510878741741,
0.04320160672068596,
-0.10791771113872528,
0.22123174369335175,
0.1419113427400589,
0.015243827365338802,
0.033310920000076294,
-0.020387038588523865,
-0.02745671756565571,
0.09325937926769257,
0.23595763742923737,
0.050473615527153015,
-0.038238245993852615,
-0.01804281957447529,
0.09704475849866867,
0.05107864737510681,
0.028350409120321274,
0.1843218356370926,
0.10277441143989563,
-0.060872633010149,
0.03882038965821266,
0.037039656192064285,
-0.010920085944235325,
-0.10037194192409515,
0.13043472170829773,
-0.022794194519519806,
0.07715292274951935,
-0.0382806658744812,
0.10387621074914932,
0.08545687794685364,
-0.1687387079000473,
0.07523951679468155,
-0.09466032683849335,
-0.11316058784723282,
-0.07370827347040176,
-0.06486910581588745,
-0.050816696137189865,
-0.11011575162410736,
0.02165874093770981,
-0.08040174841880798,
0.05939963459968567,
0.17645764350891113,
0.0369044728577137,
-0.010622300207614899,
0.19723065197467804,
-0.014212717302143574,
0.011388525366783142,
0.10153508931398392,
0.022337481379508972,
-0.046732347458601,
-0.12606090307235718,
-0.02593657560646534,
-0.06619791686534882,
-0.02526526339352131,
-0.0444984957575798,
-0.022433195263147354,
-0.05746745690703392,
-0.027785873040556908,
-0.06626231223344803,
-0.09566786140203476,
0.018755046650767326,
-0.021235423162579536,
0.041104402393102646,
-0.11470507830381393,
0.030971014872193336,
-0.020734170451760292,
-0.017691517248749733,
0.26706764101982117,
-0.062735415995121,
-0.03237578272819519,
-0.09937850385904312,
0.19088950753211975,
-0.03233347833156586,
0.10111202299594879,
-0.06584399193525314,
-0.08785472065210342,
0.010661041364073753,
0.18332532048225403,
0.34375134110450745,
-0.2551864683628082,
-0.014671317301690578,
-0.033769913017749786,
0.005852305795997381,
0.031287696212530136,
0.1898612231016159,
0.03759590908885002,
-0.004532466642558575,
-0.12143553793430328,
-0.04796743392944336,
-0.0678483247756958,
-0.03958548232913017,
-0.03992974013090134,
0.10200723260641098,
0.12811563909053802,
0.04889565333724022,
-0.0886411964893341,
0.18202529847621918,
-0.024926554411649704,
-0.19295907020568848,
0.09307811409235,
-0.2681970000267029,
-0.15629461407661438,
-0.005886706057935953,
0.12827889621257782,
0.0037831515073776245,
0.06872901320457458,
-0.07290210574865341,
-0.08135783672332764,
-0.09530816227197647,
0.01128493994474411,
-0.1085650622844696,
-0.11532306671142578,
0.08123001456260681,
-0.10991034656763077,
0.16048763692378998,
-0.03934657946228981,
0.038544584065675735,
0.08638375997543335,
0.010190047323703766,
0.030724629759788513,
0.011189698241651058,
0.05088271200656891,
-0.028522418811917305,
-0.08300800621509552,
0.0928727239370346,
-0.06282311677932739,
-0.09399949759244919,
0.06573107838630676,
-0.15466558933258057,
0.049245286732912064,
-0.05743793398141861,
-0.09340282529592514,
-0.060611870139837265,
0.021785195916891098,
-0.11926601082086563,
0.08289197087287903,
0.14208561182022095,
0.016423501074314117,
0.09744908660650253,
-0.07712052762508392,
0.024689801037311554,
0.06109236553311348,
0.07083610445261002,
-0.1264457404613495,
-0.15361618995666504,
-0.025847794488072395,
0.11746436357498169,
-0.014401513151824474,
-0.11265918612480164,
-0.05094429850578308,
-0.05951177328824997,
0.008908338844776154,
-0.04463579133152962,
0.0858580470085144,
0.17259129881858826,
0.05977718532085419,
-0.04865558072924614,
-0.20376138389110565,
0.0054877218790352345,
0.12470693141222,
-0.099235400557518,
-0.07186322659254074
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | ssaryssane/ssary-solar-10.7B | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T07:16:04+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
60,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.04654794931411743,
0.16618601977825165,
-0.005445904564112425,
0.01853804849088192,
0.0981811136007309,
0.011998992413282394,
0.06433123350143433,
0.11398410052061081,
-0.0230073444545269,
0.11406639218330383,
0.03047988750040531,
0.10172267258167267,
0.11317981779575348,
0.14841650426387787,
-0.002152352826669812,
-0.22403094172477722,
0.050844956189394,
-0.12105348706245422,
-0.033293843269348145,
0.11749980598688126,
0.1483822613954544,
-0.09928343445062637,
0.07274559140205383,
-0.029687678441405296,
-0.012143402360379696,
-0.030057786032557487,
-0.05890674889087677,
-0.046214159578084946,
0.04651786759495735,
0.06640566885471344,
0.06770290434360504,
0.0071083661168813705,
0.09012923389673233,
-0.2696533799171448,
0.018959321081638336,
0.07145345956087112,
-0.002759667346253991,
0.06957992166280746,
0.06404146552085876,
-0.07107418030500412,
0.10337356477975845,
-0.05106033384799957,
0.14650006592273712,
0.08365883678197861,
-0.09081148356199265,
-0.1895141303539276,
-0.08866965025663376,
0.09882009029388428,
0.17572562396526337,
0.04925641790032387,
-0.02320658043026924,
0.09761467576026917,
-0.08769196271896362,
0.015438909642398357,
0.04981724172830582,
-0.07620415836572647,
-0.05378096550703049,
0.05986575037240982,
0.07907199114561081,
0.06627275794744492,
-0.12434766441583633,
-0.02885502204298973,
0.005009706597775221,
0.010980482213199139,
0.0769270583987236,
0.01728810742497444,
0.146672785282135,
0.0338633768260479,
-0.12615777552127838,
-0.04880760237574577,
0.09869225323200226,
0.03395522013306618,
-0.04422314465045929,
-0.24749068915843964,
-0.03152675926685333,
-0.030810698866844177,
-0.029386121779680252,
-0.03716538846492767,
0.04340358078479767,
-0.007673026993870735,
0.08638741075992584,
-0.0060646249912679195,
-0.07403432577848434,
-0.03937075287103653,
0.06169692054390907,
0.0672287791967392,
0.02999979443848133,
-0.013745363801717758,
0.010938193649053574,
0.11620724946260452,
0.1095694974064827,
-0.12054188549518585,
-0.05555335059762001,
-0.06393084675073624,
-0.08656639605760574,
-0.040790557861328125,
0.034162238240242004,
0.03456587344408035,
0.05349370837211609,
0.25305667519569397,
0.015654386952519417,
0.059652652591466904,
0.034477248787879944,
0.007892133668065071,
0.05848940089344978,
0.11044429242610931,
-0.06018859148025513,
-0.10444226115942001,
-0.02648012898862362,
0.08843598514795303,
0.008199662901461124,
-0.03287925571203232,
-0.05088530853390694,
0.06019928678870201,
0.01946467161178589,
0.11926145106554031,
0.09061790257692337,
0.010536285117268562,
-0.07121123373508453,
-0.061038948595523834,
0.1891259253025055,
-0.16544590890407562,
0.04322727024555206,
0.035097137093544006,
-0.03903156518936157,
0.00019933005387429148,
0.013914269395172596,
0.016625655815005302,
-0.025983380153775215,
0.09017423540353775,
-0.054113563150167465,
-0.04145489260554314,
-0.11186197400093079,
-0.03383193537592888,
0.033762916922569275,
0.008953776210546494,
-0.035059962421655655,
-0.033713940531015396,
-0.08351044356822968,
-0.07577689737081528,
0.09320491552352905,
-0.07346344739198685,
-0.04878907650709152,
-0.01804324984550476,
-0.07530532777309418,
0.022395428270101547,
0.019394835457205772,
0.07707412540912628,
-0.02362251654267311,
0.04399976506829262,
-0.05189276114106178,
0.05863580107688904,
0.11207318305969238,
0.03570080175995827,
-0.05736649036407471,
0.06062258034944534,
-0.23834340274333954,
0.09552820026874542,
-0.07409077137708664,
0.05591456592082977,
-0.153293639421463,
-0.024439791217446327,
0.04788333550095558,
0.008784620091319084,
-0.009650949388742447,
0.13416339457035065,
-0.21702027320861816,
-0.02536402828991413,
0.1717337965965271,
-0.10057014971971512,
-0.07069246470928192,
0.05619903281331062,
-0.04835370555520058,
0.10988964140415192,
0.03825836628675461,
-0.025690359994769096,
0.06171267107129097,
-0.1267417073249817,
0.003717758459970355,
-0.05005312338471413,
-0.017048977315425873,
0.1548657864332199,
0.07182947546243668,
-0.07217690348625183,
0.07399354875087738,
0.025708531960844994,
-0.0246540866792202,
-0.04625825211405754,
-0.015164627693593502,
-0.10536660254001617,
0.014689887873828411,
-0.06369215250015259,
0.014470234513282776,
-0.020807426422834396,
-0.09071163833141327,
-0.027962757274508476,
-0.17504668235778809,
-0.03014434315264225,
0.08651752024888992,
-0.008693269453942776,
-0.01803150773048401,
-0.1178668737411499,
0.009341353550553322,
0.04177580401301384,
0.0061247628182172775,
-0.13462838530540466,
-0.04812471568584442,
0.02780051715672016,
-0.1600649207830429,
0.034652888774871826,
-0.05392369255423546,
0.04932025074958801,
0.025790516287088394,
-0.028889117762446404,
-0.026493212208151817,
0.021633783355355263,
0.005992184858769178,
-0.011999987065792084,
-0.24343903362751007,
-0.028118690475821495,
-0.024888472631573677,
0.1682123839855194,
-0.20917098224163055,
0.03546025976538658,
0.07867541164159775,
0.15366052091121674,
0.011240328662097454,
-0.04177491366863251,
0.005974748637527227,
-0.06935794651508331,
-0.02736494317650795,
-0.05875484645366669,
-0.0047869328409433365,
-0.03310677409172058,
-0.04545191675424576,
0.04568447172641754,
-0.16510973870754242,
-0.032636504620313644,
0.09776268899440765,
0.06289951503276825,
-0.13922683894634247,
-0.020621931180357933,
-0.03630133345723152,
-0.049253206700086594,
-0.04911839962005615,
-0.0605199858546257,
0.10893940925598145,
0.05891856551170349,
0.04574795812368393,
-0.05928509309887886,
-0.07568105310201645,
-0.001827909960411489,
-0.013898161239922047,
-0.017864689230918884,
0.09759635478258133,
0.0751434788107872,
-0.13251115381717682,
0.09224759042263031,
0.09603385627269745,
0.07919023185968399,
0.09113933145999908,
-0.02355697751045227,
-0.08261934667825699,
-0.045987509191036224,
0.031442027539014816,
0.020124373957514763,
0.13039541244506836,
-0.024294709786772728,
0.04352088272571564,
0.042134687304496765,
-0.019369594752788544,
0.014752166345715523,
-0.08687400817871094,
0.033972494304180145,
0.028472330421209335,
-0.016721390187740326,
0.050190530717372894,
-0.03876714035868645,
0.02440318465232849,
0.08830609917640686,
0.045322712510824203,
0.03507532551884651,
0.015493292361497879,
-0.05206458270549774,
-0.1083620935678482,
0.16405931115150452,
-0.12714070081710815,
-0.22483378648757935,
-0.13936103880405426,
0.0037376401014626026,
0.035628627985715866,
-0.015835661441087723,
0.002417160663753748,
-0.059374887496232986,
-0.12220635265111923,
-0.08858037739992142,
0.015140829607844353,
0.04942670464515686,
-0.09028962254524231,
-0.06437795609235764,
0.058117836713790894,
0.03889724239706993,
-0.14560972154140472,
0.017612040042877197,
0.04854894429445267,
-0.09789852797985077,
-0.006774199660867453,
0.08094939589500427,
0.0698540136218071,
0.1770169734954834,
0.017703235149383545,
-0.021850809454917908,
0.032354529947042465,
0.20614571869373322,
-0.13538233935832977,
0.11083246022462845,
0.13607586920261383,
-0.09041404724121094,
0.08072979003190994,
0.19951270520687103,
0.03932560607790947,
-0.10153959691524506,
0.031980328261852264,
0.02283124253153801,
-0.0284719280898571,
-0.24526868760585785,
-0.07212468236684799,
-0.004402178805321455,
-0.058010730892419815,
0.07660572230815887,
0.09286724030971527,
0.08215958625078201,
0.012304253876209259,
-0.09310996532440186,
-0.08154371380805969,
0.05942574888467789,
0.10367169976234436,
0.024584239348769188,
-0.010839897207915783,
0.08998730033636093,
-0.034100502729415894,
0.019626356661319733,
0.0853661298751831,
0.005239574704319239,
0.17840281128883362,
0.05159219726920128,
0.18830420076847076,
0.07925192266702652,
0.07219027727842331,
0.009912233799695969,
0.013080619275569916,
0.018877580761909485,
0.03300119563937187,
-0.002769160782918334,
-0.08440786600112915,
-0.02248465269804001,
0.11566436290740967,
0.06668911874294281,
0.010815348476171494,
0.015172341838479042,
-0.04104290530085564,
0.07965951412916183,
0.1831512451171875,
-0.007656289264559746,
-0.1783534437417984,
-0.057547420263290405,
0.07553383708000183,
-0.09879875183105469,
-0.09854305535554886,
-0.013454320840537548,
0.03072015568614006,
-0.17046253383159637,
0.023390959948301315,
-0.02239842526614666,
0.1106182336807251,
-0.14194999635219574,
-0.020490378141403198,
0.07218493521213531,
0.07199500501155853,
0.004729843698441982,
0.05758659541606903,
-0.16417601704597473,
0.10671813786029816,
0.008950476534664631,
0.06779605895280838,
-0.09610627591609955,
0.1008887067437172,
-0.004196076653897762,
-0.02063460275530815,
0.1393408179283142,
0.002700034761801362,
-0.06884108483791351,
-0.0763031542301178,
-0.08754398673772812,
-0.009632662869989872,
0.12754282355308533,
-0.1419651061296463,
0.08767123520374298,
-0.037212442606687546,
-0.0424150750041008,
-0.0017086371080949903,
-0.10206665843725204,
-0.11638247221708298,
-0.18888559937477112,
0.06001543253660202,
-0.13492922484874725,
0.03152317553758621,
-0.10799519717693329,
-0.032371897250413895,
-0.030304040759801865,
0.19337286055088043,
-0.23447458446025848,
-0.07199826091527939,
-0.1475764364004135,
-0.10233612358570099,
0.1443224400281906,
-0.0501345656812191,
0.08485390990972519,
-0.007241467013955116,
0.16846685111522675,
0.019060896709561348,
-0.02531743235886097,
0.0971490666270256,
-0.09173708409070969,
-0.19302815198898315,
-0.07869284600019455,
0.15662524104118347,
0.13260218501091003,
0.031680017709732056,
-0.002461588243022561,
0.036563750356435776,
-0.015421539545059204,
-0.11935004591941833,
0.015969349071383476,
0.1787186712026596,
0.06237189099192619,
0.02331034652888775,
-0.027346095070242882,
-0.11273157596588135,
-0.06900003552436829,
-0.028530338779091835,
0.03054865077137947,
0.17762407660484314,
-0.07057618349790573,
0.18207968771457672,
0.14163152873516083,
-0.05922834202647209,
-0.20400173962116241,
0.010538800619542599,
0.03055560030043125,
0.0009220078936778009,
0.02591954916715622,
-0.20123432576656342,
0.08688826113939285,
0.004683020059019327,
-0.05110127478837967,
0.13194532692432404,
-0.17217805981636047,
-0.14451217651367188,
0.0765485092997551,
0.038384392857551575,
-0.19559739530086517,
-0.12913893163204193,
-0.09174312651157379,
-0.045869920402765274,
-0.18591414391994476,
0.09569250047206879,
0.0305706188082695,
0.010893458500504494,
0.03030681423842907,
0.029179483652114868,
0.019487828016281128,
-0.0418255440890789,
0.18391458690166473,
-0.024792250245809555,
0.026594700291752815,
-0.08539514988660812,
-0.06927408277988434,
0.03743394836783409,
-0.052842434495687485,
0.07349982857704163,
-0.023486759513616562,
0.007861839607357979,
-0.10348054021596909,
-0.042148489505052567,
-0.03735732287168503,
0.015448716469109058,
-0.09657872468233109,
-0.08514349907636642,
-0.045032672584056854,
0.09675803780555725,
0.09690850973129272,
-0.033646680414676666,
-0.028050623834133148,
-0.07533035427331924,
0.04412057250738144,
0.19926515221595764,
0.1785389482975006,
0.042153384536504745,
-0.08034496754407883,
-0.004150947090238333,
-0.010121207684278488,
0.04310847446322441,
-0.20463712513446808,
0.06283636391162872,
0.05450061708688736,
0.01973269321024418,
0.11436162889003754,
-0.019565396010875702,
-0.15359151363372803,
-0.07263088971376419,
0.06303015351295471,
-0.060181066393852234,
-0.19620554149150848,
0.00867035984992981,
0.060603946447372437,
-0.16371412575244904,
-0.04535605385899544,
0.04643881320953369,
-0.005620351992547512,
-0.038163937628269196,
0.021896906197071075,
0.09194854646921158,
0.0026654244866222143,
0.07427921891212463,
0.05387866869568825,
0.0827430784702301,
-0.10537070035934448,
0.08090532571077347,
0.08839722722768784,
-0.08452684432268143,
0.023530138656497,
0.10478579998016357,
-0.059433579444885254,
-0.03440561518073082,
0.020135708153247833,
0.08153781294822693,
0.01775863952934742,
-0.040019966661930084,
0.013229827396571636,
-0.10452935844659805,
0.05954122915863991,
0.08839859813451767,
0.032507482916116714,
0.016702456399798393,
0.03425082191824913,
0.04607953503727913,
-0.07238735258579254,
0.12142276018857956,
0.031868141144514084,
0.017129309475421906,
-0.036505792289972305,
-0.040896978229284286,
0.019542274996638298,
-0.03214648738503456,
-0.005015232600271702,
-0.03023446537554264,
-0.07695909589529037,
-0.014793801121413708,
-0.1626158058643341,
-0.011131818406283855,
-0.05648450180888176,
0.010329355485737324,
0.03204665705561638,
-0.032609567046165466,
0.008124498650431633,
0.009250079281628132,
-0.07695289701223373,
-0.0663459524512291,
-0.020460480824112892,
0.09540658444166183,
-0.16213038563728333,
0.022481130436062813,
0.08244425803422928,
-0.12187694013118744,
0.09281346201896667,
0.016204802319407463,
-0.006236857734620571,
0.025038830935955048,
-0.1475188434123993,
0.034843120723962784,
-0.03386561945080757,
0.010836300440132618,
0.04373383894562721,
-0.21569781005382538,
-0.00004886732858722098,
-0.033673107624053955,
-0.06639216095209122,
-0.009451326914131641,
-0.03672455996274948,
-0.11508306115865707,
0.1058407872915268,
0.007236586883664131,
-0.08753558248281479,
-0.03186136856675148,
0.029325377196073532,
0.0838974118232727,
-0.021959776058793068,
0.15145497024059296,
-0.008370938710868359,
0.07429654151201248,
-0.16209737956523895,
-0.018623165786266327,
-0.006028574425727129,
0.022658247500658035,
-0.01664556935429573,
-0.01111356820911169,
0.044031109660863876,
-0.022746501490473747,
0.17925859987735748,
-0.030318550765514374,
0.02272745408117771,
0.06815794110298157,
0.019072026014328003,
-0.030184008181095123,
0.10406795144081116,
0.04094860330224037,
0.02014910988509655,
0.018591465428471565,
0.003289656015112996,
-0.04647882282733917,
-0.03173251822590828,
-0.19407226145267487,
0.07288651913404465,
0.15608493983745575,
0.09729263186454773,
-0.016707008704543114,
0.07954329252243042,
-0.10199416428804398,
-0.1109243705868721,
0.12477338314056396,
-0.04797708988189697,
-0.002418199321255088,
-0.07150927931070328,
0.13247236609458923,
0.1437523066997528,
-0.1859612911939621,
0.07269313186407089,
-0.0699717253446579,
-0.04708027467131615,
-0.10980689525604248,
-0.19441905617713928,
-0.05561789125204086,
-0.049456022679805756,
-0.016053348779678345,
-0.04698808491230011,
0.07504211366176605,
0.054538097232580185,
0.006766852922737598,
-0.0023397188633680344,
0.06506035476922989,
-0.031050674617290497,
-0.0037882844917476177,
0.032597362995147705,
0.06591679900884628,
0.012734474614262581,
-0.030802709981799126,
0.016619903966784477,
-0.013545602560043335,
0.045626189559698105,
0.06578011065721512,
0.04976864159107208,
-0.02938537672162056,
0.014603170566260815,
-0.038539156317710876,
-0.10249634087085724,
0.043612558394670486,
-0.024421939626336098,
-0.0789753645658493,
0.15477414429187775,
0.023680059239268303,
0.007779473438858986,
-0.020137663930654526,
0.23901568353176117,
-0.0738423764705658,
-0.0964353010058403,
-0.14737580716609955,
0.10557299107313156,
-0.038081806153059006,
0.05800395458936691,
0.04625935107469559,
-0.10226529091596603,
0.018044332042336464,
0.1338089406490326,
0.16182038187980652,
-0.039008259773254395,
0.020095856860280037,
0.031135575845837593,
0.00566398398950696,
-0.03622615709900856,
0.04847532883286476,
0.06906453520059586,
0.16569648683071136,
-0.04632584750652313,
0.09100406616926193,
0.0019041687482967973,
-0.09579581767320633,
-0.038361791521310806,
0.11069868505001068,
-0.016052277758717537,
0.019335128366947174,
-0.05818064883351326,
0.11742528527975082,
-0.06386786699295044,
-0.23783175647258759,
0.06453443318605423,
-0.0684293657541275,
-0.13765870034694672,
-0.02378307841718197,
0.08207765966653824,
-0.012955902144312859,
0.027587108314037323,
0.0730307325720787,
-0.07240920513868332,
0.201939657330513,
0.03798431158065796,
-0.05499868467450142,
-0.055047210305929184,
0.0805421993136406,
-0.10008571296930313,
0.2739645540714264,
0.01557221356779337,
0.04601577669382095,
0.10384146869182587,
-0.009341772645711899,
-0.13838784396648407,
0.019836371764540672,
0.09581108391284943,
-0.10502193123102188,
0.04196618124842644,
0.19815568625926971,
-0.0014755994779989123,
0.12389086186885834,
0.07657600939273834,
-0.07551808655261993,
0.0478031262755394,
-0.08054235577583313,
-0.06760486960411072,
-0.09260394424200058,
0.09703279286623001,
-0.07772123068571091,
0.14251399040222168,
0.13876807689666748,
-0.05074559152126312,
0.012724342755973339,
-0.031311117112636566,
0.044293127954006195,
-0.00010600237874314189,
0.10321761667728424,
0.004272161517292261,
-0.1832672357559204,
0.024692710489034653,
0.005650998093187809,
0.10749758034944534,
-0.16033467650413513,
-0.09566054493188858,
0.042343202978372574,
0.003505636239424348,
-0.0672195628285408,
0.1290110945701599,
0.05665452033281326,
0.04342988133430481,
-0.03997718170285225,
-0.03521440550684929,
-0.0060732318088412285,
0.13561366498470306,
-0.10713256150484085,
0.0009933578548952937
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | automatic-speech-recognition | SpideyDLK/wav2vec2-large-xls-r-300m-sinhala-test2 | [
"transformers",
"tensorboard",
"safetensors",
"wav2vec2",
"automatic-speech-recognition",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-08T07:16:59+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #tensorboard #safetensors #wav2vec2 #automatic-speech-recognition #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #tensorboard #safetensors #wav2vec2 #automatic-speech-recognition #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
51,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #wav2vec2 #automatic-speech-recognition #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06918960809707642,
0.13210147619247437,
-0.0040207370184361935,
0.023134203627705574,
0.11738458275794983,
0.003100133500993252,
0.06489233672618866,
0.1062328964471817,
-0.018454808741807938,
0.11934409290552139,
0.02399194799363613,
0.10645237565040588,
0.10633884370326996,
0.1783033311367035,
-0.006676932331174612,
-0.20753470063209534,
0.05159076303243637,
-0.1328369528055191,
-0.006210802122950554,
0.12206359207630157,
0.12859149277210236,
-0.12210913747549057,
0.0661126896739006,
-0.03582390025258064,
-0.006673390511423349,
-0.036393795162439346,
-0.05692959576845169,
-0.05386972799897194,
0.06668701767921448,
0.062350641936063766,
0.060644831508398056,
0.018570519983768463,
0.09333337843418121,
-0.2811316251754761,
0.022820472717285156,
0.08144836127758026,
0.006955916993319988,
0.06573166698217392,
0.07077054679393768,
-0.07532189786434174,
0.07820203900337219,
-0.06897217780351639,
0.14961306750774384,
0.07984381169080734,
-0.09012829512357712,
-0.1924186497926712,
-0.08871616423130035,
0.0939040556550026,
0.18705229461193085,
0.05621805414557457,
-0.031164970248937607,
0.13427454233169556,
-0.06805921345949173,
0.01878679171204567,
0.0681581050157547,
-0.07620836049318314,
-0.052718792110681534,
0.06274411827325821,
0.07032535970211029,
0.09331566095352173,
-0.13174302875995636,
-0.00581010989844799,
0.02852778322994709,
0.010611804202198982,
0.10465876758098602,
0.019570648670196533,
0.12078016996383667,
0.03880659490823746,
-0.14213255047798157,
-0.04347489774227142,
0.08553854376077652,
0.040430404245853424,
-0.053023193031549454,
-0.25508078932762146,
-0.01728249341249466,
-0.03535711020231247,
-0.03508080542087555,
-0.050225600600242615,
0.04455358535051346,
-0.02446228265762329,
0.07527779787778854,
-0.005772874690592289,
-0.07288429886102676,
-0.049391333013772964,
0.07827333360910416,
0.07604440301656723,
0.027127955108880997,
-0.02607012540102005,
0.012724175117909908,
0.11518751829862595,
0.11186537146568298,
-0.11170512437820435,
-0.052296292036771774,
-0.06167195737361908,
-0.09369548410177231,
-0.047245271503925323,
0.03096519224345684,
0.04075217619538307,
0.05507682263851166,
0.21005180478096008,
0.004100160673260689,
0.05025365948677063,
0.030208947136998177,
0.013425402343273163,
0.06431768089532852,
0.09155748784542084,
-0.0652301162481308,
-0.12225554138422012,
-0.02715214155614376,
0.10966562479734421,
0.009606653824448586,
-0.03282571956515312,
-0.04075070470571518,
0.0665077269077301,
0.030208082869648933,
0.12366250902414322,
0.0723525807261467,
0.018685176968574524,
-0.07855737954378128,
-0.06267400830984116,
0.1677972972393036,
-0.1649521440267563,
0.03285328298807144,
0.02912791818380356,
-0.050073519349098206,
-0.008440917357802391,
0.01682254858314991,
0.021022414788603783,
-0.018704243004322052,
0.08882031589746475,
-0.054653100669384,
-0.03264474496245384,
-0.11321555823087692,
-0.05006399378180504,
0.028676055371761322,
0.006981914862990379,
-0.03174450621008873,
-0.04053306579589844,
-0.10819326341152191,
-0.07601769268512726,
0.07845603674650192,
-0.06794282793998718,
-0.04567456990480423,
-0.03693155571818352,
-0.077850341796875,
0.013987138867378235,
-0.001372430007904768,
0.11866221576929092,
-0.028359893709421158,
0.049781348556280136,
-0.06040623039007187,
0.07331450283527374,
0.1427365392446518,
0.027582714334130287,
-0.05536656826734543,
0.05209227278828621,
-0.22961750626564026,
0.10650996118783951,
-0.0820845440030098,
0.039568543434143066,
-0.16523221135139465,
-0.01437871903181076,
0.04151884838938713,
0.02703598327934742,
-0.011580551974475384,
0.13367699086666107,
-0.20120634138584137,
-0.03629620373249054,
0.17902998626232147,
-0.11463885754346848,
-0.08275967836380005,
0.05660289525985718,
-0.05534304678440094,
0.12154120951890945,
0.04968025162816048,
-0.015457268804311752,
0.02872299961745739,
-0.14586561918258667,
-0.015341621823608875,
-0.06385710090398788,
-0.031775522977113724,
0.15648432075977325,
0.058627333492040634,
-0.05283202603459358,
0.06168147549033165,
0.01965263858437538,
-0.018219612538814545,
-0.04959159716963768,
-0.03271770104765892,
-0.09723224490880966,
0.011255990713834763,
-0.0728980302810669,
0.023943135514855385,
-0.031872402876615524,
-0.09092787653207779,
-0.03651702031493187,
-0.15960368514060974,
0.006672970950603485,
0.09574975073337555,
-0.005800875835120678,
-0.02275932766497135,
-0.11338774859905243,
-0.010310402140021324,
0.020829740911722183,
-0.0006964936037547886,
-0.14685183763504028,
-0.05314113572239876,
0.017828308045864105,
-0.16250769793987274,
0.031012238934636116,
-0.03655901551246643,
0.04738416150212288,
0.03556562215089798,
-0.03982981666922569,
-0.03375418856739998,
0.019630931317806244,
0.022369354963302612,
-0.010214408859610558,
-0.2756194770336151,
-0.015468244440853596,
-0.043052829802036285,
0.16435527801513672,
-0.2469322234392166,
0.04182727262377739,
0.07295827567577362,
0.1338571161031723,
0.015705497935414314,
-0.03647774085402489,
0.028713135048747063,
-0.06289805471897125,
-0.030222538858652115,
-0.06501726806163788,
-0.007188703399151564,
-0.039097823202610016,
-0.04806915298104286,
0.04462466016411781,
-0.16899824142456055,
-0.033922191709280014,
0.1186266764998436,
0.04557104408740997,
-0.15134701132774353,
-0.04948775842785835,
-0.04092395305633545,
-0.056753676384687424,
-0.06932670623064041,
-0.0517798475921154,
0.10663432627916336,
0.05747092142701149,
0.05196038633584976,
-0.05911761149764061,
-0.06484735757112503,
0.00799498613923788,
-0.01853559911251068,
-0.023748042061924934,
0.07913291454315186,
0.06702018529176712,
-0.11829525977373123,
0.09312599897384644,
0.08573136478662491,
0.07933273166418076,
0.10508506000041962,
-0.0014733473071828485,
-0.09117123484611511,
-0.025300826877355576,
0.029316658154129982,
0.016105778515338898,
0.14908336102962494,
-0.04350128397345543,
0.04314031824469566,
0.040114615112543106,
-0.01687462255358696,
0.008028145879507065,
-0.09918303042650223,
0.030367493629455566,
0.026081476360559464,
-0.012195796705782413,
0.041467417031526566,
-0.05302301421761513,
0.021834537386894226,
0.10195169597864151,
0.03181454911828041,
0.04113520681858063,
0.011278065852820873,
-0.050533477216959,
-0.11812540888786316,
0.17222443222999573,
-0.10861039906740189,
-0.2369978129863739,
-0.12320686131715775,
-0.01618431694805622,
0.02991701476275921,
-0.015134924091398716,
0.01900940015912056,
-0.06770696491003036,
-0.11834623664617538,
-0.09672471135854721,
0.04564153030514717,
0.06599046289920807,
-0.08051323890686035,
-0.055777665227651596,
0.06501153111457825,
0.048011794686317444,
-0.13664643466472626,
0.02571168728172779,
0.03327706828713417,
-0.08857693523168564,
0.00793769583106041,
0.08559047430753708,
0.06839455664157867,
0.18071474134922028,
0.01134483702480793,
-0.023087946698069572,
0.017521869391202927,
0.19720622897148132,
-0.14027054607868195,
0.10202740132808685,
0.13801661133766174,
-0.07145930081605911,
0.07873693108558655,
0.2032429575920105,
0.039016321301460266,
-0.10376140475273132,
0.039679598063230515,
0.036421533674001694,
-0.025852223858237267,
-0.24745285511016846,
-0.08099643886089325,
0.00836301501840353,
-0.0664474293589592,
0.0802333801984787,
0.08307429403066635,
0.09203000366687775,
0.023238254711031914,
-0.1043974831700325,
-0.07363210618495941,
0.05418974906206131,
0.11036353558301926,
-0.004034504294395447,
-0.011317858472466469,
0.09753942489624023,
-0.020273780450224876,
0.02676866576075554,
0.08875394612550735,
0.012205728329718113,
0.18836407363414764,
0.050518929958343506,
0.14771167933940887,
0.09208200126886368,
0.053752463310956955,
0.016467519104480743,
0.010000402107834816,
0.017887894064188004,
0.02435637265443802,
-0.014350295066833496,
-0.08589190989732742,
-0.006933859083801508,
0.1298609972000122,
0.027646880596876144,
0.04127250239253044,
0.013248836621642113,
-0.04125351831316948,
0.08765199780464172,
0.17516882717609406,
0.013442369177937508,
-0.20506484806537628,
-0.06488820165395737,
0.0686659887433052,
-0.08813467621803284,
-0.10374542325735092,
-0.021716099232435226,
0.04023343697190285,
-0.1762947142124176,
0.02770446240901947,
-0.025082001462578773,
0.0983029454946518,
-0.12493812292814255,
-0.01920684240758419,
0.0476171039044857,
0.06939635425806046,
-0.018209589645266533,
0.0625329241156578,
-0.17832936346530914,
0.13725855946540833,
0.012600419111549854,
0.07603015750646591,
-0.0920197069644928,
0.0829358845949173,
0.010243658907711506,
-0.008985995315015316,
0.14880549907684326,
-0.002428766805678606,
-0.056611087173223495,
-0.10275979340076447,
-0.09291432052850723,
-0.01180565357208252,
0.11795864999294281,
-0.11873860657215118,
0.09995509684085846,
-0.017298342660069466,
-0.043639615178108215,
0.0016699014231562614,
-0.12897762656211853,
-0.1380222588777542,
-0.17400150001049042,
0.041601065546274185,
-0.12252611666917801,
0.04249255359172821,
-0.10634490847587585,
-0.05313412845134735,
-0.058118730783462524,
0.19448153674602509,
-0.2263878583908081,
-0.07106572389602661,
-0.1503530591726303,
-0.06515897810459137,
0.11819497495889664,
-0.042735762894153595,
0.08508200198411942,
0.017862383276224136,
0.19214710593223572,
0.010283242911100388,
-0.013114631175994873,
0.10883224755525589,
-0.10211063176393509,
-0.21299202740192413,
-0.10015871375799179,
0.13945214450359344,
0.13517092168331146,
0.038856618106365204,
0.002108179498463869,
0.030881604179739952,
-0.006152692716568708,
-0.11462404578924179,
0.028862472623586655,
0.18585458397865295,
0.10306477546691895,
0.03526908904314041,
-0.03260820358991623,
-0.14471980929374695,
-0.08779244124889374,
-0.045098960399627686,
0.017435450106859207,
0.19264571368694305,
-0.07120641320943832,
0.17354503273963928,
0.15474873781204224,
-0.053835928440093994,
-0.20943360030651093,
0.03015606477856636,
0.036211419850587845,
0.0007652041967958212,
0.05587008595466614,
-0.19489167630672455,
0.0909743532538414,
0.0033501458819955587,
-0.057322751730680466,
0.12121490389108658,
-0.17501963675022125,
-0.15013514459133148,
0.07031099498271942,
0.07301220297813416,
-0.17921873927116394,
-0.12142012268304825,
-0.09439031779766083,
-0.04026462882757187,
-0.11460573226213455,
0.07970702648162842,
-0.016233494505286217,
0.010252374224364758,
0.032961323857307434,
0.018216567113995552,
0.010428756475448608,
-0.04740371182560921,
0.1864585429430008,
-0.003947122488170862,
0.04788469523191452,
-0.07597782462835312,
-0.06253167986869812,
0.045070283114910126,
-0.06455249339342117,
0.0716865211725235,
-0.00903246272355318,
0.006079745013266802,
-0.1052967831492424,
-0.06088602915406227,
-0.03328738734126091,
0.02272024378180504,
-0.07930614799261093,
-0.09432698786258698,
-0.03726235777139664,
0.10006307810544968,
0.09058371931314468,
-0.03892482817173004,
-0.06462740153074265,
-0.08978539705276489,
0.028800709173083305,
0.21877005696296692,
0.177296444773674,
0.05685123801231384,
-0.066028892993927,
-0.00540707865729928,
-0.01588953658938408,
0.053271859884262085,
-0.2026120126247406,
0.0566285103559494,
0.035300228744745255,
0.033545590937137604,
0.11711569130420685,
-0.026464059948921204,
-0.16407892107963562,
-0.048686347901821136,
0.05304291099309921,
-0.07358507066965103,
-0.17289869487285614,
0.014132710173726082,
0.07088939845561981,
-0.1477956771850586,
-0.023786291480064392,
0.04775075986981392,
-0.017420068383216858,
-0.03159533068537712,
0.006238185800611973,
0.08124099671840668,
0.01671770215034485,
0.09224288910627365,
0.053469255566596985,
0.09704500436782837,
-0.10683690756559372,
0.06699982285499573,
0.07745448499917984,
-0.10474617779254913,
0.03967198729515076,
0.0603945255279541,
-0.06895622611045837,
-0.03619396686553955,
0.033563096076250076,
0.08692663908004761,
0.04178347438573837,
-0.060071151703596115,
0.0073408023454248905,
-0.10486608743667603,
0.06092875450849533,
0.1210157498717308,
0.04285310208797455,
0.0076990588568151,
0.036018576472997665,
0.04045969620347023,
-0.09288305044174194,
0.12451037764549255,
0.04114879295229912,
0.028287222608923912,
-0.05418051406741142,
-0.028997255489230156,
0.03649618849158287,
-0.03188192844390869,
-0.01566455140709877,
-0.04152749106287956,
-0.06663620471954346,
-0.010323094204068184,
-0.16889281570911407,
0.006573607679456472,
-0.05270812287926674,
0.008401375263929367,
0.021295055747032166,
-0.03304858133196831,
0.005127503536641598,
0.019244063645601273,
-0.07131489366292953,
-0.052214257419109344,
-0.006754601374268532,
0.10161449760198593,
-0.17169132828712463,
0.014349433593451977,
0.0744767114520073,
-0.12469461560249329,
0.08815638720989227,
0.018520260229706764,
0.0005999338463880122,
0.03465453162789345,
-0.13307695090770721,
0.043367430567741394,
-0.006723123602569103,
0.011691853404045105,
0.048354603350162506,
-0.21661832928657532,
-0.0025545719545334578,
-0.04856108874082565,
-0.055710889399051666,
-0.006375120021402836,
-0.02562650851905346,
-0.11432337760925293,
0.10399775207042694,
0.010540200397372246,
-0.0755159854888916,
-0.02542583830654621,
0.037674929946660995,
0.0969945415854454,
-0.03298725560307503,
0.16065140068531036,
-0.01863807439804077,
0.06254526972770691,
-0.1797095239162445,
-0.018202031031250954,
-0.01975269988179207,
0.023043567314743996,
-0.03248249739408493,
-0.008440588600933552,
0.05180126056075096,
-0.023841936141252518,
0.20870842039585114,
-0.022057142108678818,
0.033427316695451736,
0.06674833595752716,
-0.021141132339835167,
-0.02877473458647728,
0.1086326614022255,
0.054397158324718475,
0.012029323726892471,
0.03175004944205284,
0.006914193741977215,
-0.04090225324034691,
-0.004564614500850439,
-0.1556052416563034,
0.07673801481723785,
0.17203287780284882,
0.0805397778749466,
-0.00828546192497015,
0.06094660609960556,
-0.11003988236188889,
-0.11399497091770172,
0.10722645372152328,
-0.05822233483195305,
-0.014757114462554455,
-0.05772337689995766,
0.14011409878730774,
0.15646083652973175,
-0.19130073487758636,
0.06022409349679947,
-0.06736859679222107,
-0.04819837212562561,
-0.10633485019207001,
-0.17335662245750427,
-0.061282314360141754,
-0.0583864226937294,
-0.01613355241715908,
-0.05076048895716667,
0.06713438034057617,
0.08348768949508667,
0.02054755762219429,
0.016258614137768745,
0.0817527249455452,
-0.02199946530163288,
0.007656866684556007,
0.034995537251234055,
0.06331320106983185,
0.0073803807608783245,
-0.04667557775974274,
0.009565448388457298,
0.0006085589993745089,
0.035281602293252945,
0.04957476258277893,
0.037472013384103775,
-0.026353945955634117,
0.007689491845667362,
-0.02916470356285572,
-0.11019428819417953,
0.04115133360028267,
-0.026625385507941246,
-0.06341774761676788,
0.1439228504896164,
0.031860120594501495,
-0.008713874034583569,
-0.025656426325440407,
0.25211021304130554,
-0.07529866695404053,
-0.08892348408699036,
-0.1387489140033722,
0.13557645678520203,
-0.031552400439977646,
0.06481313705444336,
0.037692490965127945,
-0.11259825527667999,
0.03179538995027542,
0.1362704634666443,
0.1458069533109665,
-0.049145035445690155,
0.019655266776680946,
0.013711978681385517,
0.0032459446229040623,
-0.04005579650402069,
0.04973040521144867,
0.06590425968170166,
0.12457112967967987,
-0.05082963407039642,
0.08012272417545319,
-0.0028764382004737854,
-0.10040896385908127,
-0.02852385863661766,
0.12230420112609863,
-0.003029873361811042,
0.019506774842739105,
-0.0761401429772377,
0.12728425860404968,
-0.043905097991228104,
-0.2665610611438751,
0.06613168120384216,
-0.0650629922747612,
-0.14912083745002747,
-0.022557994350790977,
0.05126400291919708,
-0.008650023490190506,
0.026705266907811165,
0.06785756349563599,
-0.0670214518904686,
0.18420551717281342,
0.03873218223452568,
-0.05507900193333626,
-0.058854296803474426,
0.07306438684463501,
-0.09833692759275436,
0.2929907441139221,
0.00751500902697444,
0.05993965268135071,
0.09920700639486313,
-0.029096059501171112,
-0.13847678899765015,
0.031734831631183624,
0.08172675222158432,
-0.07410130649805069,
0.055994872003793716,
0.21827135980129242,
-0.008840959519147873,
0.11804516613483429,
0.07454971224069595,
-0.09561564773321152,
0.05016838759183884,
-0.10613930225372314,
-0.09673135727643967,
-0.08329153805971146,
0.09532807767391205,
-0.05763502046465874,
0.14755868911743164,
0.1186022087931633,
-0.04606860503554344,
0.02281493879854679,
-0.018614748492836952,
0.048749152570962906,
0.0023650694638490677,
0.12439922988414764,
0.020209291949868202,
-0.19710010290145874,
0.026845410466194153,
-0.008902255445718765,
0.10291280597448349,
-0.2202581763267517,
-0.09718955308198929,
0.04764820635318756,
0.0019112902227789164,
-0.05895697697997093,
0.12370198965072632,
0.055919989943504333,
0.04170476272702217,
-0.04714735969901085,
-0.028212912380695343,
-0.002841046778485179,
0.16146929562091827,
-0.11127673834562302,
0.0008471902110613883
] |
null | null | transformers | # MiquMaid-v2-70B 2.3bpw
## Description
Exllama quant of [NeverSleep/MiquMaid-v2-70B](https://huggingface.co/NeverSleep/MiquMaid-v2-70B)
## Other quants:
EXL2: [4bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-4bpw-exl2), [3.5bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-3.5bpw-exl2), [3bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-3bpw-exl2), [2.4bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-2.4bpw-exl2), [2.3bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-2.3bpw-exl2)
2.4bpw is probably the most you can fit in a 24gb card
GGUF:
[2bit Imatrix GGUF](https://huggingface.co/Kooten/MiquMaid-v2-70B-Imatrix-GGUF)
## Prompt format: Alpaca
```
### Instruction:
{system prompt}
### Input:
{input}
### Response:
{reply}
```
## Contact
Kooten on discord
[ko-fi.com/kooten](https://ko-fi.com/kooten) | {"license": "cc-by-nc-4.0", "tags": ["not-for-all-audiences", "nsfw"]} | text-generation | Kooten/MiquMaid-v2-70B-2.3bpw-exl2 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"not-for-all-audiences",
"nsfw",
"conversational",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T07:19:49+00:00 | [] | [] | TAGS
#transformers #safetensors #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # MiquMaid-v2-70B 2.3bpw
## Description
Exllama quant of NeverSleep/MiquMaid-v2-70B
## Other quants:
EXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw
2.4bpw is probably the most you can fit in a 24gb card
GGUF:
2bit Imatrix GGUF
## Prompt format: Alpaca
## Contact
Kooten on discord
URL | [
"# MiquMaid-v2-70B 2.3bpw",
"## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B",
"## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF",
"## Prompt format: Alpaca",
"## Contact\nKooten on discord\n\nURL"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# MiquMaid-v2-70B 2.3bpw",
"## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B",
"## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF",
"## Prompt format: Alpaca",
"## Contact\nKooten on discord\n\nURL"
] | [
75,
14,
21,
60,
8,
7
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# MiquMaid-v2-70B 2.3bpw## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF## Prompt format: Alpaca## Contact\nKooten on discord\n\nURL"
] | [
-0.05462871864438057,
0.03968195989727974,
-0.005613294430077076,
0.08636385947465897,
0.025682266801595688,
0.02309575304389,
0.14916153252124786,
0.12353517860174179,
0.0012638714397326112,
0.010506183840334415,
0.047497447580099106,
0.11128713935613632,
0.04715196043252945,
0.16344191133975983,
-0.05409565195441246,
-0.09677650779485703,
0.029876993969082832,
0.031548578292131424,
0.07817255705595016,
0.11553706228733063,
0.0658516064286232,
-0.03080986626446247,
0.05545118451118469,
-0.045411936938762665,
-0.10771387070417404,
-0.010061044245958328,
-0.026401611045002937,
-0.06609615683555603,
0.04724499210715294,
0.08093319833278656,
0.03973068669438362,
0.10466296225786209,
-0.032508958131074905,
-0.1831509917974472,
0.027630116790533066,
0.03857865184545517,
-0.061096105724573135,
0.0267206821590662,
0.030109457671642303,
-0.08377084136009216,
0.0065761748701334,
0.005248923785984516,
-0.06772267818450928,
0.07283280044794083,
-0.09085682779550552,
-0.07497118413448334,
-0.049437809735536575,
0.1174791231751442,
0.07380672544240952,
0.027549488469958305,
-0.0027588405646383762,
0.09631910920143127,
-0.011719651520252228,
0.08221626281738281,
0.23221468925476074,
-0.25614553689956665,
-0.01979360170662403,
0.0890415832400322,
0.026884563267230988,
0.09330549091100693,
-0.07278577238321304,
0.03233679383993149,
0.024116825312376022,
-0.028410155326128006,
0.07978902757167816,
-0.020944666117429733,
0.061588484793901443,
-0.03237708657979965,
-0.08920841664075851,
0.008524853736162186,
0.14593851566314697,
0.038776643574237823,
-0.06243681162595749,
-0.05040939524769783,
-0.08031398802995682,
-0.10788575559854507,
-0.04940611124038696,
0.04408194124698639,
0.010442037135362625,
-0.011499276384711266,
-0.04487871751189232,
-0.0016920858761295676,
-0.10727041214704514,
-0.04793059453368187,
-0.12919725477695465,
0.1742779016494751,
0.007015387527644634,
0.029429687187075615,
0.009926079772412777,
0.03476538509130478,
-0.08510544151067734,
-0.08646319061517715,
-0.08939836174249649,
-0.06029944866895676,
0.06816742569208145,
0.0006137778982520103,
0.0060682594776153564,
0.08153722435235977,
0.08979740738868713,
0.14426957070827484,
-0.0704265683889389,
0.08852799236774445,
0.007694023661315441,
0.06494704633951187,
-0.0303485207259655,
-0.11917492747306824,
-0.017188720405101776,
-0.05255572125315666,
0.053298965096473694,
0.030588636174798012,
0.08478690683841705,
-0.022324586287140846,
-0.10261798650026321,
0.004418323747813702,
-0.0005285231163725257,
0.041180528700351715,
0.01837020553648472,
0.05679277330636978,
-0.05014805868268013,
-0.02483997866511345,
0.18683604896068573,
-0.06382892280817032,
0.019208135083317757,
0.05785587430000305,
0.028209103271365166,
0.046868324279785156,
0.06624944508075714,
-0.01025584526360035,
-0.023176806047558784,
0.051996294409036636,
-0.030430644750595093,
0.016438469290733337,
-0.01693536527454853,
-0.05762525647878647,
0.08360191434621811,
-0.0402279794216156,
-0.021577052772045135,
-0.18609608709812164,
-0.15185877680778503,
0.0892271026968956,
-0.006130347028374672,
-0.04489123448729515,
0.01642603799700737,
0.006869292818009853,
-0.03160620480775833,
-0.004961084108799696,
-0.051594555377960205,
0.016852615401148796,
-0.07684045284986496,
0.08750616759061813,
0.08656437695026398,
0.13206596672534943,
-0.11523281037807465,
-0.017806435003876686,
-0.045896016061306,
0.04772521182894707,
-0.08400578796863556,
-0.0024859292898327112,
-0.12114755064249039,
-0.02985631301999092,
-0.06567087769508362,
-0.029246525838971138,
-0.05078355222940445,
0.045153841376304626,
-0.009391160681843758,
0.09400949627161026,
-0.06958302110433578,
-0.042647525668144226,
0.26255661249160767,
-0.14600145816802979,
-0.10760916024446487,
0.11409296840429306,
0.05329018831253052,
-0.06403595954179764,
0.046564288437366486,
0.08476655930280685,
0.13308656215667725,
-0.11144395917654037,
-0.10333004593849182,
0.06986727565526962,
-0.009180082939565182,
-0.09095463901758194,
0.0901874378323555,
0.07936210185289383,
0.009030201472342014,
0.029110323637723923,
0.13304628431797028,
0.02607777900993824,
0.009620578959584236,
-0.058235108852386475,
-0.06468204408884048,
-0.08666034787893295,
0.08282995969057083,
-0.01714908704161644,
-0.040114808827638626,
-0.1246294230222702,
-0.0957668349146843,
0.013373511843383312,
0.07090982049703598,
-0.018171891570091248,
-0.016591599211096764,
-0.13972999155521393,
0.11216485500335693,
-0.10984674096107483,
0.02836526557803154,
-0.047588132321834564,
-0.029098233208060265,
-0.03523528203368187,
-0.023582853376865387,
0.048695243895053864,
0.0683092251420021,
0.09301873296499252,
0.07169021666049957,
-0.09902350604534149,
0.03379334881901741,
0.12266159057617188,
0.025415392592549324,
-0.004210139624774456,
-0.08425669372081757,
0.05573258548974991,
-0.024155626073479652,
0.14293517172336578,
-0.09245239943265915,
0.046327002346515656,
0.15534357726573944,
0.08614455163478851,
0.0037099584005773067,
-0.043744076043367386,
0.06829946488142014,
0.012366159819066525,
-0.04692970961332321,
0.00313390651717782,
0.07419553399085999,
-0.008569195866584778,
-0.009615495800971985,
0.11989374458789825,
-0.2579943537712097,
0.18477946519851685,
0.18155112862586975,
-0.024251218885183334,
-0.031190145760774612,
-0.05466989800333977,
0.014213918708264828,
-0.01914878562092781,
-0.019718367606401443,
-0.06306243687868118,
0.024342920631170273,
-0.022862058132886887,
0.13323643803596497,
-0.10007791221141815,
-0.048873744904994965,
0.05767920985817909,
-0.05574214458465576,
-0.009546788409352303,
0.043473806232213974,
0.07671518623828888,
-0.12288188189268112,
0.14978539943695068,
0.17981182038784027,
-0.07564140111207962,
0.12440188229084015,
0.026164503768086433,
-0.010237730108201504,
-0.034573014825582504,
0.04586231708526611,
0.03651653230190277,
0.07613583654165268,
-0.07134894281625748,
0.05716903507709503,
0.05664889141917229,
-0.013583994470536709,
0.031486596912145615,
-0.1392097920179367,
-0.03454868122935295,
0.04010531306266785,
-0.09133186936378479,
-0.06856313347816467,
0.0029930544551461935,
-0.012123853899538517,
0.09634728729724884,
-0.051172226667404175,
0.0151793472468853,
0.02667129598557949,
0.04594183713197708,
-0.08153323829174042,
0.14566658437252045,
-0.09569376707077026,
-0.3116782605648041,
-0.0847119465470314,
-0.1284981220960617,
-0.05758877843618393,
0.023157233372330666,
0.09245860576629639,
-0.05296801030635834,
-0.04427652060985565,
-0.10441756993532181,
-0.06468174606561661,
-0.006327823735773563,
-0.003543524770066142,
0.013461503200232983,
-0.009649187326431274,
0.07351000607013702,
-0.08925090730190277,
-0.008700436912477016,
0.08201100677251816,
-0.0483907051384449,
0.16451780498027802,
-0.010318737477064133,
0.09112001955509186,
0.1174730658531189,
-0.00532990787178278,
-0.025101320818066597,
-0.03433813899755478,
0.12245602905750275,
-0.05349602550268173,
0.0660184770822525,
0.12651114165782928,
-0.012847972102463245,
0.06820431351661682,
0.08263979107141495,
0.0313698872923851,
-0.0638093501329422,
-0.050279825925827026,
0.017591871321201324,
-0.12045831978321075,
-0.1337604522705078,
-0.04014673829078674,
-0.04213780164718628,
0.15704859793186188,
0.04371748864650726,
0.06263991445302963,
0.0461680069565773,
0.12747672200202942,
-0.06614802032709122,
-0.015180259943008423,
0.06685416400432587,
0.08398330956697464,
0.11973760277032852,
0.024804122745990753,
0.15748140215873718,
-0.08496750891208649,
-0.015092714689671993,
0.11926478892564774,
0.11178149282932281,
0.025954745709896088,
-0.007506332825869322,
0.10218018293380737,
0.07102223485708237,
0.021347589790821075,
0.11311580240726471,
0.07328968495130539,
-0.0026367136742919683,
-0.04651924595236778,
-0.03390103206038475,
-0.06663548946380615,
-0.0007911762222647667,
0.025753671303391457,
-0.0025532294530421495,
-0.04050863906741142,
0.02812269516289234,
-0.034900639206171036,
0.08135387301445007,
0.026569759473204613,
0.08195865154266357,
-0.21843664348125458,
-0.03823065385222435,
0.07766195386648178,
-0.0073872837238013744,
-0.038870785385370255,
0.02819693461060524,
0.0839802548289299,
0.031762074679136276,
0.1666204035282135,
-0.017943214625120163,
0.06954346597194672,
-0.07792740315198898,
0.02270338125526905,
-0.03558695316314697,
0.06303701549768448,
-0.009828463196754456,
0.10124894231557846,
-0.2702677845954895,
0.05406534671783447,
0.03911640867590904,
0.012933675199747086,
-0.0512402206659317,
-0.025828396901488304,
0.012298676185309887,
-0.0057602692395448685,
0.08831238001585007,
0.011912876740098,
0.07159589231014252,
-0.14081422984600067,
-0.08613098412752151,
0.012502682395279408,
0.07563789933919907,
0.05049887299537659,
0.049809254705905914,
0.03506065160036087,
-0.05274786055088043,
-0.002272301120683551,
-0.039415597915649414,
-0.047275226563215256,
-0.1106230691075325,
0.0988166332244873,
0.1548018902540207,
-0.032320261001586914,
-0.023816099390387535,
-0.02496068924665451,
-0.20748291909694672,
0.21939581632614136,
-0.12452900409698486,
-0.08882945775985718,
-0.04944700375199318,
-0.0832827240228653,
-0.001870198524557054,
-0.020995788276195526,
0.03721954673528671,
-0.03974827006459236,
0.11568465083837509,
-0.03733474016189575,
-0.06860154122114182,
0.06377389281988144,
-0.12131496518850327,
-0.1256176382303238,
-0.032848916947841644,
0.14714723825454712,
-0.09806981682777405,
0.02652599848806858,
0.031563229858875275,
-0.00908026471734047,
-0.012118726968765259,
-0.13069817423820496,
0.09224153310060501,
0.023035690188407898,
-0.02026773802936077,
0.019051091745495796,
0.009422855451703072,
-0.044820696115493774,
-0.0038737673312425613,
-0.04855204373598099,
0.11819975823163986,
0.3375561535358429,
-0.033253688365221024,
0.10430677235126495,
0.11823908984661102,
-0.03925016149878502,
-0.23273508250713348,
-0.15197767317295074,
-0.08917058259248734,
-0.08026974648237228,
-0.02618008479475975,
-0.0776616781949997,
0.05277220904827118,
0.12756672501564026,
-0.041231296956539154,
0.19919653236865997,
-0.13271255791187286,
-0.11089973896741867,
0.018949029967188835,
0.11006651818752289,
0.2567412853240967,
-0.17722545564174652,
-0.06200508773326874,
-0.0752960667014122,
-0.17756403982639313,
0.1902105063199997,
-0.14214211702346802,
0.10697191208600998,
-0.02453598938882351,
0.05200010538101196,
-0.015589326620101929,
-0.0803237110376358,
0.11616364866495132,
-0.0658508762717247,
0.00044632505159825087,
-0.12797310948371887,
0.0690278485417366,
0.04081951826810837,
-0.018960893154144287,
0.02049086056649685,
-0.2321067899465561,
-0.00988951325416565,
-0.06565215438604355,
-0.05583958327770233,
-0.038727905601263046,
-0.00023237272398546338,
-0.043735552579164505,
-0.038992878049612045,
-0.047801773995161057,
-0.010120686143636703,
0.013254605233669281,
-0.034426990896463394,
-0.02884223498404026,
-0.0705920085310936,
0.09139885008335114,
0.3066074550151825,
0.0989837571978569,
-0.07417727261781693,
-0.0633426308631897,
-0.06453460454940796,
-0.0971173569560051,
0.007349807303398848,
-0.1491622030735016,
0.060295578092336655,
0.03762001171708107,
0.005227247253060341,
0.06503812223672867,
0.04151170700788498,
-0.03544652462005615,
0.02620224840939045,
0.1514519602060318,
-0.14216360449790955,
-0.024759771302342415,
0.005389028694480658,
0.10857832431793213,
0.006834217347204685,
0.017205974087119102,
0.13893648982048035,
-0.017017720267176628,
0.0167778842151165,
0.047297947108745575,
0.024538926780223846,
-0.0012479564175009727,
0.11745572090148926,
0.03670969605445862,
0.0777256041765213,
-0.11292732506990433,
0.06440133601427078,
-0.03650728613138199,
-0.04162565618753433,
-0.003168125869706273,
0.06716066598892212,
-0.08876374363899231,
-0.12810565531253815,
0.005565598141402006,
-0.04745018482208252,
-0.05508878454566002,
-0.06173035502433777,
-0.03690164536237717,
-0.13767486810684204,
0.028230922296643257,
0.1638973355293274,
0.01357676088809967,
-0.02242996357381344,
0.0844157487154007,
0.03446313738822937,
-0.07712150365114212,
0.08417782932519913,
-0.019561080262064934,
0.09278254210948944,
-0.12609489262104034,
0.04749179631471634,
-0.0016132327727973461,
-0.007301103323698044,
-0.04669464752078056,
0.035576242953538895,
-0.10568154603242874,
-0.03433302417397499,
-0.1491684913635254,
0.10262365639209747,
-0.10474497824907303,
0.002713537309318781,
-0.01652722992002964,
0.012324493378400803,
-0.020932689309120178,
-0.041724491864442825,
-0.06852147728204727,
0.010543332435190678,
0.020175276324152946,
-0.02043377049267292,
-0.10335575044155121,
-0.027218539267778397,
0.025746464729309082,
-0.02979753166437149,
0.05672663450241089,
0.025861550122499466,
-0.05336223542690277,
-0.007038215175271034,
-0.19472067058086395,
-0.01883937045931816,
0.08995362371206284,
0.05368582159280777,
-0.03851822763681412,
0.007796904072165489,
0.04419377073645592,
0.05177628993988037,
0.026168838143348694,
0.0377337820827961,
0.09331557899713516,
-0.06873610615730286,
0.026792455464601517,
-0.0799652710556984,
-0.06387046724557877,
-0.03302504122257233,
-0.013131218031048775,
0.13698022067546844,
-0.06809591501951218,
0.11491597443819046,
-0.07088657468557358,
-0.010772456414997578,
-0.13536681234836578,
0.025279566645622253,
0.033773016184568405,
-0.16981492936611176,
-0.04467831924557686,
-0.032458242028951645,
0.02186114341020584,
-0.03443259000778198,
0.17976628243923187,
-0.08073955774307251,
-0.13388532400131226,
0.03063264675438404,
-0.08720866590738297,
-0.02676357701420784,
-0.009222250431776047,
0.22291381657123566,
0.06203266605734825,
-0.0063768913969397545,
-0.11695804446935654,
0.017065254971385002,
0.07553274184465408,
-0.1263376772403717,
0.08154652267694473,
0.10888063162565231,
-0.08232927322387695,
0.05225128307938576,
0.05267905816435814,
-0.06263347715139389,
-0.06022016331553459,
0.009968842379748821,
-0.1582767814397812,
0.07725675404071808,
-0.0008461984689347446,
0.07696793973445892,
0.11914816498756409,
-0.03652351349592209,
0.0013217931846156716,
-0.05899305269122124,
-0.02301083132624626,
-0.12008960545063019,
-0.058694206178188324,
-0.09711109101772308,
-0.08988483250141144,
0.0170946903526783,
-0.06860893964767456,
0.01091998815536499,
0.05576353520154953,
0.05810963362455368,
0.021176479756832123,
0.14136949181556702,
-0.02093879133462906,
0.0041021620854735374,
-0.011083738878369331,
-0.024960894137620926,
-0.031291790306568146,
0.06412894278764725,
-0.055553846061229706,
0.0652904212474823,
-0.024883359670639038,
0.03777208924293518,
0.031222950667142868,
0.039708737283945084,
0.015064368955790997,
-0.06872664391994476,
-0.09646310657262802,
-0.04163780063390732,
0.03577916696667671,
0.06667640805244446,
0.12303807586431503,
-0.027082694694399834,
0.0007769117364659905,
0.01965719275176525,
0.0820627361536026,
-0.01876005157828331,
-0.10589835792779922,
-0.014882689341902733,
0.15236398577690125,
-0.08563089370727539,
0.06225680932402611,
-0.019777506589889526,
-0.026622522622346878,
0.023485954850912094,
0.24315142631530762,
0.1641516387462616,
-0.03592509776353836,
0.056933674961328506,
0.015109721571207047,
0.019536802545189857,
0.06707826256752014,
0.08861327916383743,
0.0621817409992218,
0.17947249114513397,
-0.027069346979260445,
-0.0028321072459220886,
0.041956085711717606,
-0.0036036004312336445,
-0.07690445333719254,
0.061242856085300446,
-0.023169662803411484,
0.01405516266822815,
-0.0695575699210167,
0.00041224309825338423,
-0.07419848442077637,
0.039684638381004333,
0.0016632917104288936,
-0.10801953077316284,
-0.055633995682001114,
0.003512996481731534,
0.08153297007083893,
0.03434872627258301,
-0.017731865867972374,
-0.020918745547533035,
-0.01529934722930193,
-0.019051745533943176,
-0.01521812193095684,
-0.17024369537830353,
0.06499679386615753,
-0.03513602167367935,
0.0019809079822152853,
0.06183760613203049,
-0.03531571850180626,
0.1325581669807434,
0.10879559814929962,
-0.02639136277139187,
-0.05971791595220566,
0.17516623437404633,
0.006382128689438105,
-0.1746118813753128,
0.04125582426786423,
0.1287490874528885,
-0.04347047582268715,
0.026052162051200867,
0.09089497476816177,
-0.05651354789733887,
-0.04561867564916611,
0.09519978612661362,
-0.06933765113353729,
-0.01193737331777811,
0.03914382681250572,
-0.0352400541305542,
0.07780278474092484,
0.05588942766189575,
-0.03939811512827873,
-0.04485276713967323,
-0.01597669906914234,
0.07273998111486435,
-0.03802225738763809,
-0.05596733093261719,
-0.01691693812608719,
-0.1827869564294815,
-0.04122013971209526,
0.07182426005601883,
0.12928318977355957,
-0.31272661685943604,
-0.012806331738829613,
-0.08651796728372574,
0.0002568863274063915,
-0.1467311978340149,
0.053951580077409744,
0.18526926636695862,
0.03199469670653343,
-0.06653238087892532,
-0.11336275935173035,
-0.015688130632042885,
0.09567563235759735,
-0.09332531690597534,
-0.08264816552400589
] |
null | null | transformers | # MiquMaid-v2-70B-DPO 2.3bpw
## Description
Exllama quant of [NeverSleep/MiquMaid-v2-70B-DPO](https://huggingface.co/NeverSleep/MiquMaid-v2-70B-DPO)
## Other quants:
EXL2: [4bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-4bpw-exl2), [3.5bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-3.5bpw-exl2), [3bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-3bpw-exl2), [2.4bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-2.4bpw-exl2), [2.3bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-2.3bpw-exl2)
2.4bpw is probably the most you can fit in a 24gb card
GGUF:
[2bit Imatrix GGUF](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-Imatrix-GGUF)
## Prompt format: Alpaca
```
### Instruction:
{system prompt}
### Input:
{input}
### Response:
{reply}
```
## Contact
Kooten on discord
[ko-fi.com/kooten](https://ko-fi.com/kooten) | {"license": "cc-by-nc-4.0", "tags": ["not-for-all-audiences", "nsfw"]} | text-generation | Kooten/MiquMaid-v2-70B-DPO-2.3bpw-exl2 | [
"transformers",
"pytorch",
"llama",
"text-generation",
"not-for-all-audiences",
"nsfw",
"conversational",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T07:21:08+00:00 | [] | [] | TAGS
#transformers #pytorch #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # MiquMaid-v2-70B-DPO 2.3bpw
## Description
Exllama quant of NeverSleep/MiquMaid-v2-70B-DPO
## Other quants:
EXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw
2.4bpw is probably the most you can fit in a 24gb card
GGUF:
2bit Imatrix GGUF
## Prompt format: Alpaca
## Contact
Kooten on discord
URL | [
"# MiquMaid-v2-70B-DPO 2.3bpw",
"## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B-DPO",
"## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF",
"## Prompt format: Alpaca",
"## Contact\nKooten on discord\n\nURL"
] | [
"TAGS\n#transformers #pytorch #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# MiquMaid-v2-70B-DPO 2.3bpw",
"## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B-DPO",
"## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF",
"## Prompt format: Alpaca",
"## Contact\nKooten on discord\n\nURL"
] | [
74,
17,
24,
60,
8,
7
] | [
"passage: TAGS\n#transformers #pytorch #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# MiquMaid-v2-70B-DPO 2.3bpw## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B-DPO## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF## Prompt format: Alpaca## Contact\nKooten on discord\n\nURL"
] | [
-0.044714588671922684,
0.07053855806589127,
-0.006472108885645866,
0.08963311463594437,
0.046541228890419006,
0.027045022696256638,
0.12088601291179657,
0.14405423402786255,
0.014738352969288826,
0.006500504445284605,
0.06115616858005524,
0.1293952763080597,
0.06298977881669998,
0.15900544822216034,
-0.039994820952415466,
-0.15927733480930328,
0.01640900783240795,
0.06320153176784515,
0.0956837609410286,
0.12400859594345093,
0.06877097487449646,
-0.024632887914776802,
0.04888942092657089,
-0.031176328659057617,
-0.12982696294784546,
-0.0037487235385924578,
-0.0506441704928875,
-0.06482784450054169,
0.04970645532011986,
0.09308187663555145,
0.02998536266386509,
0.089790940284729,
-0.03316275775432587,
-0.17720210552215576,
0.02257847785949707,
0.032468657940626144,
-0.05208562687039375,
0.03586138039827347,
0.04599694535136223,
-0.04875147342681885,
0.03879562392830849,
0.028177529573440552,
-0.07431315630674362,
0.0729152262210846,
-0.10254853963851929,
-0.08800757676362991,
-0.05403418093919754,
0.1414429098367691,
0.06856164336204529,
0.04026023671030998,
-0.007424216251820326,
0.11284463852643967,
-0.038222525268793106,
0.07409295439720154,
0.22914549708366394,
-0.24434968829154968,
-0.02202177606523037,
0.08194029331207275,
0.03905249014496803,
0.05363250523805618,
-0.06481838971376419,
0.005072473082691431,
0.011236696504056454,
-0.005235403310507536,
0.020380577072501183,
-0.033922139555215836,
0.04958230257034302,
-0.023361684754490852,
-0.092148058116436,
-0.011144550517201424,
0.15172739326953888,
0.06018036603927612,
-0.046841494739055634,
-0.03283561393618584,
-0.07858943194150925,
-0.1273341327905655,
-0.039209939539432526,
0.03246459737420082,
-0.0028519246261566877,
-0.014426540583372116,
-0.04466380923986435,
-0.0004573810438159853,
-0.10552771389484406,
-0.042330529540777206,
-0.1401955634355545,
0.16399845480918884,
0.0013027612585574389,
0.02683587744832039,
0.0038250654470175505,
0.046993646770715714,
-0.08782261610031128,
-0.08559836447238922,
-0.0844353586435318,
-0.08929897844791412,
0.039896927773952484,
-0.0016872873529791832,
-0.009874574840068817,
0.10184069722890854,
0.07948913425207138,
0.17464835941791534,
-0.028718510642647743,
0.07543449103832245,
0.02931916154921055,
0.06542092561721802,
0.003190812421962619,
-0.07868587970733643,
-0.03375476226210594,
-0.04018646851181984,
0.03132643178105354,
0.013874782249331474,
0.045864757150411606,
-0.022642817348241806,
-0.11314968019723892,
-0.020659005269408226,
-0.007755500730127096,
0.021798541769385338,
0.041311681270599365,
0.0518466979265213,
-0.062447480857372284,
-0.06253232806921005,
0.18441244959831238,
-0.04893317446112633,
0.022176625207066536,
0.058921072632074356,
-0.00404808996245265,
0.1027211844921112,
0.04744579643011093,
-0.002559461398050189,
-0.057597555220127106,
0.0551440604031086,
-0.03330235928297043,
0.027849454432725906,
-0.023407474160194397,
-0.04932582750916481,
0.09060642123222351,
-0.061918191611766815,
-0.01466177124530077,
-0.17030565440654755,
-0.13505227863788605,
0.07027854025363922,
-0.003088889177888632,
-0.05849997699260712,
-0.03979896381497383,
0.0024417780805379152,
-0.01822763681411743,
-0.008036751300096512,
-0.060455355793237686,
0.03377052769064903,
-0.06996500492095947,
0.08760222047567368,
0.07477381080389023,
0.12423359602689743,
-0.11942136287689209,
0.001507976558059454,
-0.03883212432265282,
0.02713092416524887,
-0.07426518201828003,
0.017832962796092033,
-0.1143360584974289,
-0.028976278379559517,
-0.08244732022285461,
-0.04963108152151108,
-0.028856903314590454,
0.03800131753087044,
-0.004482584074139595,
0.10808398574590683,
-0.0579034648835659,
-0.05771133676171303,
0.24289356172084808,
-0.09080570936203003,
-0.06684145331382751,
0.12702830135822296,
0.055049557238817215,
-0.08301563560962677,
0.04759910702705383,
0.07158645242452621,
0.11311239004135132,
-0.09917943924665451,
-0.07686859369277954,
0.08124679327011108,
-0.011259030550718307,
-0.0801607072353363,
0.0948600322008133,
0.05091918259859085,
0.03686495125293732,
0.03263501822948456,
0.15713757276535034,
0.032842230051755905,
-0.0026894633192569017,
-0.06811990588903427,
-0.0505693256855011,
-0.061133190989494324,
0.07610224187374115,
0.007235870696604252,
-0.029025916010141373,
-0.0955430418252945,
-0.11065088957548141,
-0.0008554991218261421,
0.07657531648874283,
-0.0035722858738154173,
-0.00687751080840826,
-0.1199212297797203,
0.09803527593612671,
-0.08372977375984192,
0.030327534303069115,
-0.07454391568899155,
0.001842405297793448,
-0.030397696420550346,
-0.00667579984292388,
0.08002500236034393,
0.10323717445135117,
0.07235824316740036,
0.03423256054520607,
-0.07776124030351639,
0.03508930653333664,
0.0942356064915657,
0.0033862662967294455,
-0.004293294157832861,
-0.06750502437353134,
0.027540184557437897,
-0.012848722748458385,
0.11927790939807892,
-0.0780339166522026,
0.025479687377810478,
0.13342560827732086,
0.10253150761127472,
-0.026553625240921974,
-0.034959763288497925,
0.06550068408250809,
0.03188753500580788,
-0.04265051335096359,
0.016608811914920807,
0.08760169893503189,
-0.01807459630072117,
-0.016226330772042274,
0.09340644627809525,
-0.20978322625160217,
0.11592710763216019,
0.17338787019252777,
-0.06286834180355072,
-0.030191119760274887,
-0.06200527399778366,
-0.009397633373737335,
-0.025612125173211098,
0.010581940412521362,
-0.05591259151697159,
0.07372822612524033,
-0.014301981776952744,
0.12630917131900787,
-0.08401960134506226,
-0.053442589938640594,
0.052640896290540695,
-0.04561567306518555,
0.001098610577173531,
0.06777717918157578,
0.10902096331119537,
-0.11511655896902084,
0.15832236409187317,
0.1409941166639328,
-0.08714363723993301,
0.1329921931028366,
0.02742917276918888,
-0.03060336783528328,
-0.03351878747344017,
0.016035187989473343,
0.008075536228716373,
0.05638837814331055,
-0.11611217260360718,
0.05596216395497322,
0.08016131818294525,
-0.01117769069969654,
0.025623200461268425,
-0.13141393661499023,
-0.041388221085071564,
0.020703991875052452,
-0.07004982978105545,
-0.08307671546936035,
-0.002492465078830719,
0.005527769681066275,
0.10200861096382141,
-0.014292672276496887,
-0.002460112562403083,
0.03936536982655525,
0.04267275333404541,
-0.07194366306066513,
0.14058230817317963,
-0.09675328433513641,
-0.30754080414772034,
-0.08008468151092529,
-0.14468052983283997,
-0.08231412619352341,
0.017124861478805542,
0.09344231337308884,
-0.06911429017782211,
-0.031049128621816635,
-0.04818037152290344,
-0.04481799155473709,
-0.0453900471329689,
-0.01138200145214796,
0.011800087988376617,
-0.01709608919918537,
0.057255640625953674,
-0.07991204410791397,
-0.021313998848199844,
0.05823058262467384,
-0.041680093854665756,
0.16690468788146973,
-0.02455214038491249,
0.09032610058784485,
0.1382235288619995,
0.01771741360425949,
-0.01567610539495945,
-0.03264303877949715,
0.15724681317806244,
-0.06558819115161896,
0.04896187037229538,
0.140899196267128,
-0.010204381309449673,
0.07770119607448578,
0.09174099564552307,
0.04033017158508301,
-0.052346374839544296,
-0.052001953125,
0.030335262417793274,
-0.09790725260972977,
-0.14655302464962006,
-0.03747549280524254,
-0.06361988186836243,
0.14526258409023285,
0.055247507989406586,
0.06039135903120041,
0.06819937378168106,
0.11635667830705643,
-0.07641355693340302,
0.0010627417359501123,
0.030577881261706352,
0.09089512377977371,
0.16761288046836853,
0.021745484322309494,
0.15401703119277954,
-0.05987175181508064,
-0.03688066452741623,
0.12736542522907257,
0.13192865252494812,
0.0447261780500412,
0.008578271605074406,
0.10670933872461319,
0.0575421079993248,
0.015538107603788376,
0.09585168957710266,
0.045998990535736084,
0.013541074469685555,
-0.031909361481666565,
-0.04078533127903938,
-0.059278424829244614,
-0.00827273353934288,
0.04253223165869713,
0.07121993601322174,
-0.07730409502983093,
0.03060041554272175,
-0.0871773362159729,
0.06747172772884369,
0.06749456375837326,
0.09035712480545044,
-0.20828305184841156,
-0.02903691865503788,
0.0492992028594017,
-0.027686186134815216,
-0.04108947142958641,
0.019237011671066284,
0.0033983096946030855,
-0.02265339158475399,
0.16490133106708527,
-0.009565490297973156,
0.09337092936038971,
-0.07613679766654968,
0.02411770075559616,
-0.0338226854801178,
0.054226770997047424,
0.0034263692796230316,
0.11376647651195526,
-0.29268181324005127,
0.07485878467559814,
0.02288718894124031,
0.0005749412812292576,
-0.07035329192876816,
-0.00891693402081728,
0.013402647338807583,
-0.0376628078520298,
0.08995207399129868,
0.012298673391342163,
0.11349622160196304,
-0.13562147319316864,
-0.06898555159568787,
0.017747193574905396,
0.059164248406887054,
0.03426086902618408,
0.03861125186085701,
0.049048591405153275,
-0.04077602177858353,
-0.011266120709478855,
-0.001795804244466126,
-0.03881014138460159,
-0.0936688631772995,
0.12383025884628296,
0.1194177195429802,
-0.020970359444618225,
0.013628321699798107,
-0.033023688942193985,
-0.1898318976163864,
0.22514505684375763,
-0.11713287979364395,
-0.09892069548368454,
-0.06058783829212189,
-0.05795872211456299,
0.019991809502243996,
-0.03735589236021042,
0.01892710290849209,
-0.05369899794459343,
0.0896204337477684,
-0.033495981246232986,
-0.08422456681728363,
0.08101078867912292,
-0.10428692400455475,
-0.1338534951210022,
-0.03536622226238251,
0.15668708086013794,
-0.07063838094472885,
0.03965840861201286,
0.019966838881373405,
-0.012665345333516598,
-0.0438053160905838,
-0.12635159492492676,
0.10424704849720001,
0.007329963613301516,
-0.002632468705996871,
0.016515517607331276,
0.017716681584715843,
-0.011487540788948536,
-0.03943308815360069,
-0.03442687541246414,
0.14255033433437347,
0.35915517807006836,
-0.034397631883621216,
0.11800651252269745,
0.11835189163684845,
-0.03979350998997688,
-0.21248693764209747,
-0.16489779949188232,
-0.03939265012741089,
-0.0809093713760376,
-0.05781354382634163,
-0.13845770061016083,
0.04036715626716614,
0.13433125615119934,
-0.034980885684490204,
0.1884947270154953,
-0.1458730846643448,
-0.0987638533115387,
0.04070794954895973,
0.054591789841651917,
0.2429225593805313,
-0.1580864042043686,
-0.059810806065797806,
-0.06034215912222862,
-0.20395290851593018,
0.2224174439907074,
-0.03631289675831795,
0.12511390447616577,
-0.0393255241215229,
0.09837166965007782,
-0.001993323676288128,
-0.05944513529539108,
0.12456733733415604,
-0.024411717429757118,
-0.029704274609684944,
-0.11768456548452377,
0.04352441430091858,
0.006544982548803091,
-0.007476132363080978,
0.010971295647323132,
-0.17724232375621796,
-0.028789140284061432,
-0.13762985169887543,
-0.05964351445436478,
-0.07759278267621994,
0.017694750800728798,
-0.0324295274913311,
-0.05358017235994339,
-0.05304780602455139,
-0.002339698374271393,
0.014565674588084221,
-0.02174234390258789,
-0.012300857342779636,
-0.07258806377649307,
0.07574020326137543,
0.3016442656517029,
0.09257855266332626,
-0.07774484902620316,
-0.09384124726057053,
-0.06768631935119629,
-0.09478170424699783,
0.020963100716471672,
-0.11858249455690384,
0.03639422729611397,
0.07870140671730042,
0.00818940345197916,
0.06286131590604782,
0.04487607628107071,
-0.06004851311445236,
0.01976076327264309,
0.12061292678117752,
-0.13924053311347961,
0.0035109096206724644,
-0.012904836796224117,
0.06426538527011871,
0.0004830602847505361,
-0.010837472975254059,
0.13555209338665009,
-0.007770462427288294,
-0.013880827464163303,
0.05861411988735199,
0.018588431179523468,
-0.020132580772042274,
0.11367584019899368,
0.03427526354789734,
0.07374835014343262,
-0.13012085855007172,
0.06280937790870667,
-0.011710122227668762,
-0.057938285171985626,
-0.009553780779242516,
0.10470626503229141,
-0.06942444294691086,
-0.1281229704618454,
-0.017565634101629257,
-0.08156058937311172,
-0.10000215470790863,
-0.03873807564377785,
-0.024182287976145744,
-0.11852511018514633,
0.06595028936862946,
0.11983434110879898,
0.021855168044567108,
-0.023626646026968956,
0.06423848867416382,
0.040888722985982895,
-0.06889340281486511,
0.05863850936293602,
-0.008758666925132275,
0.08217542618513107,
-0.10086691379547119,
0.05555346608161926,
-0.007061339914798737,
0.0461638867855072,
-0.04201734438538551,
0.010972502641379833,
-0.12206320464611053,
-0.02888454869389534,
-0.1396256387233734,
0.05564318224787712,
-0.11448029428720474,
-0.0042108469642698765,
-0.01952638477087021,
-0.0009550842805765569,
-0.02884681709110737,
-0.03973059728741646,
-0.09098395705223083,
0.0050004697404801846,
0.020015783607959747,
-0.006559096276760101,
-0.09409793466329575,
-0.027230864390730858,
0.028968509286642075,
-0.024230733513832092,
0.07300527393817902,
0.05105408653616905,
-0.05333970859646797,
-0.01904074102640152,
-0.1517544686794281,
-0.03361266851425171,
0.07926446199417114,
0.07087060064077377,
-0.024411149322986603,
0.024184443056583405,
0.04434391111135483,
0.04079227149486542,
0.03555969148874283,
0.030358070507645607,
0.07711880654096603,
-0.09001129120588303,
-0.016469324007630348,
-0.07010436058044434,
-0.1000484898686409,
-0.05186765268445015,
-0.009646378457546234,
0.1260780245065689,
-0.030950384214520454,
0.10655119270086288,
-0.04905868321657181,
0.027633417397737503,
-0.13733385503292084,
0.02048070728778839,
0.029975077137351036,
-0.14781266450881958,
-0.007679146248847246,
-0.056988589465618134,
0.030674370005726814,
-0.02854304574429989,
0.16360093653202057,
-0.07858380675315857,
-0.1033821702003479,
0.028543874621391296,
-0.07002395391464233,
-0.035602178424596786,
-0.01797136850655079,
0.19583596289157867,
0.07752490043640137,
-0.012725140899419785,
-0.0937030166387558,
0.02738000452518463,
0.055913809686899185,
-0.07584147155284882,
0.0886782854795456,
0.08252093940973282,
-0.028579875826835632,
0.04358023405075073,
0.06532984226942062,
-0.0727643370628357,
-0.08326202630996704,
0.00553511269390583,
-0.1215699315071106,
0.10331056267023087,
-0.01954471319913864,
0.0893598422408104,
0.08785654604434967,
-0.04223296046257019,
0.01773318089544773,
-0.06809408217668533,
-0.03367454931139946,
-0.11106155812740326,
-0.08674868196249008,
-0.08289938420057297,
-0.09571036696434021,
0.030324719846248627,
-0.07679879665374756,
0.03228994458913803,
0.059942539781332016,
0.04798736050724983,
-0.015749964863061905,
0.08724889904260635,
-0.016599778085947037,
-0.033341094851493835,
0.010021810419857502,
-0.02546333521604538,
-0.020232195034623146,
0.031406063586473465,
-0.034418266266584396,
0.06700176000595093,
-0.01307188905775547,
0.046294916421175,
0.025574248284101486,
0.01975582167506218,
-0.0037201971281319857,
-0.06899359077215195,
-0.09355884790420532,
-0.034980468451976776,
0.03559292480349541,
0.07919174432754517,
0.15640795230865479,
-0.033441729843616486,
0.015186140313744545,
0.016482267528772354,
0.07712971419095993,
-0.027210939675569534,
-0.08388170599937439,
-0.03649117797613144,
0.1982543021440506,
-0.08536490052938461,
0.05232378467917442,
-0.014084680005908012,
-0.021845277398824692,
-0.009982340969145298,
0.24571792781352997,
0.1558847576379776,
-0.03991904482245445,
0.03136739507317543,
0.044981200248003006,
0.023012852296233177,
0.07660641521215439,
0.0905439704656601,
0.07994207739830017,
0.18143093585968018,
-0.050377506762742996,
0.0053182682022452354,
0.028514400124549866,
0.0007951001753099263,
-0.08011437207460403,
0.059871990233659744,
0.01776491105556488,
-0.01853647455573082,
-0.06392320245504379,
0.021462487056851387,
-0.128445565700531,
0.06996779888868332,
-0.005404614377766848,
-0.11880183964967728,
-0.07562223076820374,
0.00618683872744441,
0.0508873276412487,
0.04571069777011871,
0.012938856147229671,
-0.03052581287920475,
-0.010501690208911896,
0.020472561940550804,
-0.01956459879875183,
-0.17715047299861908,
0.05631408467888832,
0.005441918969154358,
-0.000989497289992869,
0.0295205470174551,
-0.03956368565559387,
0.15653616189956665,
0.10544408112764359,
0.008605197072029114,
-0.050191234797239304,
0.13058732450008392,
0.020257728174328804,
-0.16821886599063873,
0.039294056594371796,
0.1280258446931839,
-0.033985938876867294,
-0.00006575257430085912,
0.0934705138206482,
-0.06436701864004135,
-0.05077739059925079,
0.10798081755638123,
-0.029545823112130165,
-0.01736445166170597,
0.06301417201757431,
-0.05442693084478378,
0.07241501659154892,
0.06543491035699844,
-0.0500086210668087,
-0.07034324109554291,
-0.044883646070957184,
0.05321793258190155,
-0.02452039159834385,
-0.05559171736240387,
-0.057204216718673706,
-0.1698412299156189,
-0.052262261509895325,
0.053311221301555634,
0.1249486654996872,
-0.24341890215873718,
-0.006247243378311396,
-0.0860835537314415,
0.0004986964631825686,
-0.12966571748256683,
0.06122725456953049,
0.09838581085205078,
0.011236891150474548,
-0.05131477862596512,
-0.10252511501312256,
-0.019671428948640823,
0.0791243314743042,
-0.1230047270655632,
-0.07160782068967819
] |
null | null | transformers | # MiquMaid-v2-70B-DPO 4bpw
## Description
Exllama quant of [NeverSleep/MiquMaid-v2-70B-DPO](https://huggingface.co/NeverSleep/MiquMaid-v2-70B-DPO)
## Other quants:
EXL2: [4bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-4bpw-exl2), [3.5bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-3.5bpw-exl2), [3bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-3bpw-exl2), [2.4bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-2.4bpw-exl2), [2.3bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-2.3bpw-exl2)
2.4bpw is probably the most you can fit in a 24gb card
GGUF:
[2bit Imatrix GGUF](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-Imatrix-GGUF)
## Prompt format: Alpaca
```
### Instruction:
{system prompt}
### Input:
{input}
### Response:
{reply}
```
## Contact
Kooten on discord
[ko-fi.com/kooten](https://ko-fi.com/kooten) | {"license": "cc-by-nc-4.0", "tags": ["not-for-all-audiences", "nsfw"]} | text-generation | Kooten/MiquMaid-v2-70B-DPO-4bpw-exl2 | [
"transformers",
"pytorch",
"llama",
"text-generation",
"not-for-all-audiences",
"nsfw",
"conversational",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T07:30:38+00:00 | [] | [] | TAGS
#transformers #pytorch #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # MiquMaid-v2-70B-DPO 4bpw
## Description
Exllama quant of NeverSleep/MiquMaid-v2-70B-DPO
## Other quants:
EXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw
2.4bpw is probably the most you can fit in a 24gb card
GGUF:
2bit Imatrix GGUF
## Prompt format: Alpaca
## Contact
Kooten on discord
URL | [
"# MiquMaid-v2-70B-DPO 4bpw",
"## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B-DPO",
"## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF",
"## Prompt format: Alpaca",
"## Contact\nKooten on discord\n\nURL"
] | [
"TAGS\n#transformers #pytorch #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# MiquMaid-v2-70B-DPO 4bpw",
"## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B-DPO",
"## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF",
"## Prompt format: Alpaca",
"## Contact\nKooten on discord\n\nURL"
] | [
74,
17,
24,
60,
8,
7
] | [
"passage: TAGS\n#transformers #pytorch #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# MiquMaid-v2-70B-DPO 4bpw## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B-DPO## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF## Prompt format: Alpaca## Contact\nKooten on discord\n\nURL"
] | [
-0.04644676670432091,
0.06905650347471237,
-0.006369173061102629,
0.09042017161846161,
0.047347068786621094,
0.026605168357491493,
0.12129374593496323,
0.14315111935138702,
0.014627628959715366,
0.006336309481412172,
0.060429323464632034,
0.1284266710281372,
0.0626446008682251,
0.15729092061519623,
-0.0392300970852375,
-0.1599978357553482,
0.015019461512565613,
0.06435201317071915,
0.09702903777360916,
0.12389644235372543,
0.06778998672962189,
-0.026428906247019768,
0.049677979201078415,
-0.031134648248553276,
-0.13069579005241394,
-0.003331628628075123,
-0.05093158408999443,
-0.06481458991765976,
0.05071186274290085,
0.09342139959335327,
0.030195655301213264,
0.08912765979766846,
-0.03379698097705841,
-0.175348699092865,
0.023106563836336136,
0.03129938989877701,
-0.05267338827252388,
0.03619316592812538,
0.04472493752837181,
-0.048057835549116135,
0.03929080069065094,
0.026945261284708977,
-0.075498066842556,
0.07221364974975586,
-0.10294999182224274,
-0.08929858356714249,
-0.05306236818432808,
0.1406671702861786,
0.0672849714756012,
0.042136646807193756,
-0.008511453866958618,
0.11213847249746323,
-0.03820280358195305,
0.07446464151144028,
0.2272040992975235,
-0.24635003507137299,
-0.021739866584539413,
0.08054785430431366,
0.038457900285720825,
0.05478234589099884,
-0.06685396283864975,
0.005746540613472462,
0.012364517897367477,
-0.005924376193434,
0.01999732293188572,
-0.03361017629504204,
0.048092395067214966,
-0.022363081574440002,
-0.0918666198849678,
-0.0105274124071002,
0.15269674360752106,
0.06020362302660942,
-0.04590100795030594,
-0.032740525901317596,
-0.07875090837478638,
-0.12744921445846558,
-0.039817873388528824,
0.033469054847955704,
-0.0030316186603158712,
-0.015541131608188152,
-0.04544256627559662,
-0.0002928159374278039,
-0.10643107444047928,
-0.041079211980104446,
-0.1410113275051117,
0.16265693306922913,
0.0016911743441596627,
0.026098022237420082,
0.003551999805495143,
0.046803414821624756,
-0.08784632384777069,
-0.08544237166643143,
-0.08285140991210938,
-0.08801098167896271,
0.04040305316448212,
-0.001165185123682022,
-0.009809217415750027,
0.10344649106264114,
0.07855737209320068,
0.17306648194789886,
-0.02941589616239071,
0.07419101893901825,
0.03061412088572979,
0.06543432921171188,
0.002621312625706196,
-0.07872060686349869,
-0.03515704348683357,
-0.0410807766020298,
0.031855493783950806,
0.012921330519020557,
0.04565628245472908,
-0.023285413160920143,
-0.11403121054172516,
-0.021091321483254433,
-0.009021035395562649,
0.021226193755865097,
0.040569450706243515,
0.05262055993080139,
-0.06149494647979736,
-0.06144549325108528,
0.1858980804681778,
-0.04824770987033844,
0.022376226261258125,
0.05848727002739906,
-0.004674564581364393,
0.10410351306200027,
0.04727393388748169,
-0.0033934174571186304,
-0.05805998668074608,
0.055190857499837875,
-0.03450365364551544,
0.027631642296910286,
-0.02378392405807972,
-0.04997659847140312,
0.09032437205314636,
-0.06118850037455559,
-0.013780569657683372,
-0.16977757215499878,
-0.1346013844013214,
0.06920924782752991,
-0.0046393307857215405,
-0.057764723896980286,
-0.039836399257183075,
0.0016554485773667693,
-0.018505292013287544,
-0.008226707577705383,
-0.060180895030498505,
0.03363193944096565,
-0.06990969181060791,
0.08760694414377213,
0.07565318048000336,
0.12543350458145142,
-0.11965396255254745,
0.0020303758792579174,
-0.039431195706129074,
0.026080386713147163,
-0.07444153726100922,
0.017006250098347664,
-0.11360727995634079,
-0.026763200759887695,
-0.08075693249702454,
-0.05049572512507439,
-0.02787816897034645,
0.038458969444036484,
-0.0046360730193555355,
0.10894215852022171,
-0.05706582963466644,
-0.059176843613386154,
0.24394376575946808,
-0.09001754224300385,
-0.0675416886806488,
0.1283743530511856,
0.05509573593735695,
-0.08288798481225967,
0.04968947917222977,
0.07293882966041565,
0.11220444738864899,
-0.09920275956392288,
-0.07897584140300751,
0.08202799409627914,
-0.011124574579298496,
-0.08113528788089752,
0.09324955940246582,
0.05226298049092293,
0.03726081922650337,
0.03277285769581795,
0.16015088558197021,
0.03315340727567673,
-0.0033591832034289837,
-0.06818566471338272,
-0.05035815015435219,
-0.060364384204149246,
0.07670265436172485,
0.008411850780248642,
-0.027853697538375854,
-0.0955849140882492,
-0.11081656068563461,
-0.0018407582538202405,
0.07643625140190125,
-0.003808574751019478,
-0.007042072247713804,
-0.11988542228937149,
0.09860610961914062,
-0.08431075513362885,
0.030534401535987854,
-0.07410287111997604,
0.0018009860068559647,
-0.031286194920539856,
-0.005432157777249813,
0.0811082273721695,
0.10398299992084503,
0.07237140089273453,
0.03410398215055466,
-0.07853773236274719,
0.03451237827539444,
0.09517441689968109,
0.002592475153505802,
-0.005579901393502951,
-0.06802248954772949,
0.028330156579613686,
-0.012391743250191212,
0.1195276528596878,
-0.08033236861228943,
0.025248754769563675,
0.13460321724414825,
0.10169269144535065,
-0.026195334270596504,
-0.03396874666213989,
0.06583473831415176,
0.033120084553956985,
-0.04175065830349922,
0.017414839938282967,
0.08775833249092102,
-0.01821061596274376,
-0.01708090491592884,
0.0931672602891922,
-0.21018573641777039,
0.11587024480104446,
0.1743914633989334,
-0.06498851627111435,
-0.02952040545642376,
-0.06295245885848999,
-0.008836954832077026,
-0.025662142783403397,
0.011829815804958344,
-0.05600219964981079,
0.0754731297492981,
-0.0157193373888731,
0.1265556514263153,
-0.08388042449951172,
-0.05341155081987381,
0.052586495876312256,
-0.04451216757297516,
0.0030763521790504456,
0.06995439529418945,
0.10785765945911407,
-0.1132221594452858,
0.15843208134174347,
0.14048916101455688,
-0.08623234927654266,
0.1342327743768692,
0.02710869163274765,
-0.03057202138006687,
-0.03284471854567528,
0.014494323171675205,
0.007302282843738794,
0.05731755122542381,
-0.11842694878578186,
0.05457044765353203,
0.08014906942844391,
-0.012231960892677307,
0.026148827746510506,
-0.13130798935890198,
-0.041042447090148926,
0.020197927951812744,
-0.06961576640605927,
-0.0844624936580658,
-0.001935245469212532,
0.0051458440721035,
0.10185524076223373,
-0.014874766580760479,
-0.002055926714092493,
0.038969628512859344,
0.04297409579157829,
-0.0709800273180008,
0.14033782482147217,
-0.09656988829374313,
-0.3066079616546631,
-0.0803823247551918,
-0.14415043592453003,
-0.0817282423377037,
0.016433846205472946,
0.09379468113183975,
-0.0701383575797081,
-0.030534593388438225,
-0.0475073978304863,
-0.045373786240816116,
-0.04463600739836693,
-0.011426780372858047,
0.009467605501413345,
-0.017857851460576057,
0.056181080639362335,
-0.07972937077283859,
-0.020921099931001663,
0.05751101300120354,
-0.04194076731801033,
0.1678701490163803,
-0.02438369020819664,
0.0915750190615654,
0.13780845701694489,
0.017972661182284355,
-0.016100287437438965,
-0.03309692442417145,
0.15850713849067688,
-0.06741360574960709,
0.050333715975284576,
0.14121335744857788,
-0.008974883705377579,
0.07715511322021484,
0.09150806814432144,
0.04131568968296051,
-0.05215676128864288,
-0.05244860053062439,
0.03022785857319832,
-0.097344771027565,
-0.14737333357334137,
-0.0378730334341526,
-0.0640517845749855,
0.14547155797481537,
0.05489320307970047,
0.05971897020936012,
0.06923962384462357,
0.11639398336410522,
-0.07709486782550812,
0.0023638715501874685,
0.028515661135315895,
0.08973821252584457,
0.17000845074653625,
0.02026033215224743,
0.15407803654670715,
-0.0606195405125618,
-0.03826061263680458,
0.12693186104297638,
0.1314786821603775,
0.045955292880535126,
0.007944581098854542,
0.10924795269966125,
0.05701237544417381,
0.017258860170841217,
0.09667086601257324,
0.046098820865154266,
0.012954002246260643,
-0.03285377845168114,
-0.040217723697423935,
-0.059389784932136536,
-0.005913493689149618,
0.042309269309043884,
0.07064083963632584,
-0.07769019156694412,
0.02984972670674324,
-0.08867993950843811,
0.0673200786113739,
0.06441158801317215,
0.09003761410713196,
-0.20849324762821198,
-0.02992071770131588,
0.04827633127570152,
-0.027335915714502335,
-0.04112113267183304,
0.019475549459457397,
0.004770760424435139,
-0.02262500487267971,
0.16354256868362427,
-0.008895487524569035,
0.09425846487283707,
-0.07757965475320816,
0.02507471852004528,
-0.03523316606879234,
0.05439484119415283,
0.0023457815404981375,
0.11449393630027771,
-0.2918599843978882,
0.07582505792379379,
0.023314490914344788,
0.0012276594061404467,
-0.07052648812532425,
-0.010207179002463818,
0.013882055878639221,
-0.034718673676252365,
0.0901244506239891,
0.012438245117664337,
0.11421339213848114,
-0.13491208851337433,
-0.06831399351358414,
0.018464000895619392,
0.05840577557682991,
0.0349113866686821,
0.03806408867239952,
0.050281524658203125,
-0.04098742827773094,
-0.011960378848016262,
-0.003586642909795046,
-0.03816379979252815,
-0.09426084905862808,
0.12283581495285034,
0.12159112840890884,
-0.019400738179683685,
0.014618787914514542,
-0.03383609279990196,
-0.19241802394390106,
0.22267019748687744,
-0.11631155759096146,
-0.0990121141076088,
-0.06024293228983879,
-0.05829605460166931,
0.020255519077181816,
-0.0381576232612133,
0.01897526904940605,
-0.05281871184706688,
0.0883147343993187,
-0.033136241137981415,
-0.08289738744497299,
0.08133358508348465,
-0.10381626337766647,
-0.13379605114459991,
-0.034744199365377426,
0.1557082235813141,
-0.07043249160051346,
0.03930508345365524,
0.01958887279033661,
-0.012698178179562092,
-0.04660229757428169,
-0.12687496840953827,
0.10464207828044891,
0.005121081601828337,
-0.002601879183202982,
0.01565508171916008,
0.016671134158968925,
-0.011074880138039589,
-0.03946428745985031,
-0.03481132909655571,
0.1423601508140564,
0.35763075947761536,
-0.0339338481426239,
0.11625292897224426,
0.12072649598121643,
-0.03970842808485031,
-0.2128467708826065,
-0.1643359512090683,
-0.0416446328163147,
-0.07984194904565811,
-0.05990324914455414,
-0.1397559642791748,
0.04150431230664253,
0.13324101269245148,
-0.03415882587432861,
0.19096927344799042,
-0.14427873492240906,
-0.09891446679830551,
0.040286749601364136,
0.053915467113256454,
0.2451668232679367,
-0.1578064262866974,
-0.059641093015670776,
-0.060271572321653366,
-0.20136107504367828,
0.21934112906455994,
-0.038508519530296326,
0.1254277378320694,
-0.039567794650793076,
0.1007058322429657,
-0.002317465143278241,
-0.05951494723558426,
0.12377520650625229,
-0.023141780868172646,
-0.028445562347769737,
-0.11745912581682205,
0.04279402643442154,
0.007250812370330095,
-0.006112854462116957,
0.01075492799282074,
-0.17772872745990753,
-0.029193278402090073,
-0.13894276320934296,
-0.059055034071207047,
-0.07789207994937897,
0.018466319888830185,
-0.032297417521476746,
-0.05342421308159828,
-0.053447071462869644,
-0.0018394328653812408,
0.013208162970840931,
-0.02125684730708599,
-0.012754383496940136,
-0.07367511838674545,
0.07575365155935287,
0.29926881194114685,
0.09205321967601776,
-0.0795254334807396,
-0.09419336169958115,
-0.06699013710021973,
-0.09462294727563858,
0.022283926606178284,
-0.1190989688038826,
0.03506796434521675,
0.08005673438310623,
0.008496243506669998,
0.06271721422672272,
0.04441560059785843,
-0.05945681408047676,
0.020567039027810097,
0.1208619624376297,
-0.1396447718143463,
0.002770765218883753,
-0.01386578194797039,
0.06085022911429405,
0.0014808432897552848,
-0.00932159274816513,
0.13522863388061523,
-0.008964493870735168,
-0.01375904493033886,
0.05763800069689751,
0.017567263916134834,
-0.019581550732254982,
0.11292410641908646,
0.03593772277235985,
0.07312403619289398,
-0.13113315403461456,
0.06227600574493408,
-0.010140580125153065,
-0.06087154150009155,
-0.010526877827942371,
0.1051144078373909,
-0.07064200192689896,
-0.12954294681549072,
-0.01851516030728817,
-0.07937530428171158,
-0.10015391558408737,
-0.039711881428956985,
-0.02498324029147625,
-0.11733373254537582,
0.0656353086233139,
0.1202923059463501,
0.021431051194667816,
-0.024201160296797752,
0.0635220855474472,
0.04049062356352806,
-0.0697946771979332,
0.05972917005419731,
-0.009029263630509377,
0.08210357278585434,
-0.10026324540376663,
0.056725118309259415,
-0.007556890603154898,
0.04659852012991905,
-0.04206955060362816,
0.010912511497735977,
-0.1222345381975174,
-0.029154933989048004,
-0.13856512308120728,
0.05458979308605194,
-0.11419469863176346,
-0.0046210624277591705,
-0.019820652902126312,
-0.00031494468566961586,
-0.028095893561840057,
-0.04015031084418297,
-0.09113653004169464,
0.0059418026357889175,
0.022432716563344002,
-0.005477967206388712,
-0.09419668465852737,
-0.027275707572698593,
0.029129697009921074,
-0.02314028888940811,
0.07347982376813889,
0.05101842060685158,
-0.05335640162229538,
-0.019004682078957558,
-0.1502988040447235,
-0.03239204362034798,
0.07845284789800644,
0.07175181806087494,
-0.02373453415930271,
0.026781048625707626,
0.0446876659989357,
0.040196746587753296,
0.03672241047024727,
0.030681585893034935,
0.07553961127996445,
-0.09002012014389038,
-0.015301886014640331,
-0.06937103718519211,
-0.09953287988901138,
-0.05252215266227722,
-0.009011564776301384,
0.12413270771503448,
-0.03197924420237541,
0.1056065708398819,
-0.04909102991223335,
0.027423663064837456,
-0.1370212584733963,
0.02081146091222763,
0.029845934361219406,
-0.1496959626674652,
-0.006880047731101513,
-0.05901682376861572,
0.031370438635349274,
-0.028788993135094643,
0.16387051343917847,
-0.07810865342617035,
-0.10397236049175262,
0.027627112343907356,
-0.06955862790346146,
-0.03780260309576988,
-0.01863793469965458,
0.19606934487819672,
0.07767161726951599,
-0.013180472888052464,
-0.09442797303199768,
0.02831275574862957,
0.05446578189730644,
-0.07327965646982193,
0.08605897426605225,
0.08183206617832184,
-0.027250660583376884,
0.043587591499090195,
0.06545001268386841,
-0.07225538790225983,
-0.08333423733711243,
0.003726643044501543,
-0.12132780253887177,
0.10432349145412445,
-0.020077280700206757,
0.08851893246173859,
0.08799950778484344,
-0.04072103276848793,
0.01856432482600212,
-0.06785811483860016,
-0.03454233705997467,
-0.1094784364104271,
-0.08605699986219406,
-0.08265437185764313,
-0.0956774652004242,
0.030323902145028114,
-0.07698007673025131,
0.03162778913974762,
0.05978146567940712,
0.048156559467315674,
-0.017655223608016968,
0.0899934470653534,
-0.01561000756919384,
-0.03311045840382576,
0.011788095347583294,
-0.025170205160975456,
-0.01997991092503071,
0.029793737456202507,
-0.033783409744501114,
0.06704971194267273,
-0.014576822519302368,
0.044692330062389374,
0.02526051364839077,
0.018841538578271866,
-0.0034625304397195578,
-0.06826424598693848,
-0.0928659662604332,
-0.03558668866753578,
0.03649220988154411,
0.08033668249845505,
0.15678779780864716,
-0.03397746384143829,
0.015076669864356518,
0.015631316229701042,
0.07815948873758316,
-0.027574598789215088,
-0.08332843333482742,
-0.035558611154556274,
0.1978200227022171,
-0.08578452467918396,
0.05133093148469925,
-0.015142149291932583,
-0.02114446647465229,
-0.00793991144746542,
0.24854402244091034,
0.15722061693668365,
-0.040044721215963364,
0.031300682574510574,
0.04459945112466812,
0.023063266649842262,
0.07644317299127579,
0.09012192487716675,
0.08018463104963303,
0.18161940574645996,
-0.05073779448866844,
0.005434007383882999,
0.02759838104248047,
0.000705020793247968,
-0.08142122626304626,
0.059782423079013824,
0.018318818882107735,
-0.019371159374713898,
-0.06391309201717377,
0.021794551983475685,
-0.128355473279953,
0.06926099210977554,
-0.004509282764047384,
-0.11827269196510315,
-0.07526135444641113,
0.0054476032964885235,
0.05292195454239845,
0.04792305454611778,
0.012942680157721043,
-0.030346445739269257,
-0.010075491853058338,
0.021668286994099617,
-0.019004857167601585,
-0.17709168791770935,
0.056254927068948746,
0.005602172110229731,
-0.0030614288989454508,
0.027611874043941498,
-0.04015858471393585,
0.15660524368286133,
0.10575133562088013,
0.00849032774567604,
-0.04976148158311844,
0.13094083964824677,
0.01960224285721779,
-0.169062077999115,
0.03937619552016258,
0.12717700004577637,
-0.03223618119955063,
-0.0018831146880984306,
0.09252047538757324,
-0.06245819851756096,
-0.0493307001888752,
0.10854639112949371,
-0.029410231858491898,
-0.01744406670331955,
0.0635363832116127,
-0.05378352105617523,
0.07308464497327805,
0.06606873124837875,
-0.04988853260874748,
-0.07064728438854218,
-0.04445139691233635,
0.05181637033820152,
-0.023535815998911858,
-0.05790677294135094,
-0.05711677297949791,
-0.1689419001340866,
-0.05284210294485092,
0.053685273975133896,
0.12549011409282684,
-0.24432633817195892,
-0.004932210315018892,
-0.08727642148733139,
0.00022428696684073657,
-0.12954051792621613,
0.06038738042116165,
0.09986879676580429,
0.010964027605950832,
-0.050395190715789795,
-0.10188592225313187,
-0.01950676552951336,
0.07851249724626541,
-0.12314824014902115,
-0.07194638252258301
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bert-finetuned-ner
This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0662
- Precision: 0.9272
- Recall: 0.9472
- F1: 0.9371
- Accuracy: 0.9850
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
| 0.0811 | 1.0 | 1756 | 0.0764 | 0.9101 | 0.9335 | 0.9217 | 0.9809 |
| 0.0408 | 2.0 | 3512 | 0.0595 | 0.9268 | 0.9465 | 0.9366 | 0.9852 |
| 0.0231 | 3.0 | 5268 | 0.0662 | 0.9272 | 0.9472 | 0.9371 | 0.9850 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.2.0+rocm5.7
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["precision", "recall", "f1", "accuracy"], "base_model": "bert-base-cased", "model-index": [{"name": "bert-finetuned-ner", "results": []}]} | token-classification | Ankush-Chander/bert-finetuned-ner | [
"transformers",
"safetensors",
"bert",
"token-classification",
"generated_from_trainer",
"base_model:bert-base-cased",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T07:32:02+00:00 | [] | [] | TAGS
#transformers #safetensors #bert #token-classification #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| bert-finetuned-ner
==================
This model is a fine-tuned version of bert-base-cased on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0662
* Precision: 0.9272
* Recall: 0.9472
* F1: 0.9371
* Accuracy: 0.9850
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.2.0+rocm5.7
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.2.0+rocm5.7\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #safetensors #bert #token-classification #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.2.0+rocm5.7\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
64,
98,
4,
34
] | [
"passage: TAGS\n#transformers #safetensors #bert #token-classification #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.2.0+rocm5.7\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.09831220656633377,
0.07958109676837921,
-0.0017806809628382325,
0.10985765606164932,
0.1635071039199829,
0.02668670192360878,
0.12369365245103836,
0.10207930952310562,
-0.08563597500324249,
0.02737467922270298,
0.12613767385482788,
0.15009385347366333,
-0.002090511843562126,
0.14714820683002472,
-0.06575959920883179,
-0.227696031332016,
0.020761754363775253,
0.028197679668664932,
-0.06721656024456024,
0.11695695668458939,
0.1059551015496254,
-0.1372283399105072,
0.09770841151475906,
-0.007583102211356163,
-0.18529614806175232,
0.025115426629781723,
0.035012517124414444,
-0.0585329644382,
0.1359691172838211,
0.03129078075289726,
0.1287764608860016,
0.011531610041856766,
0.10812941938638687,
-0.20152011513710022,
0.005949531681835651,
0.058524660766124725,
0.0046082753688097,
0.07858985662460327,
0.03663022071123123,
-0.003846408799290657,
0.08306239545345306,
-0.06684644520282745,
0.06854445487260818,
0.021774547174572945,
-0.1226971223950386,
-0.251228392124176,
-0.07796281576156616,
0.04706045240163803,
0.10559286922216415,
0.08409972488880157,
-0.0007842931663617492,
0.14462101459503174,
-0.09228267520666122,
0.07928992807865143,
0.22557079792022705,
-0.3082197308540344,
-0.05891966074705124,
0.060818616300821304,
0.01539008878171444,
0.06830534338951111,
-0.11324869096279144,
-0.02810603380203247,
0.06987427920103073,
0.024819832295179367,
0.1306712031364441,
-0.03405775502324104,
-0.0776149332523346,
0.01877436973154545,
-0.14108368754386902,
-0.01149131078273058,
0.15255188941955566,
0.05792757496237755,
-0.050984300673007965,
-0.031330276280641556,
-0.05626202002167702,
-0.12809400260448456,
-0.0408359058201313,
-0.028026068583130836,
0.0489673838019371,
-0.030673682689666748,
-0.0688982829451561,
0.023592116311192513,
-0.09784766286611557,
-0.09277608245611191,
-0.050392307341098785,
0.17091898620128632,
0.04078972339630127,
0.010546734556555748,
-0.0011532314820215106,
0.11111114174127579,
-0.0357256680727005,
-0.12244255840778351,
0.004505884367972612,
0.014938686974346638,
0.005830916576087475,
-0.06361766159534454,
-0.053266365081071854,
0.007415841333568096,
0.0258750282227993,
0.17494477331638336,
-0.05387221649289131,
0.03584132716059685,
0.03333302214741707,
0.025917133316397667,
-0.10353977233171463,
0.16010048985481262,
-0.04349268227815628,
-0.04203726351261139,
0.018759533762931824,
0.0643944963812828,
0.028222639113664627,
0.011084962636232376,
-0.11055567860603333,
0.019174499437212944,
0.11307826638221741,
0.015185169875621796,
-0.09060557186603546,
0.07508856058120728,
-0.0546051524579525,
-0.004069817252457142,
0.03469835966825485,
-0.08581904321908951,
0.030779007822275162,
0.0033215414732694626,
-0.06643234938383102,
-0.07760277390480042,
0.02261275053024292,
0.026584671810269356,
0.016296520829200745,
0.11040935665369034,
-0.08942226320505142,
0.009115955792367458,
-0.09381615370512009,
-0.11263544857501984,
0.00576390465721488,
-0.043470580130815506,
0.046425655484199524,
-0.11200307309627533,
-0.17048469185829163,
-0.003327329643070698,
0.05448651686310768,
-0.023394711315631866,
-0.044402897357940674,
-0.04104434326291084,
-0.07164280116558075,
0.0024075936526060104,
-0.02608470618724823,
0.10447226464748383,
-0.07039155066013336,
0.08840996026992798,
0.0596042163670063,
0.06084148585796356,
-0.0463496670126915,
0.03923950716853142,
-0.11302606761455536,
0.019132113084197044,
-0.19155292212963104,
0.0070922039449214935,
-0.07125571370124817,
0.06404747813940048,
-0.08579554408788681,
-0.0829005017876625,
0.02184605784714222,
0.010096593759953976,
0.06342214345932007,
0.09752793610095978,
-0.12846878170967102,
-0.06587778031826019,
0.14662116765975952,
-0.0955214723944664,
-0.14418543875217438,
0.10392648726701736,
-0.06167716905474663,
0.049651652574539185,
0.07337011396884918,
0.15121819078922272,
0.06468922644853592,
-0.08912701159715652,
0.01035927701741457,
-0.0004947098204866052,
0.0585702583193779,
-0.05677284672856331,
0.08260942250490189,
0.006433199159801006,
-0.041778068989515305,
0.02444491907954216,
-0.06828764826059341,
0.0540994293987751,
-0.09889411926269531,
-0.08369879424571991,
-0.03560865670442581,
-0.1062127873301506,
0.04174242913722992,
0.05841950327157974,
0.06485086679458618,
-0.11640206724405289,
-0.08762300759553909,
0.08292786777019501,
0.09157507121562958,
-0.055724360048770905,
0.010680424980819225,
-0.07375026494264603,
0.06610315293073654,
-0.05842623859643936,
-0.027566615492105484,
-0.1517084836959839,
-0.05076796934008598,
0.01159333810210228,
0.005720422603189945,
-0.002198622329160571,
-0.0049454024992883205,
0.07086014002561569,
0.08458349108695984,
-0.07688698172569275,
-0.04046803340315819,
-0.015726834535598755,
0.024006670340895653,
-0.12931688129901886,
-0.21010348200798035,
-0.030865121632814407,
-0.024879533797502518,
0.13485302031040192,
-0.24041251838207245,
0.03838163986802101,
-0.026699386537075043,
0.0951351672410965,
0.029864676296710968,
-0.00036433737841434777,
-0.06039324775338173,
0.07267769426107407,
-0.04008305445313454,
-0.05298847705125809,
0.05118989571928978,
-0.003443983616307378,
-0.07553650438785553,
-0.049552951008081436,
-0.11379539221525192,
0.2189609557390213,
0.13224920630455017,
-0.09600858390331268,
-0.08326144516468048,
-0.009493527933955193,
-0.04393612965941429,
-0.026706242933869362,
-0.052721891552209854,
0.01735427789390087,
0.1137830913066864,
-0.028113091364502907,
0.14701581001281738,
-0.07145506143569946,
-0.03112766332924366,
0.018180936574935913,
-0.04727376252412796,
0.02389395795762539,
0.10091965645551682,
0.09951212257146835,
-0.11091522127389908,
0.15640319883823395,
0.18213573098182678,
-0.08614228665828705,
0.10173646360635757,
-0.03848882392048836,
-0.06185241416096687,
-0.02120809257030487,
-0.014201388694345951,
-0.003633060259744525,
0.12251677364110947,
-0.11200214922428131,
0.0093354731798172,
0.015072330832481384,
0.02760210447013378,
0.006324430461972952,
-0.2107950747013092,
-0.035067714750766754,
0.03462888300418854,
-0.03742165490984917,
-0.004375382326543331,
-0.03180399164557457,
-0.009406514465808868,
0.098078653216362,
0.002870784141123295,
-0.10468444228172302,
0.0482088141143322,
0.0031739231199026108,
-0.08274620026350021,
0.21664440631866455,
-0.08428800851106644,
-0.09631215780973434,
-0.12349293380975723,
-0.08644216507673264,
-0.04164620488882065,
0.029276132583618164,
0.06836877763271332,
-0.07629434019327164,
-0.049164846539497375,
-0.08743561804294586,
0.015333281829953194,
0.038622550666332245,
0.034459277987480164,
0.008426466025412083,
-0.0018090428784489632,
0.08852152526378632,
-0.09922076761722565,
-0.015258999541401863,
-0.05254410579800606,
-0.0719466283917427,
0.036156270653009415,
0.02906910702586174,
0.10737273842096329,
0.13829897344112396,
-0.02189031057059765,
-0.0017261133762076497,
-0.030263138934969902,
0.23663821816444397,
-0.04884570091962814,
-0.036892399191856384,
0.1300402283668518,
-0.009556509554386139,
0.03274412825703621,
0.1452403962612152,
0.06103111431002617,
-0.09891704469919205,
0.022533036768436432,
0.0442742183804512,
-0.021306093782186508,
-0.1996358186006546,
-0.039486031979322433,
-0.02920493483543396,
-0.024826163426041603,
0.09690818935632706,
0.028207534924149513,
0.025133104994893074,
0.07466696947813034,
0.024991165846586227,
0.06459787487983704,
-0.014715719036757946,
0.07478879392147064,
0.10304789245128632,
0.04754674434661865,
0.12922842800617218,
-0.038911644369363785,
-0.0603669099509716,
0.032352663576602936,
-0.01348694134503603,
0.20135556161403656,
0.020582858473062515,
0.07886270433664322,
0.05629081651568413,
0.17929990589618683,
0.00009217452316079289,
0.07850615680217743,
-0.002551690209656954,
-0.05333138629794121,
-0.014020326547324657,
-0.04586435854434967,
-0.030925557017326355,
0.04106966033577919,
-0.10406935214996338,
0.07385329157114029,
-0.12800739705562592,
0.0022258071694523096,
0.05889572948217392,
0.23990023136138916,
0.058063872158527374,
-0.3375164866447449,
-0.09350112825632095,
0.019316742196679115,
-0.023261025547981262,
-0.026688789948821068,
0.03228925168514252,
0.09419069439172745,
-0.06512091308832169,
0.029513631016016006,
-0.05241502448916435,
0.07909289747476578,
-0.005823995918035507,
0.04521613568067551,
0.07235081493854523,
0.09434757381677628,
-0.00672544352710247,
0.07449168711900711,
-0.2796105742454529,
0.2733730673789978,
0.00964664202183485,
0.07838501036167145,
-0.04930786043405533,
0.00016207255248446018,
0.03154172748327255,
0.11284030228853226,
0.07163410633802414,
-0.018261220306158066,
-0.07035346329212189,
-0.22684162855148315,
-0.03889385610818863,
0.03974045813083649,
0.08010976016521454,
-0.04366300627589226,
0.10165977478027344,
-0.043552156537771225,
0.0015019563725218177,
0.09350906312465668,
-0.012703068554401398,
-0.09185571223497391,
-0.07647749781608582,
-0.03553386405110359,
0.027860049158334732,
0.006049425341188908,
-0.09330032765865326,
-0.09915915876626968,
-0.11140696704387665,
0.15474753081798553,
-0.03177804499864578,
-0.014658923260867596,
-0.10409601032733917,
0.07422810047864914,
0.05613882839679718,
-0.0812893658876419,
0.05203460529446602,
0.013103777542710304,
0.07580825686454773,
0.04868819937109947,
-0.058193933218717575,
0.13389132916927338,
-0.08395720273256302,
-0.1748361885547638,
-0.06688990443944931,
0.09698296338319778,
0.027166277170181274,
0.04544954001903534,
0.00661328062415123,
0.012588665820658207,
-0.01367996260523796,
-0.08048291504383087,
0.012647667899727821,
-0.00832938402891159,
0.07232210785150528,
0.030864635482430458,
-0.07590785622596741,
-0.008995486423373222,
-0.05229681357741356,
-0.03295600041747093,
0.14627055823802948,
0.2813584506511688,
-0.0935591533780098,
-0.007586196064949036,
0.0692455992102623,
-0.05796591937541962,
-0.20352278649806976,
0.033373404294252396,
0.02923092059791088,
-0.005136498250067234,
0.05508413910865784,
-0.13770164549350739,
0.15373097360134125,
0.11468848586082458,
-0.02976273000240326,
0.07699833810329437,
-0.26201972365379333,
-0.12825094163417816,
0.1569330394268036,
0.15561793744564056,
0.13026003539562225,
-0.1466950923204422,
-0.019213814288377762,
-0.043062154203653336,
-0.126937597990036,
0.10128626227378845,
-0.11680161207914352,
0.10005784779787064,
-0.019198328256607056,
0.041712649166584015,
0.0022957713808864355,
-0.05070838704705238,
0.12625783681869507,
0.008199193514883518,
0.12560752034187317,
-0.05868252366781235,
-0.029757140204310417,
0.0340866893529892,
-0.04979519918560982,
0.015984637662768364,
-0.08787699043750763,
0.04201898351311684,
-0.06658630818128586,
-0.0326450914144516,
-0.05865761265158653,
0.0423768125474453,
-0.031245896592736244,
-0.07298372685909271,
-0.03636746481060982,
0.038495078682899475,
0.03327082097530365,
-0.022570747882127762,
0.14979392290115356,
0.016566604375839233,
0.16188760101795197,
0.1266263723373413,
0.06572439521551132,
-0.08001720160245895,
-0.01979462057352066,
-0.0028213965706527233,
-0.03741413354873657,
0.07614663243293762,
-0.13731513917446136,
0.051124896854162216,
0.11614800244569778,
0.004790155217051506,
0.1432119905948639,
0.07780785113573074,
-0.021545501425862312,
-0.00005621669697575271,
0.07511106133460999,
-0.1590983122587204,
-0.07822761684656143,
0.001803574850782752,
-0.041799772530794144,
-0.11258093267679214,
0.07965211570262909,
0.10361629724502563,
-0.07746175676584244,
-0.002373819937929511,
-0.018608761951327324,
0.00994702335447073,
-0.06501707434654236,
0.20243220031261444,
0.06924531608819962,
0.05142158269882202,
-0.09208018332719803,
0.07286699116230011,
0.03882827237248421,
-0.03912424296140671,
-0.006344042718410492,
0.02625451423227787,
-0.08935263752937317,
-0.04799299314618111,
0.07955136895179749,
0.19467055797576904,
-0.05721511319279671,
-0.05971342325210571,
-0.13465510308742523,
-0.12911435961723328,
0.0554274246096611,
0.1631699502468109,
0.11761815845966339,
0.014311357401311398,
-0.012652755714952946,
0.015274551697075367,
-0.1164470911026001,
0.09927800297737122,
0.028081901371479034,
0.08513196557760239,
-0.16623134911060333,
0.12174510955810547,
-0.0032519109081476927,
0.01142369955778122,
-0.028274059295654297,
0.04644166678190231,
-0.1297406554222107,
-0.00044845693628303707,
-0.13689187169075012,
-0.027784962207078934,
-0.03345385193824768,
0.017223849892616272,
0.01727762259542942,
-0.06910945475101471,
-0.06752181053161621,
0.024323362857103348,
-0.10739050060510635,
-0.014295108616352081,
0.043654199689626694,
0.06958591192960739,
-0.12543997168540955,
-0.04295589402318001,
0.023413607850670815,
-0.06479630619287491,
0.06155102699995041,
0.03493325039744377,
0.03628024831414223,
0.06638278812170029,
-0.18405510485172272,
0.011325341649353504,
0.07504988461732864,
0.005022465717047453,
0.06571654975414276,
-0.09838075190782547,
-0.006624607834964991,
0.00573468254879117,
0.04993539676070213,
0.01387250330299139,
0.08707065135240555,
-0.13288824260234833,
-0.013765361160039902,
-0.0331207700073719,
-0.07930240035057068,
-0.04703912511467934,
0.008963709697127342,
0.09997568279504776,
-0.012822726741433144,
0.21962100267410278,
-0.09374181926250458,
0.003843106562271714,
-0.20471127331256866,
0.0032314436975866556,
-0.01876039057970047,
-0.10668116062879562,
-0.1527889519929886,
-0.05297112837433815,
0.04883605241775513,
-0.045002978295087814,
0.1543465405702591,
0.007233229465782642,
0.04542522504925728,
0.03630175068974495,
-0.05399436131119728,
0.0684366375207901,
0.03212660551071167,
0.2310352921485901,
0.03936995193362236,
-0.03480613976716995,
0.03814487159252167,
0.04848698899149895,
0.11425748467445374,
0.06748522073030472,
0.1610034704208374,
0.17476633191108704,
-0.04728342592716217,
0.09418213367462158,
0.03955711051821709,
-0.07828104496002197,
-0.13258174061775208,
0.026465823873877525,
-0.04574745148420334,
0.07432672381401062,
-0.017915846779942513,
0.2030239701271057,
0.09288463741540909,
-0.16761986911296844,
0.009297623299062252,
-0.07391824573278427,
-0.07106336951255798,
-0.11996874213218689,
-0.024797193706035614,
-0.102790467441082,
-0.18031400442123413,
0.0009336341172456741,
-0.1104981079697609,
-0.002584097906947136,
0.12614591419696808,
0.000010933542398561258,
-0.0085369311273098,
0.16794057190418243,
0.004105341155081987,
0.046494413167238235,
0.01941482536494732,
-0.0026755905710160732,
-0.038760945200920105,
-0.08471111953258514,
-0.08526057749986649,
0.0007255972013808787,
-0.03665401041507721,
0.016057318076491356,
-0.06815817207098007,
-0.05357777699828148,
0.04904535412788391,
-0.006722364574670792,
-0.09589654207229614,
0.020281026139855385,
0.019624289125204086,
0.04802772402763367,
0.044985704123973846,
0.005353051703423262,
0.01728653721511364,
0.003871762193739414,
0.22798381745815277,
-0.0813646912574768,
-0.0775085985660553,
-0.10765101760625839,
0.28505003452301025,
0.054986622184515,
0.03174348548054695,
0.01904674991965294,
-0.0822770893573761,
0.026446331292390823,
0.20790594816207886,
0.1803666055202484,
-0.07997119426727295,
-0.0010371820535510778,
-0.011297213844954967,
-0.02032775990664959,
-0.04458765313029289,
0.0976206436753273,
0.12819844484329224,
0.005587501917034388,
-0.0756351500749588,
-0.03554091602563858,
-0.03365280479192734,
-0.005781602580100298,
-0.039627742022275925,
0.048702362924814224,
0.03974958881735802,
0.010212685912847519,
-0.045678503811359406,
0.04721447452902794,
-0.014153587631881237,
-0.1008584052324295,
0.08149441331624985,
-0.172096386551857,
-0.14768104255199432,
-0.020609702914953232,
0.10317273437976837,
-0.000815954408608377,
0.05726301670074463,
-0.038060035556554794,
-0.0070727793499827385,
0.06073908135294914,
-0.01601472496986389,
-0.07332497835159302,
-0.11527873575687408,
0.08691602945327759,
-0.07021123915910721,
0.2546461820602417,
-0.03555218502879143,
0.05831902101635933,
0.13294416666030884,
0.03353669494390488,
-0.08822346478700638,
0.09399806708097458,
0.0399518683552742,
-0.09599117934703827,
0.03225810080766678,
0.0595274344086647,
-0.04613848030567169,
0.12879236042499542,
0.03742818906903267,
-0.15423095226287842,
0.015357974916696548,
-0.06811739504337311,
-0.08722546696662903,
-0.05075886845588684,
-0.04422324150800705,
-0.05477740615606308,
0.13149824738502502,
0.19017383456230164,
-0.040086958557367325,
0.011851784773170948,
-0.05931057780981064,
0.047826431691646576,
0.06787673383951187,
0.027176009491086006,
-0.038958124816417694,
-0.2349729984998703,
0.04255380108952522,
0.0802200511097908,
-0.022448787465691566,
-0.24378438293933868,
-0.08288711309432983,
-0.004735493566840887,
-0.05937906727194786,
-0.09344840049743652,
0.08427915722131729,
0.10915159434080124,
0.057827312499284744,
-0.06589742004871368,
-0.13251543045043945,
-0.08427954465150833,
0.16011826694011688,
-0.13063779473304749,
-0.11201208829879761
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.8.2 | {"library_name": "peft", "base_model": "models/mixtral-lora-merged-3675"} | null | atom-team/mixtral | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:models/mixtral-lora-merged-3675",
"region:us"
] | 2024-02-08T07:33:31+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #safetensors #arxiv-1910.09700 #base_model-models/mixtral-lora-merged-3675 #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.8.2 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
"TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-models/mixtral-lora-merged-3675 #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
41,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-models/mixtral-lora-merged-3675 #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2"
] | [
-0.13101492822170258,
0.1977311372756958,
-0.0025078733451664448,
0.029466183856129646,
0.081293486058712,
0.022473450750112534,
0.05339458957314491,
0.13120678067207336,
0.019189337268471718,
0.10636375099420547,
0.07228220999240875,
0.11489839851856232,
0.10707191377878189,
0.20940837264060974,
0.002700644778087735,
-0.15539640188217163,
0.026914995163679123,
-0.08242541551589966,
0.005858516786247492,
0.13198167085647583,
0.14440156519412994,
-0.10225237905979156,
0.08192561566829681,
-0.015183054842054844,
-0.0016733095981180668,
-0.03414000943303108,
-0.06631128489971161,
-0.01602485217154026,
0.048453204333782196,
0.02907000482082367,
0.06350649148225784,
-0.013226686045527458,
0.09106961637735367,
-0.25972220301628113,
0.019663603976368904,
0.041013721376657486,
0.007989229634404182,
0.08513255417346954,
0.10030190646648407,
-0.0406324528157711,
0.12033858895301819,
-0.0263143889605999,
0.13889268040657043,
0.09410740435123444,
-0.08800096064805984,
-0.23342028260231018,
-0.06892313808202744,
0.07254067808389664,
0.18939967453479767,
0.08773273229598999,
-0.046128418296575546,
0.1435544490814209,
-0.07082024961709976,
0.02490893378853798,
0.03403259813785553,
-0.09760366380214691,
-0.06731971353292465,
0.07125566899776459,
0.1358734667301178,
0.0598730742931366,
-0.12067443877458572,
-0.0386224091053009,
0.03252159431576729,
0.04403804615139961,
0.05983646959066391,
0.0049911197274923325,
0.16185303032398224,
0.02935427613556385,
-0.14766210317611694,
-0.052667926996946335,
0.14256635308265686,
0.013645825907588005,
-0.04723500832915306,
-0.2280578315258026,
-0.0021501381415873766,
-0.09305530786514282,
-0.022975390776991844,
-0.056481439620256424,
0.031086212024092674,
0.011463532224297523,
0.11900284141302109,
-0.040517233312129974,
-0.09511276334524155,
-0.026932165026664734,
0.09264659136533737,
0.05543569102883339,
0.026164144277572632,
-0.020906001329421997,
0.010387646965682507,
0.12473484873771667,
0.08206702768802643,
-0.132618710398674,
-0.06588873267173767,
-0.07720967382192612,
-0.04610569775104523,
-0.037289395928382874,
0.04002426192164421,
0.014472106471657753,
0.06846144795417786,
0.26505598425865173,
-0.02029813453555107,
0.06050390005111694,
0.06244012713432312,
0.016832174733281136,
0.0431690514087677,
0.10661473870277405,
-0.03170400112867355,
-0.1593373864889145,
-0.011377793736755848,
0.10070987790822983,
-0.004669223912060261,
-0.026946602389216423,
-0.04436321556568146,
0.031105775386095047,
0.04166543856263161,
0.11515448987483978,
0.11435416340827942,
-0.017075518146157265,
-0.08104534447193146,
-0.059587180614471436,
0.19574452936649323,
-0.15625593066215515,
0.04441101476550102,
0.023490658029913902,
-0.0037279969546943903,
-0.060040492564439774,
0.00729017797857523,
0.02002430520951748,
-0.024640636518597603,
0.06284961849451065,
-0.06603945791721344,
-0.04311089217662811,
-0.12869054079055786,
-0.028303029015660286,
0.0324375182390213,
0.01847953163087368,
-0.043697651475667953,
-0.047483935952186584,
-0.08449075371026993,
-0.11137789487838745,
0.10846428573131561,
-0.05334038287401199,
-0.05164244771003723,
-0.025944653898477554,
-0.08608156442642212,
0.02421456016600132,
0.029157333076000214,
0.07602086663246155,
-0.028060581535100937,
0.0502014234662056,
0.004483413882553577,
0.05928840488195419,
0.08043729513883591,
0.029381971806287766,
-0.08406870067119598,
0.06602605432271957,
-0.19463136792182922,
0.0812702551484108,
-0.08290868997573853,
0.03260008245706558,
-0.16592726111412048,
-0.009332958608865738,
0.014816541224718094,
0.028421951457858086,
0.041616469621658325,
0.16521021723747253,
-0.2179841250181198,
-0.02222711592912674,
0.15535415709018707,
-0.11152077466249466,
-0.13592948019504547,
0.04521584138274193,
-0.041875630617141724,
0.1806858330965042,
0.029378147795796394,
0.004225794691592455,
0.10260098427534103,
-0.1636650711297989,
-0.028925521299242973,
-0.02582375518977642,
0.002881903201341629,
0.07583331316709518,
0.09217541664838791,
-0.09111859649419785,
-0.003951460588723421,
0.01135129202157259,
-0.06259612739086151,
-0.01047542691230774,
-0.043156735599040985,
-0.10418127477169037,
0.0023868363350629807,
-0.0874781683087349,
0.029290299862623215,
0.002868240000680089,
-0.09690071642398834,
-0.01100537646561861,
-0.15707425773143768,
-0.06267304718494415,
0.0913093164563179,
0.0026698862202465534,
-0.026342831552028656,
-0.10748342424631119,
0.061338890343904495,
-0.04379025474190712,
-0.026033224537968636,
-0.1426049917936325,
-0.029324669390916824,
0.016415691003203392,
-0.14086776971817017,
-0.009963298216462135,
-0.11982240527868271,
0.06747676432132721,
0.005592965520918369,
-0.049551334232091904,
-0.04575556144118309,
-0.004673891235142946,
0.0009939040755853057,
-0.05664048343896866,
-0.2308792620897293,
-0.03159021958708763,
-0.053069647401571274,
0.1603100299835205,
-0.22309228777885437,
0.04230746626853943,
0.010031469166278839,
0.11534100025892258,
0.0008413580362685025,
-0.06878545880317688,
0.023622600361704826,
-0.06527799367904663,
-0.024931980296969414,
-0.07050229609012604,
-0.004422832280397415,
0.00006741979450453073,
-0.03186860308051109,
0.015775837004184723,
-0.1093505322933197,
-0.05427003651857376,
0.0999450832605362,
0.06517232209444046,
-0.1536945253610611,
0.005525955930352211,
-0.03795458376407623,
-0.05948485806584358,
-0.07661230862140656,
-0.06677410751581192,
0.08856852352619171,
0.054768096655607224,
0.03870801255106926,
-0.07646259665489197,
-0.0700889453291893,
0.004684647545218468,
-0.025432420894503593,
-0.00998449232429266,
0.11883088946342468,
0.07231995463371277,
-0.10346900671720505,
0.09281810373067856,
0.07692910730838776,
0.019747447222471237,
0.06868889927864075,
-0.029149392619729042,
-0.10550101101398468,
-0.034301210194826126,
0.05926057696342468,
0.009600146673619747,
0.17228583991527557,
-0.06775636225938797,
0.05972784757614136,
0.04765337333083153,
-0.042905379086732864,
0.048712633550167084,
-0.08628479391336441,
0.00751006044447422,
0.0037271364126354456,
-0.015682194381952286,
0.032474976032972336,
-0.019204335287213326,
0.007287394255399704,
0.0763130709528923,
0.057271890342235565,
0.026161588728427887,
0.023074930533766747,
-0.03486791253089905,
-0.1449144333600998,
0.18194033205509186,
-0.0932818204164505,
-0.23786014318466187,
-0.1608998030424118,
0.06268294900655746,
0.049509789794683456,
-0.01406962238252163,
0.02737160585820675,
-0.05342062562704086,
-0.10175805538892746,
-0.08431752026081085,
0.006789942272007465,
0.03593280911445618,
-0.057547006756067276,
-0.0728028267621994,
0.04363740608096123,
0.04319525137543678,
-0.11838316917419434,
0.02784152142703533,
0.06703835725784302,
-0.009470147080719471,
0.0015534788835793734,
0.0567084476351738,
0.09636937826871872,
0.1850050687789917,
-0.005182886961847544,
0.0014779146295040846,
0.06118274852633476,
0.2763042449951172,
-0.15845952928066254,
0.10828164964914322,
0.14994429051876068,
-0.06798402220010757,
0.0672493502497673,
0.18259242177009583,
0.024703584611415863,
-0.098201222717762,
0.02672748640179634,
0.026626382023096085,
-0.018708480522036552,
-0.2757023870944977,
-0.054310716688632965,
-0.017667515203356743,
-0.08828048408031464,
0.07153863459825516,
0.08527760207653046,
0.0748993530869484,
0.04248740151524544,
-0.05753131955862045,
-0.10994327813386917,
0.025023043155670166,
0.1067877784371376,
-0.008801638148725033,
0.00446580583229661,
0.08406289666891098,
-0.04795113205909729,
0.005415075924247503,
0.08983971178531647,
-0.019522588700056076,
0.1402515023946762,
0.057115089148283005,
0.09596806764602661,
0.08467389643192291,
0.10162512212991714,
-0.008127155713737011,
0.03185635805130005,
0.01766301691532135,
0.023285550996661186,
0.02367042936384678,
-0.08749783039093018,
0.008606689050793648,
0.10867205262184143,
0.02986925095319748,
0.026278166100382805,
0.015574407763779163,
-0.047301020473241806,
0.03536445274949074,
0.19625237584114075,
0.028903447091579437,
-0.2186414748430252,
-0.08728162944316864,
0.05074770748615265,
-0.07661672681570053,
-0.15892520546913147,
-0.009963764809072018,
0.02579100802540779,
-0.16415250301361084,
0.016761276870965958,
-0.04246778413653374,
0.1034054234623909,
-0.08006279915571213,
-0.043603286147117615,
0.1107393130660057,
0.047019485384225845,
-0.018765360116958618,
0.05094430223107338,
-0.19289334118366241,
0.11026762425899506,
0.02847425825893879,
0.0750589594244957,
-0.0832187831401825,
0.09794435650110245,
0.0011466024443507195,
-0.01897674985229969,
0.16589128971099854,
-0.0006307333242148161,
-0.05539817363023758,
-0.08618653565645218,
-0.09352366626262665,
-0.001195865566842258,
0.07726653665304184,
-0.1258138120174408,
0.08134608715772629,
-0.03841320425271988,
-0.024108177050948143,
-0.006677200552076101,
-0.08129014819860458,
-0.13410641252994537,
-0.1518969088792801,
0.05159905552864075,
-0.09347356110811234,
0.02762426808476448,
-0.0855739414691925,
-0.054644625633955,
0.017700407654047012,
0.18797622621059418,
-0.2110428363084793,
-0.10849736630916595,
-0.14470817148685455,
-0.11216353625059128,
0.16443723440170288,
-0.041359953582286835,
0.08290141820907593,
-0.0023965956643223763,
0.15866471827030182,
0.007233109325170517,
-0.014305345714092255,
0.08850935846567154,
-0.09320814162492752,
-0.1894121617078781,
-0.04994163662195206,
0.16294056177139282,
0.14235933125019073,
0.028337107971310616,
-0.004777108784765005,
0.03238183632493019,
-0.06955365836620331,
-0.11247636377811432,
0.028386369347572327,
0.16045831143856049,
0.0734521672129631,
-0.01041861716657877,
-0.024638548493385315,
-0.11285308003425598,
-0.05786707252264023,
-0.04446154832839966,
-0.0042221685871481895,
0.20263570547103882,
-0.06641823053359985,
0.14407522976398468,
0.1081942617893219,
-0.057455044239759445,
-0.2093385010957718,
0.03897496312856674,
0.04321196675300598,
0.02686900459229946,
0.04257962852716446,
-0.1786588877439499,
0.09077389538288116,
-0.011423757299780846,
-0.08405761420726776,
0.1709817349910736,
-0.18146245181560516,
-0.13602550327777863,
0.1133841872215271,
0.024468930438160896,
-0.2106456309556961,
-0.14228284358978271,
-0.10256621241569519,
-0.020102160051465034,
-0.131020650267601,
0.03646630421280861,
0.00843360647559166,
0.009016519412398338,
0.013174978084862232,
0.016587262973189354,
0.03977164253592491,
-0.05440682917833328,
0.2123889923095703,
-0.04032447561621666,
-0.0015092899557203054,
-0.051026616245508194,
-0.06904590874910355,
0.01947328820824623,
-0.052425943315029144,
0.12255710363388062,
-0.01066295150667429,
0.038956720381975174,
-0.17159363627433777,
-0.04383876919746399,
-0.05246352404356003,
0.03895635902881622,
-0.08951245248317719,
-0.08032957464456558,
-0.044815752655267715,
0.09103373438119888,
0.08740467578172684,
-0.02187979593873024,
0.001431094016879797,
-0.0945679247379303,
0.08160622417926788,
0.2003055065870285,
0.19929108023643494,
0.07689174264669418,
-0.0600191205739975,
0.03015722706913948,
-0.0383666455745697,
0.045674629509449005,
-0.20761774480342865,
0.04048549011349678,
0.06357874721288681,
0.020152758806943893,
0.06619903445243835,
-0.010531431995332241,
-0.1545747071504593,
-0.07659731060266495,
0.0865154042840004,
-0.05856602266430855,
-0.16735705733299255,
-0.029562734067440033,
0.019681362435221672,
-0.21360661089420319,
-0.04482606053352356,
0.03614777699112892,
-0.014610647223889828,
-0.03971840813755989,
0.02138878032565117,
0.0792410746216774,
-0.029100600630044937,
0.10173190385103226,
0.08677118271589279,
0.09361756592988968,
-0.09577502310276031,
0.0540512353181839,
0.07381163537502289,
-0.03735879063606262,
0.0303103756159544,
0.12108102440834045,
-0.042098067700862885,
-0.047559816390275955,
0.0844283252954483,
0.12116990238428116,
0.004014563746750355,
-0.05773986876010895,
0.0030187270604074,
-0.04667292535305023,
0.055711325258016586,
0.1007411926984787,
0.03509727865457535,
0.003129531629383564,
0.07456441968679428,
0.02741217613220215,
-0.09097830951213837,
0.12143363803625107,
0.061856482177972794,
0.025242092087864876,
-0.052831243723630905,
-0.03685421124100685,
-0.016836080700159073,
-0.01032479852437973,
-0.019589820876717567,
0.0005761717911809683,
-0.08398573100566864,
0.0024573898408561945,
-0.12761583924293518,
0.020915241912007332,
-0.0783599317073822,
0.004759123083204031,
0.03426789864897728,
-0.05177300050854683,
0.001585514983162284,
-0.0014150525676086545,
-0.07152964919805527,
-0.05701761692762375,
-0.01604541577398777,
0.07963787764310837,
-0.1410338133573532,
0.03742758929729462,
0.0771525427699089,
-0.10488380491733551,
0.06718673557043076,
-0.008882639929652214,
0.010464935563504696,
-0.004274084232747555,
-0.14188849925994873,
0.054760802537202835,
-0.025702927261590958,
-0.008299757726490498,
0.002415982075035572,
-0.2037140280008316,
-0.005624857265502214,
-0.030392928048968315,
-0.06326981633901596,
0.020993901416659355,
0.004203697666525841,
-0.12133044004440308,
0.10446561872959137,
0.0033300265204161406,
-0.05639326572418213,
-0.022678829729557037,
0.03786991536617279,
0.09204836189746857,
-0.0039931414648890495,
0.13086743652820587,
-0.030242284759879112,
0.07970283925533295,
-0.17776314914226532,
-0.00963819120079279,
-0.017845582216978073,
0.05854746699333191,
-0.02458040788769722,
-0.037012360990047455,
0.06112007051706314,
-0.024624526500701904,
0.16929787397384644,
-0.004461493343114853,
0.07084934413433075,
0.050757329910993576,
0.011618639342486858,
0.04295482113957405,
0.07509218901395798,
0.0652996301651001,
-0.014202962629497051,
-0.0024032057262957096,
0.042233794927597046,
-0.001630362356081605,
-0.05475115403532982,
-0.16186872124671936,
0.06023145094513893,
0.17688106000423431,
0.05708583444356918,
0.028910692781209946,
0.011576597578823566,
-0.11690264940261841,
-0.07473988831043243,
0.10095632076263428,
-0.022362764924764633,
-0.03282218053936958,
-0.06385776400566101,
0.21606099605560303,
0.13195925951004028,
-0.19866184890270233,
0.07420742511749268,
-0.060870226472616196,
-0.043505582958459854,
-0.14142519235610962,
-0.16896039247512817,
-0.059916526079177856,
-0.052756838500499725,
-0.025999469682574272,
-0.05566401034593582,
0.043776120990514755,
0.03945796564221382,
-0.0036206282675266266,
-0.027506230399012566,
0.10849332809448242,
0.019861700013279915,
-0.03196585550904274,
0.043966758996248245,
0.056847892701625824,
0.03978142514824867,
-0.0926104187965393,
0.007514936849474907,
0.004839611239731312,
0.017875799909234047,
0.06984900683164597,
0.02004077285528183,
-0.07123644649982452,
0.028792861849069595,
-0.02239285781979561,
-0.12389541417360306,
0.04022122174501419,
-0.0030571455135941505,
-0.022295033559203148,
0.1483316719532013,
0.040285613387823105,
0.010083826258778572,
-0.01606481522321701,
0.22836735844612122,
-0.07590840011835098,
-0.08120842278003693,
-0.1406882256269455,
0.07508563250303268,
-0.07930915802717209,
0.02232973463833332,
0.02518034726381302,
-0.1214710921049118,
0.016601432114839554,
0.16918830573558807,
0.12263420969247818,
-0.014735880307853222,
0.004936723504215479,
0.0460653118789196,
0.003278210759162903,
-0.0412067286670208,
0.01963464356958866,
0.051352545619010925,
0.19309279322624207,
-0.07498933374881744,
0.05766104534268379,
-0.019642358645796776,
-0.07979892194271088,
-0.014974438585340977,
0.09325068444013596,
-0.013131575658917427,
-0.004796124529093504,
-0.06122472882270813,
0.15332958102226257,
-0.07543076574802399,
-0.20604664087295532,
0.06262116134166718,
-0.060897305607795715,
-0.13970817625522614,
-0.044138357043266296,
0.03737650439143181,
-0.02695734053850174,
0.0010530882282182574,
0.06304259598255157,
-0.040355149656534195,
0.17689870297908783,
0.027813667431473732,
-0.04348531365394592,
-0.09173320978879929,
0.05760841816663742,
-0.14841116964817047,
0.2779867947101593,
0.025322308763861656,
0.06305278837680817,
0.11509562283754349,
-0.022274641320109367,
-0.1497436761856079,
0.015726730227470398,
0.10988197475671768,
-0.0642421618103981,
0.07382756471633911,
0.15995332598686218,
0.007870757021009922,
0.13070426881313324,
0.06532178074121475,
-0.04532612860202789,
0.03705069422721863,
-0.0933978408575058,
-0.045079149305820465,
-0.127914160490036,
0.08077869564294815,
-0.0918973907828331,
0.1593232899904251,
0.11781656742095947,
-0.0726831778883934,
0.007786177564412355,
-0.02154405787587166,
0.09158952534198761,
0.014309068210422993,
0.11296945065259933,
0.009344341233372688,
-0.1933489441871643,
0.03936775401234627,
0.015302582643926144,
0.09600092470645905,
-0.20864363014698029,
-0.05405039340257645,
0.04063950106501579,
-0.022894294932484627,
-0.07125745713710785,
0.11842647939920425,
0.03379382565617561,
0.027111928910017014,
-0.037399426102638245,
-0.042318105697631836,
0.011778431013226509,
0.1492074877023697,
-0.1120511069893837,
-0.017204953357577324
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-classification | trtd56/practical_nlp_course_3 | [
"transformers",
"safetensors",
"distilbert",
"text-classification",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T07:35:19+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #distilbert #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #distilbert #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
48,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #distilbert #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.07302619516849518,
0.15942901372909546,
-0.0037264563143253326,
0.025167323648929596,
0.1172078475356102,
0.008749904111027718,
0.07480382919311523,
0.10722988098859787,
-0.02045559324324131,
0.12543776631355286,
0.039410512894392014,
0.10394789278507233,
0.11057476699352264,
0.19190818071365356,
-0.002767085563391447,
-0.20705340802669525,
0.06408926099538803,
-0.1160135567188263,
0.01611061953008175,
0.12252140045166016,
0.14288343489170074,
-0.10968661308288574,
0.07077977806329727,
-0.03890860080718994,
-0.02716642990708351,
-0.03298870474100113,
-0.06216486915946007,
-0.05656079575419426,
0.06601378321647644,
0.0591726079583168,
0.0694754421710968,
0.024590101093053818,
0.08098499476909637,
-0.28943222761154175,
0.019404316321015358,
0.07737266272306442,
0.0035657489206641912,
0.06202864274382591,
0.07916183769702911,
-0.07883433997631073,
0.10480276495218277,
-0.05524575710296631,
0.15781880915164948,
0.07489447295665741,
-0.0982506051659584,
-0.18013636767864227,
-0.08504346013069153,
0.09836910665035248,
0.17099453508853912,
0.05479501187801361,
-0.03633468225598335,
0.14033685624599457,
-0.08061470836400986,
0.01687866821885109,
0.06772492825984955,
-0.0681740939617157,
-0.05254287272691727,
0.05607175827026367,
0.07204149663448334,
0.09469678997993469,
-0.13314193487167358,
-0.00799859780818224,
0.04471778869628906,
0.01626773551106453,
0.10993628203868866,
0.023736488074064255,
0.12237779796123505,
0.02965191937983036,
-0.14525189995765686,
-0.06473848968744278,
0.11550955474376678,
0.035839054733514786,
-0.060212019830942154,
-0.24547284841537476,
-0.0033295010216534138,
-0.034091219305992126,
-0.026645053178071976,
-0.04267369583249092,
0.0422113798558712,
-0.030090559273958206,
0.09037542343139648,
0.006197172217071056,
-0.06834427267313004,
-0.051925547420978546,
0.09237229824066162,
0.06009524688124657,
0.026748334988951683,
-0.027690796181559563,
0.02246314287185669,
0.12014351040124893,
0.1042904481291771,
-0.11278197169303894,
-0.06414325535297394,
-0.06594311445951462,
-0.08891893923282623,
-0.04987602308392525,
0.034507665783166885,
0.07233929634094238,
0.045586712658405304,
0.20240989327430725,
0.004374812822788954,
0.051542725414037704,
0.027702245861291885,
0.0163713525980711,
0.06719023734331131,
0.06825068593025208,
-0.05008118972182274,
-0.12756529450416565,
-0.03760921210050583,
0.11946272850036621,
0.0017542883288115263,
-0.03342404216527939,
-0.0370357446372509,
0.06106860190629959,
0.04892996326088905,
0.12214437127113342,
0.0646679550409317,
0.018874529749155045,
-0.07587674260139465,
-0.046467430889606476,
0.18032068014144897,
-0.15827614068984985,
0.02335406094789505,
0.01725550927221775,
-0.0531628392636776,
-0.033948007971048355,
0.018618647009134293,
0.009236144833266735,
-0.029387421905994415,
0.1005609855055809,
-0.06605342775583267,
-0.04097803682088852,
-0.10903192311525345,
-0.055316012352705,
0.03402786701917648,
-0.025065315887331963,
-0.02789347805082798,
-0.040971264243125916,
-0.1248450055718422,
-0.07442043721675873,
0.06903292238712311,
-0.06419821083545685,
-0.06718063354492188,
-0.040221910923719406,
-0.06102044880390167,
0.014937658794224262,
0.0008266915683634579,
0.1269739717245102,
-0.02968521974980831,
0.04848431050777435,
-0.0539008304476738,
0.06914420425891876,
0.13718481361865997,
0.03272920101881027,
-0.06809919327497482,
0.06580659747123718,
-0.21232743561267853,
0.10933514684438705,
-0.09396476298570633,
0.026792176067829132,
-0.16038449108600616,
-0.02306288480758667,
0.03069896623492241,
0.039205435663461685,
-0.01574552245438099,
0.1448679268360138,
-0.1747746467590332,
-0.03626937046647072,
0.18672682344913483,
-0.12991686165332794,
-0.09265255182981491,
0.06183605268597603,
-0.0648084431886673,
0.13347013294696808,
0.05529501289129257,
-0.01992315798997879,
0.05587787926197052,
-0.13651202619075775,
-0.023517979308962822,
-0.058770496398210526,
-0.011057188734412193,
0.15450166165828705,
0.06303975731134415,
-0.04996807500720024,
0.024645399302244186,
0.017310835421085358,
-0.024148117750883102,
-0.04886231571435928,
-0.03430904448032379,
-0.09810014069080353,
0.005970593076199293,
-0.07982048392295837,
0.025509681552648544,
-0.02279755286872387,
-0.08887400478124619,
-0.040562164038419724,
-0.15593992173671722,
0.009587006643414497,
0.0986250564455986,
0.0006499737501144409,
-0.029481856152415276,
-0.09914560616016388,
0.0014640848385170102,
0.016265012323856354,
-0.010709897615015507,
-0.1529860496520996,
-0.05147454887628555,
0.025713054463267326,
-0.16740785539150238,
0.02983911894261837,
-0.04416975751519203,
0.03472619876265526,
0.04469497501850128,
-0.047529187053442,
-0.02975785918533802,
0.015605244785547256,
0.02078833244740963,
-0.024411868304014206,
-0.25051596760749817,
-0.013653411529958248,
-0.051656268537044525,
0.17981497943401337,
-0.25592783093452454,
0.04935307428240776,
0.0690855160355568,
0.12038503587245941,
0.005616906564682722,
-0.04484110698103905,
0.038755834102630615,
-0.05312656611204147,
-0.04079194739460945,
-0.06756321340799332,
-0.004968787543475628,
-0.03330003470182419,
-0.04708937928080559,
0.040533605962991714,
-0.18370530009269714,
-0.026839453727006912,
0.11585007607936859,
0.06803574413061142,
-0.17149686813354492,
-0.07743752747774124,
-0.034665726125240326,
-0.05996506288647652,
-0.08542647957801819,
-0.056485775858163834,
0.09173574298620224,
0.04302561655640602,
0.055119626224040985,
-0.07221351563930511,
-0.0563325397670269,
0.015307560563087463,
-0.011831860989332199,
-0.032375045120716095,
0.08966241031885147,
0.07603370398283005,
-0.12257120013237,
0.10713227838277817,
0.06915293633937836,
0.06829847395420074,
0.10371299833059311,
0.006018918938934803,
-0.0951351672410965,
-0.012076831422746181,
0.028954172506928444,
0.013578351587057114,
0.14422492682933807,
-0.07140666991472244,
0.03330845758318901,
0.04359918460249901,
-0.027328653261065483,
0.009608421474695206,
-0.10246647149324417,
0.018117014318704605,
0.03343784064054489,
-0.008881162852048874,
0.017250988632440567,
-0.05481864511966705,
0.014968239702284336,
0.10633815079927444,
0.03211374580860138,
0.027500580996274948,
0.01981731504201889,
-0.040416620671749115,
-0.12751449644565582,
0.1772654801607132,
-0.09383377432823181,
-0.2552470862865448,
-0.13026653230190277,
-0.009479007683694363,
0.045126691460609436,
-0.010854403488337994,
0.019198866561055183,
-0.05917074531316757,
-0.1081017553806305,
-0.10490734130144119,
0.026286281645298004,
0.054074980318546295,
-0.08816048502922058,
-0.064018115401268,
0.05169869586825371,
0.0385097898542881,
-0.12403316795825958,
0.021811455488204956,
0.046125855296850204,
-0.07025353610515594,
0.00821257010102272,
0.052987806499004364,
0.08472178876399994,
0.1826072335243225,
0.007897963747382164,
-0.016298603266477585,
0.008750800043344498,
0.2144501805305481,
-0.1484457403421402,
0.092045359313488,
0.14109621942043304,
-0.06516804546117783,
0.08377774804830551,
0.20131921768188477,
0.030504774302244186,
-0.09844772517681122,
0.03905881568789482,
0.03513709455728531,
-0.0375148244202137,
-0.24395905435085297,
-0.0748228207230568,
0.0031239830423146486,
-0.06623414903879166,
0.10724245756864548,
0.08736731112003326,
0.1171678826212883,
0.05268942564725876,
-0.11185546219348907,
-0.06449731439352036,
0.05344700068235397,
0.12066427618265152,
-0.028124094009399414,
0.0008641352178528905,
0.09650425612926483,
-0.02977217361330986,
0.02383269928395748,
0.09186029434204102,
0.018334977328777313,
0.1854310929775238,
0.04487955570220947,
0.1315774768590927,
0.08984522521495819,
0.06165572628378868,
0.01767764426767826,
0.01994951255619526,
0.022676948457956314,
0.028990833088755608,
-0.022242991253733635,
-0.0817873626947403,
-0.00921230111271143,
0.14159180223941803,
0.026489878073334694,
0.03602421656250954,
0.001440341817215085,
-0.04777481406927109,
0.07105493545532227,
0.16661210358142853,
0.012482628226280212,
-0.22979335486888885,
-0.06520283222198486,
0.07564391940832138,
-0.07074891030788422,
-0.11627703160047531,
-0.013096708804368973,
0.024812309071421623,
-0.18332423269748688,
0.04349841922521591,
-0.024669349193572998,
0.1018587276339531,
-0.11199972778558731,
-0.02344847284257412,
0.035318560898303986,
0.06107853353023529,
-0.035138774663209915,
0.07848566025495529,
-0.20783106982707977,
0.1402515470981598,
0.007242240011692047,
0.06469187885522842,
-0.10684854537248611,
0.08134520798921585,
0.020340995863080025,
0.006346969865262508,
0.1665121465921402,
-0.005634299945086241,
-0.072713203728199,
-0.09345488250255585,
-0.07864519953727722,
-0.017188850790262222,
0.0979963019490242,
-0.11784757673740387,
0.09015297889709473,
-0.007544329855591059,
-0.03196582943201065,
-0.0007019630284048617,
-0.12950846552848816,
-0.13376227021217346,
-0.18478168547153473,
0.04834262654185295,
-0.12510578334331512,
0.041554566472768784,
-0.10858581960201263,
-0.060765668749809265,
-0.041379012167453766,
0.19413886964321136,
-0.20414148271083832,
-0.08119912445545197,
-0.14911502599716187,
-0.0672706589102745,
0.11254695802927017,
-0.03948867693543434,
0.08191721886396408,
0.008871423080563545,
0.2073923498392105,
-0.004810879472643137,
0.0006135239964351058,
0.09140623360872269,
-0.09588538110256195,
-0.2094263732433319,
-0.0959051325917244,
0.13635295629501343,
0.13115985691547394,
0.04470321163535118,
0.00023247375793289393,
0.02411508932709694,
-0.0018883526790887117,
-0.11162916570901871,
0.03426937386393547,
0.15202432870864868,
0.10249507427215576,
0.044034719467163086,
-0.0260105412453413,
-0.13932733237743378,
-0.1056612879037857,
-0.054744839668273926,
0.013206261210143566,
0.1903214454650879,
-0.0706305131316185,
0.1657869964838028,
0.1536196768283844,
-0.06531279534101486,
-0.21233291923999786,
0.03679078444838524,
0.030905993655323982,
-0.00751135777682066,
0.04347773641347885,
-0.2047269195318222,
0.07352772355079651,
0.01412410382181406,
-0.05716951563954353,
0.1305869072675705,
-0.17576472461223602,
-0.14771407842636108,
0.09065452963113785,
0.07857703417539597,
-0.2075619101524353,
-0.12917637825012207,
-0.0950717106461525,
-0.05231890827417374,
-0.10034287720918655,
0.09251669049263,
-0.0036216825246810913,
0.005252200644463301,
0.036232154816389084,
0.01758572831749916,
0.01728934422135353,
-0.05098523199558258,
0.19524237513542175,
-0.00017524124996270984,
0.05021730437874794,
-0.07728931307792664,
-0.07839185744524002,
0.03842216357588768,
-0.06752927601337433,
0.08417709171772003,
-0.02161126770079136,
0.0039355861954391,
-0.11725787818431854,
-0.06764968484640121,
-0.04570414870977402,
0.03315238282084465,
-0.08949651569128036,
-0.09646400064229965,
-0.0555412657558918,
0.10287721455097198,
0.09537502378225327,
-0.03549838066101074,
-0.06785823404788971,
-0.09521738439798355,
0.05743926018476486,
0.2211635708808899,
0.18752726912498474,
0.07758046686649323,
-0.07665256410837173,
-0.008446265943348408,
-0.02362825535237789,
0.05575858801603317,
-0.2147134691476822,
0.04626009985804558,
0.03838435187935829,
0.030744675546884537,
0.1351434588432312,
-0.022784622386097908,
-0.16072605550289154,
-0.04722895845770836,
0.05541609972715378,
-0.07028964161872864,
-0.15762348473072052,
0.003693870734423399,
0.08388359844684601,
-0.15567344427108765,
-0.05364304408431053,
0.030349692329764366,
-0.03299986198544502,
-0.02724997140467167,
0.002993965055793524,
0.08165504038333893,
0.02525121532380581,
0.10604418069124222,
0.06794179975986481,
0.11212385445833206,
-0.10361232608556747,
0.07820820808410645,
0.08721207082271576,
-0.11143109202384949,
0.03750693425536156,
0.059706296771764755,
-0.06430401653051376,
-0.03306615725159645,
0.028105957433581352,
0.08702781051397324,
0.02858729287981987,
-0.07410863786935806,
0.0023060773964971304,
-0.11285153776407242,
0.06773319095373154,
0.13773435354232788,
0.037572041153907776,
0.009064391255378723,
0.04253077879548073,
0.030666319653391838,
-0.1025259718298912,
0.11677869409322739,
0.04715273529291153,
0.03828616067767143,
-0.053534768521785736,
-0.002754961373284459,
0.04357896372675896,
-0.015574077144265175,
-0.017309002578258514,
-0.03927738964557648,
-0.06638500094413757,
-0.009345067664980888,
-0.16059128940105438,
0.027963994070887566,
-0.06438141316175461,
0.011313637718558311,
0.015024027787148952,
-0.02930280566215515,
0.006326301023364067,
0.010901868343353271,
-0.07644513994455338,
-0.04005778953433037,
-0.0025265931617468596,
0.11033432930707932,
-0.16255317628383636,
0.006753581576049328,
0.08725008368492126,
-0.12882095575332642,
0.07888396829366684,
-0.003228981513530016,
-0.008663777261972427,
0.019871357828378677,
-0.1389452964067459,
0.06426677107810974,
-0.007317067123949528,
0.006886337883770466,
0.024405626580119133,
-0.20780570805072784,
0.002691886154934764,
-0.049495045095682144,
-0.06124653294682503,
-0.003442719578742981,
-0.03931323438882828,
-0.11277955025434494,
0.10321920365095139,
0.017737101763486862,
-0.08050814270973206,
-0.018862100318074226,
0.05358913913369179,
0.11278057098388672,
-0.053978823125362396,
0.14271147549152374,
-0.018007846549153328,
0.05715036392211914,
-0.1816556304693222,
-0.017987793311476707,
-0.017368610948324203,
0.016075139865279198,
-0.03470727428793907,
-0.008873502723872662,
0.05237460881471634,
-0.01958826184272766,
0.22800102829933167,
-0.023029034957289696,
0.01981639862060547,
0.06532696634531021,
0.0016252564964815974,
-0.010984939523041248,
0.09684767574071884,
0.048498742282390594,
0.015143456868827343,
0.0203377865254879,
0.013252451084554195,
-0.04566340893507004,
-0.008616970852017403,
-0.12847718596458435,
0.08234056085348129,
0.1677752137184143,
0.08175479620695114,
-0.006052352488040924,
0.047567352652549744,
-0.11316590011119843,
-0.09173060953617096,
0.10132203251123428,
-0.03303218260407448,
-0.013127516023814678,
-0.05242474004626274,
0.1442553550004959,
0.15683847665786743,
-0.1846613585948944,
0.0673123374581337,
-0.06864999234676361,
-0.058019280433654785,
-0.10558338463306427,
-0.17708730697631836,
-0.0631738007068634,
-0.033932529389858246,
-0.009048123843967915,
-0.060769032686948776,
0.06745719909667969,
0.10813924670219421,
0.01437336578965187,
0.004817943554371595,
0.08580505102872849,
-0.03281113877892494,
0.006333827041089535,
0.04443316161632538,
0.052908364683389664,
0.015542974695563316,
-0.06320759654045105,
0.004275370854884386,
0.006610610987991095,
0.0376921184360981,
0.055147934705019,
0.030873596668243408,
-0.0092905443161726,
0.007207514252513647,
-0.020693093538284302,
-0.10057692229747772,
0.04111333191394806,
-0.025823315605521202,
-0.047910936176776886,
0.1509503871202469,
0.020467912778258324,
-0.003414076054468751,
-0.022258523851633072,
0.2298767864704132,
-0.06479788571596146,
-0.07484833151102066,
-0.13822507858276367,
0.14135941863059998,
-0.03916965425014496,
0.05368134006857872,
0.049936410039663315,
-0.10397564619779587,
0.03804606944322586,
0.14477981626987457,
0.14261196553707123,
-0.034453462809324265,
0.008940902538597584,
0.009526451118290424,
0.004399977158755064,
-0.02350606769323349,
0.05356355383992195,
0.04485337436199188,
0.11325705051422119,
-0.06528755277395248,
0.09648586809635162,
-0.005538135301321745,
-0.09084830433130264,
-0.019364742562174797,
0.1391776204109192,
0.002899263286963105,
0.024846963584423065,
-0.08323919028043747,
0.12169293314218521,
-0.06053123623132706,
-0.2529715597629547,
0.06497339904308319,
-0.06441039592027664,
-0.1503337174654007,
-0.019829563796520233,
0.015622834675014019,
-0.0025740223936736584,
0.022466620430350304,
0.06178610026836395,
-0.06470615416765213,
0.15161879360675812,
0.03660573810338974,
-0.07138057053089142,
-0.07539889216423035,
0.07816334068775177,
-0.08136013150215149,
0.30430659651756287,
0.007375960238277912,
0.05443240702152252,
0.09480572491884232,
-0.03690790757536888,
-0.13316340744495392,
0.0335354208946228,
0.09097745269536972,
-0.047231536358594894,
0.06487338244915009,
0.20800761878490448,
-0.011225960217416286,
0.11401397734880447,
0.07447969168424606,
-0.08660271763801575,
0.05096733942627907,
-0.0917983278632164,
-0.09906064718961716,
-0.0893944799900055,
0.0902828648686409,
-0.059031881392002106,
0.1506001204252243,
0.12994202971458435,
-0.04605574533343315,
0.005047217011451721,
-0.022221196442842484,
0.05354851856827736,
-0.0026379574555903673,
0.11034536361694336,
0.03008626215159893,
-0.19489215314388275,
0.03033076599240303,
-0.00037526662345044315,
0.10122878104448318,
-0.25035029649734497,
-0.08561131358146667,
0.03936697542667389,
-0.007475157734006643,
-0.057129982858896255,
0.12413015216588974,
0.054405856877565384,
0.047805771231651306,
-0.05493326112627983,
-0.05230220779776573,
-0.007250586990267038,
0.1655176728963852,
-0.10096944123506546,
-0.0014428504509851336
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# RewardModel_RobertaBase
This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.5009
- F1: 0.7738
- Roc Auc: 0.7738
- Accuracy: 0.7698
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | F1 | Roc Auc | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:------:|:-------:|:--------:|
| No log | 1.0 | 93 | 0.5440 | 0.7331 | 0.7341 | 0.7302 |
| 0.648 | 2.0 | 186 | 0.5009 | 0.7738 | 0.7738 | 0.7698 |
| 0.5515 | 3.0 | 279 | 0.4938 | 0.7545 | 0.7560 | 0.75 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "mit", "tags": ["generated_from_trainer"], "metrics": ["f1", "accuracy"], "base_model": "roberta-base", "model-index": [{"name": "RewardModel_RobertaBase", "results": []}]} | text-classification | RajuEEE/RewardModel_RobertaBase | [
"transformers",
"tensorboard",
"safetensors",
"roberta",
"text-classification",
"generated_from_trainer",
"base_model:roberta-base",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T07:40:00+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #roberta #text-classification #generated_from_trainer #base_model-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us
| RewardModel\_RobertaBase
========================
This model is a fine-tuned version of roberta-base on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.5009
* F1: 0.7738
* Roc Auc: 0.7738
* Accuracy: 0.7698
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #roberta #text-classification #generated_from_trainer #base_model-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
63,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #roberta #text-classification #generated_from_trainer #base_model-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.09373882412910461,
0.07018259912729263,
-0.0017258531879633665,
0.10201310366392136,
0.1709260493516922,
0.01937614567577839,
0.1633339524269104,
0.10257398337125778,
-0.09508547186851501,
0.045202918350696564,
0.12393975257873535,
0.14906087517738342,
-0.001141179003752768,
0.12444236129522324,
-0.07543622702360153,
-0.23876160383224487,
0.0013685035519301891,
0.030273087322711945,
-0.0894056037068367,
0.11395230889320374,
0.10148957371711731,
-0.13388310372829437,
0.08646070212125778,
-0.012859152629971504,
-0.20780037343502045,
0.02956921048462391,
0.04382127895951271,
-0.05731358379125595,
0.14230020344257355,
0.04548192396759987,
0.14787669479846954,
0.033983658999204636,
0.08708725869655609,
-0.19165241718292236,
0.018879028037190437,
0.05999885872006416,
-0.014579074457287788,
0.0855783075094223,
0.04522363841533661,
-0.030190907418727875,
0.09458091855049133,
-0.09394030272960663,
0.06502729654312134,
0.024751000106334686,
-0.1354130506515503,
-0.18897195160388947,
-0.07449539005756378,
0.03463162109255791,
0.08625391125679016,
0.08124381303787231,
-0.01819031499326229,
0.1398644894361496,
-0.0679207295179367,
0.09572023153305054,
0.20540255308151245,
-0.3081384599208832,
-0.06944693624973297,
0.06106884032487869,
0.026300452649593353,
0.08839083462953568,
-0.11551932245492935,
-0.00033925511525012553,
0.07825645804405212,
0.019451474770903587,
0.11718824505805969,
-0.03312164545059204,
-0.06642556935548782,
0.014713405631482601,
-0.14314214885234833,
0.004298370331525803,
0.13327844440937042,
0.04192034900188446,
-0.04315769299864769,
-0.03659830242395401,
-0.0586593896150589,
-0.13667884469032288,
-0.041744112968444824,
-0.026217401027679443,
0.04146284982562065,
-0.028280865401029587,
-0.10367295891046524,
-0.018155556172132492,
-0.11400198191404343,
-0.08044528216123581,
-0.062128838151693344,
0.17377541959285736,
0.03427509590983391,
-0.0007933186134323478,
-0.024326790124177933,
0.09534341096878052,
-0.03579001873731613,
-0.12332232296466827,
0.021965421736240387,
0.02619268372654915,
-0.006273171864449978,
-0.0732455849647522,
-0.0655897781252861,
-0.08780793845653534,
0.02541438303887844,
0.15293020009994507,
-0.059917330741882324,
0.046599265187978745,
0.017280390486121178,
0.04579300805926323,
-0.09711708128452301,
0.16025204956531525,
-0.057647328823804855,
-0.01852126605808735,
0.024835243821144104,
0.06324584782123566,
0.034284960478544235,
-0.0028902338817715645,
-0.1239895150065422,
0.013214895501732826,
0.11581912636756897,
0.017588943243026733,
-0.07545897364616394,
0.08391236513853073,
-0.042739078402519226,
0.005324601661413908,
0.010028446093201637,
-0.08933284133672714,
0.02927257865667343,
0.005272254813462496,
-0.05889496952295303,
-0.061657849699258804,
0.029435984790325165,
0.01566355675458908,
0.017506862059235573,
0.10719793289899826,
-0.09799482673406601,
0.020201385021209717,
-0.09235439449548721,
-0.13413070142269135,
0.006978468503803015,
-0.06818676739931107,
0.032240256667137146,
-0.11608883738517761,
-0.15162542462348938,
-0.019727738574147224,
0.04308861121535301,
-0.027858207002282143,
-0.01635660231113434,
-0.05674222111701965,
-0.0819830596446991,
0.010535611771047115,
-0.010500628501176834,
0.08932168781757355,
-0.055144038051366806,
0.09909746795892715,
0.057345446199178696,
0.06807012856006622,
-0.055060967803001404,
0.033790312707424164,
-0.10016993433237076,
0.02008179947733879,
-0.21053680777549744,
0.021917499601840973,
-0.06402577459812164,
0.07385976612567902,
-0.07679137587547302,
-0.07890663295984268,
-0.007231888826936483,
0.024717651307582855,
0.07672986388206482,
0.09309037029743195,
-0.15364451706409454,
-0.07216495275497437,
0.168548583984375,
-0.0981474369764328,
-0.134349524974823,
0.12071125954389572,
-0.06998039036989212,
0.06697141379117966,
0.06903176009654999,
0.18188366293907166,
0.06764013320207596,
-0.08882448077201843,
0.003841668600216508,
-0.011767532676458359,
0.042900193482637405,
-0.04883940517902374,
0.05661807581782341,
0.004772299434989691,
-0.00762628810480237,
0.010473229922354221,
-0.030047357082366943,
0.052933648228645325,
-0.09686901420354843,
-0.07917876541614532,
-0.035802632570266724,
-0.11050312221050262,
0.056224193423986435,
0.059637509286403656,
0.0770082175731659,
-0.1252211034297943,
-0.08627196401357651,
0.10153578966856003,
0.07430264353752136,
-0.0705232173204422,
0.016678689047694206,
-0.07443589717149734,
0.07421053946018219,
-0.06706999987363815,
-0.03213009238243103,
-0.16332975029945374,
-0.057601574808359146,
-0.006213085725903511,
0.03601903095841408,
0.0313536636531353,
0.015948791056871414,
0.08043375611305237,
0.07630778104066849,
-0.07379975914955139,
-0.027321988716721535,
-0.003479655599221587,
0.015064068138599396,
-0.12822319567203522,
-0.2097322642803192,
-0.006460902746766806,
-0.03681747615337372,
0.15102826058864594,
-0.2583620846271515,
0.05415629222989082,
-0.0016765603795647621,
0.08388826996088028,
0.04006648063659668,
0.0002731861313804984,
-0.04011308774352074,
0.07656420767307281,
-0.049669623374938965,
-0.05483416095376015,
0.050541605800390244,
0.0033116997219622135,
-0.08934319764375687,
-0.038019806146621704,
-0.15937112271785736,
0.1980612874031067,
0.14025461673736572,
-0.1157793328166008,
-0.11183708161115646,
0.003289021784439683,
-0.04013761505484581,
-0.022547954693436623,
-0.05059899017214775,
0.012092139571905136,
0.12241854518651962,
-0.023849748075008392,
0.15147778391838074,
-0.06979215145111084,
-0.032694850116968155,
0.020706545561552048,
-0.06180998682975769,
0.014778048731386662,
0.11152289062738419,
0.10131791979074478,
-0.11897477507591248,
0.15242885053157806,
0.14526396989822388,
-0.11241286247968674,
0.14786162972450256,
-0.029510632157325745,
-0.056375421583652496,
-0.021122585982084274,
0.0030578637961298227,
0.017933106049895287,
0.11596319824457169,
-0.10682672262191772,
-0.010856210254132748,
-0.002567627467215061,
0.005183768924325705,
0.022982103750109673,
-0.2253468632698059,
-0.04189823940396309,
0.0325823649764061,
-0.03297324478626251,
0.014202611520886421,
-0.02355038933455944,
-0.006162881385535002,
0.10547137260437012,
-0.0018219944322481751,
-0.07934758067131042,
0.03855657950043678,
0.0009786243317648768,
-0.0916619524359703,
0.22430738806724548,
-0.07018067687749863,
-0.11027919501066208,
-0.13504423201084137,
-0.05791395157575607,
-0.03997121378779411,
0.03380218520760536,
0.05692209675908089,
-0.08647309988737106,
-0.0353410467505455,
-0.0931202620267868,
0.016347942873835564,
0.024809477850794792,
0.03382374346256256,
-0.011710588820278645,
0.01370810903608799,
0.07876428961753845,
-0.10560913383960724,
-0.0041988305747509,
-0.05743642896413803,
-0.07816098630428314,
0.04491204768419266,
0.028418438509106636,
0.12241573631763458,
0.1538962870836258,
-0.04167892411351204,
-0.007699419744312763,
-0.04037807136774063,
0.21819175779819489,
-0.06740274280309677,
-0.01961972564458847,
0.11540335416793823,
-0.020141271874308586,
0.03903917968273163,
0.1366535723209381,
0.06673532724380493,
-0.0987861156463623,
0.03335164487361908,
0.045948777347803116,
-0.032028503715991974,
-0.21919235587120056,
-0.03123149648308754,
-0.03129665553569794,
-0.01160092931240797,
0.08377633988857269,
0.035835787653923035,
0.04408920928835869,
0.08141378313302994,
0.03243423253297806,
0.07252753525972366,
-0.014539295807480812,
0.07030351459980011,
0.09847883135080338,
0.04630044102668762,
0.13561972975730896,
-0.05746789276599884,
-0.07542263716459274,
0.02600897289812565,
-0.01442425511777401,
0.19699200987815857,
0.022814664989709854,
0.1171041801571846,
0.05876244604587555,
0.1459486335515976,
0.011778706684708595,
0.06700699031352997,
0.00027739451616071165,
-0.06572216749191284,
0.00022258654644247144,
-0.04765701666474342,
-0.01648285798728466,
0.0366230271756649,
-0.09139755368232727,
0.0551845021545887,
-0.11918995529413223,
0.012536914087831974,
0.06536757200956345,
0.2000620812177658,
0.058761607855558395,
-0.3271774351596832,
-0.09925386309623718,
0.02197856828570366,
-0.01632809452712536,
-0.015571444295346737,
0.022952085360884666,
0.12121009081602097,
-0.04656165465712547,
0.028606988489627838,
-0.06662972271442413,
0.07557319849729538,
-0.03139039874076843,
0.04008318856358528,
0.04136588051915169,
0.10224351286888123,
-0.030037520453333855,
0.06605198979377747,
-0.2869971990585327,
0.28074705600738525,
0.019785255193710327,
0.08328815549612045,
-0.04230407252907753,
-0.015317193232476711,
0.0324607752263546,
0.09799319505691528,
0.06175024434924126,
-0.027615852653980255,
-0.08418594300746918,
-0.21020570397377014,
-0.03350641950964928,
0.0289694145321846,
0.11445201933383942,
-0.02581067569553852,
0.10968644171953201,
-0.03375612571835518,
0.004423551727086306,
0.0971614271402359,
-0.03331423178315163,
-0.09054809808731079,
-0.08771692961454391,
-0.032372742891311646,
0.018570352345705032,
-0.030239608138799667,
-0.08084264397621155,
-0.10365689545869827,
-0.13115598261356354,
0.16930320858955383,
-0.036741141229867935,
-0.012729907408356667,
-0.10901793837547302,
0.0870051309466362,
0.045793306082487106,
-0.08592052757740021,
0.04780092090368271,
0.016065575182437897,
0.08153633028268814,
0.01874871365725994,
-0.05956201255321503,
0.13533571362495422,
-0.06702467054128647,
-0.16703510284423828,
-0.06704700738191605,
0.09469249099493027,
0.024941204115748405,
0.04411697760224342,
0.0007678403053432703,
0.012681377120316029,
-0.007461427245289087,
-0.07887478172779083,
0.02854038216173649,
-0.028546009212732315,
0.05351853370666504,
0.02323789894580841,
-0.06470664590597153,
-0.018266096711158752,
-0.06010375916957855,
-0.028364941477775574,
0.16767460107803345,
0.29009780287742615,
-0.08946128934621811,
-0.020182475447654724,
0.058877598494291306,
-0.06535152345895767,
-0.22278092801570892,
0.07816309481859207,
0.0165525134652853,
0.005272659473121166,
0.05072006210684776,
-0.14107675850391388,
0.12254244089126587,
0.10101060569286346,
-0.023566273972392082,
0.09887875616550446,
-0.27038154006004333,
-0.1449773758649826,
0.13684332370758057,
0.17360550165176392,
0.1305563747882843,
-0.16740688681602478,
-0.028277531266212463,
-0.039162494242191315,
-0.08739452064037323,
0.10421086102724075,
-0.1450052559375763,
0.11268087476491928,
-0.0007786698406562209,
0.05591399595141411,
0.006640353240072727,
-0.058769069612026215,
0.11979648470878601,
-0.012654376216232777,
0.127639040350914,
-0.0643840879201889,
-0.04260893911123276,
0.06553687900304794,
-0.04449276626110077,
0.012948448769748211,
-0.09054006636142731,
0.033109527081251144,
-0.03950582072138786,
-0.03423316031694412,
-0.05454662814736366,
0.04774418845772743,
-0.0397665798664093,
-0.07199177891016006,
-0.05106927827000618,
0.032967016100883484,
0.01737184077501297,
-0.01699371077120304,
0.1565360128879547,
-0.005812722258269787,
0.17884862422943115,
0.1268020123243332,
0.08847977966070175,
-0.04829975590109825,
0.007005832623690367,
0.014146102592349052,
-0.042484913021326065,
0.05986150726675987,
-0.15999145805835724,
0.03674427419900894,
0.11473383009433746,
0.015982119366526604,
0.15145215392112732,
0.08485907316207886,
-0.03709380328655243,
0.02559431828558445,
0.08582992851734161,
-0.160331130027771,
-0.1126520186662674,
-0.006826663389801979,
-0.06441926956176758,
-0.10604707896709442,
0.079326331615448,
0.118343785405159,
-0.07785560935735703,
0.0012756469659507275,
-0.013586741872131824,
-0.0016199147794395685,
-0.05151192843914032,
0.1856248825788498,
0.09926223754882812,
0.04740285500884056,
-0.07200188934803009,
0.07025835663080215,
0.04423109069466591,
-0.06486254930496216,
0.010816583409905434,
0.05203075706958771,
-0.07278124243021011,
-0.05112501233816147,
0.06961476057767868,
0.2116849571466446,
-0.062455445528030396,
-0.04764954000711441,
-0.16130295395851135,
-0.11383278667926788,
0.041200801730155945,
0.22283785045146942,
0.10008261352777481,
0.006688037421554327,
-0.021594662219285965,
0.025992028415203094,
-0.14021077752113342,
0.10440990328788757,
0.033301785588264465,
0.09025460481643677,
-0.15042918920516968,
0.1686716079711914,
-0.005546333733946085,
0.007726375013589859,
-0.03334248438477516,
0.04438682645559311,
-0.13833051919937134,
-0.0016990398289635777,
-0.13116075098514557,
-0.01798604428768158,
-0.03202866017818451,
0.005589840468019247,
0.00839859526604414,
-0.06036832928657532,
-0.07298502326011658,
0.0074756997637450695,
-0.10575123131275177,
-0.00967662874609232,
0.043647538870573044,
0.0488039068877697,
-0.12922105193138123,
-0.0344497486948967,
0.019121350720524788,
-0.06170383468270302,
0.06250397861003876,
0.018987074494361877,
0.02971678227186203,
0.06796275079250336,
-0.2043450027704239,
0.033510416746139526,
0.07110732793807983,
-0.010897410102188587,
0.05252470821142197,
-0.07481835782527924,
-0.006415215320885181,
-0.010326402261853218,
0.07289522886276245,
0.023196475580334663,
0.07365758717060089,
-0.12151110917329788,
0.02161586843430996,
-0.04340115934610367,
-0.06696807593107224,
-0.059685517102479935,
0.03363588824868202,
0.08047094196081161,
-0.006967449095100164,
0.1920611411333084,
-0.1094469428062439,
0.009848746471107006,
-0.2099749594926834,
0.00679992837831378,
-0.010750231333076954,
-0.1126059889793396,
-0.11291332542896271,
-0.05474075302481651,
0.056844014674425125,
-0.06013955548405647,
0.1440676748752594,
0.019870014861226082,
0.025159072130918503,
0.041001077741384506,
-0.03523211181163788,
0.04419238120317459,
0.043387725949287415,
0.2341066598892212,
0.028850888833403587,
-0.03970082849264145,
0.005689963698387146,
0.05789414048194885,
0.11753252148628235,
0.07435894757509232,
0.1824411153793335,
0.16428223252296448,
-0.07740338146686554,
0.10383164137601852,
0.04746042191982269,
-0.05034247785806656,
-0.10359599441289902,
0.05974293872714043,
-0.047598786652088165,
0.060570839792490005,
-0.026714658364653587,
0.1755073219537735,
0.1310650110244751,
-0.14900106191635132,
0.010590322315692902,
-0.06279108673334122,
-0.0831713080406189,
-0.12213876843452454,
-0.03612959757447243,
-0.10635454207658768,
-0.14828146994113922,
0.006284309551119804,
-0.11368397623300552,
-0.004442843608558178,
0.100802943110466,
0.00829640869051218,
-0.01825559325516224,
0.1868530511856079,
0.02146235853433609,
0.041865888983011246,
0.04355306178331375,
-0.0016038956819102168,
-0.0299723818898201,
-0.08774278312921524,
-0.08196187019348145,
0.005543894600123167,
-0.03287115320563316,
0.014677242375910282,
-0.05569355934858322,
-0.03762190043926239,
0.048447929322719574,
-0.016988450661301613,
-0.10721141844987869,
0.01549484208226204,
0.04016749933362007,
0.053429797291755676,
0.03788156062364578,
0.010375593788921833,
0.001887332065962255,
-0.004200102295726538,
0.23178938031196594,
-0.07053878903388977,
-0.07595696300268173,
-0.09803228080272675,
0.2541307806968689,
0.04162134975194931,
0.031038878485560417,
0.00033371380413882434,
-0.1046786680817604,
0.04426239803433418,
0.22354766726493835,
0.20859064161777496,
-0.08328669518232346,
0.01546599343419075,
-0.013500235974788666,
-0.010300575755536556,
-0.03267408534884453,
0.09668884426355362,
0.10130433738231659,
0.021478675305843353,
-0.08370475471019745,
-0.056252621114254,
-0.03251209855079651,
-0.004530253820121288,
-0.03663622960448265,
0.06112964078783989,
0.046331144869327545,
0.022691190242767334,
-0.05143909901380539,
0.06409921497106552,
-0.042782071977853775,
-0.11741778999567032,
0.06339316815137863,
-0.19959035515785217,
-0.14396563172340393,
-0.019642602652311325,
0.10631020367145538,
-0.015604009851813316,
0.06447510421276093,
-0.036354538053274155,
-0.009930199943482876,
0.03150143101811409,
-0.017200060188770294,
-0.07686488330364227,
-0.06721015274524689,
0.056656476110219955,
-0.08917520195245743,
0.21312372386455536,
-0.04664086177945137,
0.04045092687010765,
0.1443788707256317,
0.041046615689992905,
-0.06752461194992065,
0.10147040337324142,
0.04185163974761963,
-0.07850762456655502,
0.04117405414581299,
0.08889197558164597,
-0.04831518977880478,
0.1045442596077919,
0.05436667054891586,
-0.15492720901966095,
0.03749195486307144,
-0.09204734861850739,
-0.08836627006530762,
-0.0606393963098526,
-0.03996272757649422,
-0.06251857429742813,
0.1275290995836258,
0.1888735592365265,
-0.030511973425745964,
0.024951571598649025,
-0.047104932367801666,
0.020924778655171394,
0.07843248546123505,
0.04158955439925194,
-0.04449788108468056,
-0.2473830282688141,
0.03190385922789574,
0.08339628577232361,
-0.015324379317462444,
-0.3083210289478302,
-0.07171683758497238,
-0.007791276555508375,
-0.04802301898598671,
-0.09636399894952774,
0.0889403373003006,
0.13900849223136902,
0.06271837651729584,
-0.0621466301381588,
-0.11781445145606995,
-0.07734454423189163,
0.16548651456832886,
-0.13382549583911896,
-0.10848048329353333
] |
null | null | transformers | # WestLake 7B v2 laser - AWQ
- Model creator: [Common Sense](https://huggingface.co/senseable)
- Original model: [WestLake 7B v2](https://huggingface.co/senseable/WestLake-7B-v2)
- Fine Tuning: [cognitivecomputations](https://huggingface.co/cognitivecomputations/WestLake-7B-v2-laser)
It follows the implementation of [laserRMT](https://github.com/cognitivecomputations/laserRMT)

## Model description
This repo contains AWQ model files for [Common Sense's WestLake 7B v2](https://huggingface.co/senseable/WestLake-7B-v2).
These files were quantised using hardware kindly provided by [SolidRusT Networks](https://solidrust.net/).
## How to use
### Install the necessary packages
```bash
pip install --upgrade autoawq autoawq-kernels
```
### Example Python code
```bash
from awq import AutoAWQForCausalLM
from transformers import AutoTokenizer, TextStreamer
quant_path = "/srv/home/shaun/repos/samantha-1.1-westlake-7b-laser-AWQ"
# Load model
model = AutoAWQForCausalLM.from_quantized(quant_path,
fuse_layers=True)
tokenizer = AutoTokenizer.from_pretrained(quant_path,
trust_remote_code=True)
streamer = TextStreamer(tokenizer,
skip_prompt=True,
skip_special_tokens=True)
# Convert prompt to tokens
prompt_template = """\
<|system|>
</s>
<|user|>
{prompt}</s>
<|assistant|>"""
prompt = "You're standing on the surface of the Earth. "\
"You walk one mile south, one mile west and one mile north. "\
"You end up exactly where you started. Where are you?"
tokens = tokenizer(prompt_template.format(prompt=prompt),
return_tensors='pt').input_ids.cuda()
# Generate output
generation_output = model.generate(tokens,
streamer=streamer,
max_new_tokens=512)
```
### About AWQ
AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.
AWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead.
It is supported by:
- [Text Generation Webui](https://github.com/oobabooga/text-generation-webui) - using Loader: AutoAWQ
- [vLLM](https://github.com/vllm-project/vllm) - version 0.2.2 or later for support for all model types.
- [Hugging Face Text Generation Inference (TGI)](https://github.com/huggingface/text-generation-inference)
- [Transformers](https://huggingface.co/docs/transformers) version 4.35.0 and later, from any code or client that supports Transformers
- [AutoAWQ](https://github.com/casper-hansen/AutoAWQ) - for use from Python code
## Prompt template: ChatML
```plaintext
<|im_start|>system
{system_message}<|im_end|>
<|im_start|>user
{prompt}<|im_end|>
<|im_start|>assistant
```
Also working with Basic Mistral format:
```plaintext
<|system|>
</s>
<|user|>
{prompt}</s>
<|assistant|>
```
| {"language": ["en"], "license": "apache-2.0", "library_name": "transformers", "tags": ["finetuned", "quantized", "4-bit", "AWQ", "transformers", "pytorch", "mistral", "text-generation", "conversational", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us"], "model_name": "WestLake 7B v2", "base_model": "senseable/WestLake-7B-v2", "model_creator": "Common Sense", "model_type": "mistral", "pipeline_tag": "text-generation", "prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant\n", "quantized_by": "Suparious"} | text-generation | solidrust/WestLake-7B-v2-laser-AWQ | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"finetuned",
"quantized",
"4-bit",
"AWQ",
"pytorch",
"conversational",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us",
"en",
"base_model:senseable/WestLake-7B-v2"
] | 2024-02-08T07:40:37+00:00 | [] | [
"en"
] | TAGS
#transformers #safetensors #mistral #text-generation #finetuned #quantized #4-bit #AWQ #pytorch #conversational #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us #en #base_model-senseable/WestLake-7B-v2
| # WestLake 7B v2 laser - AWQ
- Model creator: Common Sense
- Original model: WestLake 7B v2
- Fine Tuning: cognitivecomputations
It follows the implementation of laserRMT
!image/png
## Model description
This repo contains AWQ model files for Common Sense's WestLake 7B v2.
These files were quantised using hardware kindly provided by SolidRusT Networks.
## How to use
### Install the necessary packages
### Example Python code
### About AWQ
AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.
AWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead.
It is supported by:
- Text Generation Webui - using Loader: AutoAWQ
- vLLM - version 0.2.2 or later for support for all model types.
- Hugging Face Text Generation Inference (TGI)
- Transformers version 4.35.0 and later, from any code or client that supports Transformers
- AutoAWQ - for use from Python code
## Prompt template: ChatML
Also working with Basic Mistral format:
| [
"# WestLake 7B v2 laser - AWQ\n\n- Model creator: Common Sense\n- Original model: WestLake 7B v2\n- Fine Tuning: cognitivecomputations\n\nIt follows the implementation of laserRMT\n\n!image/png",
"## Model description\n\nThis repo contains AWQ model files for Common Sense's WestLake 7B v2.\n\nThese files were quantised using hardware kindly provided by SolidRusT Networks.",
"## How to use",
"### Install the necessary packages",
"### Example Python code",
"### About AWQ\n\nAWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.\n\nAWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead.\n\nIt is supported by:\n\n- Text Generation Webui - using Loader: AutoAWQ\n- vLLM - version 0.2.2 or later for support for all model types.\n- Hugging Face Text Generation Inference (TGI)\n- Transformers version 4.35.0 and later, from any code or client that supports Transformers\n- AutoAWQ - for use from Python code",
"## Prompt template: ChatML\n\n\n\nAlso working with Basic Mistral format:"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #finetuned #quantized #4-bit #AWQ #pytorch #conversational #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us #en #base_model-senseable/WestLake-7B-v2 \n",
"# WestLake 7B v2 laser - AWQ\n\n- Model creator: Common Sense\n- Original model: WestLake 7B v2\n- Fine Tuning: cognitivecomputations\n\nIt follows the implementation of laserRMT\n\n!image/png",
"## Model description\n\nThis repo contains AWQ model files for Common Sense's WestLake 7B v2.\n\nThese files were quantised using hardware kindly provided by SolidRusT Networks.",
"## How to use",
"### Install the necessary packages",
"### Example Python code",
"### About AWQ\n\nAWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.\n\nAWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead.\n\nIt is supported by:\n\n- Text Generation Webui - using Loader: AutoAWQ\n- vLLM - version 0.2.2 or later for support for all model types.\n- Hugging Face Text Generation Inference (TGI)\n- Transformers version 4.35.0 and later, from any code or client that supports Transformers\n- AutoAWQ - for use from Python code",
"## Prompt template: ChatML\n\n\n\nAlso working with Basic Mistral format:"
] | [
94,
51,
40,
4,
7,
6,
180,
16
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #finetuned #quantized #4-bit #AWQ #pytorch #conversational #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us #en #base_model-senseable/WestLake-7B-v2 \n# WestLake 7B v2 laser - AWQ\n\n- Model creator: Common Sense\n- Original model: WestLake 7B v2\n- Fine Tuning: cognitivecomputations\n\nIt follows the implementation of laserRMT\n\n!image/png## Model description\n\nThis repo contains AWQ model files for Common Sense's WestLake 7B v2.\n\nThese files were quantised using hardware kindly provided by SolidRusT Networks.## How to use### Install the necessary packages### Example Python code### About AWQ\n\nAWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.\n\nAWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead.\n\nIt is supported by:\n\n- Text Generation Webui - using Loader: AutoAWQ\n- vLLM - version 0.2.2 or later for support for all model types.\n- Hugging Face Text Generation Inference (TGI)\n- Transformers version 4.35.0 and later, from any code or client that supports Transformers\n- AutoAWQ - for use from Python code## Prompt template: ChatML\n\n\n\nAlso working with Basic Mistral format:"
] | [
-0.0980864092707634,
0.12693631649017334,
-0.0013324454193934798,
0.04546245187520981,
0.0433187372982502,
0.00407678447663784,
-0.002246956340968609,
0.10694032162427902,
0.02901236154139042,
0.09408833086490631,
-0.0423109196126461,
0.01920299604535103,
0.08753874897956848,
0.07817245274782181,
0.05041585490107536,
-0.15843163430690765,
0.044295575469732285,
-0.027123086154460907,
0.025082455947995186,
0.047732509672641754,
0.09485828131437302,
-0.03027651458978653,
0.070233553647995,
0.01129111647605896,
-0.06392426043748856,
-0.023896474391222,
-0.012797595001757145,
0.0022156541235744953,
0.04896502196788788,
0.08367230743169785,
-0.028792155906558037,
0.007101691327989101,
0.049071766436100006,
-0.13768504559993744,
0.007125254720449448,
0.09443359076976776,
0.011238876730203629,
0.050749797374010086,
0.08719843626022339,
0.011164016090333462,
0.09920549392700195,
-0.1299535036087036,
0.027583610266447067,
0.07036704570055008,
-0.025715477764606476,
-0.1720079630613327,
-0.05081258714199066,
-0.002304266905412078,
0.056978482753038406,
0.04151855409145355,
0.011180075816810131,
0.0743548721075058,
-0.05355340614914894,
0.06088464707136154,
0.10722341388463974,
-0.18349087238311768,
-0.01809404045343399,
0.10923239588737488,
0.04892860725522041,
0.0977168083190918,
-0.08987821638584137,
0.037225086241960526,
0.015671849250793457,
-0.013409359380602837,
0.057395294308662415,
0.011627218686044216,
0.08923277258872986,
0.007437821011990309,
-0.11019620299339294,
-0.028212865814566612,
0.058589670807123184,
0.020106010138988495,
-0.05304073542356491,
-0.11634135991334915,
-0.08578556776046753,
-0.047633446753025055,
0.0004867137467954308,
-0.022985495626926422,
0.02197396010160446,
0.023580551147460938,
0.03770197555422783,
-0.07072444260120392,
-0.04054437205195427,
-0.036240287125110626,
-0.049757834523916245,
0.1003926694393158,
0.037350740283727646,
0.041501179337501526,
0.06455139815807343,
0.10504895448684692,
-0.22571732103824615,
-0.08678664267063141,
-0.09543044120073318,
-0.03542377054691315,
-0.09936686605215073,
0.0179402194917202,
-0.01118266861885786,
0.052862510085105896,
-0.020906629040837288,
0.1354508250951767,
-0.04388510435819626,
0.051305271685123444,
0.050826285034418106,
-0.02564132958650589,
-0.05973349139094353,
0.1360923945903778,
-0.02860494889318943,
-0.20460663735866547,
0.06978961825370789,
0.0023339316248893738,
0.09047587215900421,
-0.01348428800702095,
-0.058674056082963943,
-0.09736448526382446,
-0.001392486272379756,
0.02126491814851761,
-0.009565480053424835,
0.010800722055137157,
-0.08155101537704468,
-0.08469688892364502,
0.22445005178451538,
-0.08218057453632355,
0.07652883231639862,
0.02112036570906639,
-0.04068886488676071,
0.06340818107128143,
0.06855854392051697,
-0.011299327947199345,
-0.048755235970020294,
0.031152985990047455,
-0.006254993844777346,
-0.02456030622124672,
-0.10117457807064056,
-0.10652483999729156,
0.045312199741601944,
-0.02218038961291313,
-0.02545573189854622,
-0.07740100473165512,
-0.21224388480186462,
0.008413955569267273,
-0.005606673192232847,
-0.05672486498951912,
-0.029379526153206825,
0.036017484962940216,
-0.041216541081666946,
0.012533674016594887,
0.0343807190656662,
0.07650920748710632,
-0.052922703325748444,
0.02991456165909767,
0.03812054917216301,
0.08855129033327103,
-0.049180321395397186,
0.007024036254733801,
-0.017288116738200188,
-0.005444840528070927,
-0.1152411624789238,
0.13459748029708862,
-0.10082878917455673,
0.03046996332705021,
-0.0710475891828537,
-0.03215065225958824,
-0.011090769432485104,
-0.05774817615747452,
0.015329143032431602,
0.04259735345840454,
-0.18648211658000946,
-0.033553704619407654,
0.04713241755962372,
-0.1300065666437149,
-0.002499829512089491,
0.15656732022762299,
0.035608790814876556,
-0.034518785774707794,
0.1288319081068039,
0.015926875174045563,
0.2616461515426636,
-0.15831977128982544,
-0.07033351808786392,
-0.00827767327427864,
0.027153698727488518,
0.018850650638341904,
0.10260064154863358,
0.021258093416690826,
-0.018254151567816734,
0.045391879975795746,
-0.06153706833720207,
0.013235130347311497,
0.021460959687829018,
-0.05524250864982605,
-0.0872487723827362,
-0.09943564236164093,
0.03231191262602806,
0.004505580756813288,
-0.04881918430328369,
-0.02608717978000641,
-0.09495566040277481,
-0.03897658362984657,
0.16435663402080536,
-0.007214790675789118,
-0.036768633872270584,
-0.10837627947330475,
0.07982797175645828,
-0.09753567725419998,
0.007404322270303965,
-0.0856071412563324,
-0.05069059878587723,
0.0940369963645935,
-0.10587803274393082,
-0.032047100365161896,
-0.046507205814123154,
0.06772710382938385,
0.08447615057229996,
-0.059587374329566956,
0.018436236307024956,
0.006566643714904785,
-0.009196829050779343,
-0.0558767169713974,
-0.07703624665737152,
-0.023969674482941628,
-0.0224740132689476,
0.05836932361125946,
-0.13666708767414093,
0.03556562215089798,
0.15362973511219025,
0.1083020269870758,
0.02930537797510624,
-0.07414055615663528,
0.017494332045316696,
-0.03083542175590992,
0.027394060045480728,
-0.10036875307559967,
0.02254851721227169,
0.010099709965288639,
-0.032282505184412,
0.06391789019107819,
-0.14838847517967224,
0.02185671590268612,
0.07583966106176376,
0.22406774759292603,
0.016823826357722282,
0.01916959322988987,
-0.006926196161657572,
-0.06178800016641617,
-0.09870677441358566,
-0.10961828380823135,
0.09015165269374847,
-0.02314312756061554,
0.04395541548728943,
-0.11019935458898544,
-0.05609181150794029,
-0.010982531122863293,
0.0651967003941536,
0.025985147804021835,
0.12283620238304138,
-0.011369197629392147,
-0.08494211733341217,
0.05069887638092041,
0.0397462360560894,
-0.05485595390200615,
0.16917604207992554,
0.03600436449050903,
-0.09238070994615555,
-0.01561541948467493,
0.005785706453025341,
0.07739362120628357,
0.13267607986927032,
-0.03149675950407982,
0.08397781103849411,
0.04527561739087105,
-0.01358442660421133,
0.02263888344168663,
-0.1103646382689476,
0.0806688740849495,
0.030088035389780998,
-0.10439234972000122,
0.06566787511110306,
-0.020085284486413002,
-0.0014424881665036082,
0.05740761011838913,
-0.018631858751177788,
0.12573575973510742,
-0.005154851824045181,
-0.016192084178328514,
-0.10290052741765976,
0.11421840637922287,
-0.08728915452957153,
-0.15164650976657867,
-0.187673881649971,
-0.0028076544404029846,
-0.06350020319223404,
-0.027840441092848778,
0.0021387797314673662,
-0.0019059756305068731,
-0.10683363676071167,
-0.08241543173789978,
0.07635002583265305,
0.016859296709299088,
0.011244128458201885,
-0.02164149470627308,
-0.00024413217033725232,
0.06725005805492401,
-0.12483465671539307,
0.015669651329517365,
0.05341500788927078,
-0.1533706784248352,
-0.012933988124132156,
0.008305140770971775,
0.06403826177120209,
0.05165278539061546,
-0.018207285553216934,
-0.05188685283064842,
0.008383885957300663,
0.220100536942482,
-0.06888259947299957,
0.13586114346981049,
0.15922975540161133,
0.056929152458906174,
0.06520986557006836,
0.050646308809518814,
0.014415987767279148,
-0.05324899032711983,
0.01975296437740326,
0.012238798663020134,
-0.052277494221925735,
-0.1273297518491745,
-0.1200142651796341,
-0.04599538818001747,
-0.01909295655786991,
0.07405462861061096,
0.0785093903541565,
0.005981009919196367,
0.05031925439834595,
-0.061812106519937515,
0.05689511448144913,
0.0236346535384655,
0.07979298382997513,
0.08138035237789154,
0.04290294274687767,
0.0830255001783371,
-0.09153401851654053,
0.09748686850070953,
0.1621691733598709,
0.0934426486492157,
0.15552374720573425,
-0.11035274714231491,
0.16057434678077698,
-0.014100950211286545,
0.15936604142189026,
0.1120019257068634,
0.059790559113025665,
-0.04948638007044792,
0.025447409600019455,
-0.017474208027124405,
-0.0877835750579834,
-0.012973064556717873,
0.056235171854496,
0.014907116070389748,
0.04517201706767082,
0.02520102448761463,
0.025443170219659805,
0.03960626572370529,
0.11565249413251877,
-0.010703779757022858,
-0.22308333218097687,
-0.12554219365119934,
0.04654435068368912,
-0.011666658334434032,
-0.037141744047403336,
-0.03290528804063797,
0.18201105296611786,
-0.070372074842453,
0.07027841359376907,
-0.007835042662918568,
0.08608562499284744,
-0.0936359092593193,
-0.029334206134080887,
0.05275171995162964,
0.21681201457977295,
0.03466539829969406,
0.0759371966123581,
-0.191274955868721,
-0.0004863012582063675,
0.03931101784110069,
0.12692579627037048,
-0.07475945353507996,
0.033249251544475555,
0.0657370537519455,
0.04183446615934372,
0.05607965961098671,
0.01291235163807869,
0.013928769156336784,
-0.02435927838087082,
-0.19338248670101166,
0.0435827299952507,
0.008444159291684628,
-0.00296818558126688,
0.03817567601799965,
-0.02646682783961296,
-0.015425188466906548,
-0.03164815530180931,
-0.021305272355675697,
-0.07605275511741638,
-0.1988847851753235,
0.041463453322649,
0.03305940702557564,
0.011939704418182373,
-0.089814692735672,
-0.04774526506662369,
-0.07164070010185242,
0.13270463049411774,
-0.13350945711135864,
-0.0782904103398323,
-0.1332215815782547,
-0.055553484708070755,
0.13041189312934875,
-0.07526823878288269,
0.04181050881743431,
-0.02078910917043686,
0.1669176071882248,
-0.02295420505106449,
-0.08696726709604263,
0.006822419352829456,
-0.1129717007279396,
-0.17093047499656677,
0.005135956220328808,
0.10933432728052139,
0.013940226286649704,
0.08399536460638046,
0.0019073350122198462,
-0.0003211467119399458,
-0.030045269057154655,
-0.15238049626350403,
-0.053340353071689606,
0.2932609021663666,
-0.05991537496447563,
0.03579557687044144,
-0.06242877244949341,
-0.09587409347295761,
-0.034734468907117844,
-0.03873105347156525,
0.04979308322072029,
0.15963107347488403,
-0.09634863585233688,
0.12095098942518234,
0.05404186248779297,
-0.050894543528556824,
-0.16274727880954742,
0.01260063424706459,
0.06767790019512177,
0.07543446123600006,
-0.0038775510620325804,
-0.0987030416727066,
0.08546207845211029,
0.10947967320680618,
-0.019815826788544655,
0.08303333818912506,
-0.19124427437782288,
-0.12702934443950653,
-0.060922879725694656,
0.053230058401823044,
0.016657399013638496,
-0.08855085074901581,
-0.0689099133014679,
-0.04931321367621422,
-0.1564008742570877,
0.14363588392734528,
-0.038978978991508484,
0.10899398475885391,
-0.013558808714151382,
0.0069526005536317825,
0.005804711952805519,
-0.0408807247877121,
0.19711367785930634,
-0.05469931289553642,
0.008679568767547607,
-0.04363024979829788,
0.1350148320198059,
0.0268036387860775,
-0.06688524782657623,
0.08773135393857956,
-0.058249332010746,
0.06348391622304916,
-0.1405297964811325,
-0.0439191497862339,
-0.058171652257442474,
0.0024766973219811916,
-0.023416053503751755,
-0.018706142902374268,
-0.10263190418481827,
0.06696391850709915,
0.05744485184550285,
-0.0459757037460804,
-0.13874410092830658,
-0.007784686982631683,
-0.011861125007271767,
0.11093549430370331,
0.12632422149181366,
-0.10109280049800873,
-0.06675666570663452,
-0.032087501138448715,
0.0016177230281755328,
0.09082812070846558,
-0.13120131194591522,
0.07156616449356079,
0.1200217455625534,
-0.009123842231929302,
0.04374051094055176,
-0.010540634393692017,
-0.10594839602708817,
-0.004256097134202719,
0.061461713165044785,
-0.10952325165271759,
-0.21807284653186798,
-0.03255137801170349,
0.061584945768117905,
-0.059864092618227005,
-0.036501627415418625,
0.10179376602172852,
-0.060098372399806976,
-0.011365406215190887,
0.02939435839653015,
0.05884118005633354,
-0.005616792012006044,
0.18908590078353882,
-0.01033756136894226,
0.0492289662361145,
-0.08795846998691559,
0.12732146680355072,
0.03737049177289009,
-0.05745808407664299,
-0.020167721435427666,
0.10839667916297913,
-0.0971575379371643,
-0.0317387580871582,
-0.052994102239608765,
0.003147671464830637,
-0.011733329854905605,
-0.0724201574921608,
0.011960904113948345,
-0.049500253051519394,
0.037833742797374725,
0.11483947187662125,
0.037853747606277466,
0.029364371672272682,
-0.020208628848195076,
0.004393231589347124,
-0.06631702929735184,
0.08171632140874863,
-0.010433937422931194,
0.0383424274623394,
-0.10992957651615143,
0.03907065466046333,
0.015793489292263985,
-0.01281745545566082,
-0.016803720965981483,
-0.003806941444054246,
-0.05758829414844513,
-0.05326183885335922,
-0.12786294519901276,
0.03925975784659386,
-0.03352588787674904,
0.026064664125442505,
-0.023175109177827835,
-0.02947537787258625,
-0.017787126824259758,
0.09004824608564377,
-0.018136033788323402,
-0.06495221704244614,
-0.05128564313054085,
0.01737363263964653,
-0.041573382914066315,
0.02976803667843342,
0.07470472902059555,
-0.05729309096932411,
0.11340045183897018,
-0.033159833401441574,
-0.0018984891939908266,
-0.011745678260922432,
-0.1536692976951599,
-0.003585249651223421,
0.026849936693906784,
0.040208589285612106,
0.013676024973392487,
-0.14521080255508423,
0.035292476415634155,
0.03259000554680824,
-0.021517710760235786,
-0.0232949610799551,
0.1412527710199356,
-0.07049474865198135,
0.038734354078769684,
-0.014042016118764877,
0.017177635803818703,
-0.05822434276342392,
0.0020172670483589172,
0.08221199363470078,
0.0448935404419899,
0.10238049924373627,
-0.030874840915203094,
-0.03542177379131317,
-0.11715075373649597,
-0.015658441931009293,
-0.008877106942236423,
-0.03318742290139198,
-0.03841036185622215,
0.024016784504055977,
0.0568883903324604,
0.053695715963840485,
0.15606050193309784,
-0.059398092329502106,
0.04337039217352867,
0.011712828651070595,
0.00663224421441555,
-0.02421729266643524,
-0.033857449889183044,
0.11025679111480713,
0.038643818348646164,
0.03181212767958641,
0.03940371796488762,
0.021128209307789803,
-0.002313589444383979,
0.003909908700734377,
0.017014985904097557,
0.11964210867881775,
0.02245998941361904,
0.04977545142173767,
0.019663754850625992,
-0.03801629692316055,
-0.1550539880990982,
0.061143845319747925,
-0.10033506900072098,
0.04052891954779625,
-0.06951326131820679,
0.09518447518348694,
0.12139754742383957,
-0.17003116011619568,
0.06727751344442368,
0.03153957426548004,
-0.05794987455010414,
-0.10743553191423416,
-0.1640043705701828,
-0.060674361884593964,
-0.11519645899534225,
-0.03746015578508377,
-0.0716540589928627,
0.042206961661577225,
0.05258242413401604,
0.061864644289016724,
0.00804649107158184,
0.09720388799905777,
-0.12905985116958618,
-0.0319782979786396,
0.03972933813929558,
-0.0006768167950212955,
0.00787789560854435,
0.10328059643507004,
-0.01402550283819437,
0.05125020444393158,
0.04242921248078346,
0.10658018290996552,
0.04761981591582298,
0.041604768484830856,
0.017216036096215248,
0.004475066903978586,
-0.016729727387428284,
0.013785367831587791,
-0.024874525144696236,
-0.0527033694088459,
0.17566727101802826,
0.04991212487220764,
-0.033869948238134384,
0.018828045576810837,
0.17943289875984192,
-0.052582722157239914,
-0.13730739057064056,
-0.15516088902950287,
0.03488905727863312,
-0.07690931856632233,
-0.006955588236451149,
0.01474031899124384,
-0.10523143410682678,
0.02033216319978237,
0.18249322474002838,
0.03965001180768013,
-0.035682372748851776,
0.007291954476386309,
-0.024781549349427223,
-0.029444988816976547,
-0.015703462064266205,
0.04540136083960533,
0.04946684092283249,
0.2719123661518097,
-0.03477431833744049,
0.00569771695882082,
-0.0029002451337873936,
-0.0278121717274189,
-0.0737295150756836,
0.063436359167099,
-0.11628366261720657,
0.020021965727210045,
-0.039348285645246506,
-0.002305326983332634,
0.04985582455992699,
-0.12434229999780655,
0.010562969371676445,
-0.09364210814237595,
-0.09874211251735687,
0.005371697712689638,
0.02379531040787697,
-0.03245147690176964,
-0.025050073862075806,
0.004214998800307512,
0.047856371849775314,
0.037628307938575745,
-0.010347412899136543,
-0.12027550488710403,
-0.0458856076002121,
0.032882243394851685,
0.11531560122966766,
0.16490188241004944,
0.014983042143285275,
-0.023497454822063446,
0.1116999089717865,
-0.018346430733799934,
-0.2154763638973236,
0.08650823682546616,
-0.03150589391589165,
-0.19058114290237427,
-0.0004069410788360983,
0.18399451673030853,
-0.010779906064271927,
0.1337585151195526,
0.03311466425657272,
0.05252014845609665,
0.005348144099116325,
0.009751550853252411,
0.040111348032951355,
-0.057802390307188034,
0.007639855146408081,
-0.10525578260421753,
0.1250208467245102,
0.14632675051689148,
0.002697748364880681,
0.014268733561038971,
-0.08083826303482056,
0.07487506419420242,
-0.025722835212945938,
0.06869379431009293,
0.01008952409029007,
-0.15398071706295013,
0.02238684706389904,
0.027611836791038513,
0.040297091007232666,
-0.09521009027957916,
-0.10161691159009933,
-0.02049851231276989,
-0.04063209518790245,
-0.022808518260717392,
0.12560349702835083,
0.11241108179092407,
0.030854227021336555,
-0.029835086315870285,
-0.23970574140548706,
0.0006784906145185232,
0.06992953270673752,
-0.138535276055336,
-0.055578213185071945
] |
null | null | null |
# **Q-Learning** Agent playing1 **FrozenLake-v1**
This is a trained model of a **Q-Learning** agent playing **FrozenLake-v1** .
## Usage
```python
model = load_from_hub(repo_id="Atozzio/q-FrozenLake-v1-4x4-noSlippery", filename="q-learning.pkl")
# Don't forget to check if you need to add additional attributes (is_slippery=False etc)
env = gym.make(model["env_id"])
```
| {"tags": ["FrozenLake-v1-4x4-no_slippery", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "q-FrozenLake-v1-4x4-noSlippery", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "FrozenLake-v1-4x4-no_slippery", "type": "FrozenLake-v1-4x4-no_slippery"}, "metrics": [{"type": "mean_reward", "value": "1.00 +/- 0.00", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | Atozzio/q-FrozenLake-v1-4x4-noSlippery | [
"FrozenLake-v1-4x4-no_slippery",
"q-learning",
"reinforcement-learning",
"custom-implementation",
"model-index",
"region:us"
] | 2024-02-08T07:48:55+00:00 | [] | [] | TAGS
#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us
|
# Q-Learning Agent playing1 FrozenLake-v1
This is a trained model of a Q-Learning agent playing FrozenLake-v1 .
## Usage
| [
"# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage"
] | [
"TAGS\n#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n",
"# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage"
] | [
40,
39
] | [
"passage: TAGS\n#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage"
] | [
0.04578453302383423,
-0.08074592798948288,
-0.00430759321898222,
0.10720831900835037,
0.05034215748310089,
-0.040469273924827576,
0.11997015029191971,
0.018999949097633362,
0.20601962506771088,
-0.010012076236307621,
0.1455274522304535,
0.007022971753031015,
-0.006192410364747047,
0.1867983490228653,
0.04572829231619835,
-0.26324528455734253,
0.01831899583339691,
-0.09495259821414948,
-0.07281816750764847,
0.11870454251766205,
0.05470194295048714,
-0.01901467889547348,
-0.0007633853238075972,
0.056141503155231476,
-0.0673527717590332,
0.0007737681735306978,
0.031996939331293106,
-0.012976245954632759,
0.19804789125919342,
-0.02254498563706875,
0.06641989201307297,
0.054705578833818436,
0.0758768692612648,
-0.1998077929019928,
0.0358855277299881,
-0.04215473681688309,
-0.09439758956432343,
-0.03934839740395546,
-0.018780618906021118,
0.05878105387091637,
0.053356342017650604,
0.03858819976449013,
0.058354366570711136,
0.09384993463754654,
-0.0773480236530304,
0.04328357055783272,
0.04280758649110794,
0.024811049923300743,
0.04589218273758888,
-0.0237203948199749,
-0.027002155780792236,
0.08246652781963348,
-0.22182892262935638,
0.10318073630332947,
-0.010159241035580635,
-0.5270710587501526,
-0.00633762264624238,
0.24088262021541595,
0.11517096310853958,
0.05707438662648201,
-0.06903956830501556,
0.10566288232803345,
0.03913382440805435,
-0.007209456991404295,
0.03210983797907829,
0.02150118350982666,
0.12817370891571045,
0.06009242683649063,
-0.09581366181373596,
0.040699947625398636,
0.13722525537014008,
0.012822695076465607,
0.020306183025240898,
-0.08888901025056839,
0.0410032719373703,
-0.03461858257651329,
-0.007679527159780264,
-0.09758518636226654,
0.05478060990571976,
0.012466507963836193,
-0.0934976264834404,
-0.09247440844774246,
-0.04236573353409767,
-0.06708304584026337,
0.11252415925264359,
0.046419668942689896,
-0.0874939113855362,
0.03884070739150047,
-0.06760413944721222,
0.05918780341744423,
-0.16863860189914703,
0.02074250765144825,
-0.06627868115901947,
-0.09376336634159088,
-0.11799788475036621,
-0.01683047041296959,
-0.07946427166461945,
0.009092256426811218,
0.056664444506168365,
0.1447116881608963,
0.22076484560966492,
0.06690320372581482,
0.09728849679231644,
0.07456006109714508,
0.06531001627445221,
0.1538129299879074,
0.10918238013982773,
0.019075315445661545,
-0.015266558155417442,
0.0948706716299057,
-0.06445580720901489,
-0.1351388692855835,
-0.15579092502593994,
0.005488025024533272,
0.0983937531709671,
0.08871900290250778,
-0.044080477207899094,
-0.006702381651848555,
-0.024641724303364754,
0.08566431701183319,
-0.11314457654953003,
-0.024612564593553543,
-0.002267979085445404,
0.06882024556398392,
-0.024801667779684067,
0.020378148183226585,
-0.06242705136537552,
0.12715265154838562,
0.04222423583269119,
-0.059924717992544174,
-0.055308472365140915,
-0.03053177334368229,
-0.014276440255343914,
-0.027539284899830818,
0.02446848154067993,
-0.07659092545509338,
0.04767750948667526,
-0.16766095161437988,
-0.042871296405792236,
-0.04784649610519409,
0.025697942823171616,
-0.03907240927219391,
-0.13557587563991547,
-0.17699143290519714,
-0.048906855285167694,
-0.022438718006014824,
0.03549358621239662,
-0.038111843168735504,
0.006551501806825399,
-0.006318534724414349,
-0.1583600640296936,
0.09783563017845154,
0.09784027189016342,
-0.03643378987908363,
-0.02749447710812092,
0.056263517588377,
-0.07194498926401138,
0.1561182290315628,
-0.21054518222808838,
-0.054014235734939575,
-0.044764336198568344,
-0.06595750898122787,
0.19673264026641846,
0.012690845876932144,
-0.01202624011784792,
0.19873127341270447,
-0.29073721170425415,
-0.06078760325908661,
0.12533614039421082,
-0.07834373414516449,
-0.0936407670378685,
0.06941844522953033,
-0.04206686094403267,
0.023345354944467545,
0.046047765761613846,
0.36345911026000977,
-0.02069227211177349,
-0.16197136044502258,
-0.021782705560326576,
0.13971707224845886,
-0.1184760183095932,
0.059895481914281845,
0.04240793362259865,
0.12543781101703644,
-0.04250509291887283,
-0.018672896549105644,
-0.09023164212703705,
0.05999075248837471,
-0.05241934582591057,
-0.09016361832618713,
-0.03393383324146271,
-0.07645075023174286,
0.13294468820095062,
-0.0629684180021286,
0.05601520463824272,
-0.03255095332860947,
-0.07133250683546066,
-0.050324998795986176,
-0.016492370516061783,
0.04460815340280533,
0.05951254442334175,
-0.12794871628284454,
0.11029167473316193,
0.13025271892547607,
-0.0006193425506353378,
-0.07498852163553238,
-0.17872096598148346,
0.003240168560296297,
0.009576505981385708,
0.039837226271629333,
0.17141658067703247,
0.12209978699684143,
0.033295199275016785,
0.008770671673119068,
-0.06389404833316803,
-0.18276847898960114,
0.058129217475652695,
-0.056212130934000015,
-0.14230976998806,
-0.052409034222364426,
-0.0728459507226944,
0.017381802201271057,
-0.0859743058681488,
-0.017379917204380035,
0.021926190704107285,
0.006908397190272808,
0.02990424446761608,
-0.026645656675100327,
-0.049561817198991776,
0.021254703402519226,
0.06490101665258408,
-0.0037617047782987356,
0.12023693323135376,
0.008277264423668385,
-0.18308481574058533,
0.07930773496627808,
0.08478537946939468,
0.09196605533361435,
0.013250201940536499,
0.02685922384262085,
-0.021522263064980507,
-0.08061408251523972,
-0.054420311003923416,
0.02957955375313759,
0.11417073011398315,
0.1317172348499298,
0.2361993044614792,
0.08753683418035507,
0.04697408527135849,
-0.02164587564766407,
-0.016415923833847046,
0.002810494042932987,
-0.06318057328462601,
-0.029935607686638832,
0.10614971816539764,
0.05865858122706413,
-0.067733034491539,
-0.04576427489519119,
0.09590928256511688,
0.02732124738395214,
0.21205885708332062,
-0.03342745825648308,
0.01286078616976738,
-0.10957037657499313,
-0.06550975888967514,
-0.031982194632291794,
0.09201868623495102,
0.09498392790555954,
0.009755023755133152,
-0.022056059911847115,
-0.04259001836180687,
0.0012916827108711004,
-0.1334889680147171,
-0.10375088453292847,
0.026475343853235245,
0.013400445692241192,
-0.11206940561532974,
0.11674030870199203,
-0.11352457851171494,
0.039504457265138626,
0.06024791672825813,
-0.13837239146232605,
0.04428480193018913,
-0.029713207855820656,
-0.07886212319135666,
0.16866780817508698,
-0.11075661331415176,
-0.094340018928051,
-0.08831550180912018,
0.004082420375198126,
0.0075836325995624065,
-0.03922267258167267,
-0.009283260442316532,
-0.19952571392059326,
-0.005375816952437162,
-0.03544965013861656,
0.013616434298455715,
-0.06988783925771713,
-0.11287739872932434,
-0.010957922786474228,
0.07084179669618607,
-0.043388739228248596,
-0.07803605496883392,
0.007967432029545307,
-0.08923084288835526,
-0.10623309016227722,
0.028189711272716522,
0.019765101373195648,
-0.022883659228682518,
0.16152891516685486,
0.01816628873348236,
0.05626589432358742,
-0.03298520669341087,
0.30665266513824463,
-0.038163769990205765,
0.08371731638908386,
-0.02993497997522354,
-0.07433546334505081,
0.06130730360746384,
-0.022327827289700508,
0.06086638569831848,
-0.020221687853336334,
-0.02362890914082527,
0.0077952733263373375,
-0.08579335361719131,
-0.18365982174873352,
-0.05417544022202492,
0.03724347800016403,
0.195254847407341,
0.031118987128138542,
0.01910330168902874,
-0.0488768145442009,
-0.010547760874032974,
0.1665220558643341,
-0.10005921125411987,
0.04030545800924301,
-0.05366240441799164,
0.11506262421607971,
-0.08640182018280029,
0.06195629760622978,
0.020486772060394287,
0.04266135022044182,
-0.04877188801765442,
0.09486009180545807,
0.0826394334435463,
0.1121082529425621,
-0.02206910029053688,
0.046257395297288895,
0.019012698903679848,
0.07383184134960175,
0.11073657125234604,
0.0368414968252182,
-0.0729052945971489,
0.001982470043003559,
-0.006313489284366369,
-0.039427030831575394,
0.11933320760726929,
0.17963355779647827,
-0.11991413682699203,
-0.05106910318136215,
0.27167606353759766,
0.0031242913100868464,
0.19481229782104492,
-0.01315275114029646,
0.043591804802417755,
-0.04484925419092178,
0.04572054371237755,
-0.05338600277900696,
-0.04086209088563919,
0.2094656229019165,
0.08045925945043564,
-0.17165091633796692,
-0.08549032360315323,
-0.05912299454212189,
0.07081323862075806,
0.10728751868009567,
0.0013539529172703624,
-0.04156802222132683,
0.0004610282776411623,
0.0014198932331055403,
0.08339415490627289,
-0.14520122110843658,
0.11816094070672989,
-0.03172019124031067,
0.05612684786319733,
0.017555562779307365,
-0.045326150953769684,
0.04264266416430473,
0.07474290579557419,
0.26618310809135437,
0.0904107540845871,
-0.040318213403224945,
-0.0892091691493988,
-0.12260187417268753,
0.010461576282978058,
0.029102616012096405,
-0.03534553572535515,
0.0037547778338193893,
-0.020087555050849915,
0.0318896509706974,
0.008264793083071709,
0.016230624169111252,
-0.08987458795309067,
-0.03175399824976921,
-0.027736429125070572,
-0.023839212954044342,
0.10733365267515182,
-0.09495144337415695,
-0.1444292515516281,
-0.15713949501514435,
0.04191131144762039,
-0.0766405463218689,
-0.056593164801597595,
-0.054507751017808914,
-0.05239389091730118,
-0.0311186034232378,
-0.03773957118391991,
0.09099467098712921,
-0.0021037792321294546,
0.14807306230068207,
-0.1920108050107956,
-0.04220759496092796,
0.051812779158353806,
-0.07607918977737427,
-0.08729588985443115,
0.03410962224006653,
0.12136995792388916,
0.05116051807999611,
0.11504370719194412,
0.013609255664050579,
0.09567681699991226,
0.0045484392903745174,
-0.06713183224201202,
0.15302421152591705,
-0.14069625735282898,
-0.27875974774360657,
-0.03836318850517273,
0.016946332529187202,
0.1615200787782669,
-0.05613167956471443,
0.031766023486852646,
0.3335736393928528,
0.27782970666885376,
-0.1428707242012024,
0.25916144251823425,
0.019178593531250954,
0.004398873541504145,
-0.19130495190620422,
-0.10125631093978882,
0.025324683636426926,
0.04740457236766815,
0.12032642960548401,
-0.14564448595046997,
-0.010732659138739109,
-0.04543145373463631,
-0.025908485054969788,
0.10386138409376144,
-0.12300799041986465,
-0.07263197749853134,
0.07765276730060577,
0.039809420704841614,
0.1808302253484726,
0.03932500258088112,
0.0014799144119024277,
0.13626977801322937,
0.06612244248390198,
0.019124457612633705,
0.05216038227081299,
0.08028066903352737,
-0.018944554030895233,
0.14207926392555237,
0.05448179319500923,
-0.02551644667983055,
0.052681710571050644,
-0.0054580713622272015,
-0.03219012916088104,
0.015605825930833817,
-0.183198019862175,
-0.10147556662559509,
-0.0561356320977211,
-0.10798973590135574,
-0.04978342354297638,
0.056853994727134705,
-0.12395523488521576,
-0.007896827533841133,
-0.03841273859143257,
0.03718273714184761,
-0.07831971347332001,
-0.09360362589359283,
-0.036494381725788116,
0.1351792961359024,
0.07210618257522583,
0.04471297934651375,
0.035655103623867035,
-0.07390819489955902,
0.07097936421632767,
0.21671734750270844,
0.08159157633781433,
0.028919655829668045,
-0.19545674324035645,
-0.024042490869760513,
-0.0803457647562027,
0.06306298077106476,
-0.08856996893882751,
-0.016788700595498085,
0.11923003196716309,
0.08616556972265244,
0.05413002520799637,
0.09640096127986908,
-0.045083072036504745,
0.021686913445591927,
0.02684609219431877,
-0.15131035447120667,
-0.18501274287700653,
-0.08534606546163559,
-0.03519878163933754,
0.11561143398284912,
-0.06398691236972809,
0.10897188633680344,
-0.13615410029888153,
0.010051886551082134,
-0.006060056854039431,
0.02693452313542366,
-0.03596206381917,
-0.11251141875982285,
0.15348562598228455,
0.11999429017305374,
-0.06767056882381439,
0.03127254918217659,
-0.09527092427015305,
-0.04423454403877258,
0.12686803936958313,
-0.013623855076730251,
-0.0371493324637413,
-0.054547641426324844,
-0.03628576174378395,
0.15247689187526703,
-0.03436964750289917,
0.008244883269071579,
-0.041229065507650375,
-0.18217355012893677,
0.0798322781920433,
0.09045056998729706,
0.019827889278531075,
-0.031874191015958786,
-0.09797266125679016,
-0.010231015272438526,
-0.0011165260802954435,
0.11730700731277466,
-0.10696814209222794,
-0.10933240503072739,
-0.15144047141075134,
0.06713984161615372,
-0.0007159380475059152,
0.18502596020698547,
-0.06394898891448975,
-0.08904669433832169,
-0.12429379671812057,
0.02344517596065998,
-0.0027384376153349876,
-0.042264558374881744,
0.01618490368127823,
0.07992301136255264,
-0.04095321521162987,
0.02075677551329136,
-0.06651144474744797,
0.06372585147619247,
-0.11786920577287674,
0.09625071287155151,
0.01063506118953228,
0.016993753612041473,
-0.0417880080640316,
-0.01618220843374729,
0.039470795542001724,
-0.057925306260585785,
0.07921463251113892,
0.011758086271584034,
0.0010938759660348296,
0.10196787863969803,
-0.0034960443153977394,
0.06409632414579391,
-0.05372481048107147,
-0.023290161043405533,
0.06578411161899567,
-0.05874887853860855,
-0.03370826691389084,
-0.1573946475982666,
-0.0709633082151413,
0.020051732659339905,
-0.04775108024477959,
0.002077929675579071,
0.03673801198601723,
0.062159497290849686,
-0.06937079131603241,
-0.12125655263662338,
-0.043812792748212814,
-0.028638383373618126,
0.021301284432411194,
0.10829301923513412,
-0.07526551932096481,
0.1547859013080597,
-0.052787959575653076,
-0.00020603960729204118,
0.07437096536159515,
0.04048224538564682,
0.01393822580575943,
-0.10422444343566895,
-0.04698587954044342,
-0.11035211384296417,
0.1502903699874878,
-0.007902312092483044,
-0.03533121198415756,
0.03719403222203255,
-0.11946307867765427,
-0.1572723090648651,
0.03418220207095146,
0.10199101269245148,
0.0448341928422451,
0.025807438418269157,
0.027079269289970398,
-0.04042419046163559,
-0.021270349621772766,
-0.07034418731927872,
0.0882953479886055,
-0.12085357308387756,
-0.09669415652751923,
0.09555385261774063,
0.12178351730108261,
-0.0036850625183433294,
-0.07441367954015732,
0.11554073542356491,
-0.021787192672491074,
0.05525410920381546,
-0.02971339225769043,
0.10308072715997696,
0.0796005055308342,
-0.12273547053337097,
0.005693064536899328,
-0.036891788244247437,
-0.0741485133767128,
-0.12975730001926422,
0.019545545801520348,
-0.061916105449199677,
-0.13383042812347412,
0.12179028987884521,
-0.09376577287912369,
0.030037038028240204,
-0.10506992787122726,
0.021338803693652153,
0.01864001713693142,
0.061665527522563934,
-0.10988292098045349,
0.08575301617383957,
0.13424484431743622,
-0.043199893087148666,
-0.07184189558029175,
-0.12455986440181732,
-0.05022053420543671,
-0.04231856390833855,
-0.13957437872886658,
-0.11600435525178909,
0.0100301094353199,
-0.023418782278895378,
-0.05818291753530502,
0.0015462689334526658,
-0.03659068048000336,
0.008594646118581295,
0.021907730028033257,
0.04032021388411522,
-0.02693161368370056,
0.05134565755724907,
-0.057569269090890884,
-0.052510857582092285,
0.11489357799291611,
0.04113486409187317,
-0.03561042994260788,
-0.052359987050294876,
0.12997733056545258,
-0.11959461867809296,
0.07662346214056015,
-0.020313527435064316,
0.017129231244325638,
-0.06435854732990265,
0.17131924629211426,
0.11673715710639954,
-0.1367570012807846,
-0.005008010193705559,
-0.08210669457912445,
0.020409544929862022,
0.023555370047688484,
0.13693512976169586,
-0.03411718085408211,
-0.0012358218664303422,
-0.1580323874950409,
0.018575575202703476,
-0.18557456135749817,
-0.03716109320521355,
0.04671547934412956,
0.09917585551738739,
0.15293832123279572,
-0.0034432117827236652,
-0.1263325810432434,
0.10424192249774933,
-0.2118520885705948,
0.0907607227563858,
0.05121984705328941,
-0.11874113976955414,
-0.06765396893024445,
-0.06795281916856766,
0.1198519766330719,
0.009196433238685131,
0.2040700763463974,
-0.013615905307233334,
-0.09132910519838333,
-0.07060808688402176,
-0.01980910450220108,
-0.030524181202054024,
0.09714830666780472,
0.041414931416511536,
0.04653804749250412,
0.12821412086486816,
0.00368314771912992,
0.07533777505159378,
0.060310911387205124,
0.02759413793683052,
-0.012300663627684116,
0.04076618701219559,
0.08261215686798096,
-0.14588621258735657,
-0.1659701019525528,
0.1326720416545868,
0.025149408727884293,
0.11792458593845367,
0.03658788278698921,
-0.1549617499113083,
0.06687124073505402,
0.2523096203804016,
-0.11147607117891312,
0.02505038119852543,
0.12737524509429932,
-0.0366884209215641,
0.0672016367316246,
0.1144871786236763,
-0.02633814327418804,
-0.05217865854501724,
-0.011363590136170387,
0.10233135521411896,
0.028660254552960396,
-0.04646271467208862,
-0.02340836264193058,
-0.03373933956027031,
-0.019070526584982872,
-0.011738128960132599,
-0.0909019410610199,
-0.1543993502855301,
-0.10471053421497345,
-0.16619662940502167,
0.04399140924215317,
-0.04626438021659851,
0.13418889045715332,
0.09469578415155411,
-0.012723101302981377,
0.04568437114357948,
0.028575526550412178,
0.07275456190109253,
0.07916246354579926,
-0.02939477376639843,
-0.036159269511699677
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | akashAD/phi-1_5-finetuned-query-classify | [
"transformers",
"safetensors",
"phi",
"text-generation",
"custom_code",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T07:50:54+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #phi #text-generation #custom_code #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #phi #text-generation #custom_code #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
51,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #phi #text-generation #custom_code #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.07441530376672745,
0.13596796989440918,
-0.0039025098085403442,
0.027505790814757347,
0.12266930937767029,
0.005486504640430212,
0.0640522912144661,
0.10635033994913101,
-0.02424517273902893,
0.12324316799640656,
0.022558843716979027,
0.10996841639280319,
0.10686804354190826,
0.18769802153110504,
-0.005247652996331453,
-0.20407868921756744,
0.052618369460105896,
-0.1339786946773529,
-0.005346748046576977,
0.12320936471223831,
0.12734514474868774,
-0.11981219798326492,
0.07206547260284424,
-0.04122542217373848,
-0.006963358726352453,
-0.03489234298467636,
-0.05797455459833145,
-0.048964742571115494,
0.06430605798959732,
0.06026479974389076,
0.0595458559691906,
0.01642736792564392,
0.09416768699884415,
-0.2770709991455078,
0.02226952090859413,
0.08239509910345078,
0.005618265364319086,
0.06627916544675827,
0.07169265300035477,
-0.07693344354629517,
0.08495151996612549,
-0.0664825439453125,
0.14687298238277435,
0.07982858270406723,
-0.09438081085681915,
-0.1880333423614502,
-0.09057903289794922,
0.09343822300434113,
0.18868158757686615,
0.059760916978120804,
-0.030585795640945435,
0.13118816912174225,
-0.06512762606143951,
0.01845625787973404,
0.06895748525857925,
-0.07347869127988815,
-0.05359777435660362,
0.06372594088315964,
0.0695996955037117,
0.09846188873052597,
-0.12773703038692474,
-0.009679428301751614,
0.0320039726793766,
0.013185882940888405,
0.10711356997489929,
0.015842726454138756,
0.12049184739589691,
0.03668152913451195,
-0.14256839454174042,
-0.048429738730192184,
0.08701564371585846,
0.036932192742824554,
-0.0556931346654892,
-0.24876362085342407,
-0.020193742588162422,
-0.038236599415540695,
-0.035524725914001465,
-0.04393884912133217,
0.045244693756103516,
-0.02178187482059002,
0.08087658137083054,
-0.0036660772748291492,
-0.06963636726140976,
-0.05113578215241432,
0.08353596180677414,
0.07143381237983704,
0.028143590316176414,
-0.026850108057260513,
0.01174293551594019,
0.11898057907819748,
0.11549804359674454,
-0.11481842398643494,
-0.051060616970062256,
-0.06272978335618973,
-0.08522246032953262,
-0.04741254821419716,
0.03236919641494751,
0.04775122553110123,
0.05512697622179985,
0.21314972639083862,
-0.0013204539427533746,
0.04992509260773659,
0.03274988383054733,
0.01066634152084589,
0.06731437146663666,
0.08677016198635101,
-0.06419411301612854,
-0.13046035170555115,
-0.02145533822476864,
0.11218346655368805,
0.01231331005692482,
-0.0314481221139431,
-0.038787633180618286,
0.06697112321853638,
0.030200589448213577,
0.12535712122917175,
0.07337819784879684,
0.02016271837055683,
-0.07914318144321442,
-0.06071794033050537,
0.17465178668498993,
-0.16488726437091827,
0.031738489866256714,
0.025622278451919556,
-0.050521157681941986,
-0.018783841282129288,
0.01899137534201145,
0.016399497166275978,
-0.02034466527402401,
0.0874326080083847,
-0.057896632701158524,
-0.03757679834961891,
-0.11592794209718704,
-0.05162535607814789,
0.026356054469943047,
0.005852686706930399,
-0.030844759196043015,
-0.04131776839494705,
-0.11982329934835434,
-0.07797703891992569,
0.07937561720609665,
-0.06747156381607056,
-0.04716470465064049,
-0.03360702842473984,
-0.07821470499038696,
0.012420105747878551,
0.0004091960145160556,
0.11594925820827484,
-0.030400289222598076,
0.05072459205985069,
-0.05188135802745819,
0.07153815776109695,
0.14497826993465424,
0.027334017679095268,
-0.06379645317792892,
0.056507088243961334,
-0.23278117179870605,
0.10387995839118958,
-0.07912862300872803,
0.04115021973848343,
-0.16289231181144714,
-0.01469900831580162,
0.03982052579522133,
0.02627355046570301,
-0.006934128236025572,
0.1390630006790161,
-0.19293013215065002,
-0.03536631911993027,
0.18081475794315338,
-0.11522960662841797,
-0.08808764070272446,
0.052698906511068344,
-0.054467298090457916,
0.12280778586864471,
0.050916753709316254,
-0.02316596917808056,
0.030826477333903313,
-0.1417923867702484,
-0.01687287911772728,
-0.06468956172466278,
-0.025315633043646812,
0.15359161794185638,
0.05711430311203003,
-0.051229238510131836,
0.052433162927627563,
0.020726939663290977,
-0.021120568737387657,
-0.04788779839873314,
-0.03371148556470871,
-0.09498532861471176,
0.009590674191713333,
-0.0753822848200798,
0.01856466382741928,
-0.029109936207532883,
-0.09291869401931763,
-0.035687901079654694,
-0.15461255609989166,
0.005384957883507013,
0.09633282572031021,
-0.0055311573669314384,
-0.025023676455020905,
-0.10565053671598434,
-0.004781804513186216,
0.016855718567967415,
-0.00016373902326449752,
-0.15187151730060577,
-0.05655598267912865,
0.019132478162646294,
-0.16505438089370728,
0.02767706662416458,
-0.047599732875823975,
0.045580726116895676,
0.04182368889451027,
-0.03936924785375595,
-0.03521854057908058,
0.018329832702875137,
0.020268244668841362,
-0.01465248316526413,
-0.2745935916900635,
-0.018261034041643143,
-0.04086336866021156,
0.17035499215126038,
-0.2475264072418213,
0.04439546912908554,
0.059322256594896317,
0.1307375431060791,
0.01173730194568634,
-0.03705551475286484,
0.03161298856139183,
-0.06255152821540833,
-0.033266592770814896,
-0.0667189359664917,
-0.009362515062093735,
-0.03631554916501045,
-0.03938153013586998,
0.03835300728678703,
-0.17000539600849152,
-0.03406575322151184,
0.11603987962007523,
0.04540814831852913,
-0.15477602183818817,
-0.05056281015276909,
-0.04004296287894249,
-0.05738358572125435,
-0.07204438000917435,
-0.05216284841299057,
0.09743187576532364,
0.05571887642145157,
0.05545826256275177,
-0.05951985344290733,
-0.061445724219083786,
0.009554018266499043,
-0.02068808674812317,
-0.01867086999118328,
0.08103828877210617,
0.07077015191316605,
-0.11522045731544495,
0.09927672892808914,
0.08799781650304794,
0.08139653503894806,
0.10005104541778564,
0.0010666352463886142,
-0.09291157126426697,
-0.02339044213294983,
0.027750657871365547,
0.014336124062538147,
0.14687193930149078,
-0.04000900313258171,
0.0429966077208519,
0.04120675474405289,
-0.01584675721824169,
0.008143718354403973,
-0.09446796029806137,
0.02997143194079399,
0.02818182110786438,
-0.010246568359434605,
0.037614114582538605,
-0.056816551834344864,
0.019303709268569946,
0.10318583995103836,
0.03345242142677307,
0.04412994161248207,
0.009559271857142448,
-0.04857930168509483,
-0.11974377185106277,
0.1767151653766632,
-0.11110951006412506,
-0.23173309862613678,
-0.12149907648563385,
-0.01399032212793827,
0.02910485304892063,
-0.011633564718067646,
0.02006695233285427,
-0.06405475735664368,
-0.1171390563249588,
-0.09921693801879883,
0.045627593994140625,
0.07059439271688461,
-0.08641253411769867,
-0.06378761678934097,
0.06134819984436035,
0.04600827768445015,
-0.13516096770763397,
0.02277415804564953,
0.03902184218168259,
-0.08909334987401962,
0.007867258042097092,
0.07907920330762863,
0.07120607048273087,
0.17945720255374908,
0.012182043865323067,
-0.024257373064756393,
0.019671371206641197,
0.20478034019470215,
-0.13766378164291382,
0.10145840793848038,
0.14393620193004608,
-0.06286554783582687,
0.08066798746585846,
0.20545852184295654,
0.036268092691898346,
-0.1057758554816246,
0.044006094336509705,
0.03648979216814041,
-0.02651887945830822,
-0.24340394139289856,
-0.08019697666168213,
0.004161624237895012,
-0.06197261065244675,
0.08161719888448715,
0.08306818455457687,
0.09198566526174545,
0.02785661816596985,
-0.1081320270895958,
-0.06691340357065201,
0.05036139488220215,
0.11249116063117981,
-0.008557078428566456,
-0.007815919816493988,
0.09523359686136246,
-0.02225065603852272,
0.029176659882068634,
0.09147068858146667,
0.01374965999275446,
0.18282483518123627,
0.045852772891521454,
0.14848409593105316,
0.09157159924507141,
0.059395745396614075,
0.01233623269945383,
0.01314469799399376,
0.019094863906502724,
0.026712998747825623,
-0.015145753510296345,
-0.08685000985860825,
-0.012303872965276241,
0.1268911212682724,
0.010885220021009445,
0.04597875103354454,
0.0076150596141815186,
-0.04230163246393204,
0.08450151234865189,
0.17545753717422485,
0.01328173466026783,
-0.21406996250152588,
-0.06688741594552994,
0.06981010735034943,
-0.08051439374685287,
-0.10911136865615845,
-0.024429909884929657,
0.03406251221895218,
-0.18049609661102295,
0.02387341856956482,
-0.025180401280522346,
0.10069414228200912,
-0.12370731681585312,
-0.018827902153134346,
0.052628833800554276,
0.07052139192819595,
-0.018700284883379936,
0.06386490911245346,
-0.17778280377388,
0.13549263775348663,
0.013200430199503899,
0.07557245343923569,
-0.09068016707897186,
0.08482389152050018,
0.0111467270180583,
-0.002043683547526598,
0.1468254029750824,
-0.0010637122904881835,
-0.05408002436161041,
-0.11050406098365784,
-0.0906725600361824,
-0.011339336633682251,
0.11465787142515182,
-0.12593887746334076,
0.10165182501077652,
-0.016582757234573364,
-0.044178079813718796,
-0.0030248011462390423,
-0.12813955545425415,
-0.14044401049613953,
-0.17314541339874268,
0.04187968373298645,
-0.13014033436775208,
0.0451013408601284,
-0.10672678053379059,
-0.05035872012376785,
-0.05017208307981491,
0.19719818234443665,
-0.21763156354427338,
-0.07621806859970093,
-0.15351133048534393,
-0.06420157849788666,
0.11623851954936981,
-0.04613782465457916,
0.08647869527339935,
0.012962628155946732,
0.18781377375125885,
0.014061033725738525,
-0.015962716192007065,
0.10993410646915436,
-0.10395599156618118,
-0.21440888941287994,
-0.10220180451869965,
0.13403694331645966,
0.13545545935630798,
0.03708728775382042,
0.00035940390080213547,
0.03232092037796974,
-0.007850716821849346,
-0.11358384788036346,
0.023570599034428596,
0.18197759985923767,
0.11685380339622498,
0.037179335951805115,
-0.034665536135435104,
-0.13531899452209473,
-0.0839521661400795,
-0.042324043810367584,
0.008525622077286243,
0.18976294994354248,
-0.06857912987470627,
0.1652597337961197,
0.15934355556964874,
-0.055173277854919434,
-0.21036414802074432,
0.0313970185816288,
0.033629804849624634,
0.0021239151246845722,
0.05604655668139458,
-0.20132838189601898,
0.0957157164812088,
0.00788893923163414,
-0.057729288935661316,
0.12271249294281006,
-0.18383583426475525,
-0.14666838943958282,
0.0679788589477539,
0.07568002492189407,
-0.18666845560073853,
-0.12836617231369019,
-0.09530680626630783,
-0.04426150023937225,
-0.1240062341094017,
0.0767902210354805,
-0.019116053357720375,
0.009703016839921474,
0.03049294650554657,
0.017553992569446564,
0.010632803663611412,
-0.04766656085848808,
0.18440434336662292,
-0.005318623501807451,
0.050052255392074585,
-0.07833196222782135,
-0.05977580323815346,
0.04439995810389519,
-0.06766178458929062,
0.07768969982862473,
-0.011583259329199791,
0.012072126381099224,
-0.10826653987169266,
-0.05835650488734245,
-0.03404201939702034,
0.024099772796034813,
-0.08059826493263245,
-0.09612218290567398,
-0.037487708032131195,
0.09951330721378326,
0.09140417724847794,
-0.03928857669234276,
-0.06511175632476807,
-0.08731205761432648,
0.032564677298069,
0.21537625789642334,
0.17581914365291595,
0.05872897058725357,
-0.06627403944730759,
-0.004332480486482382,
-0.013938636519014835,
0.0518467053771019,
-0.20769350230693817,
0.054770588874816895,
0.037577200680971146,
0.03502080589532852,
0.11540760844945908,
-0.02692747861146927,
-0.15991008281707764,
-0.04947725683450699,
0.054892536252737045,
-0.07749488949775696,
-0.16381342709064484,
0.014566776342689991,
0.0698530450463295,
-0.15249542891979218,
-0.023638471961021423,
0.04465564340353012,
-0.019799569621682167,
-0.033337272703647614,
0.003100043162703514,
0.08220522850751877,
0.016223285347223282,
0.09557998180389404,
0.05498101934790611,
0.09563100337982178,
-0.10772447288036346,
0.06952618062496185,
0.07929857820272446,
-0.10225345939397812,
0.03691693767905235,
0.06498466432094574,
-0.07187005877494812,
-0.035860974341630936,
0.04298849776387215,
0.09064827114343643,
0.03834117203950882,
-0.05791795626282692,
0.006768029183149338,
-0.1019822508096695,
0.058791015297174454,
0.11939579248428345,
0.043310075998306274,
0.008954020217061043,
0.03660706803202629,
0.039979103952646255,
-0.09563424438238144,
0.12395429611206055,
0.04702746868133545,
0.03306521847844124,
-0.05115185305476189,
-0.030670443549752235,
0.033765602856874466,
-0.03013032302260399,
-0.016100779175758362,
-0.04014170542359352,
-0.06690575927495956,
-0.012527769431471825,
-0.17551913857460022,
0.004580955021083355,
-0.05460330843925476,
0.004814730025827885,
0.01808975264430046,
-0.030708830803632736,
0.006335208658128977,
0.01802607998251915,
-0.07054036855697632,
-0.05601222440600395,
-0.007826367393136024,
0.10339196026325226,
-0.17446856200695038,
0.014777671545743942,
0.07612703740596771,
-0.12507562339305878,
0.0856013149023056,
0.019583139568567276,
0.0037027799990028143,
0.030126722529530525,
-0.12974882125854492,
0.04562569037079811,
-0.008532330393791199,
0.0110886599868536,
0.04916682839393616,
-0.2144635170698166,
-0.00006298656080616638,
-0.048669952899217606,
-0.06047982722520828,
-0.008075869642198086,
-0.022350680083036423,
-0.11814062297344208,
0.10634202510118484,
0.011746841482818127,
-0.07358573377132416,
-0.025499174371361732,
0.039712198078632355,
0.09615115076303482,
-0.03705960139632225,
0.15897074341773987,
-0.017435546964406967,
0.06202258542180061,
-0.18289825320243835,
-0.022887926548719406,
-0.01893971487879753,
0.02133350446820259,
-0.041365109384059906,
-0.009923189878463745,
0.053397808223962784,
-0.0207329411059618,
0.20566853880882263,
-0.017383437603712082,
0.03804474696516991,
0.06314582377672195,
-0.013027135282754898,
-0.014618457295000553,
0.10731750726699829,
0.048464275896549225,
0.011181623674929142,
0.02981831505894661,
0.010703807696700096,
-0.03417062386870384,
-0.005388245452195406,
-0.1523161679506302,
0.077239029109478,
0.16865576803684235,
0.08118049800395966,
-0.008775852620601654,
0.054247766733169556,
-0.11114349961280823,
-0.11427582800388336,
0.0913289338350296,
-0.056236591190099716,
-0.013130106031894684,
-0.05957726389169693,
0.14178913831710815,
0.15433260798454285,
-0.1893889456987381,
0.06022936478257179,
-0.06838760524988174,
-0.04934924468398094,
-0.10593041032552719,
-0.172664612531662,
-0.05883026123046875,
-0.05516333505511284,
-0.0200477484613657,
-0.055225640535354614,
0.065329410135746,
0.09025082737207413,
0.016052771359682083,
0.013859162107110023,
0.08703679591417313,
-0.017933277413249016,
0.0062139625661075115,
0.030186165124177933,
0.06423898041248322,
0.011570695787668228,
-0.04630725830793381,
0.008485673926770687,
-0.0025986286345869303,
0.032309141010046005,
0.051715679466724396,
0.03794151917099953,
-0.025615144520998,
0.009215989150106907,
-0.028480391949415207,
-0.11046841740608215,
0.039721209555864334,
-0.025451408699154854,
-0.0637175589799881,
0.14837713539600372,
0.02840440161526203,
-0.0067321667447686195,
-0.023941997438669205,
0.2540489137172699,
-0.07580790668725967,
-0.08472823351621628,
-0.1372121125459671,
0.14538848400115967,
-0.03086796961724758,
0.06126725301146507,
0.039360348135232925,
-0.11396372318267822,
0.03298129141330719,
0.1398884356021881,
0.14529621601104736,
-0.051987502723932266,
0.017118316143751144,
0.014656228013336658,
0.0029644023161381483,
-0.038789160549640656,
0.05346855893731117,
0.06594305485486984,
0.1253807246685028,
-0.05399900674819946,
0.08479224145412445,
-0.004516707267612219,
-0.10068650543689728,
-0.03351442143321037,
0.1215558722615242,
-0.0051100910641252995,
0.021343640983104706,
-0.07243939489126205,
0.12681175768375397,
-0.04128829762339592,
-0.26241305470466614,
0.06438590586185455,
-0.06042512133717537,
-0.14796359837055206,
-0.0250939279794693,
0.042521025985479355,
-0.005583211313933134,
0.02851193957030773,
0.06835387647151947,
-0.06482896208763123,
0.1890970915555954,
0.037300046533346176,
-0.05160627141594887,
-0.06661834567785263,
0.07295969873666763,
-0.10394251346588135,
0.2989484369754791,
0.006831952836364508,
0.056051596999168396,
0.10134656727313995,
-0.031050942838191986,
-0.14268989861011505,
0.031161298975348473,
0.08590974658727646,
-0.06741449981927872,
0.055566806346178055,
0.2119295448064804,
-0.00888585951179266,
0.10927559435367584,
0.07252391427755356,
-0.08921248465776443,
0.04874938353896141,
-0.10592100024223328,
-0.09221301972866058,
-0.08717742562294006,
0.09235067665576935,
-0.0555795282125473,
0.14914315938949585,
0.12058565765619278,
-0.04728509485721588,
0.021935351192951202,
-0.021421335637569427,
0.050473760813474655,
0.004591043572872877,
0.12314941734075546,
0.022128529846668243,
-0.19706779718399048,
0.026956038549542427,
-0.0006028058123774827,
0.10154236108064651,
-0.21685825288295746,
-0.09420528262853622,
0.04897189885377884,
0.00334720266982913,
-0.06139703094959259,
0.1250954121351242,
0.05242267996072769,
0.041330695152282715,
-0.046910692006349564,
-0.030460618436336517,
-0.00651969201862812,
0.1657109558582306,
-0.10882015526294708,
-0.004513995256274939
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# output_llama2_instruct
This model is a fine-tuned version of [NousResearch/Llama-2-7b-hf](https://huggingface.co/NousResearch/Llama-2-7b-hf) on the generator dataset.
It achieves the following results on the evaluation set:
- Loss: 1.6746
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 4
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: constant
- lr_scheduler_warmup_steps: 20
- training_steps: 500
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| 1.9193 | 0.0 | 20 | 1.8124 |
| 1.8847 | 0.01 | 40 | 1.7835 |
| 1.813 | 0.01 | 60 | 1.7710 |
| 1.9026 | 0.01 | 80 | 1.7605 |
| 1.8352 | 0.01 | 100 | 1.7449 |
| 1.7327 | 0.02 | 120 | 1.7100 |
| 1.8625 | 0.02 | 140 | 1.7055 |
| 1.9379 | 0.02 | 160 | 1.7008 |
| 1.8597 | 0.02 | 180 | 1.6971 |
| 1.8703 | 0.03 | 200 | 1.6943 |
| 1.7749 | 0.03 | 220 | 1.6902 |
| 1.7645 | 0.03 | 240 | 1.6898 |
| 1.7894 | 0.04 | 260 | 1.6886 |
| 1.8492 | 0.04 | 280 | 1.6870 |
| 1.7331 | 0.04 | 300 | 1.6841 |
| 1.7278 | 0.04 | 320 | 1.6830 |
| 1.6963 | 0.05 | 340 | 1.6826 |
| 1.8226 | 0.05 | 360 | 1.6813 |
| 1.8246 | 0.05 | 380 | 1.6797 |
| 1.8577 | 0.05 | 400 | 1.6778 |
| 1.6691 | 0.06 | 420 | 1.6779 |
| 1.7001 | 0.06 | 440 | 1.6774 |
| 1.9008 | 0.06 | 460 | 1.6770 |
| 1.7856 | 0.07 | 480 | 1.6733 |
| 1.7674 | 0.07 | 500 | 1.6746 |
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1 | {"library_name": "peft", "tags": ["trl", "sft", "generated_from_trainer"], "datasets": ["generator"], "base_model": "NousResearch/Llama-2-7b-hf", "model-index": [{"name": "output_llama2_instruct", "results": []}]} | null | lillybak/output_llama2_instruct | [
"peft",
"tensorboard",
"safetensors",
"trl",
"sft",
"generated_from_trainer",
"dataset:generator",
"base_model:NousResearch/Llama-2-7b-hf",
"region:us"
] | 2024-02-08T07:51:11+00:00 | [] | [] | TAGS
#peft #tensorboard #safetensors #trl #sft #generated_from_trainer #dataset-generator #base_model-NousResearch/Llama-2-7b-hf #region-us
| output\_llama2\_instruct
========================
This model is a fine-tuned version of NousResearch/Llama-2-7b-hf on the generator dataset.
It achieves the following results on the evaluation set:
* Loss: 1.6746
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0002
* train\_batch\_size: 4
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: constant
* lr\_scheduler\_warmup\_steps: 20
* training\_steps: 500
### Training results
### Framework versions
* PEFT 0.8.2
* Transformers 4.37.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: constant\n* lr\\_scheduler\\_warmup\\_steps: 20\n* training\\_steps: 500",
"### Training results",
"### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#peft #tensorboard #safetensors #trl #sft #generated_from_trainer #dataset-generator #base_model-NousResearch/Llama-2-7b-hf #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: constant\n* lr\\_scheduler\\_warmup\\_steps: 20\n* training\\_steps: 500",
"### Training results",
"### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
55,
114,
4,
39
] | [
"passage: TAGS\n#peft #tensorboard #safetensors #trl #sft #generated_from_trainer #dataset-generator #base_model-NousResearch/Llama-2-7b-hf #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: constant\n* lr\\_scheduler\\_warmup\\_steps: 20\n* training\\_steps: 500### Training results### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.13017165660858154,
0.09099465608596802,
-0.0017913709161803126,
0.11236614733934402,
0.1390979140996933,
0.012024123221635818,
0.10543175041675568,
0.12358928471803665,
-0.07744388282299042,
0.06816596537828445,
0.14004254341125488,
0.09483623504638672,
0.02650539018213749,
0.17675596475601196,
-0.04498128965497017,
-0.2292870283126831,
-0.005220260005444288,
0.00031239280360750854,
-0.06475698947906494,
0.1270563155412674,
0.06840843707323074,
-0.13823255896568298,
0.07991071790456772,
-0.03257494047284126,
-0.18518611788749695,
0.013754434883594513,
0.010690347291529179,
-0.003954917658120394,
0.13161717355251312,
-0.008838155306875706,
0.11700642108917236,
0.033168721944093704,
0.10757036507129669,
-0.20200517773628235,
0.01838613674044609,
0.0795687586069107,
0.015075968578457832,
0.08947349339723587,
0.0597257986664772,
-0.015691623091697693,
0.09712906181812286,
-0.1054496020078659,
0.042751625180244446,
0.018544958904385567,
-0.1512666791677475,
-0.21017156541347504,
-0.09718998521566391,
0.02001197449862957,
0.08794121444225311,
0.07627463340759277,
-0.017029564827680588,
0.141261488199234,
-0.06762395054101944,
0.0930202528834343,
0.25319015979766846,
-0.2647899091243744,
-0.09398025274276733,
0.034773342311382294,
0.02815721370279789,
0.12497938424348831,
-0.13250304758548737,
-0.008634839206933975,
0.05634164437651634,
0.03660309687256813,
0.12911571562290192,
-0.024130839854478836,
-0.06600774079561234,
0.022723514586687088,
-0.1387379765510559,
0.0038192991632968187,
0.11807149648666382,
0.04825948178768158,
-0.028817184269428253,
-0.007465936243534088,
-0.07594949752092361,
-0.18440485000610352,
-0.04713205620646477,
-0.02759041078388691,
0.047773268073797226,
-0.049055568873882294,
-0.06537694483995438,
-0.038376584649086,
-0.09687183797359467,
-0.09018396586179733,
-0.015627330169081688,
0.13662204146385193,
0.029674790799617767,
0.009994328953325748,
-0.012937972322106361,
0.12373210489749908,
-0.02497219853103161,
-0.13787104189395905,
0.0183339174836874,
0.028121907263994217,
-0.050201475620269775,
-0.055565979331731796,
-0.051098257303237915,
-0.05912045016884804,
0.02499462477862835,
0.10811279714107513,
-0.11483490467071533,
0.07846242934465408,
0.027270520105957985,
0.03585895150899887,
-0.10056810826063156,
0.11020925641059875,
-0.07894505560398102,
-0.03155117109417915,
0.0019328388152644038,
0.09816295653581619,
0.009252145886421204,
-0.006704163737595081,
-0.07778798788785934,
0.04627036303281784,
0.09485888481140137,
0.005908279679715633,
-0.059973131865262985,
0.03267715126276016,
-0.05938952416181564,
0.0035593186039477587,
-0.010138509795069695,
-0.09509250521659851,
0.037905529141426086,
0.018856173381209373,
-0.06015368551015854,
-0.06889256834983826,
-0.0016130066942423582,
0.01947939582169056,
0.012686870992183685,
0.09947725385427475,
-0.11165942251682281,
0.04139838367700577,
-0.09646181017160416,
-0.10963635891675949,
-0.0003020008734893054,
-0.09736506640911102,
0.0058286916464567184,
-0.06837150454521179,
-0.18211449682712555,
-0.04720905050635338,
0.05991590395569801,
-0.0688217282295227,
-0.009089008904993534,
-0.0669877752661705,
-0.08624278008937836,
0.006234634667634964,
-0.0061441343277692795,
0.13295596837997437,
-0.0720500499010086,
0.11740190535783768,
0.010730749927461147,
0.06352727860212326,
-0.034522876143455505,
0.02324567176401615,
-0.08312670141458511,
0.05041271448135376,
-0.16963501274585724,
0.0380663126707077,
-0.07144483178853989,
0.05580960959196091,
-0.10144287347793579,
-0.07350873947143555,
-0.04454862326383591,
-0.000013369571206567343,
0.10788333415985107,
0.1416245698928833,
-0.23018522560596466,
-0.04733661562204361,
0.17467516660690308,
-0.06468960642814636,
-0.10441341251134872,
0.11144784092903137,
-0.061136990785598755,
0.069703608751297,
0.05556371435523033,
0.2346910685300827,
0.06791071593761444,
-0.09512002766132355,
0.021255258470773697,
-0.027774443849921227,
0.07203938066959381,
-0.02901647426187992,
0.04493526369333267,
-0.002211357932537794,
0.013456754386425018,
0.010094497352838516,
-0.044908326119184494,
0.03246024623513222,
-0.10870838165283203,
-0.07639623433351517,
-0.02891865186393261,
-0.11671330779790878,
0.04087952524423599,
0.06346730887889862,
0.05967417731881142,
-0.11445116251707077,
-0.08125585317611694,
0.0729861706495285,
0.08685842156410217,
-0.06347980350255966,
0.023541755974292755,
-0.04520662873983383,
0.07322319597005844,
-0.06661801040172577,
-0.03455647453665733,
-0.1606399267911911,
-0.05598130077123642,
-0.0006938714650459588,
0.03380802273750305,
-0.012719337828457355,
0.00858750008046627,
0.1050848439335823,
0.08008013665676117,
-0.07399877905845642,
-0.028961729258298874,
-0.030832357704639435,
0.01634456031024456,
-0.13313455879688263,
-0.22868900001049042,
-0.025350037962198257,
-0.039931099861860275,
0.14973914623260498,
-0.24967777729034424,
0.03069368004798889,
0.009287594817578793,
0.09772752970457077,
0.04363121837377548,
-0.06143045425415039,
-0.018756939098238945,
0.06442733108997345,
-0.016328275203704834,
-0.07701224088668823,
0.058227792382240295,
-0.00811673142015934,
-0.07129468023777008,
-0.051821693778038025,
-0.1277426779270172,
0.1588500440120697,
0.09788785129785538,
-0.0067500644363462925,
-0.11719320714473724,
-0.020060693845152855,
-0.06383365392684937,
-0.03538419306278229,
-0.04425198584794998,
0.02921447716653347,
0.1123788133263588,
0.016713010147213936,
0.12442520260810852,
-0.08055159449577332,
-0.04556199535727501,
0.025817446410655975,
-0.03826596215367317,
0.039349060505628586,
0.14547277987003326,
0.1027207225561142,
-0.07989784330129623,
0.1390988677740097,
0.11244852095842361,
-0.07140138000249863,
0.14649251103401184,
-0.0369589701294899,
-0.08692900091409683,
-0.03879717364907265,
0.028814056888222694,
0.0012511368840932846,
0.17377014458179474,
-0.03895764797925949,
0.019398188218474388,
-0.007708882447332144,
0.017723284661769867,
0.02365024760365486,
-0.22803376615047455,
-0.05716189742088318,
0.01945422776043415,
-0.04769536852836609,
-0.0250412505120039,
-0.029777314513921738,
-0.011037188582122326,
0.10990245640277863,
-0.007081951480358839,
-0.06582652032375336,
0.006760128308087587,
0.010014157742261887,
-0.07502580434083939,
0.21414074301719666,
-0.08808839321136475,
-0.0855826735496521,
-0.07326254993677139,
-0.015361536294221878,
-0.03271793574094772,
0.004977849777787924,
0.041550878435373306,
-0.08203784376382828,
-0.0026689679361879826,
-0.1139020100235939,
-0.04412128031253815,
0.02698346972465515,
0.004534939303994179,
-0.022214684635400772,
-0.023764317855238914,
0.08485632389783859,
-0.08543433994054794,
0.009595354087650776,
-0.055720582604408264,
-0.044568032026290894,
0.04764220491051674,
0.049532700330019,
0.12557588517665863,
0.13754487037658691,
0.000525072799064219,
0.01579950749874115,
-0.018498092889785767,
0.24570807814598083,
-0.09135231375694275,
0.019037345424294472,
0.07833889871835709,
-0.0026684908661991358,
0.05489165708422661,
0.12830370664596558,
0.06843432784080505,
-0.11229722946882248,
0.012228096835315228,
0.05882595106959343,
-0.037751853466033936,
-0.18663731217384338,
-0.01884530670940876,
-0.04170246049761772,
-0.040537621825933456,
0.07521028071641922,
0.03727954626083374,
-0.0199802964925766,
0.04873175919055939,
0.048182081431150436,
0.0005681146867573261,
-0.03831374645233154,
0.04508824646472931,
0.006840724963694811,
0.032386429607868195,
0.10545129328966141,
-0.026940466836094856,
-0.041670091450214386,
0.03084706887602806,
0.012186707928776741,
0.22409260272979736,
-0.023174114525318146,
0.09783320873975754,
0.05835004895925522,
0.18484921753406525,
-0.02964317426085472,
0.0657220333814621,
0.004627258516848087,
-0.06602093577384949,
-0.009344421327114105,
-0.04782475531101227,
-0.017251955345273018,
0.029783444479107857,
-0.05189066007733345,
0.05219132453203201,
-0.08584146201610565,
-0.0025030469987541437,
0.06418026983737946,
0.26638224720954895,
0.07273949682712555,
-0.31660255789756775,
-0.06702245771884918,
0.014224578626453876,
-0.007274703588336706,
-0.038082778453826904,
0.016444502398371696,
0.18470731377601624,
-0.059880580753088,
0.012792659923434258,
-0.07502897083759308,
0.07352998107671738,
-0.013636808842420578,
0.04062502831220627,
0.051792845129966736,
0.11996368318796158,
-0.03951842710375786,
0.04631970822811127,
-0.2604597210884094,
0.30264589190483093,
0.025303836911916733,
0.08832623809576035,
-0.030990978702902794,
-0.03442686051130295,
0.02733892761170864,
0.05401914194226265,
0.09112713485956192,
0.003653946565464139,
-0.06787262111902237,
-0.214732363820076,
-0.08236999064683914,
0.04263701289892197,
0.11280691623687744,
-0.05232715606689453,
0.10208609700202942,
-0.0006966638029552996,
0.0072728474624454975,
0.0548417828977108,
-0.04034193977713585,
-0.08744407445192337,
-0.06335712969303131,
-0.01688171736896038,
0.0034978960175067186,
-0.044043418020009995,
-0.07434258610010147,
-0.09827132523059845,
-0.08216052502393723,
0.1160680502653122,
-0.025355348363518715,
-0.023843752220273018,
-0.13643048703670502,
0.07975209504365921,
0.08744776993989944,
-0.06329968571662903,
0.008118526078760624,
0.03822785243391991,
0.053574468940496445,
0.037632834166288376,
-0.019467130303382874,
0.1306694895029068,
-0.05711135268211365,
-0.17024321854114532,
-0.06931424140930176,
0.11161413043737411,
0.07476911693811417,
0.058564525097608566,
-0.021650848910212517,
0.03694190829992294,
-0.005916919559240341,
-0.08603953570127487,
0.06317586451768875,
-0.024188630282878876,
0.057122208178043365,
0.008149828761816025,
-0.05729208141565323,
0.09012230485677719,
-0.05358031392097473,
-0.028842654079198837,
0.13064824044704437,
0.31485891342163086,
-0.08558482676744461,
0.03264236077666283,
0.03124338574707508,
-0.06772368401288986,
-0.19668380916118622,
0.08333989977836609,
0.06519432365894318,
0.016796134412288666,
0.08949461579322815,
-0.16891968250274658,
0.09010940790176392,
0.13112011551856995,
-0.0156309362500906,
0.15544620156288147,
-0.31041982769966125,
-0.1200411319732666,
0.08292413502931595,
0.15709896385669708,
0.08511045575141907,
-0.15657640993595123,
-0.028380515053868294,
0.00005091768616694026,
-0.07645901292562485,
0.06837619096040726,
-0.15608984231948853,
0.11113076657056808,
-0.00573023222386837,
0.05454253405332565,
0.021024275571107864,
-0.06299839168787003,
0.14018867909908295,
-0.008233503438532352,
0.12830699980258942,
-0.033351704478263855,
0.008242348209023476,
0.03902037814259529,
-0.05522875860333443,
0.0013965257676318288,
-0.07211942225694656,
0.03716721758246422,
-0.07075811922550201,
-0.0061570629477500916,
-0.098596952855587,
0.021584797650575638,
-0.04201142489910126,
-0.05860880762338638,
-0.021860742941498756,
0.055050741881132126,
0.053298626095056534,
-0.009122257120907307,
0.10538554936647415,
-0.023381490260362625,
0.1870332658290863,
0.1251845508813858,
0.055642176419496536,
-0.05341647192835808,
-0.03857783228158951,
0.006203611381351948,
-0.01715310662984848,
0.03639397770166397,
-0.15383197367191315,
0.025116363540291786,
0.13901947438716888,
0.034951355308294296,
0.12664192914962769,
0.0662769302725792,
-0.05241917818784714,
0.011303907260298729,
0.06615012884140015,
-0.16791053116321564,
-0.13011260330677032,
0.002876459388062358,
-0.06412556022405624,
-0.11590497940778732,
0.045871552079916,
0.11993513256311417,
-0.06709654629230499,
-0.000989678781479597,
-0.013591308146715164,
0.034042757004499435,
-0.05220678821206093,
0.23170354962348938,
0.0757964476943016,
0.05951707437634468,
-0.08639376610517502,
0.061736904084682465,
0.023848488926887512,
-0.04562723636627197,
0.029844049364328384,
0.06965100765228271,
-0.0788385197520256,
-0.026924293488264084,
0.106993168592453,
0.17938005924224854,
0.0023221843875944614,
-0.029517268761992455,
-0.15274299681186676,
-0.09386172890663147,
0.05236073210835457,
0.19277048110961914,
0.0644853413105011,
0.007374440785497427,
-0.01989470049738884,
0.010413303971290588,
-0.1294172704219818,
0.09920388460159302,
0.0506916344165802,
0.07825277745723724,
-0.12677372992038727,
0.16413183510303497,
-0.02087428793311119,
0.0212709940969944,
-0.03063862770795822,
0.04772619158029556,
-0.13095024228096008,
0.009952571243047714,
-0.1414404958486557,
-0.042798981070518494,
-0.007475427817553282,
-0.010217860341072083,
0.0027819122187793255,
-0.06174574792385101,
-0.07062938064336777,
0.010230621322989464,
-0.12607605755329132,
-0.023438576608896255,
0.03758550435304642,
0.025413457304239273,
-0.12107666581869125,
-0.027591072022914886,
0.03209853544831276,
-0.0636485293507576,
0.06078792363405228,
0.027056124061346054,
0.023382719606161118,
0.0560220442712307,
-0.16727130115032196,
0.02712647244334221,
0.03414669632911682,
-0.0252641960978508,
0.05864813178777695,
-0.07378300279378891,
-0.02963363565504551,
-0.03460480272769928,
0.05943389609456062,
0.021246133372187614,
0.06557978689670563,
-0.11581513285636902,
0.019365767017006874,
-0.046793147921562195,
-0.06646285206079483,
-0.054091088473796844,
0.021246129646897316,
0.05415232852101326,
0.0362093560397625,
0.1558142751455307,
-0.08803924918174744,
0.029498349875211716,
-0.22248776257038116,
-0.02034737728536129,
0.00579687487334013,
-0.09266579896211624,
-0.08807787299156189,
-0.03689205273985863,
0.09708250313997269,
-0.056820522993803024,
0.10375139862298965,
0.006330639589577913,
0.03768323361873627,
0.027019185945391655,
-0.07553499937057495,
-0.003909089602530003,
0.030158452689647675,
0.18378694355487823,
0.011252524331212044,
-0.04346555098891258,
0.054418958723545074,
0.06296195089817047,
0.07965601235628128,
0.10203129053115845,
0.22018001973628998,
0.14247721433639526,
-0.011499806307256222,
0.0812484622001648,
0.012270096689462662,
-0.08515191078186035,
-0.09428107738494873,
0.09987671673297882,
-0.05953009054064751,
0.07454203814268112,
-0.040017202496528625,
0.19493308663368225,
0.09821643680334091,
-0.1761724352836609,
0.025567056611180305,
-0.05940790846943855,
-0.10529542714357376,
-0.11135810613632202,
-0.03241019695997238,
-0.08673594892024994,
-0.1530812829732895,
-0.0011722473427653313,
-0.1169605553150177,
0.01939057931303978,
0.13991378247737885,
0.017919743433594704,
0.00845604669302702,
0.19902074337005615,
0.06023385748267174,
0.03935873135924339,
0.04982159659266472,
0.016684694215655327,
-0.017386361956596375,
-0.07892806828022003,
-0.10045497864484787,
0.020900394767522812,
-0.04287451505661011,
0.034921810030937195,
-0.06799342483282089,
-0.04637578874826431,
0.042629119008779526,
-0.00687576225027442,
-0.10364029556512833,
0.017837470397353172,
0.020751021802425385,
0.04833489656448364,
0.023759806528687477,
0.03434325009584427,
-0.0038818379398435354,
-0.025718413293361664,
0.26114729046821594,
-0.07455221563577652,
-0.025338465347886086,
-0.11045248806476593,
0.27962809801101685,
0.03426988422870636,
0.009116359986364841,
0.008314642123878002,
-0.10912881791591644,
0.029993953183293343,
0.19377732276916504,
0.17871662974357605,
-0.1447213888168335,
0.0001222216960741207,
-0.02090114913880825,
-0.004913925658911467,
-0.048542916774749756,
0.10501398891210556,
0.08847755193710327,
0.0015257165068760514,
-0.10211747884750366,
-0.036073144525289536,
-0.03991929069161415,
-0.026316598057746887,
-0.05713674798607826,
0.04901999235153198,
0.050717081874608994,
0.03550947457551956,
-0.07268822193145752,
0.07956857979297638,
-0.04820942133665085,
-0.15494494140148163,
0.06509900093078613,
-0.1908179074525833,
-0.1717606782913208,
-0.0215581264346838,
0.06417504698038101,
-0.011731285601854324,
0.05568881332874298,
-0.037734825164079666,
0.004998796619474888,
0.0650126188993454,
-0.03305327892303467,
-0.03026517480611801,
-0.10729440301656723,
0.0762828141450882,
-0.13300667703151703,
0.22477173805236816,
-0.051442548632621765,
0.0035011242143809795,
0.12153198570013046,
0.0156365018337965,
-0.08063510805368423,
0.08711157739162445,
0.05140496790409088,
-0.09995441883802414,
0.0033168483059853315,
0.123443104326725,
-0.043150484561920166,
0.07701774686574936,
0.05625130236148834,
-0.12771926820278168,
0.006772032473236322,
-0.05328434333205223,
-0.04425916075706482,
-0.0394635908305645,
-0.03753950819373131,
-0.04726850986480713,
0.13841579854488373,
0.1976427435874939,
-0.034196335822343826,
0.04658418893814087,
-0.07057204842567444,
0.03551538288593292,
0.05345093086361885,
0.06869060546159744,
-0.042099643498659134,
-0.25172290205955505,
0.032135866582393646,
0.0705474391579628,
-0.032222967594861984,
-0.25400757789611816,
-0.07615812867879868,
0.006751911714673042,
-0.050789956003427505,
-0.05723167210817337,
0.09773135185241699,
0.09221780300140381,
0.06563235819339752,
-0.06120911240577698,
-0.10276708751916885,
-0.06078539788722992,
0.16739384829998016,
-0.1234743520617485,
-0.07787147164344788
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
<details><summary>See axolotl config</summary>
axolotl version: `0.4.0`
```yaml
base_model: 152334H/miqu-1-70b-sf
model_type: LlamaForCausalLM
tokenizer_type: LlamaTokenizer
is_llama_derived_model: true
load_in_8bit: false
load_in_4bit: true
strict: false
datasets:
- path: teknium/OpenHermes-2.5
type: sharegpt
conversation: chatml
dataset_prepared_path: hermes-prepped
val_set_size: 0
output_dir: ./qlora-hermes
adapter: qlora
lora_model_dir:
sequence_len: 4096
sample_packing: true
pad_to_sequence_len: true
lora_r: 32
lora_alpha: 16
lora_dropout: 0.05
lora_target_modules:
lora_target_linear: true
lora_fan_in_fan_out:
wandb_project: huggingface
wandb_entity: 152334h
wandb_watch:
wandb_name: hermes2-miqu
wandb_log_model:
gradient_accumulation_steps: 16
micro_batch_size: 2
num_epochs: 3
optimizer: paged_adamw_8bit
lr_scheduler: cosine
learning_rate: 0.0001
train_on_inputs: false
group_by_length: false
bf16: auto
fp16:
tf32: false
gradient_checkpointing: true
early_stopping_patience:
resume_from_checkpoint:
local_rank:
logging_steps: 1
xformers_attention:
flash_attention: true
warmup_steps: 10
evals_per_epoch: 4
eval_table_size:
eval_sample_packing: false
saves_per_epoch: 1
debug:
deepspeed:
weight_decay: 0.05
fsdp:
fsdp_config:
save_safetensors: true
resize_token_embeddings_to_32x: true
lora_modules_to_save:
- embed_tokens
- lm_head
special_tokens:
eos_token: "<|im_end|>"
tokens:
- "<|im_start|>"
- "<|im_end|>"
```
</details><br>
# qlora-hermes
This model is a fine-tuned version of [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf) on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 2
- eval_batch_size: 2
- seed: 42
- distributed_type: multi-GPU
- num_devices: 6
- gradient_accumulation_steps: 16
- total_train_batch_size: 192
- total_eval_batch_size: 12
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_steps: 10
- num_epochs: 3
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.38.0.dev0
- Pytorch 2.2.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.0 | {"library_name": "peft", "tags": ["generated_from_trainer"], "base_model": "152334H/miqu-1-70b-sf", "model-index": [{"name": "qlora-hermes", "results": []}]} | null | 152334H/miqu-1-70b-hermes2.5-qlora | [
"peft",
"safetensors",
"llama",
"generated_from_trainer",
"base_model:152334H/miqu-1-70b-sf",
"4-bit",
"region:us"
] | 2024-02-08T07:53:23+00:00 | [] | [] | TAGS
#peft #safetensors #llama #generated_from_trainer #base_model-152334H/miqu-1-70b-sf #4-bit #region-us
|
<img src="URL alt="Built with Axolotl" width="200" height="32"/>
<details><summary>See axolotl config</summary>
axolotl version: '0.4.0'
</details><br>
# qlora-hermes
This model is a fine-tuned version of 152334H/miqu-1-70b-sf on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 2
- eval_batch_size: 2
- seed: 42
- distributed_type: multi-GPU
- num_devices: 6
- gradient_accumulation_steps: 16
- total_train_batch_size: 192
- total_eval_batch_size: 12
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_steps: 10
- num_epochs: 3
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.38.0.dev0
- Pytorch 2.2.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.0 | [
"# qlora-hermes\n\nThis model is a fine-tuned version of 152334H/miqu-1-70b-sf on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 2\n- eval_batch_size: 2\n- seed: 42\n- distributed_type: multi-GPU\n- num_devices: 6\n- gradient_accumulation_steps: 16\n- total_train_batch_size: 192\n- total_eval_batch_size: 12\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_steps: 10\n- num_epochs: 3",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.38.0.dev0\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.0"
] | [
"TAGS\n#peft #safetensors #llama #generated_from_trainer #base_model-152334H/miqu-1-70b-sf #4-bit #region-us \n",
"# qlora-hermes\n\nThis model is a fine-tuned version of 152334H/miqu-1-70b-sf on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 2\n- eval_batch_size: 2\n- seed: 42\n- distributed_type: multi-GPU\n- num_devices: 6\n- gradient_accumulation_steps: 16\n- total_train_batch_size: 192\n- total_eval_batch_size: 12\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_steps: 10\n- num_epochs: 3",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.38.0.dev0\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.0"
] | [
44,
36,
6,
12,
8,
3,
157,
4,
44
] | [
"passage: TAGS\n#peft #safetensors #llama #generated_from_trainer #base_model-152334H/miqu-1-70b-sf #4-bit #region-us \n# qlora-hermes\n\nThis model is a fine-tuned version of 152334H/miqu-1-70b-sf on the None dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 2\n- eval_batch_size: 2\n- seed: 42\n- distributed_type: multi-GPU\n- num_devices: 6\n- gradient_accumulation_steps: 16\n- total_train_batch_size: 192\n- total_eval_batch_size: 12\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_steps: 10\n- num_epochs: 3### Training results### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.38.0.dev0\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.0"
] | [
-0.1500214785337448,
0.1520698070526123,
-0.0025102978106588125,
0.08215036988258362,
0.11982252448797226,
0.027967600151896477,
0.09121426194906235,
0.14306072890758514,
-0.07972313463687897,
0.10937033593654633,
0.09621958434581757,
0.009688599966466427,
0.057564977556467056,
0.14809472858905792,
-0.015221117064356804,
-0.22698207199573517,
0.01501032430678606,
-0.0176819059997797,
-0.08116529881954193,
0.09054985642433167,
0.100681833922863,
-0.08525625616312027,
0.06813614070415497,
0.019324982538819313,
-0.12454506009817123,
-0.015462699346244335,
-0.04203560948371887,
-0.04781626909971237,
0.07855764031410217,
0.02650066837668419,
0.07128702849149704,
0.00014018750516697764,
0.10815025866031647,
-0.1894863396883011,
-0.009222924709320068,
0.07579643279314041,
0.03670375421643257,
0.09330011904239655,
0.05512760207056999,
0.010780800133943558,
0.07392227649688721,
-0.12927694618701935,
0.076941579580307,
0.016943560913205147,
-0.08241458237171173,
-0.17745399475097656,
-0.07848164439201355,
0.07717347145080566,
0.10950750857591629,
0.10341648012399673,
-0.001111328136175871,
0.15871833264827728,
-0.05421440675854683,
0.05869658663868904,
0.15593035519123077,
-0.27833548188209534,
-0.08027756214141846,
0.05302436277270317,
0.04565911740064621,
0.0625375509262085,
-0.11524023115634918,
-0.02447974868118763,
0.047955792397260666,
0.02550739422440529,
0.07221999764442444,
0.021255137398838997,
0.07843837887048721,
-0.020709501579403877,
-0.11247222125530243,
-0.040797363966703415,
0.1675233691930771,
0.07883712649345398,
-0.06258384883403778,
-0.10546696931123734,
-0.04555943235754967,
-0.15568134188652039,
-0.013199013657867908,
-0.011012520641088486,
0.015497659333050251,
-0.04002230241894722,
-0.07601245492696762,
-0.031066156923770905,
-0.06609233468770981,
-0.07254812866449356,
0.0380510650575161,
0.13833998143672943,
0.04284502938389778,
0.011080729775130749,
-0.00802616961300373,
0.10690929740667343,
-0.024769727140665054,
-0.1390705406665802,
-0.02933083102107048,
-0.01598123461008072,
-0.08499334752559662,
-0.046849243342876434,
-0.04689455032348633,
-0.00607773894444108,
-0.016402140259742737,
0.14418262243270874,
-0.05630737170577049,
0.06932006776332855,
0.05902571976184845,
0.0026026873383671045,
-0.029078029096126556,
0.12717927992343903,
-0.0521094985306263,
-0.04731016606092453,
0.0006929785595275462,
0.14027726650238037,
0.017236731946468353,
-0.010543887503445148,
-0.07196276634931564,
-0.032480109483003616,
0.08268812298774719,
0.04499872028827667,
-0.04081084951758385,
0.004936324432492256,
-0.03849930688738823,
-0.03632790222764015,
0.08574222028255463,
-0.1045427918434143,
0.03551987186074257,
-0.006613048259168863,
-0.07922253012657166,
-0.03726516291499138,
0.002338231075555086,
0.004275178536772728,
-0.02140718325972557,
0.08039290457963943,
-0.09908708184957504,
-0.033172380179166794,
-0.0690493956208229,
-0.05114291235804558,
0.00005110864731250331,
-0.02215173840522766,
0.0011224399786442518,
-0.08402989059686661,
-0.15905337035655975,
-0.03780020773410797,
0.04507984593510628,
-0.08178907632827759,
-0.07824260741472244,
-0.02441311813890934,
-0.0822712779045105,
0.039781972765922546,
0.0059441495686769485,
0.1402820646762848,
-0.04395832493901253,
0.09352440387010574,
0.025443026795983315,
0.019231565296649933,
0.04469474405050278,
0.026446286588907242,
-0.061880726367235184,
0.05954268202185631,
-0.1033036857843399,
0.07911541312932968,
-0.07700029760599136,
0.008239426650106907,
-0.11511129140853882,
-0.10041103512048721,
-0.02876698412001133,
-0.029522204771637917,
0.08431906253099442,
0.12392765283584595,
-0.12811380624771118,
-0.025455258786678314,
0.16622094810009003,
-0.07187715917825699,
-0.09846806526184082,
0.10036476701498032,
-0.023015618324279785,
-0.018565980717539787,
0.03525383397936821,
0.11643455922603607,
0.15938441455364227,
-0.12098591029644012,
-0.03600774332880974,
0.030067266896367073,
0.10415611416101456,
-0.009746216237545013,
0.11270696669816971,
-0.011778205633163452,
0.004179583862423897,
0.0005776035832241178,
-0.05255885794758797,
0.012284095399081707,
-0.07527995854616165,
-0.08548665791749954,
-0.050306014716625214,
-0.1067311018705368,
0.028962282463908195,
0.015210752375423908,
0.041054196655750275,
-0.06374815106391907,
-0.14819590747356415,
0.018010396510362625,
0.15613999962806702,
-0.04873744398355484,
-0.004828257951885462,
-0.07939979434013367,
0.0686296671628952,
-0.0863020122051239,
-0.030611462891101837,
-0.1768079698085785,
-0.0802340880036354,
0.05675366148352623,
-0.0746898502111435,
-0.008358767256140709,
-0.002868554089218378,
0.06964166462421417,
0.10073959827423096,
-0.06904631108045578,
-0.04417429119348526,
-0.106163889169693,
-0.019285527989268303,
-0.10658586770296097,
-0.14493916928768158,
-0.06782956421375275,
-0.038087397813797,
0.2136157751083374,
-0.2392093986272812,
0.004368985071778297,
-0.007086862809956074,
0.13690118491649628,
0.03241380676627159,
-0.0708446204662323,
0.0009430334903299809,
0.03520796447992325,
-0.005963173229247332,
-0.10373280197381973,
0.02698439732193947,
0.01190634910017252,
-0.11086822301149368,
-0.08229086548089981,
-0.14298862218856812,
0.07103262096643448,
0.07417843490839005,
0.1333095282316208,
-0.09254497289657593,
-0.052690424025058746,
-0.06860808283090591,
-0.048786457628011703,
-0.04543724283576012,
-0.019605228677392006,
0.14744050800800323,
0.03530074656009674,
0.12369772046804428,
-0.07479896396398544,
-0.06458844244480133,
0.025961747393012047,
0.021167587488889694,
-0.02681610733270645,
0.0795871838927269,
0.03120390884578228,
-0.11954866349697113,
0.07245134562253952,
0.07882877439260483,
-0.06873829662799835,
0.1263887584209442,
-0.056807443499565125,
-0.1033322811126709,
-0.04449697211384773,
0.03309495374560356,
0.020128216594457626,
0.1379171460866928,
-0.056534379720687866,
0.026252035051584244,
0.03259255364537239,
0.023282010108232498,
0.027165105566382408,
-0.15536104142665863,
-0.014838716015219688,
0.03627882897853851,
-0.024372616782784462,
-0.026146305724978447,
-0.017784828320145607,
-0.011906362138688564,
0.07598830759525299,
0.030038151890039444,
-0.01139130163937807,
0.0178024023771286,
-0.009655916132032871,
-0.07169961929321289,
0.18224294483661652,
-0.0955759808421135,
-0.10638406127691269,
-0.14631639420986176,
0.07329743355512619,
-0.06462734192609787,
-0.017629489302635193,
0.0023893709294497967,
-0.05936388671398163,
-0.035509899258613586,
-0.09677989780902863,
-0.04989242181181908,
-0.06844537705183029,
0.00794602744281292,
0.056782450526952744,
0.01044414285570383,
0.11310262978076935,
-0.10913681238889694,
0.01723913662135601,
0.01525556668639183,
-0.04485519230365753,
-0.008315798826515675,
0.034710872918367386,
0.09214968234300613,
0.08930545300245285,
-0.0034351670183241367,
0.018724896013736725,
-0.018628114834427834,
0.27041521668434143,
-0.09224487841129303,
-0.015529119409620762,
0.11564923077821732,
0.029001720249652863,
0.06059420481324196,
0.0910734310746193,
0.031347088515758514,
-0.08093449473381042,
0.013534093275666237,
0.0506737157702446,
-0.020562533289194107,
-0.20563408732414246,
-0.02842901088297367,
-0.026579298079013824,
-0.0673050805926323,
0.1184663325548172,
0.05039774253964424,
-0.019847040995955467,
0.07315821945667267,
-0.030902624130249023,
0.06133735552430153,
-0.03796278312802315,
0.09337113797664642,
0.02576962299644947,
0.07358361035585403,
0.09867250174283981,
-0.016573894768953323,
-0.006058932282030582,
0.057599350810050964,
0.030335357412695885,
0.229780375957489,
-0.050633300095796585,
0.10840215533971786,
0.013973228633403778,
0.16502897441387177,
-0.015415438450872898,
0.045324068516492844,
0.010699840262532234,
-0.005571752320975065,
-0.00040179098141379654,
-0.06996306777000427,
-0.042000725865364075,
0.04420166090130806,
0.03935646638274193,
0.06022636964917183,
-0.10764804482460022,
0.058639395982027054,
-0.004725232720375061,
0.27159252762794495,
0.06867989152669907,
-0.3182535767555237,
-0.10333457589149475,
0.017542237415909767,
-0.011555486358702183,
-0.0739990621805191,
0.013523951172828674,
0.14839254319667816,
-0.11885179579257965,
0.05531264841556549,
-0.07056199759244919,
0.07670888304710388,
-0.04341241344809532,
-0.005128094926476479,
0.058998048305511475,
0.1070270836353302,
0.0010377619182690978,
0.08617430925369263,
-0.1807606816291809,
0.20652078092098236,
0.02085522934794426,
0.11400788277387619,
-0.05080849677324295,
0.03916265070438385,
0.012838653288781643,
0.03714454919099808,
0.10094259679317474,
-0.0004088591958861798,
-0.037148453295230865,
-0.18693576753139496,
-0.11735223978757858,
0.039052970707416534,
0.09389205276966095,
-0.07956358045339584,
0.1003585010766983,
-0.03594355657696724,
-0.0020933051127940416,
0.02313617616891861,
0.008683732710778713,
-0.13102498650550842,
-0.1330285668373108,
0.02694576606154442,
0.012620429508388042,
-0.025464849546551704,
-0.09393968433141708,
-0.09587644785642624,
-0.001457562786526978,
0.14397591352462769,
-0.004981765989214182,
-0.0547139048576355,
-0.13710543513298035,
0.06677308678627014,
0.1425914764404297,
-0.06779374927282333,
0.012679342180490494,
0.008868565782904625,
0.15652577579021454,
0.027588795870542526,
-0.030350590124726295,
0.04959724098443985,
-0.0634319856762886,
-0.18156863749027252,
-0.05603465437889099,
0.1478288471698761,
0.022271160036325455,
0.04289206862449646,
-0.006241494789719582,
0.016523776575922966,
-0.0016018096357584,
-0.08620039373636246,
0.04215623438358307,
0.0453152097761631,
0.08118046820163727,
0.025334373116493225,
-0.05166882649064064,
0.07125817239284515,
-0.03624846041202545,
-0.014753754250705242,
0.08687543123960495,
0.3016160726547241,
-0.0762288048863411,
0.0070522199384868145,
0.06745196878910065,
-0.038933075964450836,
-0.11736326664686203,
0.010307405143976212,
0.13835176825523376,
0.04061850532889366,
0.02957049198448658,
-0.18130390346050262,
0.05775834247469902,
0.1316252052783966,
-0.025516560301184654,
0.05381859838962555,
-0.29767709970474243,
-0.11659165471792221,
0.054065704345703125,
0.10223519057035446,
-0.012816975824534893,
-0.13886988162994385,
-0.06351490318775177,
-0.03460272401571274,
-0.11192011088132858,
0.07887354493141174,
-0.0669502317905426,
0.11313356459140778,
-0.006685066968202591,
0.05583377555012703,
0.036363277584314346,
-0.04158634692430496,
0.19000400602817535,
0.017345571890473366,
0.0609884038567543,
-0.02095966599881649,
0.03358083218336105,
0.060516197234392166,
-0.08647378534078598,
0.03945475444197655,
-0.06949488073587418,
0.08675757795572281,
-0.16984574496746063,
-0.015544643625617027,
-0.06309463083744049,
0.02738819271326065,
-0.05486231669783592,
-0.054068680852651596,
-0.032135143876075745,
0.06261119246482849,
0.06273926794528961,
-0.03063887357711792,
0.06432322412729263,
0.03729792684316635,
0.09629359096288681,
0.12525124847888947,
0.061391692608594894,
0.027886053547263145,
-0.12720952928066254,
-0.012612085789442062,
-0.017612630501389503,
0.08026789128780365,
-0.12562346458435059,
0.012487536296248436,
0.11159706860780716,
0.03409793972969055,
0.11961784213781357,
0.021248556673526764,
-0.0680868849158287,
0.009393182583153248,
0.043727364391088486,
-0.08959385752677917,
-0.14443570375442505,
-0.03458993136882782,
0.02613329142332077,
-0.15741592645645142,
-0.011058610863983631,
0.12561984360218048,
-0.06161996349692345,
-0.013249905779957771,
-0.009302179329097271,
0.0059175798669457436,
-0.026230158284306526,
0.20360137522220612,
0.03375210240483284,
0.08889847993850708,
-0.06313465535640717,
0.11230882257223129,
0.07715565711259842,
-0.07786242663860321,
0.05519939959049225,
0.029726635664701462,
-0.08840359002351761,
-0.014793920330703259,
0.021558042615652084,
0.10003571212291718,
-0.0032671275548636913,
-0.030720200389623642,
-0.07019344717264175,
-0.06954359263181686,
0.051232341676950455,
0.006660443264991045,
0.027583308517932892,
-0.00887605082243681,
-0.027816012501716614,
0.02303726226091385,
-0.14597617089748383,
0.10632573068141937,
0.06023821234703064,
0.07817850261926651,
-0.1561557501554489,
0.09291234612464905,
-0.014221529476344585,
0.02380085550248623,
0.001940797083079815,
0.01914733089506626,
-0.08475308120250702,
-0.025921113789081573,
-0.12998296320438385,
0.0049752951599657536,
-0.018120359629392624,
0.0031260494142770767,
-0.01918979175388813,
-0.0335967130959034,
-0.038178302347660065,
0.03553055226802826,
-0.06860994547605515,
-0.07794489711523056,
0.0067658619955182076,
0.04882149398326874,
-0.11265008896589279,
-0.0018294304609298706,
0.02360427938401699,
-0.11602186411619186,
0.08117305487394333,
0.05048610270023346,
0.047216542065143585,
-0.0039507378824055195,
-0.014403646811842918,
0.025869734585285187,
0.011026354506611824,
0.021502135321497917,
0.04750831797719002,
-0.10529675334692001,
0.005240983795374632,
-0.04587278515100479,
0.020062465220689774,
0.010782669298350811,
0.048858627676963806,
-0.12498483806848526,
-0.04668943211436272,
-0.027298377826809883,
-0.02584155462682247,
-0.06361403316259384,
0.037035081535577774,
0.09168703109025955,
0.021808048710227013,
0.15704850852489471,
-0.059865210205316544,
0.026295345276594162,
-0.2065853476524353,
-0.03464778512716293,
0.005588431376963854,
-0.01395587157458067,
-0.050044938921928406,
-0.024836335331201553,
0.08524069935083389,
-0.06058796867728233,
0.0672733262181282,
-0.03493508696556091,
0.09562569856643677,
0.03269931301474571,
-0.06332488358020782,
0.021139420568943024,
0.007250790484249592,
0.19183649122714996,
0.079411581158638,
-0.005007435102015734,
0.09587086737155914,
-0.02361256815493107,
0.059416063129901886,
0.08940549939870834,
0.11228849738836288,
0.13281480967998505,
0.005154024343937635,
0.07566247880458832,
0.04497382417321205,
-0.11570417881011963,
-0.16689133644104004,
0.07860316336154938,
-0.019132385030388832,
0.10581078380346298,
-0.04207084700465202,
0.12356027215719223,
0.09570827335119247,
-0.1820344179868698,
0.025308040902018547,
-0.06021982058882713,
-0.10968908667564392,
-0.08428890258073807,
-0.08064552396535873,
-0.07859005033969879,
-0.09752029925584793,
0.024106772616505623,
-0.11250639706850052,
0.007086173165589571,
0.09567717462778091,
0.005774081684648991,
-0.0036469032056629658,
0.15665562450885773,
-0.036079999059438705,
0.009181221015751362,
0.06511872261762619,
0.025209806859493256,
0.004662359599024057,
-0.02415814995765686,
-0.07595572620630264,
0.05708723142743111,
-0.02266697958111763,
0.08924610912799835,
-0.05286261439323425,
-0.011017589829862118,
0.053366243839263916,
0.042805951088666916,
-0.07918809354305267,
0.02497486211359501,
0.0012267830315977335,
0.035899534821510315,
0.054452698677778244,
0.03646509349346161,
0.006272910162806511,
-0.05513671413064003,
0.26433390378952026,
-0.07845448702573776,
-0.02520180307328701,
-0.14125685393810272,
0.17390014231204987,
0.013301352970302105,
-0.009898477233946323,
0.06228340417146683,
-0.12151020765304565,
-0.011926569975912571,
0.1217491626739502,
0.11623599380254745,
-0.07960852235555649,
-0.029875829815864563,
0.019406164065003395,
-0.026321563869714737,
-0.07217379659414291,
0.1114272028207779,
0.08900032192468643,
0.0047347755171358585,
-0.05634136125445366,
0.03027859702706337,
-0.012221585027873516,
-0.026786640286445618,
-0.06099938228726387,
0.09671073406934738,
-0.00003353312422404997,
0.014207400381565094,
-0.040357042104005814,
0.06190072000026703,
0.014145356602966785,
-0.1547871232032776,
0.04605773091316223,
-0.14394249022006989,
-0.20956116914749146,
-0.020742328837513924,
0.02145969122648239,
-0.004475997295230627,
0.07697245478630066,
-0.023602882400155067,
-0.0023388497065752745,
0.13078910112380981,
-0.029897641390562057,
-0.022001011297106743,
-0.11167273670434952,
0.07586019486188889,
-0.024793339893221855,
0.2374502718448639,
-0.002475164132192731,
0.05265539512038231,
0.11828248202800751,
0.0036042979918420315,
-0.14633360505104065,
0.03281978890299797,
0.0885171890258789,
-0.06579948216676712,
0.042402684688568115,
0.17678382992744446,
-0.0488554947078228,
0.12952090799808502,
0.05019840970635414,
-0.1065400242805481,
-0.044652216136455536,
-0.022352617233991623,
0.01629551872611046,
-0.07789932936429977,
-0.007030094042420387,
-0.05663105845451355,
0.17154085636138916,
0.20502914488315582,
-0.05340396612882614,
-0.0243475753813982,
-0.03641606494784355,
0.06385692954063416,
0.05362526699900627,
0.09400438517332077,
-0.012625988572835922,
-0.21500487625598907,
0.01080812606960535,
0.018525822088122368,
0.03615795075893402,
-0.25931403040885925,
-0.0868786945939064,
0.0488823801279068,
-0.06224776431918144,
-0.03052632510662079,
0.11874061822891235,
0.03748580440878868,
0.025895478203892708,
-0.053101543337106705,
-0.09375286847352982,
-0.06743715703487396,
0.1268063187599182,
-0.15252086520195007,
-0.07289893180131912
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | mertllc/mms-tts-tur-thirties-male | [
"transformers",
"safetensors",
"vits",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-08T07:57:19+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #vits #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #vits #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
34,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #vits #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.054659612476825714,
0.21414990723133087,
-0.0031807427294552326,
0.026865221560001373,
0.1250888854265213,
0.00032571866177022457,
0.04081440716981888,
0.12862813472747803,
-0.02167222462594509,
0.11129128932952881,
0.03218022361397743,
0.09727001935243607,
0.10339263826608658,
0.16586677730083466,
0.03691011667251587,
-0.21517004072666168,
0.009132993407547474,
-0.09292528033256531,
0.018077509477734566,
0.10867427289485931,
0.13162045180797577,
-0.10489460080862045,
0.07603627443313599,
-0.03790099918842316,
-0.017673974856734276,
-0.0003223843814339489,
-0.0923151820898056,
-0.070840984582901,
0.06550594419240952,
0.06909013539552689,
0.06122942641377449,
0.009997012093663216,
0.10145736485719681,
-0.29726552963256836,
0.01687687449157238,
0.08279260247945786,
-0.004506718832999468,
0.06148726865649223,
0.0646374449133873,
-0.08339887112379074,
0.1029256209731102,
-0.08559336513280869,
0.13652671873569489,
0.08214850723743439,
-0.06937385350465775,
-0.21391066908836365,
-0.06977995485067368,
0.0987061932682991,
0.12011658400297165,
0.06274435669183731,
-0.02326560579240322,
0.1522950381040573,
-0.06972704082727432,
0.012022249400615692,
0.1361677050590515,
-0.09713108092546463,
-0.05137801170349121,
0.049987345933914185,
0.11240657418966293,
0.10166463255882263,
-0.1353231519460678,
0.007596791721880436,
0.04457303136587143,
0.023097742348909378,
0.09194746613502502,
0.020738936960697174,
0.0916183590888977,
0.04564107209444046,
-0.13860996067523956,
-0.03957565128803253,
0.10889606922864914,
0.03478158637881279,
-0.05796414613723755,
-0.21188515424728394,
-0.0026691502425819635,
-0.026535477489233017,
-0.023307178169488907,
-0.05803702771663666,
0.045833978801965714,
-0.03317271173000336,
0.067923404276371,
-0.042256616055965424,
-0.10016343742609024,
-0.03838508576154709,
0.0836847797036171,
0.06997206062078476,
0.013808192685246468,
-0.026154542341828346,
0.03861820325255394,
0.11874474585056305,
0.037009406834840775,
-0.10824361443519592,
-0.0663856491446495,
-0.06518013030290604,
-0.09711762517690659,
-0.04532422870397568,
0.04776853322982788,
0.01869308575987816,
0.030892416834831238,
0.20719914138317108,
-0.0024066849146038294,
0.040300752967596054,
0.01544452179223299,
0.00820851232856512,
0.05608583986759186,
0.09020276367664337,
-0.057233426719903946,
-0.13989022374153137,
-0.04616677761077881,
0.08976847678422928,
-0.00493787182494998,
-0.03551584109663963,
-0.04997507110238075,
0.048379965126514435,
0.05169600620865822,
0.1267518699169159,
0.08646857738494873,
-0.012898874469101429,
-0.05273304134607315,
-0.025197435170412064,
0.22986702620983124,
-0.14503952860832214,
0.04801303148269653,
-0.016220765188336372,
-0.026413746178150177,
-0.04562145099043846,
0.037146687507629395,
0.02893291600048542,
-0.0071297562681138515,
0.09902069717645645,
-0.055000074207782745,
-0.03897455707192421,
-0.10056453198194504,
-0.03981734439730644,
0.04000834375619888,
-0.0014343701768666506,
-0.011925416998565197,
-0.07901987433433533,
-0.1033727377653122,
-0.04151687026023865,
0.0622556135058403,
-0.06062569096684456,
-0.03672588989138603,
0.014433487318456173,
-0.0646335631608963,
-0.011868113651871681,
-0.0046113538555800915,
0.10713792592287064,
-0.03111988678574562,
0.041085705161094666,
-0.03385680913925171,
0.05467362701892853,
0.10134078562259674,
0.03396330401301384,
-0.0692443996667862,
0.05283360555768013,
-0.2253323644399643,
0.0846395194530487,
-0.1103181466460228,
0.040045637637376785,
-0.1649162620306015,
-0.04362662881612778,
0.01545786950737238,
0.01223697792738676,
0.010682502761483192,
0.11813149601221085,
-0.18765069544315338,
-0.02040630392730236,
0.13456352055072784,
-0.09486816823482513,
-0.10925174504518509,
0.07470420002937317,
-0.04261988773941994,
0.14796192944049835,
0.04623936489224434,
-0.017894135788083076,
0.07337126135826111,
-0.16546636819839478,
-0.06534566730260849,
-0.015944186598062515,
-0.01140376552939415,
0.13805019855499268,
0.06177884340286255,
-0.05833873897790909,
0.06357681751251221,
0.02317901886999607,
-0.022351879626512527,
-0.04479735344648361,
-0.05049646645784378,
-0.10716529190540314,
-0.006589649710804224,
-0.0877491682767868,
0.049144841730594635,
-0.008710972033441067,
-0.07987060397863388,
-0.032660458236932755,
-0.18162156641483307,
0.03565994277596474,
0.08912748098373413,
0.006954456213861704,
-0.008257697336375713,
-0.07709750533103943,
0.012575463391840458,
-0.027584582567214966,
-0.010441360995173454,
-0.16807158291339874,
-0.045059818774461746,
0.045085642486810684,
-0.1683385670185089,
0.03666726127266884,
-0.05383622646331787,
0.057435907423496246,
0.04089425876736641,
-0.0608406662940979,
-0.012410139665007591,
-0.020455263555049896,
0.02037479542195797,
-0.03554835915565491,
-0.19715940952301025,
-0.04920884966850281,
-0.033720988780260086,
0.15323609113693237,
-0.2512565553188324,
0.03701164573431015,
0.04283377155661583,
0.1445688009262085,
-0.004499740432947874,
-0.041343484073877335,
0.021006079390645027,
-0.05124713480472565,
-0.04886976629495621,
-0.064845971763134,
-0.003489583032205701,
-0.029771825298666954,
-0.04689984768629074,
0.014419492334127426,
-0.17416127026081085,
-0.03588438406586647,
0.09719391912221909,
0.1012604832649231,
-0.15479636192321777,
-0.018018238246440887,
-0.046819429844617844,
-0.06501296907663345,
-0.08719377964735031,
-0.0634685754776001,
0.12365260720252991,
0.04887883737683296,
0.044603388756513596,
-0.07642911374568939,
-0.06516730040311813,
0.02209198847413063,
0.00037755590165033937,
-0.03342745080590248,
0.07709765434265137,
0.06420876830816269,
-0.09495706856250763,
0.07597044855356216,
0.0879693329334259,
0.07397416979074478,
0.09690815210342407,
0.017737112939357758,
-0.10766889899969101,
-0.025353191420435905,
0.025884538888931274,
0.02590569481253624,
0.14766225218772888,
-0.052133310586214066,
0.03766921907663345,
0.047928281128406525,
-0.048178963363170624,
0.018924955278635025,
-0.09172655642032623,
0.02477680705487728,
0.03108147345483303,
-0.0051895990036427975,
0.04569429159164429,
-0.04261132329702377,
0.0015583503991365433,
0.07553404569625854,
0.0439009927213192,
0.054722823202610016,
0.004550157580524683,
-0.014615098014473915,
-0.09760808199644089,
0.16303586959838867,
-0.09686829894781113,
-0.2844827473163605,
-0.15191766619682312,
0.025421515107154846,
0.038875505328178406,
-0.02202117070555687,
0.031196635216474533,
-0.0685606598854065,
-0.10619828850030899,
-0.10253546386957169,
-0.0007893215515650809,
0.021664658561348915,
-0.07999464124441147,
-0.07771245390176773,
0.07423610240221024,
0.04034431278705597,
-0.14601534605026245,
0.03843066841363907,
0.05174413323402405,
-0.05686575174331665,
-0.020990731194615364,
0.08788161724805832,
0.11919383704662323,
0.15064425766468048,
-0.01956579089164734,
-0.029653063043951988,
0.02179299294948578,
0.18913501501083374,
-0.13056331872940063,
0.10870491713285446,
0.1331699639558792,
-0.0433298796415329,
0.08741360157728195,
0.17486868798732758,
0.02946310304105282,
-0.08184187114238739,
0.04125521704554558,
0.04271497204899788,
-0.0446363128721714,
-0.2628204822540283,
-0.0587831549346447,
0.013565518893301487,
-0.07289978116750717,
0.09574431926012039,
0.09441626816987991,
0.13101495802402496,
0.03733300045132637,
-0.07704862952232361,
-0.042284153401851654,
-0.0007691121427342296,
0.11566338688135147,
-0.04729871824383736,
-0.00864650122821331,
0.08112052828073502,
-0.04204992949962616,
0.0042695761658251286,
0.101866215467453,
0.024085933342576027,
0.18680992722511292,
0.02045324817299843,
0.1325864940881729,
0.06266885250806808,
0.07362587004899979,
-0.00304698059335351,
0.021530818194150925,
0.04571235924959183,
0.016793522983789444,
-0.004352389834821224,
-0.10109587758779526,
0.004940509796142578,
0.14031140506267548,
0.044244058430194855,
0.029351718723773956,
0.0023038540966808796,
-0.025745723396539688,
0.059172797948122025,
0.16894783079624176,
-0.014623390510678291,
-0.20305828750133514,
-0.07212355732917786,
0.07476779818534851,
-0.05524183437228203,
-0.12190999835729599,
-0.03604535013437271,
0.03974858298897743,
-0.17753031849861145,
0.03411399945616722,
-0.020660564303398132,
0.09808827936649323,
-0.0960298478603363,
-0.025731271132826805,
0.017328539863228798,
0.08463997393846512,
-0.017630890011787415,
0.09686511754989624,
-0.15011048316955566,
0.12523487210273743,
0.03229980170726776,
0.0898485779762268,
-0.11468798667192459,
0.08304145932197571,
-0.009098101407289505,
0.016468055546283722,
0.18883956968784332,
-0.00914006493985653,
-0.043279051780700684,
-0.0765409916639328,
-0.09724772721529007,
-0.016675574705004692,
0.12457696348428726,
-0.11865599453449249,
0.08336363732814789,
-0.006434252485632896,
-0.05090279504656792,
0.010499227792024612,
-0.11436042934656143,
-0.17895425856113434,
-0.19684189558029175,
0.061690423637628555,
-0.10233647376298904,
0.01922602578997612,
-0.1105671152472496,
-0.06737665832042694,
-0.029828263446688652,
0.2358294576406479,
-0.14021140336990356,
-0.07348582148551941,
-0.1486395299434662,
-0.049397800117731094,
0.1688835471868515,
-0.039627790451049805,
0.07352027297019958,
-0.014237076044082642,
0.21156272292137146,
-0.0005727469106204808,
-0.0019497170578688383,
0.0662601962685585,
-0.09127254039049149,
-0.17042554914951324,
-0.0796523243188858,
0.1408538520336151,
0.1185344010591507,
0.05187511071562767,
-0.00005241960025159642,
0.008437353186309338,
-0.01933823712170124,
-0.11107131093740463,
-0.005973829887807369,
0.13854430615901947,
0.06674695014953613,
0.03547331318259239,
-0.05006469413638115,
-0.10860110819339752,
-0.06920936703681946,
-0.058358483016490936,
0.05175930634140968,
0.18184207379817963,
-0.1009909063577652,
0.17350798845291138,
0.15878215432167053,
-0.07211574912071228,
-0.21567314863204956,
0.039191193878650665,
0.04846473038196564,
-0.014512532390654087,
0.04614531248807907,
-0.1829945594072342,
0.09505120664834976,
0.015141540206968784,
-0.052736036479473114,
0.12199369817972183,
-0.15728448331356049,
-0.15639621019363403,
0.06087431684136391,
0.04970995709300041,
-0.23623821139335632,
-0.1441342532634735,
-0.08822641521692276,
-0.06784138828516006,
-0.14815589785575867,
0.07915012538433075,
-0.019972164183855057,
0.011897586286067963,
0.04091079905629158,
0.013740893453359604,
0.023185279220342636,
-0.055776987224817276,
0.18284909427165985,
-0.0035617330577224493,
0.014864614233374596,
-0.06912479549646378,
-0.058035630732774734,
0.0975092425942421,
-0.05838471278548241,
0.1184525191783905,
-0.003918026573956013,
0.013672815635800362,
-0.08212041109800339,
-0.05343952775001526,
-0.046617619693279266,
0.05752236396074295,
-0.08050531893968582,
-0.11092408001422882,
-0.04487094283103943,
0.08938708156347275,
0.07764840126037598,
-0.033286161720752716,
-0.010930746793746948,
-0.07634644955396652,
0.10063119232654572,
0.19033774733543396,
0.17030654847621918,
0.018113715574145317,
-0.07677590847015381,
0.015532949939370155,
-0.03924742713570595,
0.04019718989729881,
-0.2505480647087097,
0.03877655416727066,
0.0529145747423172,
0.0354921817779541,
0.1059221550822258,
-0.02500346675515175,
-0.17749741673469543,
-0.0438142865896225,
0.06573881208896637,
-0.045354213565588,
-0.22390563786029816,
-0.009726951830089092,
0.09943331032991409,
-0.1914641559123993,
-0.015451330691576004,
0.02838914282619953,
-0.04480560123920441,
-0.02868090756237507,
0.0007889526314102113,
0.0600614957511425,
0.015805870294570923,
0.09190283715724945,
0.07423794269561768,
0.09749054163694382,
-0.08805927634239197,
0.09811163693666458,
0.10723351687192917,
-0.09035424888134003,
0.03553062304854393,
0.06695880740880966,
-0.0467107780277729,
-0.04594837874174118,
0.05199020728468895,
0.04819667339324951,
0.01212578546255827,
-0.0561964213848114,
0.010319532826542854,
-0.04872706159949303,
0.04633839800953865,
0.10621411353349686,
0.028242740780115128,
-0.03058992512524128,
0.06704547256231308,
0.03252853453159332,
-0.1153404489159584,
0.09847725927829742,
0.012868257239460945,
0.03807265684008598,
-0.06272068619728088,
-0.015808504074811935,
0.04865187034010887,
0.027409857138991356,
-0.01764598675072193,
-0.025427930057048798,
-0.035527609288692474,
-0.015147317200899124,
-0.15422900021076202,
-0.012660279870033264,
-0.07294544577598572,
0.007333413697779179,
0.006807927042245865,
-0.03955657035112381,
-0.0043836915865540504,
0.029364487156271935,
-0.07081043720245361,
-0.06899864971637726,
-0.0017123379511758685,
0.10014908015727997,
-0.16123399138450623,
0.0016520773060619831,
0.07378670573234558,
-0.10700937360525131,
0.06776659190654755,
-0.009028629399836063,
0.006400149781256914,
0.021102426573634148,
-0.1615109145641327,
0.05426544323563576,
-0.010029333643615246,
0.02013414539396763,
0.032934170216321945,
-0.16248436272144318,
0.0024488656781613827,
-0.047329291701316833,
-0.022390197962522507,
-0.004845738876610994,
-0.04656189680099487,
-0.11974798142910004,
0.07715073227882385,
-0.01184067688882351,
-0.05094744265079498,
-0.01612357795238495,
0.05293868109583855,
0.08231643587350845,
-0.03882661834359169,
0.09632368385791779,
-0.005011113826185465,
0.05959545075893402,
-0.17253276705741882,
-0.02932477742433548,
-0.0432354174554348,
0.014331330545246601,
0.01743181422352791,
-0.009555062279105186,
0.03874485567212105,
-0.00935265514999628,
0.22544825077056885,
-0.03915993124246597,
0.16461394727230072,
0.055936723947525024,
-0.0032888432033360004,
0.0007776605198159814,
0.06758615374565125,
0.05568486079573631,
0.03412187471985817,
0.00899792555719614,
0.02200561948120594,
-0.023325180634856224,
-0.006471368949860334,
-0.1553903967142105,
0.02697177603840828,
0.14716137945652008,
0.0745159387588501,
0.006664956454187632,
0.07025619596242905,
-0.1267581284046173,
-0.11370917409658432,
0.09592846781015396,
-0.02568071521818638,
0.008476621471345425,
-0.07835444062948227,
0.12778781354427338,
0.14673273265361786,
-0.14686504006385803,
0.06517019122838974,
-0.053687721490859985,
-0.05600763112306595,
-0.09034380316734314,
-0.10879118740558624,
-0.06126067787408829,
-0.04308179020881653,
0.004678911529481411,
-0.042684826999902725,
0.055097613483667374,
0.04954573139548302,
-0.014461824670433998,
0.004931987728923559,
0.12391652166843414,
-0.006120255216956139,
0.001201988779939711,
0.03766126185655594,
0.03769403696060181,
0.024755796417593956,
-0.059261444956064224,
0.030717262998223305,
0.021477915346622467,
0.034908585250377655,
0.059853747487068176,
0.037230484187603,
-0.045039307326078415,
0.028804119676351547,
0.0020213082898408175,
-0.10957802832126617,
0.023749636486172676,
-0.012328135780990124,
-0.06936221569776535,
0.12969832122325897,
0.03471869230270386,
0.009512413293123245,
-0.037131089717149734,
0.23728759586811066,
-0.062090300023555756,
-0.08014962822198868,
-0.12913139164447784,
0.09616934508085251,
-0.013530191034078598,
0.057892732322216034,
0.03356536477804184,
-0.12210189551115036,
0.0036616562865674496,
0.13605539500713348,
0.11633196473121643,
-0.0003361425769980997,
0.012180927209556103,
0.044184453785419464,
0.004239979665726423,
-0.06263455748558044,
0.044461920857429504,
0.06619330495595932,
0.12273700535297394,
-0.07938622683286667,
0.07410858571529388,
0.00435013510286808,
-0.08385829627513885,
-0.0399140790104866,
0.1140546128153801,
-0.03326992690563202,
0.03303933143615723,
-0.041518088430166245,
0.10997304320335388,
-0.059399381279945374,
-0.3032641112804413,
0.03540288656949997,
-0.10066618025302887,
-0.1533578634262085,
-0.01690032333135605,
0.06605888903141022,
-0.02134985849261284,
0.01722477562725544,
0.06963877379894257,
-0.058587364852428436,
0.1905425637960434,
0.03258530795574188,
-0.07860512286424637,
-0.059183377772569656,
0.05133861303329468,
-0.0791650041937828,
0.302468478679657,
0.00626079086214304,
0.03169599175453186,
0.10508318990468979,
-0.028644336387515068,
-0.16361252963542938,
0.02362491562962532,
0.1140698790550232,
-0.08390003442764282,
0.08627496659755707,
0.19878266751766205,
-0.019539451226592064,
0.11435621976852417,
0.05704843997955322,
-0.06186779960989952,
0.0524447038769722,
-0.03936922550201416,
-0.052163589745759964,
-0.09776037186384201,
0.06190723925828934,
-0.06178663671016693,
0.15432539582252502,
0.09593082964420319,
-0.05059736222028732,
-0.006600235588848591,
-0.05587591603398323,
0.04507772624492645,
0.018967149779200554,
0.12800532579421997,
0.012484090402722359,
-0.17696550488471985,
0.032744914293289185,
0.0010579711524769664,
0.11208613961935043,
-0.24666742980480194,
-0.08353681117296219,
0.09015431255102158,
-0.019416818395256996,
-0.05258401483297348,
0.09870866686105728,
0.0722413882613182,
0.04240792989730835,
-0.04463369399309158,
-0.10492048412561417,
-0.019366342574357986,
0.1493324637413025,
-0.14043603837490082,
-0.014699541963636875
] |
null | null | peft |
见[nenekochan/Yi-6B-yoruno](https://huggingface.co/nenekochan/Yi-6B-yoruno)
## Training procedure
The following `bitsandbytes` quantization config was used during training:
- quant_method: QuantizationMethod.BITS_AND_BYTES
- load_in_8bit: False
- load_in_4bit: True
- llm_int8_threshold: 6.0
- llm_int8_skip_modules: None
- llm_int8_enable_fp32_cpu_offload: False
- llm_int8_has_fp16_weight: False
- bnb_4bit_quant_type: nf4
- bnb_4bit_use_double_quant: True
- bnb_4bit_compute_dtype: float16
### Framework versions
- PEFT 0.4.0
| {"library_name": "peft"} | null | nenekochan/Yi-6B-yoruno-peft | [
"peft",
"safetensors",
"region:us"
] | 2024-02-08T07:58:00+00:00 | [] | [] | TAGS
#peft #safetensors #region-us
|
见nenekochan/Yi-6B-yoruno
## Training procedure
The following 'bitsandbytes' quantization config was used during training:
- quant_method: QuantizationMethod.BITS_AND_BYTES
- load_in_8bit: False
- load_in_4bit: True
- llm_int8_threshold: 6.0
- llm_int8_skip_modules: None
- llm_int8_enable_fp32_cpu_offload: False
- llm_int8_has_fp16_weight: False
- bnb_4bit_quant_type: nf4
- bnb_4bit_use_double_quant: True
- bnb_4bit_compute_dtype: float16
### Framework versions
- PEFT 0.4.0
| [
"## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- quant_method: QuantizationMethod.BITS_AND_BYTES\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: True\n- bnb_4bit_compute_dtype: float16",
"### Framework versions\n\n\n- PEFT 0.4.0"
] | [
"TAGS\n#peft #safetensors #region-us \n",
"## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- quant_method: QuantizationMethod.BITS_AND_BYTES\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: True\n- bnb_4bit_compute_dtype: float16",
"### Framework versions\n\n\n- PEFT 0.4.0"
] | [
14,
171,
11
] | [
"passage: TAGS\n#peft #safetensors #region-us \n## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- quant_method: QuantizationMethod.BITS_AND_BYTES\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: True\n- bnb_4bit_compute_dtype: float16### Framework versions\n\n\n- PEFT 0.4.0"
] | [
-0.07106870412826538,
0.10301944613456726,
-0.0031449291855096817,
0.11761511862277985,
0.07894829660654068,
0.029084622859954834,
0.09514864534139633,
0.1070803627371788,
0.011923941783607006,
0.10402517765760422,
0.11686863750219345,
0.0311274416744709,
0.044409990310668945,
0.1680542379617691,
-0.029187044128775597,
0.021361641585826874,
0.05824750289320946,
0.001016590977087617,
0.02547486498951912,
0.08671349287033081,
0.050070613622665405,
-0.0623609684407711,
0.020427333191037178,
-0.08069638162851334,
-0.11878520995378494,
-0.0019016225123777986,
0.008239333517849445,
0.013481639325618744,
0.041257575154304504,
0.011339358054101467,
0.07282711565494537,
0.01679760403931141,
-0.028027959167957306,
-0.20891529321670532,
-0.005915659014135599,
0.11312521249055862,
-0.01912444457411766,
0.06477703154087067,
-0.07619012892246246,
0.10523948818445206,
-0.11571042984724045,
-0.07131016999483109,
0.008807492442429066,
0.017765101045370102,
-0.08407202363014221,
-0.1300964206457138,
-0.058977048844099045,
0.04614601284265518,
0.031672798097133636,
0.07060392200946808,
-0.033049438148736954,
0.2171044647693634,
-0.12011413276195526,
0.09194282442331314,
0.08645506203174591,
-0.27316814661026,
-0.03454873710870743,
0.08750693500041962,
-0.003796280361711979,
0.16939686238765717,
-0.09119383990764618,
-0.09190423041582108,
0.0759226605296135,
0.03673073649406433,
0.00898042879998684,
-0.011962521821260452,
-0.11792829632759094,
-0.0010948270792141557,
-0.16259686648845673,
-0.030024703592061996,
0.1348256766796112,
0.03521540015935898,
-0.05807793140411377,
-0.048125870525836945,
-0.11560747027397156,
-0.3561999499797821,
0.02992681972682476,
-0.03774404525756836,
-0.07173077762126923,
0.03653568774461746,
0.0049706329591572285,
-0.004394709598273039,
-0.02871449664235115,
-0.058682337403297424,
-0.0314779132604599,
0.08967381715774536,
0.05619971081614494,
0.02001192420721054,
0.012135723605751991,
0.10044709593057632,
-0.1336159110069275,
-0.03792774677276611,
-0.053006626665592194,
-0.03514793887734413,
-0.03165796399116516,
-0.020258767530322075,
-0.046172600239515305,
0.15590542554855347,
0.09846581518650055,
0.12038073688745499,
-0.21120816469192505,
0.11559801548719406,
-0.036426953971385956,
0.0337064266204834,
-0.029200581833720207,
0.04567511007189751,
-0.10886981338262558,
0.11020497232675552,
0.039240363985300064,
0.1650589406490326,
0.046761903911828995,
-0.018622493371367455,
-0.037541840225458145,
0.014079985208809376,
0.17084471881389618,
0.024685656651854515,
-0.11467830836772919,
0.025600148364901543,
-0.12473675608634949,
0.009395832195878029,
0.07549718767404556,
-0.08082740753889084,
0.0229786429554224,
0.05144358053803444,
-0.01792220026254654,
-0.046509236097335815,
0.09714775532484055,
-0.05106880143284798,
-0.025610487908124924,
-0.045668669044971466,
-0.10080651938915253,
-0.03890708461403847,
-0.08392015099525452,
-0.1441115438938141,
0.05483119189739227,
-0.15847191214561462,
-0.0017803781665861607,
-0.07478249818086624,
-0.018328623846173286,
0.05499095842242241,
-0.02098158746957779,
-0.06439688056707382,
0.08810152858495712,
-0.07433485239744186,
-0.1660166084766388,
-0.02343829721212387,
0.015723656862974167,
-0.04078752547502518,
-0.02594427578151226,
0.08887698501348495,
0.05904361233115196,
0.09966548532247543,
-0.1459677666425705,
-0.02160639688372612,
-0.04162564501166344,
0.08309686183929443,
0.05560307577252388,
0.08624803274869919,
-0.09706581383943558,
-0.03256930410861969,
-0.05032377690076828,
-0.050308264791965485,
-0.10368280857801437,
-0.033318229019641876,
0.14297541975975037,
0.10459276288747787,
-0.14178889989852905,
0.013136128894984722,
0.11144199967384338,
-0.041209638118743896,
-0.09017908573150635,
0.16741515696048737,
-0.03990575298666954,
0.08549914509057999,
-0.026763655245304108,
0.08241212368011475,
0.23713569343090057,
-0.13048624992370605,
-0.024632461369037628,
0.09538858383893967,
0.07374116778373718,
0.0312599278986454,
0.004286326467990875,
0.07596864551305771,
-0.11167963594198227,
0.02401646412909031,
0.07846604287624359,
0.024533314630389214,
-0.07241396605968475,
-0.055625129491090775,
-0.03661003336310387,
-0.06825026124715805,
0.11852622032165527,
0.016749326139688492,
-0.008109763264656067,
-0.09399045258760452,
-0.08775175362825394,
0.10363989323377609,
0.10236191749572754,
-0.02668185532093048,
-0.013363858684897423,
-0.1449737548828125,
0.022408735007047653,
-0.06566299498081207,
0.012224221602082253,
-0.103211410343647,
-0.005752402823418379,
0.10549987107515335,
-0.06817784160375595,
-0.007317970972508192,
-0.000422046723542735,
0.05371445044875145,
0.07335196435451508,
-0.049768149852752686,
-0.005582983139902353,
-0.03367186337709427,
0.03291329741477966,
-0.09083148837089539,
-0.07625112682580948,
0.03897610679268837,
-0.025674445554614067,
0.2338678240776062,
-0.1358279436826706,
0.044136736541986465,
0.12842217087745667,
-0.009811420924961567,
0.007620756980031729,
-0.0376976877450943,
-0.04811282083392143,
0.09878850728273392,
-0.017867760732769966,
-0.03553615137934685,
0.029412828385829926,
0.03942086175084114,
-0.036487482488155365,
-0.1509622484445572,
-0.12332437187433243,
0.048259127885103226,
0.12359749525785446,
0.12550632655620575,
-0.06842632591724396,
-0.024142909795045853,
-0.03116304613649845,
-0.023059893399477005,
0.032976940274238586,
-0.04660741984844208,
-0.00004349072332843207,
0.000685713195707649,
0.06437879055738449,
-0.11484771966934204,
-0.030586637556552887,
0.07636719942092896,
-0.04256966710090637,
-0.04653681442141533,
0.11897410452365875,
0.002328504342585802,
-0.151204451918602,
0.08192235976457596,
0.10253848135471344,
-0.12519817054271698,
0.08712726086378098,
-0.0024461611174046993,
-0.011877349577844143,
-0.10322790592908859,
0.19338726997375488,
0.03303766995668411,
0.14151236414909363,
-0.10498815774917603,
0.10689322650432587,
-0.023417053744196892,
0.009041592478752136,
0.04436955228447914,
-0.19756552577018738,
0.0022661241237074137,
-0.03828301653265953,
-0.07202835381031036,
-0.055971041321754456,
-0.007762339431792498,
0.005850424524396658,
0.0435878224670887,
-0.013235111720860004,
0.0691029280424118,
0.11663621664047241,
-0.014020192436873913,
-0.08760034292936325,
0.1998598277568817,
-0.23286540806293488,
-0.2555730938911438,
-0.18461279571056366,
0.03491222485899925,
-0.09567965567111969,
-0.03246871009469032,
-0.027642836794257164,
-0.0684155523777008,
0.03419042378664017,
-0.12216172367334366,
-0.05967012792825699,
0.03104294463992119,
0.020974934101104736,
0.027350256219506264,
0.012072287499904633,
0.16685807704925537,
-0.07622408866882324,
0.023235663771629333,
0.046058159321546555,
-0.0511644147336483,
0.1259877234697342,
-0.04961244761943817,
0.003985917661339045,
0.1118801161646843,
-0.029584357514977455,
0.00048195343697443604,
0.010580276139080524,
0.33397340774536133,
0.009621374309062958,
0.049453701823949814,
0.071914441883564,
0.013423635624349117,
0.051338065415620804,
0.11203967779874802,
0.00871092826128006,
-0.09545348584651947,
0.07801233232021332,
0.04285778850317001,
-0.07203903049230576,
-0.11581925302743912,
-0.04376877844333649,
-0.05157924070954323,
0.04532875493168831,
0.06866645067930222,
0.06282487511634827,
0.06192708760499954,
0.06651562452316284,
0.021540408954024315,
0.04673471301794052,
0.021837694570422173,
0.005031702108681202,
0.11196956038475037,
-0.01951059326529503,
0.05300875008106232,
-0.02185061387717724,
0.010642015375196934,
0.047685105353593826,
0.14003460109233856,
0.0307359267026186,
-0.07994404435157776,
0.00015640341734979302,
0.06053526699542999,
0.31345176696777344,
0.004323047120124102,
0.09050001949071884,
-0.05614371597766876,
-0.023853985592722893,
-0.01047220453619957,
-0.04666100814938545,
-0.051958873867988586,
0.025141693651676178,
-0.05900508537888527,
0.09278569370508194,
-0.002405908890068531,
0.009244444780051708,
0.09831178188323975,
0.08668886870145798,
0.20596913993358612,
-0.288271963596344,
-0.11751549690961838,
-0.007562580052763224,
0.1195870190858841,
-0.09617788344621658,
0.023789038881659508,
0.23636633157730103,
0.04941682890057564,
-0.08938766270875931,
-0.03514132648706436,
0.0012481332523748279,
0.007844117470085621,
0.014998700469732285,
0.09669895470142365,
0.12863437831401825,
-0.016489433124661446,
0.07277210056781769,
-0.2873733639717102,
0.04284977540373802,
0.0683988705277443,
0.05091038718819618,
-0.00448841555044055,
-0.0043410006910562515,
-0.07807594537734985,
-0.034628864377737045,
0.057895727455616,
0.0013592237373813987,
0.1758994311094284,
-0.2727965712547302,
-0.0908685028553009,
0.010349121876060963,
0.11216004192829132,
0.08983417600393295,
0.06355089694261551,
0.02161535993218422,
0.032137174159288406,
0.06505721062421799,
0.06528796255588531,
-0.06034553050994873,
-0.09242486953735352,
-0.008640426211059093,
0.17407171428203583,
-0.12493615597486496,
-0.07730723917484283,
-0.038384292274713516,
-0.030065219849348068,
0.020547423511743546,
-0.17923149466514587,
-0.05387422814965248,
-0.04440838471055031,
0.0033069506753236055,
0.1274687647819519,
-0.03926428034901619,
0.010757848620414734,
-0.03821999207139015,
0.05851481109857559,
-0.043973349034786224,
-0.06522496044635773,
0.10495975613594055,
-0.05692810192704201,
-0.15493375062942505,
-0.06129281595349312,
0.15074363350868225,
0.07016142457723618,
-0.018208954483270645,
-0.06866665929555893,
-0.04332505166530609,
0.031072242185473442,
-0.12977232038974762,
0.004341110587120056,
0.08092086017131805,
-0.0801481381058693,
0.05868237465620041,
-0.12158645689487457,
0.15184436738491058,
-0.06923013925552368,
0.09224583953619003,
0.058139778673648834,
0.3274558484554291,
-0.09572373330593109,
0.008758529089391232,
0.11772012710571289,
-0.010017837397754192,
-0.2571812868118286,
0.033951401710510254,
0.039133813232183456,
0.029824374243617058,
-0.021642884239554405,
-0.1129603162407875,
0.04136572405695915,
0.075788713991642,
-0.0018813018687069416,
0.15325812995433807,
-0.3202837109565735,
-0.057492610067129135,
0.04593157023191452,
0.07848251610994339,
0.08965754508972168,
-0.06845510005950928,
0.022022541612386703,
0.029466820880770683,
-0.008047026582062244,
0.16064143180847168,
-0.16008618474006653,
0.08297662436962128,
0.010013296268880367,
0.019916784018278122,
0.009331114590168,
-0.04953455179929733,
0.1319161206483841,
-0.037793826311826706,
0.07903478294610977,
0.026902969926595688,
-0.03797008469700813,
0.061329860240221024,
-0.07415678352117538,
0.032377906143665314,
-0.049353789538145065,
0.10570456832647324,
0.02841108851134777,
0.013913610950112343,
-0.06798121333122253,
-0.0037756613455712795,
-0.07431980222463608,
-0.0709119364619255,
-0.11931586265563965,
0.09336390346288681,
0.0029058647342026234,
-0.03944862633943558,
-0.05761529877781868,
0.05854353681206703,
0.06571576744318008,
0.4496440887451172,
-0.05412083864212036,
-0.060058288276195526,
0.10084106028079987,
0.10671714693307877,
-0.03587163984775543,
0.0812513679265976,
-0.09865907579660416,
0.051652710884809494,
0.11590348929166794,
0.0012981008039787412,
0.11800885945558548,
0.07096850872039795,
-0.12184671312570572,
-0.004885968752205372,
0.04133068397641182,
-0.1656176745891571,
-0.087092824280262,
-0.013287798501551151,
0.0424092672765255,
-0.09313523024320602,
0.043646689504384995,
0.10728555917739868,
-0.04979395121335983,
0.038099758327007294,
0.02066994458436966,
0.05182371288537979,
-0.09947187453508377,
0.14607267081737518,
0.021439524367451668,
0.08508527278900146,
-0.06433648616075516,
0.09846194833517075,
0.03767048940062523,
-0.007839014753699303,
0.035949211567640305,
-0.016642555594444275,
-0.11980006843805313,
-0.004057018086314201,
-0.02602505497634411,
-0.11379426717758179,
0.12506268918514252,
-0.04463735595345497,
-0.03856837749481201,
-0.10204677283763885,
-0.012970251962542534,
0.13385118544101715,
0.041775014251470566,
0.09880035370588303,
0.0012777693336829543,
0.018174344673752785,
-0.14622025191783905,
0.11359555274248123,
-0.040791161358356476,
0.022848688066005707,
-0.14989185333251953,
0.05453961342573166,
-0.019831465557217598,
0.044390078634023666,
-0.024905001744627953,
-0.003573039313778281,
-0.20598548650741577,
0.018713273108005524,
-0.02660278044641018,
0.021833447739481926,
0.037389468401670456,
0.028355339542031288,
0.01318410411477089,
0.04337342455983162,
-0.037080809473991394,
0.03858664631843567,
-0.012804670259356499,
-0.03301037847995758,
0.04498952254652977,
-0.00980586837977171,
-0.04898493364453316,
-0.062090374529361725,
0.0502467043697834,
-0.10824364423751831,
0.043004218488931656,
-0.01107618398964405,
-0.06465465575456619,
0.07479476183652878,
0.0018642247887328267,
0.033559150993824005,
0.10201934725046158,
0.042132195085287094,
0.03664269298315048,
-0.09352704137563705,
0.030679976567626,
0.00010357295104768127,
-0.03533155843615532,
0.03923627361655235,
0.13568224012851715,
-0.03898732364177704,
-0.06906324625015259,
-0.1280597299337387,
0.017529631033539772,
-0.04167097806930542,
0.015785278752446175,
0.13825371861457825,
0.056835927069187164,
0.08967248350381851,
-0.11021006107330322,
-0.025862758979201317,
-0.13358385860919952,
-0.0820988193154335,
0.04309192672371864,
-0.06415961682796478,
-0.09186573326587677,
-0.005190418567508459,
0.0668218582868576,
-0.0070160566829144955,
0.13278470933437347,
-0.10315757989883423,
-0.07774126529693604,
-0.04114126041531563,
-0.213951975107193,
-0.12442468106746674,
0.017301371321082115,
0.2559864819049835,
0.02367987670004368,
-0.06601918488740921,
-0.1014888733625412,
0.0014963280409574509,
0.06823115795850754,
0.11706733703613281,
0.027446214109659195,
0.10479676723480225,
-0.1161564290523529,
0.09809886664152145,
0.04459574073553085,
-0.0425020307302475,
0.11045181006193161,
0.2856730818748474,
-0.08184609562158585,
0.013946060091257095,
-0.08711475878953934,
0.12311910837888718,
0.04050900414586067,
-0.10936293005943298,
-0.009998939000070095,
-0.04684196785092354,
-0.14291627705097198,
-0.11003296822309494,
0.03281180560588837,
-0.08770381659269333,
-0.1563962996006012,
-0.03560052067041397,
-0.11916966736316681,
-0.0867830216884613,
0.04229974001646042,
0.04879738762974739,
-0.04313160851597786,
0.23358528316020966,
-0.04910087212920189,
0.03100656345486641,
-0.010186132974922657,
0.009149479679763317,
-0.01078015100210905,
0.011218048632144928,
-0.09162609279155731,
0.13489553332328796,
0.01901191845536232,
0.09668480604887009,
0.011699262075126171,
0.08108872175216675,
0.05472370237112045,
-0.031209953129291534,
-0.05093284696340561,
-0.009045024402439594,
0.014580848626792431,
-0.0622774139046669,
0.08270999044179916,
0.05512401461601257,
-0.0883616954088211,
-0.06525047868490219,
-0.0006380858249031007,
-0.061964549124240875,
-0.04443986341357231,
-0.14025896787643433,
0.29310035705566406,
-0.029434934258461,
0.12343154847621918,
0.0016597460489720106,
-0.07727916538715363,
-0.07096627354621887,
0.14703793823719025,
0.11947073042392731,
-0.10750815272331238,
0.0018296999623999,
0.051035262644290924,
-0.00565975159406662,
-0.10516169667243958,
0.129768505692482,
0.09303049743175507,
0.022132650017738342,
0.022644035518169403,
-0.011083212681114674,
-0.03499656915664673,
0.011222516186535358,
-0.021507829427719116,
-0.04165023937821388,
0.013380970805883408,
0.06741439551115036,
-0.15023232996463776,
-0.029560500755906105,
-0.05163979530334473,
-0.11016134172677994,
0.14259237051010132,
-0.1369706690311432,
-0.07398684322834015,
-0.024451198056340218,
-0.06473227590322495,
-0.11218223720788956,
0.020734895020723343,
-0.10676635801792145,
0.056231025606393814,
0.0453394316136837,
-0.05034762620925903,
0.036552634090185165,
-0.022246576845645905,
-0.01763385348021984,
0.01585932821035385,
0.13084733486175537,
-0.005882604513317347,
0.04575900360941887,
0.1376887708902359,
-0.041927266865968704,
-0.07532758265733719,
0.14404064416885376,
0.0364975742995739,
-0.05503920093178749,
-0.13279147446155548,
0.01697050966322422,
-0.017365043982863426,
0.14101430773735046,
0.0431353896856308,
-0.06413869559764862,
-0.007196525111794472,
-0.22700166702270508,
-0.011869627982378006,
-0.1462576538324356,
-0.060399845242500305,
-0.05463402345776558,
0.10757226496934891,
0.14706166088581085,
-0.06568368524312973,
0.0327160619199276,
-0.026373034343123436,
0.05088541656732559,
-0.0400797575712204,
0.06967458128929138,
0.05282207578420639,
-0.16074872016906738,
0.08458733558654785,
-0.050889529287815094,
0.0163042601197958,
-0.3154277205467224,
0.005113633349537849,
0.01782517321407795,
-0.01679665967822075,
-0.05641482397913933,
0.1412428319454193,
0.04208378866314888,
0.0680277869105339,
-0.07719820737838745,
-0.26971518993377686,
-0.07400535047054291,
0.14451761543750763,
0.02062358520925045,
-0.08044587075710297
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | YashRawal225/New-3-7b-chat-finetune-german500 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T07:58:06+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #mistral #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
56,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.05921921506524086,
0.15253323316574097,
-0.004925556480884552,
0.01970141939818859,
0.09812989830970764,
0.008722675032913685,
0.07155127823352814,
0.11091651022434235,
-0.02038503810763359,
0.11541511863470078,
0.03161177039146423,
0.09504877775907516,
0.11244720220565796,
0.1593349277973175,
0.0006018498679623008,
-0.22924894094467163,
0.050943523645401,
-0.12565383315086365,
-0.028005311265587807,
0.1202453151345253,
0.14323006570339203,
-0.10873830318450928,
0.07482945919036865,
-0.03924073651432991,
-0.006830108352005482,
-0.03327549248933792,
-0.06254202127456665,
-0.05196645110845566,
0.05287102237343788,
0.06693000346422195,
0.07382122427225113,
0.0121690658852458,
0.09054198116064072,
-0.27071383595466614,
0.02402324043214321,
0.07869837433099747,
-0.00047617589007131755,
0.07642106711864471,
0.049837369471788406,
-0.08698169887065887,
0.07614438980817795,
-0.060363397002220154,
0.14962489902973175,
0.07956483215093613,
-0.09049813449382782,
-0.19196605682373047,
-0.07841940224170685,
0.10002946108579636,
0.18888257443904877,
0.05783533677458763,
-0.02747977338731289,
0.11718999594449997,
-0.08618196099996567,
0.013946855440735817,
0.06651762872934341,
-0.05830651894211769,
-0.055825375020504,
0.07012750208377838,
0.08251979202032089,
0.08537944406270981,
-0.13050076365470886,
-0.011774240992963314,
0.015172234736382961,
0.00940374843776226,
0.0883294939994812,
0.017624128609895706,
0.13745273649692535,
0.04126768559217453,
-0.1351923644542694,
-0.04287068545818329,
0.09870852530002594,
0.035997726023197174,
-0.04835180938243866,
-0.24833782017230988,
-0.023138362914323807,
-0.039952121675014496,
-0.03223174810409546,
-0.0381147637963295,
0.04236193001270294,
-0.01381280180066824,
0.07635250687599182,
-0.0030598659068346024,
-0.08292017132043839,
-0.042900193482637405,
0.07140932232141495,
0.06195797771215439,
0.025352943688631058,
-0.016651969403028488,
0.0064301020465791225,
0.12258180975914001,
0.11147689074277878,
-0.12772345542907715,
-0.053019966930150986,
-0.06414514780044556,
-0.08524893969297409,
-0.04640465974807739,
0.03045455552637577,
0.03743596002459526,
0.047410931438207626,
0.2386423945426941,
0.0032438088674098253,
0.054757438600063324,
0.046099163591861725,
0.014072372578084469,
0.06632840633392334,
0.10764557868242264,
-0.05884917825460434,
-0.09735266119241714,
-0.030795203521847725,
0.10186740756034851,
0.006704956758767366,
-0.041407015174627304,
-0.05594591051340103,
0.06964502483606339,
0.020676078274846077,
0.1224241703748703,
0.07868597656488419,
0.002938423305749893,
-0.07543925195932388,
-0.06281042098999023,
0.18152743577957153,
-0.1571107804775238,
0.0444292388856411,
0.03200872242450714,
-0.03442244604229927,
-0.009351148270070553,
0.00990392453968525,
0.02681080251932144,
-0.02011663094162941,
0.09737543761730194,
-0.05644093081355095,
-0.033681318163871765,
-0.11296935379505157,
-0.0371013842523098,
0.030811145901679993,
0.01213210541754961,
-0.029025491327047348,
-0.0342867337167263,
-0.0882277637720108,
-0.0636090338230133,
0.09107700735330582,
-0.07191670686006546,
-0.04744245857000351,
-0.017612621188163757,
-0.07794062048196793,
0.022423118352890015,
0.017721612006425858,
0.09050743281841278,
-0.021899394690990448,
0.03913994878530502,
-0.056751471012830734,
0.06101011112332344,
0.11571475863456726,
0.028108863160014153,
-0.058606795966625214,
0.06155762821435928,
-0.2421950101852417,
0.10317995399236679,
-0.07758963108062744,
0.051325954496860504,
-0.1530446857213974,
-0.026070065796375275,
0.03956404700875282,
0.012061306275427341,
-0.008345595560967922,
0.1417774260044098,
-0.2185831218957901,
-0.03138069063425064,
0.1676056981086731,
-0.10102425515651703,
-0.07971794903278351,
0.06269615143537521,
-0.05407082289457321,
0.11134804040193558,
0.04596652463078499,
-0.023191405460238457,
0.05842197686433792,
-0.14511504769325256,
-0.00791724119335413,
-0.04188765957951546,
-0.017894908785820007,
0.16635635495185852,
0.07102048397064209,
-0.06073606386780739,
0.07092984020709991,
0.019934939220547676,
-0.016795052215456963,
-0.04869792237877846,
-0.028511613607406616,
-0.10498060286045074,
0.011810078285634518,
-0.059134796261787415,
0.02167343720793724,
-0.021296551451086998,
-0.09382132440805435,
-0.029188871383666992,
-0.17379464209079742,
-0.0012200147612020373,
0.08734307438135147,
-0.010546354576945305,
-0.02201107330620289,
-0.11164727807044983,
0.008580547757446766,
0.03398929536342621,
0.0007392297266051173,
-0.13708379864692688,
-0.059298936277627945,
0.02737307921051979,
-0.16233380138874054,
0.02912268228828907,
-0.05535917729139328,
0.046022266149520874,
0.040077272802591324,
-0.03548351675271988,
-0.0344831608235836,
0.01168955210596323,
0.011000183410942554,
-0.01812567003071308,
-0.25495970249176025,
-0.017501724883913994,
-0.02502158097922802,
0.17353887856006622,
-0.22721131145954132,
0.04271984100341797,
0.07614967226982117,
0.14550280570983887,
0.0073052942752838135,
-0.034482456743717194,
0.014565827324986458,
-0.07198352366685867,
-0.03167816624045372,
-0.06257235258817673,
-0.010083765722811222,
-0.03872835263609886,
-0.06014038994908333,
0.04782424867153168,
-0.16939696669578552,
-0.03236479312181473,
0.10534932464361191,
0.06398996710777283,
-0.14835967123508453,
-0.030286256223917007,
-0.0393594354391098,
-0.047035153955221176,
-0.06618485599756241,
-0.054856978356838226,
0.12015452980995178,
0.05620792135596275,
0.04745647683739662,
-0.07151947915554047,
-0.07490099221467972,
0.007241961546242237,
-0.019977761432528496,
-0.0163256898522377,
0.09354335069656372,
0.06967450678348541,
-0.12794628739356995,
0.09154868870973587,
0.0982460081577301,
0.08392132818698883,
0.10398648679256439,
-0.015390566550195217,
-0.08757331967353821,
-0.041474130004644394,
0.023933125659823418,
0.014664852991700172,
0.1483616679906845,
-0.016296299174427986,
0.054420776665210724,
0.0360836423933506,
-0.013510678894817829,
0.01076538860797882,
-0.09628108888864517,
0.02706051431596279,
0.02971329540014267,
-0.015405743382871151,
0.03466423228383064,
-0.04367179423570633,
0.019455796107649803,
0.09001301974058151,
0.041830018162727356,
0.0396038182079792,
0.010561688803136349,
-0.04398298263549805,
-0.11032342165708542,
0.17876994609832764,
-0.12373854219913483,
-0.2460412234067917,
-0.13813963532447815,
0.010937176644802094,
0.04738753288984299,
-0.011057097464799881,
0.006951550021767616,
-0.06640941649675369,
-0.1170244961977005,
-0.09733203053474426,
0.01991088129580021,
0.04529648274183273,
-0.07728998363018036,
-0.06572148203849792,
0.06318122148513794,
0.037644270807504654,
-0.13899093866348267,
0.023945696651935577,
0.0469096377491951,
-0.0813174769282341,
-0.0011905812425538898,
0.07709334045648575,
0.06798645853996277,
0.17623907327651978,
0.014159789308905602,
-0.023712651804089546,
0.025652561336755753,
0.21002908051013947,
-0.14298869669437408,
0.1094568595290184,
0.1327279806137085,
-0.08898334950208664,
0.08212688565254211,
0.20222385227680206,
0.0385010726749897,
-0.10506977140903473,
0.03657889738678932,
0.027060477063059807,
-0.02792542427778244,
-0.24959829449653625,
-0.06908850371837616,
0.001758498721756041,
-0.053698375821113586,
0.06916391849517822,
0.08716317266225815,
0.09721273928880692,
0.016790922731161118,
-0.10066783428192139,
-0.0790279284119606,
0.05001477152109146,
0.10897587984800339,
-0.001458899350836873,
-0.014394176192581654,
0.09075857698917389,
-0.02953648567199707,
0.01689162664115429,
0.09213569760322571,
0.0019032615236938,
0.1793205291032791,
0.052213337272405624,
0.17340974509716034,
0.07910763472318649,
0.06269825994968414,
0.021207094192504883,
0.006816241890192032,
0.02095629647374153,
0.01695442944765091,
-0.004212336614727974,
-0.0863528773188591,
-0.0027415938675403595,
0.1203664243221283,
0.050876569002866745,
0.03059028834104538,
0.014285655692219734,
-0.03054206818342209,
0.08466528356075287,
0.177787184715271,
0.001063879462890327,
-0.1876421719789505,
-0.07282958924770355,
0.07934894412755966,
-0.08512143790721893,
-0.10675539821386337,
-0.029639042913913727,
0.040873926132917404,
-0.17292065918445587,
0.01861744187772274,
-0.020119842141866684,
0.10806277394294739,
-0.12885749340057373,
-0.017452897503972054,
0.055447377264499664,
0.06997017562389374,
-0.009931124746799469,
0.06633757054805756,
-0.1625119000673294,
0.1177479475736618,
0.01653103344142437,
0.06594116985797882,
-0.09538834542036057,
0.095417320728302,
-0.006962447427213192,
0.007516060955822468,
0.1403670459985733,
0.010755252093076706,
-0.0641925036907196,
-0.0961010679602623,
-0.10299893468618393,
-0.010606445372104645,
0.1309773176908493,
-0.14660196006298065,
0.08697716891765594,
-0.02743646875023842,
-0.0437387153506279,
0.0037594304885715246,
-0.12246467173099518,
-0.13224415481090546,
-0.18235477805137634,
0.05769521743059158,
-0.13171130418777466,
0.040173836052417755,
-0.1089821308851242,
-0.04585907980799675,
-0.021465247496962547,
0.1977471560239792,
-0.23280778527259827,
-0.06815840303897858,
-0.15394872426986694,
-0.08265888690948486,
0.1454220414161682,
-0.04706942290067673,
0.08337214589118958,
0.000301246385788545,
0.19080647826194763,
0.020952312275767326,
-0.017133628949522972,
0.1067209243774414,
-0.09975022822618484,
-0.20161914825439453,
-0.09120959788560867,
0.15868841111660004,
0.13963958621025085,
0.038726504892110825,
-0.004869744647294283,
0.032236017286777496,
-0.021885421127080917,
-0.12115032970905304,
0.02010788396000862,
0.17255425453186035,
0.08749033510684967,
0.026468761265277863,
-0.028463367372751236,
-0.11846643686294556,
-0.07225121557712555,
-0.03745346516370773,
0.02470988966524601,
0.1813775599002838,
-0.07139390707015991,
0.18551595509052277,
0.14274363219738007,
-0.054879751056432724,
-0.19840270280838013,
0.02148755080997944,
0.04472679644823074,
0.0060237692669034,
0.03174281120300293,
-0.20237314701080322,
0.09144619107246399,
0.0006281035020947456,
-0.05034751072525978,
0.13383205235004425,
-0.18327344954013824,
-0.15106844902038574,
0.061150215566158295,
0.04303572699427605,
-0.19199669361114502,
-0.1237611323595047,
-0.08872545510530472,
-0.046805474907159805,
-0.1568751484155655,
0.1029038056731224,
0.0011325168889015913,
0.007591354660689831,
0.03782656043767929,
0.024313677102327347,
0.012553532607853413,
-0.041947584599256516,
0.19289998710155487,
-0.02507353574037552,
0.034427378326654434,
-0.0793621614575386,
-0.06381990760564804,
0.06411149352788925,
-0.057697590440511703,
0.0750909373164177,
-0.025500034913420677,
0.015388053841888905,
-0.10115842521190643,
-0.047956179827451706,
-0.029484452679753304,
0.01986371912062168,
-0.09421123564243317,
-0.09366033226251602,
-0.04838487133383751,
0.0944879949092865,
0.08926530182361603,
-0.037268105894327164,
-0.033034052699804306,
-0.07874293625354767,
0.04173892363905907,
0.17448031902313232,
0.18235735595226288,
0.045147113502025604,
-0.07717937231063843,
-0.0013610349269583821,
-0.014655699953436852,
0.04845907539129257,
-0.22060799598693848,
0.06062275543808937,
0.045259539037942886,
0.01552091259509325,
0.11744016408920288,
-0.020618194714188576,
-0.1619492471218109,
-0.0666290745139122,
0.06087447330355644,
-0.06730270385742188,
-0.1811886727809906,
0.00352504407055676,
0.0753183513879776,
-0.16591353714466095,
-0.03711319714784622,
0.04232833534479141,
-0.011535273864865303,
-0.04050648957490921,
0.013207654468715191,
0.08094717562198639,
0.0073035703971982,
0.07697968184947968,
0.05389590561389923,
0.09186159074306488,
-0.10275198519229889,
0.07336891442537308,
0.08092255145311356,
-0.08580191433429718,
0.029650582000613213,
0.0956844761967659,
-0.0660475566983223,
-0.03553546592593193,
0.039692267775535583,
0.08463539928197861,
0.025261107832193375,
-0.04666709899902344,
0.003693421371281147,
-0.09922701120376587,
0.05857077240943909,
0.11215036362409592,
0.035282451659440994,
0.011146705597639084,
0.03799959644675255,
0.04474346339702606,
-0.07786709815263748,
0.11944296956062317,
0.024733934551477432,
0.020655835047364235,
-0.04009570553898811,
-0.040743377059698105,
0.03469119220972061,
-0.027051862329244614,
-0.011984582990407944,
-0.035381630063056946,
-0.07329677045345306,
-0.014250458218157291,
-0.16089624166488647,
-0.006425157655030489,
-0.039050452411174774,
0.006492188666015863,
0.0227071400731802,
-0.03757927939295769,
0.008156952448189259,
0.012379756197333336,
-0.06891508400440216,
-0.05483170598745346,
-0.0225595161318779,
0.09499263763427734,
-0.16361327469348907,
0.02182857319712639,
0.08322018384933472,
-0.12078364938497543,
0.09284685552120209,
0.016550488770008087,
0.002410374814644456,
0.028476644307374954,
-0.15792103111743927,
0.04754367470741272,
-0.020290223881602287,
0.012727295979857445,
0.04053649678826332,
-0.2180718630552292,
-0.005482743959873915,
-0.04065772518515587,
-0.055209364742040634,
-0.008002875372767448,
-0.03194994851946831,
-0.11256447434425354,
0.09542836248874664,
0.010766619816422462,
-0.0858173593878746,
-0.029525602236390114,
0.032997291535139084,
0.07880192995071411,
-0.02688010409474373,
0.15163032710552216,
-0.004930328112095594,
0.07543973624706268,
-0.17439891397953033,
-0.02280678227543831,
-0.009784235619008541,
0.02145213820040226,
-0.02418927662074566,
-0.016610441729426384,
0.04521343484520912,
-0.027311841025948524,
0.18978725373744965,
-0.02763848751783371,
0.047156915068626404,
0.06419318169355392,
0.01327395811676979,
-0.016141459345817566,
0.11109550297260284,
0.05755641311407089,
0.024413742125034332,
0.02059282548725605,
0.0006552583072334528,
-0.04046328365802765,
-0.012729931622743607,
-0.18779614567756653,
0.06844497472047806,
0.14769941568374634,
0.09005311876535416,
-0.014767808839678764,
0.06981590390205383,
-0.09979446232318878,
-0.11724765598773956,
0.10648569464683533,
-0.06312347948551178,
-0.011802246794104576,
-0.06541955471038818,
0.14070585370063782,
0.1514706313610077,
-0.1892511397600174,
0.06684626638889313,
-0.06704412400722504,
-0.05669668689370155,
-0.11357752978801727,
-0.1923627108335495,
-0.05791294202208519,
-0.05011613294482231,
-0.018368201330304146,
-0.05373769626021385,
0.06899537891149521,
0.057158127427101135,
0.011277895420789719,
0.008883214555680752,
0.0839093029499054,
-0.009658100083470345,
0.001425864058546722,
0.031231271103024483,
0.06669623404741287,
0.016144385561347008,
-0.0304893609136343,
0.01806715875864029,
-0.003015234600752592,
0.033999331295490265,
0.059489116072654724,
0.036065202206373215,
-0.028380198404192924,
0.013694645836949348,
-0.03632815182209015,
-0.11369726806879044,
0.043240632861852646,
-0.028342511504888535,
-0.07773103564977646,
0.13286112248897552,
0.026473212987184525,
0.005609886720776558,
-0.022322779521346092,
0.2495104819536209,
-0.07400858402252197,
-0.09536818414926529,
-0.1448878049850464,
0.11703428626060486,
-0.04134928435087204,
0.06479805707931519,
0.03765689954161644,
-0.10748469084501266,
0.018750222399830818,
0.12525403499603271,
0.1550474315881729,
-0.04537956044077873,
0.019106155261397362,
0.02858782559633255,
0.004584235139191151,
-0.04013598710298538,
0.05142189934849739,
0.06933367252349854,
0.14214643836021423,
-0.05173535272479057,
0.08858583122491837,
0.0017827433766797185,
-0.10212727636098862,
-0.04129546508193016,
0.11294585466384888,
-0.012940747663378716,
0.016553698107600212,
-0.05866444855928421,
0.1253037303686142,
-0.059382375329732895,
-0.23649652302265167,
0.061238259077072144,
-0.07580125331878662,
-0.14206883311271667,
-0.02515989914536476,
0.0734870657324791,
-0.015550101175904274,
0.026368482038378716,
0.07198820263147354,
-0.07507873326539993,
0.18898127973079681,
0.03871531784534454,
-0.05198408663272858,
-0.05836968496441841,
0.07604995369911194,
-0.117560975253582,
0.2752254605293274,
0.01097069587558508,
0.05294901132583618,
0.10413134098052979,
-0.02049596607685089,
-0.13178466260433197,
0.024117950350046158,
0.09550730884075165,
-0.08813395351171494,
0.04131056368350983,
0.21484604477882385,
-0.005940921604633331,
0.1187596246600151,
0.07743308693170547,
-0.07539036870002747,
0.047102998942136765,
-0.1141449362039566,
-0.0771128386259079,
-0.08687382191419601,
0.09549140185117722,
-0.0675748735666275,
0.14216206967830658,
0.12683449685573578,
-0.054658904671669006,
0.010759806260466576,
-0.02898469939827919,
0.045599378645420074,
0.0063186027109622955,
0.10157246887683868,
0.009957551956176758,
-0.18577666580677032,
0.02454824559390545,
0.017152229323983192,
0.10993915796279907,
-0.1806284487247467,
-0.09123970568180084,
0.04470835253596306,
0.0021878182888031006,
-0.06369121372699738,
0.12484876811504364,
0.057084910571575165,
0.04630184918642044,
-0.044473882764577866,
-0.029204387217760086,
-0.0060947248712182045,
0.1420498490333557,
-0.10524781048297882,
-0.003831128589808941
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | RajuEEE/LlaMa_FineTunedModel | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-08T08:00:43+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
31,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06646376848220825,
0.2168014943599701,
-0.00225935154594481,
0.023818302899599075,
0.1271018385887146,
-0.001635765191167593,
0.04218708351254463,
0.13324736058712006,
-0.020175931975245476,
0.11144465953111649,
0.046588581055402756,
0.09377603232860565,
0.09928803145885468,
0.18404334783554077,
0.04859916493296623,
-0.2059975117444992,
0.007056170143187046,
-0.09090408682823181,
0.014076028019189835,
0.1116579994559288,
0.13719257712364197,
-0.10291384905576706,
0.08272874355316162,
-0.04045208916068077,
-0.02019004337489605,
0.00012576708104461432,
-0.09259183704853058,
-0.07032395154237747,
0.06885425746440887,
0.06264153122901917,
0.051234472543001175,
0.001456156256608665,
0.09140396863222122,
-0.2864592671394348,
0.017265573143959045,
0.08406311273574829,
0.0027674848679453135,
0.06290827691555023,
0.07236549258232117,
-0.07389893382787704,
0.11328595131635666,
-0.08021481335163116,
0.13019037246704102,
0.08625296503305435,
-0.062064990401268005,
-0.23071379959583282,
-0.07525765895843506,
0.0963398814201355,
0.12251301854848862,
0.06215599179267883,
-0.022921854630112648,
0.15455181896686554,
-0.06248689442873001,
0.012971068732440472,
0.1294165402650833,
-0.11526761949062347,
-0.05572471022605896,
0.061741601675748825,
0.11775490641593933,
0.10740239918231964,
-0.14110268652439117,
-0.0017287094378843904,
0.04900608956813812,
0.029121357947587967,
0.08589313924312592,
0.022661056369543076,
0.12003941088914871,
0.04652795568108559,
-0.13695219159126282,
-0.04037507623434067,
0.12011898308992386,
0.038862764835357666,
-0.06446044892072678,
-0.2168138176202774,
-0.006778308190405369,
-0.0601806715130806,
-0.014732478186488152,
-0.07019448280334473,
0.039128515869379044,
-0.02470310963690281,
0.07317749410867691,
-0.04465159401297569,
-0.1063927412033081,
-0.0421026237308979,
0.0892222449183464,
0.07748593389987946,
0.011527054943144321,
-0.02519804798066616,
0.04627908393740654,
0.13455867767333984,
0.05402068421244621,
-0.10399353504180908,
-0.07017925381660461,
-0.06942764669656754,
-0.09420394152402878,
-0.04035796597599983,
0.056760527193546295,
0.031942449510097504,
0.02665667235851288,
0.22703726589679718,
0.016653569415211678,
0.04155244305729866,
0.0224777739495039,
0.01032855175435543,
0.043662428855895996,
0.0955500528216362,
-0.05303520709276199,
-0.15660029649734497,
-0.04072032496333122,
0.09077946096658707,
-0.0027527001220732927,
-0.036689214408397675,
-0.03966725245118141,
0.03849169611930847,
0.06843466311693192,
0.13122352957725525,
0.07552056759595871,
-0.017929591238498688,
-0.04813180863857269,
-0.030096933245658875,
0.23523783683776855,
-0.1493375599384308,
0.04426715523004532,
-0.02271856553852558,
-0.01804111897945404,
-0.03908449783921242,
0.03597262129187584,
0.022118929773569107,
-0.000004518366949923802,
0.09706240892410278,
-0.058981191366910934,
-0.05378659814596176,
-0.10168042778968811,
-0.03272576630115509,
0.04088849574327469,
-0.013975566253066063,
-0.010589460842311382,
-0.09025166928768158,
-0.09490354359149933,
-0.04766594246029854,
0.05537205561995506,
-0.05123869329690933,
-0.03770573064684868,
0.009465423412621021,
-0.08151785284280777,
-0.005444355774670839,
-0.005417742300778627,
0.10699385404586792,
-0.03222226724028587,
0.04445803165435791,
-0.027600755915045738,
0.05225523188710213,
0.09919606149196625,
0.031576547771692276,
-0.0773419588804245,
0.0561848059296608,
-0.22559374570846558,
0.07503069192171097,
-0.11481974273920059,
0.04335082694888115,
-0.1704932004213333,
-0.042439818382263184,
0.005444696638733149,
0.0139949731528759,
0.013206101022660732,
0.12720820307731628,
-0.19255615770816803,
-0.01654396951198578,
0.13260798156261444,
-0.09212633967399597,
-0.118110790848732,
0.07884611934423447,
-0.029701577499508858,
0.1624738723039627,
0.04682036489248276,
-0.027025915682315826,
0.09224298596382141,
-0.16434773802757263,
-0.07092688232660294,
-0.00949116237461567,
-0.01727987825870514,
0.12109188735485077,
0.07512219995260239,
-0.05991523340344429,
0.046571120619773865,
0.02832140028476715,
-0.038078423589468,
-0.04424772411584854,
-0.050857074558734894,
-0.10884185880422592,
-0.01070026308298111,
-0.08987759798765182,
0.04065500199794769,
-0.01250192429870367,
-0.07916021347045898,
-0.029885273426771164,
-0.18612512946128845,
-0.0030564051121473312,
0.10038342326879501,
0.0035033065360039473,
-0.005652366206049919,
-0.08666291832923889,
0.026358824223279953,
-0.03112892620265484,
-0.008404186926782131,
-0.16764774918556213,
-0.04399421438574791,
0.046902090311050415,
-0.16094985604286194,
0.020117372274398804,
-0.06413903087377548,
0.06334125250577927,
0.03641495108604431,
-0.05590536445379257,
-0.0248766727745533,
-0.01730942726135254,
0.011945613659918308,
-0.05083848536014557,
-0.18994836509227753,
-0.056277405470609665,
-0.037882111966609955,
0.149809330701828,
-0.25956398248672485,
0.032966937869787216,
0.051140617579221725,
0.14649195969104767,
0.00406361510977149,
-0.05115427449345589,
0.01429014839231968,
-0.05360214412212372,
-0.054652128368616104,
-0.06746816635131836,
-0.006135428790003061,
-0.027576493099331856,
-0.05147203803062439,
0.019243421033024788,
-0.1755700707435608,
-0.021410830318927765,
0.09424154460430145,
0.12876708805561066,
-0.1486445665359497,
-0.018640631809830666,
-0.048725154250860214,
-0.06339836865663528,
-0.0715010017156601,
-0.07038594037294388,
0.10712739825248718,
0.0513901449739933,
0.04796046018600464,
-0.07435787469148636,
-0.07092321664094925,
0.02726263552904129,
0.006906150374561548,
-0.03382374346256256,
0.08727246522903442,
0.05199531093239784,
-0.09209315478801727,
0.0756213590502739,
0.1092359870672226,
0.07177663594484329,
0.09363535046577454,
0.01574566215276718,
-0.11756632477045059,
-0.028492970392107964,
0.036266472190618515,
0.02740776725113392,
0.1465986967086792,
-0.05952361226081848,
0.04016614332795143,
0.04494241625070572,
-0.04170418903231621,
0.022319864481687546,
-0.08787637203931808,
0.024075502529740334,
0.025203049182891846,
-0.0034381982404738665,
0.06284574419260025,
-0.02525499276816845,
-0.0050758360885083675,
0.07016654312610626,
0.047779910266399384,
0.04621000960469246,
0.009655474685132504,
-0.01720241829752922,
-0.1047825813293457,
0.16950392723083496,
-0.0951867327094078,
-0.269941508769989,
-0.17632324993610382,
0.026197833940386772,
0.04035249724984169,
-0.022378476336598396,
0.031619444489479065,
-0.07056326419115067,
-0.10630585998296738,
-0.1060405746102333,
-0.002429972169920802,
0.01714223250746727,
-0.06364088505506516,
-0.0741225928068161,
0.07348573952913284,
0.04382912442088127,
-0.14902326464653015,
0.038552410900592804,
0.055694397538900375,
-0.057955220341682434,
-0.0233661737293005,
0.09118817001581192,
0.12397737801074982,
0.14583967626094818,
-0.021366750821471214,
-0.028626007959246635,
0.029004426673054695,
0.19620531797409058,
-0.13469526171684265,
0.10371150821447372,
0.13814030587673187,
-0.04545360431075096,
0.08360563963651657,
0.1560150384902954,
0.029186224564909935,
-0.08317049592733383,
0.05044832453131676,
0.04082648828625679,
-0.043159641325473785,
-0.2666129767894745,
-0.0534592866897583,
0.012832709588110447,
-0.06255637854337692,
0.09786593168973923,
0.10183793306350708,
0.11542957276105881,
0.034910861402750015,
-0.07166364789009094,
-0.043925940990448,
-0.0058974819257855415,
0.11737963557243347,
-0.05490213260054588,
-0.012639665976166725,
0.07686592638492584,
-0.05086168646812439,
0.005355054512619972,
0.10266812145709991,
0.02973790094256401,
0.17442677915096283,
0.020399179309606552,
0.11231429129838943,
0.06195578724145889,
0.08633565157651901,
0.0007386076031252742,
0.02951662428677082,
0.05147615820169449,
0.017203815281391144,
-0.002300140680745244,
-0.10421168059110641,
-0.006156572140753269,
0.1449710875749588,
0.028103826567530632,
0.029669636860489845,
-0.0018948549404740334,
-0.005003341939300299,
0.05121048167347908,
0.1746254414319992,
-0.011592294089496136,
-0.22072425484657288,
-0.0845772922039032,
0.06936841458082199,
-0.06218599155545235,
-0.12968985736370087,
-0.026130788028240204,
0.045467354357242584,
-0.17519839107990265,
0.026703642681241035,
-0.027433741837739944,
0.0919293761253357,
-0.09345759451389313,
-0.02221956104040146,
0.03687324374914169,
0.084866963326931,
-0.014529162086546421,
0.08703910559415817,
-0.14498743414878845,
0.11886418610811234,
0.02978132851421833,
0.09024628251791,
-0.11081171780824661,
0.07909037172794342,
-0.007550720125436783,
0.009180475026369095,
0.19379350543022156,
-0.011335089802742004,
-0.03514958545565605,
-0.08774717897176743,
-0.11210042238235474,
-0.013537433929741383,
0.12687496840953827,
-0.1243172138929367,
0.08773399889469147,
-0.015198243781924248,
-0.044079482555389404,
0.00937260314822197,
-0.12100647389888763,
-0.17273177206516266,
-0.19628387689590454,
0.05585884302854538,
-0.09575839340686798,
0.025643249973654747,
-0.11914430558681488,
-0.07089093327522278,
-0.02952558360993862,
0.241120383143425,
-0.1745356321334839,
-0.06510113179683685,
-0.1468164622783661,
-0.046294767409563065,
0.1662203073501587,
-0.04437198117375374,
0.0718095526099205,
-0.0208172257989645,
0.20345525443553925,
0.005988610442727804,
-0.004939318168908358,
0.06724198162555695,
-0.08892562240362167,
-0.16873881220817566,
-0.06771010160446167,
0.1510489284992218,
0.11680185794830322,
0.04907919466495514,
-0.002248800592496991,
0.0011772146681323647,
-0.016943959519267082,
-0.1137804463505745,
-0.0033210667315870523,
0.16037839651107788,
0.03878779336810112,
0.025986969470977783,
-0.05243593826889992,
-0.08797456324100494,
-0.06899320334196091,
-0.06853509694337845,
0.06221301481127739,
0.19590823352336884,
-0.10376439243555069,
0.1700313836336136,
0.147536963224411,
-0.07305635511875153,
-0.23175598680973053,
0.035342130810022354,
0.04983805492520332,
0.0014306638622656465,
0.04886869341135025,
-0.18252557516098022,
0.10521943867206573,
0.019543392583727837,
-0.05505957826972008,
0.13485197722911835,
-0.1557481735944748,
-0.1552847921848297,
0.0722852572798729,
0.03904085233807564,
-0.22423844039440155,
-0.1354004591703415,
-0.09622503817081451,
-0.05825018882751465,
-0.14065024256706238,
0.06054598465561867,
-0.002136280992999673,
0.015948504209518433,
0.03500790148973465,
-0.0015643214574083686,
0.027123261243104935,
-0.058935679495334625,
0.18609118461608887,
-0.004065449349582195,
0.020676052197813988,
-0.060264769941568375,
-0.0478842556476593,
0.09839435666799545,
-0.06130504235625267,
0.12208222597837448,
0.004057085141539574,
0.01594383642077446,
-0.10362856835126877,
-0.048314861953258514,
-0.04328322783112526,
0.05154227837920189,
-0.07548051327466965,
-0.10070807486772537,
-0.043625857681035995,
0.08841723203659058,
0.07005169242620468,
-0.03383097052574158,
0.00549331633374095,
-0.07189501076936722,
0.10019614547491074,
0.17795267701148987,
0.17573626339435577,
0.009926567785441875,
-0.07241068035364151,
0.01677953451871872,
-0.04142116755247116,
0.044231921434402466,
-0.2513144314289093,
0.03756171092391014,
0.06098250672221184,
0.029438555240631104,
0.09217222779989243,
-0.020435843616724014,
-0.1820858269929886,
-0.04050002992153168,
0.08094815909862518,
-0.05452597141265869,
-0.22617179155349731,
-0.019085140898823738,
0.0954197570681572,
-0.2020406424999237,
-0.007372708059847355,
0.03995226323604584,
-0.048725228756666183,
-0.023169852793216705,
0.00010950004070764408,
0.06317184865474701,
0.002471912419423461,
0.09773622453212738,
0.0735151618719101,
0.09715340286493301,
-0.08337292820215225,
0.10562895983457565,
0.10150538384914398,
-0.09572599828243256,
0.03605884686112404,
0.06754924356937408,
-0.05300498008728027,
-0.043293699622154236,
0.03665391728281975,
0.033023297786712646,
0.005234600510448217,
-0.060321882367134094,
0.013913018628954887,
-0.036497246474027634,
0.044923391193151474,
0.08326134830713272,
0.03754979372024536,
-0.013354414142668247,
0.06462216377258301,
0.03401726484298706,
-0.10898099094629288,
0.10366570204496384,
0.01731540448963642,
0.04105307161808014,
-0.08384523540735245,
-0.019968897104263306,
0.035425446927547455,
0.030576206743717194,
-0.01765924133360386,
-0.02306121215224266,
-0.02860277332365513,
-0.01614218018949032,
-0.14299540221691132,
-0.023106401786208153,
-0.07243485748767853,
0.006181265693157911,
0.014656842686235905,
-0.031884219497442245,
-0.011233693920075893,
0.02475680410861969,
-0.06979699432849884,
-0.07426341623067856,
-0.006949664559215307,
0.09833318740129471,
-0.15115703642368317,
0.008848577737808228,
0.06907843053340912,
-0.11088496446609497,
0.08190931379795074,
-0.008411259390413761,
0.016245156526565552,
0.022527478635311127,
-0.15448406338691711,
0.05601610988378525,
0.0008648968650959432,
0.01916889287531376,
0.025886621326208115,
-0.16471809148788452,
0.004104440100491047,
-0.04661374166607857,
-0.02149827405810356,
-0.00004464812809601426,
-0.02647159807384014,
-0.12325995415449142,
0.06858719140291214,
-0.015622655861079693,
-0.035931166261434555,
-0.02701525390148163,
0.0539589487016201,
0.07888586074113846,
-0.027474910020828247,
0.10445091128349304,
-0.008690856397151947,
0.04941811040043831,
-0.16801609098911285,
-0.02470702864229679,
-0.04982255399227142,
0.019377702847123146,
0.009884213097393513,
-0.007693959400057793,
0.04183054715394974,
-0.00976533442735672,
0.21883612871170044,
-0.05075952783226967,
0.1607085019350052,
0.05847611650824547,
-0.017352959141135216,
-0.0007513365126214921,
0.06180921941995621,
0.05997028574347496,
0.04658793285489082,
0.009480604901909828,
0.023740366101264954,
-0.022450892254710197,
-0.006695089396089315,
-0.15932634472846985,
0.01890849508345127,
0.14999441802501678,
0.06301083415746689,
0.024745315313339233,
0.05866100639104843,
-0.12775006890296936,
-0.12135478109121323,
0.09311001747846603,
-0.026755332946777344,
0.00928465835750103,
-0.08245618641376495,
0.1358020007610321,
0.14980104565620422,
-0.14000412821769714,
0.05256148427724838,
-0.06134212389588356,
-0.05217423290014267,
-0.10388828068971634,
-0.12032219022512436,
-0.05887215584516525,
-0.053666237741708755,
0.002330566756427288,
-0.03760887682437897,
0.054546963423490524,
0.03344334661960602,
-0.009351172484457493,
-0.00022941511997487396,
0.13597318530082703,
-0.019751882180571556,
-0.0028988157864660025,
0.048313532024621964,
0.03693558648228645,
0.02373051457107067,
-0.05275435373187065,
0.02940409444272518,
0.02539868652820587,
0.032232340425252914,
0.06546790152788162,
0.033412106335163116,
-0.047448933124542236,
0.03804153576493263,
-0.0025254099164158106,
-0.11207924783229828,
0.019641218706965446,
-0.00460948096588254,
-0.0742158442735672,
0.1268945336341858,
0.0407399944961071,
0.010224059224128723,
-0.03741471841931343,
0.24361543357372284,
-0.06653323769569397,
-0.06378097087144852,
-0.13251738250255585,
0.10491154342889786,
-0.0027236645109951496,
0.06476365029811859,
0.023412218317389488,
-0.1284150779247284,
0.005243356805294752,
0.13858191668987274,
0.12181595712900162,
0.0045748427510261536,
0.009228081442415714,
0.0518609918653965,
0.0025186820421367884,
-0.06998204439878464,
0.054019294679164886,
0.06992026418447495,
0.12919506430625916,
-0.07847554981708527,
0.07680778950452805,
0.0006860480643808842,
-0.08370215445756912,
-0.02947772853076458,
0.11312682181596756,
-0.0409729965031147,
0.03491825982928276,
-0.047444481402635574,
0.10916327685117722,
-0.05787910893559456,
-0.29412412643432617,
0.02350960113108158,
-0.09588567912578583,
-0.15202060341835022,
-0.018367812037467957,
0.05944539234042168,
-0.02624768204987049,
0.018029648810625076,
0.06971040368080139,
-0.06011629104614258,
0.20098382234573364,
0.0335683599114418,
-0.07864278554916382,
-0.0664360448718071,
0.04837050288915634,
-0.06564252078533173,
0.2949807047843933,
0.008418165147304535,
0.02863333560526371,
0.10770907253026962,
-0.03253700211644173,
-0.18271861970424652,
0.010723991319537163,
0.1133992001414299,
-0.08056149631738663,
0.08200647681951523,
0.19000613689422607,
-0.012578671798110008,
0.1209007054567337,
0.05294662341475487,
-0.047376248985528946,
0.04217283055186272,
-0.03389401361346245,
-0.051268599927425385,
-0.10752558708190918,
0.058453381061553955,
-0.05909625440835953,
0.15447644889354706,
0.10152646154165268,
-0.05671518296003342,
-0.004550917539745569,
-0.05555408447980881,
0.04875178262591362,
0.01804669201374054,
0.12263146042823792,
0.02951994352042675,
-0.1865430772304535,
0.032826557755470276,
-0.01144319772720337,
0.10186848044395447,
-0.25588861107826233,
-0.08421015739440918,
0.08833149075508118,
-0.011924264021217823,
-0.05105875805020332,
0.10560628771781921,
0.057650718837976456,
0.04243382066488266,
-0.043439045548439026,
-0.10480839014053345,
-0.02186836116015911,
0.14663739502429962,
-0.1469624787569046,
-0.025013303384184837
] |
null | null | transformers | # Darcy-7b - AWQ
- Model creator: [gmonsoon](https://huggingface.co/gmonsoon)
- Original model: [Darcy-7b](https://huggingface.co/gmonsoon/Darcy-7b)
## Model description
Darcy-7b is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing).
- [macadeliccc/WestLake-7B-v2-laser-truthy-dpo](https://huggingface.co/macadeliccc/WestLake-7B-v2-laser-truthy-dpo)
- [FelixChao/WestSeverus-7B-DPO-v2](https://huggingface.co/FelixChao/WestSeverus-7B-DPO-v2)
- [FelixChao/Faraday-7B](https://huggingface.co/FelixChao/Faraday-7B)
### About AWQ
AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.
AWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead.
It is supported by:
- [Text Generation Webui](https://github.com/oobabooga/text-generation-webui) - using Loader: AutoAWQ
- [vLLM](https://github.com/vllm-project/vllm) - version 0.2.2 or later for support for all model types.
- [Hugging Face Text Generation Inference (TGI)](https://github.com/huggingface/text-generation-inference)
- [Transformers](https://huggingface.co/docs/transformers) version 4.35.0 and later, from any code or client that supports Transformers
- [AutoAWQ](https://github.com/casper-hansen/AutoAWQ) - for use from Python code
| {"license": "apache-2.0", "tags": ["merge", "mergekit", "lazymergekit", "macadeliccc/WestLake-7B-v2-laser-truthy-dpo", "FelixChao/WestSeverus-7B-DPO-v2", "FelixChao/Faraday-7B"], "model_name": "Darcy-7b", "base_model": ["macadeliccc/WestLake-7B-v2-laser-truthy-dpo", "FelixChao/WestSeverus-7B-DPO-v2", "FelixChao/Faraday-7B"], "pipeline_tag": "text-generation", "model_type": "mistral", "model_creator": "gmonsoon", "quantized_by": "Suparious"} | text-generation | solidrust/Darcy-7b-AWQ | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"merge",
"mergekit",
"lazymergekit",
"macadeliccc/WestLake-7B-v2-laser-truthy-dpo",
"FelixChao/WestSeverus-7B-DPO-v2",
"FelixChao/Faraday-7B",
"base_model:macadeliccc/WestLake-7B-v2-laser-truthy-dpo",
"base_model:FelixChao/WestSeverus-7B-DPO-v2",
"base_model:FelixChao/Faraday-7B",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"4-bit",
"region:us"
] | 2024-02-08T08:02:55+00:00 | [] | [] | TAGS
#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #macadeliccc/WestLake-7B-v2-laser-truthy-dpo #FelixChao/WestSeverus-7B-DPO-v2 #FelixChao/Faraday-7B #base_model-macadeliccc/WestLake-7B-v2-laser-truthy-dpo #base_model-FelixChao/WestSeverus-7B-DPO-v2 #base_model-FelixChao/Faraday-7B #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us
| # Darcy-7b - AWQ
- Model creator: gmonsoon
- Original model: Darcy-7b
## Model description
Darcy-7b is a merge of the following models using LazyMergekit.
- macadeliccc/WestLake-7B-v2-laser-truthy-dpo
- FelixChao/WestSeverus-7B-DPO-v2
- FelixChao/Faraday-7B
### About AWQ
AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.
AWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead.
It is supported by:
- Text Generation Webui - using Loader: AutoAWQ
- vLLM - version 0.2.2 or later for support for all model types.
- Hugging Face Text Generation Inference (TGI)
- Transformers version 4.35.0 and later, from any code or client that supports Transformers
- AutoAWQ - for use from Python code
| [
"# Darcy-7b - AWQ\n\n- Model creator: gmonsoon\n- Original model: Darcy-7b",
"## Model description\n\nDarcy-7b is a merge of the following models using LazyMergekit.\n\n- macadeliccc/WestLake-7B-v2-laser-truthy-dpo\n- FelixChao/WestSeverus-7B-DPO-v2\n- FelixChao/Faraday-7B",
"### About AWQ\n\nAWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.\n\nAWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead.\n\nIt is supported by:\n\n- Text Generation Webui - using Loader: AutoAWQ\n- vLLM - version 0.2.2 or later for support for all model types.\n- Hugging Face Text Generation Inference (TGI)\n- Transformers version 4.35.0 and later, from any code or client that supports Transformers\n- AutoAWQ - for use from Python code"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #macadeliccc/WestLake-7B-v2-laser-truthy-dpo #FelixChao/WestSeverus-7B-DPO-v2 #FelixChao/Faraday-7B #base_model-macadeliccc/WestLake-7B-v2-laser-truthy-dpo #base_model-FelixChao/WestSeverus-7B-DPO-v2 #base_model-FelixChao/Faraday-7B #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n",
"# Darcy-7b - AWQ\n\n- Model creator: gmonsoon\n- Original model: Darcy-7b",
"## Model description\n\nDarcy-7b is a merge of the following models using LazyMergekit.\n\n- macadeliccc/WestLake-7B-v2-laser-truthy-dpo\n- FelixChao/WestSeverus-7B-DPO-v2\n- FelixChao/Faraday-7B",
"### About AWQ\n\nAWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.\n\nAWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead.\n\nIt is supported by:\n\n- Text Generation Webui - using Loader: AutoAWQ\n- vLLM - version 0.2.2 or later for support for all model types.\n- Hugging Face Text Generation Inference (TGI)\n- Transformers version 4.35.0 and later, from any code or client that supports Transformers\n- AutoAWQ - for use from Python code"
] | [
185,
23,
70,
180
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #macadeliccc/WestLake-7B-v2-laser-truthy-dpo #FelixChao/WestSeverus-7B-DPO-v2 #FelixChao/Faraday-7B #base_model-macadeliccc/WestLake-7B-v2-laser-truthy-dpo #base_model-FelixChao/WestSeverus-7B-DPO-v2 #base_model-FelixChao/Faraday-7B #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n# Darcy-7b - AWQ\n\n- Model creator: gmonsoon\n- Original model: Darcy-7b## Model description\n\nDarcy-7b is a merge of the following models using LazyMergekit.\n\n- macadeliccc/WestLake-7B-v2-laser-truthy-dpo\n- FelixChao/WestSeverus-7B-DPO-v2\n- FelixChao/Faraday-7B### About AWQ\n\nAWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.\n\nAWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead.\n\nIt is supported by:\n\n- Text Generation Webui - using Loader: AutoAWQ\n- vLLM - version 0.2.2 or later for support for all model types.\n- Hugging Face Text Generation Inference (TGI)\n- Transformers version 4.35.0 and later, from any code or client that supports Transformers\n- AutoAWQ - for use from Python code"
] | [
-0.10054432600736618,
0.07320374250411987,
-0.0023403677623718977,
0.027308613061904907,
0.04516207054257393,
-0.009079454466700554,
0.019972069188952446,
0.09441859275102615,
0.044770270586013794,
0.11647698283195496,
-0.03810491785407066,
-0.021520934998989105,
0.10576808452606201,
0.08443276584148407,
0.005869466811418533,
-0.16278822720050812,
0.03109472244977951,
-0.04223784804344177,
0.01988982781767845,
0.04990557208657265,
0.10619362443685532,
-0.05503137782216072,
0.0959005281329155,
0.025885311886668205,
0.009488346055150032,
0.010521622374653816,
-0.022765561938285828,
-0.04390442371368408,
0.06472085416316986,
0.10787311941385269,
0.04529504477977753,
0.0015832791104912758,
0.01893492415547371,
-0.20129337906837463,
0.00790136307477951,
0.046059925109148026,
0.007627369370311499,
0.06199047714471817,
0.07679109275341034,
-0.025456294417381287,
0.06131179258227348,
-0.1139802485704422,
0.06096956506371498,
0.07115686684846878,
-0.07247505336999893,
-0.08866322785615921,
-0.07623399794101715,
0.05068041384220123,
0.09281949698925018,
0.02852841652929783,
-0.0019709118641912937,
0.06111171841621399,
-0.001176031888462603,
0.04202613607048988,
0.10134664922952652,
-0.2612704932689667,
-0.035946592688560486,
0.0733318105340004,
0.048014335334300995,
0.04016152396798134,
-0.058663565665483475,
0.04681209474802017,
0.029693586751818657,
0.006110073998570442,
0.05505082756280899,
-0.014931334182620049,
0.14390447735786438,
-0.011343328282237053,
-0.09247606992721558,
-0.007486815098673105,
0.06158941611647606,
0.02284410037100315,
-0.06263935565948486,
-0.16474132239818573,
-0.08118519932031631,
-0.018436865881085396,
-0.04253409430384636,
-0.05311085656285286,
0.019837193191051483,
0.0008324343361891806,
0.02198316715657711,
-0.04808623343706131,
-0.04832879453897476,
-0.014183061197400093,
-0.07683191448450089,
0.12201514095067978,
0.020793495699763298,
0.011501508764922619,
0.011346814222633839,
0.08014246821403503,
-0.16594716906547546,
-0.13042306900024414,
-0.06746115535497665,
-0.042415838688611984,
-0.09534711390733719,
-0.02525760605931282,
-0.03646141290664673,
0.013024453073740005,
0.017026741057634354,
0.1644822508096695,
-0.029019195586442947,
0.048156678676605225,
0.019345831125974655,
-0.015691936016082764,
-0.008650710806250572,
0.1539210081100464,
-0.016527635976672173,
-0.19969694316387177,
0.035037148743867874,
0.03212741017341614,
0.11138706654310226,
-0.005467715673148632,
-0.024934057146310806,
-0.09945754706859589,
-0.0363323837518692,
0.046581607311964035,
0.02336050570011139,
-0.02182776853442192,
-0.08451070636510849,
-0.06267129629850388,
0.20886372029781342,
-0.12335740029811859,
0.038649093359708786,
0.0006342903361655772,
-0.06884261965751648,
0.007118355017155409,
0.060183364897966385,
0.03201070427894592,
-0.0342409610748291,
0.06038924306631088,
-0.04719960317015648,
-0.02768777869641781,
-0.062067288905382156,
-0.10355400294065475,
0.027005959302186966,
0.004555363208055496,
-0.04098628833889961,
-0.11922212690114975,
-0.17561441659927368,
-0.034113358706235886,
-0.010755816474556923,
-0.058629002422094345,
-0.0016037411987781525,
0.01795201003551483,
-0.0006546293152496219,
0.017667176201939583,
0.01213165931403637,
0.04032189026474953,
-0.03464845195412636,
0.028966981917619705,
0.0331275649368763,
0.06135891377925873,
0.00006143813516246155,
0.01708461344242096,
-0.07102582603693008,
0.03582485392689705,
-0.2065410017967224,
0.12445516884326935,
-0.09316238760948181,
0.02215074934065342,
-0.12211182713508606,
-0.022865794599056244,
-0.02609560824930668,
-0.016300881281495094,
0.05366959795355797,
0.08784077316522598,
-0.14145402610301971,
-0.05044906586408615,
0.05482969060540199,
-0.13409382104873657,
-0.06342824548482895,
0.11217949539422989,
0.021810106933116913,
0.011908686719834805,
0.11036109179258347,
0.04208352416753769,
0.2452857941389084,
-0.14815540611743927,
-0.07557808607816696,
-0.014569433405995369,
0.07746148854494095,
0.0991237461566925,
0.07893505692481995,
-0.03339304402470589,
-0.039833828806877136,
0.044498953968286514,
-0.024438783526420593,
0.008537248708307743,
0.010469021275639534,
-0.05834438279271126,
-0.04893583804368973,
-0.07475720345973969,
0.03663436323404312,
-0.04186883568763733,
-0.0281011164188385,
-0.032501015812158585,
-0.08392707258462906,
-0.02746979519724846,
0.14431782066822052,
0.014417651109397411,
-0.026066157966852188,
-0.10479087382555008,
0.07497183233499527,
-0.06307180970907211,
-0.0009250182774849236,
-0.12086008489131927,
-0.0916871726512909,
0.07710396498441696,
-0.14550401270389557,
-0.010700457729399204,
-0.06677387654781342,
0.05972613766789436,
0.08331078290939331,
-0.05046062543988228,
-0.018590262159705162,
0.011542478576302528,
-0.013806699775159359,
-0.050304118543863297,
-0.0620855987071991,
-0.050380952656269073,
-0.030582580715417862,
0.17027989029884338,
-0.11197216808795929,
0.03212112560868263,
0.03193439170718193,
0.13174273073673248,
0.029272383078932762,
-0.018115166574716568,
0.025390034541487694,
0.009093977510929108,
0.013836728408932686,
-0.04737910255789757,
0.014918125234544277,
-0.000990218948572874,
-0.08750230073928833,
0.10246419161558151,
-0.14765162765979767,
0.03464783728122711,
0.08323709666728973,
0.21165339648723602,
-0.03893313184380531,
-0.04096091538667679,
-0.0003655132604762912,
-0.04375080764293671,
-0.07668758928775787,
-0.08912082016468048,
0.14711634814739227,
0.022040311247110367,
0.09588292986154556,
-0.0928259789943695,
-0.037093162536621094,
-0.009945164434611797,
0.02739768475294113,
-0.01309878472238779,
0.12433868646621704,
-0.025756416842341423,
-0.12392280250787735,
0.02190624177455902,
0.12491124123334885,
0.03546518087387085,
0.13642369210720062,
0.035364605486392975,
-0.07349816709756851,
-0.041829913854599,
0.0019686142913997173,
0.03591802343726158,
0.07566927373409271,
0.05571121722459793,
0.05750564485788345,
0.045590683817863464,
-0.016489652916789055,
0.02813579887151718,
-0.08980903774499893,
0.07735277712345123,
0.030460122972726822,
-0.03863604739308357,
0.06581401824951172,
0.0026153356302529573,
0.012043336406350136,
0.08350548148155212,
0.021041739732027054,
0.03457212820649147,
-0.034636154770851135,
-0.023682095110416412,
-0.10253659635782242,
0.1183890700340271,
-0.09440533816814423,
-0.1771293431520462,
-0.19906263053417206,
-0.029672758653759956,
-0.07675572484731674,
-0.0047151269391179085,
0.018202491104602814,
-0.005761680193245411,
-0.096336230635643,
-0.0899169072508812,
0.04918128252029419,
0.023006439208984375,
-0.04162795841693878,
-0.0023557820823043585,
0.0027067663613706827,
0.06867509335279465,
-0.1269298642873764,
-0.01523511577397585,
0.037149570882320404,
-0.08929000794887543,
-0.0035128004383295774,
0.008937139064073563,
0.08771465718746185,
0.07068754732608795,
0.028327887877821922,
-0.04518979415297508,
-0.027038251981139183,
0.1897466629743576,
-0.05224618315696716,
0.10352449864149094,
0.18267689645290375,
0.022335879504680634,
0.0714445561170578,
0.11035334318876266,
0.01997709460556507,
-0.06196709722280502,
0.02995561622083187,
0.0038139158859848976,
-0.01759410835802555,
-0.13407181203365326,
-0.13216286897659302,
-0.013854116201400757,
-0.031393177807331085,
0.07939818501472473,
0.054250650107860565,
-0.057924628257751465,
0.028410974889993668,
-0.08924977481365204,
0.012448664754629135,
0.060302767902612686,
0.06832106411457062,
0.15402750670909882,
0.014024881646037102,
0.10619054734706879,
-0.06725180149078369,
0.018370335921645164,
0.10886988788843155,
0.03796740248799324,
0.0881923958659172,
-0.02908765710890293,
0.10495627671480179,
-0.014697201550006866,
0.10137849301099777,
0.07943743467330933,
0.02063021995127201,
0.004260699264705181,
0.025284016504883766,
-0.02061089128255844,
-0.10387036949396133,
-0.05318750813603401,
0.0655960813164711,
0.005730692762881517,
0.05592659115791321,
-0.006479853298515081,
0.00866636261343956,
0.04831738397479057,
0.17491787672042847,
0.05075577646493912,
-0.20517803728580475,
-0.12101491540670395,
0.05476010963320732,
-0.03759539872407913,
-0.05469313636422157,
-0.0058937277644872665,
0.06113746762275696,
-0.10477624088525772,
0.11322105675935745,
0.01462409645318985,
0.08792421221733093,
-0.0478208065032959,
-0.015915682539343834,
-0.0071217212826013565,
0.11621162295341492,
0.010743406601250172,
0.06888924539089203,
-0.19021834433078766,
0.030398927628993988,
0.04677104577422142,
0.08819692581892014,
-0.01886446215212345,
0.0653044730424881,
0.0532863512635231,
0.03827429935336113,
0.058321766555309296,
0.007814045995473862,
0.0321323424577713,
-0.12801946699619293,
-0.1796388328075409,
0.03791111335158348,
0.04356297478079796,
-0.09423363953828812,
0.08896332234144211,
-0.021890053525567055,
-0.047012344002723694,
-0.05275255814194679,
0.003231225535273552,
-0.10615067929029465,
-0.15022552013397217,
0.02596294693648815,
-0.01877729594707489,
0.04003916308283806,
-0.08571282774209976,
-0.04514054208993912,
-0.12433759868144989,
0.11446907371282578,
-0.10923372954130173,
-0.08328425139188766,
-0.11145562678575516,
-0.04651270806789398,
0.14380523562431335,
-0.09087616205215454,
0.037027642130851746,
-0.004707263316959143,
0.16456235945224762,
-0.056730031967163086,
-0.09047240763902664,
0.036615338176488876,
-0.10873997956514359,
-0.13177956640720367,
-0.014195163734257221,
0.1404781937599182,
0.015922285616397858,
0.07011246681213379,
0.012288684956729412,
0.05496110022068024,
-0.017112987115979195,
-0.10409041494131088,
-0.03231721371412277,
0.22610275447368622,
-0.00857507437467575,
0.09625435620546341,
-0.09287721663713455,
-0.1442864090204239,
-0.065693698823452,
-0.008356497623026371,
0.06683389097452164,
0.21277226507663727,
-0.05375342816114426,
0.08638779819011688,
0.11987419426441193,
-0.0443481020629406,
-0.1602465659379959,
-0.03733085095882416,
0.0701504573225975,
0.06780079007148743,
0.023337416350841522,
-0.1139359250664711,
0.09710606187582016,
0.10769643634557724,
-0.026100341230630875,
0.04113689437508583,
-0.22053681313991547,
-0.1398201435804367,
0.016047116369009018,
0.005333439912647009,
-0.01158133614808321,
-0.0800291895866394,
-0.08935252577066422,
-0.06832239031791687,
-0.1589815616607666,
0.09511853754520416,
-0.013727156445384026,
0.055543337017297745,
0.007168126758188009,
0.011611451394855976,
0.0307615976780653,
-0.014388354495167732,
0.17908978462219238,
-0.056635480374097824,
-0.0005593809182755649,
-0.03424288332462311,
0.0874369740486145,
-0.0054894015192985535,
-0.08537833392620087,
0.09000395238399506,
-0.013270355761051178,
0.04434751719236374,
-0.17486058175563812,
-0.03342068940401077,
-0.035971615463495255,
0.0561286136507988,
-0.048880353569984436,
-0.010769317857921124,
-0.07057639956474304,
0.05330546572804451,
0.04173034057021141,
-0.03577886149287224,
-0.046280961483716965,
-0.04263315722346306,
0.03645374998450279,
0.1126805916428566,
0.07699353992938995,
-0.04795009642839432,
-0.05415588617324829,
-0.020724186673760414,
-0.021257450804114342,
0.07354090362787247,
-0.0671912208199501,
0.05649831146001816,
0.1048993393778801,
0.006341265048831701,
0.08285725861787796,
-0.03247066214680672,
-0.10944443941116333,
-0.0131126893684268,
0.057824715971946716,
-0.10298143327236176,
-0.1023256704211235,
-0.05618266388773918,
0.05134196951985359,
-0.05282604321837425,
-0.07820243388414383,
0.16194657981395721,
-0.030161257833242416,
-0.0275714173913002,
0.021248724311590195,
0.06767324358224869,
-0.006529041565954685,
0.1566808670759201,
-0.014078846201300621,
0.03690072149038315,
-0.08758672326803207,
0.08922320604324341,
0.09093873202800751,
-0.0989217534661293,
-0.017125139012932777,
0.08460249751806259,
-0.0907139852643013,
-0.03194506838917732,
-0.1279796063899994,
0.11993146687746048,
-0.09335754811763763,
-0.05982580780982971,
-0.000983662437647581,
-0.07951100170612335,
0.03199170529842377,
0.15969982743263245,
0.052143342792987823,
-0.0008001461974345148,
0.008942930027842522,
-0.014988863840699196,
0.002910047536715865,
0.10687220841646194,
0.021281758323311806,
0.05330638960003853,
-0.05593378469347954,
0.055344611406326294,
0.018180958926677704,
-0.005094817839562893,
-0.011616382747888565,
-0.02015622705221176,
-0.07212073355913162,
-0.04067246615886688,
-0.1647748202085495,
0.019830575212836266,
-0.03980131447315216,
0.0005109974299557507,
-0.03786981850862503,
-0.02005990780889988,
-0.03412238508462906,
0.02843611314892769,
-0.020345233380794525,
-0.06446687132120132,
-0.06436514854431152,
0.057619012892246246,
-0.11234083771705627,
0.012615321204066277,
0.09049385786056519,
-0.0773371234536171,
0.07977285981178284,
0.01911071687936783,
0.032023973762989044,
-0.015943730250000954,
-0.044700607657432556,
-0.02474374882876873,
0.008670222945511341,
0.04758498817682266,
0.038904640823602676,
-0.1938176304101944,
0.026764480397105217,
0.006246109493076801,
-0.06716383993625641,
-0.005295672453939915,
0.13356859982013702,
-0.09307495504617691,
-0.006465724669396877,
0.010517498478293419,
-0.015425712801516056,
-0.07297860831022263,
-0.011109315790235996,
0.08939635753631592,
0.06052989140152931,
0.09968804568052292,
-0.05017302185297012,
0.007269273977726698,
-0.15662652254104614,
-0.03822169080376625,
-0.040551941841840744,
-0.027550090104341507,
-0.010915896855294704,
-0.01021081767976284,
0.04813408851623535,
0.011152449063956738,
0.11492300778627396,
-0.0496072880923748,
0.06157957389950752,
0.04491325840353966,
-0.017469830811023712,
-0.06433672457933426,
-0.0037338444963097572,
0.09772103279829025,
0.07297670841217041,
0.03679509460926056,
0.029811425134539604,
0.03429160639643669,
0.0029389981646090746,
-0.02621692605316639,
0.018795354291796684,
0.09661857038736343,
0.058304980397224426,
0.04507998749613762,
0.05875305458903313,
-0.051920004189014435,
-0.05594545975327492,
0.018413346260786057,
-0.09018819779157639,
0.03552919626235962,
-0.016259919852018356,
0.07055026292800903,
0.13337594270706177,
-0.2044622153043747,
0.06761319190263748,
-0.0094636594876647,
-0.025942346081137657,
-0.09787815809249878,
-0.08162789046764374,
-0.0814177542924881,
-0.04227651283144951,
-0.029401110485196114,
-0.09828372299671173,
0.00880490243434906,
0.060212165117263794,
0.012676925398409367,
0.01147877611219883,
0.1352035254240036,
-0.1546923965215683,
-0.0643545389175415,
0.0331195704638958,
0.025642625987529755,
0.0381038561463356,
0.0911877304315567,
-0.04190041869878769,
0.02421378716826439,
0.048699751496315,
0.050752248615026474,
0.033234622329473495,
0.03977163881063461,
0.02735540084540844,
-0.012218584306538105,
-0.04105563834309578,
0.00511751277372241,
-0.014297505840659142,
0.0041540805250406265,
0.10222659260034561,
0.06903243064880371,
-0.004749976564198732,
-0.006575575564056635,
0.23180484771728516,
-0.06388597935438156,
-0.09684845805168152,
-0.13503241539001465,
0.14631229639053345,
-0.015844570472836494,
0.03373910114169121,
0.012820192612707615,
-0.10410457104444504,
-0.012220197357237339,
0.1287630945444107,
0.07796867936849594,
-0.06524819880723953,
0.009014046750962734,
-0.009031723253428936,
-0.006519464310258627,
0.014311213977634907,
0.022926457226276398,
0.08338990807533264,
0.24863353371620178,
-0.05020926147699356,
0.05630863457918167,
-0.01815476082265377,
-0.04567960277199745,
-0.10387399792671204,
0.1411287486553192,
-0.0802764818072319,
0.013138029724359512,
-0.01935175433754921,
0.061583034694194794,
0.05545696243643761,
-0.18514400720596313,
-0.019322436302900314,
-0.1339026391506195,
-0.11815197765827179,
0.0041089304722845554,
0.01648755557835102,
-0.01020895317196846,
0.04213559255003929,
0.013084465637803078,
0.012673438526690006,
0.06728392839431763,
-0.016876595094799995,
-0.03249169513583183,
-0.07750612497329712,
0.04570415988564491,
0.01938120648264885,
0.11682376265525818,
0.02932092361152172,
0.021551521494984627,
0.10503336787223816,
0.009274255484342575,
-0.1403745859861374,
0.08876858651638031,
0.018567530438303947,
-0.11469994485378265,
0.028407327830791473,
0.15072616934776306,
0.002149202162399888,
0.14925383031368256,
0.09794987738132477,
-0.014004421420395374,
0.04651004076004028,
0.0025891137775033712,
0.0063171712681651115,
-0.06338407099246979,
0.06299257278442383,
-0.08335253596305847,
0.14477364718914032,
0.16276177763938904,
-0.020615771412849426,
0.02219698764383793,
-0.04962088167667389,
0.05442604050040245,
0.007241276558488607,
0.09664534032344818,
-0.0011695814318954945,
-0.1534472405910492,
0.035638000816106796,
0.021039988845586777,
0.05327530950307846,
-0.1982194036245346,
-0.09460926800966263,
-0.023299511522054672,
-0.025119086727499962,
-0.002483449410647154,
0.12492118775844574,
0.08135535567998886,
0.00038294022670015693,
-0.021006856113672256,
-0.154148668050766,
0.014960134401917458,
0.07960831373929977,
-0.12914197146892548,
-0.06571963429450989
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# SMIDS_3x_beit_large_Adamax_lr001_fold2
This model is a fine-tuned version of [microsoft/beit-large-patch16-224](https://huggingface.co/microsoft/beit-large-patch16-224) on the imagefolder dataset.
It achieves the following results on the evaluation set:
- Loss: 1.2566
- Accuracy: 0.8785
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.001
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 50
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|
| 0.4861 | 1.0 | 450 | 0.5758 | 0.7937 |
| 0.471 | 2.0 | 900 | 0.4153 | 0.8286 |
| 0.2546 | 3.0 | 1350 | 0.5147 | 0.8170 |
| 0.438 | 4.0 | 1800 | 0.3988 | 0.8319 |
| 0.5075 | 5.0 | 2250 | 0.3343 | 0.8669 |
| 0.3039 | 6.0 | 2700 | 0.3548 | 0.8702 |
| 0.3013 | 7.0 | 3150 | 0.3367 | 0.8636 |
| 0.2694 | 8.0 | 3600 | 0.3849 | 0.8636 |
| 0.2787 | 9.0 | 4050 | 0.4740 | 0.8436 |
| 0.1686 | 10.0 | 4500 | 0.4075 | 0.8586 |
| 0.1552 | 11.0 | 4950 | 0.5130 | 0.8569 |
| 0.1072 | 12.0 | 5400 | 0.5022 | 0.8719 |
| 0.0735 | 13.0 | 5850 | 0.5368 | 0.8702 |
| 0.0949 | 14.0 | 6300 | 0.5410 | 0.8636 |
| 0.0733 | 15.0 | 6750 | 0.8280 | 0.8669 |
| 0.0079 | 16.0 | 7200 | 0.6797 | 0.8702 |
| 0.1273 | 17.0 | 7650 | 0.7963 | 0.8686 |
| 0.0013 | 18.0 | 8100 | 0.8007 | 0.8602 |
| 0.0492 | 19.0 | 8550 | 0.5349 | 0.8752 |
| 0.0161 | 20.0 | 9000 | 0.8632 | 0.8619 |
| 0.0683 | 21.0 | 9450 | 0.6745 | 0.8719 |
| 0.0031 | 22.0 | 9900 | 0.7968 | 0.8652 |
| 0.0491 | 23.0 | 10350 | 0.7553 | 0.8669 |
| 0.0163 | 24.0 | 10800 | 0.8260 | 0.8769 |
| 0.0161 | 25.0 | 11250 | 0.8713 | 0.8652 |
| 0.0231 | 26.0 | 11700 | 0.9006 | 0.8785 |
| 0.0001 | 27.0 | 12150 | 0.7668 | 0.8835 |
| 0.0005 | 28.0 | 12600 | 0.9973 | 0.8819 |
| 0.0015 | 29.0 | 13050 | 0.8626 | 0.8952 |
| 0.006 | 30.0 | 13500 | 0.8797 | 0.8902 |
| 0.0008 | 31.0 | 13950 | 0.8543 | 0.8985 |
| 0.0 | 32.0 | 14400 | 0.9436 | 0.8902 |
| 0.0002 | 33.0 | 14850 | 0.8985 | 0.8918 |
| 0.0001 | 34.0 | 15300 | 1.0603 | 0.8869 |
| 0.0005 | 35.0 | 15750 | 1.1369 | 0.8852 |
| 0.0 | 36.0 | 16200 | 1.0524 | 0.8852 |
| 0.0001 | 37.0 | 16650 | 1.1134 | 0.8835 |
| 0.0 | 38.0 | 17100 | 1.0243 | 0.8835 |
| 0.0 | 39.0 | 17550 | 1.1383 | 0.8835 |
| 0.0046 | 40.0 | 18000 | 1.2573 | 0.8802 |
| 0.0 | 41.0 | 18450 | 1.0366 | 0.8852 |
| 0.0 | 42.0 | 18900 | 1.1028 | 0.8802 |
| 0.0 | 43.0 | 19350 | 1.1434 | 0.8802 |
| 0.0 | 44.0 | 19800 | 1.2147 | 0.8769 |
| 0.0 | 45.0 | 20250 | 1.2232 | 0.8802 |
| 0.0 | 46.0 | 20700 | 1.2154 | 0.8819 |
| 0.0 | 47.0 | 21150 | 1.2480 | 0.8785 |
| 0.0049 | 48.0 | 21600 | 1.2687 | 0.8785 |
| 0.0 | 49.0 | 22050 | 1.2538 | 0.8785 |
| 0.0 | 50.0 | 22500 | 1.2566 | 0.8785 |
### Framework versions
- Transformers 4.32.1
- Pytorch 2.0.1
- Datasets 2.12.0
- Tokenizers 0.13.2
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["imagefolder"], "metrics": ["accuracy"], "base_model": "microsoft/beit-large-patch16-224", "model-index": [{"name": "SMIDS_3x_beit_large_Adamax_lr001_fold2", "results": [{"task": {"type": "image-classification", "name": "Image Classification"}, "dataset": {"name": "imagefolder", "type": "imagefolder", "config": "default", "split": "test", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.8785357737104825, "name": "Accuracy"}]}]}]} | image-classification | onizukal/SMIDS_3x_beit_large_Adamax_lr001_fold2 | [
"transformers",
"pytorch",
"beit",
"image-classification",
"generated_from_trainer",
"dataset:imagefolder",
"base_model:microsoft/beit-large-patch16-224",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T08:03:09+00:00 | [] | [] | TAGS
#transformers #pytorch #beit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/beit-large-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| SMIDS\_3x\_beit\_large\_Adamax\_lr001\_fold2
============================================
This model is a fine-tuned version of microsoft/beit-large-patch16-224 on the imagefolder dataset.
It achieves the following results on the evaluation set:
* Loss: 1.2566
* Accuracy: 0.8785
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.001
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* num\_epochs: 50
### Training results
### Framework versions
* Transformers 4.32.1
* Pytorch 2.0.1
* Datasets 2.12.0
* Tokenizers 0.13.2
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.32.1\n* Pytorch 2.0.1\n* Datasets 2.12.0\n* Tokenizers 0.13.2"
] | [
"TAGS\n#transformers #pytorch #beit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/beit-large-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.32.1\n* Pytorch 2.0.1\n* Datasets 2.12.0\n* Tokenizers 0.13.2"
] | [
81,
115,
4,
30
] | [
"passage: TAGS\n#transformers #pytorch #beit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/beit-large-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50### Training results### Framework versions\n\n\n* Transformers 4.32.1\n* Pytorch 2.0.1\n* Datasets 2.12.0\n* Tokenizers 0.13.2"
] | [
-0.1293599158525467,
0.1724882274866104,
-0.0023454553447663784,
0.13587836921215057,
0.11250235140323639,
0.015344180166721344,
0.13944171369075775,
0.16905122995376587,
-0.08231265097856522,
0.04725457355380058,
0.1399839073419571,
0.13659004867076874,
0.046719420701265335,
0.19427457451820374,
-0.05236957222223282,
-0.2601667046546936,
0.04119328409433365,
0.03234807401895523,
-0.02075078710913658,
0.12314869463443756,
0.09301083534955978,
-0.13055187463760376,
0.11641614139080048,
0.030263151973485947,
-0.1994711011648178,
-0.03690062463283539,
-0.00742433313280344,
-0.06729142367839813,
0.10523199290037155,
-0.003170925425365567,
0.06897550076246262,
0.03744976967573166,
0.0835329070687294,
-0.13024620711803436,
0.0019813377875834703,
0.043246712535619736,
0.0064185261726379395,
0.10353662818670273,
0.05471734330058098,
-0.015182994306087494,
0.07030405849218369,
-0.06875743716955185,
0.06702885776758194,
0.009385199286043644,
-0.11313743144273758,
-0.26980340480804443,
-0.10223843902349472,
0.07157823443412781,
0.08217991888523102,
0.068179190158844,
0.008332090452313423,
0.1646922081708908,
-0.015027978457510471,
0.10447341948747635,
0.23090173304080963,
-0.2640359699726105,
-0.055165741592645645,
0.0294360164552927,
0.014914325438439846,
0.06473758071660995,
-0.10603974759578705,
-0.018760167062282562,
0.02059783600270748,
0.044390130788087845,
0.1409236192703247,
-0.010635782033205032,
-0.02811739780008793,
-0.021928580477833748,
-0.10847067832946777,
-0.08875154703855515,
0.18579065799713135,
0.058072153478860855,
-0.04802494868636131,
-0.07736620306968689,
-0.07186304777860641,
-0.17165552079677582,
-0.041741833090782166,
0.009797174483537674,
0.04154014587402344,
-0.04674985632300377,
-0.10634559392929077,
-0.030910290777683258,
-0.07805538177490234,
-0.051461391150951385,
-0.023120464757084846,
0.1351369321346283,
0.03383360058069229,
0.05723920464515686,
-0.03597215935587883,
0.09929350018501282,
0.0073938071727752686,
-0.17543649673461914,
-0.028137801215052605,
-0.0017277015140280128,
0.015125435777008533,
-0.01991228759288788,
-0.030262885615229607,
-0.06523662805557251,
-0.001314454828388989,
0.1489848643541336,
-0.06082969531416893,
0.060913555324077606,
-0.007318461779505014,
0.04020942002534866,
-0.048562191426754,
0.1864238679409027,
-0.02870783396065235,
-0.01659870520234108,
0.0207351166754961,
0.08822518587112427,
0.06836054474115372,
-0.036532942205667496,
-0.12525734305381775,
0.03098621405661106,
0.12811045348644257,
0.0029374780133366585,
-0.021369412541389465,
0.05286439135670662,
-0.0643334686756134,
-0.0587083138525486,
0.09228596836328506,
-0.08908151835203171,
0.034838590770959854,
-0.010374085046350956,
-0.084370456635952,
-0.06779533624649048,
0.027354132384061813,
0.01850878819823265,
-0.0004323708708398044,
0.07165450602769852,
-0.09108522534370422,
0.014963540248572826,
-0.06533985584974289,
-0.10081785917282104,
0.016071073710918427,
-0.1107335016131401,
0.012424856424331665,
-0.09663169085979462,
-0.19710472226142883,
0.006891206838190556,
0.07721206545829773,
-0.05610070005059242,
-0.06814593821763992,
-0.03639180585741997,
-0.07652068138122559,
0.04154945909976959,
-0.011949662119150162,
0.07310567051172256,
-0.0747697651386261,
0.0913747176527977,
0.022405456751585007,
0.08734618872404099,
-0.05631003528833389,
0.0460691936314106,
-0.1024455726146698,
0.04984736442565918,
-0.19827407598495483,
0.0799676924943924,
-0.0493633896112442,
0.0617455318570137,
-0.09557844698429108,
-0.10542625933885574,
0.03370443359017372,
-0.05006959289312363,
0.06861566007137299,
0.0974634513258934,
-0.1729826033115387,
-0.057957619428634644,
0.1353495568037033,
-0.09663163125514984,
-0.14797286689281464,
0.10109983384609222,
-0.050693199038505554,
0.01928282529115677,
0.047161467373371124,
0.21422064304351807,
0.06321150809526443,
-0.09143578261137009,
-0.02580295503139496,
-0.03329068049788475,
0.04440530762076378,
-0.06494591385126114,
0.10175396502017975,
0.027680888772010803,
0.05365157872438431,
0.023984158411622047,
-0.032899804413318634,
0.03856229409575462,
-0.08387355506420135,
-0.10054522752761841,
-0.05070185661315918,
-0.08565592765808105,
0.039397966116666794,
0.05604296177625656,
0.05994046851992607,
-0.10856401175260544,
-0.09017187356948853,
0.04204317927360535,
0.0943065956234932,
-0.07395226508378983,
0.028962817043066025,
-0.09000826627016068,
0.11608705669641495,
-0.08325600624084473,
-0.02390553615987301,
-0.1791052222251892,
-0.04241684824228287,
0.040629271417856216,
-0.01629319041967392,
-0.006912850774824619,
-0.048891667276620865,
0.07074255496263504,
0.08783093094825745,
-0.05235742777585983,
-0.05203414335846901,
-0.05522594600915909,
0.008213330060243607,
-0.1105295866727829,
-0.1776295155286789,
-0.08015653491020203,
-0.0380605012178421,
0.14977632462978363,
-0.15268713235855103,
0.022225622087717056,
0.061116840690374374,
0.12500962615013123,
0.059797611087560654,
-0.04723487049341202,
-0.007436409126967192,
0.021452799439430237,
-0.05571167171001434,
-0.08678608387708664,
0.05719248577952385,
0.03528200834989548,
-0.07155010849237442,
-0.019102152436971664,
-0.10049699991941452,
0.1498662531375885,
0.13190734386444092,
-0.0015375686343759298,
-0.04512632265686989,
-0.01160994078963995,
-0.06610778719186783,
-0.030441991984844208,
-0.04081778973340988,
0.018804829567670822,
0.10142559558153152,
0.01744643971323967,
0.14419154822826385,
-0.09178037941455841,
-0.036961425095796585,
0.053544968366622925,
-0.028453968465328217,
-0.0331195667386055,
0.07361359149217606,
0.02190210297703743,
-0.14263916015625,
0.15015269815921783,
0.14882412552833557,
-0.04894813522696495,
0.12402692437171936,
-0.036747027188539505,
-0.0615357980132103,
-0.044876549392938614,
-0.037704430520534515,
0.014213677495718002,
0.1403394341468811,
-0.08333314955234528,
-0.005919712595641613,
0.05630137771368027,
0.019257593899965286,
-0.007085299585014582,
-0.18072617053985596,
0.0006808378966525197,
0.03521978110074997,
-0.04604950174689293,
-0.02278841845691204,
-0.014470276422798634,
0.0007941273506730795,
0.09172741323709488,
0.019804218783974648,
-0.07102026045322418,
0.05166372284293175,
0.010580740869045258,
-0.05623085796833038,
0.16415521502494812,
-0.07910753041505814,
-0.19727325439453125,
-0.11776646971702576,
-0.08754957467317581,
-0.10735819488763809,
0.013021474704146385,
0.06737184524536133,
-0.050448641180992126,
-0.04938974231481552,
-0.10206248611211777,
-0.04453543201088905,
0.021900271996855736,
0.02429220825433731,
0.05370878055691719,
-0.008031168952584267,
0.08405356109142303,
-0.09224440902471542,
-0.03291117399930954,
-0.014789600856602192,
0.018657125532627106,
0.06682770699262619,
0.018715238198637962,
0.11069032549858093,
0.08161229640245438,
-0.02844928205013275,
0.05646483600139618,
-0.01682325080037117,
0.2655041813850403,
-0.06765957176685333,
-0.006789656355977058,
0.13932959735393524,
-0.013368978165090084,
0.08428963273763657,
0.1268712729215622,
0.04151352122426033,
-0.09555158019065857,
-0.013173693791031837,
-0.00024822441628202796,
-0.05275752767920494,
-0.1537386178970337,
-0.04163756221532822,
-0.045641690492630005,
-0.0021682933438569307,
0.13930507004261017,
0.03818075731396675,
0.02474883571267128,
0.07807637751102448,
0.020041609182953835,
0.05664918199181557,
-0.017527885735034943,
0.10406769812107086,
0.08156019449234009,
0.06448414921760559,
0.13368317484855652,
-0.03653626888990402,
-0.019387291744351387,
0.05662747099995613,
0.04215037450194359,
0.20423758029937744,
-0.02541770040988922,
0.14701254665851593,
0.02641657367348671,
0.19307395815849304,
0.017521383240818977,
0.0728468969464302,
-0.014410126954317093,
0.0009393728105351329,
-0.019274147227406502,
-0.04702805355191231,
-0.06427313387393951,
0.03288881108164787,
-0.016649875789880753,
0.05632343888282776,
-0.09356046468019485,
0.039105307310819626,
0.059592608362436295,
0.30666422843933105,
0.06539998203516006,
-0.4122132360935211,
-0.09836560487747192,
0.012291035614907742,
0.0009865236934274435,
-0.055195607244968414,
-0.0072626820765435696,
0.0979013666510582,
-0.09949664771556854,
0.08215389400720596,
-0.09418605268001556,
0.08514873683452606,
-0.0845724418759346,
0.020298872143030167,
0.07689075917005539,
0.056060366332530975,
0.013226890936493874,
0.05964293330907822,
-0.21821673214435577,
0.24971400201320648,
0.018467964604496956,
0.04422129690647125,
-0.08908867090940475,
0.010060982778668404,
0.033364444971084595,
0.059161990880966187,
0.08554306626319885,
0.005977867171168327,
-0.09024009108543396,
-0.18880225718021393,
-0.1258762925863266,
0.0005427713040262461,
0.06169470399618149,
-0.036699384450912476,
0.09451829642057419,
-0.018175894394516945,
-0.012127134948968887,
0.021332256495952606,
0.0005201056483201683,
-0.03501477465033531,
-0.103630930185318,
0.02024604007601738,
0.034688886255025864,
-0.012138742953538895,
-0.06473075598478317,
-0.11475593596696854,
-0.03554871678352356,
0.16192500293254852,
0.05505121126770973,
-0.07524240761995316,
-0.1408705860376358,
0.07218684256076813,
0.07781627029180527,
-0.0855332687497139,
0.039305757731199265,
-0.016779718920588493,
0.14986851811408997,
0.020937321707606316,
-0.08943228423595428,
0.10178638249635696,
-0.05869165062904358,
-0.17860572040081024,
-0.041185978800058365,
0.09929849207401276,
0.007366738747805357,
0.05263189971446991,
0.004192214459180832,
0.06014186516404152,
-0.035002902150154114,
-0.0584394596517086,
0.06681792438030243,
-0.0073097143322229385,
0.10614755749702454,
-0.014883637428283691,
0.00864378735423088,
0.029195772483944893,
-0.04613848030567169,
0.00009839441918302327,
0.1684505194425583,
0.24079899489879608,
-0.10403203964233398,
0.060546230524778366,
0.03012177161872387,
-0.030879246070981026,
-0.18261685967445374,
0.010319743305444717,
0.07656802982091904,
-0.0001991603203350678,
0.04173794388771057,
-0.16060468554496765,
0.055176541209220886,
0.10514935851097107,
-0.043303944170475006,
0.08152011036872864,
-0.2768779397010803,
-0.11840421706438065,
0.0923023670911789,
0.138164222240448,
0.0691317543387413,
-0.13107311725616455,
-0.04327763617038727,
-0.041234806180000305,
-0.17335952818393707,
0.13665583729743958,
-0.05704028159379959,
0.11501350998878479,
-0.039327461272478104,
0.08051838725805283,
0.014901114627718925,
-0.056082114577293396,
0.14561402797698975,
0.005515002179890871,
0.08661133795976639,
-0.07185279577970505,
-0.0014093852369114757,
0.10643326491117477,
-0.10252601653337479,
0.07192501425743103,
-0.0869532898068428,
0.06187514215707779,
-0.10810889303684235,
-0.0037693935446441174,
-0.07425615191459656,
0.013987713493406773,
-0.013397954404354095,
-0.048907287418842316,
-0.0448833703994751,
0.03488645330071449,
0.06301422417163849,
-0.018155096098780632,
0.20988906919956207,
0.06445588916540146,
0.0862940326333046,
0.1728745847940445,
0.05397673323750496,
-0.10576145350933075,
-0.09408308565616608,
-0.04430058225989342,
-0.029343122616410255,
0.059755485504865646,
-0.13705183565616608,
0.053009506314992905,
0.12004052102565765,
0.013443393632769585,
0.1280696988105774,
0.05582417547702789,
-0.030783196911215782,
0.035687193274497986,
0.06206676363945007,
-0.1721130907535553,
-0.08640376478433609,
-0.010029762983322144,
0.030597826465964317,
-0.13003188371658325,
0.045725177973508835,
0.12137939780950546,
-0.0593545101583004,
-0.014887568540871143,
-0.004342919681221247,
0.03682979568839073,
-0.009421703405678272,
0.15946903824806213,
0.047883741557598114,
0.05509158596396446,
-0.11808934807777405,
0.11348052322864532,
0.057328153401613235,
-0.0728185623884201,
0.032391179352998734,
0.05030714347958565,
-0.10392948985099792,
-0.021465230733156204,
0.031419817358255386,
0.14932547509670258,
-0.06275127828121185,
-0.045640427619218826,
-0.13568063080310822,
-0.091814324259758,
0.06645428389310837,
0.07967224717140198,
0.0933644250035286,
0.01663324609398842,
-0.03539150580763817,
-0.013165266253054142,
-0.10855977237224579,
0.10982618480920792,
0.04324139654636383,
0.09105362743139267,
-0.17992232739925385,
0.054193608462810516,
-0.0015555275604128838,
0.07246194779872894,
-0.021836427971720695,
-0.00042325531831011176,
-0.08788467198610306,
0.003508437890559435,
-0.10813499987125397,
0.02464236691594124,
-0.052905477583408356,
0.006243168842047453,
-0.02064651995897293,
-0.0580705925822258,
-0.06364380568265915,
0.024784497916698456,
-0.11918067932128906,
-0.053243763744831085,
0.02146504819393158,
0.031834639608860016,
-0.12016978859901428,
-0.04392008110880852,
0.020345089957118034,
-0.08986733108758926,
0.09774119406938553,
0.06029992923140526,
-0.008077923208475113,
0.00773270707577467,
0.0036002967972308397,
-0.02274298295378685,
0.0666942149400711,
0.007561622653156519,
0.08597277849912643,
-0.1152612566947937,
-0.0221384409815073,
0.01634843461215496,
-0.004547150805592537,
0.017726117745041847,
0.15840598940849304,
-0.12086156010627747,
-0.0003179961640853435,
-0.014678256586194038,
-0.06600851565599442,
-0.06344839930534363,
0.06893838196992874,
0.10903503000736237,
0.02346671372652054,
0.21181334555149078,
-0.054371658712625504,
0.015811823308467865,
-0.20995409786701202,
-0.011581460013985634,
0.005185890011489391,
-0.1388559192419052,
-0.10497695952653885,
-0.03237957879900932,
0.06376256048679352,
-0.07031478732824326,
0.11765085160732269,
0.03525954857468605,
0.02161695808172226,
0.02906344085931778,
0.025029366835951805,
-0.0031726681627333164,
0.013450034894049168,
0.16309522092342377,
0.014403261244297028,
-0.028442582115530968,
0.12852592766284943,
0.028986822813749313,
0.09334488213062286,
0.11778779327869415,
0.17672526836395264,
0.11388354748487473,
0.04729508236050606,
0.09055530279874802,
0.05202596262097359,
-0.025968270376324654,
-0.22174733877182007,
0.03601896017789841,
-0.03978736698627472,
0.1488790065050125,
-0.0030294209718704224,
0.15902450680732727,
0.0920415073633194,
-0.18360793590545654,
0.040488436818122864,
-0.03700747340917587,
-0.0790853351354599,
-0.08454839885234833,
-0.12155362963676453,
-0.10311590880155563,
-0.15089921653270721,
0.002945262473076582,
-0.1040843203663826,
0.023338600993156433,
0.11202728003263474,
-0.008582104928791523,
-0.009919910691678524,
0.116677425801754,
-0.02631515823304653,
0.026041926816105843,
0.03836518153548241,
0.00608045794069767,
-0.059937771409749985,
-0.044151950627565384,
-0.08065995573997498,
0.014101422391831875,
0.032313644886016846,
0.05599058046936989,
-0.03235676884651184,
-0.007023791316896677,
0.03841041401028633,
-0.010091220960021019,
-0.12353866547346115,
0.01347822230309248,
0.005028906278312206,
0.05164548382163048,
0.0008541525457985699,
0.012780209071934223,
0.03201600909233093,
-0.015217483974993229,
0.19341084361076355,
-0.07325411587953568,
-0.027416478842496872,
-0.1228807121515274,
0.17896701395511627,
0.0026140701957046986,
-0.04994320869445801,
0.05295133590698242,
-0.09137362241744995,
-0.020702529698610306,
0.15485265851020813,
0.1892986297607422,
-0.07158271223306656,
-0.016520513221621513,
-0.017527583986520767,
-0.013897030614316463,
-0.022615507245063782,
0.09919055551290512,
0.0991419330239296,
-0.0069245584309101105,
-0.0751221776008606,
-0.028980256989598274,
-0.06606413424015045,
-0.034512959420681,
-0.03850788250565529,
0.06925404816865921,
-0.004570751916617155,
0.0070457919500768185,
-0.07483471930027008,
0.04310325160622597,
-0.02210995741188526,
-0.06085818260908127,
0.06226903945207596,
-0.21256737411022186,
-0.17790570855140686,
0.006773421075195074,
0.07538973540067673,
0.0015973751433193684,
0.0461571104824543,
-0.009913075715303421,
0.018662674352526665,
0.07594356685876846,
-0.02225665934383869,
-0.08672447502613068,
-0.09593749046325684,
0.10812120139598846,
-0.13375911116600037,
0.2528570294380188,
-0.03883460536599159,
0.03583916276693344,
0.12127543240785599,
0.041867125779390335,
-0.1335451751947403,
0.03351692110300064,
0.03981999680399895,
-0.032485269010066986,
0.00548918079584837,
0.14240407943725586,
-0.03740047290921211,
0.07958021014928818,
0.0458458811044693,
-0.1027912050485611,
-0.03964604437351227,
-0.04966754838824272,
-0.011354409158229828,
-0.024445757269859314,
-0.054610975086688995,
-0.036348532885313034,
0.13227923214435577,
0.17175258696079254,
-0.042096637189388275,
-0.023690558969974518,
-0.06475082784891129,
0.030860183760523796,
0.07729368656873703,
-0.03295742720365524,
-0.052064236253499985,
-0.23603148758411407,
0.0024359924718737602,
0.05229694023728371,
-0.013576737605035305,
-0.20701472461223602,
-0.110505111515522,
0.0060418094508349895,
-0.05801977962255478,
-0.07628542929887772,
0.09231390058994293,
0.06255589425563812,
0.035103797912597656,
-0.06320928037166595,
0.038133736699819565,
-0.07872021943330765,
0.14179112017154694,
-0.14508864283561707,
-0.07859515398740768
] |
null | null | transformers |
# Caution: This model may output adult content.

# FISH - Fusion of Intuitive Semantic Heuristics
For this model I took the following LoRAs
[SeanWu25/Mixtral_8x7b_Medicine](https://huggingface.co/SeanWu25/Mixtral_8x7b_Medicine)
[SeanWu25/Mixtral_8x7b_WuKurtz](https://huggingface.co/SeanWu25/Mixtral_8x7b_WuKurtz)
and
[wandb/Mixtral-8x7b-Remixtral](https://huggingface.co/wandb/Mixtral-8x7b-Remixtral)
And merged them onto their base model.
I then did a simple linear merge between them, an experimental unreleased 8x7B model, and an unreleased model that was a intermediate step in creating [Envoid/BondBurger-8x7B](https://huggingface.co/Envoid/BondBurger-8x7B?not-for-all-audiences=true)
The end results are a surprisingly good model for role palying style entertainment.
At first I was disappointed with the results but ended up settling on the following sampler parameters which really bring it to life.

## It does eccentuate characters that have particularly aggressive personalities.
## Because this is part of the whole SensualNousInstruct family of models it still suffers from the same tokenizer/special tokens weirdness.
## This model has only been tested in Q8 GGUF form due to hardware limitations.
It responds well to [INST] do a thing [/INST] instruct style formatting (although uses the ChatML special tokens) | {"license": "cc-by-nc-4.0", "tags": ["not-for-all-audiences"]} | text-generation | Envoid/Fish-8x7B | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"not-for-all-audiences",
"conversational",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T08:04:10+00:00 | [] | [] | TAGS
#transformers #safetensors #mixtral #text-generation #not-for-all-audiences #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Caution: This model may output adult content.
 | [
"# Caution: This model may output adult content.\n\n\n"
] | [
"TAGS\n#transformers #safetensors #mixtral #text-generation #not-for-all-audiences #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Caution: This model may output adult content.\n\n\n"
] | [
71,
16,
178,
15,
34,
55
] | [
"passage: TAGS\n#transformers #safetensors #mixtral #text-generation #not-for-all-audiences #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Caution: This model may output adult content.\n\n\n"
] | [
-0.027962669730186462,
-0.09273865073919296,
-0.0029221107251942158,
0.06155562028288841,
0.07195129990577698,
0.034416768699884415,
0.05274803936481476,
0.09207116812467575,
0.011995775625109673,
0.10066066682338715,
0.017931897193193436,
-0.05877932533621788,
0.02225128933787346,
0.12874852120876312,
0.030603058636188507,
-0.15354986488819122,
0.08388220518827438,
-0.026117080822587013,
0.2332351803779602,
0.04552400857210159,
0.08110646903514862,
-0.04602866992354393,
0.06945642828941345,
0.03312006965279579,
0.038026951253414154,
-0.017315270379185677,
0.010943925008177757,
0.043922506272792816,
0.09804262965917587,
0.05743424594402313,
0.052056677639484406,
0.0010279157431796193,
-0.020354852080345154,
-0.16529259085655212,
0.0351385623216629,
0.03151443228125572,
-0.012523061595857143,
0.014589007012546062,
0.06963373720645905,
0.008806143887341022,
0.16603441536426544,
-0.02964664250612259,
0.0005002597463317215,
0.05966819450259209,
-0.09538251906633377,
-0.05096154659986496,
-0.14125259220600128,
0.1123105138540268,
0.04751170799136162,
0.04635421931743622,
-0.03589881584048271,
0.07033513486385345,
-0.047676630318164825,
0.07654419541358948,
0.28892016410827637,
-0.16585496068000793,
-0.022255372256040573,
0.0014622985618188977,
0.14726205170154572,
-0.032997023314237595,
-0.034289974719285965,
-0.025899669155478477,
0.04533492773771286,
0.013627792708575726,
-0.024941064417362213,
-0.02243860252201557,
0.0803944319486618,
-0.04979199916124344,
-0.0592193640768528,
0.0007066805264912546,
0.05990408733487129,
0.08545005321502686,
-0.10079620033502579,
-0.12853853404521942,
-0.01451102364808321,
-0.05819559097290039,
-0.060386884957551956,
-0.02389456331729889,
0.01710585504770279,
0.01920558325946331,
0.09727593511343002,
-0.027083121240139008,
-0.09317653626203537,
0.02871316485106945,
-0.05257125198841095,
0.12907198071479797,
-0.024245774373412132,
0.00748381856828928,
-0.03601273149251938,
0.02726876735687256,
-0.22348980605602264,
-0.05081799253821373,
-0.13372279703617096,
-0.05522046610713005,
-0.03591533750295639,
-0.06831754744052887,
-0.06362858414649963,
-0.0061597698368132114,
0.031288836151361465,
0.14199866354465485,
-0.07436269521713257,
0.05354287847876549,
-0.056214187294244766,
0.023108573630452156,
0.05650442838668823,
0.09144125878810883,
-0.025375157594680786,
-0.08650892227888107,
0.057806238532066345,
0.0769817903637886,
0.11599691212177277,
-0.022008033469319344,
-0.037250254303216934,
0.04429970309138298,
0.03287629783153534,
0.010976867750287056,
0.06048137694597244,
0.06141233444213867,
-0.09769038110971451,
0.055058348923921585,
0.21803072094917297,
-0.0659320205450058,
0.030617298558354378,
0.036456070840358734,
-0.03624027222394943,
0.09790162742137909,
0.018443798646330833,
-0.007409373298287392,
-0.037721771746873856,
0.17313547432422638,
-0.050377268344163895,
-0.0036509300116449594,
-0.051383960992097855,
-0.07743504643440247,
0.08317018300294876,
0.05514529347419739,
-0.04566924273967743,
-0.12629102170467377,
-0.1913955807685852,
-0.07376964390277863,
-0.013235590420663357,
-0.025375206023454666,
0.0321216844022274,
-0.03187962621450424,
-0.060960400849580765,
0.07250691205263138,
-0.015528195537626743,
-0.10958671569824219,
-0.03461769223213196,
-0.02680780738592148,
0.008946364745497704,
0.02718096785247326,
-0.057056963443756104,
0.009425977244973183,
-0.12453217059373856,
0.06183069944381714,
-0.19640687108039856,
0.08071577548980713,
-0.01821441389620304,
-0.0004049667913932353,
-0.05200863257050514,
-0.04201395809650421,
-0.08318319916725159,
0.05952531844377518,
0.024959221482276917,
0.15277253091335297,
-0.11926675587892532,
-0.020640823990106583,
0.09365688264369965,
-0.1498144418001175,
-0.12479134649038315,
0.1192605122923851,
-0.013533633202314377,
-0.00009457671694690362,
0.09097174555063248,
0.04302574694156647,
0.0023164127487689257,
-0.08749662339687347,
-0.14476624131202698,
-0.019295621663331985,
-0.05813858285546303,
0.17463897168636322,
0.04563331976532936,
-0.03091043047606945,
0.05976450815796852,
-0.004768773447722197,
-0.08101852983236313,
-0.007536850869655609,
-0.0018156932201236486,
-0.027246126905083656,
-0.03059447556734085,
-0.005132410209625959,
0.03967749699950218,
-0.015006048604846,
-0.021448565647006035,
-0.06609488278627396,
-0.11523142457008362,
-0.05299878120422363,
0.13778157532215118,
-0.011229309253394604,
0.07984878867864609,
-0.11980567872524261,
0.14096181094646454,
0.01479868683964014,
0.037264350801706314,
-0.12092752009630203,
-0.09321825206279755,
0.02898470312356949,
-0.07700107246637344,
-0.028611380606889725,
-0.023793544620275497,
0.037462420761585236,
-0.00689657311886549,
-0.04507584869861603,
-0.05419439822435379,
-0.023933984339237213,
-0.006124522536993027,
-0.04497619345784187,
-0.16493719816207886,
-0.027849793434143066,
-0.09210700541734695,
0.11294253170490265,
-0.11015252023935318,
-0.03663020581007004,
0.007015560753643513,
0.05963341146707535,
0.04134169965982437,
-0.061744119971990585,
-0.008975405246019363,
0.018157316371798515,
-0.06038583070039749,
-0.013510343618690968,
0.02473226934671402,
0.014294056221842766,
-0.07209992408752441,
0.10063707083463669,
-0.21575859189033508,
-0.1429048478603363,
0.07410170882940292,
0.02610507421195507,
-0.10916658490896225,
0.0660490170121193,
0.01976832002401352,
-0.033549144864082336,
-0.03872697800397873,
-0.1192144826054573,
0.08668851107358932,
0.04387783631682396,
0.08799230307340622,
-0.04042268916964531,
-0.01590188592672348,
-0.0021624460350722075,
-0.029345756396651268,
-0.026023458689451218,
0.032481998205184937,
-0.05978945642709732,
-0.15959212183952332,
0.008319447748363018,
0.0014356843894347548,
-0.0015199152985587716,
0.08707902580499649,
0.011531469412147999,
-0.05682835727930069,
-0.053226929157972336,
0.059938978403806686,
0.05038819834589958,
0.03641330823302269,
-0.06594046950340271,
0.01878151297569275,
0.007560137193650007,
-0.009262221865355968,
0.016174420714378357,
-0.05027196183800697,
0.006604484282433987,
0.04712173715233803,
-0.011688723228871822,
-0.022422049194574356,
0.03641221672296524,
-0.007040892727673054,
0.09251819550991058,
0.04498051106929779,
-0.009642237797379494,
0.012983668595552444,
-0.05665747821331024,
-0.17053647339344025,
0.14758510887622833,
-0.08711142092943192,
-0.17565277218818665,
-0.10815826058387756,
-0.07757512480020523,
-0.02007955126464367,
0.022019168362021446,
0.05877628177404404,
-0.04845033958554268,
-0.06536474078893661,
-0.13848528265953064,
0.028620658442378044,
0.018542423844337463,
-0.06259358674287796,
0.07599703222513199,
-0.07564559578895569,
0.0382569395005703,
-0.06306837499141693,
-0.01706089824438095,
-0.0009237998747266829,
-0.015374613925814629,
0.012127661146223545,
0.004655653610825539,
0.07258470356464386,
0.1098582074046135,
0.05361606180667877,
-0.05300702527165413,
-0.044416263699531555,
0.16283771395683289,
-0.0714002475142479,
0.0812777429819107,
0.15072262287139893,
-0.06616876274347305,
0.07174363732337952,
0.122523233294487,
0.015745172277092934,
-0.01612764596939087,
0.027198277413845062,
0.0019498264882713556,
-0.06357195973396301,
-0.16487941145896912,
-0.06949084252119064,
-0.0738849863409996,
0.06881074607372284,
0.006232395302504301,
0.029839137569069862,
0.04227131977677345,
0.018328256905078888,
-0.0966753140091896,
0.009421166963875294,
0.01335009653121233,
0.11046483367681503,
0.11695617437362671,
0.03795308619737625,
0.03717785328626633,
-0.049978818744421005,
0.006857176311314106,
0.07964599132537842,
-0.07402583211660385,
0.12272388488054276,
-0.03995635360479355,
0.13135623931884766,
0.052434470504522324,
-0.03507811576128006,
-0.017366444692015648,
0.04400159418582916,
-0.06589590013027191,
-0.04475131258368492,
-0.02910398691892624,
-0.08370405435562134,
0.0033106678165495396,
0.10185325145721436,
-0.045663926750421524,
0.010145016014575958,
-0.09225842356681824,
0.030107636004686356,
0.04444432631134987,
0.10506018251180649,
0.02837938629090786,
-0.12060296535491943,
-0.10356797277927399,
0.04967683553695679,
-0.018882224336266518,
-0.010959267616271973,
-0.017651692032814026,
0.05846270173788071,
-0.10810080170631409,
0.13225020468235016,
0.0191909521818161,
0.07601805776357651,
-0.013660662807524204,
0.011837009340524673,
-0.0043706404976546764,
0.039511024951934814,
-0.0017931417096406221,
0.030565766617655754,
-0.13648980855941772,
0.05736588314175606,
0.006600244902074337,
0.10764211416244507,
0.012391619384288788,
0.01163721363991499,
0.08446250855922699,
0.049177829176187515,
0.10731616616249084,
0.03844665363430977,
-0.10396800190210342,
-0.06772240996360779,
0.030572058632969856,
-0.016589054837822914,
0.03521864861249924,
0.028916234150528908,
0.10962649434804916,
-0.03355938196182251,
-0.011847240850329399,
0.011012897826731205,
0.08297856897115707,
-0.1086740493774414,
-0.12636226415634155,
0.048661332577466965,
0.017363952472805977,
-0.013079519383609295,
-0.07853803783655167,
-0.035965610295534134,
-0.09585685282945633,
0.1308833211660385,
-0.008954489603638649,
-0.043182119727134705,
-0.06790019571781158,
0.015367995947599411,
0.046829741448163986,
-0.09850361943244934,
0.006189652718603611,
-0.012318323366343975,
0.07981310784816742,
-0.04993058368563652,
-0.07462189346551895,
0.031297676265239716,
-0.12427963316440582,
-0.21201999485492706,
-0.06539353728294373,
0.12216196954250336,
0.05781092122197151,
0.06529247760772705,
-0.0045030806213617325,
0.0004514467727858573,
0.03164046257734299,
-0.07716046273708344,
-0.015240448527038097,
0.1683066487312317,
-0.06075882539153099,
0.0881352648139,
0.04495597258210182,
-0.12110243737697601,
-0.06632538884878159,
-0.0718272402882576,
0.14513453841209412,
0.30147016048431396,
-0.07332193106412888,
0.1086195558309555,
0.0661511942744255,
-0.09698247164487839,
-0.17642942070960999,
-0.05307071655988693,
0.04804316535592079,
-0.04478120431303978,
0.03926070034503937,
-0.0794559121131897,
0.1065446138381958,
0.01854817382991314,
-0.0163759533315897,
0.0025027028750628233,
-0.17657902836799622,
-0.08810503035783768,
0.02238818258047104,
0.008775441907346249,
0.2457713633775711,
-0.13348299264907837,
-0.07125608623027802,
-0.038016434758901596,
-0.1747833788394928,
0.03633042797446251,
-0.008778008632361889,
0.04352109506726265,
0.0016262163408100605,
0.05149054154753685,
0.0357905812561512,
-0.039137694984674454,
0.12684094905853271,
-0.02587411180138588,
0.022066233679652214,
-0.09568099677562714,
-0.051275551319122314,
-0.01533491536974907,
-0.09123151004314423,
0.12068409472703934,
-0.021957481279969215,
-0.020422913134098053,
-0.10067787766456604,
0.010322906076908112,
-0.09224212169647217,
0.05854610726237297,
-0.0524522103369236,
0.011631074361503124,
-0.13343064486980438,
0.059733420610427856,
0.11978514492511749,
0.013469414785504341,
0.07536349445581436,
-0.05423298850655556,
0.1003231406211853,
0.03508839011192322,
0.11877427250146866,
-0.0037288006860762835,
-0.09012577682733536,
-0.011636929586529732,
-0.04640066996216774,
0.03280644863843918,
-0.04454420134425163,
0.03415410593152046,
0.08651147782802582,
-0.014084422960877419,
0.15687689185142517,
0.027772020548582077,
-0.12966816127300262,
0.06781493872404099,
0.12557710707187653,
-0.0774206593632698,
-0.13249996304512024,
-0.015958311036229134,
0.07755004614591599,
-0.11863192915916443,
-0.04108010604977608,
0.12551967799663544,
-0.05015560984611511,
0.0019361086888238788,
-0.008048130199313164,
0.07483627647161484,
-0.004312095232307911,
0.050266459584236145,
-0.015642473474144936,
0.03655310347676277,
-0.07645463198423386,
-0.003903120057657361,
0.021237840875983238,
-0.11679860204458237,
0.039248596876859665,
0.10738146305084229,
-0.05143611878156662,
-0.0532340481877327,
-0.05122995376586914,
0.09614089876413345,
-0.08724484592676163,
-0.09014809131622314,
-0.015159041620790958,
-0.0806877538561821,
0.053002599626779556,
0.11781832575798035,
0.0017846740083768964,
0.05052824318408966,
0.03565850481390953,
0.05699577555060387,
-0.03554508462548256,
0.04864482581615448,
0.044923629611730576,
0.07540950179100037,
-0.11153716593980789,
0.03592553734779358,
-0.0007583242258988321,
0.03662031888961792,
-0.0017079267418012023,
-0.00975324772298336,
-0.05045868083834648,
-0.05696796253323555,
-0.0958794429898262,
0.07333625853061676,
-0.15787678956985474,
-0.014188718050718307,
0.006117440294474363,
0.043004438281059265,
-0.015523198060691357,
0.016320811584591866,
-0.06217681244015694,
-0.03481196239590645,
-0.017822712659835815,
0.052212487906217575,
-0.13854463398456573,
0.01682278700172901,
0.0367799736559391,
-0.05572132021188736,
0.07703682035207748,
0.03128949925303459,
-0.020658446475863457,
0.022485950961709023,
-0.18848711252212524,
-0.021234141662716866,
-0.005364328157156706,
0.018968403339385986,
-0.023671317845582962,
-0.08099190890789032,
0.00023441269877366722,
-0.007325715385377407,
-0.04956427589058876,
0.011139364913105965,
0.07722090929746628,
-0.08686818182468414,
-0.035470567643642426,
0.001891372725367546,
0.006125778425484896,
-0.05974892899394035,
0.015097470954060555,
0.021488208323717117,
-0.00016080524073913693,
0.15058550238609314,
-0.07908248901367188,
0.039840713143348694,
-0.1415993571281433,
-0.003361042821779847,
0.03187192976474762,
-0.01147888321429491,
-0.03302701935172081,
-0.008474990725517273,
0.015756642445921898,
-0.0019767959602177143,
0.07391540706157684,
-0.08260171115398407,
-0.06265781074762344,
0.028778987005352974,
-0.013068655505776405,
0.03651086613535881,
-0.027951961383223534,
0.023745233193039894,
-0.037773214280605316,
-0.0853433683514595,
-0.10274511575698853,
-0.04101169481873512,
0.021782631054520607,
-0.04064098373055458,
0.20138409733772278,
0.014561978168785572,
0.11461712419986725,
0.12190591543912888,
0.0006698775105178356,
-0.02380630373954773,
-0.026571806520223618,
-0.09742789715528488,
-0.07280761748552322,
0.07233329862356186,
0.00567071745172143,
0.16745145618915558,
0.13060003519058228,
-0.0405484102666378,
0.10187938064336777,
-0.1256466954946518,
0.004117666278034449,
-0.03696731850504875,
-0.1036740243434906,
-0.061612293124198914,
-0.09107518941164017,
0.01601429097354412,
-0.12888768315315247,
0.0742528960108757,
0.047592032700777054,
0.008997183293104172,
-0.039647478610277176,
0.141333669424057,
-0.06742072850465775,
-0.05272401124238968,
0.08322214335203171,
-0.02525906451046467,
-0.005211948417127132,
0.15098188817501068,
0.00020584040612448007,
0.0646880641579628,
-0.017980113625526428,
0.051948487758636475,
0.07878194749355316,
0.08453558385372162,
0.05147448927164078,
-0.030897092074155807,
-0.08479125797748566,
0.03559255972504616,
0.04927791282534599,
0.05007787421345711,
0.16762831807136536,
0.03514731302857399,
0.03396236151456833,
-0.0043799313716590405,
0.15724237263202667,
-0.03093789517879486,
-0.005451838485896587,
-0.13694408535957336,
0.16524863243103027,
-0.009941394440829754,
-0.033917129039764404,
-0.04325030371546745,
-0.09824903309345245,
0.02338116243481636,
0.13979309797286987,
0.08343514800071716,
0.002039288869127631,
0.012067156843841076,
-0.05640098825097084,
0.002380533143877983,
0.016274873167276382,
0.0757908746600151,
0.03183738887310028,
0.23827865719795227,
-0.01688271202147007,
0.12216901034116745,
-0.05082092806696892,
0.03232899680733681,
-0.00926410872489214,
0.07395899295806885,
0.0011166505282744765,
0.01017765048891306,
-0.060757722705602646,
0.04558291658759117,
-0.06482741236686707,
-0.18464630842208862,
-0.020355941727757454,
-0.0836295336484909,
-0.06069530174136162,
-0.008190368302166462,
0.105513796210289,
0.08821069449186325,
0.10205590724945068,
0.02757064253091812,
-0.009202973917126656,
0.0969046950340271,
-0.027868418022990227,
-0.07286269962787628,
-0.022213010117411613,
0.012175437994301319,
-0.08861726522445679,
0.15313392877578735,
0.02725301869213581,
0.06369617581367493,
0.10872997343540192,
0.028395365923643112,
-0.09793722629547119,
0.061043936759233475,
0.013411949388682842,
-0.06909336894750595,
0.04567812755703926,
0.19999772310256958,
0.001984203467145562,
0.035208988934755325,
0.10776310414075851,
-0.06897114962339401,
0.06128881499171257,
0.13105401396751404,
-0.04080430790781975,
-0.10451571643352509,
0.07646781951189041,
-0.10300562530755997,
0.1395595222711563,
0.14485645294189453,
-0.04664772003889084,
-0.05240744724869728,
-0.007423624861985445,
-0.004683988634496927,
0.03557118400931358,
0.16122432053089142,
-0.006502735428512096,
-0.1486814320087433,
0.08531323075294495,
-0.010652258060872555,
0.09448907524347305,
-0.21434016525745392,
-0.07500652968883514,
-0.046092528849840164,
-0.016082635149359703,
-0.025028163567185402,
0.08500050008296967,
0.0745810940861702,
-0.016772598028182983,
-0.024816462770104408,
-0.06961581110954285,
0.0031424653716385365,
0.1315770149230957,
-0.0037238351069390774,
-0.06840700656175613
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# emotion_model_1
This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the imagefolder dataset.
It achieves the following results on the evaluation set:
- Loss: 1.5356
- Accuracy: 0.4437
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 64
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 10
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 2.0785 | 1.0 | 10 | 2.0617 | 0.125 |
| 2.0054 | 2.0 | 20 | 1.9826 | 0.275 |
| 1.8694 | 3.0 | 30 | 1.8516 | 0.325 |
| 1.7212 | 4.0 | 40 | 1.7082 | 0.3812 |
| 1.6101 | 5.0 | 50 | 1.6297 | 0.4375 |
| 1.5409 | 6.0 | 60 | 1.5981 | 0.4188 |
| 1.4801 | 7.0 | 70 | 1.5526 | 0.4437 |
| 1.433 | 8.0 | 80 | 1.5574 | 0.4813 |
| 1.4056 | 9.0 | 90 | 1.5094 | 0.5062 |
| 1.3797 | 10.0 | 100 | 1.5232 | 0.4688 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["imagefolder"], "metrics": ["accuracy"], "base_model": "google/vit-base-patch16-224-in21k", "model-index": [{"name": "emotion_model_1", "results": [{"task": {"type": "image-classification", "name": "Image Classification"}, "dataset": {"name": "imagefolder", "type": "imagefolder", "config": "default", "split": "train", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.44375, "name": "Accuracy"}]}]}]} | image-classification | citradiani/emotion_model_1 | [
"transformers",
"safetensors",
"vit",
"image-classification",
"generated_from_trainer",
"dataset:imagefolder",
"base_model:google/vit-base-patch16-224-in21k",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T08:04:21+00:00 | [] | [] | TAGS
#transformers #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| emotion\_model\_1
=================
This model is a fine-tuned version of google/vit-base-patch16-224-in21k on the imagefolder dataset.
It achieves the following results on the evaluation set:
* Loss: 1.5356
* Accuracy: 0.4437
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* gradient\_accumulation\_steps: 4
* total\_train\_batch\_size: 64
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* num\_epochs: 10
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
82,
144,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 10### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.11263416707515717,
0.14639002084732056,
-0.0023347162641584873,
0.0888909175992012,
0.14138729870319366,
0.03139035031199455,
0.11286097019910812,
0.12491486966609955,
-0.0684804618358612,
0.11771126836538315,
0.12876048684120178,
0.10762254148721695,
0.06416081637144089,
0.15837398171424866,
-0.012195452116429806,
-0.2798991799354553,
0.018411889672279358,
-0.01517258957028389,
-0.11016029119491577,
0.12241557985544205,
0.07580315321683884,
-0.12449830025434494,
0.09262896329164505,
0.0005246457294560969,
-0.1417904794216156,
-0.02783804014325142,
-0.015312002040445805,
-0.04088376462459564,
0.10572696477174759,
0.02993898093700409,
0.08795706182718277,
0.027417423203587532,
0.10044478625059128,
-0.23019297420978546,
0.008242331445217133,
0.08005969971418381,
0.00202979426831007,
0.08548326790332794,
0.10506167262792587,
-0.0053331744857132435,
0.1260325014591217,
-0.10286325216293335,
0.06285823881626129,
0.047724854201078415,
-0.09430777281522751,
-0.22811076045036316,
-0.08376503735780716,
0.08029728382825851,
0.12700186669826508,
0.07024449855089188,
-0.01812668703496456,
0.07701065391302109,
-0.08652091771364212,
0.07569721341133118,
0.2122918963432312,
-0.27187830209732056,
-0.08126580715179443,
0.05188830569386482,
0.023145422339439392,
0.05771065875887871,
-0.1282895803451538,
-0.0025637473445385695,
0.04293740540742874,
-0.0007365276105701923,
0.12336785346269608,
0.030457550659775734,
0.0492582730948925,
0.0008973057847470045,
-0.1503068059682846,
-0.04889225214719772,
0.114815354347229,
0.11795369535684586,
-0.02139466442167759,
-0.09087862074375153,
-0.051153603941202164,
-0.20157219469547272,
-0.047200027853250504,
0.007150442339479923,
0.041553594172000885,
-0.046424806118011475,
-0.0784698948264122,
0.027225269004702568,
-0.0729704424738884,
-0.0805383250117302,
0.039457160979509354,
0.10539969056844711,
0.0652443990111351,
-0.020090697333216667,
0.026216989383101463,
0.1180444285273552,
0.05465051904320717,
-0.1511811465024948,
-0.0009983843192458153,
0.009027903899550438,
-0.06814196705818176,
-0.019747527316212654,
0.0013122770469635725,
-0.0013713239459320903,
0.03210482373833656,
0.13296520709991455,
-0.028446638956665993,
0.058864086866378784,
0.04040234163403511,
0.028181683272123337,
-0.09678678214550018,
0.15896373987197876,
-0.08435394614934921,
-0.08629417419433594,
-0.00903131440281868,
0.12707673013210297,
0.03954126685857773,
0.003598628332838416,
-0.08126597851514816,
0.025159843266010284,
0.10991561412811279,
0.04095623642206192,
-0.006924701854586601,
0.04249408468604088,
-0.07256943732500076,
-0.02524522691965103,
0.07021674513816833,
-0.07497803866863251,
0.04439563676714897,
0.02856682613492012,
-0.06409891694784164,
-0.02705894038081169,
0.02480218932032585,
-0.002490783343091607,
0.001697049243375659,
0.11548973619937897,
-0.10698775947093964,
-0.011147618293762207,
-0.07794956862926483,
-0.090180903673172,
0.0270009133964777,
-0.08176721632480621,
0.007299825549125671,
-0.09643402695655823,
-0.11289111524820328,
-0.04173837602138519,
0.06544730812311172,
-0.05467548221349716,
-0.0724305734038353,
-0.05637528747320175,
-0.09172429144382477,
0.044362593442201614,
0.010354026220738888,
0.10166795551776886,
-0.0613301619887352,
0.10839592665433884,
0.00910210981965065,
0.07410727441310883,
0.0770958662033081,
0.043651290237903595,
-0.07345853745937347,
0.05854402109980583,
-0.17300888895988464,
0.04659843072295189,
-0.08882956206798553,
0.07665956020355225,
-0.13133393228054047,
-0.10625853389501572,
-0.014858040027320385,
-0.025541529059410095,
0.07228405028581619,
0.14633683860301971,
-0.135535329580307,
-0.0730307549238205,
0.1642940193414688,
-0.08770953863859177,
-0.1342461258172989,
0.12483952194452286,
-0.006822817958891392,
-0.05459025874733925,
0.026704726740717888,
0.14338792860507965,
0.08479925245046616,
-0.10560530424118042,
-0.028940100222826004,
-0.03200159966945648,
0.08506885170936584,
0.003855678718537092,
0.11260386556386948,
-0.008235981687903404,
-0.03566497936844826,
0.010659340769052505,
-0.07964567095041275,
0.08405420929193497,
-0.10216405242681503,
-0.08832670748233795,
-0.04018781706690788,
-0.09085588157176971,
0.04981645569205284,
0.060894761234521866,
0.027274563908576965,
-0.07853864133358002,
-0.1349344104528427,
-0.013714869506657124,
0.11356905102729797,
-0.08314316719770432,
-0.0021930201910436153,
-0.052122995257377625,
0.10236294567584991,
-0.07159513235092163,
0.0012610068079084158,
-0.11425551772117615,
-0.07137782126665115,
0.03905215486884117,
-0.06920428574085236,
-0.011560195125639439,
-0.03986148536205292,
0.06450406461954117,
0.091584712266922,
-0.06898240745067596,
-0.0922471135854721,
-0.0695490837097168,
-0.0015480000292882323,
-0.07423476129770279,
-0.2514783442020416,
-0.06672534346580505,
-0.01934506930410862,
0.1693340688943863,
-0.26427143812179565,
0.019072219729423523,
-0.010897038504481316,
0.12155988812446594,
0.04385413974523544,
-0.05445731803774834,
-0.0005740208434872329,
0.02472585253417492,
-0.04851865768432617,
-0.09086538106203079,
0.03608616441488266,
-0.004423175472766161,
-0.08896113932132721,
-0.02830936759710312,
-0.09175287187099457,
0.14701780676841736,
0.11352074891328812,
0.0031616282649338245,
-0.09266263246536255,
-0.044207863509655,
-0.07249965518712997,
-0.04880519211292267,
-0.04066028445959091,
0.028441010043025017,
0.053174663335084915,
0.010674280114471912,
0.11204065382480621,
-0.07892509549856186,
-0.027663039043545723,
0.05529102310538292,
-0.0001453310833312571,
-0.029889751225709915,
0.13140930235385895,
0.11941652745008469,
-0.09170433133840561,
0.13612139225006104,
0.1445479393005371,
-0.05966619774699211,
0.1072651818394661,
-0.04147646203637123,
-0.09395210444927216,
-0.030016805976629257,
0.037450630217790604,
0.027705855667591095,
0.14305511116981506,
-0.13417665660381317,
0.004668020643293858,
0.02000676840543747,
0.006990382913500071,
-0.005290776491165161,
-0.17861677706241608,
-0.025196338072419167,
0.03608254715800285,
-0.05273238569498062,
-0.003481445834040642,
-0.026237161830067635,
-0.0233281459659338,
0.0949050635099411,
0.022979648783802986,
-0.0397622212767601,
-0.001483505591750145,
0.0056907664984464645,
-0.08065425604581833,
0.20813681185245514,
-0.08701495081186295,
-0.13766001164913177,
-0.12290365993976593,
0.020144876092672348,
-0.058213021606206894,
0.002051406307145953,
0.03514114022254944,
-0.11825311183929443,
-0.04884336516261101,
-0.076883465051651,
0.033684633672237396,
-0.012119374237954617,
0.04438469186425209,
0.004303527530282736,
0.019862663000822067,
0.07823160290718079,
-0.07948385179042816,
0.01827317476272583,
-0.010384725406765938,
-0.04328303784132004,
0.03024294786155224,
0.04809220880270004,
0.1236550509929657,
0.13536299765110016,
0.017000291496515274,
0.02123376540839672,
-0.01858403906226158,
0.1991461217403412,
-0.09349862486124039,
0.01010766439139843,
0.10810049623250961,
0.03161110356450081,
0.04961896315217018,
0.15594172477722168,
0.05098717659711838,
-0.08501771092414856,
0.022539310157299042,
0.04696161299943924,
-0.010230314917862415,
-0.20468637347221375,
-0.03739798814058304,
-0.038991399109363556,
-0.0027955195400863886,
0.14033320546150208,
0.04404687136411667,
-0.003086442593485117,
0.07716789841651917,
-0.028968283906579018,
0.02795397862792015,
-0.005904383026063442,
0.07988911867141724,
0.040851980447769165,
0.045385658740997314,
0.10334049165248871,
-0.02789837121963501,
-0.015042003244161606,
0.03592659533023834,
-0.004527559503912926,
0.22594600915908813,
-0.017982499673962593,
0.11404768377542496,
0.04662087932229042,
0.19184358417987823,
0.0062676044180989265,
0.05315234884619713,
0.019322581589221954,
-0.035311970859766006,
0.011621087789535522,
-0.060331929475069046,
-0.02991160936653614,
0.049943577498197556,
0.02768442966043949,
0.06721828877925873,
-0.12033843249082565,
0.049965016543865204,
0.05527234077453613,
0.26810720562934875,
0.08619958162307739,
-0.34674355387687683,
-0.10310164093971252,
0.0074105230160057545,
-0.03011810965836048,
-0.04583440348505974,
0.025142591446638107,
0.11878542602062225,
-0.083377905189991,
0.06493034958839417,
-0.08956445753574371,
0.0727739930152893,
-0.06370449811220169,
-0.009777327999472618,
0.08290155231952667,
0.10797320306301117,
0.00020565080922096968,
0.07431364804506302,
-0.21445868909358978,
0.27579572796821594,
-0.007224635221064091,
0.05974392220377922,
-0.03987380117177963,
0.02084764465689659,
0.0346556231379509,
0.06210324913263321,
0.12016696482896805,
0.0042913369834423065,
-0.027388056740164757,
-0.19374661147594452,
-0.10991548001766205,
0.012148571200668812,
0.10466791689395905,
-0.09372532367706299,
0.11221829056739807,
-0.03685104846954346,
-0.04724780470132828,
0.04044114798307419,
-0.035194214433431625,
-0.11664053797721863,
-0.10671470314264297,
-0.004431837238371372,
-0.03313375264406204,
0.0705779418349266,
-0.10010486096143723,
-0.10355743020772934,
-0.10003864020109177,
0.15852241218090057,
-0.09208190441131592,
-0.02104826830327511,
-0.13741369545459747,
0.11179312318563461,
0.12075566500425339,
-0.08876339346170425,
0.07297070324420929,
-0.021552637219429016,
0.12182997167110443,
0.03075912781059742,
-0.03938739001750946,
0.10550835728645325,
-0.09451151639223099,
-0.21337531507015228,
-0.06346573680639267,
0.12969131767749786,
0.021576015278697014,
0.04728051647543907,
-0.019832175225019455,
0.019474752247333527,
-0.009377488866448402,
-0.08745493739843369,
0.06387338042259216,
0.04454511031508446,
0.04861735925078392,
0.034051358699798584,
-0.042453061789274216,
0.008305559866130352,
-0.05229688063263893,
-0.047152746468782425,
0.10545968264341354,
0.27973297238349915,
-0.1085096225142479,
0.009212692268192768,
0.04718635603785515,
-0.036515794694423676,
-0.1746189296245575,
0.02091968059539795,
0.09898395836353302,
0.015777181833982468,
0.02134227752685547,
-0.18278776109218597,
0.0897112786769867,
0.09193647652864456,
-0.031215477734804153,
0.09691400825977325,
-0.3013317584991455,
-0.11666201055049896,
0.07917776703834534,
0.15326906740665436,
-0.007217583246529102,
-0.16519373655319214,
-0.05582680553197861,
-0.0031748181208968163,
-0.08805415779352188,
0.08901648223400116,
-0.007133519742637873,
0.10602835565805435,
-0.029648292809724808,
-0.02198748290538788,
0.022244691848754883,
-0.06838976591825485,
0.14284832775592804,
-0.02690867893397808,
0.07554071396589279,
-0.029980549588799477,
0.0066247121430933475,
0.008648390881717205,
-0.08054888993501663,
0.031709346920251846,
-0.08477739989757538,
0.05882315710186958,
-0.08010473102331161,
-0.015085134655237198,
-0.08356174826622009,
0.030605347827076912,
-0.046695999801158905,
-0.04093770310282707,
-0.03865404427051544,
0.057089753448963165,
0.08086489140987396,
-0.0031086159870028496,
0.16293217241764069,
0.023965798318386078,
0.12400404363870621,
0.07362338900566101,
0.03660866245627403,
-0.014318861998617649,
-0.09749405831098557,
-0.03604046627879143,
-0.018592186272144318,
0.06747906655073166,
-0.1573304682970047,
0.01599252223968506,
0.11746915429830551,
0.04137099161744118,
0.13556070625782013,
0.055864401161670685,
-0.047139644622802734,
0.0004563053953461349,
0.09016501158475876,
-0.11321057379245758,
-0.1579945832490921,
-0.032161615788936615,
-0.021676242351531982,
-0.14608436822891235,
0.0314783975481987,
0.07656628638505936,
-0.08269017189741135,
0.0003056011046282947,
-0.0004837718151975423,
0.051179204136133194,
-0.00984780304133892,
0.1770128756761551,
0.06954038143157959,
0.07326216250658035,
-0.0904860869050026,
0.10704070329666138,
0.05643029510974884,
-0.16170679032802582,
0.006165422964841127,
0.0651981383562088,
-0.08678799867630005,
-0.02693914622068405,
0.06406639516353607,
0.09380809962749481,
-0.0033883326686918736,
-0.03152918815612793,
-0.11291111260652542,
-0.11929427832365036,
0.07038484513759613,
0.07122254371643066,
0.06183524802327156,
0.02438974753022194,
-0.00525142066180706,
0.03709089383482933,
-0.1138390600681305,
0.13000933825969696,
0.06775228679180145,
0.10212933272123337,
-0.20969921350479126,
0.06807780265808105,
0.012882844544947147,
0.01769593544304371,
-0.00968673825263977,
0.038077425211668015,
-0.12464894354343414,
-0.023315932601690292,
-0.05814618989825249,
-0.008069823496043682,
-0.07849766314029694,
0.00864370446652174,
-0.0029925392009317875,
-0.057790014892816544,
-0.0487656444311142,
0.015841299667954445,
-0.09464287012815475,
-0.06349177658557892,
0.005673701409250498,
0.08239588886499405,
-0.11252865940332413,
-0.008657319471240044,
0.034235015511512756,
-0.11794769018888474,
0.09641645848751068,
0.02132323756814003,
0.053797826170921326,
0.017525935545563698,
-0.11775258928537369,
0.025252746418118477,
0.06006407365202904,
-0.013082542456686497,
0.030234934762120247,
-0.1444271057844162,
0.009145824238657951,
-0.04890082776546478,
-0.011412082239985466,
-0.016803691163659096,
0.06030597537755966,
-0.13998641073703766,
-0.01853449083864689,
-0.044179655611515045,
-0.04505607858300209,
-0.05151282623410225,
0.05575404688715935,
0.06534609943628311,
-0.028122950345277786,
0.17909787595272064,
-0.07727182656526566,
0.0325523316860199,
-0.23738513886928558,
-0.007016455754637718,
-0.023932071402668953,
-0.0712592750787735,
-0.10594896972179413,
-0.016041036695241928,
0.07149717956781387,
-0.05540713667869568,
0.06895162165164948,
-0.019317790865898132,
0.04057838022708893,
0.02759246900677681,
-0.02360231988132,
0.03160146251320839,
0.043351273983716965,
0.1774180680513382,
0.023874277248978615,
-0.022174453362822533,
0.046406541019678116,
0.012398078106343746,
0.09061861038208008,
0.06566467881202698,
0.14264678955078125,
0.14933612942695618,
-0.03950485214591026,
0.08997886627912521,
0.04410441592335701,
-0.1179364025592804,
-0.1770010143518448,
0.12407281249761581,
-0.08336970955133438,
0.14038316905498505,
-0.016798779368400574,
0.1741590052843094,
0.11985959112644196,
-0.19023534655570984,
0.016076959669589996,
-0.019744299352169037,
-0.07539407908916473,
-0.0812983512878418,
-0.10367027670145035,
-0.09444411098957062,
-0.20454151928424835,
0.02525196596980095,
-0.10095613449811935,
0.017247607931494713,
0.056997235864400864,
0.018473932519555092,
0.011072014458477497,
0.17417003214359283,
0.07866358011960983,
0.015484889037907124,
0.07074783742427826,
0.04015457257628441,
-0.05136377736926079,
-0.022284863516688347,
-0.08747921884059906,
0.024789225310087204,
-0.05461863428354263,
0.04074038565158844,
-0.05418992042541504,
-0.08291029930114746,
0.09097734838724136,
0.0409088060259819,
-0.100664883852005,
0.03297772631049156,
-0.016156677156686783,
0.032940782606601715,
0.0719316303730011,
0.0020084779243916273,
0.009621089324355125,
-0.028848983347415924,
0.22469830513000488,
-0.0816410630941391,
-0.011637894436717033,
-0.12034131586551666,
0.20911577343940735,
0.008383158594369888,
-0.00953043531626463,
0.027300698682665825,
-0.08759496361017227,
0.0077950358390808105,
0.176945760846138,
0.14571921527385712,
-0.018432090058922768,
-0.023917322978377342,
0.02704750746488571,
-0.021104328334331512,
-0.046158917248249054,
0.09156399965286255,
0.09887616336345673,
0.04283025488257408,
-0.06317553669214249,
-0.021271731704473495,
-0.04715567082166672,
-0.04752405360341072,
-0.02438933029770851,
0.0545518584549427,
0.028165578842163086,
-0.004878654144704342,
-0.046230413019657135,
0.08932939171791077,
-0.032798923552036285,
-0.10115387290716171,
0.10806617885828018,
-0.18250951170921326,
-0.17183148860931396,
-0.05493626371026039,
0.09610933065414429,
0.00902244821190834,
0.047612424939870834,
-0.003546001622453332,
-0.029295939952135086,
0.09762834012508392,
-0.006959816440939903,
-0.07118354737758636,
-0.11834220588207245,
0.04036738723516464,
-0.034768108278512955,
0.2568790018558502,
-0.03563639894127846,
-0.017168814316391945,
0.12653648853302002,
0.035510700196027756,
-0.13079005479812622,
0.03506341204047203,
0.0731743574142456,
-0.09501080960035324,
0.050512995570898056,
0.14811526238918304,
-0.0329064279794693,
0.12201227247714996,
0.037286076694726944,
-0.09222105145454407,
-0.000491227547172457,
-0.08313700556755066,
-0.056138165295124054,
-0.06058241054415703,
0.0178908109664917,
-0.03705599531531334,
0.14732931554317474,
0.20752622187137604,
-0.05916857346892357,
-0.023254821076989174,
-0.05968793109059334,
0.03828112781047821,
0.05963090807199478,
0.11739694327116013,
0.006382807157933712,
-0.24037835001945496,
0.029462555423378944,
-0.011882128193974495,
0.023698637261986732,
-0.21002861857414246,
-0.088201604783535,
0.018034059554338455,
-0.04161333292722702,
-0.10795049369335175,
0.10735246539115906,
0.09615069627761841,
0.042521655559539795,
-0.05813954398036003,
-0.09358683228492737,
-0.04965652525424957,
0.17412620782852173,
-0.15819603204727173,
-0.08212201297283173
] |
null | null | flair | ## Polish Flair Model -- Part-of-Speech
| {"language": "pl", "tags": ["flair", "token-classification", "sequence-tagger-model"], "widget": [{"text": "Jan Brzechwa - polski poeta i adwokat \u017cydowskiego pochodzenia, autor bajek i wierszy dla dzieci, satyrycznych tekst\u00f3w dla doros\u0142ych, a tak\u017ce t\u0142umacz literatury rosyjskiej."}]} | token-classification | clarin-knext/morpho-flair-pos | [
"flair",
"pytorch",
"token-classification",
"sequence-tagger-model",
"pl",
"region:us"
] | 2024-02-08T08:08:06+00:00 | [] | [
"pl"
] | TAGS
#flair #pytorch #token-classification #sequence-tagger-model #pl #region-us
| ## Polish Flair Model -- Part-of-Speech
| [
"## Polish Flair Model -- Part-of-Speech"
] | [
"TAGS\n#flair #pytorch #token-classification #sequence-tagger-model #pl #region-us \n",
"## Polish Flair Model -- Part-of-Speech"
] | [
30,
14
] | [
"passage: TAGS\n#flair #pytorch #token-classification #sequence-tagger-model #pl #region-us \n## Polish Flair Model -- Part-of-Speech"
] | [
-0.05035844445228577,
-0.037941236048936844,
-0.010546279139816761,
0.06368713080883026,
0.05229058489203453,
0.05889217555522919,
0.09561524540185928,
0.01800938881933689,
0.13973259925842285,
0.05123975872993469,
0.09403879195451736,
0.09986227750778198,
0.010160268284380436,
0.04150402545928955,
0.03212425485253334,
-0.4114713668823242,
0.09381762892007828,
-0.01611085794866085,
-0.022534193471074104,
0.11901513487100601,
0.12116197496652603,
-0.02372170053422451,
-0.03373882174491882,
0.06367407739162445,
0.006519515533000231,
0.05249233916401863,
-0.0007086675614118576,
-0.10252624750137329,
0.12958861887454987,
-0.027865350246429443,
0.0952858105301857,
-0.02410871721804142,
-0.007280990015715361,
-0.12442309409379959,
0.005499038379639387,
-0.0524294488132,
-0.0020653766114264727,
0.02966432459652424,
0.07233255356550217,
-0.1261504888534546,
0.21178384125232697,
0.007746044080704451,
0.053460732102394104,
0.010968281887471676,
-0.10302191972732544,
-0.23616065084934235,
0.032010920345783234,
0.08705711364746094,
0.02872292697429657,
0.04683348909020424,
-0.06194738671183586,
0.1019667387008667,
-0.17741553485393524,
0.0766698345541954,
0.05931508168578148,
-0.2106138914823532,
-0.055455487221479416,
0.01379404217004776,
0.041831035166978836,
0.08534687012434006,
-0.1276150494813919,
0.04512431100010872,
-0.06829146295785904,
0.009369132108986378,
-0.09831273555755615,
-0.047782283276319504,
0.000304466113448143,
0.04404885694384575,
-0.18312573432922363,
-0.009772942401468754,
0.22559517621994019,
-0.033500440418720245,
0.0034911837428808212,
-0.10010713338851929,
0.01184877846390009,
-0.00469949608668685,
-0.003491220297291875,
-0.016288846731185913,
-0.04187208414077759,
0.028766730800271034,
0.20935356616973877,
0.02400275506079197,
-0.06479865312576294,
0.017017779871821404,
-0.031162763014435768,
0.35040900111198425,
0.02422202192246914,
0.06338823586702347,
-0.09461500495672226,
-0.04918165132403374,
-0.02290383167564869,
-0.06206992268562317,
0.05704288184642792,
-0.054154541343450546,
-0.02101046033203602,
0.04845644161105156,
0.02094372548162937,
0.058285657316446304,
0.057476505637168884,
0.13665848970413208,
-0.00426221638917923,
0.04639792442321777,
0.08702091127634048,
0.12698979675769806,
0.09790831804275513,
0.09534823894500732,
0.01094975695014,
-0.06154170259833336,
-0.07248144596815109,
-0.11934387683868408,
0.01671060360968113,
-0.07234596461057663,
-0.1100076213479042,
-0.03623068705201149,
-0.09824999421834946,
0.04558710381388664,
-0.04530775547027588,
0.03755776584148407,
-0.022652655839920044,
-0.04658964276313782,
0.00688914442434907,
-0.02776607871055603,
-0.018859988078475,
0.051868099719285965,
-0.005029828753322363,
0.2554292678833008,
-0.15150544047355652,
0.013435562141239643,
-0.07190439105033875,
0.14790914952754974,
-0.00471567502245307,
0.057454317808151245,
0.034991052001714706,
-0.1042662262916565,
0.00027555288397707045,
-0.09424266964197159,
0.02828717976808548,
-0.07140248268842697,
0.053542107343673706,
0.011343351565301418,
-0.047216031700372696,
-0.014955987222492695,
-0.02920193411409855,
-0.05383524298667908,
-0.008543831296265125,
0.03425491973757744,
-0.005347715690732002,
-0.09433692693710327,
-0.08897849172353745,
0.11799851804971695,
-0.14655916392803192,
0.016503063961863518,
-0.14212282001972198,
0.0364646390080452,
-0.13544613122940063,
-0.05649088695645332,
-0.24804186820983887,
-0.08202316612005234,
-0.07551798969507217,
-0.013532489538192749,
-0.043037861585617065,
-0.0857383981347084,
0.027189284563064575,
0.030565274879336357,
-0.10148029774427414,
0.13094137609004974,
-0.13761021196842194,
-0.09563969820737839,
0.08700579404830933,
-0.15426267683506012,
0.07974261790513992,
0.18631060421466827,
0.02350069023668766,
0.00016419558960478753,
0.03314794600009918,
0.3473256826400757,
0.05705982446670532,
-0.10463660955429077,
0.10208728164434433,
0.07353125512599945,
-0.07941240817308426,
0.062365252524614334,
0.14307190477848053,
-0.08815187215805054,
-0.06556281447410583,
0.027392467483878136,
-0.0029354386497288942,
-0.004691184964030981,
-0.007639802992343903,
-0.029744362458586693,
0.054512497037649155,
0.01009282749146223,
0.22173821926116943,
-0.009159988723695278,
-0.006957715842872858,
-0.13592909276485443,
-0.08920204639434814,
-0.09430702775716782,
0.10352107882499695,
0.02541329711675644,
0.01974737085402012,
-0.039833661168813705,
0.17427478730678558,
0.08359494060277939,
-0.03463531658053398,
-0.12222621589899063,
0.06654863059520721,
-0.0016485670348629355,
0.032617535442113876,
0.02881730906665325,
0.02470632828772068,
0.051595959812402725,
-0.04482705518603325,
-0.022178227081894875,
0.02629452385008335,
-0.0025798864662647247,
0.009529816918075085,
0.006419707555323839,
-0.19035106897354126,
-0.015978222712874413,
-0.05978287383913994,
0.049019958823919296,
-0.08626580983400345,
-0.043445441871881485,
0.08981242775917053,
0.06649481505155563,
-0.07574772089719772,
-0.0067657362669706345,
-0.011250306852161884,
0.07007921487092972,
0.003048966871574521,
0.06439395993947983,
0.12139144539833069,
-0.015660753473639488,
0.004380745347589254,
0.04942680522799492,
-0.07153112441301346,
0.06739541888237,
0.13226599991321564,
-0.1789792776107788,
-0.017380470409989357,
-0.13775531947612762,
-0.044439107179641724,
0.03797801956534386,
0.04819408431649208,
-0.028686562553048134,
0.12326832860708237,
0.0030446562450379133,
0.006133840885013342,
-0.04210387170314789,
0.01684490777552128,
0.022472180426120758,
0.0016965409740805626,
-0.014794853515923023,
0.1760273426771164,
0.0924944207072258,
-0.0574079267680645,
0.12999488413333893,
0.11454407125711441,
0.030733341351151466,
0.2400905340909958,
-0.0012702507665380836,
-0.029153458774089813,
0.004412034060806036,
-0.004921914543956518,
-0.06877782195806503,
0.23338009417057037,
-0.2722398042678833,
0.02201305329799652,
0.05555930361151695,
-0.019892493262887,
0.037430208176374435,
-0.09364065527915955,
-0.08096104115247726,
-0.05641871690750122,
-0.03273792564868927,
-0.24840863049030304,
0.060468968003988266,
-0.004813906271010637,
0.1168818399310112,
-0.006359854247421026,
-0.14458966255187988,
0.012507098726928234,
0.005278334021568298,
-0.03899219259619713,
0.09661319851875305,
-0.14152950048446655,
-0.21985262632369995,
-0.0888451412320137,
-0.07131245732307434,
-0.06662904471158981,
-0.027309684082865715,
0.026097813621163368,
-0.13762982189655304,
-0.006912102457135916,
0.1113525852560997,
0.21135084331035614,
-0.21194259822368622,
-0.03532121703028679,
-0.1925642341375351,
-0.003605227917432785,
-0.13403289020061493,
-0.05679244175553322,
-0.05316184461116791,
-0.09813278913497925,
0.17511868476867676,
0.047567691653966904,
-0.06583467870950699,
0.0720704197883606,
0.1893775910139084,
0.0646929070353508,
0.06197436526417732,
-0.08078009635210037,
0.1947081834077835,
-0.18873369693756104,
-0.0032725296914577484,
0.0952005609869957,
-0.12015163153409958,
0.023103199899196625,
0.144256129860878,
0.08298062533140182,
0.01819053292274475,
-0.06418830156326294,
-0.008801178075373173,
-0.08530471473932266,
-0.1170315146446228,
-0.19129841029644012,
-0.11400353163480759,
0.13787485659122467,
-0.014513199217617512,
0.039759617298841476,
0.004074459429830313,
-0.028483383357524872,
-0.026885876432061195,
-0.26934394240379333,
0.03944752737879753,
0.015460425056517124,
0.19007450342178345,
-0.016989370808005333,
0.054311785846948624,
-0.02625405043363571,
-0.10088738054037094,
0.0913182869553566,
0.006922412663698196,
-0.018525628373026848,
0.16291935741901398,
0.020182592794299126,
0.023289278149604797,
0.19593532383441925,
0.08236438781023026,
0.12234646081924438,
0.026774292811751366,
0.03955600783228874,
-0.0037312712520360947,
-0.07984363287687302,
0.10297366231679916,
-0.02036677859723568,
0.19754157960414886,
-0.06406770646572113,
-0.050569940358400345,
-0.2159966677427292,
0.1086510717868805,
0.03299945220351219,
0.08725738525390625,
-0.021533479914069176,
-0.013613141141831875,
0.03157920762896538,
0.03219347074627876,
-0.0432480089366436,
0.08999508619308472,
-0.05385585501790047,
-0.12852883338928223,
0.07950867712497711,
-0.009592687711119652,
0.060870274901390076,
0.03147813677787781,
0.08843754976987839,
-0.036729324609041214,
-0.030795378610491753,
-0.04089080169796944,
0.10754192620515823,
-0.12138084322214127,
0.28343722224235535,
-0.01941801980137825,
-0.017529133707284927,
-0.06537989526987076,
-0.04232016205787659,
0.06253698468208313,
0.05601182579994202,
0.18567007780075073,
0.0411607027053833,
-0.07371678948402405,
-0.1837856024503708,
-0.050989631563425064,
0.015762651339173317,
0.17031383514404297,
-0.028815865516662598,
0.0040299613028764725,
0.0030045693274587393,
0.0643400177359581,
-0.008910092525184155,
-0.04622797295451164,
-0.00872009713202715,
-0.06874149292707443,
0.07577607035636902,
0.028325622901320457,
-0.10501333326101303,
0.0764092281460762,
-0.12256216257810593,
-0.13325868546962738,
0.04949985072016716,
-0.026228083297610283,
-0.08538546413183212,
-0.10725939273834229,
-0.14393365383148193,
0.17194823920726776,
-0.059128377586603165,
-0.021710282191634178,
-0.04139548912644386,
0.02957470901310444,
-0.09471810609102249,
-0.059984926134347916,
0.08554638177156448,
-0.01943209208548069,
-0.03659011796116829,
0.021235359832644463,
0.22657041251659393,
0.03939751908183098,
0.07031571120023727,
0.02530144341289997,
0.08109430223703384,
-0.06196770444512367,
-0.1298322230577469,
0.06900041550397873,
-0.15459857881069183,
-0.1490524858236313,
0.10043734312057495,
0.033339232206344604,
0.06524241715669632,
-0.12667223811149597,
-0.007912795059382915,
0.18618406355381012,
0.202214315533638,
-0.05340445041656494,
0.19579239189624786,
0.10182509571313858,
-0.02655031718313694,
-0.19348962604999542,
-0.07411957532167435,
-0.09573959559202194,
-0.013372319750487804,
-0.027267375960946083,
-0.1557588130235672,
0.07167302817106247,
0.040509629994630814,
-0.022386634722352028,
0.11435544490814209,
-0.23524534702301025,
-0.06271400302648544,
0.20968715846538544,
-0.06744992733001709,
0.2567947506904602,
-0.011594387702643871,
-0.03321729972958565,
-0.016558445990085602,
0.00624161958694458,
0.07761946320533752,
0.054041165858507156,
0.051839377731084824,
-0.0015417995164170861,
0.046344444155693054,
0.05286191776394844,
-0.008508243598043919,
0.1840289831161499,
0.03035036288201809,
-0.05427560210227966,
-0.03471359238028526,
-0.19028250873088837,
0.050076842308044434,
0.05323927477002144,
-0.0029057173524051905,
-0.008941035717725754,
-0.0591464526951313,
-0.28032004833221436,
-0.06058885529637337,
-0.12495074421167374,
0.08197169750928879,
-0.041632965207099915,
-0.057343740016222,
-0.05068851634860039,
0.0034606046974658966,
-0.10757871717214584,
-0.017413092777132988,
0.11009597778320312,
-0.06578326970338821,
0.08842293173074722,
-0.1156492829322815,
0.08003511279821396,
0.025101518258452415,
-0.23087887465953827,
0.0024969528894871473,
-0.053848277777433395,
0.07499194145202637,
0.013510053046047688,
-0.052276212722063065,
0.11400061845779419,
0.04016691818833351,
-0.015059887431561947,
0.07455138117074966,
-0.04148651659488678,
-0.011290911585092545,
0.05182911455631256,
-0.19949088990688324,
-0.03670680522918701,
-0.01665981486439705,
-0.15246011316776276,
0.08061899989843369,
0.00966565776616335,
0.08912382274866104,
0.014728769659996033,
-0.028487220406532288,
0.014564860612154007,
-0.04527045786380768,
-0.10242176055908203,
0.06631780415773392,
0.08753294497728348,
0.03856845572590828,
-0.10592202097177505,
0.07196855545043945,
-0.007666274905204773,
-0.13472871482372284,
-0.070844367146492,
0.16396816074848175,
-0.08176029473543167,
-0.058523695915937424,
-0.035664211958646774,
0.01536356657743454,
-0.05891166999936104,
-0.03427964821457863,
-0.052817147225141525,
-0.20580802857875824,
0.02296259067952633,
0.1756434291601181,
0.11423919349908829,
-0.0018322821706533432,
-0.1148550808429718,
0.001504837186075747,
0.0006292599136941135,
0.022988885641098022,
0.04832680523395538,
-0.07040427625179291,
-0.03781347721815109,
0.12173118442296982,
0.02966105192899704,
0.06528263539075851,
-0.07439873367547989,
-0.1014258861541748,
-0.15639351308345795,
0.09274012595415115,
0.022870728746056557,
-0.04012517258524895,
-0.02137848548591137,
-0.016581833362579346,
-0.008164064027369022,
-0.03417125716805458,
-0.02684236504137516,
-0.03460869938135147,
-0.0922638401389122,
0.0818224772810936,
0.03648698702454567,
0.05321777984499931,
-0.01874331571161747,
0.006254127714782953,
0.0759749561548233,
-0.023266375064849854,
0.13999386131763458,
0.19686229526996613,
-0.09118223935365677,
0.10253044217824936,
-0.11360391229391098,
-0.005259087309241295,
0.0385298915207386,
-0.001296738744713366,
-0.02928207255899906,
0.005912467837333679,
-0.00944026280194521,
-0.02411280758678913,
0.006586436647921801,
0.09251061081886292,
0.01474232692271471,
-0.06294658780097961,
0.02708233892917633,
0.03975668177008629,
-0.23339170217514038,
0.000716951850336045,
-0.0634632334113121,
0.08630799502134323,
0.044719208031892776,
0.11092139035463333,
0.029959360137581825,
0.097026526927948,
0.02307308465242386,
0.009252536110579967,
-0.030829623341560364,
-0.1132243275642395,
-0.03359498456120491,
-0.005170233082026243,
0.060372550040483475,
-0.041298795491456985,
0.20418722927570343,
0.06735196709632874,
0.03864594176411629,
0.0315295048058033,
0.1638958901166916,
-0.07288289815187454,
0.04239322245121002,
0.20682060718536377,
0.0713522732257843,
-0.015939848497509956,
-0.032790809869766235,
0.004336249083280563,
-0.02426091581583023,
0.1853058934211731,
0.0787394791841507,
0.1312253326177597,
0.21831905841827393,
0.021068431437015533,
0.07649321854114532,
0.04703734442591667,
-0.04437753185629845,
-0.07401061803102493,
0.16469812393188477,
0.019643669947981834,
-0.03209804743528366,
0.22190611064434052,
0.0827108845114708,
-0.025106335058808327,
0.09492665529251099,
-0.029429683461785316,
-0.0739864632487297,
-0.12008123844861984,
-0.12298780679702759,
0.03135455772280693,
-0.014296706765890121,
-0.02209966443479061,
-0.09591492265462875,
-0.011029220186173916,
0.17383895814418793,
0.06633847206830978,
-0.021252015605568886,
-0.04073715582489967,
-0.0640232190489769,
-0.09621608257293701,
0.14193430542945862,
-0.034291595220565796,
0.06917029619216919,
-0.07720480114221573,
-0.06243932247161865,
-0.05333844944834709,
-0.045827046036720276,
0.008700287900865078,
-0.011175722815096378,
-0.09380948543548584,
-0.09573883563280106,
-0.20992429554462433,
-0.09357866644859314,
0.0032437408808618784,
0.02573622576892376,
0.04057632386684418,
0.16024266183376312,
0.05159405991435051,
-0.013136574067175388,
-0.017010854557156563,
0.13466186821460724,
0.0438971109688282,
-0.11116588860750198,
-0.022701896727085114,
0.1418018490076065,
-0.09654334932565689,
0.09032461047172546,
-0.03947578743100166,
0.019956376403570175,
-0.024654416367411613,
0.2212703377008438,
0.27310967445373535,
0.013388645835220814,
0.018121400848031044,
0.03720363602042198,
0.03143814578652382,
0.15236273407936096,
0.07383204251527786,
0.02668502926826477,
0.23173117637634277,
-0.033888448029756546,
0.028953135013580322,
-0.17056281864643097,
-0.008167666383087635,
-0.05974593758583069,
0.11065938323736191,
0.12106261402368546,
-0.11625828593969345,
-0.105011485517025,
0.19374747574329376,
-0.185593843460083,
0.17733894288539886,
-0.09078248590230942,
-0.16583849489688873,
-0.11546684056520462,
-0.013322343118488789,
0.04943598806858063,
0.1135810911655426,
0.0807109996676445,
-0.09064453095197678,
-0.14784111082553864,
-0.02461598627269268,
0.06881017982959747,
-0.16959504783153534,
-0.13859032094478607,
0.06808578968048096,
-0.04927487671375275,
0.08359096199274063,
-0.003199564991518855,
0.11931931972503662,
0.047464046627283096,
0.11110589653253555,
0.06885301321744919,
-0.059626225382089615,
0.056692466139793396,
-0.040322527289390564,
-0.057360272854566574,
0.022537127137184143,
0.010731746442615986,
-0.045215874910354614,
0.05866113677620888,
-0.0953088030219078,
0.06301386654376984,
0.08972669392824173,
-0.10197558254003525,
0.06477820128202438,
0.14770084619522095,
-0.12146434932947159,
-0.011873575858771801,
0.10634484887123108,
0.016191154718399048,
-0.09092428535223007,
-0.06763330101966858,
-0.04023827984929085,
0.03266846016049385,
-0.12343493849039078,
-0.08507751673460007,
-0.07075997442007065,
-0.07239314168691635,
-0.044050972908735275,
-0.09385629743337631,
-0.03406060114502907,
-0.03855740651488304,
-0.07704507559537888,
0.09376388788223267,
-0.051036763936281204,
0.01617167331278324,
-0.04556584730744362,
-0.032196637243032455,
0.025646997615695,
0.047098565846681595,
0.09995488077402115,
0.08353474736213684,
-0.05905599519610405,
-0.03238213434815407
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# hubert_3
This model is a fine-tuned version of [rinna/japanese-hubert-base](https://huggingface.co/rinna/japanese-hubert-base) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0584
- Wer: 0.1836
- Cer: 0.0589
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.00027
- train_batch_size: 32
- eval_batch_size: 32
- seed: 4
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 1000
- num_epochs: 40
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|
| 11.7882 | 1.0 | 60 | 9.4723 | 0.9944 | 0.9975 |
| 6.3705 | 2.0 | 120 | 5.9403 | 0.9944 | 0.9975 |
| 4.7536 | 3.0 | 180 | 4.5186 | 0.9944 | 0.9975 |
| 3.5501 | 4.0 | 240 | 3.3998 | 0.9944 | 0.9975 |
| 2.9838 | 5.0 | 300 | 2.9474 | 0.9944 | 0.9975 |
| 2.4703 | 6.0 | 360 | 2.2713 | 0.9944 | 0.9975 |
| 1.6962 | 7.0 | 420 | 1.6251 | 1.0 | 0.7253 |
| 1.228 | 8.0 | 480 | 1.1017 | 0.9547 | 0.5135 |
| 0.8611 | 9.0 | 540 | 0.8005 | 0.7756 | 0.4432 |
| 0.8303 | 10.0 | 600 | 0.7502 | 0.7823 | 0.4729 |
| 0.8582 | 11.0 | 660 | 0.7064 | 0.7771 | 0.4860 |
| 0.6971 | 12.0 | 720 | 0.6950 | 0.7878 | 0.4560 |
| 0.6735 | 13.0 | 780 | 0.7334 | 0.7545 | 0.4045 |
| 0.643 | 14.0 | 840 | 0.5926 | 0.7637 | 0.4227 |
| 0.6702 | 15.0 | 900 | 0.5447 | 0.6832 | 0.3137 |
| 0.909 | 16.0 | 960 | 0.5550 | 0.6955 | 0.3118 |
| 0.5869 | 17.0 | 1020 | 0.5446 | 0.7760 | 0.3850 |
| 1.2649 | 18.0 | 1080 | 0.4579 | 0.7066 | 0.3267 |
| 0.475 | 19.0 | 1140 | 0.4564 | 0.6187 | 0.2544 |
| 0.4629 | 20.0 | 1200 | 0.4068 | 0.6024 | 0.2213 |
| 0.4432 | 21.0 | 1260 | 0.3811 | 0.5987 | 0.2510 |
| 0.5371 | 22.0 | 1320 | 0.3753 | 0.5679 | 0.1950 |
| 0.3914 | 23.0 | 1380 | 0.3413 | 0.5879 | 0.2578 |
| 0.389 | 24.0 | 1440 | 0.3278 | 0.5130 | 0.1947 |
| 0.349 | 25.0 | 1500 | 0.2986 | 0.4811 | 0.1626 |
| 0.3343 | 26.0 | 1560 | 0.3187 | 0.4607 | 0.1637 |
| 0.2964 | 27.0 | 1620 | 0.2471 | 0.4236 | 0.1679 |
| 0.2935 | 28.0 | 1680 | 0.2539 | 0.4206 | 0.1476 |
| 0.2287 | 29.0 | 1740 | 0.2014 | 0.3576 | 0.1343 |
| 0.2223 | 30.0 | 1800 | 0.1745 | 0.3309 | 0.1146 |
| 0.2107 | 31.0 | 1860 | 0.1532 | 0.3049 | 0.1001 |
| 0.2042 | 32.0 | 1920 | 0.1219 | 0.2893 | 0.1260 |
| 0.1734 | 33.0 | 1980 | 0.1088 | 0.2596 | 0.0942 |
| 0.1543 | 34.0 | 2040 | 0.0889 | 0.2200 | 0.0723 |
| 0.1954 | 35.0 | 2100 | 0.0943 | 0.2285 | 0.0745 |
| 0.125 | 36.0 | 2160 | 0.0749 | 0.1996 | 0.0652 |
| 0.132 | 37.0 | 2220 | 0.0734 | 0.2029 | 0.0653 |
| 0.1136 | 38.0 | 2280 | 0.0656 | 0.1951 | 0.0624 |
| 0.1081 | 39.0 | 2340 | 0.0606 | 0.1881 | 0.0599 |
| 0.1512 | 40.0 | 2400 | 0.0584 | 0.1836 | 0.0589 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.14.6
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["wer"], "base_model": "rinna/japanese-hubert-base", "model-index": [{"name": "hubert_3", "results": []}]} | automatic-speech-recognition | tndklab/hubert_3 | [
"transformers",
"safetensors",
"hubert",
"automatic-speech-recognition",
"generated_from_trainer",
"base_model:rinna/japanese-hubert-base",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-08T08:08:11+00:00 | [] | [] | TAGS
#transformers #safetensors #hubert #automatic-speech-recognition #generated_from_trainer #base_model-rinna/japanese-hubert-base #license-apache-2.0 #endpoints_compatible #region-us
| hubert\_3
=========
This model is a fine-tuned version of rinna/japanese-hubert-base on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0584
* Wer: 0.1836
* Cer: 0.0589
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.00027
* train\_batch\_size: 32
* eval\_batch\_size: 32
* seed: 4
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 1000
* num\_epochs: 40
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.14.6
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.00027\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 40",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #safetensors #hubert #automatic-speech-recognition #generated_from_trainer #base_model-rinna/japanese-hubert-base #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.00027\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 40",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
66,
116,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #hubert #automatic-speech-recognition #generated_from_trainer #base_model-rinna/japanese-hubert-base #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.00027\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 40### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
-0.09173012524843216,
0.09753581881523132,
-0.002268433105200529,
0.08795773237943649,
0.11690832674503326,
-0.006308482028543949,
0.13874924182891846,
0.14967137575149536,
-0.06223523989319801,
0.0717792958021164,
0.09587037563323975,
0.11641683429479599,
0.028505124151706696,
0.14004866778850555,
-0.061861928552389145,
-0.2815195322036743,
0.04142084717750549,
0.009475534781813622,
-0.021610809490084648,
0.11318209022283554,
0.08808096498250961,
-0.12036760151386261,
0.08195795118808746,
0.0001943513343576342,
-0.11887261271476746,
0.023263178765773773,
0.022525135427713394,
-0.09988653659820557,
0.12679792940616608,
0.0156675074249506,
0.07208100706338882,
0.02776748314499855,
0.0942627564072609,
-0.23148901760578156,
0.007013789843767881,
0.03721413388848305,
0.03435900807380676,
0.057064302265644073,
0.035803161561489105,
0.0013926717219874263,
0.13732589781284332,
-0.09731332212686539,
0.05002696439623833,
0.03928503021597862,
-0.10265222936868668,
-0.24431492388248444,
-0.06670405715703964,
0.031392816454172134,
0.10268692672252655,
0.09802559018135071,
-0.023447951301932335,
0.1005905345082283,
-0.052827514708042145,
0.09347347915172577,
0.24921731650829315,
-0.3225368559360504,
-0.05030716210603714,
-0.026850305497646332,
0.04947593808174133,
0.07230194658041,
-0.11016204953193665,
-0.006582057103514671,
0.05545319989323616,
0.021369827911257744,
0.09869102388620377,
-0.03928906470537186,
-0.08691676706075668,
0.003520981175825,
-0.12803605198860168,
-0.009522308595478535,
0.15134941041469574,
0.04732118174433708,
-0.04922305792570114,
-0.0988488420844078,
-0.05544499307870865,
-0.12655219435691833,
-0.06156229227781296,
-0.02607361413538456,
0.04703068733215332,
-0.04079762473702431,
-0.05788389593362808,
-0.026762064546346664,
-0.0743294507265091,
-0.09070565551519394,
-0.027144640684127808,
0.19488981366157532,
0.04721546173095703,
-0.0057227639481425285,
-0.012746556662023067,
0.06743088364601135,
-0.0134345144033432,
-0.15095452964305878,
-0.012431556358933449,
0.0354500450193882,
0.00004407810774864629,
-0.0031188654247671366,
-0.031319648027420044,
-0.0010736922267824411,
0.04708646610379219,
0.14758428931236267,
-0.08994968980550766,
0.07454955577850342,
-0.016625545918941498,
0.008427969180047512,
-0.11374900490045547,
0.19003106653690338,
-0.038991544395685196,
-0.05752455070614815,
0.008760372176766396,
0.09287495166063309,
0.04999830573797226,
-0.011718463152647018,
-0.0856480821967125,
0.0033941015135496855,
0.09683375805616379,
0.045919470489025116,
-0.07764066010713577,
0.07141678780317307,
-0.02625289373099804,
0.008434263058006763,
0.00783713348209858,
-0.1257794350385666,
0.025419699028134346,
0.02287869155406952,
-0.0748286247253418,
-0.026715628802776337,
0.004710938781499863,
0.008290283381938934,
-0.011497050523757935,
0.06835852563381195,
-0.06428215652704239,
0.014870373532176018,
-0.053224436938762665,
-0.10002423822879791,
0.0075582959689199924,
-0.08814346790313721,
0.024131761863827705,
-0.10687275975942612,
-0.14276330173015594,
-0.010983435437083244,
0.029419517144560814,
-0.04168085753917694,
-0.007551536429673433,
-0.09470120817422867,
-0.09496590495109558,
0.02922406792640686,
-0.026666300371289253,
0.054633744060993195,
-0.0801088958978653,
0.09954870492219925,
0.07294397801160812,
0.07949765026569366,
-0.01832636259496212,
0.03808273375034332,
-0.09908255934715271,
0.017011070623993874,
-0.1747879981994629,
0.051769860088825226,
-0.07437849044799805,
0.03257453814148903,
-0.10904520004987717,
-0.08209055662155151,
0.017029523849487305,
0.026016082614660263,
0.060774415731430054,
0.12691140174865723,
-0.16747450828552246,
-0.07883498072624207,
0.16872289776802063,
-0.11083705723285675,
-0.10524352639913559,
0.11685076355934143,
-0.03468942269682884,
0.051718272268772125,
0.06783930957317352,
0.24990293383598328,
0.027482395991683006,
-0.13420774042606354,
-0.005767857190221548,
-0.02196403592824936,
0.0546327568590641,
0.007524170912802219,
0.06044254079461098,
-0.009222651831805706,
-0.00023110203619580716,
0.03190312534570694,
-0.04916176572442055,
0.028916236013174057,
-0.07458914071321487,
-0.09135790914297104,
-0.03792129456996918,
-0.1007719337940216,
0.008929913863539696,
0.041723694652318954,
0.0564374141395092,
-0.12008356302976608,
-0.09216495603322983,
0.015666579827666283,
0.10263463854789734,
-0.10394590348005295,
0.03978688269853592,
-0.10640110820531845,
0.05472877249121666,
-0.004226681776344776,
-0.013389642350375652,
-0.1508830189704895,
0.024222243577241898,
0.039624348282814026,
-0.022664230316877365,
0.04080881550908089,
-0.05560138076543808,
0.07775872945785522,
0.05736104026436806,
-0.06509434431791306,
-0.05885343998670578,
-0.0068183839321136475,
0.01472111139446497,
-0.06813449412584305,
-0.20360055565834045,
-0.009748599492013454,
-0.03640567883849144,
0.10073123872280121,
-0.1817445605993271,
0.0170403104275465,
-0.01055268757045269,
0.08644984662532806,
0.03660668805241585,
-0.010267344303429127,
-0.010661916807293892,
0.06709841638803482,
-0.028315259143710136,
-0.05164510756731033,
0.04298640042543411,
-0.0046584890224039555,
-0.08996172249317169,
0.01555861346423626,
-0.15070770680904388,
0.13183431327342987,
0.13790293037891388,
-0.014171818271279335,
-0.0687435194849968,
0.014432831667363644,
-0.03196124732494354,
-0.03753609582781792,
-0.0264663714915514,
0.015720266848802567,
0.16653023660182953,
-0.02932858280837536,
0.1343267410993576,
-0.0897044986486435,
-0.0024536987766623497,
0.03309502825140953,
-0.03999408334493637,
-0.007556812837719917,
0.12335296720266342,
0.05433278530836105,
-0.0683242604136467,
0.12305499613285065,
0.11417744308710098,
-0.09670346230268478,
0.14502157270908356,
-0.05843406915664673,
-0.07403652369976044,
-0.022630959749221802,
0.018070247024297714,
0.00600780313834548,
0.12222142517566681,
-0.13003472983837128,
-0.023592844605445862,
0.008283120580017567,
0.010387578047811985,
0.009763860143721104,
-0.2031649351119995,
-0.015434821136295795,
0.014188098721206188,
-0.09932393580675125,
-0.0020908440928906202,
0.016525529325008392,
-0.007388044614344835,
0.11031235754489899,
-0.017890015617012978,
-0.11046471446752548,
0.001460065133869648,
-0.014588464051485062,
-0.06747505813837051,
0.17794671654701233,
-0.10573484748601913,
-0.1700480580329895,
-0.10392111539840698,
-0.05903562903404236,
-0.04371786490082741,
0.023586057126522064,
0.07546375691890717,
-0.1107940524816513,
-0.045271214097738266,
-0.11370741575956345,
0.016087118536233902,
0.019144371151924133,
0.033305395394563675,
0.007780811283737421,
0.010221940465271473,
0.06895462423563004,
-0.11052653193473816,
-0.012650557793676853,
-0.04698512703180313,
-0.039614155888557434,
0.017997050657868385,
0.036026231944561005,
0.11247999966144562,
0.1283041089773178,
0.0013678977265954018,
0.019512353464961052,
-0.04041753336787224,
0.18506643176078796,
-0.07628168165683746,
-0.020741188898682594,
0.14104661345481873,
-0.0071521420031785965,
0.019617173820734024,
0.15960446000099182,
0.0391821563243866,
-0.10917816311120987,
0.010591194033622742,
0.01848209835588932,
-0.01929337903857231,
-0.21057388186454773,
-0.04192965850234032,
-0.03868960961699486,
0.018793068826198578,
0.08516384661197662,
0.039009541273117065,
0.02484137937426567,
0.017740806564688683,
0.02763432264328003,
0.004385386593639851,
0.018917266279459,
0.06830818206071854,
0.1265171468257904,
0.03211795911192894,
0.10871356725692749,
-0.04144623875617981,
-0.05821407213807106,
0.026454703882336617,
-0.001465276931412518,
0.19705265760421753,
0.022173400968313217,
0.1351967751979828,
0.035826727747917175,
0.16499100625514984,
0.01451431680470705,
0.04978492856025696,
0.011519786901772022,
-0.02371995337307453,
-0.011118464171886444,
-0.0657329112291336,
-0.03228626772761345,
0.061467066407203674,
-0.027910951524972916,
0.05206309258937836,
-0.11621400713920593,
0.0170576274394989,
0.052141569554805756,
0.2871558368206024,
0.05527804419398308,
-0.2948623299598694,
-0.07769196480512619,
0.019142530858516693,
-0.07517949491739273,
-0.013473100028932095,
0.07551316916942596,
0.13227719068527222,
-0.05983365327119827,
0.05453535169363022,
-0.04569857940077782,
0.0716869980096817,
-0.044115982949733734,
0.038101278245449066,
0.04113573580980301,
0.07164500653743744,
0.006625987123697996,
0.0381261482834816,
-0.27802664041519165,
0.288179486989975,
0.014866815879940987,
0.09326330572366714,
-0.04076603800058365,
0.003944884520024061,
0.04464956745505333,
0.01717643439769745,
0.1287069171667099,
-0.041258424520492554,
-0.1363379955291748,
-0.17893770337104797,
-0.06725917756557465,
0.038635194301605225,
0.1342097520828247,
0.0025654418859630823,
0.10420241951942444,
-0.03664867579936981,
-0.02528414875268936,
0.06043391674757004,
-0.08680287003517151,
-0.11252504587173462,
-0.07980164885520935,
-0.029510242864489555,
0.09242746233940125,
0.01306084357202053,
-0.06451861560344696,
-0.08345769345760345,
-0.08493857830762863,
0.10327062010765076,
-0.04263439029455185,
-0.013434123247861862,
-0.09825600683689117,
0.012213967740535736,
0.12417297810316086,
-0.08231674134731293,
0.052165254950523376,
0.019495980814099312,
0.06465613096952438,
0.03624008968472481,
-0.051102008670568466,
0.10873910784721375,
-0.07540406286716461,
-0.1770329475402832,
-0.04781349003314972,
0.15370988845825195,
0.02200717106461525,
0.05239177122712135,
0.0010610577883198857,
0.02171286754310131,
0.005903421901166439,
-0.0653776153922081,
0.01964952051639557,
0.02962729148566723,
0.03162752091884613,
0.02799364924430847,
-0.07770542055368423,
-0.04299350082874298,
-0.10544320195913315,
-0.03937257453799248,
0.14804090559482574,
0.2834817171096802,
-0.0764881819486618,
0.05183029547333717,
0.07867469638586044,
-0.049764011055231094,
-0.18896804749965668,
-0.009499862790107727,
0.03338634595274925,
0.017521461471915245,
0.0038896393962204456,
-0.15072591602802277,
0.0765661671757698,
0.06579651683568954,
-0.029943693429231644,
0.08856473118066788,
-0.2821694016456604,
-0.14399811625480652,
0.14051909744739532,
0.12209795415401459,
0.09281822293996811,
-0.15339522063732147,
-0.03820895403623581,
-0.014706489630043507,
-0.055251382291316986,
0.0645846426486969,
-0.06658072024583817,
0.1330498605966568,
-0.013450204394757748,
0.04966289550065994,
0.016433222219347954,
-0.04023461788892746,
0.12099653482437134,
0.000038160727854119614,
0.08210732787847519,
-0.039089035242795944,
-0.009870065376162529,
-0.011194546706974506,
-0.04813765361905098,
0.06609944999217987,
-0.1094643622636795,
0.03951897472143173,
-0.04490438103675842,
-0.03469699248671532,
-0.0738678053021431,
0.030607281252741814,
-0.008649339899420738,
-0.05645357817411423,
-0.04175305366516113,
0.03489479795098305,
0.04318990930914879,
-0.0011553462827578187,
0.15516532957553864,
-0.03394355624914169,
0.12306391447782516,
0.13710035383701324,
0.09723059833049774,
-0.048827335238456726,
-0.005981668829917908,
0.013776122592389584,
-0.03733835741877556,
0.07230759412050247,
-0.13580767810344696,
0.04710191488265991,
0.11098539084196091,
0.021499676629900932,
0.15334239602088928,
0.04808046296238899,
-0.04257708415389061,
0.022653184831142426,
0.0647047758102417,
-0.14594687521457672,
-0.11550014466047287,
-0.006759272422641516,
-0.04843197390437126,
-0.06245528906583786,
0.06371572613716125,
0.12381541728973389,
-0.07819782942533493,
-0.0005059727118350565,
-0.01876121759414673,
0.02629268355667591,
-0.0492081418633461,
0.1967565417289734,
0.044958654791116714,
0.0357600562274456,
-0.10407272726297379,
0.09960625320672989,
0.024784140288829803,
-0.11093418300151825,
0.052275773137807846,
0.08235488831996918,
-0.08657857775688171,
-0.03547720983624458,
0.022550810128450394,
0.12887628376483917,
0.009030940011143684,
-0.07853376865386963,
-0.15019869804382324,
-0.12461800128221512,
0.05505309998989105,
0.18467535078525543,
0.07552381604909897,
0.013057252392172813,
-0.04122151806950569,
0.028578048571944237,
-0.11522581428289413,
0.09629198908805847,
0.05248928815126419,
0.05026855692267418,
-0.14534889161586761,
0.10218799114227295,
0.021936966106295586,
0.018909551203250885,
-0.027450112625956535,
0.022279372438788414,
-0.1224793791770935,
0.020570894703269005,
-0.12058836966753006,
0.0086763771250844,
-0.05358406901359558,
-0.0005979237030260265,
0.012145508080720901,
-0.0748271569609642,
-0.0767352506518364,
0.024857226759195328,
-0.09799793362617493,
-0.014635005965828896,
0.01675512082874775,
0.06837843358516693,
-0.1236090213060379,
-0.033721357583999634,
0.02995295636355877,
-0.08535094559192657,
0.07687114924192429,
0.07976188510656357,
-0.024186940863728523,
0.08459411561489105,
-0.12377522140741348,
-0.006831163540482521,
0.08598519116640091,
0.007265097927302122,
0.029682707041502,
-0.12548978626728058,
-0.014905132353305817,
0.014671355485916138,
0.061280325055122375,
0.005490925628691912,
0.09432734549045563,
-0.11191783845424652,
0.008676229976117611,
-0.05690357834100723,
-0.061498381197452545,
-0.05929345265030861,
0.015987643972039223,
0.1241886168718338,
0.004487868398427963,
0.1804867684841156,
-0.11607658863067627,
0.020748723298311234,
-0.1815391480922699,
0.009584914892911911,
-0.02703392691910267,
-0.11739905178546906,
-0.12340468168258667,
-0.04194317013025284,
0.07511329650878906,
-0.06181667745113373,
0.12672561407089233,
-0.015050411224365234,
0.04766678437590599,
0.033060166984796524,
-0.10607031732797623,
0.004489517770707607,
0.03951707109808922,
0.24505417048931122,
0.04260401800274849,
-0.03622348979115486,
0.05322723463177681,
0.008970922790467739,
0.0927039235830307,
0.12098120898008347,
0.14769044518470764,
0.19807802140712738,
0.010931980796158314,
0.13466013967990875,
0.07373592257499695,
-0.06322270631790161,
-0.13031192123889923,
0.07881354540586472,
-0.0564148984849453,
0.08894024789333344,
-0.012076811864972115,
0.2338039129972458,
0.12910789251327515,
-0.15768444538116455,
0.04607858136296272,
-0.03565528988838196,
-0.08109916746616364,
-0.1362142711877823,
-0.043475694954395294,
-0.1109418272972107,
-0.17462298274040222,
0.028168002143502235,
-0.11045163124799728,
0.05451599508523941,
0.05498727411031723,
0.020937178283929825,
-0.0011952047934755683,
0.16685377061367035,
0.0011370168067514896,
0.012429732829332352,
0.09064298123121262,
0.01429649069905281,
-0.05530838668346405,
-0.05643376708030701,
-0.08916272968053818,
0.0184713713824749,
-0.030545935034751892,
0.014913254417479038,
-0.02873213402926922,
-0.07936207205057144,
0.04987114667892456,
-0.03282861411571503,
-0.09000902622938156,
0.01910609006881714,
0.020875193178653717,
0.07849105447530746,
0.06474315375089645,
0.04683881625533104,
-0.04012449458241463,
0.0133198332041502,
0.24334761500358582,
-0.10131824761629105,
-0.09647179394960403,
-0.11021372675895691,
0.2704285979270935,
0.03846428543329239,
0.01381708960980177,
0.0012350364122539759,
-0.0681174024939537,
-0.023836759850382805,
0.22938239574432373,
0.19141457974910736,
-0.05657661333680153,
0.0011859415099024773,
-0.03271492198109627,
-0.0012880782596766949,
-0.058337971568107605,
0.07671643048524857,
0.12846583127975464,
0.06721188127994537,
-0.04000505805015564,
-0.05823863670229912,
-0.04485905170440674,
-0.02906004525721073,
-0.049123141914606094,
0.08538230508565903,
0.006623160094022751,
-0.029543431475758553,
-0.04132581874728203,
0.06208541989326477,
-0.07958737760782242,
-0.13136255741119385,
0.027719371020793915,
-0.1999620795249939,
-0.14168809354305267,
0.0034088182728737593,
0.0909871906042099,
0.03565341606736183,
0.028242923319339752,
-0.02159699611365795,
0.005353692453354597,
0.0606829859316349,
-0.009067302569746971,
-0.07043687254190445,
-0.0789424479007721,
0.05893644317984581,
-0.11554358899593353,
0.21370400488376617,
-0.019741812720894814,
0.05027848854660988,
0.10033243894577026,
0.06605987995862961,
-0.0688798576593399,
0.12654335796833038,
0.05647124722599983,
-0.12622742354869843,
0.035646870732307434,
0.14371395111083984,
-0.04410228505730629,
0.15121431648731232,
0.05499360337853432,
-0.11926031857728958,
0.024510838091373444,
-0.029469378292560577,
-0.08862017095088959,
-0.07036270946264267,
-0.030973736196756363,
-0.0647972822189331,
0.13452869653701782,
0.17541199922561646,
-0.0438782274723053,
0.012413428165018559,
-0.044474173337221146,
0.04572923481464386,
0.05821070447564125,
0.0365120992064476,
-0.03576777130365372,
-0.25743749737739563,
0.002920084400102496,
0.02254822850227356,
-0.013578027486801147,
-0.24242760241031647,
-0.09077506512403488,
0.0002413891052128747,
-0.0440206341445446,
-0.09370031952857971,
0.07917453348636627,
0.09535598009824753,
0.036743246018886566,
-0.039810895919799805,
-0.11688435077667236,
-0.014421962201595306,
0.181536465883255,
-0.16679702699184418,
-0.09423426538705826
] |
null | null | peft |
- PEFT 0.6.0
| {"library_name": "peft", "base_model": "upstage/SOLAR-10.7B-v1.0"} | null | RefalMachine/solar_10.7_saiga_lora | [
"peft",
"base_model:upstage/SOLAR-10.7B-v1.0",
"region:us"
] | 2024-02-08T08:08:35+00:00 | [] | [] | TAGS
#peft #base_model-upstage/SOLAR-10.7B-v1.0 #region-us
|
- PEFT 0.6.0
| [] | [
"TAGS\n#peft #base_model-upstage/SOLAR-10.7B-v1.0 #region-us \n"
] | [
26
] | [
"passage: TAGS\n#peft #base_model-upstage/SOLAR-10.7B-v1.0 #region-us \n"
] | [
-0.06483274698257446,
-0.06749211251735687,
-0.005817529745399952,
-0.07230091840028763,
0.06281585991382599,
0.03871985152363777,
0.10206230729818344,
0.02653687447309494,
0.18447671830654144,
0.07009454816579819,
0.1486044079065323,
0.07906594127416611,
0.02179965190589428,
0.19211606681346893,
0.05834417790174484,
-0.30529820919036865,
0.14255818724632263,
0.00707847299054265,
0.14578086137771606,
0.039379071444272995,
-0.014302029274404049,
-0.07158858329057693,
0.03439493477344513,
-0.11016834527254105,
-0.10250727832317352,
0.058492373675107956,
-0.009732650592923164,
-0.028748972341418266,
0.11960484087467194,
-0.014602717943489552,
0.07658187299966812,
0.02482982911169529,
0.0006115915603004396,
-0.11360903084278107,
-0.007882757112383842,
-0.018895598128437996,
-0.1099146157503128,
0.10786719620227814,
0.05289127305150032,
0.10603280365467072,
0.20080801844596863,
0.07315001636743546,
0.001950033474713564,
0.011908839456737041,
-0.18557344377040863,
-0.11819825321435928,
-0.06749377399682999,
0.07366213202476501,
0.0774373710155487,
-0.02253606729209423,
0.033627454191446304,
0.20790763199329376,
-0.04788445681333542,
0.002632509684190154,
0.11991661041975021,
-0.23014917969703674,
0.04272707179188728,
0.20380623638629913,
0.0035462479572743177,
0.15013152360916138,
0.0009587286040186882,
0.03895283862948418,
0.13151036202907562,
-0.0471380315721035,
-0.06556395441293716,
-0.0002823015092872083,
0.06817204505205154,
0.12053453177213669,
-0.09965858608484268,
-0.11271687597036362,
0.3803555369377136,
0.09113167226314545,
-0.022949563339352608,
0.06369981169700623,
-0.13314220309257507,
-0.03381814435124397,
-0.007953598164021969,
0.0272854994982481,
-0.0056738960556685925,
0.09244609624147415,
0.18195737898349762,
-0.025237180292606354,
-0.10357845574617386,
-0.055167101323604584,
-0.09988439828157425,
0.3390737473964691,
-0.0033337168861180544,
0.0994938537478447,
-0.0963561087846756,
0.13155919313430786,
-0.05021853744983673,
-0.08173144608736038,
0.06336979568004608,
-0.11213140189647675,
0.01767946034669876,
0.08944899588823318,
-0.07518801093101501,
0.12460453063249588,
0.015556410886347294,
0.2184406816959381,
0.058668989688158035,
0.009117296896874905,
0.09040901064872742,
0.11996625363826752,
0.0724344402551651,
0.047938983887434006,
-0.04155386611819267,
0.007099990267306566,
0.03570336475968361,
-0.07167874276638031,
0.13185320794582367,
-0.04989172890782356,
-0.07943924516439438,
-0.0642053559422493,
-0.05006573349237442,
0.018955819308757782,
0.0716150626540184,
-0.10858608782291412,
-0.09640861302614212,
-0.006984218023717403,
0.29435476660728455,
0.014015342108905315,
-0.08878901600837708,
0.015143615193665028,
0.030437596142292023,
0.06634815037250519,
-0.024818766862154007,
-0.00418038759380579,
0.16485993564128876,
-0.06059996411204338,
-0.1422342211008072,
-0.06108125299215317,
-0.06782624870538712,
0.02591945417225361,
0.03048744983971119,
-0.005242772866040468,
0.10476576536893845,
-0.1237572431564331,
-0.11530614644289017,
0.04055888578295708,
-0.006096868775784969,
-0.16439594328403473,
-0.08320944756269455,
0.12448593229055405,
-0.05381425842642784,
0.009148171171545982,
-0.06899908185005188,
0.04141974076628685,
-0.05403544753789902,
0.017117710784077644,
0.03879697620868683,
0.07743482291698456,
-0.18517497181892395,
-0.01157817430794239,
-0.09569110721349716,
0.07674764096736908,
-0.23120298981666565,
-0.047632161527872086,
-0.06543441861867905,
0.14880521595478058,
-0.033629994839429855,
-0.009099035523831844,
-0.20927295088768005,
0.014804627746343613,
0.051234111189842224,
0.1632211059331894,
-0.14187650382518768,
-0.010852549225091934,
0.0029395699966698885,
-0.15084140002727509,
-0.07488890737295151,
-0.10489184409379959,
0.055524248629808426,
0.050692006945610046,
-0.07580892741680145,
0.18403473496437073,
-0.1485700011253357,
-0.03040197119116783,
0.038277339190244675,
0.11744364351034164,
-0.04530394822359085,
-0.21223099529743195,
0.12313429266214371,
-0.02247006446123123,
-0.3138956129550934,
-0.003323838347569108,
-0.03300749883055687,
0.07275338470935822,
-0.0801982432603836,
-0.004696969874203205,
-0.045963868498802185,
-0.0063640098087489605,
-0.07768693566322327,
0.0426015704870224,
0.06649910658597946,
-0.014043412171304226,
0.03632098808884621,
0.13077154755592346,
0.11468655616044998,
0.05582307651638985,
0.04387601092457771,
0.004500130657106638,
0.11664556711912155,
-0.17891724407672882,
-0.023139705881476402,
-0.11551718413829803,
0.019753701984882355,
0.057633377611637115,
0.08009167015552521,
-0.06106870248913765,
0.11157741397619247,
0.04049397259950638,
-0.07731398195028305,
-0.017393101006746292,
-0.04126708582043648,
-0.042877305299043655,
0.042224232107400894,
-0.03754101321101189,
-0.11061976104974747,
-0.11822160333395004,
-0.0624898336827755,
0.017698252573609352,
-0.132603719830513,
-0.010193025693297386,
-0.03133384883403778,
0.06261052191257477,
-0.02993715927004814,
-0.0012084722984582186,
0.048918016254901886,
-0.0032530981115996838,
0.0036436442751437426,
-0.02767051011323929,
0.03720097616314888,
-0.009951284155249596,
-0.053667232394218445,
0.045382045209407806,
-0.036206915974617004,
0.059967041015625,
0.09589726477861404,
-0.1444898098707199,
0.006986146327108145,
0.011353706941008568,
-0.012191057205200195,
-0.01850694976747036,
0.13853469491004944,
-0.0895809605717659,
-0.21748453378677368,
-0.10768850892782211,
-0.01167730987071991,
-0.036750420928001404,
-0.004441575147211552,
0.0014887625584378839,
0.0003375336527824402,
-0.05948392674326897,
0.09686169028282166,
0.2581774592399597,
-0.03703383728861809,
0.06796817481517792,
0.24459956586360931,
0.11505602300167084,
0.0072680930607020855,
-0.05849694833159447,
-0.1676790863275528,
0.002958063967525959,
0.01386290229856968,
-0.060542043298482895,
0.2447528839111328,
-0.12088193744421005,
-0.004838956985622644,
0.08523343503475189,
-0.06330211460590363,
0.14527028799057007,
-0.09573397040367126,
-0.10710596293210983,
-0.06111620366573334,
0.07404182851314545,
-0.13550832867622375,
0.054874829947948456,
-0.06791821867227554,
0.04847199469804764,
-0.004893225617706776,
-0.069199338555336,
0.012129784561693668,
0.00041885548853315413,
-0.05295775830745697,
0.15489588677883148,
-0.03530820086598396,
-0.061302945017814636,
-0.17509964108467102,
0.12554048001766205,
0.08078322559595108,
0.03058622218668461,
-0.04842965304851532,
-0.07379470020532608,
-0.04554806277155876,
0.006061905063688755,
0.07991603016853333,
-0.1003677025437355,
0.04164258763194084,
0.06845992058515549,
-0.029039530083537102,
-0.005017103627324104,
-0.04464732110500336,
0.02714737318456173,
-0.06764040142297745,
0.026123248040676117,
0.09623407572507858,
-0.12273786216974258,
0.06955782324075699,
0.12263170629739761,
0.08097522705793381,
-0.00543251633644104,
0.08693471550941467,
0.3090808689594269,
-0.10241376608610153,
-0.01750842295587063,
0.22084642946720123,
0.06672252714633942,
-0.060791388154029846,
0.09880859404802322,
0.09504423290491104,
-0.12348990142345428,
-0.02172943949699402,
-0.1135643944144249,
-0.11265823990106583,
-0.16027554869651794,
-0.0984187200665474,
-0.07916464656591415,
-0.038754455745220184,
-0.01063135452568531,
0.0777021273970604,
0.09660246223211288,
0.11373791843652725,
-0.019434921443462372,
-0.06446122378110886,
-0.18080899119377136,
0.04930289089679718,
0.07030103355646133,
-0.030653469264507294,
0.07881137728691101,
-0.11564989387989044,
0.054437268525362015,
0.06870304048061371,
0.06236737594008446,
0.05171265825629234,
0.19388417899608612,
-0.1151016503572464,
0.006796098779886961,
0.17557287216186523,
0.07904418557882309,
0.12418877333402634,
0.024383369833230972,
-0.07462010532617569,
-0.013403046876192093,
-0.02653372660279274,
-0.04990820959210396,
0.08941224217414856,
-0.11727767437696457,
-0.03275684267282486,
0.012237473390996456,
-0.27074941992759705,
-0.02266460843384266,
-0.027867550030350685,
0.15126711130142212,
-0.15473823249340057,
-0.046685751527547836,
0.09594705700874329,
0.0029495481867343187,
-0.0584033727645874,
0.030555784702301025,
0.04577462002635002,
-0.06639561057090759,
0.03888459876179695,
0.04793861135840416,
0.008257733657956123,
0.07151667028665543,
0.03760693967342377,
-0.013276845216751099,
0.005910684820264578,
-0.002160864183679223,
0.021207936108112335,
-0.06237469241023064,
0.24330179393291473,
0.0014929120661690831,
-0.10534156113862991,
0.023643720895051956,
-0.029320651665329933,
-0.09961745887994766,
0.13767749071121216,
0.04844130575656891,
0.013945086859166622,
-0.08121956884860992,
-0.07728705555200577,
-0.08031905442476273,
0.04595201462507248,
0.03537088260054588,
-0.1012100800871849,
-0.030760077759623528,
0.0402519553899765,
0.05974670872092247,
-0.019724445417523384,
0.08839298039674759,
-0.08160920441150665,
0.009760478511452675,
0.01800018735229969,
-0.08688464015722275,
-0.012509194202721119,
-0.05041201040148735,
-0.007081549149006605,
-0.06144506856799126,
-0.10147185623645782,
-0.07389876991510391,
-0.02801421284675598,
-0.07057300955057144,
-0.05595795437693596,
0.2428489327430725,
-0.052721377462148666,
0.057009581476449966,
-0.005434415768831968,
-0.15246357023715973,
0.0776752308011055,
-0.2021179050207138,
0.1397310197353363,
-0.13785089552402496,
0.007539375685155392,
0.01579749584197998,
0.13889051973819733,
-0.009868939407169819,
0.05188767984509468,
-0.061964068561792374,
0.04965178668498993,
-0.1453644186258316,
-0.1011054664850235,
0.1000019833445549,
-0.1202540397644043,
0.017855728045105934,
0.028666380792856216,
0.04283863306045532,
0.19925901293754578,
0.0847022533416748,
-0.054671622812747955,
-0.03981555253267288,
0.2872854769229889,
-0.07426895201206207,
0.08139975368976593,
0.12522581219673157,
0.015486636199057102,
-0.12671694159507751,
-0.0479712076485157,
-0.1394456923007965,
-0.011287250556051731,
0.061549749225378036,
-0.27310287952423096,
0.0772022008895874,
0.15375584363937378,
-0.06726932525634766,
0.05224788188934326,
-0.25188255310058594,
-0.013067594729363918,
0.17563629150390625,
-0.02954414300620556,
0.4037812650203705,
-0.13885600864887238,
-0.12138570100069046,
0.006686338223516941,
-0.21247032284736633,
0.02474556677043438,
0.019272420555353165,
-0.013933214358985424,
-0.07093660533428192,
-0.02407042495906353,
0.037790220230817795,
0.0031322455033659935,
0.26509636640548706,
-0.047945261001586914,
0.028926709666848183,
-0.05116455629467964,
-0.06705578416585922,
0.03322160243988037,
0.02085425890982151,
0.009410316124558449,
0.1496959626674652,
0.056193187832832336,
-0.26853951811790466,
0.04545442759990692,
-0.08424293994903564,
-0.027164112776517868,
-0.000684787577483803,
-0.036247070878744125,
0.0027821932453662157,
-0.014754501171410084,
-0.1296020895242691,
-0.0015607376117259264,
0.32664769887924194,
-0.007427219767123461,
0.14338259398937225,
0.04346732422709465,
-0.024002714082598686,
-0.08274754136800766,
0.09537393599748611,
-0.06952955573797226,
-0.04297981783747673,
0.11821122467517853,
-0.21870091557502747,
-0.06310757994651794,
0.005701492074877024,
0.03345939889550209,
-0.08571425825357437,
0.08285060524940491,
-0.04760076850652695,
0.013425195589661598,
0.16613559424877167,
-0.19683097302913666,
-0.13703593611717224,
-0.051705148071050644,
0.07395339757204056,
0.06988521665334702,
0.2227270007133484,
0.0692393109202385,
-0.013149088248610497,
-0.05806303396821022,
0.016482515260577202,
-0.01762229949235916,
-0.06555264443159103,
0.0022581920493394136,
0.12411420047283173,
0.044070854783058167,
-0.12714263796806335,
0.1077854335308075,
0.0553167425096035,
0.08761455118656158,
-0.08660001307725906,
0.036114197224378586,
-0.005688804667443037,
-0.05154223367571831,
-0.07098440080881119,
-0.17306765913963318,
-0.22401346266269684,
-0.05483974888920784,
-0.061699677258729935,
-0.049688711762428284,
0.05132006108760834,
0.048181816935539246,
0.0815325528383255,
-0.024337833747267723,
0.0099272346124053,
0.03798292949795723,
0.07296915352344513,
-0.16068552434444427,
-0.1545608639717102,
0.054870907217264175,
-0.08752300590276718,
-0.14927934110164642,
-0.07499323785305023,
0.07146461308002472,
-0.09369296580553055,
-0.0934004858136177,
-0.12715484201908112,
0.04061734303832054,
-0.02547648176550865,
-0.13123296201229095,
-0.06684274971485138,
-0.04602637141942978,
0.024521224200725555,
-0.08362280577421188,
-0.08421293646097183,
0.018558625131845474,
-0.11949391663074493,
0.01216824259608984,
-0.014980743639171124,
0.01739502139389515,
0.018765566870570183,
0.1496751755475998,
0.10357877612113953,
-0.05437363311648369,
0.06645531952381134,
0.14547418057918549,
0.11221052706241608,
0.09521616250276566,
-0.0664844959974289,
0.00011787811672547832,
0.07362651824951172,
-0.02427438460290432,
0.02559869922697544,
0.0820465236902237,
-0.033545758575201035,
-0.005585681181401014,
-0.05225710570812225,
-0.022081848233938217,
0.03840937837958336,
-0.07410874962806702,
-0.05176355689764023,
-0.04443126916885376,
-0.09711979329586029,
-0.028587955981492996,
-0.04834691807627678,
0.026124199852347374,
0.033971890807151794,
-0.05946870893239975,
0.06180686131119728,
-0.05240452289581299,
-0.10307358205318451,
0.007262640167027712,
-0.029853126034140587,
-0.04113887622952461,
-0.004830288700759411,
0.03053942881524563,
0.04692051187157631,
0.09694849699735641,
0.2910834550857544,
-0.1269485205411911,
0.05987376347184181,
-0.04438069835305214,
0.24882180988788605,
0.1662714034318924,
0.012907366268336773,
0.44027069211006165,
0.13869956135749817,
0.05155595764517784,
-0.030508529394865036,
0.12292912602424622,
-0.04518717899918556,
-0.007385023403912783,
0.020814895629882812,
0.08734874427318573,
-0.1622742861509323,
0.09957464784383774,
0.032486990094184875,
-0.07150827348232269,
0.10402429103851318,
0.013065893203020096,
-0.01056518591940403,
-0.06128927692770958,
-0.0005592253291979432,
0.1689346879720688,
0.14233003556728363,
-0.01739235408604145,
-0.017513541504740715,
0.0573112778365612,
0.02344306744635105,
-0.08449359238147736,
-0.07823538035154343,
0.01851218193769455,
-0.2925812005996704,
0.05156448855996132,
-0.08636578172445297,
-0.0052759707905352116,
0.3088139593601227,
-0.03129149600863457,
0.0010523020755499601,
0.031176801770925522,
0.16199597716331482,
-0.05711657181382179,
-0.05011753737926483,
0.016967415809631348,
-0.009012405760586262,
-0.15996652841567993,
-0.10201755911111832,
0.031197642907500267,
-0.05974094942212105,
-0.055021364241838455,
-0.03251253440976143,
0.022550666704773903,
0.014574510976672173,
0.014115294441580772,
-0.07269603759050369,
-0.014967354014515877,
-0.0005734374281018972,
-0.10366055369377136,
0.14863769710063934,
-0.03988127037882805,
0.03576371446251869,
-0.022846264764666557,
0.26281946897506714,
-0.10013621300458908,
0.12033334374427795,
0.012469213455915451,
0.08026184886693954,
-0.10968228429555893,
0.11241127550601959,
0.004621094558387995,
-0.10879184305667877,
-0.06422177702188492,
0.11529987305402756,
0.08952199667692184,
-0.02903035469353199,
0.04175647720694542,
0.13887478411197662,
0.011731777340173721,
0.04165593162178993,
0.12981459498405457,
0.08182945102453232,
0.07878459990024567,
-0.08359827846288681,
-0.005663351155817509,
-0.01975712925195694,
0.028818292543292046,
-0.027457602322101593,
0.02426818571984768,
0.01773497462272644,
-0.1673511564731598,
-0.08665613830089569,
0.11427482217550278,
-0.029190130531787872,
0.03208428621292114,
0.021980587393045425,
-0.12498664110898972,
-0.06480853259563446,
-0.09467045217752457,
0.0948532298207283,
-0.028616508468985558,
0.026108909398317337,
-0.12760597467422485,
-0.0074180555529892445,
0.02747167833149433,
0.07078221440315247,
-0.1715291291475296,
-0.24105778336524963,
0.0871356651186943,
-0.037883199751377106,
0.0614156574010849,
-0.036758191883563995,
0.07257413864135742,
-0.0006313854246400297,
0.03389989957213402,
-0.08922875672578812,
0.02049221284687519,
0.05975357070565224,
-0.12126701325178146,
-0.060468532145023346,
0.0494573749601841,
0.10185249894857407,
-0.025992801412940025,
0.006163603160530329,
-0.08571828901767731,
-0.04688110202550888,
0.12372100353240967,
-0.0189700685441494,
0.03014969453215599,
-0.041027460247278214,
-0.12418027967214584,
0.10909076035022736,
-0.008599084801971912,
-0.10074981302022934,
-0.0788172036409378,
-0.0670599639415741,
0.026603523641824722,
0.04355804994702339,
-0.025241289287805557,
-0.037811484187841415,
0.06725604832172394,
-0.040587879717350006,
-0.022851185873150826,
0.007342976983636618,
0.038123033940792084,
0.015111060813069344,
-0.06829354912042618,
0.05549819394946098,
-0.04272347688674927,
0.002939540194347501,
0.04223529249429703,
-0.010539565235376358,
-0.016520747914910316,
-0.1567634642124176,
0.03002692572772503,
0.07097426801919937,
-0.06872686743736267,
-0.058046355843544006
] |
null | null | transformers |
This model is the fine-tuned version of [Helsinki-NLP/opus-mt-ja-en](https://huggingface.co/Helsinki-NLP/opus-mt-ja-en) on bsd_ja_en dataset.
This will translate Japanese sentences to English sentences. | {"language": ["ja", "en"], "license": "mit", "library_name": "transformers", "datasets": ["bsd_ja_en"], "metrics": ["sacrebleu"], "pipeline_tag": "translation", "widget": [{"text": "\u304a\u306f\u3044\u3088\u3002"}, {"text": "\u50d5\u306e\u56fd\u3067\u3000\u3068\u3066\u3082\u3000\u7dba\u9e97\u306a\u3000\u6240\u304c\u3000\u6709\u308a\u307e\u3059\u3002"}]} | translation | minkhantycc/translation-en-ja | [
"transformers",
"safetensors",
"marian",
"text2text-generation",
"translation",
"ja",
"en",
"dataset:bsd_ja_en",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T08:08:41+00:00 | [] | [
"ja",
"en"
] | TAGS
#transformers #safetensors #marian #text2text-generation #translation #ja #en #dataset-bsd_ja_en #license-mit #autotrain_compatible #endpoints_compatible #region-us
|
This model is the fine-tuned version of Helsinki-NLP/opus-mt-ja-en on bsd_ja_en dataset.
This will translate Japanese sentences to English sentences. | [] | [
"TAGS\n#transformers #safetensors #marian #text2text-generation #translation #ja #en #dataset-bsd_ja_en #license-mit #autotrain_compatible #endpoints_compatible #region-us \n"
] | [
62
] | [
"passage: TAGS\n#transformers #safetensors #marian #text2text-generation #translation #ja #en #dataset-bsd_ja_en #license-mit #autotrain_compatible #endpoints_compatible #region-us \n"
] | [
-0.040520910173654556,
0.062146931886672974,
-0.006390941794961691,
0.009642377495765686,
0.12279313802719116,
0.0009974618442356586,
0.16769470274448395,
0.09651128947734833,
-0.023115329444408417,
-0.026978198438882828,
0.15866708755493164,
0.19700929522514343,
0.0026607210747897625,
0.1805523782968521,
-0.11586539447307587,
-0.17436519265174866,
0.09847486019134521,
0.016374394297599792,
-0.03023165836930275,
0.08660262078046799,
0.12164264917373657,
-0.047575462609529495,
0.10978963971138,
-0.06219690665602684,
-0.10504372417926788,
0.03933229297399521,
0.04453056678175926,
-0.12114686518907547,
0.09200050681829453,
0.08354554325342178,
0.10570180416107178,
0.09209483116865158,
-0.009051731787621975,
-0.19120405614376068,
0.02312258444726467,
-0.024761851876974106,
-0.08028032630681992,
0.033771079033613205,
0.055052585899829865,
-0.05682744085788727,
0.018932459875941277,
-0.022849028930068016,
-0.04262286424636841,
0.08135929703712463,
-0.15362374484539032,
-0.09349203109741211,
-0.06815687566995621,
-0.04024147614836693,
0.09412457793951035,
0.05284615978598595,
-0.0024610201362520456,
0.09724847227334976,
-0.045134760439395905,
0.08896605670452118,
0.09542176872491837,
-0.3223557770252228,
0.007569167762994766,
0.05711587518453598,
0.07771091163158417,
0.07066676020622253,
0.021302763372659683,
0.10153672844171524,
0.0800161212682724,
-0.028507214039564133,
-0.018094055354595184,
-0.0938156247138977,
-0.08457324653863907,
0.03939294442534447,
-0.045065149664878845,
-0.025372382253408432,
0.2698329985141754,
-0.05875655263662338,
0.013896865770220757,
-0.05511712655425072,
-0.06246171519160271,
0.001517186057753861,
-0.037012405693531036,
-0.02052829973399639,
-0.0036029487382620573,
0.08175652474164963,
0.03661715239286423,
-0.014514106325805187,
-0.15026596188545227,
-0.006664519663900137,
-0.20466512441635132,
0.08593261241912842,
-0.0025234816130250692,
0.03658396005630493,
-0.1920982003211975,
0.009640058502554893,
0.035728346556425095,
-0.12481456995010376,
-0.008125322870910168,
-0.09593117237091064,
0.07429010421037674,
-0.006102791987359524,
-0.01516781747341156,
-0.10085006803274155,
0.14348174631595612,
0.1057550385594368,
0.007252941373735666,
0.021738091483712196,
-0.11216732114553452,
0.08131101727485657,
0.03543027862906456,
0.04548100382089615,
0.025508325546979904,
-0.08433865010738373,
0.08066127449274063,
-0.1317269653081894,
0.04678446426987648,
-0.028750449419021606,
-0.16970781981945038,
-0.039279673248529434,
0.009675518609583378,
0.1330023854970932,
0.027444466948509216,
0.06914021074771881,
-0.07122443616390228,
0.042901989072561264,
0.09403333067893982,
-0.051278337836265564,
0.021079404279589653,
0.01926788128912449,
0.04909422993659973,
0.014537111856043339,
0.016539432108402252,
0.03469406068325043,
-0.027260802686214447,
0.029335318133234978,
-0.043903931975364685,
-0.03851752355694771,
-0.026851413771510124,
-0.09528467804193497,
0.07196556776762009,
-0.047742217779159546,
0.0756470113992691,
-0.20941609144210815,
-0.1857433170080185,
0.007871310226619244,
0.00928929727524519,
-0.05090617015957832,
0.023984713479876518,
-0.06200353428721428,
-0.02483414299786091,
0.05229126662015915,
-0.10658270120620728,
-0.027613431215286255,
-0.07900252938270569,
0.06673916429281235,
0.0011337768519297242,
0.06966336816549301,
-0.14082157611846924,
0.026221321895718575,
-0.12191606312990189,
-0.014179154299199581,
-0.047542642802000046,
0.061362918466329575,
-0.07597272098064423,
0.16129732131958008,
-0.04405475780367851,
0.017277052626013756,
-0.026248689740896225,
0.08202652633190155,
-0.028086919337511063,
0.22878526151180267,
-0.19881656765937805,
-0.04498118907213211,
0.2485979199409485,
-0.1277807354927063,
-0.20883060991764069,
0.07618517428636551,
-0.009116948582231998,
0.1372034251689911,
0.1059291884303093,
0.19516794383525848,
-0.00485962862148881,
-0.03998906910419464,
0.08257995545864105,
0.060069769620895386,
-0.06868898123502731,
-0.09558010846376419,
0.029977187514305115,
-0.009767935611307621,
-0.08012328296899796,
0.03128964081406593,
0.07533672451972961,
0.08990923315286636,
-0.025385485962033272,
-0.04837285727262497,
0.0009896584087982774,
-0.033110201358795166,
0.006493048742413521,
0.008953277952969074,
0.0627848282456398,
-0.1163681223988533,
0.02868000790476799,
-0.00479743629693985,
-0.030779585242271423,
0.008712070994079113,
0.010875132866203785,
-0.08894973248243332,
0.023197025060653687,
-0.04511600732803345,
0.0526917390525341,
-0.06881437450647354,
-0.04550919309258461,
-0.022116364911198616,
0.06185434013605118,
0.0021997790317982435,
0.08441659063100815,
0.06357551366090775,
-0.0184557493776083,
-0.004546676762402058,
-0.009550623595714569,
0.21378682553768158,
0.0722583532333374,
-0.04505869373679161,
-0.1125437393784523,
0.09950889647006989,
-0.06447675079107285,
-0.0030813885387033224,
-0.09143266826868057,
0.016834093257784843,
0.07671348005533218,
0.10022418200969696,
0.026771632954478264,
0.0702647864818573,
-0.014325687661767006,
0.03193819522857666,
-0.08315501362085342,
-0.005230179987847805,
0.07660795748233795,
0.009009601548314095,
-0.07031489163637161,
0.27189305424690247,
-0.1590966284275055,
0.3066866993904114,
0.19512759149074554,
-0.15355339646339417,
0.039017561823129654,
-0.011941596865653992,
0.01979534886777401,
0.007090630475431681,
0.048890143632888794,
0.014110559597611427,
-0.034424200654029846,
0.0018783487612381577,
0.17873677611351013,
-0.06579184532165527,
-0.004000382032245398,
0.02888021431863308,
-0.10576363652944565,
-0.08368532359600067,
0.06109335273504257,
0.0015040753642097116,
-0.31491419672966003,
0.20768098533153534,
0.21476103365421295,
0.07933924347162247,
0.20410875976085663,
-0.01771486923098564,
0.01951887272298336,
0.01536147203296423,
0.002721268916502595,
0.014800201170146465,
0.02145783044397831,
-0.11685584485530853,
-0.03414057567715645,
0.07447126507759094,
0.05158592760562897,
0.06220107525587082,
-0.08708423376083374,
-0.06752067059278488,
0.022194556891918182,
-0.0133946118876338,
-0.018451610580086708,
0.07740891724824905,
-0.007144222501665354,
0.10278166085481644,
-0.03962818905711174,
-0.01255897805094719,
0.09475847333669662,
-0.012760058045387268,
-0.11140666157007217,
0.17470306158065796,
-0.17594139277935028,
-0.29622459411621094,
-0.17870138585567474,
-0.1739180088043213,
0.0017276568105444312,
0.07419012486934662,
0.14227290451526642,
-0.0783582478761673,
-0.06512866914272308,
-0.013666896149516106,
0.029521621763706207,
-0.019112346693873405,
-0.025330111384391785,
-0.06442602723836899,
0.08304231613874435,
-0.034716133028268814,
-0.09580695629119873,
-0.04295642673969269,
0.04784557968378067,
-0.007027393206954002,
0.12320274114608765,
-0.1356850117444992,
0.13458748161792755,
0.04635399952530861,
-0.0037560765631496906,
0.0171689260751009,
-0.06003420427441597,
0.13223740458488464,
-0.06098713353276253,
-0.007060793228447437,
0.1799274981021881,
-0.021743902936577797,
0.00872825551778078,
0.19792138040065765,
-0.0027260074857622385,
-0.08945447206497192,
0.017095988616347313,
-0.08529504388570786,
-0.07383094727993011,
-0.2521422505378723,
-0.11810151487588882,
-0.08884795755147934,
0.11113769561052322,
0.037162039428949356,
0.045164886862039566,
0.06117107719182968,
0.09943123906850815,
-0.005733078811317682,
0.011170039884746075,
0.07069817930459976,
0.09597756713628769,
0.19222518801689148,
-0.019650548696517944,
0.12654516100883484,
-0.10202467441558838,
-0.0838940292596817,
0.10520704090595245,
0.08758342266082764,
0.023533686995506287,
0.1174258291721344,
0.10503843426704407,
0.06556878983974457,
0.09625566005706787,
0.1203811764717102,
0.13363392651081085,
0.09424565732479095,
-0.02911198139190674,
-0.04302585870027542,
-0.029481487348675728,
-0.0891106128692627,
0.043706264346838,
-0.0641469657421112,
-0.11063505709171295,
-0.054368164390325546,
-0.05852976068854332,
0.07182735204696655,
0.0012799383839592338,
0.05822037160396576,
-0.21443265676498413,
0.006959805730730295,
0.14061661064624786,
0.030213557183742523,
-0.12183304876089096,
0.09973544627428055,
0.05166270583868027,
-0.08161936700344086,
0.10355771332979202,
-0.028207851573824883,
0.11475033313035965,
-0.025113284587860107,
0.056707337498664856,
-0.10070133209228516,
-0.1455564796924591,
0.02400144375860691,
0.13189589977264404,
-0.4086610674858093,
0.21307992935180664,
0.031041596084833145,
0.021992672234773636,
-0.05501187965273857,
-0.004610815551131964,
-0.02685737796127796,
0.16523313522338867,
0.14649654924869537,
-0.012973539531230927,
-0.15038587152957916,
-0.07546145468950272,
-0.012137897312641144,
0.04163177311420441,
0.066420778632164,
0.032798636704683304,
-0.0011523122666403651,
-0.04877179116010666,
0.006625636015087366,
-0.03046393394470215,
0.04374143108725548,
-0.07550659775733948,
-0.1870516538619995,
0.024591274559497833,
0.08493461459875107,
0.12753801047801971,
-0.030452940613031387,
0.018902063369750977,
-0.08944576233625412,
0.15449799597263336,
-0.10364305973052979,
-0.04773203283548355,
-0.12110647559165955,
-0.09298920631408691,
0.021007360890507698,
-0.06249268725514412,
0.04275199770927429,
-0.05394674465060234,
0.023813756182789803,
-0.08260060846805573,
-0.17364725470542908,
0.1282915323972702,
-0.13448941707611084,
-0.035065434873104095,
-0.09141235053539276,
0.07521622627973557,
-0.07364408671855927,
-0.017464466392993927,
0.0652516782283783,
0.022144658491015434,
-0.07884182035923004,
-0.06047336384654045,
-0.0010151787428185344,
0.0022937317844480276,
0.08955015987157822,
-0.011445707641541958,
-0.08804067224264145,
-0.15880167484283447,
-0.0018632583087310195,
-0.08294473588466644,
0.21112214028835297,
0.23935149610042572,
-0.053687967360019684,
0.14439265429973602,
0.1918874830007553,
-0.09725308418273926,
-0.3363737165927887,
-0.1290995180606842,
-0.16217128932476044,
-0.0014172900700941682,
-0.004304489120841026,
-0.07844746112823486,
0.015928776934742928,
0.0024002166464924812,
-0.06416738033294678,
0.02065363898873329,
-0.15123653411865234,
-0.0914226770401001,
0.1764916181564331,
0.03808075562119484,
0.3759640157222748,
-0.16956226527690887,
-0.10868722945451736,
-0.13072174787521362,
-0.20385868847370148,
0.13298529386520386,
-0.13410814106464386,
0.03502790629863739,
0.036408793181180954,
0.05389256030321121,
0.01012403704226017,
-0.05182188004255295,
0.11754797399044037,
-0.07096924632787704,
0.035044584423303604,
-0.12828628718852997,
-0.02415478229522705,
0.06440971791744232,
-0.010948233306407928,
0.029879247769713402,
-0.10574273020029068,
0.02894512191414833,
-0.08188935369253159,
-0.02491808868944645,
-0.02419823780655861,
0.06542201340198517,
0.0006988161476328969,
-0.06083725392818451,
-0.005424969829618931,
-0.038599442690610886,
0.02156081423163414,
-0.007075082045048475,
0.2879123389720917,
-0.05915623530745506,
0.1082751601934433,
0.16221237182617188,
0.14576350152492523,
-0.12483397126197815,
0.2132476568222046,
-0.04733164608478546,
-0.09737595915794373,
0.05093875154852867,
-0.10777924209833145,
0.0813143402338028,
0.07136013358831406,
-0.08135096728801727,
0.07965272665023804,
0.0647633820772171,
-0.0013730853097513318,
0.01961931586265564,
0.16888071596622467,
-0.17127709090709686,
-0.04207318648695946,
-0.02508770301938057,
0.05355588346719742,
0.12436317652463913,
0.09397682547569275,
0.17960521578788757,
0.011178974062204361,
-0.03231154754757881,
-0.023094307631254196,
0.044997915625572205,
-0.06867953389883041,
0.09543154388666153,
0.01455117017030716,
0.013584437780082226,
-0.1329749971628189,
0.10139923542737961,
-0.003089115023612976,
-0.13354040682315826,
0.039013445377349854,
0.07517928630113602,
-0.14916172623634338,
-0.1350233554840088,
0.033103879541158676,
0.2082640677690506,
-0.13394513726234436,
-0.09623449295759201,
-0.048330437391996384,
-0.13497933745384216,
0.016852399334311485,
0.19026657938957214,
0.04849069565534592,
0.0838267058134079,
0.008680107071995735,
-0.06503613293170929,
-0.018100272864103317,
0.029527336359024048,
-0.013025752268731594,
0.037439677864313126,
-0.1091773733496666,
-0.011339934542775154,
-0.0634637326002121,
0.0586024634540081,
-0.07840428501367569,
0.01625930890440941,
-0.16307133436203003,
0.017780596390366554,
-0.1560612916946411,
0.007220305502414703,
-0.10787802189588547,
-0.03583746403455734,
0.0004203059943392873,
-0.0638309046626091,
-0.049188774079084396,
-0.04861263558268547,
-0.0840715691447258,
0.027974089607596397,
-0.040424153208732605,
0.0815345048904419,
-0.09382564574480057,
-0.046767544001340866,
0.04909136891365051,
-0.00496765086427331,
0.07747089862823486,
0.08406579494476318,
-0.06345363706350327,
0.09478060156106949,
-0.18605947494506836,
-0.05650695785880089,
0.120273657143116,
-0.0023551962804049253,
0.07771047949790955,
0.03964285925030708,
0.000541458954103291,
0.1439259797334671,
-0.005705804564058781,
0.03785217925906181,
0.05329137668013573,
-0.07162688672542572,
0.023677289485931396,
-0.03279098868370056,
-0.09276454150676727,
-0.020986678078770638,
-0.011998763307929039,
0.09278853237628937,
-0.03660758584737778,
0.1528995931148529,
-0.09855365753173828,
0.03943656384944916,
-0.018458254635334015,
0.016524173319339752,
-0.009772351942956448,
-0.1863870769739151,
-0.13406530022621155,
-0.08350774645805359,
-0.0050292303785681725,
-0.009364256635308266,
0.27034488320350647,
0.046276576817035675,
-0.025639904662966728,
0.050373293459415436,
-0.039033565670251846,
0.02298055589199066,
0.03802747279405594,
0.28892457485198975,
0.048717889934778214,
-0.0314827635884285,
-0.16034623980522156,
0.07418426871299744,
0.009073931723833084,
-0.03939737752079964,
-0.004355020821094513,
0.13547256588935852,
-0.035260725766420364,
0.11244190484285355,
0.00383016187697649,
0.00562724843621254,
-0.027543410658836365,
-0.1264832764863968,
-0.015889836475253105,
0.016713151708245277,
0.010991323739290237,
0.038398463279008865,
0.18686209619045258,
-0.06517106294631958,
-0.011704285629093647,
-0.0841129869222641,
-0.022629225626587868,
-0.18781998753547668,
-0.1431158483028412,
-0.11940751224756241,
-0.10831724852323532,
-0.0000017516766774861026,
-0.09073789417743683,
0.03489413484930992,
-0.0018620891496539116,
0.05162052810192108,
-0.06975476443767548,
0.058224789798259735,
0.04108554124832153,
-0.07799220085144043,
0.01580415666103363,
-0.0060288188979029655,
0.04584192484617233,
0.0008558520348742604,
-0.021320903673768044,
-0.07510600239038467,
-0.010299666784703732,
-0.029577111825346947,
0.068926602602005,
-0.011562023311853409,
0.05225716903805733,
-0.12231000512838364,
-0.09585229307413101,
-0.043043479323387146,
0.09255696833133698,
-0.03190125524997711,
0.14600388705730438,
0.02610073797404766,
-0.01773083582520485,
0.09983571618795395,
0.20606733858585358,
-0.019708722829818726,
-0.11606980115175247,
-0.031766217201948166,
0.15622206032276154,
0.05274520069360733,
0.1786915361881256,
-0.05199723690748215,
0.008183654397726059,
-0.02944040857255459,
0.29180002212524414,
0.27896150946617126,
-0.05500331521034241,
0.016485989093780518,
-0.005466020200401545,
0.03225724771618843,
0.06089168041944504,
0.10939198732376099,
0.05022205784916878,
0.2243027538061142,
-0.039958640933036804,
0.014125843532383442,
-0.007329340558499098,
-0.00795038789510727,
-0.1059984639286995,
0.1219300627708435,
-0.037178389728069305,
-0.05136574059724808,
-0.006208393257111311,
0.12428643554449081,
-0.12143871933221817,
0.0738014429807663,
-0.040168046951293945,
-0.13022832572460175,
-0.016356073319911957,
-0.022957399487495422,
0.20278623700141907,
0.0013716417597606778,
0.037715207785367966,
-0.014771519228816032,
-0.04236553609371185,
0.0414736233651638,
0.006700839847326279,
-0.16461962461471558,
0.034142911434173584,
-0.014191298745572567,
-0.09122490137815475,
0.09503758698701859,
-0.016948988661170006,
0.0011926810257136822,
0.09638042002916336,
0.035086508840322495,
-0.08576446026563644,
0.11413680016994476,
0.015628837049007416,
-0.005168532487004995,
0.04948257654905319,
-0.03460518270730972,
-0.023124005645513535,
-0.04081863909959793,
0.09306305646896362,
-0.17417262494564056,
0.08670192211866379,
-0.011664698831737041,
-0.1033000722527504,
-0.014790213666856289,
0.05974866449832916,
-0.061054859310388565,
0.057885974645614624,
-0.0005534444935619831,
-0.010273692198097706,
0.0204556155949831,
-0.04401419311761856,
0.042719002813100815,
0.0421263724565506,
-0.02340705506503582,
-0.023612674325704575,
-0.09042287617921829,
-0.10164904594421387,
0.14461679756641388,
0.04846037179231644,
-0.23451948165893555,
0.052960947155952454,
-0.12182650715112686,
0.02964623272418976,
-0.15142399072647095,
0.10707315057516098,
0.12841646373271942,
0.011059591546654701,
-0.027323324233293533,
-0.19299229979515076,
0.0632367953658104,
0.10254776477813721,
-0.03778931871056557,
-0.11308155208826065
] |
null | null | flair | ## Polish Flair Model -- NER
| {"language": "pl", "tags": ["flair", "token-classification", "sequence-tagger-model"], "widget": [{"text": "Jan Brzechwa - polski poeta i adwokat \u017cydowskiego pochodzenia, autor bajek i wierszy dla dzieci, satyrycznych tekst\u00f3w dla doros\u0142ych, a tak\u017ce t\u0142umacz literatury rosyjskiej."}]} | token-classification | clarin-knext/morpho-flair-ner | [
"flair",
"pytorch",
"token-classification",
"sequence-tagger-model",
"pl",
"region:us"
] | 2024-02-08T08:12:07+00:00 | [] | [
"pl"
] | TAGS
#flair #pytorch #token-classification #sequence-tagger-model #pl #region-us
| ## Polish Flair Model -- NER
| [
"## Polish Flair Model -- NER"
] | [
"TAGS\n#flair #pytorch #token-classification #sequence-tagger-model #pl #region-us \n",
"## Polish Flair Model -- NER"
] | [
30,
9
] | [
"passage: TAGS\n#flair #pytorch #token-classification #sequence-tagger-model #pl #region-us \n## Polish Flair Model -- NER"
] | [
-0.03771134465932846,
-0.03943900018930435,
-0.010174674913287163,
0.07979106158018112,
0.020937331020832062,
0.06407680362462997,
0.05683210864663124,
0.0080938171595335,
0.11943959444761276,
0.08812148869037628,
0.14531303942203522,
0.10004604607820511,
-0.022733677178621292,
0.09909739345312119,
0.035872794687747955,
-0.37102010846138,
0.09340322017669678,
0.01615891046822071,
-0.05567781254649162,
0.14803780615329742,
0.05753672868013382,
-0.027101390063762665,
-0.014619898051023483,
0.054841045290231705,
-0.018896013498306274,
0.04070658981800079,
-0.05214330926537514,
-0.05382240563631058,
0.1128356084227562,
-0.0615239255130291,
0.1508791297674179,
0.004013254772871733,
-0.01300971768796444,
-0.14023835957050323,
0.017160795629024506,
-0.054064784198999405,
-0.061748992651700974,
0.0686250850558281,
0.1319541037082672,
-0.15455178916454315,
0.2210318148136139,
-0.02418479323387146,
0.028442399576306343,
-0.00005364972457755357,
-0.1380181461572647,
-0.15137095749378204,
-0.014303099364042282,
0.12107576429843903,
0.042073916643857956,
0.03295950964093208,
-0.03030707873404026,
0.1345042884349823,
-0.21378566324710846,
0.038679659366607666,
0.1291813999414444,
-0.21939823031425476,
-0.055440936237573624,
0.10563647747039795,
-0.004520352464169264,
0.06557643413543701,
-0.08346334844827652,
0.049519121646881104,
-0.022583134472370148,
-0.020149072632193565,
-0.02539493888616562,
-0.018281733617186546,
0.009624544531106949,
0.06292911618947983,
-0.19257745146751404,
-0.03453587740659714,
0.14325226843357086,
0.0070975469425320625,
0.03840754181146622,
-0.00714239664375782,
-0.03486749902367592,
-0.05717717111110687,
-0.023125059902668,
-0.0028800885193049908,
-0.03292695805430412,
-0.006302658002823591,
0.26715049147605896,
0.06090281158685684,
-0.02771889604628086,
0.011001885868608952,
-0.07604467868804932,
0.30720043182373047,
0.03235085681080818,
0.08044152706861496,
-0.12382914125919342,
-0.019776897504925728,
-0.09521832317113876,
-0.05968090146780014,
0.08978400379419327,
-0.06458434462547302,
-0.07946904003620148,
0.010821749456226826,
0.016602545976638794,
0.12280833721160889,
0.07467862963676453,
0.2388511300086975,
0.03404088690876961,
0.06878969073295593,
0.13548120856285095,
0.12755811214447021,
0.044054266065359116,
0.010909130796790123,
-0.012339810840785503,
-0.08312239497900009,
-0.0171683207154274,
-0.13248775899410248,
0.001818948658183217,
-0.07348378002643585,
-0.10531266778707504,
-0.055205751210451126,
-0.039701953530311584,
0.02879178524017334,
-0.020259398967027664,
0.0033887105528265238,
-0.039398204535245895,
-0.045560985803604126,
0.06298556178808212,
-0.00973974447697401,
-0.0018246837425976992,
0.06146596744656563,
-0.030356457456946373,
0.2059175819158554,
-0.16403895616531372,
0.002045328263193369,
-0.011921214871108532,
0.19298039376735687,
-0.060687895864248276,
0.016408927738666534,
0.0014912323094904423,
-0.0734303742647171,
0.034390345215797424,
-0.12864580750465393,
0.018682440742850304,
-0.059983041137456894,
-0.02399197779595852,
0.053905803710222244,
0.02826579473912716,
-0.04020058736205101,
-0.0028932795394212008,
-0.01599775068461895,
-0.01494627632200718,
0.01678065024316311,
-0.010283348150551319,
-0.056642044335603714,
-0.09513358026742935,
0.059790052473545074,
-0.13288427889347076,
0.033297739923000336,
-0.12122499942779541,
0.01481608860194683,
-0.1343211978673935,
0.014697492122650146,
-0.25413596630096436,
-0.11363296955823898,
-0.12540706992149353,
0.00913239922374487,
-0.058199960738420486,
-0.0889040157198906,
-0.000020510928152361885,
-0.03160668909549713,
-0.061017874628305435,
0.13078998029232025,
-0.13961222767829895,
-0.08217252790927887,
0.04953719303011894,
-0.15041466057300568,
0.03235960006713867,
0.11445982754230499,
0.002321717794984579,
-0.04658655822277069,
-0.03281686082482338,
0.3229322135448456,
0.08875089883804321,
-0.022760741412639618,
0.08591527491807938,
0.036167655140161514,
-0.10675781965255737,
-0.023992544040083885,
0.12553702294826508,
0.0014023531693965197,
-0.12904007732868195,
0.04186377301812172,
-0.07864208519458771,
0.007000922691076994,
-0.02876504883170128,
-0.04989699274301529,
0.059593554586172104,
0.006843453273177147,
0.2251795381307602,
0.006914746016263962,
0.055735088884830475,
-0.11082236468791962,
-0.05867975950241089,
-0.0779426321387291,
0.07507801055908203,
0.0784425437450409,
0.0026295639108866453,
0.003633009735494852,
0.21083058416843414,
0.03606381639838219,
-0.03333793580532074,
-0.15968286991119385,
-0.10279128700494766,
0.008095761761069298,
-0.10070259869098663,
-0.004450594075024128,
0.0345277413725853,
0.057989928871393204,
-0.012211378663778305,
-0.0029033345635980368,
0.00392092065885663,
-0.01357603631913662,
0.00999816320836544,
0.01625671237707138,
-0.1747792661190033,
-0.04807274416089058,
-0.04720522463321686,
0.07178888469934464,
-0.11746899038553238,
-0.013381109572947025,
0.06492210924625397,
0.11109042912721634,
-0.08474822342395782,
0.01612643525004387,
-0.007402435410767794,
0.07567283511161804,
0.000765566190239042,
0.05128280445933342,
0.13852697610855103,
-0.06037436053156853,
-0.046183954924345016,
-0.016319191083312035,
-0.027198202908039093,
0.0819450095295906,
0.11019387096166611,
-0.179920956492424,
-0.04117487743496895,
-0.11106175929307938,
-0.06084359064698219,
0.02092614769935608,
0.06957128643989563,
0.06421487033367157,
-0.000638629833701998,
0.033182404935359955,
-0.006187768653035164,
-0.04906972497701645,
-0.038578618317842484,
0.042293649166822433,
-0.00005852067261002958,
-0.011471443809568882,
0.20439015328884125,
0.206423819065094,
-0.10354796051979065,
0.13957169651985168,
0.12387573719024658,
-0.018670199438929558,
0.13557806611061096,
0.011147004552185535,
-0.03177458420395851,
-0.0014852306339889765,
-0.0054148780182003975,
-0.009598706848919392,
0.21520113945007324,
-0.14923295378684998,
0.04859697073698044,
0.0697602704167366,
-0.007621692959219217,
0.040140897035598755,
-0.10525914281606674,
-0.09224190562963486,
-0.06671212613582611,
-0.005606913473457098,
-0.1984565705060959,
0.07448898255825043,
-0.020574241876602173,
0.12660184502601624,
-0.0008833115571178496,
-0.1755349338054657,
-0.0010873511200770736,
0.03545314446091652,
-0.02861844189465046,
0.14397072792053223,
-0.0781208723783493,
-0.15229563415050507,
-0.08902761340141296,
-0.028055492788553238,
-0.13105134665966034,
-0.05893843248486519,
-0.00933364313095808,
-0.1038096696138382,
0.00029278494184836745,
0.08769543468952179,
0.11881647258996964,
-0.21572907269001007,
0.004328190814703703,
-0.18984940648078918,
-0.0037834574468433857,
-0.12331120669841766,
-0.02193436771631241,
-0.042179860174655914,
-0.11446963995695114,
0.1793067306280136,
0.08209202438592911,
-0.10498927533626556,
0.1227816641330719,
0.18144582211971283,
0.027403568848967552,
0.05723174661397934,
-0.037824179977178574,
0.2174157202243805,
-0.1542561948299408,
0.04387504607439041,
0.07959122955799103,
-0.09491505473852158,
0.04888923093676567,
0.11782164126634598,
0.1095445305109024,
-0.01921665109694004,
-0.08207105845212936,
0.03760986402630806,
-0.10611066222190857,
-0.13742990791797638,
-0.18908008933067322,
-0.12936614453792572,
0.11639370024204254,
0.021891053766012192,
0.04596749693155289,
0.03119952231645584,
0.007246337831020355,
0.021691298112273216,
-0.2281372994184494,
-0.010787713341414928,
0.0357055701315403,
0.15265792608261108,
0.03765658661723137,
0.031548283994197845,
-0.022239865735173225,
-0.09886478632688522,
0.0768381804227829,
0.01902618259191513,
0.07617945969104767,
0.1556222140789032,
0.022320039570331573,
0.06334199756383896,
0.19060352444648743,
0.11897289007902145,
0.12431982904672623,
0.02267281897366047,
0.0022890467662364244,
-0.013698441907763481,
-0.052436575293540955,
0.1261099874973297,
-0.03343109041452408,
0.03466223552823067,
-0.06867730617523193,
-0.027596021071076393,
-0.2099262923002243,
0.11170841753482819,
-0.0580645427107811,
0.08075122535228729,
-0.10957829654216766,
0.030093012377619743,
0.00989033654332161,
0.06438013911247253,
-0.0031562219373881817,
0.05175376310944557,
-0.12348318099975586,
-0.10942786186933517,
0.10989899933338165,
-0.027687715366482735,
0.1081710159778595,
0.04148057475686073,
0.07233183085918427,
0.009295905940234661,
0.019060738384723663,
-0.026886513456702232,
0.11480985581874847,
-0.0724189355969429,
0.30532771348953247,
-0.019582845270633698,
-0.07026210427284241,
-0.031118301674723625,
-0.09105366468429565,
0.035341981798410416,
0.12957267463207245,
0.1746760457754135,
0.06468271464109421,
-0.11109376698732376,
-0.12528330087661743,
-0.05719304084777832,
0.014724919572472572,
0.11807609349489212,
-0.07431425899267197,
-0.004056679550558329,
0.036165572702884674,
0.06289654225111008,
-0.046066656708717346,
0.06314714252948761,
0.012853221967816353,
-0.05246904864907265,
0.05152333155274391,
-0.050535064190626144,
-0.14960256218910217,
0.06331755965948105,
-0.12147097289562225,
-0.11414168775081635,
0.08476588875055313,
-0.022311881184577942,
-0.026474831625819206,
-0.13281533122062683,
-0.09473028033971786,
0.17798715829849243,
-0.07922832667827606,
-0.0176071934401989,
-0.06038394570350647,
0.016308732330799103,
-0.10019313544034958,
-0.12466295063495636,
0.07061030715703964,
-0.021405549719929695,
-0.026489006355404854,
-0.007086559664458036,
0.14325805008411407,
0.03137762472033501,
0.0591849721968174,
-0.0046478440053761005,
0.09432245045900345,
-0.0658707395195961,
-0.16201063990592957,
0.09148313850164413,
-0.15067419409751892,
-0.08629876375198364,
0.054617609828710556,
0.04185330122709274,
0.134198397397995,
-0.05861013010144234,
0.013868400827050209,
0.15512605011463165,
0.25306880474090576,
-0.07981466501951218,
0.16673700511455536,
0.0721479058265686,
-0.020959077402949333,
-0.2105787694454193,
-0.0291287824511528,
-0.07535407692193985,
-0.04567068815231323,
0.011648887768387794,
-0.13494917750358582,
0.11276061087846756,
0.11434191465377808,
-0.022861303761601448,
0.10635305941104889,
-0.22845414280891418,
-0.04167792946100235,
0.14990603923797607,
-0.025748882442712784,
0.33821046352386475,
-0.0015371774788945913,
-0.016681579872965813,
0.017216794192790985,
-0.0811842679977417,
0.12555570900440216,
0.06606686115264893,
0.03037283569574356,
-0.06957440823316574,
0.03109646402299404,
0.027083469554781914,
-0.040919654071331024,
0.18054480850696564,
0.08240672945976257,
0.004256286658346653,
-0.0066063362173736095,
-0.15718050301074982,
0.0899377390742302,
0.0014254922280088067,
-0.01922917179763317,
0.02143819071352482,
-0.045477528125047684,
-0.23389625549316406,
-0.03328106179833412,
-0.11945287883281708,
0.12480290979146957,
-0.048463162034749985,
-0.034425314515829086,
-0.060034509748220444,
0.038511697202920914,
-0.07647807896137238,
-0.019475363194942474,
0.18254853785037994,
-0.0034681339748203754,
0.09666342288255692,
-0.154423788189888,
0.01977900229394436,
0.03301467373967171,
-0.20013728737831116,
-0.014348906464874744,
-0.03061322309076786,
0.06467141211032867,
-0.054848216474056244,
-0.056809160858392715,
0.1702757030725479,
0.02841491438448429,
-0.045670036226511,
0.1010141372680664,
-0.04435563459992409,
-0.0714666098356247,
0.09795061498880386,
-0.23913545906543732,
0.0028090698178857565,
0.016042174771428108,
-0.18465450406074524,
0.1109628677368164,
0.014789901673793793,
0.029277388006448746,
-0.006422065198421478,
-0.01835397444665432,
0.07661575078964233,
-0.04506764933466911,
-0.09461359679698944,
0.03934590891003609,
0.0794440507888794,
0.05042623355984688,
-0.10142267495393753,
0.046860650181770325,
0.043867941945791245,
-0.072186678647995,
-0.09964749217033386,
0.1764192432165146,
-0.12382546067237854,
-0.06926453113555908,
0.05237431451678276,
0.12913069128990173,
-0.16611090302467346,
-0.013350552879273891,
-0.042092785239219666,
-0.14499622583389282,
0.032873015850782394,
0.18684546649456024,
0.12425050139427185,
-0.020642345771193504,
-0.06588947772979736,
-0.013509861193597317,
-0.013235912658274174,
0.019722262397408485,
0.017726385965943336,
0.01226401049643755,
-0.06699362397193909,
0.00642326008528471,
0.025754135102033615,
0.06984370201826096,
-0.07875426113605499,
-0.05679142847657204,
-0.18470405042171478,
0.06814853847026825,
-0.014521343633532524,
-0.04944970831274986,
-0.0540793277323246,
-0.004999183118343353,
-0.054820716381073,
-0.03618334233760834,
-0.05809955671429634,
-0.05079140141606331,
-0.08317184448242188,
0.05589424818754196,
0.01856839470565319,
0.057878948748111725,
-0.010065704584121704,
0.008388837799429893,
0.10059089213609695,
-0.03382015600800514,
0.1041388213634491,
0.13696356117725372,
0.004840532783418894,
0.11707964539527893,
-0.08060089498758316,
-0.012240809388458729,
0.008719420060515404,
0.022819023579359055,
0.020409636199474335,
-0.003005032893270254,
-0.005001664161682129,
-0.03859210014343262,
-0.05098247155547142,
0.09772700071334839,
-0.07085631787776947,
-0.06328875571489334,
-0.02647746354341507,
0.017323626205325127,
-0.23368793725967407,
0.0041717905551195145,
-0.08492331951856613,
0.09834568202495575,
0.04223523661494255,
0.07682909816503525,
0.06759960204362869,
0.10950469970703125,
-0.011934582144021988,
-0.02156263403594494,
-0.017685402184724808,
-0.13215191662311554,
-0.02308407798409462,
-0.011601848527789116,
0.04914361983537674,
-0.04195060580968857,
0.24747933447360992,
0.08474640548229218,
0.019610662013292313,
0.02939266711473465,
0.1649068146944046,
-0.03527233749628067,
0.03714601323008537,
0.15677833557128906,
0.07822903990745544,
-0.021745765581727028,
-0.018029849976301193,
0.04356304556131363,
-0.04153510183095932,
0.11700379848480225,
0.1819300353527069,
0.11208481341600418,
0.1202036514878273,
-0.03260164335370064,
0.057423513382673264,
0.01785079762339592,
-0.08949320018291473,
-0.038469016551971436,
0.10697463154792786,
0.023412493988871574,
0.05521624535322189,
0.1521114706993103,
0.15692593157291412,
-0.06547697633504868,
0.10983153432607651,
0.000328326626913622,
-0.05504964664578438,
-0.120646633207798,
-0.21175862848758698,
0.008664209395647049,
-0.05376899614930153,
0.008553901687264442,
-0.07247801870107651,
-0.06433059275150299,
0.19364584982395172,
0.05651175603270531,
0.012647896073758602,
-0.03401834890246391,
0.0166479405015707,
-0.040298447012901306,
0.07080967724323273,
-0.021541349589824677,
-0.004950907547026873,
-0.02188526839017868,
-0.03876616060733795,
-0.08785415440797806,
-0.08650341629981995,
-0.011957230977714062,
-0.02449297346174717,
-0.06354515254497528,
-0.04435613006353378,
-0.1513243019580841,
-0.09867426753044128,
0.010130972601473331,
-0.005823118612170219,
0.005520687438547611,
0.10135961323976517,
0.01807621493935585,
-0.0007955647888593376,
-0.022848695516586304,
0.1272982954978943,
0.0922088623046875,
-0.04189787805080414,
0.000527294585481286,
0.18342670798301697,
-0.059975747019052505,
0.08819530159235,
-0.014435073360800743,
0.047394558787345886,
-0.02194855734705925,
0.23298008739948273,
0.23783233761787415,
-0.007292172405868769,
0.0023991165217012167,
0.05074595287442207,
0.025381319224834442,
0.16028302907943726,
0.07130024582147598,
0.007701194379478693,
0.2542094588279724,
-0.07986745983362198,
-0.013820704072713852,
-0.1955336183309555,
0.0008457649382762611,
-0.07822384685277939,
0.14529956877231598,
0.17243199050426483,
-0.10683132708072662,
-0.14026398956775665,
0.16894036531448364,
-0.22182123363018036,
0.16903012990951538,
0.03272901475429535,
-0.1260138601064682,
-0.14928509294986725,
-0.03585410863161087,
0.07745954394340515,
0.03181421384215355,
0.08923989534378052,
-0.12006756663322449,
-0.15967443585395813,
-0.1105002835392952,
0.07997255027294159,
-0.174130380153656,
-0.16554011404514313,
0.05605284124612808,
0.05822077393531799,
0.08703406900167465,
-0.012324735522270203,
0.09514007717370987,
0.056177061051130295,
0.06479151546955109,
-0.013508131727576256,
-0.07499122619628906,
0.05608809366822243,
-0.0027613313868641853,
-0.09672769159078598,
-0.009060238488018513,
0.013073291629552841,
-0.14555101096630096,
0.08909046649932861,
-0.12050917744636536,
0.02274351567029953,
0.09924700856208801,
-0.08967780321836472,
0.059890005737543106,
0.14240802824497223,
-0.07970528304576874,
-0.0015918626450002193,
0.12727206945419312,
-0.006383480038493872,
-0.06312895566225052,
-0.09300482273101807,
-0.020056787878274918,
0.031293462961912155,
-0.06306634098291397,
-0.0724574476480484,
-0.042737264186143875,
-0.06230088695883751,
0.024608517065644264,
-0.07831469923257828,
0.017436273396015167,
-0.02204200066626072,
-0.10093991458415985,
0.0905577763915062,
-0.04486663267016411,
0.026962678879499435,
-0.013319761492311954,
-0.028975630179047585,
-0.005477134138345718,
-0.05402349308133125,
0.07128262519836426,
0.04624168947339058,
-0.05433722585439682,
-0.06424877047538757
] |
null | null | flair | ## Polish Flair Model -- Noun and Verb Phrases
| {"language": "pl", "tags": ["flair", "token-classification", "sequence-tagger-model"], "widget": [{"text": "Jan Brzechwa - polski poeta i adwokat \u017cydowskiego pochodzenia, autor bajek i wierszy dla dzieci, satyrycznych tekst\u00f3w dla doros\u0142ych, a tak\u017ce t\u0142umacz literatury rosyjskiej."}]} | token-classification | clarin-knext/morpho-flair-chunk | [
"flair",
"pytorch",
"token-classification",
"sequence-tagger-model",
"pl",
"region:us"
] | 2024-02-08T08:12:17+00:00 | [] | [
"pl"
] | TAGS
#flair #pytorch #token-classification #sequence-tagger-model #pl #region-us
| ## Polish Flair Model -- Noun and Verb Phrases
| [
"## Polish Flair Model -- Noun and Verb Phrases"
] | [
"TAGS\n#flair #pytorch #token-classification #sequence-tagger-model #pl #region-us \n",
"## Polish Flair Model -- Noun and Verb Phrases"
] | [
30,
15
] | [
"passage: TAGS\n#flair #pytorch #token-classification #sequence-tagger-model #pl #region-us \n## Polish Flair Model -- Noun and Verb Phrases"
] | [
0.011189489625394344,
-0.09204304218292236,
-0.010384518653154373,
0.04451918229460716,
0.051706843078136444,
0.06485366821289062,
0.046523187309503555,
0.012322756461799145,
0.17093777656555176,
0.013790138997137547,
0.13328073918819427,
0.10259414464235306,
-0.030974626541137695,
-0.06999489665031433,
0.027539536356925964,
-0.3708871006965637,
0.08805481344461441,
-0.007723347283899784,
-0.06577328592538834,
0.15143148601055145,
0.1123742088675499,
-0.01352839544415474,
-0.005553105846047401,
0.05971209704875946,
-0.022375890985131264,
0.057617682963609695,
0.013611985370516777,
-0.065944142639637,
0.12791430950164795,
-0.04038199037313461,
0.12035757303237915,
-0.04934278130531311,
-0.009220074862241745,
-0.15687112510204315,
-0.00941114779561758,
-0.05461447685956955,
-0.023310862481594086,
0.01594165340065956,
0.0476776659488678,
-0.13936454057693481,
0.18510349094867706,
-0.016729895025491714,
0.0773644670844078,
-0.009772054851055145,
-0.11877000331878662,
-0.2028358280658722,
0.036677826195955276,
0.08447322249412537,
0.004539909306913614,
0.0862400159239769,
-0.07673995941877365,
0.1091281995177269,
-0.2133992612361908,
0.02515675500035286,
-0.00120295281521976,
-0.16111086308956146,
-0.04984161630272865,
0.00775827094912529,
0.06304121017456055,
0.08164002746343613,
-0.149583101272583,
0.043878111988306046,
-0.05071097984910011,
0.053161267191171646,
-0.06382419914007187,
-0.04922790080308914,
0.03006351739168167,
0.08754914999008179,
-0.18101544678211212,
-0.04152144119143486,
0.20709647238254547,
-0.02239440195262432,
0.0028717645909637213,
-0.09618321061134338,
-0.0029317308217287064,
0.028088750317692757,
0.007235860452055931,
-0.01668882556259632,
-0.0358833372592926,
0.05183645337820053,
0.21950678527355194,
0.023636117577552795,
-0.06823660433292389,
0.04384459927678108,
-0.0641927570104599,
0.2697998285293579,
0.028161995112895966,
0.08499658107757568,
-0.09105593711137772,
-0.04520069435238838,
-0.028715867549180984,
-0.04606698825955391,
0.08541538566350937,
-0.06731165945529938,
-0.031771447509527206,
0.02882053330540657,
-0.015261550433933735,
0.07120094448328018,
0.05451184883713722,
0.11351703107357025,
-0.008792909793555737,
0.008802568539977074,
0.07991751283407211,
0.13455022871494293,
0.13641582429409027,
0.09916091710329056,
0.013244926929473877,
-0.06642968952655792,
-0.10900046676397324,
-0.16277050971984863,
-0.007478732615709305,
-0.024345237761735916,
-0.11124350875616074,
-0.03163904696702957,
-0.09439614415168762,
0.04546471685171127,
-0.03529830276966095,
0.05309842899441719,
-0.023825502023100853,
-0.045449141412973404,
-0.033006347715854645,
-0.007212605327367783,
-0.025552351027727127,
0.054011404514312744,
-0.03629167750477791,
0.24063041806221008,
-0.2040623128414154,
0.033122263848781586,
-0.05545591190457344,
0.08652637153863907,
-0.049385033547878265,
0.04902399331331253,
0.012920720502734184,
-0.11294936388731003,
0.03602450713515282,
-0.055406730622053146,
0.05070653557777405,
-0.08104278892278671,
0.012561092153191566,
0.013854566030204296,
-0.014057321473956108,
-0.045425571501255035,
-0.004198057577013969,
-0.02616836503148079,
0.0010988269932568073,
0.029668251052498817,
0.014650438912212849,
-0.08004114776849747,
-0.0816691666841507,
0.1351243257522583,
-0.17572402954101562,
-0.017494460567831993,
-0.11272906512022018,
0.008369256742298603,
-0.2001987248659134,
-0.06274464726448059,
-0.25874802470207214,
-0.07941830158233643,
-0.05586427450180054,
0.030474888160824776,
-0.05942567437887192,
-0.06071413308382034,
-0.005028846208006144,
0.040249090641736984,
-0.15858854353427887,
0.1392287015914917,
-0.16446253657341003,
-0.07637699693441391,
0.09133590012788773,
-0.14015665650367737,
0.04247625917196274,
0.17305633425712585,
-0.00008320815686602145,
0.07110971957445145,
0.05379441753029823,
0.31612563133239746,
-0.019316980615258217,
-0.11233192682266235,
0.13822120428085327,
0.060222938656806946,
-0.07456544041633606,
0.1388036459684372,
0.1447240710258484,
-0.07424482703208923,
-0.06340736150741577,
0.016017641872167587,
-0.07009533792734146,
0.04094209894537926,
0.0027876696549355984,
-0.06039942055940628,
0.013750948011875153,
0.03758864104747772,
0.24653911590576172,
-0.011032755486667156,
0.0027476069517433643,
-0.13416986167430878,
-0.08872698992490768,
-0.0657946839928627,
0.0813087448477745,
0.06002446264028549,
-0.01031133346259594,
-0.03994510695338249,
0.14778898656368256,
0.10133359581232071,
-0.02322102151811123,
-0.11369618773460388,
0.0019813149701803923,
-0.0185944102704525,
0.037550684064626694,
-0.009562691673636436,
0.06601947546005249,
0.056947071105241776,
-0.055982522666454315,
-0.017065852880477905,
0.03758314624428749,
-0.04190448671579361,
0.02445388399064541,
0.01530668418854475,
-0.13972775638103485,
-0.003429500153288245,
-0.04785076901316643,
0.09293139725923538,
-0.0714167058467865,
-0.016750384122133255,
0.11911375820636749,
0.07518871873617172,
-0.07875902205705643,
0.0203612819314003,
-0.045732609927654266,
0.06961365789175034,
0.024614466354250908,
0.05753479152917862,
0.15257064998149872,
-0.05197328329086304,
-0.00723747443407774,
0.06445251405239105,
-0.008329945616424084,
0.0040871393866837025,
0.12977637350559235,
-0.23490220308303833,
-0.017801426351070404,
-0.18478092551231384,
-0.06997566670179367,
0.06508784741163254,
0.011201724410057068,
-0.09548120945692062,
0.10371074825525284,
0.022776182740926743,
-0.05899783968925476,
-0.06247825548052788,
0.025708412751555443,
0.005062991753220558,
-0.00467658881098032,
-0.0424395352602005,
0.15739881992340088,
0.09890032559633255,
-0.11063595861196518,
0.17533232271671295,
0.23705416917800903,
0.0627768337726593,
0.22958208620548248,
-0.014520405791699886,
-0.024583108723163605,
-0.027189260348677635,
-0.001881931209936738,
-0.09569060057401657,
0.19257989525794983,
-0.23416262865066528,
0.05333369970321655,
0.05596400797367096,
-0.04678263142704964,
0.0026385642122477293,
-0.07186613231897354,
-0.10135943442583084,
-0.06294585019350052,
-0.0018458556151017547,
-0.12965546548366547,
0.0632811188697815,
0.027480242773890495,
0.10900815576314926,
0.007164526730775833,
-0.13670042157173157,
0.05726250261068344,
0.03274860233068466,
-0.044450026005506516,
0.06962387263774872,
-0.17405644059181213,
-0.21786223351955414,
-0.051051363348960876,
-0.021381979808211327,
-0.05621573328971863,
-0.05026701092720032,
0.01100490614771843,
-0.14317749440670013,
0.00046510720858350396,
0.12879620492458344,
0.2771083116531372,
-0.173888698220253,
-0.09244900941848755,
-0.16826625168323517,
0.04237241670489311,
-0.14612114429473877,
-0.021674323827028275,
-0.07884033024311066,
-0.08892542868852615,
0.1119137704372406,
0.03366825357079506,
-0.09191238880157471,
0.051633406430482864,
0.17080247402191162,
0.06477949768304825,
0.023952914401888847,
-0.08174162358045578,
0.3006591796875,
-0.19051823019981384,
0.045644428580999374,
0.10025914758443832,
-0.08554010093212128,
0.025347081944346428,
0.13441261649131775,
0.07208064943552017,
0.04063086956739426,
-0.06588046252727509,
0.03768289461731911,
-0.058353886008262634,
-0.10244549810886383,
-0.216264545917511,
-0.08408164232969284,
0.08335094898939133,
0.013625942170619965,
0.05254437401890755,
0.04897775501012802,
0.013703841716051102,
-0.01657794415950775,
-0.287625789642334,
0.03316333517432213,
0.012067103758454323,
0.2575603425502777,
-0.026469524949789047,
0.0366809256374836,
0.009756498970091343,
-0.10635341703891754,
0.049379218369722366,
0.04131283611059189,
-0.03478904813528061,
0.18744538724422455,
0.0012300878297537565,
0.07984837889671326,
0.18109166622161865,
0.05119027942419052,
0.04845970496535301,
0.005859500262886286,
0.003750763600692153,
-0.03313935920596123,
-0.0830879956483841,
0.1093124970793724,
-0.01889754645526409,
0.15342096984386444,
-0.049774184823036194,
-0.02666764333844185,
-0.24731339514255524,
0.17075595259666443,
-0.023370439186692238,
0.061751097440719604,
0.019014304503798485,
0.016064850613474846,
0.048941608518362045,
0.011758354492485523,
-0.05319942161440849,
0.08668985217809677,
-0.052244678139686584,
-0.18145014345645905,
0.038431260734796524,
-0.03909042477607727,
0.08938118070363998,
0.0953572615981102,
0.1131114661693573,
-0.001883943215943873,
-0.053077660501003265,
-0.0035645863972604275,
0.10328266769647598,
-0.12513911724090576,
0.2643187642097473,
0.014725013636052608,
-0.05449698865413666,
-0.08283215016126633,
-0.07443435490131378,
0.022200386971235275,
0.06618717312812805,
0.1936284899711609,
0.05910463258624077,
-0.07214154303073883,
-0.11705977469682693,
-0.015654223039746284,
0.006881617940962315,
0.11938590556383133,
-0.0510014109313488,
-0.008930833078920841,
-0.02487550862133503,
0.09917297959327698,
-0.03346752002835274,
0.1227608323097229,
0.0053555904887616634,
-0.08125922828912735,
0.033305324614048004,
-0.004228218924254179,
-0.1011892557144165,
0.06003878638148308,
-0.08195953071117401,
-0.09473017603158951,
-0.00241048913449049,
-0.024108396843075752,
-0.07241380214691162,
-0.12008751928806305,
-0.1278984546661377,
0.1377926915884018,
-0.04139385372400284,
-0.0668388158082962,
-0.054441187530756,
0.03876625373959541,
-0.1393597573041916,
-0.06612633913755417,
0.06251750886440277,
0.0046793026849627495,
-0.02192912995815277,
0.011081879027187824,
0.1583355814218521,
0.026348581537604332,
0.0825771912932396,
0.024425456300377846,
0.08695068210363388,
-0.06480178982019424,
-0.13008534908294678,
0.04683791473507881,
-0.1885223537683487,
-0.12250811606645584,
0.06583041697740555,
-0.09504438936710358,
0.11983851343393326,
-0.12459119409322739,
-0.04346142336726189,
0.22315248847007751,
0.11857432126998901,
-0.04598686471581459,
0.18604379892349243,
0.08723580092191696,
-0.029963921755552292,
-0.22915896773338318,
-0.014061294496059418,
-0.0630345493555069,
-0.042919695377349854,
0.04488477483391762,
-0.15270483493804932,
0.0912347137928009,
0.08920018374919891,
-0.006973117124289274,
0.047226015478372574,
-0.2318822741508484,
-0.02301548607647419,
0.20255166292190552,
-0.05995362997055054,
0.17555133998394012,
0.005182808730751276,
-0.024508006870746613,
0.009237024933099747,
-0.07113020122051239,
0.14300338923931122,
0.04006442800164223,
0.03725314885377884,
-0.0036640537437051535,
0.09809145331382751,
0.03540293127298355,
-0.002931230468675494,
0.19728055596351624,
0.06395548582077026,
-0.023824388161301613,
-0.00922100618481636,
-0.2399931102991104,
0.09206882864236832,
0.06061733886599541,
-0.03504092991352081,
-0.03492402657866478,
-0.08199410885572433,
-0.2547008991241455,
-0.027416205033659935,
-0.15237301588058472,
0.08695194870233536,
-0.07082813233137131,
-0.08171293884515762,
-0.04876447841525078,
0.06479711085557938,
-0.0913870707154274,
-0.07815281301736832,
0.09309620410203934,
-0.06663773208856583,
0.11323609948158264,
-0.11197200417518616,
-0.002473645145073533,
0.09125365316867828,
-0.27911296486854553,
0.011316554620862007,
-0.04714757576584816,
0.08757137507200241,
-0.003659820416942239,
-0.04040120914578438,
0.08961991965770721,
-0.014805983752012253,
0.023368045687675476,
0.05694010853767395,
-0.005545181222259998,
-0.06729292124509811,
0.028049787506461143,
-0.14887090027332306,
-0.021288102492690086,
-0.031835827976465225,
-0.12816424667835236,
0.08928148448467255,
-0.11218740791082382,
0.09830760210752487,
0.031158344820141792,
-0.045651763677597046,
0.03221292048692703,
-0.03485732525587082,
-0.11279910057783127,
0.03258324787020683,
0.06664406508207321,
0.043575990945100784,
-0.07774993777275085,
0.05057014152407646,
-0.01689589023590088,
-0.15029750764369965,
-0.03528490290045738,
0.1697346568107605,
-0.07329380512237549,
-0.07080122083425522,
-0.04816574603319168,
0.08073843270540237,
-0.1341339349746704,
-0.006728657521307468,
0.01150891650468111,
-0.22360733151435852,
0.0024280017241835594,
0.15427613258361816,
0.11851871013641357,
0.019454369321465492,
-0.09576524049043655,
-0.029886405915021896,
0.03864623233675957,
0.046974021941423416,
0.017306743189692497,
-0.11620505899190903,
0.023060409352183342,
0.07405932992696762,
0.03786274418234825,
0.07018886506557465,
-0.08023080229759216,
-0.08533124625682831,
-0.16152308881282806,
0.08582261204719543,
-0.043514978140592575,
-0.057848088443279266,
-0.014727778732776642,
-0.042910363525152206,
-0.0008179565193131566,
-0.018262188881635666,
-0.022553816437721252,
-0.07325956225395203,
-0.0939721018075943,
0.08794265240430832,
0.06465411931276321,
0.0696071982383728,
0.02340824529528618,
0.001197259989567101,
0.10459587723016739,
-0.03360005095601082,
0.08374663442373276,
0.19069437682628632,
-0.08632217347621918,
0.12270965427160263,
-0.12573730945587158,
0.0016633907798677683,
0.031020071357488632,
-0.01956823840737343,
0.007031771820038557,
0.01574230194091797,
-0.004038272425532341,
-0.03271941840648651,
0.016665039584040642,
0.11211439967155457,
-0.008834573440253735,
-0.052338939160108566,
0.02186889946460724,
0.13539719581604004,
-0.27434781193733215,
-0.005423991475254297,
-0.07644406706094742,
0.10320136696100235,
-0.0021467162296175957,
0.10137663036584854,
0.03650223836302757,
0.10621736198663712,
0.038751792162656784,
0.00883209053426981,
-0.03255537897348404,
-0.07065055519342422,
-0.049853578209877014,
-0.018188554793596268,
0.0216671209782362,
-0.02143966220319271,
0.23915086686611176,
0.09461962431669235,
0.0377318374812603,
0.08719956874847412,
0.23154087364673615,
-0.0877106636762619,
0.03962910175323486,
0.11158183217048645,
0.0664232075214386,
-0.02429528906941414,
-0.07885675877332687,
0.00035340810427442193,
-0.04967352747917175,
0.13947321474552155,
0.047584693878889084,
0.1822699010372162,
0.2432783842086792,
0.01782188192009926,
0.08005635440349579,
0.06357678771018982,
0.03797582909464836,
-0.03941703960299492,
0.20003026723861694,
0.015073077753186226,
-0.08958879113197327,
0.2532445788383484,
0.08673454821109772,
-0.07616684585809708,
0.08658517897129059,
-0.03871768340468407,
-0.07971552014350891,
-0.08759558945894241,
-0.13631217181682587,
0.044449642300605774,
-0.002773349406197667,
-0.02212543413043022,
-0.06319229304790497,
-0.02197340317070484,
0.12021955102682114,
0.04910910502076149,
-0.04270341619849205,
-0.040265388786792755,
-0.03366803750395775,
-0.12302055954933167,
0.08899201452732086,
0.007354328874498606,
0.10682694613933563,
-0.08256661146879196,
-0.049247439950704575,
-0.08245468884706497,
-0.036084745079278946,
0.00958541315048933,
0.013461167924106121,
-0.09213869273662567,
-0.13500744104385376,
-0.22852882742881775,
-0.09260830283164978,
0.006460551172494888,
0.010879147797822952,
-0.043244026601314545,
0.1431642770767212,
0.027216825634241104,
0.012136148288846016,
-0.00722777284681797,
0.14918851852416992,
0.05774258077144623,
-0.1250983476638794,
-0.06242769584059715,
0.11548171937465668,
-0.08481652289628983,
0.10859762877225876,
-0.020945586264133453,
0.029384098947048187,
-0.03298891708254814,
0.2091919630765915,
0.2781068682670593,
-0.03878007456660271,
-0.0018350993050262332,
0.04920591786503792,
0.03116365522146225,
0.16351166367530823,
0.06490854173898697,
-0.02315208874642849,
0.25235483050346375,
-0.1086672693490982,
0.07271046936511993,
-0.18188589811325073,
-0.03885788470506668,
-0.05246230587363243,
0.12625229358673096,
0.13541822135448456,
-0.09388045966625214,
-0.12962999939918518,
0.1907455325126648,
-0.17652975022792816,
0.1891506016254425,
0.013179965317249298,
-0.13285396993160248,
-0.08432227373123169,
-0.03630603477358818,
0.09515025466680527,
0.08251907676458359,
0.13396859169006348,
-0.06262560188770294,
-0.16864003241062164,
-0.02407218888401985,
0.05723457783460617,
-0.16529706120491028,
-0.15256425738334656,
0.12959372997283936,
0.022203795611858368,
0.09094677865505219,
-0.004697701428085566,
0.10029029101133347,
0.06736771017313004,
0.09837382286787033,
0.06881847977638245,
-0.06995519250631332,
0.049234263598918915,
-0.02922034077346325,
-0.11631672829389572,
-0.07640036195516586,
0.02851487696170807,
-0.06753525882959366,
0.0801931694149971,
-0.16043472290039062,
0.059413593262434006,
0.12159265577793121,
-0.10343889147043228,
0.06736420840024948,
0.15771527588367462,
-0.12012521922588348,
-0.03295108675956726,
0.10559921711683273,
0.04549647867679596,
-0.06810209900140762,
-0.10125360637903214,
-0.036136507987976074,
0.05580339580774307,
-0.09687017649412155,
-0.04867375269532204,
-0.0611279159784317,
-0.046709757298231125,
0.062065087258815765,
-0.1343936175107956,
-0.03885022923350334,
-0.07227747142314911,
-0.04549426585435867,
0.11675848066806793,
-0.04123202711343765,
0.023640546947717667,
-0.030374886468052864,
-0.022502224892377853,
0.028725579380989075,
0.03766762465238571,
0.11284596472978592,
0.06038947403430939,
-0.010775831528007984,
0.001966105541214347
] |
null | null | flair | ## Polish Flair Model -- Agreement Phrases
| {"language": "pl", "tags": ["flair", "token-classification", "sequence-tagger-model"], "widget": [{"text": "Jan Brzechwa - polski poeta i adwokat \u017cydowskiego pochodzenia, autor bajek i wierszy dla dzieci, satyrycznych tekst\u00f3w dla doros\u0142ych, a tak\u017ce t\u0142umacz literatury rosyjskiej."}]} | token-classification | clarin-knext/morpho-flair-agp | [
"flair",
"pytorch",
"token-classification",
"sequence-tagger-model",
"pl",
"region:us"
] | 2024-02-08T08:12:22+00:00 | [] | [
"pl"
] | TAGS
#flair #pytorch #token-classification #sequence-tagger-model #pl #region-us
| ## Polish Flair Model -- Agreement Phrases
| [
"## Polish Flair Model -- Agreement Phrases"
] | [
"TAGS\n#flair #pytorch #token-classification #sequence-tagger-model #pl #region-us \n",
"## Polish Flair Model -- Agreement Phrases"
] | [
30,
11
] | [
"passage: TAGS\n#flair #pytorch #token-classification #sequence-tagger-model #pl #region-us \n## Polish Flair Model -- Agreement Phrases"
] | [
-0.057900361716747284,
-0.004142229910939932,
-0.00849542859941721,
0.027784233912825584,
0.02315455488860607,
0.050423670560121536,
0.04848071187734604,
0.011507323011755943,
0.2095947265625,
0.05480126664042473,
0.09817160665988922,
0.1307573765516281,
0.0019333267118781805,
0.03982694447040558,
-0.0638568326830864,
-0.3065306842327118,
0.08891941606998444,
0.04391275346279144,
-0.0819375142455101,
0.14213897287845612,
0.15083178877830505,
-0.028851382434368134,
-0.02483413554728031,
0.047148481011390686,
-0.014371775090694427,
0.059691138565540314,
-0.020018626004457474,
-0.06746498495340347,
0.16137021780014038,
0.004761748481541872,
0.15664321184158325,
0.022679856047034264,
-0.09115342050790787,
-0.17972666025161743,
0.03384937718510628,
-0.054611433297395706,
-0.021217206493020058,
0.05370878055691719,
0.13719025254249573,
-0.14197203516960144,
0.17848315834999084,
-0.008671597577631474,
0.07253085821866989,
0.03913861885666847,
-0.11802627891302109,
-0.17859739065170288,
-0.009474510326981544,
0.10548493266105652,
0.05771581456065178,
-0.039318423718214035,
-0.06089727580547333,
0.12104013562202454,
-0.16385622322559357,
0.040026359260082245,
0.005436378065496683,
-0.2342100739479065,
-0.015851590782403946,
0.089994877576828,
-0.03192750737071037,
0.049637265503406525,
-0.06567363440990448,
0.05241142213344574,
-0.03494919836521149,
-0.013227101415395737,
-0.08395019918680191,
-0.03391978144645691,
0.05016368255019188,
0.03471306711435318,
-0.14361916482448578,
-0.0702684074640274,
0.2663300335407257,
0.008092324249446392,
0.0025929072871804237,
-0.12160523980855942,
-0.04262376204133034,
0.0497407503426075,
0.00685254717245698,
-0.03893055394291878,
-0.04742567241191864,
0.034393586218357086,
0.23180721700191498,
0.070089191198349,
-0.055092498660087585,
0.052860088646411896,
-0.1227787435054779,
0.3824501931667328,
0.045601263642311096,
0.06679098308086395,
-0.10044372081756592,
-0.018040375784039497,
-0.15803363919258118,
-0.0744338110089302,
0.0729740634560585,
-0.09503806382417679,
-0.04777342453598976,
0.02971472777426243,
-0.050699423998594284,
0.10692166537046432,
0.07919526845216751,
0.2087699919939041,
-0.03120500221848488,
0.020082099363207817,
0.10966376960277557,
0.12201299518346786,
0.12261314690113068,
0.048557862639427185,
-0.05553632229566574,
-0.07540135830640793,
-0.05753280222415924,
-0.13734038174152374,
0.03958200663328171,
-0.0523492805659771,
-0.08347101509571075,
-0.02441321685910225,
-0.03619007021188736,
0.03346468135714531,
-0.034691695123910904,
0.05212631821632385,
-0.0845910906791687,
0.0009890171932056546,
0.04676032438874245,
-0.016987212002277374,
0.0036478303372859955,
0.05255100876092911,
-0.039090804755687714,
0.2358703464269638,
-0.14777445793151855,
0.04119855910539627,
-0.034261204302310944,
0.1681826263666153,
-0.025357317179441452,
0.016455667093396187,
0.025596190243959427,
-0.0823657438158989,
0.06668035686016083,
-0.08752503991127014,
0.05523074418306351,
-0.04642263054847717,
0.09627124667167664,
0.05132409185171127,
-0.03699043393135071,
-0.020732128992676735,
-0.038382720202207565,
-0.02420431561768055,
0.07151626795530319,
-0.0015820523258298635,
-0.019205451011657715,
-0.08449260145425797,
-0.087879478931427,
0.10642506182193756,
-0.1626511514186859,
-0.024200720712542534,
-0.11878973990678787,
0.049574438482522964,
-0.16155178844928741,
-0.016969820484519005,
-0.23994077742099762,
-0.10068298131227493,
-0.09159573167562485,
0.05343124642968178,
-0.015154899097979069,
-0.06558723002672195,
0.02750226855278015,
0.01022691372781992,
-0.05647877976298332,
0.12077990919351578,
-0.031166773289442062,
-0.08768925070762634,
0.0773560106754303,
-0.16170383989810944,
0.019412562251091003,
0.17513006925582886,
0.026379436254501343,
-0.03594374284148216,
0.03932548686861992,
0.32670852541923523,
-0.020948311313986778,
-0.03691568225622177,
0.0302619356662035,
0.02854311652481556,
-0.060158539563417435,
0.06058108061552048,
0.13241010904312134,
-0.013330675661563873,
-0.06918053328990936,
0.04034922644495964,
0.048068925738334656,
0.02594001777470112,
-0.03747149556875229,
-0.03946219012141228,
0.0023638897109776735,
0.027603967115283012,
0.19526931643486023,
0.006809343118220568,
-0.03894717991352081,
-0.18320906162261963,
-0.08451806753873825,
-0.13430091738700867,
0.0759861096739769,
0.07983256131410599,
-0.04816802218556404,
-0.014120858162641525,
0.22840003669261932,
-0.025260569527745247,
-0.03142755478620529,
-0.09859617799520493,
-0.08778984099626541,
-0.022447509691119194,
0.029341433197259903,
-0.005607477854937315,
0.10852208733558655,
0.03779168799519539,
-0.06921042501926422,
-0.014073916710913181,
0.018472306430339813,
0.043215397745370865,
-0.03120936080813408,
0.06329911202192307,
-0.17650547623634338,
-0.02658839151263237,
-0.061623092740774155,
0.16080112755298615,
-0.07588405907154083,
0.015602747909724712,
0.13760536909103394,
0.19987112283706665,
-0.08925512433052063,
0.058051273226737976,
0.04241933673620224,
0.11862121522426605,
-0.027486568316817284,
0.07311868667602539,
0.13722927868366241,
-0.030921483412384987,
-0.07088153809309006,
0.0883282721042633,
-0.056052304804325104,
0.1260746270418167,
0.16273103654384613,
-0.2056637853384018,
-0.029216472059488297,
-0.14567388594150543,
-0.08527713268995285,
0.05758332088589668,
0.04193507134914398,
0.039196182042360306,
0.0051712351851165295,
0.016745202243328094,
0.006224409211426973,
-0.0767049565911293,
-0.04513444006443024,
0.04321729764342308,
-0.07455664873123169,
-0.05537192150950432,
0.1625889390707016,
0.08150521665811539,
-0.10340888053178787,
0.1733153909444809,
0.17421020567417145,
0.06863812357187271,
0.19226057827472687,
-0.001554598449729383,
0.0019291732460260391,
0.007264425978064537,
0.03131650388240814,
-0.06280286610126495,
0.18054553866386414,
-0.21495461463928223,
0.05458187684416771,
0.09282582253217697,
-0.006250158883631229,
0.007585084997117519,
-0.15129230916500092,
-0.0542314313352108,
-0.027677837759256363,
-0.021373862400650978,
-0.15699031949043274,
0.08915803581476212,
0.009025524370372295,
0.1621047556400299,
-0.04303367808461189,
-0.1515868455171585,
0.03090255707502365,
0.0182074885815382,
-0.03942398354411125,
0.07759209722280502,
-0.12845385074615479,
-0.2068658173084259,
-0.07046987861394882,
-0.022049633786082268,
-0.18789373338222504,
-0.05785322189331055,
0.033713970333337784,
-0.10969402641057968,
-0.03129188343882561,
0.13656286895275116,
0.17168036103248596,
-0.1945541799068451,
-0.01422318909317255,
-0.21551786363124847,
0.039122119545936584,
-0.11467556655406952,
-0.07385372370481491,
-0.11265628784894943,
-0.09180926531553268,
0.12013193964958191,
0.09746333211660385,
-0.10153952986001968,
0.08855772018432617,
0.18211866915225983,
0.001161977299489081,
0.020712634548544884,
-0.07768675684928894,
0.17114916443824768,
-0.14412783086299896,
0.05213576555252075,
0.11622567474842072,
-0.12810997664928436,
0.04569632560014725,
0.13294002413749695,
0.06353240460157394,
-0.024044979363679886,
-0.04716882109642029,
-0.014974198304116726,
-0.09874581545591354,
-0.1373426914215088,
-0.20233425498008728,
-0.12540356814861298,
0.09268280863761902,
-0.05146319046616554,
0.01056754868477583,
0.03089040331542492,
0.04897693917155266,
-0.06457850337028503,
-0.29196035861968994,
0.011540642008185387,
0.05376829952001572,
0.2673935890197754,
-0.015654923394322395,
0.06518130749464035,
-0.04155584052205086,
-0.06854785978794098,
0.10048333555459976,
0.04041580483317375,
0.0010160538367927074,
0.15171878039836884,
-0.021403394639492035,
0.050850216299295425,
0.203422412276268,
0.08385224640369415,
0.12185920774936676,
-0.031893834471702576,
0.041362397372722626,
-0.020660964772105217,
-0.060416411608457565,
0.025768466293811798,
-0.004179669078439474,
0.09627564996480942,
-0.04028617590665817,
-0.06528469920158386,
-0.31084954738616943,
0.1590721309185028,
0.006490962579846382,
0.14904439449310303,
-0.07530723512172699,
-0.00021365053544286638,
0.028456328436732292,
0.041940830647945404,
-0.04422585293650627,
0.053117264062166214,
-0.11527656763792038,
-0.11409375071525574,
0.1437811255455017,
-0.02285991795361042,
0.1113407164812088,
0.04061402380466461,
0.05119864642620087,
-0.0659688413143158,
-0.038417354226112366,
-0.06659494340419769,
0.09473398327827454,
-0.1774769276380539,
0.31648194789886475,
0.005159089807420969,
-0.1062120720744133,
-0.06673861294984818,
-0.04992055520415306,
0.0517037957906723,
0.1946837604045868,
0.12337006628513336,
0.04081536829471588,
-0.08295289427042007,
-0.12262272089719772,
-0.04645780846476555,
0.058009590953588486,
0.10223034769296646,
-0.04889580234885216,
0.0030781158711761236,
0.020036978647112846,
0.10311707109212875,
-0.03718148171901703,
0.06856852024793625,
0.01610828936100006,
-0.07906757295131683,
0.024990489706397057,
0.06546294689178467,
-0.033711981028318405,
0.04126366600394249,
-0.07819429785013199,
-0.12307475507259369,
0.04114504158496857,
0.0015427874168381095,
-0.11291149258613586,
-0.08922464400529861,
-0.14469078183174133,
0.11864307522773743,
-0.011434895917773247,
-0.023198546841740608,
-0.08023105561733246,
0.010393613949418068,
-0.1297301948070526,
-0.05510415509343147,
0.08390278369188309,
-0.02874240092933178,
-0.07007865607738495,
-0.010321516543626785,
0.21750260889530182,
-0.042821187525987625,
0.04190315678715706,
-0.014718891121447086,
0.10152320563793182,
-0.08791535347700119,
-0.08481799066066742,
0.050397057086229324,
-0.13103537261486053,
-0.0028876184951514006,
0.11796482652425766,
-0.010068141855299473,
0.11062503606081009,
-0.09962636977434158,
-0.043587520718574524,
0.13489046692848206,
0.23728229105472565,
-0.05816096067428589,
0.12541450560092926,
0.10600278526544571,
0.02983582764863968,
-0.18372802436351776,
-0.05347072705626488,
-0.08068456500768661,
-0.020552681758999825,
-0.019497282803058624,
-0.13128453493118286,
0.030434884130954742,
0.1042407974600792,
-0.03611895814538002,
0.07872200757265091,
-0.1509118527173996,
-0.0478992834687233,
0.15583521127700806,
-0.07091102004051208,
0.2692074179649353,
-0.007214666809886694,
-0.0000189628863154212,
0.02407858334481716,
-0.09914664179086685,
0.1197250708937645,
0.08226843923330307,
0.011493971571326256,
-0.004197725094854832,
0.11802183836698532,
0.0524667426943779,
-0.03590147942304611,
0.25032687187194824,
0.04048579931259155,
-0.006609512493014336,
-0.04688505455851555,
-0.10252515971660614,
0.058873213827610016,
0.03679509833455086,
0.018594950437545776,
-0.06614946573972702,
-0.07095477730035782,
-0.294604629278183,
-0.050967395305633545,
-0.08328086882829666,
0.10506343841552734,
-0.051651161164045334,
-0.10276177525520325,
-0.07501907646656036,
0.007522886153310537,
-0.1281009018421173,
-0.061898358166217804,
0.17928674817085266,
-0.07078540325164795,
0.1682293862104416,
-0.08649642765522003,
0.10582072287797928,
-0.08542364090681076,
-0.2138414829969406,
0.042162902653217316,
-0.05459149554371834,
0.08358027040958405,
-0.09649848192930222,
-0.05378282815217972,
0.12693119049072266,
-0.0005433239857666194,
-0.011958148330450058,
0.0839356854557991,
-0.019878389313817024,
-0.03289026767015457,
0.09979574382305145,
-0.23514170944690704,
-0.08447245508432388,
0.01653088442981243,
-0.09863452613353729,
0.12382657825946808,
-0.0007080621435306966,
0.04933773726224899,
0.013312408700585365,
-0.02180349826812744,
0.04466588795185089,
-0.04286263883113861,
-0.12050554156303406,
0.03391283005475998,
0.08025314658880234,
0.028204435482621193,
-0.09511698782444,
0.04920337349176407,
-0.04074200987815857,
-0.24007374048233032,
-0.08769374340772629,
0.05723905563354492,
-0.10894805192947388,
-0.0622674897313118,
-0.06569328159093857,
0.11113342642784119,
-0.1740535944700241,
-0.09780827909708023,
-0.06293514370918274,
-0.16715502738952637,
0.01908857747912407,
0.12881477177143097,
0.13140937685966492,
0.012254730798304081,
-0.021159622818231583,
-0.07388924807310104,
0.020782433450222015,
0.036306608468294144,
0.04605810344219208,
-0.016347402706742287,
-0.06379016488790512,
-0.006950202863663435,
0.011207069270312786,
0.04583409056067467,
-0.06865116208791733,
-0.06568705290555954,
-0.14134930074214935,
0.06007867306470871,
-0.030149443075060844,
-0.09768568724393845,
-0.09175083041191101,
-0.03782135248184204,
-0.026277663186192513,
-0.08494317531585693,
-0.01951274275779724,
-0.06292455643415451,
-0.08265446871519089,
0.10462231189012527,
0.04865887016057968,
0.11705633997917175,
0.0022869480308145285,
-0.018360348418354988,
0.10960297286510468,
-0.06947986036539078,
0.13433727622032166,
0.13977453112602234,
-0.11423131823539734,
0.09733761847019196,
-0.030061403289437294,
0.010845833458006382,
0.06301925331354141,
0.005844520404934883,
0.026435088366270065,
-0.00019177861395291984,
0.012470371089875698,
-0.008508964441716671,
-0.06301382929086685,
0.11303598433732986,
0.0069277286529541016,
-0.06799115985631943,
0.044273246079683304,
0.10726907849311829,
-0.22486981749534607,
0.011844572611153126,
-0.09855415672063828,
0.08993034064769745,
0.09309148043394089,
0.12179116159677505,
0.04931240528821945,
0.07272788137197495,
-0.022475581616163254,
0.007067936006933451,
-0.03957147151231766,
-0.15477891266345978,
-0.013908187858760357,
-0.01873951405286789,
0.029109923169016838,
-0.0021855675149708986,
0.28088557720184326,
0.0921916514635086,
0.043653905391693115,
0.09019163995981216,
0.178927481174469,
-0.016638396307826042,
0.05349426344037056,
0.09345009922981262,
0.03188658133149147,
0.022666573524475098,
-0.11826810985803604,
0.046388812363147736,
-0.02257932350039482,
0.04619767889380455,
0.07959829270839691,
0.13163810968399048,
0.17908190190792084,
-0.03377045691013336,
0.054349761456251144,
0.053939417004585266,
-0.02848576381802559,
-0.11810919642448425,
0.16434015333652496,
0.029874563217163086,
-0.017532290890812874,
0.22291001677513123,
0.15461887419223785,
-0.09477221220731735,
0.13510599732398987,
-0.03357874974608421,
-0.016168400645256042,
-0.10925812274217606,
-0.11051727086305618,
-0.000580480438657105,
-0.09536946564912796,
0.016299426555633545,
-0.05611075833439827,
-0.026699256151914597,
0.23211660981178284,
0.0814795047044754,
-0.025980792939662933,
-0.07317766547203064,
-0.022149236872792244,
-0.06921329349279404,
0.055895064026117325,
0.013505296781659126,
0.04391179606318474,
-0.11895626783370972,
-0.06589871644973755,
-0.10288331657648087,
-0.07416098564863205,
0.0025908327661454678,
0.04260432720184326,
-0.12327691912651062,
-0.11309793591499329,
-0.16445781290531158,
-0.07747455686330795,
-0.00009795460209716111,
-0.008201969787478447,
0.007408846635371447,
0.029473833739757538,
0.012894940562546253,
0.03716215863823891,
0.008789930492639542,
0.17000442743301392,
0.04046683758497238,
-0.08501631766557693,
-0.08569347858428955,
0.1839667707681656,
-0.09210920333862305,
0.10569430142641068,
0.04371228814125061,
-0.006691804155707359,
0.0014780706260353327,
0.15564917027950287,
0.24183779954910278,
0.07702524214982986,
-0.012856527231633663,
-0.014980876818299294,
0.06665822863578796,
0.1329217553138733,
0.02532215416431427,
-0.0034216642379760742,
0.23336274921894073,
-0.12276865541934967,
0.039338868111371994,
-0.1841534972190857,
0.01538696326315403,
-0.04963115230202675,
0.1546938419342041,
0.14413116872310638,
-0.15015392005443573,
-0.10466159880161285,
0.1837267279624939,
-0.055560801178216934,
0.18081393837928772,
0.019655482843518257,
-0.20280279219150543,
-0.07806850224733353,
-0.05655089020729065,
0.08775103837251663,
0.004902788437902927,
0.12055818736553192,
-0.09998733550310135,
-0.1783989816904068,
-0.007870802655816078,
0.05807097628712654,
-0.16384238004684448,
-0.17604520916938782,
0.12015831470489502,
0.06222708895802498,
-0.03710325434803963,
0.02099183388054371,
0.12427778542041779,
0.09403979033231735,
0.06731437146663666,
0.03393270820379257,
-0.08236232399940491,
0.07093531638383865,
0.04456457495689392,
-0.1609305888414383,
-0.06160341948270798,
0.050206173211336136,
-0.11968278884887695,
0.03548922389745712,
-0.21473608911037445,
0.056695710867643356,
0.010073436424136162,
-0.10711353272199631,
0.03459387272596359,
0.1795300841331482,
-0.06941073387861252,
-0.0021261745132505894,
0.10277761518955231,
0.016517480835318565,
-0.06192461773753166,
-0.0790850818157196,
-0.024025101214647293,
0.06112043187022209,
-0.06932763010263443,
-0.030072543770074844,
-0.08565878123044968,
-0.05075583606958389,
0.028130512684583664,
-0.08481575548648834,
-0.056708887219429016,
-0.02553166262805462,
-0.05867702513933182,
0.09955089539289474,
-0.06932730227708817,
0.0008606023038737476,
0.002106092404574156,
-0.03131014481186867,
-0.002042192965745926,
0.006041534710675478,
0.11007444560527802,
0.06821853667497635,
-0.012165430001914501,
-0.03360646590590477
] |
null | null | null |
This repositories enable third-party libraries integrated with [huggingface_hub](https://github.com/huggingface/huggingface_hub/) to create
their own docker so that the widgets on the hub can work as the `transformers` one do.
The hardware to run the API will be provided by Hugging Face for now.
The `docker_images/common` folder is intended to be a starter point for all new libs that
want to be integrated.
### Adding a new container from a new lib.
1. Copy the `docker_images/common` folder into your library's name `docker_images/example`.
2. Edit:
- `docker_images/example/requirements.txt`
- `docker_images/example/app/main.py`
- `docker_images/example/app/pipelines/{task_name}.py`
to implement the desired functionality. All required code is marked with `IMPLEMENT_THIS` markup.
3. Remove:
- Any pipeline files in `docker_images/example/app/pipelines/` that are not used.
- Any tests associated with deleted pipelines in `docker_images/example/tests`.
- Any imports of the pipelines you deleted from `docker_images/example/app/pipelines/__init__.py`
4. Feel free to customize anything required by your lib everywhere you want. The only real requirements, are to honor the HTTP endpoints, in the same fashion as the `common` folder for all your supported tasks.
5. Edit `example/tests/test_api.py` to add TESTABLE_MODELS.
6. Pass the test suite `pytest -sv --rootdir docker_images/example/ docker_images/example/`
7. Submit your PR and enjoy !
### Going the full way
Doing the first 7 steps is good enough to get started, however in the process
you can anticipate some problems corrections early on. Maintainers will help you
along the way if you don't feel confident to follow those steps yourself
1. Test your creation within a docker
```python
./manage.py docker MY_MODEL
```
should work and responds on port 8000. `curl -X POST -d "test" http://localhost:8000` for instance if
the pipeline deals with simple text.
If it doesn't work out of the box and/or docker is slow for some reason you
can test locally (using your local python environment) with :
`./manage.py start MY_MODEL`
2. Test your docker uses cache properly.
When doing subsequent docker launch with the same model_id, the docker should start up very fast and not redownload the whole model file. If you see the model/repo being downloaded over and over, it means the cache is not being used correctly.
You can edit the `docker_images/{framework}/Dockerfile` and add an environment variable (by default it assumes `HUGGINGFACE_HUB_CACHE`), or your code directly to put
the model files in the `/data` folder.
3. Add a docker test.
Edit the `tests/test_dockers.py` file to add a new test with your new framework
in it (`def test_{framework}(self):` for instance). As a basic you should have 1 line per task in this test function with a real working model on the hub. Those tests are relatively slow but will check automatically that correct errors are replied by your API and that the cache works properly. To run those tests your can simply do:
```bash
RUN_DOCKER_TESTS=1 pytest -sv tests/test_dockers.py::DockerImageTests::test_{framework}
```
### Modifying files within `api-inference-community/{routes,validation,..}.py`.
If you ever come across a bug within `api-inference-community/` package or want to update it
the development process is slightly more involved.
- First, make sure you need to change this package, each framework is very autonomous
so if your code can get away by being standalone go that way first as it's much simpler.
- If you can make the change only in `api-inference-community` without depending on it
that's also a great option. Make sure to add the proper tests to your PR.
- Finally, the best way to go is to develop locally using `manage.py` command:
- Do the necessary modifications within `api-inference-community` first.
- Install it locally in your environment with `pip install -e .`
- Install your package dependencies locally.
- Run your webserver locally: `./manage.py start --framework example --task audio-source-separation --model-id MY_MODEL`
- When everything is working, you will need to split your PR in two, 1 for the `api-inference-community` part.
The second one will be for your package specific modifications and will only land once the `api-inference-community` tag has landed.
- This workflow is still work in progress, don't hesitate to ask questions to maintainers.
Another similar command `./manage.py docker --framework example --task audio-source-separation --model-id MY_MODEL`
Will launch the server, but this time in a protected, controlled docker environment making sure the behavior
will be exactly the one in the API.
### Available tasks
- **Automatic speech recognition**: Input is a file, output is a dict of understood words being said within the file
- **Text generation**: Input is a text, output is a dict of generated text
- **Image recognition**: Input is an image, output is a dict of generated text
- **Question answering**: Input is a question + some context, output is a dict containing necessary information to locate the answer to the `question` within the `context`.
- **Audio source separation**: Input is some audio, and the output is n audio files that sum up to the original audio but contain individual sources of sound (either speakers or instruments for instant).
- **Token classification**: Input is some text, and the output is a list of entities mentioned in the text. Entities can be anything remarkable like locations, organisations, persons, times etc...
- **Text to speech**: Input is some text, and the output is an audio file saying the text...
- **Sentence Similarity**: Input is some sentence and a list of reference sentences, and the list of similarity scores.
| {} | null | tulayaka/x | [
"region:us"
] | 2024-02-08T08:17:52+00:00 | [] | [] | TAGS
#region-us
|
This repositories enable third-party libraries integrated with huggingface_hub to create
their own docker so that the widgets on the hub can work as the 'transformers' one do.
The hardware to run the API will be provided by Hugging Face for now.
The 'docker_images/common' folder is intended to be a starter point for all new libs that
want to be integrated.
### Adding a new container from a new lib.
1. Copy the 'docker_images/common' folder into your library's name 'docker_images/example'.
2. Edit:
- 'docker_images/example/URL'
- 'docker_images/example/app/URL'
- 'docker_images/example/app/pipelines/{task_name}.py'
to implement the desired functionality. All required code is marked with 'IMPLEMENT_THIS' markup.
3. Remove:
- Any pipeline files in 'docker_images/example/app/pipelines/' that are not used.
- Any tests associated with deleted pipelines in 'docker_images/example/tests'.
- Any imports of the pipelines you deleted from 'docker_images/example/app/pipelines/__init__.py'
4. Feel free to customize anything required by your lib everywhere you want. The only real requirements, are to honor the HTTP endpoints, in the same fashion as the 'common' folder for all your supported tasks.
5. Edit 'example/tests/test_api.py' to add TESTABLE_MODELS.
6. Pass the test suite 'pytest -sv --rootdir docker_images/example/ docker_images/example/'
7. Submit your PR and enjoy !
### Going the full way
Doing the first 7 steps is good enough to get started, however in the process
you can anticipate some problems corrections early on. Maintainers will help you
along the way if you don't feel confident to follow those steps yourself
1. Test your creation within a docker
should work and responds on port 8000. 'curl -X POST -d "test" http://localhost:8000' for instance if
the pipeline deals with simple text.
If it doesn't work out of the box and/or docker is slow for some reason you
can test locally (using your local python environment) with :
'./URL start MY_MODEL'
2. Test your docker uses cache properly.
When doing subsequent docker launch with the same model_id, the docker should start up very fast and not redownload the whole model file. If you see the model/repo being downloaded over and over, it means the cache is not being used correctly.
You can edit the 'docker_images/{framework}/Dockerfile' and add an environment variable (by default it assumes 'HUGGINGFACE_HUB_CACHE'), or your code directly to put
the model files in the '/data' folder.
3. Add a docker test.
Edit the 'tests/test_dockers.py' file to add a new test with your new framework
in it ('def test_{framework}(self):' for instance). As a basic you should have 1 line per task in this test function with a real working model on the hub. Those tests are relatively slow but will check automatically that correct errors are replied by your API and that the cache works properly. To run those tests your can simply do:
### Modifying files within 'api-inference-community/{routes,validation,..}.py'.
If you ever come across a bug within 'api-inference-community/' package or want to update it
the development process is slightly more involved.
- First, make sure you need to change this package, each framework is very autonomous
so if your code can get away by being standalone go that way first as it's much simpler.
- If you can make the change only in 'api-inference-community' without depending on it
that's also a great option. Make sure to add the proper tests to your PR.
- Finally, the best way to go is to develop locally using 'URL' command:
- Do the necessary modifications within 'api-inference-community' first.
- Install it locally in your environment with 'pip install -e .'
- Install your package dependencies locally.
- Run your webserver locally: './URL start --framework example --task audio-source-separation --model-id MY_MODEL'
- When everything is working, you will need to split your PR in two, 1 for the 'api-inference-community' part.
The second one will be for your package specific modifications and will only land once the 'api-inference-community' tag has landed.
- This workflow is still work in progress, don't hesitate to ask questions to maintainers.
Another similar command './URL docker --framework example --task audio-source-separation --model-id MY_MODEL'
Will launch the server, but this time in a protected, controlled docker environment making sure the behavior
will be exactly the one in the API.
### Available tasks
- Automatic speech recognition: Input is a file, output is a dict of understood words being said within the file
- Text generation: Input is a text, output is a dict of generated text
- Image recognition: Input is an image, output is a dict of generated text
- Question answering: Input is a question + some context, output is a dict containing necessary information to locate the answer to the 'question' within the 'context'.
- Audio source separation: Input is some audio, and the output is n audio files that sum up to the original audio but contain individual sources of sound (either speakers or instruments for instant).
- Token classification: Input is some text, and the output is a list of entities mentioned in the text. Entities can be anything remarkable like locations, organisations, persons, times etc...
- Text to speech: Input is some text, and the output is an audio file saying the text...
- Sentence Similarity: Input is some sentence and a list of reference sentences, and the list of similarity scores.
| [
"### Adding a new container from a new lib.\n\n\n1. Copy the 'docker_images/common' folder into your library's name 'docker_images/example'.\n2. Edit:\n - 'docker_images/example/URL'\n - 'docker_images/example/app/URL'\n - 'docker_images/example/app/pipelines/{task_name}.py' \n\n to implement the desired functionality. All required code is marked with 'IMPLEMENT_THIS' markup.\n3. Remove:\n - Any pipeline files in 'docker_images/example/app/pipelines/' that are not used.\n - Any tests associated with deleted pipelines in 'docker_images/example/tests'.\n - Any imports of the pipelines you deleted from 'docker_images/example/app/pipelines/__init__.py'\n\n4. Feel free to customize anything required by your lib everywhere you want. The only real requirements, are to honor the HTTP endpoints, in the same fashion as the 'common' folder for all your supported tasks.\n5. Edit 'example/tests/test_api.py' to add TESTABLE_MODELS.\n6. Pass the test suite 'pytest -sv --rootdir docker_images/example/ docker_images/example/'\n7. Submit your PR and enjoy !",
"### Going the full way\n\nDoing the first 7 steps is good enough to get started, however in the process \nyou can anticipate some problems corrections early on. Maintainers will help you\nalong the way if you don't feel confident to follow those steps yourself\n\n1. Test your creation within a docker\n\n\n\nshould work and responds on port 8000. 'curl -X POST -d \"test\" http://localhost:8000' for instance if \nthe pipeline deals with simple text.\n\nIf it doesn't work out of the box and/or docker is slow for some reason you\ncan test locally (using your local python environment) with :\n\n'./URL start MY_MODEL'\n\n\n2. Test your docker uses cache properly.\n\nWhen doing subsequent docker launch with the same model_id, the docker should start up very fast and not redownload the whole model file. If you see the model/repo being downloaded over and over, it means the cache is not being used correctly.\nYou can edit the 'docker_images/{framework}/Dockerfile' and add an environment variable (by default it assumes 'HUGGINGFACE_HUB_CACHE'), or your code directly to put\nthe model files in the '/data' folder.\n\n3. Add a docker test.\n\nEdit the 'tests/test_dockers.py' file to add a new test with your new framework\nin it ('def test_{framework}(self):' for instance). As a basic you should have 1 line per task in this test function with a real working model on the hub. Those tests are relatively slow but will check automatically that correct errors are replied by your API and that the cache works properly. To run those tests your can simply do:",
"### Modifying files within 'api-inference-community/{routes,validation,..}.py'.\n\nIf you ever come across a bug within 'api-inference-community/' package or want to update it\nthe development process is slightly more involved.\n\n- First, make sure you need to change this package, each framework is very autonomous\n so if your code can get away by being standalone go that way first as it's much simpler.\n- If you can make the change only in 'api-inference-community' without depending on it\n that's also a great option. Make sure to add the proper tests to your PR.\n- Finally, the best way to go is to develop locally using 'URL' command:\n- Do the necessary modifications within 'api-inference-community' first.\n- Install it locally in your environment with 'pip install -e .'\n- Install your package dependencies locally.\n- Run your webserver locally: './URL start --framework example --task audio-source-separation --model-id MY_MODEL'\n- When everything is working, you will need to split your PR in two, 1 for the 'api-inference-community' part.\n The second one will be for your package specific modifications and will only land once the 'api-inference-community' tag has landed.\n- This workflow is still work in progress, don't hesitate to ask questions to maintainers.\n\nAnother similar command './URL docker --framework example --task audio-source-separation --model-id MY_MODEL'\nWill launch the server, but this time in a protected, controlled docker environment making sure the behavior\nwill be exactly the one in the API.",
"### Available tasks\n\n- Automatic speech recognition: Input is a file, output is a dict of understood words being said within the file\n- Text generation: Input is a text, output is a dict of generated text\n- Image recognition: Input is an image, output is a dict of generated text\n- Question answering: Input is a question + some context, output is a dict containing necessary information to locate the answer to the 'question' within the 'context'.\n- Audio source separation: Input is some audio, and the output is n audio files that sum up to the original audio but contain individual sources of sound (either speakers or instruments for instant).\n- Token classification: Input is some text, and the output is a list of entities mentioned in the text. Entities can be anything remarkable like locations, organisations, persons, times etc...\n- Text to speech: Input is some text, and the output is an audio file saying the text...\n- Sentence Similarity: Input is some sentence and a list of reference sentences, and the list of similarity scores."
] | [
"TAGS\n#region-us \n",
"### Adding a new container from a new lib.\n\n\n1. Copy the 'docker_images/common' folder into your library's name 'docker_images/example'.\n2. Edit:\n - 'docker_images/example/URL'\n - 'docker_images/example/app/URL'\n - 'docker_images/example/app/pipelines/{task_name}.py' \n\n to implement the desired functionality. All required code is marked with 'IMPLEMENT_THIS' markup.\n3. Remove:\n - Any pipeline files in 'docker_images/example/app/pipelines/' that are not used.\n - Any tests associated with deleted pipelines in 'docker_images/example/tests'.\n - Any imports of the pipelines you deleted from 'docker_images/example/app/pipelines/__init__.py'\n\n4. Feel free to customize anything required by your lib everywhere you want. The only real requirements, are to honor the HTTP endpoints, in the same fashion as the 'common' folder for all your supported tasks.\n5. Edit 'example/tests/test_api.py' to add TESTABLE_MODELS.\n6. Pass the test suite 'pytest -sv --rootdir docker_images/example/ docker_images/example/'\n7. Submit your PR and enjoy !",
"### Going the full way\n\nDoing the first 7 steps is good enough to get started, however in the process \nyou can anticipate some problems corrections early on. Maintainers will help you\nalong the way if you don't feel confident to follow those steps yourself\n\n1. Test your creation within a docker\n\n\n\nshould work and responds on port 8000. 'curl -X POST -d \"test\" http://localhost:8000' for instance if \nthe pipeline deals with simple text.\n\nIf it doesn't work out of the box and/or docker is slow for some reason you\ncan test locally (using your local python environment) with :\n\n'./URL start MY_MODEL'\n\n\n2. Test your docker uses cache properly.\n\nWhen doing subsequent docker launch with the same model_id, the docker should start up very fast and not redownload the whole model file. If you see the model/repo being downloaded over and over, it means the cache is not being used correctly.\nYou can edit the 'docker_images/{framework}/Dockerfile' and add an environment variable (by default it assumes 'HUGGINGFACE_HUB_CACHE'), or your code directly to put\nthe model files in the '/data' folder.\n\n3. Add a docker test.\n\nEdit the 'tests/test_dockers.py' file to add a new test with your new framework\nin it ('def test_{framework}(self):' for instance). As a basic you should have 1 line per task in this test function with a real working model on the hub. Those tests are relatively slow but will check automatically that correct errors are replied by your API and that the cache works properly. To run those tests your can simply do:",
"### Modifying files within 'api-inference-community/{routes,validation,..}.py'.\n\nIf you ever come across a bug within 'api-inference-community/' package or want to update it\nthe development process is slightly more involved.\n\n- First, make sure you need to change this package, each framework is very autonomous\n so if your code can get away by being standalone go that way first as it's much simpler.\n- If you can make the change only in 'api-inference-community' without depending on it\n that's also a great option. Make sure to add the proper tests to your PR.\n- Finally, the best way to go is to develop locally using 'URL' command:\n- Do the necessary modifications within 'api-inference-community' first.\n- Install it locally in your environment with 'pip install -e .'\n- Install your package dependencies locally.\n- Run your webserver locally: './URL start --framework example --task audio-source-separation --model-id MY_MODEL'\n- When everything is working, you will need to split your PR in two, 1 for the 'api-inference-community' part.\n The second one will be for your package specific modifications and will only land once the 'api-inference-community' tag has landed.\n- This workflow is still work in progress, don't hesitate to ask questions to maintainers.\n\nAnother similar command './URL docker --framework example --task audio-source-separation --model-id MY_MODEL'\nWill launch the server, but this time in a protected, controlled docker environment making sure the behavior\nwill be exactly the one in the API.",
"### Available tasks\n\n- Automatic speech recognition: Input is a file, output is a dict of understood words being said within the file\n- Text generation: Input is a text, output is a dict of generated text\n- Image recognition: Input is an image, output is a dict of generated text\n- Question answering: Input is a question + some context, output is a dict containing necessary information to locate the answer to the 'question' within the 'context'.\n- Audio source separation: Input is some audio, and the output is n audio files that sum up to the original audio but contain individual sources of sound (either speakers or instruments for instant).\n- Token classification: Input is some text, and the output is a list of entities mentioned in the text. Entities can be anything remarkable like locations, organisations, persons, times etc...\n- Text to speech: Input is some text, and the output is an audio file saying the text...\n- Sentence Similarity: Input is some sentence and a list of reference sentences, and the list of similarity scores."
] | [
6,
339,
383,
403,
248
] | [
"passage: TAGS\n#region-us \n### Adding a new container from a new lib.\n\n\n1. Copy the 'docker_images/common' folder into your library's name 'docker_images/example'.\n2. Edit:\n - 'docker_images/example/URL'\n - 'docker_images/example/app/URL'\n - 'docker_images/example/app/pipelines/{task_name}.py' \n\n to implement the desired functionality. All required code is marked with 'IMPLEMENT_THIS' markup.\n3. Remove:\n - Any pipeline files in 'docker_images/example/app/pipelines/' that are not used.\n - Any tests associated with deleted pipelines in 'docker_images/example/tests'.\n - Any imports of the pipelines you deleted from 'docker_images/example/app/pipelines/__init__.py'\n\n4. Feel free to customize anything required by your lib everywhere you want. The only real requirements, are to honor the HTTP endpoints, in the same fashion as the 'common' folder for all your supported tasks.\n5. Edit 'example/tests/test_api.py' to add TESTABLE_MODELS.\n6. Pass the test suite 'pytest -sv --rootdir docker_images/example/ docker_images/example/'\n7. Submit your PR and enjoy !",
"passage: ### Going the full way\n\nDoing the first 7 steps is good enough to get started, however in the process \nyou can anticipate some problems corrections early on. Maintainers will help you\nalong the way if you don't feel confident to follow those steps yourself\n\n1. Test your creation within a docker\n\n\n\nshould work and responds on port 8000. 'curl -X POST -d \"test\" http://localhost:8000' for instance if \nthe pipeline deals with simple text.\n\nIf it doesn't work out of the box and/or docker is slow for some reason you\ncan test locally (using your local python environment) with :\n\n'./URL start MY_MODEL'\n\n\n2. Test your docker uses cache properly.\n\nWhen doing subsequent docker launch with the same model_id, the docker should start up very fast and not redownload the whole model file. If you see the model/repo being downloaded over and over, it means the cache is not being used correctly.\nYou can edit the 'docker_images/{framework}/Dockerfile' and add an environment variable (by default it assumes 'HUGGINGFACE_HUB_CACHE'), or your code directly to put\nthe model files in the '/data' folder.\n\n3. Add a docker test.\n\nEdit the 'tests/test_dockers.py' file to add a new test with your new framework\nin it ('def test_{framework}(self):' for instance). As a basic you should have 1 line per task in this test function with a real working model on the hub. Those tests are relatively slow but will check automatically that correct errors are replied by your API and that the cache works properly. To run those tests your can simply do:### Modifying files within 'api-inference-community/{routes,validation,..}.py'.\n\nIf you ever come across a bug within 'api-inference-community/' package or want to update it\nthe development process is slightly more involved.\n\n- First, make sure you need to change this package, each framework is very autonomous\n so if your code can get away by being standalone go that way first as it's much simpler.\n- If you can make the change only in 'api-inference-community' without depending on it\n that's also a great option. Make sure to add the proper tests to your PR.\n- Finally, the best way to go is to develop locally using 'URL' command:\n- Do the necessary modifications within 'api-inference-community' first.\n- Install it locally in your environment with 'pip install -e .'\n- Install your package dependencies locally.\n- Run your webserver locally: './URL start --framework example --task audio-source-separation --model-id MY_MODEL'\n- When everything is working, you will need to split your PR in two, 1 for the 'api-inference-community' part.\n The second one will be for your package specific modifications and will only land once the 'api-inference-community' tag has landed.\n- This workflow is still work in progress, don't hesitate to ask questions to maintainers.\n\nAnother similar command './URL docker --framework example --task audio-source-separation --model-id MY_MODEL'\nWill launch the server, but this time in a protected, controlled docker environment making sure the behavior\nwill be exactly the one in the API."
] | [
-0.12909293174743652,
0.011799586936831474,
-0.0068489969708025455,
0.02549906261265278,
0.08808673918247223,
0.014246681705117226,
-0.00301073445007205,
0.13764344155788422,
0.08684596419334412,
0.11226636916399002,
0.018378490582108498,
0.07764600962400436,
-0.0035516484640538692,
0.07103075832128525,
0.04279500991106033,
-0.0757765844464302,
0.03692891448736191,
-0.06640168279409409,
0.11738801002502441,
0.08658209443092346,
0.05261355638504028,
-0.06759922206401825,
0.05722218006849289,
-0.015953481197357178,
0.0010158992372453213,
0.015655245631933212,
0.00466141989454627,
0.017895426601171494,
0.03546207770705223,
0.005618926137685776,
0.02172519639134407,
-0.021142560988664627,
-0.03912462294101715,
-0.12670770287513733,
0.026961732655763626,
0.10643775016069412,
0.06959378719329834,
0.04949459806084633,
0.11370687186717987,
0.03211177885532379,
0.07876233756542206,
-0.06831661611795425,
0.0447913333773613,
0.06762556731700897,
-0.052590519189834595,
-0.12862545251846313,
-0.10521373897790909,
-0.021216167137026787,
0.15367478132247925,
0.1403459906578064,
-0.0034431484527885914,
0.17562861740589142,
-0.07014354318380356,
0.07209707796573639,
0.19701844453811646,
-0.03716157376766205,
-0.02038039080798626,
0.020092647522687912,
0.10145297646522522,
0.028718244284391403,
-0.0549699142575264,
-0.05493972450494766,
-0.06047171726822853,
0.07664312422275543,
-0.06319737434387207,
-0.07377275824546814,
-0.006540096364915371,
-0.08716341853141785,
-0.06807257980108261,
-0.06598220765590668,
0.10008356720209122,
-0.047964755445718765,
-0.1256023645401001,
-0.17199036478996277,
-0.06514744460582733,
-0.11767323315143585,
-0.029961008578538895,
-0.022979648783802986,
0.008212976157665253,
0.005690095014870167,
0.0912175327539444,
-0.11096368730068207,
-0.10545501112937927,
-0.0542602464556694,
0.07156427204608917,
0.03326895087957382,
0.04811100289225578,
-0.011718903668224812,
-0.028151538223028183,
0.16508330404758453,
0.038572512567043304,
-0.043461866676807404,
-0.10312636196613312,
-0.07908780127763748,
-0.06407731771469116,
-0.03817470744252205,
-0.021382102742791176,
-0.13814914226531982,
0.030839942395687103,
0.2473209798336029,
-0.04202371463179588,
0.07510561496019363,
-0.06625425815582275,
0.0425308421254158,
0.0007355171255767345,
0.1541990488767624,
0.04523700848221779,
0.0016413391567766666,
0.06160367652773857,
-0.002972274087369442,
0.03880449756979942,
0.018553374335169792,
0.056130629032850266,
-0.00992905255407095,
0.01034590508788824,
0.11275815963745117,
0.05435729771852493,
0.03452874720096588,
-0.05517509952187538,
-0.025417177006602287,
0.23119817674160004,
-0.10433077067136765,
0.06098870187997818,
0.0516391322016716,
0.014180745929479599,
0.010549318045377731,
0.08784478157758713,
-0.07146956026554108,
-0.14021697640419006,
0.09447035193443298,
0.002273636870086193,
0.049228984862565994,
-0.0977184846997261,
-0.058051466941833496,
0.04227963835000992,
-0.009182725101709366,
-0.039420560002326965,
-0.08637423813343048,
-0.13147209584712982,
-0.08075027167797089,
0.03550955653190613,
0.0011100037954747677,
0.032222725450992584,
0.10286477208137512,
-0.036444660276174545,
-0.057234976440668106,
-0.000027017667889595032,
-0.012574579566717148,
-0.057662080973386765,
-0.0068462565541267395,
-0.014815174974501133,
-0.032459162175655365,
0.0936836376786232,
0.018985725939273834,
-0.11523696780204773,
0.05627651885151863,
-0.2895485758781433,
0.0358341708779335,
-0.08914953470230103,
0.014393189921975136,
-0.09330853819847107,
-0.03819578140974045,
0.028324326500296593,
0.010177010670304298,
0.03708932548761368,
0.11535629630088806,
-0.07404778897762299,
0.040878891944885254,
0.06170569360256195,
-0.09868772327899933,
-0.04952182620763779,
0.08428609371185303,
0.04487362504005432,
0.03971504420042038,
0.07981251180171967,
0.0111142098903656,
0.21246841549873352,
-0.20482346415519714,
-0.02242935821413994,
-0.0031702443957328796,
-0.05554912984371185,
0.0006692791357636452,
0.008700018748641014,
-0.054145026952028275,
0.02658109739422798,
0.05358392745256424,
-0.07861481606960297,
-0.015633679926395416,
0.04461418464779854,
-0.007808358408510685,
-0.03207312524318695,
-0.017090002074837685,
-0.10579119622707367,
-0.016345681622624397,
-0.03687908127903938,
-0.03437318280339241,
-0.07553005963563919,
-0.10190731287002563,
0.13335643708705902,
0.031110763549804688,
0.0728149563074112,
-0.051907774060964584,
0.11570781469345093,
-0.032447703182697296,
0.013225522823631763,
-0.16304439306259155,
-0.08131484687328339,
0.06372957676649094,
-0.10779188573360443,
-0.002723490819334984,
-0.07653598487377167,
0.003857026342302561,
0.050864193588495255,
0.013219140470027924,
-0.03828365355730057,
-0.022013867273926735,
-0.049769654870033264,
-0.07571806758642197,
-0.04081796854734421,
-0.04952584579586983,
-0.04687943682074547,
0.0194380059838295,
-0.09174703061580658,
0.015448663383722305,
0.03868234157562256,
-0.0017926236614584923,
0.053320445120334625,
-0.09086482971906662,
0.05121014267206192,
-0.02148800902068615,
0.005092867650091648,
-0.04569930583238602,
-0.007331458851695061,
0.02243809588253498,
0.07010437548160553,
0.05673959106206894,
-0.08332963287830353,
-0.10441026091575623,
0.04790544509887695,
0.07552050799131393,
-0.08784569799900055,
0.04672062769532204,
-0.04284748435020447,
-0.02667277306318283,
-0.09062261879444122,
-0.056026168167591095,
0.06605177372694016,
0.12266238778829575,
0.0843610092997551,
-0.039255522191524506,
-0.033472612500190735,
-0.0130547396838665,
-0.0033428948372602463,
0.06666910648345947,
0.0755189061164856,
0.07078532129526138,
0.042581453919410706,
-0.0016965530812740326,
0.08121640980243683,
-0.09301133453845978,
0.015580793842673302,
0.025645669549703598,
-0.12136481702327728,
0.009197425097227097,
0.10570710897445679,
0.021124746650457382,
0.048788882791996,
-0.006915288977324963,
0.061535436660051346,
0.042035773396492004,
0.03805689886212349,
0.029777487739920616,
-0.05467480793595314,
0.01852024346590042,
0.0026438822969794273,
-0.012215767987072468,
0.02298389934003353,
0.06420513242483139,
0.03540096431970596,
0.021321555599570274,
-0.0566810704767704,
0.05956181883811951,
-0.013779458589851856,
-0.02397380955517292,
-0.09830693900585175,
0.12019355595111847,
-0.06110330671072006,
-0.2337251901626587,
-0.14261043071746826,
-0.013705259189009666,
-0.06572361290454865,
-0.007671143859624863,
0.11544515192508698,
-0.025792768225073814,
-0.08189306408166885,
-0.09648416936397552,
0.0016337510896846652,
0.10557964444160461,
-0.12005367875099182,
-0.1598762571811676,
0.022118177264928818,
0.08740170300006866,
-0.07436268776655197,
-0.03275657817721367,
0.07274501025676727,
0.007625652477145195,
0.06229032948613167,
0.09284722805023193,
0.047137223184108734,
-0.010308641940355301,
0.06437955796718597,
-0.054507382214069366,
0.029411355033516884,
0.17272153496742249,
-0.07873702794313431,
0.07478159666061401,
0.262856662273407,
-0.013958968222141266,
0.08733853697776794,
-0.014507727697491646,
-0.012648005038499832,
-0.04658565670251846,
0.032342810183763504,
0.06418773531913757,
-0.014462847262620926,
-0.18990060687065125,
-0.030916467308998108,
-0.054931171238422394,
0.023171424865722656,
0.09522442519664764,
0.0681360736489296,
-0.05815652757883072,
0.04427136480808258,
-0.07714438438415527,
-0.09683874249458313,
-0.030116120353341103,
0.13385538756847382,
0.05803672969341278,
-0.07457093894481659,
0.026353586465120316,
-0.06320163607597351,
0.02940443530678749,
0.08486182987689972,
0.060175999999046326,
0.03996260091662407,
-0.028704028576612473,
0.08866320550441742,
0.01043243333697319,
0.09622274339199066,
-0.046598635613918304,
0.10171191394329071,
-0.013271531090140343,
0.02012314461171627,
0.015435821376740932,
-0.0792354941368103,
-0.02333529107272625,
0.05102575570344925,
-0.006919452920556068,
-0.07617660611867905,
-0.030176708474755287,
-0.028748754411935806,
0.04702381789684296,
0.21190664172172546,
0.014079338870942593,
-0.1371176838874817,
-0.044286008924245834,
-0.04339431971311569,
-0.057461079210042953,
-0.1285933554172516,
-0.004498995840549469,
0.04857417568564415,
-0.1328052580356598,
-0.0291083212941885,
-0.05212491750717163,
0.08509758114814758,
-0.010299501940608025,
0.024216391146183014,
0.08043956756591797,
0.0693240761756897,
0.03720557317137718,
0.0699099525809288,
-0.13596704602241516,
-0.08394696563482285,
-0.024524841457605362,
0.017112549394369125,
-0.04711836576461792,
0.03060402348637581,
0.043796390295028687,
0.008909701369702816,
0.12964583933353424,
-0.00027388916350901127,
-0.08016970753669739,
-0.10285758972167969,
-0.09723687171936035,
0.014244590885937214,
0.02513493224978447,
-0.020074378699064255,
0.05178612470626831,
-0.008263844065368176,
-0.012233739718794823,
-0.03982082009315491,
-0.06566578149795532,
-0.09758962690830231,
-0.08679676055908203,
0.06192309036850929,
-0.03517252206802368,
0.001112842932343483,
-0.03691538795828819,
0.02148364670574665,
0.02507447451353073,
0.1993718445301056,
-0.06016727164387703,
-0.04186708852648735,
-0.10084645450115204,
-0.10408147424459457,
0.11461373418569565,
-0.07183897495269775,
0.012697786092758179,
-0.04647450149059296,
0.07969462126493454,
-0.01276067178696394,
-0.021838247776031494,
0.03851074352860451,
-0.12840934097766876,
-0.17900046706199646,
-0.03192222863435745,
0.11532694101333618,
-0.017647303640842438,
0.004858773201704025,
-0.00903765857219696,
0.0010086394613608718,
-0.04375693202018738,
-0.08775680512189865,
-0.03596193343400955,
0.10315225273370743,
0.06626942753791809,
0.01827477663755417,
-0.11243204772472382,
0.10427671670913696,
-0.10090681165456772,
0.00009451434016227722,
0.11910680681467056,
0.22445711493492126,
-0.03282956779003143,
0.0774226039648056,
0.015956467017531395,
-0.05603110045194626,
-0.1974475383758545,
0.008534284308552742,
0.02708199992775917,
-0.0488884337246418,
0.06538049131631851,
-0.15611672401428223,
0.19237318634986877,
0.09820492565631866,
-0.026220306754112244,
0.2157711684703827,
-0.19411350786685944,
-0.11520789563655853,
0.05564539134502411,
0.028184406459331512,
-0.06711465865373611,
-0.11612392961978912,
0.004495477303862572,
-0.04396119713783264,
-0.012784229591488838,
0.023416902869939804,
-0.06510088592767715,
0.08182366192340851,
-0.027212435379624367,
0.0003061210736632347,
0.050628308206796646,
-0.09510467201471329,
0.1057601124048233,
-0.08927737921476364,
0.0743936076760292,
-0.03446986526250839,
0.07906529307365417,
0.00018109043594449759,
-0.11265520006418228,
0.18870240449905396,
-0.09025204181671143,
0.07335609197616577,
-0.0970625951886177,
-0.009401289746165276,
-0.062414176762104034,
0.08934113383293152,
-0.045069195330142975,
0.01343267597258091,
-0.06411025673151016,
0.015709925442934036,
0.05908171832561493,
-0.0321546345949173,
0.08276732265949249,
-0.020942851901054382,
-0.044672295451164246,
0.1671447455883026,
-0.041397836059331894,
0.009908381849527359,
-0.21200960874557495,
0.029614940285682678,
-0.0028974879533052444,
0.05953836441040039,
-0.021732548251748085,
0.04231404513120651,
0.0641598030924797,
0.000990179367363453,
0.019616059958934784,
-0.03974825516343117,
-0.10183869302272797,
-0.03205769881606102,
0.03371548280119896,
-0.1062193363904953,
-0.026568222790956497,
-0.004302912391722202,
0.1369534730911255,
-0.08867776393890381,
-0.0870705097913742,
0.13172096014022827,
0.016963649541139603,
-0.031140010803937912,
0.026838520541787148,
0.07678503543138504,
-0.021252721548080444,
0.14181524515151978,
-0.039641641080379486,
0.022896312177181244,
-0.055349837988615036,
0.05534735321998596,
0.02221072092652321,
-0.14971230924129486,
-0.00362576637417078,
0.16188573837280273,
-0.07286117970943451,
-0.11628513038158417,
0.014334233477711678,
0.027238959446549416,
0.020294440910220146,
-0.008075624704360962,
-0.027379680424928665,
0.004890727810561657,
0.03544793277978897,
0.027401182800531387,
0.044709332287311554,
-0.012044413015246391,
0.009068252518773079,
-0.018050387501716614,
-0.06144776940345764,
0.08283716440200806,
0.03782960772514343,
0.10558013617992401,
-0.10232491791248322,
-0.006393212825059891,
0.032753217965364456,
0.05791565775871277,
0.019682632759213448,
-0.0018907282501459122,
-0.053098879754543304,
0.038574423640966415,
-0.18844258785247803,
0.05321120098233223,
-0.05080579221248627,
-0.010526927188038826,
0.013266811147332191,
0.06560781598091125,
0.03576721251010895,
0.01859447918832302,
-0.06943836808204651,
-0.09535971283912659,
-0.07292267680168152,
0.05796078220009804,
-0.11608593165874481,
0.020241204649209976,
0.0838126540184021,
-0.11177940666675568,
0.06619079411029816,
0.04435757175087929,
-0.06531215459108353,
-0.008674506098031998,
-0.015478517860174179,
-0.025828294456005096,
0.03325117379426956,
0.048228152096271515,
-0.024652838706970215,
-0.1709040403366089,
0.020926330238580704,
0.009794226847589016,
-0.009883606806397438,
-0.02488682046532631,
0.14794766902923584,
-0.15746977925300598,
0.06230761110782623,
0.059746596962213516,
-0.084592804312706,
-0.038531772792339325,
-0.024015260860323906,
-0.0024866582825779915,
0.07847808301448822,
0.16918529570102692,
-0.030323665589094162,
0.09544284641742706,
-0.08579977601766586,
-0.016731595620512962,
0.008102868683636189,
-0.002282199449837208,
0.030485114082694054,
-0.014534560963511467,
0.047952428460121155,
-0.04280761629343033,
0.10464734584093094,
-0.030506229028105736,
0.043909892439842224,
0.03304222598671913,
0.00029226019978523254,
-0.07235223054885864,
-0.002885896246880293,
-0.06085234507918358,
-0.05458800122141838,
0.013093722984194756,
-0.011956865899264812,
-0.03177719563245773,
0.020933888852596283,
0.0016162917017936707,
0.10641340166330338,
0.19420480728149414,
0.06184078007936478,
0.05812804773449898,
-0.04897695779800415,
-0.04384739324450493,
-0.06726640462875366,
-0.10585804283618927,
-0.011315564624965191,
0.08261468261480331,
-0.0905957967042923,
0.12918004393577576,
0.06838840246200562,
-0.06920407712459564,
0.09141718596220016,
-0.0381120927631855,
-0.017919039353728294,
-0.06219814345240593,
-0.10859377682209015,
-0.029101010411977768,
-0.0498177707195282,
-0.016562538221478462,
-0.039988402277231216,
0.02197638526558876,
0.013695775531232357,
0.013477044180035591,
-0.06068423390388489,
0.21997113525867462,
-0.05751945450901985,
-0.11756130307912827,
0.012833335436880589,
-0.0015727453865110874,
-0.01677299290895462,
0.11186125874519348,
0.04209522157907486,
0.010982410982251167,
0.031185386702418327,
0.0857880711555481,
0.09999047219753265,
-0.006555566564202309,
0.043207164853811264,
-0.08895371854305267,
-0.04943275824189186,
-0.005556450225412846,
0.047679804265499115,
-0.015472782775759697,
0.18521176278591156,
0.005852784030139446,
-0.0039075263775885105,
-0.02577798441052437,
0.1837548315525055,
-0.06678973138332367,
-0.07787539809942245,
-0.0914103239774704,
0.2844671607017517,
0.03747621923685074,
-0.058760400861501694,
-0.04622262716293335,
-0.102070152759552,
0.015229590237140656,
0.1911465972661972,
0.13771018385887146,
-0.03397466242313385,
0.0077761514112353325,
0.06521761417388916,
0.008624008856713772,
-0.06701507419347763,
0.09034346044063568,
0.09355078637599945,
0.19372883439064026,
-0.019169144332408905,
0.1388523280620575,
-0.00924932025372982,
-0.04273969680070877,
-0.11291113495826721,
0.03405144438147545,
-0.06959990411996841,
-0.0012004263699054718,
-0.044386126101017,
0.04663902521133423,
-0.02785906195640564,
-0.13960951566696167,
-0.005818648263812065,
0.03904656693339348,
-0.07688853144645691,
0.03516385331749916,
-0.01697290875017643,
-0.04791511222720146,
0.05843318998813629,
-0.017403986304998398,
-0.035030484199523926,
0.1438632309436798,
-0.0844525396823883,
0.01715608686208725,
-0.09246376901865005,
0.08468262851238251,
-0.08816403895616531,
0.13836416602134705,
0.051286421716213226,
0.04753860458731651,
0.06337814033031464,
-0.013759936206042767,
-0.16228342056274414,
0.06286415457725525,
0.0417650043964386,
-0.15385988354682922,
0.027557726949453354,
0.1485796868801117,
-0.024211596697568893,
0.07183332741260529,
0.041173145174980164,
-0.03912520781159401,
-0.010958965867757797,
-0.025326360017061234,
0.03422362357378006,
-0.13373613357543945,
0.036834366619586945,
-0.10737380385398865,
0.10830089449882507,
0.15537768602371216,
0.00356169231235981,
0.01937628909945488,
-0.07347855716943741,
0.057900570333004,
0.013742057606577873,
-0.02098925970494747,
-0.03505339473485947,
-0.11050818860530853,
0.10253024101257324,
0.09297399967908859,
0.06087562069296837,
-0.18538904190063477,
-0.06195726990699768,
0.10141730308532715,
-0.035438790917396545,
0.000989777036011219,
0.10179801285266876,
-0.062181733548641205,
0.0017073655035346746,
0.0014009890146553516,
-0.05445663258433342,
0.02169325016438961,
0.09699070453643799,
-0.09410366415977478,
-0.0768285021185875
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | humung/polyglot-ko-12.8b-vlending-v0.6 | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-08T08:19:31+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
31,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06646376848220825,
0.2168014943599701,
-0.00225935154594481,
0.023818302899599075,
0.1271018385887146,
-0.001635765191167593,
0.04218708351254463,
0.13324736058712006,
-0.020175931975245476,
0.11144465953111649,
0.046588581055402756,
0.09377603232860565,
0.09928803145885468,
0.18404334783554077,
0.04859916493296623,
-0.2059975117444992,
0.007056170143187046,
-0.09090408682823181,
0.014076028019189835,
0.1116579994559288,
0.13719257712364197,
-0.10291384905576706,
0.08272874355316162,
-0.04045208916068077,
-0.02019004337489605,
0.00012576708104461432,
-0.09259183704853058,
-0.07032395154237747,
0.06885425746440887,
0.06264153122901917,
0.051234472543001175,
0.001456156256608665,
0.09140396863222122,
-0.2864592671394348,
0.017265573143959045,
0.08406311273574829,
0.0027674848679453135,
0.06290827691555023,
0.07236549258232117,
-0.07389893382787704,
0.11328595131635666,
-0.08021481335163116,
0.13019037246704102,
0.08625296503305435,
-0.062064990401268005,
-0.23071379959583282,
-0.07525765895843506,
0.0963398814201355,
0.12251301854848862,
0.06215599179267883,
-0.022921854630112648,
0.15455181896686554,
-0.06248689442873001,
0.012971068732440472,
0.1294165402650833,
-0.11526761949062347,
-0.05572471022605896,
0.061741601675748825,
0.11775490641593933,
0.10740239918231964,
-0.14110268652439117,
-0.0017287094378843904,
0.04900608956813812,
0.029121357947587967,
0.08589313924312592,
0.022661056369543076,
0.12003941088914871,
0.04652795568108559,
-0.13695219159126282,
-0.04037507623434067,
0.12011898308992386,
0.038862764835357666,
-0.06446044892072678,
-0.2168138176202774,
-0.006778308190405369,
-0.0601806715130806,
-0.014732478186488152,
-0.07019448280334473,
0.039128515869379044,
-0.02470310963690281,
0.07317749410867691,
-0.04465159401297569,
-0.1063927412033081,
-0.0421026237308979,
0.0892222449183464,
0.07748593389987946,
0.011527054943144321,
-0.02519804798066616,
0.04627908393740654,
0.13455867767333984,
0.05402068421244621,
-0.10399353504180908,
-0.07017925381660461,
-0.06942764669656754,
-0.09420394152402878,
-0.04035796597599983,
0.056760527193546295,
0.031942449510097504,
0.02665667235851288,
0.22703726589679718,
0.016653569415211678,
0.04155244305729866,
0.0224777739495039,
0.01032855175435543,
0.043662428855895996,
0.0955500528216362,
-0.05303520709276199,
-0.15660029649734497,
-0.04072032496333122,
0.09077946096658707,
-0.0027527001220732927,
-0.036689214408397675,
-0.03966725245118141,
0.03849169611930847,
0.06843466311693192,
0.13122352957725525,
0.07552056759595871,
-0.017929591238498688,
-0.04813180863857269,
-0.030096933245658875,
0.23523783683776855,
-0.1493375599384308,
0.04426715523004532,
-0.02271856553852558,
-0.01804111897945404,
-0.03908449783921242,
0.03597262129187584,
0.022118929773569107,
-0.000004518366949923802,
0.09706240892410278,
-0.058981191366910934,
-0.05378659814596176,
-0.10168042778968811,
-0.03272576630115509,
0.04088849574327469,
-0.013975566253066063,
-0.010589460842311382,
-0.09025166928768158,
-0.09490354359149933,
-0.04766594246029854,
0.05537205561995506,
-0.05123869329690933,
-0.03770573064684868,
0.009465423412621021,
-0.08151785284280777,
-0.005444355774670839,
-0.005417742300778627,
0.10699385404586792,
-0.03222226724028587,
0.04445803165435791,
-0.027600755915045738,
0.05225523188710213,
0.09919606149196625,
0.031576547771692276,
-0.0773419588804245,
0.0561848059296608,
-0.22559374570846558,
0.07503069192171097,
-0.11481974273920059,
0.04335082694888115,
-0.1704932004213333,
-0.042439818382263184,
0.005444696638733149,
0.0139949731528759,
0.013206101022660732,
0.12720820307731628,
-0.19255615770816803,
-0.01654396951198578,
0.13260798156261444,
-0.09212633967399597,
-0.118110790848732,
0.07884611934423447,
-0.029701577499508858,
0.1624738723039627,
0.04682036489248276,
-0.027025915682315826,
0.09224298596382141,
-0.16434773802757263,
-0.07092688232660294,
-0.00949116237461567,
-0.01727987825870514,
0.12109188735485077,
0.07512219995260239,
-0.05991523340344429,
0.046571120619773865,
0.02832140028476715,
-0.038078423589468,
-0.04424772411584854,
-0.050857074558734894,
-0.10884185880422592,
-0.01070026308298111,
-0.08987759798765182,
0.04065500199794769,
-0.01250192429870367,
-0.07916021347045898,
-0.029885273426771164,
-0.18612512946128845,
-0.0030564051121473312,
0.10038342326879501,
0.0035033065360039473,
-0.005652366206049919,
-0.08666291832923889,
0.026358824223279953,
-0.03112892620265484,
-0.008404186926782131,
-0.16764774918556213,
-0.04399421438574791,
0.046902090311050415,
-0.16094985604286194,
0.020117372274398804,
-0.06413903087377548,
0.06334125250577927,
0.03641495108604431,
-0.05590536445379257,
-0.0248766727745533,
-0.01730942726135254,
0.011945613659918308,
-0.05083848536014557,
-0.18994836509227753,
-0.056277405470609665,
-0.037882111966609955,
0.149809330701828,
-0.25956398248672485,
0.032966937869787216,
0.051140617579221725,
0.14649195969104767,
0.00406361510977149,
-0.05115427449345589,
0.01429014839231968,
-0.05360214412212372,
-0.054652128368616104,
-0.06746816635131836,
-0.006135428790003061,
-0.027576493099331856,
-0.05147203803062439,
0.019243421033024788,
-0.1755700707435608,
-0.021410830318927765,
0.09424154460430145,
0.12876708805561066,
-0.1486445665359497,
-0.018640631809830666,
-0.048725154250860214,
-0.06339836865663528,
-0.0715010017156601,
-0.07038594037294388,
0.10712739825248718,
0.0513901449739933,
0.04796046018600464,
-0.07435787469148636,
-0.07092321664094925,
0.02726263552904129,
0.006906150374561548,
-0.03382374346256256,
0.08727246522903442,
0.05199531093239784,
-0.09209315478801727,
0.0756213590502739,
0.1092359870672226,
0.07177663594484329,
0.09363535046577454,
0.01574566215276718,
-0.11756632477045059,
-0.028492970392107964,
0.036266472190618515,
0.02740776725113392,
0.1465986967086792,
-0.05952361226081848,
0.04016614332795143,
0.04494241625070572,
-0.04170418903231621,
0.022319864481687546,
-0.08787637203931808,
0.024075502529740334,
0.025203049182891846,
-0.0034381982404738665,
0.06284574419260025,
-0.02525499276816845,
-0.0050758360885083675,
0.07016654312610626,
0.047779910266399384,
0.04621000960469246,
0.009655474685132504,
-0.01720241829752922,
-0.1047825813293457,
0.16950392723083496,
-0.0951867327094078,
-0.269941508769989,
-0.17632324993610382,
0.026197833940386772,
0.04035249724984169,
-0.022378476336598396,
0.031619444489479065,
-0.07056326419115067,
-0.10630585998296738,
-0.1060405746102333,
-0.002429972169920802,
0.01714223250746727,
-0.06364088505506516,
-0.0741225928068161,
0.07348573952913284,
0.04382912442088127,
-0.14902326464653015,
0.038552410900592804,
0.055694397538900375,
-0.057955220341682434,
-0.0233661737293005,
0.09118817001581192,
0.12397737801074982,
0.14583967626094818,
-0.021366750821471214,
-0.028626007959246635,
0.029004426673054695,
0.19620531797409058,
-0.13469526171684265,
0.10371150821447372,
0.13814030587673187,
-0.04545360431075096,
0.08360563963651657,
0.1560150384902954,
0.029186224564909935,
-0.08317049592733383,
0.05044832453131676,
0.04082648828625679,
-0.043159641325473785,
-0.2666129767894745,
-0.0534592866897583,
0.012832709588110447,
-0.06255637854337692,
0.09786593168973923,
0.10183793306350708,
0.11542957276105881,
0.034910861402750015,
-0.07166364789009094,
-0.043925940990448,
-0.0058974819257855415,
0.11737963557243347,
-0.05490213260054588,
-0.012639665976166725,
0.07686592638492584,
-0.05086168646812439,
0.005355054512619972,
0.10266812145709991,
0.02973790094256401,
0.17442677915096283,
0.020399179309606552,
0.11231429129838943,
0.06195578724145889,
0.08633565157651901,
0.0007386076031252742,
0.02951662428677082,
0.05147615820169449,
0.017203815281391144,
-0.002300140680745244,
-0.10421168059110641,
-0.006156572140753269,
0.1449710875749588,
0.028103826567530632,
0.029669636860489845,
-0.0018948549404740334,
-0.005003341939300299,
0.05121048167347908,
0.1746254414319992,
-0.011592294089496136,
-0.22072425484657288,
-0.0845772922039032,
0.06936841458082199,
-0.06218599155545235,
-0.12968985736370087,
-0.026130788028240204,
0.045467354357242584,
-0.17519839107990265,
0.026703642681241035,
-0.027433741837739944,
0.0919293761253357,
-0.09345759451389313,
-0.02221956104040146,
0.03687324374914169,
0.084866963326931,
-0.014529162086546421,
0.08703910559415817,
-0.14498743414878845,
0.11886418610811234,
0.02978132851421833,
0.09024628251791,
-0.11081171780824661,
0.07909037172794342,
-0.007550720125436783,
0.009180475026369095,
0.19379350543022156,
-0.011335089802742004,
-0.03514958545565605,
-0.08774717897176743,
-0.11210042238235474,
-0.013537433929741383,
0.12687496840953827,
-0.1243172138929367,
0.08773399889469147,
-0.015198243781924248,
-0.044079482555389404,
0.00937260314822197,
-0.12100647389888763,
-0.17273177206516266,
-0.19628387689590454,
0.05585884302854538,
-0.09575839340686798,
0.025643249973654747,
-0.11914430558681488,
-0.07089093327522278,
-0.02952558360993862,
0.241120383143425,
-0.1745356321334839,
-0.06510113179683685,
-0.1468164622783661,
-0.046294767409563065,
0.1662203073501587,
-0.04437198117375374,
0.0718095526099205,
-0.0208172257989645,
0.20345525443553925,
0.005988610442727804,
-0.004939318168908358,
0.06724198162555695,
-0.08892562240362167,
-0.16873881220817566,
-0.06771010160446167,
0.1510489284992218,
0.11680185794830322,
0.04907919466495514,
-0.002248800592496991,
0.0011772146681323647,
-0.016943959519267082,
-0.1137804463505745,
-0.0033210667315870523,
0.16037839651107788,
0.03878779336810112,
0.025986969470977783,
-0.05243593826889992,
-0.08797456324100494,
-0.06899320334196091,
-0.06853509694337845,
0.06221301481127739,
0.19590823352336884,
-0.10376439243555069,
0.1700313836336136,
0.147536963224411,
-0.07305635511875153,
-0.23175598680973053,
0.035342130810022354,
0.04983805492520332,
0.0014306638622656465,
0.04886869341135025,
-0.18252557516098022,
0.10521943867206573,
0.019543392583727837,
-0.05505957826972008,
0.13485197722911835,
-0.1557481735944748,
-0.1552847921848297,
0.0722852572798729,
0.03904085233807564,
-0.22423844039440155,
-0.1354004591703415,
-0.09622503817081451,
-0.05825018882751465,
-0.14065024256706238,
0.06054598465561867,
-0.002136280992999673,
0.015948504209518433,
0.03500790148973465,
-0.0015643214574083686,
0.027123261243104935,
-0.058935679495334625,
0.18609118461608887,
-0.004065449349582195,
0.020676052197813988,
-0.060264769941568375,
-0.0478842556476593,
0.09839435666799545,
-0.06130504235625267,
0.12208222597837448,
0.004057085141539574,
0.01594383642077446,
-0.10362856835126877,
-0.048314861953258514,
-0.04328322783112526,
0.05154227837920189,
-0.07548051327466965,
-0.10070807486772537,
-0.043625857681035995,
0.08841723203659058,
0.07005169242620468,
-0.03383097052574158,
0.00549331633374095,
-0.07189501076936722,
0.10019614547491074,
0.17795267701148987,
0.17573626339435577,
0.009926567785441875,
-0.07241068035364151,
0.01677953451871872,
-0.04142116755247116,
0.044231921434402466,
-0.2513144314289093,
0.03756171092391014,
0.06098250672221184,
0.029438555240631104,
0.09217222779989243,
-0.020435843616724014,
-0.1820858269929886,
-0.04050002992153168,
0.08094815909862518,
-0.05452597141265869,
-0.22617179155349731,
-0.019085140898823738,
0.0954197570681572,
-0.2020406424999237,
-0.007372708059847355,
0.03995226323604584,
-0.048725228756666183,
-0.023169852793216705,
0.00010950004070764408,
0.06317184865474701,
0.002471912419423461,
0.09773622453212738,
0.0735151618719101,
0.09715340286493301,
-0.08337292820215225,
0.10562895983457565,
0.10150538384914398,
-0.09572599828243256,
0.03605884686112404,
0.06754924356937408,
-0.05300498008728027,
-0.043293699622154236,
0.03665391728281975,
0.033023297786712646,
0.005234600510448217,
-0.060321882367134094,
0.013913018628954887,
-0.036497246474027634,
0.044923391193151474,
0.08326134830713272,
0.03754979372024536,
-0.013354414142668247,
0.06462216377258301,
0.03401726484298706,
-0.10898099094629288,
0.10366570204496384,
0.01731540448963642,
0.04105307161808014,
-0.08384523540735245,
-0.019968897104263306,
0.035425446927547455,
0.030576206743717194,
-0.01765924133360386,
-0.02306121215224266,
-0.02860277332365513,
-0.01614218018949032,
-0.14299540221691132,
-0.023106401786208153,
-0.07243485748767853,
0.006181265693157911,
0.014656842686235905,
-0.031884219497442245,
-0.011233693920075893,
0.02475680410861969,
-0.06979699432849884,
-0.07426341623067856,
-0.006949664559215307,
0.09833318740129471,
-0.15115703642368317,
0.008848577737808228,
0.06907843053340912,
-0.11088496446609497,
0.08190931379795074,
-0.008411259390413761,
0.016245156526565552,
0.022527478635311127,
-0.15448406338691711,
0.05601610988378525,
0.0008648968650959432,
0.01916889287531376,
0.025886621326208115,
-0.16471809148788452,
0.004104440100491047,
-0.04661374166607857,
-0.02149827405810356,
-0.00004464812809601426,
-0.02647159807384014,
-0.12325995415449142,
0.06858719140291214,
-0.015622655861079693,
-0.035931166261434555,
-0.02701525390148163,
0.0539589487016201,
0.07888586074113846,
-0.027474910020828247,
0.10445091128349304,
-0.008690856397151947,
0.04941811040043831,
-0.16801609098911285,
-0.02470702864229679,
-0.04982255399227142,
0.019377702847123146,
0.009884213097393513,
-0.007693959400057793,
0.04183054715394974,
-0.00976533442735672,
0.21883612871170044,
-0.05075952783226967,
0.1607085019350052,
0.05847611650824547,
-0.017352959141135216,
-0.0007513365126214921,
0.06180921941995621,
0.05997028574347496,
0.04658793285489082,
0.009480604901909828,
0.023740366101264954,
-0.022450892254710197,
-0.006695089396089315,
-0.15932634472846985,
0.01890849508345127,
0.14999441802501678,
0.06301083415746689,
0.024745315313339233,
0.05866100639104843,
-0.12775006890296936,
-0.12135478109121323,
0.09311001747846603,
-0.026755332946777344,
0.00928465835750103,
-0.08245618641376495,
0.1358020007610321,
0.14980104565620422,
-0.14000412821769714,
0.05256148427724838,
-0.06134212389588356,
-0.05217423290014267,
-0.10388828068971634,
-0.12032219022512436,
-0.05887215584516525,
-0.053666237741708755,
0.002330566756427288,
-0.03760887682437897,
0.054546963423490524,
0.03344334661960602,
-0.009351172484457493,
-0.00022941511997487396,
0.13597318530082703,
-0.019751882180571556,
-0.0028988157864660025,
0.048313532024621964,
0.03693558648228645,
0.02373051457107067,
-0.05275435373187065,
0.02940409444272518,
0.02539868652820587,
0.032232340425252914,
0.06546790152788162,
0.033412106335163116,
-0.047448933124542236,
0.03804153576493263,
-0.0025254099164158106,
-0.11207924783229828,
0.019641218706965446,
-0.00460948096588254,
-0.0742158442735672,
0.1268945336341858,
0.0407399944961071,
0.010224059224128723,
-0.03741471841931343,
0.24361543357372284,
-0.06653323769569397,
-0.06378097087144852,
-0.13251738250255585,
0.10491154342889786,
-0.0027236645109951496,
0.06476365029811859,
0.023412218317389488,
-0.1284150779247284,
0.005243356805294752,
0.13858191668987274,
0.12181595712900162,
0.0045748427510261536,
0.009228081442415714,
0.0518609918653965,
0.0025186820421367884,
-0.06998204439878464,
0.054019294679164886,
0.06992026418447495,
0.12919506430625916,
-0.07847554981708527,
0.07680778950452805,
0.0006860480643808842,
-0.08370215445756912,
-0.02947772853076458,
0.11312682181596756,
-0.0409729965031147,
0.03491825982928276,
-0.047444481402635574,
0.10916327685117722,
-0.05787910893559456,
-0.29412412643432617,
0.02350960113108158,
-0.09588567912578583,
-0.15202060341835022,
-0.018367812037467957,
0.05944539234042168,
-0.02624768204987049,
0.018029648810625076,
0.06971040368080139,
-0.06011629104614258,
0.20098382234573364,
0.0335683599114418,
-0.07864278554916382,
-0.0664360448718071,
0.04837050288915634,
-0.06564252078533173,
0.2949807047843933,
0.008418165147304535,
0.02863333560526371,
0.10770907253026962,
-0.03253700211644173,
-0.18271861970424652,
0.010723991319537163,
0.1133992001414299,
-0.08056149631738663,
0.08200647681951523,
0.19000613689422607,
-0.012578671798110008,
0.1209007054567337,
0.05294662341475487,
-0.047376248985528946,
0.04217283055186272,
-0.03389401361346245,
-0.051268599927425385,
-0.10752558708190918,
0.058453381061553955,
-0.05909625440835953,
0.15447644889354706,
0.10152646154165268,
-0.05671518296003342,
-0.004550917539745569,
-0.05555408447980881,
0.04875178262591362,
0.01804669201374054,
0.12263146042823792,
0.02951994352042675,
-0.1865430772304535,
0.032826557755470276,
-0.01144319772720337,
0.10186848044395447,
-0.25588861107826233,
-0.08421015739440918,
0.08833149075508118,
-0.011924264021217823,
-0.05105875805020332,
0.10560628771781921,
0.057650718837976456,
0.04243382066488266,
-0.043439045548439026,
-0.10480839014053345,
-0.02186836116015911,
0.14663739502429962,
-0.1469624787569046,
-0.025013303384184837
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# face_emotion_recognizer
This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the imagefolder dataset.
It achieves the following results on the evaluation set:
- Loss: 1.7251
- Accuracy: 0.4188
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| No log | 1.0 | 40 | 1.9125 | 0.4125 |
| No log | 2.0 | 80 | 1.7183 | 0.4188 |
| No log | 3.0 | 120 | 1.6596 | 0.4125 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["imagefolder"], "metrics": ["accuracy"], "base_model": "google/vit-base-patch16-224-in21k", "model-index": [{"name": "face_emotion_recognizer", "results": [{"task": {"type": "image-classification", "name": "Image Classification"}, "dataset": {"name": "imagefolder", "type": "imagefolder", "config": "default", "split": "train", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.41875, "name": "Accuracy"}]}]}]} | image-classification | rendy-k/face_emotion_recognizer | [
"transformers",
"tensorboard",
"safetensors",
"vit",
"image-classification",
"generated_from_trainer",
"dataset:imagefolder",
"base_model:google/vit-base-patch16-224-in21k",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T08:20:24+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| face\_emotion\_recognizer
=========================
This model is a fine-tuned version of google/vit-base-patch16-224-in21k on the imagefolder dataset.
It achieves the following results on the evaluation set:
* Loss: 1.7251
* Accuracy: 0.4188
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
86,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.1165519431233406,
0.1323481947183609,
-0.002589212963357568,
0.12103591114282608,
0.14071886241436005,
0.005188590846955776,
0.139705091714859,
0.1393575817346573,
-0.07173100113868713,
0.08054278045892715,
0.14876572787761688,
0.13114391267299652,
0.030345233157277107,
0.18607407808303833,
-0.049024634063243866,
-0.22308041155338287,
0.026657916605472565,
0.04680468887090683,
-0.04804718494415283,
0.12098725140094757,
0.08876056969165802,
-0.136783629655838,
0.11763095110654831,
0.024404896423220634,
-0.20159287750720978,
-0.0073507847264409065,
0.027889007702469826,
-0.05720923840999603,
0.11639901250600815,
0.0372127927839756,
0.09106536954641342,
0.029656393453478813,
0.052939463406801224,
-0.1505318433046341,
0.010273020714521408,
0.07449565082788467,
-0.009181397967040539,
0.09532501548528671,
0.056325219571590424,
0.01216465886682272,
0.012989888899028301,
-0.09169132262468338,
0.040743809193372726,
0.02650708518922329,
-0.11310267448425293,
-0.2335074096918106,
-0.08633823692798615,
0.05792929232120514,
0.07824236899614334,
0.07111525535583496,
-0.0028688220772892237,
0.14341887831687927,
-0.0077273850329220295,
0.09747498482465744,
0.23024308681488037,
-0.2733115255832672,
-0.07762603461742401,
0.03984279930591583,
0.01806679368019104,
0.0786927193403244,
-0.10086648166179657,
0.01203940249979496,
0.05890904366970062,
0.016274133697152138,
0.15192632377147675,
-0.005154120735824108,
-0.011423624120652676,
-0.025894833728671074,
-0.12619882822036743,
-0.06362619996070862,
0.19216342270374298,
0.08863335847854614,
-0.04748018458485603,
-0.08144276589155197,
-0.07883680611848831,
-0.13996151089668274,
-0.046018678694963455,
-0.010846867226064205,
0.05688237398862839,
-0.034195058047771454,
-0.06462031602859497,
-0.03434380143880844,
-0.09859934449195862,
-0.06979476660490036,
-0.014257092028856277,
0.09733863919973373,
0.05384160950779915,
0.013481386937201023,
-0.01868523471057415,
0.08221779018640518,
-0.03878959268331528,
-0.14499463140964508,
-0.008276320062577724,
0.017720259726047516,
0.025156017392873764,
-0.031013010069727898,
-0.024850167334079742,
-0.11179403960704803,
0.020884696394205093,
0.1096930205821991,
-0.06981636583805084,
0.05741824582219124,
-0.01986425742506981,
0.05167732760310173,
-0.11173566430807114,
0.19050046801567078,
-0.04807506874203682,
0.017628708854317665,
0.0384475514292717,
0.10408278554677963,
0.05014200508594513,
-0.002247139113023877,
-0.10652609914541245,
0.019732769578695297,
0.12015318870544434,
0.006774095818400383,
-0.03543185815215111,
0.08127669245004654,
-0.06318386644124985,
-0.02999456599354744,
0.07963120937347412,
-0.08609656244516373,
0.02561771497130394,
-0.006639227736741304,
-0.05349783971905708,
-0.056894127279520035,
0.04623960331082344,
-0.01125720702111721,
-0.012166672386229038,
0.041881050914525986,
-0.10123040527105331,
0.012187008745968342,
-0.06722663342952728,
-0.10699499398469925,
0.013600936159491539,
-0.11544870585203171,
0.01468565035611391,
-0.12324962019920349,
-0.13767485320568085,
-0.013022191822528839,
0.06042776256799698,
-0.03130399435758591,
-0.05168246105313301,
-0.04133772477507591,
-0.08047802001237869,
0.026949135586619377,
0.00474946666508913,
0.0476473793387413,
-0.05773691460490227,
0.08816976845264435,
0.046120986342430115,
0.07698237895965576,
-0.0200846865773201,
0.04633376747369766,
-0.08758579194545746,
0.05836107209324837,
-0.2048470675945282,
0.03652292862534523,
-0.060007691383361816,
0.08667204529047012,
-0.12017237395048141,
-0.08741101622581482,
0.001324677374213934,
-0.020554782822728157,
0.06439582258462906,
0.10773269087076187,
-0.1391441524028778,
-0.05766208469867706,
0.16931891441345215,
-0.10291086882352829,
-0.15647275745868683,
0.11446194350719452,
-0.02963820844888687,
0.028713759034872055,
0.054429810494184494,
0.19654783606529236,
0.07940111309289932,
-0.10970704257488251,
-0.007916901260614395,
-0.03198160231113434,
0.03547891601920128,
-0.053363315761089325,
0.07670293748378754,
-0.0010508630657568574,
-0.010933038778603077,
0.022506600245833397,
-0.09485632926225662,
0.062171220779418945,
-0.07373321801424026,
-0.08518882840871811,
-0.06488033384084702,
-0.08821411430835724,
0.042589738965034485,
0.05897778272628784,
0.06478328257799149,
-0.10155005753040314,
-0.09094174206256866,
0.026573829352855682,
0.08071461319923401,
-0.09372373670339584,
0.018021924421191216,
-0.08160650730133057,
0.11190575361251831,
-0.1092778816819191,
0.000008168375643435866,
-0.13401943445205688,
-0.03215564042329788,
0.0498514249920845,
-0.06003656983375549,
-0.008765999227762222,
-0.037847939878702164,
0.07368729263544083,
0.06209544464945793,
-0.0644625648856163,
-0.07336944341659546,
-0.039851248264312744,
-0.003029915038496256,
-0.09838735312223434,
-0.1925726681947708,
-0.02513929456472397,
-0.027600707486271858,
0.1039295420050621,
-0.21682867407798767,
0.041051603853702545,
0.052195560187101364,
0.10093782097101212,
0.058563221246004105,
-0.031774893403053284,
0.0028878164011985064,
0.018625428900122643,
-0.040098875761032104,
-0.08946429938077927,
0.06265918165445328,
0.013831252232193947,
-0.06855639815330505,
0.007058938033878803,
-0.10018382221460342,
0.17463821172714233,
0.12972548604011536,
-0.03350052982568741,
-0.06420483440160751,
-0.005071709398180246,
-0.044098902493715286,
-0.03558432310819626,
-0.03513684868812561,
0.007663978263735771,
0.08250456303358078,
-0.008927363902330399,
0.1608663648366928,
-0.10497762262821198,
-0.026302268728613853,
0.05800237879157066,
-0.030466927215456963,
-0.03949883207678795,
0.08908171206712723,
0.06932540237903595,
-0.13753823935985565,
0.14723944664001465,
0.16730903089046478,
-0.06741046905517578,
0.12553443014621735,
-0.047188639640808105,
-0.06164444237947464,
-0.023551583290100098,
0.040897972881793976,
0.033628374338150024,
0.1286846399307251,
-0.11797640472650528,
-0.012314875610172749,
0.02303740568459034,
0.003500061808153987,
-0.007407682482153177,
-0.20054784417152405,
-0.009796563535928726,
0.038185447454452515,
-0.06102989241480827,
0.027892744168639183,
-0.005803098436444998,
-0.021509118378162384,
0.0851721540093422,
0.008919332176446915,
-0.043705690652132034,
0.04730094224214554,
0.010855515487492085,
-0.07035112380981445,
0.1955263912677765,
-0.0842631608247757,
-0.21514293551445007,
-0.13248616456985474,
-0.024147169664502144,
-0.08066541701555252,
0.021724386140704155,
0.058367207646369934,
-0.09517484158277512,
-0.05706837400794029,
-0.10546117275953293,
-0.013749380595982075,
0.029593992978334427,
0.0398063138127327,
0.04288084805011749,
-0.00281329988501966,
0.1309029757976532,
-0.09919765591621399,
-0.008200990967452526,
-0.010339846834540367,
-0.025180267170071602,
0.04924343153834343,
0.018049616366624832,
0.12053176015615463,
0.08755850046873093,
-0.025614114478230476,
0.03396854177117348,
-0.0214299988001585,
0.24250087141990662,
-0.07208780199289322,
-0.0024155564606189728,
0.15030238032341003,
0.019598156213760376,
0.06855606287717819,
0.1315159797668457,
0.038798414170742035,
-0.10375498235225677,
0.009199691005051136,
0.022889085114002228,
-0.027197055518627167,
-0.18673646450042725,
-0.017095867544412613,
-0.03906961902976036,
-0.0035322189796715975,
0.1513357013463974,
0.056140925735235214,
0.05903208255767822,
0.09312549978494644,
0.0006175086600705981,
0.08896785974502563,
-0.005170219112187624,
0.08734356611967087,
0.10983558744192123,
0.04651792719960213,
0.11035038530826569,
-0.04346156492829323,
-0.028565166518092155,
0.032922662794589996,
0.01758773997426033,
0.2258596420288086,
0.0003586753737181425,
0.17372086644172668,
0.04850974678993225,
0.18914565443992615,
0.01696128025650978,
0.05553898960351944,
-0.02192135900259018,
-0.027752922847867012,
-0.009555643424391747,
-0.054417580366134644,
-0.019942231476306915,
0.035226259380578995,
-0.048270661383867264,
0.06504014879465103,
-0.09280800074338913,
0.0419655367732048,
0.06283409148454666,
0.26434022188186646,
0.0380631685256958,
-0.378702849149704,
-0.0942588523030281,
-0.004967894870787859,
-0.014036581851541996,
-0.06143450736999512,
0.003137258579954505,
0.14408515393733978,
-0.06129135563969612,
0.05984009802341461,
-0.10333281010389328,
0.08121863752603531,
-0.05010557919740677,
0.021148210391402245,
0.0789564773440361,
0.08750289678573608,
0.008120178245007992,
0.056124892085790634,
-0.24874113500118256,
0.2579561173915863,
0.014995359815657139,
0.06339027732610703,
-0.04643380641937256,
0.012992629781365395,
0.035095397382974625,
0.10368803888559341,
0.1100526675581932,
-0.005544720217585564,
-0.014705805107951164,
-0.17761093378067017,
-0.08956919610500336,
0.006401136517524719,
0.0726199746131897,
-0.0459887757897377,
0.08232353627681732,
-0.03024131804704666,
-0.022677764296531677,
0.05050493776798248,
-0.0014745898079127073,
-0.08953549712896347,
-0.09405893832445145,
-0.007272412069141865,
0.04140733182430267,
0.013170710764825344,
-0.09514454752206802,
-0.09799089282751083,
-0.10376235097646713,
0.13376003503799438,
-0.015235471539199352,
-0.0410282276570797,
-0.12001312524080276,
0.08464493602514267,
0.05925063416361809,
-0.0925467386841774,
0.07910946011543274,
-0.028126895427703857,
0.13516655564308167,
0.02964305877685547,
-0.06258728355169296,
0.10928386449813843,
-0.06065966561436653,
-0.17449912428855896,
-0.04698718711733818,
0.1059400737285614,
-0.018479451537132263,
0.025778569281101227,
0.000657293654512614,
0.027830444276332855,
-0.01027535367757082,
-0.060087110847234726,
0.0593758150935173,
0.011301524937152863,
0.05873953551054001,
-0.013374959118664265,
-0.020165514200925827,
0.009660349227488041,
-0.06461471319198608,
-0.03063140995800495,
0.13468734920024872,
0.24747052788734436,
-0.0963728278875351,
0.005279590375721455,
0.019442733377218246,
-0.0512479804456234,
-0.1964186131954193,
0.04600051790475845,
0.06668557971715927,
0.0007783871842548251,
0.032420262694358826,
-0.1539030820131302,
0.07249294966459274,
0.0820283517241478,
-0.030655430629849434,
0.09359892457723618,
-0.2658444344997406,
-0.13340798020362854,
0.079572394490242,
0.18429149687290192,
0.06702034920454025,
-0.1460414081811905,
-0.05429479107260704,
-0.010957435704767704,
-0.0931641086935997,
0.09393112361431122,
-0.05726751312613487,
0.10618710517883301,
-0.029550114646553993,
0.00345842563547194,
0.005723967216908932,
-0.05772731825709343,
0.12989334762096405,
-0.03286895528435707,
0.10694976896047592,
-0.05779419466853142,
-0.010257193818688393,
0.07606762647628784,
-0.0770319402217865,
0.06319648772478104,
-0.09104589372873306,
0.06237801909446716,
-0.06327982991933823,
-0.01627415418624878,
-0.0712723657488823,
0.032036855816841125,
-0.019336121156811714,
-0.02621064893901348,
-0.05051662400364876,
0.024870507419109344,
0.05124181881546974,
-0.0019198822556063533,
0.19941365718841553,
0.04869740828871727,
0.08958563953638077,
0.14059363305568695,
0.044340092688798904,
-0.07648671418428421,
-0.10006602108478546,
-0.028351210057735443,
-0.028011120855808258,
0.08589788526296616,
-0.1828649938106537,
0.04973273351788521,
0.09633278101682663,
0.009273367002606392,
0.14378303289413452,
0.04722395911812782,
-0.033773940056562424,
0.019764520227909088,
0.07209594547748566,
-0.1538659781217575,
-0.16054999828338623,
-0.03131340444087982,
-0.019614113494753838,
-0.11687782406806946,
0.062024060636758804,
0.11152350157499313,
-0.08392760902643204,
0.003525851061567664,
-0.007933086715638638,
0.01631566695868969,
-0.002999776741489768,
0.1635337620973587,
0.08045348525047302,
0.04529793933033943,
-0.09153736382722855,
0.09853369742631912,
0.05268013849854469,
-0.10285185277462006,
0.022494876757264137,
0.027226150035858154,
-0.10345995426177979,
-0.03807835653424263,
0.06693289428949356,
0.14395561814308167,
-0.00031244769343174994,
-0.04989601671695709,
-0.14539380371570587,
-0.09481298923492432,
0.05762571468949318,
0.12344630062580109,
0.09325812757015228,
0.01586191914975643,
-0.011235766112804413,
-0.00020352630235720426,
-0.1038084328174591,
0.11889138072729111,
0.03174576163291931,
0.09757985174655914,
-0.21840932965278625,
0.05673594027757645,
0.017680710181593895,
0.032842427492141724,
-0.019639741629362106,
0.02971670962870121,
-0.09907104074954987,
-0.016162855550646782,
-0.06241505593061447,
0.041620977222919464,
-0.0362090989947319,
0.005332853179425001,
-0.006198203191161156,
-0.06902054697275162,
-0.06133676692843437,
0.040940284729003906,
-0.10044094175100327,
-0.04602493718266487,
0.03630116954445839,
0.07000453025102615,
-0.10183359682559967,
-0.030191749334335327,
0.02564276196062565,
-0.08066321909427643,
0.07990361005067825,
0.013267865404486656,
0.0007760329754091799,
0.023885568603873253,
-0.10031827539205551,
0.01085787732154131,
0.08448714017868042,
0.0025364812463521957,
0.029831187799572945,
-0.10417286306619644,
0.005509154871106148,
-0.0018078949069604278,
0.0015824481379240751,
-0.00699615478515625,
0.10553402453660965,
-0.13373318314552307,
-0.023934436962008476,
-0.038442570716142654,
-0.033885180950164795,
-0.058301892131567,
0.062355756759643555,
0.0863015279173851,
-0.0019346697954460979,
0.2008863240480423,
-0.08752173185348511,
0.0009649299900047481,
-0.2239176630973816,
0.004300398752093315,
-0.00459726806730032,
-0.13536016643047333,
-0.12566527724266052,
-0.027888081967830658,
0.05250002071261406,
-0.07203811407089233,
0.09599132835865021,
0.014680854976177216,
0.005211042705923319,
0.03579120710492134,
0.0036060374695807695,
0.0003007452469319105,
0.02634577453136444,
0.18441936373710632,
-0.00768197700381279,
-0.021245935931801796,
0.071430504322052,
0.018561314791440964,
0.1166866347193718,
0.08388660103082657,
0.10136276483535767,
0.1624317318201065,
-0.04277002811431885,
0.10427474975585938,
0.05028262734413147,
-0.02266664244234562,
-0.1724112331867218,
0.09977339208126068,
-0.0740966945886612,
0.1424914449453354,
-0.012829973362386227,
0.1628469079732895,
0.12104769796133041,
-0.15894842147827148,
0.027248984202742577,
-0.029555557295680046,
-0.07317375391721725,
-0.07103520631790161,
-0.14324024319648743,
-0.11685261875391006,
-0.1852739304304123,
0.015131738968193531,
-0.09863927960395813,
0.006812008563429117,
0.07533568143844604,
-0.008002867922186852,
-0.022034848108887672,
0.20631399750709534,
0.04982669651508331,
-0.0007084652897901833,
0.06960950791835785,
0.0003986731462646276,
-0.06790174543857574,
-0.05858716368675232,
-0.08376044780015945,
0.03767655789852142,
-0.0086924834176898,
0.03314293548464775,
-0.02936629392206669,
-0.008451088331639767,
0.04848368465900421,
-0.0003856987750623375,
-0.11167265474796295,
0.017148099839687347,
0.015848618000745773,
0.010576000437140465,
0.004783138632774353,
0.004587474744766951,
0.007634635549038649,
-0.00909678265452385,
0.1827525496482849,
-0.05469149351119995,
-0.007317462470382452,
-0.1184140220284462,
0.12718592584133148,
0.027091052383184433,
-0.015206380747258663,
0.02833990380167961,
-0.07893683016300201,
0.025580253452062607,
0.2181570827960968,
0.14632192254066467,
-0.019876830279827118,
-0.0011197753483429551,
-0.007078149355947971,
-0.019761933013796806,
-0.02904939092695713,
0.09375658631324768,
0.09360506385564804,
-0.04280382767319679,
-0.056142549961805344,
-0.023884721100330353,
-0.04786456376314163,
-0.01621694304049015,
-0.03726089000701904,
0.03732604905962944,
0.019226886332035065,
0.015958746895194054,
-0.06462281197309494,
0.04245367273688316,
0.01779000274837017,
-0.06993922591209412,
0.0887090265750885,
-0.19616198539733887,
-0.1390528827905655,
-0.03285215422511101,
0.09985480457544327,
-0.006918862462043762,
0.030521206557750702,
-0.021889934316277504,
0.015433182939887047,
0.0682930126786232,
-0.023958226665854454,
-0.08292330801486969,
-0.09767954051494598,
0.05181692913174629,
-0.13078364729881287,
0.24557243287563324,
-0.03272780776023865,
0.008027543313801289,
0.11222197860479355,
0.018246298655867577,
-0.11880267411470413,
0.05279287323355675,
0.026029562577605247,
-0.037948038429021835,
0.025545816868543625,
0.11117175966501236,
-0.021020427346229553,
0.10836688429117203,
0.03453453257679939,
-0.08959631621837616,
-0.017209624871611595,
-0.05268954485654831,
-0.04066532105207443,
-0.0561504140496254,
-0.023979486897587776,
-0.06903007626533508,
0.12464779615402222,
0.17412135004997253,
-0.04269969463348389,
-0.026938294991850853,
-0.06298869103193283,
0.03780514374375343,
0.09051655977964401,
0.017172006890177727,
-0.014079885557293892,
-0.22937323153018951,
0.01721780002117157,
0.022884642705321312,
-0.004662036430090666,
-0.2158614993095398,
-0.1107148826122284,
-0.016841817647218704,
-0.05281655862927437,
-0.08923662453889847,
0.08724746853113174,
0.11710873246192932,
0.050875917077064514,
-0.060574136674404144,
-0.04439282417297363,
-0.06994428485631943,
0.15989649295806885,
-0.12416505813598633,
-0.08547929674386978
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# Classifier for Academic Text Contents
This model is a fine-tuned version of [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2) on a collection of Linguistics publications.
It achieves the following results on the evaluation set:
- Loss: 0.4181
- Accuracy: 0.9193
## Model description
The model is fine-tuned with academic publications in Linguistics, to classify texts in publications into 4 classes as a filter to other tasks.
The 4 classes:
- 0: out of scope - materials that are of low significance, eg. page number and page header, noise from OCR/pdf-to-text convertion
- 1: main text - texts that are the main texts of the publication, to be used for down-stream tasks
- 2: examples - texts that are captions of the figures, or quotes or excerpts
- 3: references - references of the publication, excluding in-text citations
## Intended uses & limitations
Intended uses:
- to extract main text in academic texts for down-stream tasks
Limitations:
- training and evaluation data is limited to English, and academic texts in Linguistics
## Try it yourself with the following examples (not in training/ evaluation data)
Excerpts from Chomsky, N. (2014). Aspects of the Theory of Syntax (No. 11). MIT press.
retrieved from https://apps.dtic.mil/sti/pdfs/AD0616323.pdf
- In the case of (ioii) and (1 lii), the passive transformation will
apply to the embedded sentence, and in all four cases other
operations will give the final surface forms of (8) and (g).
- (10) (i) Noun Phrase — Verb — Noun Phrase — Sentence
(/ — persuaded — a specialist — a specialist will examine
John)
(ii) Noun Phrase — Verb — Noun Phrase — Sentence
(/ — persuaded — John — a specialist will examine John)
- (13) S
Det
Predicate-Phrase
[+Definite] nom VP
their
F1...Fm Det N
destroy [+Definite] G, ... G,
the property
- 184 SOME RESIDUAL PROBLEMS
- Peshkovskii, A. M. (1956). Russkii Sintaksis v Nauchnom Osveshchenii.
Moscow.
## Problematic cases
Definitions or findings written in point form are challenging for the model. For example:
- (2) (i) the string (1) is a Sentence (S); frighten the boy is a Verb
Phrase (VP) consisting of the Verb (V) frighten and the
Noun Phrase (NP) the boy; sincerity is also an NP; the
NP the boy consists of the Determiner (Det) the, followed
by a Noun (N); the NP sincerity consists of just an N;
the is, furthermore, an Article (Art); may is a Verbal
Auxiliary (Aux) and, furthermore, a Modal (M).
- (v) specification of a function m such that m(i) is an integer
associated with the grammar G4 as its value (with, let us
say, lower value indicated by higher number)
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 10
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.5772 | 1.0 | 762 | 0.3256 | 0.9062 |
| 0.2692 | 2.0 | 1524 | 0.3038 | 0.9163 |
| 0.217 | 3.0 | 2286 | 0.3109 | 0.9180 |
| 0.1773 | 4.0 | 3048 | 0.3160 | 0.9209 |
| 0.1619 | 5.0 | 3810 | 0.3440 | 0.9206 |
| 0.1329 | 6.0 | 4572 | 0.3675 | 0.9160 |
| 0.1165 | 7.0 | 5334 | 0.3770 | 0.9209 |
| 0.0943 | 8.0 | 6096 | 0.4012 | 0.9203 |
| 0.085 | 9.0 | 6858 | 0.4166 | 0.9196 |
| 0.0811 | 10.0 | 7620 | 0.4181 | 0.9193 |
### Framework versions
- Transformers 4.34.1
- Pytorch 2.1.0+cpu
- Datasets 2.14.7
- Tokenizers 0.14.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "sentence-transformers/all-MiniLM-L6-v2", "model-index": [{"name": "new_classifier_model", "results": []}]} | text-classification | howanching-clara/classifier_for_academic_texts | [
"transformers",
"pytorch",
"bert",
"text-classification",
"generated_from_trainer",
"base_model:sentence-transformers/all-MiniLM-L6-v2",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T08:21:32+00:00 | [] | [] | TAGS
#transformers #pytorch #bert #text-classification #generated_from_trainer #base_model-sentence-transformers/all-MiniLM-L6-v2 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| Classifier for Academic Text Contents
=====================================
This model is a fine-tuned version of sentence-transformers/all-MiniLM-L6-v2 on a collection of Linguistics publications.
It achieves the following results on the evaluation set:
* Loss: 0.4181
* Accuracy: 0.9193
Model description
-----------------
The model is fine-tuned with academic publications in Linguistics, to classify texts in publications into 4 classes as a filter to other tasks.
The 4 classes:
* 0: out of scope - materials that are of low significance, eg. page number and page header, noise from OCR/pdf-to-text convertion
* 1: main text - texts that are the main texts of the publication, to be used for down-stream tasks
* 2: examples - texts that are captions of the figures, or quotes or excerpts
* 3: references - references of the publication, excluding in-text citations
Intended uses & limitations
---------------------------
Intended uses:
* to extract main text in academic texts for down-stream tasks
Limitations:
* training and evaluation data is limited to English, and academic texts in Linguistics
Try it yourself with the following examples (not in training/ evaluation data)
------------------------------------------------------------------------------
Excerpts from Chomsky, N. (2014). Aspects of the Theory of Syntax (No. 11). MIT press.
retrieved from URL
* In the case of (ioii) and (1 lii), the passive transformation will
apply to the embedded sentence, and in all four cases other
operations will give the final surface forms of (8) and (g).
* (10) (i) Noun Phrase — Verb — Noun Phrase — Sentence
(/ — persuaded — a specialist — a specialist will examine
John)
(ii) Noun Phrase — Verb — Noun Phrase — Sentence
(/ — persuaded — John — a specialist will examine John)
* (13) S
Det
Predicate-Phrase
[+Definite] nom VP
their
F1...Fm Det N
destroy [+Definite] G, ... G,
the property
* 184 SOME RESIDUAL PROBLEMS
* Peshkovskii, A. M. (1956). Russkii Sintaksis v Nauchnom Osveshchenii.
Moscow.
Problematic cases
-----------------
Definitions or findings written in point form are challenging for the model. For example:
* (2) (i) the string (1) is a Sentence (S); frighten the boy is a Verb
Phrase (VP) consisting of the Verb (V) frighten and the
Noun Phrase (NP) the boy; sincerity is also an NP; the
NP the boy consists of the Determiner (Det) the, followed
by a Noun (N); the NP sincerity consists of just an N;
the is, furthermore, an Article (Art); may is a Verbal
Auxiliary (Aux) and, furthermore, a Modal (M).
* (v) specification of a function m such that m(i) is an integer
associated with the grammar G4 as its value (with, let us
say, lower value indicated by higher number)
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 10
### Training results
### Framework versions
* Transformers 4.34.1
* Pytorch 2.1.0+cpu
* Datasets 2.14.7
* Tokenizers 0.14.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.34.1\n* Pytorch 2.1.0+cpu\n* Datasets 2.14.7\n* Tokenizers 0.14.1"
] | [
"TAGS\n#transformers #pytorch #bert #text-classification #generated_from_trainer #base_model-sentence-transformers/all-MiniLM-L6-v2 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.34.1\n* Pytorch 2.1.0+cpu\n* Datasets 2.14.7\n* Tokenizers 0.14.1"
] | [
72,
98,
4,
35
] | [
"passage: TAGS\n#transformers #pytorch #bert #text-classification #generated_from_trainer #base_model-sentence-transformers/all-MiniLM-L6-v2 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10### Training results### Framework versions\n\n\n* Transformers 4.34.1\n* Pytorch 2.1.0+cpu\n* Datasets 2.14.7\n* Tokenizers 0.14.1"
] | [
-0.06626410037279129,
0.05729692429304123,
-0.002721177414059639,
0.12302853167057037,
0.1616990864276886,
0.02693645842373371,
0.1230107992887497,
0.12679098546504974,
-0.11922870576381683,
0.019220709800720215,
0.1196640133857727,
0.11732310801744461,
0.0012075247941538692,
0.09206071496009827,
-0.04243756830692291,
-0.2841152250766754,
0.004184043500572443,
0.03495683893561363,
-0.07869654148817062,
0.1298476904630661,
0.11212750524282455,
-0.1255912333726883,
0.08667004108428955,
-0.007276800461113453,
-0.1905706375837326,
0.01402669120579958,
-0.005418430548161268,
-0.04890555888414383,
0.12675358355045319,
0.034629177302122116,
0.09172321110963821,
-0.0014469856396317482,
0.08055715262889862,
-0.1674576699733734,
0.003547224448993802,
0.052326373755931854,
0.0104834558442235,
0.07618117332458496,
0.016294732689857483,
-0.001395216677337885,
0.1277778148651123,
-0.09828493744134903,
0.05063273012638092,
0.01866881549358368,
-0.12678764760494232,
-0.2292105257511139,
-0.05727003142237663,
0.05450267344713211,
0.09678319096565247,
0.09818489104509354,
-0.012991700321435928,
0.12525172531604767,
-0.09850279986858368,
0.10014642775058746,
0.23495447635650635,
-0.29390865564346313,
-0.05990445986390114,
0.03383501246571541,
0.022937392815947533,
0.08327136188745499,
-0.08056376129388809,
-0.028854483738541603,
0.0451069250702858,
0.04365501552820206,
0.1270359605550766,
-0.037360526621341705,
-0.06620066612958908,
0.017369741573929787,
-0.1487015336751938,
-0.033598706126213074,
0.17556172609329224,
0.03694518283009529,
-0.03327501565217972,
-0.03870595246553421,
-0.060879651457071304,
-0.14441081881523132,
-0.058852747082710266,
0.005226850509643555,
0.05966946855187416,
-0.03134937584400177,
-0.06656067073345184,
-0.017543906345963478,
-0.08959241211414337,
-0.07017982751131058,
-0.07139334827661514,
0.1578710824251175,
0.03291720896959305,
-0.004327891394495964,
-0.03842111676931381,
0.09434878081083298,
-0.01821153797209263,
-0.13382591307163239,
0.008489040657877922,
0.025740113109350204,
0.0309313852339983,
-0.03485650569200516,
-0.08724761009216309,
-0.012346606701612473,
0.0024024350568652153,
0.11927998065948486,
-0.026735911145806313,
0.045297954231500626,
0.03946443647146225,
0.04172024130821228,
-0.08957929164171219,
0.19281931221485138,
-0.030166326090693474,
-0.026168253272771835,
0.018095728009939194,
0.06286252290010452,
0.04175751656293869,
-0.009112240746617317,
-0.14437568187713623,
0.02646925486624241,
0.10048674046993256,
0.03112994134426117,
-0.07245568931102753,
0.07269540429115295,
-0.03985115885734558,
-0.04412761330604553,
0.022669216617941856,
-0.10124258697032928,
0.025338798761367798,
0.000015254268873832189,
-0.07456987351179123,
0.007211396936327219,
0.010184702463448048,
0.015151018276810646,
-0.03829144686460495,
0.10016708821058273,
-0.09187082201242447,
0.02160049043595791,
-0.09000834077596664,
-0.12291362136602402,
0.006790508981794119,
-0.05373179540038109,
0.03619490936398506,
-0.10919756442308426,
-0.17396801710128784,
0.001740714069455862,
0.05333390459418297,
-0.034731172025203705,
-0.08608538657426834,
-0.059671320021152496,
-0.07397791743278503,
0.008583844639360905,
-0.008524990640580654,
0.1059824749827385,
-0.06107039377093315,
0.08680804818868637,
0.06631360203027725,
0.06759577989578247,
-0.04785306751728058,
0.058353252708911896,
-0.11518443375825882,
0.018383512273430824,
-0.15451191365718842,
0.052587635815143585,
-0.030887391418218613,
0.08692534267902374,
-0.09322213381528854,
-0.11583758890628815,
0.02392706088721752,
-0.00032252384698949754,
0.06132667884230614,
0.0966453105211258,
-0.1479233354330063,
-0.09618701040744781,
0.15968401730060577,
-0.07181429117918015,
-0.12611420452594757,
0.1254243552684784,
-0.044680770486593246,
0.029776060953736305,
0.08477777987718582,
0.18356181681156158,
0.07141204923391342,
-0.05109593644738197,
0.047694530338048935,
0.023717904463410378,
0.08513648808002472,
-0.05245712399482727,
0.09592220187187195,
0.012168514542281628,
0.00855723675340414,
0.03368173912167549,
-0.05405299365520477,
0.04586159065365791,
-0.08485223352909088,
-0.08619280159473419,
-0.04138605296611786,
-0.08884791284799576,
0.038471248000860214,
0.059350814670324326,
0.0831807404756546,
-0.11455920338630676,
-0.08681527525186539,
0.08826802670955658,
0.09594020247459412,
-0.07566255331039429,
0.04073210060596466,
-0.07477773725986481,
0.06354459375143051,
-0.02698538266122341,
-0.0044916002079844475,
-0.16796210408210754,
0.016030970960855484,
0.011839322745800018,
0.006217938382178545,
0.045998841524124146,
0.007033385802060366,
0.059390172362327576,
0.054863911122083664,
-0.06246112659573555,
-0.00924015510827303,
-0.037584591656923294,
-0.005793295335024595,
-0.12105662375688553,
-0.1833854764699936,
-0.054725322872400284,
-0.0134309446439147,
0.12671349942684174,
-0.2042308896780014,
0.03850284963846207,
-0.009404038079082966,
0.05850556492805481,
0.00947477389127016,
0.005179318133741617,
-0.05832652375102043,
0.07742148637771606,
-0.02972649224102497,
-0.05526361241936684,
0.09414354711771011,
0.015467294491827488,
-0.08298082649707794,
-0.036024309694767,
-0.12762439250946045,
0.13946551084518433,
0.14647948741912842,
-0.10626325011253357,
-0.0449017770588398,
0.003642809111624956,
-0.03983274847269058,
-0.012317701242864132,
-0.015074453316628933,
0.006864252034574747,
0.18105584383010864,
-0.004003508482128382,
0.16474077105522156,
-0.08658084273338318,
-0.05037865787744522,
0.013812288641929626,
-0.04445173591375351,
0.018305379897356033,
0.13070525228977203,
0.08138447254896164,
-0.10924933850765228,
0.1373414546251297,
0.17291614413261414,
-0.10040856897830963,
0.12652748823165894,
-0.04712989926338196,
-0.05229196324944496,
-0.0203066635876894,
-0.05030661076307297,
-0.00918572023510933,
0.06355040520429611,
-0.14546163380146027,
-0.010024270974099636,
0.022208774462342262,
0.031511206179857254,
0.021588487550616264,
-0.19586361944675446,
-0.01882002502679825,
0.057880502194166183,
-0.05030839145183563,
-0.01895105466246605,
-0.02718471735715866,
-0.009671756066381931,
0.10296984016895294,
0.021938079968094826,
-0.06381995230913162,
0.03008057177066803,
-0.004714563954621553,
-0.07895077019929886,
0.20462679862976074,
-0.08817511051893234,
-0.1606895476579666,
-0.11664015054702759,
-0.06059427931904793,
-0.05669273063540459,
0.017767267301678658,
0.06897760927677155,
-0.09592112898826599,
-0.03554605692625046,
-0.0937984511256218,
0.033948954194784164,
-0.02992112748324871,
0.012849043123424053,
0.026974275708198547,
-0.0017936958465725183,
0.05763079226016998,
-0.11622045934200287,
-0.019069820642471313,
-0.03770377114415169,
-0.0693015530705452,
0.020039238035678864,
0.00713313976302743,
0.06650891155004501,
0.1664610058069229,
-0.007277391385287046,
0.016539903357625008,
-0.03810062259435654,
0.2635018229484558,
-0.04956301301717758,
-0.03015931136906147,
0.17467008531093597,
0.0025740324053913355,
0.04463782534003258,
0.11172476410865784,
0.04926407337188721,
-0.08759594708681107,
0.0167077723890543,
0.02455565519630909,
-0.03122984804213047,
-0.1864151507616043,
-0.06720656156539917,
-0.06612378358840942,
0.004193207714706659,
0.1022268608212471,
0.020309248939156532,
-0.010061043314635754,
0.0511770062148571,
0.017426757141947746,
0.09117800742387772,
-0.03588569164276123,
0.07556559890508652,
0.1779043972492218,
0.048678919672966,
0.14639440178871155,
-0.04355927184224129,
-0.06421809643507004,
0.03330899029970169,
-0.02630680613219738,
0.1927066296339035,
0.015189696103334427,
0.15107229351997375,
0.03631512075662613,
0.12947165966033936,
0.00419862475246191,
0.06679623574018478,
-0.030022311955690384,
-0.03249145671725273,
-0.030685944482684135,
-0.04587911069393158,
-0.04831283912062645,
0.017880506813526154,
-0.04970403015613556,
0.05216289311647415,
-0.1405600756406784,
-0.016131103038787842,
0.05832076072692871,
0.2836815416812897,
0.042220208793878555,
-0.317014217376709,
-0.11701294034719467,
0.03718749061226845,
-0.04509438946843147,
0.0022531484719365835,
0.03219398483633995,
0.07680407166481018,
-0.08819252252578735,
0.024857735261321068,
-0.04793820157647133,
0.10203561186790466,
-0.03726176545023918,
0.06422670185565948,
0.054516859352588654,
0.08239399641752243,
0.024896766990423203,
0.0890020951628685,
-0.2998943626880646,
0.25542742013931274,
0.0005551534122787416,
0.058187734335660934,
-0.06296438723802567,
0.005514351185411215,
0.027770215645432472,
0.08982875198125839,
0.05186464637517929,
0.0007692122599110007,
-0.040371041744947433,
-0.1795760840177536,
-0.021428119391202927,
0.03912516310811043,
0.09710916876792908,
-0.05420144647359848,
0.10002495348453522,
-0.06147124618291855,
0.0036851733457297087,
0.06709060072898865,
0.037168897688388824,
-0.031246770173311234,
-0.0999818742275238,
-0.0007693592342548072,
0.023939261212944984,
-0.02125542238354683,
-0.06136823818087578,
-0.10553821921348572,
-0.10623560100793839,
0.1633371114730835,
-0.016760962083935738,
-0.04970061779022217,
-0.09926901012659073,
0.06657445430755615,
0.05913399159908295,
-0.08777419477701187,
0.03685334697365761,
-0.0026642547454684973,
0.08333506435155869,
0.03876464068889618,
-0.08332961052656174,
0.11292779445648193,
-0.07371414452791214,
-0.16743682324886322,
-0.03264399245381355,
0.1317111849784851,
0.005992817226797342,
0.07479313760995865,
-0.009503370150923729,
0.01265768799930811,
-0.03639370948076248,
-0.08679903298616409,
-0.005471502896398306,
-0.014904465526342392,
0.06569001078605652,
0.018926767632365227,
-0.06747503578662872,
0.051643650978803635,
-0.05301329866051674,
-0.01743524894118309,
0.2070150077342987,
0.2675187289714813,
-0.0757564976811409,
0.025722695514559746,
0.04573848843574524,
-0.06841713190078735,
-0.201844722032547,
0.005271166563034058,
0.04696127772331238,
0.00465068593621254,
0.03241533413529396,
-0.17461277544498444,
0.10904061794281006,
0.0941312238574028,
-0.009938876144587994,
0.07800254225730896,
-0.3412962555885315,
-0.12830576300621033,
0.11553801596164703,
0.1180771142244339,
0.11809388548135757,
-0.14125700294971466,
-0.011382093653082848,
-0.052033957093954086,
-0.13219931721687317,
0.14530044794082642,
-0.09463542699813843,
0.1327507197856903,
-0.021514609456062317,
0.06658466160297394,
0.006502086296677589,
-0.03987734764814377,
0.122139111161232,
0.009865229949355125,
0.09365622699260712,
-0.05784690007567406,
-0.054860059171915054,
0.045620497316122055,
-0.037194062024354935,
0.017547331750392914,
-0.12820255756378174,
0.041993144899606705,
-0.10783881694078445,
-0.025235049426555634,
-0.06859548389911652,
0.012425837107002735,
-0.03484056144952774,
-0.06471897661685944,
-0.0472869910299778,
0.04064195603132248,
0.057755596935749054,
-0.001440458232536912,
0.1531394124031067,
-0.010515333153307438,
0.1249808818101883,
0.14922088384628296,
0.07912479341030121,
-0.09745261818170547,
-0.044430818408727646,
-0.012593811377882957,
-0.008053487166762352,
0.07248307764530182,
-0.15581104159355164,
0.05325690284371376,
0.13212557137012482,
0.020207872614264488,
0.1345575451850891,
0.08271981030702591,
-0.0036793916951864958,
-0.004111338872462511,
0.04798528552055359,
-0.1479772925376892,
-0.08284913003444672,
-0.02251046523451805,
-0.0403289869427681,
-0.11720392853021622,
0.03807383030653,
0.12810811400413513,
-0.0684819370508194,
-0.018810270354151726,
-0.006174910347908735,
0.010099575854837894,
-0.06791670620441437,
0.18164797127246857,
0.05692015588283539,
0.055528610944747925,
-0.10710791498422623,
0.07724476605653763,
0.06102386862039566,
-0.08063919097185135,
0.019762953743338585,
0.09923380613327026,
-0.07799118012189865,
-0.05807903781533241,
0.07154753804206848,
0.1313544660806656,
-0.04424607381224632,
-0.03096017614006996,
-0.14093439280986786,
-0.14305607974529266,
0.07424840331077576,
0.11439487338066101,
0.12665528059005737,
0.000868814589921385,
-0.07976311445236206,
0.018652278929948807,
-0.11522533744573593,
0.09152930974960327,
0.031101500615477562,
0.05441885441541672,
-0.1297539323568344,
0.1702450066804886,
-0.00312820659019053,
0.05128241702914238,
-0.02432454563677311,
0.013237458653748035,
-0.11660795658826828,
0.012896474450826645,
-0.126207634806633,
-0.02623242698609829,
-0.019839998334646225,
0.010949408635497093,
-0.007100717630237341,
-0.04250405728816986,
-0.06938384473323822,
0.015034795738756657,
-0.11714501678943634,
-0.03477025777101517,
0.009139858186244965,
0.05355104058980942,
-0.09328561276197433,
-0.02878507785499096,
0.029222581535577774,
-0.06314448267221451,
0.06612057983875275,
0.0431644581258297,
0.010262181051075459,
0.056859660893678665,
-0.10881316661834717,
0.005789049901068211,
0.05069616064429283,
0.027547912672162056,
0.05662646144628525,
-0.10514523088932037,
-0.0017973858630284667,
0.010135985910892487,
0.061554450541734695,
0.02078542485833168,
0.09081225097179413,
-0.1339755803346634,
-0.007861204445362091,
-0.025996387004852295,
-0.08405718207359314,
-0.05773463472723961,
0.010963963344693184,
0.04840271547436714,
0.01883549615740776,
0.19789768755435944,
-0.07946091890335083,
0.030901361256837845,
-0.2057185173034668,
-0.0037363937590271235,
-0.0067390077747404575,
-0.11936251074075699,
-0.14048708975315094,
-0.06900785863399506,
0.05193018913269043,
-0.05138961225748062,
0.13659289479255676,
0.030954552814364433,
0.04778816178441048,
0.04428485035896301,
-0.010430504567921162,
0.048754919320344925,
0.025263650342822075,
0.23353028297424316,
0.047434285283088684,
-0.03797684982419014,
0.06017482280731201,
0.03657859191298485,
0.10810365527868271,
0.1240040585398674,
0.203605055809021,
0.15430817008018494,
-0.011152404360473156,
0.10786225646734238,
0.0288098081946373,
-0.04004877433180809,
-0.14890176057815552,
0.029120435938239098,
-0.002123425481840968,
0.11845674365758896,
-0.027233509346842766,
0.2089162915945053,
0.08852830529212952,
-0.1813291609287262,
0.04833567515015602,
-0.07292435318231583,
-0.08120936900377274,
-0.10673798620700836,
-0.07317758351564407,
-0.09141078591346741,
-0.17365339398384094,
-0.010090789757668972,
-0.126870259642601,
0.020366095006465912,
0.11323462426662445,
0.001670068479143083,
-0.02803795225918293,
0.12632431089878082,
-0.017337678000330925,
0.0019154901383444667,
0.06979362666606903,
-0.015073578804731369,
-0.024683021008968353,
-0.07767292112112045,
-0.06379014998674393,
-0.00042398611549288034,
-0.009929592721164227,
0.03764894977211952,
-0.030929604545235634,
-0.04579257220029831,
0.01985282078385353,
-0.03704997897148132,
-0.10709215700626373,
0.005448869429528713,
0.026367060840129852,
0.0664324164390564,
0.053484465926885605,
-0.004695338662713766,
-0.0028352485969662666,
-0.007303845603018999,
0.2392081469297409,
-0.09637332707643509,
-0.10709735751152039,
-0.08603508025407791,
0.2625310719013214,
0.05316736921668053,
0.006649557966738939,
0.02618466317653656,
-0.06506161391735077,
-0.027326783165335655,
0.25936201214790344,
0.21879640221595764,
-0.10486051440238953,
-0.003218392375856638,
0.002279361942782998,
-0.006074531935155392,
-0.0065450966358184814,
0.11931474506855011,
0.1444437950849533,
0.026891617104411125,
-0.09405772387981415,
-0.028866039589047432,
-0.045986827462911606,
-0.01677050068974495,
-0.05881199613213539,
0.09130625426769257,
0.05536171793937683,
-0.0058433134108781815,
-0.015526246279478073,
0.041434142738580704,
-0.0717262551188469,
-0.05534111335873604,
0.03392894193530083,
-0.21158471703529358,
-0.1585388481616974,
-0.012173227965831757,
0.08636754006147385,
0.025509528815746307,
0.07799503952264786,
-0.029478181153535843,
-0.0024528673384338617,
0.0669880285859108,
-0.01414299476891756,
-0.07748018950223923,
-0.10600616782903671,
0.11952424794435501,
-0.10476464778184891,
0.18175458908081055,
-0.03651602193713188,
0.06496620923280716,
0.12141723930835724,
0.05851544067263603,
-0.06780765950679779,
0.06251354515552521,
0.03950352221727371,
-0.06933678686618805,
0.03993766009807587,
0.08630946278572083,
-0.039497461169958115,
0.08289840817451477,
0.05077784135937691,
-0.13830867409706116,
0.005758864805102348,
-0.02090795338153839,
-0.04673752188682556,
-0.03160817548632622,
-0.060462936758995056,
-0.07853007316589355,
0.12492823600769043,
0.21502582728862762,
-0.046995989978313446,
-0.019459791481494904,
-0.06359145790338516,
0.023374125361442566,
0.05572754889726639,
0.03202204033732414,
-0.06404484063386917,
-0.22231453657150269,
0.0007466327515430748,
0.06786741316318512,
-0.011489639058709145,
-0.25431185960769653,
-0.08812163770198822,
0.006257942877709866,
-0.04606156051158905,
-0.09567051380872726,
0.09138581901788712,
0.08528212457895279,
0.027247270569205284,
-0.06461837142705917,
-0.09736510366201401,
-0.0777638852596283,
0.1514366865158081,
-0.15475617349147797,
-0.10269635915756226
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.8.2 | {"library_name": "peft", "base_model": "tiiuae/falcon-7b-instruct"} | null | madhiarasan/hr_qna | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:tiiuae/falcon-7b-instruct",
"region:us"
] | 2024-02-08T08:21:44+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #safetensors #arxiv-1910.09700 #base_model-tiiuae/falcon-7b-instruct #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.8.2 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
"TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-tiiuae/falcon-7b-instruct #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
39,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-tiiuae/falcon-7b-instruct #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2"
] | [
-0.11734078079462051,
0.20820164680480957,
-0.002760490635409951,
0.034437160938978195,
0.08834109455347061,
0.0204817745834589,
0.05106845498085022,
0.12554429471492767,
-0.010636605322360992,
0.10709162056446075,
0.0669572725892067,
0.10690649598836899,
0.11333406716585159,
0.22011126577854156,
0.0018083102768287063,
-0.18448278307914734,
0.027663569897413254,
-0.09058341383934021,
0.0011518682586029172,
0.12491855025291443,
0.14306765794754028,
-0.10104523599147797,
0.08355220407247543,
-0.015142649412155151,
-0.0067593310959637165,
-0.03543645888566971,
-0.06823195517063141,
-0.0214981846511364,
0.03986876457929611,
0.03997591510415077,
0.06164052337408066,
-0.00845589954406023,
0.09297027438879013,
-0.2600044906139374,
0.019212251529097557,
0.04829448089003563,
-0.0012513687834143639,
0.08909079432487488,
0.09872765094041824,
-0.038300834596157074,
0.11861186474561691,
-0.027645057067275047,
0.14010697603225708,
0.08736509829759598,
-0.08881652355194092,
-0.21763654053211212,
-0.06756764650344849,
0.09159471839666367,
0.18884752690792084,
0.07710597664117813,
-0.038632847368717194,
0.12676239013671875,
-0.07497035712003708,
0.023656994104385376,
0.030665580183267593,
-0.08688811957836151,
-0.06844764947891235,
0.05670436471700668,
0.11523283272981644,
0.060738325119018555,
-0.12972652912139893,
-0.036145299673080444,
0.02803785726428032,
0.03652189299464226,
0.07825840264558792,
0.008712342008948326,
0.1569414585828781,
0.02849903330206871,
-0.14243187010288239,
-0.04446979984641075,
0.13761988282203674,
0.024349292740225792,
-0.04069056734442711,
-0.22757668793201447,
-0.007875264622271061,
-0.09177695214748383,
-0.02792537584900856,
-0.05287682265043259,
0.03472858667373657,
0.01073219534009695,
0.11962463706731796,
-0.03602224960923195,
-0.09367720037698746,
-0.01926487870514393,
0.09299173206090927,
0.04539812356233597,
0.02439892664551735,
-0.017480986192822456,
0.005412148777395487,
0.12425576150417328,
0.07897710055112839,
-0.13216376304626465,
-0.058012548834085464,
-0.07757290452718735,
-0.04606996849179268,
-0.037848249077796936,
0.0474659763276577,
0.04347943887114525,
0.05706462636590004,
0.25204044580459595,
-0.028847239911556244,
0.06091305613517761,
0.06764272600412369,
0.018567340448498726,
0.04833822324872017,
0.09788120537996292,
-0.05105748400092125,
-0.16349248588085175,
-0.010711210779845715,
0.09655610471963882,
-0.004537667613476515,
-0.028424235060811043,
-0.05277970805764198,
0.043664027005434036,
0.03272606432437897,
0.10913430899381638,
0.10617480427026749,
-0.012899474240839481,
-0.07856442034244537,
-0.06300023198127747,
0.21868327260017395,
-0.15423104166984558,
0.048134300857782364,
0.022934982553124428,
-0.008047125302255154,
-0.04165388271212578,
0.010160334408283234,
0.01743599772453308,
-0.03021341748535633,
0.07781604677438736,
-0.06899859011173248,
-0.043093565851449966,
-0.1244254857301712,
-0.028180714696645737,
0.0282726027071476,
0.0035621020942926407,
-0.0349862203001976,
-0.03905573487281799,
-0.07636228203773499,
-0.09671091288328171,
0.10624849796295166,
-0.06087924912571907,
-0.0570172443985939,
-0.02860010415315628,
-0.09400805085897446,
0.025472750887274742,
0.027243241667747498,
0.07497099786996841,
-0.028231341391801834,
0.0424661822617054,
-0.012435317039489746,
0.06297753006219864,
0.07383302599191666,
0.031020499765872955,
-0.06957783550024033,
0.06122349202632904,
-0.19468949735164642,
0.07909191399812698,
-0.07783868908882141,
0.03272624686360359,
-0.15836361050605774,
-0.008405922912061214,
0.01527400966733694,
0.023668667301535606,
0.031263645738363266,
0.16287365555763245,
-0.21567176282405853,
-0.026980040594935417,
0.15905427932739258,
-0.10402932018041611,
-0.11932123452425003,
0.036806248128414154,
-0.041929468512535095,
0.164528027176857,
0.02560739777982235,
-0.006919456645846367,
0.101661816239357,
-0.160472571849823,
-0.02574020065367222,
-0.016291329637169838,
-0.004099872428923845,
0.08382543176412582,
0.08616536110639572,
-0.08437233418226242,
0.024365490302443504,
0.014519273303449154,
-0.04861389845609665,
-0.019686639308929443,
-0.0424431636929512,
-0.10493219643831253,
0.00982362125068903,
-0.08487288653850555,
0.018341809511184692,
-0.004678139928728342,
-0.09071797877550125,
-0.003867991967126727,
-0.15798506140708923,
-0.04791360720992088,
0.0856880396604538,
0.004394585732370615,
-0.02406136505305767,
-0.10503757745027542,
0.04861176386475563,
-0.0343891941010952,
-0.02259737439453602,
-0.13956843316555023,
-0.022136427462100983,
0.018811620771884918,
-0.13170069456100464,
-0.0075032273307442665,
-0.11644557863473892,
0.0661802887916565,
0.013129648752510548,
-0.053224533796310425,
-0.04385121539235115,
-0.0013948752311989665,
0.006464752834290266,
-0.05481293797492981,
-0.24181492626667023,
-0.030536426231265068,
-0.05113111436367035,
0.1532227247953415,
-0.22319543361663818,
0.04145006835460663,
0.03755781799554825,
0.12765850126743317,
0.004889278206974268,
-0.06728103011846542,
0.023435447365045547,
-0.07114197313785553,
-0.027438007295131683,
-0.07537958025932312,
-0.005312131252139807,
-0.001255533774383366,
-0.03440462797880173,
0.01847194693982601,
-0.12120147049427032,
-0.041046593338251114,
0.10160775482654572,
0.06494741141796112,
-0.1551491618156433,
-0.00039427538285963237,
-0.04045442119240761,
-0.06156684085726738,
-0.07303357124328613,
-0.06866736710071564,
0.09614787250757217,
0.054180826991796494,
0.04242325946688652,
-0.07318048179149628,
-0.07518955320119858,
0.010552770458161831,
-0.022736676037311554,
-0.01681915856897831,
0.1113535538315773,
0.0780043751001358,
-0.10704155266284943,
0.09712038189172745,
0.07562727481126785,
0.03677409514784813,
0.08657705783843994,
-0.024839719757437706,
-0.10636695474386215,
-0.03040473349392414,
0.047899890691041946,
0.012615017592906952,
0.1704326868057251,
-0.06202604994177818,
0.05550690367817879,
0.04573486000299454,
-0.03990234062075615,
0.048711009323596954,
-0.0919276773929596,
0.011583262123167515,
0.00940575823187828,
-0.013554096221923828,
0.0226000864058733,
-0.027275582775473595,
0.009116844274103642,
0.079106904566288,
0.054089102894067764,
0.02953213080763817,
0.0241659227758646,
-0.03355247899889946,
-0.13610456883907318,
0.1810297966003418,
-0.099529929459095,
-0.2415156364440918,
-0.16298052668571472,
0.058940399438142776,
0.05297185480594635,
-0.015385766513645649,
0.01677166111767292,
-0.053791940212249756,
-0.10769476741552353,
-0.08647291362285614,
-0.0014049314195290208,
0.03062673844397068,
-0.056574124842882156,
-0.0666913092136383,
0.0445503331720829,
0.04421960189938545,
-0.1220373660326004,
0.03130371868610382,
0.0626428872346878,
-0.017061052843928337,
-0.0034029216039925814,
0.0596851147711277,
0.09030809998512268,
0.1851290464401245,
-0.0046744938008487225,
0.00021215260494500399,
0.05979183316230774,
0.2779941260814667,
-0.15687233209609985,
0.12215100228786469,
0.1351831704378128,
-0.07280522584915161,
0.07658097892999649,
0.18960417807102203,
0.03309699892997742,
-0.09664555639028549,
0.023184947669506073,
0.02723373845219612,
-0.020357714965939522,
-0.26056692004203796,
-0.05391215160489082,
-0.017888054251670837,
-0.08452443033456802,
0.07397046685218811,
0.0895681157708168,
0.08042965084314346,
0.03575718402862549,
-0.06749409437179565,
-0.0944945365190506,
0.030778443440794945,
0.10439611971378326,
-0.031735196709632874,
0.005476905964314938,
0.08204668760299683,
-0.0392841137945652,
0.010351463221013546,
0.09800240397453308,
-0.010100106708705425,
0.1521749645471573,
0.0564655102789402,
0.10680194199085236,
0.08092774450778961,
0.09539657086133957,
-0.0052260784432291985,
0.03660063445568085,
0.014500129967927933,
0.025488533079624176,
0.015615060925483704,
-0.08435112237930298,
0.022825926542282104,
0.11183862388134003,
0.034905415028333664,
0.031148940324783325,
0.01996702142059803,
-0.042640745639801025,
0.04672888666391373,
0.1868085414171219,
0.01611233688890934,
-0.21257570385932922,
-0.08143191039562225,
0.059313833713531494,
-0.08114693313837051,
-0.14790968596935272,
-0.01207645982503891,
0.037354230880737305,
-0.16320572793483734,
0.017727559432387352,
-0.037746917456388474,
0.1037052795290947,
-0.09459785372018814,
-0.041316598653793335,
0.11052058637142181,
0.05967207998037338,
-0.017985213547945023,
0.04620477557182312,
-0.17665433883666992,
0.11644484102725983,
0.030528616160154343,
0.07711417973041534,
-0.08602524548768997,
0.10444454103708267,
0.0000051857355174433906,
-0.00811526458710432,
0.16608795523643494,
0.0029618169646710157,
-0.045948661863803864,
-0.08756425976753235,
-0.09929921478033066,
-0.012035014107823372,
0.09140215069055557,
-0.1369546353816986,
0.07538863271474838,
-0.027216481044888496,
-0.030222872272133827,
-0.004024986177682877,
-0.092646524310112,
-0.13049808144569397,
-0.16985760629177094,
0.05329817906022072,
-0.09450704604387283,
0.023748332634568214,
-0.08739619702100754,
-0.05406913161277771,
0.005525992251932621,
0.18045683205127716,
-0.22867515683174133,
-0.1045476645231247,
-0.15230704843997955,
-0.10835380107164383,
0.16008785367012024,
-0.041446931660175323,
0.08645131438970566,
0.0008634133846499026,
0.16138069331645966,
0.013235794380307198,
-0.015416119247674942,
0.09868823736906052,
-0.09570268541574478,
-0.18909385800361633,
-0.05569318309426308,
0.16112017631530762,
0.14340999722480774,
0.029500320553779602,
-0.012264356948435307,
0.028235074132680893,
-0.05768236890435219,
-0.12141615152359009,
0.02324010245501995,
0.16108977794647217,
0.07501135021448135,
-0.016943715512752533,
-0.022342151030898094,
-0.10978677123785019,
-0.053877923637628555,
-0.0397106371819973,
-0.010958577506244183,
0.19116513431072235,
-0.07318787276744843,
0.15455874800682068,
0.11030265688896179,
-0.056548912078142166,
-0.20802472531795502,
0.036502011120319366,
0.04605567455291748,
0.02085144817829132,
0.04067692160606384,
-0.18539230525493622,
0.08975748717784882,
-0.011804462410509586,
-0.07642050087451935,
0.1684717833995819,
-0.16506439447402954,
-0.13687142729759216,
0.10286331176757812,
0.030901620164513588,
-0.22392518818378448,
-0.13503752648830414,
-0.09907813370227814,
-0.020382920280098915,
-0.13290058076381683,
0.05146527290344238,
-0.00040726392762735486,
0.006460164673626423,
0.02338598482310772,
0.005979300010949373,
0.026668768376111984,
-0.05003682151436806,
0.20502403378486633,
-0.030186811462044716,
0.007683186791837215,
-0.048794638365507126,
-0.08358537405729294,
0.02691052481532097,
-0.04846952110528946,
0.10678016394376755,
-0.0039045207668095827,
0.029312437400221825,
-0.16362617909908295,
-0.03971630334854126,
-0.05068541690707207,
0.02673187479376793,
-0.09168225526809692,
-0.08633565902709961,
-0.0407903678715229,
0.09282858669757843,
0.09546621888875961,
-0.026953555643558502,
-0.0034133356530219316,
-0.08851537853479385,
0.062139350920915604,
0.21209976077079773,
0.19339919090270996,
0.063711978495121,
-0.05808789283037186,
0.017779141664505005,
-0.0337221585214138,
0.045908644795417786,
-0.21906277537345886,
0.04234527423977852,
0.05514661967754364,
0.017928361892700195,
0.0756688266992569,
-0.009760073386132717,
-0.15216703712940216,
-0.07274731993675232,
0.08505624532699585,
-0.05825933814048767,
-0.16530942916870117,
-0.0289873369038105,
0.016428479924798012,
-0.20418566465377808,
-0.03811594098806381,
0.029709363356232643,
-0.017782993614673615,
-0.037604957818984985,
0.020899564027786255,
0.08174248784780502,
-0.023955313488841057,
0.09953639656305313,
0.08600252866744995,
0.09164861589670181,
-0.10287722945213318,
0.06654666364192963,
0.07328788191080093,
-0.03730612248182297,
0.029470540583133698,
0.11463844031095505,
-0.04918389394879341,
-0.03605872392654419,
0.08037924766540527,
0.10282153636217117,
0.021972520276904106,
-0.05755160003900528,
0.009237207472324371,
-0.04880226403474808,
0.05845468491315842,
0.10334239155054092,
0.026590446010231972,
0.0031887979712337255,
0.05871282517910004,
0.034214846789836884,
-0.0902101993560791,
0.1125803142786026,
0.05955227464437485,
0.0165397971868515,
-0.05734850838780403,
-0.038214895874261856,
-0.011950135231018066,
-0.018017224967479706,
-0.02064218558371067,
-0.001972436672076583,
-0.08667390793561935,
-0.007954687811434269,
-0.10027481615543365,
0.022494280710816383,
-0.07858970016241074,
0.008052089251577854,
0.02931516058743,
-0.04746434837579727,
0.0013534465106204152,
0.0018911621300503612,
-0.07022151350975037,
-0.0531441904604435,
-0.011296980082988739,
0.08028347790241241,
-0.13379159569740295,
0.041163623332977295,
0.0720820501446724,
-0.10512391477823257,
0.0722789391875267,
-0.0080735357478261,
0.009603062644600868,
0.002738350536674261,
-0.1515064686536789,
0.057879623025655746,
-0.02471158094704151,
-0.008804970420897007,
0.018293293192982674,
-0.19362454116344452,
-0.006663484498858452,
-0.03933658450841904,
-0.06608349084854126,
0.007965613156557083,
-0.011620837263762951,
-0.1199798509478569,
0.09844714403152466,
0.002606139751151204,
-0.061128001660108566,
-0.02566545084118843,
0.04066308215260506,
0.10586780309677124,
-0.020529067143797874,
0.133222758769989,
-0.02299213595688343,
0.07189196348190308,
-0.1741039901971817,
-0.006793564185500145,
-0.012046009302139282,
0.04731685668230057,
-0.027232524007558823,
-0.03056574985384941,
0.06071736291050911,
-0.021982496604323387,
0.17421258985996246,
-0.012680399231612682,
0.07032647728919983,
0.05198749527335167,
0.00997571088373661,
0.021897418424487114,
0.08211468160152435,
0.06096404418349266,
-0.007212643977254629,
-0.001939830370247364,
0.032961152493953705,
-0.005106652621179819,
-0.04529997706413269,
-0.16330303251743317,
0.05992535874247551,
0.15794123709201813,
0.051688313484191895,
0.02587483450770378,
0.022463064640760422,
-0.11597660183906555,
-0.0829419270157814,
0.11386226117610931,
-0.02665914222598076,
-0.03476277366280556,
-0.0695861279964447,
0.17911197245121002,
0.13906370103359222,
-0.1985146403312683,
0.06898098438978195,
-0.05230652168393135,
-0.04704243689775467,
-0.13746999204158783,
-0.17320163547992706,
-0.05675959959626198,
-0.04932331293821335,
-0.027015792205929756,
-0.06207159534096718,
0.051294248551130295,
0.04572872072458267,
0.0028256417717784643,
-0.01840038411319256,
0.10395382344722748,
0.016976751387119293,
-0.02532275579869747,
0.04699812829494476,
0.06548187136650085,
0.038627032190561295,
-0.09174156934022903,
0.007587284315377474,
-0.002293705940246582,
0.01970020681619644,
0.06891614943742752,
0.01807936653494835,
-0.060284461826086044,
0.024095797911286354,
-0.0205357875674963,
-0.12168102711439133,
0.040349312126636505,
-0.015099799260497093,
-0.04164285212755203,
0.15001407265663147,
0.036165233701467514,
0.008809513412415981,
-0.02104180119931698,
0.22528551518917084,
-0.0845077857375145,
-0.07293173670768738,
-0.14992713928222656,
0.061421964317560196,
-0.06753890216350555,
0.032307662069797516,
0.031194649636745453,
-0.12372498214244843,
0.008965158835053444,
0.16953149437904358,
0.12706954777240753,
-0.017235754057765007,
0.011755096726119518,
0.04631460830569267,
0.0054819295182824135,
-0.039178598672151566,
0.02386985719203949,
0.04763784632086754,
0.17436441779136658,
-0.06912772357463837,
0.060602426528930664,
-0.010066114366054535,
-0.08422619104385376,
-0.016610965132713318,
0.1004536971449852,
-0.006028721109032631,
0.002527001779526472,
-0.0653572604060173,
0.1435474455356598,
-0.07789696753025055,
-0.21248222887516022,
0.062445711344480515,
-0.062453895807266235,
-0.13928431272506714,
-0.04283962398767471,
0.03173775225877762,
-0.02029750682413578,
0.003866987768560648,
0.07300805300474167,
-0.045953720808029175,
0.18954069912433624,
0.03739549592137337,
-0.052604615688323975,
-0.08307439088821411,
0.05310605838894844,
-0.15448454022407532,
0.2798924148082733,
0.020205844193696976,
0.04607383534312248,
0.10831371694803238,
-0.018147770315408707,
-0.14655283093452454,
0.009704295545816422,
0.10396505147218704,
-0.06918948888778687,
0.058933284133672714,
0.16969238221645355,
0.003520707134157419,
0.12444666028022766,
0.05716300383210182,
-0.0549672469496727,
0.03483917936682701,
-0.09883621335029602,
-0.048986852169036865,
-0.11174394935369492,
0.07913976162672043,
-0.08493151515722275,
0.16219210624694824,
0.11736402660608292,
-0.0692162811756134,
0.0009490751544944942,
-0.01712220162153244,
0.08669418841600418,
0.010318588465452194,
0.11391915380954742,
0.013223540037870407,
-0.18779438734054565,
0.03298543021082878,
0.009580777026712894,
0.10570746660232544,
-0.18654847145080566,
-0.05411550775170326,
0.0407252237200737,
-0.019413217902183533,
-0.08267369121313095,
0.1180468425154686,
0.038627687841653824,
0.03224669024348259,
-0.038653165102005005,
-0.03813322260975838,
0.008183742873370647,
0.14366917312145233,
-0.11073201894760132,
-0.012064279057085514
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# Whisper Tiny Hu v6
This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on the Common Voice 16.0 dataset.
It achieves the following results on the evaluation set:
- Loss: 0.1880
- Wer Ortho: 11.4928
- Wer: 10.6137
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 64
- eval_batch_size: 32
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 256
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: constant_with_warmup
- lr_scheduler_warmup_steps: 100
- training_steps: 2000
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer Ortho | Wer |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:-------:|
| 0.0036 | 5.35 | 1000 | 0.1855 | 12.1161 | 11.2185 |
| 0.0004 | 10.71 | 2000 | 0.1880 | 11.4928 | 10.6137 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.0
| {"language": ["hu"], "license": "apache-2.0", "tags": ["hf-asr-leaderboard", "generated_from_trainer"], "datasets": ["mozilla-foundation/common_voice_16_0"], "metrics": ["wer"], "base_model": "openai/whisper-tiny", "widget": [{"example_title": "Sample 1", "src": "https://huggingface.co/datasets/Hungarians/samples/resolve/main/Sample1.flac"}, {"example_title": "Sample 2", "src": "https://huggingface.co/datasets/Hungarians/samples/resolve/main/Sample2.flac"}], "pipeline_tag": "automatic-speech-recognition", "model-index": [{"name": "Whisper Tiny Hungarian", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "Common Voice 16.0 - Hungarian", "type": "mozilla-foundation/common_voice_16_0", "config": "hu", "split": "test", "args": "hu"}, "metrics": [{"type": "wer", "value": 10.6137, "name": "Wer"}]}]}]} | automatic-speech-recognition | sarpba/whisper-tiny-cv16-hu-v6 | [
"transformers",
"tensorboard",
"safetensors",
"whisper",
"automatic-speech-recognition",
"hf-asr-leaderboard",
"generated_from_trainer",
"hu",
"dataset:mozilla-foundation/common_voice_16_0",
"base_model:openai/whisper-tiny",
"license:apache-2.0",
"model-index",
"endpoints_compatible",
"region:us"
] | 2024-02-08T08:26:42+00:00 | [] | [
"hu"
] | TAGS
#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #hf-asr-leaderboard #generated_from_trainer #hu #dataset-mozilla-foundation/common_voice_16_0 #base_model-openai/whisper-tiny #license-apache-2.0 #model-index #endpoints_compatible #region-us
| Whisper Tiny Hu v6
==================
This model is a fine-tuned version of openai/whisper-tiny on the Common Voice 16.0 dataset.
It achieves the following results on the evaluation set:
* Loss: 0.1880
* Wer Ortho: 11.4928
* Wer: 10.6137
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 64
* eval\_batch\_size: 32
* seed: 42
* gradient\_accumulation\_steps: 4
* total\_train\_batch\_size: 256
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: constant\_with\_warmup
* lr\_scheduler\_warmup\_steps: 100
* training\_steps: 2000
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 256\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: constant\\_with\\_warmup\n* lr\\_scheduler\\_warmup\\_steps: 100\n* training\\_steps: 2000\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #hf-asr-leaderboard #generated_from_trainer #hu #dataset-mozilla-foundation/common_voice_16_0 #base_model-openai/whisper-tiny #license-apache-2.0 #model-index #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 256\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: constant\\_with\\_warmup\n* lr\\_scheduler\\_warmup\\_steps: 100\n* training\\_steps: 2000\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.0"
] | [
103,
165,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #hf-asr-leaderboard #generated_from_trainer #hu #dataset-mozilla-foundation/common_voice_16_0 #base_model-openai/whisper-tiny #license-apache-2.0 #model-index #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 256\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: constant\\_with\\_warmup\n* lr\\_scheduler\\_warmup\\_steps: 100\n* training\\_steps: 2000\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.0"
] | [
-0.14243119955062866,
0.14677533507347107,
-0.004138108342885971,
0.04630911350250244,
0.08839862793684006,
0.01278104167431593,
0.10172461718320847,
0.14272424578666687,
-0.05008114129304886,
0.11648859083652496,
0.10124356299638748,
0.057230547070503235,
0.09126835316419601,
0.15348954498767853,
-0.027727574110031128,
-0.272977739572525,
0.025550082325935364,
-0.02578171342611313,
-0.137018084526062,
0.10279074311256409,
0.08875799179077148,
-0.11056794226169586,
0.03512335941195488,
0.016408000141382217,
-0.07130160182714462,
-0.03464224562048912,
-0.056546393781900406,
-0.03682020306587219,
0.09456316381692886,
0.02768874354660511,
0.06117561087012291,
0.05553913488984108,
0.09026221930980682,
-0.2580570578575134,
0.009138387627899647,
0.05942381173372269,
0.03654531389474869,
0.06596384942531586,
0.0969148576259613,
-0.026872945949435234,
0.10079549252986908,
-0.07278366386890411,
0.04021989181637764,
0.06022520363330841,
-0.1045256108045578,
-0.27399519085884094,
-0.07640380412340164,
0.04239221662282944,
0.11397232115268707,
0.047236789017915726,
-0.03696173429489136,
0.062499672174453735,
-0.06102501228451729,
0.08218396455049515,
0.20779310166835785,
-0.24741500616073608,
-0.07749054580926895,
-0.012077871710062027,
0.036416519433259964,
0.0315360426902771,
-0.1200818344950676,
-0.016925819218158722,
0.02094312198460102,
0.007986465469002724,
0.08546756207942963,
0.025944510474801064,
0.000556702318135649,
0.005177158396691084,
-0.12138914316892624,
-0.03432491794228554,
0.11619098484516144,
0.08914878964424133,
-0.023886172100901604,
-0.10857858508825302,
-0.017394233494997025,
-0.18865235149860382,
-0.062253374606370926,
0.027532652020454407,
0.033053699880838394,
-0.0424916073679924,
-0.0746532753109932,
0.037874191999435425,
-0.060893464833498,
-0.08312937617301941,
0.05824873596429825,
0.12269829958677292,
0.03606207296252251,
-0.03147288039326668,
0.022741839289665222,
0.10440313071012497,
0.04203193634748459,
-0.17733994126319885,
-0.03346215561032295,
0.03176354989409447,
-0.1174362525343895,
-0.0252736434340477,
-0.03027510643005371,
0.03151846304535866,
0.07119467109441757,
0.14748622477054596,
-0.02883671037852764,
0.09485355019569397,
0.03629928454756737,
0.0075072962790727615,
-0.0775892361998558,
0.16348226368427277,
-0.07863745838403702,
-0.08821571618318558,
-0.04054835066199303,
0.14425700902938843,
-0.00762028805911541,
-0.012542108073830605,
-0.0468255952000618,
0.041998930275440216,
0.12571796774864197,
0.03631908819079399,
-0.01556059904396534,
0.04292496666312218,
-0.07192890346050262,
-0.004428742919117212,
-0.01620166003704071,
-0.11254757642745972,
0.028684575110673904,
0.06492722034454346,
-0.07708225399255753,
-0.03356487676501274,
-0.010678457096219063,
0.012611690908670425,
-0.001689599477685988,
0.08899058401584625,
-0.06490892171859741,
-0.028492383658885956,
-0.08453866094350815,
-0.07037331163883209,
0.030716579407453537,
-0.042422231286764145,
-0.0038376948796212673,
-0.03435855731368065,
-0.11089339852333069,
-0.06421682238578796,
0.06918731331825256,
-0.06966957449913025,
-0.06552265584468842,
-0.08094801008701324,
-0.10615796595811844,
0.05712622404098511,
-0.008859091438353062,
0.1559782177209854,
-0.06250371783971786,
0.09417196363210678,
0.03625757619738579,
0.059213023632764816,
0.09594444930553436,
0.044649191200733185,
-0.03516188636422157,
0.0757732018828392,
-0.13681857287883759,
0.09803936630487442,
-0.11980951577425003,
0.08210117369890213,
-0.13342808187007904,
-0.09295595437288284,
-0.014945033006370068,
0.0037756378296762705,
0.10346754640340805,
0.14313797652721405,
-0.19659607112407684,
-0.07214123010635376,
0.1871005892753601,
-0.06616300344467163,
-0.1357300728559494,
0.1205623671412468,
-0.009002829901874065,
-0.01307968981564045,
0.02235300838947296,
0.1521737277507782,
0.1097230464220047,
-0.07101098448038101,
-0.0065728360787034035,
-0.0320599228143692,
0.13025100529193878,
0.04210146516561508,
0.0814410001039505,
-0.036204881966114044,
0.03146187588572502,
0.012929807417094707,
-0.035954542458057404,
0.032630614936351776,
-0.0973772257566452,
-0.09426435828208923,
-0.01004711538553238,
-0.09566563367843628,
-0.01566600799560547,
0.05741136893630028,
0.01933949813246727,
-0.08229960501194,
-0.13184010982513428,
-0.01008653361350298,
0.10014883428812027,
-0.09302596747875214,
-0.011453974060714245,
-0.09362102299928665,
0.07204768806695938,
0.008794821798801422,
0.0001810347312130034,
-0.121515192091465,
-0.046223852783441544,
0.027926282957196236,
-0.05723113939166069,
-0.010695712640881538,
-0.03973473981022835,
0.09901951998472214,
0.05710393562912941,
-0.034371841698884964,
-0.06973395496606827,
-0.020667636767029762,
-0.0005486845620907843,
-0.05526154488325119,
-0.24281325936317444,
-0.07543343305587769,
-0.01832668110728264,
0.2283385545015335,
-0.22879533469676971,
0.017929183319211006,
0.0580364428460598,
0.1291993260383606,
0.0257123913615942,
-0.056698136031627655,
0.02466380037367344,
0.03781471401453018,
-0.013370211236178875,
-0.08100055158138275,
0.02615009807050228,
0.011685596778988838,
-0.12777361273765564,
0.016128109768033028,
-0.13812057673931122,
0.10228820890188217,
0.08737988024950027,
0.06271098554134369,
-0.09640822559595108,
-0.0661427304148674,
-0.06870975345373154,
-0.0498642735183239,
0.007240546867251396,
-0.002153271110728383,
0.13104553520679474,
0.03304886445403099,
0.0931701809167862,
-0.08160152286291122,
-0.056839268654584885,
0.029974518343806267,
0.003946295939385891,
0.007296563126146793,
0.1500643938779831,
0.04754627123475075,
-0.05057291314005852,
0.09384206682443619,
0.022460751235485077,
-0.06286096572875977,
0.14118395745754242,
-0.08734297007322311,
-0.08698573708534241,
-0.035209134221076965,
0.059221312403678894,
0.04356648772954941,
0.1387706845998764,
-0.14443936944007874,
0.00010328378266422078,
0.025056233629584312,
0.00044072451419197023,
0.018541622906923294,
-0.1767907738685608,
-0.008001635782420635,
0.037136010825634,
-0.07647234201431274,
-0.025580499321222305,
-0.00810660608112812,
-0.007162631489336491,
0.08229474723339081,
-0.0001341039896942675,
-0.07421761751174927,
-0.04169495031237602,
-0.03575710207223892,
-0.0792546272277832,
0.1965368688106537,
-0.08159339427947998,
-0.10021606087684631,
-0.11807636171579361,
-0.0001625541044631973,
-0.011971679516136646,
-0.012115910649299622,
0.019308174028992653,
-0.08318838477134705,
-0.03476879745721817,
-0.08932164311408997,
-0.0009617132600396872,
0.015299992635846138,
0.029600035399198532,
0.024452755227684975,
0.012889285571873188,
0.09442805498838425,
-0.09631393104791641,
0.022384898737072945,
-0.023459702730178833,
-0.043760646134614944,
0.039202574640512466,
0.02740732580423355,
0.09747256338596344,
0.1397784799337387,
0.049768928438425064,
0.0369420051574707,
-0.013091273605823517,
0.1895257532596588,
-0.11692862212657928,
0.023675573989748955,
0.09186121821403503,
0.0066034384071826935,
0.04547766596078873,
0.16010858118534088,
0.047142356634140015,
-0.09890870004892349,
0.008862207643687725,
0.047532495111227036,
-0.00864645279943943,
-0.20094558596611023,
-0.01776067167520523,
-0.06353839486837387,
-0.02608010731637478,
0.10162394493818283,
0.026750987395644188,
-0.034457944333553314,
0.03979169577360153,
-0.022372672334313393,
-0.04625774547457695,
0.031934674829244614,
0.044408172369003296,
0.02650904282927513,
0.05122533068060875,
0.11862532049417496,
-0.00706331105902791,
-0.03157714009284973,
0.01162389013916254,
0.0009042552555911243,
0.21950672566890717,
-0.034642480313777924,
0.190625861287117,
0.05106949433684349,
0.13291695713996887,
-0.013049288652837276,
0.06991342455148697,
-0.008994406089186668,
-0.02085679955780506,
0.02338862605392933,
-0.06313083320856094,
-0.0316944494843483,
0.0369252972304821,
0.05588056892156601,
0.026709869503974915,
-0.09478992968797684,
0.03300297632813454,
0.043941136449575424,
0.33440327644348145,
0.09486742317676544,
-0.3089078366756439,
-0.08503775298595428,
0.011136320419609547,
-0.04904932156205177,
-0.05163193866610527,
0.012255425564944744,
0.12065152823925018,
-0.07135654240846634,
0.06865589320659637,
-0.057848863303661346,
0.08423502743244171,
-0.06425418704748154,
0.006441895384341478,
0.06994904577732086,
0.10736604779958725,
-0.001469415263272822,
0.06027552857995033,
-0.21135535836219788,
0.28323662281036377,
-0.0065331971272826195,
0.09221751987934113,
-0.03944461792707443,
0.04136209934949875,
0.03222097083926201,
-0.030795177444815636,
0.07119818776845932,
-0.006590749137103558,
-0.08908262103796005,
-0.1746496707201004,
-0.10960152000188828,
0.037644002586603165,
0.12597770988941193,
-0.06323839724063873,
0.13397005200386047,
-0.02518278919160366,
-0.03737211227416992,
0.04803144559264183,
-0.07486819475889206,
-0.09140638262033463,
-0.10870199650526047,
0.03224873170256615,
-0.005687114782631397,
0.07576649636030197,
-0.10348790884017944,
-0.09188171476125717,
-0.07967658340930939,
0.14732065796852112,
-0.10350954532623291,
-0.03847706690430641,
-0.13528519868850708,
0.053504880517721176,
0.18248696625232697,
-0.06087314337491989,
0.027738790959119797,
0.002276927698403597,
0.13451328873634338,
0.038260262459516525,
0.005140556953847408,
0.10134247690439224,
-0.0845322385430336,
-0.21986746788024902,
-0.04995512589812279,
0.16678471863269806,
0.0392109714448452,
0.05414729192852974,
-0.020234541967511177,
0.028585098683834076,
-0.01669318974018097,
-0.08833418786525726,
0.06771460920572281,
0.012364585883915424,
-0.0037217256613075733,
0.01874253712594509,
-0.03875811770558357,
0.03561326861381531,
-0.057189054787158966,
-0.03516962751746178,
0.08576603978872299,
0.3027970790863037,
-0.07388251274824142,
0.03273683786392212,
0.0448392815887928,
-0.03514765575528145,
-0.16139280796051025,
-0.0015444033779203892,
0.1355006992816925,
0.041845593601465225,
-0.038784559816122055,
-0.21061992645263672,
0.057417090982198715,
0.07467024773359299,
-0.026526836678385735,
0.10803399235010147,
-0.30249515175819397,
-0.13636861741542816,
0.10108187049627304,
0.07336997240781784,
-0.013964290730655193,
-0.16124774515628815,
-0.06976965069770813,
-0.01017523929476738,
-0.09803283959627151,
0.042055681347846985,
-0.022326800972223282,
0.11129560321569443,
-0.002390566747635603,
0.02069912478327751,
0.02691536955535412,
-0.06567401438951492,
0.15598765015602112,
-0.006943207234144211,
0.056308895349502563,
-0.011092034168541431,
0.059826355427503586,
0.01094332616776228,
-0.07093971222639084,
0.012521300464868546,
-0.12743845582008362,
0.03037245199084282,
-0.12256719172000885,
-0.02613501250743866,
-0.08762499690055847,
0.032620739191770554,
-0.04799316078424454,
-0.018209978938102722,
0.00566110759973526,
0.060904648154973984,
0.08239485323429108,
0.012114602141082287,
0.06228246912360191,
-0.06671717017889023,
0.1436275988817215,
0.1221325546503067,
0.11398342996835709,
-0.039080291986465454,
-0.05500391498208046,
0.011106656864285469,
0.00965858343988657,
0.0343056246638298,
-0.09201440960168839,
0.0383606031537056,
0.1443883776664734,
0.04480605944991112,
0.13074861466884613,
0.05855632945895195,
-0.08145365118980408,
-0.02020266093313694,
0.0820155069231987,
-0.07770629227161407,
-0.15536747872829437,
-0.007080255076289177,
0.052664484828710556,
-0.1649787575006485,
0.010906017385423183,
0.0903831422328949,
-0.0344722643494606,
0.0034476048313081264,
0.004240283742547035,
0.05376608297228813,
-0.01385167520493269,
0.22450792789459229,
0.02950645051896572,
0.10986261069774628,
-0.09996030479669571,
0.06424151360988617,
0.02808423340320587,
-0.07750052213668823,
0.037380944937467575,
0.07982391864061356,
-0.058144181966781616,
-0.016165990382432938,
0.05783737078309059,
0.09028933197259903,
0.049222636967897415,
-0.04818558320403099,
-0.1266298145055771,
-0.14787113666534424,
0.07231588661670685,
0.12138731777667999,
0.02726699598133564,
0.04323809966444969,
0.007852552458643913,
0.01869247294962406,
-0.08620497584342957,
0.11593252420425415,
0.09771889448165894,
0.07313250750303268,
-0.10118886083364487,
0.1141442134976387,
-0.017710763961076736,
-0.030046118423342705,
0.004520625341683626,
0.01086603943258524,
-0.11872057616710663,
0.001895099994726479,
-0.1144508495926857,
-0.017801158130168915,
-0.05781983211636543,
-0.001397922053001821,
0.018627077341079712,
-0.05168436840176582,
-0.058777034282684326,
0.006300312466919422,
-0.11727652698755264,
-0.04987000674009323,
-0.02314756065607071,
0.06909637153148651,
-0.12017951160669327,
-0.008108686655759811,
0.05002174153923988,
-0.14165745675563812,
0.09961359202861786,
0.034380100667476654,
0.017919518053531647,
0.012350756675004959,
-0.10915081202983856,
-0.0016179573722183704,
0.0034669560845941305,
-0.0007366062491200864,
0.03516564890742302,
-0.17520521581172943,
-0.0008565241587348282,
-0.03290553018450737,
0.0018122941255569458,
-0.010240777395665646,
0.02946420945227146,
-0.1178964227437973,
-0.005987456068396568,
-0.028174981474876404,
-0.025692220777273178,
-0.052372269332408905,
0.048842065036296844,
0.06566482037305832,
0.013735875487327576,
0.15323789417743683,
-0.07322123646736145,
0.031683299690485,
-0.23339638113975525,
0.0035442246589809656,
-0.004834332969039679,
-0.063534215092659,
-0.0682792067527771,
-0.011455361731350422,
0.10876061022281647,
-0.06455393135547638,
0.09228811413049698,
-0.06510309875011444,
0.01371926162391901,
0.019556304439902306,
-0.10456070303916931,
0.03546690568327904,
0.06401855498552322,
0.18751618266105652,
0.04901789501309395,
-0.01593591272830963,
0.0685926154255867,
-0.010825896635651588,
0.03124818578362465,
0.06566872447729111,
0.1657360941171646,
0.14620152115821838,
0.023680947721004486,
0.07989080995321274,
0.08612070232629776,
-0.139185830950737,
-0.09384439885616302,
0.17768315970897675,
-0.08733955770730972,
0.13232505321502686,
-0.027857109904289246,
0.2073417454957962,
0.11743910610675812,
-0.18971525132656097,
0.05428708717226982,
-0.05714058503508568,
-0.08897305279970169,
-0.09153832495212555,
-0.07612662762403488,
-0.08994060009717941,
-0.15715017914772034,
0.027607833966612816,
-0.10881887376308441,
0.03067765198647976,
0.04917897284030914,
0.04633837193250656,
0.043743159621953964,
0.12372896075248718,
0.06651753187179565,
0.027149153873324394,
0.09869404882192612,
0.020884757861495018,
-0.025686830282211304,
-0.01863262802362442,
-0.12046821415424347,
0.04792281612753868,
-0.009997840039432049,
0.05595597252249718,
-0.04317633435130119,
-0.08518891781568527,
0.05731315538287163,
0.026886790990829468,
-0.110914446413517,
0.028933009132742882,
-0.02910648100078106,
0.05052808299660683,
0.04174446687102318,
0.02993181347846985,
0.0022240146063268185,
-0.02520333230495453,
0.22160466015338898,
-0.09115325659513474,
-0.07677311450242996,
-0.14983052015304565,
0.22557875514030457,
-0.00982489250600338,
-0.0022427719086408615,
0.020018398761749268,
-0.07500683516263962,
-0.0018521088641136885,
0.1534961313009262,
0.1446126252412796,
-0.05717942863702774,
-0.024931015446782112,
-0.0019664200954139233,
-0.012339228764176369,
-0.04054148122668266,
0.0734802782535553,
0.10437379777431488,
0.04276932775974274,
-0.051767732948064804,
-0.02739623375236988,
-0.02050110138952732,
-0.06936196982860565,
-0.02235378697514534,
0.10454369336366653,
0.02344231680035591,
0.004816438537091017,
-0.03355049341917038,
0.1098841056227684,
-0.08514050394296646,
-0.14731858670711517,
0.03011978417634964,
-0.17306548357009888,
-0.1977406144142151,
-0.055167149752378464,
0.05360228568315506,
0.027678020298480988,
0.054895929992198944,
0.0028048320673406124,
-0.03347775712609291,
0.07775657624006271,
0.0016731454525142908,
-0.013340412639081478,
-0.09346159547567368,
0.06050192937254906,
-0.10235296189785004,
0.20950168371200562,
-0.03588886931538582,
0.001830519177019596,
0.1328027844429016,
0.028680918738245964,
-0.09580608457326889,
0.035328492522239685,
0.0889936164021492,
-0.15039896965026855,
0.04842972010374069,
0.17765185236930847,
-0.03591141849756241,
0.11968402564525604,
0.04508071765303612,
-0.09382152557373047,
0.0077223824337124825,
-0.07696934789419174,
-0.0435873381793499,
-0.06808783113956451,
-0.003976591397076845,
-0.029567038640379906,
0.13572360575199127,
0.2247610241174698,
-0.07740779966115952,
-0.005421951413154602,
-0.053013306111097336,
0.027087880298495293,
0.023756401613354683,
0.0908539667725563,
-0.05058443546295166,
-0.2790159583091736,
0.021350689232349396,
0.002257472835481167,
0.016044102609157562,
-0.2005104273557663,
-0.06563837081193924,
0.030172282829880714,
-0.0424177423119545,
-0.04436466842889786,
0.12145119160413742,
0.09934737533330917,
0.05905985087156296,
-0.058494240045547485,
-0.04445585235953331,
-0.03614041581749916,
0.17592498660087585,
-0.1677306741476059,
-0.06294208019971848
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# SciBERT_TwoWayLoss_25K_bs64_P10_N5
This model is a fine-tuned version of [allenai/scibert_scivocab_uncased](https://huggingface.co/allenai/scibert_scivocab_uncased) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 15.1250
- Accuracy: 0.7066
- Precision: 0.0321
- Recall: 0.9982
- F1: 0.0622
- Hamming: 0.2934
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- training_steps: 25000
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 | Hamming |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|:---------:|:------:|:------:|:-------:|
| 28.5732 | 0.16 | 5000 | 26.4288 | 0.6945 | 0.0307 | 0.9910 | 0.0595 | 0.3055 |
| 19.8755 | 0.32 | 10000 | 18.9620 | 0.7010 | 0.0315 | 0.9959 | 0.0610 | 0.2990 |
| 17.1294 | 0.47 | 15000 | 16.5587 | 0.7021 | 0.0316 | 0.9970 | 0.0613 | 0.2979 |
| 15.8209 | 0.63 | 20000 | 15.4919 | 0.7053 | 0.0320 | 0.9982 | 0.0620 | 0.2947 |
| 15.4304 | 0.79 | 25000 | 15.1250 | 0.7066 | 0.0321 | 0.9982 | 0.0622 | 0.2934 |
### Framework versions
- Transformers 4.33.3
- Pytorch 2.2.0.dev20231002
- Datasets 2.7.1
- Tokenizers 0.13.3
| {"tags": ["generated_from_trainer"], "metrics": ["accuracy", "precision", "recall", "f1"], "base_model": "allenai/scibert_scivocab_uncased", "model-index": [{"name": "SciBERT_TwoWayLoss_25K_bs64_P10_N5", "results": []}]} | text-classification | bdpc/SciBERT_TwoWayLoss_25K_bs64_P10_N5 | [
"transformers",
"pytorch",
"bert",
"text-classification",
"generated_from_trainer",
"base_model:allenai/scibert_scivocab_uncased",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T08:26:53+00:00 | [] | [] | TAGS
#transformers #pytorch #bert #text-classification #generated_from_trainer #base_model-allenai/scibert_scivocab_uncased #autotrain_compatible #endpoints_compatible #region-us
| SciBERT\_TwoWayLoss\_25K\_bs64\_P10\_N5
=======================================
This model is a fine-tuned version of allenai/scibert\_scivocab\_uncased on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 15.1250
* Accuracy: 0.7066
* Precision: 0.0321
* Recall: 0.9982
* F1: 0.0622
* Hamming: 0.2934
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 64
* eval\_batch\_size: 64
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* training\_steps: 25000
### Training results
### Framework versions
* Transformers 4.33.3
* Pytorch 2.2.0.dev20231002
* Datasets 2.7.1
* Tokenizers 0.13.3
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 25000",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.33.3\n* Pytorch 2.2.0.dev20231002\n* Datasets 2.7.1\n* Tokenizers 0.13.3"
] | [
"TAGS\n#transformers #pytorch #bert #text-classification #generated_from_trainer #base_model-allenai/scibert_scivocab_uncased #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 25000",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.33.3\n* Pytorch 2.2.0.dev20231002\n* Datasets 2.7.1\n* Tokenizers 0.13.3"
] | [
61,
116,
4,
35
] | [
"passage: TAGS\n#transformers #pytorch #bert #text-classification #generated_from_trainer #base_model-allenai/scibert_scivocab_uncased #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 25000### Training results### Framework versions\n\n\n* Transformers 4.33.3\n* Pytorch 2.2.0.dev20231002\n* Datasets 2.7.1\n* Tokenizers 0.13.3"
] | [
-0.11322607100009918,
0.060777418315410614,
-0.0026239193975925446,
0.12863731384277344,
0.16494037210941315,
0.034882139414548874,
0.12129004299640656,
0.1317269653081894,
-0.06993754208087921,
0.027023080736398697,
0.1219538226723671,
0.15389613807201385,
0.019535236060619354,
0.1288287192583084,
-0.05040936917066574,
-0.2954252064228058,
-0.008254884742200375,
0.020648682489991188,
-0.07212471216917038,
0.1365307867527008,
0.08932062238454819,
-0.13636687397956848,
0.09187042713165283,
-0.01680370420217514,
-0.15832801163196564,
0.005222628824412823,
0.012866973876953125,
-0.04602166265249252,
0.15064255893230438,
0.013846359215676785,
0.1189182698726654,
0.021367304027080536,
0.11824069917201996,
-0.1792900562286377,
0.008058745414018631,
0.041857752948999405,
0.023648666217923164,
0.0888923779129982,
0.05999043583869934,
-0.00555133493617177,
0.12414117157459259,
-0.08065593987703323,
0.06060557812452316,
0.020101040601730347,
-0.13494589924812317,
-0.19134284555912018,
-0.07120350003242493,
0.03545016795396805,
0.0741691142320633,
0.08992337435483932,
-0.005595020949840546,
0.10724600404500961,
-0.08764242380857468,
0.11453878879547119,
0.2332475632429123,
-0.2594282031059265,
-0.08017909526824951,
0.013292266987264156,
0.008211282081902027,
0.0905754342675209,
-0.10896365344524384,
-0.011909586377441883,
0.03143075481057167,
0.0389772467315197,
0.12452062219381332,
-0.03229856863617897,
-0.12019116431474686,
0.026946507394313812,
-0.13748127222061157,
-0.028064263984560966,
0.10466268658638,
0.023422803729772568,
-0.041496630758047104,
-0.034736815840005875,
-0.06828556954860687,
-0.16236047446727753,
-0.04020315036177635,
-0.015795815736055374,
0.03500084578990936,
-0.04661891236901283,
-0.07649854570627213,
-0.029312092810869217,
-0.08775650709867477,
-0.100187748670578,
-0.03286212682723999,
0.17967993021011353,
0.038783181458711624,
0.01577591709792614,
-0.008044502697885036,
0.12906219065189362,
0.018989160656929016,
-0.15190033614635468,
0.010758263990283012,
0.010509027168154716,
-0.043031610548496246,
-0.04464249312877655,
-0.05548499897122383,
-0.017599666491150856,
-0.02184409648180008,
0.1312403529882431,
-0.03323938325047493,
0.05366754159331322,
0.041950833052396774,
0.018694913014769554,
-0.10090630501508713,
0.19318729639053345,
-0.0502639040350914,
-0.023743368685245514,
-0.02012118697166443,
0.08920057117938995,
0.002730759559199214,
-0.027036570012569427,
-0.1046731024980545,
-0.011172428727149963,
0.09965861588716507,
0.02381865680217743,
-0.08251074701547623,
0.06916748732328415,
-0.04818912222981453,
-0.03853327035903931,
0.0038917900528758764,
-0.09896790981292725,
0.034227486699819565,
0.016581598669290543,
-0.10405836254358292,
-0.02213030308485031,
0.02316940203309059,
0.009434331208467484,
-0.014818849973380566,
0.12858407199382782,
-0.09141708165407181,
0.038857992738485336,
-0.09326663613319397,
-0.11637327820062637,
-0.008447324857115746,
-0.09047342836856842,
0.007662117946892977,
-0.07835703343153,
-0.19567836821079254,
-0.025875968858599663,
0.04875972121953964,
-0.05383988842368126,
-0.03479090705513954,
-0.061677396297454834,
-0.05904849246144295,
0.023653248324990273,
-0.00849335640668869,
0.1314145177602768,
-0.05521006137132645,
0.1178673729300499,
0.045114897191524506,
0.07462507486343384,
-0.01810470223426819,
0.05661189556121826,
-0.10872122645378113,
0.009518253616988659,
-0.18335303664207458,
0.06206457316875458,
-0.05587933212518692,
0.04428922012448311,
-0.08276736736297607,
-0.10938357561826706,
0.01876484975218773,
0.003043351462110877,
0.08788173645734787,
0.13856202363967896,
-0.1759018748998642,
-0.07562753558158875,
0.14319124817848206,
-0.069814532995224,
-0.09670642763376236,
0.11442127823829651,
-0.06456783413887024,
0.04590245336294174,
0.06256859749555588,
0.15931060910224915,
0.07717397809028625,
-0.06697686016559601,
0.028535889461636543,
-0.005596464034169912,
0.08086281269788742,
-0.0012793760979548097,
0.07282522320747375,
0.013714738190174103,
-0.004456198774278164,
0.025151154026389122,
-0.06514649838209152,
0.03856390342116356,
-0.10957116633653641,
-0.09793584048748016,
-0.02958599291741848,
-0.10500084608793259,
0.08673115074634552,
0.07492080330848694,
0.07892110198736191,
-0.09251224249601364,
-0.08564925193786621,
0.054892346262931824,
0.09751357138156891,
-0.06162001192569733,
0.019611041992902756,
-0.06151682138442993,
0.04471088573336601,
-0.013761008158326149,
-0.023802893236279488,
-0.18823201954364777,
-0.03574512526392937,
0.017593922093510628,
0.04552995041012764,
0.021116726100444794,
0.0031089973635971546,
0.0917237251996994,
0.08538768440485,
-0.06958133727312088,
-0.03744582459330559,
-0.04535222426056862,
-0.007747768424451351,
-0.13555070757865906,
-0.2003425657749176,
-0.06675176322460175,
-0.02088693156838417,
0.12856626510620117,
-0.21431955695152283,
0.035600461065769196,
-0.016811754554510117,
0.0836050808429718,
0.021805817261338234,
-0.024481460452079773,
-0.03502786532044411,
0.07205364853143692,
-0.03312648460268974,
-0.05410710349678993,
0.06996027380228043,
-0.014238432049751282,
-0.08394847810268402,
-0.05206623300909996,
-0.11408158391714096,
0.14756686985492706,
0.10202006250619888,
-0.10793370008468628,
-0.09324514120817184,
0.0007269465713761747,
-0.05460747331380844,
-0.0323854424059391,
-0.04804670810699463,
0.026561986654996872,
0.16634872555732727,
0.0038699847646057606,
0.14905424416065216,
-0.060196273028850555,
-0.04074978083372116,
0.015804186463356018,
-0.017987245693802834,
0.037342771887779236,
0.13722078502178192,
0.09458201378583908,
-0.09243571758270264,
0.12698480486869812,
0.12046117335557938,
-0.0861004963517189,
0.1520955115556717,
-0.022361459210515022,
-0.07943549007177353,
-0.01694793440401554,
-0.025549590587615967,
-0.0011909212917089462,
0.103069968521595,
-0.11971636116504669,
-0.01557076070457697,
0.012601003050804138,
0.01231086440384388,
-0.0015634495066478848,
-0.20715603232383728,
-0.04048435762524605,
0.043850045651197433,
-0.0508701354265213,
-0.026047809049487114,
-0.009740485809743404,
0.007031307090073824,
0.11952681839466095,
0.009918609634041786,
-0.0871233269572258,
0.01646457426249981,
-0.003474967088550329,
-0.06978889554738998,
0.21154293417930603,
-0.06995129585266113,
-0.12596924602985382,
-0.1165476143360138,
-0.060161907225847244,
-0.05534742400050163,
0.02832852676510811,
0.03613685443997383,
-0.09506747126579285,
-0.00798797607421875,
-0.060322701930999756,
0.029195012524724007,
-0.003424200462177396,
0.03930642455816269,
-0.01490772794932127,
-0.016507508233189583,
0.06259230524301529,
-0.09047063440084457,
-0.008162518963217735,
-0.0595448799431324,
-0.07876808941364288,
0.03639571741223335,
0.03120119497179985,
0.11328773200511932,
0.1435122936964035,
-0.019797615706920624,
0.016932815313339233,
-0.02346927672624588,
0.2299499362707138,
-0.06935667991638184,
-0.008976425044238567,
0.14298376441001892,
-0.023075329139828682,
0.048242971301078796,
0.12353657931089401,
0.06769799441099167,
-0.08370150625705719,
0.011910113506019115,
0.04629865288734436,
-0.03530509024858475,
-0.21056555211544037,
-0.034967102110385895,
-0.050691526383161545,
0.008335963822901249,
0.10279776155948639,
0.015001166611909866,
0.025609470903873444,
0.07128633558750153,
0.019670791923999786,
0.03835856914520264,
-0.02970518171787262,
0.06067961826920509,
0.0901007205247879,
0.0431673638522625,
0.1372215747833252,
-0.02518496848642826,
-0.0694611445069313,
0.04033368080854416,
-0.020800715312361717,
0.1948581337928772,
-0.019927211105823517,
0.10732104629278183,
0.024585334584116936,
0.15417076647281647,
-0.014753337018191814,
0.08637111634016037,
0.007886157371103764,
-0.04216023534536362,
-0.01257680170238018,
-0.0436798632144928,
-0.059678055346012115,
0.02775055728852749,
-0.059515926986932755,
0.06577367335557938,
-0.15086594223976135,
0.011933721601963043,
0.05094826966524124,
0.27103206515312195,
0.05955052748322487,
-0.32154691219329834,
-0.10834003239870071,
0.010266409255564213,
-0.037466540932655334,
-0.02870401181280613,
0.011249309405684471,
0.12272464483976364,
-0.10746078938245773,
0.02955225668847561,
-0.07184460759162903,
0.08864541351795197,
-0.057241667062044144,
0.05337073281407356,
0.07820457220077515,
0.08838637173175812,
-0.006902041379362345,
0.07470817863941193,
-0.27974164485931396,
0.28423112630844116,
0.010132982395589352,
0.06270936131477356,
-0.07220378518104553,
-0.008340820670127869,
0.04852967709302902,
0.08022208511829376,
0.06920047849416733,
-0.014217490330338478,
-0.04305253177881241,
-0.23541653156280518,
-0.05979810655117035,
0.018604930490255356,
0.11607114225625992,
-0.07001346349716187,
0.11053061485290527,
-0.04097829759120941,
0.0002958417753688991,
0.05691635236144066,
-0.0389217734336853,
-0.055206455290317535,
-0.07953943312168121,
0.0037517710588872433,
-0.0009975082939490676,
-0.014158784411847591,
-0.05482397973537445,
-0.12082913517951965,
-0.06597607582807541,
0.15569958090782166,
-0.026311486959457397,
-0.03751857206225395,
-0.1341594010591507,
0.08815846592187881,
0.09949356317520142,
-0.09004699438810349,
0.02234192006289959,
0.01594715751707554,
0.07048142701387405,
0.038474712520837784,
-0.05912115052342415,
0.11224541068077087,
-0.07337679713964462,
-0.19993382692337036,
-0.0601157546043396,
0.11236320436000824,
0.05113566666841507,
0.0716264545917511,
-0.02059956081211567,
0.01991426758468151,
-0.023640226572752,
-0.08404037356376648,
0.03629358112812042,
-0.018087320029735565,
0.0543324276804924,
0.04123779013752937,
-0.06430694460868835,
0.053876183927059174,
-0.06506627798080444,
-0.008732454851269722,
0.16010209918022156,
0.2523648738861084,
-0.09806618094444275,
0.030738646164536476,
0.033863365650177,
-0.0664454847574234,
-0.19941622018814087,
0.03526909276843071,
0.07292694598436356,
0.01901070959866047,
0.05625315010547638,
-0.20593160390853882,
0.09140648692846298,
0.09049583226442337,
-0.012497663497924805,
0.07558238506317139,
-0.310346782207489,
-0.12393680959939957,
0.11138509958982468,
0.1300848126411438,
0.1174929216504097,
-0.14140585064888,
-0.01922472007572651,
-0.007784165441989899,
-0.08759327232837677,
0.08772243559360504,
-0.06579653173685074,
0.12727227807044983,
-0.03275291994214058,
0.08580881357192993,
0.033120717853307724,
-0.04518772289156914,
0.1192910447716713,
0.009985420852899551,
0.09586546570062637,
-0.05176251009106636,
-0.04436599835753441,
0.024653473868966103,
-0.04766470938920975,
0.011246348731219769,
-0.08042080700397491,
0.04455604404211044,
-0.11781403422355652,
-0.022008780390024185,
-0.09382589161396027,
0.024981683120131493,
-0.03826994076371193,
-0.07163041085004807,
-0.01095743477344513,
0.04221881553530693,
0.0438716858625412,
-0.003962178714573383,
0.1222144216299057,
-0.016202056780457497,
0.15396028757095337,
0.11465810239315033,
0.09542932361364365,
-0.015033443458378315,
-0.030365345999598503,
-0.006263168528676033,
-0.002981423167511821,
0.055116597563028336,
-0.1183537095785141,
0.033677343279123306,
0.1461094617843628,
0.03357096016407013,
0.12753614783287048,
0.08549758791923523,
-0.006084870081394911,
0.0044884770177304745,
0.05544697865843773,
-0.17808513343334198,
-0.05874479189515114,
-0.02114073559641838,
-0.061082493513822556,
-0.13142633438110352,
0.05058847367763519,
0.10777252167463303,
-0.060270681977272034,
-0.014111949130892754,
-0.01303863525390625,
0.013012341223657131,
-0.042949553579092026,
0.2220647931098938,
0.05663270130753517,
0.06122646853327751,
-0.10052894800901413,
0.06530968099832535,
0.0401872843503952,
-0.05076230317354202,
0.0058869291096925735,
0.10765733569860458,
-0.08396819978952408,
-0.037401504814624786,
0.06080351024866104,
0.14739389717578888,
-0.07089882344007492,
-0.017031153663992882,
-0.1576898843050003,
-0.11114457994699478,
0.06962037831544876,
0.16495351493358612,
0.10285767167806625,
0.013337933458387852,
-0.07101059705018997,
0.026593228802084923,
-0.11536216735839844,
0.09768413752317429,
0.061899639666080475,
0.07126546651124954,
-0.1460626721382141,
0.1830959916114807,
-0.020120395347476006,
0.05162522941827774,
-0.028596648946404457,
0.01752403751015663,
-0.11278122663497925,
0.016018057242035866,
-0.12450063228607178,
-0.04697561264038086,
-0.008143682964146137,
0.00040639942744746804,
-0.01271379180252552,
-0.07477424293756485,
-0.04768175259232521,
0.00399434007704258,
-0.11826067417860031,
-0.024190546944737434,
0.0232811588793993,
0.036590952426195145,
-0.11958090960979462,
-0.04284306988120079,
0.03558968007564545,
-0.07975448668003082,
0.07793085277080536,
0.06264517456293106,
0.019132662564516068,
0.05159357935190201,
-0.12546886503696442,
0.011315478943288326,
0.03279691934585571,
0.004662573337554932,
0.0581335723400116,
-0.06980524957180023,
-0.004519004374742508,
-0.03386951982975006,
0.06470746546983719,
0.020947132259607315,
0.10305336117744446,
-0.12205538898706436,
0.031088564544916153,
-0.00942651741206646,
-0.08270712196826935,
-0.06419926136732101,
0.05769544094800949,
0.07322192192077637,
0.02902405336499214,
0.16693571209907532,
-0.08954121917486191,
0.054934460669755936,
-0.22038534283638,
0.000462504627648741,
-0.006070960778743029,
-0.11815790832042694,
-0.13026545941829681,
-0.08069711923599243,
0.08441128581762314,
-0.05040693283081055,
0.09182167798280716,
0.019842715933918953,
0.09124478697776794,
0.02529192343354225,
-0.03740745037794113,
0.04545897617936134,
0.027339642867445946,
0.18770326673984528,
0.04359619691967964,
-0.04953617975115776,
0.07975684851408005,
0.05436698719859123,
0.08791255205869675,
0.14704445004463196,
0.21894457936286926,
0.13401897251605988,
-0.01479929219931364,
0.08611303567886353,
0.042673710733652115,
-0.068903349339962,
-0.17377261817455292,
-0.0035070248413830996,
-0.03677574545145035,
0.08222942799329758,
-0.04507683962583542,
0.21632568538188934,
0.054710496217012405,
-0.17000488936901093,
0.04853948950767517,
-0.04924093559384346,
-0.1008920669555664,
-0.12498663365840912,
-0.02550434321165085,
-0.07603474706411362,
-0.14206084609031677,
-0.00008562617585994303,
-0.11107003688812256,
0.03972528129816055,
0.10213038325309753,
0.010073383338749409,
-0.013304433785378933,
0.1656186878681183,
0.02041034772992134,
0.03724338486790657,
0.06758622825145721,
0.009690772742033005,
-0.006895836442708969,
-0.0789174884557724,
-0.08696281164884567,
-0.012760933488607407,
-0.016731731593608856,
0.025197722017765045,
-0.07745826989412308,
-0.06846438348293304,
0.022722812369465828,
-0.006325817666947842,
-0.10981941968202591,
0.014879039488732815,
0.012316560372710228,
0.07395362854003906,
0.056870412081480026,
0.007348461076617241,
0.00760895200073719,
-0.0234462171792984,
0.25182974338531494,
-0.09724720567464828,
-0.04704252630472183,
-0.10327094048261642,
0.29568055272102356,
0.036001238971948624,
0.006410595029592514,
0.02602032572031021,
-0.069056436419487,
-0.00943495612591505,
0.22953197360038757,
0.1826784461736679,
-0.11231271177530289,
-0.006816741079092026,
0.005394464358687401,
-0.005542530212551355,
-0.004393824841827154,
0.10261720418930054,
0.10069281607866287,
0.03940023109316826,
-0.0999714732170105,
-0.029353978112339973,
-0.02733774110674858,
-0.032340195029973984,
-0.03412627428770065,
0.08298652619123459,
0.03910793364048004,
0.018813783302903175,
-0.0580991730093956,
0.048812367022037506,
-0.06800634413957596,
-0.10046184062957764,
0.06464620679616928,
-0.21189992129802704,
-0.17116159200668335,
-0.02743125893175602,
0.0640559121966362,
0.021503208205103874,
0.07696720212697983,
-0.024388860911130905,
-0.008803535252809525,
0.07173354923725128,
-0.010118621401488781,
-0.0671629086136818,
-0.1097874715924263,
0.1146639883518219,
-0.10249804705381393,
0.20237480103969574,
-0.05075293406844139,
0.04824726656079292,
0.11651717126369476,
0.043045882135629654,
-0.07014910131692886,
0.04409252107143402,
0.05867602676153183,
-0.09748433530330658,
0.022207433357834816,
0.1396719068288803,
-0.04462052136659622,
0.08248202502727509,
0.036560509353876114,
-0.15821373462677002,
0.015652917325496674,
-0.06613995879888535,
-0.07067205011844635,
-0.02833997644484043,
-0.04141415283083916,
-0.03831064701080322,
0.13927972316741943,
0.2508515417575836,
-0.0312089491635561,
0.010292841121554375,
-0.07527371495962143,
0.01374831423163414,
0.05917297303676605,
0.05077211186289787,
-0.06651241332292557,
-0.2361927330493927,
0.001194559969007969,
0.06720447540283203,
-0.015717636793851852,
-0.23350706696510315,
-0.09611000120639801,
0.010241828858852386,
-0.05309658870100975,
-0.08297353237867355,
0.0917973741889,
0.060633979737758636,
0.043233472853899,
-0.04176126793026924,
-0.07289251685142517,
-0.06847694516181946,
0.16616292297840118,
-0.1596599966287613,
-0.08117108792066574
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# results
This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0004
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 100
- training_steps: 100
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1 | {"license": "apache-2.0", "library_name": "peft", "tags": ["trl", "sft", "generated_from_trainer"], "base_model": "mistralai/Mistral-7B-v0.1", "model-index": [{"name": "results", "results": []}]} | null | mahiatlinux/mistral_custom | [
"peft",
"tensorboard",
"safetensors",
"trl",
"sft",
"generated_from_trainer",
"base_model:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"region:us"
] | 2024-02-08T08:27:08+00:00 | [] | [] | TAGS
#peft #tensorboard #safetensors #trl #sft #generated_from_trainer #base_model-mistralai/Mistral-7B-v0.1 #license-apache-2.0 #region-us
|
# results
This model is a fine-tuned version of mistralai/Mistral-7B-v0.1 on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0004
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 100
- training_steps: 100
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1 | [
"# results\n\nThis model is a fine-tuned version of mistralai/Mistral-7B-v0.1 on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0004\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 16\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 100\n- training_steps: 100",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#peft #tensorboard #safetensors #trl #sft #generated_from_trainer #base_model-mistralai/Mistral-7B-v0.1 #license-apache-2.0 #region-us \n",
"# results\n\nThis model is a fine-tuned version of mistralai/Mistral-7B-v0.1 on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0004\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 16\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 100\n- training_steps: 100",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
55,
31,
6,
12,
8,
3,
126,
4,
39
] | [
"passage: TAGS\n#peft #tensorboard #safetensors #trl #sft #generated_from_trainer #base_model-mistralai/Mistral-7B-v0.1 #license-apache-2.0 #region-us \n# results\n\nThis model is a fine-tuned version of mistralai/Mistral-7B-v0.1 on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0004\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 16\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 100\n- training_steps: 100### Training results### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.12153369188308716,
0.1396321952342987,
-0.003012345638126135,
0.08958746492862701,
0.13440477848052979,
0.03158366680145264,
0.09080924093723297,
0.14380402863025665,
-0.0017380182398483157,
0.10009001940488815,
0.09431161731481552,
0.02520378679037094,
0.06190046668052673,
0.14543327689170837,
-0.012632617726922035,
-0.2338891476392746,
0.008532081730663776,
-0.07490881532430649,
-0.062450870871543884,
0.09877856820821762,
0.10114763677120209,
-0.08618303388357162,
0.06680269539356232,
-0.015391707420349121,
-0.09909109771251678,
-0.011549761518836021,
-0.040031079202890396,
-0.04291630908846855,
0.10016318410634995,
-0.004782082047313452,
0.08486933261156082,
0.0023931243922561407,
0.13882721960544586,
-0.22115731239318848,
0.0003695101186167449,
0.06848983466625214,
0.04881998151540756,
0.09337960183620453,
0.0715380311012268,
-0.010369517840445042,
0.041575200855731964,
-0.14197304844856262,
0.09203334897756577,
0.027021074667572975,
-0.09222845733165741,
-0.13687938451766968,
-0.11300069093704224,
0.0759778544306755,
0.10277117788791656,
0.08416976034641266,
0.008291813544929028,
0.08607238531112671,
-0.08641985058784485,
0.05256224423646927,
0.2147638350725174,
-0.2526836395263672,
-0.05748845264315605,
0.05915059521794319,
0.04659619182348251,
0.0896710753440857,
-0.1178930327296257,
-0.022278232499957085,
0.038648441433906555,
0.02018960937857628,
0.08391706645488739,
0.01481724251061678,
-0.03338978812098503,
0.02304411679506302,
-0.12305077165365219,
-0.037875112146139145,
0.13970854878425598,
0.03968651220202446,
-0.04684550687670708,
-0.12996193766593933,
-0.02800651825964451,
-0.12939667701721191,
-0.013432957231998444,
0.005007794592529535,
0.01926899328827858,
-0.035717230290174484,
-0.02355899289250374,
-0.03454531729221344,
-0.061844658106565475,
-0.07960189133882523,
0.03190121054649353,
0.05948999151587486,
0.03037170134484768,
0.017282115295529366,
0.00019527909171301872,
0.12513382732868195,
-0.0005170022486709058,
-0.12716057896614075,
-0.015628524124622345,
-0.019833268597722054,
-0.11732549965381622,
-0.04559716209769249,
-0.021756300702691078,
0.021780073642730713,
0.004328363109380007,
0.14993953704833984,
-0.01780916191637516,
0.07906097173690796,
0.05170699954032898,
0.010053611360490322,
-0.02337079867720604,
0.14390243589878082,
-0.06594522297382355,
-0.06755010038614273,
-0.01743311807513237,
0.1350577473640442,
0.00837484747171402,
-0.024031471461057663,
-0.07528851926326752,
-0.026843970641493797,
0.057907190173864365,
0.09606823325157166,
-0.02376220002770424,
0.009480414912104607,
-0.0663214921951294,
-0.021479256451129913,
0.07170277833938599,
-0.13639305531978607,
0.04132835194468498,
0.016039445996284485,
-0.09386128187179565,
-0.05695444718003273,
0.009726577438414097,
-0.002689003013074398,
-0.03329329565167427,
0.08364108204841614,
-0.0754941776394844,
-0.021911529824137688,
-0.06860765069723129,
-0.04760437458753586,
0.017286039888858795,
-0.08514592796564102,
-0.022088441997766495,
-0.04377994313836098,
-0.201422780752182,
-0.04995529726147652,
0.04190875589847565,
-0.0927731990814209,
-0.044519126415252686,
-0.04512828588485718,
-0.07187692075967789,
0.026383087038993835,
-0.005083616357296705,
0.1403970867395401,
-0.056769538670778275,
0.07746706157922745,
-0.0478479340672493,
0.04138429835438728,
0.0730016753077507,
0.04121099039912224,
-0.06584125012159348,
0.05777120217680931,
-0.11067259311676025,
0.08342797309160233,
-0.1136123463511467,
0.0454581081867218,
-0.14596107602119446,
-0.07717966288328171,
0.030150016769766808,
-0.03512546420097351,
0.07782036811113358,
0.1424405872821808,
-0.1881290227174759,
-0.0005873769405297935,
0.13327881693840027,
-0.048990651965141296,
-0.07660822570323944,
0.10400617122650146,
-0.03981144726276398,
0.01056157611310482,
0.04166784510016441,
0.162796750664711,
0.1233106330037117,
-0.15212059020996094,
0.004067274741828442,
0.01833202876150608,
0.07632067054510117,
0.03276568651199341,
0.07446209341287613,
-0.028876377269625664,
0.021041201427578926,
0.011804845184087753,
-0.08413056284189224,
0.007899723015725613,
-0.06854699552059174,
-0.09453695267438889,
-0.047198109328746796,
-0.09340731054544449,
0.05500417202711105,
0.011309178546071053,
0.005779705476015806,
-0.05622830241918564,
-0.11678057163953781,
0.015122268348932266,
0.15107637643814087,
-0.04767065495252609,
-0.006376910023391247,
-0.06804771721363068,
0.06865677982568741,
0.0008268525707535446,
-0.025389481335878372,
-0.18421801924705505,
-0.10723301768302917,
0.042094871401786804,
-0.06779710948467255,
-0.0022385569754987955,
-0.006838399916887283,
0.06962369382381439,
0.06756143271923065,
-0.03992696851491928,
-0.02856799028813839,
-0.061713531613349915,
-0.005764655768871307,
-0.08684862405061722,
-0.1868169605731964,
-0.06171578913927078,
-0.04559832438826561,
0.1874782145023346,
-0.2331634908914566,
0.0021165141370147467,
-0.014906305819749832,
0.12949734926223755,
0.0388384610414505,
-0.06236688792705536,
0.018259475007653236,
0.02210121788084507,
0.009262505918741226,
-0.11667554825544357,
0.030066417530179024,
-0.012467478401958942,
-0.09740111231803894,
-0.02661338821053505,
-0.11057085543870926,
0.0063048917800188065,
0.053129103034734726,
0.1447739452123642,
-0.11017444729804993,
-0.06115854158997536,
-0.05602551996707916,
-0.041377954185009,
-0.07617456465959549,
-0.002163642318919301,
0.17016690969467163,
0.04210701957345009,
0.10294503718614578,
-0.05891720578074455,
-0.07184285670518875,
0.005384587682783604,
0.011282378807663918,
-0.004047605209052563,
0.10855261236429214,
0.0433477908372879,
-0.10534760355949402,
0.07762916386127472,
0.10953549295663834,
-0.03368935361504555,
0.11755528301000595,
-0.06324557960033417,
-0.09849198162555695,
-0.03093319945037365,
0.04923999309539795,
0.005353488493710756,
0.14552122354507446,
-0.037522055208683014,
0.0067176963202655315,
0.03847895562648773,
0.01899995096027851,
0.006386986467987299,
-0.16654369235038757,
-0.028250539675354958,
0.02097848244011402,
-0.052555229514837265,
-0.035182636231184006,
-0.022673340514302254,
0.015154004096984863,
0.07932545989751816,
0.03484515845775604,
-0.038931671530008316,
0.0039900122210383415,
-0.029033750295639038,
-0.07468361407518387,
0.17075961828231812,
-0.12109611183404922,
-0.1134629026055336,
-0.12240651994943619,
0.05923975259065628,
-0.011820471845567226,
-0.025300923734903336,
-0.013471572659909725,
-0.07453935593366623,
-0.0559641569852829,
-0.10124556720256805,
-0.032117798924446106,
-0.02085292525589466,
-0.025189600884914398,
0.0389159694314003,
0.020036282017827034,
0.07083959132432938,
-0.11512143909931183,
0.02220803312957287,
-0.0031348553020507097,
-0.06792145222425461,
-0.0057808998972177505,
0.041876666247844696,
0.0873798206448555,
0.127597376704216,
0.016041574999690056,
0.006767618004232645,
-0.032902028411626816,
0.22783240675926208,
-0.09551060199737549,
0.012937054969370365,
0.09262112528085709,
0.005281118676066399,
0.06639231741428375,
0.14272892475128174,
0.028383558616042137,
-0.09207399189472198,
0.022700389847159386,
0.049112606793642044,
-0.017077380791306496,
-0.2533135712146759,
-0.038206156343221664,
-0.03302323445677757,
-0.033309340476989746,
0.11362570524215698,
0.057492002844810486,
-0.008644391782581806,
0.039881397038698196,
-0.01816445402801037,
-0.003333010943606496,
0.008082617074251175,
0.06960760802030563,
0.03340393304824829,
0.03015006333589554,
0.07543762773275375,
-0.027742993086576462,
0.0022741518914699554,
0.06438882648944855,
0.044120658189058304,
0.29575657844543457,
-0.03962038457393646,
0.17298050224781036,
0.0033566236961632967,
0.1677374690771103,
-0.02847503125667572,
0.040016718208789825,
0.014581667259335518,
0.0011706154327839613,
-0.005133992061018944,
-0.0626712366938591,
-0.02468392252922058,
0.03272488713264465,
0.009819277562201023,
0.029266152530908585,
-0.07073234021663666,
0.024791967123746872,
0.026903949677944183,
0.2712201178073883,
0.061974119395017624,
-0.2750551700592041,
-0.07236917316913605,
0.03366383537650108,
-0.021303577348589897,
-0.0720909908413887,
0.016964849084615707,
0.1423918753862381,
-0.13358329236507416,
0.042158640921115875,
-0.06328415870666504,
0.08731792122125626,
-0.05518873780965805,
-0.013272762298583984,
0.06864801049232483,
0.10973742604255676,
-0.013488917611539364,
0.08734849095344543,
-0.15464311838150024,
0.21550695598125458,
0.015235479921102524,
0.08746779710054398,
-0.07019758224487305,
0.030629586428403854,
0.016354726627469063,
0.061265721917152405,
0.1341438740491867,
0.007439968176186085,
-0.0520675927400589,
-0.17390909790992737,
-0.12417248636484146,
0.008409256115555763,
0.11133243143558502,
-0.06324554234743118,
0.06900037080049515,
-0.059455305337905884,
-0.010690179653465748,
0.02432444505393505,
-0.09800790995359421,
-0.11747992783784866,
-0.13704179227352142,
0.0286701787263155,
-0.016782809048891068,
-0.017585689201951027,
-0.08311816304922104,
-0.08440335839986801,
0.00024936426780186594,
0.12316916882991791,
-0.04789776727557182,
-0.05422794818878174,
-0.1497361660003662,
0.014344773255288601,
0.17397694289684296,
-0.0538545697927475,
0.021277854219079018,
0.016593117266893387,
0.12092607468366623,
0.0342293456196785,
-0.05140982195734978,
0.060673151165246964,
-0.06858734786510468,
-0.1920088678598404,
-0.07735971361398697,
0.13342078030109406,
0.05349098891019821,
0.05360347777605057,
-0.01672503910958767,
0.029732901602983475,
0.0017961047124117613,
-0.09621672332286835,
0.03151802718639374,
0.09506846964359283,
0.04702330380678177,
0.0573401004076004,
-0.053778763860464096,
0.050235215574502945,
-0.04589756950736046,
-0.02720494195818901,
0.09999649226665497,
0.2438729852437973,
-0.10010778903961182,
0.12686119973659515,
0.07637409120798111,
-0.056688446551561356,
-0.18655286729335785,
0.022913752123713493,
0.11145725846290588,
0.04020083695650101,
0.0600481778383255,
-0.17212243378162384,
0.06305274367332458,
0.11361216753721237,
-0.03316708281636238,
0.09498106688261032,
-0.35978472232818604,
-0.11962989717721939,
0.05805901810526848,
0.08012959361076355,
-0.020978834480047226,
-0.1316465437412262,
-0.05215331166982651,
-0.0179120060056448,
-0.1118120402097702,
0.0585462711751461,
-0.046739835292100906,
0.10604161024093628,
-0.009520748630166054,
0.08855481445789337,
0.040107764303684235,
-0.042363423854112625,
0.1795477271080017,
0.02123935893177986,
0.06959881633520126,
-0.05365299433469772,
-0.0007448808173649013,
0.05850326642394066,
-0.08058799058198929,
0.04847735911607742,
-0.06802096962928772,
0.07293976843357086,
-0.1470731496810913,
-0.01077137142419815,
-0.057468168437480927,
0.034401245415210724,
-0.05170920863747597,
-0.055801551789045334,
-0.04646892100572586,
0.0833139419555664,
0.06522870063781738,
-0.03316597640514374,
0.06841899454593658,
0.024440860375761986,
0.06951680034399033,
0.10087340325117111,
0.06491019576787949,
0.028098272159695625,
-0.12715096771717072,
0.006231504492461681,
-0.017095835879445076,
0.058259133249521255,
-0.15023034811019897,
0.016738876700401306,
0.12216747552156448,
0.050826750695705414,
0.1278025507926941,
0.02329026535153389,
-0.07665100693702698,
-0.0038055696059018373,
0.04223282262682915,
-0.0936022400856018,
-0.15192954242229462,
0.004168127663433552,
0.005024230573326349,
-0.14338183403015137,
-0.015198149718344212,
0.10183949768543243,
-0.04594364017248154,
-0.013515940867364407,
-0.011384577490389347,
0.04800111800432205,
-0.008460878394544125,
0.19636036455631256,
0.04437712952494621,
0.0766880139708519,
-0.07992282509803772,
0.1295921504497528,
0.08508314937353134,
-0.05965736508369446,
0.043092772364616394,
0.06457661092281342,
-0.09843666106462479,
-0.017733631655573845,
0.05598589777946472,
0.1067247986793518,
-0.00009650301217334345,
-0.025035768747329712,
-0.05188044533133507,
-0.09342309087514877,
0.057276397943496704,
0.04726673662662506,
0.027749331668019295,
0.0019959481433033943,
-0.02431119792163372,
0.003975483123213053,
-0.11687524616718292,
0.1022472083568573,
0.06208689138293266,
0.056229639798402786,
-0.13993290066719055,
0.05732792988419533,
-0.0114178117364645,
0.029328620061278343,
-0.007542021106928587,
0.017633700743317604,
-0.09418858587741852,
-0.010631839744746685,
-0.14946526288986206,
0.014621681533753872,
-0.02673276886343956,
0.014858881011605263,
-0.010701489634811878,
-0.04510131850838661,
-0.012745439074933529,
0.03283536806702614,
-0.07310597598552704,
-0.06795894354581833,
-0.011544876731932163,
0.06287618726491928,
-0.13515831530094147,
-0.01885024458169937,
0.041507817804813385,
-0.10307524353265762,
0.08713041990995407,
0.04800519347190857,
0.030285799875855446,
0.0184400025755167,
-0.0952976867556572,
0.0005314411828294396,
0.0021154240239411592,
0.018921729177236557,
0.062060028314590454,
-0.12810668349266052,
-0.027050746604800224,
-0.06511060148477554,
0.0025858799926936626,
0.016823088750243187,
0.02083151787519455,
-0.12989066541194916,
-0.023513004183769226,
-0.03537072241306305,
-0.049521785229444504,
-0.06751929223537445,
0.05345892533659935,
0.10605968534946442,
0.032351285219192505,
0.1333322674036026,
-0.05789482966065407,
0.05755194276571274,
-0.18046247959136963,
-0.043431833386421204,
-0.0015827195020392537,
-0.005373233929276466,
-0.07080087065696716,
-0.015281668864190578,
0.08754762262105942,
-0.05475883558392525,
0.09088809788227081,
-0.03619358688592911,
0.07029847055673599,
0.0209562536329031,
-0.0241877268999815,
-0.0067165992222726345,
0.029794329777359962,
0.13891245424747467,
0.06077821925282478,
-0.018537884578108788,
0.07340367138385773,
-0.03497197851538658,
0.030575953423976898,
0.057677675038576126,
0.14877302944660187,
0.13338392972946167,
0.023081034421920776,
0.04373065382242203,
0.05363920331001282,
-0.1434546411037445,
-0.12392795085906982,
0.1257718801498413,
-0.03357642889022827,
0.10012198239564896,
-0.06589829176664352,
0.16589294373989105,
0.09256797283887863,
-0.1970539689064026,
0.0527045764029026,
-0.07253187149763107,
-0.11795143038034439,
-0.10029861330986023,
-0.07518915832042694,
-0.05721741542220116,
-0.1018209457397461,
0.016576390713453293,
-0.10067302733659744,
0.056052159518003464,
0.09952298551797867,
0.015161420218646526,
0.023766659200191498,
0.12044978141784668,
-0.020763583481311798,
0.01181025616824627,
0.07085555791854858,
0.03816124051809311,
0.007484677713364363,
-0.047483887523412704,
-0.06218796223402023,
0.0494275726377964,
-0.02548776939511299,
0.06651875376701355,
-0.031789135187864304,
0.003050651168450713,
0.039438407868146896,
-0.0027509713545441628,
-0.06697465479373932,
0.02519492618739605,
0.009458610787987709,
0.01202342752367258,
0.06302258372306824,
0.06650834530591965,
0.012780141085386276,
-0.04423517733812332,
0.31267794966697693,
-0.0799495056271553,
-0.05950658395886421,
-0.1434411257505417,
0.1675068587064743,
0.009301701560616493,
-0.003216314595192671,
0.0631847083568573,
-0.11648684740066528,
-0.013261285610496998,
0.13993167877197266,
0.15412527322769165,
-0.08671760559082031,
-0.025823133066296577,
-0.01806446723639965,
-0.01255171187222004,
-0.06161457300186157,
0.1173015609383583,
0.08180782943964005,
0.04086177423596382,
-0.06530707329511642,
0.029600225389003754,
0.005076270550489426,
-0.05079854279756546,
-0.09485138952732086,
0.08128495514392853,
-0.008702835068106651,
0.011383907869458199,
-0.06013408303260803,
0.08472981303930283,
-0.015206673182547092,
-0.1950434148311615,
0.055728185921907425,
-0.15521512925624847,
-0.19617478549480438,
-0.03482077643275261,
0.08472596853971481,
-0.0017012425232678652,
0.060426607728004456,
-0.001769322669133544,
-0.007487575989216566,
0.14709974825382233,
-0.022152280434966087,
-0.03911445289850235,
-0.1102588027715683,
0.0718156024813652,
-0.03827303647994995,
0.23480333387851715,
0.008239395916461945,
0.04933217167854309,
0.09848984330892563,
0.01064758189022541,
-0.16927722096443176,
0.01849677413702011,
0.1081060841679573,
-0.06269580125808716,
0.02728581428527832,
0.1670040488243103,
-0.05104502663016319,
0.09338700026273727,
0.047452010214328766,
-0.07737889885902405,
-0.014611819759011269,
-0.02164958044886589,
0.0080002686008811,
-0.08338459581136703,
-0.0015897375997155905,
-0.04182068631052971,
0.170494943857193,
0.19778700172901154,
-0.05346231907606125,
0.005580282770097256,
-0.052094172686338425,
0.01788310520350933,
0.03955734148621559,
0.07630371302366257,
-0.02067345753312111,
-0.16329683363437653,
0.013834623619914055,
0.03226402401924133,
0.04592142999172211,
-0.2005518674850464,
-0.09359388798475266,
0.0337136946618557,
-0.0463278666138649,
-0.03549389913678169,
0.1321127712726593,
0.04518696293234825,
0.02347993105649948,
-0.025173353031277657,
-0.10024198889732361,
-0.03577529639005661,
0.13542377948760986,
-0.13548563420772552,
-0.04423985257744789
] |
null | null | transformers | <!-- markdownlint-disable first-line-h1 -->
<!-- markdownlint-disable html -->
<div align="center">
<h1>
<img src="image/huozi-logo.jpg" width="30" /> 活字通用大模型
</h1>
</div>
</p>
<div align="center">
<a href="https://github.com/HIT-SCIR/huozi/pulls">
<image src="https://img.shields.io/badge/PRs-welcome-brightgreen">
</a>
<a href="https://github.com/HIT-SCIR/huozi/pulls">
<image src="https://img.shields.io/badge/License-Apache_2.0-green.svg">
</a>
<!-- <h4 align="center">
<p>
<b>中文</b> |
<a href="https://github.com/HIT-SCIR/huozi/blob/main/README_EN.md">English</a>
<p>
</h4> -->
</div>
## 🔖 目录
|章节|说明|
|---|---|
|[💁🏻♂ 开源清单](#-开源清单)|本仓库开源项目清单|
|[💡 模型介绍](#-模型介绍)|简要介绍活字模型结构和训练过程|
|[📥 模型下载](#-模型下载)|活字模型下载链接|
|[💻 模型推理](#-模型推理)|活字模型推理样例,包括vLLM推理加速、llama.cpp量化推理等框架的使用流程|
|[📈 模型性能](#-模型性能)|活字模型在主流评测任务上的性能|
|[🗂 生成样例](#-生成样例)|活字模型实际生成效果样例|
## 💁🏻♂ 开源清单

- **活字 3.0**: [[模型权重](#-模型下载)]
- 活字3.0为一个稀疏混合专家模型,支持32K上下文,具有丰富的中、英文知识和强大的数学推理、代码生成能力。活字3.0较旧版活字具有更强的指令遵循能力和安全性。
- **中文MT-Bench**: [[数据集](data/mt-bench-zh/)]
- 本数据集是英文MT-Bench对话能力评测数据集的中文版。它包含了一系列多轮对话问题,每一组问题都经过了精心的人工校对,并为适应中文语境进行了必要的调整。
- **《ChatGPT 调研报告》**: [[PDF](https://github.com/HIT-SCIR/huozi/blob/main/pdf/chatgpt_book.pdf)]
- 哈工大自然语言处理研究所组织多位老师和同学撰写了本调研报告,从技术原理、应用场景、未来发展等方面对ChatGPT进行了尽量详尽的介绍及总结。
- **活字 2.0**: [[模型权重](https://huggingface.co/HIT-SCIR/huozi-7b-rlhf)] [[RLHF数据](data/huozi-rlhf/huozi_rlhf_data.csv)]
- 在活字1.0基础上,通过人类反馈的强化学习(RLHF)进一步优化了模型回复质量,使其更加符合人类偏好。相较于上一个版本平均长度明显提高,遵从指令的能力更强,逻辑更加清晰。
- 16.9k 人工标注的偏好数据,回复来自活字模型,可以用于训练奖励模型。
- **活字 1.0**: [[模型权重](https://huggingface.co/HIT-SCIR/huozi-7b-sft)]
- 在Bloom模型的基础上,在大约 150 亿 tokens 上进行指令微调训练得到的模型,具有更强的指令遵循能力、更好的安全性。
## 💡 模型介绍
大规模语言模型(LLM)在自然语言处理领域取得了显著的进展,并在广泛的应用场景中展现了其强大的潜力。这一技术不仅吸引了学术界的广泛关注,也成为了工业界的热点。在此背景下,哈尔滨工业大学社会计算与信息检索研究中心(HIT-SCIR)近期推出了最新成果——**活字3.0**,致力于为自然语言处理的研究和实际应用提供更多可能性和选择。
活字3.0是基于Chinese-Mixtral-8x7B,在大约30万行指令数据上微调得到的模型。该模型支持**32K上下文**,能够有效处理长文本。活字3.0继承了基座模型丰富的**中英文知识**,并在**数学推理**、**代码生成**等任务上具有强大性能。经过指令微调,活字3.0还在**指令遵循能力**和**安全性**方面实现了显著提升。
此外,我们开源了**中文MT-Bench数据集**。这是一个中文开放问题集,包括80组对话任务,用于评估模型的多轮对话和指令遵循能力。该数据集是根据原始MT-Bench翻译得来的,每组问题均经过人工校对和中文语境下的适当调整。我们还对原始MT-Bench中的部分错误答案进行了修正。
> [!IMPORTANT]
> 活字系列模型仍然可能生成包含事实性错误的误导性回复或包含偏见/歧视的有害内容,请谨慎鉴别和使用生成的内容,请勿将生成的有害内容传播至互联网。
### 模型结构
活字3.0是一个稀疏混合专家模型(SMoE),使用了Mixtral-8x7B的模型结构。它区别于LLaMA、BLOOM等常见模型,活字3.0的每个前馈神经网络(FFN)层被替换为了“专家层”,该层包含8个FFN和一个“路由器”。这种设计使得模型在推理过程中,可以独立地将每个Token路由到最适合处理它的两个专家中。活字3.0共拥有46.7B个参数,但得益于其稀疏激活的特性,实际推理时仅需激活13B参数,有效提升了计算效率和处理速度。

### 训练过程
由于Mixtral-8x7B词表不支持中文,因此对中文的编解码效率较低,限制了中文场景下的实用性。我们首先基于Mixtral-8x7B进行了中文扩词表增量预训练,显著提高了模型对中文的编解码效率,并使模型具备了强大的中文生成和理解能力。这项成果名为[Chinese-Mixtral-8x7B](https://github.com/HIT-SCIR/Chinese-Mixtral-8x7B),我们已于2024年1月18日开源了其模型权重和训练代码。基于此,我们进一步对模型进行指令微调,最终推出了活字3.0。这一版本的中文编码、指令遵循、安全回复等能力都有显著提升。
## 📥 模型下载
|模型名称|文件大小|下载地址|备注|
|:---:|:---:|:---:|:---:|
|huozi3|88GB|[🤗HuggingFace](https://huggingface.co/HIT-SCIR/huozi3)<br>[ModelScope](https://modelscope.cn/models/HIT-SCIR/huozi3/summary)|活字3.0 完整模型|
|huozi3-gguf|25GB|[🤗HuggingFace](https://huggingface.co/HIT-SCIR/huozi3-gguf)<br>[ModelScope](https://modelscope.cn/models/HIT-SCIR/huozi3-gguf/summary)|活字3.0 GGUF版本,适用于llama.cpp等推理框架|
|huozi3-awq|24GB|[🤗HuggingFace](https://huggingface.co/HIT-SCIR/huozi3-awq)<br>[ModelScope](https://modelscope.cn/models/HIT-SCIR/huozi3-awq/summary)|活字3.0 AWQ版本,适用于AutoAWQ等推理框架|
如果您希望微调活字3.0或Chinese-Mixtral-8x7B,请参考[此处训练代码](https://github.com/HIT-SCIR/Chinese-Mixtral-8x7B?tab=readme-ov-file#%E5%BE%AE%E8%B0%83)。
## 💻 模型推理
### Quick Start
活字3.0采用ChatML格式的prompt模板,格式为:
```
<|beginofutterance|>系统
{system prompt}<|endofutterance|>
<|beginofutterance|>用户
{input}<|endofutterance|>
<|beginofutterance|>助手
{output}<|endofutterance|>
```
使用活字3.0进行推理的示例代码如下:
```python
# quickstart.py
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "HIT-SCIR/huozi3"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
model_id,
attn_implementation="flash_attention_2",
torch_dtype=torch.bfloat16,
device_map="auto",
)
text = """<|beginofutterance|>系统
你是一个智能助手<|endofutterance|>
<|beginofutterance|>用户
请你用python写一段快速排序的代码<|endofutterance|>
<|beginofutterance|>助手
"""
inputs = tokenizer(text, return_tensors="pt").to(0)
outputs = model.generate(
**inputs,
eos_token_id=57001,
temperature=0.8,
top_p=0.9,
max_new_tokens=2048,
)
print(tokenizer.decode(outputs[0], skip_special_tokens=False))
```
活字3.0支持全部Mixtral模型生态,包括Transformers、vLLM、llama.cpp、AutoAWQ、Text generation web UI等框架。
如果您在下载模型时遇到网络问题,可以使用我们在[ModelScope](#modelscope-模型推理)上提供的检查点。
<details>
<summary>
#### Transformers 模型推理 + 流式生成
</summary>
transformers支持为tokenizer添加聊天模板,并支持流式生成。示例代码如下:
```python
# example/transformers-stream/stream.py
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer
model_id = "HIT-SCIR/huozi3"
model = AutoModelForCausalLM.from_pretrained(
model_id,
attn_implementation="flash_attention_2",
torch_dtype=torch.bfloat16,
device_map="auto",
)
tokenizer = AutoTokenizer.from_pretrained(model_id)
tokenizer.chat_template = """{% for message in messages %}{{'<|beginofutterance|>' + message['role'] + '\n' + message['content']}}{% if (loop.last and add_generation_prompt) or not loop.last %}{{ '<|endofutterance|>' + '\n'}}{% endif %}{% endfor %}
{% if add_generation_prompt and messages[-1]['role'] != '助手' %}{{ '<|beginofutterance|>助手\n' }}{% endif %}"""
chat = [
{"role": "系统", "content": "你是一个智能助手"},
{"role": "用户", "content": "请你用python写一段快速排序的代码"},
]
inputs = tokenizer.apply_chat_template(
chat,
tokenize=True,
add_generation_prompt=True,
return_tensors="pt",
).to(0)
stream_output = model.generate(
inputs,
streamer=TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True),
eos_token_id=57001,
temperature=0.8,
top_p=0.9,
max_new_tokens=2048,
)
```
</details>
<details>
<summary>
#### ModelScope 模型推理
</summary>
ModelScope的接口与Transformers非常相似,只需将transformers替换为modelscope即可:
```diff
# example/modelscope-generate/generate.py
import torch
- from transformers import AutoModelForCausalLM, AutoTokenizer
+ from modelscope import AutoTokenizer, AutoModelForCausalLM
model_id = "HIT-SCIR/huozi3"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
model_id,
attn_implementation="flash_attention_2",
torch_dtype=torch.bfloat16,
device_map="auto",
)
text = """<|beginofutterance|>系统
你是一个智能助手<|endofutterance|>
<|beginofutterance|>用户
请你用python写一段快速排序的代码<|endofutterance|>
<|beginofutterance|>助手
"""
inputs = tokenizer(text, return_tensors="pt").to(0)
outputs = model.generate(
**inputs,
eos_token_id=57001,
temperature=0.8,
top_p=0.9,
max_new_tokens=2048,
)
print(tokenizer.decode(outputs[0], skip_special_tokens=False))
```
</details>
<details>
<summary>
#### vLLM 推理加速
</summary>
活字3.0支持通过vLLM实现推理加速,示例代码如下:
```python
# example/vllm-generate/generate.py
from vllm import LLM, SamplingParams
prompts = [
"""<|beginofutterance|>系统
你是一个智能助手<|endofutterance|>
<|beginofutterance|>用户
请你用python写一段快速排序的代码<|endofutterance|>
<|beginofutterance|>助手
""",
]
sampling_params = SamplingParams(
temperature=0.8, top_p=0.95, stop_token_ids=[57001], max_tokens=2048
)
llm = LLM(
model="HIT-SCIR/huozi3",
tensor_parallel_size=4,
)
outputs = llm.generate(prompts, sampling_params)
for output in outputs:
prompt = output.prompt
generated_text = output.outputs[0].text
print(generated_text)
```
</details>
<details>
<summary>
#### 部署 OpenAI API Server
</summary>
活字3.0可以部署为支持OpenAI API协议的服务,这使得活字3.0可以直接通过OpenAI API进行调用。
环境准备:
```shell
$ pip install vllm openai
```
启动服务:
```shell
$ python -m vllm.entrypoints.openai.api_server --model /path/to/huozi3/checkpoint --served-model-name huozi --chat-template template.jinja --tensor-parallel-size 8 --response-role 助手 --max-model-len 2048
```
使用OpenAI API发送请求:
```python
# example/openai-api/openai-client.py
from openai import OpenAI
openai_api_key = "EMPTY"
openai_api_base = "http://localhost:8000/v1"
client = OpenAI(
api_key=openai_api_key,
base_url=openai_api_base,
)
chat_response = client.chat.completions.create(
model="huozi",
messages=[
{"role": "系统", "content": "你是一个智能助手"},
{"role": "用户", "content": "请你用python写一段快速排序的代码"},
],
extra_body={"stop_token_ids": [57001]},
)
print("Chat response:", chat_response.choices[0].message.content)
```
下面是一个使用OpenAI API + Gradio + 流式生成的示例代码:
```python
# example/openai-api/openai-client-gradio.py
from openai import OpenAI
import gradio as gr
openai_api_key = "EMPTY"
openai_api_base = "http://localhost:8000/v1"
client = OpenAI(
api_key=openai_api_key,
base_url=openai_api_base,
)
def predict(message, history):
history_openai_format = [
{"role": "系统", "content": "你是一个智能助手"},
]
for human, assistant in history:
history_openai_format.append({"role": "用户", "content": human})
history_openai_format.append({"role": "助手", "content": assistant})
history_openai_format.append({"role": "用户", "content": message})
models = client.models.list()
stream = client.chat.completions.create(
model=models.data[0].id,
messages=history_openai_format,
temperature=0.8,
stream=True,
extra_body={"repetition_penalty": 1, "stop_token_ids": [57001]},
)
partial_message = ""
for chunk in stream:
partial_message += chunk.choices[0].delta.content or ""
yield partial_message
gr.ChatInterface(predict).queue().launch()
```
</details>
### 量化推理
活字3.0支持量化推理,下表为活字3.0在各个量化框架下显存占用量:
|量化方法|显存占用|
|:---:|:---:|
|无|95GB|
|AWQ|32GB|
|GGUF(q4_0)|28GB|
|GGUF(q2_k)|18GB|
|GGUF(q2_k, offload 16层)|9.6GB|
<details>
<summary>
#### GGUF 格式
</summary>
GGUF格式旨在快速加载和保存模型,由llama.cpp团队推出。我们已经提供了[GGUF格式的活字3.0](https://huggingface.co/HIT-SCIR/huozi3-gguf)。
您也可以手动将HuggingFace格式的活字3.0转换到GGUF格式,以使用其他的量化方法。
##### Step 1 环境准备
首先需要下载llama.cpp的源码。我们在仓库中提供了llama.cpp的submodule,这个版本的llama.cpp已经过测试,可以成功进行推理:
```shell
$ git clone --recurse-submodules https://github.com/HIT-SCIR/huozi
$ cd examples/llama.cpp
```
您也可以下载最新版本的llama.cpp源码:
```shell
$ git clone https://github.com/ggerganov/llama.cpp.git
$ cd llama.cpp
```
然后需要进行编译。根据您的硬件平台,编译命令有细微差异:
```shell
$ make # 用于纯CPU推理
$ make LLAMA_CUBLAS=1 # 用于GPU推理
$ LLAMA_METAL=1 make # 用于Apple Silicon,暂未经过测试
```
##### Step 2 格式转换(可选)
以下命令需要在`llama.cpp/`目录下:
```shell
# 转换为GGUF格式
$ python convert.py --outfile /path/to/huozi-gguf/huozi3.gguf /path/to/huozi3
# 进行GGUF格式的q4_0量化
$ quantize /path/to/huozi-gguf/huozi3.gguf /path/to/huozi-gguf/huozi3-q4_0.gguf q4_0
```
##### Step 3 开始推理
以下命令需要在`llama.cpp/`目录下:
```shell
$ main -m /path/to/huozi-gguf/huozi3-q4_0.gguf --color --interactive-first -c 2048 -t 6 --temp 0.2 --repeat_penalty 1.1 -ngl 999 --in-prefix "<|beginofutterance|>用户\n" --in-suffix "<|endofutterance|>\n<|beginofutterance|>助手" -r "<|endofutterance|>"
```
`-ngl`参数表示向GPU中offload的层数,降低这个值可以缓解GPU显存压力。经过我们的实际测试,q2_k量化的模型offload 16层,显存占用可降低至9.6GB,可在消费级GPU上运行模型:
```shell
$ main -m /path/to/huozi-gguf/huozi3-q2_k.gguf --color --interactive-first -c 2048 -t 6 --temp 0.2 --repeat_penalty 1.1 -ngl 16 --in-prefix "<|beginofutterance|>用户\n" --in-suffix "<|endofutterance|>\n<|beginofutterance|>助手" -r "<|endofutterance|>"
```
关于`main`的更多参数,可以参考llama.cpp的[官方文档](https://github.com/ggerganov/llama.cpp/tree/master/examples/main)。
</details>
<details>
<summary>
#### AWQ 格式
</summary>
AWQ是一种量化模型的存储格式。我们已经提供了[AWQ格式的活字3.0](https://huggingface.co/HIT-SCIR/huozi3-awq),您也可以手动将HuggingFace格式的活字3.0转换到AWQ格式。
##### Step 1 格式转换(可选)
```python
# example/autoawq-generate/quant.py
from awq import AutoAWQForCausalLM
from transformers import AutoTokenizer
model_path = "/path/to/huozi3"
quant_path = "/path/to/save/huozi3-awq"
modules_to_not_convert = ["gate"]
quant_config = {
"zero_point": True,
"q_group_size": 128,
"w_bit": 4,
"version": "GEMM",
"modules_to_not_convert": modules_to_not_convert,
}
model = AutoAWQForCausalLM.from_pretrained(
model_path,
safetensors=True,
**{"low_cpu_mem_usage": True},
)
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
model.quantize(
tokenizer,
quant_config=quant_config,
modules_to_not_convert=modules_to_not_convert,
)
model.save_quantized(quant_path)
tokenizer.save_pretrained(quant_path)
print(f'Model is quantized and saved at "{quant_path}"')
```
##### Step 2 开始推理
在获取到AWQ格式的模型权重后,可以使用AutoAWQForCausalLM代替AutoModelForCausalLM加载模型。示例代码如下:
```diff
# example/autoawq-generate/generate.py
import torch
+ from awq import AutoAWQForCausalLM
from transformers import AutoTokenizer, TextStreamer
- model_id = "HIT-SCIR/huozi3"
+ model_id = "HIT-SCIR/huozi3-awq" # or model_id = "/path/to/saved/huozi3-awq"
+ model = AutoAWQForCausalLM.from_quantized(model_id, fuse_layers=True)
- model = AutoModelForCausalLM.from_pretrained(
- model_id,
- attn_implementation="flash_attention_2",
- torch_dtype=torch.bfloat16,
- device_map="auto",
- )
tokenizer = AutoTokenizer.from_pretrained(model_id)
tokenizer.chat_template = """{% for message in messages %}{{'<|beginofutterance|>' + message['role'] + '\n' + message['content']}}{% if (loop.last and add_generation_prompt) or not loop.last %}{{ '<|endofutterance|>' + '\n'}}{% endif %}{% endfor %}
{% if add_generation_prompt and messages[-1]['role'] != '助手' %}{{ '<|beginofutterance|>助手\n' }}{% endif %}"""
chat = [
{"role": "系统", "content": "你是一个智能助手"},
{"role": "用户", "content": "请你用python写一段快速排序的代码"},
]
inputs = tokenizer.apply_chat_template(
chat,
tokenize=True,
add_generation_prompt=True,
return_tensors="pt",
).to(0)
stream_output = model.generate(
inputs,
streamer=TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True),
eos_token_id=57001,
temperature=0.8,
top_p=0.9,
max_new_tokens=2048,
)
```
</details>
## 📈 模型性能

针对大模型综合能力评价,我们分别使用以下评测数据集对活字3.0进行评测:
- C-Eval:一个全面的中文基础模型评估套件。它包含了13948个多项选择题,涵盖了52个不同的学科和四个难度级别。
- CMMLU:一个综合性的中文评估基准,专门用于评估语言模型在中文语境下的知识和推理能力,涵盖了从基础学科到高级专业水平的67个主题。
- GAOKAO:一个以中国高考题目为数据集,旨在提供和人类对齐的,直观,高效地测评大模型语言理解能力、逻辑推理能力的测评框架。
- MMLU:一个包含57个多选任务的英文评测数据集,涵盖了初等数学、美国历史、计算机科学、法律等,难度覆盖高中水平到专家水平,是目前主流的LLM评测数据集之一。
- HellaSwag:一个极具挑战的英文NLI评测数据集,每一个问题都需要对上下文进行深入理解,而不能基于常识进行回答。
- GSM8K:一个高质量的小学数学应用题的数据集,这些问题需要 2 到 8 个步骤来解决,解决方案主要涉及使用基本算术运算,可用于评价多步数学推理能力。
- HumanEval:一个由 164 个原创编程问题组成的数据集,通过衡量从文档字符串生成程序的功能正确性,来够评估语言理解、算法和简单的数学能力。
- MT-Bench:一个开放的英文问题集,包括80个多轮对话任务,用于评估聊天机器人的多轮对话和指令遵循能力,并通过大模型裁判(GPT-4)对模型回答进行打分。
- MT-Bench-zh:我们根据MT-Bench翻译得来的中文问题集,每组问题均经过人工校对和中文语境下的适当调整。我们已在[此处](data/mt-bench-zh/)开源MT-Bench-zh数据集。
- MT-Bench-safety:我们手工构造的安全数据集,包括暴力、色情、敏感等风险内容。该数据集为封闭数据集。
活字3.0在推理时仅激活13B参数。下表为活字3.0与其他13B规模的中文模型以及旧版活字在各个评测数据集上的结果:
<!-- | 模型名称 | 模型结构 | C-Eval<br>(中文) | CMMLU<br>(中文) | GAOKAO<br>(中文) | MT-Bench-zh<br>(中文对话) | MT-Bench-safety<br>(中文安全) | MMLU<br>(英文) | HellaSwag<br>(英文) | MT-Bench<br>(英文对话) | GSM8K<br>(数学) | HumanEval<br>(代码) |
|---------------------------------------------|---------|--------------|-------------|---------------|--------------------------|-----------------------------|------------|------------------|-----------------------|-------------|-----------------|
| baichuan-inc/Baichuan2-13B-Chat v2 | Baichuan| 56.13 | 58.50 | 48.99 | 6.74 | 8.30 | 54.50 | 51.19 | 6.59 | 25.17 | 20.12 |
| wangrongsheng/Aurora-Plus | Mixtral | 47.67 | 48.75 | 35.05 | 5.47 | 6.70 | 67.80 | 78.27 | 7.13 | 66.26 | 27.44 |
| TigerResearch/tigerbot-13b-chat-v5 | LLaMA | 49.78 | 51.28 | 41.31 | 5.98 | 7.63 | 56.34 | 35.17 | 4.88 | 66.19 | 14.63 |
| hfl/chinese-alpaca-2-13b | LLaMA | 43.47 | 44.53 | 25.94 | 5.77 | 8.13 | 53.05 | 56.85 | 6.24 | 32.75 | 14.02 |
| 活字1.0 | BLOOM | 37.27 | 36.24 | 19.72 | 4.48 | 7.18 | 39.68 | 33.21 | 4.34 | 21.99 | 13.41 |
| 活字2.0 | BLOOM | 32.05 | 34.68 | 22.97 | 5.08 | 6.68 | 38.04 | 33.34 | 4.79 | 19.86 | 6.71 |
| **活字3.0(最新版本)** | Mixtral | 51.82 | 51.06 | 41.21 | 6.29 | 7.58 | 69.48 | 65.18 | 7.62 | 65.81 | 40.85 | -->

> 我们在C-Eval、CMMLU、MMLU采用5-shot,GSM8K采用4-shot,HellaSwag、HumanEval采用0-shot,HumanEval采用pass@1指标。所有测试均采用greedy策略。
>
> 我们使用OpenCompass作为评测框架,commit hash为[4c87e77](https://github.com/open-compass/opencompass/tree/4c87e777d855636b9eda7ec87bcbbf12b62caed3)。评测代码位于[此处](./evaluate/)。
根据上表中的测试结果,活字3.0较旧版活字取得了巨大的性能提升。在中文知识方面,活字3.0达到了与Tigerbot-13B-chat-v5相当的性能,并是在中文对话和指令遵循方面表现得更加优秀。在英文知识方面,得益于原版Mixtral-8x7B的强大性能,活字3.0超过了Baichuan2-13B-Chat v2和LLaMA系列的扩词表模型,并在英文对话和指令遵循能力上达到了较高水平。在数学推理和代码生成任务上,活字3.0均展现出强大的性能,这说明活字3.0对复杂问题的深层次理解、多步推理、以及结构化信息处理等方面具有较强水平。由于我们采用了较高质量的代码数据集,活字3.0的代码生成能力也超越了同为Mixtral结构的Aurora-Plus模型。
## 🗂 生成样例
下面是活字3.0在MT-Bench-zh评测集上的生成效果展示,并与活字2.0(RLHF版本)进行对比:





## <img src="https://cdn.jsdelivr.net/gh/LightChen233/blog-img/folders.png" width="25" /> 开源协议
对本仓库源码的使用遵循开源许可协议 [Apache 2.0](https://github.com/HIT-SCIR/huozi/blob/main/LICENSE)。
活字支持商用。如果将活字模型或其衍生品用作商业用途,请您按照如下方式联系许可方,以进行登记并向许可方申请书面授权:联系邮箱:<[email protected]>。
## <img src="https://cdn.jsdelivr.net/gh/LightChen233/blog-img/notes.png" width="25" /> Citation
### 活字大模型
```latex
@misc{huozi,
author = {Huozi-Team}.
title = {Huozi: Leveraging Large Language Models for Enhanced Open-Domain Chatting}
year = {2024},
publisher = {GitHub},
journal = {GitHub repository}
howpublished = {\url{https://github.com/HIT-SCIR/huozi}}
}
```
## <img src="https://cdn.jsdelivr.net/gh/LightChen233/blog-img/star.png" width="25" /> Star History
[](https://star-history.com/#HIT-SCIR/huozi&Date)
| {} | text-generation | HIT-SCIR/huozi3 | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T08:27:24+00:00 | [] | [] | TAGS
#transformers #safetensors #mixtral #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
活字通用大模型
=========
目录
--
开源清单
------

* 活字 3.0: [模型权重]
+ 活字3.0为一个稀疏混合专家模型,支持32K上下文,具有丰富的中、英文知识和强大的数学推理、代码生成能力。活字3.0较旧版活字具有更强的指令遵循能力和安全性。
* 中文MT-Bench: [数据集]
+ 本数据集是英文MT-Bench对话能力评测数据集的中文版。它包含了一系列多轮对话问题,每一组问题都经过了精心的人工校对,并为适应中文语境进行了必要的调整。
* 《ChatGPT 调研报告》: [PDF]
+ 哈工大自然语言处理研究所组织多位老师和同学撰写了本调研报告,从技术原理、应用场景、未来发展等方面对ChatGPT进行了尽量详尽的介绍及总结。
* 活字 2.0: [模型权重] [RLHF数据]
+ 在活字1.0基础上,通过人类反馈的强化学习(RLHF)进一步优化了模型回复质量,使其更加符合人类偏好。相较于上一个版本平均长度明显提高,遵从指令的能力更强,逻辑更加清晰。
+ 16.9k 人工标注的偏好数据,回复来自活字模型,可以用于训练奖励模型。
* 活字 1.0: [模型权重]
+ 在Bloom模型的基础上,在大约 150 亿 tokens 上进行指令微调训练得到的模型,具有更强的指令遵循能力、更好的安全性。
模型介绍
----
大规模语言模型(LLM)在自然语言处理领域取得了显著的进展,并在广泛的应用场景中展现了其强大的潜力。这一技术不仅吸引了学术界的广泛关注,也成为了工业界的热点。在此背景下,哈尔滨工业大学社会计算与信息检索研究中心(HIT-SCIR)近期推出了最新成果——活字3.0,致力于为自然语言处理的研究和实际应用提供更多可能性和选择。
活字3.0是基于Chinese-Mixtral-8x7B,在大约30万行指令数据上微调得到的模型。该模型支持32K上下文,能够有效处理长文本。活字3.0继承了基座模型丰富的中英文知识,并在数学推理、代码生成等任务上具有强大性能。经过指令微调,活字3.0还在指令遵循能力和安全性方面实现了显著提升。
此外,我们开源了中文MT-Bench数据集。这是一个中文开放问题集,包括80组对话任务,用于评估模型的多轮对话和指令遵循能力。该数据集是根据原始MT-Bench翻译得来的,每组问题均经过人工校对和中文语境下的适当调整。我们还对原始MT-Bench中的部分错误答案进行了修正。
>
> [!IMPORTANT]
> 活字系列模型仍然可能生成包含事实性错误的误导性回复或包含偏见/歧视的有害内容,请谨慎鉴别和使用生成的内容,请勿将生成的有害内容传播至互联网。
>
>
>
### 模型结构
活字3.0是一个稀疏混合专家模型(SMoE),使用了Mixtral-8x7B的模型结构。它区别于LLaMA、BLOOM等常见模型,活字3.0的每个前馈神经网络(FFN)层被替换为了“专家层”,该层包含8个FFN和一个“路由器”。这种设计使得模型在推理过程中,可以独立地将每个Token路由到最适合处理它的两个专家中。活字3.0共拥有46.7B个参数,但得益于其稀疏激活的特性,实际推理时仅需激活13B参数,有效提升了计算效率和处理速度。

### 训练过程
由于Mixtral-8x7B词表不支持中文,因此对中文的编解码效率较低,限制了中文场景下的实用性。我们首先基于Mixtral-8x7B进行了中文扩词表增量预训练,显著提高了模型对中文的编解码效率,并使模型具备了强大的中文生成和理解能力。这项成果名为Chinese-Mixtral-8x7B,我们已于2024年1月18日开源了其模型权重和训练代码。基于此,我们进一步对模型进行指令微调,最终推出了活字3.0。这一版本的中文编码、指令遵循、安全回复等能力都有显著提升。
模型下载
----
如果您希望微调活字3.0或Chinese-Mixtral-8x7B,请参考此处训练代码。
模型推理
----
### Quick Start
活字3.0采用ChatML格式的prompt模板,格式为:
使用活字3.0进行推理的示例代码如下:
活字3.0支持全部Mixtral模型生态,包括Transformers、vLLM、URL、AutoAWQ、Text generation web UI等框架。
如果您在下载模型时遇到网络问题,可以使用我们在ModelScope上提供的检查点。
#### Transformers 模型推理 + 流式生成
transformers支持为tokenizer添加聊天模板,并支持流式生成。示例代码如下:
#### ModelScope 模型推理
ModelScope的接口与Transformers非常相似,只需将transformers替换为modelscope即可:
#### vLLM 推理加速
活字3.0支持通过vLLM实现推理加速,示例代码如下:
#### 部署 OpenAI API Server
活字3.0可以部署为支持OpenAI API协议的服务,这使得活字3.0可以直接通过OpenAI API进行调用。
环境准备:
启动服务:
使用OpenAI API发送请求:
下面是一个使用OpenAI API + Gradio + 流式生成的示例代码:
### 量化推理
活字3.0支持量化推理,下表为活字3.0在各个量化框架下显存占用量:
#### GGUF 格式
GGUF格式旨在快速加载和保存模型,由llama.cpp团队推出。我们已经提供了GGUF格式的活字3.0。
您也可以手动将HuggingFace格式的活字3.0转换到GGUF格式,以使用其他的量化方法。
##### Step 1 环境准备
首先需要下载llama.cpp的源码。我们在仓库中提供了llama.cpp的submodule,这个版本的llama.cpp已经过测试,可以成功进行推理:
您也可以下载最新版本的llama.cpp源码:
然后需要进行编译。根据您的硬件平台,编译命令有细微差异:
##### Step 2 格式转换(可选)
以下命令需要在'URL'目录下:
##### Step 3 开始推理
以下命令需要在'URL'目录下:
'-ngl'参数表示向GPU中offload的层数,降低这个值可以缓解GPU显存压力。经过我们的实际测试,q2\_k量化的模型offload 16层,显存占用可降低至9.6GB,可在消费级GPU上运行模型:
关于'main'的更多参数,可以参考llama.cpp的官方文档。
#### AWQ 格式
AWQ是一种量化模型的存储格式。我们已经提供了AWQ格式的活字3.0,您也可以手动将HuggingFace格式的活字3.0转换到AWQ格式。
##### Step 1 格式转换(可选)
##### Step 2 开始推理
在获取到AWQ格式的模型权重后,可以使用AutoAWQForCausalLM代替AutoModelForCausalLM加载模型。示例代码如下:
模型性能
----

针对大模型综合能力评价,我们分别使用以下评测数据集对活字3.0进行评测:
* C-Eval:一个全面的中文基础模型评估套件。它包含了13948个多项选择题,涵盖了52个不同的学科和四个难度级别。
* CMMLU:一个综合性的中文评估基准,专门用于评估语言模型在中文语境下的知识和推理能力,涵盖了从基础学科到高级专业水平的67个主题。
* GAOKAO:一个以中国高考题目为数据集,旨在提供和人类对齐的,直观,高效地测评大模型语言理解能力、逻辑推理能力的测评框架。
* MMLU:一个包含57个多选任务的英文评测数据集,涵盖了初等数学、美国历史、计算机科学、法律等,难度覆盖高中水平到专家水平,是目前主流的LLM评测数据集之一。
* HellaSwag:一个极具挑战的英文NLI评测数据集,每一个问题都需要对上下文进行深入理解,而不能基于常识进行回答。
* GSM8K:一个高质量的小学数学应用题的数据集,这些问题需要 2 到 8 个步骤来解决,解决方案主要涉及使用基本算术运算,可用于评价多步数学推理能力。
* HumanEval:一个由 164 个原创编程问题组成的数据集,通过衡量从文档字符串生成程序的功能正确性,来够评估语言理解、算法和简单的数学能力。
* MT-Bench:一个开放的英文问题集,包括80个多轮对话任务,用于评估聊天机器人的多轮对话和指令遵循能力,并通过大模型裁判(GPT-4)对模型回答进行打分。
* MT-Bench-zh:我们根据MT-Bench翻译得来的中文问题集,每组问题均经过人工校对和中文语境下的适当调整。我们已在此处开源MT-Bench-zh数据集。
* MT-Bench-safety:我们手工构造的安全数据集,包括暴力、色情、敏感等风险内容。该数据集为封闭数据集。
活字3.0在推理时仅激活13B参数。下表为活字3.0与其他13B规模的中文模型以及旧版活字在各个评测数据集上的结果:

>
> 我们在C-Eval、CMMLU、MMLU采用5-shot,GSM8K采用4-shot,HellaSwag、HumanEval采用0-shot,HumanEval采用pass@1指标。所有测试均采用greedy策略。
>
>
> 我们使用OpenCompass作为评测框架,commit hash为4c87e77。评测代码位于此处。
>
>
>
根据上表中的测试结果,活字3.0较旧版活字取得了巨大的性能提升。在中文知识方面,活字3.0达到了与Tigerbot-13B-chat-v5相当的性能,并是在中文对话和指令遵循方面表现得更加优秀。在英文知识方面,得益于原版Mixtral-8x7B的强大性能,活字3.0超过了Baichuan2-13B-Chat v2和LLaMA系列的扩词表模型,并在英文对话和指令遵循能力上达到了较高水平。在数学推理和代码生成任务上,活字3.0均展现出强大的性能,这说明活字3.0对复杂问题的深层次理解、多步推理、以及结构化信息处理等方面具有较强水平。由于我们采用了较高质量的代码数据集,活字3.0的代码生成能力也超越了同为Mixtral结构的Aurora-Plus模型。
生成样例
----
下面是活字3.0在MT-Bench-zh评测集上的生成效果展示,并与活字2.0(RLHF版本)进行对比:





<img src="URL width="25" /> 开源协议
--------------------------------
对本仓库源码的使用遵循开源许可协议 Apache 2.0。
活字支持商用。如果将活字模型或其衍生品用作商业用途,请您按照如下方式联系许可方,以进行登记并向许可方申请书面授权:联系邮箱:[jngao@URL](mailto:jngao@URL)。
<img src="URL width="25" /> Citation
------------------------------------
### 活字大模型
<img src="URL width="25" /> Star History
----------------------------------------
,使用了Mixtral-8x7B的模型结构。它区别于LLaMA、BLOOM等常见模型,活字3.0的每个前馈神经网络(FFN)层被替换为了“专家层”,该层包含8个FFN和一个“路由器”。这种设计使得模型在推理过程中,可以独立地将每个Token路由到最适合处理它的两个专家中。活字3.0共拥有46.7B个参数,但得益于其稀疏激活的特性,实际推理时仅需激活13B参数,有效提升了计算效率和处理速度。\n\n\n",
"### 训练过程\n\n\n由于Mixtral-8x7B词表不支持中文,因此对中文的编解码效率较低,限制了中文场景下的实用性。我们首先基于Mixtral-8x7B进行了中文扩词表增量预训练,显著提高了模型对中文的编解码效率,并使模型具备了强大的中文生成和理解能力。这项成果名为Chinese-Mixtral-8x7B,我们已于2024年1月18日开源了其模型权重和训练代码。基于此,我们进一步对模型进行指令微调,最终推出了活字3.0。这一版本的中文编码、指令遵循、安全回复等能力都有显著提升。\n\n\n模型下载\n----\n\n\n\n如果您希望微调活字3.0或Chinese-Mixtral-8x7B,请参考此处训练代码。\n\n\n模型推理\n----",
"### Quick Start\n\n\n活字3.0采用ChatML格式的prompt模板,格式为:\n\n\n使用活字3.0进行推理的示例代码如下:\n\n\n活字3.0支持全部Mixtral模型生态,包括Transformers、vLLM、URL、AutoAWQ、Text generation web UI等框架。\n\n\n如果您在下载模型时遇到网络问题,可以使用我们在ModelScope上提供的检查点。",
"#### Transformers 模型推理 + 流式生成\n\n\n\ntransformers支持为tokenizer添加聊天模板,并支持流式生成。示例代码如下:",
"#### ModelScope 模型推理\n\n\n\nModelScope的接口与Transformers非常相似,只需将transformers替换为modelscope即可:",
"#### vLLM 推理加速\n\n\n\n活字3.0支持通过vLLM实现推理加速,示例代码如下:",
"#### 部署 OpenAI API Server\n\n\n\n活字3.0可以部署为支持OpenAI API协议的服务,这使得活字3.0可以直接通过OpenAI API进行调用。\n\n\n环境准备:\n\n\n启动服务:\n\n\n使用OpenAI API发送请求:\n\n\n下面是一个使用OpenAI API + Gradio + 流式生成的示例代码:",
"### 量化推理\n\n\n活字3.0支持量化推理,下表为活字3.0在各个量化框架下显存占用量:",
"#### GGUF 格式\n\n\n\nGGUF格式旨在快速加载和保存模型,由llama.cpp团队推出。我们已经提供了GGUF格式的活字3.0。\n\n\n您也可以手动将HuggingFace格式的活字3.0转换到GGUF格式,以使用其他的量化方法。",
"##### Step 1 环境准备\n\n\n首先需要下载llama.cpp的源码。我们在仓库中提供了llama.cpp的submodule,这个版本的llama.cpp已经过测试,可以成功进行推理:\n\n\n您也可以下载最新版本的llama.cpp源码:\n\n\n然后需要进行编译。根据您的硬件平台,编译命令有细微差异:",
"##### Step 2 格式转换(可选)\n\n\n以下命令需要在'URL'目录下:",
"##### Step 3 开始推理\n\n\n以下命令需要在'URL'目录下:\n\n\n'-ngl'参数表示向GPU中offload的层数,降低这个值可以缓解GPU显存压力。经过我们的实际测试,q2\\_k量化的模型offload 16层,显存占用可降低至9.6GB,可在消费级GPU上运行模型:\n\n\n关于'main'的更多参数,可以参考llama.cpp的官方文档。",
"#### AWQ 格式\n\n\n\nAWQ是一种量化模型的存储格式。我们已经提供了AWQ格式的活字3.0,您也可以手动将HuggingFace格式的活字3.0转换到AWQ格式。",
"##### Step 1 格式转换(可选)",
"##### Step 2 开始推理\n\n\n在获取到AWQ格式的模型权重后,可以使用AutoAWQForCausalLM代替AutoModelForCausalLM加载模型。示例代码如下:\n\n\n\n模型性能\n----\n\n\n\n\n\n针对大模型综合能力评价,我们分别使用以下评测数据集对活字3.0进行评测:\n\n\n* C-Eval:一个全面的中文基础模型评估套件。它包含了13948个多项选择题,涵盖了52个不同的学科和四个难度级别。\n* CMMLU:一个综合性的中文评估基准,专门用于评估语言模型在中文语境下的知识和推理能力,涵盖了从基础学科到高级专业水平的67个主题。\n* GAOKAO:一个以中国高考题目为数据集,旨在提供和人类对齐的,直观,高效地测评大模型语言理解能力、逻辑推理能力的测评框架。\n* MMLU:一个包含57个多选任务的英文评测数据集,涵盖了初等数学、美国历史、计算机科学、法律等,难度覆盖高中水平到专家水平,是目前主流的LLM评测数据集之一。\n* HellaSwag:一个极具挑战的英文NLI评测数据集,每一个问题都需要对上下文进行深入理解,而不能基于常识进行回答。\n* GSM8K:一个高质量的小学数学应用题的数据集,这些问题需要 2 到 8 个步骤来解决,解决方案主要涉及使用基本算术运算,可用于评价多步数学推理能力。\n* HumanEval:一个由 164 个原创编程问题组成的数据集,通过衡量从文档字符串生成程序的功能正确性,来够评估语言理解、算法和简单的数学能力。\n* MT-Bench:一个开放的英文问题集,包括80个多轮对话任务,用于评估聊天机器人的多轮对话和指令遵循能力,并通过大模型裁判(GPT-4)对模型回答进行打分。\n* MT-Bench-zh:我们根据MT-Bench翻译得来的中文问题集,每组问题均经过人工校对和中文语境下的适当调整。我们已在此处开源MT-Bench-zh数据集。\n* MT-Bench-safety:我们手工构造的安全数据集,包括暴力、色情、敏感等风险内容。该数据集为封闭数据集。\n\n\n活字3.0在推理时仅激活13B参数。下表为活字3.0与其他13B规模的中文模型以及旧版活字在各个评测数据集上的结果:\n\n\n\n\n\n\n> \n> 我们在C-Eval、CMMLU、MMLU采用5-shot,GSM8K采用4-shot,HellaSwag、HumanEval采用0-shot,HumanEval采用pass@1指标。所有测试均采用greedy策略。\n> \n> \n> 我们使用OpenCompass作为评测框架,commit hash为4c87e77。评测代码位于此处。\n> \n> \n> \n\n\n根据上表中的测试结果,活字3.0较旧版活字取得了巨大的性能提升。在中文知识方面,活字3.0达到了与Tigerbot-13B-chat-v5相当的性能,并是在中文对话和指令遵循方面表现得更加优秀。在英文知识方面,得益于原版Mixtral-8x7B的强大性能,活字3.0超过了Baichuan2-13B-Chat v2和LLaMA系列的扩词表模型,并在英文对话和指令遵循能力上达到了较高水平。在数学推理和代码生成任务上,活字3.0均展现出强大的性能,这说明活字3.0对复杂问题的深层次理解、多步推理、以及结构化信息处理等方面具有较强水平。由于我们采用了较高质量的代码数据集,活字3.0的代码生成能力也超越了同为Mixtral结构的Aurora-Plus模型。\n\n\n生成样例\n----\n\n\n下面是活字3.0在MT-Bench-zh评测集上的生成效果展示,并与活字2.0(RLHF版本)进行对比:\n\n\n\n\n\n\n\n\n\n<img src=\"URL width=\"25\" /> 开源协议\n--------------------------------\n\n\n对本仓库源码的使用遵循开源许可协议 Apache 2.0。\n\n\n活字支持商用。如果将活字模型或其衍生品用作商业用途,请您按照如下方式联系许可方,以进行登记并向许可方申请书面授权:联系邮箱:[jngao@URL](mailto:jngao@URL)。\n\n\n<img src=\"URL width=\"25\" /> Citation\n------------------------------------",
"### 活字大模型\n\n\n<img src=\"URL width=\"25\" /> Star History\n----------------------------------------\n\n\n,使用了Mixtral-8x7B的模型结构。它区别于LLaMA、BLOOM等常见模型,活字3.0的每个前馈神经网络(FFN)层被替换为了“专家层”,该层包含8个FFN和一个“路由器”。这种设计使得模型在推理过程中,可以独立地将每个Token路由到最适合处理它的两个专家中。活字3.0共拥有46.7B个参数,但得益于其稀疏激活的特性,实际推理时仅需激活13B参数,有效提升了计算效率和处理速度。\n\n\n",
"### 训练过程\n\n\n由于Mixtral-8x7B词表不支持中文,因此对中文的编解码效率较低,限制了中文场景下的实用性。我们首先基于Mixtral-8x7B进行了中文扩词表增量预训练,显著提高了模型对中文的编解码效率,并使模型具备了强大的中文生成和理解能力。这项成果名为Chinese-Mixtral-8x7B,我们已于2024年1月18日开源了其模型权重和训练代码。基于此,我们进一步对模型进行指令微调,最终推出了活字3.0。这一版本的中文编码、指令遵循、安全回复等能力都有显著提升。\n\n\n模型下载\n----\n\n\n\n如果您希望微调活字3.0或Chinese-Mixtral-8x7B,请参考此处训练代码。\n\n\n模型推理\n----",
"### Quick Start\n\n\n活字3.0采用ChatML格式的prompt模板,格式为:\n\n\n使用活字3.0进行推理的示例代码如下:\n\n\n活字3.0支持全部Mixtral模型生态,包括Transformers、vLLM、URL、AutoAWQ、Text generation web UI等框架。\n\n\n如果您在下载模型时遇到网络问题,可以使用我们在ModelScope上提供的检查点。",
"#### Transformers 模型推理 + 流式生成\n\n\n\ntransformers支持为tokenizer添加聊天模板,并支持流式生成。示例代码如下:",
"#### ModelScope 模型推理\n\n\n\nModelScope的接口与Transformers非常相似,只需将transformers替换为modelscope即可:",
"#### vLLM 推理加速\n\n\n\n活字3.0支持通过vLLM实现推理加速,示例代码如下:",
"#### 部署 OpenAI API Server\n\n\n\n活字3.0可以部署为支持OpenAI API协议的服务,这使得活字3.0可以直接通过OpenAI API进行调用。\n\n\n环境准备:\n\n\n启动服务:\n\n\n使用OpenAI API发送请求:\n\n\n下面是一个使用OpenAI API + Gradio + 流式生成的示例代码:",
"### 量化推理\n\n\n活字3.0支持量化推理,下表为活字3.0在各个量化框架下显存占用量:",
"#### GGUF 格式\n\n\n\nGGUF格式旨在快速加载和保存模型,由llama.cpp团队推出。我们已经提供了GGUF格式的活字3.0。\n\n\n您也可以手动将HuggingFace格式的活字3.0转换到GGUF格式,以使用其他的量化方法。",
"##### Step 1 环境准备\n\n\n首先需要下载llama.cpp的源码。我们在仓库中提供了llama.cpp的submodule,这个版本的llama.cpp已经过测试,可以成功进行推理:\n\n\n您也可以下载最新版本的llama.cpp源码:\n\n\n然后需要进行编译。根据您的硬件平台,编译命令有细微差异:",
"##### Step 2 格式转换(可选)\n\n\n以下命令需要在'URL'目录下:",
"##### Step 3 开始推理\n\n\n以下命令需要在'URL'目录下:\n\n\n'-ngl'参数表示向GPU中offload的层数,降低这个值可以缓解GPU显存压力。经过我们的实际测试,q2\\_k量化的模型offload 16层,显存占用可降低至9.6GB,可在消费级GPU上运行模型:\n\n\n关于'main'的更多参数,可以参考llama.cpp的官方文档。",
"#### AWQ 格式\n\n\n\nAWQ是一种量化模型的存储格式。我们已经提供了AWQ格式的活字3.0,您也可以手动将HuggingFace格式的活字3.0转换到AWQ格式。",
"##### Step 1 格式转换(可选)",
"##### Step 2 开始推理\n\n\n在获取到AWQ格式的模型权重后,可以使用AutoAWQForCausalLM代替AutoModelForCausalLM加载模型。示例代码如下:\n\n\n\n模型性能\n----\n\n\n\n\n\n针对大模型综合能力评价,我们分别使用以下评测数据集对活字3.0进行评测:\n\n\n* C-Eval:一个全面的中文基础模型评估套件。它包含了13948个多项选择题,涵盖了52个不同的学科和四个难度级别。\n* CMMLU:一个综合性的中文评估基准,专门用于评估语言模型在中文语境下的知识和推理能力,涵盖了从基础学科到高级专业水平的67个主题。\n* GAOKAO:一个以中国高考题目为数据集,旨在提供和人类对齐的,直观,高效地测评大模型语言理解能力、逻辑推理能力的测评框架。\n* MMLU:一个包含57个多选任务的英文评测数据集,涵盖了初等数学、美国历史、计算机科学、法律等,难度覆盖高中水平到专家水平,是目前主流的LLM评测数据集之一。\n* HellaSwag:一个极具挑战的英文NLI评测数据集,每一个问题都需要对上下文进行深入理解,而不能基于常识进行回答。\n* GSM8K:一个高质量的小学数学应用题的数据集,这些问题需要 2 到 8 个步骤来解决,解决方案主要涉及使用基本算术运算,可用于评价多步数学推理能力。\n* HumanEval:一个由 164 个原创编程问题组成的数据集,通过衡量从文档字符串生成程序的功能正确性,来够评估语言理解、算法和简单的数学能力。\n* MT-Bench:一个开放的英文问题集,包括80个多轮对话任务,用于评估聊天机器人的多轮对话和指令遵循能力,并通过大模型裁判(GPT-4)对模型回答进行打分。\n* MT-Bench-zh:我们根据MT-Bench翻译得来的中文问题集,每组问题均经过人工校对和中文语境下的适当调整。我们已在此处开源MT-Bench-zh数据集。\n* MT-Bench-safety:我们手工构造的安全数据集,包括暴力、色情、敏感等风险内容。该数据集为封闭数据集。\n\n\n活字3.0在推理时仅激活13B参数。下表为活字3.0与其他13B规模的中文模型以及旧版活字在各个评测数据集上的结果:\n\n\n\n\n\n\n> \n> 我们在C-Eval、CMMLU、MMLU采用5-shot,GSM8K采用4-shot,HellaSwag、HumanEval采用0-shot,HumanEval采用pass@1指标。所有测试均采用greedy策略。\n> \n> \n> 我们使用OpenCompass作为评测框架,commit hash为4c87e77。评测代码位于此处。\n> \n> \n> \n\n\n根据上表中的测试结果,活字3.0较旧版活字取得了巨大的性能提升。在中文知识方面,活字3.0达到了与Tigerbot-13B-chat-v5相当的性能,并是在中文对话和指令遵循方面表现得更加优秀。在英文知识方面,得益于原版Mixtral-8x7B的强大性能,活字3.0超过了Baichuan2-13B-Chat v2和LLaMA系列的扩词表模型,并在英文对话和指令遵循能力上达到了较高水平。在数学推理和代码生成任务上,活字3.0均展现出强大的性能,这说明活字3.0对复杂问题的深层次理解、多步推理、以及结构化信息处理等方面具有较强水平。由于我们采用了较高质量的代码数据集,活字3.0的代码生成能力也超越了同为Mixtral结构的Aurora-Plus模型。\n\n\n生成样例\n----\n\n\n下面是活字3.0在MT-Bench-zh评测集上的生成效果展示,并与活字2.0(RLHF版本)进行对比:\n\n\n\n\n\n\n\n\n\n<img src=\"URL width=\"25\" /> 开源协议\n--------------------------------\n\n\n对本仓库源码的使用遵循开源许可协议 Apache 2.0。\n\n\n活字支持商用。如果将活字模型或其衍生品用作商业用途,请您按照如下方式联系许可方,以进行登记并向许可方申请书面授权:联系邮箱:[jngao@URL](mailto:jngao@URL)。\n\n\n<img src=\"URL width=\"25\" /> Citation\n------------------------------------",
"### 活字大模型\n\n\n<img src=\"URL width=\"25\" /> Star History\n----------------------------------------\n\n\n,使用了Mixtral-8x7B的模型结构。它区别于LLaMA、BLOOM等常见模型,活字3.0的每个前馈神经网络(FFN)层被替换为了“专家层”,该层包含8个FFN和一个“路由器”。这种设计使得模型在推理过程中,可以独立地将每个Token路由到最适合处理它的两个专家中。活字3.0共拥有46.7B个参数,但得益于其稀疏激活的特性,实际推理时仅需激活13B参数,有效提升了计算效率和处理速度。\n\n\n### 训练过程\n\n\n由于Mixtral-8x7B词表不支持中文,因此对中文的编解码效率较低,限制了中文场景下的实用性。我们首先基于Mixtral-8x7B进行了中文扩词表增量预训练,显著提高了模型对中文的编解码效率,并使模型具备了强大的中文生成和理解能力。这项成果名为Chinese-Mixtral-8x7B,我们已于2024年1月18日开源了其模型权重和训练代码。基于此,我们进一步对模型进行指令微调,最终推出了活字3.0。这一版本的中文编码、指令遵循、安全回复等能力都有显著提升。\n\n\n模型下载\n----\n\n\n\n如果您希望微调活字3.0或Chinese-Mixtral-8x7B,请参考此处训练代码。\n\n\n模型推理\n----### Quick Start\n\n\n活字3.0采用ChatML格式的prompt模板,格式为:\n\n\n使用活字3.0进行推理的示例代码如下:\n\n\n活字3.0支持全部Mixtral模型生态,包括Transformers、vLLM、URL、AutoAWQ、Text generation web UI等框架。\n\n\n如果您在下载模型时遇到网络问题,可以使用我们在ModelScope上提供的检查点。",
"passage: #### Transformers 模型推理 + 流式生成\n\n\n\ntransformers支持为tokenizer添加聊天模板,并支持流式生成。示例代码如下:#### ModelScope 模型推理\n\n\n\nModelScope的接口与Transformers非常相似,只需将transformers替换为modelscope即可:#### vLLM 推理加速\n\n\n\n活字3.0支持通过vLLM实现推理加速,示例代码如下:#### 部署 OpenAI API Server\n\n\n\n活字3.0可以部署为支持OpenAI API协议的服务,这使得活字3.0可以直接通过OpenAI API进行调用。\n\n\n环境准备:\n\n\n启动服务:\n\n\n使用OpenAI API发送请求:\n\n\n下面是一个使用OpenAI API + Gradio + 流式生成的示例代码:### 量化推理\n\n\n活字3.0支持量化推理,下表为活字3.0在各个量化框架下显存占用量:#### GGUF 格式\n\n\n\nGGUF格式旨在快速加载和保存模型,由llama.cpp团队推出。我们已经提供了GGUF格式的活字3.0。\n\n\n您也可以手动将HuggingFace格式的活字3.0转换到GGUF格式,以使用其他的量化方法。##### Step 1 环境准备\n\n\n首先需要下载llama.cpp的源码。我们在仓库中提供了llama.cpp的submodule,这个版本的llama.cpp已经过测试,可以成功进行推理:\n\n\n您也可以下载最新版本的llama.cpp源码:\n\n\n然后需要进行编译。根据您的硬件平台,编译命令有细微差异:##### Step 2 格式转换(可选)\n\n\n以下命令需要在'URL'目录下:##### Step 3 开始推理\n\n\n以下命令需要在'URL'目录下:\n\n\n'-ngl'参数表示向GPU中offload的层数,降低这个值可以缓解GPU显存压力。经过我们的实际测试,q2\\_k量化的模型offload 16层,显存占用可降低至9.6GB,可在消费级GPU上运行模型:\n\n\n关于'main'的更多参数,可以参考llama.cpp的官方文档。#### AWQ 格式\n\n\n\nAWQ是一种量化模型的存储格式。我们已经提供了AWQ格式的活字3.0,您也可以手动将HuggingFace格式的活字3.0转换到AWQ格式。##### Step 1 格式转换(可选)"
] | [
-0.059320516884326935,
0.08864933997392654,
-0.013755923137068748,
0.008774982765316963,
0.07806988060474396,
0.02856220118701458,
0.07317875325679779,
0.08475907146930695,
-0.019097542390227318,
0.015560531057417393,
0.05081143230199814,
0.05529959499835968,
0.07889559864997864,
0.13392439484596252,
0.033511124551296234,
-0.17502135038375854,
0.03880435973405838,
-0.006619702093303204,
-0.04705941677093506,
0.0540950633585453,
0.08075754344463348,
-0.026997581124305725,
0.05565928667783737,
0.02342212200164795,
-0.042195528745651245,
-0.010278076864778996,
-0.01531115174293518,
-0.013959983363747597,
0.028185654431581497,
0.0655902624130249,
0.0375203974545002,
0.05222390592098236,
0.03729080408811569,
-0.060152024030685425,
0.0264256801456213,
0.03733232989907265,
0.02737855166196823,
0.04983152821660042,
-0.005267640110105276,
0.0718914344906807,
0.0827551931142807,
0.01741916313767433,
-0.000678104639519006,
0.05467507988214493,
-0.03127512335777283,
-0.0660170167684555,
-0.01809520460665226,
0.0036833705380558968,
0.08636029809713364,
0.04206205904483795,
0.02610166184604168,
0.06607961654663086,
-0.008015727624297142,
0.05058753490447998,
0.011613811366260052,
-0.16148635745048523,
-0.06690721213817596,
0.062404878437519073,
0.08442962169647217,
0.0569191575050354,
0.002155123045668006,
0.013680992648005486,
0.026265176013112068,
0.01683887280523777,
0.011843858286738396,
-0.027865327894687653,
0.023589005693793297,
-0.04004316031932831,
-0.07095954567193985,
-0.00460395123809576,
0.1947258710861206,
-0.030574709177017212,
-0.05019579827785492,
-0.09033536911010742,
-0.039188090711832047,
-0.042273618280887604,
-0.002720605581998825,
-0.032969266176223755,
0.00613204063847661,
0.04007822275161743,
0.07039010524749756,
-0.1094236820936203,
-0.05318272113800049,
-0.06254825741052628,
0.024131473153829575,
0.06539618223905563,
0.04379420727491379,
0.044534534215927124,
-0.022577140480279922,
0.09555362164974213,
0.0003320049727335572,
-0.06408091634511948,
-0.050497204065322876,
-0.09585270285606384,
-0.022834718227386475,
0.025044184178113937,
-0.006222237832844257,
-0.07225942611694336,
0.05162206292152405,
0.1008884459733963,
0.012033037841320038,
0.06644408404827118,
0.044551968574523926,
0.01035541482269764,
0.02613529935479164,
0.0914163738489151,
-0.022430136799812317,
-0.05328106880187988,
0.040454987436532974,
0.04258309677243233,
0.017639748752117157,
-0.014256222173571587,
-0.05130772292613983,
-0.028602279722690582,
-0.0637022852897644,
0.05651380121707916,
-0.010373052209615707,
0.045942068099975586,
0.012816141359508038,
-0.04965168237686157,
0.15746034681797028,
-0.06667697429656982,
-0.022725578397512436,
-0.01209181733429432,
-0.024207033216953278,
0.06882493942975998,
0.03703952953219414,
-0.01984698325395584,
-0.05623982846736908,
-0.077411949634552,
-0.020134752616286278,
-0.03742588311433792,
-0.07779772579669952,
-0.019541917368769646,
-0.02148308791220188,
-0.024879254400730133,
0.005166993010789156,
-0.07872964441776276,
-0.13707700371742249,
-0.008828200399875641,
0.0892975777387619,
-0.02237950637936592,
-0.03711850196123123,
0.039175279438495636,
-0.00383267505094409,
-0.009612007066607475,
-0.007052907254546881,
0.040157776325941086,
-0.012343964539468288,
0.017010405659675598,
0.03679246827960014,
0.029935188591480255,
-0.07729128748178482,
0.014444458298385143,
-0.02648867294192314,
0.02388831600546837,
-0.1392887830734253,
0.08625823259353638,
-0.04790148884057999,
-0.004273050464689732,
-0.031737878918647766,
0.0025148564018309116,
-0.05462455004453659,
0.016441501677036285,
0.03486201912164688,
0.05613916367292404,
-0.10910998284816742,
0.006120170466601849,
0.07123933732509613,
-0.11295615881681442,
-0.03491534665226936,
0.03950294852256775,
0.027485735714435577,
0.07459858059883118,
0.0469941720366478,
0.030661625787615776,
0.21312116086483002,
-0.10133920609951019,
0.007625811733305454,
0.06464500725269318,
-0.046497590839862823,
-0.014319310896098614,
0.040598466992378235,
0.06465591490268707,
-0.054688211530447006,
0.06389933079481125,
-0.1082274541258812,
0.05982356145977974,
0.002613141667097807,
-0.018976274877786636,
-0.01540466770529747,
-0.07414175570011139,
-0.042712390422821045,
-0.014181076548993587,
0.0078002456575632095,
0.04956934228539467,
-0.04147179424762726,
-0.03137790039181709,
0.12751060724258423,
-0.03411904722452164,
-0.022031156346201897,
-0.11512210220098495,
0.02640681341290474,
0.007494086399674416,
0.030972659587860107,
-0.06842799484729767,
-0.027528466656804085,
0.05457685515284538,
-0.03468863293528557,
0.0249976497143507,
0.05601680651307106,
0.028365347534418106,
0.057703468948602676,
0.0045438166707754135,
0.00515707116574049,
0.003433451522141695,
-0.018298069015145302,
-0.024548228830099106,
-0.04322701320052147,
-0.025725260376930237,
-0.039412036538124084,
0.08725360780954361,
-0.06443426012992859,
0.06481927633285522,
-0.03572540730237961,
0.006938271224498749,
-0.0010512960143387318,
-0.024852806702256203,
0.03435387462377548,
-0.05131630226969719,
-0.03384096175432205,
-0.041433826088905334,
0.003148056101053953,
0.06189277768135071,
-0.02278796210885048,
0.0838904082775116,
-0.041233837604522705,
0.047425802797079086,
0.05976877361536026,
0.08692524582147598,
-0.027429644018411636,
-0.04592543840408325,
-0.022865062579512596,
-0.010623798705637455,
0.009516945108771324,
-0.030938735231757164,
0.13298112154006958,
0.014669068157672882,
0.08212056756019592,
-0.04598285257816315,
0.01964128389954567,
0.008198092691600323,
-0.002081450307741761,
0.0367940329015255,
-0.017739057540893555,
0.09438411146402359,
-0.031071413308382034,
-0.005866842344403267,
0.03398944064974785,
-0.027787640690803528,
0.12477842718362808,
0.006796770729124546,
-0.029618805274367332,
-0.026781819760799408,
0.036957014352083206,
0.005057459697127342,
0.08556064963340759,
-0.09417732059955597,
0.0011768894037231803,
0.008218133822083473,
-0.028917528688907623,
0.07574845850467682,
-0.05850674957036972,
0.03394132852554321,
-0.0010089526185765862,
-0.02830253541469574,
0.0733250305056572,
0.0030063835438340902,
-0.057415738701820374,
0.026695845648646355,
-0.024858426302671432,
0.03330576792359352,
-0.0077058738097548485,
-0.0017068565357476473,
-0.03526655584573746,
0.06542648375034332,
-0.048676714301109314,
-0.12056775391101837,
-0.09634685516357422,
0.004689395893365145,
-0.0232196357101202,
0.027966946363449097,
0.009482135064899921,
-0.0478178896009922,
-0.038086943328380585,
-0.04829685762524605,
-0.04761223495006561,
0.002869181102141738,
-0.02081764116883278,
0.036045223474502563,
0.04372350126504898,
0.038913544267416,
-0.09430284798145294,
0.008252283558249474,
0.013801803812384605,
-0.06554822623729706,
-0.0018556955037638545,
-0.030701693147420883,
0.03107057884335518,
0.10469521582126617,
0.008406104519963264,
0.00878933910280466,
0.012334243394434452,
0.1123477891087532,
-0.03620392456650734,
0.06734786182641983,
0.1842508763074875,
0.021153172478079796,
0.011210004799067974,
0.03738773986697197,
0.0021342902909964323,
-0.049383532255887985,
0.01627766154706478,
-0.03035227581858635,
-0.04563596844673157,
-0.13951237499713898,
-0.029252152889966965,
-0.04103581979870796,
0.0017471401952207088,
0.019686589017510414,
0.03872847557067871,
-0.003534710966050625,
0.09744106233119965,
-0.02292640507221222,
0.04856478422880173,
0.07082650065422058,
0.04444582015275955,
0.01625729724764824,
-0.045431628823280334,
0.05750800669193268,
-0.04287518560886383,
0.03976760432124138,
0.07196776568889618,
0.07731573283672333,
0.08639425039291382,
-0.03995908051729202,
0.04246220737695694,
0.032729338854551315,
0.06883363425731659,
0.05249829590320587,
0.07217121869325638,
0.00045350799337029457,
0.01871664635837078,
-0.02980988658964634,
-0.04171472042798996,
-0.08655203878879547,
0.03556238114833832,
-0.005381215363740921,
-0.05660988762974739,
-0.01431288756430149,
0.13682717084884644,
0.008906703442335129,
0.1258237659931183,
-0.014811479486525059,
-0.1708761751651764,
-0.06044436991214752,
0.012035559862852097,
-0.010105898603796959,
-0.0833984985947609,
0.028312336653470993,
0.0748693197965622,
-0.04860524833202362,
0.006125779822468758,
-0.021037383005023003,
0.06782335042953491,
-0.04343242943286896,
0.011909420602023602,
0.048664260655641556,
0.10722233355045319,
0.03202913701534271,
0.03824113309383392,
-0.10407322645187378,
0.03193887695670128,
0.013056975789368153,
0.07869914174079895,
-0.06719443202018738,
0.05270584672689438,
0.058965932577848434,
0.04173365980386734,
0.06023979187011719,
-0.017529523000121117,
-0.07779558002948761,
-0.030590172857046127,
-0.1224176436662674,
0.06566153466701508,
0.02683316543698311,
0.014145273715257645,
0.03468307852745056,
-0.04034458473324776,
0.012119567021727562,
-0.0353768914937973,
-0.006379281170666218,
-0.11831216514110565,
-0.1213495135307312,
0.03888503462076187,
0.05675075203180313,
-0.005384240299463272,
-0.036620110273361206,
0.01843293011188507,
0.04923776164650917,
0.14557528495788574,
-0.09255094826221466,
-0.058415088802576065,
-0.06400039047002792,
-0.035445909947156906,
0.09614018350839615,
-0.018483199179172516,
0.010376283898949623,
-0.04079688340425491,
0.0902426689863205,
0.006665271706879139,
-0.05632583051919937,
-0.013461843132972717,
-0.10063840448856354,
-0.0483492836356163,
-0.02805083990097046,
0.026998985558748245,
0.003256097435951233,
0.014257719740271568,
0.024562399834394455,
-0.06650126725435257,
-0.0007912227883934975,
-0.09732884913682938,
-0.04665551707148552,
0.1068134680390358,
-0.042013220489025116,
0.03068508207798004,
-0.11176417768001556,
-0.05211317166686058,
-0.020858455449342728,
0.008084147237241268,
0.0032730852253735065,
0.08393057435750961,
-0.034369006752967834,
0.05749097466468811,
0.10769854485988617,
-0.030789613723754883,
-0.16640809178352356,
-0.05310593917965889,
0.08450475335121155,
0.02076004631817341,
-0.025243638083338737,
-0.17120890319347382,
0.058409105986356735,
0.053062036633491516,
-0.02054203301668167,
0.05958849936723709,
-0.13319668173789978,
-0.06438033282756805,
0.04070926457643509,
0.005282554775476456,
0.08001033961772919,
-0.09109532833099365,
-0.021421868354082108,
-0.024654140695929527,
-0.05071781948208809,
0.028997667133808136,
-0.05148795619606972,
0.08576373755931854,
-0.011699354276061058,
-0.05823444947600365,
0.03157057985663414,
-0.0416845828294754,
0.10030241310596466,
-0.06175072491168976,
0.013384964317083359,
-0.03639916703104973,
0.04738503694534302,
0.05215011164546013,
-0.0558018758893013,
0.12568965554237366,
-0.06410360336303711,
0.0449371263384819,
-0.08780062198638916,
-0.028445135802030563,
-0.0004527531564235687,
0.006503199227154255,
0.013842694461345673,
-0.029834523797035217,
-0.027627229690551758,
0.016823092475533485,
-0.001023089629597962,
0.03379601612687111,
-0.02679935283958912,
-0.00908273458480835,
-0.021592004224658012,
0.16911205649375916,
0.07429522275924683,
-0.0717533528804779,
-0.010191839188337326,
-0.012503057718276978,
0.014621286652982235,
0.07781543582677841,
-0.05094008892774582,
0.03068608045578003,
0.0327288955450058,
-0.034147944301366806,
0.023193541914224625,
0.017462966963648796,
-0.025134962052106857,
0.02575208805501461,
0.047711264342069626,
-0.04817718267440796,
-0.03184635937213898,
-0.04900442436337471,
0.053505413234233856,
-0.023491833359003067,
0.03940754011273384,
0.10442662239074707,
-0.025250522419810295,
0.00844292901456356,
0.016948256641626358,
0.032163675874471664,
-0.014480914920568466,
0.11187034845352173,
-0.0008220621384680271,
0.001641426351852715,
-0.040300507098436356,
0.048226047307252884,
0.03128281235694885,
0.018588589504361153,
0.022222701460123062,
0.04079747945070267,
-0.06925006955862045,
-0.06100677698850632,
-0.09215206652879715,
-0.015359051525592804,
0.015621347352862358,
-0.05632329359650612,
-0.014041869901120663,
-0.03476547449827194,
-0.019926542416214943,
-0.016305595636367798,
0.007310332730412483,
0.008058086037635803,
-0.020283225923776627,
0.025799337774515152,
-0.07557022571563721,
0.030932592228055,
0.0005930897314101458,
0.035341776907444,
-0.0908442884683609,
0.031415704637765884,
0.051428306847810745,
0.033543381839990616,
-0.03868996351957321,
-0.024384019896388054,
-0.035437747836112976,
0.004892154596745968,
-0.12812574207782745,
0.04041164368391037,
-0.04394005984067917,
-0.024643221870064735,
0.025169184431433678,
-0.009662196040153503,
-0.03549414128065109,
0.043775372207164764,
-0.004194849636405706,
0.014073116704821587,
-0.042796023190021515,
0.03390877693891525,
-0.012577135115861893,
0.03205350786447525,
0.0026153691578656435,
-0.07321186363697052,
0.052187614142894745,
0.016781440004706383,
-0.02993115410208702,
-0.012626787647604942,
-0.04110727459192276,
-0.015382192097604275,
0.0038374904543161392,
0.05246385931968689,
-0.02976207621395588,
-0.004439336247742176,
0.03155537322163582,
0.02612709067761898,
-0.005536410491913557,
-0.03602486848831177,
0.10001476854085922,
-0.031727541238069534,
0.0492696575820446,
0.017184816300868988,
-0.0006542392075061798,
-0.02169555053114891,
0.003685850417241454,
0.07085587084293365,
0.06861061602830887,
0.04993700981140137,
-0.02127763070166111,
0.00396906491369009,
-0.09615585952997208,
0.023212233558297157,
0.0042485729791224,
0.015602169558405876,
-0.014145166613161564,
-0.033878620713949203,
0.03390569984912872,
-0.0010301790898665786,
0.17733153700828552,
0.004810132551938295,
-0.004435096867382526,
-0.005721437744796276,
-0.056209951639175415,
-0.003954371437430382,
-0.00028704878059215844,
0.02866174280643463,
0.05969276651740074,
0.04621398448944092,
-0.016204044222831726,
0.0071555450558662415,
-0.0033531635999679565,
-0.09284184128046036,
0.03572619706392288,
0.029391810297966003,
-0.0024325568228960037,
0.09483860433101654,
0.013525133952498436,
-0.027976319193840027,
-0.03884736821055412,
0.0552828274667263,
-0.07330651581287384,
0.05895116925239563,
-0.048565998673439026,
0.034523915499448776,
0.10038460046052933,
-0.054101958870887756,
0.014889845624566078,
0.0017990642227232456,
-0.04105709493160248,
-0.10243893414735794,
-0.07572958618402481,
-0.03710642457008362,
-0.08904482424259186,
-0.019314363598823547,
-0.05067919194698334,
-0.017222166061401367,
-0.05741746351122856,
0.014741592109203339,
-0.005898754112422466,
0.08193471282720566,
0.003473693737760186,
-0.03825787827372551,
0.012255800887942314,
0.0011946512386202812,
0.015757113695144653,
-0.009667569771409035,
-0.03409869223833084,
0.012269620783627033,
0.04250378534197807,
0.05834171175956726,
0.04190920293331146,
0.02045203000307083,
0.015886841341853142,
0.02579655684530735,
-0.012481234967708588,
-0.028647037222981453,
-0.0081242136657238,
-0.029504362493753433,
0.0554092712700367,
0.008360481821000576,
-0.04904722422361374,
0.012019671499729156,
0.09899382293224335,
-0.0149290282279253,
-0.0668165385723114,
-0.13288292288780212,
0.14009329676628113,
-0.010315557941794395,
-0.00037270598113536835,
-0.007202584762126207,
0.0025448547676205635,
-0.02856174111366272,
0.1788823902606964,
0.12456651777029037,
-0.0831371322274208,
-0.022003378719091415,
0.02485952153801918,
-0.00218014232814312,
-0.04492395371198654,
0.13763512670993805,
0.05384470522403717,
0.1776137351989746,
0.02534066140651703,
0.0042562708258628845,
0.004672480747103691,
-0.002485183533281088,
-0.06314113736152649,
-0.009015313349664211,
-0.04827521741390228,
0.008480940014123917,
0.008537642657756805,
0.03420848399400711,
-0.0393252819776535,
-0.10461098700761795,
-0.006278656888753176,
0.022045664489269257,
-0.05470506101846695,
-0.008027279749512672,
-0.00717634754255414,
-0.0052752140909433365,
0.035218264907598495,
-0.022259997203946114,
0.01446650642901659,
0.10088604688644409,
-0.030799653381109238,
-0.03259093686938286,
-0.0048542264848947525,
0.06629076600074768,
-0.031180165708065033,
0.18357892334461212,
0.005674783140420914,
0.013961950317025185,
0.026710018515586853,
-0.016733938828110695,
-0.1124507486820221,
0.06008249893784523,
-0.01891210675239563,
-0.10216343402862549,
0.011280958540737629,
0.033476367592811584,
-0.04244779795408249,
0.08954140543937683,
0.024928580969572067,
-0.017046570777893066,
-0.00028791045770049095,
0.03454514220356941,
-0.009553528390824795,
-0.047861285507678986,
0.03167346864938736,
-0.06865792721509933,
0.1387280970811844,
0.09889830648899078,
-0.008151094429194927,
-0.010654127225279808,
-0.04260719195008278,
0.08044053614139557,
0.008871886879205704,
0.04032620042562485,
-0.021607769653201103,
-0.12265051156282425,
-0.00923824217170477,
-0.0025760773569345474,
0.03059500642120838,
-0.08021560311317444,
-0.02103385142982006,
-0.014285175129771233,
-0.005159138701856136,
-0.018425947055220604,
0.05298589915037155,
0.06793031096458435,
0.05738428980112076,
-0.01793704926967621,
-0.014933864586055279,
-0.012195191346108913,
0.03387029469013214,
-0.13344058394432068,
-0.07212861627340317
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# roberta-large-squad-model1
This model is a fine-tuned version of [roberta-large](https://huggingface.co/roberta-large) on the squad dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 16
- seed: 83
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| {"license": "mit", "tags": ["generated_from_trainer"], "datasets": ["varun-v-rao/squad"], "base_model": "roberta-large", "model-index": [{"name": "roberta-large-squad-model1", "results": []}]} | question-answering | varun-v-rao/roberta-large-squad-model1 | [
"transformers",
"tensorboard",
"safetensors",
"roberta",
"question-answering",
"generated_from_trainer",
"dataset:varun-v-rao/squad",
"base_model:roberta-large",
"license:mit",
"endpoints_compatible",
"region:us"
] | 2024-02-08T08:29:10+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #roberta #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-roberta-large #license-mit #endpoints_compatible #region-us
|
# roberta-large-squad-model1
This model is a fine-tuned version of roberta-large on the squad dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 16
- seed: 83
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| [
"# roberta-large-squad-model1\n\nThis model is a fine-tuned version of roberta-large on the squad dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 83\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #roberta #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-roberta-large #license-mit #endpoints_compatible #region-us \n",
"# roberta-large-squad-model1\n\nThis model is a fine-tuned version of roberta-large on the squad dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 83\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
70,
34,
6,
12,
8,
3,
90,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #roberta #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-roberta-large #license-mit #endpoints_compatible #region-us \n# roberta-large-squad-model1\n\nThis model is a fine-tuned version of roberta-large on the squad dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 83\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3### Training results### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
-0.0637679249048233,
0.06048141419887543,
-0.0020348625257611275,
0.07675410807132721,
0.17139829695224762,
0.01945500075817108,
0.13432429730892181,
0.08449210226535797,
-0.12447147071361542,
0.05579020828008652,
0.07491812855005264,
0.07167085260152817,
0.02059265598654747,
0.11354649811983109,
-0.0393010713160038,
-0.2305970937013626,
0.004248541779816151,
-0.013202819041907787,
-0.10484462976455688,
0.10241935402154922,
0.09873408824205399,
-0.12535110116004944,
0.05878431722521782,
-0.009872988797724247,
-0.19624318182468414,
0.04400775954127312,
-0.005292105488479137,
-0.027827885001897812,
0.10868871957063675,
0.015709288418293,
0.13552533090114594,
0.006983448285609484,
0.13973818719387054,
-0.22157275676727295,
0.014904133975505829,
0.08204883337020874,
0.027777839452028275,
0.06332731246948242,
0.03854108601808548,
-0.004585756454616785,
0.10923676937818527,
-0.1450890302658081,
0.09781220555305481,
0.024685410782694817,
-0.08374745398759842,
-0.1771799474954605,
-0.08305242657661438,
0.03979884833097458,
0.09349049627780914,
0.09815889596939087,
-0.013598131015896797,
0.16721278429031372,
-0.1282738298177719,
0.08182687312364578,
0.20424365997314453,
-0.2777280807495117,
-0.09293796122074127,
0.10002999007701874,
0.055979300290346146,
0.09482908248901367,
-0.11891916394233704,
-0.012525889091193676,
0.05175163224339485,
0.025706822052598,
0.10147318989038467,
-0.029027700424194336,
-0.0853506326675415,
0.012883114628493786,
-0.15216265618801117,
0.017948489636182785,
0.14127738773822784,
0.057920053601264954,
-0.03051222302019596,
-0.05495232343673706,
-0.0462135411798954,
-0.03877570852637291,
-0.026099922135472298,
-0.05660919472575188,
0.050336804240942,
-0.05574708431959152,
-0.11030557006597519,
-0.03839307650923729,
-0.08590047061443329,
-0.0737432986497879,
-0.014531160704791546,
0.11112257838249207,
0.049961384385824203,
0.009948359802365303,
-0.05448532849550247,
0.08766660839319229,
-0.022265465930104256,
-0.09593310952186584,
0.001701662316918373,
0.004490464460104704,
-0.08742234110832214,
-0.07523825019598007,
-0.038617465645074844,
-0.04809524118900299,
0.03956828638911247,
0.16609245538711548,
-0.07044409215450287,
0.0521649606525898,
0.02079407498240471,
0.008514391258358955,
-0.039516840130090714,
0.13046525418758392,
-0.06166519224643707,
-0.06227092444896698,
0.004236666020005941,
0.08008377999067307,
0.007540220394730568,
0.017135227099061012,
-0.07969237118959427,
0.0038226195611059666,
0.09079137444496155,
0.05478081479668617,
-0.06589976698160172,
0.03394685313105583,
-0.008249753154814243,
-0.005318718496710062,
-0.03329651802778244,
-0.10797587782144547,
0.03835781663656235,
-0.0020344755612313747,
-0.05530969426035881,
-0.02501177415251732,
0.005538006313145161,
0.014888317324221134,
0.008598155342042446,
0.11109539866447449,
-0.09331762045621872,
0.01240761112421751,
-0.10452568531036377,
-0.09643247723579407,
0.0017110987100750208,
-0.07328543812036514,
0.0017114953370764852,
-0.08025563508272171,
-0.16247403621673584,
-0.04912114515900612,
0.03365691751241684,
-0.04040851444005966,
-0.01685640960931778,
-0.05236375704407692,
-0.07751556485891342,
-0.01587343029677868,
0.004207214806228876,
0.15154510736465454,
-0.04107236489653587,
0.07790088653564453,
0.019696028903126717,
0.04160788655281067,
0.005041320342570543,
0.020890485495328903,
-0.09198177605867386,
0.019156580790877342,
-0.1558554768562317,
0.037772487848997116,
-0.07884923368692398,
0.0630635917186737,
-0.10589438676834106,
-0.100817009806633,
-0.014920747838914394,
0.006056969985365868,
0.05371463671326637,
0.08840479701757431,
-0.15815065801143646,
-0.03473145142197609,
0.19430235028266907,
-0.08164931833744049,
-0.08340386301279068,
0.10911449790000916,
-0.06561553478240967,
0.052787937223911285,
0.07147445529699326,
0.16752514243125916,
0.06376536935567856,
-0.11678893119096756,
0.00021703547099605203,
-0.02687249332666397,
0.041013848036527634,
-0.005614893510937691,
0.04975210502743721,
-0.0012866755714640021,
0.01751994714140892,
-0.000017607717381906696,
-0.06327968090772629,
0.017641441896557808,
-0.1074364185333252,
-0.07286351174116135,
-0.04335376247763634,
-0.10356947034597397,
-0.0009102601907216012,
0.048483237624168396,
0.045135777443647385,
-0.09435488283634186,
-0.09263837337493896,
0.13234956562519073,
0.11440830677747726,
-0.05781753733754158,
0.0002612823445815593,
-0.0725824162364006,
0.03240474313497543,
-0.061239853501319885,
-0.02816605567932129,
-0.18077056109905243,
-0.11481433361768723,
0.0017730252584442496,
-0.016674470156431198,
0.044529084116220474,
0.05167754739522934,
0.07112592458724976,
0.06533171236515045,
-0.060500193387269974,
-0.006417426280677319,
-0.06346052885055542,
0.006064029410481453,
-0.09429934620857239,
-0.216837540268898,
-0.025993941351771355,
-0.027909867465496063,
0.16246289014816284,
-0.2725488841533661,
0.028578078374266624,
-0.05196400731801987,
0.11803010106086731,
0.034756213426589966,
-0.035168007016181946,
-0.028003167361021042,
0.06039124354720116,
-0.008095004595816135,
-0.07482816278934479,
0.041196852922439575,
-0.015906058251857758,
-0.06446974724531174,
-0.08294601738452911,
-0.14336785674095154,
0.08206277340650558,
0.089591845870018,
-0.006277989596128464,
-0.10651954263448715,
0.007293181028217077,
-0.06861009448766708,
-0.03194041922688484,
-0.08803127706050873,
0.012328729033470154,
0.12323237210512161,
-0.010293208062648773,
0.12139037251472473,
-0.058043286204338074,
-0.05440355837345123,
-0.0025881193578243256,
-0.027334904298186302,
0.029980940744280815,
0.0993540808558464,
0.11464843899011612,
-0.13334783911705017,
0.09912742674350739,
0.09557780623435974,
-0.09325043112039566,
0.143319234251976,
-0.04676549881696701,
-0.07297486811876297,
-0.025783585384488106,
0.01012913603335619,
0.001362006994895637,
0.13358978927135468,
-0.08194665610790253,
0.0018739814404398203,
0.00817881990224123,
0.010753918439149857,
0.028123771771788597,
-0.18198806047439575,
-0.03270857408642769,
0.010725516825914383,
-0.02930140495300293,
0.00838156882673502,
-0.015077073127031326,
0.030277803540229797,
0.10154569894075394,
0.009350347332656384,
-0.027469592168927193,
0.002786611905321479,
-0.0061066728085279465,
-0.08116274327039719,
0.20584610104560852,
-0.07465584576129913,
-0.0821484848856926,
-0.0942186638712883,
0.02742120996117592,
-0.03337593376636505,
-0.026884861290454865,
0.02922859974205494,
-0.11541232466697693,
-0.04385094717144966,
-0.09257343411445618,
0.015844907611608505,
-0.011495835147798061,
-0.005241473205387592,
0.017219042405486107,
0.030244912952184677,
0.08006007224321365,
-0.12587140500545502,
0.01142045482993126,
-0.059866610914468765,
-0.13475827872753143,
0.01708567701280117,
0.052363406866788864,
0.1298079937696457,
0.11936894804239273,
-0.02140049822628498,
0.008748303167521954,
-0.033448465168476105,
0.20971554517745972,
-0.06629102677106857,
-0.003926669247448444,
0.10580170899629593,
0.013798223808407784,
0.03548724949359894,
0.13108594715595245,
0.0479530394077301,
-0.10534605383872986,
0.04485022649168968,
0.09678596258163452,
-0.029652703553438187,
-0.23833321034908295,
-0.031371332705020905,
-0.038085177540779114,
-0.0795687586069107,
0.06090159714221954,
0.0410047248005867,
0.009288541041314602,
0.051320742815732956,
0.008042797446250916,
0.03559785336256027,
-0.013739737682044506,
0.08197683095932007,
0.08801943808794022,
0.03690426051616669,
0.12312410771846771,
-0.04967915639281273,
-0.06569138914346695,
0.04570836201310158,
-0.022572804242372513,
0.31193622946739197,
-0.0023162723518908024,
0.04638084024190903,
0.08345504105091095,
0.12166322022676468,
-0.017956282943487167,
0.028743170201778412,
0.009295370429754257,
-0.03934117779135704,
0.0021820319816470146,
-0.052357181906700134,
0.008129695430397987,
0.019702177494764328,
-0.04195287823677063,
0.06273713707923889,
-0.07193166017532349,
0.04672468453645706,
0.05177740007638931,
0.2258027344942093,
0.026918558403849602,
-0.2781989872455597,
-0.07410989701747894,
0.015714135020971298,
-0.03598323464393616,
-0.0185773316770792,
0.017787504941225052,
0.13931432366371155,
-0.10609214752912521,
0.02705913782119751,
-0.05801532045006752,
0.08513305336236954,
-0.008732453919947147,
0.011606899090111256,
0.03288509324193001,
0.1589563935995102,
-0.019131287932395935,
0.07698146253824234,
-0.22475965321063995,
0.24353311955928802,
0.010047630406916142,
0.12748944759368896,
-0.03278028592467308,
0.00043611673754639924,
0.0177006833255291,
0.07502555102109909,
0.08543416857719421,
-0.0071572959423065186,
-0.035522203892469406,
-0.1694675236940384,
-0.0308417696505785,
0.061850205063819885,
0.12362013012170792,
-0.0185234397649765,
0.10648444294929504,
-0.04961678385734558,
0.00789907481521368,
0.06431064754724503,
-0.06734037399291992,
-0.16725340485572815,
-0.10060332715511322,
-0.036987096071243286,
-0.0018137793522328138,
-0.07825274020433426,
-0.09023364633321762,
-0.0905454009771347,
-0.05830913782119751,
0.16390247642993927,
0.007027298212051392,
-0.008672825060784817,
-0.12087811529636383,
0.11963450163602829,
0.08919849246740341,
-0.05922014266252518,
0.006865548901259899,
0.01897973008453846,
0.0858386978507042,
0.03418687731027603,
-0.05268378555774689,
0.06954411417245865,
-0.07681982964277267,
-0.1525668054819107,
-0.06262712180614471,
0.1028299555182457,
0.06834801286458969,
0.044596485793590546,
0.0005874195485375822,
0.027074331417679787,
0.015733620151877403,
-0.09615715593099594,
0.0023727649822831154,
0.0676547959446907,
0.06330197304487228,
0.062239110469818115,
-0.08893098682165146,
-0.014183836057782173,
-0.03811654448509216,
-0.029948174953460693,
0.12757079303264618,
0.22752870619297028,
-0.08470267802476883,
0.030588775873184204,
0.060667626559734344,
-0.07489867508411407,
-0.18081022799015045,
0.10594646632671356,
0.07453153282403946,
0.003772685769945383,
0.07306596636772156,
-0.146053284406662,
0.16629858314990997,
0.1251004934310913,
-0.013251347467303276,
0.03465365245938301,
-0.3092174828052521,
-0.12851689755916595,
0.07404396682977676,
0.1551697999238968,
0.07268938422203064,
-0.16375984251499176,
-0.01690095290541649,
-0.02209869772195816,
-0.12605538964271545,
0.09987764805555344,
-0.16794288158416748,
0.10013370215892792,
0.0041718631982803345,
0.07670361548662186,
0.010712242685258389,
-0.05062194913625717,
0.11949034035205841,
0.016916364431381226,
0.11201798170804977,
-0.056703537702560425,
-0.011810554191470146,
0.11040069907903671,
-0.04762199893593788,
0.01627139002084732,
-0.045536842197179794,
0.05802400782704353,
-0.09113116562366486,
-0.031906407326459885,
-0.0770440474152565,
0.0545109324157238,
-0.055162131786346436,
-0.06682243198156357,
-0.06245694309473038,
0.0665862038731575,
0.04476320743560791,
-0.02243771031498909,
0.06164724379777908,
-0.00871226191520691,
0.1436610221862793,
0.028713945299386978,
0.10560567677021027,
0.0005851754685863853,
-0.05227431282401085,
0.005306599661707878,
-0.017876284196972847,
0.052071746438741684,
-0.12902097404003143,
0.02232339233160019,
0.11718159914016724,
0.04415230453014374,
0.14562632143497467,
0.05394809693098068,
-0.047779861837625504,
0.015215244144201279,
0.05619780346751213,
-0.10376646369695663,
-0.18842868506908417,
0.02960083819925785,
-0.06774076819419861,
-0.1334599107503891,
0.04739333316683769,
0.09404352307319641,
-0.06100023537874222,
-0.007369932718575001,
-0.02169531024992466,
0.017879551276564598,
-0.03987057879567146,
0.17981630563735962,
0.0640096664428711,
0.057160452008247375,
-0.07791311293840408,
0.07804717123508453,
0.061807144433259964,
-0.04517628252506256,
0.02464412897825241,
0.0803600400686264,
-0.08192752301692963,
-0.021301934495568275,
0.0891268253326416,
0.22436821460723877,
-0.049689289182424545,
-0.03758343681693077,
-0.1173919066786766,
-0.10501397401094437,
0.03365137428045273,
0.15400639176368713,
0.05956720933318138,
-0.03694460168480873,
-0.015923628583550453,
0.05200346186757088,
-0.16183006763458252,
0.09407412260770798,
0.03945237025618553,
0.07784730941057205,
-0.1328531950712204,
0.11440379917621613,
0.00952839758247137,
0.0067662023939192295,
-0.016207292675971985,
0.0387064553797245,
-0.13161610066890717,
-0.01234298013150692,
-0.13311266899108887,
-0.02607416920363903,
-0.021830987185239792,
0.004434108734130859,
0.001096248161047697,
-0.05562781170010567,
-0.06716573983430862,
0.03918957710266113,
-0.07261990010738373,
-0.0393458716571331,
0.03116193786263466,
0.05658550560474396,
-0.1471695452928543,
-0.0029094030614942312,
0.017090341076254845,
-0.07325499504804611,
0.05875342711806297,
0.031570449471473694,
0.039581455290317535,
0.0441170409321785,
-0.1834612637758255,
-0.008904739283025265,
0.03219933807849884,
0.008822552859783173,
0.07257779687643051,
-0.07466999441385269,
-0.009182603098452091,
-0.01573760248720646,
0.08903341740369797,
0.02613246440887451,
0.030011840164661407,
-0.1088147908449173,
0.030826283618807793,
-0.08370143920183182,
-0.06796539574861526,
-0.05346433073282242,
0.02352597750723362,
0.07596638798713684,
0.029681380838155746,
0.1814735233783722,
-0.10074089467525482,
0.03733329847455025,
-0.22197555005550385,
-0.032255612313747406,
-0.0014086572919040918,
-0.035034701228141785,
-0.08020996302366257,
-0.04083830118179321,
0.06560046970844269,
-0.06249186769127846,
0.11749440431594849,
0.011350894346833229,
0.07184107601642609,
0.05058027803897858,
-0.05206527188420296,
0.003875725669786334,
0.01674506440758705,
0.18532204627990723,
0.04941215366125107,
-0.02357914298772812,
0.05172795429825783,
0.0018762183608487248,
0.06742450594902039,
0.018409911543130875,
0.21351882815361023,
0.19742219150066376,
-0.06241742894053459,
0.04243826866149902,
0.07000678032636642,
-0.08211080729961395,
-0.10061166435480118,
0.13064837455749512,
-0.03043983317911625,
0.0635499507188797,
-0.04336075857281685,
0.1772441416978836,
0.1590132862329483,
-0.16645720601081848,
0.024375373497605324,
-0.07190193235874176,
-0.09645920246839523,
-0.11553840339183807,
-0.008413562551140785,
-0.10048727691173553,
-0.12135361135005951,
0.0383211150765419,
-0.1266583949327469,
0.010648264549672604,
0.10826802998781204,
0.022974759340286255,
0.011551203206181526,
0.17465220391750336,
0.0025491537526249886,
0.04376503452658653,
0.026868222281336784,
-0.0021096430718898773,
-0.01707765832543373,
-0.05307962745428085,
-0.045859284698963165,
0.03777660056948662,
-0.021645784378051758,
0.060742419213056564,
-0.05070222169160843,
-0.033573783934116364,
0.03382755443453789,
-0.026671184226870537,
-0.06744984537363052,
0.02812059409916401,
0.03611025586724281,
0.03882720321416855,
0.03322266787290573,
0.04656968638300896,
-0.01626545377075672,
-0.03425198048353195,
0.2745743691921234,
-0.06287752836942673,
-0.12228616327047348,
-0.12136800587177277,
0.22010406851768494,
0.0284440740942955,
0.0015088565414771438,
0.03321676701307297,
-0.11510290205478668,
0.024318739771842957,
0.18536187708377838,
0.1860838532447815,
-0.06514247506856918,
-0.0065948013216257095,
-0.011026911437511444,
-0.021280253306031227,
-0.06473614275455475,
0.11351223289966583,
0.10608681291341782,
0.0521523617208004,
-0.05339915305376053,
-0.023039719089865685,
-0.019458752125501633,
-0.017849422991275787,
-0.06643008440732956,
0.031084582209587097,
0.05755646154284477,
0.015309078618884087,
-0.03148091584444046,
0.09048087894916534,
-0.006700634490698576,
-0.19803625345230103,
0.07272376865148544,
-0.15240271389484406,
-0.16721943020820618,
-0.030524566769599915,
0.0980539470911026,
-0.03160231560468674,
0.0515747107565403,
-0.03589045628905296,
-0.020624633878469467,
0.07971547544002533,
-0.018075089901685715,
-0.05075007304549217,
-0.1052158772945404,
0.08546394854784012,
-0.07660026103258133,
0.21677543222904205,
-0.02671978995203972,
0.0616280771791935,
0.13126614689826965,
-0.0022941711358726025,
-0.09889823198318481,
0.07430129498243332,
0.055249929428100586,
-0.08529496192932129,
0.02208465151488781,
0.13561148941516876,
-0.05069142207503319,
0.1000942587852478,
0.05542796477675438,
-0.14257057011127472,
0.013866260647773743,
-0.060214564204216,
-0.045402880758047104,
-0.0887414962053299,
0.010571417398750782,
-0.07878521829843521,
0.14259599149227142,
0.20607811212539673,
-0.02484050765633583,
0.039399806410074234,
-0.07863815128803253,
0.03884773701429367,
0.06596435606479645,
0.10366103798151016,
-0.037208832800388336,
-0.2209898680448532,
0.030970066785812378,
0.004751273430883884,
0.0007169866585172713,
-0.29124921560287476,
-0.08258510380983353,
0.021827874705195427,
-0.03950433060526848,
-0.06882144510746002,
0.10219112783670425,
0.12549902498722076,
0.06219564378261566,
-0.045490335673093796,
-0.17936471104621887,
-0.05722680687904358,
0.16748693585395813,
-0.1421896070241928,
-0.0597928948700428
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | mertllc/mms-tts-tur-fifties_female | [
"transformers",
"safetensors",
"vits",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-08T08:29:32+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #vits #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #vits #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
34,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #vits #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.054659612476825714,
0.21414990723133087,
-0.0031807427294552326,
0.026865221560001373,
0.1250888854265213,
0.00032571866177022457,
0.04081440716981888,
0.12862813472747803,
-0.02167222462594509,
0.11129128932952881,
0.03218022361397743,
0.09727001935243607,
0.10339263826608658,
0.16586677730083466,
0.03691011667251587,
-0.21517004072666168,
0.009132993407547474,
-0.09292528033256531,
0.018077509477734566,
0.10867427289485931,
0.13162045180797577,
-0.10489460080862045,
0.07603627443313599,
-0.03790099918842316,
-0.017673974856734276,
-0.0003223843814339489,
-0.0923151820898056,
-0.070840984582901,
0.06550594419240952,
0.06909013539552689,
0.06122942641377449,
0.009997012093663216,
0.10145736485719681,
-0.29726552963256836,
0.01687687449157238,
0.08279260247945786,
-0.004506718832999468,
0.06148726865649223,
0.0646374449133873,
-0.08339887112379074,
0.1029256209731102,
-0.08559336513280869,
0.13652671873569489,
0.08214850723743439,
-0.06937385350465775,
-0.21391066908836365,
-0.06977995485067368,
0.0987061932682991,
0.12011658400297165,
0.06274435669183731,
-0.02326560579240322,
0.1522950381040573,
-0.06972704082727432,
0.012022249400615692,
0.1361677050590515,
-0.09713108092546463,
-0.05137801170349121,
0.049987345933914185,
0.11240657418966293,
0.10166463255882263,
-0.1353231519460678,
0.007596791721880436,
0.04457303136587143,
0.023097742348909378,
0.09194746613502502,
0.020738936960697174,
0.0916183590888977,
0.04564107209444046,
-0.13860996067523956,
-0.03957565128803253,
0.10889606922864914,
0.03478158637881279,
-0.05796414613723755,
-0.21188515424728394,
-0.0026691502425819635,
-0.026535477489233017,
-0.023307178169488907,
-0.05803702771663666,
0.045833978801965714,
-0.03317271173000336,
0.067923404276371,
-0.042256616055965424,
-0.10016343742609024,
-0.03838508576154709,
0.0836847797036171,
0.06997206062078476,
0.013808192685246468,
-0.026154542341828346,
0.03861820325255394,
0.11874474585056305,
0.037009406834840775,
-0.10824361443519592,
-0.0663856491446495,
-0.06518013030290604,
-0.09711762517690659,
-0.04532422870397568,
0.04776853322982788,
0.01869308575987816,
0.030892416834831238,
0.20719914138317108,
-0.0024066849146038294,
0.040300752967596054,
0.01544452179223299,
0.00820851232856512,
0.05608583986759186,
0.09020276367664337,
-0.057233426719903946,
-0.13989022374153137,
-0.04616677761077881,
0.08976847678422928,
-0.00493787182494998,
-0.03551584109663963,
-0.04997507110238075,
0.048379965126514435,
0.05169600620865822,
0.1267518699169159,
0.08646857738494873,
-0.012898874469101429,
-0.05273304134607315,
-0.025197435170412064,
0.22986702620983124,
-0.14503952860832214,
0.04801303148269653,
-0.016220765188336372,
-0.026413746178150177,
-0.04562145099043846,
0.037146687507629395,
0.02893291600048542,
-0.0071297562681138515,
0.09902069717645645,
-0.055000074207782745,
-0.03897455707192421,
-0.10056453198194504,
-0.03981734439730644,
0.04000834375619888,
-0.0014343701768666506,
-0.011925416998565197,
-0.07901987433433533,
-0.1033727377653122,
-0.04151687026023865,
0.0622556135058403,
-0.06062569096684456,
-0.03672588989138603,
0.014433487318456173,
-0.0646335631608963,
-0.011868113651871681,
-0.0046113538555800915,
0.10713792592287064,
-0.03111988678574562,
0.041085705161094666,
-0.03385680913925171,
0.05467362701892853,
0.10134078562259674,
0.03396330401301384,
-0.0692443996667862,
0.05283360555768013,
-0.2253323644399643,
0.0846395194530487,
-0.1103181466460228,
0.040045637637376785,
-0.1649162620306015,
-0.04362662881612778,
0.01545786950737238,
0.01223697792738676,
0.010682502761483192,
0.11813149601221085,
-0.18765069544315338,
-0.02040630392730236,
0.13456352055072784,
-0.09486816823482513,
-0.10925174504518509,
0.07470420002937317,
-0.04261988773941994,
0.14796192944049835,
0.04623936489224434,
-0.017894135788083076,
0.07337126135826111,
-0.16546636819839478,
-0.06534566730260849,
-0.015944186598062515,
-0.01140376552939415,
0.13805019855499268,
0.06177884340286255,
-0.05833873897790909,
0.06357681751251221,
0.02317901886999607,
-0.022351879626512527,
-0.04479735344648361,
-0.05049646645784378,
-0.10716529190540314,
-0.006589649710804224,
-0.0877491682767868,
0.049144841730594635,
-0.008710972033441067,
-0.07987060397863388,
-0.032660458236932755,
-0.18162156641483307,
0.03565994277596474,
0.08912748098373413,
0.006954456213861704,
-0.008257697336375713,
-0.07709750533103943,
0.012575463391840458,
-0.027584582567214966,
-0.010441360995173454,
-0.16807158291339874,
-0.045059818774461746,
0.045085642486810684,
-0.1683385670185089,
0.03666726127266884,
-0.05383622646331787,
0.057435907423496246,
0.04089425876736641,
-0.0608406662940979,
-0.012410139665007591,
-0.020455263555049896,
0.02037479542195797,
-0.03554835915565491,
-0.19715940952301025,
-0.04920884966850281,
-0.033720988780260086,
0.15323609113693237,
-0.2512565553188324,
0.03701164573431015,
0.04283377155661583,
0.1445688009262085,
-0.004499740432947874,
-0.041343484073877335,
0.021006079390645027,
-0.05124713480472565,
-0.04886976629495621,
-0.064845971763134,
-0.003489583032205701,
-0.029771825298666954,
-0.04689984768629074,
0.014419492334127426,
-0.17416127026081085,
-0.03588438406586647,
0.09719391912221909,
0.1012604832649231,
-0.15479636192321777,
-0.018018238246440887,
-0.046819429844617844,
-0.06501296907663345,
-0.08719377964735031,
-0.0634685754776001,
0.12365260720252991,
0.04887883737683296,
0.044603388756513596,
-0.07642911374568939,
-0.06516730040311813,
0.02209198847413063,
0.00037755590165033937,
-0.03342745080590248,
0.07709765434265137,
0.06420876830816269,
-0.09495706856250763,
0.07597044855356216,
0.0879693329334259,
0.07397416979074478,
0.09690815210342407,
0.017737112939357758,
-0.10766889899969101,
-0.025353191420435905,
0.025884538888931274,
0.02590569481253624,
0.14766225218772888,
-0.052133310586214066,
0.03766921907663345,
0.047928281128406525,
-0.048178963363170624,
0.018924955278635025,
-0.09172655642032623,
0.02477680705487728,
0.03108147345483303,
-0.0051895990036427975,
0.04569429159164429,
-0.04261132329702377,
0.0015583503991365433,
0.07553404569625854,
0.0439009927213192,
0.054722823202610016,
0.004550157580524683,
-0.014615098014473915,
-0.09760808199644089,
0.16303586959838867,
-0.09686829894781113,
-0.2844827473163605,
-0.15191766619682312,
0.025421515107154846,
0.038875505328178406,
-0.02202117070555687,
0.031196635216474533,
-0.0685606598854065,
-0.10619828850030899,
-0.10253546386957169,
-0.0007893215515650809,
0.021664658561348915,
-0.07999464124441147,
-0.07771245390176773,
0.07423610240221024,
0.04034431278705597,
-0.14601534605026245,
0.03843066841363907,
0.05174413323402405,
-0.05686575174331665,
-0.020990731194615364,
0.08788161724805832,
0.11919383704662323,
0.15064425766468048,
-0.01956579089164734,
-0.029653063043951988,
0.02179299294948578,
0.18913501501083374,
-0.13056331872940063,
0.10870491713285446,
0.1331699639558792,
-0.0433298796415329,
0.08741360157728195,
0.17486868798732758,
0.02946310304105282,
-0.08184187114238739,
0.04125521704554558,
0.04271497204899788,
-0.0446363128721714,
-0.2628204822540283,
-0.0587831549346447,
0.013565518893301487,
-0.07289978116750717,
0.09574431926012039,
0.09441626816987991,
0.13101495802402496,
0.03733300045132637,
-0.07704862952232361,
-0.042284153401851654,
-0.0007691121427342296,
0.11566338688135147,
-0.04729871824383736,
-0.00864650122821331,
0.08112052828073502,
-0.04204992949962616,
0.0042695761658251286,
0.101866215467453,
0.024085933342576027,
0.18680992722511292,
0.02045324817299843,
0.1325864940881729,
0.06266885250806808,
0.07362587004899979,
-0.00304698059335351,
0.021530818194150925,
0.04571235924959183,
0.016793522983789444,
-0.004352389834821224,
-0.10109587758779526,
0.004940509796142578,
0.14031140506267548,
0.044244058430194855,
0.029351718723773956,
0.0023038540966808796,
-0.025745723396539688,
0.059172797948122025,
0.16894783079624176,
-0.014623390510678291,
-0.20305828750133514,
-0.07212355732917786,
0.07476779818534851,
-0.05524183437228203,
-0.12190999835729599,
-0.03604535013437271,
0.03974858298897743,
-0.17753031849861145,
0.03411399945616722,
-0.020660564303398132,
0.09808827936649323,
-0.0960298478603363,
-0.025731271132826805,
0.017328539863228798,
0.08463997393846512,
-0.017630890011787415,
0.09686511754989624,
-0.15011048316955566,
0.12523487210273743,
0.03229980170726776,
0.0898485779762268,
-0.11468798667192459,
0.08304145932197571,
-0.009098101407289505,
0.016468055546283722,
0.18883956968784332,
-0.00914006493985653,
-0.043279051780700684,
-0.0765409916639328,
-0.09724772721529007,
-0.016675574705004692,
0.12457696348428726,
-0.11865599453449249,
0.08336363732814789,
-0.006434252485632896,
-0.05090279504656792,
0.010499227792024612,
-0.11436042934656143,
-0.17895425856113434,
-0.19684189558029175,
0.061690423637628555,
-0.10233647376298904,
0.01922602578997612,
-0.1105671152472496,
-0.06737665832042694,
-0.029828263446688652,
0.2358294576406479,
-0.14021140336990356,
-0.07348582148551941,
-0.1486395299434662,
-0.049397800117731094,
0.1688835471868515,
-0.039627790451049805,
0.07352027297019958,
-0.014237076044082642,
0.21156272292137146,
-0.0005727469106204808,
-0.0019497170578688383,
0.0662601962685585,
-0.09127254039049149,
-0.17042554914951324,
-0.0796523243188858,
0.1408538520336151,
0.1185344010591507,
0.05187511071562767,
-0.00005241960025159642,
0.008437353186309338,
-0.01933823712170124,
-0.11107131093740463,
-0.005973829887807369,
0.13854430615901947,
0.06674695014953613,
0.03547331318259239,
-0.05006469413638115,
-0.10860110819339752,
-0.06920936703681946,
-0.058358483016490936,
0.05175930634140968,
0.18184207379817963,
-0.1009909063577652,
0.17350798845291138,
0.15878215432167053,
-0.07211574912071228,
-0.21567314863204956,
0.039191193878650665,
0.04846473038196564,
-0.014512532390654087,
0.04614531248807907,
-0.1829945594072342,
0.09505120664834976,
0.015141540206968784,
-0.052736036479473114,
0.12199369817972183,
-0.15728448331356049,
-0.15639621019363403,
0.06087431684136391,
0.04970995709300041,
-0.23623821139335632,
-0.1441342532634735,
-0.08822641521692276,
-0.06784138828516006,
-0.14815589785575867,
0.07915012538433075,
-0.019972164183855057,
0.011897586286067963,
0.04091079905629158,
0.013740893453359604,
0.023185279220342636,
-0.055776987224817276,
0.18284909427165985,
-0.0035617330577224493,
0.014864614233374596,
-0.06912479549646378,
-0.058035630732774734,
0.0975092425942421,
-0.05838471278548241,
0.1184525191783905,
-0.003918026573956013,
0.013672815635800362,
-0.08212041109800339,
-0.05343952775001526,
-0.046617619693279266,
0.05752236396074295,
-0.08050531893968582,
-0.11092408001422882,
-0.04487094283103943,
0.08938708156347275,
0.07764840126037598,
-0.033286161720752716,
-0.010930746793746948,
-0.07634644955396652,
0.10063119232654572,
0.19033774733543396,
0.17030654847621918,
0.018113715574145317,
-0.07677590847015381,
0.015532949939370155,
-0.03924742713570595,
0.04019718989729881,
-0.2505480647087097,
0.03877655416727066,
0.0529145747423172,
0.0354921817779541,
0.1059221550822258,
-0.02500346675515175,
-0.17749741673469543,
-0.0438142865896225,
0.06573881208896637,
-0.045354213565588,
-0.22390563786029816,
-0.009726951830089092,
0.09943331032991409,
-0.1914641559123993,
-0.015451330691576004,
0.02838914282619953,
-0.04480560123920441,
-0.02868090756237507,
0.0007889526314102113,
0.0600614957511425,
0.015805870294570923,
0.09190283715724945,
0.07423794269561768,
0.09749054163694382,
-0.08805927634239197,
0.09811163693666458,
0.10723351687192917,
-0.09035424888134003,
0.03553062304854393,
0.06695880740880966,
-0.0467107780277729,
-0.04594837874174118,
0.05199020728468895,
0.04819667339324951,
0.01212578546255827,
-0.0561964213848114,
0.010319532826542854,
-0.04872706159949303,
0.04633839800953865,
0.10621411353349686,
0.028242740780115128,
-0.03058992512524128,
0.06704547256231308,
0.03252853453159332,
-0.1153404489159584,
0.09847725927829742,
0.012868257239460945,
0.03807265684008598,
-0.06272068619728088,
-0.015808504074811935,
0.04865187034010887,
0.027409857138991356,
-0.01764598675072193,
-0.025427930057048798,
-0.035527609288692474,
-0.015147317200899124,
-0.15422900021076202,
-0.012660279870033264,
-0.07294544577598572,
0.007333413697779179,
0.006807927042245865,
-0.03955657035112381,
-0.0043836915865540504,
0.029364487156271935,
-0.07081043720245361,
-0.06899864971637726,
-0.0017123379511758685,
0.10014908015727997,
-0.16123399138450623,
0.0016520773060619831,
0.07378670573234558,
-0.10700937360525131,
0.06776659190654755,
-0.009028629399836063,
0.006400149781256914,
0.021102426573634148,
-0.1615109145641327,
0.05426544323563576,
-0.010029333643615246,
0.02013414539396763,
0.032934170216321945,
-0.16248436272144318,
0.0024488656781613827,
-0.047329291701316833,
-0.022390197962522507,
-0.004845738876610994,
-0.04656189680099487,
-0.11974798142910004,
0.07715073227882385,
-0.01184067688882351,
-0.05094744265079498,
-0.01612357795238495,
0.05293868109583855,
0.08231643587350845,
-0.03882661834359169,
0.09632368385791779,
-0.005011113826185465,
0.05959545075893402,
-0.17253276705741882,
-0.02932477742433548,
-0.0432354174554348,
0.014331330545246601,
0.01743181422352791,
-0.009555062279105186,
0.03874485567212105,
-0.00935265514999628,
0.22544825077056885,
-0.03915993124246597,
0.16461394727230072,
0.055936723947525024,
-0.0032888432033360004,
0.0007776605198159814,
0.06758615374565125,
0.05568486079573631,
0.03412187471985817,
0.00899792555719614,
0.02200561948120594,
-0.023325180634856224,
-0.006471368949860334,
-0.1553903967142105,
0.02697177603840828,
0.14716137945652008,
0.0745159387588501,
0.006664956454187632,
0.07025619596242905,
-0.1267581284046173,
-0.11370917409658432,
0.09592846781015396,
-0.02568071521818638,
0.008476621471345425,
-0.07835444062948227,
0.12778781354427338,
0.14673273265361786,
-0.14686504006385803,
0.06517019122838974,
-0.053687721490859985,
-0.05600763112306595,
-0.09034380316734314,
-0.10879118740558624,
-0.06126067787408829,
-0.04308179020881653,
0.004678911529481411,
-0.042684826999902725,
0.055097613483667374,
0.04954573139548302,
-0.014461824670433998,
0.004931987728923559,
0.12391652166843414,
-0.006120255216956139,
0.001201988779939711,
0.03766126185655594,
0.03769403696060181,
0.024755796417593956,
-0.059261444956064224,
0.030717262998223305,
0.021477915346622467,
0.034908585250377655,
0.059853747487068176,
0.037230484187603,
-0.045039307326078415,
0.028804119676351547,
0.0020213082898408175,
-0.10957802832126617,
0.023749636486172676,
-0.012328135780990124,
-0.06936221569776535,
0.12969832122325897,
0.03471869230270386,
0.009512413293123245,
-0.037131089717149734,
0.23728759586811066,
-0.062090300023555756,
-0.08014962822198868,
-0.12913139164447784,
0.09616934508085251,
-0.013530191034078598,
0.057892732322216034,
0.03356536477804184,
-0.12210189551115036,
0.0036616562865674496,
0.13605539500713348,
0.11633196473121643,
-0.0003361425769980997,
0.012180927209556103,
0.044184453785419464,
0.004239979665726423,
-0.06263455748558044,
0.044461920857429504,
0.06619330495595932,
0.12273700535297394,
-0.07938622683286667,
0.07410858571529388,
0.00435013510286808,
-0.08385829627513885,
-0.0399140790104866,
0.1140546128153801,
-0.03326992690563202,
0.03303933143615723,
-0.041518088430166245,
0.10997304320335388,
-0.059399381279945374,
-0.3032641112804413,
0.03540288656949997,
-0.10066618025302887,
-0.1533578634262085,
-0.01690032333135605,
0.06605888903141022,
-0.02134985849261284,
0.01722477562725544,
0.06963877379894257,
-0.058587364852428436,
0.1905425637960434,
0.03258530795574188,
-0.07860512286424637,
-0.059183377772569656,
0.05133861303329468,
-0.0791650041937828,
0.302468478679657,
0.00626079086214304,
0.03169599175453186,
0.10508318990468979,
-0.028644336387515068,
-0.16361252963542938,
0.02362491562962532,
0.1140698790550232,
-0.08390003442764282,
0.08627496659755707,
0.19878266751766205,
-0.019539451226592064,
0.11435621976852417,
0.05704843997955322,
-0.06186779960989952,
0.0524447038769722,
-0.03936922550201416,
-0.052163589745759964,
-0.09776037186384201,
0.06190723925828934,
-0.06178663671016693,
0.15432539582252502,
0.09593082964420319,
-0.05059736222028732,
-0.006600235588848591,
-0.05587591603398323,
0.04507772624492645,
0.018967149779200554,
0.12800532579421997,
0.012484090402722359,
-0.17696550488471985,
0.032744914293289185,
0.0010579711524769664,
0.11208613961935043,
-0.24666742980480194,
-0.08353681117296219,
0.09015431255102158,
-0.019416818395256996,
-0.05258401483297348,
0.09870866686105728,
0.0722413882613182,
0.04240792989730835,
-0.04463369399309158,
-0.10492048412561417,
-0.019366342574357986,
0.1493324637413025,
-0.14043603837490082,
-0.014699541963636875
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | automatic-speech-recognition | SiRoZaRuPa/EN-300m-clean-0208 | [
"transformers",
"tensorboard",
"safetensors",
"wav2vec2",
"automatic-speech-recognition",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-08T08:31:59+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #tensorboard #safetensors #wav2vec2 #automatic-speech-recognition #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #tensorboard #safetensors #wav2vec2 #automatic-speech-recognition #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
51,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #wav2vec2 #automatic-speech-recognition #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06918960809707642,
0.13210147619247437,
-0.0040207370184361935,
0.023134203627705574,
0.11738458275794983,
0.003100133500993252,
0.06489233672618866,
0.1062328964471817,
-0.018454808741807938,
0.11934409290552139,
0.02399194799363613,
0.10645237565040588,
0.10633884370326996,
0.1783033311367035,
-0.006676932331174612,
-0.20753470063209534,
0.05159076303243637,
-0.1328369528055191,
-0.006210802122950554,
0.12206359207630157,
0.12859149277210236,
-0.12210913747549057,
0.0661126896739006,
-0.03582390025258064,
-0.006673390511423349,
-0.036393795162439346,
-0.05692959576845169,
-0.05386972799897194,
0.06668701767921448,
0.062350641936063766,
0.060644831508398056,
0.018570519983768463,
0.09333337843418121,
-0.2811316251754761,
0.022820472717285156,
0.08144836127758026,
0.006955916993319988,
0.06573166698217392,
0.07077054679393768,
-0.07532189786434174,
0.07820203900337219,
-0.06897217780351639,
0.14961306750774384,
0.07984381169080734,
-0.09012829512357712,
-0.1924186497926712,
-0.08871616423130035,
0.0939040556550026,
0.18705229461193085,
0.05621805414557457,
-0.031164970248937607,
0.13427454233169556,
-0.06805921345949173,
0.01878679171204567,
0.0681581050157547,
-0.07620836049318314,
-0.052718792110681534,
0.06274411827325821,
0.07032535970211029,
0.09331566095352173,
-0.13174302875995636,
-0.00581010989844799,
0.02852778322994709,
0.010611804202198982,
0.10465876758098602,
0.019570648670196533,
0.12078016996383667,
0.03880659490823746,
-0.14213255047798157,
-0.04347489774227142,
0.08553854376077652,
0.040430404245853424,
-0.053023193031549454,
-0.25508078932762146,
-0.01728249341249466,
-0.03535711020231247,
-0.03508080542087555,
-0.050225600600242615,
0.04455358535051346,
-0.02446228265762329,
0.07527779787778854,
-0.005772874690592289,
-0.07288429886102676,
-0.049391333013772964,
0.07827333360910416,
0.07604440301656723,
0.027127955108880997,
-0.02607012540102005,
0.012724175117909908,
0.11518751829862595,
0.11186537146568298,
-0.11170512437820435,
-0.052296292036771774,
-0.06167195737361908,
-0.09369548410177231,
-0.047245271503925323,
0.03096519224345684,
0.04075217619538307,
0.05507682263851166,
0.21005180478096008,
0.004100160673260689,
0.05025365948677063,
0.030208947136998177,
0.013425402343273163,
0.06431768089532852,
0.09155748784542084,
-0.0652301162481308,
-0.12225554138422012,
-0.02715214155614376,
0.10966562479734421,
0.009606653824448586,
-0.03282571956515312,
-0.04075070470571518,
0.0665077269077301,
0.030208082869648933,
0.12366250902414322,
0.0723525807261467,
0.018685176968574524,
-0.07855737954378128,
-0.06267400830984116,
0.1677972972393036,
-0.1649521440267563,
0.03285328298807144,
0.02912791818380356,
-0.050073519349098206,
-0.008440917357802391,
0.01682254858314991,
0.021022414788603783,
-0.018704243004322052,
0.08882031589746475,
-0.054653100669384,
-0.03264474496245384,
-0.11321555823087692,
-0.05006399378180504,
0.028676055371761322,
0.006981914862990379,
-0.03174450621008873,
-0.04053306579589844,
-0.10819326341152191,
-0.07601769268512726,
0.07845603674650192,
-0.06794282793998718,
-0.04567456990480423,
-0.03693155571818352,
-0.077850341796875,
0.013987138867378235,
-0.001372430007904768,
0.11866221576929092,
-0.028359893709421158,
0.049781348556280136,
-0.06040623039007187,
0.07331450283527374,
0.1427365392446518,
0.027582714334130287,
-0.05536656826734543,
0.05209227278828621,
-0.22961750626564026,
0.10650996118783951,
-0.0820845440030098,
0.039568543434143066,
-0.16523221135139465,
-0.01437871903181076,
0.04151884838938713,
0.02703598327934742,
-0.011580551974475384,
0.13367699086666107,
-0.20120634138584137,
-0.03629620373249054,
0.17902998626232147,
-0.11463885754346848,
-0.08275967836380005,
0.05660289525985718,
-0.05534304678440094,
0.12154120951890945,
0.04968025162816048,
-0.015457268804311752,
0.02872299961745739,
-0.14586561918258667,
-0.015341621823608875,
-0.06385710090398788,
-0.031775522977113724,
0.15648432075977325,
0.058627333492040634,
-0.05283202603459358,
0.06168147549033165,
0.01965263858437538,
-0.018219612538814545,
-0.04959159716963768,
-0.03271770104765892,
-0.09723224490880966,
0.011255990713834763,
-0.0728980302810669,
0.023943135514855385,
-0.031872402876615524,
-0.09092787653207779,
-0.03651702031493187,
-0.15960368514060974,
0.006672970950603485,
0.09574975073337555,
-0.005800875835120678,
-0.02275932766497135,
-0.11338774859905243,
-0.010310402140021324,
0.020829740911722183,
-0.0006964936037547886,
-0.14685183763504028,
-0.05314113572239876,
0.017828308045864105,
-0.16250769793987274,
0.031012238934636116,
-0.03655901551246643,
0.04738416150212288,
0.03556562215089798,
-0.03982981666922569,
-0.03375418856739998,
0.019630931317806244,
0.022369354963302612,
-0.010214408859610558,
-0.2756194770336151,
-0.015468244440853596,
-0.043052829802036285,
0.16435527801513672,
-0.2469322234392166,
0.04182727262377739,
0.07295827567577362,
0.1338571161031723,
0.015705497935414314,
-0.03647774085402489,
0.028713135048747063,
-0.06289805471897125,
-0.030222538858652115,
-0.06501726806163788,
-0.007188703399151564,
-0.039097823202610016,
-0.04806915298104286,
0.04462466016411781,
-0.16899824142456055,
-0.033922191709280014,
0.1186266764998436,
0.04557104408740997,
-0.15134701132774353,
-0.04948775842785835,
-0.04092395305633545,
-0.056753676384687424,
-0.06932670623064041,
-0.0517798475921154,
0.10663432627916336,
0.05747092142701149,
0.05196038633584976,
-0.05911761149764061,
-0.06484735757112503,
0.00799498613923788,
-0.01853559911251068,
-0.023748042061924934,
0.07913291454315186,
0.06702018529176712,
-0.11829525977373123,
0.09312599897384644,
0.08573136478662491,
0.07933273166418076,
0.10508506000041962,
-0.0014733473071828485,
-0.09117123484611511,
-0.025300826877355576,
0.029316658154129982,
0.016105778515338898,
0.14908336102962494,
-0.04350128397345543,
0.04314031824469566,
0.040114615112543106,
-0.01687462255358696,
0.008028145879507065,
-0.09918303042650223,
0.030367493629455566,
0.026081476360559464,
-0.012195796705782413,
0.041467417031526566,
-0.05302301421761513,
0.021834537386894226,
0.10195169597864151,
0.03181454911828041,
0.04113520681858063,
0.011278065852820873,
-0.050533477216959,
-0.11812540888786316,
0.17222443222999573,
-0.10861039906740189,
-0.2369978129863739,
-0.12320686131715775,
-0.01618431694805622,
0.02991701476275921,
-0.015134924091398716,
0.01900940015912056,
-0.06770696491003036,
-0.11834623664617538,
-0.09672471135854721,
0.04564153030514717,
0.06599046289920807,
-0.08051323890686035,
-0.055777665227651596,
0.06501153111457825,
0.048011794686317444,
-0.13664643466472626,
0.02571168728172779,
0.03327706828713417,
-0.08857693523168564,
0.00793769583106041,
0.08559047430753708,
0.06839455664157867,
0.18071474134922028,
0.01134483702480793,
-0.023087946698069572,
0.017521869391202927,
0.19720622897148132,
-0.14027054607868195,
0.10202740132808685,
0.13801661133766174,
-0.07145930081605911,
0.07873693108558655,
0.2032429575920105,
0.039016321301460266,
-0.10376140475273132,
0.039679598063230515,
0.036421533674001694,
-0.025852223858237267,
-0.24745285511016846,
-0.08099643886089325,
0.00836301501840353,
-0.0664474293589592,
0.0802333801984787,
0.08307429403066635,
0.09203000366687775,
0.023238254711031914,
-0.1043974831700325,
-0.07363210618495941,
0.05418974906206131,
0.11036353558301926,
-0.004034504294395447,
-0.011317858472466469,
0.09753942489624023,
-0.020273780450224876,
0.02676866576075554,
0.08875394612550735,
0.012205728329718113,
0.18836407363414764,
0.050518929958343506,
0.14771167933940887,
0.09208200126886368,
0.053752463310956955,
0.016467519104480743,
0.010000402107834816,
0.017887894064188004,
0.02435637265443802,
-0.014350295066833496,
-0.08589190989732742,
-0.006933859083801508,
0.1298609972000122,
0.027646880596876144,
0.04127250239253044,
0.013248836621642113,
-0.04125351831316948,
0.08765199780464172,
0.17516882717609406,
0.013442369177937508,
-0.20506484806537628,
-0.06488820165395737,
0.0686659887433052,
-0.08813467621803284,
-0.10374542325735092,
-0.021716099232435226,
0.04023343697190285,
-0.1762947142124176,
0.02770446240901947,
-0.025082001462578773,
0.0983029454946518,
-0.12493812292814255,
-0.01920684240758419,
0.0476171039044857,
0.06939635425806046,
-0.018209589645266533,
0.0625329241156578,
-0.17832936346530914,
0.13725855946540833,
0.012600419111549854,
0.07603015750646591,
-0.0920197069644928,
0.0829358845949173,
0.010243658907711506,
-0.008985995315015316,
0.14880549907684326,
-0.002428766805678606,
-0.056611087173223495,
-0.10275979340076447,
-0.09291432052850723,
-0.01180565357208252,
0.11795864999294281,
-0.11873860657215118,
0.09995509684085846,
-0.017298342660069466,
-0.043639615178108215,
0.0016699014231562614,
-0.12897762656211853,
-0.1380222588777542,
-0.17400150001049042,
0.041601065546274185,
-0.12252611666917801,
0.04249255359172821,
-0.10634490847587585,
-0.05313412845134735,
-0.058118730783462524,
0.19448153674602509,
-0.2263878583908081,
-0.07106572389602661,
-0.1503530591726303,
-0.06515897810459137,
0.11819497495889664,
-0.042735762894153595,
0.08508200198411942,
0.017862383276224136,
0.19214710593223572,
0.010283242911100388,
-0.013114631175994873,
0.10883224755525589,
-0.10211063176393509,
-0.21299202740192413,
-0.10015871375799179,
0.13945214450359344,
0.13517092168331146,
0.038856618106365204,
0.002108179498463869,
0.030881604179739952,
-0.006152692716568708,
-0.11462404578924179,
0.028862472623586655,
0.18585458397865295,
0.10306477546691895,
0.03526908904314041,
-0.03260820358991623,
-0.14471980929374695,
-0.08779244124889374,
-0.045098960399627686,
0.017435450106859207,
0.19264571368694305,
-0.07120641320943832,
0.17354503273963928,
0.15474873781204224,
-0.053835928440093994,
-0.20943360030651093,
0.03015606477856636,
0.036211419850587845,
0.0007652041967958212,
0.05587008595466614,
-0.19489167630672455,
0.0909743532538414,
0.0033501458819955587,
-0.057322751730680466,
0.12121490389108658,
-0.17501963675022125,
-0.15013514459133148,
0.07031099498271942,
0.07301220297813416,
-0.17921873927116394,
-0.12142012268304825,
-0.09439031779766083,
-0.04026462882757187,
-0.11460573226213455,
0.07970702648162842,
-0.016233494505286217,
0.010252374224364758,
0.032961323857307434,
0.018216567113995552,
0.010428756475448608,
-0.04740371182560921,
0.1864585429430008,
-0.003947122488170862,
0.04788469523191452,
-0.07597782462835312,
-0.06253167986869812,
0.045070283114910126,
-0.06455249339342117,
0.0716865211725235,
-0.00903246272355318,
0.006079745013266802,
-0.1052967831492424,
-0.06088602915406227,
-0.03328738734126091,
0.02272024378180504,
-0.07930614799261093,
-0.09432698786258698,
-0.03726235777139664,
0.10006307810544968,
0.09058371931314468,
-0.03892482817173004,
-0.06462740153074265,
-0.08978539705276489,
0.028800709173083305,
0.21877005696296692,
0.177296444773674,
0.05685123801231384,
-0.066028892993927,
-0.00540707865729928,
-0.01588953658938408,
0.053271859884262085,
-0.2026120126247406,
0.0566285103559494,
0.035300228744745255,
0.033545590937137604,
0.11711569130420685,
-0.026464059948921204,
-0.16407892107963562,
-0.048686347901821136,
0.05304291099309921,
-0.07358507066965103,
-0.17289869487285614,
0.014132710173726082,
0.07088939845561981,
-0.1477956771850586,
-0.023786291480064392,
0.04775075986981392,
-0.017420068383216858,
-0.03159533068537712,
0.006238185800611973,
0.08124099671840668,
0.01671770215034485,
0.09224288910627365,
0.053469255566596985,
0.09704500436782837,
-0.10683690756559372,
0.06699982285499573,
0.07745448499917984,
-0.10474617779254913,
0.03967198729515076,
0.0603945255279541,
-0.06895622611045837,
-0.03619396686553955,
0.033563096076250076,
0.08692663908004761,
0.04178347438573837,
-0.060071151703596115,
0.0073408023454248905,
-0.10486608743667603,
0.06092875450849533,
0.1210157498717308,
0.04285310208797455,
0.0076990588568151,
0.036018576472997665,
0.04045969620347023,
-0.09288305044174194,
0.12451037764549255,
0.04114879295229912,
0.028287222608923912,
-0.05418051406741142,
-0.028997255489230156,
0.03649618849158287,
-0.03188192844390869,
-0.01566455140709877,
-0.04152749106287956,
-0.06663620471954346,
-0.010323094204068184,
-0.16889281570911407,
0.006573607679456472,
-0.05270812287926674,
0.008401375263929367,
0.021295055747032166,
-0.03304858133196831,
0.005127503536641598,
0.019244063645601273,
-0.07131489366292953,
-0.052214257419109344,
-0.006754601374268532,
0.10161449760198593,
-0.17169132828712463,
0.014349433593451977,
0.0744767114520073,
-0.12469461560249329,
0.08815638720989227,
0.018520260229706764,
0.0005999338463880122,
0.03465453162789345,
-0.13307695090770721,
0.043367430567741394,
-0.006723123602569103,
0.011691853404045105,
0.048354603350162506,
-0.21661832928657532,
-0.0025545719545334578,
-0.04856108874082565,
-0.055710889399051666,
-0.006375120021402836,
-0.02562650851905346,
-0.11432337760925293,
0.10399775207042694,
0.010540200397372246,
-0.0755159854888916,
-0.02542583830654621,
0.037674929946660995,
0.0969945415854454,
-0.03298725560307503,
0.16065140068531036,
-0.01863807439804077,
0.06254526972770691,
-0.1797095239162445,
-0.018202031031250954,
-0.01975269988179207,
0.023043567314743996,
-0.03248249739408493,
-0.008440588600933552,
0.05180126056075096,
-0.023841936141252518,
0.20870842039585114,
-0.022057142108678818,
0.033427316695451736,
0.06674833595752716,
-0.021141132339835167,
-0.02877473458647728,
0.1086326614022255,
0.054397158324718475,
0.012029323726892471,
0.03175004944205284,
0.006914193741977215,
-0.04090225324034691,
-0.004564614500850439,
-0.1556052416563034,
0.07673801481723785,
0.17203287780284882,
0.0805397778749466,
-0.00828546192497015,
0.06094660609960556,
-0.11003988236188889,
-0.11399497091770172,
0.10722645372152328,
-0.05822233483195305,
-0.014757114462554455,
-0.05772337689995766,
0.14011409878730774,
0.15646083652973175,
-0.19130073487758636,
0.06022409349679947,
-0.06736859679222107,
-0.04819837212562561,
-0.10633485019207001,
-0.17335662245750427,
-0.061282314360141754,
-0.0583864226937294,
-0.01613355241715908,
-0.05076048895716667,
0.06713438034057617,
0.08348768949508667,
0.02054755762219429,
0.016258614137768745,
0.0817527249455452,
-0.02199946530163288,
0.007656866684556007,
0.034995537251234055,
0.06331320106983185,
0.0073803807608783245,
-0.04667557775974274,
0.009565448388457298,
0.0006085589993745089,
0.035281602293252945,
0.04957476258277893,
0.037472013384103775,
-0.026353945955634117,
0.007689491845667362,
-0.02916470356285572,
-0.11019428819417953,
0.04115133360028267,
-0.026625385507941246,
-0.06341774761676788,
0.1439228504896164,
0.031860120594501495,
-0.008713874034583569,
-0.025656426325440407,
0.25211021304130554,
-0.07529866695404053,
-0.08892348408699036,
-0.1387489140033722,
0.13557645678520203,
-0.031552400439977646,
0.06481313705444336,
0.037692490965127945,
-0.11259825527667999,
0.03179538995027542,
0.1362704634666443,
0.1458069533109665,
-0.049145035445690155,
0.019655266776680946,
0.013711978681385517,
0.0032459446229040623,
-0.04005579650402069,
0.04973040521144867,
0.06590425968170166,
0.12457112967967987,
-0.05082963407039642,
0.08012272417545319,
-0.0028764382004737854,
-0.10040896385908127,
-0.02852385863661766,
0.12230420112609863,
-0.003029873361811042,
0.019506774842739105,
-0.0761401429772377,
0.12728425860404968,
-0.043905097991228104,
-0.2665610611438751,
0.06613168120384216,
-0.0650629922747612,
-0.14912083745002747,
-0.022557994350790977,
0.05126400291919708,
-0.008650023490190506,
0.026705266907811165,
0.06785756349563599,
-0.0670214518904686,
0.18420551717281342,
0.03873218223452568,
-0.05507900193333626,
-0.058854296803474426,
0.07306438684463501,
-0.09833692759275436,
0.2929907441139221,
0.00751500902697444,
0.05993965268135071,
0.09920700639486313,
-0.029096059501171112,
-0.13847678899765015,
0.031734831631183624,
0.08172675222158432,
-0.07410130649805069,
0.055994872003793716,
0.21827135980129242,
-0.008840959519147873,
0.11804516613483429,
0.07454971224069595,
-0.09561564773321152,
0.05016838759183884,
-0.10613930225372314,
-0.09673135727643967,
-0.08329153805971146,
0.09532807767391205,
-0.05763502046465874,
0.14755868911743164,
0.1186022087931633,
-0.04606860503554344,
0.02281493879854679,
-0.018614748492836952,
0.048749152570962906,
0.0023650694638490677,
0.12439922988414764,
0.020209291949868202,
-0.19710010290145874,
0.026845410466194153,
-0.008902255445718765,
0.10291280597448349,
-0.2202581763267517,
-0.09718955308198929,
0.04764820635318756,
0.0019112902227789164,
-0.05895697697997093,
0.12370198965072632,
0.055919989943504333,
0.04170476272702217,
-0.04714735969901085,
-0.028212912380695343,
-0.002841046778485179,
0.16146929562091827,
-0.11127673834562302,
0.0008471902110613883
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
Quantization of the model Galactica-6.7b-evol-instruct in 4 bits using GPTQ.
- **Quantizated with:** GPTQ
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** GeorgiaTechResearchInstitute/galactica-6.7b-evol-instruct-70k
- **Paper [optional]:** https://galactica.org/static/paper.pdf
- **Demo :** https://github.com/paperswithcode/galai/blob/main/docs/model_card.md
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | EtienneDu91/galactica-6.7b-evol-instruct-70k-GPTQ-4b | [
"transformers",
"safetensors",
"opt",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"4-bit",
"region:us"
] | 2024-02-08T08:32:21+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #opt #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
Quantization of the model Galactica-6.7b-evol-instruct in 4 bits using GPTQ.
- Quantizated with: GPTQ
### Model Sources [optional]
- Repository: GeorgiaTechResearchInstitute/galactica-6.7b-evol-instruct-70k
- Paper [optional]: URL
- Demo : URL
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\nQuantization of the model Galactica-6.7b-evol-instruct in 4 bits using GPTQ.\n\n- Quantizated with: GPTQ",
"### Model Sources [optional]\n\n\n\n- Repository: GeorgiaTechResearchInstitute/galactica-6.7b-evol-instruct-70k\n- Paper [optional]: URL\n- Demo : URL",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #opt #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\nQuantization of the model Galactica-6.7b-evol-instruct in 4 bits using GPTQ.\n\n- Quantizated with: GPTQ",
"### Model Sources [optional]\n\n\n\n- Repository: GeorgiaTechResearchInstitute/galactica-6.7b-evol-instruct-70k\n- Paper [optional]: URL\n- Demo : URL",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
59,
6,
3,
37,
47,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #opt #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n# Model Card for Model ID## Model Details### Model Description\n\nQuantization of the model Galactica-6.7b-evol-instruct in 4 bits using GPTQ.\n\n- Quantizated with: GPTQ### Model Sources [optional]\n\n\n\n- Repository: GeorgiaTechResearchInstitute/galactica-6.7b-evol-instruct-70k\n- Paper [optional]: URL\n- Demo : URL## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06492895632982254,
0.2485455572605133,
-0.0022158771753311157,
0.027863509953022003,
0.13087362051010132,
0.0049267541617155075,
0.03436320647597313,
0.13001832365989685,
-0.0205960962921381,
0.10478580743074417,
0.03861735761165619,
0.08583404868841171,
0.10517625510692596,
0.18005163967609406,
0.04986954107880592,
-0.2165023237466812,
-0.014399087987840176,
-0.0876489132642746,
0.01078783068805933,
0.11461411416530609,
0.12630201876163483,
-0.09554307162761688,
0.07664137333631516,
-0.04331670701503754,
-0.0212095957249403,
0.004955760668963194,
-0.09595382213592529,
-0.06799504905939102,
0.05141330882906914,
0.05261100083589554,
0.018406733870506287,
0.007685211952775717,
0.10473617911338806,
-0.2953396141529083,
0.01226802822202444,
0.08000114560127258,
0.0014606722397729754,
0.0610983669757843,
0.0668618381023407,
-0.08207937330007553,
0.08165228366851807,
-0.10964649170637131,
0.12205932289361954,
0.08432561904191971,
-0.05529686063528061,
-0.19763576984405518,
-0.08130604773759842,
0.0838431864976883,
0.10587812960147858,
0.07578179240226746,
-0.03294218331575394,
0.15090222656726837,
-0.04800757020711899,
0.02440529689192772,
0.15004071593284607,
-0.1211579218506813,
-0.0515468455851078,
0.05410098657011986,
0.1139674186706543,
0.10398968309164047,
-0.1315048485994339,
0.018510688096284866,
0.05249796062707901,
0.02789142169058323,
0.09079307317733765,
0.027026602998375893,
0.08273075520992279,
0.04613037034869194,
-0.1284024566411972,
-0.028089752420783043,
0.1263481229543686,
0.036739181727170944,
-0.06293926388025284,
-0.1955985128879547,
0.008643078617751598,
-0.05083083733916283,
-0.013437643647193909,
-0.06252213567495346,
0.0404779426753521,
-0.027179155498743057,
0.05353834107518196,
-0.058536823838949203,
-0.10539493709802628,
-0.04831704497337341,
0.07853515446186066,
0.09718143194913864,
0.0198779609054327,
-0.03411445766687393,
0.047345053404569626,
0.1289806216955185,
0.030611442402005196,
-0.10666277259588242,
-0.06450263410806656,
-0.046178191900253296,
-0.10292468219995499,
-0.040517281740903854,
0.05287964642047882,
0.02055947110056877,
0.021123554557561874,
0.20616213977336884,
0.008333525620400906,
0.03860843554139137,
0.02781750075519085,
0.0015696946065872908,
0.043245017528533936,
0.0893787294626236,
-0.057707831263542175,
-0.16357874870300293,
-0.038922019302845,
0.08591561019420624,
-0.0035445126704871655,
-0.0325944721698761,
-0.03160560503602028,
0.0177516620606184,
0.07113640010356903,
0.12600551545619965,
0.08671972900629044,
-0.014846844598650932,
-0.06445138901472092,
-0.022134842351078987,
0.2153104692697525,
-0.1450158655643463,
0.0439218245446682,
-0.03885809704661369,
-0.016067083925008774,
-0.06181516870856285,
0.03861593082547188,
0.024919943884015083,
-0.007290075067430735,
0.08399653434753418,
-0.06556542962789536,
-0.05277950316667557,
-0.08073796331882477,
-0.024238526821136475,
0.031584929674863815,
-0.00478577334433794,
-0.008569258265197277,
-0.07201912999153137,
-0.11629980802536011,
-0.03869177773594856,
0.04762246832251549,
-0.06689810752868652,
-0.05239380896091461,
0.0190404262393713,
-0.07902634888887405,
-0.010866610333323479,
0.00015827923198230565,
0.09595482796430588,
-0.02887842245399952,
0.038564618676900864,
-0.010551038198173046,
0.051502350717782974,
0.1076425239443779,
0.034210432320833206,
-0.06958814710378647,
0.03882521018385887,
-0.20272697508335114,
0.0806896761059761,
-0.12076292932033539,
0.053703512996435165,
-0.17122943699359894,
-0.04632825031876564,
-0.0032711264211684465,
0.0008966894820332527,
0.022770041599869728,
0.12240201979875565,
-0.18527869880199432,
-0.02537393942475319,
0.134786918759346,
-0.07595961540937424,
-0.10570789128541946,
0.0873759463429451,
-0.0171673446893692,
0.14522524178028107,
0.05499045178294182,
-0.01983044669032097,
0.08054640889167786,
-0.1483544409275055,
-0.08002115786075592,
-0.02358884923160076,
0.0007626526057720184,
0.13583992421627045,
0.06707373261451721,
-0.06700074672698975,
0.0587092749774456,
0.029756374657154083,
-0.01889154128730297,
-0.040976911783218384,
-0.05536157637834549,
-0.10046778619289398,
-0.008522582240402699,
-0.08755437284708023,
0.05426979809999466,
-0.009525632485747337,
-0.06865072250366211,
-0.03851805627346039,
-0.19272181391716003,
-0.010993778705596924,
0.09536363929510117,
-0.0008582004229538143,
-0.0067699141800403595,
-0.07801692187786102,
0.009217973798513412,
-0.026443153619766235,
-0.012683901004493237,
-0.16617785394191742,
-0.04106809198856354,
0.05342451483011246,
-0.1369878649711609,
0.04101516678929329,
-0.08488772809505463,
0.05973641946911812,
0.036548882722854614,
-0.06463591009378433,
-0.004788040649145842,
-0.01906176283955574,
0.013302210718393326,
-0.052462685853242874,
-0.18609179556369781,
-0.03655979782342911,
-0.033972930163145065,
0.15957142412662506,
-0.24987731873989105,
0.04351238161325455,
0.073219895362854,
0.14409023523330688,
0.00030963862081989646,
-0.05762028694152832,
0.01850152015686035,
-0.06148192659020424,
-0.04061427712440491,
-0.07406282424926758,
-0.01253951620310545,
-0.02804293856024742,
-0.043072640895843506,
-0.00347367231734097,
-0.1587100327014923,
-0.028672434389591217,
0.08423908799886703,
0.13092041015625,
-0.1531660407781601,
-0.01811714470386505,
-0.039327703416347504,
-0.06858926266431808,
-0.07899434119462967,
-0.06594744324684143,
0.12263375520706177,
0.04306042566895485,
0.04224539175629616,
-0.07489849627017975,
-0.07524267584085464,
0.02249656617641449,
0.01579158939421177,
-0.039258383214473724,
0.0769464373588562,
0.044059667736291885,
-0.10905511677265167,
0.07553675025701523,
0.07370628416538239,
0.0696975365281105,
0.1086159348487854,
0.024719463661313057,
-0.11701945960521698,
-0.029369890689849854,
0.02178591676056385,
0.027365226298570633,
0.1380978673696518,
-0.05976833775639534,
0.040027204900979996,
0.040211401879787445,
-0.04863779991865158,
0.03011356107890606,
-0.08894205093383789,
0.025579407811164856,
0.028165921568870544,
-0.01072653941810131,
0.04263006150722504,
-0.03538178652524948,
-0.008995004929602146,
0.061642836779356,
0.04646572470664978,
0.07278472930192947,
0.004873939324170351,
-0.013297013007104397,
-0.103795126080513,
0.17068937420845032,
-0.09602290391921997,
-0.27382200956344604,
-0.16123943030834198,
0.04153357073664665,
0.04143509268760681,
-0.030150050297379494,
0.027459239587187767,
-0.07586228847503662,
-0.10338939726352692,
-0.10172665864229202,
-0.00353231024928391,
0.008670157752931118,
-0.054876912385225296,
-0.05266117304563522,
0.07189930230379105,
0.03736471012234688,
-0.14203627407550812,
0.04179505631327629,
0.05507484823465347,
-0.06300780922174454,
-0.020047439262270927,
0.07856033742427826,
0.10928533226251602,
0.13151700794696808,
-0.023316236212849617,
-0.019464988261461258,
0.027009909972548485,
0.20198030769824982,
-0.13435843586921692,
0.10620138794183731,
0.1207163855433464,
-0.03308330103754997,
0.08555427938699722,
0.16315048933029175,
0.024155938997864723,
-0.08394128084182739,
0.05320923775434494,
0.04162675514817238,
-0.029296373948454857,
-0.2678546905517578,
-0.0525396354496479,
0.005394009407609701,
-0.06484492868185043,
0.10552100837230682,
0.09473647177219391,
0.129742830991745,
0.040994904935359955,
-0.07315371930599213,
-0.02293115109205246,
0.00988229550421238,
0.11298087984323502,
-0.04737129807472229,
-0.005112816113978624,
0.07838605344295502,
-0.04210381582379341,
0.004054614342749119,
0.09776412695646286,
0.036744557321071625,
0.2158859372138977,
0.013524997048079967,
0.13470372557640076,
0.05956872180104256,
0.08099354058504105,
0.003303050994873047,
0.016346804797649384,
0.04988337308168411,
0.010645149275660515,
0.0025323256850242615,
-0.10107843577861786,
0.021917134523391724,
0.13247229158878326,
0.030932223424315453,
0.01749274879693985,
0.017652170732617378,
-0.02348579466342926,
0.039652835577726364,
0.18976560235023499,
-0.0010028131073340774,
-0.21514444053173065,
-0.08023948967456818,
0.06921038031578064,
-0.0659533366560936,
-0.11935911327600479,
-0.038654424250125885,
0.031003031879663467,
-0.17360688745975494,
0.04555166885256767,
-0.03321617469191551,
0.0897010788321495,
-0.09760976582765579,
-0.035299208015203476,
0.04012395069003105,
0.09753531217575073,
-0.024182191118597984,
0.09763652831315994,
-0.11361934244632721,
0.12975051999092102,
0.031402233988046646,
0.09609731286764145,
-0.11231588572263718,
0.08478008210659027,
-0.002979439217597246,
0.00881533045321703,
0.1840788573026657,
-0.009907939471304417,
-0.023870427161455154,
-0.08187396824359894,
-0.10489872097969055,
-0.01089415606111288,
0.1357319951057434,
-0.13113026320934296,
0.0890079140663147,
-0.020998157560825348,
-0.045237019658088684,
0.007372410036623478,
-0.1453072875738144,
-0.20731477439403534,
-0.20752806961536407,
0.06615724414587021,
-0.09264487773180008,
0.03128935769200325,
-0.11631037294864655,
-0.08094912767410278,
-0.022121910005807877,
0.26553788781166077,
-0.15618367493152618,
-0.06964822113513947,
-0.14604680240154266,
-0.018293270841240883,
0.18186378479003906,
-0.04497719183564186,
0.07957420498132706,
-0.013350976631045341,
0.20435534417629242,
0.012792674824595451,
0.0007172205368988216,
0.058652568608522415,
-0.08344025909900665,
-0.17071440815925598,
-0.0612877793610096,
0.14965946972370148,
0.12372659891843796,
0.04655003547668457,
0.0029554287903010845,
0.014267233200371265,
-0.022170092910528183,
-0.1151617169380188,
-0.0013988284626975656,
0.1495891809463501,
0.0687088668346405,
0.045841995626688004,
-0.041796404868364334,
-0.12090694159269333,
-0.07692667096853256,
-0.06831911951303482,
0.06695525348186493,
0.21507754921913147,
-0.10190587490797043,
0.155456081032753,
0.14125904440879822,
-0.06450420618057251,
-0.20033672451972961,
0.04388466849923134,
0.056798357516527176,
0.0045372419990599155,
0.07068607956171036,
-0.1864013373851776,
0.10455002635717392,
0.0199896227568388,
-0.052815984934568405,
0.1185351312160492,
-0.13611294329166412,
-0.15408265590667725,
0.04744667187333107,
0.04176997393369675,
-0.24628230929374695,
-0.12480732798576355,
-0.0983353704214096,
-0.06423445045948029,
-0.1583910584449768,
0.0702686533331871,
0.00837394967675209,
0.030196262523531914,
0.02382057160139084,
0.006100391037762165,
0.025549037382006645,
-0.05419508367776871,
0.17923560738563538,
-0.010596845299005508,
0.004268846940249205,
-0.059283312410116196,
-0.0334007553756237,
0.11472573131322861,
-0.04685860127210617,
0.11613255739212036,
-0.00962455477565527,
0.02967512421309948,
-0.06417036801576614,
-0.05456084758043289,
-0.044567935168743134,
0.04627130925655365,
-0.07167719304561615,
-0.09282892942428589,
-0.05126693844795227,
0.08878768235445023,
0.05713209882378578,
-0.04615040495991707,
-0.020299891009926796,
-0.060972921550273895,
0.0975649282336235,
0.155148446559906,
0.1737409383058548,
0.0161178819835186,
-0.06491278111934662,
0.013826942071318626,
-0.04425448924303055,
0.021211933344602585,
-0.24403193593025208,
0.03733118623495102,
0.062450066208839417,
0.048766445368528366,
0.0968855693936348,
-0.024077188223600388,
-0.18118073046207428,
-0.038755953311920166,
0.08083999902009964,
-0.0505918525159359,
-0.2227422595024109,
-0.028493260964751244,
0.07252851873636246,
-0.20957234501838684,
0.0040898737497627735,
0.0360654853284359,
-0.03985421732068062,
-0.03546028584241867,
0.0018042523879557848,
0.06499318033456802,
0.009270250797271729,
0.10887293517589569,
0.07213980704545975,
0.09173528850078583,
-0.08453360944986343,
0.10641037672758102,
0.09756974130868912,
-0.04977111890912056,
0.029455387964844704,
0.08264753967523575,
-0.05187100172042847,
-0.03900254890322685,
0.03896890953183174,
-0.007215480785816908,
0.031354568898677826,
-0.03291191905736923,
0.0075553120113909245,
-0.02691829204559326,
0.038990531116724014,
0.06549962610006332,
0.03218192234635353,
-0.02545514702796936,
0.05270187556743622,
0.031682129949331284,
-0.11229922622442245,
0.10374309867620468,
0.03917780518531799,
0.035554319620132446,
-0.08885916322469711,
-0.01895715296268463,
0.0358644537627697,
0.027121029794216156,
-0.017861470580101013,
-0.022513361647725105,
-0.022946421056985855,
-0.02310868538916111,
-0.10881394892930984,
-0.021490564569830894,
-0.09043014049530029,
0.0028869726229459047,
0.005350644234567881,
-0.06039374694228172,
-0.01852746307849884,
0.03205212950706482,
-0.06339004635810852,
-0.0830668956041336,
-0.003625177778303623,
0.09971921890974045,
-0.14507953822612762,
0.0053719813004136086,
0.08759342133998871,
-0.10501892864704132,
0.08465040475130081,
-0.021577173843979836,
0.02612544223666191,
0.02488415688276291,
-0.16002321243286133,
0.046240612864494324,
0.0019471963169053197,
0.01463718805462122,
0.01802578754723072,
-0.16204708814620972,
0.0007690702914260328,
-0.04921474680304527,
-0.006799951661378145,
0.002664298051968217,
-0.02063259482383728,
-0.12063552439212799,
0.044290680438280106,
-0.04020993039011955,
-0.03329206630587578,
-0.029817819595336914,
0.05140003189444542,
0.09700947999954224,
-0.0377107635140419,
0.09666676074266434,
-0.006891360506415367,
0.051564961671829224,
-0.17360949516296387,
-0.015296084806323051,
-0.05560685321688652,
0.004855169449001551,
0.021521922200918198,
-0.005525266285985708,
0.04495210573077202,
0.001319396891631186,
0.20610783994197845,
-0.05998735874891281,
0.16146863996982574,
0.05786270275712013,
-0.001460517873056233,
0.014333384111523628,
0.05717949941754341,
0.05696938559412956,
0.04399082809686661,
0.00512533588334918,
0.021960027515888214,
-0.01965939998626709,
-0.011000879108905792,
-0.16320595145225525,
0.003632795764133334,
0.13145266473293304,
0.047458238899707794,
0.01905948296189308,
0.056696195155382156,
-0.13483136892318726,
-0.12415964901447296,
0.10982432216405869,
-0.032900188118219376,
0.0032863160595297813,
-0.07277778536081314,
0.12107477337121964,
0.14846166968345642,
-0.14045655727386475,
0.04984106123447418,
-0.047337014228105545,
-0.06848020851612091,
-0.08632142841815948,
-0.11678571254014969,
-0.06294292956590652,
-0.048108819872140884,
0.00959003809839487,
-0.04291105270385742,
0.05765237286686897,
0.04305584728717804,
-0.009613477624952793,
-0.00020685087656602263,
0.12547777593135834,
0.003763723885640502,
-0.02322978526353836,
0.057856831699609756,
0.03214063495397568,
0.0195488128811121,
-0.053230926394462585,
0.02792608179152012,
0.023323003202676773,
0.05284693092107773,
0.06512542068958282,
0.024561945348978043,
-0.06767189502716064,
0.0351460799574852,
-0.0004094343166798353,
-0.11319027096033096,
0.015064947307109833,
-0.002866577124223113,
-0.06629398465156555,
0.11099530011415482,
0.03544718772172928,
0.006719656754285097,
-0.030281636863946915,
0.2548395097255707,
-0.06513017416000366,
-0.07665016502141953,
-0.12283258140087128,
0.060199882835149765,
0.0005356904002837837,
0.05289331078529358,
0.03171338886022568,
-0.13703499734401703,
0.012473145499825478,
0.1301085203886032,
0.10793294757604599,
0.013116789050400257,
0.0200324859470129,
0.04951508343219757,
-0.0019402988255023956,
-0.05413200706243515,
0.026965836063027382,
0.05942260101437569,
0.11847151815891266,
-0.08084212243556976,
0.0781920924782753,
0.01071203127503395,
-0.08089379966259003,
-0.03797081857919693,
0.08860687166452408,
-0.03579205647110939,
0.03835368901491165,
-0.041456274688243866,
0.1295255571603775,
-0.0324859581887722,
-0.28693777322769165,
0.01727973110973835,
-0.10542989522218704,
-0.17160893976688385,
-0.013368119485676289,
0.04967804625630379,
-0.03058108687400818,
0.011019009165465832,
0.07319226861000061,
-0.04279545694589615,
0.173733651638031,
0.03543631359934807,
-0.08847320824861526,
-0.055184874683618546,
0.04690372198820114,
-0.07436775416135788,
0.297113299369812,
-0.005647643469274044,
0.03685363382101059,
0.10905656963586807,
-0.03822219371795654,
-0.1767616868019104,
0.00989824254065752,
0.09395764768123627,
-0.06546325236558914,
0.09915579855442047,
0.20224615931510925,
-0.016049250960350037,
0.0950351357460022,
0.06572679430246353,
-0.03523799031972885,
0.044779498130083084,
-0.041741643100976944,
-0.0367664135992527,
-0.11017907410860062,
0.05451679974794388,
-0.06131463497877121,
0.1515970379114151,
0.0918351337313652,
-0.06291328370571136,
-0.005600310862064362,
-0.05213579535484314,
0.0442471019923687,
0.025478530675172806,
0.14182385802268982,
0.025127029046416283,
-0.20204679667949677,
0.024187231436371803,
0.005215669982135296,
0.10994672030210495,
-0.24092787504196167,
-0.08275723457336426,
0.07658125460147858,
-0.02126235142350197,
-0.06236131489276886,
0.11044321954250336,
0.08412113040685654,
0.02551858127117157,
-0.033279940485954285,
-0.14078496396541595,
-0.026988914236426353,
0.13672243058681488,
-0.1737520545721054,
-0.019667701795697212
] |
null | null | null | <!-- markdownlint-disable first-line-h1 -->
<!-- markdownlint-disable html -->
<div align="center">
<h1>
<img src="image/huozi-logo.jpg" width="30" /> 活字通用大模型
</h1>
</div>
</p>
<div align="center">
<a href="https://github.com/HIT-SCIR/huozi/pulls">
<image src="https://img.shields.io/badge/PRs-welcome-brightgreen">
</a>
<a href="https://github.com/HIT-SCIR/huozi/pulls">
<image src="https://img.shields.io/badge/License-Apache_2.0-green.svg">
</a>
<!-- <h4 align="center">
<p>
<b>中文</b> |
<a href="https://github.com/HIT-SCIR/huozi/blob/main/README_EN.md">English</a>
<p>
</h4> -->
</div>
## 🔖 目录
|章节|说明|
|---|---|
|[💁🏻♂ 开源清单](#-开源清单)|本仓库开源项目清单|
|[💡 模型介绍](#-模型介绍)|简要介绍活字模型结构和训练过程|
|[📥 模型下载](#-模型下载)|活字模型下载链接|
|[💻 模型推理](#-模型推理)|活字模型推理样例,包括vLLM推理加速、llama.cpp量化推理等框架的使用流程|
|[📈 模型性能](#-模型性能)|活字模型在主流评测任务上的性能|
|[🗂 生成样例](#-生成样例)|活字模型实际生成效果样例|
## 💁🏻♂ 开源清单

- **活字 3.0**: [[模型权重](#-模型下载)]
- 活字3.0为一个稀疏混合专家模型,支持32K上下文,具有丰富的中、英文知识和强大的数学推理、代码生成能力。活字3.0较旧版活字具有更强的指令遵循能力和安全性。
- **中文MT-Bench**: [[数据集](data/mt-bench-zh/)]
- 本数据集是英文MT-Bench对话能力评测数据集的中文版。它包含了一系列多轮对话问题,每一组问题都经过了精心的人工校对,并为适应中文语境进行了必要的调整。
- **《ChatGPT 调研报告》**: [[PDF](https://github.com/HIT-SCIR/huozi/blob/main/pdf/chatgpt_book.pdf)]
- 哈工大自然语言处理研究所组织多位老师和同学撰写了本调研报告,从技术原理、应用场景、未来发展等方面对ChatGPT进行了尽量详尽的介绍及总结。
- **活字 2.0**: [[模型权重](https://huggingface.co/HIT-SCIR/huozi-7b-rlhf)] [[RLHF数据](data/huozi-rlhf/huozi_rlhf_data.csv)]
- 在活字1.0基础上,通过人类反馈的强化学习(RLHF)进一步优化了模型回复质量,使其更加符合人类偏好。相较于上一个版本平均长度明显提高,遵从指令的能力更强,逻辑更加清晰。
- 16.9k 人工标注的偏好数据,回复来自活字模型,可以用于训练奖励模型。
- **活字 1.0**: [[模型权重](https://huggingface.co/HIT-SCIR/huozi-7b-sft)]
- 在Bloom模型的基础上,在大约 150 亿 tokens 上进行指令微调训练得到的模型,具有更强的指令遵循能力、更好的安全性。
## 💡 模型介绍
大规模语言模型(LLM)在自然语言处理领域取得了显著的进展,并在广泛的应用场景中展现了其强大的潜力。这一技术不仅吸引了学术界的广泛关注,也成为了工业界的热点。在此背景下,哈尔滨工业大学社会计算与信息检索研究中心(HIT-SCIR)近期推出了最新成果——**活字3.0**,致力于为自然语言处理的研究和实际应用提供更多可能性和选择。
活字3.0是基于Chinese-Mixtral-8x7B,在大约30万行指令数据上微调得到的模型。该模型支持**32K上下文**,能够有效处理长文本。活字3.0继承了基座模型丰富的**中英文知识**,并在**数学推理**、**代码生成**等任务上具有强大性能。经过指令微调,活字3.0还在**指令遵循能力**和**安全性**方面实现了显著提升。
此外,我们开源了**中文MT-Bench数据集**。这是一个中文开放问题集,包括80组对话任务,用于评估模型的多轮对话和指令遵循能力。该数据集是根据原始MT-Bench翻译得来的,每组问题均经过人工校对和中文语境下的适当调整。我们还对原始MT-Bench中的部分错误答案进行了修正。
> [!IMPORTANT]
> 活字系列模型仍然可能生成包含事实性错误的误导性回复或包含偏见/歧视的有害内容,请谨慎鉴别和使用生成的内容,请勿将生成的有害内容传播至互联网。
### 模型结构
活字3.0是一个稀疏混合专家模型(SMoE),使用了Mixtral-8x7B的模型结构。它区别于LLaMA、BLOOM等常见模型,活字3.0的每个前馈神经网络(FFN)层被替换为了“专家层”,该层包含8个FFN和一个“路由器”。这种设计使得模型在推理过程中,可以独立地将每个Token路由到最适合处理它的两个专家中。活字3.0共拥有46.7B个参数,但得益于其稀疏激活的特性,实际推理时仅需激活13B参数,有效提升了计算效率和处理速度。

### 训练过程
由于Mixtral-8x7B词表不支持中文,因此对中文的编解码效率较低,限制了中文场景下的实用性。我们首先基于Mixtral-8x7B进行了中文扩词表增量预训练,显著提高了模型对中文的编解码效率,并使模型具备了强大的中文生成和理解能力。这项成果名为[Chinese-Mixtral-8x7B](https://github.com/HIT-SCIR/Chinese-Mixtral-8x7B),我们已于2024年1月18日开源了其模型权重和训练代码。基于此,我们进一步对模型进行指令微调,最终推出了活字3.0。这一版本的中文编码、指令遵循、安全回复等能力都有显著提升。
## 📥 模型下载
|模型名称|文件大小|下载地址|备注|
|:---:|:---:|:---:|:---:|
|huozi3|88GB|[🤗HuggingFace](https://huggingface.co/HIT-SCIR/huozi3)<br>[ModelScope](https://modelscope.cn/models/HIT-SCIR/huozi3/summary)|活字3.0 完整模型|
|huozi3-gguf|25GB|[🤗HuggingFace](https://huggingface.co/HIT-SCIR/huozi3-gguf)<br>[ModelScope](https://modelscope.cn/models/HIT-SCIR/huozi3-gguf/summary)|活字3.0 GGUF版本,适用于llama.cpp等推理框架|
|huozi3-awq|24GB|[🤗HuggingFace](https://huggingface.co/HIT-SCIR/huozi3-awq)<br>[ModelScope](https://modelscope.cn/models/HIT-SCIR/huozi3-awq/summary)|活字3.0 AWQ版本,适用于AutoAWQ等推理框架|
如果您希望微调活字3.0或Chinese-Mixtral-8x7B,请参考[此处训练代码](https://github.com/HIT-SCIR/Chinese-Mixtral-8x7B?tab=readme-ov-file#%E5%BE%AE%E8%B0%83)。
## 💻 模型推理
### Quick Start
活字3.0采用ChatML格式的prompt模板,格式为:
```
<|beginofutterance|>系统
{system prompt}<|endofutterance|>
<|beginofutterance|>用户
{input}<|endofutterance|>
<|beginofutterance|>助手
{output}<|endofutterance|>
```
使用活字3.0进行推理的示例代码如下:
```python
# quickstart.py
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "HIT-SCIR/huozi3"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
model_id,
attn_implementation="flash_attention_2",
torch_dtype=torch.bfloat16,
device_map="auto",
)
text = """<|beginofutterance|>系统
你是一个智能助手<|endofutterance|>
<|beginofutterance|>用户
请你用python写一段快速排序的代码<|endofutterance|>
<|beginofutterance|>助手
"""
inputs = tokenizer(text, return_tensors="pt").to(0)
outputs = model.generate(
**inputs,
eos_token_id=57001,
temperature=0.8,
top_p=0.9,
max_new_tokens=2048,
)
print(tokenizer.decode(outputs[0], skip_special_tokens=False))
```
活字3.0支持全部Mixtral模型生态,包括Transformers、vLLM、llama.cpp、AutoAWQ、Text generation web UI等框架。
如果您在下载模型时遇到网络问题,可以使用我们在[ModelScope](#modelscope-模型推理)上提供的检查点。
<details>
<summary>
#### Transformers 模型推理 + 流式生成
</summary>
transformers支持为tokenizer添加聊天模板,并支持流式生成。示例代码如下:
```python
# example/transformers-stream/stream.py
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer
model_id = "HIT-SCIR/huozi3"
model = AutoModelForCausalLM.from_pretrained(
model_id,
attn_implementation="flash_attention_2",
torch_dtype=torch.bfloat16,
device_map="auto",
)
tokenizer = AutoTokenizer.from_pretrained(model_id)
tokenizer.chat_template = """{% for message in messages %}{{'<|beginofutterance|>' + message['role'] + '\n' + message['content']}}{% if (loop.last and add_generation_prompt) or not loop.last %}{{ '<|endofutterance|>' + '\n'}}{% endif %}{% endfor %}
{% if add_generation_prompt and messages[-1]['role'] != '助手' %}{{ '<|beginofutterance|>助手\n' }}{% endif %}"""
chat = [
{"role": "系统", "content": "你是一个智能助手"},
{"role": "用户", "content": "请你用python写一段快速排序的代码"},
]
inputs = tokenizer.apply_chat_template(
chat,
tokenize=True,
add_generation_prompt=True,
return_tensors="pt",
).to(0)
stream_output = model.generate(
inputs,
streamer=TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True),
eos_token_id=57001,
temperature=0.8,
top_p=0.9,
max_new_tokens=2048,
)
```
</details>
<details>
<summary>
#### ModelScope 模型推理
</summary>
ModelScope的接口与Transformers非常相似,只需将transformers替换为modelscope即可:
```diff
# example/modelscope-generate/generate.py
import torch
- from transformers import AutoModelForCausalLM, AutoTokenizer
+ from modelscope import AutoTokenizer, AutoModelForCausalLM
model_id = "HIT-SCIR/huozi3"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
model_id,
attn_implementation="flash_attention_2",
torch_dtype=torch.bfloat16,
device_map="auto",
)
text = """<|beginofutterance|>系统
你是一个智能助手<|endofutterance|>
<|beginofutterance|>用户
请你用python写一段快速排序的代码<|endofutterance|>
<|beginofutterance|>助手
"""
inputs = tokenizer(text, return_tensors="pt").to(0)
outputs = model.generate(
**inputs,
eos_token_id=57001,
temperature=0.8,
top_p=0.9,
max_new_tokens=2048,
)
print(tokenizer.decode(outputs[0], skip_special_tokens=False))
```
</details>
<details>
<summary>
#### vLLM 推理加速
</summary>
活字3.0支持通过vLLM实现推理加速,示例代码如下:
```python
# example/vllm-generate/generate.py
from vllm import LLM, SamplingParams
prompts = [
"""<|beginofutterance|>系统
你是一个智能助手<|endofutterance|>
<|beginofutterance|>用户
请你用python写一段快速排序的代码<|endofutterance|>
<|beginofutterance|>助手
""",
]
sampling_params = SamplingParams(
temperature=0.8, top_p=0.95, stop_token_ids=[57001], max_tokens=2048
)
llm = LLM(
model="HIT-SCIR/huozi3",
tensor_parallel_size=4,
)
outputs = llm.generate(prompts, sampling_params)
for output in outputs:
prompt = output.prompt
generated_text = output.outputs[0].text
print(generated_text)
```
</details>
<details>
<summary>
#### 部署 OpenAI API Server
</summary>
活字3.0可以部署为支持OpenAI API协议的服务,这使得活字3.0可以直接通过OpenAI API进行调用。
环境准备:
```shell
$ pip install vllm openai
```
启动服务:
```shell
$ python -m vllm.entrypoints.openai.api_server --model /path/to/huozi3/checkpoint --served-model-name huozi --chat-template template.jinja --tensor-parallel-size 8 --response-role 助手 --max-model-len 2048
```
使用OpenAI API发送请求:
```python
# example/openai-api/openai-client.py
from openai import OpenAI
openai_api_key = "EMPTY"
openai_api_base = "http://localhost:8000/v1"
client = OpenAI(
api_key=openai_api_key,
base_url=openai_api_base,
)
chat_response = client.chat.completions.create(
model="huozi",
messages=[
{"role": "系统", "content": "你是一个智能助手"},
{"role": "用户", "content": "请你用python写一段快速排序的代码"},
],
extra_body={"stop_token_ids": [57001]},
)
print("Chat response:", chat_response.choices[0].message.content)
```
下面是一个使用OpenAI API + Gradio + 流式生成的示例代码:
```python
# example/openai-api/openai-client-gradio.py
from openai import OpenAI
import gradio as gr
openai_api_key = "EMPTY"
openai_api_base = "http://localhost:8000/v1"
client = OpenAI(
api_key=openai_api_key,
base_url=openai_api_base,
)
def predict(message, history):
history_openai_format = [
{"role": "系统", "content": "你是一个智能助手"},
]
for human, assistant in history:
history_openai_format.append({"role": "用户", "content": human})
history_openai_format.append({"role": "助手", "content": assistant})
history_openai_format.append({"role": "用户", "content": message})
models = client.models.list()
stream = client.chat.completions.create(
model=models.data[0].id,
messages=history_openai_format,
temperature=0.8,
stream=True,
extra_body={"repetition_penalty": 1, "stop_token_ids": [57001]},
)
partial_message = ""
for chunk in stream:
partial_message += chunk.choices[0].delta.content or ""
yield partial_message
gr.ChatInterface(predict).queue().launch()
```
</details>
### 量化推理
活字3.0支持量化推理,下表为活字3.0在各个量化框架下显存占用量:
|量化方法|显存占用|
|:---:|:---:|
|无|95GB|
|AWQ|32GB|
|GGUF(q4_0)|28GB|
|GGUF(q2_k)|18GB|
|GGUF(q2_k, offload 16层)|9.6GB|
<details>
<summary>
#### GGUF 格式
</summary>
GGUF格式旨在快速加载和保存模型,由llama.cpp团队推出。我们已经提供了[GGUF格式的活字3.0](https://huggingface.co/HIT-SCIR/huozi3-gguf)。
您也可以手动将HuggingFace格式的活字3.0转换到GGUF格式,以使用其他的量化方法。
##### Step 1 环境准备
首先需要下载llama.cpp的源码。我们在仓库中提供了llama.cpp的submodule,这个版本的llama.cpp已经过测试,可以成功进行推理:
```shell
$ git clone --recurse-submodules https://github.com/HIT-SCIR/huozi
$ cd examples/llama.cpp
```
您也可以下载最新版本的llama.cpp源码:
```shell
$ git clone https://github.com/ggerganov/llama.cpp.git
$ cd llama.cpp
```
然后需要进行编译。根据您的硬件平台,编译命令有细微差异:
```shell
$ make # 用于纯CPU推理
$ make LLAMA_CUBLAS=1 # 用于GPU推理
$ LLAMA_METAL=1 make # 用于Apple Silicon,暂未经过测试
```
##### Step 2 格式转换(可选)
以下命令需要在`llama.cpp/`目录下:
```shell
# 转换为GGUF格式
$ python convert.py --outfile /path/to/huozi-gguf/huozi3.gguf /path/to/huozi3
# 进行GGUF格式的q4_0量化
$ quantize /path/to/huozi-gguf/huozi3.gguf /path/to/huozi-gguf/huozi3-q4_0.gguf q4_0
```
##### Step 3 开始推理
以下命令需要在`llama.cpp/`目录下:
```shell
$ main -m /path/to/huozi-gguf/huozi3-q4_0.gguf --color --interactive-first -c 2048 -t 6 --temp 0.2 --repeat_penalty 1.1 -ngl 999 --in-prefix "<|beginofutterance|>用户\n" --in-suffix "<|endofutterance|>\n<|beginofutterance|>助手" -r "<|endofutterance|>"
```
`-ngl`参数表示向GPU中offload的层数,降低这个值可以缓解GPU显存压力。经过我们的实际测试,q2_k量化的模型offload 16层,显存占用可降低至9.6GB,可在消费级GPU上运行模型:
```shell
$ main -m /path/to/huozi-gguf/huozi3-q2_k.gguf --color --interactive-first -c 2048 -t 6 --temp 0.2 --repeat_penalty 1.1 -ngl 16 --in-prefix "<|beginofutterance|>用户\n" --in-suffix "<|endofutterance|>\n<|beginofutterance|>助手" -r "<|endofutterance|>"
```
关于`main`的更多参数,可以参考llama.cpp的[官方文档](https://github.com/ggerganov/llama.cpp/tree/master/examples/main)。
</details>
<details>
<summary>
#### AWQ 格式
</summary>
AWQ是一种量化模型的存储格式。我们已经提供了[AWQ格式的活字3.0](https://huggingface.co/HIT-SCIR/huozi3-awq),您也可以手动将HuggingFace格式的活字3.0转换到AWQ格式。
##### Step 1 格式转换(可选)
```python
# example/autoawq-generate/quant.py
from awq import AutoAWQForCausalLM
from transformers import AutoTokenizer
model_path = "/path/to/huozi3"
quant_path = "/path/to/save/huozi3-awq"
modules_to_not_convert = ["gate"]
quant_config = {
"zero_point": True,
"q_group_size": 128,
"w_bit": 4,
"version": "GEMM",
"modules_to_not_convert": modules_to_not_convert,
}
model = AutoAWQForCausalLM.from_pretrained(
model_path,
safetensors=True,
**{"low_cpu_mem_usage": True},
)
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
model.quantize(
tokenizer,
quant_config=quant_config,
modules_to_not_convert=modules_to_not_convert,
)
model.save_quantized(quant_path)
tokenizer.save_pretrained(quant_path)
print(f'Model is quantized and saved at "{quant_path}"')
```
##### Step 2 开始推理
在获取到AWQ格式的模型权重后,可以使用AutoAWQForCausalLM代替AutoModelForCausalLM加载模型。示例代码如下:
```diff
# example/autoawq-generate/generate.py
import torch
+ from awq import AutoAWQForCausalLM
from transformers import AutoTokenizer, TextStreamer
- model_id = "HIT-SCIR/huozi3"
+ model_id = "HIT-SCIR/huozi3-awq" # or model_id = "/path/to/saved/huozi3-awq"
+ model = AutoAWQForCausalLM.from_quantized(model_id, fuse_layers=True)
- model = AutoModelForCausalLM.from_pretrained(
- model_id,
- attn_implementation="flash_attention_2",
- torch_dtype=torch.bfloat16,
- device_map="auto",
- )
tokenizer = AutoTokenizer.from_pretrained(model_id)
tokenizer.chat_template = """{% for message in messages %}{{'<|beginofutterance|>' + message['role'] + '\n' + message['content']}}{% if (loop.last and add_generation_prompt) or not loop.last %}{{ '<|endofutterance|>' + '\n'}}{% endif %}{% endfor %}
{% if add_generation_prompt and messages[-1]['role'] != '助手' %}{{ '<|beginofutterance|>助手\n' }}{% endif %}"""
chat = [
{"role": "系统", "content": "你是一个智能助手"},
{"role": "用户", "content": "请你用python写一段快速排序的代码"},
]
inputs = tokenizer.apply_chat_template(
chat,
tokenize=True,
add_generation_prompt=True,
return_tensors="pt",
).to(0)
stream_output = model.generate(
inputs,
streamer=TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True),
eos_token_id=57001,
temperature=0.8,
top_p=0.9,
max_new_tokens=2048,
)
```
</details>
## 📈 模型性能

针对大模型综合能力评价,我们分别使用以下评测数据集对活字3.0进行评测:
- C-Eval:一个全面的中文基础模型评估套件。它包含了13948个多项选择题,涵盖了52个不同的学科和四个难度级别。
- CMMLU:一个综合性的中文评估基准,专门用于评估语言模型在中文语境下的知识和推理能力,涵盖了从基础学科到高级专业水平的67个主题。
- GAOKAO:一个以中国高考题目为数据集,旨在提供和人类对齐的,直观,高效地测评大模型语言理解能力、逻辑推理能力的测评框架。
- MMLU:一个包含57个多选任务的英文评测数据集,涵盖了初等数学、美国历史、计算机科学、法律等,难度覆盖高中水平到专家水平,是目前主流的LLM评测数据集之一。
- HellaSwag:一个极具挑战的英文NLI评测数据集,每一个问题都需要对上下文进行深入理解,而不能基于常识进行回答。
- GSM8K:一个高质量的小学数学应用题的数据集,这些问题需要 2 到 8 个步骤来解决,解决方案主要涉及使用基本算术运算,可用于评价多步数学推理能力。
- HumanEval:一个由 164 个原创编程问题组成的数据集,通过衡量从文档字符串生成程序的功能正确性,来够评估语言理解、算法和简单的数学能力。
- MT-Bench:一个开放的英文问题集,包括80个多轮对话任务,用于评估聊天机器人的多轮对话和指令遵循能力,并通过大模型裁判(GPT-4)对模型回答进行打分。
- MT-Bench-zh:我们根据MT-Bench翻译得来的中文问题集,每组问题均经过人工校对和中文语境下的适当调整。我们已在[此处](data/mt-bench-zh/)开源MT-Bench-zh数据集。
- MT-Bench-safety:我们手工构造的安全数据集,包括暴力、色情、敏感等风险内容。该数据集为封闭数据集。
活字3.0在推理时仅激活13B参数。下表为活字3.0与其他13B规模的中文模型以及旧版活字在各个评测数据集上的结果:
<!-- | 模型名称 | 模型结构 | C-Eval<br>(中文) | CMMLU<br>(中文) | GAOKAO<br>(中文) | MT-Bench-zh<br>(中文对话) | MT-Bench-safety<br>(中文安全) | MMLU<br>(英文) | HellaSwag<br>(英文) | MT-Bench<br>(英文对话) | GSM8K<br>(数学) | HumanEval<br>(代码) |
|---------------------------------------------|---------|--------------|-------------|---------------|--------------------------|-----------------------------|------------|------------------|-----------------------|-------------|-----------------|
| baichuan-inc/Baichuan2-13B-Chat v2 | Baichuan| 56.13 | 58.50 | 48.99 | 6.74 | 8.30 | 54.50 | 51.19 | 6.59 | 25.17 | 20.12 |
| wangrongsheng/Aurora-Plus | Mixtral | 47.67 | 48.75 | 35.05 | 5.47 | 6.70 | 67.80 | 78.27 | 7.13 | 66.26 | 27.44 |
| TigerResearch/tigerbot-13b-chat-v5 | LLaMA | 49.78 | 51.28 | 41.31 | 5.98 | 7.63 | 56.34 | 35.17 | 4.88 | 66.19 | 14.63 |
| hfl/chinese-alpaca-2-13b | LLaMA | 43.47 | 44.53 | 25.94 | 5.77 | 8.13 | 53.05 | 56.85 | 6.24 | 32.75 | 14.02 |
| 活字1.0 | BLOOM | 37.27 | 36.24 | 19.72 | 4.48 | 7.18 | 39.68 | 33.21 | 4.34 | 21.99 | 13.41 |
| 活字2.0 | BLOOM | 32.05 | 34.68 | 22.97 | 5.08 | 6.68 | 38.04 | 33.34 | 4.79 | 19.86 | 6.71 |
| **活字3.0(最新版本)** | Mixtral | 51.82 | 51.06 | 41.21 | 6.29 | 7.58 | 69.48 | 65.18 | 7.62 | 65.81 | 40.85 | -->

> 我们在C-Eval、CMMLU、MMLU采用5-shot,GSM8K采用4-shot,HellaSwag、HumanEval采用0-shot,HumanEval采用pass@1指标。所有测试均采用greedy策略。
>
> 我们使用OpenCompass作为评测框架,commit hash为[4c87e77](https://github.com/open-compass/opencompass/tree/4c87e777d855636b9eda7ec87bcbbf12b62caed3)。评测代码位于[此处](./evaluate/)。
根据上表中的测试结果,活字3.0较旧版活字取得了巨大的性能提升。在中文知识方面,活字3.0达到了与Tigerbot-13B-chat-v5相当的性能,并是在中文对话和指令遵循方面表现得更加优秀。在英文知识方面,得益于原版Mixtral-8x7B的强大性能,活字3.0超过了Baichuan2-13B-Chat v2和LLaMA系列的扩词表模型,并在英文对话和指令遵循能力上达到了较高水平。在数学推理和代码生成任务上,活字3.0均展现出强大的性能,这说明活字3.0对复杂问题的深层次理解、多步推理、以及结构化信息处理等方面具有较强水平。由于我们采用了较高质量的代码数据集,活字3.0的代码生成能力也超越了同为Mixtral结构的Aurora-Plus模型。
## 🗂 生成样例
下面是活字3.0在MT-Bench-zh评测集上的生成效果展示,并与活字2.0(RLHF版本)进行对比:





## <img src="https://cdn.jsdelivr.net/gh/LightChen233/blog-img/folders.png" width="25" /> 开源协议
对本仓库源码的使用遵循开源许可协议 [Apache 2.0](https://github.com/HIT-SCIR/huozi/blob/main/LICENSE)。
活字支持商用。如果将活字模型或其衍生品用作商业用途,请您按照如下方式联系许可方,以进行登记并向许可方申请书面授权:联系邮箱:<[email protected]>。
## <img src="https://cdn.jsdelivr.net/gh/LightChen233/blog-img/notes.png" width="25" /> Citation
### 活字大模型
```latex
@misc{huozi,
author = {Huozi-Team}.
title = {Huozi: Leveraging Large Language Models for Enhanced Open-Domain Chatting}
year = {2024},
publisher = {GitHub},
journal = {GitHub repository}
howpublished = {\url{https://github.com/HIT-SCIR/huozi}}
}
```
## <img src="https://cdn.jsdelivr.net/gh/LightChen233/blog-img/star.png" width="25" /> Star History
[](https://star-history.com/#HIT-SCIR/huozi&Date)
| {} | null | HIT-SCIR/huozi3-gguf | [
"gguf",
"region:us"
] | 2024-02-08T08:33:16+00:00 | [] | [] | TAGS
#gguf #region-us
|
活字通用大模型
=========
目录
--
开源清单
------

* 活字 3.0: [模型权重]
+ 活字3.0为一个稀疏混合专家模型,支持32K上下文,具有丰富的中、英文知识和强大的数学推理、代码生成能力。活字3.0较旧版活字具有更强的指令遵循能力和安全性。
* 中文MT-Bench: [数据集]
+ 本数据集是英文MT-Bench对话能力评测数据集的中文版。它包含了一系列多轮对话问题,每一组问题都经过了精心的人工校对,并为适应中文语境进行了必要的调整。
* 《ChatGPT 调研报告》: [PDF]
+ 哈工大自然语言处理研究所组织多位老师和同学撰写了本调研报告,从技术原理、应用场景、未来发展等方面对ChatGPT进行了尽量详尽的介绍及总结。
* 活字 2.0: [模型权重] [RLHF数据]
+ 在活字1.0基础上,通过人类反馈的强化学习(RLHF)进一步优化了模型回复质量,使其更加符合人类偏好。相较于上一个版本平均长度明显提高,遵从指令的能力更强,逻辑更加清晰。
+ 16.9k 人工标注的偏好数据,回复来自活字模型,可以用于训练奖励模型。
* 活字 1.0: [模型权重]
+ 在Bloom模型的基础上,在大约 150 亿 tokens 上进行指令微调训练得到的模型,具有更强的指令遵循能力、更好的安全性。
模型介绍
----
大规模语言模型(LLM)在自然语言处理领域取得了显著的进展,并在广泛的应用场景中展现了其强大的潜力。这一技术不仅吸引了学术界的广泛关注,也成为了工业界的热点。在此背景下,哈尔滨工业大学社会计算与信息检索研究中心(HIT-SCIR)近期推出了最新成果——活字3.0,致力于为自然语言处理的研究和实际应用提供更多可能性和选择。
活字3.0是基于Chinese-Mixtral-8x7B,在大约30万行指令数据上微调得到的模型。该模型支持32K上下文,能够有效处理长文本。活字3.0继承了基座模型丰富的中英文知识,并在数学推理、代码生成等任务上具有强大性能。经过指令微调,活字3.0还在指令遵循能力和安全性方面实现了显著提升。
此外,我们开源了中文MT-Bench数据集。这是一个中文开放问题集,包括80组对话任务,用于评估模型的多轮对话和指令遵循能力。该数据集是根据原始MT-Bench翻译得来的,每组问题均经过人工校对和中文语境下的适当调整。我们还对原始MT-Bench中的部分错误答案进行了修正。
>
> [!IMPORTANT]
> 活字系列模型仍然可能生成包含事实性错误的误导性回复或包含偏见/歧视的有害内容,请谨慎鉴别和使用生成的内容,请勿将生成的有害内容传播至互联网。
>
>
>
### 模型结构
活字3.0是一个稀疏混合专家模型(SMoE),使用了Mixtral-8x7B的模型结构。它区别于LLaMA、BLOOM等常见模型,活字3.0的每个前馈神经网络(FFN)层被替换为了“专家层”,该层包含8个FFN和一个“路由器”。这种设计使得模型在推理过程中,可以独立地将每个Token路由到最适合处理它的两个专家中。活字3.0共拥有46.7B个参数,但得益于其稀疏激活的特性,实际推理时仅需激活13B参数,有效提升了计算效率和处理速度。

### 训练过程
由于Mixtral-8x7B词表不支持中文,因此对中文的编解码效率较低,限制了中文场景下的实用性。我们首先基于Mixtral-8x7B进行了中文扩词表增量预训练,显著提高了模型对中文的编解码效率,并使模型具备了强大的中文生成和理解能力。这项成果名为Chinese-Mixtral-8x7B,我们已于2024年1月18日开源了其模型权重和训练代码。基于此,我们进一步对模型进行指令微调,最终推出了活字3.0。这一版本的中文编码、指令遵循、安全回复等能力都有显著提升。
模型下载
----
如果您希望微调活字3.0或Chinese-Mixtral-8x7B,请参考此处训练代码。
模型推理
----
### Quick Start
活字3.0采用ChatML格式的prompt模板,格式为:
使用活字3.0进行推理的示例代码如下:
活字3.0支持全部Mixtral模型生态,包括Transformers、vLLM、URL、AutoAWQ、Text generation web UI等框架。
如果您在下载模型时遇到网络问题,可以使用我们在ModelScope上提供的检查点。
#### Transformers 模型推理 + 流式生成
transformers支持为tokenizer添加聊天模板,并支持流式生成。示例代码如下:
#### ModelScope 模型推理
ModelScope的接口与Transformers非常相似,只需将transformers替换为modelscope即可:
#### vLLM 推理加速
活字3.0支持通过vLLM实现推理加速,示例代码如下:
#### 部署 OpenAI API Server
活字3.0可以部署为支持OpenAI API协议的服务,这使得活字3.0可以直接通过OpenAI API进行调用。
环境准备:
启动服务:
使用OpenAI API发送请求:
下面是一个使用OpenAI API + Gradio + 流式生成的示例代码:
### 量化推理
活字3.0支持量化推理,下表为活字3.0在各个量化框架下显存占用量:
#### GGUF 格式
GGUF格式旨在快速加载和保存模型,由llama.cpp团队推出。我们已经提供了GGUF格式的活字3.0。
您也可以手动将HuggingFace格式的活字3.0转换到GGUF格式,以使用其他的量化方法。
##### Step 1 环境准备
首先需要下载llama.cpp的源码。我们在仓库中提供了llama.cpp的submodule,这个版本的llama.cpp已经过测试,可以成功进行推理:
您也可以下载最新版本的llama.cpp源码:
然后需要进行编译。根据您的硬件平台,编译命令有细微差异:
##### Step 2 格式转换(可选)
以下命令需要在'URL'目录下:
##### Step 3 开始推理
以下命令需要在'URL'目录下:
'-ngl'参数表示向GPU中offload的层数,降低这个值可以缓解GPU显存压力。经过我们的实际测试,q2\_k量化的模型offload 16层,显存占用可降低至9.6GB,可在消费级GPU上运行模型:
关于'main'的更多参数,可以参考llama.cpp的官方文档。
#### AWQ 格式
AWQ是一种量化模型的存储格式。我们已经提供了AWQ格式的活字3.0,您也可以手动将HuggingFace格式的活字3.0转换到AWQ格式。
##### Step 1 格式转换(可选)
##### Step 2 开始推理
在获取到AWQ格式的模型权重后,可以使用AutoAWQForCausalLM代替AutoModelForCausalLM加载模型。示例代码如下:
模型性能
----

针对大模型综合能力评价,我们分别使用以下评测数据集对活字3.0进行评测:
* C-Eval:一个全面的中文基础模型评估套件。它包含了13948个多项选择题,涵盖了52个不同的学科和四个难度级别。
* CMMLU:一个综合性的中文评估基准,专门用于评估语言模型在中文语境下的知识和推理能力,涵盖了从基础学科到高级专业水平的67个主题。
* GAOKAO:一个以中国高考题目为数据集,旨在提供和人类对齐的,直观,高效地测评大模型语言理解能力、逻辑推理能力的测评框架。
* MMLU:一个包含57个多选任务的英文评测数据集,涵盖了初等数学、美国历史、计算机科学、法律等,难度覆盖高中水平到专家水平,是目前主流的LLM评测数据集之一。
* HellaSwag:一个极具挑战的英文NLI评测数据集,每一个问题都需要对上下文进行深入理解,而不能基于常识进行回答。
* GSM8K:一个高质量的小学数学应用题的数据集,这些问题需要 2 到 8 个步骤来解决,解决方案主要涉及使用基本算术运算,可用于评价多步数学推理能力。
* HumanEval:一个由 164 个原创编程问题组成的数据集,通过衡量从文档字符串生成程序的功能正确性,来够评估语言理解、算法和简单的数学能力。
* MT-Bench:一个开放的英文问题集,包括80个多轮对话任务,用于评估聊天机器人的多轮对话和指令遵循能力,并通过大模型裁判(GPT-4)对模型回答进行打分。
* MT-Bench-zh:我们根据MT-Bench翻译得来的中文问题集,每组问题均经过人工校对和中文语境下的适当调整。我们已在此处开源MT-Bench-zh数据集。
* MT-Bench-safety:我们手工构造的安全数据集,包括暴力、色情、敏感等风险内容。该数据集为封闭数据集。
活字3.0在推理时仅激活13B参数。下表为活字3.0与其他13B规模的中文模型以及旧版活字在各个评测数据集上的结果:

>
> 我们在C-Eval、CMMLU、MMLU采用5-shot,GSM8K采用4-shot,HellaSwag、HumanEval采用0-shot,HumanEval采用pass@1指标。所有测试均采用greedy策略。
>
>
> 我们使用OpenCompass作为评测框架,commit hash为4c87e77。评测代码位于此处。
>
>
>
根据上表中的测试结果,活字3.0较旧版活字取得了巨大的性能提升。在中文知识方面,活字3.0达到了与Tigerbot-13B-chat-v5相当的性能,并是在中文对话和指令遵循方面表现得更加优秀。在英文知识方面,得益于原版Mixtral-8x7B的强大性能,活字3.0超过了Baichuan2-13B-Chat v2和LLaMA系列的扩词表模型,并在英文对话和指令遵循能力上达到了较高水平。在数学推理和代码生成任务上,活字3.0均展现出强大的性能,这说明活字3.0对复杂问题的深层次理解、多步推理、以及结构化信息处理等方面具有较强水平。由于我们采用了较高质量的代码数据集,活字3.0的代码生成能力也超越了同为Mixtral结构的Aurora-Plus模型。
生成样例
----
下面是活字3.0在MT-Bench-zh评测集上的生成效果展示,并与活字2.0(RLHF版本)进行对比:





<img src="URL width="25" /> 开源协议
--------------------------------
对本仓库源码的使用遵循开源许可协议 Apache 2.0。
活字支持商用。如果将活字模型或其衍生品用作商业用途,请您按照如下方式联系许可方,以进行登记并向许可方申请书面授权:联系邮箱:[jngao@URL](mailto:jngao@URL)。
<img src="URL width="25" /> Citation
------------------------------------
### 活字大模型
<img src="URL width="25" /> Star History
----------------------------------------
,使用了Mixtral-8x7B的模型结构。它区别于LLaMA、BLOOM等常见模型,活字3.0的每个前馈神经网络(FFN)层被替换为了“专家层”,该层包含8个FFN和一个“路由器”。这种设计使得模型在推理过程中,可以独立地将每个Token路由到最适合处理它的两个专家中。活字3.0共拥有46.7B个参数,但得益于其稀疏激活的特性,实际推理时仅需激活13B参数,有效提升了计算效率和处理速度。\n\n\n",
"### 训练过程\n\n\n由于Mixtral-8x7B词表不支持中文,因此对中文的编解码效率较低,限制了中文场景下的实用性。我们首先基于Mixtral-8x7B进行了中文扩词表增量预训练,显著提高了模型对中文的编解码效率,并使模型具备了强大的中文生成和理解能力。这项成果名为Chinese-Mixtral-8x7B,我们已于2024年1月18日开源了其模型权重和训练代码。基于此,我们进一步对模型进行指令微调,最终推出了活字3.0。这一版本的中文编码、指令遵循、安全回复等能力都有显著提升。\n\n\n模型下载\n----\n\n\n\n如果您希望微调活字3.0或Chinese-Mixtral-8x7B,请参考此处训练代码。\n\n\n模型推理\n----",
"### Quick Start\n\n\n活字3.0采用ChatML格式的prompt模板,格式为:\n\n\n使用活字3.0进行推理的示例代码如下:\n\n\n活字3.0支持全部Mixtral模型生态,包括Transformers、vLLM、URL、AutoAWQ、Text generation web UI等框架。\n\n\n如果您在下载模型时遇到网络问题,可以使用我们在ModelScope上提供的检查点。",
"#### Transformers 模型推理 + 流式生成\n\n\n\ntransformers支持为tokenizer添加聊天模板,并支持流式生成。示例代码如下:",
"#### ModelScope 模型推理\n\n\n\nModelScope的接口与Transformers非常相似,只需将transformers替换为modelscope即可:",
"#### vLLM 推理加速\n\n\n\n活字3.0支持通过vLLM实现推理加速,示例代码如下:",
"#### 部署 OpenAI API Server\n\n\n\n活字3.0可以部署为支持OpenAI API协议的服务,这使得活字3.0可以直接通过OpenAI API进行调用。\n\n\n环境准备:\n\n\n启动服务:\n\n\n使用OpenAI API发送请求:\n\n\n下面是一个使用OpenAI API + Gradio + 流式生成的示例代码:",
"### 量化推理\n\n\n活字3.0支持量化推理,下表为活字3.0在各个量化框架下显存占用量:",
"#### GGUF 格式\n\n\n\nGGUF格式旨在快速加载和保存模型,由llama.cpp团队推出。我们已经提供了GGUF格式的活字3.0。\n\n\n您也可以手动将HuggingFace格式的活字3.0转换到GGUF格式,以使用其他的量化方法。",
"##### Step 1 环境准备\n\n\n首先需要下载llama.cpp的源码。我们在仓库中提供了llama.cpp的submodule,这个版本的llama.cpp已经过测试,可以成功进行推理:\n\n\n您也可以下载最新版本的llama.cpp源码:\n\n\n然后需要进行编译。根据您的硬件平台,编译命令有细微差异:",
"##### Step 2 格式转换(可选)\n\n\n以下命令需要在'URL'目录下:",
"##### Step 3 开始推理\n\n\n以下命令需要在'URL'目录下:\n\n\n'-ngl'参数表示向GPU中offload的层数,降低这个值可以缓解GPU显存压力。经过我们的实际测试,q2\\_k量化的模型offload 16层,显存占用可降低至9.6GB,可在消费级GPU上运行模型:\n\n\n关于'main'的更多参数,可以参考llama.cpp的官方文档。",
"#### AWQ 格式\n\n\n\nAWQ是一种量化模型的存储格式。我们已经提供了AWQ格式的活字3.0,您也可以手动将HuggingFace格式的活字3.0转换到AWQ格式。",
"##### Step 1 格式转换(可选)",
"##### Step 2 开始推理\n\n\n在获取到AWQ格式的模型权重后,可以使用AutoAWQForCausalLM代替AutoModelForCausalLM加载模型。示例代码如下:\n\n\n\n模型性能\n----\n\n\n\n\n\n针对大模型综合能力评价,我们分别使用以下评测数据集对活字3.0进行评测:\n\n\n* C-Eval:一个全面的中文基础模型评估套件。它包含了13948个多项选择题,涵盖了52个不同的学科和四个难度级别。\n* CMMLU:一个综合性的中文评估基准,专门用于评估语言模型在中文语境下的知识和推理能力,涵盖了从基础学科到高级专业水平的67个主题。\n* GAOKAO:一个以中国高考题目为数据集,旨在提供和人类对齐的,直观,高效地测评大模型语言理解能力、逻辑推理能力的测评框架。\n* MMLU:一个包含57个多选任务的英文评测数据集,涵盖了初等数学、美国历史、计算机科学、法律等,难度覆盖高中水平到专家水平,是目前主流的LLM评测数据集之一。\n* HellaSwag:一个极具挑战的英文NLI评测数据集,每一个问题都需要对上下文进行深入理解,而不能基于常识进行回答。\n* GSM8K:一个高质量的小学数学应用题的数据集,这些问题需要 2 到 8 个步骤来解决,解决方案主要涉及使用基本算术运算,可用于评价多步数学推理能力。\n* HumanEval:一个由 164 个原创编程问题组成的数据集,通过衡量从文档字符串生成程序的功能正确性,来够评估语言理解、算法和简单的数学能力。\n* MT-Bench:一个开放的英文问题集,包括80个多轮对话任务,用于评估聊天机器人的多轮对话和指令遵循能力,并通过大模型裁判(GPT-4)对模型回答进行打分。\n* MT-Bench-zh:我们根据MT-Bench翻译得来的中文问题集,每组问题均经过人工校对和中文语境下的适当调整。我们已在此处开源MT-Bench-zh数据集。\n* MT-Bench-safety:我们手工构造的安全数据集,包括暴力、色情、敏感等风险内容。该数据集为封闭数据集。\n\n\n活字3.0在推理时仅激活13B参数。下表为活字3.0与其他13B规模的中文模型以及旧版活字在各个评测数据集上的结果:\n\n\n\n\n\n\n> \n> 我们在C-Eval、CMMLU、MMLU采用5-shot,GSM8K采用4-shot,HellaSwag、HumanEval采用0-shot,HumanEval采用pass@1指标。所有测试均采用greedy策略。\n> \n> \n> 我们使用OpenCompass作为评测框架,commit hash为4c87e77。评测代码位于此处。\n> \n> \n> \n\n\n根据上表中的测试结果,活字3.0较旧版活字取得了巨大的性能提升。在中文知识方面,活字3.0达到了与Tigerbot-13B-chat-v5相当的性能,并是在中文对话和指令遵循方面表现得更加优秀。在英文知识方面,得益于原版Mixtral-8x7B的强大性能,活字3.0超过了Baichuan2-13B-Chat v2和LLaMA系列的扩词表模型,并在英文对话和指令遵循能力上达到了较高水平。在数学推理和代码生成任务上,活字3.0均展现出强大的性能,这说明活字3.0对复杂问题的深层次理解、多步推理、以及结构化信息处理等方面具有较强水平。由于我们采用了较高质量的代码数据集,活字3.0的代码生成能力也超越了同为Mixtral结构的Aurora-Plus模型。\n\n\n生成样例\n----\n\n\n下面是活字3.0在MT-Bench-zh评测集上的生成效果展示,并与活字2.0(RLHF版本)进行对比:\n\n\n\n\n\n\n\n\n\n<img src=\"URL width=\"25\" /> 开源协议\n--------------------------------\n\n\n对本仓库源码的使用遵循开源许可协议 Apache 2.0。\n\n\n活字支持商用。如果将活字模型或其衍生品用作商业用途,请您按照如下方式联系许可方,以进行登记并向许可方申请书面授权:联系邮箱:[jngao@URL](mailto:jngao@URL)。\n\n\n<img src=\"URL width=\"25\" /> Citation\n------------------------------------",
"### 活字大模型\n\n\n<img src=\"URL width=\"25\" /> Star History\n----------------------------------------\n\n\n,使用了Mixtral-8x7B的模型结构。它区别于LLaMA、BLOOM等常见模型,活字3.0的每个前馈神经网络(FFN)层被替换为了“专家层”,该层包含8个FFN和一个“路由器”。这种设计使得模型在推理过程中,可以独立地将每个Token路由到最适合处理它的两个专家中。活字3.0共拥有46.7B个参数,但得益于其稀疏激活的特性,实际推理时仅需激活13B参数,有效提升了计算效率和处理速度。\n\n\n",
"### 训练过程\n\n\n由于Mixtral-8x7B词表不支持中文,因此对中文的编解码效率较低,限制了中文场景下的实用性。我们首先基于Mixtral-8x7B进行了中文扩词表增量预训练,显著提高了模型对中文的编解码效率,并使模型具备了强大的中文生成和理解能力。这项成果名为Chinese-Mixtral-8x7B,我们已于2024年1月18日开源了其模型权重和训练代码。基于此,我们进一步对模型进行指令微调,最终推出了活字3.0。这一版本的中文编码、指令遵循、安全回复等能力都有显著提升。\n\n\n模型下载\n----\n\n\n\n如果您希望微调活字3.0或Chinese-Mixtral-8x7B,请参考此处训练代码。\n\n\n模型推理\n----",
"### Quick Start\n\n\n活字3.0采用ChatML格式的prompt模板,格式为:\n\n\n使用活字3.0进行推理的示例代码如下:\n\n\n活字3.0支持全部Mixtral模型生态,包括Transformers、vLLM、URL、AutoAWQ、Text generation web UI等框架。\n\n\n如果您在下载模型时遇到网络问题,可以使用我们在ModelScope上提供的检查点。",
"#### Transformers 模型推理 + 流式生成\n\n\n\ntransformers支持为tokenizer添加聊天模板,并支持流式生成。示例代码如下:",
"#### ModelScope 模型推理\n\n\n\nModelScope的接口与Transformers非常相似,只需将transformers替换为modelscope即可:",
"#### vLLM 推理加速\n\n\n\n活字3.0支持通过vLLM实现推理加速,示例代码如下:",
"#### 部署 OpenAI API Server\n\n\n\n活字3.0可以部署为支持OpenAI API协议的服务,这使得活字3.0可以直接通过OpenAI API进行调用。\n\n\n环境准备:\n\n\n启动服务:\n\n\n使用OpenAI API发送请求:\n\n\n下面是一个使用OpenAI API + Gradio + 流式生成的示例代码:",
"### 量化推理\n\n\n活字3.0支持量化推理,下表为活字3.0在各个量化框架下显存占用量:",
"#### GGUF 格式\n\n\n\nGGUF格式旨在快速加载和保存模型,由llama.cpp团队推出。我们已经提供了GGUF格式的活字3.0。\n\n\n您也可以手动将HuggingFace格式的活字3.0转换到GGUF格式,以使用其他的量化方法。",
"##### Step 1 环境准备\n\n\n首先需要下载llama.cpp的源码。我们在仓库中提供了llama.cpp的submodule,这个版本的llama.cpp已经过测试,可以成功进行推理:\n\n\n您也可以下载最新版本的llama.cpp源码:\n\n\n然后需要进行编译。根据您的硬件平台,编译命令有细微差异:",
"##### Step 2 格式转换(可选)\n\n\n以下命令需要在'URL'目录下:",
"##### Step 3 开始推理\n\n\n以下命令需要在'URL'目录下:\n\n\n'-ngl'参数表示向GPU中offload的层数,降低这个值可以缓解GPU显存压力。经过我们的实际测试,q2\\_k量化的模型offload 16层,显存占用可降低至9.6GB,可在消费级GPU上运行模型:\n\n\n关于'main'的更多参数,可以参考llama.cpp的官方文档。",
"#### AWQ 格式\n\n\n\nAWQ是一种量化模型的存储格式。我们已经提供了AWQ格式的活字3.0,您也可以手动将HuggingFace格式的活字3.0转换到AWQ格式。",
"##### Step 1 格式转换(可选)",
"##### Step 2 开始推理\n\n\n在获取到AWQ格式的模型权重后,可以使用AutoAWQForCausalLM代替AutoModelForCausalLM加载模型。示例代码如下:\n\n\n\n模型性能\n----\n\n\n\n\n\n针对大模型综合能力评价,我们分别使用以下评测数据集对活字3.0进行评测:\n\n\n* C-Eval:一个全面的中文基础模型评估套件。它包含了13948个多项选择题,涵盖了52个不同的学科和四个难度级别。\n* CMMLU:一个综合性的中文评估基准,专门用于评估语言模型在中文语境下的知识和推理能力,涵盖了从基础学科到高级专业水平的67个主题。\n* GAOKAO:一个以中国高考题目为数据集,旨在提供和人类对齐的,直观,高效地测评大模型语言理解能力、逻辑推理能力的测评框架。\n* MMLU:一个包含57个多选任务的英文评测数据集,涵盖了初等数学、美国历史、计算机科学、法律等,难度覆盖高中水平到专家水平,是目前主流的LLM评测数据集之一。\n* HellaSwag:一个极具挑战的英文NLI评测数据集,每一个问题都需要对上下文进行深入理解,而不能基于常识进行回答。\n* GSM8K:一个高质量的小学数学应用题的数据集,这些问题需要 2 到 8 个步骤来解决,解决方案主要涉及使用基本算术运算,可用于评价多步数学推理能力。\n* HumanEval:一个由 164 个原创编程问题组成的数据集,通过衡量从文档字符串生成程序的功能正确性,来够评估语言理解、算法和简单的数学能力。\n* MT-Bench:一个开放的英文问题集,包括80个多轮对话任务,用于评估聊天机器人的多轮对话和指令遵循能力,并通过大模型裁判(GPT-4)对模型回答进行打分。\n* MT-Bench-zh:我们根据MT-Bench翻译得来的中文问题集,每组问题均经过人工校对和中文语境下的适当调整。我们已在此处开源MT-Bench-zh数据集。\n* MT-Bench-safety:我们手工构造的安全数据集,包括暴力、色情、敏感等风险内容。该数据集为封闭数据集。\n\n\n活字3.0在推理时仅激活13B参数。下表为活字3.0与其他13B规模的中文模型以及旧版活字在各个评测数据集上的结果:\n\n\n\n\n\n\n> \n> 我们在C-Eval、CMMLU、MMLU采用5-shot,GSM8K采用4-shot,HellaSwag、HumanEval采用0-shot,HumanEval采用pass@1指标。所有测试均采用greedy策略。\n> \n> \n> 我们使用OpenCompass作为评测框架,commit hash为4c87e77。评测代码位于此处。\n> \n> \n> \n\n\n根据上表中的测试结果,活字3.0较旧版活字取得了巨大的性能提升。在中文知识方面,活字3.0达到了与Tigerbot-13B-chat-v5相当的性能,并是在中文对话和指令遵循方面表现得更加优秀。在英文知识方面,得益于原版Mixtral-8x7B的强大性能,活字3.0超过了Baichuan2-13B-Chat v2和LLaMA系列的扩词表模型,并在英文对话和指令遵循能力上达到了较高水平。在数学推理和代码生成任务上,活字3.0均展现出强大的性能,这说明活字3.0对复杂问题的深层次理解、多步推理、以及结构化信息处理等方面具有较强水平。由于我们采用了较高质量的代码数据集,活字3.0的代码生成能力也超越了同为Mixtral结构的Aurora-Plus模型。\n\n\n生成样例\n----\n\n\n下面是活字3.0在MT-Bench-zh评测集上的生成效果展示,并与活字2.0(RLHF版本)进行对比:\n\n\n\n\n\n\n\n\n\n<img src=\"URL width=\"25\" /> 开源协议\n--------------------------------\n\n\n对本仓库源码的使用遵循开源许可协议 Apache 2.0。\n\n\n活字支持商用。如果将活字模型或其衍生品用作商业用途,请您按照如下方式联系许可方,以进行登记并向许可方申请书面授权:联系邮箱:[jngao@URL](mailto:jngao@URL)。\n\n\n<img src=\"URL width=\"25\" /> Citation\n------------------------------------",
"### 活字大模型\n\n\n<img src=\"URL width=\"25\" /> Star History\n----------------------------------------\n\n\n,使用了Mixtral-8x7B的模型结构。它区别于LLaMA、BLOOM等常见模型,活字3.0的每个前馈神经网络(FFN)层被替换为了“专家层”,该层包含8个FFN和一个“路由器”。这种设计使得模型在推理过程中,可以独立地将每个Token路由到最适合处理它的两个专家中。活字3.0共拥有46.7B个参数,但得益于其稀疏激活的特性,实际推理时仅需激活13B参数,有效提升了计算效率和处理速度。\n\n\n### 训练过程\n\n\n由于Mixtral-8x7B词表不支持中文,因此对中文的编解码效率较低,限制了中文场景下的实用性。我们首先基于Mixtral-8x7B进行了中文扩词表增量预训练,显著提高了模型对中文的编解码效率,并使模型具备了强大的中文生成和理解能力。这项成果名为Chinese-Mixtral-8x7B,我们已于2024年1月18日开源了其模型权重和训练代码。基于此,我们进一步对模型进行指令微调,最终推出了活字3.0。这一版本的中文编码、指令遵循、安全回复等能力都有显著提升。\n\n\n模型下载\n----\n\n\n\n如果您希望微调活字3.0或Chinese-Mixtral-8x7B,请参考此处训练代码。\n\n\n模型推理\n----### Quick Start\n\n\n活字3.0采用ChatML格式的prompt模板,格式为:\n\n\n使用活字3.0进行推理的示例代码如下:\n\n\n活字3.0支持全部Mixtral模型生态,包括Transformers、vLLM、URL、AutoAWQ、Text generation web UI等框架。\n\n\n如果您在下载模型时遇到网络问题,可以使用我们在ModelScope上提供的检查点。#### Transformers 模型推理 + 流式生成\n\n\n\ntransformers支持为tokenizer添加聊天模板,并支持流式生成。示例代码如下:",
"passage: #### ModelScope 模型推理\n\n\n\nModelScope的接口与Transformers非常相似,只需将transformers替换为modelscope即可:#### vLLM 推理加速\n\n\n\n活字3.0支持通过vLLM实现推理加速,示例代码如下:#### 部署 OpenAI API Server\n\n\n\n活字3.0可以部署为支持OpenAI API协议的服务,这使得活字3.0可以直接通过OpenAI API进行调用。\n\n\n环境准备:\n\n\n启动服务:\n\n\n使用OpenAI API发送请求:\n\n\n下面是一个使用OpenAI API + Gradio + 流式生成的示例代码:### 量化推理\n\n\n活字3.0支持量化推理,下表为活字3.0在各个量化框架下显存占用量:#### GGUF 格式\n\n\n\nGGUF格式旨在快速加载和保存模型,由llama.cpp团队推出。我们已经提供了GGUF格式的活字3.0。\n\n\n您也可以手动将HuggingFace格式的活字3.0转换到GGUF格式,以使用其他的量化方法。##### Step 1 环境准备\n\n\n首先需要下载llama.cpp的源码。我们在仓库中提供了llama.cpp的submodule,这个版本的llama.cpp已经过测试,可以成功进行推理:\n\n\n您也可以下载最新版本的llama.cpp源码:\n\n\n然后需要进行编译。根据您的硬件平台,编译命令有细微差异:##### Step 2 格式转换(可选)\n\n\n以下命令需要在'URL'目录下:##### Step 3 开始推理\n\n\n以下命令需要在'URL'目录下:\n\n\n'-ngl'参数表示向GPU中offload的层数,降低这个值可以缓解GPU显存压力。经过我们的实际测试,q2\\_k量化的模型offload 16层,显存占用可降低至9.6GB,可在消费级GPU上运行模型:\n\n\n关于'main'的更多参数,可以参考llama.cpp的官方文档。#### AWQ 格式\n\n\n\nAWQ是一种量化模型的存储格式。我们已经提供了AWQ格式的活字3.0,您也可以手动将HuggingFace格式的活字3.0转换到AWQ格式。##### Step 1 格式转换(可选)"
] | [
-0.05140579491853714,
0.1153252124786377,
-0.014505773782730103,
0.004634128883481026,
0.07512851059436798,
0.040776170790195465,
0.06442595273256302,
0.09249566495418549,
-0.01056581735610962,
0.013626687228679657,
0.045796893537044525,
0.043534405529499054,
0.06642831861972809,
0.13158360123634338,
0.04666101932525635,
-0.19021764397621155,
0.023004328832030296,
-0.025518082082271576,
-0.07714027911424637,
0.046685028821229935,
0.06895988434553146,
-0.0168229341506958,
0.05024337023496628,
0.014110537245869637,
-0.04396500065922737,
0.0060064662247896194,
-0.034365568310022354,
-0.010293282568454742,
0.01630323752760887,
0.07183556258678436,
0.02886786125600338,
0.05432203412055969,
0.04307940602302551,
-0.05653747171163559,
0.031160064041614532,
0.028007451444864273,
0.013693110086023808,
0.04054386168718338,
-0.007929732091724873,
0.06358020752668381,
0.12943725287914276,
0.013165931217372417,
-0.01301414892077446,
0.04943776875734329,
-0.022102687507867813,
-0.054315101355314255,
-0.03538959473371506,
0.007069728337228298,
0.06277623027563095,
0.03356514126062393,
0.024605989456176758,
0.04709082096815109,
0.002716798335313797,
0.046786338090896606,
0.03096805140376091,
-0.14037856459617615,
-0.07047814875841141,
0.09017457813024521,
0.06906762719154358,
0.05360688269138336,
-0.009957607835531235,
0.034835103899240494,
0.025939669460058212,
0.023470113053917885,
-0.022274665534496307,
-0.02627018466591835,
0.028619779273867607,
-0.025815222412347794,
-0.07470498979091644,
-0.006234446540474892,
0.18019096553325653,
-0.021759118884801865,
-0.06153257191181183,
-0.08820497989654541,
-0.026557672768831253,
-0.02738357149064541,
-0.011498481035232544,
-0.024758199229836464,
0.017333433032035828,
0.03287845104932785,
0.05244605988264084,
-0.1277412623167038,
-0.04676587134599686,
-0.06115878000855446,
0.019964758306741714,
0.09347358345985413,
0.04200713336467743,
0.037110310047864914,
-0.004552809521555901,
0.07747411727905273,
-0.0007833875715732574,
-0.06031415984034538,
-0.05065321922302246,
-0.08659225702285767,
-0.010334660299122334,
0.029733940958976746,
-0.043133094906806946,
-0.08411821722984314,
0.07099585235118866,
0.10039563477039337,
0.005014580674469471,
0.06336316466331482,
0.022173497825860977,
0.011787496507167816,
0.030992578715085983,
0.0878564715385437,
-0.027377165853977203,
-0.03639073297381401,
0.029808029532432556,
0.039047207683324814,
0.01645507849752903,
-0.007807815447449684,
-0.058583229780197144,
-0.02319958247244358,
-0.07287417352199554,
0.059958189725875854,
-0.009985614567995071,
0.04087914526462555,
0.03044603392481804,
-0.05191490054130554,
0.14608390629291534,
-0.0616072341799736,
-0.029053185135126114,
-0.01709141954779625,
-0.031327638775110245,
0.06478781998157501,
0.04083763062953949,
-0.0219193696975708,
-0.04180173575878143,
-0.07529415190219879,
-0.009956080466508865,
-0.03345141187310219,
-0.06772715598344803,
-0.013142744079232216,
-0.01559019647538662,
-0.012214615941047668,
0.013366706669330597,
-0.08278633654117584,
-0.1296030879020691,
-0.007138976827263832,
0.08299019187688828,
-0.01691552624106407,
-0.028112230822443962,
0.02786429598927498,
-0.012177659198641777,
-0.016686469316482544,
-0.005193451419472694,
0.057533908635377884,
-0.009085164405405521,
0.010685080662369728,
0.03906349837779999,
0.035455889999866486,
-0.07900936901569366,
0.0049947574734687805,
-0.0020755676086992025,
0.01562468707561493,
-0.11451868712902069,
0.08028502762317657,
-0.06717613339424133,
-0.010682784020900726,
-0.020891960710287094,
-0.014998164027929306,
-0.05461573600769043,
0.020865444093942642,
0.037041157484054565,
0.04603830724954605,
-0.09742981195449829,
-0.006307884585112333,
0.07891075313091278,
-0.10591732710599899,
-0.03385859727859497,
0.03852909803390503,
0.03021979331970215,
0.02391956001520157,
0.035794615745544434,
0.0441599041223526,
0.21195729076862335,
-0.09675685316324234,
0.0055824583396315575,
0.058867841958999634,
-0.030975963920354843,
-0.0019215578213334084,
0.042644012719392776,
0.06832948327064514,
-0.041365403681993484,
0.06710591912269592,
-0.10435006022453308,
0.07152290642261505,
-0.008048663847148418,
-0.020781081169843674,
-0.013936728239059448,
-0.07916802912950516,
-0.02963719144463539,
-0.004683244042098522,
0.009376553818583488,
0.037056174129247665,
-0.038258299231529236,
-0.043561384081840515,
0.13881899416446686,
-0.03228648751974106,
-0.02206563577055931,
-0.11001648753881454,
0.031669266521930695,
0.01078240666538477,
0.028724271804094315,
-0.059518977999687195,
-0.04719612002372742,
0.035887058824300766,
-0.013177331537008286,
0.040612392127513885,
0.08261917531490326,
0.027047988027334213,
0.056231528520584106,
-0.0037224600091576576,
0.012844317592680454,
-0.022338559851050377,
-0.022260352969169617,
0.010445677675306797,
-0.026010088622570038,
-0.024559710174798965,
-0.049059703946113586,
0.061265043914318085,
-0.05424408242106438,
0.05625663325190544,
-0.044668447226285934,
0.015355044044554234,
-0.012415267527103424,
-0.027674641460180283,
0.043928876519203186,
-0.045325394719839096,
-0.022146251052618027,
-0.03604808449745178,
0.008366087451577187,
0.053636252880096436,
-0.02167041227221489,
0.06992306560277939,
-0.006594466976821423,
0.0533430278301239,
0.06019691377878189,
0.08195224404335022,
-0.018551664426922798,
-0.03520383685827255,
-0.030722089111804962,
0.008153055794537067,
0.025111034512519836,
-0.011310306377708912,
0.14656588435173035,
0.014822260476648808,
0.07115386426448822,
-0.0396839901804924,
0.022113189101219177,
0.009338615462183952,
-0.002713186666369438,
0.029170174151659012,
-0.033178724348545074,
0.09098464250564575,
-0.03276786953210831,
-0.021299876272678375,
0.032929904758930206,
-0.031176956370472908,
0.14042265713214874,
-0.002204596996307373,
-0.035378795117139816,
-0.03318342566490173,
0.028604403138160706,
0.0040982672944664955,
0.07039467245340347,
-0.07269904017448425,
-0.00484805554151535,
0.0038572531193494797,
-0.011856229975819588,
0.08956608921289444,
-0.0494878776371479,
0.023204589262604713,
-0.0009989039972424507,
-0.03414066880941391,
0.0647798404097557,
0.025446875020861626,
-0.06176217645406723,
0.020336661487817764,
-0.031083522364497185,
0.042604960501194,
-0.02800406888127327,
-0.009194920770823956,
-0.03096899762749672,
0.061494193971157074,
-0.0413079708814621,
-0.1399371474981308,
-0.09522682428359985,
-0.001011008396744728,
-0.04380398243665695,
0.027506664395332336,
0.009866883978247643,
-0.04897557944059372,
-0.03016771376132965,
-0.03749652951955795,
-0.02518496662378311,
-0.02162337116897106,
-0.01995113492012024,
0.0466369166970253,
0.038476571440696716,
0.04632420092821121,
-0.09770426154136658,
0.011962946504354477,
0.011988550424575806,
-0.08146660774946213,
-0.007237798534333706,
-0.020097099244594574,
0.03980410844087601,
0.06663407385349274,
-0.0015028156340122223,
0.018869824707508087,
0.01232574600726366,
0.11574241518974304,
-0.03353789448738098,
0.044191647320985794,
0.17148864269256592,
0.027326716110110283,
0.00778968445956707,
-0.0004619825631380081,
0.0025261901319026947,
-0.06390957534313202,
0.014158807694911957,
-0.016016963869333267,
-0.04224500060081482,
-0.12087693810462952,
-0.041934769600629807,
-0.040066201239824295,
0.0011296691372990608,
0.021001111716032028,
0.039191946387290955,
-0.016622798517346382,
0.0848798081278801,
-0.009136003442108631,
0.0232947189360857,
0.043308258056640625,
0.030628962442278862,
0.028634164482355118,
-0.052651695907115936,
0.050370752811431885,
-0.04126128554344177,
0.0376705601811409,
0.08196010440587997,
0.07413125783205032,
0.07676415145397186,
-0.043119363486766815,
0.022077256813645363,
0.04316046088933945,
0.04533722996711731,
0.04780237376689911,
0.05490609258413315,
-0.006047761999070644,
0.0025366591289639473,
-0.03099440410733223,
-0.025006543844938278,
-0.08341113477945328,
0.019983887672424316,
0.0389326810836792,
-0.07058173418045044,
-0.015573029406368732,
0.10071825236082077,
0.01087600365281105,
0.08828438818454742,
-0.029876016080379486,
-0.14282558858394623,
-0.04661189019680023,
0.008939488790929317,
-0.014087250456213951,
-0.06118585914373398,
0.030891751870512962,
0.0746210440993309,
-0.038294561207294464,
0.024792548269033432,
-0.02093162201344967,
0.061348140239715576,
-0.0344337522983551,
0.005733913742005825,
0.025751357898116112,
0.10917401313781738,
0.041210416704416275,
0.04686422646045685,
-0.0858454704284668,
0.024741224944591522,
0.0055557419545948505,
0.0590411052107811,
-0.048839692026376724,
0.04962829500436783,
0.08898866176605225,
0.04183918237686157,
0.06936627626419067,
-0.011071981862187386,
-0.08813048899173737,
-0.016022590920329094,
-0.10163605213165283,
0.07720743119716644,
0.04396983981132507,
0.003740733489394188,
0.037715233862400055,
-0.03282224014401436,
0.006410738918930292,
-0.041067082434892654,
0.01646365597844124,
-0.11234034597873688,
-0.11955112218856812,
0.034959159791469574,
0.046026166528463364,
-0.00896061584353447,
-0.042337898164987564,
0.011721227318048477,
0.001478562131524086,
0.14353808760643005,
-0.023302515968680382,
-0.03717700019478798,
-0.060849469155073166,
0.005152903497219086,
0.10218539834022522,
-0.01732744835317135,
0.013411364518105984,
-0.025319915264844894,
0.08563324809074402,
0.004889491014182568,
-0.05852381885051727,
-0.03428163379430771,
-0.09529036283493042,
-0.04739375784993172,
-0.012200267054140568,
0.03486878052353859,
-0.012625042349100113,
0.016200680285692215,
0.01742769218981266,
-0.05133563280105591,
0.015405390411615372,
-0.10728831589221954,
-0.024759739637374878,
0.0897768884897232,
-0.04990702122449875,
0.039649270474910736,
-0.10817565768957138,
-0.015269830822944641,
-0.017693687230348587,
0.01835867017507553,
0.03073960728943348,
0.07775215804576874,
-0.030572552233934402,
0.04888700321316719,
0.09710454940795898,
-0.04144403710961342,
-0.16739536821842194,
-0.06552132964134216,
0.09198746085166931,
0.024697458371520042,
-0.01814839616417885,
-0.188677117228508,
0.04157894104719162,
0.04468696564435959,
-0.008519267663359642,
0.07405409216880798,
-0.13058888912200928,
-0.05183473229408264,
0.03217015415430069,
0.01326520275324583,
0.08261115849018097,
-0.08261390775442123,
-0.01626255176961422,
-0.02207241766154766,
-0.03764466568827629,
0.025418084114789963,
-0.029953191056847572,
0.08980469405651093,
-0.015491647645831108,
-0.043986666947603226,
0.029889758676290512,
-0.03384371101856232,
0.10761787742376328,
-0.07437136024236679,
0.005458155646920204,
-0.03739837184548378,
0.048921793699264526,
0.044158466160297394,
-0.04282119870185852,
0.11090857535600662,
-0.07344797253608704,
0.057919345796108246,
-0.07575658708810806,
-0.028989454731345177,
-0.001501379068940878,
0.007236809469759464,
0.016847245395183563,
-0.01849246583878994,
-0.030794456601142883,
0.02251620776951313,
0.0029203849844634533,
0.036122605204582214,
-0.03423462435603142,
-0.018457025289535522,
-0.022975759580731392,
0.1377089023590088,
0.052931610494852066,
-0.06341332197189331,
-0.013686333782970905,
-0.019396526739001274,
0.00932271033525467,
0.06469491124153137,
-0.061101216822862625,
0.0294716265052557,
0.03495151549577713,
-0.030542638152837753,
0.02026517316699028,
0.017809998244047165,
-0.021770065650343895,
0.03515876829624176,
0.05730893835425377,
-0.05362052470445633,
-0.028709711506962776,
-0.05701728165149689,
0.030345134437084198,
-0.022656140848994255,
0.036124348640441895,
0.10589541494846344,
-0.015355415642261505,
0.010459613986313343,
0.02084319293498993,
0.018930355086922646,
-0.0181356742978096,
0.09105142951011658,
0.014063818380236626,
-0.013239286839962006,
-0.032015129923820496,
0.050749365240335464,
0.00234078336507082,
0.024465490132570267,
0.021754765883088112,
0.04497937858104706,
-0.05546874552965164,
-0.06474318355321884,
-0.09624253213405609,
0.010860487818717957,
0.024784250184893608,
-0.04973963648080826,
-0.02388923056423664,
-0.00798911415040493,
-0.017334945499897003,
-0.025196488946676254,
0.021238751709461212,
-0.009648861363530159,
-0.03243473172187805,
0.038872599601745605,
-0.06732794642448425,
0.026329735293984413,
-0.009553661569952965,
0.039311107248067856,
-0.08677786588668823,
0.018571069464087486,
0.045320942997932434,
0.0335826501250267,
-0.040736645460128784,
-0.02845924161374569,
-0.020437732338905334,
0.0018744743429124355,
-0.11242455244064331,
0.03374467045068741,
-0.040998127311468124,
-0.014499797485768795,
0.026461558416485786,
-0.021425660699605942,
-0.044673897325992584,
0.04075043648481369,
-0.017007378861308098,
0.003598053939640522,
-0.04344823583960533,
0.033070191740989685,
-0.007283146493136883,
0.03438253700733185,
-0.005024517886340618,
-0.06306484341621399,
0.042711615562438965,
0.0070465668104588985,
-0.029197555035352707,
0.019516577944159508,
-0.043798401951789856,
-0.03551418334245682,
0.01571907475590706,
0.05490114167332649,
-0.03529305011034012,
-0.00702363345772028,
0.041527580469846725,
0.03235841542482376,
0.006161754950881004,
-0.034205589443445206,
0.08449883759021759,
-0.031460437923669815,
0.028387034311890602,
0.0137871615588665,
0.00023393519222736359,
-0.01587541587650776,
-0.007630970329046249,
0.07310144603252411,
0.08810555189847946,
0.04671776667237282,
-0.02511662058532238,
0.017069991677999496,
-0.09502191841602325,
0.02357759140431881,
0.005114045459777117,
0.01243460550904274,
0.004023693967610598,
-0.03440435230731964,
0.03452888876199722,
-0.010223506018519402,
0.18213096261024475,
-0.007461824454367161,
-0.00096084363758564,
-0.011016391217708588,
-0.04987460374832153,
0.0009843956213444471,
-0.001926494762301445,
0.05506081134080887,
0.06059756129980087,
0.047815218567848206,
-0.02065914496779442,
0.0048877703957259655,
-0.006336643826216459,
-0.10795065760612488,
0.034630268812179565,
0.025377023965120316,
0.012141418643295765,
0.09640626609325409,
0.030299853533506393,
-0.02703629434108734,
-0.024084482342004776,
0.0707450658082962,
-0.06324602663516998,
0.06473408639431,
-0.0425044447183609,
0.0040446617640554905,
0.08741829544305801,
-0.06814660131931305,
0.02606765367090702,
0.014392741955816746,
-0.02855014055967331,
-0.081507109105587,
-0.054207704961299896,
-0.04521312937140465,
-0.09722332656383514,
0.008222825825214386,
-0.05787396430969238,
-0.013750012964010239,
-0.04861002415418625,
0.020960956811904907,
-0.005067589692771435,
0.07866496592760086,
-0.010683659464120865,
-0.05139058083295822,
0.013243434019386768,
-0.01637750305235386,
0.013444656506180763,
-0.0003629261627793312,
-0.037512317299842834,
0.017749730497598648,
0.04494556039571762,
0.055468376725912094,
0.029708191752433777,
0.014716562815010548,
0.0077024600468575954,
0.006636985577642918,
-0.0074274372309446335,
-0.026298608630895615,
-0.013140968047082424,
-0.01886230707168579,
0.05214446038007736,
0.007981957867741585,
-0.04754804074764252,
0.011165481060743332,
0.08290088176727295,
-0.016050653532147408,
-0.07096920907497406,
-0.1341545730829239,
0.11162830889225006,
-0.00862804800271988,
0.006203753873705864,
-0.011937067843973637,
0.007194166071712971,
-0.039522700011730194,
0.1641179323196411,
0.13169121742248535,
-0.0780065655708313,
-0.01708020269870758,
0.051611967384815216,
-0.0008402634412050247,
-0.02978813275694847,
0.14037495851516724,
0.058796707540750504,
0.15286093950271606,
0.04018450900912285,
-0.022225379943847656,
-0.0050382171757519245,
0.001513252966105938,
-0.027591772377490997,
0.002698375843465328,
-0.04059671610593796,
-0.00030101905576884747,
0.007892176508903503,
0.04596634954214096,
-0.039709217846393585,
-0.11982013285160065,
0.004866333678364754,
0.0017816312611103058,
-0.05631870776414871,
-0.008481668308377266,
0.003457909682765603,
0.008723512291908264,
0.04116275906562805,
-0.023202117532491684,
0.01863507181406021,
0.08767466247081757,
-0.03455633670091629,
-0.028047338128089905,
-0.01240714080631733,
0.05649081617593765,
-0.03528088331222534,
0.16637180745601654,
0.0008319197222590446,
0.009747437201440334,
0.02206646278500557,
-0.019928568974137306,
-0.09664107859134674,
0.0613490492105484,
-0.01928851567208767,
-0.09764289855957031,
0.0026751530822366476,
0.055634379386901855,
-0.042201049625873566,
0.08702285587787628,
0.0226045623421669,
-0.012935122475028038,
0.002569104079157114,
0.02428688295185566,
-0.01547269057482481,
-0.04054424539208412,
0.043923377990722656,
-0.08454832434654236,
0.1281239539384842,
0.09368838369846344,
0.0044517358765006065,
-0.009691642597317696,
-0.0376763790845871,
0.09419676661491394,
0.015741722658276558,
0.044265687465667725,
-0.026406968012452126,
-0.10182860493659973,
-0.038198575377464294,
-0.004380721598863602,
0.018803894519805908,
-0.08440011739730835,
-0.015626445412635803,
-0.003847220679745078,
-0.00009567057713866234,
-0.0005995715036988258,
0.061955079436302185,
0.05638768896460533,
0.052244000136852264,
-0.01000757236033678,
-0.01232198067009449,
-0.004087312147021294,
0.020427923649549484,
-0.14069275557994843,
-0.06620076298713684
] |
null | null | transformers | <!-- markdownlint-disable first-line-h1 -->
<!-- markdownlint-disable html -->
<div align="center">
<h1>
<img src="image/huozi-logo.jpg" width="30" /> 活字通用大模型
</h1>
</div>
</p>
<div align="center">
<a href="https://github.com/HIT-SCIR/huozi/pulls">
<image src="https://img.shields.io/badge/PRs-welcome-brightgreen">
</a>
<a href="https://github.com/HIT-SCIR/huozi/pulls">
<image src="https://img.shields.io/badge/License-Apache_2.0-green.svg">
</a>
<!-- <h4 align="center">
<p>
<b>中文</b> |
<a href="https://github.com/HIT-SCIR/huozi/blob/main/README_EN.md">English</a>
<p>
</h4> -->
</div>
## 🔖 目录
|章节|说明|
|---|---|
|[💁🏻♂ 开源清单](#-开源清单)|本仓库开源项目清单|
|[💡 模型介绍](#-模型介绍)|简要介绍活字模型结构和训练过程|
|[📥 模型下载](#-模型下载)|活字模型下载链接|
|[💻 模型推理](#-模型推理)|活字模型推理样例,包括vLLM推理加速、llama.cpp量化推理等框架的使用流程|
|[📈 模型性能](#-模型性能)|活字模型在主流评测任务上的性能|
|[🗂 生成样例](#-生成样例)|活字模型实际生成效果样例|
## 💁🏻♂ 开源清单

- **活字 3.0**: [[模型权重](#-模型下载)]
- 活字3.0为一个稀疏混合专家模型,支持32K上下文,具有丰富的中、英文知识和强大的数学推理、代码生成能力。活字3.0较旧版活字具有更强的指令遵循能力和安全性。
- **中文MT-Bench**: [[数据集](data/mt-bench-zh/)]
- 本数据集是英文MT-Bench对话能力评测数据集的中文版。它包含了一系列多轮对话问题,每一组问题都经过了精心的人工校对,并为适应中文语境进行了必要的调整。
- **《ChatGPT 调研报告》**: [[PDF](https://github.com/HIT-SCIR/huozi/blob/main/pdf/chatgpt_book.pdf)]
- 哈工大自然语言处理研究所组织多位老师和同学撰写了本调研报告,从技术原理、应用场景、未来发展等方面对ChatGPT进行了尽量详尽的介绍及总结。
- **活字 2.0**: [[模型权重](https://huggingface.co/HIT-SCIR/huozi-7b-rlhf)] [[RLHF数据](data/huozi-rlhf/huozi_rlhf_data.csv)]
- 在活字1.0基础上,通过人类反馈的强化学习(RLHF)进一步优化了模型回复质量,使其更加符合人类偏好。相较于上一个版本平均长度明显提高,遵从指令的能力更强,逻辑更加清晰。
- 16.9k 人工标注的偏好数据,回复来自活字模型,可以用于训练奖励模型。
- **活字 1.0**: [[模型权重](https://huggingface.co/HIT-SCIR/huozi-7b-sft)]
- 在Bloom模型的基础上,在大约 150 亿 tokens 上进行指令微调训练得到的模型,具有更强的指令遵循能力、更好的安全性。
## 💡 模型介绍
大规模语言模型(LLM)在自然语言处理领域取得了显著的进展,并在广泛的应用场景中展现了其强大的潜力。这一技术不仅吸引了学术界的广泛关注,也成为了工业界的热点。在此背景下,哈尔滨工业大学社会计算与信息检索研究中心(HIT-SCIR)近期推出了最新成果——**活字3.0**,致力于为自然语言处理的研究和实际应用提供更多可能性和选择。
活字3.0是基于Chinese-Mixtral-8x7B,在大约30万行指令数据上微调得到的模型。该模型支持**32K上下文**,能够有效处理长文本。活字3.0继承了基座模型丰富的**中英文知识**,并在**数学推理**、**代码生成**等任务上具有强大性能。经过指令微调,活字3.0还在**指令遵循能力**和**安全性**方面实现了显著提升。
此外,我们开源了**中文MT-Bench数据集**。这是一个中文开放问题集,包括80组对话任务,用于评估模型的多轮对话和指令遵循能力。该数据集是根据原始MT-Bench翻译得来的,每组问题均经过人工校对和中文语境下的适当调整。我们还对原始MT-Bench中的部分错误答案进行了修正。
> [!IMPORTANT]
> 活字系列模型仍然可能生成包含事实性错误的误导性回复或包含偏见/歧视的有害内容,请谨慎鉴别和使用生成的内容,请勿将生成的有害内容传播至互联网。
### 模型结构
活字3.0是一个稀疏混合专家模型(SMoE),使用了Mixtral-8x7B的模型结构。它区别于LLaMA、BLOOM等常见模型,活字3.0的每个前馈神经网络(FFN)层被替换为了“专家层”,该层包含8个FFN和一个“路由器”。这种设计使得模型在推理过程中,可以独立地将每个Token路由到最适合处理它的两个专家中。活字3.0共拥有46.7B个参数,但得益于其稀疏激活的特性,实际推理时仅需激活13B参数,有效提升了计算效率和处理速度。

### 训练过程
由于Mixtral-8x7B词表不支持中文,因此对中文的编解码效率较低,限制了中文场景下的实用性。我们首先基于Mixtral-8x7B进行了中文扩词表增量预训练,显著提高了模型对中文的编解码效率,并使模型具备了强大的中文生成和理解能力。这项成果名为[Chinese-Mixtral-8x7B](https://github.com/HIT-SCIR/Chinese-Mixtral-8x7B),我们已于2024年1月18日开源了其模型权重和训练代码。基于此,我们进一步对模型进行指令微调,最终推出了活字3.0。这一版本的中文编码、指令遵循、安全回复等能力都有显著提升。
## 📥 模型下载
|模型名称|文件大小|下载地址|备注|
|:---:|:---:|:---:|:---:|
|huozi3|88GB|[🤗HuggingFace](https://huggingface.co/HIT-SCIR/huozi3)<br>[ModelScope](https://modelscope.cn/models/HIT-SCIR/huozi3/summary)|活字3.0 完整模型|
|huozi3-gguf|25GB|[🤗HuggingFace](https://huggingface.co/HIT-SCIR/huozi3-gguf)<br>[ModelScope](https://modelscope.cn/models/HIT-SCIR/huozi3-gguf/summary)|活字3.0 GGUF版本,适用于llama.cpp等推理框架|
|huozi3-awq|24GB|[🤗HuggingFace](https://huggingface.co/HIT-SCIR/huozi3-awq)<br>[ModelScope](https://modelscope.cn/models/HIT-SCIR/huozi3-awq/summary)|活字3.0 AWQ版本,适用于AutoAWQ等推理框架|
如果您希望微调活字3.0或Chinese-Mixtral-8x7B,请参考[此处训练代码](https://github.com/HIT-SCIR/Chinese-Mixtral-8x7B?tab=readme-ov-file#%E5%BE%AE%E8%B0%83)。
## 💻 模型推理
### Quick Start
活字3.0采用ChatML格式的prompt模板,格式为:
```
<|beginofutterance|>系统
{system prompt}<|endofutterance|>
<|beginofutterance|>用户
{input}<|endofutterance|>
<|beginofutterance|>助手
{output}<|endofutterance|>
```
使用活字3.0进行推理的示例代码如下:
```python
# quickstart.py
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "HIT-SCIR/huozi3"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
model_id,
attn_implementation="flash_attention_2",
torch_dtype=torch.bfloat16,
device_map="auto",
)
text = """<|beginofutterance|>系统
你是一个智能助手<|endofutterance|>
<|beginofutterance|>用户
请你用python写一段快速排序的代码<|endofutterance|>
<|beginofutterance|>助手
"""
inputs = tokenizer(text, return_tensors="pt").to(0)
outputs = model.generate(
**inputs,
eos_token_id=57001,
temperature=0.8,
top_p=0.9,
max_new_tokens=2048,
)
print(tokenizer.decode(outputs[0], skip_special_tokens=False))
```
活字3.0支持全部Mixtral模型生态,包括Transformers、vLLM、llama.cpp、AutoAWQ、Text generation web UI等框架。
如果您在下载模型时遇到网络问题,可以使用我们在[ModelScope](#modelscope-模型推理)上提供的检查点。
<details>
<summary>
#### Transformers 模型推理 + 流式生成
</summary>
transformers支持为tokenizer添加聊天模板,并支持流式生成。示例代码如下:
```python
# example/transformers-stream/stream.py
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer
model_id = "HIT-SCIR/huozi3"
model = AutoModelForCausalLM.from_pretrained(
model_id,
attn_implementation="flash_attention_2",
torch_dtype=torch.bfloat16,
device_map="auto",
)
tokenizer = AutoTokenizer.from_pretrained(model_id)
tokenizer.chat_template = """{% for message in messages %}{{'<|beginofutterance|>' + message['role'] + '\n' + message['content']}}{% if (loop.last and add_generation_prompt) or not loop.last %}{{ '<|endofutterance|>' + '\n'}}{% endif %}{% endfor %}
{% if add_generation_prompt and messages[-1]['role'] != '助手' %}{{ '<|beginofutterance|>助手\n' }}{% endif %}"""
chat = [
{"role": "系统", "content": "你是一个智能助手"},
{"role": "用户", "content": "请你用python写一段快速排序的代码"},
]
inputs = tokenizer.apply_chat_template(
chat,
tokenize=True,
add_generation_prompt=True,
return_tensors="pt",
).to(0)
stream_output = model.generate(
inputs,
streamer=TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True),
eos_token_id=57001,
temperature=0.8,
top_p=0.9,
max_new_tokens=2048,
)
```
</details>
<details>
<summary>
#### ModelScope 模型推理
</summary>
ModelScope的接口与Transformers非常相似,只需将transformers替换为modelscope即可:
```diff
# example/modelscope-generate/generate.py
import torch
- from transformers import AutoModelForCausalLM, AutoTokenizer
+ from modelscope import AutoTokenizer, AutoModelForCausalLM
model_id = "HIT-SCIR/huozi3"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
model_id,
attn_implementation="flash_attention_2",
torch_dtype=torch.bfloat16,
device_map="auto",
)
text = """<|beginofutterance|>系统
你是一个智能助手<|endofutterance|>
<|beginofutterance|>用户
请你用python写一段快速排序的代码<|endofutterance|>
<|beginofutterance|>助手
"""
inputs = tokenizer(text, return_tensors="pt").to(0)
outputs = model.generate(
**inputs,
eos_token_id=57001,
temperature=0.8,
top_p=0.9,
max_new_tokens=2048,
)
print(tokenizer.decode(outputs[0], skip_special_tokens=False))
```
</details>
<details>
<summary>
#### vLLM 推理加速
</summary>
活字3.0支持通过vLLM实现推理加速,示例代码如下:
```python
# example/vllm-generate/generate.py
from vllm import LLM, SamplingParams
prompts = [
"""<|beginofutterance|>系统
你是一个智能助手<|endofutterance|>
<|beginofutterance|>用户
请你用python写一段快速排序的代码<|endofutterance|>
<|beginofutterance|>助手
""",
]
sampling_params = SamplingParams(
temperature=0.8, top_p=0.95, stop_token_ids=[57001], max_tokens=2048
)
llm = LLM(
model="HIT-SCIR/huozi3",
tensor_parallel_size=4,
)
outputs = llm.generate(prompts, sampling_params)
for output in outputs:
prompt = output.prompt
generated_text = output.outputs[0].text
print(generated_text)
```
</details>
<details>
<summary>
#### 部署 OpenAI API Server
</summary>
活字3.0可以部署为支持OpenAI API协议的服务,这使得活字3.0可以直接通过OpenAI API进行调用。
环境准备:
```shell
$ pip install vllm openai
```
启动服务:
```shell
$ python -m vllm.entrypoints.openai.api_server --model /path/to/huozi3/checkpoint --served-model-name huozi --chat-template template.jinja --tensor-parallel-size 8 --response-role 助手 --max-model-len 2048
```
使用OpenAI API发送请求:
```python
# example/openai-api/openai-client.py
from openai import OpenAI
openai_api_key = "EMPTY"
openai_api_base = "http://localhost:8000/v1"
client = OpenAI(
api_key=openai_api_key,
base_url=openai_api_base,
)
chat_response = client.chat.completions.create(
model="huozi",
messages=[
{"role": "系统", "content": "你是一个智能助手"},
{"role": "用户", "content": "请你用python写一段快速排序的代码"},
],
extra_body={"stop_token_ids": [57001]},
)
print("Chat response:", chat_response.choices[0].message.content)
```
下面是一个使用OpenAI API + Gradio + 流式生成的示例代码:
```python
# example/openai-api/openai-client-gradio.py
from openai import OpenAI
import gradio as gr
openai_api_key = "EMPTY"
openai_api_base = "http://localhost:8000/v1"
client = OpenAI(
api_key=openai_api_key,
base_url=openai_api_base,
)
def predict(message, history):
history_openai_format = [
{"role": "系统", "content": "你是一个智能助手"},
]
for human, assistant in history:
history_openai_format.append({"role": "用户", "content": human})
history_openai_format.append({"role": "助手", "content": assistant})
history_openai_format.append({"role": "用户", "content": message})
models = client.models.list()
stream = client.chat.completions.create(
model=models.data[0].id,
messages=history_openai_format,
temperature=0.8,
stream=True,
extra_body={"repetition_penalty": 1, "stop_token_ids": [57001]},
)
partial_message = ""
for chunk in stream:
partial_message += chunk.choices[0].delta.content or ""
yield partial_message
gr.ChatInterface(predict).queue().launch()
```
</details>
### 量化推理
活字3.0支持量化推理,下表为活字3.0在各个量化框架下显存占用量:
|量化方法|显存占用|
|:---:|:---:|
|无|95GB|
|AWQ|32GB|
|GGUF(q4_0)|28GB|
|GGUF(q2_k)|18GB|
|GGUF(q2_k, offload 16层)|9.6GB|
<details>
<summary>
#### GGUF 格式
</summary>
GGUF格式旨在快速加载和保存模型,由llama.cpp团队推出。我们已经提供了[GGUF格式的活字3.0](https://huggingface.co/HIT-SCIR/huozi3-gguf)。
您也可以手动将HuggingFace格式的活字3.0转换到GGUF格式,以使用其他的量化方法。
##### Step 1 环境准备
首先需要下载llama.cpp的源码。我们在仓库中提供了llama.cpp的submodule,这个版本的llama.cpp已经过测试,可以成功进行推理:
```shell
$ git clone --recurse-submodules https://github.com/HIT-SCIR/huozi
$ cd examples/llama.cpp
```
您也可以下载最新版本的llama.cpp源码:
```shell
$ git clone https://github.com/ggerganov/llama.cpp.git
$ cd llama.cpp
```
然后需要进行编译。根据您的硬件平台,编译命令有细微差异:
```shell
$ make # 用于纯CPU推理
$ make LLAMA_CUBLAS=1 # 用于GPU推理
$ LLAMA_METAL=1 make # 用于Apple Silicon,暂未经过测试
```
##### Step 2 格式转换(可选)
以下命令需要在`llama.cpp/`目录下:
```shell
# 转换为GGUF格式
$ python convert.py --outfile /path/to/huozi-gguf/huozi3.gguf /path/to/huozi3
# 进行GGUF格式的q4_0量化
$ quantize /path/to/huozi-gguf/huozi3.gguf /path/to/huozi-gguf/huozi3-q4_0.gguf q4_0
```
##### Step 3 开始推理
以下命令需要在`llama.cpp/`目录下:
```shell
$ main -m /path/to/huozi-gguf/huozi3-q4_0.gguf --color --interactive-first -c 2048 -t 6 --temp 0.2 --repeat_penalty 1.1 -ngl 999 --in-prefix "<|beginofutterance|>用户\n" --in-suffix "<|endofutterance|>\n<|beginofutterance|>助手" -r "<|endofutterance|>"
```
`-ngl`参数表示向GPU中offload的层数,降低这个值可以缓解GPU显存压力。经过我们的实际测试,q2_k量化的模型offload 16层,显存占用可降低至9.6GB,可在消费级GPU上运行模型:
```shell
$ main -m /path/to/huozi-gguf/huozi3-q2_k.gguf --color --interactive-first -c 2048 -t 6 --temp 0.2 --repeat_penalty 1.1 -ngl 16 --in-prefix "<|beginofutterance|>用户\n" --in-suffix "<|endofutterance|>\n<|beginofutterance|>助手" -r "<|endofutterance|>"
```
关于`main`的更多参数,可以参考llama.cpp的[官方文档](https://github.com/ggerganov/llama.cpp/tree/master/examples/main)。
</details>
<details>
<summary>
#### AWQ 格式
</summary>
AWQ是一种量化模型的存储格式。我们已经提供了[AWQ格式的活字3.0](https://huggingface.co/HIT-SCIR/huozi3-awq),您也可以手动将HuggingFace格式的活字3.0转换到AWQ格式。
##### Step 1 格式转换(可选)
```python
# example/autoawq-generate/quant.py
from awq import AutoAWQForCausalLM
from transformers import AutoTokenizer
model_path = "/path/to/huozi3"
quant_path = "/path/to/save/huozi3-awq"
modules_to_not_convert = ["gate"]
quant_config = {
"zero_point": True,
"q_group_size": 128,
"w_bit": 4,
"version": "GEMM",
"modules_to_not_convert": modules_to_not_convert,
}
model = AutoAWQForCausalLM.from_pretrained(
model_path,
safetensors=True,
**{"low_cpu_mem_usage": True},
)
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
model.quantize(
tokenizer,
quant_config=quant_config,
modules_to_not_convert=modules_to_not_convert,
)
model.save_quantized(quant_path)
tokenizer.save_pretrained(quant_path)
print(f'Model is quantized and saved at "{quant_path}"')
```
##### Step 2 开始推理
在获取到AWQ格式的模型权重后,可以使用AutoAWQForCausalLM代替AutoModelForCausalLM加载模型。示例代码如下:
```diff
# example/autoawq-generate/generate.py
import torch
+ from awq import AutoAWQForCausalLM
from transformers import AutoTokenizer, TextStreamer
- model_id = "HIT-SCIR/huozi3"
+ model_id = "HIT-SCIR/huozi3-awq" # or model_id = "/path/to/saved/huozi3-awq"
+ model = AutoAWQForCausalLM.from_quantized(model_id, fuse_layers=True)
- model = AutoModelForCausalLM.from_pretrained(
- model_id,
- attn_implementation="flash_attention_2",
- torch_dtype=torch.bfloat16,
- device_map="auto",
- )
tokenizer = AutoTokenizer.from_pretrained(model_id)
tokenizer.chat_template = """{% for message in messages %}{{'<|beginofutterance|>' + message['role'] + '\n' + message['content']}}{% if (loop.last and add_generation_prompt) or not loop.last %}{{ '<|endofutterance|>' + '\n'}}{% endif %}{% endfor %}
{% if add_generation_prompt and messages[-1]['role'] != '助手' %}{{ '<|beginofutterance|>助手\n' }}{% endif %}"""
chat = [
{"role": "系统", "content": "你是一个智能助手"},
{"role": "用户", "content": "请你用python写一段快速排序的代码"},
]
inputs = tokenizer.apply_chat_template(
chat,
tokenize=True,
add_generation_prompt=True,
return_tensors="pt",
).to(0)
stream_output = model.generate(
inputs,
streamer=TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True),
eos_token_id=57001,
temperature=0.8,
top_p=0.9,
max_new_tokens=2048,
)
```
</details>
## 📈 模型性能

针对大模型综合能力评价,我们分别使用以下评测数据集对活字3.0进行评测:
- C-Eval:一个全面的中文基础模型评估套件。它包含了13948个多项选择题,涵盖了52个不同的学科和四个难度级别。
- CMMLU:一个综合性的中文评估基准,专门用于评估语言模型在中文语境下的知识和推理能力,涵盖了从基础学科到高级专业水平的67个主题。
- GAOKAO:一个以中国高考题目为数据集,旨在提供和人类对齐的,直观,高效地测评大模型语言理解能力、逻辑推理能力的测评框架。
- MMLU:一个包含57个多选任务的英文评测数据集,涵盖了初等数学、美国历史、计算机科学、法律等,难度覆盖高中水平到专家水平,是目前主流的LLM评测数据集之一。
- HellaSwag:一个极具挑战的英文NLI评测数据集,每一个问题都需要对上下文进行深入理解,而不能基于常识进行回答。
- GSM8K:一个高质量的小学数学应用题的数据集,这些问题需要 2 到 8 个步骤来解决,解决方案主要涉及使用基本算术运算,可用于评价多步数学推理能力。
- HumanEval:一个由 164 个原创编程问题组成的数据集,通过衡量从文档字符串生成程序的功能正确性,来够评估语言理解、算法和简单的数学能力。
- MT-Bench:一个开放的英文问题集,包括80个多轮对话任务,用于评估聊天机器人的多轮对话和指令遵循能力,并通过大模型裁判(GPT-4)对模型回答进行打分。
- MT-Bench-zh:我们根据MT-Bench翻译得来的中文问题集,每组问题均经过人工校对和中文语境下的适当调整。我们已在[此处](data/mt-bench-zh/)开源MT-Bench-zh数据集。
- MT-Bench-safety:我们手工构造的安全数据集,包括暴力、色情、敏感等风险内容。该数据集为封闭数据集。
活字3.0在推理时仅激活13B参数。下表为活字3.0与其他13B规模的中文模型以及旧版活字在各个评测数据集上的结果:
<!-- | 模型名称 | 模型结构 | C-Eval<br>(中文) | CMMLU<br>(中文) | GAOKAO<br>(中文) | MT-Bench-zh<br>(中文对话) | MT-Bench-safety<br>(中文安全) | MMLU<br>(英文) | HellaSwag<br>(英文) | MT-Bench<br>(英文对话) | GSM8K<br>(数学) | HumanEval<br>(代码) |
|---------------------------------------------|---------|--------------|-------------|---------------|--------------------------|-----------------------------|------------|------------------|-----------------------|-------------|-----------------|
| baichuan-inc/Baichuan2-13B-Chat v2 | Baichuan| 56.13 | 58.50 | 48.99 | 6.74 | 8.30 | 54.50 | 51.19 | 6.59 | 25.17 | 20.12 |
| wangrongsheng/Aurora-Plus | Mixtral | 47.67 | 48.75 | 35.05 | 5.47 | 6.70 | 67.80 | 78.27 | 7.13 | 66.26 | 27.44 |
| TigerResearch/tigerbot-13b-chat-v5 | LLaMA | 49.78 | 51.28 | 41.31 | 5.98 | 7.63 | 56.34 | 35.17 | 4.88 | 66.19 | 14.63 |
| hfl/chinese-alpaca-2-13b | LLaMA | 43.47 | 44.53 | 25.94 | 5.77 | 8.13 | 53.05 | 56.85 | 6.24 | 32.75 | 14.02 |
| 活字1.0 | BLOOM | 37.27 | 36.24 | 19.72 | 4.48 | 7.18 | 39.68 | 33.21 | 4.34 | 21.99 | 13.41 |
| 活字2.0 | BLOOM | 32.05 | 34.68 | 22.97 | 5.08 | 6.68 | 38.04 | 33.34 | 4.79 | 19.86 | 6.71 |
| **活字3.0(最新版本)** | Mixtral | 51.82 | 51.06 | 41.21 | 6.29 | 7.58 | 69.48 | 65.18 | 7.62 | 65.81 | 40.85 | -->

> 我们在C-Eval、CMMLU、MMLU采用5-shot,GSM8K采用4-shot,HellaSwag、HumanEval采用0-shot,HumanEval采用pass@1指标。所有测试均采用greedy策略。
>
> 我们使用OpenCompass作为评测框架,commit hash为[4c87e77](https://github.com/open-compass/opencompass/tree/4c87e777d855636b9eda7ec87bcbbf12b62caed3)。评测代码位于[此处](./evaluate/)。
根据上表中的测试结果,活字3.0较旧版活字取得了巨大的性能提升。在中文知识方面,活字3.0达到了与Tigerbot-13B-chat-v5相当的性能,并是在中文对话和指令遵循方面表现得更加优秀。在英文知识方面,得益于原版Mixtral-8x7B的强大性能,活字3.0超过了Baichuan2-13B-Chat v2和LLaMA系列的扩词表模型,并在英文对话和指令遵循能力上达到了较高水平。在数学推理和代码生成任务上,活字3.0均展现出强大的性能,这说明活字3.0对复杂问题的深层次理解、多步推理、以及结构化信息处理等方面具有较强水平。由于我们采用了较高质量的代码数据集,活字3.0的代码生成能力也超越了同为Mixtral结构的Aurora-Plus模型。
## 🗂 生成样例
下面是活字3.0在MT-Bench-zh评测集上的生成效果展示,并与活字2.0(RLHF版本)进行对比:





## <img src="https://cdn.jsdelivr.net/gh/LightChen233/blog-img/folders.png" width="25" /> 开源协议
对本仓库源码的使用遵循开源许可协议 [Apache 2.0](https://github.com/HIT-SCIR/huozi/blob/main/LICENSE)。
活字支持商用。如果将活字模型或其衍生品用作商业用途,请您按照如下方式联系许可方,以进行登记并向许可方申请书面授权:联系邮箱:<[email protected]>。
## <img src="https://cdn.jsdelivr.net/gh/LightChen233/blog-img/notes.png" width="25" /> Citation
### 活字大模型
```latex
@misc{huozi,
author = {Huozi-Team}.
title = {Huozi: Leveraging Large Language Models for Enhanced Open-Domain Chatting}
year = {2024},
publisher = {GitHub},
journal = {GitHub repository}
howpublished = {\url{https://github.com/HIT-SCIR/huozi}}
}
```
## <img src="https://cdn.jsdelivr.net/gh/LightChen233/blog-img/star.png" width="25" /> Star History
[](https://star-history.com/#HIT-SCIR/huozi&Date)
| {} | text-generation | HIT-SCIR/huozi3-awq | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"4-bit",
"region:us"
] | 2024-02-08T08:33:32+00:00 | [] | [] | TAGS
#transformers #safetensors #mixtral #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us
|
活字通用大模型
=========
目录
--
开源清单
------

* 活字 3.0: [模型权重]
+ 活字3.0为一个稀疏混合专家模型,支持32K上下文,具有丰富的中、英文知识和强大的数学推理、代码生成能力。活字3.0较旧版活字具有更强的指令遵循能力和安全性。
* 中文MT-Bench: [数据集]
+ 本数据集是英文MT-Bench对话能力评测数据集的中文版。它包含了一系列多轮对话问题,每一组问题都经过了精心的人工校对,并为适应中文语境进行了必要的调整。
* 《ChatGPT 调研报告》: [PDF]
+ 哈工大自然语言处理研究所组织多位老师和同学撰写了本调研报告,从技术原理、应用场景、未来发展等方面对ChatGPT进行了尽量详尽的介绍及总结。
* 活字 2.0: [模型权重] [RLHF数据]
+ 在活字1.0基础上,通过人类反馈的强化学习(RLHF)进一步优化了模型回复质量,使其更加符合人类偏好。相较于上一个版本平均长度明显提高,遵从指令的能力更强,逻辑更加清晰。
+ 16.9k 人工标注的偏好数据,回复来自活字模型,可以用于训练奖励模型。
* 活字 1.0: [模型权重]
+ 在Bloom模型的基础上,在大约 150 亿 tokens 上进行指令微调训练得到的模型,具有更强的指令遵循能力、更好的安全性。
模型介绍
----
大规模语言模型(LLM)在自然语言处理领域取得了显著的进展,并在广泛的应用场景中展现了其强大的潜力。这一技术不仅吸引了学术界的广泛关注,也成为了工业界的热点。在此背景下,哈尔滨工业大学社会计算与信息检索研究中心(HIT-SCIR)近期推出了最新成果——活字3.0,致力于为自然语言处理的研究和实际应用提供更多可能性和选择。
活字3.0是基于Chinese-Mixtral-8x7B,在大约30万行指令数据上微调得到的模型。该模型支持32K上下文,能够有效处理长文本。活字3.0继承了基座模型丰富的中英文知识,并在数学推理、代码生成等任务上具有强大性能。经过指令微调,活字3.0还在指令遵循能力和安全性方面实现了显著提升。
此外,我们开源了中文MT-Bench数据集。这是一个中文开放问题集,包括80组对话任务,用于评估模型的多轮对话和指令遵循能力。该数据集是根据原始MT-Bench翻译得来的,每组问题均经过人工校对和中文语境下的适当调整。我们还对原始MT-Bench中的部分错误答案进行了修正。
>
> [!IMPORTANT]
> 活字系列模型仍然可能生成包含事实性错误的误导性回复或包含偏见/歧视的有害内容,请谨慎鉴别和使用生成的内容,请勿将生成的有害内容传播至互联网。
>
>
>
### 模型结构
活字3.0是一个稀疏混合专家模型(SMoE),使用了Mixtral-8x7B的模型结构。它区别于LLaMA、BLOOM等常见模型,活字3.0的每个前馈神经网络(FFN)层被替换为了“专家层”,该层包含8个FFN和一个“路由器”。这种设计使得模型在推理过程中,可以独立地将每个Token路由到最适合处理它的两个专家中。活字3.0共拥有46.7B个参数,但得益于其稀疏激活的特性,实际推理时仅需激活13B参数,有效提升了计算效率和处理速度。

### 训练过程
由于Mixtral-8x7B词表不支持中文,因此对中文的编解码效率较低,限制了中文场景下的实用性。我们首先基于Mixtral-8x7B进行了中文扩词表增量预训练,显著提高了模型对中文的编解码效率,并使模型具备了强大的中文生成和理解能力。这项成果名为Chinese-Mixtral-8x7B,我们已于2024年1月18日开源了其模型权重和训练代码。基于此,我们进一步对模型进行指令微调,最终推出了活字3.0。这一版本的中文编码、指令遵循、安全回复等能力都有显著提升。
模型下载
----
如果您希望微调活字3.0或Chinese-Mixtral-8x7B,请参考此处训练代码。
模型推理
----
### Quick Start
活字3.0采用ChatML格式的prompt模板,格式为:
使用活字3.0进行推理的示例代码如下:
活字3.0支持全部Mixtral模型生态,包括Transformers、vLLM、URL、AutoAWQ、Text generation web UI等框架。
如果您在下载模型时遇到网络问题,可以使用我们在ModelScope上提供的检查点。
#### Transformers 模型推理 + 流式生成
transformers支持为tokenizer添加聊天模板,并支持流式生成。示例代码如下:
#### ModelScope 模型推理
ModelScope的接口与Transformers非常相似,只需将transformers替换为modelscope即可:
#### vLLM 推理加速
活字3.0支持通过vLLM实现推理加速,示例代码如下:
#### 部署 OpenAI API Server
活字3.0可以部署为支持OpenAI API协议的服务,这使得活字3.0可以直接通过OpenAI API进行调用。
环境准备:
启动服务:
使用OpenAI API发送请求:
下面是一个使用OpenAI API + Gradio + 流式生成的示例代码:
### 量化推理
活字3.0支持量化推理,下表为活字3.0在各个量化框架下显存占用量:
#### GGUF 格式
GGUF格式旨在快速加载和保存模型,由llama.cpp团队推出。我们已经提供了GGUF格式的活字3.0。
您也可以手动将HuggingFace格式的活字3.0转换到GGUF格式,以使用其他的量化方法。
##### Step 1 环境准备
首先需要下载llama.cpp的源码。我们在仓库中提供了llama.cpp的submodule,这个版本的llama.cpp已经过测试,可以成功进行推理:
您也可以下载最新版本的llama.cpp源码:
然后需要进行编译。根据您的硬件平台,编译命令有细微差异:
##### Step 2 格式转换(可选)
以下命令需要在'URL'目录下:
##### Step 3 开始推理
以下命令需要在'URL'目录下:
'-ngl'参数表示向GPU中offload的层数,降低这个值可以缓解GPU显存压力。经过我们的实际测试,q2\_k量化的模型offload 16层,显存占用可降低至9.6GB,可在消费级GPU上运行模型:
关于'main'的更多参数,可以参考llama.cpp的官方文档。
#### AWQ 格式
AWQ是一种量化模型的存储格式。我们已经提供了AWQ格式的活字3.0,您也可以手动将HuggingFace格式的活字3.0转换到AWQ格式。
##### Step 1 格式转换(可选)
##### Step 2 开始推理
在获取到AWQ格式的模型权重后,可以使用AutoAWQForCausalLM代替AutoModelForCausalLM加载模型。示例代码如下:
模型性能
----

针对大模型综合能力评价,我们分别使用以下评测数据集对活字3.0进行评测:
* C-Eval:一个全面的中文基础模型评估套件。它包含了13948个多项选择题,涵盖了52个不同的学科和四个难度级别。
* CMMLU:一个综合性的中文评估基准,专门用于评估语言模型在中文语境下的知识和推理能力,涵盖了从基础学科到高级专业水平的67个主题。
* GAOKAO:一个以中国高考题目为数据集,旨在提供和人类对齐的,直观,高效地测评大模型语言理解能力、逻辑推理能力的测评框架。
* MMLU:一个包含57个多选任务的英文评测数据集,涵盖了初等数学、美国历史、计算机科学、法律等,难度覆盖高中水平到专家水平,是目前主流的LLM评测数据集之一。
* HellaSwag:一个极具挑战的英文NLI评测数据集,每一个问题都需要对上下文进行深入理解,而不能基于常识进行回答。
* GSM8K:一个高质量的小学数学应用题的数据集,这些问题需要 2 到 8 个步骤来解决,解决方案主要涉及使用基本算术运算,可用于评价多步数学推理能力。
* HumanEval:一个由 164 个原创编程问题组成的数据集,通过衡量从文档字符串生成程序的功能正确性,来够评估语言理解、算法和简单的数学能力。
* MT-Bench:一个开放的英文问题集,包括80个多轮对话任务,用于评估聊天机器人的多轮对话和指令遵循能力,并通过大模型裁判(GPT-4)对模型回答进行打分。
* MT-Bench-zh:我们根据MT-Bench翻译得来的中文问题集,每组问题均经过人工校对和中文语境下的适当调整。我们已在此处开源MT-Bench-zh数据集。
* MT-Bench-safety:我们手工构造的安全数据集,包括暴力、色情、敏感等风险内容。该数据集为封闭数据集。
活字3.0在推理时仅激活13B参数。下表为活字3.0与其他13B规模的中文模型以及旧版活字在各个评测数据集上的结果:

>
> 我们在C-Eval、CMMLU、MMLU采用5-shot,GSM8K采用4-shot,HellaSwag、HumanEval采用0-shot,HumanEval采用pass@1指标。所有测试均采用greedy策略。
>
>
> 我们使用OpenCompass作为评测框架,commit hash为4c87e77。评测代码位于此处。
>
>
>
根据上表中的测试结果,活字3.0较旧版活字取得了巨大的性能提升。在中文知识方面,活字3.0达到了与Tigerbot-13B-chat-v5相当的性能,并是在中文对话和指令遵循方面表现得更加优秀。在英文知识方面,得益于原版Mixtral-8x7B的强大性能,活字3.0超过了Baichuan2-13B-Chat v2和LLaMA系列的扩词表模型,并在英文对话和指令遵循能力上达到了较高水平。在数学推理和代码生成任务上,活字3.0均展现出强大的性能,这说明活字3.0对复杂问题的深层次理解、多步推理、以及结构化信息处理等方面具有较强水平。由于我们采用了较高质量的代码数据集,活字3.0的代码生成能力也超越了同为Mixtral结构的Aurora-Plus模型。
生成样例
----
下面是活字3.0在MT-Bench-zh评测集上的生成效果展示,并与活字2.0(RLHF版本)进行对比:





<img src="URL width="25" /> 开源协议
--------------------------------
对本仓库源码的使用遵循开源许可协议 Apache 2.0。
活字支持商用。如果将活字模型或其衍生品用作商业用途,请您按照如下方式联系许可方,以进行登记并向许可方申请书面授权:联系邮箱:[jngao@URL](mailto:jngao@URL)。
<img src="URL width="25" /> Citation
------------------------------------
### 活字大模型
<img src="URL width="25" /> Star History
----------------------------------------
,使用了Mixtral-8x7B的模型结构。它区别于LLaMA、BLOOM等常见模型,活字3.0的每个前馈神经网络(FFN)层被替换为了“专家层”,该层包含8个FFN和一个“路由器”。这种设计使得模型在推理过程中,可以独立地将每个Token路由到最适合处理它的两个专家中。活字3.0共拥有46.7B个参数,但得益于其稀疏激活的特性,实际推理时仅需激活13B参数,有效提升了计算效率和处理速度。\n\n\n",
"### 训练过程\n\n\n由于Mixtral-8x7B词表不支持中文,因此对中文的编解码效率较低,限制了中文场景下的实用性。我们首先基于Mixtral-8x7B进行了中文扩词表增量预训练,显著提高了模型对中文的编解码效率,并使模型具备了强大的中文生成和理解能力。这项成果名为Chinese-Mixtral-8x7B,我们已于2024年1月18日开源了其模型权重和训练代码。基于此,我们进一步对模型进行指令微调,最终推出了活字3.0。这一版本的中文编码、指令遵循、安全回复等能力都有显著提升。\n\n\n模型下载\n----\n\n\n\n如果您希望微调活字3.0或Chinese-Mixtral-8x7B,请参考此处训练代码。\n\n\n模型推理\n----",
"### Quick Start\n\n\n活字3.0采用ChatML格式的prompt模板,格式为:\n\n\n使用活字3.0进行推理的示例代码如下:\n\n\n活字3.0支持全部Mixtral模型生态,包括Transformers、vLLM、URL、AutoAWQ、Text generation web UI等框架。\n\n\n如果您在下载模型时遇到网络问题,可以使用我们在ModelScope上提供的检查点。",
"#### Transformers 模型推理 + 流式生成\n\n\n\ntransformers支持为tokenizer添加聊天模板,并支持流式生成。示例代码如下:",
"#### ModelScope 模型推理\n\n\n\nModelScope的接口与Transformers非常相似,只需将transformers替换为modelscope即可:",
"#### vLLM 推理加速\n\n\n\n活字3.0支持通过vLLM实现推理加速,示例代码如下:",
"#### 部署 OpenAI API Server\n\n\n\n活字3.0可以部署为支持OpenAI API协议的服务,这使得活字3.0可以直接通过OpenAI API进行调用。\n\n\n环境准备:\n\n\n启动服务:\n\n\n使用OpenAI API发送请求:\n\n\n下面是一个使用OpenAI API + Gradio + 流式生成的示例代码:",
"### 量化推理\n\n\n活字3.0支持量化推理,下表为活字3.0在各个量化框架下显存占用量:",
"#### GGUF 格式\n\n\n\nGGUF格式旨在快速加载和保存模型,由llama.cpp团队推出。我们已经提供了GGUF格式的活字3.0。\n\n\n您也可以手动将HuggingFace格式的活字3.0转换到GGUF格式,以使用其他的量化方法。",
"##### Step 1 环境准备\n\n\n首先需要下载llama.cpp的源码。我们在仓库中提供了llama.cpp的submodule,这个版本的llama.cpp已经过测试,可以成功进行推理:\n\n\n您也可以下载最新版本的llama.cpp源码:\n\n\n然后需要进行编译。根据您的硬件平台,编译命令有细微差异:",
"##### Step 2 格式转换(可选)\n\n\n以下命令需要在'URL'目录下:",
"##### Step 3 开始推理\n\n\n以下命令需要在'URL'目录下:\n\n\n'-ngl'参数表示向GPU中offload的层数,降低这个值可以缓解GPU显存压力。经过我们的实际测试,q2\\_k量化的模型offload 16层,显存占用可降低至9.6GB,可在消费级GPU上运行模型:\n\n\n关于'main'的更多参数,可以参考llama.cpp的官方文档。",
"#### AWQ 格式\n\n\n\nAWQ是一种量化模型的存储格式。我们已经提供了AWQ格式的活字3.0,您也可以手动将HuggingFace格式的活字3.0转换到AWQ格式。",
"##### Step 1 格式转换(可选)",
"##### Step 2 开始推理\n\n\n在获取到AWQ格式的模型权重后,可以使用AutoAWQForCausalLM代替AutoModelForCausalLM加载模型。示例代码如下:\n\n\n\n模型性能\n----\n\n\n\n\n\n针对大模型综合能力评价,我们分别使用以下评测数据集对活字3.0进行评测:\n\n\n* C-Eval:一个全面的中文基础模型评估套件。它包含了13948个多项选择题,涵盖了52个不同的学科和四个难度级别。\n* CMMLU:一个综合性的中文评估基准,专门用于评估语言模型在中文语境下的知识和推理能力,涵盖了从基础学科到高级专业水平的67个主题。\n* GAOKAO:一个以中国高考题目为数据集,旨在提供和人类对齐的,直观,高效地测评大模型语言理解能力、逻辑推理能力的测评框架。\n* MMLU:一个包含57个多选任务的英文评测数据集,涵盖了初等数学、美国历史、计算机科学、法律等,难度覆盖高中水平到专家水平,是目前主流的LLM评测数据集之一。\n* HellaSwag:一个极具挑战的英文NLI评测数据集,每一个问题都需要对上下文进行深入理解,而不能基于常识进行回答。\n* GSM8K:一个高质量的小学数学应用题的数据集,这些问题需要 2 到 8 个步骤来解决,解决方案主要涉及使用基本算术运算,可用于评价多步数学推理能力。\n* HumanEval:一个由 164 个原创编程问题组成的数据集,通过衡量从文档字符串生成程序的功能正确性,来够评估语言理解、算法和简单的数学能力。\n* MT-Bench:一个开放的英文问题集,包括80个多轮对话任务,用于评估聊天机器人的多轮对话和指令遵循能力,并通过大模型裁判(GPT-4)对模型回答进行打分。\n* MT-Bench-zh:我们根据MT-Bench翻译得来的中文问题集,每组问题均经过人工校对和中文语境下的适当调整。我们已在此处开源MT-Bench-zh数据集。\n* MT-Bench-safety:我们手工构造的安全数据集,包括暴力、色情、敏感等风险内容。该数据集为封闭数据集。\n\n\n活字3.0在推理时仅激活13B参数。下表为活字3.0与其他13B规模的中文模型以及旧版活字在各个评测数据集上的结果:\n\n\n\n\n\n\n> \n> 我们在C-Eval、CMMLU、MMLU采用5-shot,GSM8K采用4-shot,HellaSwag、HumanEval采用0-shot,HumanEval采用pass@1指标。所有测试均采用greedy策略。\n> \n> \n> 我们使用OpenCompass作为评测框架,commit hash为4c87e77。评测代码位于此处。\n> \n> \n> \n\n\n根据上表中的测试结果,活字3.0较旧版活字取得了巨大的性能提升。在中文知识方面,活字3.0达到了与Tigerbot-13B-chat-v5相当的性能,并是在中文对话和指令遵循方面表现得更加优秀。在英文知识方面,得益于原版Mixtral-8x7B的强大性能,活字3.0超过了Baichuan2-13B-Chat v2和LLaMA系列的扩词表模型,并在英文对话和指令遵循能力上达到了较高水平。在数学推理和代码生成任务上,活字3.0均展现出强大的性能,这说明活字3.0对复杂问题的深层次理解、多步推理、以及结构化信息处理等方面具有较强水平。由于我们采用了较高质量的代码数据集,活字3.0的代码生成能力也超越了同为Mixtral结构的Aurora-Plus模型。\n\n\n生成样例\n----\n\n\n下面是活字3.0在MT-Bench-zh评测集上的生成效果展示,并与活字2.0(RLHF版本)进行对比:\n\n\n\n\n\n\n\n\n\n<img src=\"URL width=\"25\" /> 开源协议\n--------------------------------\n\n\n对本仓库源码的使用遵循开源许可协议 Apache 2.0。\n\n\n活字支持商用。如果将活字模型或其衍生品用作商业用途,请您按照如下方式联系许可方,以进行登记并向许可方申请书面授权:联系邮箱:[jngao@URL](mailto:jngao@URL)。\n\n\n<img src=\"URL width=\"25\" /> Citation\n------------------------------------",
"### 活字大模型\n\n\n<img src=\"URL width=\"25\" /> Star History\n----------------------------------------\n\n\n,使用了Mixtral-8x7B的模型结构。它区别于LLaMA、BLOOM等常见模型,活字3.0的每个前馈神经网络(FFN)层被替换为了“专家层”,该层包含8个FFN和一个“路由器”。这种设计使得模型在推理过程中,可以独立地将每个Token路由到最适合处理它的两个专家中。活字3.0共拥有46.7B个参数,但得益于其稀疏激活的特性,实际推理时仅需激活13B参数,有效提升了计算效率和处理速度。\n\n\n",
"### 训练过程\n\n\n由于Mixtral-8x7B词表不支持中文,因此对中文的编解码效率较低,限制了中文场景下的实用性。我们首先基于Mixtral-8x7B进行了中文扩词表增量预训练,显著提高了模型对中文的编解码效率,并使模型具备了强大的中文生成和理解能力。这项成果名为Chinese-Mixtral-8x7B,我们已于2024年1月18日开源了其模型权重和训练代码。基于此,我们进一步对模型进行指令微调,最终推出了活字3.0。这一版本的中文编码、指令遵循、安全回复等能力都有显著提升。\n\n\n模型下载\n----\n\n\n\n如果您希望微调活字3.0或Chinese-Mixtral-8x7B,请参考此处训练代码。\n\n\n模型推理\n----",
"### Quick Start\n\n\n活字3.0采用ChatML格式的prompt模板,格式为:\n\n\n使用活字3.0进行推理的示例代码如下:\n\n\n活字3.0支持全部Mixtral模型生态,包括Transformers、vLLM、URL、AutoAWQ、Text generation web UI等框架。\n\n\n如果您在下载模型时遇到网络问题,可以使用我们在ModelScope上提供的检查点。",
"#### Transformers 模型推理 + 流式生成\n\n\n\ntransformers支持为tokenizer添加聊天模板,并支持流式生成。示例代码如下:",
"#### ModelScope 模型推理\n\n\n\nModelScope的接口与Transformers非常相似,只需将transformers替换为modelscope即可:",
"#### vLLM 推理加速\n\n\n\n活字3.0支持通过vLLM实现推理加速,示例代码如下:",
"#### 部署 OpenAI API Server\n\n\n\n活字3.0可以部署为支持OpenAI API协议的服务,这使得活字3.0可以直接通过OpenAI API进行调用。\n\n\n环境准备:\n\n\n启动服务:\n\n\n使用OpenAI API发送请求:\n\n\n下面是一个使用OpenAI API + Gradio + 流式生成的示例代码:",
"### 量化推理\n\n\n活字3.0支持量化推理,下表为活字3.0在各个量化框架下显存占用量:",
"#### GGUF 格式\n\n\n\nGGUF格式旨在快速加载和保存模型,由llama.cpp团队推出。我们已经提供了GGUF格式的活字3.0。\n\n\n您也可以手动将HuggingFace格式的活字3.0转换到GGUF格式,以使用其他的量化方法。",
"##### Step 1 环境准备\n\n\n首先需要下载llama.cpp的源码。我们在仓库中提供了llama.cpp的submodule,这个版本的llama.cpp已经过测试,可以成功进行推理:\n\n\n您也可以下载最新版本的llama.cpp源码:\n\n\n然后需要进行编译。根据您的硬件平台,编译命令有细微差异:",
"##### Step 2 格式转换(可选)\n\n\n以下命令需要在'URL'目录下:",
"##### Step 3 开始推理\n\n\n以下命令需要在'URL'目录下:\n\n\n'-ngl'参数表示向GPU中offload的层数,降低这个值可以缓解GPU显存压力。经过我们的实际测试,q2\\_k量化的模型offload 16层,显存占用可降低至9.6GB,可在消费级GPU上运行模型:\n\n\n关于'main'的更多参数,可以参考llama.cpp的官方文档。",
"#### AWQ 格式\n\n\n\nAWQ是一种量化模型的存储格式。我们已经提供了AWQ格式的活字3.0,您也可以手动将HuggingFace格式的活字3.0转换到AWQ格式。",
"##### Step 1 格式转换(可选)",
"##### Step 2 开始推理\n\n\n在获取到AWQ格式的模型权重后,可以使用AutoAWQForCausalLM代替AutoModelForCausalLM加载模型。示例代码如下:\n\n\n\n模型性能\n----\n\n\n\n\n\n针对大模型综合能力评价,我们分别使用以下评测数据集对活字3.0进行评测:\n\n\n* C-Eval:一个全面的中文基础模型评估套件。它包含了13948个多项选择题,涵盖了52个不同的学科和四个难度级别。\n* CMMLU:一个综合性的中文评估基准,专门用于评估语言模型在中文语境下的知识和推理能力,涵盖了从基础学科到高级专业水平的67个主题。\n* GAOKAO:一个以中国高考题目为数据集,旨在提供和人类对齐的,直观,高效地测评大模型语言理解能力、逻辑推理能力的测评框架。\n* MMLU:一个包含57个多选任务的英文评测数据集,涵盖了初等数学、美国历史、计算机科学、法律等,难度覆盖高中水平到专家水平,是目前主流的LLM评测数据集之一。\n* HellaSwag:一个极具挑战的英文NLI评测数据集,每一个问题都需要对上下文进行深入理解,而不能基于常识进行回答。\n* GSM8K:一个高质量的小学数学应用题的数据集,这些问题需要 2 到 8 个步骤来解决,解决方案主要涉及使用基本算术运算,可用于评价多步数学推理能力。\n* HumanEval:一个由 164 个原创编程问题组成的数据集,通过衡量从文档字符串生成程序的功能正确性,来够评估语言理解、算法和简单的数学能力。\n* MT-Bench:一个开放的英文问题集,包括80个多轮对话任务,用于评估聊天机器人的多轮对话和指令遵循能力,并通过大模型裁判(GPT-4)对模型回答进行打分。\n* MT-Bench-zh:我们根据MT-Bench翻译得来的中文问题集,每组问题均经过人工校对和中文语境下的适当调整。我们已在此处开源MT-Bench-zh数据集。\n* MT-Bench-safety:我们手工构造的安全数据集,包括暴力、色情、敏感等风险内容。该数据集为封闭数据集。\n\n\n活字3.0在推理时仅激活13B参数。下表为活字3.0与其他13B规模的中文模型以及旧版活字在各个评测数据集上的结果:\n\n\n\n\n\n\n> \n> 我们在C-Eval、CMMLU、MMLU采用5-shot,GSM8K采用4-shot,HellaSwag、HumanEval采用0-shot,HumanEval采用pass@1指标。所有测试均采用greedy策略。\n> \n> \n> 我们使用OpenCompass作为评测框架,commit hash为4c87e77。评测代码位于此处。\n> \n> \n> \n\n\n根据上表中的测试结果,活字3.0较旧版活字取得了巨大的性能提升。在中文知识方面,活字3.0达到了与Tigerbot-13B-chat-v5相当的性能,并是在中文对话和指令遵循方面表现得更加优秀。在英文知识方面,得益于原版Mixtral-8x7B的强大性能,活字3.0超过了Baichuan2-13B-Chat v2和LLaMA系列的扩词表模型,并在英文对话和指令遵循能力上达到了较高水平。在数学推理和代码生成任务上,活字3.0均展现出强大的性能,这说明活字3.0对复杂问题的深层次理解、多步推理、以及结构化信息处理等方面具有较强水平。由于我们采用了较高质量的代码数据集,活字3.0的代码生成能力也超越了同为Mixtral结构的Aurora-Plus模型。\n\n\n生成样例\n----\n\n\n下面是活字3.0在MT-Bench-zh评测集上的生成效果展示,并与活字2.0(RLHF版本)进行对比:\n\n\n\n\n\n\n\n\n\n<img src=\"URL width=\"25\" /> 开源协议\n--------------------------------\n\n\n对本仓库源码的使用遵循开源许可协议 Apache 2.0。\n\n\n活字支持商用。如果将活字模型或其衍生品用作商业用途,请您按照如下方式联系许可方,以进行登记并向许可方申请书面授权:联系邮箱:[jngao@URL](mailto:jngao@URL)。\n\n\n<img src=\"URL width=\"25\" /> Citation\n------------------------------------",
"### 活字大模型\n\n\n<img src=\"URL width=\"25\" /> Star History\n----------------------------------------\n\n\n,使用了Mixtral-8x7B的模型结构。它区别于LLaMA、BLOOM等常见模型,活字3.0的每个前馈神经网络(FFN)层被替换为了“专家层”,该层包含8个FFN和一个“路由器”。这种设计使得模型在推理过程中,可以独立地将每个Token路由到最适合处理它的两个专家中。活字3.0共拥有46.7B个参数,但得益于其稀疏激活的特性,实际推理时仅需激活13B参数,有效提升了计算效率和处理速度。\n\n\n### 训练过程\n\n\n由于Mixtral-8x7B词表不支持中文,因此对中文的编解码效率较低,限制了中文场景下的实用性。我们首先基于Mixtral-8x7B进行了中文扩词表增量预训练,显著提高了模型对中文的编解码效率,并使模型具备了强大的中文生成和理解能力。这项成果名为Chinese-Mixtral-8x7B,我们已于2024年1月18日开源了其模型权重和训练代码。基于此,我们进一步对模型进行指令微调,最终推出了活字3.0。这一版本的中文编码、指令遵循、安全回复等能力都有显著提升。\n\n\n模型下载\n----\n\n\n\n如果您希望微调活字3.0或Chinese-Mixtral-8x7B,请参考此处训练代码。\n\n\n模型推理\n----### Quick Start\n\n\n活字3.0采用ChatML格式的prompt模板,格式为:\n\n\n使用活字3.0进行推理的示例代码如下:\n\n\n活字3.0支持全部Mixtral模型生态,包括Transformers、vLLM、URL、AutoAWQ、Text generation web UI等框架。\n\n\n如果您在下载模型时遇到网络问题,可以使用我们在ModelScope上提供的检查点。",
"passage: #### Transformers 模型推理 + 流式生成\n\n\n\ntransformers支持为tokenizer添加聊天模板,并支持流式生成。示例代码如下:#### ModelScope 模型推理\n\n\n\nModelScope的接口与Transformers非常相似,只需将transformers替换为modelscope即可:#### vLLM 推理加速\n\n\n\n活字3.0支持通过vLLM实现推理加速,示例代码如下:#### 部署 OpenAI API Server\n\n\n\n活字3.0可以部署为支持OpenAI API协议的服务,这使得活字3.0可以直接通过OpenAI API进行调用。\n\n\n环境准备:\n\n\n启动服务:\n\n\n使用OpenAI API发送请求:\n\n\n下面是一个使用OpenAI API + Gradio + 流式生成的示例代码:### 量化推理\n\n\n活字3.0支持量化推理,下表为活字3.0在各个量化框架下显存占用量:#### GGUF 格式\n\n\n\nGGUF格式旨在快速加载和保存模型,由llama.cpp团队推出。我们已经提供了GGUF格式的活字3.0。\n\n\n您也可以手动将HuggingFace格式的活字3.0转换到GGUF格式,以使用其他的量化方法。##### Step 1 环境准备\n\n\n首先需要下载llama.cpp的源码。我们在仓库中提供了llama.cpp的submodule,这个版本的llama.cpp已经过测试,可以成功进行推理:\n\n\n您也可以下载最新版本的llama.cpp源码:\n\n\n然后需要进行编译。根据您的硬件平台,编译命令有细微差异:##### Step 2 格式转换(可选)\n\n\n以下命令需要在'URL'目录下:##### Step 3 开始推理\n\n\n以下命令需要在'URL'目录下:\n\n\n'-ngl'参数表示向GPU中offload的层数,降低这个值可以缓解GPU显存压力。经过我们的实际测试,q2\\_k量化的模型offload 16层,显存占用可降低至9.6GB,可在消费级GPU上运行模型:\n\n\n关于'main'的更多参数,可以参考llama.cpp的官方文档。#### AWQ 格式\n\n\n\nAWQ是一种量化模型的存储格式。我们已经提供了AWQ格式的活字3.0,您也可以手动将HuggingFace格式的活字3.0转换到AWQ格式。##### Step 1 格式转换(可选)"
] | [
-0.05876411497592926,
0.09481782466173172,
-0.013685280457139015,
0.006109001114964485,
0.08035314828157425,
0.02607966959476471,
0.07301564514636993,
0.08384080976247787,
-0.0266827791929245,
0.012907098047435284,
0.04938654601573944,
0.05221274495124817,
0.08053438365459442,
0.14356258511543274,
0.03541846573352814,
-0.1741069108247757,
0.04114098846912384,
-0.016387274488806725,
-0.04859021678566933,
0.05464690178632736,
0.08054212480783463,
-0.02907858043909073,
0.051493704319000244,
0.017233403399586678,
-0.039360567927360535,
-0.0072915032505989075,
-0.019430061802268028,
-0.016874339431524277,
0.026603084057569504,
0.06279340386390686,
0.03119783289730549,
0.05776682496070862,
0.040227361023426056,
-0.06287889182567596,
0.02780282124876976,
0.039929091930389404,
0.031062383204698563,
0.050356656312942505,
-0.00706237880513072,
0.07389208674430847,
0.08148212730884552,
0.018355239182710648,
0.0008715635631233454,
0.05449631065130234,
-0.028268788009881973,
-0.06111709028482437,
-0.018019601702690125,
-0.0034477263689041138,
0.08414584398269653,
0.04145463928580284,
0.024466071277856827,
0.056777868419885635,
-0.009548427537083626,
0.04943928122520447,
0.008897434920072556,
-0.15397609770298004,
-0.06714247167110443,
0.05855478346347809,
0.07925503700971603,
0.05869365856051445,
0.0004624258726835251,
0.018094217404723167,
0.023553790524601936,
0.017689822241663933,
0.01574309542775154,
-0.024465596303343773,
0.02212243340909481,
-0.0368150919675827,
-0.06540517508983612,
-0.013297145254909992,
0.19415971636772156,
-0.031388431787490845,
-0.05331895500421524,
-0.0914805680513382,
-0.042526088654994965,
-0.03852040693163872,
-0.005609557963907719,
-0.03252214938402176,
0.00948520191013813,
0.041121382266283035,
0.059163063764572144,
-0.1128583773970604,
-0.05216589570045471,
-0.06028105318546295,
0.02654699608683586,
0.07172413915395737,
0.04325072094798088,
0.04443323612213135,
-0.020910387858748436,
0.0988035798072815,
0.008466935716569424,
-0.06742150336503983,
-0.049310747534036636,
-0.08970896899700165,
-0.023104142397642136,
0.030311085283756256,
-0.010173363611102104,
-0.07674622535705566,
0.05419702082872391,
0.10183945298194885,
0.014319393783807755,
0.0655364841222763,
0.040184441953897476,
0.008724257349967957,
0.02434387244284153,
0.0944681391119957,
-0.019340170547366142,
-0.0626983791589737,
0.03924129530787468,
0.051618706434965134,
0.016145257279276848,
-0.02013619989156723,
-0.050397343933582306,
-0.026513265445828438,
-0.06132567301392555,
0.05794985219836235,
-0.0062666889280080795,
0.047057412564754486,
0.01819666661322117,
-0.049623697996139526,
0.14867940545082092,
-0.073030486702919,
-0.024029700085520744,
-0.011643038131296635,
-0.02837558463215828,
0.06325143575668335,
0.03907621651887894,
-0.02517692930996418,
-0.05900828540325165,
-0.08136291056871414,
-0.016648875549435616,
-0.04213843494653702,
-0.08201846480369568,
-0.01602196879684925,
-0.02427659183740616,
-0.02722110226750374,
-0.00008536456152796745,
-0.07798534631729126,
-0.15221446752548218,
-0.012687008827924728,
0.08794896304607391,
-0.01709459349513054,
-0.039261043071746826,
0.03245473653078079,
-0.004739575553685427,
-0.007780164014548063,
-0.012593112885951996,
0.045955754816532135,
-0.01393021922558546,
0.0212099626660347,
0.031332869082689285,
0.027367588132619858,
-0.07699437439441681,
0.016726922243833542,
-0.02747504971921444,
0.02136060781776905,
-0.1324019432067871,
0.09388814866542816,
-0.04555235803127289,
-0.004510522820055485,
-0.030180927366018295,
0.006079336162656546,
-0.055863309651613235,
0.01969456672668457,
0.03251754492521286,
0.05782301723957062,
-0.10810942947864532,
0.00008559506386518478,
0.06365300714969635,
-0.11115209758281708,
-0.029000718146562576,
0.037541668862104416,
0.027327222749590874,
0.07392030954360962,
0.053269073367118835,
0.030811434611678123,
0.22001993656158447,
-0.0914575457572937,
0.008493931964039803,
0.0625397264957428,
-0.039241526275873184,
-0.009809576906263828,
0.03784123435616493,
0.06476415693759918,
-0.056885361671447754,
0.0699821412563324,
-0.10860354453325272,
0.06170693412423134,
0.005605973768979311,
-0.02074246108531952,
-0.014899179339408875,
-0.07567930966615677,
-0.04200787842273712,
-0.01332929264754057,
0.004902202636003494,
0.048893626779317856,
-0.043616991490125656,
-0.03283772990107536,
0.13409587740898132,
-0.03522660583257675,
-0.024261007085442543,
-0.11548177152872086,
0.02688085474073887,
0.008192765526473522,
0.036761656403541565,
-0.07052212953567505,
-0.03212612867355347,
0.05182921886444092,
-0.03190960735082626,
0.027880623936653137,
0.049595847725868225,
0.02992655709385872,
0.05969485640525818,
0.008761437609791756,
0.0002356476616114378,
-0.0032550166361033916,
-0.02175113558769226,
-0.020566407591104507,
-0.03958188369870186,
-0.026913661509752274,
-0.039641231298446655,
0.08385860919952393,
-0.057308752089738846,
0.0667286366224289,
-0.03659357503056526,
0.008925306610763073,
0.0029517649672925472,
-0.027433134615421295,
0.034592799842357635,
-0.05432076007127762,
-0.03548414260149002,
-0.04217401146888733,
0.0017054141499102116,
0.060128334909677505,
-0.01794879510998726,
0.08462503552436829,
-0.04114022105932236,
0.05537667125463486,
0.05649912357330322,
0.09640898555517197,
-0.022517584264278412,
-0.04257815331220627,
-0.02496519312262535,
-0.007535780780017376,
0.005643825978040695,
-0.033413805067539215,
0.12946616113185883,
0.015494586899876595,
0.08426713943481445,
-0.043868288397789,
0.02217237465083599,
0.00748229818418622,
0.0015322251711040735,
0.03800797462463379,
-0.02040226384997368,
0.08443927019834518,
-0.021687671542167664,
-0.0016752008814364672,
0.03836734592914581,
-0.0237509123980999,
0.13220903277397156,
0.0025903619825839996,
-0.03288887441158295,
-0.023426497355103493,
0.03498772904276848,
0.001037506852298975,
0.08539588749408722,
-0.09129905700683594,
-0.00039364222902804613,
0.006969362031668425,
-0.027790674939751625,
0.07809425890445709,
-0.059017688035964966,
0.037461746484041214,
-0.0007794335251674056,
-0.029771186411380768,
0.07918456196784973,
-0.001740102656185627,
-0.052296604961156845,
0.0305052250623703,
-0.023542767390608788,
0.027936246246099472,
-0.011083273217082024,
-0.004295621067285538,
-0.033769719302654266,
0.06896239519119263,
-0.04555712267756462,
-0.1238548681139946,
-0.10155307501554489,
0.005631839390844107,
-0.028599046170711517,
0.029226167127490044,
0.009389293380081654,
-0.04930678755044937,
-0.036978792399168015,
-0.05345489829778671,
-0.0466601587831974,
0.010784510523080826,
-0.01835945062339306,
0.039853740483522415,
0.04688549041748047,
0.041759803891181946,
-0.09248657524585724,
0.00859467126429081,
0.016048215329647064,
-0.07155169546604156,
-0.006582000758498907,
-0.02298668771982193,
0.0356617346405983,
0.10543262213468552,
0.0031751571223139763,
0.008329201489686966,
0.012921277433633804,
0.10774405300617218,
-0.03785422071814537,
0.06357379257678986,
0.18202611804008484,
0.019351134076714516,
0.011812274344265461,
0.03921068087220192,
-0.00010539637878537178,
-0.050248511135578156,
0.01667700707912445,
-0.031865525990724564,
-0.044475555419921875,
-0.13560234010219574,
-0.02969539910554886,
-0.039045050740242004,
0.007051459979265928,
0.025256628170609474,
0.041013699024915695,
-0.009950743988156319,
0.09633466601371765,
-0.02397848851978779,
0.049855928868055344,
0.07925064861774445,
0.03943339362740517,
0.013681685552001,
-0.049542710185050964,
0.05907139182090759,
-0.04581200331449509,
0.033267777413129807,
0.0710124522447586,
0.07885682582855225,
0.08591587841510773,
-0.04601209983229637,
0.04559452831745148,
0.03319263830780983,
0.06375846266746521,
0.04932177811861038,
0.07096284627914429,
0.003415021114051342,
0.02214045636355877,
-0.031014522537589073,
-0.042511697858572006,
-0.09021389484405518,
0.03276994451880455,
-0.004167080856859684,
-0.05735289305448532,
-0.01629999838769436,
0.1345573514699936,
0.012132924981415272,
0.13096651434898376,
-0.016075385734438896,
-0.1801685392856598,
-0.05918344855308533,
0.01540380995720625,
-0.014802636578679085,
-0.08348043262958527,
0.029705964028835297,
0.07846545428037643,
-0.04695955663919449,
0.006265506148338318,
-0.021114394068717957,
0.06950525939464569,
-0.047381382435560226,
0.013277146965265274,
0.04551204666495323,
0.11117736250162125,
0.033728718757629395,
0.03963320702314377,
-0.10537557303905487,
0.02526996284723282,
0.01164116058498621,
0.07830831408500671,
-0.06562130898237228,
0.055054422467947006,
0.06645441055297852,
0.04009534418582916,
0.05499923974275589,
-0.02075936086475849,
-0.08796963840723038,
-0.03247752785682678,
-0.1265045553445816,
0.07049684226512909,
0.029228702187538147,
0.007499409839510918,
0.04031297564506531,
-0.04254380613565445,
0.013575805351138115,
-0.040006570518016815,
-0.008947915397584438,
-0.11781232059001923,
-0.1257818043231964,
0.03532278165221214,
0.05714389681816101,
-0.0010479986667633057,
-0.041431985795497894,
0.01611940935254097,
0.05090676620602608,
0.1508808583021164,
-0.08644434809684753,
-0.05597427114844322,
-0.06692858040332794,
-0.02887796424329281,
0.08985653519630432,
-0.01909000426530838,
0.005996709689497948,
-0.0385812409222126,
0.09544891119003296,
0.005407341755926609,
-0.060061976313591,
-0.018916353583335876,
-0.09571436047554016,
-0.049240659922361374,
-0.028131861239671707,
0.022521313279867172,
0.000026796013116836548,
0.01121903769671917,
0.02094685286283493,
-0.06605963408946991,
0.0022880099713802338,
-0.09390750527381897,
-0.045949120074510574,
0.11311960965394974,
-0.04733860120177269,
0.03640663996338844,
-0.11057385802268982,
-0.06688585132360458,
-0.02182316780090332,
0.012587005272507668,
0.006388772279024124,
0.08207748830318451,
-0.029412774369120598,
0.06156962364912033,
0.11187544465065002,
-0.029654646292328835,
-0.1771329939365387,
-0.05076576769351959,
0.08042895793914795,
0.023111604154109955,
-0.022563792765140533,
-0.1804644763469696,
0.05503212288022041,
0.04832097142934799,
-0.024956751614809036,
0.06403253972530365,
-0.13853925466537476,
-0.06443976610898972,
0.0480860136449337,
0.007111556828022003,
0.08165955543518066,
-0.08876204490661621,
-0.022284742444753647,
-0.032246291637420654,
-0.03880226984620094,
0.02686707302927971,
-0.05327668413519859,
0.08781684935092926,
-0.010811097919940948,
-0.053194694221019745,
0.0337652787566185,
-0.04151558130979538,
0.1003669947385788,
-0.0671612024307251,
0.01608673669397831,
-0.03821983188390732,
0.055570848286151886,
0.04639048129320145,
-0.061327435076236725,
0.12789836525917053,
-0.06787239015102386,
0.04854108765721321,
-0.08722209930419922,
-0.02621522918343544,
-0.0006697685457766056,
0.007488863542675972,
0.017287831753492355,
-0.026739731431007385,
-0.028287462890148163,
0.014484849758446217,
-0.0009555828291922808,
0.03557722270488739,
-0.02493773028254509,
-0.00652979826554656,
-0.01772075891494751,
0.17322221398353577,
0.06830696761608124,
-0.07212565094232559,
-0.010356264188885689,
-0.013489754870533943,
0.015356771647930145,
0.07938875257968903,
-0.05805885046720505,
0.0315520204603672,
0.03092469461262226,
-0.03630847856402397,
0.027853410691022873,
0.015626221895217896,
-0.027424294501543045,
0.027008431032299995,
0.05269036442041397,
-0.05085059255361557,
-0.03367950767278671,
-0.051614850759506226,
0.056355513632297516,
-0.028053179383277893,
0.03823434188961983,
0.1114409863948822,
-0.026395471766591072,
0.012246821075677872,
0.0166720412671566,
0.034657031297683716,
-0.012858975678682327,
0.11576013267040253,
-0.002703034318983555,
0.0019600759260356426,
-0.04081074893474579,
0.05825522169470787,
0.03229593113064766,
0.014740144833922386,
0.0226206686347723,
0.04226105660200119,
-0.07054662704467773,
-0.058811917901039124,
-0.09568054974079132,
-0.011163083836436272,
0.02592238411307335,
-0.05799565836787224,
-0.019294261932373047,
-0.034256380051374435,
-0.015381456352770329,
-0.023956473916769028,
0.005732007324695587,
0.01097339577972889,
-0.02139316126704216,
0.029920250177383423,
-0.06831862777471542,
0.03522592782974243,
0.0027535795234143734,
0.035095177590847015,
-0.08792208135128021,
0.02808433771133423,
0.054819170385599136,
0.03733470290899277,
-0.040224701166152954,
-0.0303640253841877,
-0.028331201523542404,
0.009061722084879875,
-0.13267149031162262,
0.03846151381731033,
-0.042245570570230484,
-0.024997256696224213,
0.025977231562137604,
-0.00515158474445343,
-0.03474532812833786,
0.04314378648996353,
-0.006944872904568911,
0.010675568133592606,
-0.04312538355588913,
0.033035919070243835,
-0.017676200717687607,
0.03760131821036339,
0.0037422184832394123,
-0.07564146816730499,
0.05365428701043129,
0.013539206236600876,
-0.030256226658821106,
-0.009939955547451973,
-0.041690245270729065,
-0.024624867364764214,
0.004186462610960007,
0.05057064816355705,
-0.030277004465460777,
0.0008412143215537071,
0.03308509290218353,
0.028900736942887306,
-0.005083378404378891,
-0.040253765881061554,
0.10484524071216583,
-0.03157294541597366,
0.05206746608018875,
0.020011551678180695,
0.001718839630484581,
-0.021946098655462265,
0.004159093834459782,
0.06600460410118103,
0.06866812705993652,
0.05530190095305443,
-0.022144882008433342,
0.001865672878921032,
-0.10110974311828613,
0.02348283864557743,
0.005159699823707342,
0.009437039494514465,
-0.02282087504863739,
-0.034554626792669296,
0.03382909297943115,
-0.004302048124372959,
0.173444002866745,
0.00848334189504385,
-0.005235570482909679,
-0.006306715775281191,
-0.057200852781534195,
-0.003709801472723484,
0.0021933766547590494,
0.022864528000354767,
0.06408777087926865,
0.044701799750328064,
-0.015838660299777985,
0.0011936863884329796,
-0.001726352027617395,
-0.09439557045698166,
0.027980193495750427,
0.030611515045166016,
0.008954260498285294,
0.09822128713130951,
0.01415751501917839,
-0.02480558305978775,
-0.04244313761591911,
0.0623016357421875,
-0.07514170557260513,
0.0636577382683754,
-0.0498967207968235,
0.03034200519323349,
0.10545504093170166,
-0.055225297808647156,
0.01620541885495186,
0.0032247379422187805,
-0.03545834869146347,
-0.10362406075000763,
-0.0769638866186142,
-0.03791984170675278,
-0.0896037295460701,
-0.01724117621779442,
-0.04906086623668671,
-0.017837952822446823,
-0.04912228509783745,
0.013349751010537148,
-0.0034705782309174538,
0.08002873510122299,
0.0022200793027877808,
-0.040786609053611755,
0.015218321233987808,
-0.002257064450532198,
0.014188665896654129,
-0.00480240024626255,
-0.04192284867167473,
0.01692270115017891,
0.04533352702856064,
0.06562313437461853,
0.041802290827035904,
0.023700382560491562,
0.018184740096330643,
0.024437224492430687,
-0.011838145554065704,
-0.027296897023916245,
-0.009246790781617165,
-0.03136183321475983,
0.05188366025686264,
0.010336894541978836,
-0.050348684191703796,
0.013010745868086815,
0.10652954131364822,
-0.01603534445166588,
-0.06802605092525482,
-0.13362576067447662,
0.1374061405658722,
-0.007859395816922188,
0.0010499423369765282,
-0.007302235346287489,
0.0005499641411006451,
-0.03189614787697792,
0.17692683637142181,
0.13029739260673523,
-0.08764059841632843,
-0.023010972887277603,
0.02237706258893013,
-0.003368584904819727,
-0.04554629325866699,
0.1430967003107071,
0.04990725219249725,
0.172064870595932,
0.03017861768603325,
0.005683703348040581,
0.006440362893044949,
-0.009369034320116043,
-0.05443340167403221,
-0.0003233952447772026,
-0.05669066309928894,
0.014083265326917171,
0.0007404517382383347,
0.03178728371858597,
-0.039610862731933594,
-0.1132802814245224,
-0.005930236540734768,
0.018692027777433395,
-0.05690278112888336,
-0.007862056605517864,
0.0025101788341999054,
-0.004896192345768213,
0.03488852456212044,
-0.021922821179032326,
0.020150816068053246,
0.09304726123809814,
-0.034743234515190125,
-0.02984979748725891,
-0.006681990809738636,
0.0644446462392807,
-0.0322665236890316,
0.1895783245563507,
0.003842892125248909,
0.005682679824531078,
0.02343561500310898,
-0.018858274444937706,
-0.11909715831279755,
0.0592271126806736,
-0.019731253385543823,
-0.0940096527338028,
0.010557488538324833,
0.04110215604305267,
-0.04456828162074089,
0.09426715970039368,
0.030472153797745705,
-0.013058120384812355,
-0.0018576704896986485,
0.03059018775820732,
-0.012179136276245117,
-0.047637879848480225,
0.024609286338090897,
-0.06491567194461823,
0.13363716006278992,
0.09713023155927658,
-0.003686778014525771,
-0.006261945236474276,
-0.04795190691947937,
0.07883556932210922,
0.0060691083781421185,
0.03883010894060135,
-0.019560681656003,
-0.12407370656728745,
-0.01673206314444542,
-0.0005330881103873253,
0.028648609295487404,
-0.0905115157365799,
-0.01983395218849182,
-0.011553607881069183,
-0.004039477556943893,
-0.013700045645236969,
0.0627937763929367,
0.07274423539638519,
0.0580732636153698,
-0.018212981522083282,
-0.02175496146082878,
-0.008828938007354736,
0.0333956740796566,
-0.13212445378303528,
-0.07172344624996185
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# wav2vec2-base-finetuned-ks
This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 1.1376
- Accuracy: 0.8210
- F1: 0.8209
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 3e-05
- train_batch_size: 32
- eval_batch_size: 32
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 128
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 50
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
| 1.3731 | 0.99 | 35 | 1.3532 | 0.3767 | 0.2859 |
| 1.3039 | 2.0 | 71 | 1.2740 | 0.4237 | 0.3434 |
| 1.2185 | 2.99 | 106 | 1.1573 | 0.5020 | 0.4423 |
| 1.0887 | 4.0 | 142 | 1.1107 | 0.5013 | 0.4389 |
| 1.0183 | 4.99 | 177 | 1.0801 | 0.5610 | 0.5348 |
| 0.8625 | 6.0 | 213 | 0.9364 | 0.6373 | 0.6285 |
| 0.7487 | 6.99 | 248 | 0.9735 | 0.6048 | 0.5867 |
| 0.6151 | 8.0 | 284 | 0.8946 | 0.6698 | 0.6735 |
| 0.5081 | 8.99 | 319 | 0.8748 | 0.6797 | 0.6855 |
| 0.4559 | 10.0 | 355 | 0.8701 | 0.6850 | 0.6832 |
| 0.4347 | 10.99 | 390 | 0.8887 | 0.7003 | 0.7040 |
| 0.2845 | 12.0 | 426 | 0.8715 | 0.7129 | 0.7145 |
| 0.275 | 12.99 | 461 | 0.8846 | 0.7268 | 0.7263 |
| 0.2301 | 14.0 | 497 | 0.8651 | 0.7261 | 0.7324 |
| 0.1657 | 14.99 | 532 | 0.8573 | 0.7473 | 0.7473 |
| 0.1593 | 16.0 | 568 | 0.8472 | 0.7420 | 0.7443 |
| 0.1398 | 16.99 | 603 | 0.7433 | 0.7825 | 0.7829 |
| 0.1318 | 18.0 | 639 | 0.7989 | 0.7739 | 0.7768 |
| 0.1425 | 18.99 | 674 | 0.7967 | 0.7759 | 0.7788 |
| 0.1116 | 20.0 | 710 | 0.8969 | 0.7659 | 0.7650 |
| 0.0716 | 20.99 | 745 | 0.9783 | 0.7434 | 0.7480 |
| 0.0909 | 22.0 | 781 | 0.9413 | 0.7593 | 0.7626 |
| 0.0691 | 22.99 | 816 | 0.9298 | 0.7832 | 0.7832 |
| 0.068 | 24.0 | 852 | 0.9522 | 0.7725 | 0.7744 |
| 0.0416 | 24.99 | 887 | 0.9624 | 0.7686 | 0.7746 |
| 0.0569 | 26.0 | 923 | 0.9376 | 0.7832 | 0.7832 |
| 0.0369 | 26.99 | 958 | 1.0163 | 0.7845 | 0.7843 |
| 0.0482 | 28.0 | 994 | 1.0013 | 0.7931 | 0.7895 |
| 0.0497 | 28.99 | 1029 | 1.1005 | 0.7725 | 0.7713 |
| 0.0427 | 30.0 | 1065 | 1.0346 | 0.7891 | 0.7901 |
| 0.0252 | 30.99 | 1100 | 1.0611 | 0.7871 | 0.7883 |
| 0.0268 | 32.0 | 1136 | 1.0436 | 0.7944 | 0.7962 |
| 0.022 | 32.99 | 1171 | 1.0217 | 0.8031 | 0.8012 |
| 0.0127 | 34.0 | 1207 | 1.0936 | 0.7971 | 0.7969 |
| 0.0153 | 34.99 | 1242 | 1.0777 | 0.8097 | 0.8055 |
| 0.0062 | 36.0 | 1278 | 1.2379 | 0.7699 | 0.7751 |
| 0.0081 | 36.99 | 1313 | 1.0697 | 0.7977 | 0.7987 |
| 0.0072 | 38.0 | 1349 | 1.1284 | 0.7997 | 0.8001 |
| 0.0105 | 38.99 | 1384 | 1.0593 | 0.8137 | 0.8136 |
| 0.0102 | 40.0 | 1420 | 1.0805 | 0.8130 | 0.8126 |
| 0.0088 | 40.99 | 1455 | 1.1237 | 0.8110 | 0.8115 |
| 0.0073 | 42.0 | 1491 | 1.0980 | 0.8170 | 0.8167 |
| 0.0046 | 42.99 | 1526 | 1.1584 | 0.8044 | 0.8049 |
| 0.0061 | 44.0 | 1562 | 1.1517 | 0.8110 | 0.8114 |
| 0.0021 | 44.99 | 1597 | 1.1564 | 0.8064 | 0.8074 |
| 0.0073 | 46.0 | 1633 | 1.1214 | 0.8183 | 0.8183 |
| 0.002 | 46.99 | 1668 | 1.1376 | 0.8210 | 0.8209 |
| 0.0064 | 48.0 | 1704 | 1.1283 | 0.8210 | 0.8208 |
| 0.0072 | 48.99 | 1739 | 1.1271 | 0.8203 | 0.8201 |
| 0.0019 | 49.3 | 1750 | 1.1273 | 0.8203 | 0.8201 |
### Framework versions
- Transformers 4.36.2
- Pytorch 2.1.1+cu121
- Datasets 2.16.1
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["accuracy", "f1"], "base_model": "facebook/wav2vec2-base", "model-index": [{"name": "wav2vec2-base-finetuned-ks", "results": []}]} | audio-classification | Rest970828/wav2vec2-base-finetuned-ks | [
"transformers",
"safetensors",
"wav2vec2",
"audio-classification",
"generated_from_trainer",
"base_model:facebook/wav2vec2-base",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-08T08:33:35+00:00 | [] | [] | TAGS
#transformers #safetensors #wav2vec2 #audio-classification #generated_from_trainer #base_model-facebook/wav2vec2-base #license-apache-2.0 #endpoints_compatible #region-us
| wav2vec2-base-finetuned-ks
==========================
This model is a fine-tuned version of facebook/wav2vec2-base on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 1.1376
* Accuracy: 0.8210
* F1: 0.8209
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 3e-05
* train\_batch\_size: 32
* eval\_batch\_size: 32
* seed: 42
* gradient\_accumulation\_steps: 4
* total\_train\_batch\_size: 128
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* num\_epochs: 50
### Training results
### Framework versions
* Transformers 4.36.2
* Pytorch 2.1.1+cu121
* Datasets 2.16.1
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.36.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #safetensors #wav2vec2 #audio-classification #generated_from_trainer #base_model-facebook/wav2vec2-base #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.36.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.0"
] | [
62,
144,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #wav2vec2 #audio-classification #generated_from_trainer #base_model-facebook/wav2vec2-base #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50### Training results### Framework versions\n\n\n* Transformers 4.36.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.0"
] | [
-0.14329606294631958,
0.0944584459066391,
-0.0022311178036034107,
0.08478180319070816,
0.1544645130634308,
0.010041214525699615,
0.14536669850349426,
0.0822194367647171,
-0.12154621630907059,
0.06591782718896866,
0.09069937467575073,
0.09433561563491821,
0.03134075924754143,
0.12583814561367035,
-0.04935247078537941,
-0.24673005938529968,
0.013826980255544186,
0.020587053149938583,
-0.12821044027805328,
0.14044195413589478,
0.07720711082220078,
-0.12269282341003418,
0.05294593423604965,
0.008910865522921085,
-0.1707790046930313,
-0.017431942746043205,
0.015711113810539246,
-0.07290342450141907,
0.13628634810447693,
0.007723227608948946,
0.13258734345436096,
0.05690648779273033,
0.11714906245470047,
-0.17471566796302795,
0.01055447943508625,
0.07857298105955124,
0.009965820237994194,
0.09440531581640244,
0.09321106970310211,
0.008280651643872261,
0.06260558217763901,
-0.07569112628698349,
0.05356650799512863,
0.03140972927212715,
-0.12077333778142929,
-0.2639411687850952,
-0.0938318520784378,
0.07605935633182526,
0.1104937270283699,
0.09089509397745132,
-0.010443031787872314,
0.10445994138717651,
-0.0850260853767395,
0.08852753043174744,
0.27392786741256714,
-0.2812771201133728,
-0.06198515370488167,
-0.0013552801683545113,
0.043372876942157745,
0.05028853565454483,
-0.1128656268119812,
-0.023874932900071144,
0.06050385162234306,
0.03401755169034004,
0.13954509794712067,
0.002102506347000599,
-0.036853279918432236,
-0.018537994474172592,
-0.15403759479522705,
-0.05855377018451691,
0.0957561805844307,
0.05793336033821106,
-0.06089678406715393,
-0.038865841925144196,
-0.06755189597606659,
-0.21967218816280365,
-0.03489858657121658,
0.015136335045099258,
0.02902447246015072,
-0.06681374460458755,
-0.10805652290582657,
0.042930711060762405,
-0.08376827090978622,
-0.10934964567422867,
0.013760301284492016,
0.18381810188293457,
0.04740281403064728,
0.01146617904305458,
-0.009911233559250832,
0.13207918405532837,
0.0545964390039444,
-0.17675872147083282,
0.03179645910859108,
0.03960496559739113,
-0.057996224611997604,
-0.02133110538125038,
-0.036096666008234024,
0.020624903962016106,
-0.010178283788263798,
0.13896159827709198,
-0.11050126701593399,
0.01931173726916313,
0.037883210927248,
0.04022083804011345,
-0.09533505141735077,
0.1459318846464157,
-0.09338754415512085,
-0.03689642995595932,
0.006101858802139759,
0.12253120541572571,
0.040630489587783813,
-0.015905266627669334,
-0.08212090283632278,
0.0014730909606441855,
0.1238703727722168,
0.03293205797672272,
-0.038560234010219574,
0.026864957064390182,
-0.05644448101520538,
-0.020442768931388855,
0.07831872254610062,
-0.08953030407428741,
0.016809092834591866,
0.03962886333465576,
-0.06576482206583023,
-0.02513684332370758,
0.008896786719560623,
0.008394808508455753,
0.0028075953014194965,
0.15322978794574738,
-0.08129680901765823,
-0.007674035616219044,
-0.07429137825965881,
-0.10275669395923615,
0.029907749965786934,
-0.07216504216194153,
0.013325104489922523,
-0.07856810092926025,
-0.13080339133739471,
-0.01374809443950653,
0.03427092730998993,
-0.021571384742856026,
-0.08312662690877914,
-0.023117635399103165,
-0.1236586719751358,
0.02862902730703354,
-0.03495476394891739,
0.13404691219329834,
-0.06606525927782059,
0.13826416432857513,
0.048842668533325195,
0.05841084197163582,
0.01685965247452259,
0.05305621773004532,
-0.08198712766170502,
0.05120133236050606,
-0.19099372625350952,
0.026836881414055824,
-0.05751221626996994,
0.01870867609977722,
-0.08243401348590851,
-0.133474200963974,
0.029372472316026688,
-0.013039892539381981,
0.09803380072116852,
0.10127893090248108,
-0.1566808521747589,
-0.10776615887880325,
0.1474255621433258,
-0.11635399609804153,
-0.11490186303853989,
0.1300329715013504,
-0.009240997023880482,
-0.04292716458439827,
0.038931991904973984,
0.11186691373586655,
0.09793559461832047,
-0.1134110540151596,
-0.03871142491698265,
-0.0343942791223526,
0.0800730511546135,
-0.02757652848958969,
0.104102224111557,
0.026543980464339256,
0.02502436377108097,
-0.007300384808331728,
-0.06753582507371902,
0.06651028245687485,
-0.11295481771230698,
-0.08162646740674973,
-0.03079274110496044,
-0.09859296679496765,
0.07833755016326904,
0.06372994929552078,
0.02181006781756878,
-0.07748332619667053,
-0.10505567491054535,
0.06006499007344246,
0.11350986361503601,
-0.07046259939670563,
0.013569497503340244,
-0.06519906222820282,
0.10136585682630539,
-0.0747886449098587,
-0.03336417302489281,
-0.19545504450798035,
-0.03160068020224571,
0.013096877373754978,
-0.030781200155615807,
-0.007035934366285801,
-0.014456209726631641,
0.05953913554549217,
0.11500831693410873,
-0.068149633705616,
-0.09803280234336853,
-0.09745489805936813,
-0.003533271374180913,
-0.07773184031248093,
-0.23460513353347778,
-0.09386546909809113,
-0.015374066308140755,
0.1377600133419037,
-0.19293415546417236,
0.0186480600386858,
0.010428251698613167,
0.10795459151268005,
0.03563258796930313,
-0.024908147752285004,
-0.019723666831851006,
0.09531471133232117,
-0.010208334773778915,
-0.05729398876428604,
0.03778998926281929,
0.010027398355305195,
-0.07027188688516617,
-0.020303303375840187,
-0.11505406349897385,
0.20148532092571259,
0.11837408691644669,
-0.026044271886348724,
-0.08850489556789398,
0.019121142104268074,
-0.09079567342996597,
-0.05133875086903572,
-0.06514270603656769,
0.00988877471536398,
0.14667388796806335,
0.030223123729228973,
0.13619935512542725,
-0.09851551800966263,
-0.038424041122198105,
0.05275348946452141,
-0.007000318728387356,
0.03265739604830742,
0.11021076887845993,
0.07953207939863205,
-0.0719224065542221,
0.1387217938899994,
0.15070658922195435,
-0.09534663707017899,
0.132640078663826,
-0.08198074251413345,
-0.09732658416032791,
-0.02228783443570137,
-0.011494534090161324,
0.010729263536632061,
0.15049342811107635,
-0.06904773414134979,
0.019922031089663506,
0.01805451326072216,
0.017602233216166496,
-0.005193319637328386,
-0.22743916511535645,
-0.027325892820954323,
0.026759520173072815,
-0.059844497591257095,
-0.08028405904769897,
-0.023708036169409752,
0.025929490104317665,
0.10895277559757233,
-0.015105362981557846,
-0.093205027282238,
0.01485749427229166,
0.002404323546215892,
-0.07450275868177414,
0.1945847123861313,
-0.05915718153119087,
-0.12412989884614944,
-0.13247740268707275,
-0.005267052445560694,
-0.03755445033311844,
-0.006777768023312092,
0.04528011381626129,
-0.08673740923404694,
-0.02249211259186268,
-0.04233551025390625,
0.04166838526725769,
0.03524314612150192,
0.04186669737100601,
0.02147744782269001,
0.019610313698649406,
0.09586971998214722,
-0.09540770947933197,
0.026147952303290367,
-0.035335928201675415,
-0.04982329159975052,
0.027105296030640602,
0.08359746634960175,
0.08644624054431915,
0.1611497402191162,
-0.0029717334546148777,
-0.007661997340619564,
-0.019819648936390877,
0.20959019660949707,
-0.1003064215183258,
-0.02349526807665825,
0.161838099360466,
-0.031725648790597916,
0.04922737553715706,
0.15458717942237854,
0.06918913871049881,
-0.07015909999608994,
0.00023244648764375597,
0.031952183693647385,
-0.03142717853188515,
-0.23403559625148773,
-0.040617555379867554,
-0.03532365709543228,
0.005851421505212784,
0.08529473841190338,
0.009502808563411236,
-0.007183260750025511,
0.04252811148762703,
-0.030042031779885292,
0.028405824676156044,
-0.01283994223922491,
0.043834391981363297,
0.0450848713517189,
0.05290897190570831,
0.12811724841594696,
-0.023416006937623024,
-0.04598177969455719,
0.01135964971035719,
-0.012335021048784256,
0.21013396978378296,
-0.041208457201719284,
0.11421577632427216,
0.054887160658836365,
0.16604264080524445,
0.02856779657304287,
0.11026956886053085,
0.0068249087780714035,
-0.03409165143966675,
0.01769855245947838,
-0.057657450437545776,
-0.02657318487763405,
-0.002701336285099387,
-0.033940572291612625,
0.07861834764480591,
-0.1540181189775467,
0.016023384407162666,
0.01821930892765522,
0.3221757709980011,
0.07859990000724792,
-0.32906320691108704,
-0.12499392777681351,
-0.017885301262140274,
-0.034330666065216064,
-0.028994621708989143,
0.02291417308151722,
0.15470419824123383,
-0.07818129658699036,
0.049631088972091675,
-0.06679819524288177,
0.0810665488243103,
-0.041699208319187164,
0.0259222611784935,
0.10540765523910522,
0.09439842402935028,
-0.0011267708614468575,
0.048551470041275024,
-0.2134244292974472,
0.2984776496887207,
0.009084708988666534,
0.07963499426841736,
-0.03325297683477402,
0.005121932830661535,
0.02148209512233734,
0.04791201278567314,
0.06069137901067734,
-0.014757621102035046,
-0.03936994448304176,
-0.22272352874279022,
-0.10246032476425171,
0.012762938626110554,
0.1095055341720581,
-0.04130656644701958,
0.1141056939959526,
-0.027223961427807808,
-0.028254598379135132,
0.05754915252327919,
-0.06622659415006638,
-0.07649001479148865,
-0.052976761013269424,
0.0030829166062176228,
0.013693577609956264,
0.04953506216406822,
-0.1065705418586731,
-0.1410062611103058,
-0.0891680046916008,
0.12794804573059082,
-0.08005046099424362,
-0.039475586265325546,
-0.11846678704023361,
0.09113793075084686,
0.1379612684249878,
-0.07057806104421616,
0.05419241636991501,
0.023610370233654976,
0.12541912496089935,
0.011590740643441677,
-0.042743273079395294,
0.0912412628531456,
-0.09042048454284668,
-0.2575738728046417,
-0.04455538094043732,
0.1749536246061325,
0.03680936619639397,
0.06991684436798096,
-0.025531524792313576,
0.02250073477625847,
0.023929385468363762,
-0.07795688509941101,
0.028681829571723938,
-0.0365845188498497,
0.05684462562203407,
0.03176630288362503,
-0.016409719362854958,
-0.02692202292382717,
-0.04783123731613159,
-0.02631465531885624,
0.11981154978275299,
0.3081210255622864,
-0.0830603539943695,
0.021778006106615067,
0.07951425015926361,
-0.02280590496957302,
-0.18325477838516235,
0.019834553822875023,
0.12583687901496887,
0.028731219470500946,
0.03826802596449852,
-0.16612966358661652,
0.09225527942180634,
0.08101123571395874,
-0.038520071655511856,
0.08764757961034775,
-0.2979062497615814,
-0.1254022866487503,
0.08466175943613052,
0.12751103937625885,
0.05270726606249809,
-0.13922083377838135,
-0.03291456401348114,
-0.009244779124855995,
-0.12478812783956528,
0.12218125909566879,
-0.0943320170044899,
0.11336015909910202,
-0.013384264893829823,
0.033663854002952576,
0.017028389498591423,
-0.0627661719918251,
0.11438983678817749,
-0.0027479452546685934,
0.08907585591077805,
-0.005295301787555218,
0.012322835624217987,
0.07858987897634506,
-0.0432116836309433,
-0.013626505620777607,
-0.06402693688869476,
0.04131615534424782,
-0.042745452374219894,
-0.012819744646549225,
-0.1133083775639534,
0.021989518776535988,
-0.061752088367938995,
-0.046860046684741974,
-0.0052502756007015705,
0.038301266729831696,
0.027612511068582535,
-0.028950285166502,
0.12451034039258957,
0.02801761031150818,
0.15641158819198608,
0.12385642528533936,
0.08187928050756454,
-0.029310571029782295,
-0.08608661592006683,
-0.017403244972229004,
-0.028037531301379204,
0.07671492546796799,
-0.11277936398983002,
0.0343284010887146,
0.12747196853160858,
0.06878556311130524,
0.10388585180044174,
0.07245408743619919,
-0.04964325949549675,
0.00315496395342052,
0.06915983557701111,
-0.1428353488445282,
-0.07242829352617264,
-0.01933976449072361,
0.0011254705023020506,
-0.13610419631004333,
0.057974766939878464,
0.11307238787412643,
-0.07962232828140259,
-0.011080202646553516,
0.007315858732908964,
0.010573037900030613,
-0.05667593330144882,
0.23846620321273804,
0.05860648676753044,
0.10054559260606766,
-0.11517708003520966,
0.07412544637918472,
0.04172616824507713,
-0.13504533469676971,
-0.004182741511613131,
0.05120040476322174,
-0.08042296767234802,
-0.01184081844985485,
0.052806656807661057,
0.08513970673084259,
-0.02995835803449154,
-0.06254154443740845,
-0.14345692098140717,
-0.14754265546798706,
0.0671052560210228,
0.1745375096797943,
0.06934100389480591,
0.04761794954538345,
-0.021216480061411858,
0.023840203881263733,
-0.13726572692394257,
0.1213754191994667,
0.06078412011265755,
0.08451281487941742,
-0.18810786306858063,
0.15838254988193512,
0.017810622230172157,
0.041281089186668396,
-0.019447078928351402,
0.039971642196178436,
-0.10302646458148956,
0.01825854741036892,
-0.11902976781129837,
-0.025530975311994553,
-0.027797434478998184,
-0.0011839353246614337,
-0.024350864812731743,
-0.049615610390901566,
-0.05960807576775551,
0.03242045268416405,
-0.09957876056432724,
-0.02449805848300457,
0.01080678217113018,
0.03098907135426998,
-0.1382845789194107,
-0.01825684681534767,
0.045750170946121216,
-0.11256040632724762,
0.0939115583896637,
0.054252274334430695,
0.036623764783144,
0.0465189591050148,
-0.1189408004283905,
-0.025902099907398224,
0.06838485598564148,
-0.0021362348925322294,
0.05138549953699112,
-0.13997292518615723,
0.0036562730092555285,
-0.031161438673734665,
0.016453927382826805,
0.0004601225664373487,
0.06690430641174316,
-0.12569516897201538,
-0.002035943092778325,
-0.014034031890332699,
-0.03039376437664032,
-0.05661316215991974,
0.014308490790426731,
0.09793733060359955,
0.006342905107885599,
0.18959179520606995,
-0.07935471832752228,
0.04351237788796425,
-0.2341158539056778,
-0.0025773560628294945,
-0.03194309026002884,
-0.1038207858800888,
-0.13151462376117706,
-0.024487273767590523,
0.08251014351844788,
-0.050533223897218704,
0.0684746727347374,
-0.04233698546886444,
0.08781363815069199,
0.02762993425130844,
-0.04144960641860962,
0.008492464199662209,
0.05892118811607361,
0.18628761172294617,
0.03180402144789696,
-0.027879241853952408,
0.049772921949625015,
0.012621515430510044,
0.07771363109350204,
0.0939621776342392,
0.18213920295238495,
0.14036914706230164,
-0.002506556222215295,
0.06507489830255508,
0.07371217012405396,
-0.10442809760570526,
-0.18412554264068604,
0.04101592302322388,
-0.07082685828208923,
0.10868649929761887,
-0.0030544260516762733,
0.18965159356594086,
0.09563251584768295,
-0.1720227599143982,
0.04120253026485443,
-0.039988793432712555,
-0.0722573846578598,
-0.0953788086771965,
-0.000583604269195348,
-0.061685942113399506,
-0.15953415632247925,
0.01777053251862526,
-0.11510877311229706,
0.011878297664225101,
0.09592913091182709,
0.006117450539022684,
0.011493266560137272,
0.17238731682300568,
0.06589628010988235,
0.0323459692299366,
0.06698664277791977,
0.033851053565740585,
-0.024324262514710426,
-0.0219526756554842,
-0.08526752144098282,
0.04975767433643341,
-0.03419750928878784,
0.03602558374404907,
-0.06845836341381073,
-0.10145688056945801,
0.08008893579244614,
0.026611702516674995,
-0.11254002153873444,
0.03611269220709801,
-0.012338816188275814,
0.08835428208112717,
0.06424643099308014,
-0.0005317772156558931,
0.03878258541226387,
-0.019424352794885635,
0.2255677431821823,
-0.08730094879865646,
-0.06941650807857513,
-0.11960361152887344,
0.25884705781936646,
-0.00018439882842358202,
-0.025666335597634315,
0.05361571162939072,
-0.07849572598934174,
-0.021378375589847565,
0.15528357028961182,
0.15373677015304565,
-0.052879225462675095,
-0.006265481933951378,
0.0024409270845353603,
-0.018024416640400887,
-0.05919964611530304,
0.09541859477758408,
0.13513973355293274,
0.10524224489927292,
-0.09411939233541489,
-0.04958600923418999,
-0.07134201377630234,
-0.03368202969431877,
0.007204351481050253,
0.0661809891462326,
0.007838917896151543,
-0.00932739395648241,
-0.04020944610238075,
0.0777726024389267,
-0.05493592470884323,
-0.14165957272052765,
0.055851951241493225,
-0.20009459555149078,
-0.1840747594833374,
-0.03373230621218681,
0.10198963433504105,
0.03117610700428486,
0.055664680898189545,
-0.009792769327759743,
-0.02263724058866501,
0.0827297791838646,
-0.01750810258090496,
-0.029811527580022812,
-0.09163596481084824,
0.06894691288471222,
-0.07256923615932465,
0.17885224521160126,
-0.04531532898545265,
0.08185125142335892,
0.10344581305980682,
0.06330130994319916,
-0.10126449167728424,
0.04554620757699013,
0.06406737864017487,
-0.15804579854011536,
0.008054179139435291,
0.18270112574100494,
-0.03964165970683098,
0.0942121222615242,
0.022526191547513008,
-0.15793287754058838,
-0.002970695262774825,
-0.04876558110117912,
-0.053093716502189636,
-0.041342537850141525,
-0.03144976496696472,
-0.01925945095717907,
0.12608404457569122,
0.19501784443855286,
-0.05411801487207413,
0.002011211821809411,
-0.06181305646896362,
0.022172512486577034,
0.0774928629398346,
0.06859230250120163,
-0.007937147282063961,
-0.28911182284355164,
0.04213311895728111,
0.023907020688056946,
-0.009419955313205719,
-0.23447610437870026,
-0.07524372637271881,
0.03470039740204811,
-0.07465633749961853,
-0.1164964884519577,
0.06967634707689285,
0.03465425968170166,
0.057195693254470825,
-0.04257576912641525,
-0.033681076020002365,
-0.07996956259012222,
0.17124062776565552,
-0.18892541527748108,
-0.09381520748138428
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# roberta-base-squad-model1
This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the squad dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 16
- seed: 62
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| {"license": "mit", "tags": ["generated_from_trainer"], "datasets": ["varun-v-rao/squad"], "base_model": "roberta-base", "model-index": [{"name": "roberta-base-squad-model1", "results": []}]} | question-answering | varun-v-rao/roberta-base-squad-model1 | [
"transformers",
"tensorboard",
"safetensors",
"roberta",
"question-answering",
"generated_from_trainer",
"dataset:varun-v-rao/squad",
"base_model:roberta-base",
"license:mit",
"endpoints_compatible",
"region:us"
] | 2024-02-08T08:37:53+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #roberta #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-roberta-base #license-mit #endpoints_compatible #region-us
|
# roberta-base-squad-model1
This model is a fine-tuned version of roberta-base on the squad dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 16
- seed: 62
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| [
"# roberta-base-squad-model1\n\nThis model is a fine-tuned version of roberta-base on the squad dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 62\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #roberta #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-roberta-base #license-mit #endpoints_compatible #region-us \n",
"# roberta-base-squad-model1\n\nThis model is a fine-tuned version of roberta-base on the squad dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 62\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
69,
32,
6,
12,
8,
3,
90,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #roberta #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-roberta-base #license-mit #endpoints_compatible #region-us \n# roberta-base-squad-model1\n\nThis model is a fine-tuned version of roberta-base on the squad dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 62\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3### Training results### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
-0.05920407176017761,
0.044143032282590866,
-0.0017031957395374775,
0.0719393938779831,
0.1878899335861206,
0.01474237721413374,
0.14789164066314697,
0.07593148201704025,
-0.14720875024795532,
0.052473295480012894,
0.06580695509910583,
0.08583996444940567,
0.011615472845733166,
0.09776519238948822,
-0.03962094709277153,
-0.23033282160758972,
0.0051600332371890545,
-0.00785842165350914,
-0.12911322712898254,
0.09627899527549744,
0.09951755404472351,
-0.11870641261339188,
0.05999325215816498,
-0.007106272969394922,
-0.21964314579963684,
0.05520259961485863,
0.0073182606138288975,
-0.035673968493938446,
0.10662443935871124,
0.02048671245574951,
0.1388801634311676,
0.0017827902920544147,
0.1354358047246933,
-0.2111060619354248,
0.01794152520596981,
0.0902932807803154,
0.020737159997224808,
0.06484099477529526,
0.02303813025355339,
-0.010786986909806728,
0.12536932528018951,
-0.1277768760919571,
0.0929955467581749,
0.01982739008963108,
-0.0913863331079483,
-0.1673450917005539,
-0.09124065935611725,
0.027908334508538246,
0.0852721557021141,
0.10314677655696869,
-0.013250358402729034,
0.1769126057624817,
-0.11856980621814728,
0.0782354325056076,
0.19249936938285828,
-0.277554452419281,
-0.10495509207248688,
0.10687453299760818,
0.04900875687599182,
0.09101510047912598,
-0.1139296218752861,
-0.002589658834040165,
0.050848498940467834,
0.03165986016392708,
0.08054650574922562,
-0.026290979236364365,
-0.08398357778787613,
0.007838387973606586,
-0.14674495160579681,
0.03364168480038643,
0.12515489757061005,
0.05435545742511749,
-0.03285912051796913,
-0.022019945085048676,
-0.04309344291687012,
-0.027842774987220764,
-0.039893876761198044,
-0.059488918632268906,
0.04911809042096138,
-0.04822070524096489,
-0.1344042420387268,
-0.03354596719145775,
-0.0821462944149971,
-0.0704788938164711,
-0.025642147287726402,
0.1358613818883896,
0.04385233670473099,
0.011767160147428513,
-0.0661073625087738,
0.08187387883663177,
-0.015216432511806488,
-0.10425706952810287,
0.01548024546355009,
0.007686886005103588,
-0.07965569198131561,
-0.07559341192245483,
-0.05668747425079346,
-0.058964237570762634,
0.04078443720936775,
0.1827663779258728,
-0.052045051008462906,
0.05613407492637634,
0.005129185970872641,
0.00998068880289793,
-0.04983026161789894,
0.124854676425457,
-0.06642118096351624,
-0.06701140850782394,
0.015555244870483875,
0.062465935945510864,
0.01587161235511303,
0.0092360470443964,
-0.08105818927288055,
0.0128678223118186,
0.07945549488067627,
0.034938789904117584,
-0.06087939441204071,
0.039081376045942307,
0.00038929429138079286,
-0.006554193329066038,
-0.03501320630311966,
-0.09977206587791443,
0.03622742369771004,
-0.011558075435459614,
-0.061519477516412735,
-0.015470115467905998,
0.012847460806369781,
0.025404594838619232,
0.018192652612924576,
0.11123178899288177,
-0.09882994741201401,
0.01497410237789154,
-0.11273729801177979,
-0.09998726844787598,
-0.0022826939821243286,
-0.06570396572351456,
0.005749686621129513,
-0.08404431492090225,
-0.14283771812915802,
-0.04401077330112457,
0.04553960636258125,
-0.04737871140241623,
-0.008941266685724258,
-0.051560260355472565,
-0.08721626549959183,
-0.004240528214722872,
0.006750693079084158,
0.16396494209766388,
-0.039522115141153336,
0.06936739385128021,
0.03290165960788727,
0.035741787403821945,
-0.027545413002371788,
0.021631799638271332,
-0.08869779855012894,
0.014418006874620914,
-0.1821782886981964,
0.03552158176898956,
-0.09128449857234955,
0.06912414729595184,
-0.10201865434646606,
-0.08826133608818054,
-0.017443975433707237,
0.007607975043356419,
0.0778508111834526,
0.0736289843916893,
-0.16529680788516998,
-0.03693598136305809,
0.1643420308828354,
-0.08959919959306717,
-0.08750048279762268,
0.09162123501300812,
-0.06815098226070404,
0.06759396940469742,
0.07292032241821289,
0.1590459793806076,
0.05313257500529289,
-0.1089048832654953,
-0.005676796659827232,
-0.038953665643930435,
0.04879327118396759,
-0.01157404389232397,
0.039513394236564636,
0.008161625824868679,
0.028423694893717766,
0.005766968242824078,
-0.06566178053617477,
0.020193278789520264,
-0.10995318740606308,
-0.07380589097738266,
-0.04332321882247925,
-0.09719963371753693,
-0.03246841952204704,
0.05307084694504738,
0.05978289991617203,
-0.08959009498357773,
-0.09310127794742584,
0.15328942239284515,
0.10735861212015152,
-0.06066516786813736,
0.01715754345059395,
-0.08365309238433838,
0.04763574153184891,
-0.06992438435554504,
-0.026395412161946297,
-0.19176116585731506,
-0.11364080756902695,
-0.0076235206797719,
0.019122010096907616,
0.058674369007349014,
0.040168821811676025,
0.0713391974568367,
0.056692395359277725,
-0.05251602455973625,
0.013455440290272236,
-0.06690102070569992,
0.008773928508162498,
-0.09593174606561661,
-0.23959651589393616,
-0.022159315645694733,
-0.02886180952191353,
0.18088237941265106,
-0.2561984956264496,
0.021028432995080948,
-0.051591262221336365,
0.1338895559310913,
0.03250269219279289,
-0.037960268557071686,
-0.04593309015035629,
0.06843789666891098,
-0.00402615824714303,
-0.07337833195924759,
0.04264301434159279,
-0.004196495749056339,
-0.06372623890638351,
-0.08295358717441559,
-0.1381240040063858,
0.0823015421628952,
0.10047256201505661,
-0.03742733597755432,
-0.10044776648283005,
0.026895305141806602,
-0.06661789119243622,
-0.02893480844795704,
-0.06774736940860748,
0.021439753472805023,
0.12761276960372925,
-0.011738781817257404,
0.12217473983764648,
-0.0530833937227726,
-0.04623865336179733,
0.0018936933483928442,
-0.03461573272943497,
0.03842094540596008,
0.08716137707233429,
0.12797512114048004,
-0.1249345988035202,
0.09208451211452484,
0.08226097375154495,
-0.10400818288326263,
0.13355273008346558,
-0.04260821267962456,
-0.06729771941900253,
-0.029339689761400223,
-0.01346998754888773,
0.007581404410302639,
0.14006729423999786,
-0.08006589114665985,
0.0023425049148499966,
-0.0018699460197240114,
0.014442786574363708,
0.04217331483960152,
-0.18395574390888214,
-0.029776329174637794,
0.0027488695923238993,
-0.017582273110747337,
0.0045397463254630566,
-0.007695501204580069,
0.018322531133890152,
0.10055606812238693,
0.011826523579657078,
-0.032228462398052216,
-0.00449928967282176,
-0.007410059217363596,
-0.07996059209108353,
0.20286878943443298,
-0.07939238101243973,
-0.04763556644320488,
-0.0855972170829773,
-0.014908738434314728,
-0.03948814049363136,
-0.03064996749162674,
0.028197219595313072,
-0.11883339285850525,
-0.04719208925962448,
-0.0830560028553009,
0.01513221487402916,
0.0006268945289775729,
0.0006955474382266402,
0.032473303377628326,
0.027002213522791862,
0.0895044207572937,
-0.13179346919059753,
0.012228624895215034,
-0.08175776153802872,
-0.13485033810138702,
0.007477977778762579,
0.05810458958148956,
0.12608271837234497,
0.11847425997257233,
-0.018807223066687584,
0.011787175200879574,
-0.025303814560174942,
0.23700706660747528,
-0.06229249760508537,
-0.011875903233885765,
0.10275198519229889,
0.009286347776651382,
0.030426859855651855,
0.12544475495815277,
0.05696769058704376,
-0.11604319512844086,
0.04420899599790573,
0.1065066009759903,
-0.03293987736105919,
-0.23317460715770721,
-0.03007524088025093,
-0.03699331358075142,
-0.11217384040355682,
0.04648536071181297,
0.040745921432971954,
0.003704485483467579,
0.05586346983909607,
0.026382258161902428,
0.056950636208057404,
-0.02458801306784153,
0.07984784245491028,
0.08505643904209137,
0.05023965612053871,
0.12872247397899628,
-0.05858362466096878,
-0.06409995257854462,
0.036785103380680084,
-0.023268602788448334,
0.30611932277679443,
-0.003194541437551379,
0.02064412273466587,
0.0973794087767601,
0.11507151275873184,
-0.018531503155827522,
0.018962625414133072,
0.0021297538187354803,
-0.04871685802936554,
0.0005249847308732569,
-0.05502072721719742,
0.010564741678535938,
0.0066042207181453705,
-0.06755615770816803,
0.060621801763772964,
-0.06443599611520767,
0.04844127967953682,
0.05495206266641617,
0.2165038287639618,
0.011145939119160175,
-0.2693665027618408,
-0.07910964637994766,
0.010891124606132507,
-0.02937241829931736,
-0.008684366941452026,
0.006190172396600246,
0.1099800318479538,
-0.09165090322494507,
0.02408842369914055,
-0.059976015239953995,
0.09234083443880081,
0.005025257356464863,
0.009440446272492409,
0.041086796671152115,
0.17407438158988953,
-0.015712877735495567,
0.059943173080682755,
-0.22912906110286713,
0.24130567908287048,
0.009661818854510784,
0.14135821163654327,
-0.021076537668704987,
0.0010172688635066152,
0.022401290014386177,
0.0732278823852539,
0.05415679141879082,
-0.00618040282279253,
-0.054281000047922134,
-0.16239657998085022,
-0.013110999949276447,
0.06863716244697571,
0.1233988031744957,
-0.007599725853651762,
0.11039022356271744,
-0.045904599130153656,
0.011567531153559685,
0.06919721513986588,
-0.05753093585371971,
-0.18327045440673828,
-0.09722457826137543,
-0.036869991570711136,
-0.00814721081405878,
-0.09209788590669632,
-0.08381492644548416,
-0.09279836714267731,
-0.0684744194149971,
0.1534993052482605,
0.023951366543769836,
0.012224894016981125,
-0.1190793439745903,
0.14311130344867706,
0.0888378843665123,
-0.05937375873327255,
0.010828502476215363,
0.012954184785485268,
0.07480176538228989,
0.04113459587097168,
-0.06824241578578949,
0.06901117414236069,
-0.07591524720191956,
-0.13865867257118225,
-0.059184759855270386,
0.08130259811878204,
0.07291480898857117,
0.03894585743546486,
0.0012654238380491734,
0.032292190939188004,
0.0171640757471323,
-0.09178958833217621,
-0.0004689711204264313,
0.04954390227794647,
0.04849055036902428,
0.0451129712164402,
-0.08028063178062439,
-0.019253617152571678,
-0.04500412940979004,
-0.02338770404458046,
0.11237820982933044,
0.23282849788665771,
-0.07677523046731949,
0.014730319380760193,
0.05299776792526245,
-0.08238568902015686,
-0.17802469432353973,
0.12847864627838135,
0.0805501863360405,
-0.0015270111616700888,
0.06081932410597801,
-0.14533087611198425,
0.18876534700393677,
0.11574002355337143,
-0.013154588639736176,
0.03231699392199516,
-0.2861500382423401,
-0.13510377705097198,
0.07024767994880676,
0.16018283367156982,
0.09751645475625992,
-0.15983925759792328,
-0.01589246653020382,
-0.03355143591761589,
-0.13936258852481842,
0.10921438783407211,
-0.18322789669036865,
0.10086234658956528,
0.012631269171833992,
0.07511855661869049,
-0.0007976166089065373,
-0.046365972608327866,
0.1191440150141716,
0.02238302119076252,
0.1170324981212616,
-0.055302396416664124,
-0.004532202612608671,
0.126687154173851,
-0.0362434908747673,
-0.012781620025634766,
-0.022851549088954926,
0.05164739862084389,
-0.07492601871490479,
-0.015827087685465813,
-0.07619200646877289,
0.04435383901000023,
-0.05523889884352684,
-0.059416092932224274,
-0.05689515173435211,
0.06295231729745865,
0.037599556148052216,
-0.023931678384542465,
0.07951825857162476,
-0.016693996265530586,
0.1587604582309723,
0.012718315236270428,
0.09252927452325821,
-0.006988616194576025,
-0.01997286267578602,
0.01725725643336773,
-0.019513806328177452,
0.05643235519528389,
-0.14614930748939514,
0.0163701344281435,
0.11138232052326202,
0.03670801222324371,
0.1309238076210022,
0.060136809945106506,
-0.05647926405072212,
0.02566470019519329,
0.0704585388302803,
-0.09629512578248978,
-0.15289306640625,
0.03871258720755577,
-0.06581480801105499,
-0.11635015904903412,
0.06266430020332336,
0.1136530265212059,
-0.061366867274045944,
-0.006643127650022507,
-0.014435109682381153,
-0.002966152736917138,
-0.06282666325569153,
0.1826573610305786,
0.06878664344549179,
0.04336214065551758,
-0.08226891607046127,
0.08458135277032852,
0.047043297439813614,
-0.0371459424495697,
0.02514803595840931,
0.07886187732219696,
-0.08156877011060715,
-0.016710784286260605,
0.07560644298791885,
0.24087797105312347,
-0.06980195641517639,
-0.029175743460655212,
-0.13929837942123413,
-0.0910111591219902,
0.030250297859311104,
0.17753562331199646,
0.0659310519695282,
-0.04309043660759926,
-0.031637027859687805,
0.06126068904995918,
-0.1632392704486847,
0.07800161838531494,
0.043057553470134735,
0.0792735368013382,
-0.12100381404161453,
0.10628683865070343,
0.02033848688006401,
-0.004477962851524353,
-0.02197236195206642,
0.032965537160634995,
-0.1391047090291977,
-0.0086982948705554,
-0.13443587720394135,
-0.037090595811605453,
-0.023109454661607742,
0.00008763228106545284,
-0.00018456876568961889,
-0.06437800079584122,
-0.08749247342348099,
0.0379086472094059,
-0.07702022045850754,
-0.03545030578970909,
0.035361867398023605,
0.041749268770217896,
-0.14576922357082367,
0.006456912495195866,
0.014333274215459824,
-0.07087139040231705,
0.05374062433838844,
0.04780092462897301,
0.0426480770111084,
0.06074722483754158,
-0.20754893124103546,
-0.01029976923018694,
0.02529778890311718,
-0.0008001935784704983,
0.08359251171350479,
-0.06644289195537567,
-0.0072166030295193195,
-0.014997798949480057,
0.10795681923627853,
0.0238712839782238,
0.055429037660360336,
-0.10411933064460754,
0.022975139319896698,
-0.08869729191064835,
-0.06613356620073318,
-0.05092778429389,
0.02638767659664154,
0.08506669104099274,
0.0471123568713665,
0.17704550921916962,
-0.1050257459282875,
0.026025323197245598,
-0.2157045602798462,
-0.030536163598299026,
-0.0021916108671575785,
-0.04447987303137779,
-0.09089863300323486,
-0.04662563279271126,
0.0641099289059639,
-0.06693507730960846,
0.11860539019107819,
0.006223978940397501,
0.07849550992250443,
0.053408075124025345,
-0.044410500675439835,
-0.0067949783988296986,
0.021781284362077713,
0.1898665875196457,
0.06055641174316406,
-0.005161673296242952,
0.050591081380844116,
0.014564555138349533,
0.06430260092020035,
0.0267376396805048,
0.1959640234708786,
0.1938345581293106,
-0.04258577153086662,
0.05615140125155449,
0.07922782748937607,
-0.06982813775539398,
-0.07633399963378906,
0.1539587527513504,
-0.04217853024601936,
0.055766113102436066,
-0.04934315010905266,
0.1437252312898636,
0.15938299894332886,
-0.15200258791446686,
0.018391290679574013,
-0.07586848735809326,
-0.08791964501142502,
-0.11631567776203156,
0.0007028262480162084,
-0.09537535160779953,
-0.13785521686077118,
0.0478234700858593,
-0.12586814165115356,
0.006897837854921818,
0.11644883453845978,
0.018784644082188606,
0.018595658242702484,
0.1695287674665451,
0.003322263015434146,
0.043035849928855896,
0.022705117240548134,
-0.010936332866549492,
-0.026654040440917015,
-0.04897356033325195,
-0.04287160187959671,
0.02855786122381687,
-0.012602905742824078,
0.05758990719914436,
-0.049075473099946976,
-0.034090906381607056,
0.043495118618011475,
-0.024781743064522743,
-0.07303063571453094,
0.03659426048398018,
0.040179766714572906,
0.03539351373910904,
0.043593332171440125,
0.03920977935194969,
-0.010400844737887383,
-0.03102579526603222,
0.281848281621933,
-0.06165772303938866,
-0.12835165858268738,
-0.13831403851509094,
0.22172628343105316,
0.031658921390771866,
0.0005038183880969882,
0.018291164189577103,
-0.10652194917201996,
0.017570998519659042,
0.18075472116470337,
0.19589778780937195,
-0.07911903411149979,
-0.008162182755768299,
0.001658173743635416,
-0.02492341212928295,
-0.0789145827293396,
0.12686868011951447,
0.10444708913564682,
0.03900905326008797,
-0.056203220039606094,
-0.04826309159398079,
-0.03325766697525978,
-0.023477233946323395,
-0.04690922051668167,
0.029690345749258995,
0.05973184108734131,
0.011355431750416756,
-0.028251130133867264,
0.09344025701284409,
-0.01576313003897667,
-0.2000170350074768,
0.06719236820936203,
-0.15544287860393524,
-0.17475298047065735,
-0.032118938863277435,
0.0898858904838562,
-0.030757347121834755,
0.059436243027448654,
-0.0366884283721447,
-0.0147250359877944,
0.07591898739337921,
-0.018080251291394234,
-0.031030235812067986,
-0.11050038784742355,
0.09064504504203796,
-0.07718342542648315,
0.20288817584514618,
-0.039446692913770676,
0.05949924513697624,
0.13998736441135406,
0.021349523216485977,
-0.09129375964403152,
0.06781842559576035,
0.04701370373368263,
-0.09681126475334167,
0.014569233171641827,
0.1276503950357437,
-0.05378789082169533,
0.11019841581583023,
0.058035701513290405,
-0.14445295929908752,
0.02873656526207924,
-0.06893022358417511,
-0.04701998829841614,
-0.08289454877376556,
0.0021695848554372787,
-0.09109444171190262,
0.14244791865348816,
0.20186497271060944,
-0.020748961716890335,
0.04181148484349251,
-0.08961217105388641,
0.033944517374038696,
0.06382282078266144,
0.10389484465122223,
-0.054344866424798965,
-0.2436789721250534,
0.03318328782916069,
0.028864508494734764,
-0.012104236520826817,
-0.28862953186035156,
-0.07569059729576111,
0.033162277191877365,
-0.04852790758013725,
-0.06222965195775032,
0.10676663368940353,
0.12589581310749054,
0.06323264539241791,
-0.050808604806661606,
-0.18405333161354065,
-0.05805571377277374,
0.16764014959335327,
-0.1440134197473526,
-0.06827626377344131
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.8.2 | {"library_name": "peft", "base_model": "roberta-large"} | null | shahzebnaveed/roberta-large-lora-token-cls | [
"peft",
"tensorboard",
"safetensors",
"arxiv:1910.09700",
"base_model:roberta-large",
"region:us"
] | 2024-02-08T08:41:41+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #tensorboard #safetensors #arxiv-1910.09700 #base_model-roberta-large #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.8.2 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
"TAGS\n#peft #tensorboard #safetensors #arxiv-1910.09700 #base_model-roberta-large #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
37,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#peft #tensorboard #safetensors #arxiv-1910.09700 #base_model-roberta-large #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2"
] | [
-0.1039939597249031,
0.20482034981250763,
-0.0034118767362087965,
0.03314211219549179,
0.09255602210760117,
0.017728978767991066,
0.055051419883966446,
0.12758898735046387,
-0.027070242911577225,
0.11070758104324341,
0.06898312270641327,
0.09562660753726959,
0.10716049373149872,
0.21113936603069305,
0.007183713372796774,
-0.20159584283828735,
0.026557529345154762,
-0.09130296111106873,
-0.010939025320112705,
0.1233915463089943,
0.14426888525485992,
-0.09676888585090637,
0.0782037302851677,
-0.015273685567080975,
-0.010487910360097885,
-0.031065918505191803,
-0.07200358808040619,
-0.0317814014852047,
0.041117362678050995,
0.047700706869363785,
0.05705716460943222,
-0.0013586940476670861,
0.08218035101890564,
-0.2664763331413269,
0.0191944632679224,
0.04890412464737892,
-0.004754777532070875,
0.08701634407043457,
0.09768558293581009,
-0.04667680338025093,
0.12918208539485931,
-0.03213713318109512,
0.1402621865272522,
0.07761510461568832,
-0.0931490808725357,
-0.2240004986524582,
-0.06697029620409012,
0.08887406438589096,
0.17592036724090576,
0.07530181109905243,
-0.04356363043189049,
0.12921471893787384,
-0.08742193132638931,
0.018307141959667206,
0.04706219956278801,
-0.09384248405694962,
-0.07057934999465942,
0.05448449030518532,
0.10448870062828064,
0.05547814071178436,
-0.13582220673561096,
-0.030126621946692467,
0.02428106963634491,
0.03383446857333183,
0.07735473662614822,
0.011946592479944229,
0.15323525667190552,
0.02874756045639515,
-0.1458890736103058,
-0.04291249439120293,
0.1305362433195114,
0.025253605097532272,
-0.035667721182107925,
-0.2295127958059311,
0.005993305239826441,
-0.08000753074884415,
-0.02549234963953495,
-0.04929549619555473,
0.03772243484854698,
0.0034341770224273205,
0.09849443286657333,
-0.0317983515560627,
-0.09033366292715073,
-0.010323571041226387,
0.09173685312271118,
0.04967605695128441,
0.02435820735991001,
-0.019819622859358788,
0.004128992557525635,
0.12412872910499573,
0.058921944350004196,
-0.12992464005947113,
-0.059362273663282394,
-0.0712377205491066,
-0.04831165075302124,
-0.04048529639840126,
0.04071430489420891,
0.03800312057137489,
0.058357127010822296,
0.2523353099822998,
-0.034639839082956314,
0.05927376449108124,
0.06132042780518532,
0.021214110776782036,
0.047672025859355927,
0.09729872643947601,
-0.059774089604616165,
-0.15139824151992798,
-0.016173873096704483,
0.09825541824102402,
-0.005280292592942715,
-0.02411479689180851,
-0.05055597051978111,
0.04796237871050835,
0.03494029492139816,
0.10947620868682861,
0.09765053540468216,
-0.01082166749984026,
-0.07562774419784546,
-0.05263125151395798,
0.1979546993970871,
-0.15233959257602692,
0.04601076990365982,
0.021410973742604256,
-0.013613223098218441,
-0.04385784640908241,
0.012430544011294842,
0.015030297450721264,
-0.028121626004576683,
0.09881515800952911,
-0.06724000722169876,
-0.037554781883955,
-0.11857212334871292,
-0.025931749492883682,
0.032997436821460724,
0.002055194927379489,
-0.02977088652551174,
-0.029428211972117424,
-0.06104828417301178,
-0.09519565105438232,
0.10015860944986343,
-0.06408759951591492,
-0.05965584143996239,
-0.03174522519111633,
-0.09100048989057541,
0.020218178629875183,
0.027731800451874733,
0.09526670724153519,
-0.023869642987847328,
0.04314729571342468,
-0.014820885844528675,
0.06373517215251923,
0.08209636062383652,
0.03497742861509323,
-0.07223355770111084,
0.06388592720031738,
-0.19513443112373352,
0.08564312011003494,
-0.08044575154781342,
0.02818349003791809,
-0.15791216492652893,
-0.014650962315499783,
0.004529223311692476,
0.024441370740532875,
0.030698556452989578,
0.15692739188671112,
-0.20307590067386627,
-0.03546600416302681,
0.15974530577659607,
-0.10464348644018173,
-0.11720746010541916,
0.04006120190024376,
-0.04786383733153343,
0.15701431035995483,
0.02227981574833393,
-0.0006844537565484643,
0.09774049371480942,
-0.14531569182872772,
-0.02589975669980049,
-0.022356625646352768,
-0.003461116226390004,
0.1031140610575676,
0.08493250608444214,
-0.08512715995311737,
0.03500940650701523,
0.01416793279349804,
-0.050015002489089966,
-0.027018312364816666,
-0.04546566680073738,
-0.10814785212278366,
0.007004577200859785,
-0.0798206627368927,
0.02033853717148304,
-0.004668188281357288,
-0.08182264864444733,
-0.01083294115960598,
-0.16453401744365692,
-0.0316205695271492,
0.08024539798498154,
0.005779394414275885,
-0.01845717616379261,
-0.1001630574464798,
0.041727177798748016,
-0.027415096759796143,
-0.019740106537938118,
-0.15032276511192322,
-0.022418146952986717,
0.016468903049826622,
-0.13583646714687347,
0.009541790001094341,
-0.11479885876178741,
0.06794262677431107,
0.009936687536537647,
-0.06148706004023552,
-0.035320356488227844,
-0.00667656771838665,
0.007161977235227823,
-0.050260432064533234,
-0.24270065128803253,
-0.01983138918876648,
-0.05473934859037399,
0.1541450321674347,
-0.22327177226543427,
0.03920980915427208,
0.052612174302339554,
0.13322825729846954,
0.006170748732984066,
-0.06424521654844284,
0.029439732432365417,
-0.0649518147110939,
-0.025731608271598816,
-0.07492586970329285,
-0.005526699125766754,
-0.007523299660533667,
-0.047451794147491455,
0.01708679459989071,
-0.12315928936004639,
-0.038257572799921036,
0.10115406662225723,
0.07392662018537521,
-0.16040365397930145,
-0.01635352149605751,
-0.04315457120537758,
-0.06174421310424805,
-0.08267242461442947,
-0.05987825244665146,
0.11204449087381363,
0.0526580736041069,
0.03768560662865639,
-0.07380044460296631,
-0.07091013342142105,
0.008225513622164726,
-0.02248947136104107,
-0.023980088531970978,
0.11701278388500214,
0.07608165591955185,
-0.10650042444467545,
0.09944973140954971,
0.07481130212545395,
0.03356042504310608,
0.08573590219020844,
-0.028548991307616234,
-0.10645368695259094,
-0.02958151325583458,
0.04699639230966568,
0.01621144823729992,
0.16159573197364807,
-0.06415797770023346,
0.05146804824471474,
0.04319623485207558,
-0.04055047035217285,
0.045593295246362686,
-0.0978669598698616,
0.010930454358458519,
0.009372537955641747,
-0.013553254306316376,
0.021893955767154694,
-0.02185768634080887,
0.010736804455518723,
0.0834585577249527,
0.050389502197504044,
0.038431376218795776,
0.025185950100421906,
-0.0302962027490139,
-0.12921954691410065,
0.18379591405391693,
-0.0950385183095932,
-0.24330264329910278,
-0.15909188985824585,
0.06399400532245636,
0.05629001930356026,
-0.018863031640648842,
0.021449485793709755,
-0.055848971009254456,
-0.1039910688996315,
-0.08314487338066101,
0.0009867037879303098,
0.02981632575392723,
-0.05645839497447014,
-0.06912750005722046,
0.0474056601524353,
0.045094624161720276,
-0.11655616015195847,
0.036521606147289276,
0.05584081634879112,
-0.019766483455896378,
0.0010433883871883154,
0.05892007425427437,
0.08797211199998856,
0.18107245862483978,
-0.007445742841809988,
-0.0017772519495338202,
0.04729333147406578,
0.272234708070755,
-0.1614806205034256,
0.1136476993560791,
0.12155752629041672,
-0.06932947784662247,
0.07989393174648285,
0.18709337711334229,
0.033801041543483734,
-0.0985562801361084,
0.027648909017443657,
0.03213082253932953,
-0.029106590896844864,
-0.2619863450527191,
-0.04713023081421852,
-0.013368708081543446,
-0.08671943098306656,
0.08106856048107147,
0.0918947383761406,
0.07823474705219269,
0.0351269394159317,
-0.06758866459131241,
-0.08941832929849625,
0.034386683255434036,
0.0998150184750557,
-0.021676138043403625,
0.006013612728565931,
0.08409041166305542,
-0.03670215979218483,
0.008112802170217037,
0.09592808037996292,
-0.009813585318624973,
0.1675621122121811,
0.051696181297302246,
0.10953974723815918,
0.08155576139688492,
0.08687058836221695,
-0.0028196803759783506,
0.02467036247253418,
0.01111510768532753,
0.023668430745601654,
0.014641867950558662,
-0.08423621207475662,
0.030964475125074387,
0.1099734753370285,
0.04519037529826164,
0.029405910521745682,
0.01704998128116131,
-0.03944063559174538,
0.05305952951312065,
0.17628929018974304,
0.010041942819952965,
-0.19909672439098358,
-0.07543018460273743,
0.06298578530550003,
-0.07613501697778702,
-0.1343839466571808,
-0.01579299196600914,
0.036332547664642334,
-0.16584038734436035,
0.022364551201462746,
-0.04270295426249504,
0.10183750838041306,
-0.08820034563541412,
-0.03958583250641823,
0.0974060520529747,
0.065524160861969,
-0.024425197392702103,
0.05598469078540802,
-0.19207489490509033,
0.12415581196546555,
0.026499830186367035,
0.07288254052400589,
-0.08608163893222809,
0.10026583820581436,
0.003628777340054512,
-0.006039144936949015,
0.16846220195293427,
0.0003823137085419148,
-0.06178602576255798,
-0.06319334357976913,
-0.0963609591126442,
-0.013644658029079437,
0.0979241281747818,
-0.13534604012966156,
0.06860578060150146,
-0.019421104341745377,
-0.027493419125676155,
-0.003631018102169037,
-0.09162600338459015,
-0.11886495351791382,
-0.175649031996727,
0.0543985441327095,
-0.09907004237174988,
0.02249843068420887,
-0.09183035790920258,
-0.063141830265522,
0.0033307450357824564,
0.17849664390087128,
-0.21520663797855377,
-0.09882555156946182,
-0.14904548227787018,
-0.08466557413339615,
0.15800197422504425,
-0.04321957379579544,
0.07990661263465881,
0.0017901521641761065,
0.16497835516929626,
0.011175336316227913,
-0.011124396696686745,
0.09807322919368744,
-0.08616279810667038,
-0.1914513111114502,
-0.057141922414302826,
0.16776838898658752,
0.13357038795948029,
0.03625514358282089,
-0.0148813147097826,
0.028175493702292442,
-0.04913958162069321,
-0.1198287233710289,
0.027265852317214012,
0.14256969094276428,
0.06313374638557434,
-0.012913008220493793,
-0.02828885242342949,
-0.100216805934906,
-0.06276635825634003,
-0.0446605421602726,
-0.0008291262201964855,
0.18697500228881836,
-0.07759082317352295,
0.1644555628299713,
0.11065690964460373,
-0.059902433305978775,
-0.2096906155347824,
0.04670947790145874,
0.05135374888777733,
0.01660807989537716,
0.04116034507751465,
-0.19450245797634125,
0.0860455110669136,
-0.002217908389866352,
-0.07091590017080307,
0.1644822359085083,
-0.1731770783662796,
-0.14085149765014648,
0.09863976389169693,
0.034298304468393326,
-0.2259451448917389,
-0.1424044370651245,
-0.1025804877281189,
-0.016194600611925125,
-0.11327741295099258,
0.055482108145952225,
-0.002983665792271495,
0.012196945026516914,
0.026774991303682327,
0.014642015099525452,
0.02555624581873417,
-0.046337421983480453,
0.20304441452026367,
-0.029367750510573387,
0.010312098078429699,
-0.04899614676833153,
-0.08684443682432175,
0.03399556130170822,
-0.0475204773247242,
0.1010676696896553,
-0.008398856967687607,
0.023709341883659363,
-0.1556209772825241,
-0.04099954292178154,
-0.05325561761856079,
0.03053264319896698,
-0.09551405161619186,
-0.08558488637208939,
-0.04807034507393837,
0.09281411021947861,
0.09263274818658829,
-0.027687547728419304,
0.0036986321210861206,
-0.09008275717496872,
0.06735283136367798,
0.2019115388393402,
0.19102486968040466,
0.07113546133041382,
-0.07406491786241531,
0.021297398954629898,
-0.031156279146671295,
0.04500454664230347,
-0.2332124561071396,
0.040704671293497086,
0.05761473625898361,
0.021980352699756622,
0.08654788881540298,
-0.010149660520255566,
-0.1571722775697708,
-0.06902245432138443,
0.08406352996826172,
-0.05299151688814163,
-0.17032818496227264,
-0.025017540901899338,
0.030846940353512764,
-0.2054910659790039,
-0.04312801733613014,
0.024670235812664032,
-0.019411824643611908,
-0.038512468338012695,
0.026068849489092827,
0.07990427315235138,
-0.016962265595793724,
0.10601039975881577,
0.08437635749578476,
0.09237690269947052,
-0.10215945541858673,
0.07657496631145477,
0.07460839301347733,
-0.04947834089398384,
0.02969328872859478,
0.11414040625095367,
-0.04612085968255997,
-0.03313629329204559,
0.07970578968524933,
0.09605006873607635,
0.028439603745937347,
-0.05356919765472412,
0.007883585058152676,
-0.058963898569345474,
0.06148695945739746,
0.11273707449436188,
0.029363805428147316,
-0.0011685312492772937,
0.055598918348550797,
0.031876448541879654,
-0.09438694268465042,
0.11104190349578857,
0.059160176664590836,
0.017740115523338318,
-0.0461704395711422,
-0.03578837588429451,
-0.005217015277594328,
-0.01794848032295704,
-0.019196022301912308,
-0.006507891695946455,
-0.09391563385725021,
-0.007593000773340464,
-0.08800359070301056,
0.024024484679102898,
-0.07152515649795532,
0.007864032872021198,
0.03127516433596611,
-0.049787335097789764,
0.0032039829529821873,
0.006508716382086277,
-0.0720214694738388,
-0.05038514733314514,
-0.0137830451130867,
0.08333924412727356,
-0.1340407282114029,
0.037225354462862015,
0.07321266829967499,
-0.10382691025733948,
0.07127141952514648,
-0.007598377764225006,
0.0074599492363631725,
0.0063698189333081245,
-0.1579129844903946,
0.05809925124049187,
-0.022206928580999374,
-0.01401431579142809,
0.017690669745206833,
-0.20336568355560303,
-0.005882448051124811,
-0.045799415558576584,
-0.055012401193380356,
0.012262297794222832,
-0.024770617485046387,
-0.12343289703130722,
0.09944003820419312,
-0.0015682264929637313,
-0.06636209040880203,
-0.01804283633828163,
0.04130443558096886,
0.10236820578575134,
-0.021642541512846947,
0.13775981962680817,
-0.026926632970571518,
0.07325421273708344,
-0.17307370901107788,
-0.0067768581211566925,
-0.014590374194085598,
0.03966771811246872,
-0.0222787968814373,
-0.023348277434706688,
0.05829951912164688,
-0.02087186463177204,
0.1779889315366745,
-0.01488038245588541,
0.06797545403242111,
0.05835094302892685,
0.011379112489521503,
0.012316826730966568,
0.08672472834587097,
0.05740960314869881,
-0.0020359300542622805,
-0.0059504942037165165,
0.031629543751478195,
-0.009405864402651787,
-0.04240699112415314,
-0.1582862138748169,
0.06472272425889969,
0.15192431211471558,
0.051984984427690506,
0.022871699184179306,
0.030489720404148102,
-0.10854712128639221,
-0.07790955901145935,
0.13081155717372894,
-0.0144780483096838,
-0.03495129942893982,
-0.06898097693920135,
0.17290367186069489,
0.14184780418872833,
-0.19934237003326416,
0.0748419240117073,
-0.05267225205898285,
-0.0500386543571949,
-0.13222338259220123,
-0.1675230860710144,
-0.06170913949608803,
-0.044278912246227264,
-0.021152134984731674,
-0.06488154828548431,
0.05007345974445343,
0.052737053483724594,
0.004427257459610701,
-0.015014896169304848,
0.10652175545692444,
0.006881861016154289,
-0.02497684955596924,
0.051470186561346054,
0.06506705284118652,
0.03417570888996124,
-0.09260488301515579,
0.013115977868437767,
-0.0033531920053064823,
0.017256205901503563,
0.0648803859949112,
0.01932196132838726,
-0.04961413890123367,
0.017261432483792305,
-0.02295573800802231,
-0.11429179459810257,
0.04195922613143921,
-0.014996354468166828,
-0.042309317737817764,
0.1460072547197342,
0.03264705464243889,
0.006016178987920284,
-0.024505073204636574,
0.23076067864894867,
-0.07842391729354858,
-0.06913387030363083,
-0.1504673957824707,
0.07234090566635132,
-0.06374874711036682,
0.03367502987384796,
0.028418606147170067,
-0.11887001246213913,
0.01652238517999649,
0.16324663162231445,
0.13158349692821503,
-0.009903703816235065,
0.011766190640628338,
0.04205421358346939,
0.0064820293337106705,
-0.03080022521317005,
0.017369039356708527,
0.049773555248975754,
0.1484447568655014,
-0.07000596821308136,
0.06211014837026596,
-0.011369502171874046,
-0.08143796771764755,
-0.02095530554652214,
0.10505586862564087,
-0.00027203213539905846,
0.0011871816823258996,
-0.07227636128664017,
0.14152862131595612,
-0.081522136926651,
-0.22550523281097412,
0.06595391035079956,
-0.07465586066246033,
-0.14673669636249542,
-0.04455196484923363,
0.028107941150665283,
-0.018711229786276817,
0.011846153996884823,
0.07712863385677338,
-0.050490763038396835,
0.17137257754802704,
0.04417232424020767,
-0.05621631443500519,
-0.08096278458833694,
0.05482756718993187,
-0.14783340692520142,
0.27974745631217957,
0.019492778927087784,
0.03967148810625076,
0.1040702536702156,
-0.01760082319378853,
-0.140618696808815,
0.008547311648726463,
0.10527709871530533,
-0.06808476150035858,
0.05705593526363373,
0.17376472055912018,
0.0027864433359354734,
0.12759272754192352,
0.05764014273881912,
-0.05279562249779701,
0.036714375019073486,
-0.10158709436655045,
-0.0494898222386837,
-0.1062898337841034,
0.08118882775306702,
-0.08451215922832489,
0.1607438027858734,
0.12619709968566895,
-0.06539805233478546,
-0.004563852678984404,
-0.019567089155316353,
0.0803888812661171,
0.00891472864896059,
0.11418142914772034,
0.013508426956832409,
-0.18090596795082092,
0.0329541452229023,
0.0060462686233222485,
0.1007387638092041,
-0.20629453659057617,
-0.06136501207947731,
0.044125135987997055,
-0.01776704378426075,
-0.07919507473707199,
0.12217075377702713,
0.04517076909542084,
0.035796068608760834,
-0.0410739965736866,
-0.05520506575703621,
0.010621641762554646,
0.14340871572494507,
-0.11024732142686844,
-0.005277007352560759
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# dit-base-finetuned-rvlcdip-finetuned-custom-first
This model is a fine-tuned version of [microsoft/dit-base-finetuned-rvlcdip](https://huggingface.co/microsoft/dit-base-finetuned-rvlcdip) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0567
- Accuracy: 0.9949
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 32
- eval_batch_size: 32
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 128
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.3686 | 1.0 | 79 | 0.2356 | 0.9746 |
| 0.0891 | 2.0 | 158 | 0.0792 | 0.9936 |
| 0.0652 | 3.0 | 237 | 0.0567 | 0.9949 |
### Framework versions
- Transformers 4.38.0.dev0
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "microsoft/dit-base-finetuned-rvlcdip", "model-index": [{"name": "dit-base-finetuned-rvlcdip-finetuned-custom-first", "results": []}]} | image-classification | stray-light/dit-base-finetuned-rvlcdip-finetuned-custom-first | [
"transformers",
"tensorboard",
"safetensors",
"beit",
"image-classification",
"generated_from_trainer",
"base_model:microsoft/dit-base-finetuned-rvlcdip",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T08:43:48+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #beit #image-classification #generated_from_trainer #base_model-microsoft/dit-base-finetuned-rvlcdip #autotrain_compatible #endpoints_compatible #region-us
| dit-base-finetuned-rvlcdip-finetuned-custom-first
=================================================
This model is a fine-tuned version of microsoft/dit-base-finetuned-rvlcdip on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0567
* Accuracy: 0.9949
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 32
* eval\_batch\_size: 32
* seed: 42
* gradient\_accumulation\_steps: 4
* total\_train\_batch\_size: 128
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.38.0.dev0
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #beit #image-classification #generated_from_trainer #base_model-microsoft/dit-base-finetuned-rvlcdip #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
68,
144,
4,
38
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #beit #image-classification #generated_from_trainer #base_model-microsoft/dit-base-finetuned-rvlcdip #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.14705434441566467,
0.08937692642211914,
-0.0010730313370004296,
0.0807885080575943,
0.15600307285785675,
0.020786412060260773,
0.11399022489786148,
0.12067710608243942,
-0.1314442902803421,
0.08374034613370895,
0.11311988532543182,
0.07884129881858826,
0.025648612529039383,
0.12451182305812836,
-0.03653258457779884,
-0.29706814885139465,
0.009279430843889713,
0.01568175107240677,
-0.17174705862998962,
0.11268578469753265,
0.0976119264960289,
-0.13073845207691193,
0.07778306305408478,
0.02275685966014862,
-0.198835089802742,
-0.004272801801562309,
-0.004471168853342533,
-0.06046237051486969,
0.12335986644029617,
0.0413169264793396,
0.13314761221408844,
0.022965963929891586,
0.12283599376678467,
-0.17228485643863678,
0.006364481523633003,
0.08021936565637589,
0.014171754941344261,
0.09921788424253464,
0.08370719850063324,
0.0033202068880200386,
0.11629591882228851,
-0.07545149326324463,
0.08135479688644409,
0.04317481443285942,
-0.11074996739625931,
-0.28493618965148926,
-0.09701182693243027,
0.06890900433063507,
0.12591904401779175,
0.07497544586658478,
-0.011595257557928562,
0.1107342317700386,
-0.0628209188580513,
0.0916205421090126,
0.2208699733018875,
-0.26221945881843567,
-0.1010948121547699,
0.039159782230854034,
0.0468328520655632,
0.020906001329421997,
-0.131456196308136,
-0.0091648418456316,
0.0659058466553688,
0.03094477578997612,
0.09684834629297256,
0.012187096290290356,
0.008473954163491726,
0.0010423039784654975,
-0.14106327295303345,
-0.028824487701058388,
0.11830006539821625,
0.07395314425230026,
-0.047220513224601746,
-0.059231288731098175,
-0.0534086674451828,
-0.19922374188899994,
-0.04069865494966507,
-0.0014344730880111456,
0.03089088201522827,
-0.07353690266609192,
-0.1387072652578354,
0.02056410349905491,
-0.09275209903717041,
-0.10843180119991302,
0.013921011239290237,
0.19070592522621155,
0.0509563609957695,
0.0017399182543158531,
-0.010807477869093418,
0.13462235033512115,
0.0419621542096138,
-0.1645711064338684,
-0.004810033366084099,
0.013877608813345432,
-0.06831131875514984,
-0.030735135078430176,
-0.047613710165023804,
-0.0320897102355957,
-0.006377944722771645,
0.17481273412704468,
-0.06828898191452026,
0.0501883365213871,
0.03790822997689247,
0.01974642276763916,
-0.09701791405677795,
0.19158749282360077,
-0.08238119632005692,
-0.040970902889966965,
-0.010530821047723293,
0.10365260392427444,
0.02018417790532112,
0.0013279286213219166,
-0.08802428096532822,
0.015243370085954666,
0.10357269644737244,
0.0373227633535862,
-0.034623436629772186,
0.04341237246990204,
-0.0352388434112072,
-0.019157379865646362,
0.08587633818387985,
-0.07796543091535568,
0.03443662077188492,
0.008207212202250957,
-0.09749506413936615,
-0.031111789867281914,
0.03514401614665985,
-0.0030838986858725548,
0.03205757588148117,
0.15203632414340973,
-0.10299045592546463,
-0.00663255900144577,
-0.11178369075059891,
-0.11646601557731628,
0.015078103169798851,
-0.037501633167266846,
0.0027926224283874035,
-0.09752139449119568,
-0.1320207566022873,
-0.031384050846099854,
0.038733430206775665,
-0.0513777919113636,
-0.03698939457535744,
-0.027967175468802452,
-0.11064604669809341,
0.03956269100308418,
0.0012953062541782856,
0.13418911397457123,
-0.04635363072156906,
0.11266475170850754,
0.08181840926408768,
0.05955911800265312,
0.008597130887210369,
0.03170302137732506,
-0.06933777034282684,
0.050630487501621246,
-0.20479369163513184,
0.03755636885762215,
-0.07539798319339752,
0.07772066444158554,
-0.09703044593334198,
-0.12983573973178864,
0.012777788564562798,
-0.003024632576853037,
0.09569501131772995,
0.11951605975627899,
-0.14220744371414185,
-0.08177801221609116,
0.14389310777187347,
-0.09254742413759232,
-0.13672234117984772,
0.10546191036701202,
-0.0230263564735651,
-0.008606445044279099,
0.028768042102456093,
0.08076053112745285,
0.07822387665510178,
-0.10071637481451035,
-0.033411938697099686,
-0.04498063027858734,
0.10884551703929901,
-0.0023731100372970104,
0.08319856226444244,
0.0016755505930632353,
0.037831079214811325,
0.011682162992656231,
-0.06853802502155304,
0.05032177269458771,
-0.11577242612838745,
-0.07899085432291031,
-0.040826987475156784,
-0.09557640552520752,
0.05081046000123024,
0.07028903067111969,
0.06742119789123535,
-0.09798838198184967,
-0.13146668672561646,
0.08525191247463226,
0.11851119995117188,
-0.07187581062316895,
0.009682674892246723,
-0.07223591208457947,
0.07182788848876953,
-0.08261606842279434,
-0.015890436246991158,
-0.1777482032775879,
-0.09425119310617447,
0.0124399084597826,
-0.009243004024028778,
-0.009517415426671505,
-0.020616615191102028,
0.08836379647254944,
0.10090069472789764,
-0.08634745329618454,
-0.0734521746635437,
-0.09381275624036789,
-0.009122940711677074,
-0.09553061425685883,
-0.22717677056789398,
-0.09841427206993103,
-0.02777756005525589,
0.14885394275188446,
-0.24351437389850616,
0.0350567102432251,
0.02668284997344017,
0.15561912953853607,
0.05782356485724449,
-0.031601328402757645,
-0.027591422200202942,
0.06144152581691742,
-0.04699760675430298,
-0.08238577842712402,
0.016571613028645515,
-0.0011891971807926893,
-0.09918805956840515,
-0.04555031657218933,
-0.11201979219913483,
0.1507948935031891,
0.1204163134098053,
-0.03706635907292366,
-0.11872615665197372,
0.006071871612221003,
-0.0912400633096695,
-0.04718702659010887,
-0.04160728305578232,
-0.018549209460616112,
0.07372654229402542,
0.027791831642389297,
0.14031919836997986,
-0.06923195719718933,
-0.054775308817625046,
0.03773587569594383,
-0.0025845549535006285,
-0.002259833738207817,
0.09742122888565063,
0.08191128820180893,
-0.052554961293935776,
0.12454758584499359,
0.11798959970474243,
-0.10348866134881973,
0.12441791594028473,
-0.05821232870221138,
-0.09293527156114578,
-0.017013253644108772,
0.004274397157132626,
0.04029322415590286,
0.16610106825828552,
-0.06670023500919342,
-0.01113931555300951,
0.01206415519118309,
0.006146637257188559,
0.01136066298931837,
-0.2127566635608673,
-0.009151879698038101,
0.03521529212594032,
-0.031652502715587616,
0.001155784702859819,
-0.016299284994602203,
-0.004161317832767963,
0.09818652272224426,
0.007708395831286907,
-0.04057593271136284,
0.00993146002292633,
0.0048009417951107025,
-0.06309521943330765,
0.21693642437458038,
-0.0745423436164856,
-0.10883227735757828,
-0.15203365683555603,
0.007780421059578657,
-0.06396202743053436,
0.01044909656047821,
0.023733921349048615,
-0.11444664001464844,
-0.040916550904512405,
-0.04626970365643501,
0.05314831808209419,
-0.005341054871678352,
0.04578861594200134,
-0.007156602572649717,
0.02841385453939438,
0.09941117465496063,
-0.10913539677858353,
0.03594665601849556,
-0.0227950531989336,
-0.06831154972314835,
0.02489641308784485,
0.04534456878900528,
0.12483304738998413,
0.13989347219467163,
-0.00609328830614686,
0.006291811820119619,
-0.022168049588799477,
0.19169017672538757,
-0.10118070989847183,
-0.01984061487019062,
0.12762907147407532,
0.0025167185813188553,
0.037859413772821426,
0.12044785171747208,
0.06702342629432678,
-0.096432164311409,
0.028212491422891617,
0.054577335715293884,
-0.019866006448864937,
-0.2075629085302353,
-0.01840631663799286,
-0.03232349455356598,
-0.0050011686980724335,
0.11065874993801117,
0.04028775170445442,
0.01870296150445938,
0.06901171803474426,
-0.02661013789474964,
0.05904626473784447,
-0.04734823852777481,
0.09140995144844055,
0.05795775353908539,
0.05350066348910332,
0.1374158263206482,
-0.03654363378882408,
-0.043395571410655975,
0.025910165160894394,
-0.01714300736784935,
0.2086234837770462,
-0.03644225373864174,
0.11356882005929947,
0.03494081273674965,
0.1619916409254074,
-0.0006312245968729258,
0.08138150721788406,
0.01931760646402836,
-0.05068901553750038,
0.022659841924905777,
-0.05770006775856018,
-0.020882023498415947,
0.04501883685588837,
-0.0070900265127420425,
0.09324855357408524,
-0.13438040018081665,
0.017003824934363365,
0.05112433061003685,
0.3011508882045746,
0.07342565804719925,
-0.35599201917648315,
-0.13596057891845703,
0.0014648035867139697,
-0.036631688475608826,
-0.04734962061047554,
0.004437162075191736,
0.13412250578403473,
-0.08967802673578262,
0.062160030007362366,
-0.0945906713604927,
0.08598148077726364,
-0.030261699110269547,
0.004936414770781994,
0.10693187266588211,
0.1047532930970192,
-0.013333620503544807,
0.05115888640284538,
-0.23132333159446716,
0.27602484822273254,
0.009466408751904964,
0.08648315072059631,
-0.02914116345345974,
0.03201530873775482,
0.04317391663789749,
0.06899914145469666,
0.051827285438776016,
-0.032697029411792755,
-0.07406638562679291,
-0.20964102447032928,
-0.06580674648284912,
0.01823333650827408,
0.13007864356040955,
-0.07640432566404343,
0.13910330832004547,
-0.024982962757349014,
-0.03319987654685974,
0.06322558224201202,
-0.05920696631073952,
-0.09725122153759003,
-0.07129363715648651,
0.004591645207256079,
-0.026728792116045952,
0.04536284878849983,
-0.111794613301754,
-0.12373141944408417,
-0.05602353438735008,
0.16891011595726013,
-0.08181123435497284,
-0.030534550547599792,
-0.14944058656692505,
0.12624816596508026,
0.14667782187461853,
-0.06796351820230484,
0.06151358410716057,
-0.0003248462453484535,
0.1419028341770172,
0.026435207575559616,
-0.027168236672878265,
0.10807111859321594,
-0.09230471402406693,
-0.25661730766296387,
-0.06268687546253204,
0.12254950404167175,
0.019967127591371536,
0.04930145666003227,
-0.028369951993227005,
0.040147699415683746,
-0.006204892415553331,
-0.08656011521816254,
0.03790632262825966,
-0.02484777383506298,
0.053303610533475876,
0.030875522643327713,
-0.041707806289196014,
0.025108577683568,
-0.03331815078854561,
-0.028569936752319336,
0.08787978440523148,
0.29851165413856506,
-0.10051726549863815,
-0.05547752231359482,
0.033644575625658035,
-0.027458520606160164,
-0.1691930890083313,
0.07951341569423676,
0.11039523780345917,
0.03354202210903168,
0.009390028193593025,
-0.17213791608810425,
0.10195019841194153,
0.1007142886519432,
-0.042673610150814056,
0.13670757412910461,
-0.2597953677177429,
-0.14540459215641022,
0.08669950813055038,
0.1405271291732788,
-0.028364455327391624,
-0.17748627066612244,
-0.051492877304553986,
-0.033434249460697174,
-0.1185353547334671,
0.08301594108343124,
-0.05592232942581177,
0.10241398960351944,
-0.015149782411754131,
0.046969641000032425,
0.017333602532744408,
-0.0596013143658638,
0.15244948863983154,
-0.023053234443068504,
0.1009129136800766,
-0.018232164904475212,
0.01240207627415657,
0.0724259540438652,
-0.0680900290608406,
0.006589498836547136,
-0.04871120676398277,
0.045302413403987885,
-0.08191905170679092,
-0.007567527238279581,
-0.1000719964504242,
0.03809270262718201,
-0.059377774596214294,
-0.05696842446923256,
-0.03425217047333717,
0.04878385737538338,
0.001784912426955998,
-0.02389848604798317,
0.17419371008872986,
0.006485839374363422,
0.18182262778282166,
0.09925572574138641,
0.056200060993433,
0.002661264967173338,
-0.058748625218868256,
0.007728843949735165,
-0.019627923145890236,
0.0839635580778122,
-0.16676440834999084,
0.01513521559536457,
0.12832966446876526,
0.05715301260352135,
0.12021464854478836,
0.07288480550050735,
-0.05259621888399124,
0.017022883519530296,
0.0996314063668251,
-0.11485493928194046,
-0.0856470838189125,
-0.017261026427149773,
0.0013630210887640715,
-0.1593123823404312,
0.07314642518758774,
0.10979631543159485,
-0.08894753456115723,
-0.003840506076812744,
0.007403126917779446,
-0.0009615360177122056,
-0.03308793902397156,
0.23218627274036407,
0.08945466578006744,
0.08251930773258209,
-0.08893081545829773,
0.07962174713611603,
0.03430144861340523,
-0.12210680544376373,
-0.008032741025090218,
0.06993850320577621,
-0.05944207310676575,
-0.010059203021228313,
0.030305467545986176,
0.09204369783401489,
-0.06678961962461472,
-0.04648309201002121,
-0.16956883668899536,
-0.11894921213388443,
0.04981080815196037,
0.13929784297943115,
0.07325994223356247,
0.03276936337351799,
-0.01104638259857893,
0.06295698136091232,
-0.14281567931175232,
0.12117858976125717,
0.0634976327419281,
0.11907697468996048,
-0.17439204454421997,
0.16596287488937378,
0.0050487336702644825,
0.02170010656118393,
-0.013911259360611439,
0.039130307734012604,
-0.11299333721399307,
-0.002932037226855755,
-0.14602534472942352,
-0.045411545783281326,
-0.033811286091804504,
-0.009178628213703632,
0.0024655719753354788,
-0.04899007827043533,
-0.06446704268455505,
0.022079117596149445,
-0.11377741396427155,
-0.052994921803474426,
0.018942950293421745,
0.023660309612751007,
-0.12245099991559982,
-0.008944259025156498,
0.035831037908792496,
-0.1181267574429512,
0.0777805894613266,
0.03250182047486305,
0.05385298654437065,
0.0489422082901001,
-0.0744706392288208,
0.008881422691047192,
0.057167474180459976,
-0.022327475249767303,
0.056975796818733215,
-0.0930384024977684,
-0.006464164238423109,
-0.047330304980278015,
0.05284123495221138,
-0.005857058800756931,
0.04432675242424011,
-0.14262866973876953,
-0.01154558826237917,
-0.029171409085392952,
-0.04197791963815689,
-0.05721081793308258,
0.042513877153396606,
0.06710673868656158,
0.021075118333101273,
0.1674746870994568,
-0.08470245450735092,
0.02330365590751171,
-0.23263046145439148,
-0.007375144865363836,
-0.025232430547475815,
-0.09756217896938324,
-0.10178637504577637,
-0.011908984743058681,
0.08210695534944534,
-0.06601128727197647,
0.06986937671899796,
-0.043289102613925934,
0.11131415516138077,
0.0492696650326252,
-0.041976410895586014,
0.02357761561870575,
0.05316740646958351,
0.18594035506248474,
0.033971358090639114,
-0.013331030495464802,
0.04958134517073631,
0.04805932939052582,
0.08418187499046326,
0.07198891788721085,
0.18151667714118958,
0.1121639534831047,
-0.03791102021932602,
0.11765439063310623,
0.06917817145586014,
-0.09539640694856644,
-0.147441565990448,
0.049661099910736084,
-0.0487530492246151,
0.11749300360679626,
-0.023402240127325058,
0.1379251778125763,
0.11891836673021317,
-0.1753424108028412,
0.009917769581079483,
-0.04209938272833824,
-0.07693721354007721,
-0.08899522572755814,
0.025014445185661316,
-0.07507406175136566,
-0.17836564779281616,
0.023029664531350136,
-0.11835870891809464,
0.0018627080135047436,
0.07878611981868744,
0.020988550037145615,
0.00827941857278347,
0.19702216982841492,
0.04592972993850708,
0.033834222704172134,
0.07769370079040527,
0.031998466700315475,
-0.011409517377614975,
-0.020996855571866035,
-0.08674319088459015,
0.012322689406573772,
-0.02646806463599205,
0.03647628426551819,
-0.07206839323043823,
-0.11138646304607391,
0.07220184057950974,
0.04920177906751633,
-0.10966673493385315,
0.0305568128824234,
0.02201632782816887,
0.06426362693309784,
0.05406564101576805,
-0.000686255341861397,
0.02922113984823227,
-0.03703409805893898,
0.2418672889471054,
-0.12022895365953445,
-0.0336117260158062,
-0.14328926801681519,
0.26146456599235535,
0.013366137631237507,
-0.01752791553735733,
0.02206483855843544,
-0.1106039360165596,
-0.017873967066407204,
0.13882246613502502,
0.14322376251220703,
-0.037534989416599274,
-0.02327977493405342,
0.020647414028644562,
-0.02364156022667885,
-0.07339213788509369,
0.08142311871051788,
0.0958797037601471,
0.07253800332546234,
-0.07599014043807983,
-0.05439934507012367,
-0.030641889199614525,
-0.05705489218235016,
-0.0014685061760246754,
0.05300144478678703,
0.01324956864118576,
0.0001531735761091113,
-0.05176512151956558,
0.07781486213207245,
-0.02168780378997326,
-0.14894969761371613,
0.11188733577728271,
-0.18345753848552704,
-0.18396654725074768,
-0.025424974039196968,
0.06472699344158173,
0.004927916917949915,
0.07713094353675842,
-0.010067468509078026,
-0.030870670452713966,
0.10901051759719849,
-0.01931731402873993,
-0.040687862783670425,
-0.1395265907049179,
0.08941851556301117,
-0.05703199282288551,
0.2206566035747528,
-0.05374085530638695,
0.01975485123693943,
0.12811799347400665,
0.030733313411474228,
-0.10780160129070282,
0.020278112962841988,
0.07561253011226654,
-0.13757453858852386,
0.031163301318883896,
0.18426759541034698,
-0.0388469398021698,
0.11371023952960968,
0.023861616849899292,
-0.1643378585577011,
0.010476162657141685,
-0.07054296880960464,
-0.060412533581256866,
-0.07911978662014008,
-0.013514254242181778,
-0.04033529758453369,
0.1301574409008026,
0.2439529299736023,
-0.049767691642045975,
-0.0048882910050451756,
-0.058341652154922485,
0.03767462447285652,
0.07939258962869644,
0.09219798445701599,
-0.02470843493938446,
-0.28349870443344116,
0.03503130376338959,
0.04247564077377319,
-0.021676234900951385,
-0.25501108169555664,
-0.08912108838558197,
0.051148273050785065,
-0.06164507567882538,
-0.06332027167081833,
0.10270777344703674,
0.0751962810754776,
0.061931174248456955,
-0.05311755836009979,
-0.07213782519102097,
-0.07539527118206024,
0.16579732298851013,
-0.1767130196094513,
-0.09318113327026367
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | inStryde/mask2former-swin-large-ade-semantic-instryde-foot | [
"transformers",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-08T08:46:55+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
26,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.08389580249786377,
0.19830818474292755,
-0.0013316317927092314,
0.02313883788883686,
0.11396584659814835,
0.01961737498641014,
0.053626976907253265,
0.14538456499576569,
0.0060051376931369305,
0.10656800121068954,
0.066679947078228,
0.09131570905447006,
0.09678101539611816,
0.20042605698108673,
0.04371999576687813,
-0.17659740149974823,
0.010636410675942898,
-0.06930278241634369,
-0.010073255747556686,
0.11651819199323654,
0.141214057803154,
-0.10151198506355286,
0.07627976685762405,
-0.03319970890879631,
-0.02870541252195835,
-0.0070160143077373505,
-0.07769215852022171,
-0.05755697935819626,
0.07573003321886063,
0.054863471537828445,
0.04207949340343475,
-0.0008347301045432687,
0.08447454124689102,
-0.2674994468688965,
0.013753628358244896,
0.07452993094921112,
0.010659529827535152,
0.05990942195057869,
0.07833302766084671,
-0.04036625102162361,
0.12881849706172943,
-0.06320446729660034,
0.13035163283348083,
0.0906217098236084,
-0.0681561604142189,
-0.24378153681755066,
-0.08239314705133438,
0.06505522131919861,
0.12533815205097198,
0.07694927603006363,
-0.02823091857135296,
0.16422191262245178,
-0.07247646898031235,
0.019290022552013397,
0.09481704235076904,
-0.1151006743311882,
-0.060644298791885376,
0.08318385481834412,
0.14101974666118622,
0.10340547561645508,
-0.1255619376897812,
-0.012289565056562424,
0.04275871813297272,
0.045979104936122894,
0.07389909774065018,
0.011339850723743439,
0.1143413558602333,
0.05629947781562805,
-0.13526225090026855,
-0.05700986459851265,
0.14547574520111084,
0.023872992023825645,
-0.057064127177000046,
-0.2138909548521042,
-0.002902575535699725,
-0.07730814069509506,
-0.011685127392411232,
-0.06846728920936584,
0.0291305985301733,
-0.01194276288151741,
0.060226380825042725,
-0.0496203787624836,
-0.09797755628824234,
-0.046314824372529984,
0.1015089675784111,
0.054820988327264786,
0.011354796588420868,
-0.01489334274083376,
0.03576440364122391,
0.13432876765727997,
0.04213530570268631,
-0.10012737661600113,
-0.07065672427415848,
-0.0701170489192009,
-0.09620913118124008,
-0.03947552293539047,
0.04272124543786049,
0.020167991518974304,
0.042202774435281754,
0.2283228635787964,
0.024096308276057243,
0.05459817871451378,
0.029667891561985016,
0.0026177873369306326,
0.03211980313062668,
0.1073630079627037,
-0.041210614144802094,
-0.188126802444458,
-0.03292805701494217,
0.0931866466999054,
-0.009821015410125256,
-0.028658604249358177,
-0.033444397151470184,
0.035014089196920395,
0.08379437029361725,
0.11821532249450684,
0.08875755965709686,
-0.012828069739043713,
-0.037612639367580414,
-0.03493109717965126,
0.2115669697523117,
-0.14141373336315155,
0.045799970626831055,
-0.022097334265708923,
-0.018195297569036484,
-0.06905751675367355,
0.030103791505098343,
0.01831657998263836,
-0.003142025787383318,
0.06966056674718857,
-0.061253178864717484,
-0.05794486775994301,
-0.11518853157758713,
-0.045523155480623245,
0.04711875319480896,
-0.024105608463287354,
-0.024469668045639992,
-0.07765042781829834,
-0.11219723522663116,
-0.06417357176542282,
0.06612563133239746,
-0.04156653955578804,
-0.03974827378988266,
0.005308232270181179,
-0.07131324708461761,
0.008387917652726173,
0.008993842639029026,
0.12122467905282974,
-0.030063031241297722,
0.05833350867033005,
-0.002476902212947607,
0.05916252359747887,
0.10643328726291656,
0.03227818012237549,
-0.08492200076580048,
0.057466037571430206,
-0.20633617043495178,
0.08371785283088684,
-0.11420095711946487,
0.034276340156793594,
-0.17048145830631256,
-0.024183684960007668,
0.008447963744401932,
0.023597201332449913,
0.023726604878902435,
0.1338067352771759,
-0.2097422182559967,
-0.016196569427847862,
0.14133213460445404,
-0.09649793803691864,
-0.12422871589660645,
0.07990546524524689,
-0.03459475561976433,
0.1747698187828064,
0.038475677371025085,
-0.019652999937534332,
0.09909367561340332,
-0.15559963881969452,
-0.05852397903800011,
-0.026064254343509674,
-0.008927824907004833,
0.08823978155851364,
0.07542291283607483,
-0.05844951793551445,
0.02285866066813469,
0.02562655322253704,
-0.04727208614349365,
-0.0268824752420187,
-0.05256075784564018,
-0.10127434879541397,
-0.023140445351600647,
-0.09642518311738968,
0.026515161618590355,
0.000058677000197349116,
-0.07310442626476288,
-0.028560271486639977,
-0.17347893118858337,
-0.02563360333442688,
0.10103316605091095,
0.004820956848561764,
-0.007559072691947222,
-0.08540112525224686,
0.022149885073304176,
-0.05362366884946823,
-0.006164622958749533,
-0.16996455192565918,
-0.03558015450835228,
0.051895126700401306,
-0.14917676150798798,
0.015460150316357613,
-0.07327745854854584,
0.07047311216592789,
0.02098717913031578,
-0.05859505757689476,
-0.03108096309006214,
0.0007694467785768211,
0.004292082041501999,
-0.06229274719953537,
-0.1903683841228485,
-0.058886781334877014,
-0.041500482708215714,
0.15720732510089874,
-0.24841000139713287,
0.0300158578902483,
0.03247617185115814,
0.13185922801494598,
0.007058668415993452,
-0.06344027817249298,
0.02096918225288391,
-0.04676475748419762,
-0.050621338188648224,
-0.06898977607488632,
-0.009901339188218117,
-0.014539826661348343,
-0.031393732875585556,
0.012980648316442966,
-0.14970256388187408,
-0.060514215379953384,
0.09452559798955917,
0.11224991828203201,
-0.14555825293064117,
0.00204002158716321,
-0.0460561066865921,
-0.07002599537372589,
-0.07487804442644119,
-0.0761631652712822,
0.07739497721195221,
0.044650159776210785,
0.049250341951847076,
-0.06317461282014847,
-0.06234706938266754,
0.023210179060697556,
0.005524294450879097,
-0.019023682922124863,
0.0948529988527298,
0.074309803545475,
-0.09122881293296814,
0.07973480224609375,
0.08461450785398483,
0.04414684325456619,
0.086973637342453,
0.005991141777485609,
-0.11396963149309158,
-0.03062884695827961,
0.037754856050014496,
0.024159027263522148,
0.15351562201976776,
-0.08692087233066559,
0.030462130904197693,
0.052177220582962036,
-0.03854219615459442,
0.03157065063714981,
-0.0923321321606636,
0.025362705811858177,
0.021495236083865166,
-0.006555700208991766,
0.05864228308200836,
-0.018769768998026848,
-0.01403577346354723,
0.06336429715156555,
0.05677810311317444,
0.044270504266023636,
0.02595379762351513,
-0.02093072421848774,
-0.1278371512889862,
0.16537296772003174,
-0.09028079360723495,
-0.2540280222892761,
-0.17074446380138397,
0.015454737469553947,
0.03706491366028786,
-0.021728800609707832,
0.039588842540979385,
-0.06286025792360306,
-0.10237989574670792,
-0.09417891502380371,
0.0029635571409016848,
0.023925531655550003,
-0.058347854763269424,
-0.0817074254155159,
0.060779985040426254,
0.04047083482146263,
-0.13689260184764862,
0.0349188968539238,
0.06170675903558731,
-0.03042641654610634,
0.0018567070364952087,
0.07321398705244064,
0.12743599712848663,
0.14838241040706635,
-0.006730219814926386,
-0.012446845881640911,
0.035035960376262665,
0.229813352227211,
-0.1490442156791687,
0.10630457103252411,
0.14053207635879517,
-0.021705523133277893,
0.06635113060474396,
0.1461038440465927,
0.023231739178299904,
-0.07546708732843399,
0.04147516191005707,
0.04027445614337921,
-0.04228919371962547,
-0.2589097023010254,
-0.05694316700100899,
-0.00946022942662239,
-0.07043391466140747,
0.09718906134366989,
0.09238530695438385,
0.11972260475158691,
0.0337289460003376,
-0.05568677559494972,
-0.025771914049983025,
-0.003401360474526882,
0.114128477871418,
-0.027640055865049362,
-0.004564122296869755,
0.07965842634439468,
-0.05878787487745285,
0.011684526689350605,
0.09941446036100388,
0.019347423687577248,
0.17601320147514343,
0.02533329278230667,
0.10681075602769852,
0.06725578010082245,
0.09347675740718842,
-0.0015635732561349869,
0.034774236381053925,
0.05337131395936012,
0.022044572979211807,
0.010453542694449425,
-0.09408048540353775,
-0.012431944720447063,
0.13713060319423676,
0.019816776737570763,
0.009031654335558414,
0.008926562033593655,
-0.01010479498654604,
0.03131420537829399,
0.20501568913459778,
0.0009575071162544191,
-0.22537250816822052,
-0.09500737488269806,
0.059459153562784195,
-0.06931101530790329,
-0.143676295876503,
-0.02094252221286297,
0.030270220711827278,
-0.17292405664920807,
0.016790566965937614,
-0.0316389761865139,
0.09112390875816345,
-0.07145322859287262,
-0.028050832450389862,
0.06891903281211853,
0.07569212466478348,
-0.012108199298381805,
0.07973295450210571,
-0.19069278240203857,
0.12254468351602554,
0.03037673607468605,
0.08605273067951202,
-0.11708726733922958,
0.07849059253931046,
-0.0019813794642686844,
-0.014807495288550854,
0.17999744415283203,
-0.014062200672924519,
-0.0586031936109066,
-0.08878950774669647,
-0.08704045414924622,
-0.011727320961654186,
0.10361312329769135,
-0.09322915226221085,
0.09586969763040543,
-0.02775636687874794,
-0.03705112263560295,
0.012418309226632118,
-0.10469507426023483,
-0.1636953055858612,
-0.18679304420948029,
0.06244563311338425,
-0.07802703976631165,
0.012347841635346413,
-0.11227322369813919,
-0.06334327906370163,
-0.01575082167983055,
0.23160123825073242,
-0.16648635268211365,
-0.07049825042486191,
-0.1498587429523468,
-0.03997112438082695,
0.17463743686676025,
-0.042160745710134506,
0.06849376112222672,
-0.021383514627814293,
0.1873992383480072,
-0.008081548847258091,
-0.013158116489648819,
0.06569221615791321,
-0.09637628495693207,
-0.16879262030124664,
-0.05748843029141426,
0.14160962402820587,
0.10863390564918518,
0.05731578543782234,
-0.0038195757661014795,
0.013171887956559658,
-0.03383830562233925,
-0.09896382689476013,
0.013824623078107834,
0.13817466795444489,
0.0034514935687184334,
0.00682973163202405,
-0.03995988517999649,
-0.07027145475149155,
-0.05825701728463173,
-0.07912654429674149,
0.057147104293107986,
0.187900573015213,
-0.09512355923652649,
0.1602867990732193,
0.12431421875953674,
-0.06468851119279861,
-0.2306901067495346,
0.03996593505144119,
0.04701630026102066,
0.007666614837944508,
0.022401191294193268,
-0.19138796627521515,
0.09788824617862701,
0.0009011493530124426,
-0.06807263940572739,
0.14616990089416504,
-0.16564498841762543,
-0.1461436152458191,
0.08002161979675293,
0.025075770914554596,
-0.22560662031173706,
-0.14821304380893707,
-0.1037549376487732,
-0.03735695406794548,
-0.13707835972309113,
0.048581719398498535,
0.02614329755306244,
0.019834673032164574,
0.025222565978765488,
0.005338077899068594,
0.029657263308763504,
-0.07272187620401382,
0.1870686560869217,
-0.020297454670071602,
0.0072362530045211315,
-0.050640691071748734,
-0.04617878794670105,
0.09227550774812698,
-0.06150037795305252,
0.11741586774587631,
0.018679620698094368,
0.018796883523464203,
-0.1431548148393631,
-0.049209367483854294,
-0.060803934931755066,
0.04456847906112671,
-0.07284719496965408,
-0.09393193572759628,
-0.04137463867664337,
0.08888561278581619,
0.07211937010288239,
-0.032792408019304276,
-0.0027768779546022415,
-0.07569456845521927,
0.09405932575464249,
0.184477761387825,
0.17357055842876434,
0.009977072477340698,
-0.07020942866802216,
0.024555526673793793,
-0.042279548943042755,
0.03349342197179794,
-0.24652716517448425,
0.03456863760948181,
0.066053606569767,
0.03803660348057747,
0.08509242534637451,
-0.016836483031511307,
-0.1781480610370636,
-0.04086102172732353,
0.08498652279376984,
-0.06206206604838371,
-0.19876568019390106,
-0.02703288197517395,
0.08424776047468185,
-0.20383712649345398,
-0.032998621463775635,
0.041543323546648026,
-0.03834589570760727,
-0.02396267279982567,
-0.002415500348433852,
0.06396626681089401,
-0.008327016606926918,
0.12156640738248825,
0.06747189164161682,
0.10266115516424179,
-0.09284433722496033,
0.08920657634735107,
0.10416955500841141,
-0.09140542894601822,
0.03545991703867912,
0.10264154523611069,
-0.05670900270342827,
-0.04460543021559715,
0.033935222774744034,
0.05925208330154419,
-0.028357384726405144,
-0.06409841030836105,
-0.000502707262057811,
-0.0359574519097805,
0.04993389546871185,
0.08058220148086548,
0.036113787442445755,
-0.01202210783958435,
0.06544706225395203,
0.028145326301455498,
-0.11693570017814636,
0.10949387401342392,
0.04405685141682625,
0.04509059712290764,
-0.07182393968105316,
-0.012280966155230999,
0.015999672934412956,
0.032540347427129745,
-0.019734015688300133,
-0.014576527290046215,
-0.03146412968635559,
-0.007561005651950836,
-0.1553635597229004,
-0.02064543403685093,
-0.06516171246767044,
0.006067827809602022,
0.022207623347640038,
-0.03830232471227646,
-0.012014663778245449,
0.01381110493093729,
-0.07979435473680496,
-0.07571027427911758,
-0.01700955256819725,
0.08539021760225296,
-0.1381402313709259,
0.006627439055591822,
0.07182712107896805,
-0.10980239510536194,
0.07347989827394485,
-0.0048679932951927185,
0.017079560086131096,
0.010923396795988083,
-0.11654401570558548,
0.04386281594634056,
-0.005810429807752371,
0.01551580335944891,
0.022556742653250694,
-0.171111062169075,
0.011553828604519367,
-0.038553636521101,
-0.03114982508122921,
0.011926400475203991,
-0.025060230866074562,
-0.11875922232866287,
0.08676479011774063,
-0.028097305446863174,
-0.037512701004743576,
-0.03292486071586609,
0.06296087801456451,
0.08736220002174377,
-0.011740099638700485,
0.09667140990495682,
-0.025766119360923767,
0.04818311333656311,
-0.1756584197282791,
-0.01910574547946453,
-0.050167568027973175,
0.02537350542843342,
-0.01759655587375164,
-0.0070639788173139095,
0.055272240191698074,
-0.004191063344478607,
0.20991376042366028,
-0.03921036794781685,
0.1548677533864975,
0.05199402943253517,
-0.009925156831741333,
0.010884369723498821,
0.05032730847597122,
0.06423956155776978,
0.031145188957452774,
0.00853167474269867,
0.04660189896821976,
-0.004552975296974182,
-0.020357951521873474,
-0.13699717819690704,
0.02791593410074711,
0.16117429733276367,
0.061918217688798904,
0.0392887257039547,
0.03704594820737839,
-0.1422400325536728,
-0.09538721293210983,
0.10306388139724731,
-0.0331864058971405,
0.014331420883536339,
-0.08317886292934418,
0.17621558904647827,
0.12328410148620605,
-0.1574767529964447,
0.0577850341796875,
-0.07234696298837662,
-0.05066767707467079,
-0.1024852767586708,
-0.11832084506750107,
-0.06293155997991562,
-0.06027044355869293,
-0.004747506696730852,
-0.042489297688007355,
0.05734556168317795,
0.026751231402158737,
-0.003270963439717889,
-0.006759525276720524,
0.12665949761867523,
-0.0249644722789526,
-0.004145825747400522,
0.04152364656329155,
0.0326087586581707,
0.019319625571370125,
-0.05872373282909393,
0.017997145652770996,
0.018602589145302773,
0.022180357947945595,
0.06835069507360458,
0.0260987039655447,
-0.059317342936992645,
0.044286735355854034,
0.00319746439345181,
-0.11313364654779434,
0.018146557733416557,
-0.00002245741598017048,
-0.05020225793123245,
0.13557326793670654,
0.04076748713850975,
0.01548024732619524,
-0.029270920902490616,
0.24342355132102966,
-0.07199113070964813,
-0.08681939542293549,
-0.13965600728988647,
0.11511493474245071,
-0.023563209921121597,
0.03755274787545204,
0.016542524099349976,
-0.12659503519535065,
0.011511262506246567,
0.18531471490859985,
0.12824349105358124,
0.012459068559110165,
-0.007656481582671404,
0.05736639350652695,
-0.0007639875984750688,
-0.05985576659440994,
0.05051197111606598,
0.0664999932050705,
0.16097788512706757,
-0.09069112688302994,
0.0652846097946167,
-0.008405503816902637,
-0.0831485390663147,
-0.027498632669448853,
0.11705785244703293,
-0.022675158455967903,
0.02148384228348732,
-0.03778035193681717,
0.11204422265291214,
-0.052532415837049484,
-0.2719486355781555,
0.02952493168413639,
-0.09503202140331268,
-0.13993041217327118,
-0.02591860294342041,
0.041448429226875305,
-0.03349510580301285,
0.01577647216618061,
0.06254769116640091,
-0.045389387756586075,
0.18837277591228485,
0.025987716391682625,
-0.08679025620222092,
-0.07755549252033234,
0.05874146893620491,
-0.08695939928293228,
0.2789687216281891,
0.003863075515255332,
0.04782010242342949,
0.12108923494815826,
-0.03053574077785015,
-0.18664880096912384,
0.014769754372537136,
0.11989909410476685,
-0.09114406257867813,
0.07780203968286514,
0.18139931559562683,
-0.005561648402363062,
0.12649618089199066,
0.04705416411161423,
-0.03877115994691849,
0.03976387158036232,
-0.02721380814909935,
-0.03821522742509842,
-0.12209630757570267,
0.05661242455244064,
-0.0612691193819046,
0.15957388281822205,
0.1158948540687561,
-0.05964287370443344,
0.001120698289014399,
-0.06126941740512848,
0.06300627440214157,
0.014774397015571594,
0.12115653604269028,
0.018452486023306847,
-0.2023056596517563,
0.05087360367178917,
-0.03283824771642685,
0.08166342973709106,
-0.254973828792572,
-0.08186668157577515,
0.07622263580560684,
-0.019022729247808456,
-0.04275642707943916,
0.12311509251594543,
0.06101066991686821,
0.03676839917898178,
-0.03853875398635864,
-0.08537755906581879,
-0.01412904355674982,
0.15376435220241547,
-0.14123432338237762,
-0.029574336484074593
] |
null | null | transformers |
# BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation
Model card for image captioning pretrained on COCO dataset - base architecture (with ViT large backbone).
|  |
|:--:|
| <b> Pull figure from BLIP official repo | Image source: https://github.com/salesforce/BLIP </b>|
## TL;DR
Authors from the [paper](https://arxiv.org/abs/2201.12086) write in the abstract:
*Vision-Language Pre-training (VLP) has advanced the performance for many vision-language tasks. However, most existing pre-trained models only excel in either understanding-based tasks or generation-based tasks. Furthermore, performance improvement has been largely achieved by scaling up the dataset with noisy image-text pairs collected from the web, which is a suboptimal source of supervision. In this paper, we propose BLIP, a new VLP framework which transfers flexibly to both vision-language understanding and generation tasks. BLIP effectively utilizes the noisy web data by bootstrapping the captions, where a captioner generates synthetic captions and a filter removes the noisy ones. We achieve state-of-the-art results on a wide range of vision-language tasks, such as image-text retrieval (+2.7% in average recall@1), image captioning (+2.8% in CIDEr), and VQA (+1.6% in VQA score). BLIP also demonstrates strong generalization ability when directly transferred to videolanguage tasks in a zero-shot manner. Code, models, and datasets are released.*
## Usage
You can use this model for conditional and un-conditional image captioning
### Using the Pytorch model
#### Running the model on CPU
<details>
<summary> Click to expand </summary>
```python
import requests
from PIL import Image
from transformers import BlipProcessor, BlipForConditionalGeneration
processor = BlipProcessor.from_pretrained("Salesforce/blip-image-captioning-large")
model = BlipForConditionalGeneration.from_pretrained("Salesforce/blip-image-captioning-large")
img_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg'
raw_image = Image.open(requests.get(img_url, stream=True).raw).convert('RGB')
# conditional image captioning
text = "a photography of"
inputs = processor(raw_image, text, return_tensors="pt")
out = model.generate(**inputs)
print(processor.decode(out[0], skip_special_tokens=True))
# unconditional image captioning
inputs = processor(raw_image, return_tensors="pt")
out = model.generate(**inputs)
print(processor.decode(out[0], skip_special_tokens=True))
```
</details>
#### Running the model on GPU
##### In full precision
<details>
<summary> Click to expand </summary>
```python
import requests
from PIL import Image
from transformers import BlipProcessor, BlipForConditionalGeneration
processor = BlipProcessor.from_pretrained("Salesforce/blip-image-captioning-large")
model = BlipForConditionalGeneration.from_pretrained("Salesforce/blip-image-captioning-large").to("cuda")
img_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg'
raw_image = Image.open(requests.get(img_url, stream=True).raw).convert('RGB')
# conditional image captioning
text = "a photography of"
inputs = processor(raw_image, text, return_tensors="pt").to("cuda")
out = model.generate(**inputs)
print(processor.decode(out[0], skip_special_tokens=True))
# unconditional image captioning
inputs = processor(raw_image, return_tensors="pt").to("cuda")
out = model.generate(**inputs)
print(processor.decode(out[0], skip_special_tokens=True))
```
</details>
##### In half precision (`float16`)
<details>
<summary> Click to expand </summary>
```python
import torch
import requests
from PIL import Image
from transformers import BlipProcessor, BlipForConditionalGeneration
processor = BlipProcessor.from_pretrained("Salesforce/blip-image-captioning-large")
model = BlipForConditionalGeneration.from_pretrained("Salesforce/blip-image-captioning-large", torch_dtype=torch.float16).to("cuda")
img_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg'
raw_image = Image.open(requests.get(img_url, stream=True).raw).convert('RGB')
# conditional image captioning
text = "a photography of"
inputs = processor(raw_image, text, return_tensors="pt").to("cuda", torch.float16)
out = model.generate(**inputs)
print(processor.decode(out[0], skip_special_tokens=True))
# >>> a photography of a woman and her dog
# unconditional image captioning
inputs = processor(raw_image, return_tensors="pt").to("cuda", torch.float16)
out = model.generate(**inputs)
print(processor.decode(out[0], skip_special_tokens=True))
>>> a woman sitting on the beach with her dog
```
</details>
## BibTex and citation info
```
@misc{https://doi.org/10.48550/arxiv.2201.12086,
doi = {10.48550/ARXIV.2201.12086},
url = {https://arxiv.org/abs/2201.12086},
author = {Li, Junnan and Li, Dongxu and Xiong, Caiming and Hoi, Steven},
keywords = {Computer Vision and Pattern Recognition (cs.CV), FOS: Computer and information sciences, FOS: Computer and information sciences},
title = {BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation},
publisher = {arXiv},
year = {2022},
copyright = {Creative Commons Attribution 4.0 International}
}
``` | {"license": "bsd-3-clause", "tags": ["image-captioning"], "pipeline_tag": "image-to-text", "languages": ["en"]} | image-to-text | gizmo-ai/blip-image-captioning-large | [
"transformers",
"pytorch",
"tf",
"safetensors",
"blip",
"text2text-generation",
"image-captioning",
"image-to-text",
"arxiv:2201.12086",
"license:bsd-3-clause",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T08:47:59+00:00 | [
"2201.12086"
] | [] | TAGS
#transformers #pytorch #tf #safetensors #blip #text2text-generation #image-captioning #image-to-text #arxiv-2201.12086 #license-bsd-3-clause #autotrain_compatible #endpoints_compatible #region-us
| BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation
========================================================================================================
Model card for image captioning pretrained on COCO dataset - base architecture (with ViT large backbone).
TL;DR
-----
Authors from the paper write in the abstract:
*Vision-Language Pre-training (VLP) has advanced the performance for many vision-language tasks. However, most existing pre-trained models only excel in either understanding-based tasks or generation-based tasks. Furthermore, performance improvement has been largely achieved by scaling up the dataset with noisy image-text pairs collected from the web, which is a suboptimal source of supervision. In this paper, we propose BLIP, a new VLP framework which transfers flexibly to both vision-language understanding and generation tasks. BLIP effectively utilizes the noisy web data by bootstrapping the captions, where a captioner generates synthetic captions and a filter removes the noisy ones. We achieve state-of-the-art results on a wide range of vision-language tasks, such as image-text retrieval (+2.7% in average recall@1), image captioning (+2.8% in CIDEr), and VQA (+1.6% in VQA score). BLIP also demonstrates strong generalization ability when directly transferred to videolanguage tasks in a zero-shot manner. Code, models, and datasets are released.*
Usage
-----
You can use this model for conditional and un-conditional image captioning
### Using the Pytorch model
#### Running the model on CPU
Click to expand
#### Running the model on GPU
##### In full precision
Click to expand
##### In half precision ('float16')
Click to expand
BibTex and citation info
------------------------
| [
"### Using the Pytorch model",
"#### Running the model on CPU\n\n\n\n Click to expand",
"#### Running the model on GPU",
"##### In full precision\n\n\n\n Click to expand",
"##### In half precision ('float16')\n\n\n\n Click to expand \n\nBibTex and citation info\n------------------------"
] | [
"TAGS\n#transformers #pytorch #tf #safetensors #blip #text2text-generation #image-captioning #image-to-text #arxiv-2201.12086 #license-bsd-3-clause #autotrain_compatible #endpoints_compatible #region-us \n",
"### Using the Pytorch model",
"#### Running the model on CPU\n\n\n\n Click to expand",
"#### Running the model on GPU",
"##### In full precision\n\n\n\n Click to expand",
"##### In half precision ('float16')\n\n\n\n Click to expand \n\nBibTex and citation info\n------------------------"
] | [
78,
9,
11,
8,
9,
27
] | [
"passage: TAGS\n#transformers #pytorch #tf #safetensors #blip #text2text-generation #image-captioning #image-to-text #arxiv-2201.12086 #license-bsd-3-clause #autotrain_compatible #endpoints_compatible #region-us \n### Using the Pytorch model#### Running the model on CPU\n\n\n\n Click to expand#### Running the model on GPU##### In full precision\n\n\n\n Click to expand##### In half precision ('float16')\n\n\n\n Click to expand \n\nBibTex and citation info\n------------------------"
] | [
-0.11166590452194214,
0.02385896071791649,
-0.0025774622336030006,
0.12677650153636932,
0.08719374984502792,
0.02891787700355053,
0.11310332268476486,
0.09582334756851196,
0.025206318125128746,
0.0451727919280529,
0.12596268951892853,
0.15332633256912231,
0.05572886019945145,
0.24887868762016296,
-0.04043459892272949,
-0.2405615597963333,
0.015875136479735374,
0.07609176635742188,
0.12668102979660034,
0.13645502924919128,
0.039823051542043686,
-0.144617959856987,
0.10721034556627274,
-0.059959281235933304,
-0.1796085089445114,
-0.014452570118010044,
0.014258426614105701,
-0.05353936925530434,
0.16589444875717163,
0.09590816497802734,
-0.04352971166372299,
0.07373030483722687,
0.0758068859577179,
-0.1134905144572258,
0.0459117516875267,
0.016550961881875992,
-0.07969265431165695,
0.14459706842899323,
0.12375155091285706,
0.023815521970391273,
0.17587772011756897,
0.02012038789689541,
-0.028779273852705956,
0.01926250010728836,
-0.05168469622731209,
-0.0990898609161377,
0.003122485475614667,
0.19304251670837402,
0.0290928203612566,
-0.019642740488052368,
0.02484707534313202,
0.1302570253610611,
0.0010200859978795052,
0.12750865519046783,
0.1990700662136078,
-0.26707902550697327,
-0.03061041608452797,
0.025120094418525696,
0.0903315618634224,
0.08117852360010147,
-0.07017969340085983,
0.040853261947631836,
0.007191153708845377,
0.013350658118724823,
0.06361158192157745,
-0.01640748605132103,
-0.026148628443479538,
-0.027378778904676437,
-0.06361749023199081,
-0.09079030156135559,
0.04408285766839981,
0.0030673141591250896,
-0.033478930592536926,
-0.0733918622136116,
-0.1420874446630478,
-0.04100378602743149,
-0.0702105313539505,
0.01013229601085186,
0.008120378479361534,
-0.02643241360783577,
0.0012567760422825813,
-0.10381131619215012,
-0.10656868666410446,
-0.13181059062480927,
-0.1244172528386116,
0.15073835849761963,
0.03491426631808281,
0.06812708079814911,
-0.03966625779867172,
0.16165152192115784,
-0.010275575332343578,
-0.08152454346418381,
0.0019306157482787967,
-0.07200398296117783,
0.0483616478741169,
0.05501438304781914,
-0.015517682768404484,
0.0747910588979721,
0.050703078508377075,
0.13489270210266113,
-0.03583954647183418,
-0.04023087024688721,
-0.0590578094124794,
0.09014618396759033,
-0.03617976978421211,
-0.024797122925519943,
-0.12840943038463593,
-0.01261854637414217,
0.10003524273633957,
0.07804078608751297,
0.05221547186374664,
-0.025579635053873062,
-0.14247965812683105,
-0.07021936029195786,
0.10144305974245071,
0.03866992145776749,
0.01519932970404625,
0.05160988122224808,
-0.04790133237838745,
-0.06403802335262299,
0.24922844767570496,
-0.08737276494503021,
-0.018383171409368515,
0.02415132336318493,
-0.03953971341252327,
-0.04905763268470764,
0.11079879850149155,
-0.07511856406927109,
-0.0895010307431221,
0.0490286722779274,
-0.0806298777461052,
-0.014323662966489792,
-0.0835508480668068,
-0.05455111712217331,
-0.003438379615545273,
-0.057465050369501114,
0.022816838696599007,
-0.14320598542690277,
-0.19942964613437653,
0.09960301220417023,
0.02300695702433586,
-0.07380712032318115,
-0.02375464141368866,
0.05823420360684395,
-0.007099035661667585,
-0.03065672144293785,
-0.08238887041807175,
0.17313678562641144,
-0.0315980538725853,
0.18945631384849548,
0.012866766192018986,
0.13672024011611938,
-0.19103610515594482,
0.044697973877191544,
-0.12144539505243301,
0.018980834633111954,
-0.031034773215651512,
0.060875408351421356,
0.022107357159256935,
-0.02433014288544655,
-0.016927627846598625,
-0.0704566240310669,
-0.03694935515522957,
-0.0037383593153208494,
0.10281915217638016,
0.11408361047506332,
-0.12211144715547562,
-0.03510735556483269,
0.19733926653862,
-0.11434237658977509,
-0.19537340104579926,
0.10688868165016174,
-0.021427420899271965,
0.03187498450279236,
0.04988126456737518,
0.08940625190734863,
-0.01933274231851101,
-0.09339039027690887,
0.029315466061234474,
0.06641430407762527,
-0.09902498871088028,
-0.10181369632482529,
0.05641856789588928,
0.10651161521673203,
0.017497239634394646,
0.055793143808841705,
-0.01561116985976696,
0.044525016099214554,
-0.05906595289707184,
-0.062923364341259,
-0.026720162481069565,
-0.057883016765117645,
0.06947268545627594,
0.05532835051417351,
0.008546188473701477,
-0.02176031842827797,
-0.05002385750412941,
0.025486450642347336,
0.08652141690254211,
-0.037638965994119644,
0.02708715759217739,
-0.07182815670967102,
0.15135350823402405,
-0.11418111622333527,
0.05749543756246567,
-0.17824599146842957,
0.04029978811740875,
0.006453216541558504,
0.04413970932364464,
0.08462794125080109,
-0.000529569573700428,
0.05573943629860878,
0.11775332689285278,
-0.029892055317759514,
-0.0016481357160955667,
0.04691829904913902,
-0.02043377421796322,
-0.1005590558052063,
-0.09611134976148605,
-0.040382612496614456,
-0.03843383118510246,
-0.012366444803774357,
-0.14773322641849518,
0.04089336469769478,
0.0859891027212143,
0.05364865064620972,
-0.019460467621684074,
0.04371977597475052,
-0.060241423547267914,
-0.055171094834804535,
-0.0654512271285057,
-0.0035822561476379633,
0.13137666881084442,
0.011402365751564503,
0.011338779702782631,
0.09960836917161942,
-0.14405089616775513,
0.20705480873584747,
0.18167276680469513,
-0.08704154193401337,
-0.011764212511479855,
-0.13883556425571442,
-0.05284974351525307,
-0.04302379861474037,
0.05139344930648804,
-0.040207188576459885,
-0.00028934847796335816,
0.029966503381729126,
0.1603582799434662,
-0.14675173163414001,
-0.028108416125178337,
0.051581766456365585,
-0.012450385838747025,
-0.0046003758907318115,
0.05258157476782799,
0.1288476288318634,
-0.16890498995780945,
0.1040821298956871,
0.1307235062122345,
0.0071649085730314255,
0.1685243397951126,
0.06965499371290207,
-0.12101784348487854,
0.01409935113042593,
-0.002964464481920004,
-0.014487005770206451,
0.17660662531852722,
-0.06374708563089371,
0.022809874266386032,
0.09448094666004181,
-0.07417766749858856,
0.07138999551534653,
-0.17500805854797363,
0.016208773478865623,
-0.03764064982533455,
-0.031590577214956284,
0.03812725096940994,
0.008127325214445591,
-0.005246035289019346,
0.1595027893781662,
-0.022911377251148224,
-0.07225310057401657,
0.060412678867578506,
-0.005250153131783009,
-0.05632934719324112,
0.15005044639110565,
-0.01934080943465233,
-0.25475382804870605,
-0.16607290506362915,
-0.12917786836624146,
-0.10491731017827988,
0.03179384022951126,
0.041541263461112976,
0.06714688986539841,
-0.034811437129974365,
-0.04984952136874199,
-0.0972379669547081,
0.09000424295663834,
-0.028733836486935616,
-0.09777703136205673,
0.04476844519376755,
0.014835636131465435,
-0.0991448238492012,
-0.017255550250411034,
-0.042333219200372696,
-0.0487126000225544,
0.12727490067481995,
0.0022932584397494793,
0.085948146879673,
0.1025322675704956,
-0.07962934672832489,
0.007924327626824379,
-0.010514923371374607,
0.07083958387374878,
-0.044501516968011856,
0.06200579181313515,
0.21397562325000763,
-0.042581621557474136,
0.05831864848732948,
0.1163770854473114,
0.032398566603660583,
-0.0838608667254448,
0.03467882424592972,
-0.08397041261196136,
-0.08417216688394547,
-0.10007651150226593,
-0.0629502460360527,
-0.06159172207117081,
0.0808950662612915,
0.1397387534379959,
0.051395971328020096,
0.051512572914361954,
0.13834679126739502,
-0.04491965100169182,
0.06207479164004326,
0.046675700694322586,
0.06926237046718597,
0.11616349965333939,
-0.02918427437543869,
0.14324770867824554,
-0.04763789474964142,
-0.028002535924315453,
0.11957390606403351,
0.12776465713977814,
0.08129172027111053,
-0.11378835886716843,
0.09275196492671967,
0.04828765615820885,
0.0857996717095375,
0.07382919639348984,
0.21249718964099884,
-0.08796725422143936,
0.027516193687915802,
-0.05884896591305733,
-0.08012860268354416,
-0.08169476687908173,
0.02376210130751133,
-0.04314330592751503,
0.011481797322630882,
-0.015925223007798195,
-0.05310347303748131,
0.036948952823877335,
0.07184307277202606,
0.010157456621527672,
-0.3211396038532257,
-0.06268860399723053,
-0.001369061297737062,
0.00975314062088728,
-0.14219869673252106,
0.003503207815811038,
0.11877397447824478,
-0.06997953355312347,
-0.03520011901855469,
-0.07577083259820938,
0.09469106048345566,
0.008124100044369698,
0.01303725503385067,
-0.02942049875855446,
0.09528230875730515,
0.0003084458876401186,
0.12792284786701202,
-0.2619419991970062,
0.1551353633403778,
0.024414535611867905,
-0.014260931871831417,
-0.13109342753887177,
0.015768099576234818,
0.05791233107447624,
0.07560893893241882,
0.11944649368524551,
-0.06685096025466919,
0.17280632257461548,
-0.12297689914703369,
-0.12194168567657471,
0.015456601977348328,
0.03275355324149132,
-0.02583022229373455,
0.010396292433142662,
-0.024135662242770195,
-0.013248185627162457,
-0.05999568849802017,
0.03322060406208038,
0.007121694274246693,
-0.1339312493801117,
0.043902087956666946,
-0.07294410467147827,
-0.027485722675919533,
-0.04996621236205101,
-0.0925576239824295,
-0.051802780479192734,
0.12819691002368927,
-0.025812827050685883,
-0.07887894660234451,
-0.0972747653722763,
0.08996112644672394,
0.041928429156541824,
-0.09864170849323273,
0.06083039194345474,
-0.08645473420619965,
0.08790338784456253,
-0.05516591668128967,
-0.15545833110809326,
0.1016850471496582,
-0.06152395159006119,
-0.15652036666870117,
-0.029306739568710327,
0.13866490125656128,
-0.08493489027023315,
-0.01765834353864193,
0.015530107542872429,
-0.015143423341214657,
-0.08170193433761597,
-0.0833376795053482,
-0.01792294532060623,
-0.0952683761715889,
0.09664591401815414,
-0.004256340675055981,
-0.13959945738315582,
-0.16175343096256256,
-0.01401050016283989,
0.022026944905519485,
0.12435495853424072,
0.24134759604930878,
-0.057832762598991394,
0.10196340084075928,
0.2557581067085266,
-0.0031087032984942198,
-0.3000974953174591,
-0.11760640889406204,
-0.042485561221838,
-0.004324517212808132,
0.01872905157506466,
-0.09226761013269424,
0.05489378422498703,
0.072666697204113,
-0.03993380442261696,
0.18581852316856384,
-0.14193911850452423,
-0.13026025891304016,
0.1479993462562561,
0.18953385949134827,
0.0907890573143959,
-0.11940007656812668,
-0.0489363819360733,
-0.0792984738945961,
-0.14312520623207092,
0.2148916870355606,
-0.03247098997235298,
0.08867485821247101,
-0.08273915946483612,
-0.0001253264199476689,
0.01390625536441803,
-0.05507713928818703,
0.08407190442085266,
-0.09787508100271225,
0.07354552298784256,
-0.13365451991558075,
-0.07486820220947266,
0.09369613230228424,
-0.006917967926710844,
0.11053521186113358,
-0.10866191238164902,
0.07941597700119019,
-0.02132938615977764,
-0.05688481032848358,
-0.09903448075056076,
-0.0023273169063031673,
-0.007188163232058287,
-0.10884898155927658,
-0.01961902342736721,
0.02168206498026848,
-0.06641718000173569,
-0.008771194145083427,
-0.05353166535496712,
-0.04481031373143196,
-0.03167607635259628,
0.274949312210083,
0.09753918647766113,
-0.14369073510169983,
0.04764076694846153,
-0.10218729078769684,
-0.051181014627218246,
0.08881424367427826,
-0.07121586054563522,
0.03587178885936737,
0.11566893756389618,
0.008505642414093018,
0.05886262282729149,
0.06266634166240692,
-0.042088259011507034,
-0.00784717220813036,
0.10235296934843063,
-0.2120855450630188,
-0.05533710867166519,
-0.029306527227163315,
0.07041808217763901,
-0.030407587066292763,
0.06376485526561737,
0.09198079258203506,
-0.028203019872307777,
-0.05643470957875252,
0.023211341351270676,
0.04298434033989906,
-0.036236342042684555,
0.1085398942232132,
0.09944486618041992,
0.03248794376850128,
-0.1099168136715889,
0.020120272412896156,
0.02630075253546238,
-0.0529741607606411,
-0.07814833521842957,
0.13164022564888,
-0.09140054136514664,
-0.10036639869213104,
-0.034076642245054245,
0.025670606642961502,
-0.059620823711156845,
-0.043598636984825134,
-0.017504526302218437,
-0.05036405101418495,
0.04628530517220497,
0.13674886524677277,
0.09059959650039673,
0.01299795787781477,
-0.039537463337183,
0.006614416837692261,
-0.09087120741605759,
0.16057832539081573,
-0.04720934480428696,
0.07495194673538208,
-0.0388764925301075,
0.12073364853858948,
0.015496434643864632,
0.13362935185432434,
-0.08657897263765335,
0.005720170214772224,
-0.06690666079521179,
0.01222214475274086,
-0.15695646405220032,
0.038913432508707047,
-0.0849018320441246,
-0.009020611643791199,
-0.034558139741420746,
0.007576033007353544,
-0.031747400760650635,
-0.015979688614606857,
-0.10252635926008224,
-0.00420484971255064,
0.01523549947887659,
0.011050837114453316,
-0.15859095752239227,
-0.00566478306427598,
0.0301375649869442,
-0.045116961002349854,
0.12980687618255615,
0.014558334834873676,
-0.02538532391190529,
0.021561939269304276,
-0.16968025267124176,
-0.035767193883657455,
0.038633596152067184,
0.07983538508415222,
0.016140379011631012,
0.048084359616041183,
0.0838681012392044,
0.06547082960605621,
-0.029947616159915924,
-0.025327473878860474,
0.1358751356601715,
-0.09280207008123398,
0.017165588214993477,
-0.07470989972352982,
-0.060399558395147324,
-0.056462571024894714,
0.05456256866455078,
-0.053777873516082764,
0.06702708452939987,
0.08408113569021225,
-0.040656790137290955,
0.008752526715397835,
-0.18308529257774353,
0.016116421669721603,
-0.025808114558458328,
-0.14089235663414001,
-0.05853842571377754,
-0.07998736947774887,
0.0430225245654583,
0.018571583554148674,
0.21747726202011108,
0.056131310760974884,
-0.07523278146982193,
-0.030850905925035477,
0.052024420350790024,
0.026873597875237465,
-0.00273060891777277,
0.2481766790151596,
-0.025641197338700294,
0.05051442235708237,
-0.07722308486700058,
0.06629785895347595,
0.020890232175588608,
0.00890614278614521,
0.09254787117242813,
0.10543397814035416,
0.009168528020381927,
0.04264584556221962,
0.10348193347454071,
-0.019281890243291855,
-0.02611200511455536,
-0.09502832591533661,
-0.06251213699579239,
0.02133796736598015,
-0.06966685503721237,
0.16320769488811493,
0.14821362495422363,
-0.10444727540016174,
0.022871896624565125,
0.023294681683182716,
-0.04221270978450775,
-0.04888465628027916,
-0.11772701889276505,
-0.045710109174251556,
-0.04654162749648094,
-0.009146470576524734,
-0.06314364075660706,
-0.05019639804959297,
0.1258738934993744,
0.0020111599005758762,
-0.08957146853208542,
0.14466825127601624,
-0.02967998757958412,
-0.031063862144947052,
0.019898055121302605,
0.042817771434783936,
-0.02385665290057659,
-0.014257446862757206,
-0.05642341449856758,
-0.0007333000539802015,
0.005448049400001764,
0.0501730777323246,
-0.01355029083788395,
0.04851750284433365,
0.005383591167628765,
-0.012142131105065346,
-0.04412110894918442,
-0.010479878634214401,
-0.019277557730674744,
0.04952653869986534,
0.1760987788438797,
-0.011533400043845177,
0.021068792790174484,
0.00914407055824995,
0.10393031686544418,
-0.022090857848525047,
-0.14667105674743652,
-0.016426289454102516,
0.21474745869636536,
-0.041807714849710464,
0.028416365385055542,
0.02229253016412258,
-0.05836828425526619,
-0.001460572937503457,
0.22410057485103607,
0.19801874458789825,
-0.20838961005210876,
0.002580324187874794,
0.0755905881524086,
0.022441545501351357,
0.04955677315592766,
0.15145224332809448,
0.06964106857776642,
0.27373939752578735,
-0.048164453357458115,
0.012994332239031792,
-0.029094157740473747,
0.015035918913781643,
0.02402958832681179,
0.05660584941506386,
0.013743261806666851,
-0.028934484347701073,
-0.0916401743888855,
-0.0038938079960644245,
-0.007027700077742338,
-0.01658187434077263,
0.01581924594938755,
-0.1112687885761261,
-0.07120001316070557,
-0.00700072105973959,
0.0194179005920887,
-0.03287206217646599,
-0.015927504748106003,
-0.048541005700826645,
0.11090004444122314,
0.00206334819085896,
0.002843801397830248,
-0.18912935256958008,
0.0631820410490036,
0.07506293058395386,
-0.09874610602855682,
0.24105538427829742,
-0.07189328223466873,
0.09182820469141006,
0.05497939512133598,
0.007739938795566559,
-0.1081935241818428,
0.10788386315107346,
-0.03459221497178078,
-0.05722992494702339,
0.06269054859876633,
0.03935655578970909,
0.031926125288009644,
-0.17952032387256622,
0.016302434727549553,
-0.07562170922756195,
0.02023269794881344,
-0.01882977783679962,
-0.05869387462735176,
-0.039364397525787354,
0.05475076287984848,
-0.06624381244182587,
0.09491024911403656,
0.07140397280454636,
-0.013332939706742764,
-0.04197130352258682,
-0.12939883768558502,
0.05365307256579399,
0.017096322029829025,
0.005585821811109781,
-0.007474787998944521,
-0.08918555080890656,
-0.023520030081272125,
-0.05408027395606041,
-0.006227187812328339,
-0.3363669514656067,
-0.015958329662680626,
-0.07825682312250137,
-0.02750428393483162,
-0.2094404101371765,
0.07478093355894089,
0.11775477975606918,
0.038186367601156235,
-0.008797294460237026,
-0.07874569296836853,
-0.04734478145837784,
0.03159322217106819,
-0.12243970483541489,
-0.07677605003118515
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [Laurie]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [En]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | Laurie/phi2_DPO | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-08T08:48:32+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by: [Laurie]
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP): [En]
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: [Laurie]\n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): [En]\n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: [Laurie]\n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): [En]\n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
31,
6,
3,
90,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: [Laurie]\n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): [En]\n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.054196134209632874,
0.20484185218811035,
-0.004303015768527985,
0.02659502811729908,
0.10846821963787079,
0.003680384950712323,
0.06099560111761093,
0.11256509274244308,
-0.007235040422528982,
0.11959873884916306,
0.028566740453243256,
0.08275841921567917,
0.1189548596739769,
0.15476387739181519,
0.003695658640936017,
-0.24076876044273376,
0.05059663951396942,
-0.09401450306177139,
0.0030361220706254244,
0.11259933561086655,
0.1343085616827011,
-0.10673597455024719,
0.0896289274096489,
-0.007599486969411373,
-0.011671803891658783,
-0.013459917157888412,
-0.07526063919067383,
-0.06647757440805435,
0.05941231921315193,
0.07617790997028351,
0.0741746798157692,
0.011243109591305256,
0.08297235518693924,
-0.28342947363853455,
0.014096211642026901,
0.08139947056770325,
0.00048522837460041046,
0.06792433559894562,
0.08203725516796112,
-0.06268644332885742,
0.1191689670085907,
-0.07379422336816788,
0.1392553597688675,
0.07333510369062424,
-0.0876007154583931,
-0.1881076991558075,
-0.06934322416782379,
0.06595537811517715,
0.13602198660373688,
0.052500803023576736,
-0.02747165597975254,
0.13948379456996918,
-0.09495876729488373,
0.006959725636988878,
0.12020473182201385,
-0.07297036051750183,
-0.05541558936238289,
0.034095801413059235,
0.10249555855989456,
0.08430936187505722,
-0.11706697940826416,
-0.00859979260712862,
0.032457176595926285,
0.020408717915415764,
0.08825699985027313,
0.016136134043335915,
0.1442171037197113,
0.03940529003739357,
-0.14096789062023163,
-0.05164537951350212,
0.09138494729995728,
0.03971962258219719,
-0.04985390603542328,
-0.23116213083267212,
-0.024738242849707603,
-0.017108628526329994,
-0.02907055802643299,
-0.042414113879203796,
0.05353047698736191,
-0.037973057478666306,
0.07635346055030823,
-0.005380644463002682,
-0.08113472163677216,
-0.02952219918370247,
0.05165145546197891,
0.06496024876832962,
0.02140076830983162,
-0.008364762179553509,
0.01716778427362442,
0.11503452062606812,
0.06978731602430344,
-0.12895606458187103,
-0.07277439534664154,
-0.07323681563138962,
-0.1025879830121994,
-0.04822540283203125,
0.033937595784664154,
0.07125266641378403,
0.04237307980656624,
0.20704352855682373,
-0.018962392583489418,
0.045796848833560944,
0.05107459798455238,
0.004804511554539204,
0.06853224337100983,
0.10670563578605652,
-0.07090245932340622,
-0.1475491225719452,
-0.05689878389239311,
0.08437176048755646,
-0.0032260592561215162,
-0.03297841176390648,
-0.04686526209115982,
0.040434472262859344,
0.029213763773441315,
0.11492762714624405,
0.08537550270557404,
-0.003091292455792427,
-0.06230713799595833,
-0.03866563364863396,
0.22142980992794037,
-0.14764712750911713,
0.04180923476815224,
0.0034967248793691397,
-0.04661210998892784,
-0.009176385588943958,
0.00813100952655077,
0.018074076622724533,
-0.03101499006152153,
0.10731850564479828,
-0.07364045828580856,
-0.038437213748693466,
-0.10992707312107086,
-0.0639590471982956,
0.029407989233732224,
-0.0009278096840716898,
-0.015165421180427074,
-0.04561445862054825,
-0.11079031229019165,
-0.0507732555270195,
0.07150499522686005,
-0.07558774948120117,
-0.059258025139570236,
0.010905029252171516,
-0.04919581860303879,
0.003817802993580699,
-0.00008406545384787023,
0.11748078465461731,
-0.033696793019771576,
0.021994875743985176,
-0.04586203023791313,
0.06896261870861053,
0.11153269559144974,
0.035643916577100754,
-0.08051832765340805,
0.07121429592370987,
-0.22423554956912994,
0.10124862939119339,
-0.0870344340801239,
0.028491005301475525,
-0.146965891122818,
-0.04262514412403107,
0.024924268946051598,
0.02438051998615265,
-0.002707310952246189,
0.1316716969013214,
-0.19725756347179413,
-0.028144391253590584,
0.15303917229175568,
-0.11286210268735886,
-0.09187445044517517,
0.06527622789144516,
-0.05454457178711891,
0.10776735097169876,
0.04794113337993622,
-0.028587447479367256,
0.06808029115200043,
-0.13729360699653625,
-0.04727635160088539,
-0.023322638124227524,
-0.01082488801330328,
0.13643363118171692,
0.06700792908668518,
-0.05375295132398605,
0.0691985934972763,
0.020448893308639526,
-0.032202113419771194,
-0.0341593399643898,
-0.03774197772145271,
-0.09413006156682968,
0.005643716547638178,
-0.0737539529800415,
0.033095184713602066,
-0.02253885753452778,
-0.08659622073173523,
-0.03163355216383934,
-0.1721361130475998,
0.04135269671678543,
0.08344686776399612,
0.00878971815109253,
-0.01921849697828293,
-0.08853428810834885,
0.01988689787685871,
-0.009819399565458298,
-0.020725302398204803,
-0.1616506725549698,
-0.050107426941394806,
0.042298708111047745,
-0.20050568878650665,
0.020499365404248238,
-0.040027737617492676,
0.05043455958366394,
0.0349392369389534,
-0.04404103383421898,
-0.0033105670008808374,
0.005217714700847864,
0.018728790804743767,
-0.027158789336681366,
-0.20083889365196228,
-0.032431282103061676,
-0.027181854471564293,
0.13650386035442352,
-0.22705727815628052,
0.028333701193332672,
0.07192924618721008,
0.14116482436656952,
-0.0021952646784484386,
-0.04146401211619377,
0.01783418469130993,
-0.0529899038374424,
-0.05282416194677353,
-0.06658631563186646,
-0.007816963829100132,
-0.032101523131132126,
-0.04268839582800865,
0.06328320503234863,
-0.18858541548252106,
-0.04381110519170761,
0.10594075173139572,
0.10121870785951614,
-0.14873114228248596,
-0.02548077329993248,
-0.04134270176291466,
-0.06483501940965652,
-0.09459517896175385,
-0.05775724723935127,
0.13649335503578186,
0.049901653081178665,
0.048421066254377365,
-0.0852949321269989,
-0.06281457096338272,
0.010642818175256252,
0.0002461542608216405,
-0.038235150277614594,
0.08905230462551117,
0.0897173210978508,
-0.10746387392282486,
0.08493547886610031,
0.07839972525835037,
0.06032858416438103,
0.09770677238702774,
0.002925068372860551,
-0.10917813330888748,
-0.026515016332268715,
0.009010876528918743,
0.01760687306523323,
0.138573557138443,
-0.04204317182302475,
0.04952899366617203,
0.0548083521425724,
-0.028478918597102165,
0.017474593594670296,
-0.10389458388090134,
0.02663234993815422,
0.04546825960278511,
-0.009118902496993542,
0.01642083190381527,
-0.03805535286664963,
0.02735576033592224,
0.08699426054954529,
0.03643663972616196,
0.028634296730160713,
0.004960530903190374,
-0.0348934642970562,
-0.10378069430589676,
0.1773453950881958,
-0.08740012347698212,
-0.28715023398399353,
-0.1395212858915329,
-0.005917896516621113,
0.048900071531534195,
-0.022870227694511414,
0.01382638979703188,
-0.05290757492184639,
-0.114642433822155,
-0.10459273308515549,
0.004664338193833828,
0.04527437314391136,
-0.07383424788713455,
-0.06815873831510544,
0.052729249000549316,
0.035394541919231415,
-0.14084014296531677,
0.02611432410776615,
0.046706847846508026,
-0.038183100521564484,
-0.009392914362251759,
0.07293549180030823,
0.10490678250789642,
0.1736450493335724,
-0.007402011658996344,
-0.02189239300787449,
0.025457460433244705,
0.24189408123493195,
-0.1432628631591797,
0.10984954237937927,
0.15489278733730316,
-0.059070128947496414,
0.10438910126686096,
0.19858647882938385,
0.024531487375497818,
-0.07874298095703125,
0.039219845086336136,
0.0411323644220829,
-0.05388592928647995,
-0.23801544308662415,
-0.06000780314207077,
-0.00015711525338701904,
-0.07549291104078293,
0.09354081749916077,
0.08894579857587814,
0.10939071327447891,
0.04515286907553673,
-0.08406683802604675,
-0.06744616478681564,
0.019352653995156288,
0.10727629065513611,
-0.026827704161405563,
0.010388152673840523,
0.08985289931297302,
-0.04497305303812027,
-0.0063971360214054585,
0.10793449729681015,
0.011718768626451492,
0.19342637062072754,
0.028250370174646378,
0.14991357922554016,
0.07147187739610672,
0.03941499814391136,
0.023529235273599625,
0.027174340561032295,
0.0283015388995409,
0.009653751738369465,
-0.013237119652330875,
-0.09016717970371246,
0.0314609631896019,
0.13576993346214294,
0.06785121560096741,
0.030110478401184082,
0.02007933519780636,
-0.03625325858592987,
0.05689863860607147,
0.17028085887432098,
0.012141929939389229,
-0.21461573243141174,
-0.04096812382340431,
0.08644591271877289,
-0.07741178572177887,
-0.12076176702976227,
-0.024002665653824806,
0.03842300921678543,
-0.1800841987133026,
0.0454949326813221,
-0.015790048986673355,
0.11236972361803055,
-0.11858978122472763,
-0.029217496514320374,
0.043445076793432236,
0.08599399775266647,
-0.03147190064191818,
0.07890886813402176,
-0.17111892998218536,
0.12127456814050674,
0.010778852738440037,
0.06265894323587418,
-0.11200644820928574,
0.09620469063520432,
0.010271726176142693,
-0.004536936525255442,
0.16734054684638977,
-0.00036840757820755243,
-0.07875418663024902,
-0.0682671070098877,
-0.08478140085935593,
-0.02276461571455002,
0.09188129007816315,
-0.1087576374411583,
0.08456107974052429,
-0.008431249298155308,
-0.041868239641189575,
0.006700813304632902,
-0.11097506433725357,
-0.1313074678182602,
-0.19485437870025635,
0.058790843933820724,
-0.1109076514840126,
-0.0006678894278593361,
-0.09896750003099442,
-0.06011313199996948,
-0.04811393842101097,
0.20001813769340515,
-0.14025837182998657,
-0.09612520039081573,
-0.15294025838375092,
-0.09738217294216156,
0.1732979267835617,
-0.04963665455579758,
0.08736883103847504,
0.00024931548978202045,
0.22158105671405792,
0.005620776209980249,
-0.00908597931265831,
0.07660334557294846,
-0.08735139667987823,
-0.1770162731409073,
-0.07553574442863464,
0.11920882016420364,
0.12599524855613708,
0.051466528326272964,
-0.0107383718714118,
0.023264845833182335,
-0.03392580524086952,
-0.11662154644727707,
0.0016911056591197848,
0.12043455988168716,
0.06845682114362717,
0.03728953376412392,
-0.005637803580611944,
-0.10317178070545197,
-0.07359828054904938,
-0.04061805456876755,
0.026015158742666245,
0.1889028400182724,
-0.08594691753387451,
0.1533898264169693,
0.13432054221630096,
-0.053626738488674164,
-0.21395084261894226,
0.033365193754434586,
0.04353572800755501,
0.0037413304671645164,
0.05579587444663048,
-0.18152093887329102,
0.08503967523574829,
0.027432706207036972,
-0.05166502296924591,
0.14662088453769684,
-0.1674591451883316,
-0.15796831250190735,
0.07806592434644699,
0.053346388041973114,
-0.21290254592895508,
-0.12140168249607086,
-0.08724400401115417,
-0.06651420146226883,
-0.1423894762992859,
0.0886000320315361,
0.013351885601878166,
-0.0017088119639083743,
0.04950578883290291,
0.032524995505809784,
0.017978202551603317,
-0.05023800954222679,
0.2136572301387787,
-0.0013972087763249874,
0.033360619097948074,
-0.0755331963300705,
-0.09439089894294739,
0.06506326049566269,
-0.05430286377668381,
0.08845636993646622,
-0.02103913575410843,
0.004903497640043497,
-0.0825686976313591,
-0.05818142369389534,
-0.05026451125741005,
0.03579851612448692,
-0.08047130703926086,
-0.10398096591234207,
-0.06391946226358414,
0.09361781924962997,
0.09515181928873062,
-0.034639839082956314,
-0.03638337552547455,
-0.08351866155862808,
0.039766453206539154,
0.2006448358297348,
0.17177192866802216,
0.05389874801039696,
-0.09489423036575317,
0.006840402260422707,
-0.018438085913658142,
0.03691910207271576,
-0.21123355627059937,
0.04707885533571243,
0.04755498096346855,
0.026349766179919243,
0.11329349875450134,
-0.01940120942890644,
-0.16173596680164337,
-0.0480998270213604,
0.056721121072769165,
-0.03497987613081932,
-0.2007400244474411,
-0.010869455523788929,
0.045929353684186935,
-0.18113857507705688,
-0.06366261839866638,
0.014949681237339973,
-0.016242405399680138,
-0.025617282837629318,
0.00973495189100504,
0.06534307450056076,
0.023916900157928467,
0.10175001621246338,
0.060897957533597946,
0.10039069503545761,
-0.11216479539871216,
0.08403176814317703,
0.09666462242603302,
-0.08674376457929611,
0.01304932776838541,
0.07204759865999222,
-0.053834524005651474,
-0.022957248613238335,
0.02666306123137474,
0.057283736765384674,
-0.0034068080130964518,
-0.06195664405822754,
-0.02049374021589756,
-0.1045675277709961,
0.0673963725566864,
0.12596629559993744,
0.040291834622621536,
-0.007705249823629856,
0.05418014898896217,
0.01790819689631462,
-0.08749373257160187,
0.11045672744512558,
0.03683001920580864,
0.03751925006508827,
-0.05941551923751831,
-0.024470645934343338,
0.04322197660803795,
0.010362990200519562,
-0.01937495730817318,
-0.025460831820964813,
-0.05474642291665077,
-0.01284365076571703,
-0.18922050297260284,
0.01718529500067234,
-0.0767764002084732,
0.007099643815308809,
0.010170014575123787,
-0.03960195183753967,
-0.018824564293026924,
0.015784690156579018,
-0.07798941433429718,
-0.05046527832746506,
-0.007164747919887304,
0.0974695160984993,
-0.1421574205160141,
0.00626219529658556,
0.09087475389242172,
-0.11800844222307205,
0.06598333269357681,
-0.023052450269460678,
-0.014316370710730553,
0.0011266916990280151,
-0.12474151700735092,
0.042620155960321426,
-0.002806516131386161,
0.020079849287867546,
0.04416791722178459,
-0.1773548126220703,
0.0054356129840016365,
-0.04030992090702057,
-0.04348345473408699,
-0.015151025727391243,
-0.08064394444227219,
-0.11696093529462814,
0.10613346099853516,
-0.0006396674434654415,
-0.08380883187055588,
-0.012203032150864601,
0.04590419679880142,
0.111353300511837,
-0.04131268709897995,
0.12284190207719803,
0.004365136381238699,
0.06566130369901657,
-0.17895303666591644,
-0.025891343131661415,
-0.016704514622688293,
0.009906311519443989,
0.02109062299132347,
-0.011865048669278622,
0.044236920773983,
-0.010623261332511902,
0.2586866617202759,
-0.026337098330259323,
0.07741348445415497,
0.06469458341598511,
0.035697877407073975,
0.01332850567996502,
0.08686363697052002,
0.06741948425769806,
0.01674581505358219,
0.0036508720368146896,
0.028528675436973572,
-0.030600380152463913,
-0.013306314125657082,
-0.15302540361881256,
0.0790591612458229,
0.147830069065094,
0.08837617933750153,
0.01121124904602766,
0.061743419617414474,
-0.1058998703956604,
-0.10455755889415741,
0.08116292953491211,
-0.03990678861737251,
-0.0047508240677416325,
-0.058698106557130814,
0.15225175023078918,
0.1493549942970276,
-0.17402169108390808,
0.08073722571134567,
-0.04040298983454704,
-0.05228932946920395,
-0.11675874888896942,
-0.15877290070056915,
-0.06669758260250092,
-0.02437772788107395,
-0.0038315868005156517,
-0.055703576654195786,
0.06390844285488129,
0.11026692390441895,
0.0010562894167378545,
0.0005360586219467223,
0.09678085893392563,
-0.018575146794319153,
-0.013244815170764923,
0.03266690671443939,
0.04919743537902832,
0.03673020377755165,
-0.0448305681347847,
0.020055722445249557,
0.007341388147324324,
0.03690418228507042,
0.05807460844516754,
0.024638215079903603,
-0.03513272851705551,
0.016517208889126778,
-0.005543774459511042,
-0.10325343906879425,
0.02057637646794319,
-0.02585965394973755,
-0.0637068971991539,
0.1282096803188324,
0.031878504902124405,
0.016980821266770363,
-0.032755643129348755,
0.20601914823055267,
-0.07007027417421341,
-0.0738086923956871,
-0.14077627658843994,
0.10834762454032898,
-0.033487554639577866,
0.06223937124013901,
0.057906173169612885,
-0.11595596373081207,
-0.003984871320426464,
0.13651598989963531,
0.13038693368434906,
-0.02642480842769146,
0.004437415394932032,
0.0294784065335989,
0.0066003757528960705,
-0.0477001890540123,
0.04628758877515793,
0.03525162488222122,
0.1484072506427765,
-0.07306617498397827,
0.07409515976905823,
0.0021855318918824196,
-0.08530484139919281,
-0.04011600464582443,
0.14084024727344513,
0.00016216411313507706,
0.030928824096918106,
-0.06527088582515717,
0.10535262525081635,
-0.07369977980852127,
-0.24339087307453156,
0.04521755129098892,
-0.08010156452655792,
-0.159461110830307,
-0.015108855441212654,
0.025179240852594376,
-0.011890014633536339,
0.02676936611533165,
0.06690191477537155,
-0.0630112960934639,
0.16340845823287964,
0.03588634356856346,
-0.08678793907165527,
-0.06021903082728386,
0.07399401813745499,
-0.08659541606903076,
0.2921910881996155,
0.012431015260517597,
0.03596409410238266,
0.10380212962627411,
-0.020363926887512207,
-0.1379791647195816,
0.02752917818725109,
0.10063077509403229,
-0.0945255309343338,
0.0693148523569107,
0.18235032260417938,
-0.012513813562691212,
0.09940995275974274,
0.07081297785043716,
-0.061855655163526535,
0.05614025145769119,
-0.08290119469165802,
-0.06322108209133148,
-0.09711064398288727,
0.05950569733977318,
-0.06315789371728897,
0.14754429459571838,
0.1215326264500618,
-0.0439717173576355,
-0.0018456866964697838,
-0.02850368805229664,
0.04159243032336235,
0.011714864522218704,
0.12162020057439804,
0.012689411640167236,
-0.1665055900812149,
0.03060060366988182,
0.0011378610506653786,
0.10888917744159698,
-0.21892568469047546,
-0.08445140719413757,
0.0515170581638813,
-0.032984357327222824,
-0.052278101444244385,
0.10494641959667206,
0.06233486160635948,
0.05129052326083183,
-0.04654242470860481,
-0.06279025971889496,
-0.01054215244948864,
0.14981770515441895,
-0.12072893232107162,
-0.010352554731070995
] |
null | null | transformers |
# Full Parameter Finetuning Qwen1.5-0.5B 16384 context length on Malaysian instructions dataset
README at https://github.com/mesolitica/malaya/tree/5.1/session/qwen2
We use exact Qwen1.5 Instruct chat template.
WandB, https://wandb.ai/huseinzol05/Qwen1.5-0.5B-4096-fpf-instructions-16k?workspace=user-huseinzol05
## how-to
```python
from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
import torch
TORCH_DTYPE = 'bfloat16'
nf4_config = BitsAndBytesConfig(
load_in_4bit=True,
bnb_4bit_quant_type='nf4',
bnb_4bit_use_double_quant=True,
bnb_4bit_compute_dtype=getattr(torch, TORCH_DTYPE)
)
tokenizer = AutoTokenizer.from_pretrained('mesolitica/malaysian-Qwen1.5-0.5B-16k-instructions')
model = AutoModelForCausalLM.from_pretrained(
'mesolitica/malaysian-Qwen1.5-0.5B-16k-instructions',
use_flash_attention_2 = True,
quantization_config = nf4_config
)
messages = [
{'role': 'user', 'content': 'KWSP tu apa'}
]
prompt = tokenizer.apply_chat_template(messages, tokenize = False)
inputs = tokenizer([prompt], return_tensors='pt', add_special_tokens=False).to('cuda')
generate_kwargs = dict(
inputs,
max_new_tokens=1024,
top_p=0.95,
top_k=50,
temperature=0.9,
do_sample=True,
num_beams=1,
)
r = model.generate(**generate_kwargs)
tokenizer.decode(r[0])
```
```text
<|im_start|>user
KWSP tu apa<|im_end|>
<|im_start|>assistant
KWSP merujuk kepada Skim Simpanan Wang Persaraan (KWSP), iaitu skim simpanan untuk ahli kumpulan berumur 20 tahun ke atas. KWSP menawarkan beberapa faedah, termasuk:
1. Akaun Simpanan Wajib - Ahli boleh menyumbang kepada KWSP melalui akaun simpanan wajib.
2. Akaun Simpanan Amanah - Ahli boleh menyumbang kepada KWSP melalui akaun amanah.
3. Akaun Simpanan Pelaburan - Ahli boleh menyumbang kepada KWSP melalui pelaburan dalam skim pelaburan KWSP atau pelaburan bersama.
4. Perolehan - Ahli boleh mendapatkan pelbagai perkhidmatan dan ganjaran daripada KWSP.
5. Akaun Simpanan untuk Penyelamat - Ahli boleh menyumbang kepada KWSP melalui akaun simpanan untuk penyelamat, yang boleh digunakan untuk bantuan kewangan atau pemberhentian pasaran kewangan.
KWSP menawarkan faedah dan perlindungan yang komprehensif untuk ahli, termasuk:
1. Akaun Simpanan Amanah - Akaun ini menyediakan faedah dan ganjaran sebelum dan selepas persaraan.
2. Akaun Simpanan - Akaun simpanan menawarkan kadar faedah yang lebih rendah daripada akaun simpanan biasa.
3. Akaun Simpanan Pelaburan - Akaun ini menawarkan kadar faedah yang lebih rendah dan pelbagai ganjaran.
4. Akaun Simpanan untuk Penyelamat - Akaun ini menawarkan kadar faedah yang lebih rendah daripada akaun simpanan biasa dan pelbagai ganjaran.
5. Perolehan - Akaun ini menyediakan faedah dan ganjaran sebelum dan selepas persaraan.
KWSP juga menawarkan pelbagai perkhidmatan dan ganjaran lain, termasuk:
1. Program Pekerjaan - KWSP menyediakan pelbagai skim pekerjaan, termasuk Skim Pencen, Skim Pekerjaan, dan Skim Penginapan.
2. Program Perubatan - KWSP menyediakan pelbagai skim perubatan, termasuk Skim Cukai Perubatan.
3. Program Keselamatan Sosial - KWSP menyediakan pelbagai program keselamatan sosial, termasuk Skim Simpanan 401 (KWSP), Skim Simpanan Amanah (SW401) dan Skim Simpanan Perumahan (RSP) untuk pemastautin berdaftar.
4. Program Penjagaan Kesihatan - KWSP menyediakan pelbagai skim penjagaan kesihatan, termasuk Skim Insurans Kesihatan Swasta (SIC), Skim Insurans Kesihatan Kumpulan Swasta (KSMC), dan Skim Insurans Kesihatan Kumpulan Swasta 1Malaysia (SIC 1M).
5. Program Pemindahan - KWSP menyediakan pelbagai skim pemindahan, termasuk Skim Pemindahan Amanah, Skim Pemindahan KWSP, dan Skim Pemindahan Persekutuan.
KWSP menawarkan pelbagai pilihan dan perlindungan untuk ahli yang ingin menyumbang kepada KWSP untuk persaraan atau bantuan kewangan atau pemberhentian pasaran kewangan.<|im_end|>
```
```python
messages = [
{'role': 'user', 'content': 'macam mana nak kod websocket guna react, contoh kod sekali'}
]
prompt = tokenizer.apply_chat_template(messages, tokenize = False)
inputs = tokenizer([prompt], return_tensors='pt', add_special_tokens=False).to('cuda')
generate_kwargs = dict(
inputs,
max_new_tokens=1024,
top_p=0.95,
top_k=50,
temperature=0.9,
do_sample=True,
num_beams=1,
)
r = model.generate(**generate_kwargs)
tokenizer.decode(r[0])
```
```text
<|im_start|>user
macam mana nak kod websocket guna react, contoh kod sekali<|im_end|>
<|im_start|>assistant
Untuk menggunakan websocket pada React untuk mencipta pelanggan untuk menghantar dan menerima pesanan, anda boleh mengikuti langkah-langkah berikut:
1. Mulakan dengan membuat fail baru, contohnya `WebsocketClient.js`, dan perpustakaan yang diperlukan seperti `ws` untuk Node.js, dan `axios` untuk JavaScript untuk membuat permintaan HTTP ke websocket. Pastikan anda telah mengimport perpustakaan dengan betul.
2. Dalam fail `WebsocketClient.js`, cipta kelas baru dengan nama `WebsocketClient`. Anda boleh menggunakan konstruktor dengan parameter seperti nama aplikasi dan senarai objek pesanan. Contoh:
```
```javascript
import { WebSocketClient, Client } from 'ws';
import axios from 'axios';
class WebsocketClient {
constructor(appName, socket) {
this.appName = appName;
this.socket = socket;
}
connect() {
this.socket.onopen = () => {
console.log('Connection successful');
};
this.socket.onclose = () => {
console.log('Connection lost');
};
this.socket.onmessage = (message) => {
// Code to handle incoming message
};
}
emit(data) {
this.socket.onmessage(data);
}
disconnect() {
this.socket.offopen();
}
}
// Contoh penggunaan
const client = new WebsocketClient('api', axios({ port: 123 }));
client.connect();
client.emit('message', { type: 'info', content: 'Hello, server! });
client.emit('message', { type: 'warning', content: 'Goodbye!'});
client.disconnect();
```
```
3. Dalam fail `WebSocketClient.js`, cipta kelas dengan nama `Client` dengan parameter seperti aplikasi dan `websocket` dari perpustakaan axios. Contoh:
```
```javascript
class Client {
constructor(appName, websocket) {
this.appName = appName;
this.websocket = websocket;
}
connect() {
this.websocket.onopen = () => {
console.log('Connection successful');
};
this.websocket.onclose = () => {
console.log('Connection lost');
};
this.websocket.onmessage = (message) => {
// Code to handle incoming message
};
}
disconnect() {
this.websocket.offopen();
}
}
// Contoh penggunaan
const client = new Client('api', axios({ port: 123 }));
client.connect();
client.emit('message', { type: 'info', content: 'Hello, server! });
client.emit('message', { type: 'warning', content: 'Goodbye!'});
client.disconnect();
```
```
Dalam contoh di atas, kami menggunakan konstruktor yang mengambil nama aplikasi dan websocket sebagai parameter untuk memulakan dan menggunakan fungsi `connect` untuk menghubungi websocket dengan ID sesi yang diinginkan. Kemudian, kami menggunakan fungsi `onmessage` untuk mengendalikan respons websocket dan fungsi `onclose` untuk menutup websocket. Apabila kami menghantar permintaan ke websocket, kami menggunakan kaedah `onmessage` untuk memproses respons tersebut. Apabila kami membuka websocket, kami menggunakan kaedah `onclose` untuk menutup websocket dan menggunakan kaedah `disconnect` untuk membunuhnya.
Dengan menggunakan kod di atas dan menggunakannya pada projek React anda, anda boleh mencipta websocket yang boleh menghantar dan menerima pesanan untuk aplikasi anda.<|im_end|>
``` | {"language": ["ms"]} | text-generation | mesolitica/malaysian-Qwen1.5-0.5B-16k-instructions | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"ms",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T08:52:44+00:00 | [] | [
"ms"
] | TAGS
#transformers #safetensors #qwen2 #text-generation #conversational #ms #autotrain_compatible #endpoints_compatible #region-us
|
# Full Parameter Finetuning Qwen1.5-0.5B 16384 context length on Malaysian instructions dataset
README at URL
We use exact Qwen1.5 Instruct chat template.
WandB, URL
## how-to
| [
"# Full Parameter Finetuning Qwen1.5-0.5B 16384 context length on Malaysian instructions dataset\n\nREADME at URL\n\nWe use exact Qwen1.5 Instruct chat template.\n\nWandB, URL",
"## how-to"
] | [
"TAGS\n#transformers #safetensors #qwen2 #text-generation #conversational #ms #autotrain_compatible #endpoints_compatible #region-us \n",
"# Full Parameter Finetuning Qwen1.5-0.5B 16384 context length on Malaysian instructions dataset\n\nREADME at URL\n\nWe use exact Qwen1.5 Instruct chat template.\n\nWandB, URL",
"## how-to"
] | [
45,
42,
4
] | [
"passage: TAGS\n#transformers #safetensors #qwen2 #text-generation #conversational #ms #autotrain_compatible #endpoints_compatible #region-us \n# Full Parameter Finetuning Qwen1.5-0.5B 16384 context length on Malaysian instructions dataset\n\nREADME at URL\n\nWe use exact Qwen1.5 Instruct chat template.\n\nWandB, URL## how-to"
] | [
-0.09238763898611069,
-0.08712953329086304,
-0.00019865110516548157,
0.024838143959641457,
0.08467838168144226,
-0.062469590455293655,
0.053060803562402725,
0.08114220947027206,
-0.07888399064540863,
0.016763845458626747,
0.030485324561595917,
-0.045858234167099,
0.06248825788497925,
0.05996016785502434,
-0.1080191507935524,
-0.22134579718112946,
0.07734183222055435,
0.011416316963732243,
0.00041946707642637193,
0.11253585666418076,
0.0985342487692833,
-0.0678483173251152,
0.12146001309156418,
0.012971808202564716,
-0.07086123526096344,
-0.00632714107632637,
-0.03229989483952522,
-0.10283646732568741,
0.09175464510917664,
0.06399711221456528,
0.08508756011724472,
0.10023800283670425,
0.005001394543796778,
-0.1652185320854187,
0.04569898173213005,
-0.0459611751139164,
-0.030835682526230812,
-0.002781553426757455,
0.003930260427296162,
0.15101125836372375,
-0.06956006586551666,
0.024232126772403717,
-0.04500385746359825,
0.07626038044691086,
-0.058122485876083374,
0.12868210673332214,
-0.06823074072599411,
0.09949439764022827,
0.19022171199321747,
0.14985796809196472,
-0.053320419043302536,
0.12030193954706192,
-0.11541443318128586,
0.07093224674463272,
0.08333796262741089,
-0.4139663875102997,
0.0003519253805279732,
0.07893839478492737,
0.08912345767021179,
0.09437542408704758,
-0.15380913019180298,
-0.010250016115605831,
0.07616978138685226,
0.008165945298969746,
-0.07178791612386703,
-0.14254891872406006,
-0.12915700674057007,
-0.025707608088850975,
-0.03770372271537781,
0.09259811788797379,
0.19197575747966766,
0.07718092203140259,
-0.02824128232896328,
-0.06594742834568024,
-0.09950072318315506,
-0.04581072926521301,
-0.042217228561639786,
-0.012166964821517467,
-0.07441437989473343,
0.08416593074798584,
-0.056425541639328,
-0.0359506793320179,
-0.12618157267570496,
-0.03926287218928337,
-0.15205718576908112,
0.04715899005532265,
0.02163085900247097,
0.028514428064227104,
-0.22158032655715942,
-0.006659917067736387,
0.005301498342305422,
-0.14339201152324677,
-0.05925361439585686,
-0.0751926526427269,
0.007392259780317545,
0.040728628635406494,
-0.0004538092762231827,
-0.014852706342935562,
0.17460840940475464,
0.05676849186420441,
-0.07786132395267487,
0.08738294988870621,
0.004938262049108744,
0.025065146386623383,
-0.04191741347312927,
0.11509127169847488,
-0.016281208023428917,
-0.10553529113531113,
0.13243468105793,
0.020707784220576286,
0.034534599632024765,
-0.03301328793168068,
-0.0788550078868866,
-0.09936820715665817,
0.023601606488227844,
0.08624694496393204,
-0.03606954962015152,
0.11580926179885864,
0.04052850604057312,
0.006389512214809656,
0.21422378718852997,
-0.09143438935279846,
-0.0628303810954094,
0.04769228398799896,
-0.009533210657536983,
-0.08207289129495621,
0.0029728792142122984,
0.0682230219244957,
-0.06501717120409012,
-0.029197707772254944,
-0.012900176458060741,
-0.039622049778699875,
0.03899411857128143,
-0.06477716565132141,
-0.03603744134306908,
0.12216514348983765,
0.034616682678461075,
-0.17936702072620392,
-0.06334494799375534,
-0.009218472056090832,
-0.08718962222337723,
0.016940925270318985,
-0.05332043766975403,
-0.02771284431219101,
-0.05093282461166382,
0.01896178163588047,
-0.03621353581547737,
0.02137596905231476,
-0.056334685534238815,
0.08287031203508377,
0.15381570160388947,
0.08959201723337173,
-0.1606018990278244,
0.012148675508797169,
-0.10295543819665909,
0.03001120686531067,
0.0006550525431521237,
0.09631841629743576,
-0.03702289238572121,
0.11092180013656616,
-0.02339027263224125,
-0.0047343019396066666,
-0.06476029753684998,
-0.020796315744519234,
0.028867758810520172,
0.1428568810224533,
-0.08686938881874084,
-0.019360488280653954,
0.23896533250808716,
-0.11266174167394638,
-0.19561654329299927,
0.13176901638507843,
0.046095266938209534,
0.018877064809203148,
0.009350747801363468,
0.1714482307434082,
0.20513959228992462,
-0.04539235308766365,
-0.007231090683490038,
0.11502350121736526,
-0.07297492027282715,
-0.06381877511739731,
0.014838368631899357,
0.05253683403134346,
0.024145280942320824,
0.06804082542657852,
0.0677819550037384,
0.07336177676916122,
-0.009485769085586071,
-0.08131401985883713,
-0.03550160303711891,
-0.10434290766716003,
0.013923045247793198,
-0.06437444686889648,
0.06463362276554108,
-0.0607033409178257,
0.041557058691978455,
-0.13208959996700287,
0.07323708385229111,
-0.0023482460528612137,
0.011964534409344196,
-0.1354415863752365,
0.006659274455159903,
0.02029535174369812,
0.053579095751047134,
-0.08905883878469467,
-0.1181129738688469,
-0.004386401269584894,
0.16568560898303986,
0.021161386743187904,
0.015853650867938995,
0.02749895304441452,
0.017805153504014015,
-0.03864657133817673,
-0.09093288332223892,
0.07913052290678024,
-0.0006456179544329643,
-0.12175469845533371,
0.033647239208221436,
0.06337957084178925,
-0.0140930050984025,
0.0663524642586708,
-0.09098612517118454,
0.03760914131999016,
0.051312509924173355,
-0.03244369477033615,
-0.010118776932358742,
0.07858642935752869,
0.09819471836090088,
0.006619476247578859,
-0.0009016540716402233,
0.02725679613649845,
0.03138410300016403,
0.056962594389915466,
-0.09522706270217896,
0.1623637080192566,
-0.18773527443408966,
0.09132833033800125,
0.12234731763601303,
-0.13280464708805084,
0.06170232221484184,
-0.06626210361719131,
-0.06748553365468979,
-0.04754343256354332,
-0.09336522221565247,
0.024277128279209137,
0.1451437771320343,
0.02119085006415844,
0.1462230235338211,
-0.06628193706274033,
-0.016195842996239662,
0.0034549732226878405,
-0.07789316773414612,
-0.0020787669345736504,
0.037395745515823364,
-0.018677158281207085,
-0.1736295223236084,
0.029554730281233788,
0.07112062722444534,
0.03434780612587929,
0.1208854541182518,
0.004644544795155525,
0.00406113313511014,
0.029726160690188408,
0.10525187104940414,
-0.052562128752470016,
0.050526853650808334,
-0.14961136877536774,
-0.05030687525868416,
0.02931840904057026,
0.059563200920820236,
0.06561428308486938,
-0.11439791321754456,
-0.04915465787053108,
-0.005177921149879694,
-0.03439909219741821,
-0.11708909273147583,
0.019515452906489372,
-0.012214585207402706,
0.07745065540075302,
-0.014247390441596508,
-0.030471613630652428,
0.018876180052757263,
-0.07414331287145615,
-0.1506182998418808,
0.15265454351902008,
-0.08050061017274857,
-0.25429147481918335,
-0.10828021913766861,
-0.11193832010030746,
-0.13598684966564178,
-0.057349223643541336,
0.11939046531915665,
-0.19202260673046112,
-0.039028387516736984,
-0.012135517783463001,
-0.03372253105044365,
0.06758124381303787,
0.057046737521886826,
0.030103912577033043,
0.023381784558296204,
0.003123261034488678,
-0.124506376683712,
-0.004618310835212469,
-0.03250516206026077,
-0.07470856606960297,
0.11969808489084244,
-0.14052481949329376,
0.1040850579738617,
0.10774574428796768,
-0.05604562163352966,
0.02979741431772709,
-0.016492033377289772,
0.3107089698314667,
-0.026782648637890816,
0.05030333623290062,
0.17445538938045502,
-0.0232832133769989,
0.022461676970124245,
0.18325316905975342,
-0.037074822932481766,
-0.08013535290956497,
0.09075203537940979,
0.039972156286239624,
-0.005678507033735514,
-0.16177590191364288,
-0.05885164812207222,
-0.0920434519648552,
0.07476828992366791,
-0.10906669497489929,
0.052673522382974625,
0.026491330936551094,
0.07527140527963638,
-0.024185514077544212,
0.017772668972611427,
0.06081216037273407,
0.045876700431108475,
0.10465487092733383,
0.038727130740880966,
0.09088549762964249,
-0.09298533946275711,
0.0028505974914878607,
0.04847100004553795,
0.13305407762527466,
0.09696200489997864,
0.006668750196695328,
0.10311280936002731,
0.03690359741449356,
0.2743629515171051,
0.1620616912841797,
0.06085933372378349,
-0.13358330726623535,
-0.09015683084726334,
0.03815729543566704,
-0.044417183846235275,
-0.05393402650952339,
-0.015876632183790207,
-0.09147060662508011,
0.009552857838571072,
-0.12327352911233902,
0.24593698978424072,
0.08101972937583923,
0.12163551896810532,
0.055135417729616165,
0.007913555018603802,
-0.11989206075668335,
0.01627810299396515,
-0.03355274721980095,
-0.02266097627580166,
0.14124204218387604,
0.21460139751434326,
-0.0797608271241188,
0.009864194318652153,
0.018748417496681213,
0.17058269679546356,
0.004750869702547789,
0.07203249633312225,
-0.13094358146190643,
-0.06671420484781265,
-0.02230975590646267,
0.11207059770822525,
-0.2950739562511444,
0.15979896485805511,
0.02874527871608734,
0.05784664675593376,
-0.007678791414946318,
-0.02856631577014923,
0.03873502090573311,
0.16986076533794403,
0.053016845136880875,
0.003080744529142976,
-0.03478679433465004,
-0.13975273072719574,
-0.07190760225057602,
0.10778507590293884,
0.08823949098587036,
0.18268448114395142,
0.055030133575201035,
-0.03738591820001602,
0.0039472575299441814,
0.0044209579937160015,
0.08859863132238388,
-0.1052423045039177,
-0.10986541956663132,
-0.015062139369547367,
0.1669759303331375,
0.035683974623680115,
-0.044557422399520874,
0.0771535113453865,
-0.06292353570461273,
0.1591041088104248,
-0.11643318086862564,
-0.10378631204366684,
-0.07512784749269485,
-0.03197168558835983,
-0.0073424517177045345,
-0.14532780647277832,
-0.0533255934715271,
-0.02741835080087185,
0.052321404218673706,
-0.048927124589681625,
-0.06111276522278786,
0.06725171953439713,
-0.11154496669769287,
-0.07598891109228134,
0.016349492594599724,
0.13738222420215607,
-0.06705502420663834,
0.02198251336812973,
0.10078483819961548,
-0.11451524496078491,
-0.059572529047727585,
-0.17298799753189087,
-0.06929413974285126,
-0.1343872994184494,
-0.02578376792371273,
-0.05089661106467247,
-0.07667037099599838,
-0.16607747972011566,
-0.09643435478210449,
-0.07696244865655899,
0.25070783495903015,
0.10618510842323303,
-0.022133180871605873,
0.03736067935824394,
0.25015294551849365,
0.04768586531281471,
-0.20457415282726288,
-0.1118050143122673,
-0.06457266956567764,
-0.013819627463817596,
-0.0982823371887207,
0.03152451291680336,
0.093079574406147,
0.024316810071468353,
0.019123775884509087,
-0.0742531344294548,
-0.15969927608966827,
-0.09788116067647934,
0.10257813334465027,
0.03411509096622467,
0.21519064903259277,
-0.1326235979795456,
-0.10073509067296982,
0.005292211193591356,
-0.3284200131893158,
0.016329685226082802,
-0.13450902700424194,
0.04716404154896736,
0.026162579655647278,
0.14716728031635284,
-0.006664978340268135,
0.007516344543546438,
0.13419288396835327,
0.00548340380191803,
-0.004883131943643093,
-0.04919615760445595,
-0.09189168363809586,
0.021743962541222572,
-0.016380945220589638,
0.06392362713813782,
-0.11276765912771225,
0.06867919117212296,
-0.11905855685472488,
0.01646001823246479,
-0.06958171725273132,
-0.012722067534923553,
0.011908085085451603,
-0.0683608278632164,
-0.00498198950663209,
-0.04910893365740776,
0.06563135236501694,
0.024827880784869194,
0.05726731941103935,
-0.07407187670469284,
-0.017233004793524742,
0.1325739473104477,
0.13700498640537262,
-0.21241307258605957,
0.07790061086416245,
-0.07987929880619049,
-0.05093048885464668,
0.09144270420074463,
-0.06591475754976273,
0.06944772601127625,
0.04904560744762421,
-0.03247905150055885,
0.08220858126878738,
0.05784735456109047,
0.04509873315691948,
0.12338411808013916,
0.07922419160604477,
-0.09697124361991882,
-0.0503511019051075,
-0.07423150539398193,
0.10953390598297119,
0.006805097218602896,
0.05359369143843651,
0.06442626565694809,
-0.03687884286046028,
-0.016998039558529854,
-0.02422429621219635,
0.030050672590732574,
-0.08258185535669327,
0.1376042515039444,
0.03142673149704933,
0.09660280495882034,
-0.13164913654327393,
0.06773222982883453,
0.02241155318915844,
-0.059700727462768555,
0.016579871997237206,
0.20879371464252472,
-0.19192218780517578,
-0.08874719589948654,
-0.027408430352807045,
0.1305292248725891,
-0.05726493522524834,
-0.10098511725664139,
-0.047252804040908813,
-0.09364376217126846,
-0.05386725440621376,
0.1285971701145172,
0.003891344415023923,
0.03189906105399132,
0.04294641315937042,
-0.004700768738985062,
-0.10378538817167282,
0.046790316700935364,
-0.022362403571605682,
0.040356334298849106,
-0.12359412759542465,
0.05565773323178291,
0.026645729318261147,
0.19720160961151123,
-0.060244809836149216,
-0.07893986254930496,
-0.10761173814535141,
0.02094755880534649,
-0.2261766940355301,
0.06604465842247009,
-0.1021956279873848,
0.001984481932595372,
-0.10044204443693161,
-0.040965888649225235,
-0.05854878947138786,
0.008543121628463268,
-0.01179672684520483,
-0.009566016495227814,
-0.05152727663516998,
-0.007462019566446543,
-0.14003457129001617,
0.026174137368798256,
0.04902535304427147,
0.013304064981639385,
0.12982037663459778,
0.0616719126701355,
-0.09448227286338806,
0.07388327270746231,
-0.024580955505371094,
-0.07804415374994278,
0.06637535989284515,
0.02496405504643917,
0.08570969104766846,
0.08194992691278458,
0.007122955750674009,
0.039995431900024414,
-0.003288281848654151,
0.047655776143074036,
0.2617895305156708,
-0.08239400386810303,
-0.035765163600444794,
-0.1686849594116211,
-0.04494025185704231,
-0.08595684170722961,
-0.03551171347498894,
0.03326176479458809,
-0.0011660525342449546,
0.13971564173698425,
-0.12997132539749146,
0.028737664222717285,
-0.09829320758581161,
-0.009999305941164494,
0.032187916338443756,
-0.07032064348459244,
-0.09603450447320938,
-0.07641822844743729,
0.03399090841412544,
-0.07978953421115875,
0.06326522678136826,
-0.09265712648630142,
0.046333763748407364,
0.051884379237890244,
-0.025938568636775017,
-0.055285606533288956,
-0.04318271204829216,
0.2962331473827362,
0.10663386434316635,
-0.008931669406592846,
0.026031574234366417,
0.030155295506119728,
0.022500013932585716,
0.0018849357729777694,
0.10902557522058487,
0.06482642143964767,
0.04433682560920715,
0.12408831715583801,
0.03166802600026131,
-0.005977313499897718,
-0.12903399765491486,
-0.1767033338546753,
-0.1415085345506668,
-0.054525334388017654,
-0.030809937044978142,
-0.07836077362298965,
0.2969629466533661,
-0.05241703987121582,
0.008253407664597034,
-0.024456875398755074,
-0.07770904153585434,
-0.1285715401172638,
-0.07382215559482574,
-0.08683932572603226,
-0.08962350338697433,
-0.009855716489255428,
-0.11993709206581116,
-0.048987191170454025,
-0.05660761520266533,
0.03556795045733452,
-0.03502172231674194,
0.2098146229982376,
-0.10378243774175644,
-0.04500764235854149,
0.003559308825060725,
-0.030207306146621704,
0.06438077986240387,
0.10782844573259354,
0.06128288432955742,
0.034530408680438995,
-0.017285162582993507,
-0.005807435605674982,
0.05694784224033356,
-0.022084450349211693,
0.05734295770525932,
-0.06982342153787613,
-0.07744424045085907,
-0.05329914763569832,
0.027557427063584328,
0.04152221605181694,
0.1444225013256073,
0.07478756457567215,
0.009822425432503223,
0.008930637501180172,
0.27461135387420654,
-0.06795064359903336,
-0.19564779102802277,
-0.12826605141162872,
0.0020641349256038666,
0.015291236340999603,
0.06111465021967888,
-0.04750522971153259,
-0.0876198336482048,
-0.04169917479157448,
0.3681924641132355,
0.1528383046388626,
-0.16506431996822357,
0.06872404366731644,
-0.030585000291466713,
0.03393520042300224,
0.021799007430672646,
0.14970235526561737,
0.18131573498249054,
0.29079800844192505,
-0.051514655351638794,
-0.09344392269849777,
-0.027766309678554535,
-0.06957872956991196,
-0.11456210166215897,
0.11570766568183899,
0.03952682390809059,
-0.0016603277763351798,
-0.09029430150985718,
0.06774071604013443,
-0.14742112159729004,
0.133450448513031,
-0.14581389725208282,
-0.08627453446388245,
-0.07826868444681168,
-0.021622540429234505,
0.08306784927845001,
0.09810290485620499,
0.015461593866348267,
0.031132476404309273,
0.09024789184331894,
0.0895250141620636,
0.008585941046476364,
-0.0803147628903389,
0.020465878769755363,
0.07375342398881912,
0.07810518145561218,
0.01493627205491066,
0.047997042536735535,
0.017140710726380348,
0.06454810500144958,
0.06264770030975342,
-0.08816572278738022,
0.2226714938879013,
0.017757030203938484,
-0.06691422313451767,
-0.029246950522065163,
0.06957434117794037,
0.018434034660458565,
-0.004924489185214043,
0.05704415217041969,
-0.15631088614463806,
0.02101750485599041,
0.044274162501096725,
0.05533917620778084,
-0.07040364295244217,
0.08680924773216248,
0.013987376354634762,
0.08620869368314743,
0.07530320435762405,
-0.05489083752036095,
0.0372614748775959,
-0.0028185273986309767,
0.02428686060011387,
-0.012960963882505894,
-0.004283457528799772,
-0.08787792921066284,
-0.15606330335140228,
-0.050551638007164,
0.002280317945405841,
-0.0031572028528898954,
-0.2354307770729065,
-0.0011945475125685334,
-0.07946296781301498,
0.04073293134570122,
-0.07486637681722641,
0.15810097754001617,
0.18355923891067505,
0.0003653982130344957,
-0.021511077880859375,
-0.1205567717552185,
0.018778212368488312,
0.08725840598344803,
-0.13702242076396942,
-0.15110045671463013
] |
null | null | stable-baselines3 |
# **DQN** Agent playing **SpaceInvadersNoFrameskip-v4**
This is a trained model of a **DQN** agent playing **SpaceInvadersNoFrameskip-v4**
using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3)
and the [RL Zoo](https://github.com/DLR-RM/rl-baselines3-zoo).
The RL Zoo is a training framework for Stable Baselines3
reinforcement learning agents,
with hyperparameter optimization and pre-trained agents included.
## Usage (with SB3 RL Zoo)
RL Zoo: https://github.com/DLR-RM/rl-baselines3-zoo<br/>
SB3: https://github.com/DLR-RM/stable-baselines3<br/>
SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib
Install the RL Zoo (with SB3 and SB3-Contrib):
```bash
pip install rl_zoo3
```
```
# Download model and save it into the logs/ folder
python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga Yukino666 -f logs/
python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/
```
If you installed the RL Zoo3 via pip (`pip install rl_zoo3`), from anywhere you can do:
```
python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga Yukino666 -f logs/
python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/
```
## Training (with the RL Zoo)
```
python -m rl_zoo3.train --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/
# Upload the model and generate video (when possible)
python -m rl_zoo3.push_to_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ -orga Yukino666
```
## Hyperparameters
```python
OrderedDict([('batch_size', 32),
('buffer_size', 50000),
('env_wrapper',
['stable_baselines3.common.atari_wrappers.AtariWrapper']),
('exploration_final_eps', 0.01),
('exploration_fraction', 0.1),
('frame_stack', 4),
('gradient_steps', 1),
('learning_rate', 0.0001),
('learning_starts', 50000),
('n_timesteps', 1000000.0),
('optimize_memory_usage', False),
('policy', 'CnnPolicy'),
('target_update_interval', 1000),
('train_freq', 4),
('normalize', False)])
```
# Environment Arguments
```python
{'render_mode': 'rgb_array'}
```
| {"library_name": "stable-baselines3", "tags": ["SpaceInvadersNoFrameskip-v4", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "DQN", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "SpaceInvadersNoFrameskip-v4", "type": "SpaceInvadersNoFrameskip-v4"}, "metrics": [{"type": "mean_reward", "value": "651.00 +/- 257.73", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | Yukino666/dqn-SpaceInvadersNoFrameskip-v4 | [
"stable-baselines3",
"SpaceInvadersNoFrameskip-v4",
"deep-reinforcement-learning",
"reinforcement-learning",
"model-index",
"region:us"
] | 2024-02-08T08:54:59+00:00 | [] | [] | TAGS
#stable-baselines3 #SpaceInvadersNoFrameskip-v4 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
|
# DQN Agent playing SpaceInvadersNoFrameskip-v4
This is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4
using the stable-baselines3 library
and the RL Zoo.
The RL Zoo is a training framework for Stable Baselines3
reinforcement learning agents,
with hyperparameter optimization and pre-trained agents included.
## Usage (with SB3 RL Zoo)
RL Zoo: URL
SB3: URL
SB3 Contrib: URL
Install the RL Zoo (with SB3 and SB3-Contrib):
If you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do:
## Training (with the RL Zoo)
## Hyperparameters
# Environment Arguments
| [
"# DQN Agent playing SpaceInvadersNoFrameskip-v4\nThis is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4\nusing the stable-baselines3 library\nand the RL Zoo.\n\nThe RL Zoo is a training framework for Stable Baselines3\nreinforcement learning agents,\nwith hyperparameter optimization and pre-trained agents included.",
"## Usage (with SB3 RL Zoo)\n\nRL Zoo: URL\nSB3: URL\nSB3 Contrib: URL\n\nInstall the RL Zoo (with SB3 and SB3-Contrib):\n\n\n\n\nIf you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do:",
"## Training (with the RL Zoo)",
"## Hyperparameters",
"# Environment Arguments"
] | [
"TAGS\n#stable-baselines3 #SpaceInvadersNoFrameskip-v4 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n",
"# DQN Agent playing SpaceInvadersNoFrameskip-v4\nThis is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4\nusing the stable-baselines3 library\nand the RL Zoo.\n\nThe RL Zoo is a training framework for Stable Baselines3\nreinforcement learning agents,\nwith hyperparameter optimization and pre-trained agents included.",
"## Usage (with SB3 RL Zoo)\n\nRL Zoo: URL\nSB3: URL\nSB3 Contrib: URL\n\nInstall the RL Zoo (with SB3 and SB3-Contrib):\n\n\n\n\nIf you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do:",
"## Training (with the RL Zoo)",
"## Hyperparameters",
"# Environment Arguments"
] | [
43,
90,
73,
9,
5,
7
] | [
"passage: TAGS\n#stable-baselines3 #SpaceInvadersNoFrameskip-v4 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# DQN Agent playing SpaceInvadersNoFrameskip-v4\nThis is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4\nusing the stable-baselines3 library\nand the RL Zoo.\n\nThe RL Zoo is a training framework for Stable Baselines3\nreinforcement learning agents,\nwith hyperparameter optimization and pre-trained agents included.## Usage (with SB3 RL Zoo)\n\nRL Zoo: URL\nSB3: URL\nSB3 Contrib: URL\n\nInstall the RL Zoo (with SB3 and SB3-Contrib):\n\n\n\n\nIf you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do:## Training (with the RL Zoo)## Hyperparameters# Environment Arguments"
] | [
0.043572068214416504,
0.2414778620004654,
-0.0026879787910729647,
0.012635791674256325,
0.05784223601222038,
0.0030472534708678722,
0.08585051447153091,
0.10650663822889328,
0.024212315678596497,
-0.001382096204906702,
0.003954293206334114,
0.17533031105995178,
0.03632635250687599,
0.13125447928905487,
-0.018073517829179764,
-0.2066594809293747,
-0.013479253277182579,
-0.06247470900416374,
-0.07153085619211197,
0.036099132150411606,
0.07206681370735168,
-0.030116932466626167,
0.036061208695173264,
-0.051406677812337875,
-0.057161085307598114,
0.036824777722358704,
-0.03157254680991173,
0.007067287806421518,
0.15158706903457642,
-0.1222257912158966,
0.12329676002264023,
0.020955175161361694,
0.1896144151687622,
-0.12332789599895477,
0.0339222252368927,
0.08982209116220474,
-0.036988191306591034,
0.013221588917076588,
0.00975361280143261,
-0.052562564611434937,
0.1590864509344101,
-0.09371145814657211,
0.07146181166172028,
0.010926910676062107,
-0.07592244446277618,
-0.1774153709411621,
-0.09356249868869781,
0.07947742193937302,
0.0617753230035305,
0.005319166928529739,
0.03726791962981224,
0.11306490749120712,
-0.020991774275898933,
0.06488905102014542,
0.11562903225421906,
-0.17549200356006622,
0.013578375801444054,
0.17859570682048798,
0.003242473118007183,
0.15767055749893188,
-0.05546637624502182,
0.019877681508660316,
0.02752300351858139,
0.04758313298225403,
0.06873945891857147,
-0.08186400681734085,
-0.1364826112985611,
-0.056155186146497726,
-0.15456219017505646,
-0.03352400287985802,
0.05195203423500061,
-0.011860138736665249,
-0.05783402919769287,
-0.010724928230047226,
-0.04010869935154915,
0.0008851495804265141,
-0.028637725859880447,
0.01805497519671917,
0.07031578570604324,
-0.01226285845041275,
0.02092539705336094,
-0.08391954004764557,
-0.0390290804207325,
-0.038563769310712814,
-0.018022390082478523,
0.12054917961359024,
0.08285853266716003,
0.0266572255641222,
-0.04135355353355408,
0.10274127870798111,
-0.07091585546731949,
-0.05454207584261894,
0.04555258899927139,
-0.03786851093173027,
-0.10615779459476471,
0.02120024710893631,
-0.05905991420149803,
0.026879185810685158,
0.09943640232086182,
0.18048083782196045,
-0.09862488508224487,
0.012620617635548115,
-0.03430783003568649,
0.08121664822101593,
-0.03196052461862564,
0.03197542577981949,
-0.0840383991599083,
-0.016251085326075554,
0.17835216224193573,
0.0030782297253608704,
0.022272996604442596,
0.002074616262689233,
-0.049819961190223694,
-0.02881433069705963,
-0.017756454646587372,
0.06631895154714584,
0.07032092660665512,
0.010587303899228573,
-0.0037596761249005795,
-0.027667716145515442,
-0.036921944469213486,
-0.05629328638315201,
-0.04952820762991905,
0.018803736194968224,
-0.04712437093257904,
-0.047942135483026505,
0.06027210131287575,
-0.005624116864055395,
0.11337806284427643,
-0.025607796385884285,
0.026316547766327858,
-0.019410157576203346,
-0.07494441419839859,
-0.13221681118011475,
-0.0304415225982666,
0.0691632330417633,
0.04371757060289383,
-0.22497159242630005,
-0.16994807124137878,
-0.008539012633264065,
0.017946386709809303,
-0.018741264939308167,
-0.11334165185689926,
0.02453240379691124,
-0.007166135590523481,
-0.049758363515138626,
-0.01601579785346985,
0.10474669933319092,
-0.020438622683286667,
0.018010856583714485,
-0.05593825876712799,
0.16603368520736694,
-0.14290283620357513,
0.031004127115011215,
-0.08706212788820267,
0.023509707301855087,
-0.21286657452583313,
0.041208744049072266,
-0.177636057138443,
0.04863585904240608,
-0.08500861376523972,
0.02327173389494419,
0.021320728585124016,
0.01968831568956375,
0.08580207824707031,
0.10143322497606277,
-0.23631145060062408,
0.05405791476368904,
0.07900930196046829,
-0.022739801555871964,
-0.04218491166830063,
0.06798892468214035,
-0.06558530032634735,
0.1382148116827011,
0.046505436301231384,
0.24831900000572205,
0.10361487418413162,
-0.2036508023738861,
0.061786454170942307,
0.0578593946993351,
-0.08880111575126648,
-0.004730981774628162,
-0.020022382959723473,
0.11598580330610275,
-0.01114928349852562,
0.03338807821273804,
-0.12186288088560104,
0.1456439197063446,
0.02738998830318451,
-0.0165485180914402,
-0.04454165697097778,
-0.1614885926246643,
0.10309953987598419,
-0.015504824928939342,
0.09532155096530914,
-0.042415786534547806,
0.0001161050095106475,
-0.011168917641043663,
0.18012429773807526,
-0.043841805309057236,
0.0007168867159634829,
0.07871408760547638,
0.10895700752735138,
0.028009075671434402,
-0.020230965688824654,
-0.20380273461341858,
-0.0423048660159111,
0.02367858961224556,
0.044489551335573196,
0.2190362960100174,
0.19936694204807281,
0.07770156860351562,
-0.022313760593533516,
-0.025487221777439117,
-0.003248062450438738,
-0.05106664076447487,
0.03467361256480217,
-0.027858436107635498,
-0.024532482028007507,
0.06065356358885765,
-0.09305168688297272,
0.02817818708717823,
-0.13112716376781464,
0.06307920068502426,
-0.17345242202281952,
0.06863926351070404,
0.021998396143317223,
-0.005436043255031109,
0.024577690288424492,
-0.011292695067822933,
-0.034188106656074524,
-0.06233125180006027,
0.07110602408647537,
0.06098933145403862,
0.014702376909554005,
0.0021991983521729708,
-0.0683600977063179,
-0.13828523457050323,
0.08231553435325623,
-0.04042381793260574,
-0.14305958151817322,
0.06392676383256912,
0.011172642931342125,
0.04875864461064339,
-0.05975872278213501,
0.016254881396889687,
0.22900153696537018,
0.05321883037686348,
0.09785865992307663,
-0.04092191904783249,
-0.022525805979967117,
-0.06617844104766846,
-0.06677833944559097,
0.09694591909646988,
0.10812206566333771,
0.060318704694509506,
-0.0030071530491113663,
0.07626225054264069,
0.10942911356687546,
-0.1035122498869896,
-0.0651884600520134,
0.03220061957836151,
-0.05973697826266289,
0.019652515649795532,
0.049140311777591705,
0.02971293032169342,
0.08619047701358795,
0.1833551675081253,
0.008245792239904404,
0.0386311337351799,
-0.025997694581747055,
0.026109617203474045,
-0.15547916293144226,
-0.03145433962345123,
0.04308181628584862,
0.00886955764144659,
-0.07408110797405243,
0.04994636029005051,
0.051439400762319565,
0.13607151806354523,
-0.08217083662748337,
-0.13170577585697174,
-0.059745315462350845,
-0.03804200142621994,
-0.04239124804735184,
0.14975430071353912,
-0.08507520705461502,
-0.19221234321594238,
-0.017164425924420357,
-0.15751953423023224,
-0.02518727444112301,
-0.005179801490157843,
0.002318724524229765,
-0.08325926214456558,
0.017780914902687073,
0.010001576505601406,
-0.03129372000694275,
-0.0684933215379715,
-0.06596160680055618,
-0.05786636844277382,
0.09124112874269485,
0.06932931393384933,
-0.12240120023488998,
-0.00961651187390089,
-0.03742414712905884,
-0.020465577021241188,
0.04516167193651199,
0.08452648669481277,
-0.007267598994076252,
0.07773483544588089,
-0.13209199905395508,
-0.06962883472442627,
0.02834828943014145,
0.2766247093677521,
0.02882981114089489,
0.004668009467422962,
0.17051753401756287,
-0.03629542142152786,
0.04912714660167694,
0.16181479394435883,
0.030781643465161324,
-0.14196757972240448,
0.07090470939874649,
-0.011341600678861141,
-0.09542687982320786,
-0.1706860214471817,
-0.10215658694505692,
-0.037867411971092224,
-0.05015881359577179,
0.05638284236192703,
0.004951419774442911,
-0.04476970434188843,
0.05910305306315422,
0.08782228082418442,
-0.017004497349262238,
-0.06151578947901726,
0.11129767447710037,
0.032263003289699554,
-0.030136963352560997,
0.08078382909297943,
-0.042354047298431396,
-0.04206389561295509,
0.0032403599470853806,
0.22643887996673584,
0.0937788337469101,
-0.01775507442653179,
-0.042567066848278046,
0.019317636266350746,
0.05095715448260307,
0.03613382205367088,
0.11312435567378998,
-0.06975842267274857,
-0.06826137751340866,
-0.035185977816581726,
0.027829548344016075,
-0.02945687249302864,
0.08205190300941467,
0.0630207508802414,
0.005563626065850258,
-0.04653681069612503,
-0.07972332090139389,
-0.04849022626876831,
0.08408913016319275,
-0.027642227709293365,
-0.10093270242214203,
0.09321888536214828,
0.048575710505247116,
0.0016974330646917224,
0.03055831417441368,
0.027994604781270027,
0.01462269201874733,
-0.07982148975133896,
-0.06775744259357452,
0.011468625627458096,
0.07076629996299744,
-0.06822766363620758,
-0.027886953204870224,
-0.19817815721035004,
0.14578363299369812,
0.010630400851368904,
0.04118429124355316,
-0.13048617541790009,
0.1209396943449974,
-0.023116756230592728,
-0.026430301368236542,
0.013811616227030754,
0.0014643745962530375,
0.08203291147947311,
-0.04806509613990784,
0.15762180089950562,
0.009528410620987415,
-0.28092408180236816,
-0.1418946087360382,
-0.08416824042797089,
-0.051183976233005524,
-0.022873088717460632,
0.014752174727618694,
0.0642135739326477,
0.01516205258667469,
0.003868846921250224,
-0.013076163828372955,
0.03185269236564636,
-0.09826882928609848,
-0.06493937969207764,
-0.04839126765727997,
-0.02250157669186592,
-0.06525848805904388,
-0.05647949501872063,
-0.0006809153710491955,
-0.17226077616214752,
0.12522587180137634,
0.11787347495555878,
-0.06451737880706787,
-0.041814323514699936,
-0.06554657220840454,
0.046191465109586716,
-0.07571537792682648,
0.0469326451420784,
0.003414976177737117,
0.019198855385184288,
-0.06806991249322891,
-0.17922484874725342,
0.016097763553261757,
-0.10899919271469116,
0.03772687539458275,
-0.05070559307932854,
0.020257100462913513,
0.08594245463609695,
0.17520126700401306,
0.05856714025139809,
0.01460097823292017,
-0.07239776104688644,
-0.07543374598026276,
-0.0017121878918260336,
-0.06344114243984222,
0.05762333422899246,
-0.009151889942586422,
-0.20333483815193176,
0.02763226442039013,
-0.11414948850870132,
0.06860900670289993,
0.3310066759586334,
0.3324824273586273,
-0.10698744654655457,
0.1177443116903305,
0.04819539934396744,
-0.042202454060316086,
-0.21051374077796936,
-0.002244179602712393,
0.012272895313799381,
0.024992236867547035,
0.13725964725017548,
-0.12924811244010925,
0.05453680083155632,
0.0794181227684021,
-0.024458877742290497,
0.01456840243190527,
-0.09078162908554077,
-0.10816970467567444,
0.20847418904304504,
0.14226987957954407,
0.04421741142868996,
-0.09421348571777344,
0.08391669392585754,
0.004295284394174814,
0.08375877887010574,
0.2107764035463333,
-0.052112679928541183,
0.10695768147706985,
0.005195184610784054,
0.19852910935878754,
0.0328996516764164,
-0.023768596351146698,
0.10834760218858719,
-0.009801650419831276,
0.07911337912082672,
0.03985166177153587,
-0.007676942739635706,
0.010487722232937813,
-0.04522453248500824,
0.014148596674203873,
-0.028376007452607155,
0.010284217074513435,
-0.2274095118045807,
0.0582297146320343,
-0.06368855386972427,
0.04604509472846985,
0.008256820961833,
-0.0999874547123909,
-0.03583388403058052,
0.06431841105222702,
0.08014573156833649,
0.01975327916443348,
0.0436067171394825,
-0.03867863491177559,
0.11051398515701294,
0.20660489797592163,
-0.009811338968575,
0.17751595377922058,
-0.0615963339805603,
0.01464168168604374,
-0.023011628538370132,
-0.04223164543509483,
-0.1462583988904953,
-0.035259708762168884,
0.03498423472046852,
0.057734888046979904,
0.015203364193439484,
0.049647457897663116,
-0.05656236410140991,
0.08498423546552658,
0.021687336266040802,
-0.041541360318660736,
0.033579520881175995,
0.08835696429014206,
0.12415177375078201,
0.010754258371889591,
-0.030121933668851852,
0.06147436052560806,
-0.08128108084201813,
-0.09446098655462265,
-0.004497923422604799,
-0.029991207644343376,
-0.1083834245800972,
0.11353230476379395,
0.16914646327495575,
0.039594944566488266,
-0.057076629251241684,
0.10688766092061996,
-0.02768099494278431,
0.10047874599695206,
0.009198128245770931,
0.06507332623004913,
-0.014091075398027897,
-0.03691792115569115,
0.10611724853515625,
-0.05442855879664421,
-0.01637818105518818,
0.07645545154809952,
-0.06522727757692337,
-0.023877469822764397,
-0.0801999643445015,
0.06034626066684723,
0.09222240000963211,
-0.16854619979858398,
-0.0639432892203331,
-0.032122284173965454,
-0.08628080040216446,
0.013965039514005184,
0.012447911314666271,
0.0710059329867363,
-0.08589600026607513,
0.06316167116165161,
-0.024337708950042725,
0.015639442950487137,
-0.03689891844987869,
0.019222697243094444,
-0.19525384902954102,
-0.002140450058504939,
-0.11280795186758041,
-0.00348020251840353,
-0.002931603929027915,
0.04463808611035347,
-0.04961875081062317,
-0.029358822852373123,
-0.0030675032176077366,
0.044366419315338135,
-0.16609135270118713,
0.002798673929646611,
-0.011639905162155628,
0.03210212290287018,
-0.0002893915225286037,
-0.0983390137553215,
0.014195028692483902,
-0.04294256120920181,
-0.04198618605732918,
0.04925514757633209,
0.009436776861548424,
0.06470516324043274,
-0.2795179784297943,
-0.14905457198619843,
0.030816160142421722,
0.0683867484331131,
0.05483196675777435,
-0.1830425262451172,
0.03568267077207565,
-0.08042316138744354,
-0.02253127470612526,
-0.037770628929138184,
0.018491698428988457,
-0.0539514496922493,
0.0018174031283706427,
-0.04225044324994087,
-0.023033907637000084,
-0.028055014088749886,
-0.07556360960006714,
0.0826747715473175,
0.12462522834539413,
0.07555580884218216,
-0.03807181864976883,
0.09595896303653717,
-0.10009756684303284,
-0.04657831788063049,
-0.04052736237645149,
-0.036951083689928055,
0.017965637147426605,
-0.0870552659034729,
0.048530060797929764,
0.05188591405749321,
0.18719671666622162,
-0.08520494401454926,
-0.058800119906663895,
-0.014255574904382229,
0.0746525228023529,
0.07849094271659851,
0.005095830652862787,
0.17779210209846497,
-0.045693784952163696,
0.05693846940994263,
0.021304311230778694,
0.046699028462171555,
0.10497613251209259,
-0.023569339886307716,
0.14490213990211487,
0.21171095967292786,
-0.037196725606918335,
-0.11048602312803268,
0.043668005615472794,
0.01745123788714409,
-0.002401199424639344,
0.05968761444091797,
0.11983796209096909,
-0.050589341670274734,
-0.10903856158256531,
0.23442286252975464,
0.054169271141290665,
-0.11218088120222092,
0.09546315670013428,
0.039532262831926346,
-0.015890996903181076,
-0.1301896870136261,
0.010444961488246918,
-0.0013640925753861666,
-0.11233190447092056,
0.03386834263801575,
-0.06087532266974449,
-0.025547027587890625,
0.11809267848730087,
0.008789865300059319,
0.03317064419388771,
-0.04139537364244461,
-0.03756232187151909,
-0.04352104663848877,
-0.04273213446140289,
-0.012549578212201595,
-0.02991986647248268,
-0.030186517164111137,
-0.07621737569570541,
-0.007770835887640715,
-0.012012424878776073,
0.030795488506555557,
-0.015285328030586243,
-0.02503054589033127,
-0.021192016080021858,
-0.06697061657905579,
-0.0026312144473195076,
-0.008178025484085083,
0.015549594536423683,
0.010121971368789673,
0.2358063906431198,
0.07042546570301056,
-0.10260069370269775,
-0.01036880537867546,
0.22197756171226501,
-0.03853277862071991,
-0.06528383493423462,
-0.07849395275115967,
0.25128230452537537,
-0.10482002794742584,
0.051095426082611084,
-0.005819917656481266,
-0.06550488620996475,
-0.07153836637735367,
0.2309868484735489,
0.13502730429172516,
-0.1677926480770111,
0.06329060345888138,
-0.0368385910987854,
-0.009490780532360077,
-0.14286863803863525,
0.16013580560684204,
0.1865294873714447,
0.09480160474777222,
-0.12259847670793533,
0.0023130534682422876,
-0.03518044203519821,
-0.018328361213207245,
-0.1660851687192917,
-0.004593863617628813,
-0.029364850372076035,
-0.0427238829433918,
-0.050771355628967285,
0.029773715883493423,
-0.15205919742584229,
-0.0927426889538765,
-0.1916799396276474,
-0.11482496559619904,
-0.12386849522590637,
-0.04549141973257065,
-0.11142764985561371,
-0.0019938007462769747,
0.02257080189883709,
-0.0641874223947525,
0.021061956882476807,
-0.0212461706250906,
-0.05887424945831299,
0.015386379323899746,
-0.08395619690418243,
0.0674985870718956,
0.06488548219203949,
0.15327942371368408,
-0.0790991559624672,
0.025424562394618988,
0.07090727984905243,
-0.057595450431108475,
-0.10164349526166916,
0.06067253649234772,
0.015708057209849358,
-0.1972588747739792,
0.007548294495791197,
0.17712996900081635,
-0.10420889407396317,
0.09745754301548004,
0.048501528799533844,
-0.012951982207596302,
0.0867827981710434,
-0.024721821770071983,
-0.016682926565408707,
-0.04852180927991867,
-0.011212974786758423,
-0.10143939405679703,
0.09892100840806961,
0.0876845121383667,
-0.0517118014395237,
0.07436849176883698,
-0.09508965909481049,
-0.04068392515182495,
0.13103286921977997,
-0.010057874955236912,
-0.08450483530759811,
-0.11667824536561966,
-0.04081142693758011,
0.09684515744447708,
-0.018041390925645828,
-0.20185889303684235,
-0.11639472097158432,
-0.11752668023109436,
-0.00014377340266946703,
-0.03563340753316879,
0.061800602823495865,
0.02430674433708191,
-0.02556120604276657,
-0.008150683715939522,
-0.17615078389644623,
-0.06614746153354645,
0.13479791581630707,
-0.10176112502813339,
-0.07456064969301224
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | unsloth/yi-34b-chat-bnb-4bit | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"4-bit",
"region:us"
] | 2024-02-08T08:55:23+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
63,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.04188906401395798,
0.1868634819984436,
-0.005423274356871843,
0.01757637970149517,
0.09925421327352524,
0.005491955671459436,
0.05528821051120758,
0.11709722131490707,
-0.0533502921462059,
0.12540937960147858,
0.04112355411052704,
0.11151822656393051,
0.11794505268335342,
0.14256049692630768,
-0.0006609223200939596,
-0.21857817471027374,
0.049333877861499786,
-0.10810572654008865,
-0.011713704094290733,
0.12273083627223969,
0.14624762535095215,
-0.10107240080833435,
0.06873463094234467,
-0.03527786582708359,
-0.018390236422419548,
-0.03754361346364021,
-0.0631043016910553,
-0.0428098663687706,
0.04095882177352905,
0.05579971522092819,
0.06504834443330765,
0.0006914962432347238,
0.08312074840068817,
-0.28029099106788635,
0.019063476473093033,
0.06934154778718948,
-0.002301337430253625,
0.0655684545636177,
0.0665774792432785,
-0.06358441710472107,
0.10711093246936798,
-0.053030017763376236,
0.13777746260166168,
0.08432129770517349,
-0.09262416511774063,
-0.18061797320842743,
-0.09176573157310486,
0.1097564846277237,
0.17784090340137482,
0.05030388385057449,
-0.027255237102508545,
0.10309778153896332,
-0.08135051280260086,
0.01753326505422592,
0.049427784979343414,
-0.08851896971464157,
-0.05635679140686989,
0.06728223711252213,
0.09259241074323654,
0.05506645515561104,
-0.12643326818943024,
-0.033265259116888046,
0.005194802302867174,
0.01665104553103447,
0.07253342121839523,
0.021623503416776657,
0.14840956032276154,
0.03535414859652519,
-0.13311921060085297,
-0.05290580168366432,
0.10409694164991379,
0.04025519639253616,
-0.04047621786594391,
-0.2443436086177826,
-0.02754354290664196,
-0.02584495209157467,
-0.034590695053339005,
-0.04391676187515259,
0.04140663892030716,
-0.0025886280927807093,
0.08679894357919693,
-0.008228464052081108,
-0.07375224679708481,
-0.03308725357055664,
0.0671553984284401,
0.062167905271053314,
0.029431181028485298,
-0.020213494077324867,
0.022710734978318214,
0.10721412301063538,
0.09491091966629028,
-0.11595078557729721,
-0.058361902832984924,
-0.06589995324611664,
-0.0719945952296257,
-0.041349031031131744,
0.0347420796751976,
0.016205165535211563,
0.07120134681463242,
0.2571125328540802,
0.01922749914228916,
0.05696013197302818,
0.028859760612249374,
0.007498756982386112,
0.051865316927433014,
0.10615350306034088,
-0.06344283372163773,
-0.11439100652933121,
-0.0198503527790308,
0.08982639759778976,
0.019948948174715042,
-0.03634757548570633,
-0.0477963387966156,
0.06657220423221588,
0.04311024025082588,
0.11142025142908096,
0.09991209954023361,
0.020174458622932434,
-0.07410643249750137,
-0.06169014424085617,
0.19804425537586212,
-0.15571331977844238,
0.03830697387456894,
0.04228408634662628,
-0.0350167416036129,
-0.020801648497581482,
0.008958199061453342,
0.02559036947786808,
-0.03405916318297386,
0.08602763712406158,
-0.0542801134288311,
-0.04678044095635414,
-0.11117270588874817,
-0.03103097900748253,
0.044144779443740845,
0.009705442003905773,
-0.03457854315638542,
-0.03633419796824455,
-0.07459833472967148,
-0.08514720946550369,
0.08634167164564133,
-0.07002478837966919,
-0.05773063376545906,
-0.02634890377521515,
-0.08299881964921951,
0.02310962788760662,
0.020002737641334534,
0.0748848021030426,
-0.025510413572192192,
0.05559871718287468,
-0.0527077317237854,
0.05521956831216812,
0.102859266102314,
0.034644488245248795,
-0.059215348213911057,
0.0586630143225193,
-0.2292465716600418,
0.08570117503404617,
-0.06799479573965073,
0.06182733550667763,
-0.15809562802314758,
-0.02371748723089695,
0.03577778860926628,
0.0046258047223091125,
-0.005437619984149933,
0.1341741383075714,
-0.2101311832666397,
-0.02308833971619606,
0.16872304677963257,
-0.09519724547863007,
-0.07186034321784973,
0.051047924906015396,
-0.04659112170338631,
0.10240385681390762,
0.033361393958330154,
0.0017350923735648394,
0.061108484864234924,
-0.10985323041677475,
-0.011671986430883408,
-0.056128498166799545,
-0.0257779061794281,
0.1382409632205963,
0.07547777146100998,
-0.07926145941019058,
0.06575100868940353,
0.02194836176931858,
-0.020947322249412537,
-0.06538521498441696,
-0.018329601734876633,
-0.1007523387670517,
0.016174660995602608,
-0.06786375492811203,
0.010926736518740654,
-0.018007369711995125,
-0.09520112723112106,
-0.0293081384152174,
-0.16955751180648804,
-0.03186587616801262,
0.08120496571063995,
-0.0039502959698438644,
-0.014517090283334255,
-0.11174587905406952,
0.025241373106837273,
0.033142950385808945,
0.004078295081853867,
-0.13219983875751495,
-0.03811480104923248,
0.03414497151970863,
-0.14914745092391968,
0.03702550381422043,
-0.07292518764734268,
0.05196194723248482,
0.015052439644932747,
-0.02801859751343727,
-0.026396771892905235,
0.02275901474058628,
0.009010270237922668,
-0.016446102410554886,
-0.23674587905406952,
-0.02596096321940422,
-0.02948179468512535,
0.16272801160812378,
-0.2067878395318985,
0.03451593220233917,
0.08174202591180801,
0.15760692954063416,
0.003921739757061005,
-0.05103260651230812,
0.018777096644043922,
-0.07056388258934021,
-0.02478284202516079,
-0.05614311248064041,
0.0029366237577050924,
-0.018842201679944992,
-0.044184111058712006,
0.028629349544644356,
-0.17749272286891937,
-0.0478653647005558,
0.09850458055734634,
0.04734146222472191,
-0.1258343607187271,
-0.02524331584572792,
-0.03737637773156166,
-0.05151774734258652,
-0.0409732386469841,
-0.06301578134298325,
0.09911505877971649,
0.06279782205820084,
0.0382394976913929,
-0.06066882237792015,
-0.07980932295322418,
-0.004368528723716736,
-0.0154745914041996,
-0.024671558290719986,
0.09488890320062637,
0.07743741571903229,
-0.13006442785263062,
0.09399931877851486,
0.0845300555229187,
0.07865235209465027,
0.08858645707368851,
-0.02114405855536461,
-0.07441538572311401,
-0.03704451024532318,
0.037285782396793365,
0.019511960446834564,
0.12384209036827087,
-0.04054240137338638,
0.04314751178026199,
0.04097121208906174,
-0.027314281091094017,
0.0179463941603899,
-0.079317107796669,
0.03420396149158478,
0.022174343466758728,
-0.015434152446687222,
0.054325349628925323,
-0.037130847573280334,
0.019326936453580856,
0.08801604062318802,
0.05892675369977951,
0.04210146516561508,
0.015272390097379684,
-0.05278733745217323,
-0.11141471564769745,
0.15833337604999542,
-0.12401816993951797,
-0.21754354238510132,
-0.13233990967273712,
0.01126034650951624,
0.02784722112119198,
-0.014460757374763489,
0.004845886025577784,
-0.06044495850801468,
-0.10889866203069687,
-0.09166782349348068,
0.006674831733107567,
0.056104786694049835,
-0.08347801119089127,
-0.060731321573257446,
0.04706362634897232,
0.04077374190092087,
-0.14247475564479828,
0.020912135019898415,
0.04298027232289314,
-0.09156464785337448,
-0.011692359112203121,
0.07930251210927963,
0.07647769898176193,
0.18610550463199615,
0.021635528653860092,
-0.02018691971898079,
0.03035759925842285,
0.22015435993671417,
-0.1371978521347046,
0.11312834918498993,
0.13443483412265778,
-0.08657801896333694,
0.08107224106788635,
0.20937806367874146,
0.04128330200910568,
-0.09665821492671967,
0.030488910153508186,
0.030000556260347366,
-0.023369858041405678,
-0.2347419559955597,
-0.07046357542276382,
-0.0001642264542169869,
-0.06512240320444107,
0.07899942249059677,
0.09564073383808136,
0.07777895033359528,
0.017650140449404716,
-0.09577023237943649,
-0.09159492701292038,
0.059776052832603455,
0.10890021175146103,
0.014759724959731102,
-0.00850776955485344,
0.08850563317537308,
-0.03501661866903305,
0.014902367256581783,
0.08669138699769974,
0.005105787422508001,
0.15939538180828094,
0.04961811378598213,
0.1777653694152832,
0.08442499488592148,
0.07212188839912415,
0.0022480690386146307,
0.007672940380871296,
0.012716339901089668,
0.041776418685913086,
-0.00586339458823204,
-0.08367537707090378,
-0.025811415165662766,
0.11007143557071686,
0.06822896003723145,
0.01674085482954979,
0.013469134457409382,
-0.048928502947092056,
0.08716683089733124,
0.17909866571426392,
0.002402160782366991,
-0.18131543695926666,
-0.05789600685238838,
0.0750206708908081,
-0.09928116202354431,
-0.10272461175918579,
-0.008900126442313194,
0.01586969569325447,
-0.16637296974658966,
0.0353267602622509,
-0.020183570683002472,
0.10838212817907333,
-0.13442763686180115,
-0.017363429069519043,
0.07578733563423157,
0.0699351504445076,
-0.0022181386593729258,
0.05766081437468529,
-0.17974431812763214,
0.09947559982538223,
0.012100325897336006,
0.0704662874341011,
-0.09766525775194168,
0.09176629781723022,
-0.009300199337303638,
-0.030800525099039078,
0.1424117237329483,
-0.004621058702468872,
-0.0713675245642662,
-0.06180211901664734,
-0.09473362565040588,
-0.011214162223041058,
0.1266127973794937,
-0.1298455148935318,
0.09194545447826385,
-0.03333241492509842,
-0.03638272359967232,
-0.010627835057675838,
-0.08796215802431107,
-0.10934466868638992,
-0.17969419062137604,
0.0595061257481575,
-0.12724877893924713,
0.03739791736006737,
-0.1058867797255516,
-0.025679782032966614,
-0.025718556717038155,
0.1798698753118515,
-0.24040640890598297,
-0.07245399057865143,
-0.14371363818645477,
-0.09313122928142548,
0.13163676857948303,
-0.04657937213778496,
0.09096090495586395,
-0.016109757125377655,
0.16015022993087769,
0.02005946636199951,
-0.019326770678162575,
0.08680126816034317,
-0.08495312184095383,
-0.19593647122383118,
-0.07017680257558823,
0.16592541337013245,
0.11997250467538834,
0.03374743461608887,
0.0002775360771920532,
0.037539996206760406,
-0.020964371040463448,
-0.1182674989104271,
0.021811965852975845,
0.15264973044395447,
0.06795535981655121,
0.009478275664150715,
-0.02395930327475071,
-0.1102604866027832,
-0.0759085938334465,
-0.028979400172829628,
0.03204840421676636,
0.17042486369609833,
-0.07147930562496185,
0.17116157710552216,
0.14203238487243652,
-0.05904092639684677,
-0.20811264216899872,
-0.0021387117449194193,
0.026254096999764442,
-0.00908663496375084,
0.010671177878975868,
-0.18749478459358215,
0.08536157757043839,
-0.002555366139858961,
-0.054554641246795654,
0.1051621288061142,
-0.162221297621727,
-0.13799940049648285,
0.08243122696876526,
0.050192203372716904,
-0.187874898314476,
-0.13649384677410126,
-0.0961209312081337,
-0.0409083291888237,
-0.16012360155582428,
0.09442240744829178,
0.02092057839035988,
0.01192572433501482,
0.031539436429739,
0.014813697896897793,
0.024217691272497177,
-0.04841739311814308,
0.1757010519504547,
-0.018160223960876465,
0.021848153322935104,
-0.09544609487056732,
-0.08094970881938934,
0.017702657729387283,
-0.05052844062447548,
0.07107838243246078,
-0.01870221272110939,
0.011255724355578423,
-0.10348273068666458,
-0.036123502999544144,
-0.042589638382196426,
0.015495449304580688,
-0.09958311915397644,
-0.08601320534944534,
-0.047096606343984604,
0.09428968280553818,
0.09659279137849808,
-0.022826362401247025,
-0.0273736622184515,
-0.07949339598417282,
0.054964661598205566,
0.20859195291996002,
0.18808089196681976,
0.04425831884145737,
-0.062306828796863556,
-0.001769690657965839,
-0.014944184571504593,
0.0417049303650856,
-0.19710716605186462,
0.05912362411618233,
0.0562397725880146,
0.02090633660554886,
0.10401801019906998,
-0.019544212147593498,
-0.15806327760219574,
-0.07712605595588684,
0.06873630732297897,
-0.0639798492193222,
-0.2012440711259842,
0.010393884032964706,
0.05963889881968498,
-0.1750810444355011,
-0.03916953504085541,
0.046524349600076675,
-0.002354544820263982,
-0.039746832102537155,
0.023544801399111748,
0.09521713852882385,
0.004076723475009203,
0.0779079720377922,
0.07060851901769638,
0.08211179822683334,
-0.09903982281684875,
0.08303046971559525,
0.0972437709569931,
-0.07399678975343704,
0.029093235731124878,
0.10163311660289764,
-0.05663156509399414,
-0.0387692004442215,
0.03549791872501373,
0.08046255260705948,
0.02705121599137783,
-0.043615687638521194,
0.011865747161209583,
-0.09553902596235275,
0.06658727675676346,
0.10204542428255081,
0.029401643201708794,
0.017935337498784065,
0.044615767896175385,
0.04629816859960556,
-0.0760611966252327,
0.12436296790838242,
0.03220634162425995,
0.015023278072476387,
-0.04068984091281891,
-0.0437593087553978,
0.0095688970759511,
-0.031515296548604965,
-0.004716753493994474,
-0.02209470048546791,
-0.08724912256002426,
-0.015015563927590847,
-0.13236366212368011,
-0.008743971586227417,
-0.06157073378562927,
0.012826182879507542,
0.028436195105314255,
-0.031039604917168617,
0.0064733498729765415,
0.004978197161108255,
-0.0703253373503685,
-0.06705103069543839,
-0.014137581922113895,
0.09520672261714935,
-0.16576512157917023,
0.026662830263376236,
0.08306188136339188,
-0.11256496608257294,
0.10015136748552322,
0.01094723865389824,
-0.006087163463234901,
0.022484684363007545,
-0.1469746232032776,
0.036669302731752396,
-0.03779744729399681,
0.008942550048232079,
0.02391820214688778,
-0.20087900757789612,
0.0007856925949454308,
-0.033994678407907486,
-0.0705132707953453,
-0.009060341864824295,
-0.025304662063717842,
-0.11176600307226181,
0.10632134974002838,
0.0007318496936932206,
-0.08119626343250275,
-0.02982376329600811,
0.03208870068192482,
0.07710310071706772,
-0.028534123674035072,
0.15100301802158356,
-0.012797199189662933,
0.06572870165109634,
-0.1587447077035904,
-0.01154392957687378,
-0.0113789988681674,
0.015301055274903774,
-0.03572189435362816,
-0.007380445022135973,
0.050975777208805084,
-0.01390343438833952,
0.17424729466438293,
-0.03530619665980339,
0.01664605177938938,
0.0663379579782486,
0.04406430572271347,
-0.03276953473687172,
0.09988830238580704,
0.04999752715229988,
0.017141716554760933,
0.009620473720133305,
0.013025123625993729,
-0.04225568845868111,
-0.036079514771699905,
-0.19024775922298431,
0.06989775598049164,
0.1866932511329651,
0.09840615093708038,
-0.020935317501425743,
0.07415685057640076,
-0.10040472447872162,
-0.09374922513961792,
0.15041497349739075,
-0.037143923342227936,
-0.006523944437503815,
-0.07367441058158875,
0.13029973208904266,
0.14549610018730164,
-0.1814785897731781,
0.06606297194957733,
-0.07180570065975189,
-0.04206360504031181,
-0.10935494303703308,
-0.19550631940364838,
-0.06178471818566322,
-0.048242729157209396,
-0.01886189728975296,
-0.0473010279238224,
0.06674393266439438,
0.060638632625341415,
0.0006839959532953799,
-0.008536890149116516,
0.06868739426136017,
-0.03323382884263992,
-0.0032462377566844225,
0.028674080967903137,
0.0611579604446888,
0.00929503608494997,
-0.03787853941321373,
0.017562363296747208,
-0.012991433031857014,
0.0538424551486969,
0.07582131028175354,
0.050257325172424316,
-0.027161642909049988,
0.01914774253964424,
-0.040576957166194916,
-0.10661055147647858,
0.04927254468202591,
-0.0285146776586771,
-0.07324735075235367,
0.15458324551582336,
0.020231405273079872,
0.0046523697674274445,
-0.012391991913318634,
0.23933540284633636,
-0.06322796642780304,
-0.1032072901725769,
-0.1446000337600708,
0.0738082081079483,
-0.04255184531211853,
0.04970810189843178,
0.03866847604513168,
-0.11122773587703705,
0.025612516328692436,
0.1497742384672165,
0.15382428467273712,
-0.041335951536893845,
0.022738350555300713,
0.03465048223733902,
0.008969487622380257,
-0.023746641352772713,
0.0378718264400959,
0.06464387476444244,
0.15299609303474426,
-0.047938305884599686,
0.07862778007984161,
0.0008240414317697287,
-0.08734137564897537,
-0.03727345913648605,
0.11393242329359055,
-0.011411378160119057,
0.014004142954945564,
-0.058753952383995056,
0.11869291216135025,
-0.0715123638510704,
-0.21872852742671967,
0.04109719768166542,
-0.07026688009500504,
-0.1331670731306076,
-0.023679273203015327,
0.07693633437156677,
-0.012184051796793938,
0.02233847789466381,
0.07813236862421036,
-0.07187089323997498,
0.18990160524845123,
0.039505671709775925,
-0.05827925726771355,
-0.05084679275751114,
0.07384473830461502,
-0.07778072357177734,
0.29490548372268677,
0.015166562050580978,
0.04005192592740059,
0.11093740910291672,
-0.0153964227065444,
-0.1418374478816986,
0.023329483345150948,
0.09676562994718552,
-0.09851957112550735,
0.052737001329660416,
0.17991116642951965,
0.0022530595306307077,
0.1274615228176117,
0.07713494449853897,
-0.08902951329946518,
0.04603179916739464,
-0.07316387444734573,
-0.07014941424131393,
-0.09849017858505249,
0.10400687903165817,
-0.08915241807699203,
0.1447027027606964,
0.121854268014431,
-0.056043028831481934,
0.011541114188730717,
-0.03429169952869415,
0.046591032296419144,
-0.003980774898082018,
0.12048541754484177,
0.010917561128735542,
-0.19038154184818268,
0.02617969550192356,
-0.02745375595986843,
0.10133756697177887,
-0.16829264163970947,
-0.08677268028259277,
0.04525073990225792,
0.01023704931139946,
-0.07288476079702377,
0.1258174479007721,
0.05900004133582115,
0.029306763783097267,
-0.04840981587767601,
-0.022562798112630844,
-0.010515614412724972,
0.14134228229522705,
-0.10410626232624054,
-0.004606697708368301
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | YashRawal225/New-3-7b-chat-finetune-german500-GGUF | [
"transformers",
"gguf",
"mistral",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T08:57:46+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #gguf #mistral #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #gguf #mistral #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
54,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #gguf #mistral #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06043652072548866,
0.1364833414554596,
-0.004403033759444952,
0.025301463901996613,
0.10390770435333252,
0.012222958728671074,
0.06422175467014313,
0.11201353371143341,
-0.02215377427637577,
0.11453705281019211,
0.0299602709710598,
0.09424004703760147,
0.10989700257778168,
0.1616695374250412,
0.0023904612753540277,
-0.2208220511674881,
0.03988941013813019,
-0.1240171417593956,
-0.02984563261270523,
0.11872946470975876,
0.1390196830034256,
-0.10718946158885956,
0.07221073657274246,
-0.03836611285805702,
-0.016214806586503983,
-0.04039283096790314,
-0.0650324672460556,
-0.0473860502243042,
0.057339537888765335,
0.07783833146095276,
0.06201384216547012,
0.011981473304331303,
0.0917908176779747,
-0.2713852822780609,
0.022272784262895584,
0.08073437213897705,
-0.00612818356603384,
0.07390498369932175,
0.047688182443380356,
-0.07891020178794861,
0.08520569652318954,
-0.05711792781949043,
0.14042367041110992,
0.08162671327590942,
-0.09282717108726501,
-0.18776856362819672,
-0.08457721024751663,
0.09580861777067184,
0.1898847371339798,
0.0587463453412056,
-0.028977539390325546,
0.1119241937994957,
-0.08120796829462051,
0.012070165015757084,
0.06472708284854889,
-0.05421328544616699,
-0.05492231622338295,
0.08680249750614166,
0.07574701309204102,
0.09384738653898239,
-0.13446687161922455,
-0.004772716201841831,
0.018588203936815262,
0.01262358482927084,
0.0950685366988182,
0.01657790318131447,
0.13131365180015564,
0.049048081040382385,
-0.140764519572258,
-0.04201579838991165,
0.09974140673875809,
0.03831565007567406,
-0.04992787167429924,
-0.24278609454631805,
-0.015016105957329273,
-0.039592236280441284,
-0.03327985107898712,
-0.04015984386205673,
0.04598761349916458,
-0.012547816149890423,
0.07089859247207642,
-0.01654037833213806,
-0.08055675029754639,
-0.04829836264252663,
0.08019156008958817,
0.06698194146156311,
0.023209763690829277,
-0.02230772003531456,
0.0047767129726707935,
0.12692083418369293,
0.08691585063934326,
-0.12325932830572128,
-0.0572965107858181,
-0.055576391518116,
-0.07894167304039001,
-0.0505097396671772,
0.025057127699255943,
0.031804412603378296,
0.04710208252072334,
0.23212438821792603,
-0.0007881721830926836,
0.051500190049409866,
0.04625837877392769,
0.01615966483950615,
0.06657599657773972,
0.10544904321432114,
-0.06659606099128723,
-0.10459265857934952,
-0.022843411192297935,
0.1002625972032547,
0.004284040071070194,
-0.03949115797877312,
-0.06561244279146194,
0.06807776540517807,
0.01189518067985773,
0.11644040048122406,
0.07568138092756271,
0.005296717863529921,
-0.07717901468276978,
-0.06888340413570404,
0.18166229128837585,
-0.1509106457233429,
0.04143797978758812,
0.024866178631782532,
-0.04749131575226784,
-0.024455977603793144,
0.014581927098333836,
0.03290419280529022,
-0.011321557685732841,
0.09354069083929062,
-0.055512748658657074,
-0.02506476268172264,
-0.11226438730955124,
-0.03890789672732353,
0.029873467981815338,
0.015495078638195992,
-0.028310107067227364,
-0.03058359958231449,
-0.09655560553073883,
-0.06690455973148346,
0.09398456662893295,
-0.0757417306303978,
-0.04778528958559036,
-0.02608492784202099,
-0.0799868032336235,
0.01741219311952591,
0.017633600160479546,
0.1161181777715683,
-0.020083341747522354,
0.045540522783994675,
-0.04894882068037987,
0.06232153996825218,
0.11703310906887054,
0.027779873460531235,
-0.051331304013729095,
0.050766848027706146,
-0.2468864619731903,
0.11047130823135376,
-0.0699729397892952,
0.04586569964885712,
-0.15732333064079285,
-0.02320648357272148,
0.03739454969763756,
0.013470029458403587,
-0.004970688838511705,
0.1318424642086029,
-0.2107630968093872,
-0.035862501710653305,
0.16487571597099304,
-0.0941334068775177,
-0.07838058471679688,
0.0592513307929039,
-0.05803397670388222,
0.11582395434379578,
0.0453321598470211,
-0.01968478038907051,
0.0389326736330986,
-0.13793522119522095,
-0.011286995373666286,
-0.0492904894053936,
-0.026583466678857803,
0.16451646387577057,
0.06756879389286041,
-0.05376594513654709,
0.06846120953559875,
0.01910865493118763,
-0.022263716906309128,
-0.04327300190925598,
-0.02934308350086212,
-0.1046091839671135,
0.008644867688417435,
-0.0626973956823349,
0.023895632475614548,
-0.02701975591480732,
-0.09059704840183258,
-0.026224544271826744,
-0.17026758193969727,
0.009946096688508987,
0.08623059093952179,
-0.005379438865929842,
-0.022709479555487633,
-0.10147675126791,
-0.005692289676517248,
0.02261645346879959,
0.0048945629969239235,
-0.1397348940372467,
-0.05365082249045372,
0.022965101525187492,
-0.15685193240642548,
0.04306052252650261,
-0.045327503234148026,
0.046764008700847626,
0.039786938577890396,
-0.04747619107365608,
-0.030519988387823105,
0.009937143884599209,
0.007571724709123373,
-0.01955103687942028,
-0.2650749087333679,
-0.016965730115771294,
-0.027991242706775665,
0.17468513548374176,
-0.2410573661327362,
0.04456106573343277,
0.06975222378969193,
0.13544012606143951,
0.003451697062700987,
-0.03641485422849655,
0.020228935405611992,
-0.07635372132062912,
-0.03188900277018547,
-0.06797868758440018,
-0.013225452974438667,
-0.033293288201093674,
-0.06513632833957672,
0.034783974289894104,
-0.15517988801002502,
-0.04300735145807266,
0.10524103790521622,
0.05128473415970802,
-0.1539568156003952,
-0.03992617130279541,
-0.04174779728055,
-0.05412213131785393,
-0.06827438622713089,
-0.048278987407684326,
0.13131935894489288,
0.04810507595539093,
0.0493318997323513,
-0.06775516271591187,
-0.07868152111768723,
0.009349185042083263,
-0.02292558178305626,
-0.012239581905305386,
0.08462435007095337,
0.07055048644542694,
-0.1269891858100891,
0.09356658905744553,
0.08098162710666656,
0.07788645476102829,
0.10919603705406189,
-0.0015403409488499165,
-0.0832747220993042,
-0.042522359639406204,
0.02857637219130993,
0.011683383025228977,
0.15302543342113495,
-0.0294245183467865,
0.05255543813109398,
0.036057356745004654,
-0.008859539404511452,
0.011850828304886818,
-0.10144154727458954,
0.02902897819876671,
0.027241898700594902,
-0.01878364384174347,
0.036192089319229126,
-0.05132675915956497,
0.007138434331864119,
0.09350678324699402,
0.045503199100494385,
0.05535358935594559,
0.012586526572704315,
-0.041250936686992645,
-0.11177234351634979,
0.17973937094211578,
-0.11823903769254684,
-0.24363362789154053,
-0.12793584167957306,
0.0007059029303491116,
0.03348066657781601,
-0.006928946822881699,
0.01571100950241089,
-0.07546469569206238,
-0.11098508536815643,
-0.08902988582849503,
0.032033950090408325,
0.03961782902479172,
-0.08039093017578125,
-0.07527642697095871,
0.0650135725736618,
0.038120221346616745,
-0.13707153499126434,
0.026021720841526985,
0.040351502597332,
-0.08186745643615723,
0.005527558270841837,
0.08146802335977554,
0.05672338604927063,
0.18210339546203613,
0.015382747165858746,
-0.020216533914208412,
0.023511972278356552,
0.2066899985074997,
-0.14384369552135468,
0.11113261431455612,
0.12521137297153473,
-0.06881777942180634,
0.08286892622709274,
0.2035301774740219,
0.04323985427618027,
-0.10859351605176926,
0.0386701263487339,
0.030059516429901123,
-0.03070959448814392,
-0.24403400719165802,
-0.0715213492512703,
0.002208442660048604,
-0.07089735567569733,
0.06957205384969711,
0.08384032547473907,
0.1109042540192604,
0.019877837970852852,
-0.10532350838184357,
-0.05141087993979454,
0.04708562418818474,
0.11190445721149445,
-0.008227919228374958,
-0.014682337641716003,
0.09348871558904648,
-0.019662166014313698,
0.023097241297364235,
0.09417568892240524,
0.0028958276379853487,
0.19007478654384613,
0.04980888590216637,
0.16618570685386658,
0.08220270276069641,
0.05510367825627327,
0.019585950300097466,
-0.006269503850489855,
0.013085569255053997,
0.014382525347173214,
-0.009213880635797977,
-0.08658449351787567,
0.0041406783275306225,
0.12387280911207199,
0.041406914591789246,
0.03364110738039017,
0.01692693866789341,
-0.0504080131649971,
0.08317570388317108,
0.16685259342193604,
0.004088270012289286,
-0.1873372346162796,
-0.07739754766225815,
0.07311257719993591,
-0.07838805764913559,
-0.10809872299432755,
-0.04073942080140114,
0.03965435177087784,
-0.17545931041240692,
0.012425045482814312,
-0.021709609776735306,
0.10526955127716064,
-0.12251711636781693,
-0.017526261508464813,
0.05529288947582245,
0.08159451931715012,
-0.01358067523688078,
0.07315830886363983,
-0.1763472855091095,
0.12572243809700012,
0.021667633205652237,
0.06820559501647949,
-0.09648541361093521,
0.09181233495473862,
-0.0006925527704879642,
0.00410801824182272,
0.13389833271503448,
0.012742117047309875,
-0.05217821151018143,
-0.10189642012119293,
-0.09275900572538376,
-0.010364187881350517,
0.13251928985118866,
-0.14939773082733154,
0.09157802164554596,
-0.01644320785999298,
-0.043866630643606186,
0.005519861355423927,
-0.12461551278829575,
-0.13678978383541107,
-0.18108223378658295,
0.06213892996311188,
-0.12908536195755005,
0.04339681938290596,
-0.103909432888031,
-0.04458380490541458,
-0.020100725814700127,
0.2116032987833023,
-0.22369344532489777,
-0.07021985948085785,
-0.15435709059238434,
-0.05292452126741409,
0.13897038996219635,
-0.045630041509866714,
0.08443552255630493,
0.003589295083656907,
0.18610624969005585,
0.020096633583307266,
-0.010717025958001614,
0.11111374944448471,
-0.1012648418545723,
-0.1989838182926178,
-0.0967506542801857,
0.153251051902771,
0.1476786732673645,
0.04360144957900047,
-0.0050810580141842365,
0.033771052956581116,
-0.024940066039562225,
-0.12379155308008194,
0.02163020707666874,
0.16985459625720978,
0.10638262331485748,
0.031244108453392982,
-0.03019586019217968,
-0.10422205179929733,
-0.07228221744298935,
-0.04733051732182503,
0.0162336602807045,
0.1811097413301468,
-0.07009261101484299,
0.18402846157550812,
0.14696069061756134,
-0.05599227547645569,
-0.18902993202209473,
0.02689802646636963,
0.04877376928925514,
0.006599194835871458,
0.030709972605109215,
-0.2184126079082489,
0.08919534832239151,
-0.001161001855507493,
-0.05516020953655243,
0.1403244435787201,
-0.18534722924232483,
-0.14838449656963348,
0.048710741102695465,
0.0478043258190155,
-0.18395769596099854,
-0.12266750633716583,
-0.09677054733037949,
-0.04542998969554901,
-0.1381692737340927,
0.09247630089521408,
0.0033229643013328314,
0.017334748059511185,
0.032229889184236526,
0.029868371784687042,
0.008680860511958599,
-0.04173145070672035,
0.1908217966556549,
-0.011501830071210861,
0.03444297984242439,
-0.07767865061759949,
-0.05841992795467377,
0.05982344225049019,
-0.06426037847995758,
0.07752633094787598,
-0.024919483810663223,
0.015127211809158325,
-0.10224231332540512,
-0.05372263118624687,
-0.02883896231651306,
0.025868847966194153,
-0.08586836606264114,
-0.09965287148952484,
-0.05096714198589325,
0.09296621382236481,
0.08098159730434418,
-0.0355696864426136,
-0.05194370076060295,
-0.0778568908572197,
0.035892367362976074,
0.1722700297832489,
0.18028423190116882,
0.03782324865460396,
-0.07421562075614929,
-0.004091201350092888,
-0.010291436687111855,
0.0496017225086689,
-0.23521865904331207,
0.05318988487124443,
0.03963218629360199,
0.030428865924477577,
0.12168171256780624,
-0.023127460852265358,
-0.16378004848957062,
-0.06184251978993416,
0.05630405247211456,
-0.06704883277416229,
-0.17083445191383362,
-0.000630140770226717,
0.07737462967634201,
-0.16328275203704834,
-0.0310327410697937,
0.04110848158597946,
-0.015949146822094917,
-0.04062995687127113,
0.011260150000452995,
0.07768364250659943,
0.006894576828926802,
0.08359560370445251,
0.05484339967370033,
0.09231801331043243,
-0.10504472255706787,
0.06287282705307007,
0.08169671893119812,
-0.08033277094364166,
0.03528609871864319,
0.08323246985673904,
-0.06819913536310196,
-0.039311472326517105,
0.04117212072014809,
0.08421476185321808,
0.020080450922250748,
-0.0398726649582386,
0.011065002530813217,
-0.08928437530994415,
0.0565703809261322,
0.10340022295713425,
0.04198940098285675,
0.007326774764806032,
0.030771195888519287,
0.0505385547876358,
-0.08703406155109406,
0.11335117369890213,
0.028310080990195274,
0.0251682810485363,
-0.035952337086200714,
-0.039654262363910675,
0.04045959934592247,
-0.02565566822886467,
-0.013029326684772968,
-0.03820253536105156,
-0.07435781508684158,
-0.01643351837992668,
-0.15688352286815643,
-0.007759492378681898,
-0.0316164493560791,
0.006815608125180006,
0.01821567304432392,
-0.04114970564842224,
0.010930243879556656,
0.018521344289183617,
-0.07067343592643738,
-0.05803371220827103,
-0.02059350721538067,
0.0964127853512764,
-0.16133065521717072,
0.019881660118699074,
0.08061572909355164,
-0.1179603785276413,
0.09199973940849304,
0.02707931399345398,
0.012555565685033798,
0.03609583154320717,
-0.15143649280071259,
0.05112122744321823,
-0.017864592373371124,
0.008806884288787842,
0.042952802032232285,
-0.2253558486700058,
0.00000970917972153984,
-0.050889916718006134,
-0.04614536464214325,
-0.007651264313608408,
-0.03057464025914669,
-0.1163935586810112,
0.09113796800374985,
-0.00620719138532877,
-0.07838267087936401,
-0.032173555344343185,
0.03994172066450119,
0.07846645265817642,
-0.023159310221672058,
0.144456148147583,
-0.0032758440356701612,
0.06922536343336105,
-0.1830170899629593,
-0.02126016654074192,
-0.013005703687667847,
0.02154187299311161,
-0.031274329870939255,
-0.018267685547471046,
0.04769599810242653,
-0.02359769120812416,
0.19292710721492767,
-0.020377924665808678,
0.05038834363222122,
0.058175038546323776,
0.014238815754652023,
-0.009736555628478527,
0.1028982549905777,
0.05889854207634926,
0.022524449974298477,
0.024915989488363266,
0.005529727321118116,
-0.02605004608631134,
-0.006936070509254932,
-0.18414975702762604,
0.064814992249012,
0.14198492467403412,
0.08173108100891113,
-0.015209355391561985,
0.060035038739442825,
-0.11191552877426147,
-0.11726744472980499,
0.12110288441181183,
-0.05822555348277092,
-0.018610043451189995,
-0.06386368721723557,
0.15004152059555054,
0.1499270498752594,
-0.19568637013435364,
0.06749383360147476,
-0.06039800867438316,
-0.05739985778927803,
-0.10258513689041138,
-0.19043412804603577,
-0.056816812604665756,
-0.05827757716178894,
-0.018971892073750496,
-0.05394699424505234,
0.07371001690626144,
0.06399813294410706,
0.011189207434654236,
0.013231812976300716,
0.08207091689109802,
-0.005056121852248907,
0.0017720245523378253,
0.026677938178181648,
0.06357316672801971,
0.009438545443117619,
-0.0345911867916584,
0.019056543707847595,
0.004688593093305826,
0.028754517436027527,
0.05849899351596832,
0.032911207526922226,
-0.02578773722052574,
0.008587337099015713,
-0.02981565333902836,
-0.11192777007818222,
0.037752166390419006,
-0.02589316852390766,
-0.06693420559167862,
0.13451255857944489,
0.022167084738612175,
-0.0014553592773154378,
-0.024997936561703682,
0.25857025384902954,
-0.07144878059625626,
-0.09280755370855331,
-0.1377401500940323,
0.12671497464179993,
-0.039189986884593964,
0.05589829757809639,
0.02988755702972412,
-0.10922381281852722,
0.026033977046608925,
0.13058888912200928,
0.14655712246894836,
-0.04922638460993767,
0.014898952096700668,
0.03076462261378765,
0.002008798299357295,
-0.032078176736831665,
0.04228813946247101,
0.07166885584592819,
0.13631084561347961,
-0.05662459135055542,
0.08064515143632889,
0.002992721274495125,
-0.10224848985671997,
-0.0341300405561924,
0.1151568740606308,
-0.007688653655350208,
0.01520447712391615,
-0.05891072750091553,
0.12546302378177643,
-0.04415013641119003,
-0.2500936985015869,
0.06200886517763138,
-0.07310980558395386,
-0.14133808016777039,
-0.023708723485469818,
0.057564280927181244,
-0.01629544422030449,
0.029828494414687157,
0.06932152062654495,
-0.07558722048997879,
0.19148772954940796,
0.03625214099884033,
-0.06222386658191681,
-0.06124694272875786,
0.07273191213607788,
-0.1111014112830162,
0.27837511897087097,
0.0041924468241631985,
0.061887096613645554,
0.10492724925279617,
-0.02171275019645691,
-0.13117121160030365,
0.027804624289274216,
0.08286241441965103,
-0.08334644138813019,
0.047949112951755524,
0.22766880691051483,
-0.007333519868552685,
0.10479136556386948,
0.07446389645338058,
-0.08290313929319382,
0.05003875866532326,
-0.10806924849748611,
-0.07840195298194885,
-0.08505915105342865,
0.09337235987186432,
-0.06431708484888077,
0.14627376198768616,
0.12489324808120728,
-0.050261951982975006,
0.009868069551885128,
-0.032556887716054916,
0.049779262393713,
0.012486230581998825,
0.1209091991186142,
0.008578131906688213,
-0.18858198821544647,
0.026575982570648193,
0.011860244907438755,
0.09852910786867142,
-0.19759181141853333,
-0.09519941359758377,
0.05197626352310181,
0.006810850463807583,
-0.05416284501552582,
0.1224554181098938,
0.05903016775846481,
0.04608935862779617,
-0.04469767585396767,
-0.029989657923579216,
-0.010168750770390034,
0.14482511579990387,
-0.10945922136306763,
-0.00364923314191401
] |
null | null | null |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
This modelcard aims to be a base template for new models. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1).
## Model Details
### Model Description
<!-- Gangster with guns in hand -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] | {} | null | issuez/ishan | [
"arxiv:1910.09700",
"region:us"
] | 2024-02-08T09:00:23+00:00 | [
"1910.09700"
] | [] | TAGS
#arxiv-1910.09700 #region-us
|
# Model Card for Model ID
This modelcard aims to be a base template for new models. It has been generated using this raw template.
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID\n\n\n\nThis modelcard aims to be a base template for new models. It has been generated using this raw template.",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#arxiv-1910.09700 #region-us \n",
"# Model Card for Model ID\n\n\n\nThis modelcard aims to be a base template for new models. It has been generated using this raw template.",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
15,
29,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#arxiv-1910.09700 #region-us \n# Model Card for Model ID\n\n\n\nThis modelcard aims to be a base template for new models. It has been generated using this raw template.## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.1066984087228775,
0.19898438453674316,
-0.002620849059894681,
0.027911467477679253,
0.09412756562232971,
0.02142420969903469,
0.05197415128350258,
0.12995286285877228,
-0.022686492651700974,
0.09772004932165146,
0.07303693890571594,
0.09985987842082977,
0.11060800403356552,
0.19985371828079224,
0.022886212915182114,
-0.19676423072814941,
0.0380873903632164,
-0.07859895378351212,
-0.0053507364355027676,
0.12146519124507904,
0.14281919598579407,
-0.09727081656455994,
0.09723988175392151,
-0.0014166681794449687,
-0.036095861345529556,
-0.032103247940540314,
-0.07407337427139282,
-0.015863366425037384,
0.04475326091051102,
0.04351950064301491,
0.06786411255598068,
-0.005140793044120073,
0.08499236404895782,
-0.25888389348983765,
0.01854773797094822,
0.04429004341363907,
-0.010532835498452187,
0.08963978290557861,
0.08659035712480545,
-0.0503561794757843,
0.133544921875,
-0.022494465112686157,
0.13020861148834229,
0.09404259920120239,
-0.09533175081014633,
-0.22292618453502655,
-0.06270451098680496,
0.08238474279642105,
0.17174527049064636,
0.08238210529088974,
-0.04212580993771553,
0.11285244673490524,
-0.08852370828390121,
0.009028825908899307,
0.026999270543456078,
-0.06745808571577072,
-0.0541369654238224,
0.06993507593870163,
0.10711865872144699,
0.058814842253923416,
-0.11843946576118469,
-0.02349478006362915,
0.02842799760401249,
0.03442508354783058,
0.06341962516307831,
0.009891163557767868,
0.16722379624843597,
0.02796054631471634,
-0.14623764157295227,
-0.045774128288030624,
0.14920903742313385,
0.03012506291270256,
-0.04215821996331215,
-0.20709463953971863,
-0.006747975014150143,
-0.08887778222560883,
-0.02198212593793869,
-0.04837455600500107,
0.049169208854436874,
0.017894232645630836,
0.1088389903306961,
-0.04266509786248207,
-0.09933976829051971,
-0.01135216560214758,
0.09378619492053986,
0.0346422903239727,
0.014278990216553211,
-0.007528449408710003,
-0.0004577430372592062,
0.1285746544599533,
0.05469144508242607,
-0.12714460492134094,
-0.06107940524816513,
-0.07115962356328964,
-0.038405947387218475,
-0.0369131825864315,
0.02906898967921734,
0.04100806638598442,
0.043637074530124664,
0.256840318441391,
-0.004247268196195364,
0.055150821805000305,
0.07816781103610992,
0.029096456244587898,
0.05928764492273331,
0.1027042493224144,
-0.05899541452527046,
-0.16195039451122284,
-0.010378924198448658,
0.08282309025526047,
-0.001994097838178277,
-0.03426366299390793,
-0.07769469916820526,
0.04409316927194595,
0.03231460228562355,
0.10194675624370575,
0.10213643312454224,
-0.00840142834931612,
-0.06900389492511749,
-0.06323889642953873,
0.2089644968509674,
-0.140780970454216,
0.04491811990737915,
0.014242668636143208,
-0.02261270396411419,
-0.03249955177307129,
0.01572984829545021,
0.02586018666625023,
-0.03010505624115467,
0.0799839049577713,
-0.07839398831129074,
-0.03527021035552025,
-0.12673307955265045,
-0.027471506968140602,
0.022427299991250038,
-0.003538058837875724,
-0.020284080877900124,
-0.028788061812520027,
-0.07818468660116196,
-0.09368987381458282,
0.11955387890338898,
-0.06332498788833618,
-0.04999423027038574,
-0.03507404401898384,
-0.08191268146038055,
0.029741330072283745,
0.037989094853401184,
0.09118469059467316,
-0.02362010069191456,
0.030886046588420868,
-0.011452100239694118,
0.06288884580135345,
0.05513912811875343,
0.03946930542588234,
-0.08444610983133316,
0.06066261604428291,
-0.20766597986221313,
0.09049645811319351,
-0.061090268194675446,
0.035395506769418716,
-0.16141952574253082,
-0.007028630934655666,
0.010053620673716068,
0.03627076745033264,
0.029977615922689438,
0.15956375002861023,
-0.21718010306358337,
-0.033392831683158875,
0.13398465514183044,
-0.10535142570734024,
-0.10920390486717224,
0.03313661739230156,
-0.05438750982284546,
0.1846301406621933,
0.021011749282479286,
-0.0004254789964761585,
0.07448633760213852,
-0.12248582392930984,
-0.023236358538269997,
-0.017138930037617683,
-0.02830614522099495,
0.0713268369436264,
0.081262968480587,
-0.09127437323331833,
0.018645839765667915,
0.012333624064922333,
-0.04975755140185356,
-0.027676379308104515,
-0.039810311049222946,
-0.1081002727150917,
-0.0050798640586435795,
-0.07177256047725677,
0.0032630744390189648,
-0.016572296619415283,
-0.08035643398761749,
0.001617362373508513,
-0.16802436113357544,
-0.02455195039510727,
0.07920140773057938,
0.00814704317599535,
-0.014599336311221123,
-0.09017311781644821,
0.05735816806554794,
-0.0606040433049202,
-0.027315329760313034,
-0.14675214886665344,
0.0052270544692873955,
0.013336344622075558,
-0.14710111916065216,
0.020299475640058517,
-0.10115809738636017,
0.06349264085292816,
0.011477353982627392,
-0.04672861844301224,
-0.04369935020804405,
0.00010367027425672859,
0.003724793205037713,
-0.053453478962183,
-0.23140744864940643,
-0.03374785929918289,
-0.044856321066617966,
0.15386763215065002,
-0.21619822084903717,
0.036581382155418396,
0.04023589566349983,
0.11894483119249344,
-0.0035392972640693188,
-0.05845631659030914,
0.02484537661075592,
-0.07670248299837112,
-0.039557602256536484,
-0.07007710635662079,
0.001572409993968904,
-0.0014461677055805922,
-0.04872008040547371,
0.016424696892499924,
-0.12396717816591263,
-0.06818132847547531,
0.11014950275421143,
0.04142379015684128,
-0.15492349863052368,
-0.0041915783658623695,
-0.030916975811123848,
-0.06000775843858719,
-0.05342598259449005,
-0.05972793325781822,
0.11303042620420456,
0.04413822293281555,
0.03973376750946045,
-0.07595038414001465,
-0.05902018025517464,
0.010918207466602325,
-0.029565786942839622,
-0.016297759488224983,
0.093429334461689,
0.0999373197555542,
-0.12134627997875214,
0.0989208072423935,
0.07267444580793381,
0.03332529962062836,
0.08463598042726517,
-0.010384823195636272,
-0.10775253921747208,
-0.031286682933568954,
0.028272075578570366,
0.002783637959510088,
0.16402263939380646,
-0.07852847874164581,
0.05485396459698677,
0.04151233285665512,
-0.02716772072017193,
0.05665498599410057,
-0.0957925021648407,
0.01761704683303833,
0.021760543808341026,
-0.005722802598029375,
0.007382235489785671,
-0.030695531517267227,
-0.00876180361956358,
0.07580762356519699,
0.06536837667226791,
0.03976568952202797,
0.033472709357738495,
-0.029367417097091675,
-0.13732430338859558,
0.1905110627412796,
-0.10305890440940857,
-0.22935202717781067,
-0.1717958003282547,
0.05177067220211029,
0.05311822518706322,
-0.006569376215338707,
0.025543777272105217,
-0.06189529225230217,
-0.10580533742904663,
-0.08138279616832733,
0.018486447632312775,
0.006763557903468609,
-0.06118743494153023,
-0.09133443981409073,
0.039310213178396225,
0.03854845091700554,
-0.130089670419693,
0.03713662177324295,
0.05565035715699196,
-0.013944907113909721,
-0.01169645506888628,
0.04666148126125336,
0.09532339870929718,
0.19625136256217957,
-0.007600754965096712,
-0.008414373733103275,
0.06695323437452316,
0.291735976934433,
-0.15341155230998993,
0.12901893258094788,
0.12389722466468811,
-0.07051796466112137,
0.08391714096069336,
0.18495109677314758,
0.03306480497121811,
-0.09837296605110168,
0.020791195333003998,
0.02582281269133091,
-0.026949184015393257,
-0.2523637115955353,
-0.05266418308019638,
-0.006489538121968508,
-0.11327216774225235,
0.0706535056233406,
0.08739753067493439,
0.09179038554430008,
0.052998367697000504,
-0.06112697720527649,
-0.09166146069765091,
-0.0003765109577216208,
0.11145967245101929,
-0.04029975086450577,
0.002805144991725683,
0.07836277037858963,
-0.040786001831293106,
0.013674819841980934,
0.09837955981492996,
0.005413474980741739,
0.16164012253284454,
0.06552805751562119,
0.13401569426059723,
0.08540262281894684,
0.07983675599098206,
0.011559084989130497,
0.0339510552585125,
0.006372304633259773,
0.017902348190546036,
0.009317909367382526,
-0.07689813524484634,
0.023254239931702614,
0.11834190040826797,
0.040373314172029495,
0.045214686542749405,
0.011671608313918114,
-0.039621613919734955,
0.03956446796655655,
0.1767490804195404,
0.016947781667113304,
-0.2155885547399521,
-0.0772833302617073,
0.06542522460222244,
-0.058402981609106064,
-0.1495116949081421,
-0.025624670088291168,
0.02235453948378563,
-0.1576213240623474,
0.0005415278719738126,
-0.028421247377991676,
0.10273677110671997,
-0.09609103202819824,
-0.04047273099422455,
0.08817856013774872,
0.0699915885925293,
-0.028443265706300735,
0.062181491404771805,
-0.17871366441249847,
0.12371177971363068,
0.03400380536913872,
0.07102521508932114,
-0.09190616011619568,
0.09926209598779678,
-0.005890274420380592,
0.013706923462450504,
0.1655038744211197,
0.015541122294962406,
-0.09426835179328918,
-0.0710771307349205,
-0.08823360502719879,
-0.013685347512364388,
0.09967034310102463,
-0.13288703560829163,
0.06852756440639496,
-0.019996264949440956,
-0.027522722259163857,
0.006831855047494173,
-0.08690175414085388,
-0.13149002194404602,
-0.18112291395664215,
0.05532965436577797,
-0.10246208310127258,
0.024886637926101685,
-0.07404468208551407,
-0.04842938482761383,
0.040009692311286926,
0.19972540438175201,
-0.21909521520137787,
-0.10020548850297928,
-0.15154099464416504,
-0.11518421769142151,
0.16108576953411102,
-0.04635784775018692,
0.09108468145132065,
-0.01019457820802927,
0.1620863974094391,
0.010745878331363201,
-0.02071799710392952,
0.1160653606057167,
-0.0854450985789299,
-0.1714930683374405,
-0.05915606766939163,
0.14893034100532532,
0.14446774125099182,
0.035233963280916214,
-0.01287093386054039,
0.031883079558610916,
-0.07141809165477753,
-0.11891574412584305,
0.03534413129091263,
0.13700434565544128,
0.06963939964771271,
-0.01262744888663292,
-0.03579355776309967,
-0.09220988303422928,
-0.0504169799387455,
-0.03974350169301033,
0.008700315840542316,
0.18124674260616302,
-0.07448253035545349,
0.15189899504184723,
0.13152532279491425,
-0.0723627433180809,
-0.20356547832489014,
0.06049336493015289,
0.0346653088927269,
0.02138940617442131,
0.01630406267940998,
-0.21584440767765045,
0.08776868134737015,
-0.006339477840811014,
-0.06874293833971024,
0.18010607361793518,
-0.17902927100658417,
-0.13895957171916962,
0.0988360270857811,
0.03516041859984398,
-0.1823381632566452,
-0.13705183565616608,
-0.09613028168678284,
-0.03228107467293739,
-0.1230158656835556,
0.05866828188300133,
0.026329705491662025,
0.015535218641161919,
0.021184591576457024,
0.029537182301282883,
0.019990645349025726,
-0.050314560532569885,
0.2066401094198227,
-0.012754418887197971,
0.013829488307237625,
-0.06200092285871506,
-0.10324833542108536,
0.04607655853033066,
-0.05281443893909454,
0.11618918925523758,
0.0008675124263390899,
0.0222539734095335,
-0.1703120321035385,
-0.034940678626298904,
-0.05094180256128311,
0.03240950033068657,
-0.0940355733036995,
-0.09862573444843292,
-0.04792311042547226,
0.0863310769200325,
0.09178805351257324,
-0.02642832137644291,
-0.0012948049698024988,
-0.10240978002548218,
0.04736471548676491,
0.19468940794467926,
0.19447913765907288,
0.056417278945446014,
-0.06639297306537628,
0.028046250343322754,
-0.03318989276885986,
0.0474521666765213,
-0.24178913235664368,
0.03477860614657402,
0.05343414843082428,
0.011909419670701027,
0.08445286750793457,
-0.003811764298006892,
-0.16544894874095917,
-0.0645582303404808,
0.08673491328954697,
-0.044566020369529724,
-0.1641440987586975,
-0.032721146941185,
0.022641237825155258,
-0.20684140920639038,
-0.04179441183805466,
0.011281585320830345,
-0.019901549443602562,
-0.0412454716861248,
0.019307231530547142,
0.07510565966367722,
-0.03287685289978981,
0.08019816875457764,
0.09813148528337479,
0.08825678378343582,
-0.10000404715538025,
0.08111211657524109,
0.06777224689722061,
-0.04150259494781494,
0.033621978014707565,
0.10420102626085281,
-0.04986701160669327,
-0.04245395585894585,
0.08457721024751663,
0.12530498206615448,
-0.023088322952389717,
-0.05413966253399849,
0.01212761178612709,
-0.04834878444671631,
0.054270725697278976,
0.10672589391469955,
0.03587748110294342,
-0.0011736709857359529,
0.050750378519296646,
0.028017738834023476,
-0.10256616771221161,
0.08914750814437866,
0.03725229576230049,
0.01791483536362648,
-0.03840089589357376,
-0.04189951717853546,
0.004631043411791325,
-0.01516848523169756,
-0.018755726516246796,
-0.0170601699501276,
-0.08432288467884064,
-0.012585177086293697,
-0.11483073979616165,
0.008729316294193268,
-0.06474046409130096,
0.0068718683905899525,
0.030621705576777458,
-0.048198994249105453,
0.002455118577927351,
0.0015593849821016192,
-0.0763937458395958,
-0.051290228962898254,
-0.013947847299277782,
0.06659863144159317,
-0.12318508327007294,
0.042245566844940186,
0.06755290925502777,
-0.0967436209321022,
0.06653253734111786,
-0.007241120561957359,
0.011410431936383247,
0.0035017048940062523,
-0.15551216900348663,
0.04931795224547386,
-0.02801262028515339,
-0.024408893659710884,
0.02252740040421486,
-0.1943521499633789,
-0.0076536573469638824,
-0.04313570633530617,
-0.0573619082570076,
-0.004662544000893831,
-0.010509601794183254,
-0.11749584227800369,
0.10912971943616867,
0.007869033142924309,
-0.06068027764558792,
-0.027412936091423035,
0.04882120341062546,
0.10086818784475327,
-0.02643461339175701,
0.13437911868095398,
-0.007259611040353775,
0.07193886488676071,
-0.16531120240688324,
-0.004601365886628628,
-0.012241186574101448,
0.0436379611492157,
-0.026195699349045753,
-0.0405074842274189,
0.046567026525735855,
-0.02435867115855217,
0.19325846433639526,
-0.022945057600736618,
0.07084392011165619,
0.04857128486037254,
0.032133232802152634,
0.015501349233090878,
0.0795411467552185,
0.07082084566354752,
-0.005705251824110746,
0.0012581591727212071,
0.03978053480386734,
0.017982542514801025,
-0.03728210926055908,
-0.1555383801460266,
0.06970943510532379,
0.13411745429039001,
0.06166819855570793,
0.04408809542655945,
0.016431381925940514,
-0.10990120470523834,
-0.0851396843791008,
0.11883285641670227,
-0.007235993165522814,
-0.03617050126194954,
-0.06722866743803024,
0.173916757106781,
0.14665542542934418,
-0.1881304681301117,
0.07379920780658722,
-0.04021890461444855,
-0.047917887568473816,
-0.1390228271484375,
-0.19778694212436676,
-0.05708994343876839,
-0.04697444662451744,
-0.031041637063026428,
-0.06054393947124481,
0.0458466075360775,
0.05282822251319885,
-0.0030727433040738106,
-0.022938158363103867,
0.09914897382259369,
0.015943726524710655,
-0.02539999410510063,
0.02896830625832081,
0.05823741853237152,
0.03165817633271217,
-0.08659189939498901,
0.015606595203280449,
0.005570207256823778,
0.012911485508084297,
0.06870248168706894,
0.02457418665289879,
-0.05173683166503906,
0.027958450838923454,
-0.022081928327679634,
-0.11900684982538223,
0.030582444742321968,
-0.008381795138120651,
-0.040503326803445816,
0.13942766189575195,
0.027071574702858925,
0.004188410937786102,
-0.01948714442551136,
0.21969179809093475,
-0.07356120645999908,
-0.06005077436566353,
-0.13661882281303406,
0.08630871772766113,
-0.0646464005112648,
0.0389556810259819,
0.019435638561844826,
-0.1268240362405777,
0.018831565976142883,
0.177330881357193,
0.1440054327249527,
-0.01947030983865261,
0.0009323913836851716,
0.04481814056634903,
0.005301912315189838,
-0.0307242963463068,
0.02539828233420849,
0.04486451670527458,
0.15326927602291107,
-0.08694307506084442,
0.06786760687828064,
-0.017319831997156143,
-0.0827098861336708,
-0.012140425853431225,
0.11545486748218536,
-0.006142809521406889,
0.0005785097600892186,
-0.06454599648714066,
0.1289747804403305,
-0.09516481310129166,
-0.20264828205108643,
0.06106545776128769,
-0.06001608073711395,
-0.13382460176944733,
-0.04726420342922211,
0.03044959343969822,
-0.012021848000586033,
0.015416436828672886,
0.06878575682640076,
-0.056695982813835144,
0.18439672887325287,
0.04437119513750076,
-0.07817023992538452,
-0.10054308921098709,
0.05510885640978813,
-0.15851499140262604,
0.2766340374946594,
0.028239939361810684,
0.029272811487317085,
0.10791875422000885,
-0.003789537586271763,
-0.14966876804828644,
0.014645657502114773,
0.09151527285575867,
-0.055869899690151215,
0.05673375353217125,
0.17499114573001862,
0.0023246139753609896,
0.11875680834054947,
0.04801327362656593,
-0.05847916379570961,
0.053358294069767,
-0.10834988206624985,
-0.05009477213025093,
-0.10316772758960724,
0.06604080647230148,
-0.08997134119272232,
0.1677444577217102,
0.12227477133274078,
-0.0657745748758316,
-0.012364364229142666,
-0.022101426497101784,
0.08321168273687363,
0.01727793924510479,
0.10294997692108154,
0.009740419685840607,
-0.16769815981388092,
0.037040822207927704,
0.015676027163863182,
0.09718561172485352,
-0.194975346326828,
-0.05438392236828804,
0.04106029495596886,
-0.019167637452483177,
-0.07238626480102539,
0.11234915256500244,
0.04907934367656708,
0.053642772138118744,
-0.04926076903939247,
-0.025769544765353203,
0.009724765084683895,
0.1444830447435379,
-0.1114528551697731,
-0.024292191490530968
] |
null | null | transformers | # zephyr-infoNCA-preference
This model is a fine-tuned version of [HuggingFaceH4/mistral-7b-sft-beta](https://huggingface.co/HuggingFaceH4/mistral-7b-sft-beta) on the [openbmb/UltraFeedback](https://huggingface.co/datasets/openbmb/UltraFeedback) dataset.
It achieves the following results on the evaluation set:
- Loss: 0.4575
- Rewards/chosen: -0.8931
- Rewards/rejected: -2.0138
- Rewards/accuracies: 0.7745
- Rewards/margins: 1.1206
- Verify/constant 1: 1.0
- Verify/constant 1len: 1000.0
- Logps/rejected: -434.5525
- Logps/chosen: -364.4662
- Verify/bz: 1.0
- Verify/gather Bz: 2.0
- Regularization/forward Kl: 2.0564
- Regularization/reverse Kl: 1.0252
- Regularization/policy Data Loss: 3.8558
- Regularization/reference Data Loss: 1.3337
- Regularization/policy Ref Data Loss Gap: 2.5221
- Mask/mask Ratio: 0.4809
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-06
- train_batch_size: 1
- eval_batch_size: 1
- seed: 42
- distributed_type: multi-GPU
- num_devices: 2
- gradient_accumulation_steps: 16
- total_train_batch_size: 32
- total_eval_batch_size: 2
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 1
### Training results
| Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Verify/constant 1 | Verify/constant 1len | Logps/rejected | Logps/chosen | Verify/bz | Verify/gather Bz | Regularization/forward Kl | Regularization/reverse Kl | Regularization/policy Data Loss | Regularization/reference Data Loss | Regularization/policy Ref Data Loss Gap | Mask/mask Ratio |
|:-------------:|:-----:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:-----------------:|:--------------------:|:--------------:|:------------:|:---------:|:----------------:|:-------------------------:|:-------------------------:|:-------------------------------:|:----------------------------------:|:---------------------------------------:|:---------------:|
| 0.6892 | 0.05 | 100 | 0.6881 | 0.0039 | -0.0063 | 0.7145 | 0.0102 | 1.0 | 1000.0 | -233.8040 | -274.7605 | 1.0 | 2.0 | 0.0009 | 0.0009 | 1.3405 | 1.3337 | 0.0068 | 0.4809 |
| 0.6259 | 0.1 | 200 | 0.6258 | -0.1279 | -0.2905 | 0.7145 | 0.1627 | 1.0 | 1000.0 | -262.2266 | -287.9373 | 1.0 | 2.0 | 0.1727 | 0.1289 | 1.6331 | 1.3337 | 0.2994 | 0.4809 |
| 0.5436 | 0.15 | 300 | 0.5495 | -0.4736 | -0.9395 | 0.7415 | 0.4659 | 1.0 | 1000.0 | -327.1224 | -322.5125 | 1.0 | 2.0 | 0.6904 | 0.3995 | 2.2940 | 1.3337 | 0.9603 | 0.4809 |
| 0.5492 | 0.21 | 400 | 0.5161 | -0.5783 | -1.2015 | 0.7545 | 0.6232 | 1.0 | 1000.0 | -353.3223 | -332.9807 | 1.0 | 2.0 | 0.9794 | 0.5146 | 2.7574 | 1.3337 | 1.4237 | 0.4809 |
| 0.521 | 0.26 | 500 | 0.4982 | -0.7257 | -1.5000 | 0.7595 | 0.7743 | 1.0 | 1000.0 | -383.1716 | -347.7220 | 1.0 | 2.0 | 1.2016 | 0.5622 | 3.0006 | 1.3337 | 1.6669 | 0.4809 |
| 0.5152 | 0.31 | 600 | 0.4887 | -0.6594 | -1.4497 | 0.7685 | 0.7903 | 1.0 | 1000.0 | -378.1454 | -341.0961 | 1.0 | 2.0 | 1.2196 | 0.6044 | 3.0235 | 1.3337 | 1.6897 | 0.4809 |
| 0.4862 | 0.36 | 700 | 0.4857 | -0.7064 | -1.5442 | 0.7655 | 0.8378 | 1.0 | 1000.0 | -387.5948 | -345.7939 | 1.0 | 2.0 | 1.2568 | 0.6231 | 3.2214 | 1.3337 | 1.8877 | 0.4809 |
| 0.4632 | 0.41 | 800 | 0.4803 | -0.6298 | -1.4654 | 0.7755 | 0.8356 | 1.0 | 1000.0 | -379.7145 | -338.1303 | 1.0 | 2.0 | 1.3128 | 0.7041 | 2.8330 | 1.3337 | 1.4993 | 0.4809 |
| 0.4912 | 0.46 | 900 | 0.4707 | -0.7165 | -1.6486 | 0.7750 | 0.9321 | 1.0 | 1000.0 | -398.0345 | -346.8000 | 1.0 | 2.0 | 1.4120 | 0.7160 | 3.0682 | 1.3337 | 1.7345 | 0.4809 |
| 0.4588 | 0.52 | 1000 | 0.4680 | -0.8531 | -1.8542 | 0.7690 | 1.0011 | 1.0 | 1000.0 | -418.5936 | -360.4624 | 1.0 | 2.0 | 1.6382 | 0.8346 | 3.5448 | 1.3337 | 2.2111 | 0.4809 |
| 0.4956 | 0.57 | 1100 | 0.4650 | -0.7990 | -1.7772 | 0.7790 | 0.9781 | 1.0 | 1000.0 | -410.8913 | -355.0567 | 1.0 | 2.0 | 1.6270 | 0.8004 | 3.5035 | 1.3337 | 2.1698 | 0.4809 |
| 0.4738 | 0.62 | 1200 | 0.4629 | -0.8068 | -1.8169 | 0.7705 | 1.0102 | 1.0 | 1000.0 | -414.8670 | -355.8280 | 1.0 | 2.0 | 1.7938 | 0.8907 | 3.6708 | 1.3337 | 2.3371 | 0.4809 |
| 0.4657 | 0.67 | 1300 | 0.4622 | -0.8659 | -1.9282 | 0.7655 | 1.0623 | 1.0 | 1000.0 | -425.9926 | -361.7412 | 1.0 | 2.0 | 1.9375 | 0.9455 | 3.7639 | 1.3337 | 2.4301 | 0.4809 |
| 0.4938 | 0.72 | 1400 | 0.4586 | -0.8258 | -1.9093 | 0.7745 | 1.0834 | 1.0 | 1000.0 | -424.0995 | -357.7357 | 1.0 | 2.0 | 1.8620 | 0.9612 | 3.5611 | 1.3337 | 2.2274 | 0.4809 |
| 0.4511 | 0.77 | 1500 | 0.4580 | -0.8174 | -1.8815 | 0.7765 | 1.0641 | 1.0 | 1000.0 | -421.3289 | -356.8928 | 1.0 | 2.0 | 1.8762 | 0.9513 | 3.6341 | 1.3337 | 2.3003 | 0.4809 |
| 0.4724 | 0.83 | 1600 | 0.4573 | -0.8790 | -1.9952 | 0.7735 | 1.1162 | 1.0 | 1000.0 | -432.6913 | -363.0503 | 1.0 | 2.0 | 2.0060 | 1.0139 | 3.7650 | 1.3337 | 2.4312 | 0.4809 |
| 0.5045 | 0.88 | 1700 | 0.4572 | -0.8903 | -2.0141 | 0.7725 | 1.1238 | 1.0 | 1000.0 | -434.5795 | -364.1794 | 1.0 | 2.0 | 2.0502 | 1.0267 | 3.8128 | 1.3337 | 2.4790 | 0.4809 |
| 0.5007 | 0.93 | 1800 | 0.4577 | -0.9008 | -2.0247 | 0.7715 | 1.1239 | 1.0 | 1000.0 | -435.6480 | -365.2350 | 1.0 | 2.0 | 2.0707 | 1.0309 | 3.8706 | 1.3337 | 2.5369 | 0.4809 |
| 0.4747 | 0.98 | 1900 | 0.4576 | -0.8929 | -2.0129 | 0.7735 | 1.1200 | 1.0 | 1000.0 | -434.4668 | -364.4426 | 1.0 | 2.0 | 2.0555 | 1.0247 | 3.8552 | 1.3337 | 2.5215 | 0.4809 |
### Framework versions
- Transformers 4.35.0
- Pytorch 2.1.0
- Datasets 2.14.6
- Tokenizers 0.14.1
| {} | text-generation | ChenDRAG/zephyr-infoNCA-preference | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"4-bit",
"region:us"
] | 2024-02-08T09:01:01+00:00 | [] | [] | TAGS
#transformers #safetensors #mistral #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us
| zephyr-infoNCA-preference
=========================
This model is a fine-tuned version of HuggingFaceH4/mistral-7b-sft-beta on the openbmb/UltraFeedback dataset.
It achieves the following results on the evaluation set:
* Loss: 0.4575
* Rewards/chosen: -0.8931
* Rewards/rejected: -2.0138
* Rewards/accuracies: 0.7745
* Rewards/margins: 1.1206
* Verify/constant 1: 1.0
* Verify/constant 1len: 1000.0
* Logps/rejected: -434.5525
* Logps/chosen: -364.4662
* Verify/bz: 1.0
* Verify/gather Bz: 2.0
* Regularization/forward Kl: 2.0564
* Regularization/reverse Kl: 1.0252
* Regularization/policy Data Loss: 3.8558
* Regularization/reference Data Loss: 1.3337
* Regularization/policy Ref Data Loss Gap: 2.5221
* Mask/mask Ratio: 0.4809
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-06
* train\_batch\_size: 1
* eval\_batch\_size: 1
* seed: 42
* distributed\_type: multi-GPU
* num\_devices: 2
* gradient\_accumulation\_steps: 16
* total\_train\_batch\_size: 32
* total\_eval\_batch\_size: 2
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: cosine
* lr\_scheduler\_warmup\_ratio: 0.1
* num\_epochs: 1
### Training results
### Framework versions
* Transformers 4.35.0
* Pytorch 2.1.0
* Datasets 2.14.6
* Tokenizers 0.14.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 1\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 2\n* gradient\\_accumulation\\_steps: 16\n* total\\_train\\_batch\\_size: 32\n* total\\_eval\\_batch\\_size: 2\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 1",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.0\n* Pytorch 2.1.0\n* Datasets 2.14.6\n* Tokenizers 0.14.1"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 1\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 2\n* gradient\\_accumulation\\_steps: 16\n* total\\_train\\_batch\\_size: 32\n* total\\_eval\\_batch\\_size: 2\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 1",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.0\n* Pytorch 2.1.0\n* Datasets 2.14.6\n* Tokenizers 0.14.1"
] | [
54,
179,
4,
30
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 1\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 2\n* gradient\\_accumulation\\_steps: 16\n* total\\_train\\_batch\\_size: 32\n* total\\_eval\\_batch\\_size: 2\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 1### Training results### Framework versions\n\n\n* Transformers 4.35.0\n* Pytorch 2.1.0\n* Datasets 2.14.6\n* Tokenizers 0.14.1"
] | [
-0.09374158829450607,
0.0865808054804802,
-0.003508933586999774,
0.06578915566205978,
0.11688736081123352,
0.038814477622509,
0.1166367456316948,
0.12111333757638931,
-0.07969483733177185,
0.11100991070270538,
0.13099730014801025,
0.09454995393753052,
0.054377395659685135,
0.1595364362001419,
-0.04189189150929451,
-0.26775169372558594,
0.02015146054327488,
-0.04553069919347763,
-0.1383098065853119,
0.10957453399896622,
0.071271151304245,
-0.12765759229660034,
0.07583056390285492,
-0.031095115467905998,
-0.11601417511701584,
-0.03676983341574669,
-0.03391733393073082,
-0.032394129782915115,
0.12565301358699799,
0.03196001052856445,
0.095780149102211,
0.05023622885346413,
0.10451655089855194,
-0.25656643509864807,
0.007278406992554665,
0.07023212313652039,
0.024007685482501984,
0.08145192265510559,
0.08441093564033508,
-0.008788242004811764,
0.1570168137550354,
-0.11987771093845367,
0.06872227787971497,
0.030749376863241196,
-0.11133337020874023,
-0.18774950504302979,
-0.05946514382958412,
0.03228650242090225,
0.11749044805765152,
0.06546664237976074,
-0.021478749811649323,
0.09551989287137985,
-0.10657543689012527,
0.08345071226358414,
0.22237414121627808,
-0.26016971468925476,
-0.07135442644357681,
0.05746198445558548,
0.03416028246283531,
0.0765567198395729,
-0.13189202547073364,
0.000288498995359987,
0.0424414798617363,
0.006401175167411566,
0.09719745069742203,
0.007153216749429703,
0.037770505994558334,
0.023383593186736107,
-0.14649826288223267,
-0.04503883793950081,
0.08802426606416702,
0.055452294647693634,
-0.010574202984571457,
-0.0875718742609024,
-0.04920126870274544,
-0.18907901644706726,
-0.04219051077961922,
-0.013078625313937664,
0.02986433170735836,
-0.05457999184727669,
-0.08906687051057816,
0.04413919895887375,
-0.07714866101741791,
-0.08968332409858704,
0.01824774406850338,
0.1480839103460312,
0.055730707943439484,
-0.019566942006349564,
0.01712970621883869,
0.11879091709852219,
0.053263671696186066,
-0.14745965600013733,
-0.0016999406507238746,
0.016004953533411026,
-0.09131720662117004,
-0.031496189534664154,
-0.024414125829935074,
0.05280110239982605,
0.059937652200460434,
0.1300731599330902,
-0.0999235138297081,
0.0733199492096901,
0.06033606082201004,
0.0034010116942226887,
-0.05825158953666687,
0.12428691983222961,
-0.06916514039039612,
-0.0730501264333725,
-0.04636475071310997,
0.10304740816354752,
0.013999581336975098,
-0.012577713467180729,
-0.07418069243431091,
0.03657260537147522,
0.11043952405452728,
0.05975958704948425,
-0.030596042051911354,
0.0494985468685627,
-0.05197145789861679,
-0.00999379437416792,
0.0434681735932827,
-0.08862054347991943,
0.04723756015300751,
0.04536442086100578,
-0.059643153101205826,
-0.02856384962797165,
-0.010483251884579659,
0.008882488124072552,
-0.006049198564141989,
0.13036185503005981,
-0.09849873185157776,
-0.01811964437365532,
-0.0754762664437294,
-0.09898308664560318,
0.022593822330236435,
-0.05795103311538696,
-0.0025721266865730286,
-0.08265341073274612,
-0.09033310413360596,
-0.033975400030612946,
0.046910449862480164,
-0.07356777042150497,
-0.06308773905038834,
-0.07908398658037186,
-0.09188234061002731,
0.04740508273243904,
0.0008964724838733673,
0.17559778690338135,
-0.06956973671913147,
0.09972748160362244,
0.015297251753509045,
0.07410239428281784,
0.09031753987073898,
0.030378419905900955,
-0.05875605717301369,
0.05575774237513542,
-0.15124833583831787,
0.06130119785666466,
-0.07987489551305771,
0.04251326993107796,
-0.11755600571632385,
-0.10724272578954697,
-0.003144079353660345,
-0.01564914919435978,
0.07825615257024765,
0.14960969984531403,
-0.11520379781723022,
-0.0841284692287445,
0.18062689900398254,
-0.08283045887947083,
-0.10655731707811356,
0.13199253380298615,
-0.019423890858888626,
-0.05828140303492546,
0.023265181109309196,
0.14191481471061707,
0.07672379910945892,
-0.11014967411756516,
-0.010495600290596485,
-0.0067551457323133945,
0.09210721403360367,
0.04021649807691574,
0.09182711690664291,
0.0012173405848443508,
-0.0024427229072898626,
0.005599942058324814,
-0.02334580384194851,
0.04900746792554855,
-0.1128782331943512,
-0.0833989605307579,
-0.026752376928925514,
-0.08934946358203888,
0.03679130971431732,
0.04734189435839653,
0.03384943678975105,
-0.10008026659488678,
-0.11841976642608643,
-0.02226584032177925,
0.10397203266620636,
-0.07381696999073029,
0.008600592613220215,
-0.04840201511979103,
0.05103060230612755,
-0.018388232216238976,
0.005293402820825577,
-0.13649380207061768,
-0.0894269198179245,
0.03519611060619354,
-0.01784038543701172,
-0.0034758427646011114,
0.008744429796934128,
0.0733824297785759,
0.09648381173610687,
-0.06883471459150314,
-0.04833348095417023,
-0.020301416516304016,
0.005432393867522478,
-0.09607035666704178,
-0.2696106731891632,
-0.03094509430229664,
-0.028388243168592453,
0.18487347662448883,
-0.2440643310546875,
0.0066866278648376465,
0.0031398034188896418,
0.11771998554468155,
0.018310144543647766,
-0.03385508805513382,
-0.013976119458675385,
0.05329594761133194,
-0.05101129412651062,
-0.06328515708446503,
0.03296639397740364,
-0.013285872526466846,
-0.10509750247001648,
-0.014827855862677097,
-0.16634561121463776,
0.10228219628334045,
0.10193686932325363,
-0.01178055815398693,
-0.11939988285303116,
-0.057552482932806015,
-0.06884437054395676,
-0.05405401811003685,
-0.030610112473368645,
0.020831912755966187,
0.1335500031709671,
0.0013732225634157658,
0.10687683522701263,
-0.06850186735391617,
-0.05531966686248779,
0.03968003764748573,
-0.01033695973455906,
-0.010296509601175785,
0.16266831755638123,
0.0963423028588295,
-0.09030944108963013,
0.12416010349988937,
0.10752961039543152,
-0.0684814304113388,
0.12479331344366074,
-0.04691418632864952,
-0.08987289667129517,
-0.03854072839021683,
0.032290827482938766,
0.023672746494412422,
0.10839231312274933,
-0.11696728318929672,
0.011548986658453941,
0.012523649260401726,
0.028666473925113678,
0.01720227301120758,
-0.1893433779478073,
-0.018200498074293137,
0.031875286251306534,
-0.0638156458735466,
0.007705983705818653,
-0.036269281059503555,
-0.0020129343029111624,
0.11255799978971481,
0.015540537424385548,
-0.057902127504348755,
-0.019257165491580963,
-0.008642349392175674,
-0.08650916069746017,
0.22700285911560059,
-0.09053503721952438,
-0.12789973616600037,
-0.09548107534646988,
0.009393737651407719,
-0.028580738231539726,
0.0042159100994467735,
0.03347095102071762,
-0.11711470037698746,
-0.03706733137369156,
-0.08248746395111084,
0.023259639739990234,
-0.030609676614403725,
0.04979529604315758,
0.006564836949110031,
0.01696793921291828,
0.03263387829065323,
-0.08377834409475327,
0.007434350438416004,
-0.026431046426296234,
-0.053761910647153854,
0.058045994490385056,
0.04748966172337532,
0.0986209362745285,
0.16690103709697723,
0.024649931117892265,
0.011010087095201015,
-0.02646651305258274,
0.17006711661815643,
-0.09340523183345795,
-0.005711109843105078,
0.05387650057673454,
0.01462982501834631,
0.038647640496492386,
0.1514652520418167,
0.04302237182855606,
-0.0846235528588295,
0.027727903798222542,
0.04807264357805252,
-0.02341052144765854,
-0.21704670786857605,
-0.04766270890831947,
-0.03825092315673828,
0.016857050359249115,
0.10270990431308746,
0.0303189754486084,
-0.010367929004132748,
0.044480908662080765,
-0.014125928282737732,
-0.003088610013946891,
0.00833717081695795,
0.056569892913103104,
0.04025040939450264,
0.03852261230349541,
0.11200740188360214,
-0.025828884914517403,
-0.0521768219769001,
0.04180397838354111,
-0.007387695834040642,
0.21862637996673584,
-0.022176468744874,
0.1402815580368042,
0.030085787177085876,
0.1396690458059311,
-0.00359214935451746,
0.06148570775985718,
0.02971069887280464,
-0.044077761471271515,
0.009248791262507439,
-0.050788696855306625,
-0.023836132138967514,
0.054146215319633484,
0.0564899779856205,
0.01883769780397415,
-0.11697529256343842,
0.05844232067465782,
0.06202102079987526,
0.24889954924583435,
0.08666163682937622,
-0.2970208525657654,
-0.09801702201366425,
0.023653453215956688,
-0.039203133434057236,
-0.012874356471002102,
0.029127659276127815,
0.16291813552379608,
-0.08728034049272537,
0.056665610522031784,
-0.05918183550238609,
0.06670404970645905,
-0.05696672573685646,
0.014242206700146198,
0.056319914758205414,
0.11252672970294952,
-0.009043720550835133,
0.06770415604114532,
-0.25010764598846436,
0.2956026792526245,
-0.006269517820328474,
0.048669926822185516,
-0.06131323054432869,
0.016983184963464737,
0.011260580271482468,
0.029177416115999222,
0.07974949479103088,
-0.004619223531335592,
-0.10744510591030121,
-0.1897789090871811,
-0.10726257413625717,
0.03397743031382561,
0.14797410368919373,
-0.07815477252006531,
0.12943977117538452,
-0.03576536476612091,
-0.027894148603081703,
0.05337340012192726,
-0.06853743642568588,
-0.07831942290067673,
-0.10502002388238907,
0.020411675795912743,
-0.03011113964021206,
0.0062726689502596855,
-0.07682807743549347,
-0.10446483641862869,
-0.11026590317487717,
0.19596898555755615,
-0.0981002226471901,
-0.01888306625187397,
-0.1295284777879715,
0.08826202899217606,
0.13799743354320526,
-0.07936575263738632,
0.02557973377406597,
-0.0004757422721013427,
0.08888070285320282,
0.01733792945742607,
-0.02682160958647728,
0.12331933528184891,
-0.09125030040740967,
-0.236156165599823,
-0.056725844740867615,
0.1407487541437149,
0.04631650075316429,
0.06976193189620972,
-0.028255973011255264,
0.029784493148326874,
-0.014054549857974052,
-0.10336292535066605,
0.06280434131622314,
0.034055281430482864,
0.058347806334495544,
0.04997030645608902,
-0.05208459496498108,
0.021547822281718254,
-0.047589682042598724,
-0.05826297029852867,
0.13860687613487244,
0.3381063938140869,
-0.09759709984064102,
0.036187879741191864,
0.07810526341199875,
-0.054714787751436234,
-0.18701986968517303,
-0.008062145672738552,
0.07945796102285385,
0.019268808886408806,
-0.0029956516809761524,
-0.1755111664533615,
0.06846778839826584,
0.09876292943954468,
-0.015769757330417633,
0.09462940692901611,
-0.3182390630245209,
-0.13771717250347137,
0.08579510450363159,
0.11075785011053085,
0.015455242246389389,
-0.18157745897769928,
-0.038252297788858414,
-0.00013113221211824566,
-0.07405944168567657,
0.08167527616024017,
-0.032394446432590485,
0.12881743907928467,
-0.008759253658354282,
0.01904122903943062,
0.024782299995422363,
-0.06896455585956573,
0.15878652036190033,
-0.01908671110868454,
0.06662772595882416,
-0.03353690356016159,
-0.003763290122151375,
-0.00009446006879443303,
-0.06961263716220856,
0.0035315898712724447,
-0.10901285707950592,
0.04481372609734535,
-0.05149662494659424,
-0.030292345210909843,
-0.07575910538434982,
0.023213941603899002,
-0.052733901888132095,
-0.059599000960588455,
-0.03700363636016846,
0.04998796433210373,
0.07207150012254715,
-0.020275283604860306,
0.11849851161241531,
-0.018482757732272148,
0.16249416768550873,
0.12361788749694824,
0.1007508859038353,
-0.011653578840196133,
-0.023667111992836,
0.0037571461871266365,
-0.011858167126774788,
0.03311389684677124,
-0.12285497784614563,
0.025357162579894066,
0.14818744361400604,
0.0339852049946785,
0.13158227503299713,
0.06583089381456375,
-0.048148345202207565,
-0.007931279949843884,
0.07773669809103012,
-0.14141419529914856,
-0.15178443491458893,
0.0010460647754371166,
-0.02582663856446743,
-0.140476793050766,
0.03235502541065216,
0.10833321511745453,
-0.05673358216881752,
0.0016230238834396005,
-0.011545279063284397,
0.06468216329813004,
-0.020738905295729637,
0.21988292038440704,
0.030890755355358124,
0.09148871153593063,
-0.09144172072410583,
0.08182516694068909,
0.04709397628903389,
-0.14964398741722107,
0.03642350435256958,
0.10723592340946198,
-0.0768512710928917,
-0.03293735161423683,
0.08544882386922836,
0.10050107538700104,
0.00278506800532341,
-0.02735172212123871,
-0.12269524484872818,
-0.14891156554222107,
0.08608924597501755,
0.10755094140768051,
0.04634270817041397,
0.047004278749227524,
-0.008645921014249325,
0.04461604356765747,
-0.1323527693748474,
0.1278623342514038,
0.061834145337343216,
0.08164410293102264,
-0.1380883753299713,
0.14385049045085907,
-0.01313534565269947,
0.009816210716962814,
-0.020665932446718216,
0.030503693968057632,
-0.14787574112415314,
-0.008112985640764236,
-0.10561439394950867,
-0.01556609757244587,
-0.05996207892894745,
0.003933335654437542,
0.015001999214291573,
-0.056885529309511185,
-0.043879397213459015,
0.0018823789432644844,
-0.09434010088443756,
-0.03714662790298462,
-0.023517237976193428,
0.06402834504842758,
-0.12430911511182785,
-0.021892856806516647,
0.03953612595796585,
-0.11159757524728775,
0.1025809496641159,
0.03664719685912132,
0.03691325709223747,
0.032698653638362885,
-0.14087563753128052,
0.03907286748290062,
0.044064633548259735,
-0.00704170111566782,
0.042517635971307755,
-0.16151610016822815,
-0.004405026324093342,
-0.04035305976867676,
0.03062516264617443,
0.019293084740638733,
0.047405850142240524,
-0.11692732572555542,
0.03561379760503769,
-0.04942760616540909,
-0.08967427909374237,
-0.056814875453710556,
0.0329449288547039,
0.05710407346487045,
-0.032565440982580185,
0.14058370888233185,
-0.0918210819363594,
0.046336740255355835,
-0.21546411514282227,
-0.011762420646846294,
-0.00668701296672225,
-0.06903732568025589,
-0.07101530581712723,
-0.033289771527051926,
0.0985458642244339,
-0.050193559378385544,
0.10634670406579971,
-0.03602820634841919,
0.04329042136669159,
0.029464324936270714,
-0.073271244764328,
0.03692247346043587,
0.07018852978944778,
0.17080850899219513,
0.050415799021720886,
-0.04027669504284859,
0.02017592266201973,
0.019523896276950836,
0.08688313513994217,
0.06830485910177231,
0.22682826220989227,
0.13057668507099152,
-0.007384416181594133,
0.085150808095932,
0.03614545613527298,
-0.15906807780265808,
-0.12790696322917938,
0.1203412190079689,
-0.10282357782125473,
0.10874339193105698,
-0.03427799046039581,
0.16880135238170624,
0.10349901765584946,
-0.1934848129749298,
0.016349680721759796,
-0.06293636560440063,
-0.09732469171285629,
-0.11127955466508865,
-0.035714566707611084,
-0.09510104358196259,
-0.169852614402771,
0.00651336507871747,
-0.12080681324005127,
0.03889879211783409,
0.06842991709709167,
0.056007690727710724,
0.025790967047214508,
0.14652755856513977,
0.07200688868761063,
0.043202344328165054,
0.04984855651855469,
0.04653747007250786,
-0.011947761289775372,
-0.026225313544273376,
-0.09713970124721527,
0.0072228084318339825,
-0.041947364807128906,
0.0462355874478817,
-0.05416034534573555,
-0.0805002897977829,
0.0847286581993103,
0.017395298928022385,
-0.0917990431189537,
0.020037174224853516,
-0.011897687800228596,
0.04687461256980896,
0.06755754351615906,
0.015428478829562664,
-0.008965367451310158,
-0.026120219379663467,
0.21451982855796814,
-0.08899791538715363,
-0.06493846327066422,
-0.11794188618659973,
0.2937953770160675,
-0.006239450071007013,
0.018630405887961388,
0.031036054715514183,
-0.06849706918001175,
-0.02497601881623268,
0.1674961894750595,
0.19818460941314697,
-0.06352739036083221,
-0.01454232819378376,
0.018547803163528442,
-0.005849586334079504,
-0.009331312030553818,
0.09362898021936417,
0.11751467734575272,
0.08447840064764023,
-0.08184275031089783,
-0.035927899181842804,
-0.023531973361968994,
-0.0382242314517498,
-0.042100150138139725,
0.05250576138496399,
0.03999210149049759,
0.005087895318865776,
-0.0373539924621582,
0.0682060495018959,
-0.07644715160131454,
-0.11217758804559708,
0.06039278954267502,
-0.20406465232372284,
-0.16178365051746368,
-0.0295164342969656,
0.08737947046756744,
-0.006437304429709911,
0.063840351998806,
0.00046400161227211356,
-0.04545655474066734,
0.08331239223480225,
-0.005981376860290766,
-0.0629948228597641,
-0.08220801502466202,
0.05515047535300255,
-0.048967473208904266,
0.1920837014913559,
-0.04341239109635353,
0.027114441618323326,
0.13457228243350983,
0.029738111421465874,
-0.10470852255821228,
0.047562375664711,
0.09300743043422699,
-0.1372644603252411,
0.047373246401548386,
0.1457965224981308,
-0.04605914652347565,
0.10505396872758865,
0.06503193825483322,
-0.1001274362206459,
0.008947654627263546,
-0.09053780138492584,
-0.04121594503521919,
-0.0350295826792717,
-0.011501185595989227,
-0.03478676453232765,
0.15443776547908783,
0.2192065715789795,
-0.048225294798612595,
0.005947508849203587,
-0.040276557207107544,
0.030163371935486794,
0.020297881215810776,
0.1365910768508911,
-0.029566984623670578,
-0.29540693759918213,
0.037136536091566086,
0.01957586221396923,
0.04135904833674431,
-0.23219265043735504,
-0.0849544107913971,
0.02575218863785267,
-0.020444950088858604,
-0.10224878042936325,
0.11793678253889084,
0.08003830909729004,
0.04651497304439545,
-0.060660868883132935,
-0.13300366699695587,
-0.04460569843649864,
0.17385706305503845,
-0.15542887151241302,
-0.08516206592321396
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bert-base-cased-wikitext2
This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 6.9775
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| 7.1964 | 1.0 | 1171 | 7.0937 |
| 7.0407 | 2.0 | 2342 | 7.0175 |
| 6.954 | 3.0 | 3513 | 6.9555 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "bert-base-cased", "model-index": [{"name": "bert-base-cased-wikitext2", "results": []}]} | fill-mask | GavinQiangLi/bert-base-cased-wikitext2 | [
"transformers",
"tensorboard",
"safetensors",
"bert",
"fill-mask",
"generated_from_trainer",
"base_model:bert-base-cased",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T09:05:34+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #bert #fill-mask #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| bert-base-cased-wikitext2
=========================
This model is a fine-tuned version of bert-base-cased on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 6.9775
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3.0
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #bert #fill-mask #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
67,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #bert #fill-mask #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.10209926962852478,
0.08588685095310211,
-0.0022868302185088396,
0.10783667862415314,
0.13204804062843323,
0.024829724803566933,
0.16159650683403015,
0.10759466141462326,
-0.06513112038373947,
0.05036148801445961,
0.13632068037986755,
0.11392592638731003,
0.018861979246139526,
0.15859436988830566,
-0.0397927425801754,
-0.2204110026359558,
0.01243942603468895,
0.029290547594428062,
-0.08428763598203659,
0.1122547909617424,
0.08683156222105026,
-0.12789596617221832,
0.08313006907701492,
0.003926153294742107,
-0.16986551880836487,
0.01678743213415146,
0.026343008503317833,
-0.053073424845933914,
0.1331562101840973,
0.02407696284353733,
0.1479429304599762,
0.014969303272664547,
0.09706030040979385,
-0.1889471709728241,
0.01507424283772707,
0.0663505271077156,
-0.0073885186575353146,
0.0775480717420578,
0.02040504664182663,
0.021201489493250847,
0.0770827978849411,
-0.08209694921970367,
0.06792661547660828,
0.01838952861726284,
-0.12002301961183548,
-0.21939057111740112,
-0.07874327152967453,
0.041087355464696884,
0.0934670940041542,
0.08174169063568115,
-0.005425270181149244,
0.13578368723392487,
-0.05444895476102829,
0.08400395512580872,
0.23104478418827057,
-0.3250315487384796,
-0.07078772783279419,
0.041609231382608414,
0.04492439329624176,
0.05203361064195633,
-0.10273600369691849,
-0.016509735956788063,
0.0607292503118515,
0.03030957467854023,
0.1409817785024643,
-0.02832370437681675,
-0.04964012652635574,
-0.0019454320427030325,
-0.14632852375507355,
-0.02167084813117981,
0.12340842187404633,
0.05296074226498604,
-0.045799583196640015,
-0.04413887858390808,
-0.06914446502923965,
-0.15059234201908112,
-0.05016249418258667,
-0.024809369817376137,
0.04678196832537651,
-0.02657415345311165,
-0.07526062428951263,
-0.012526600621640682,
-0.10705455392599106,
-0.08779503405094147,
-0.058205608278512955,
0.1516861766576767,
0.04001280665397644,
0.011092279106378555,
-0.015714110806584358,
0.09360402077436447,
-0.054410580545663834,
-0.13451354205608368,
0.020957564935088158,
0.02932620979845524,
0.0010927512776106596,
-0.04505523666739464,
-0.055746667087078094,
-0.07512685656547546,
0.030774550512433052,
0.15474222600460052,
-0.04650983586907387,
0.04587341472506523,
0.025010529905557632,
0.05310777947306633,
-0.11552425473928452,
0.15429732203483582,
-0.04737526923418045,
-0.02789655327796936,
0.017489086836576462,
0.07811328023672104,
0.05284327268600464,
-0.00014860746159683913,
-0.11785382032394409,
0.031200658529996872,
0.09066015481948853,
0.009124386124312878,
-0.06277395039796829,
0.06942610442638397,
-0.05320104584097862,
0.013931632041931152,
0.02308451011776924,
-0.08581676334142685,
0.02651439607143402,
-0.004540890455245972,
-0.04207834601402283,
-0.08166174590587616,
0.043781984597444534,
0.01638738252222538,
0.019192062318325043,
0.10758667439222336,
-0.09506863355636597,
0.0034481724724173546,
-0.09873180091381073,
-0.12269613146781921,
0.017831675708293915,
-0.08089128881692886,
0.02210608869791031,
-0.1075596809387207,
-0.16404110193252563,
0.0026403642259538174,
0.07237353920936584,
-0.024157539010047913,
-0.022727010771632195,
-0.03429555520415306,
-0.07757106423377991,
0.013788356445729733,
-0.010901730507612228,
0.10522652417421341,
-0.06381533294916153,
0.09836647659540176,
0.0449320524930954,
0.08536335080862045,
-0.04723205789923668,
0.031138382852077484,
-0.09376823157072067,
0.03424786031246185,
-0.19261451065540314,
-0.01019538938999176,
-0.07175127416849136,
0.05221229046583176,
-0.07762843370437622,
-0.08123618364334106,
-0.022588465362787247,
0.013747124001383781,
0.08558784425258636,
0.09290031343698502,
-0.16733358800411224,
-0.06571376323699951,
0.18958964943885803,
-0.09687639772891998,
-0.13365855813026428,
0.12928855419158936,
-0.04954995959997177,
0.045025534927845,
0.04364392161369324,
0.15594428777694702,
0.06269057095050812,
-0.13925118744373322,
0.007787591312080622,
0.0010086565744131804,
0.055502500385046005,
-0.04966351017355919,
0.06622454524040222,
-0.004275529179722071,
-0.00042928053881041706,
0.014735547825694084,
-0.04650343209505081,
0.053577058017253876,
-0.08351941406726837,
-0.09044919908046722,
-0.03798409923911095,
-0.10869782418012619,
0.05990836024284363,
0.05521957948803902,
0.05815882608294487,
-0.1169571653008461,
-0.10586342215538025,
0.04176175221800804,
0.07559660077095032,
-0.06458564102649689,
0.017675841227173805,
-0.07888907939195633,
0.08864249289035797,
-0.07199972867965698,
-0.024988384917378426,
-0.1449703723192215,
-0.06102145090699196,
0.01045291405171156,
-0.020711749792099,
0.008006851188838482,
0.0015341740800067782,
0.08865495026111603,
0.08617810904979706,
-0.0687909945845604,
-0.02787967585027218,
-0.038259636610746384,
0.01490863785147667,
-0.12190014123916626,
-0.2090400606393814,
-0.03447791561484337,
-0.03679769113659859,
0.13110584020614624,
-0.20725029706954956,
0.045451462268829346,
-0.036524537950754166,
0.08892511576414108,
0.03704782947897911,
-0.01306892465800047,
-0.05027230083942413,
0.0721937045454979,
-0.026033729314804077,
-0.06136570870876312,
0.052607461810112,
0.011793251149356365,
-0.08239222317934036,
-0.03924349695444107,
-0.1426243633031845,
0.1947338879108429,
0.12947820127010345,
-0.08063171058893204,
-0.09198484569787979,
0.011183440685272217,
-0.05395917966961861,
-0.0316583476960659,
-0.04325316473841667,
0.0037198446225374937,
0.1176866888999939,
-0.012637550011277199,
0.13665790855884552,
-0.0724373534321785,
-0.027687694877386093,
0.03826136142015457,
-0.05157143622636795,
0.0026811673305928707,
0.08145370334386826,
0.13151581585407257,
-0.08209016174077988,
0.15193884074687958,
0.18156658113002777,
-0.10678968578577042,
0.13714858889579773,
-0.035943277180194855,
-0.06709121912717819,
-0.02814900130033493,
0.0015533189289271832,
0.013094200752675533,
0.14823299646377563,
-0.12664894759655,
0.0013201998081058264,
0.008593611419200897,
-0.0004074955068062991,
0.014555629342794418,
-0.22908253967761993,
-0.037141427397727966,
0.03651884198188782,
-0.046520594507455826,
-0.009363294579088688,
-0.012150023132562637,
-0.016159171238541603,
0.092155322432518,
0.003375684143975377,
-0.091131292283535,
0.041047099977731705,
-0.0022101986687630415,
-0.07620758563280106,
0.2044854313135147,
-0.08195438235998154,
-0.12802819907665253,
-0.13418638706207275,
-0.07304321229457855,
-0.02642097696661949,
0.025332577526569366,
0.061225391924381256,
-0.07328761368989944,
-0.047196224331855774,
-0.09566134214401245,
-0.0101085864007473,
0.0498160645365715,
0.04275965690612793,
0.023215552791953087,
-0.00894680991768837,
0.0961121916770935,
-0.1018616110086441,
-0.0139893414452672,
-0.047138940542936325,
-0.052354998886585236,
0.033987484872341156,
0.039661310613155365,
0.1214451715350151,
0.13522334396839142,
-0.02110862545669079,
-0.00202011177316308,
-0.027738699689507484,
0.22831694781780243,
-0.06232732906937599,
-0.017600713297724724,
0.12790611386299133,
-0.02350490540266037,
0.04989365115761757,
0.137238547205925,
0.06575813889503479,
-0.09097786247730255,
0.019653385505080223,
0.04434065893292427,
-0.03270355984568596,
-0.19034159183502197,
-0.029188932850956917,
-0.042577218264341354,
-0.01688273623585701,
0.098199762403965,
0.03229266032576561,
0.02387896738946438,
0.06705469638109207,
0.03277532011270523,
0.07172002643346786,
-0.027401842176914215,
0.06409361213445663,
0.07827135920524597,
0.03843621164560318,
0.1227346807718277,
-0.04315449297428131,
-0.06255150586366653,
0.023723211139440536,
0.0009419078705832362,
0.2093757539987564,
0.023906761780381203,
0.11384072154760361,
0.07381703704595566,
0.17555828392505646,
-0.002697582356631756,
0.07907407730817795,
-0.008746251463890076,
-0.06159562990069389,
-0.005227542016655207,
-0.05578836426138878,
-0.020456140860915184,
0.03894364461302757,
-0.0768061950802803,
0.06605756282806396,
-0.10626499354839325,
0.021738534793257713,
0.04595551639795303,
0.2564658224582672,
0.05925258994102478,
-0.328666090965271,
-0.09320255368947983,
0.011608736589550972,
-0.02167290635406971,
-0.024323320016264915,
0.019844045862555504,
0.1308905929327011,
-0.04630689695477486,
0.02933957241475582,
-0.07678196579217911,
0.06903206557035446,
0.004610879812389612,
0.04281828552484512,
0.073114313185215,
0.10277190059423447,
-0.003499735612422228,
0.056867580860853195,
-0.2707573473453522,
0.2965399920940399,
0.00997297652065754,
0.08872371912002563,
-0.05734993517398834,
0.010247171856462955,
0.037500977516174316,
0.06264133751392365,
0.09126733243465424,
-0.022177891805768013,
-0.059642449021339417,
-0.19408199191093445,
-0.05587810277938843,
0.030143504962325096,
0.09104043245315552,
-0.0217283945530653,
0.10158535838127136,
-0.029353858903050423,
-0.010268899612128735,
0.0901373028755188,
0.019715912640094757,
-0.09113026410341263,
-0.0903572216629982,
-0.012094590812921524,
0.03550439327955246,
-0.050738830119371414,
-0.09442244470119476,
-0.1116785779595375,
-0.13283459842205048,
0.16071516275405884,
-0.034509725868701935,
-0.01628248393535614,
-0.0980525091290474,
0.0653165876865387,
0.06220351904630661,
-0.07656925916671753,
0.06741510331630707,
0.008890236727893353,
0.09291130304336548,
0.014632579870522022,
-0.056076936423778534,
0.12756595015525818,
-0.08829561620950699,
-0.15695901215076447,
-0.08772911131381989,
0.09658555686473846,
0.023812150582671165,
0.055089034140110016,
-0.0028325587045401335,
0.010792279615998268,
-0.009293346665799618,
-0.06765849143266678,
0.03876864165067673,
-0.019427215680480003,
0.06540418416261673,
0.0013478245818987489,
-0.07148730754852295,
0.0005583643214777112,
-0.04207152500748634,
-0.021052733063697815,
0.1477106809616089,
0.2801376283168793,
-0.0946185514330864,
0.0023549229372292757,
0.062462903559207916,
-0.05581221356987953,
-0.21889133751392365,
0.030123872682452202,
0.03690727427601814,
0.015564428642392159,
0.05480673164129257,
-0.14139050245285034,
0.11193069815635681,
0.0890912413597107,
-0.031956009566783905,
0.1208915039896965,
-0.280830979347229,
-0.13170568645000458,
0.14678390324115753,
0.16518211364746094,
0.11851324886083603,
-0.1524769812822342,
-0.02549261972308159,
-0.025513509288430214,
-0.11539236456155777,
0.08085992932319641,
-0.11590105295181274,
0.1106761172413826,
-0.009904539212584496,
0.05512208119034767,
-0.001831741421483457,
-0.07408720254898071,
0.1246870830655098,
-0.03520219773054123,
0.10653999447822571,
-0.06125889718532562,
-0.009630038402974606,
0.0550985112786293,
-0.05193086341023445,
0.014495464973151684,
-0.10413301736116409,
0.025907158851623535,
-0.017521517351269722,
-0.025595437735319138,
-0.05553548410534859,
0.047794803977012634,
-0.03694078326225281,
-0.06471715122461319,
-0.027041807770729065,
0.033794425427913666,
0.03148815408349037,
-0.018063556402921677,
0.1225803941488266,
0.019668810069561005,
0.17135699093341827,
0.11911963671445847,
0.05548541247844696,
-0.062339067459106445,
-0.0496305376291275,
0.0005261291516944766,
-0.04044659435749054,
0.07417548447847366,
-0.13007323443889618,
0.030881157144904137,
0.10995958000421524,
0.01230582408607006,
0.1448640376329422,
0.07603268325328827,
-0.045747190713882446,
0.012271547690033913,
0.08247845619916916,
-0.16187460720539093,
-0.09119239449501038,
0.008870056830346584,
-0.03730693459510803,
-0.13073395192623138,
0.05652497336268425,
0.10442039370536804,
-0.0763629600405693,
0.009667645208537579,
-0.009044189006090164,
0.017169030383229256,
-0.05788194388151169,
0.18984077870845795,
0.06534823030233383,
0.05221972614526749,
-0.06951560825109482,
0.06007074937224388,
0.022096388041973114,
-0.07578176259994507,
0.015422003343701363,
0.031023787334561348,
-0.07575872540473938,
-0.04434578865766525,
0.07139506191015244,
0.1892668604850769,
0.0004101428494323045,
-0.05827173590660095,
-0.15145601332187653,
-0.11132652312517166,
0.05031880363821983,
0.1966564655303955,
0.09139274805784225,
0.0001312941312789917,
-0.024495964869856834,
0.03196707367897034,
-0.1085885763168335,
0.10805071145296097,
0.03336583450436592,
0.0808260515332222,
-0.14129826426506042,
0.1269405633211136,
0.0015672105364501476,
0.017220819368958473,
-0.027176205068826675,
0.05744783207774162,
-0.11754970997571945,
-0.007653494831174612,
-0.12294919788837433,
-0.019684968516230583,
-0.031029270961880684,
-0.006589110009372234,
0.010441653430461884,
-0.053152237087488174,
-0.07213756442070007,
0.020206347107887268,
-0.10486947000026703,
-0.026604725047945976,
0.04683193936944008,
0.04141177982091904,
-0.12287264317274094,
-0.036652419716119766,
0.032720357179641724,
-0.06795817613601685,
0.05241988226771355,
0.025963343679904938,
0.028506867587566376,
0.055992186069488525,
-0.17965780198574066,
0.0220120120793581,
0.06670307368040085,
0.008710314519703388,
0.05080439895391464,
-0.09215405583381653,
-0.017312223091721535,
-0.012628982774913311,
0.04168844223022461,
0.007958072237670422,
0.09102199226617813,
-0.12937971949577332,
0.0058471509255468845,
-0.024261100217700005,
-0.06707391142845154,
-0.04694417491555214,
0.007129462901502848,
0.09224674105644226,
-0.0058198473416268826,
0.19995607435703278,
-0.10859742015600204,
0.022111540660262108,
-0.20270346105098724,
0.009775719605386257,
-0.006528137717396021,
-0.08900018036365509,
-0.10090076178312302,
-0.04448052868247032,
0.05049294978380203,
-0.0601014606654644,
0.1414039433002472,
0.005628677085042,
0.02248559333384037,
0.030398545786738396,
-0.05484245717525482,
0.04381261020898819,
0.025746727362275124,
0.21978971362113953,
0.019517334178090096,
-0.04206973314285278,
0.018357915803790092,
0.044082675129175186,
0.10899187624454498,
0.07574927806854248,
0.18073409795761108,
0.16068653762340546,
-0.015409502200782299,
0.11321334540843964,
0.05163443461060524,
-0.060558151453733444,
-0.13558028638362885,
0.04688490182161331,
-0.050158098340034485,
0.0974966511130333,
-0.01956983655691147,
0.2011023759841919,
0.10207760334014893,
-0.15657445788383484,
0.016974516212940216,
-0.05231526494026184,
-0.075982004404068,
-0.1170903667807579,
-0.061862360686063766,
-0.09058181941509247,
-0.1396179348230362,
0.007974976673722267,
-0.09746598452329636,
0.0006756834336556494,
0.1024397686123848,
0.0010972041636705399,
-0.006407460197806358,
0.20471949875354767,
-0.00010033214493887499,
0.05276424437761307,
0.0363563597202301,
0.012778309173882008,
-0.0373048409819603,
-0.08906097710132599,
-0.084968201816082,
-0.0019526934484019876,
-0.025706078857183456,
0.02039281465113163,
-0.06310059875249863,
-0.038683291524648666,
0.05456643924117088,
0.007322158198803663,
-0.10231509059667587,
0.00572785921394825,
0.015847641974687576,
0.049847643822431564,
0.04287450388073921,
0.008085833862423897,
0.022425364702939987,
-0.0053983586840331554,
0.1850830763578415,
-0.08182044327259064,
-0.07688719779253006,
-0.11349064856767654,
0.2190798819065094,
0.021165719255805016,
0.012131462804973125,
0.006161106750369072,
-0.08162438869476318,
0.02300076372921467,
0.23600146174430847,
0.20043587684631348,
-0.07048363983631134,
0.011836552061140537,
0.01000204961746931,
-0.015138769522309303,
-0.06519274413585663,
0.0988474190235138,
0.1234702318906784,
0.03896811977028847,
-0.07644236087799072,
-0.06255723536014557,
-0.04446043819189072,
-0.008212316781282425,
-0.05128813907504082,
0.032099399715662,
0.04076532647013664,
0.017520025372505188,
-0.04545918107032776,
0.04515508562326431,
-0.01890101656317711,
-0.12228599190711975,
0.0825340524315834,
-0.19301261007785797,
-0.14708054065704346,
-0.00575566990301013,
0.11985781788825989,
-0.011210829950869083,
0.05825086683034897,
-0.036857809871435165,
0.002166211372241378,
0.08173574507236481,
-0.02004588209092617,
-0.05864664539694786,
-0.09392084181308746,
0.06357826292514801,
-0.08664799481630325,
0.2603675127029419,
-0.03505554050207138,
0.050325509160757065,
0.13317684829235077,
0.04906455799937248,
-0.07072868943214417,
0.08548084646463394,
0.05074382945895195,
-0.09472692757844925,
0.014037487097084522,
0.08311029523611069,
-0.04714005067944527,
0.10328809916973114,
0.046015966683626175,
-0.1322982907295227,
0.020920010283589363,
-0.0833863615989685,
-0.07965494692325592,
-0.04668809846043587,
-0.019613957032561302,
-0.06767944246530533,
0.13308559358119965,
0.1875530183315277,
-0.02689715474843979,
0.0034371556248515844,
-0.04990079253911972,
0.03521425276994705,
0.07422935962677002,
0.03788664937019348,
-0.0482247993350029,
-0.23412354290485382,
0.041102733463048935,
0.04876848682761192,
-0.018865888938307762,
-0.28438878059387207,
-0.09932856261730194,
0.0017534358194097877,
-0.055851202458143234,
-0.09377887099981308,
0.06445474177598953,
0.12508763372898102,
0.06378360092639923,
-0.06340860575437546,
-0.1025419533252716,
-0.07096154242753983,
0.1498255729675293,
-0.1453854739665985,
-0.10337767750024796
] |
null | null | null |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
This modelcard aims to be a base template for new models. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1).
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] | {} | null | milk639/Goddess | [
"arxiv:1910.09700",
"region:us"
] | 2024-02-08T09:09:17+00:00 | [
"1910.09700"
] | [] | TAGS
#arxiv-1910.09700 #region-us
|
# Model Card for Model ID
This modelcard aims to be a base template for new models. It has been generated using this raw template.
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID\n\n\n\nThis modelcard aims to be a base template for new models. It has been generated using this raw template.",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#arxiv-1910.09700 #region-us \n",
"# Model Card for Model ID\n\n\n\nThis modelcard aims to be a base template for new models. It has been generated using this raw template.",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
15,
29,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#arxiv-1910.09700 #region-us \n# Model Card for Model ID\n\n\n\nThis modelcard aims to be a base template for new models. It has been generated using this raw template.## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.1066984087228775,
0.19898438453674316,
-0.002620849059894681,
0.027911467477679253,
0.09412756562232971,
0.02142420969903469,
0.05197415128350258,
0.12995286285877228,
-0.022686492651700974,
0.09772004932165146,
0.07303693890571594,
0.09985987842082977,
0.11060800403356552,
0.19985371828079224,
0.022886212915182114,
-0.19676423072814941,
0.0380873903632164,
-0.07859895378351212,
-0.0053507364355027676,
0.12146519124507904,
0.14281919598579407,
-0.09727081656455994,
0.09723988175392151,
-0.0014166681794449687,
-0.036095861345529556,
-0.032103247940540314,
-0.07407337427139282,
-0.015863366425037384,
0.04475326091051102,
0.04351950064301491,
0.06786411255598068,
-0.005140793044120073,
0.08499236404895782,
-0.25888389348983765,
0.01854773797094822,
0.04429004341363907,
-0.010532835498452187,
0.08963978290557861,
0.08659035712480545,
-0.0503561794757843,
0.133544921875,
-0.022494465112686157,
0.13020861148834229,
0.09404259920120239,
-0.09533175081014633,
-0.22292618453502655,
-0.06270451098680496,
0.08238474279642105,
0.17174527049064636,
0.08238210529088974,
-0.04212580993771553,
0.11285244673490524,
-0.08852370828390121,
0.009028825908899307,
0.026999270543456078,
-0.06745808571577072,
-0.0541369654238224,
0.06993507593870163,
0.10711865872144699,
0.058814842253923416,
-0.11843946576118469,
-0.02349478006362915,
0.02842799760401249,
0.03442508354783058,
0.06341962516307831,
0.009891163557767868,
0.16722379624843597,
0.02796054631471634,
-0.14623764157295227,
-0.045774128288030624,
0.14920903742313385,
0.03012506291270256,
-0.04215821996331215,
-0.20709463953971863,
-0.006747975014150143,
-0.08887778222560883,
-0.02198212593793869,
-0.04837455600500107,
0.049169208854436874,
0.017894232645630836,
0.1088389903306961,
-0.04266509786248207,
-0.09933976829051971,
-0.01135216560214758,
0.09378619492053986,
0.0346422903239727,
0.014278990216553211,
-0.007528449408710003,
-0.0004577430372592062,
0.1285746544599533,
0.05469144508242607,
-0.12714460492134094,
-0.06107940524816513,
-0.07115962356328964,
-0.038405947387218475,
-0.0369131825864315,
0.02906898967921734,
0.04100806638598442,
0.043637074530124664,
0.256840318441391,
-0.004247268196195364,
0.055150821805000305,
0.07816781103610992,
0.029096456244587898,
0.05928764492273331,
0.1027042493224144,
-0.05899541452527046,
-0.16195039451122284,
-0.010378924198448658,
0.08282309025526047,
-0.001994097838178277,
-0.03426366299390793,
-0.07769469916820526,
0.04409316927194595,
0.03231460228562355,
0.10194675624370575,
0.10213643312454224,
-0.00840142834931612,
-0.06900389492511749,
-0.06323889642953873,
0.2089644968509674,
-0.140780970454216,
0.04491811990737915,
0.014242668636143208,
-0.02261270396411419,
-0.03249955177307129,
0.01572984829545021,
0.02586018666625023,
-0.03010505624115467,
0.0799839049577713,
-0.07839398831129074,
-0.03527021035552025,
-0.12673307955265045,
-0.027471506968140602,
0.022427299991250038,
-0.003538058837875724,
-0.020284080877900124,
-0.028788061812520027,
-0.07818468660116196,
-0.09368987381458282,
0.11955387890338898,
-0.06332498788833618,
-0.04999423027038574,
-0.03507404401898384,
-0.08191268146038055,
0.029741330072283745,
0.037989094853401184,
0.09118469059467316,
-0.02362010069191456,
0.030886046588420868,
-0.011452100239694118,
0.06288884580135345,
0.05513912811875343,
0.03946930542588234,
-0.08444610983133316,
0.06066261604428291,
-0.20766597986221313,
0.09049645811319351,
-0.061090268194675446,
0.035395506769418716,
-0.16141952574253082,
-0.007028630934655666,
0.010053620673716068,
0.03627076745033264,
0.029977615922689438,
0.15956375002861023,
-0.21718010306358337,
-0.033392831683158875,
0.13398465514183044,
-0.10535142570734024,
-0.10920390486717224,
0.03313661739230156,
-0.05438750982284546,
0.1846301406621933,
0.021011749282479286,
-0.0004254789964761585,
0.07448633760213852,
-0.12248582392930984,
-0.023236358538269997,
-0.017138930037617683,
-0.02830614522099495,
0.0713268369436264,
0.081262968480587,
-0.09127437323331833,
0.018645839765667915,
0.012333624064922333,
-0.04975755140185356,
-0.027676379308104515,
-0.039810311049222946,
-0.1081002727150917,
-0.0050798640586435795,
-0.07177256047725677,
0.0032630744390189648,
-0.016572296619415283,
-0.08035643398761749,
0.001617362373508513,
-0.16802436113357544,
-0.02455195039510727,
0.07920140773057938,
0.00814704317599535,
-0.014599336311221123,
-0.09017311781644821,
0.05735816806554794,
-0.0606040433049202,
-0.027315329760313034,
-0.14675214886665344,
0.0052270544692873955,
0.013336344622075558,
-0.14710111916065216,
0.020299475640058517,
-0.10115809738636017,
0.06349264085292816,
0.011477353982627392,
-0.04672861844301224,
-0.04369935020804405,
0.00010367027425672859,
0.003724793205037713,
-0.053453478962183,
-0.23140744864940643,
-0.03374785929918289,
-0.044856321066617966,
0.15386763215065002,
-0.21619822084903717,
0.036581382155418396,
0.04023589566349983,
0.11894483119249344,
-0.0035392972640693188,
-0.05845631659030914,
0.02484537661075592,
-0.07670248299837112,
-0.039557602256536484,
-0.07007710635662079,
0.001572409993968904,
-0.0014461677055805922,
-0.04872008040547371,
0.016424696892499924,
-0.12396717816591263,
-0.06818132847547531,
0.11014950275421143,
0.04142379015684128,
-0.15492349863052368,
-0.0041915783658623695,
-0.030916975811123848,
-0.06000775843858719,
-0.05342598259449005,
-0.05972793325781822,
0.11303042620420456,
0.04413822293281555,
0.03973376750946045,
-0.07595038414001465,
-0.05902018025517464,
0.010918207466602325,
-0.029565786942839622,
-0.016297759488224983,
0.093429334461689,
0.0999373197555542,
-0.12134627997875214,
0.0989208072423935,
0.07267444580793381,
0.03332529962062836,
0.08463598042726517,
-0.010384823195636272,
-0.10775253921747208,
-0.031286682933568954,
0.028272075578570366,
0.002783637959510088,
0.16402263939380646,
-0.07852847874164581,
0.05485396459698677,
0.04151233285665512,
-0.02716772072017193,
0.05665498599410057,
-0.0957925021648407,
0.01761704683303833,
0.021760543808341026,
-0.005722802598029375,
0.007382235489785671,
-0.030695531517267227,
-0.00876180361956358,
0.07580762356519699,
0.06536837667226791,
0.03976568952202797,
0.033472709357738495,
-0.029367417097091675,
-0.13732430338859558,
0.1905110627412796,
-0.10305890440940857,
-0.22935202717781067,
-0.1717958003282547,
0.05177067220211029,
0.05311822518706322,
-0.006569376215338707,
0.025543777272105217,
-0.06189529225230217,
-0.10580533742904663,
-0.08138279616832733,
0.018486447632312775,
0.006763557903468609,
-0.06118743494153023,
-0.09133443981409073,
0.039310213178396225,
0.03854845091700554,
-0.130089670419693,
0.03713662177324295,
0.05565035715699196,
-0.013944907113909721,
-0.01169645506888628,
0.04666148126125336,
0.09532339870929718,
0.19625136256217957,
-0.007600754965096712,
-0.008414373733103275,
0.06695323437452316,
0.291735976934433,
-0.15341155230998993,
0.12901893258094788,
0.12389722466468811,
-0.07051796466112137,
0.08391714096069336,
0.18495109677314758,
0.03306480497121811,
-0.09837296605110168,
0.020791195333003998,
0.02582281269133091,
-0.026949184015393257,
-0.2523637115955353,
-0.05266418308019638,
-0.006489538121968508,
-0.11327216774225235,
0.0706535056233406,
0.08739753067493439,
0.09179038554430008,
0.052998367697000504,
-0.06112697720527649,
-0.09166146069765091,
-0.0003765109577216208,
0.11145967245101929,
-0.04029975086450577,
0.002805144991725683,
0.07836277037858963,
-0.040786001831293106,
0.013674819841980934,
0.09837955981492996,
0.005413474980741739,
0.16164012253284454,
0.06552805751562119,
0.13401569426059723,
0.08540262281894684,
0.07983675599098206,
0.011559084989130497,
0.0339510552585125,
0.006372304633259773,
0.017902348190546036,
0.009317909367382526,
-0.07689813524484634,
0.023254239931702614,
0.11834190040826797,
0.040373314172029495,
0.045214686542749405,
0.011671608313918114,
-0.039621613919734955,
0.03956446796655655,
0.1767490804195404,
0.016947781667113304,
-0.2155885547399521,
-0.0772833302617073,
0.06542522460222244,
-0.058402981609106064,
-0.1495116949081421,
-0.025624670088291168,
0.02235453948378563,
-0.1576213240623474,
0.0005415278719738126,
-0.028421247377991676,
0.10273677110671997,
-0.09609103202819824,
-0.04047273099422455,
0.08817856013774872,
0.0699915885925293,
-0.028443265706300735,
0.062181491404771805,
-0.17871366441249847,
0.12371177971363068,
0.03400380536913872,
0.07102521508932114,
-0.09190616011619568,
0.09926209598779678,
-0.005890274420380592,
0.013706923462450504,
0.1655038744211197,
0.015541122294962406,
-0.09426835179328918,
-0.0710771307349205,
-0.08823360502719879,
-0.013685347512364388,
0.09967034310102463,
-0.13288703560829163,
0.06852756440639496,
-0.019996264949440956,
-0.027522722259163857,
0.006831855047494173,
-0.08690175414085388,
-0.13149002194404602,
-0.18112291395664215,
0.05532965436577797,
-0.10246208310127258,
0.024886637926101685,
-0.07404468208551407,
-0.04842938482761383,
0.040009692311286926,
0.19972540438175201,
-0.21909521520137787,
-0.10020548850297928,
-0.15154099464416504,
-0.11518421769142151,
0.16108576953411102,
-0.04635784775018692,
0.09108468145132065,
-0.01019457820802927,
0.1620863974094391,
0.010745878331363201,
-0.02071799710392952,
0.1160653606057167,
-0.0854450985789299,
-0.1714930683374405,
-0.05915606766939163,
0.14893034100532532,
0.14446774125099182,
0.035233963280916214,
-0.01287093386054039,
0.031883079558610916,
-0.07141809165477753,
-0.11891574412584305,
0.03534413129091263,
0.13700434565544128,
0.06963939964771271,
-0.01262744888663292,
-0.03579355776309967,
-0.09220988303422928,
-0.0504169799387455,
-0.03974350169301033,
0.008700315840542316,
0.18124674260616302,
-0.07448253035545349,
0.15189899504184723,
0.13152532279491425,
-0.0723627433180809,
-0.20356547832489014,
0.06049336493015289,
0.0346653088927269,
0.02138940617442131,
0.01630406267940998,
-0.21584440767765045,
0.08776868134737015,
-0.006339477840811014,
-0.06874293833971024,
0.18010607361793518,
-0.17902927100658417,
-0.13895957171916962,
0.0988360270857811,
0.03516041859984398,
-0.1823381632566452,
-0.13705183565616608,
-0.09613028168678284,
-0.03228107467293739,
-0.1230158656835556,
0.05866828188300133,
0.026329705491662025,
0.015535218641161919,
0.021184591576457024,
0.029537182301282883,
0.019990645349025726,
-0.050314560532569885,
0.2066401094198227,
-0.012754418887197971,
0.013829488307237625,
-0.06200092285871506,
-0.10324833542108536,
0.04607655853033066,
-0.05281443893909454,
0.11618918925523758,
0.0008675124263390899,
0.0222539734095335,
-0.1703120321035385,
-0.034940678626298904,
-0.05094180256128311,
0.03240950033068657,
-0.0940355733036995,
-0.09862573444843292,
-0.04792311042547226,
0.0863310769200325,
0.09178805351257324,
-0.02642832137644291,
-0.0012948049698024988,
-0.10240978002548218,
0.04736471548676491,
0.19468940794467926,
0.19447913765907288,
0.056417278945446014,
-0.06639297306537628,
0.028046250343322754,
-0.03318989276885986,
0.0474521666765213,
-0.24178913235664368,
0.03477860614657402,
0.05343414843082428,
0.011909419670701027,
0.08445286750793457,
-0.003811764298006892,
-0.16544894874095917,
-0.0645582303404808,
0.08673491328954697,
-0.044566020369529724,
-0.1641440987586975,
-0.032721146941185,
0.022641237825155258,
-0.20684140920639038,
-0.04179441183805466,
0.011281585320830345,
-0.019901549443602562,
-0.0412454716861248,
0.019307231530547142,
0.07510565966367722,
-0.03287685289978981,
0.08019816875457764,
0.09813148528337479,
0.08825678378343582,
-0.10000404715538025,
0.08111211657524109,
0.06777224689722061,
-0.04150259494781494,
0.033621978014707565,
0.10420102626085281,
-0.04986701160669327,
-0.04245395585894585,
0.08457721024751663,
0.12530498206615448,
-0.023088322952389717,
-0.05413966253399849,
0.01212761178612709,
-0.04834878444671631,
0.054270725697278976,
0.10672589391469955,
0.03587748110294342,
-0.0011736709857359529,
0.050750378519296646,
0.028017738834023476,
-0.10256616771221161,
0.08914750814437866,
0.03725229576230049,
0.01791483536362648,
-0.03840089589357376,
-0.04189951717853546,
0.004631043411791325,
-0.01516848523169756,
-0.018755726516246796,
-0.0170601699501276,
-0.08432288467884064,
-0.012585177086293697,
-0.11483073979616165,
0.008729316294193268,
-0.06474046409130096,
0.0068718683905899525,
0.030621705576777458,
-0.048198994249105453,
0.002455118577927351,
0.0015593849821016192,
-0.0763937458395958,
-0.051290228962898254,
-0.013947847299277782,
0.06659863144159317,
-0.12318508327007294,
0.042245566844940186,
0.06755290925502777,
-0.0967436209321022,
0.06653253734111786,
-0.007241120561957359,
0.011410431936383247,
0.0035017048940062523,
-0.15551216900348663,
0.04931795224547386,
-0.02801262028515339,
-0.024408893659710884,
0.02252740040421486,
-0.1943521499633789,
-0.0076536573469638824,
-0.04313570633530617,
-0.0573619082570076,
-0.004662544000893831,
-0.010509601794183254,
-0.11749584227800369,
0.10912971943616867,
0.007869033142924309,
-0.06068027764558792,
-0.027412936091423035,
0.04882120341062546,
0.10086818784475327,
-0.02643461339175701,
0.13437911868095398,
-0.007259611040353775,
0.07193886488676071,
-0.16531120240688324,
-0.004601365886628628,
-0.012241186574101448,
0.0436379611492157,
-0.026195699349045753,
-0.0405074842274189,
0.046567026525735855,
-0.02435867115855217,
0.19325846433639526,
-0.022945057600736618,
0.07084392011165619,
0.04857128486037254,
0.032133232802152634,
0.015501349233090878,
0.0795411467552185,
0.07082084566354752,
-0.005705251824110746,
0.0012581591727212071,
0.03978053480386734,
0.017982542514801025,
-0.03728210926055908,
-0.1555383801460266,
0.06970943510532379,
0.13411745429039001,
0.06166819855570793,
0.04408809542655945,
0.016431381925940514,
-0.10990120470523834,
-0.0851396843791008,
0.11883285641670227,
-0.007235993165522814,
-0.03617050126194954,
-0.06722866743803024,
0.173916757106781,
0.14665542542934418,
-0.1881304681301117,
0.07379920780658722,
-0.04021890461444855,
-0.047917887568473816,
-0.1390228271484375,
-0.19778694212436676,
-0.05708994343876839,
-0.04697444662451744,
-0.031041637063026428,
-0.06054393947124481,
0.0458466075360775,
0.05282822251319885,
-0.0030727433040738106,
-0.022938158363103867,
0.09914897382259369,
0.015943726524710655,
-0.02539999410510063,
0.02896830625832081,
0.05823741853237152,
0.03165817633271217,
-0.08659189939498901,
0.015606595203280449,
0.005570207256823778,
0.012911485508084297,
0.06870248168706894,
0.02457418665289879,
-0.05173683166503906,
0.027958450838923454,
-0.022081928327679634,
-0.11900684982538223,
0.030582444742321968,
-0.008381795138120651,
-0.040503326803445816,
0.13942766189575195,
0.027071574702858925,
0.004188410937786102,
-0.01948714442551136,
0.21969179809093475,
-0.07356120645999908,
-0.06005077436566353,
-0.13661882281303406,
0.08630871772766113,
-0.0646464005112648,
0.0389556810259819,
0.019435638561844826,
-0.1268240362405777,
0.018831565976142883,
0.177330881357193,
0.1440054327249527,
-0.01947030983865261,
0.0009323913836851716,
0.04481814056634903,
0.005301912315189838,
-0.0307242963463068,
0.02539828233420849,
0.04486451670527458,
0.15326927602291107,
-0.08694307506084442,
0.06786760687828064,
-0.017319831997156143,
-0.0827098861336708,
-0.012140425853431225,
0.11545486748218536,
-0.006142809521406889,
0.0005785097600892186,
-0.06454599648714066,
0.1289747804403305,
-0.09516481310129166,
-0.20264828205108643,
0.06106545776128769,
-0.06001608073711395,
-0.13382460176944733,
-0.04726420342922211,
0.03044959343969822,
-0.012021848000586033,
0.015416436828672886,
0.06878575682640076,
-0.056695982813835144,
0.18439672887325287,
0.04437119513750076,
-0.07817023992538452,
-0.10054308921098709,
0.05510885640978813,
-0.15851499140262604,
0.2766340374946594,
0.028239939361810684,
0.029272811487317085,
0.10791875422000885,
-0.003789537586271763,
-0.14966876804828644,
0.014645657502114773,
0.09151527285575867,
-0.055869899690151215,
0.05673375353217125,
0.17499114573001862,
0.0023246139753609896,
0.11875680834054947,
0.04801327362656593,
-0.05847916379570961,
0.053358294069767,
-0.10834988206624985,
-0.05009477213025093,
-0.10316772758960724,
0.06604080647230148,
-0.08997134119272232,
0.1677444577217102,
0.12227477133274078,
-0.0657745748758316,
-0.012364364229142666,
-0.022101426497101784,
0.08321168273687363,
0.01727793924510479,
0.10294997692108154,
0.009740419685840607,
-0.16769815981388092,
0.037040822207927704,
0.015676027163863182,
0.09718561172485352,
-0.194975346326828,
-0.05438392236828804,
0.04106029495596886,
-0.019167637452483177,
-0.07238626480102539,
0.11234915256500244,
0.04907934367656708,
0.053642772138118744,
-0.04926076903939247,
-0.025769544765353203,
0.009724765084683895,
0.1444830447435379,
-0.1114528551697731,
-0.024292191490530968
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bert-base-cased-lora-591K-squad-model1
This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the squad dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 16
- seed: 51
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["varun-v-rao/squad"], "base_model": "bert-base-cased", "model-index": [{"name": "bert-base-cased-lora-591K-squad-model1", "results": []}]} | question-answering | varun-v-rao/bert-base-cased-lora-591K-squad-model1 | [
"transformers",
"tensorboard",
"safetensors",
"bert",
"question-answering",
"generated_from_trainer",
"dataset:varun-v-rao/squad",
"base_model:bert-base-cased",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-08T09:10:35+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #bert #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-bert-base-cased #license-apache-2.0 #endpoints_compatible #region-us
|
# bert-base-cased-lora-591K-squad-model1
This model is a fine-tuned version of bert-base-cased on the squad dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 16
- seed: 51
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| [
"# bert-base-cased-lora-591K-squad-model1\n\nThis model is a fine-tuned version of bert-base-cased on the squad dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 51\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #bert #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-bert-base-cased #license-apache-2.0 #endpoints_compatible #region-us \n",
"# bert-base-cased-lora-591K-squad-model1\n\nThis model is a fine-tuned version of bert-base-cased on the squad dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 51\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
73,
44,
6,
12,
8,
3,
90,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #bert #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-bert-base-cased #license-apache-2.0 #endpoints_compatible #region-us \n# bert-base-cased-lora-591K-squad-model1\n\nThis model is a fine-tuned version of bert-base-cased on the squad dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 51\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3### Training results### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
-0.09145665913820267,
0.1660051792860031,
-0.0029817998874932528,
0.09280429035425186,
0.11671546846628189,
0.004390701185911894,
0.10629177838563919,
0.15111778676509857,
-0.0675644800066948,
0.08697506785392761,
0.06444214284420013,
0.026290442794561386,
0.05784488841891289,
0.12048704922199249,
-0.031211331486701965,
-0.19845403730869293,
0.013801941648125648,
-0.011826681904494762,
-0.07118812203407288,
0.0879519060254097,
0.10054947435855865,
-0.11165314167737961,
0.07726491242647171,
-0.017734669148921967,
-0.09581656008958817,
0.04651709645986557,
-0.034399889409542084,
-0.04877108708024025,
0.08729628473520279,
0.008080222643911839,
0.08139409869909286,
0.008248022757470608,
0.13319095969200134,
-0.2516491413116455,
0.0021197765599936247,
0.07818868011236191,
0.025988714769482613,
0.08323834836483002,
0.03778867423534393,
0.006531683262437582,
0.0297104362398386,
-0.16379204392433167,
0.10031075775623322,
0.022574778646230698,
-0.07024310529232025,
-0.16653519868850708,
-0.1046714186668396,
0.06734948605298996,
0.10271524637937546,
0.08093757182359695,
0.003559364704415202,
0.14861398935317993,
-0.05486508831381798,
0.07658088207244873,
0.21914561092853546,
-0.2931159436702728,
-0.056693557649850845,
0.05101078376173973,
0.05803021043539047,
0.08137007802724838,
-0.11715271323919296,
0.0007829091628082097,
0.05580465868115425,
0.01589070074260235,
0.09507642686367035,
-0.012757755815982819,
-0.07955682277679443,
0.017287088558077812,
-0.127811998128891,
-0.026978997513651848,
0.16925756633281708,
0.05709536001086235,
-0.0457116924226284,
-0.09463652968406677,
-0.047045525163412094,
-0.06701275706291199,
-0.017166707664728165,
-0.05603046342730522,
0.044827692210674286,
-0.0559227429330349,
-0.056906457990407944,
-0.054353464394807816,
-0.08303539454936981,
-0.07804267853498459,
0.023123102262616158,
0.054408930242061615,
0.05555896461009979,
0.02121456153690815,
-0.039691582322120667,
0.0901472270488739,
-0.03046628274023533,
-0.1336132138967514,
-0.030397774651646614,
0.011007776483893394,
-0.0861649289727211,
-0.057553742080926895,
-0.006840275134891272,
-0.0244199987500906,
0.016570135951042175,
0.16560322046279907,
-0.04579909145832062,
0.06146160140633583,
-0.02578282728791237,
-0.004333978518843651,
-0.017551256343722343,
0.1397520899772644,
-0.04843012988567352,
-0.050872113555669785,
0.0058160992339253426,
0.09792908281087875,
0.022034574300050735,
-0.0027093314565718174,
-0.07942412793636322,
-0.009616612456738949,
0.08829879760742188,
0.07880283892154694,
-0.03584225848317146,
0.020688125863671303,
-0.028324462473392487,
-0.01158800721168518,
0.022667746990919113,
-0.13910521566867828,
0.058512091636657715,
-0.008570229634642601,
-0.06820516288280487,
-0.05903869867324829,
0.030984751880168915,
0.00032727644429542124,
-0.02368936501443386,
0.058350324630737305,
-0.06630437076091766,
-0.02022417262196541,
-0.07554221898317337,
-0.06268669664859772,
0.04186098650097847,
-0.07740102708339691,
-0.006977734621614218,
-0.05956925079226494,
-0.2046472132205963,
-0.023139208555221558,
0.0295933298766613,
-0.07117064297199249,
-0.02472248673439026,
-0.041124701499938965,
-0.06285157799720764,
-0.0003666800621431321,
-0.01291185524314642,
0.12166857719421387,
-0.043461721390485764,
0.06947727501392365,
0.01940402202308178,
0.04576105996966362,
0.040461134165525436,
0.03516017273068428,
-0.08838832378387451,
0.04275169223546982,
-0.1297949254512787,
0.046539608389139175,
-0.1168774664402008,
0.022309618070721626,
-0.13904216885566711,
-0.08238620311021805,
0.007696334272623062,
-0.025785380974411964,
0.07196641713380814,
0.13060395419597626,
-0.16921964287757874,
-0.00755421444773674,
0.16347715258598328,
-0.08347325772047043,
-0.10818089544773102,
0.10602134466171265,
-0.04882420599460602,
0.025995930656790733,
0.07495675981044769,
0.15402810275554657,
0.09387337416410446,
-0.17572788894176483,
-0.036799926310777664,
0.010211837477982044,
0.08337004482746124,
0.020225955173373222,
0.0713491141796112,
-0.008693939074873924,
0.036120105534791946,
0.016049526631832123,
-0.08881013095378876,
-0.02851780317723751,
-0.07378478348255157,
-0.0911094918847084,
-0.05602952092885971,
-0.09361710399389267,
0.032056622207164764,
0.03758351877331734,
0.024856338277459145,
-0.08120514452457428,
-0.1217123344540596,
0.0866333544254303,
0.12403310835361481,
-0.0540509857237339,
0.016014087945222855,
-0.08898225426673889,
0.06484893709421158,
-0.05613077059388161,
-0.022471996024250984,
-0.16844512522220612,
-0.12287922203540802,
0.04858829826116562,
-0.049300987273454666,
0.025648411363363266,
0.015061939135193825,
0.07423029839992523,
0.06112411618232727,
-0.06474551558494568,
-0.024654243141412735,
-0.07445751130580902,
0.004788283258676529,
-0.10367312282323837,
-0.18560738861560822,
-0.051945485174655914,
-0.04564296454191208,
0.1178499162197113,
-0.22507138550281525,
0.025682874023914337,
0.023752570152282715,
0.1487913727760315,
0.03628892824053764,
-0.04429004341363907,
0.0029025846160948277,
0.021238679066300392,
-0.0005296070012263954,
-0.08347602933645248,
0.01909334771335125,
-0.015277876518666744,
-0.07091713696718216,
-0.06200295686721802,
-0.11318761110305786,
0.06277792155742645,
0.07024538516998291,
0.0908038392663002,
-0.07965172827243805,
-0.017579568549990654,
-0.05002474784851074,
-0.036848776042461395,
-0.09869252890348434,
-0.03244492784142494,
0.15455545485019684,
0.022047139704227448,
0.11573846638202667,
-0.07109841704368591,
-0.07494781911373138,
0.0026120082475245,
0.0018566105281934142,
-0.024289729073643684,
0.0928829237818718,
0.048534780740737915,
-0.0985473096370697,
0.1107703149318695,
0.12034276127815247,
-0.022959990426898003,
0.10810916870832443,
-0.0666101947426796,
-0.09959910064935684,
-0.0327867791056633,
0.02447476238012314,
-0.006884296424686909,
0.15369795262813568,
-0.07778751105070114,
-0.004033091943711042,
0.027308426797389984,
0.0019260038388893008,
0.009678558446466923,
-0.15997429192066193,
-0.005516742821782827,
0.026487218216061592,
-0.061508286744356155,
-0.003420822089537978,
-0.0321362167596817,
0.02020939812064171,
0.08879369497299194,
0.02054458297789097,
-0.01931929402053356,
0.017276322469115257,
-0.0183304063975811,
-0.07922009378671646,
0.16570766270160675,
-0.09967232495546341,
-0.13470788300037384,
-0.12885813415050507,
0.044185325503349304,
-0.03491218760609627,
-0.02615731582045555,
0.02010609209537506,
-0.09248676896095276,
-0.06455094367265701,
-0.10985006392002106,
-0.026966148987412453,
-0.008190599270164967,
-0.014498086646199226,
0.06328032165765762,
0.018464745953679085,
0.0971793532371521,
-0.1366879791021347,
0.013799340464174747,
-0.007510711904615164,
-0.09601718932390213,
-0.029115937650203705,
0.04948752745985985,
0.11903464794158936,
0.08188574016094208,
-0.017727678641676903,
0.027519874274730682,
-0.0369151271879673,
0.20679697394371033,
-0.06886313110589981,
0.011110608465969563,
0.1052875816822052,
-0.009628264233469963,
0.055909737944602966,
0.13557027280330658,
0.027518251910805702,
-0.08990596979856491,
0.028264328837394714,
0.08525176346302032,
-0.016259318217635155,
-0.2606695592403412,
-0.022724421694874763,
-0.011792506091296673,
-0.03766883164644241,
0.0892973244190216,
0.06704238802194595,
0.009523928165435791,
0.03977404907345772,
-0.01579391211271286,
0.010432560928165913,
-0.0031971472781151533,
0.08382375538349152,
0.07308299094438553,
0.013873139396309853,
0.08645132929086685,
-0.04078157991170883,
-0.04023109748959541,
0.05808248370885849,
0.05081052705645561,
0.2577988803386688,
-0.012560753151774406,
0.133182093501091,
0.026073815301060677,
0.16475364565849304,
-0.04668049514293671,
0.03154384344816208,
-0.0031862773466855288,
0.009365365840494633,
0.004379604943096638,
-0.07185056060552597,
0.006717739626765251,
0.051159393042325974,
-0.038435135036706924,
0.04764655977487564,
-0.06873705983161926,
0.03177206590771675,
0.04026756435632706,
0.2614749073982239,
0.04660721495747566,
-0.2569739818572998,
-0.06636067479848862,
0.04295595735311508,
-0.03900957480072975,
-0.048842087388038635,
0.013892294839024544,
0.13997425138950348,
-0.10975039750337601,
0.05013594031333923,
-0.050070058554410934,
0.09314840286970139,
-0.02717219479382038,
-0.0007931943982839584,
0.038047607988119125,
0.08716284483671188,
-0.0022323974408209324,
0.09650150686502457,
-0.1995699405670166,
0.2180471569299698,
0.03369693458080292,
0.11500391364097595,
-0.06620896607637405,
0.036354489624500275,
-0.0005259591853246093,
0.05951935425400734,
0.15594051778316498,
-0.01667613349854946,
-0.063724085688591,
-0.17310768365859985,
-0.10255970805883408,
0.03598375618457794,
0.10021356493234634,
-0.04495720937848091,
0.0901576578617096,
-0.04052935913205147,
-0.015054555609822273,
0.046837568283081055,
-0.06069498881697655,
-0.15140685439109802,
-0.11550844460725784,
0.0076636821031570435,
0.0033643231727182865,
-0.042974911630153656,
-0.0899137482047081,
-0.10223189741373062,
-0.064478799700737,
0.1501190960407257,
-0.018007101491093636,
-0.03879678621888161,
-0.12762755155563354,
0.0694400891661644,
0.11952345073223114,
-0.06405584514141083,
0.0037040256429463625,
0.02605203539133072,
0.1380012184381485,
0.032432958483695984,
-0.07541147619485855,
0.04910273104906082,
-0.06290818005800247,
-0.16640110313892365,
-0.059590455144643784,
0.1486223340034485,
0.05736491456627846,
0.046736035495996475,
0.016614170745015144,
0.026204152032732964,
0.02571270614862442,
-0.07411963492631912,
0.02616141550242901,
0.06889284402132034,
0.0932641327381134,
0.03283137083053589,
-0.0980255976319313,
0.009374269284307957,
-0.05391751229763031,
-0.01696218177676201,
0.11473339051008224,
0.20889577269554138,
-0.09034237265586853,
0.0915900468826294,
0.08454623073339462,
-0.09082350134849548,
-0.18846139311790466,
0.06403486430644989,
0.06964169442653656,
0.011703482829034328,
0.07061884552240372,
-0.1647171974182129,
0.12139187008142471,
0.09167890250682831,
-0.03418569639325142,
0.033776625990867615,
-0.2863713800907135,
-0.1271096169948578,
0.09040322154760361,
0.10588310658931732,
-0.01738196611404419,
-0.15412573516368866,
-0.04783160984516144,
-0.015941603109240532,
-0.1338827908039093,
0.09554871171712875,
-0.13019342720508575,
0.07382066547870636,
0.005663967691361904,
0.08402927964925766,
0.0253960769623518,
-0.04499555751681328,
0.13747698068618774,
0.037892572581768036,
0.08492495864629745,
-0.05236069858074188,
0.00041205811430700123,
0.1085749939084053,
-0.07538745552301407,
0.08376414328813553,
-0.05132504552602768,
0.06735744327306747,
-0.15163522958755493,
-0.01909630373120308,
-0.051385849714279175,
0.05474051460623741,
-0.06163620948791504,
-0.0495167039334774,
-0.05323905870318413,
0.06652254611253738,
0.059205811470746994,
-0.03186574950814247,
0.0988372415304184,
0.025434697046875954,
0.08940666168928146,
0.12815137207508087,
0.10003890097141266,
0.01945003867149353,
-0.10050805658102036,
0.017391318455338478,
-0.03363418951630592,
0.05812736228108406,
-0.12432321906089783,
0.04603810980916023,
0.12549439072608948,
0.04160672798752785,
0.13752543926239014,
0.010428905487060547,
-0.07156752049922943,
-0.013552659191191196,
0.0280831977725029,
-0.11359985917806625,
-0.18974623084068298,
0.0035147438757121563,
-0.016436759382486343,
-0.15480434894561768,
0.038098227232694626,
0.10061699897050858,
-0.053535476326942444,
-0.01771377958357334,
-0.012935851700603962,
0.041097477078437805,
-0.014263797551393509,
0.1717657893896103,
0.05837281420826912,
0.0639333426952362,
-0.07203496992588043,
0.12564945220947266,
0.07758626341819763,
-0.06677673757076263,
0.06917960941791534,
0.053780753165483475,
-0.07375095784664154,
-0.024056553840637207,
0.06639482080936432,
0.19825369119644165,
0.005799584556370974,
-0.052293017506599426,
-0.09613093733787537,
-0.07053159922361374,
0.03873119130730629,
0.1374775916337967,
0.04221177101135254,
-0.0179615318775177,
-0.009610236622393131,
0.036069106310606,
-0.12783591449260712,
0.128404438495636,
0.04966358467936516,
0.060174666345119476,
-0.13455122709274292,
0.05638789385557175,
-0.00413327245041728,
0.035473328083753586,
-0.021055275574326515,
0.032466303557157516,
-0.09395206719636917,
-0.012891927734017372,
-0.1451415866613388,
0.003957478795200586,
-0.02584308199584484,
0.0034957807511091232,
-0.011020053178071976,
-0.07687291502952576,
-0.03300953656435013,
0.051427315920591354,
-0.06220962852239609,
-0.04614272341132164,
0.02082233689725399,
0.06354821473360062,
-0.18188641965389252,
-0.02623099274933338,
0.038003481924533844,
-0.08680582046508789,
0.07234078645706177,
0.03352215886116028,
0.029553474858403206,
0.02730540744960308,
-0.11043909192085266,
0.007711225189268589,
0.00918651930987835,
0.040219299495220184,
0.05664706602692604,
-0.1136547178030014,
-0.012853986583650112,
-0.02449863776564598,
0.030566509813070297,
0.01967812143266201,
0.050437748432159424,
-0.11145022511482239,
-0.021296028047800064,
-0.06210261210799217,
-0.05302957445383072,
-0.041444748640060425,
0.053153906017541885,
0.11417251825332642,
0.020950941368937492,
0.1487416923046112,
-0.07881424576044083,
0.04802417755126953,
-0.20510472357273102,
-0.020216794684529305,
0.009800033643841743,
-0.03522520512342453,
-0.08107735216617584,
-0.030871333554387093,
0.06907997280359268,
-0.06952732801437378,
0.11070915311574936,
-0.015469647012650967,
0.09876193851232529,
0.05025225877761841,
-0.0528772808611393,
-0.005251229275017977,
0.017517343163490295,
0.1625751405954361,
0.039989013224840164,
-0.01990194246172905,
0.06946595013141632,
-0.037316907197237015,
0.05274542048573494,
0.023625662550330162,
0.140838161110878,
0.16959233582019806,
-0.007616848219186068,
0.04859969764947891,
0.09470079839229584,
-0.10530734062194824,
-0.1312195062637329,
0.0915152058005333,
-0.026293747127056122,
0.09066522866487503,
-0.05672313645482063,
0.17662037909030914,
0.09835367649793625,
-0.17967025935649872,
0.04663502052426338,
-0.07205681502819061,
-0.10909485816955566,
-0.11330567300319672,
-0.06595110148191452,
-0.09021435678005219,
-0.10291185230016708,
0.031098224222660065,
-0.1252318024635315,
0.02521609328687191,
0.07703409343957901,
0.008767733350396156,
0.004325150512158871,
0.17287133634090424,
-0.03308897465467453,
0.04312005266547203,
0.05245022475719452,
0.0252174511551857,
0.0015858742408454418,
-0.04795270413160324,
-0.03206152096390724,
0.0555771067738533,
0.01933816261589527,
0.061036061495542526,
-0.02664940059185028,
0.028571734204888344,
0.022097207605838776,
-0.005393750499933958,
-0.07057905197143555,
0.007833053357899189,
0.019191360101103783,
0.03371742367744446,
0.07239913195371628,
0.055280651897192,
0.011871603317558765,
-0.03961801901459694,
0.26016485691070557,
-0.07718027383089066,
-0.05338997021317482,
-0.140303835272789,
0.14940208196640015,
0.02263382077217102,
0.0051595838740468025,
0.06329802423715591,
-0.1299966424703598,
0.0013013430871069431,
0.16523133218288422,
0.12976112961769104,
-0.032368022948503494,
-0.010496501810848713,
-0.019317762926220894,
-0.010521505028009415,
-0.05042126774787903,
0.06775251030921936,
0.09762414544820786,
0.022380562499165535,
-0.06260353326797485,
-0.01268291100859642,
0.012647919356822968,
-0.03693702816963196,
-0.08047959953546524,
0.07444462180137634,
-0.0001340600720141083,
0.02053201198577881,
-0.03882015869021416,
0.0735711008310318,
0.02208707109093666,
-0.2402333915233612,
0.044462889432907104,
-0.16980227828025818,
-0.17472687363624573,
-0.0077568963170051575,
0.10069070756435394,
-0.017239047214388847,
0.02574014477431774,
-0.010121779516339302,
0.010992661118507385,
0.14653588831424713,
-0.006594808306545019,
-0.055747486650943756,
-0.11328625679016113,
0.09564412385225296,
-0.09107719361782074,
0.24757733941078186,
0.005789581220597029,
0.06492247432470322,
0.10480242222547531,
-0.020771292969584465,
-0.1464427411556244,
0.028553389012813568,
0.08697522431612015,
-0.07312843948602676,
0.004960112273693085,
0.15227362513542175,
-0.046572767198085785,
0.12797030806541443,
0.04773876070976257,
-0.0947248563170433,
-0.02512577921152115,
-0.027830835431814194,
-0.028790893033146858,
-0.1000073105096817,
0.017531707882881165,
-0.07571202516555786,
0.15763555467128754,
0.17559567093849182,
-0.040422286838293076,
0.014893118292093277,
-0.06843883544206619,
0.03847827762365341,
0.05455382168292999,
0.05994177237153053,
0.004516283515840769,
-0.18975305557250977,
0.030369052663445473,
0.028467733412981033,
0.03564273193478584,
-0.252509206533432,
-0.09868576377630234,
0.06167072057723999,
-0.0299521554261446,
-0.05894114822149277,
0.0923638790845871,
0.10303865373134613,
0.041672054678201675,
-0.04002191126346588,
-0.147108793258667,
-0.04393041878938675,
0.1353960484266281,
-0.14820179343223572,
-0.03703603893518448
] |
null | null | transformers |
# zephyr-NCA-preference
This model is a fine-tuned version of [HuggingFaceH4/mistral-7b-sft-beta](https://huggingface.co/HuggingFaceH4/mistral-7b-sft-beta) on the [openbmb/UltraFeedback](https://huggingface.co/datasets/openbmb/UltraFeedback) dataset.
It achieves the following results on the evaluation set:
- Loss: 1.3030
- Rewards/chosen: 0.0489
- Rewards/rejected: -0.5399
- Rewards/accuracies: 0.7820
- Rewards/margins: 0.5888
- Verify/constant 1: 1.0
- Verify/constant 1len: 1000.0
- Logps/rejected: -287.1594
- Logps/chosen: -270.2584
- Verify/bz: 1.0
- Verify/gather Bz: 2.0
- Regularization/forward Kl: 0.6109
- Regularization/reverse Kl: 0.4631
- Regularization/policy Data Loss: 1.8007
- Regularization/reference Data Loss: 1.3337
- Regularization/policy Ref Data Loss Gap: 0.4670
- Mask/mask Ratio: 0.4809
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-06
- train_batch_size: 1
- eval_batch_size: 1
- seed: 42
- distributed_type: multi-GPU
- num_devices: 2
- gradient_accumulation_steps: 16
- total_train_batch_size: 32
- total_eval_batch_size: 2
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 1
### Training results
| Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Verify/constant 1 | Verify/constant 1len | Logps/rejected | Logps/chosen | Verify/bz | Verify/gather Bz | Regularization/forward Kl | Regularization/reverse Kl | Regularization/policy Data Loss | Regularization/reference Data Loss | Regularization/policy Ref Data Loss Gap | Mask/mask Ratio |
|:-------------:|:-----:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:-----------------:|:--------------------:|:--------------:|:------------:|:---------:|:----------------:|:-------------------------:|:-------------------------:|:-------------------------------:|:----------------------------------:|:---------------------------------------:|:---------------:|
| 1.3844 | 0.05 | 100 | 1.3839 | 0.0037 | -0.0061 | 0.7075 | 0.0098 | 1.0 | 1000.0 | -233.7844 | -274.7838 | 1.0 | 2.0 | 0.0009 | 0.0009 | 1.3404 | 1.3337 | 0.0067 | 0.4809 |
| 1.3593 | 0.1 | 200 | 1.3605 | -0.0445 | -0.1811 | 0.7320 | 0.1366 | 1.0 | 1000.0 | -251.2808 | -279.5988 | 1.0 | 2.0 | 0.1063 | 0.0867 | 1.4942 | 1.3337 | 0.1604 | 0.4809 |
| 1.3432 | 0.15 | 300 | 1.3399 | -0.0181 | -0.2809 | 0.7695 | 0.2628 | 1.0 | 1000.0 | -261.2633 | -276.9577 | 1.0 | 2.0 | 0.2787 | 0.2104 | 1.5199 | 1.3337 | 0.1862 | 0.4809 |
| 1.3404 | 0.21 | 400 | 1.3251 | 0.0042 | -0.3854 | 0.7720 | 0.3896 | 1.0 | 1000.0 | -271.7116 | -274.7323 | 1.0 | 2.0 | 0.5454 | 0.4274 | 1.5819 | 1.3337 | 0.2481 | 0.4809 |
| 1.3295 | 0.26 | 500 | 1.3173 | 0.0213 | -0.4300 | 0.7770 | 0.4513 | 1.0 | 1000.0 | -276.1767 | -273.0250 | 1.0 | 2.0 | 0.5684 | 0.4290 | 1.6808 | 1.3337 | 0.3471 | 0.4809 |
| 1.3187 | 0.31 | 600 | 1.3122 | 0.0267 | -0.4649 | 0.7790 | 0.4917 | 1.0 | 1000.0 | -279.6683 | -272.4786 | 1.0 | 2.0 | 0.5839 | 0.4556 | 1.7090 | 1.3337 | 0.3753 | 0.4809 |
| 1.3105 | 0.36 | 700 | 1.3106 | 0.0180 | -0.5079 | 0.7685 | 0.5259 | 1.0 | 1000.0 | -283.9655 | -273.3516 | 1.0 | 2.0 | 0.5818 | 0.4701 | 1.8137 | 1.3337 | 0.4800 | 0.4809 |
| 1.3086 | 0.41 | 800 | 1.3094 | 0.0287 | -0.5003 | 0.7820 | 0.5290 | 1.0 | 1000.0 | -283.2076 | -272.2820 | 1.0 | 2.0 | 0.5724 | 0.4410 | 1.7950 | 1.3337 | 0.4613 | 0.4809 |
| 1.3164 | 0.46 | 900 | 1.3071 | 0.0494 | -0.4863 | 0.7865 | 0.5356 | 1.0 | 1000.0 | -281.7993 | -270.2156 | 1.0 | 2.0 | 0.5937 | 0.4471 | 1.6937 | 1.3337 | 0.3599 | 0.4809 |
| 1.3065 | 0.52 | 1000 | 1.3058 | 0.0442 | -0.5122 | 0.7875 | 0.5564 | 1.0 | 1000.0 | -284.3954 | -270.7371 | 1.0 | 2.0 | 0.6214 | 0.4609 | 1.7262 | 1.3337 | 0.3925 | 0.4809 |
| 1.3274 | 0.57 | 1100 | 1.3097 | 0.0187 | -0.5605 | 0.7765 | 0.5792 | 1.0 | 1000.0 | -289.2202 | -273.2801 | 1.0 | 2.0 | 0.6048 | 0.4467 | 1.9267 | 1.3337 | 0.5930 | 0.4809 |
| 1.3128 | 0.62 | 1200 | 1.3053 | 0.0391 | -0.5393 | 0.7795 | 0.5784 | 1.0 | 1000.0 | -287.1077 | -271.2448 | 1.0 | 2.0 | 0.5974 | 0.4596 | 1.8496 | 1.3337 | 0.5159 | 0.4809 |
| 1.3018 | 0.67 | 1300 | 1.3043 | 0.0370 | -0.5532 | 0.7765 | 0.5902 | 1.0 | 1000.0 | -288.4903 | -271.4501 | 1.0 | 2.0 | 0.6164 | 0.4737 | 1.8233 | 1.3337 | 0.4896 | 0.4809 |
| 1.3137 | 0.72 | 1400 | 1.3040 | 0.0532 | -0.5183 | 0.7790 | 0.5715 | 1.0 | 1000.0 | -285.0031 | -269.8345 | 1.0 | 2.0 | 0.5985 | 0.4642 | 1.7409 | 1.3337 | 0.4072 | 0.4809 |
| 1.304 | 0.77 | 1500 | 1.3034 | 0.0489 | -0.5344 | 0.7815 | 0.5833 | 1.0 | 1000.0 | -286.6187 | -270.2639 | 1.0 | 2.0 | 0.6056 | 0.4668 | 1.7960 | 1.3337 | 0.4623 | 0.4809 |
| 1.3194 | 0.83 | 1600 | 1.3033 | 0.0496 | -0.5367 | 0.7770 | 0.5864 | 1.0 | 1000.0 | -286.8489 | -270.1884 | 1.0 | 2.0 | 0.6093 | 0.4660 | 1.7863 | 1.3337 | 0.4526 | 0.4809 |
| 1.3194 | 0.88 | 1700 | 1.3030 | 0.0498 | -0.5367 | 0.7820 | 0.5865 | 1.0 | 1000.0 | -286.8430 | -270.1689 | 1.0 | 2.0 | 0.6106 | 0.4640 | 1.7905 | 1.3337 | 0.4568 | 0.4809 |
| 1.32 | 0.93 | 1800 | 1.3031 | 0.0475 | -0.5425 | 0.7815 | 0.5901 | 1.0 | 1000.0 | -287.4280 | -270.3985 | 1.0 | 2.0 | 0.6118 | 0.4635 | 1.8042 | 1.3337 | 0.4705 | 0.4809 |
| 1.3119 | 0.98 | 1900 | 1.3030 | 0.0490 | -0.5398 | 0.7810 | 0.5888 | 1.0 | 1000.0 | -287.1560 | -270.2523 | 1.0 | 2.0 | 0.6107 | 0.4630 | 1.8007 | 1.3337 | 0.4670 | 0.4809 |
### Framework versions
- Transformers 4.35.0
- Pytorch 2.1.0
- Datasets 2.14.6
- Tokenizers 0.14.1
| {} | text-generation | ChenDRAG/zephyr-NCA-preference | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"4-bit",
"region:us"
] | 2024-02-08T09:12:01+00:00 | [] | [] | TAGS
#transformers #safetensors #mistral #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us
| zephyr-NCA-preference
=====================
This model is a fine-tuned version of HuggingFaceH4/mistral-7b-sft-beta on the openbmb/UltraFeedback dataset.
It achieves the following results on the evaluation set:
* Loss: 1.3030
* Rewards/chosen: 0.0489
* Rewards/rejected: -0.5399
* Rewards/accuracies: 0.7820
* Rewards/margins: 0.5888
* Verify/constant 1: 1.0
* Verify/constant 1len: 1000.0
* Logps/rejected: -287.1594
* Logps/chosen: -270.2584
* Verify/bz: 1.0
* Verify/gather Bz: 2.0
* Regularization/forward Kl: 0.6109
* Regularization/reverse Kl: 0.4631
* Regularization/policy Data Loss: 1.8007
* Regularization/reference Data Loss: 1.3337
* Regularization/policy Ref Data Loss Gap: 0.4670
* Mask/mask Ratio: 0.4809
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-06
* train\_batch\_size: 1
* eval\_batch\_size: 1
* seed: 42
* distributed\_type: multi-GPU
* num\_devices: 2
* gradient\_accumulation\_steps: 16
* total\_train\_batch\_size: 32
* total\_eval\_batch\_size: 2
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: cosine
* lr\_scheduler\_warmup\_ratio: 0.1
* num\_epochs: 1
### Training results
### Framework versions
* Transformers 4.35.0
* Pytorch 2.1.0
* Datasets 2.14.6
* Tokenizers 0.14.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 1\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 2\n* gradient\\_accumulation\\_steps: 16\n* total\\_train\\_batch\\_size: 32\n* total\\_eval\\_batch\\_size: 2\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 1",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.0\n* Pytorch 2.1.0\n* Datasets 2.14.6\n* Tokenizers 0.14.1"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 1\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 2\n* gradient\\_accumulation\\_steps: 16\n* total\\_train\\_batch\\_size: 32\n* total\\_eval\\_batch\\_size: 2\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 1",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.0\n* Pytorch 2.1.0\n* Datasets 2.14.6\n* Tokenizers 0.14.1"
] | [
54,
179,
4,
30
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 1\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 2\n* gradient\\_accumulation\\_steps: 16\n* total\\_train\\_batch\\_size: 32\n* total\\_eval\\_batch\\_size: 2\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 1### Training results### Framework versions\n\n\n* Transformers 4.35.0\n* Pytorch 2.1.0\n* Datasets 2.14.6\n* Tokenizers 0.14.1"
] | [
-0.09374158829450607,
0.0865808054804802,
-0.003508933586999774,
0.06578915566205978,
0.11688736081123352,
0.038814477622509,
0.1166367456316948,
0.12111333757638931,
-0.07969483733177185,
0.11100991070270538,
0.13099730014801025,
0.09454995393753052,
0.054377395659685135,
0.1595364362001419,
-0.04189189150929451,
-0.26775169372558594,
0.02015146054327488,
-0.04553069919347763,
-0.1383098065853119,
0.10957453399896622,
0.071271151304245,
-0.12765759229660034,
0.07583056390285492,
-0.031095115467905998,
-0.11601417511701584,
-0.03676983341574669,
-0.03391733393073082,
-0.032394129782915115,
0.12565301358699799,
0.03196001052856445,
0.095780149102211,
0.05023622885346413,
0.10451655089855194,
-0.25656643509864807,
0.007278406992554665,
0.07023212313652039,
0.024007685482501984,
0.08145192265510559,
0.08441093564033508,
-0.008788242004811764,
0.1570168137550354,
-0.11987771093845367,
0.06872227787971497,
0.030749376863241196,
-0.11133337020874023,
-0.18774950504302979,
-0.05946514382958412,
0.03228650242090225,
0.11749044805765152,
0.06546664237976074,
-0.021478749811649323,
0.09551989287137985,
-0.10657543689012527,
0.08345071226358414,
0.22237414121627808,
-0.26016971468925476,
-0.07135442644357681,
0.05746198445558548,
0.03416028246283531,
0.0765567198395729,
-0.13189202547073364,
0.000288498995359987,
0.0424414798617363,
0.006401175167411566,
0.09719745069742203,
0.007153216749429703,
0.037770505994558334,
0.023383593186736107,
-0.14649826288223267,
-0.04503883793950081,
0.08802426606416702,
0.055452294647693634,
-0.010574202984571457,
-0.0875718742609024,
-0.04920126870274544,
-0.18907901644706726,
-0.04219051077961922,
-0.013078625313937664,
0.02986433170735836,
-0.05457999184727669,
-0.08906687051057816,
0.04413919895887375,
-0.07714866101741791,
-0.08968332409858704,
0.01824774406850338,
0.1480839103460312,
0.055730707943439484,
-0.019566942006349564,
0.01712970621883869,
0.11879091709852219,
0.053263671696186066,
-0.14745965600013733,
-0.0016999406507238746,
0.016004953533411026,
-0.09131720662117004,
-0.031496189534664154,
-0.024414125829935074,
0.05280110239982605,
0.059937652200460434,
0.1300731599330902,
-0.0999235138297081,
0.0733199492096901,
0.06033606082201004,
0.0034010116942226887,
-0.05825158953666687,
0.12428691983222961,
-0.06916514039039612,
-0.0730501264333725,
-0.04636475071310997,
0.10304740816354752,
0.013999581336975098,
-0.012577713467180729,
-0.07418069243431091,
0.03657260537147522,
0.11043952405452728,
0.05975958704948425,
-0.030596042051911354,
0.0494985468685627,
-0.05197145789861679,
-0.00999379437416792,
0.0434681735932827,
-0.08862054347991943,
0.04723756015300751,
0.04536442086100578,
-0.059643153101205826,
-0.02856384962797165,
-0.010483251884579659,
0.008882488124072552,
-0.006049198564141989,
0.13036185503005981,
-0.09849873185157776,
-0.01811964437365532,
-0.0754762664437294,
-0.09898308664560318,
0.022593822330236435,
-0.05795103311538696,
-0.0025721266865730286,
-0.08265341073274612,
-0.09033310413360596,
-0.033975400030612946,
0.046910449862480164,
-0.07356777042150497,
-0.06308773905038834,
-0.07908398658037186,
-0.09188234061002731,
0.04740508273243904,
0.0008964724838733673,
0.17559778690338135,
-0.06956973671913147,
0.09972748160362244,
0.015297251753509045,
0.07410239428281784,
0.09031753987073898,
0.030378419905900955,
-0.05875605717301369,
0.05575774237513542,
-0.15124833583831787,
0.06130119785666466,
-0.07987489551305771,
0.04251326993107796,
-0.11755600571632385,
-0.10724272578954697,
-0.003144079353660345,
-0.01564914919435978,
0.07825615257024765,
0.14960969984531403,
-0.11520379781723022,
-0.0841284692287445,
0.18062689900398254,
-0.08283045887947083,
-0.10655731707811356,
0.13199253380298615,
-0.019423890858888626,
-0.05828140303492546,
0.023265181109309196,
0.14191481471061707,
0.07672379910945892,
-0.11014967411756516,
-0.010495600290596485,
-0.0067551457323133945,
0.09210721403360367,
0.04021649807691574,
0.09182711690664291,
0.0012173405848443508,
-0.0024427229072898626,
0.005599942058324814,
-0.02334580384194851,
0.04900746792554855,
-0.1128782331943512,
-0.0833989605307579,
-0.026752376928925514,
-0.08934946358203888,
0.03679130971431732,
0.04734189435839653,
0.03384943678975105,
-0.10008026659488678,
-0.11841976642608643,
-0.02226584032177925,
0.10397203266620636,
-0.07381696999073029,
0.008600592613220215,
-0.04840201511979103,
0.05103060230612755,
-0.018388232216238976,
0.005293402820825577,
-0.13649380207061768,
-0.0894269198179245,
0.03519611060619354,
-0.01784038543701172,
-0.0034758427646011114,
0.008744429796934128,
0.0733824297785759,
0.09648381173610687,
-0.06883471459150314,
-0.04833348095417023,
-0.020301416516304016,
0.005432393867522478,
-0.09607035666704178,
-0.2696106731891632,
-0.03094509430229664,
-0.028388243168592453,
0.18487347662448883,
-0.2440643310546875,
0.0066866278648376465,
0.0031398034188896418,
0.11771998554468155,
0.018310144543647766,
-0.03385508805513382,
-0.013976119458675385,
0.05329594761133194,
-0.05101129412651062,
-0.06328515708446503,
0.03296639397740364,
-0.013285872526466846,
-0.10509750247001648,
-0.014827855862677097,
-0.16634561121463776,
0.10228219628334045,
0.10193686932325363,
-0.01178055815398693,
-0.11939988285303116,
-0.057552482932806015,
-0.06884437054395676,
-0.05405401811003685,
-0.030610112473368645,
0.020831912755966187,
0.1335500031709671,
0.0013732225634157658,
0.10687683522701263,
-0.06850186735391617,
-0.05531966686248779,
0.03968003764748573,
-0.01033695973455906,
-0.010296509601175785,
0.16266831755638123,
0.0963423028588295,
-0.09030944108963013,
0.12416010349988937,
0.10752961039543152,
-0.0684814304113388,
0.12479331344366074,
-0.04691418632864952,
-0.08987289667129517,
-0.03854072839021683,
0.032290827482938766,
0.023672746494412422,
0.10839231312274933,
-0.11696728318929672,
0.011548986658453941,
0.012523649260401726,
0.028666473925113678,
0.01720227301120758,
-0.1893433779478073,
-0.018200498074293137,
0.031875286251306534,
-0.0638156458735466,
0.007705983705818653,
-0.036269281059503555,
-0.0020129343029111624,
0.11255799978971481,
0.015540537424385548,
-0.057902127504348755,
-0.019257165491580963,
-0.008642349392175674,
-0.08650916069746017,
0.22700285911560059,
-0.09053503721952438,
-0.12789973616600037,
-0.09548107534646988,
0.009393737651407719,
-0.028580738231539726,
0.0042159100994467735,
0.03347095102071762,
-0.11711470037698746,
-0.03706733137369156,
-0.08248746395111084,
0.023259639739990234,
-0.030609676614403725,
0.04979529604315758,
0.006564836949110031,
0.01696793921291828,
0.03263387829065323,
-0.08377834409475327,
0.007434350438416004,
-0.026431046426296234,
-0.053761910647153854,
0.058045994490385056,
0.04748966172337532,
0.0986209362745285,
0.16690103709697723,
0.024649931117892265,
0.011010087095201015,
-0.02646651305258274,
0.17006711661815643,
-0.09340523183345795,
-0.005711109843105078,
0.05387650057673454,
0.01462982501834631,
0.038647640496492386,
0.1514652520418167,
0.04302237182855606,
-0.0846235528588295,
0.027727903798222542,
0.04807264357805252,
-0.02341052144765854,
-0.21704670786857605,
-0.04766270890831947,
-0.03825092315673828,
0.016857050359249115,
0.10270990431308746,
0.0303189754486084,
-0.010367929004132748,
0.044480908662080765,
-0.014125928282737732,
-0.003088610013946891,
0.00833717081695795,
0.056569892913103104,
0.04025040939450264,
0.03852261230349541,
0.11200740188360214,
-0.025828884914517403,
-0.0521768219769001,
0.04180397838354111,
-0.007387695834040642,
0.21862637996673584,
-0.022176468744874,
0.1402815580368042,
0.030085787177085876,
0.1396690458059311,
-0.00359214935451746,
0.06148570775985718,
0.02971069887280464,
-0.044077761471271515,
0.009248791262507439,
-0.050788696855306625,
-0.023836132138967514,
0.054146215319633484,
0.0564899779856205,
0.01883769780397415,
-0.11697529256343842,
0.05844232067465782,
0.06202102079987526,
0.24889954924583435,
0.08666163682937622,
-0.2970208525657654,
-0.09801702201366425,
0.023653453215956688,
-0.039203133434057236,
-0.012874356471002102,
0.029127659276127815,
0.16291813552379608,
-0.08728034049272537,
0.056665610522031784,
-0.05918183550238609,
0.06670404970645905,
-0.05696672573685646,
0.014242206700146198,
0.056319914758205414,
0.11252672970294952,
-0.009043720550835133,
0.06770415604114532,
-0.25010764598846436,
0.2956026792526245,
-0.006269517820328474,
0.048669926822185516,
-0.06131323054432869,
0.016983184963464737,
0.011260580271482468,
0.029177416115999222,
0.07974949479103088,
-0.004619223531335592,
-0.10744510591030121,
-0.1897789090871811,
-0.10726257413625717,
0.03397743031382561,
0.14797410368919373,
-0.07815477252006531,
0.12943977117538452,
-0.03576536476612091,
-0.027894148603081703,
0.05337340012192726,
-0.06853743642568588,
-0.07831942290067673,
-0.10502002388238907,
0.020411675795912743,
-0.03011113964021206,
0.0062726689502596855,
-0.07682807743549347,
-0.10446483641862869,
-0.11026590317487717,
0.19596898555755615,
-0.0981002226471901,
-0.01888306625187397,
-0.1295284777879715,
0.08826202899217606,
0.13799743354320526,
-0.07936575263738632,
0.02557973377406597,
-0.0004757422721013427,
0.08888070285320282,
0.01733792945742607,
-0.02682160958647728,
0.12331933528184891,
-0.09125030040740967,
-0.236156165599823,
-0.056725844740867615,
0.1407487541437149,
0.04631650075316429,
0.06976193189620972,
-0.028255973011255264,
0.029784493148326874,
-0.014054549857974052,
-0.10336292535066605,
0.06280434131622314,
0.034055281430482864,
0.058347806334495544,
0.04997030645608902,
-0.05208459496498108,
0.021547822281718254,
-0.047589682042598724,
-0.05826297029852867,
0.13860687613487244,
0.3381063938140869,
-0.09759709984064102,
0.036187879741191864,
0.07810526341199875,
-0.054714787751436234,
-0.18701986968517303,
-0.008062145672738552,
0.07945796102285385,
0.019268808886408806,
-0.0029956516809761524,
-0.1755111664533615,
0.06846778839826584,
0.09876292943954468,
-0.015769757330417633,
0.09462940692901611,
-0.3182390630245209,
-0.13771717250347137,
0.08579510450363159,
0.11075785011053085,
0.015455242246389389,
-0.18157745897769928,
-0.038252297788858414,
-0.00013113221211824566,
-0.07405944168567657,
0.08167527616024017,
-0.032394446432590485,
0.12881743907928467,
-0.008759253658354282,
0.01904122903943062,
0.024782299995422363,
-0.06896455585956573,
0.15878652036190033,
-0.01908671110868454,
0.06662772595882416,
-0.03353690356016159,
-0.003763290122151375,
-0.00009446006879443303,
-0.06961263716220856,
0.0035315898712724447,
-0.10901285707950592,
0.04481372609734535,
-0.05149662494659424,
-0.030292345210909843,
-0.07575910538434982,
0.023213941603899002,
-0.052733901888132095,
-0.059599000960588455,
-0.03700363636016846,
0.04998796433210373,
0.07207150012254715,
-0.020275283604860306,
0.11849851161241531,
-0.018482757732272148,
0.16249416768550873,
0.12361788749694824,
0.1007508859038353,
-0.011653578840196133,
-0.023667111992836,
0.0037571461871266365,
-0.011858167126774788,
0.03311389684677124,
-0.12285497784614563,
0.025357162579894066,
0.14818744361400604,
0.0339852049946785,
0.13158227503299713,
0.06583089381456375,
-0.048148345202207565,
-0.007931279949843884,
0.07773669809103012,
-0.14141419529914856,
-0.15178443491458893,
0.0010460647754371166,
-0.02582663856446743,
-0.140476793050766,
0.03235502541065216,
0.10833321511745453,
-0.05673358216881752,
0.0016230238834396005,
-0.011545279063284397,
0.06468216329813004,
-0.020738905295729637,
0.21988292038440704,
0.030890755355358124,
0.09148871153593063,
-0.09144172072410583,
0.08182516694068909,
0.04709397628903389,
-0.14964398741722107,
0.03642350435256958,
0.10723592340946198,
-0.0768512710928917,
-0.03293735161423683,
0.08544882386922836,
0.10050107538700104,
0.00278506800532341,
-0.02735172212123871,
-0.12269524484872818,
-0.14891156554222107,
0.08608924597501755,
0.10755094140768051,
0.04634270817041397,
0.047004278749227524,
-0.008645921014249325,
0.04461604356765747,
-0.1323527693748474,
0.1278623342514038,
0.061834145337343216,
0.08164410293102264,
-0.1380883753299713,
0.14385049045085907,
-0.01313534565269947,
0.009816210716962814,
-0.020665932446718216,
0.030503693968057632,
-0.14787574112415314,
-0.008112985640764236,
-0.10561439394950867,
-0.01556609757244587,
-0.05996207892894745,
0.003933335654437542,
0.015001999214291573,
-0.056885529309511185,
-0.043879397213459015,
0.0018823789432644844,
-0.09434010088443756,
-0.03714662790298462,
-0.023517237976193428,
0.06402834504842758,
-0.12430911511182785,
-0.021892856806516647,
0.03953612595796585,
-0.11159757524728775,
0.1025809496641159,
0.03664719685912132,
0.03691325709223747,
0.032698653638362885,
-0.14087563753128052,
0.03907286748290062,
0.044064633548259735,
-0.00704170111566782,
0.042517635971307755,
-0.16151610016822815,
-0.004405026324093342,
-0.04035305976867676,
0.03062516264617443,
0.019293084740638733,
0.047405850142240524,
-0.11692732572555542,
0.03561379760503769,
-0.04942760616540909,
-0.08967427909374237,
-0.056814875453710556,
0.0329449288547039,
0.05710407346487045,
-0.032565440982580185,
0.14058370888233185,
-0.0918210819363594,
0.046336740255355835,
-0.21546411514282227,
-0.011762420646846294,
-0.00668701296672225,
-0.06903732568025589,
-0.07101530581712723,
-0.033289771527051926,
0.0985458642244339,
-0.050193559378385544,
0.10634670406579971,
-0.03602820634841919,
0.04329042136669159,
0.029464324936270714,
-0.073271244764328,
0.03692247346043587,
0.07018852978944778,
0.17080850899219513,
0.050415799021720886,
-0.04027669504284859,
0.02017592266201973,
0.019523896276950836,
0.08688313513994217,
0.06830485910177231,
0.22682826220989227,
0.13057668507099152,
-0.007384416181594133,
0.085150808095932,
0.03614545613527298,
-0.15906807780265808,
-0.12790696322917938,
0.1203412190079689,
-0.10282357782125473,
0.10874339193105698,
-0.03427799046039581,
0.16880135238170624,
0.10349901765584946,
-0.1934848129749298,
0.016349680721759796,
-0.06293636560440063,
-0.09732469171285629,
-0.11127955466508865,
-0.035714566707611084,
-0.09510104358196259,
-0.169852614402771,
0.00651336507871747,
-0.12080681324005127,
0.03889879211783409,
0.06842991709709167,
0.056007690727710724,
0.025790967047214508,
0.14652755856513977,
0.07200688868761063,
0.043202344328165054,
0.04984855651855469,
0.04653747007250786,
-0.011947761289775372,
-0.026225313544273376,
-0.09713970124721527,
0.0072228084318339825,
-0.041947364807128906,
0.0462355874478817,
-0.05416034534573555,
-0.0805002897977829,
0.0847286581993103,
0.017395298928022385,
-0.0917990431189537,
0.020037174224853516,
-0.011897687800228596,
0.04687461256980896,
0.06755754351615906,
0.015428478829562664,
-0.008965367451310158,
-0.026120219379663467,
0.21451982855796814,
-0.08899791538715363,
-0.06493846327066422,
-0.11794188618659973,
0.2937953770160675,
-0.006239450071007013,
0.018630405887961388,
0.031036054715514183,
-0.06849706918001175,
-0.02497601881623268,
0.1674961894750595,
0.19818460941314697,
-0.06352739036083221,
-0.01454232819378376,
0.018547803163528442,
-0.005849586334079504,
-0.009331312030553818,
0.09362898021936417,
0.11751467734575272,
0.08447840064764023,
-0.08184275031089783,
-0.035927899181842804,
-0.023531973361968994,
-0.0382242314517498,
-0.042100150138139725,
0.05250576138496399,
0.03999210149049759,
0.005087895318865776,
-0.0373539924621582,
0.0682060495018959,
-0.07644715160131454,
-0.11217758804559708,
0.06039278954267502,
-0.20406465232372284,
-0.16178365051746368,
-0.0295164342969656,
0.08737947046756744,
-0.006437304429709911,
0.063840351998806,
0.00046400161227211356,
-0.04545655474066734,
0.08331239223480225,
-0.005981376860290766,
-0.0629948228597641,
-0.08220801502466202,
0.05515047535300255,
-0.048967473208904266,
0.1920837014913559,
-0.04341239109635353,
0.027114441618323326,
0.13457228243350983,
0.029738111421465874,
-0.10470852255821228,
0.047562375664711,
0.09300743043422699,
-0.1372644603252411,
0.047373246401548386,
0.1457965224981308,
-0.04605914652347565,
0.10505396872758865,
0.06503193825483322,
-0.1001274362206459,
0.008947654627263546,
-0.09053780138492584,
-0.04121594503521919,
-0.0350295826792717,
-0.011501185595989227,
-0.03478676453232765,
0.15443776547908783,
0.2192065715789795,
-0.048225294798612595,
0.005947508849203587,
-0.040276557207107544,
0.030163371935486794,
0.020297881215810776,
0.1365910768508911,
-0.029566984623670578,
-0.29540693759918213,
0.037136536091566086,
0.01957586221396923,
0.04135904833674431,
-0.23219265043735504,
-0.0849544107913971,
0.02575218863785267,
-0.020444950088858604,
-0.10224878042936325,
0.11793678253889084,
0.08003830909729004,
0.04651497304439545,
-0.060660868883132935,
-0.13300366699695587,
-0.04460569843649864,
0.17385706305503845,
-0.15542887151241302,
-0.08516206592321396
] |
null | null | transformers |
# zephyr-NCA-reward
This model is a fine-tuned version of [HuggingFaceH4/mistral-7b-sft-beta](https://huggingface.co/HuggingFaceH4/mistral-7b-sft-beta) on the [openbmb/UltraFeedback](https://huggingface.co/datasets/openbmb/UltraFeedback) dataset.
It achieves the following results on the evaluation set:
- Loss: 1.3007
- Loss/mini Gap Loss: 1.3007
- Loss/ori Loss: 1.3007
- Loss/reward Entrophy: 0.0
- Regularization/forward Kl: 0.5698
- Regularization/reverse Kl: 0.4143
- Regularization/policy Data Loss: 1.6956
- Regularization/reference Data Loss: 1.2661
- Regularization/policy Ref Data Loss Gap: 0.4295
- Mask/mask Ratio: 0.4577
- Reward/reward A0: -0.0038
- Reward/reward A1: -0.1788
- Reward/reward A2: -0.3592
- Reward/reward A3: -0.6457
- Rewards/chosen: -0.0038
- Rewards/rejected: -0.3945
- Rewards/margins: 0.3908
- Reward/a01 Acc: 0.6449
- Reward/a02 Acc: 0.7396
- Reward/a03 Acc: 0.8344
- Rewards/accuracies: 0.7396
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-06
- train_batch_size: 1
- eval_batch_size: 1
- seed: 42
- distributed_type: multi-GPU
- num_devices: 4
- gradient_accumulation_steps: 8
- total_train_batch_size: 32
- total_eval_batch_size: 4
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 1
### Training results
| Training Loss | Epoch | Step | Validation Loss | Loss/mini Gap Loss | Loss/ori Loss | Loss/reward Entrophy | Regularization/forward Kl | Regularization/reverse Kl | Regularization/policy Data Loss | Regularization/reference Data Loss | Regularization/policy Ref Data Loss Gap | Mask/mask Ratio | Reward/reward A0 | Reward/reward A1 | Reward/reward A2 | Reward/reward A3 | Rewards/chosen | Rewards/rejected | Rewards/margins | Reward/a01 Acc | Reward/a02 Acc | Reward/a03 Acc | Rewards/accuracies |
|:-------------:|:-----:|:----:|:---------------:|:------------------:|:-------------:|:--------------------:|:-------------------------:|:-------------------------:|:-------------------------------:|:----------------------------------:|:---------------------------------------:|:---------------:|:----------------:|:----------------:|:----------------:|:----------------:|:--------------:|:----------------:|:---------------:|:--------------:|:--------------:|:--------------:|:------------------:|
| 1.3845 | 0.05 | 100 | 1.3843 | 1.3843 | 1.3843 | 0.0 | 0.0006 | 0.0006 | 1.2682 | 1.2661 | 0.0022 | 0.4577 | 0.0030 | -0.0001 | -0.0023 | -0.0049 | 0.0030 | -0.0024 | 0.0054 | 0.5932 | 0.6579 | 0.7117 | 0.6542 |
| 1.3641 | 0.11 | 200 | 1.3632 | 1.3632 | 1.3632 | 0.0 | 0.0688 | 0.0617 | 1.3653 | 1.2661 | 0.0992 | 0.4577 | -0.0453 | -0.0905 | -0.1223 | -0.1596 | -0.0453 | -0.1241 | 0.0788 | 0.6082 | 0.6791 | 0.7396 | 0.6756 |
| 1.3464 | 0.16 | 300 | 1.3430 | 1.3430 | 1.3430 | 0.0 | 0.2320 | 0.1950 | 1.3931 | 1.2661 | 0.1270 | 0.4577 | -0.0499 | -0.1410 | -0.2129 | -0.3031 | -0.0499 | -0.2190 | 0.1691 | 0.6304 | 0.6988 | 0.7671 | 0.6988 |
| 1.3387 | 0.21 | 400 | 1.3285 | 1.3285 | 1.3285 | 0.0 | 0.4617 | 0.3766 | 1.4589 | 1.2661 | 0.1928 | 0.4577 | -0.0167 | -0.1373 | -0.2414 | -0.3912 | -0.0167 | -0.2566 | 0.2399 | 0.6356 | 0.7076 | 0.7930 | 0.7120 |
| 1.3309 | 0.27 | 500 | 1.3204 | 1.3204 | 1.3204 | 0.0 | 0.4646 | 0.3825 | 1.4782 | 1.2661 | 0.2121 | 0.4577 | -0.0003 | -0.1341 | -0.2534 | -0.4304 | -0.0003 | -0.2727 | 0.2723 | 0.6372 | 0.7107 | 0.8100 | 0.7193 |
| 1.325 | 0.32 | 600 | 1.3164 | 1.3164 | 1.3164 | 0.0 | 0.5434 | 0.4317 | 1.5453 | 1.2661 | 0.2792 | 0.4577 | -0.0366 | -0.1874 | -0.3337 | -0.5403 | -0.0366 | -0.3538 | 0.3172 | 0.6335 | 0.7205 | 0.8100 | 0.7214 |
| 1.3311 | 0.37 | 700 | 1.3122 | 1.3122 | 1.3122 | 0.0 | 0.5382 | 0.4264 | 1.5599 | 1.2661 | 0.2938 | 0.4577 | -0.0042 | -0.1527 | -0.2999 | -0.5274 | -0.0042 | -0.3267 | 0.3224 | 0.6413 | 0.7200 | 0.8245 | 0.7286 |
| 1.3112 | 0.42 | 800 | 1.3086 | 1.3086 | 1.3086 | 0.0 | 0.5743 | 0.4255 | 1.6721 | 1.2661 | 0.4060 | 0.4577 | -0.0112 | -0.1685 | -0.3250 | -0.5754 | -0.0112 | -0.3563 | 0.3451 | 0.6449 | 0.7334 | 0.8287 | 0.7357 |
| 1.3156 | 0.48 | 900 | 1.3082 | 1.3082 | 1.3082 | 0.0 | 0.5717 | 0.4240 | 1.6341 | 1.2661 | 0.3680 | 0.4577 | -0.0214 | -0.1861 | -0.3578 | -0.6112 | -0.0214 | -0.3850 | 0.3637 | 0.6460 | 0.7360 | 0.8261 | 0.7360 |
| 1.3131 | 0.53 | 1000 | 1.3066 | 1.3066 | 1.3066 | 0.0 | 0.5842 | 0.4200 | 1.7286 | 1.2661 | 0.4626 | 0.4577 | -0.0454 | -0.2257 | -0.4053 | -0.6707 | -0.0454 | -0.4339 | 0.3885 | 0.6506 | 0.7422 | 0.8328 | 0.7419 |
| 1.3092 | 0.58 | 1100 | 1.3040 | 1.3040 | 1.3040 | 0.0 | 0.5668 | 0.4164 | 1.6753 | 1.2661 | 0.4092 | 0.4577 | -0.0194 | -0.1939 | -0.3686 | -0.6412 | -0.0194 | -0.4012 | 0.3818 | 0.6460 | 0.7428 | 0.8349 | 0.7412 |
| 1.3097 | 0.64 | 1200 | 1.3027 | 1.3028 | 1.3028 | 0.0 | 0.5639 | 0.4199 | 1.6401 | 1.2661 | 0.3740 | 0.4577 | -0.0002 | -0.1708 | -0.3436 | -0.6201 | -0.0002 | -0.3782 | 0.3780 | 0.6444 | 0.7422 | 0.8395 | 0.7421 |
| 1.2929 | 0.69 | 1300 | 1.3019 | 1.3019 | 1.3019 | 0.0 | 0.5674 | 0.4188 | 1.6644 | 1.2661 | 0.3983 | 0.4577 | -0.0039 | -0.1761 | -0.3536 | -0.6335 | -0.0039 | -0.3877 | 0.3838 | 0.6470 | 0.7417 | 0.8354 | 0.7414 |
| 1.3107 | 0.74 | 1400 | 1.3017 | 1.3017 | 1.3017 | 0.0 | 0.5596 | 0.4140 | 1.6506 | 1.2661 | 0.3845 | 0.4577 | 0.0060 | -0.1611 | -0.3364 | -0.6151 | 0.0060 | -0.3708 | 0.3768 | 0.6444 | 0.7422 | 0.8333 | 0.7400 |
| 1.296 | 0.8 | 1500 | 1.3013 | 1.3013 | 1.3013 | 0.0 | 0.5751 | 0.4164 | 1.7004 | 1.2661 | 0.4343 | 0.4577 | -0.0053 | -0.1799 | -0.3600 | -0.6481 | -0.0053 | -0.3960 | 0.3907 | 0.6465 | 0.7422 | 0.8349 | 0.7412 |
| 1.304 | 0.85 | 1600 | 1.3007 | 1.3007 | 1.3007 | 0.0 | 0.5724 | 0.4169 | 1.6883 | 1.2661 | 0.4222 | 0.4577 | -0.0015 | -0.1760 | -0.3549 | -0.6421 | -0.0015 | -0.3910 | 0.3895 | 0.6434 | 0.7407 | 0.8370 | 0.7403 |
| 1.3101 | 0.9 | 1700 | 1.3006 | 1.3006 | 1.3006 | 0.0 | 0.5671 | 0.4145 | 1.6800 | 1.2661 | 0.4139 | 0.4577 | 0.0013 | -0.1716 | -0.3500 | -0.6354 | 0.0013 | -0.3857 | 0.3870 | 0.6423 | 0.7396 | 0.8359 | 0.7393 |
| 1.2987 | 0.96 | 1800 | 1.3007 | 1.3008 | 1.3008 | 0.0 | 0.5698 | 0.4143 | 1.6954 | 1.2661 | 0.4293 | 0.4577 | -0.0038 | -0.1785 | -0.3590 | -0.6456 | -0.0038 | -0.3944 | 0.3906 | 0.6449 | 0.7391 | 0.8349 | 0.7396 |
### Framework versions
- Transformers 4.35.0
- Pytorch 2.0.1+cu117
- Datasets 2.14.6
- Tokenizers 0.14.1
| {} | text-generation | ChenDRAG/zephyr-NCA-reward | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"4-bit",
"region:us"
] | 2024-02-08T09:13:06+00:00 | [] | [] | TAGS
#transformers #safetensors #mistral #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us
| zephyr-NCA-reward
=================
This model is a fine-tuned version of HuggingFaceH4/mistral-7b-sft-beta on the openbmb/UltraFeedback dataset.
It achieves the following results on the evaluation set:
* Loss: 1.3007
* Loss/mini Gap Loss: 1.3007
* Loss/ori Loss: 1.3007
* Loss/reward Entrophy: 0.0
* Regularization/forward Kl: 0.5698
* Regularization/reverse Kl: 0.4143
* Regularization/policy Data Loss: 1.6956
* Regularization/reference Data Loss: 1.2661
* Regularization/policy Ref Data Loss Gap: 0.4295
* Mask/mask Ratio: 0.4577
* Reward/reward A0: -0.0038
* Reward/reward A1: -0.1788
* Reward/reward A2: -0.3592
* Reward/reward A3: -0.6457
* Rewards/chosen: -0.0038
* Rewards/rejected: -0.3945
* Rewards/margins: 0.3908
* Reward/a01 Acc: 0.6449
* Reward/a02 Acc: 0.7396
* Reward/a03 Acc: 0.8344
* Rewards/accuracies: 0.7396
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-06
* train\_batch\_size: 1
* eval\_batch\_size: 1
* seed: 42
* distributed\_type: multi-GPU
* num\_devices: 4
* gradient\_accumulation\_steps: 8
* total\_train\_batch\_size: 32
* total\_eval\_batch\_size: 4
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: cosine
* lr\_scheduler\_warmup\_ratio: 0.1
* num\_epochs: 1
### Training results
### Framework versions
* Transformers 4.35.0
* Pytorch 2.0.1+cu117
* Datasets 2.14.6
* Tokenizers 0.14.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 1\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 4\n* gradient\\_accumulation\\_steps: 8\n* total\\_train\\_batch\\_size: 32\n* total\\_eval\\_batch\\_size: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 1",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.0\n* Pytorch 2.0.1+cu117\n* Datasets 2.14.6\n* Tokenizers 0.14.1"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 1\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 4\n* gradient\\_accumulation\\_steps: 8\n* total\\_train\\_batch\\_size: 32\n* total\\_eval\\_batch\\_size: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 1",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.0\n* Pytorch 2.0.1+cu117\n* Datasets 2.14.6\n* Tokenizers 0.14.1"
] | [
54,
179,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 1\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 4\n* gradient\\_accumulation\\_steps: 8\n* total\\_train\\_batch\\_size: 32\n* total\\_eval\\_batch\\_size: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 1### Training results### Framework versions\n\n\n* Transformers 4.35.0\n* Pytorch 2.0.1+cu117\n* Datasets 2.14.6\n* Tokenizers 0.14.1"
] | [
-0.09459036588668823,
0.08349987119436264,
-0.0033134061377495527,
0.06625774502754211,
0.11839877814054489,
0.04149315506219864,
0.11561131477355957,
0.11848306655883789,
-0.07903377711772919,
0.111440509557724,
0.13056831061840057,
0.09509420394897461,
0.05297644063830376,
0.15447966754436493,
-0.041627436876297,
-0.26931649446487427,
0.01917191594839096,
-0.044229958206415176,
-0.1345263570547104,
0.1087552011013031,
0.06834726780653,
-0.12827707827091217,
0.07649292796850204,
-0.034886643290519714,
-0.11514803022146225,
-0.03469147905707359,
-0.03332795947790146,
-0.030489850789308548,
0.12708990275859833,
0.033505819737911224,
0.09606434404850006,
0.04889211058616638,
0.10462213307619095,
-0.2572297155857086,
0.007390032056719065,
0.0708279013633728,
0.020962471142411232,
0.08017878979444504,
0.07987357676029205,
-0.007295187562704086,
0.15662536025047302,
-0.12268684804439545,
0.06851626932621002,
0.031046541407704353,
-0.11294794827699661,
-0.18616190552711487,
-0.057404644787311554,
0.03267893195152283,
0.11392494291067123,
0.07028090208768845,
-0.021923642605543137,
0.09209734946489334,
-0.10718280076980591,
0.08544208854436874,
0.22218722105026245,
-0.2632473111152649,
-0.0720437616109848,
0.06060138717293739,
0.033955346792936325,
0.08007574081420898,
-0.13588182628154755,
0.0013243267312645912,
0.04234666749835014,
0.005314059089869261,
0.09516578167676926,
0.0085448632016778,
0.03456028923392296,
0.02358432672917843,
-0.14582520723342896,
-0.0407228022813797,
0.08940701931715012,
0.0588083490729332,
-0.009696507826447487,
-0.08561886101961136,
-0.049267638474702835,
-0.19607223570346832,
-0.04132387414574623,
-0.014079200103878975,
0.02995157241821289,
-0.05661816522479057,
-0.08975566178560257,
0.04745246469974518,
-0.07503843307495117,
-0.09065177291631699,
0.01789611205458641,
0.1497116982936859,
0.055104050785303116,
-0.022189924493432045,
0.02031029388308525,
0.11754148453474045,
0.04927736893296242,
-0.1489703208208084,
-0.0009899286087602377,
0.017966480925679207,
-0.09037849307060242,
-0.030646542087197304,
-0.025243189185857773,
0.05303163081407547,
0.058629315346479416,
0.13095629215240479,
-0.09374025464057922,
0.07501448690891266,
0.06566888093948364,
0.0013596077915281057,
-0.05950207635760307,
0.12245193868875504,
-0.07187268137931824,
-0.0776204913854599,
-0.047526806592941284,
0.10476558655500412,
0.010877205990254879,
-0.01383955217897892,
-0.07122877240180969,
0.03301803395152092,
0.10991654545068741,
0.056824713945388794,
-0.032920219004154205,
0.052495360374450684,
-0.05032024160027504,
-0.008047004230320454,
0.04004167765378952,
-0.09158593416213989,
0.047449011355638504,
0.045299168676137924,
-0.058484166860580444,
-0.022555196657776833,
-0.008145524188876152,
0.0055323196575045586,
-0.004182509612292051,
0.1341051459312439,
-0.09946104884147644,
-0.020844440907239914,
-0.07704806327819824,
-0.1002984270453453,
0.019793199375271797,
-0.05800246447324753,
-0.002017393708229065,
-0.08187723904848099,
-0.08346795290708542,
-0.03949904441833496,
0.04613487794995308,
-0.07260240614414215,
-0.0605640634894371,
-0.08133469521999359,
-0.09300214052200317,
0.04876815155148506,
0.0013939691707491875,
0.1710871309041977,
-0.0687604621052742,
0.09825164824724197,
0.016474440693855286,
0.07535368204116821,
0.08405464887619019,
0.031275663524866104,
-0.05552664399147034,
0.05518164485692978,
-0.15027494728565216,
0.057226236909627914,
-0.08132058382034302,
0.04363035410642624,
-0.11572565883398056,
-0.10829383134841919,
-0.005979182198643684,
-0.014963869005441666,
0.07971882075071335,
0.1469012200832367,
-0.11880210041999817,
-0.082569420337677,
0.181160569190979,
-0.08050774782896042,
-0.10704720765352249,
0.13106414675712585,
-0.018879540264606476,
-0.05798514932394028,
0.025170916691422462,
0.1381615400314331,
0.07819686084985733,
-0.10774397850036621,
-0.017450738698244095,
-0.007722171954810619,
0.09384097903966904,
0.041461627930402756,
0.08800718933343887,
0.0033695786260068417,
-0.004098130390048027,
0.006115412339568138,
-0.02167855203151703,
0.0465259775519371,
-0.11190404742956161,
-0.08416653424501419,
-0.028102176263928413,
-0.08738511800765991,
0.03599746152758598,
0.05049702525138855,
0.029639801010489464,
-0.09887426346540451,
-0.11877799034118652,
-0.020096877589821815,
0.10246593505144119,
-0.07555253058671951,
0.006580020766705275,
-0.04740884527564049,
0.04801659658551216,
-0.013706440106034279,
0.005601315759122372,
-0.1364334225654602,
-0.09174325317144394,
0.031558629125356674,
-0.016218887642025948,
-0.002267401898279786,
0.010831484571099281,
0.07562549412250519,
0.09563034772872925,
-0.06901293247938156,
-0.04832461476325989,
-0.02333582378923893,
0.007716824766248465,
-0.09592193365097046,
-0.2708592116832733,
-0.032037973403930664,
-0.025553453713655472,
0.18429909646511078,
-0.24519796669483185,
0.006296704988926649,
0.0020868817809969187,
0.11548617482185364,
0.01447226945310831,
-0.034147489815950394,
-0.015718499198555946,
0.055538684129714966,
-0.047796543687582016,
-0.061656609177589417,
0.0338117778301239,
-0.016405154019594193,
-0.1061541959643364,
-0.021232405677437782,
-0.1691109836101532,
0.10564595460891724,
0.1010228767991066,
-0.013560394756495953,
-0.12202293425798416,
-0.055359501391649246,
-0.06736621260643005,
-0.056104231625795364,
-0.027965864166617393,
0.01867040991783142,
0.13909661769866943,
0.0010971829760819674,
0.10225242376327515,
-0.06392983347177505,
-0.05255952104926109,
0.03830788657069206,
-0.005865600425750017,
-0.005750981159508228,
0.16183766722679138,
0.0978136658668518,
-0.08964818716049194,
0.12259787321090698,
0.10614115744829178,
-0.07086280733346939,
0.12316255271434784,
-0.04672571271657944,
-0.0887284129858017,
-0.039531443268060684,
0.030179398134350777,
0.023051820695400238,
0.10807683318853378,
-0.11821333318948746,
0.010052603669464588,
0.010241247713565826,
0.025983227416872978,
0.01664845645427704,
-0.1866587996482849,
-0.018769649788737297,
0.03349661827087402,
-0.06307698041200638,
0.005887387320399284,
-0.03579065203666687,
-0.004953342489898205,
0.11204490810632706,
0.01569216512143612,
-0.058501046150922775,
-0.022103555500507355,
-0.009251303039491177,
-0.08406902849674225,
0.22895731031894684,
-0.0891348198056221,
-0.12196218222379684,
-0.096208855509758,
0.010934865102171898,
-0.027906404808163643,
0.001815095660276711,
0.03129278123378754,
-0.11752051115036011,
-0.037741247564554214,
-0.0800592377781868,
0.01842617243528366,
-0.027016067877411842,
0.051240090280771255,
0.004828073550015688,
0.016022909432649612,
0.03370488062500954,
-0.08502703160047531,
0.00835044402629137,
-0.02659814991056919,
-0.05725604295730591,
0.05600631982088089,
0.04848244786262512,
0.09969574213027954,
0.1665690839290619,
0.024542344734072685,
0.010492525063455105,
-0.022753991186618805,
0.17159855365753174,
-0.09232673794031143,
-0.005815389100462198,
0.05195770040154457,
0.013415329158306122,
0.037431687116622925,
0.15554992854595184,
0.04668334126472473,
-0.08488281071186066,
0.027878347784280777,
0.04946491867303848,
-0.021775582805275917,
-0.21620014309883118,
-0.044169407337903976,
-0.03898930177092552,
0.015411839820444584,
0.10287213325500488,
0.029875338077545166,
-0.006720555946230888,
0.04693201184272766,
-0.01271937694400549,
-0.0013586226850748062,
0.00740768713876605,
0.053588397800922394,
0.045081667602062225,
0.038702331483364105,
0.11302526295185089,
-0.024786297231912613,
-0.05326512083411217,
0.03691834211349487,
-0.013778112828731537,
0.21812807023525238,
-0.01938784308731556,
0.13994409143924713,
0.030524007976055145,
0.1405211091041565,
-0.004707259126007557,
0.0640132948756218,
0.026732970029115677,
-0.04662178084254265,
0.012073094956576824,
-0.052746206521987915,
-0.024797223508358,
0.05123714357614517,
0.056409142911434174,
0.02154596708714962,
-0.11599991470575333,
0.055368129163980484,
0.05832603573799133,
0.24313393235206604,
0.08540976792573929,
-0.298745721578598,
-0.09656431525945663,
0.018888084217905998,
-0.03719763085246086,
-0.013719647191464901,
0.02920687571167946,
0.16510362923145294,
-0.08582692593336105,
0.056277960538864136,
-0.05676496401429176,
0.06687387824058533,
-0.05882565304636955,
0.016678586602211,
0.05838172882795334,
0.11369223147630692,
-0.01174535695463419,
0.067227303981781,
-0.24877376854419708,
0.2920156419277191,
-0.006226883735507727,
0.05064240097999573,
-0.05797440931200981,
0.01472498383373022,
0.011750904843211174,
0.030427010729908943,
0.07773042470216751,
-0.005303435027599335,
-0.10277339816093445,
-0.19053170084953308,
-0.10786738246679306,
0.03618469461798668,
0.14868837594985962,
-0.07872313261032104,
0.13051053881645203,
-0.03596580773591995,
-0.029934918507933617,
0.05568871274590492,
-0.06527472287416458,
-0.08208127319812775,
-0.10569208115339279,
0.01872829906642437,
-0.024520879611372948,
0.013540932908654213,
-0.07414288073778152,
-0.10486941039562225,
-0.11013667285442352,
0.19510728120803833,
-0.09833796322345734,
-0.01613868772983551,
-0.1296708732843399,
0.0859590619802475,
0.1415148377418518,
-0.07918450981378555,
0.02848142758011818,
0.0016111310105770826,
0.08342868089675903,
0.02026323415338993,
-0.02613249607384205,
0.12422234565019608,
-0.0896647572517395,
-0.23365354537963867,
-0.05529848858714104,
0.13996239006519318,
0.04630671441555023,
0.06875304877758026,
-0.029371513053774834,
0.030658472329378128,
-0.015165227465331554,
-0.10626368224620819,
0.061921484768390656,
0.034031324088573456,
0.0556124746799469,
0.05113164335489273,
-0.05488458648324013,
0.020770644769072533,
-0.046063393354415894,
-0.05746941640973091,
0.13746707141399384,
0.3321138322353363,
-0.0959765613079071,
0.03022158145904541,
0.07714002579450607,
-0.053939152508974075,
-0.18529793620109558,
-0.009612658992409706,
0.08015777915716171,
0.020736251026391983,
-0.0014307622332125902,
-0.17713847756385803,
0.07068664580583572,
0.0994337946176529,
-0.012462268583476543,
0.09435561299324036,
-0.31686270236968994,
-0.1374107301235199,
0.08690407127141953,
0.11198707669973373,
0.01786641590297222,
-0.1808450073003769,
-0.03743049129843712,
-0.00280194915831089,
-0.07708480209112167,
0.08081277459859848,
-0.03515563905239105,
0.12946033477783203,
-0.009644810110330582,
0.01606849953532219,
0.024813607335090637,
-0.06848563998937607,
0.15665771067142487,
-0.014929880388081074,
0.06901713460683823,
-0.0342157781124115,
-0.007772983051836491,
-0.00522491754963994,
-0.06743596494197845,
0.004468532744795084,
-0.10833071172237396,
0.04303190857172012,
-0.05066809803247452,
-0.029916398227214813,
-0.07645808905363083,
0.022395946085453033,
-0.05421290174126625,
-0.058536577969789505,
-0.035537101328372955,
0.048219963908195496,
0.07079815864562988,
-0.019927354529500008,
0.11708471924066544,
-0.02040914259850979,
0.16556638479232788,
0.11933333426713943,
0.09972136467695236,
-0.008372299373149872,
-0.030391931533813477,
0.006840440444648266,
-0.010310254991054535,
0.034415874630212784,
-0.11956392973661423,
0.0243118554353714,
0.15076306462287903,
0.03456885740160942,
0.1322663128376007,
0.06709526479244232,
-0.0457598939538002,
-0.01261505763977766,
0.07900403439998627,
-0.1409885138273239,
-0.1488877832889557,
0.0012397541431710124,
-0.03062387742102146,
-0.14172789454460144,
0.03306405246257782,
0.10580016672611237,
-0.05895499512553215,
0.004521420691162348,
-0.012151127681136131,
0.061577945947647095,
-0.023174727335572243,
0.2214745730161667,
0.033339112997055054,
0.0907038152217865,
-0.09166974574327469,
0.07780075818300247,
0.04752783849835396,
-0.15022043883800507,
0.035208769142627716,
0.10663340985774994,
-0.07793675363063812,
-0.03230152279138565,
0.08663896471261978,
0.09830830246210098,
0.0037120417691767216,
-0.027581151574850082,
-0.12238410115242004,
-0.14830198884010315,
0.08409141004085541,
0.10944442451000214,
0.04461313784122467,
0.04865550994873047,
-0.008464244194328785,
0.04381981119513512,
-0.13317221403121948,
0.12612591683864594,
0.057884350419044495,
0.08257605880498886,
-0.13792391121387482,
0.14545120298862457,
-0.011405766010284424,
0.005043454933911562,
-0.020090844482183456,
0.03131258487701416,
-0.14959083497524261,
-0.009170101955533028,
-0.09907711297273636,
-0.018180273473262787,
-0.058956168591976166,
0.0029661760199815035,
0.015507161617279053,
-0.056551769375801086,
-0.044664911925792694,
0.000581751053687185,
-0.09341119974851608,
-0.03824172541499138,
-0.020137647166848183,
0.06393681466579437,
-0.12374092638492584,
-0.020475558936595917,
0.04064440727233887,
-0.11175373196601868,
0.10018236190080643,
0.03678356483578682,
0.03949365019798279,
0.03153151646256447,
-0.1428423970937729,
0.03909836709499359,
0.04314694553613663,
-0.005599013529717922,
0.041975416243076324,
-0.15968754887580872,
-0.004530262667685747,
-0.04340311884880066,
0.03246104344725609,
0.01967952959239483,
0.04502886161208153,
-0.11855417490005493,
0.035769861191511154,
-0.05149351432919502,
-0.08682386577129364,
-0.05765620991587639,
0.029962483793497086,
0.05463375896215439,
-0.03377974405884743,
0.14093558490276337,
-0.09324726462364197,
0.04572669789195061,
-0.21748335659503937,
-0.01085562165826559,
-0.006686428561806679,
-0.0696132481098175,
-0.06625276058912277,
-0.034965187311172485,
0.10028935223817825,
-0.051714640110731125,
0.10330832749605179,
-0.04026497155427933,
0.03991146385669708,
0.028890833258628845,
-0.07510998100042343,
0.03299690783023834,
0.06863056868314743,
0.17205607891082764,
0.05327438563108444,
-0.040905870497226715,
0.019932230934500694,
0.021218113601207733,
0.08253244310617447,
0.06952179223299026,
0.22442275285720825,
0.12781181931495667,
-0.009253584779798985,
0.08336618542671204,
0.03884361311793327,
-0.16050535440444946,
-0.12202801555395126,
0.12256968766450882,
-0.10529916733503342,
0.10680237412452698,
-0.03611154109239578,
0.16713300347328186,
0.10087237507104874,
-0.1910608559846878,
0.01868007518351078,
-0.0628194659948349,
-0.10033576190471649,
-0.11067420244216919,
-0.03239371255040169,
-0.0946730226278305,
-0.16967427730560303,
0.006654978729784489,
-0.12122918665409088,
0.039136771112680435,
0.07225199043750763,
0.058132268488407135,
0.026044663041830063,
0.15605120360851288,
0.07229140400886536,
0.04309896379709244,
0.049990419298410416,
0.047389835119247437,
-0.012529741041362286,
-0.030456889420747757,
-0.0941929966211319,
0.00869179330766201,
-0.04403360188007355,
0.04595746845006943,
-0.055504895746707916,
-0.08068273961544037,
0.08519285172224045,
0.02283927984535694,
-0.09216637909412384,
0.019043833017349243,
-0.012574653141200542,
0.04769481346011162,
0.06827721744775772,
0.01579306274652481,
-0.009062805213034153,
-0.027743980288505554,
0.21199119091033936,
-0.08926616609096527,
-0.06567773967981339,
-0.11514440178871155,
0.28805097937583923,
-0.0026443805545568466,
0.018574947491288185,
0.026787646114826202,
-0.06703846901655197,
-0.020028138533234596,
0.16424158215522766,
0.1940123438835144,
-0.06325586140155792,
-0.01347468513995409,
0.020200418308377266,
-0.005333127919584513,
-0.007129548583179712,
0.09277672320604324,
0.11633964627981186,
0.08552375435829163,
-0.08048051595687866,
-0.03250814229249954,
-0.023351799696683884,
-0.03901699557900429,
-0.04006117582321167,
0.0479448102414608,
0.042707670480012894,
0.0049803853034973145,
-0.035914964973926544,
0.07094228267669678,
-0.07547341287136078,
-0.11217764019966125,
0.06346817314624786,
-0.20594458281993866,
-0.16254876554012299,
-0.03399503231048584,
0.09103517234325409,
-0.0028870210517197847,
0.06352841109037399,
0.00019580799562390894,
-0.04676448926329613,
0.08590064197778702,
-0.003921758849173784,
-0.06310167908668518,
-0.08238274604082108,
0.05223899334669113,
-0.04617534950375557,
0.18797612190246582,
-0.0436374731361866,
0.03136734664440155,
0.13384106755256653,
0.029067039489746094,
-0.10169881582260132,
0.050360679626464844,
0.09203716367483139,
-0.1443629115819931,
0.04775996133685112,
0.14409179985523224,
-0.045173078775405884,
0.10576803982257843,
0.062267668545246124,
-0.10055439174175262,
0.010774781927466393,
-0.08502671122550964,
-0.04008159041404724,
-0.035616759210824966,
-0.01047245878726244,
-0.03369492292404175,
0.15848363935947418,
0.22297866642475128,
-0.049236148595809937,
0.004675254225730896,
-0.04030175879597664,
0.02965495176613331,
0.021444780752062798,
0.13783562183380127,
-0.031532350927591324,
-0.29289984703063965,
0.03891545534133911,
0.016161197796463966,
0.04230399802327156,
-0.23047323524951935,
-0.08427061885595322,
0.028693770989775658,
-0.022442778572440147,
-0.10245763510465622,
0.11508750915527344,
0.08062564581632614,
0.049220193177461624,
-0.06032932549715042,
-0.1337314248085022,
-0.04428347572684288,
0.1733447015285492,
-0.1573759764432907,
-0.08859311789274216
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# results_packing
This model is a fine-tuned version of [TheBloke/Mistral-7B-Instruct-v0.1-GPTQ](https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-GPTQ) on the generator dataset.
It achieves the following results on the evaluation set:
- Loss: 0.5551
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: constant
- lr_scheduler_warmup_ratio: 0.03
- training_steps: 200
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| 0.8415 | 0.08 | 50 | 0.6991 |
| 0.6472 | 0.17 | 100 | 0.6063 |
| 0.5802 | 0.25 | 150 | 0.5596 |
| 0.5499 | 0.33 | 200 | 0.5551 |
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.2.0
- Datasets 2.16.1
- Tokenizers 0.15.1 | {"license": "apache-2.0", "library_name": "peft", "tags": ["trl", "sft", "generated_from_trainer"], "datasets": ["generator"], "base_model": "TheBloke/Mistral-7B-Instruct-v0.1-GPTQ", "model-index": [{"name": "results_packing", "results": []}]} | null | ananyarn/results_packing | [
"peft",
"safetensors",
"trl",
"sft",
"generated_from_trainer",
"dataset:generator",
"base_model:TheBloke/Mistral-7B-Instruct-v0.1-GPTQ",
"license:apache-2.0",
"region:us"
] | 2024-02-08T09:17:22+00:00 | [] | [] | TAGS
#peft #safetensors #trl #sft #generated_from_trainer #dataset-generator #base_model-TheBloke/Mistral-7B-Instruct-v0.1-GPTQ #license-apache-2.0 #region-us
| results\_packing
================
This model is a fine-tuned version of TheBloke/Mistral-7B-Instruct-v0.1-GPTQ on the generator dataset.
It achieves the following results on the evaluation set:
* Loss: 0.5551
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 4
* eval\_batch\_size: 4
* seed: 42
* gradient\_accumulation\_steps: 4
* total\_train\_batch\_size: 16
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: constant
* lr\_scheduler\_warmup\_ratio: 0.03
* training\_steps: 200
### Training results
### Framework versions
* PEFT 0.8.2
* Transformers 4.37.2
* Pytorch 2.2.0
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 4\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: constant\n* lr\\_scheduler\\_warmup\\_ratio: 0.03\n* training\\_steps: 200",
"### Training results",
"### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.2\n* Pytorch 2.2.0\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#peft #safetensors #trl #sft #generated_from_trainer #dataset-generator #base_model-TheBloke/Mistral-7B-Instruct-v0.1-GPTQ #license-apache-2.0 #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 4\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: constant\n* lr\\_scheduler\\_warmup\\_ratio: 0.03\n* training\\_steps: 200",
"### Training results",
"### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.2\n* Pytorch 2.2.0\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
64,
144,
4,
36
] | [
"passage: TAGS\n#peft #safetensors #trl #sft #generated_from_trainer #dataset-generator #base_model-TheBloke/Mistral-7B-Instruct-v0.1-GPTQ #license-apache-2.0 #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 4\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: constant\n* lr\\_scheduler\\_warmup\\_ratio: 0.03\n* training\\_steps: 200### Training results### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.2\n* Pytorch 2.2.0\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.13377125561237335,
0.08383166790008545,
-0.0018904240569099784,
0.10671833902597427,
0.14234797656536102,
0.01838228851556778,
0.1030697375535965,
0.12457898259162903,
-0.09588422626256943,
0.08616204559803009,
0.11931104958057404,
0.07477358728647232,
0.04183875024318695,
0.18180584907531738,
-0.03875081613659859,
-0.274992972612381,
0.006144147366285324,
0.0018115618731826544,
-0.11069521307945251,
0.13819946348667145,
0.09488068521022797,
-0.1136125773191452,
0.07361114770174026,
-0.011961371637880802,
-0.15589815378189087,
-0.004536728374660015,
0.012978260405361652,
-0.04543145373463631,
0.13011236488819122,
0.025189753621816635,
0.10367168486118317,
0.011508884839713573,
0.11055102199316025,
-0.18338975310325623,
0.008014867082238197,
0.08882709592580795,
0.021615421399474144,
0.10195335745811462,
0.08986746519804001,
0.017846805974841118,
0.12248973548412323,
-0.06064843386411667,
0.05602741241455078,
0.043073870241642,
-0.14009377360343933,
-0.30416494607925415,
-0.1109301820397377,
0.10068481415510178,
0.12254202365875244,
0.07829367369413376,
-0.005021500866860151,
0.13191960752010345,
-0.040971189737319946,
0.08153638988733292,
0.2750687599182129,
-0.29028981924057007,
-0.09792816638946533,
0.029324833303689957,
0.04174743592739105,
0.07086960971355438,
-0.12594860792160034,
-0.025279099121689796,
0.06633492559194565,
0.04604831710457802,
0.12510405480861664,
-0.005362286698073149,
0.004486030898988247,
0.006952430587261915,
-0.14777952432632446,
-0.03742630407214165,
0.13705897331237793,
0.05312646925449371,
-0.045090287923812866,
-0.03970468416810036,
-0.07260547578334808,
-0.19959314167499542,
-0.02875564992427826,
-0.004606133326888084,
0.023869846016168594,
-0.04679986461997032,
-0.056025803089141846,
0.002286340342834592,
-0.08246422559022903,
-0.10576459020376205,
0.009352206252515316,
0.17077603936195374,
0.04516137018799782,
0.016899827867746353,
-0.005359055940061808,
0.139296293258667,
0.011922798119485378,
-0.15559040009975433,
0.00678927730768919,
0.011488744057714939,
-0.04588254913687706,
-0.027252033352851868,
-0.021351149305701256,
0.028024524450302124,
-0.0020134339574724436,
0.18188902735710144,
-0.14078933000564575,
0.05353438854217529,
0.03714262694120407,
0.029882172122597694,
-0.11476610600948334,
0.14396357536315918,
-0.08641528338193893,
-0.031677886843681335,
-0.0159074105322361,
0.12452486157417297,
0.02994551509618759,
-0.005745079834014177,
-0.07163485139608383,
0.023305462673306465,
0.09435736387968063,
0.03718050941824913,
-0.023632530122995377,
0.018023556098341942,
-0.053799811750650406,
-0.019775232300162315,
0.08574818819761276,
-0.08943519741296768,
0.039275817573070526,
0.018320763483643532,
-0.07691895216703415,
-0.04466017708182335,
-0.0008009399753063917,
-0.0008803015807643533,
0.011055944487452507,
0.13202181458473206,
-0.09563037008047104,
0.004513903986662626,
-0.0960465744137764,
-0.100472092628479,
0.009467034600675106,
-0.03559005260467529,
-0.007143228314816952,
-0.08710110187530518,
-0.1641549915075302,
-0.030542319640517235,
0.04536396637558937,
-0.06547167152166367,
-0.05501753091812134,
-0.00811881572008133,
-0.11452669650316238,
0.02892909198999405,
-0.015487127006053925,
0.14187535643577576,
-0.06084952503442764,
0.12349686026573181,
0.03727973997592926,
0.04777878150343895,
0.00276117748580873,
0.02332514524459839,
-0.07851707935333252,
0.05260852724313736,
-0.18646302819252014,
0.010109716095030308,
-0.06371790170669556,
0.052616845816373825,
-0.12287671864032745,
-0.1201241984963417,
-0.01298441644757986,
-0.00861834641546011,
0.10422123223543167,
0.13010963797569275,
-0.16185760498046875,
-0.06831175088882446,
0.1604682058095932,
-0.10286224633455276,
-0.13106311857700348,
0.12034232914447784,
-0.008270499296486378,
0.0024883223231881857,
0.029622767120599747,
0.1370117962360382,
0.08870550245046616,
-0.11855949461460114,
-0.004278648179024458,
-0.030410392209887505,
0.0964960977435112,
-0.0009069075458683074,
0.10239942371845245,
-0.020782319828867912,
-0.007679462432861328,
0.005502782296389341,
-0.07939378917217255,
0.043942391872406006,
-0.10954231768846512,
-0.07342495769262314,
-0.03909656032919884,
-0.08367607742547989,
0.06751565635204315,
0.06739813834428787,
0.032857850193977356,
-0.10173046588897705,
-0.11546573787927628,
0.057142361998558044,
0.1203455701470375,
-0.051837258040905,
0.01785784587264061,
-0.04919830709695816,
0.10515008121728897,
-0.04736226424574852,
-0.019479278475046158,
-0.1720719039440155,
-0.04816390946507454,
0.01825050823390484,
-0.01061963476240635,
-0.03719090670347214,
-0.028561003506183624,
0.09102105349302292,
0.09559805691242218,
-0.0951160192489624,
-0.07540931552648544,
-0.08526129275560379,
-0.007338641211390495,
-0.11382509768009186,
-0.22294631600379944,
-0.08369659632444382,
-0.024733498692512512,
0.12987326085567474,
-0.23971836268901825,
0.033747922629117966,
0.011199629865586758,
0.11967382580041885,
0.045754801481962204,
-0.03716861829161644,
-0.032779593020677567,
0.0660521537065506,
-0.03557179123163223,
-0.08050254732370377,
0.032916754484176636,
-0.002567555522546172,
-0.06201588362455368,
-0.031961023807525635,
-0.129450261592865,
0.18276967108249664,
0.10501037538051605,
0.02550300396978855,
-0.11009524017572403,
-0.018428226932883263,
-0.08966686576604843,
-0.04240654781460762,
-0.05982058495283127,
-0.005292042624205351,
0.08227414637804031,
0.02613469772040844,
0.13452233374118805,
-0.08659123629331589,
-0.05490085482597351,
0.04141761362552643,
-0.015472439117729664,
0.014484124258160591,
0.12557613849639893,
0.06435872614383698,
-0.05346056818962097,
0.14668971300125122,
0.1539342999458313,
-0.0631137564778328,
0.11587904393672943,
-0.05900925397872925,
-0.09569564461708069,
-0.034192249178886414,
0.027935761958360672,
0.027503758668899536,
0.16001537442207336,
-0.010998638346791267,
0.024011747911572456,
0.015359650366008282,
0.026997968554496765,
-0.00027981854509562254,
-0.21633966267108917,
-0.033090442419052124,
0.019985243678092957,
-0.05075082927942276,
-0.020023087039589882,
-0.018528437241911888,
-0.003460153006017208,
0.10844465345144272,
0.009702985174953938,
-0.06673824042081833,
0.0034405679907649755,
0.018648194149136543,
-0.07069108635187149,
0.20846787095069885,
-0.08570559322834015,
-0.10561865568161011,
-0.13430127501487732,
0.030578985810279846,
-0.043382056057453156,
-0.009479725733399391,
0.04403933882713318,
-0.07792554795742035,
-0.02424011193215847,
-0.08643262088298798,
0.01067190058529377,
0.00565321184694767,
0.02444612607359886,
0.009812816977500916,
0.0009606247185729444,
0.10112056136131287,
-0.09432443231344223,
0.018180619925260544,
-0.01995892822742462,
-0.052885752171278,
0.023961514234542847,
0.04983692616224289,
0.09929338842630386,
0.144028440117836,
0.014918912202119827,
0.01334261056035757,
-0.029870329424738884,
0.20910438895225525,
-0.09686649590730667,
0.002883204957470298,
0.14082860946655273,
0.015975823625922203,
0.05715235695242882,
0.13375471532344818,
0.06539272516965866,
-0.09279413521289825,
0.01610318198800087,
0.04187219962477684,
-0.027814827859401703,
-0.21167545020580292,
-0.034222349524497986,
-0.036417338997125626,
0.01000512856990099,
0.10657326877117157,
0.04569251462817192,
0.0013318260898813605,
0.033558525145053864,
-0.030708739534020424,
0.011761801317334175,
-0.010164521634578705,
0.07625162601470947,
0.017063694074749947,
0.04734019562602043,
0.11753546446561813,
-0.034325871616601944,
-0.015187373384833336,
0.04357476532459259,
-0.011225772090256214,
0.21192598342895508,
-0.03674935922026634,
0.08135562390089035,
0.04052288085222244,
0.17447932064533234,
-0.01778518594801426,
0.0852239653468132,
0.006235490553081036,
-0.042901411652565,
0.0063021318055689335,
-0.06527739763259888,
-0.0028645601123571396,
0.03578849509358406,
-0.053102586418390274,
0.1019369512796402,
-0.12168946117162704,
-0.004660302773118019,
0.06021621823310852,
0.31808796525001526,
0.07754731923341751,
-0.327361136674881,
-0.12306278198957443,
0.00014236968127079308,
-0.015486782416701317,
-0.04707389324903488,
0.025856900960206985,
0.15974198281764984,
-0.08022255450487137,
0.05303235352039337,
-0.08563124388456345,
0.08056322485208511,
0.008142123930156231,
0.023346422240138054,
0.12628304958343506,
0.11101941764354706,
-0.01630471460521221,
0.04621448740363121,
-0.24061152338981628,
0.3032817244529724,
0.018349168822169304,
0.08264368027448654,
-0.025935685262084007,
0.010063808411359787,
0.023319773375988007,
0.061989590525627136,
0.06635565310716629,
-0.0032403231598436832,
-0.04802580922842026,
-0.22350697219371796,
-0.08698854595422745,
0.00837918184697628,
0.1095721572637558,
-0.07589497417211533,
0.10853540152311325,
-0.03670421987771988,
-0.018197620287537575,
0.04265962541103363,
-0.06295354664325714,
-0.07651910930871964,
-0.05060494318604469,
0.009605481289327145,
-0.008705389685928822,
0.010922182351350784,
-0.11246418207883835,
-0.11521503329277039,
-0.05600398778915405,
0.11062741279602051,
-0.08865989744663239,
-0.0337057039141655,
-0.1399073451757431,
0.08921999484300613,
0.13991396129131317,
-0.0749034732580185,
0.03648895025253296,
0.018336601555347443,
0.09125374257564545,
0.030603952705860138,
-0.030312899500131607,
0.11828825622797012,
-0.08438034355640411,
-0.2462238073348999,
-0.04572932794690132,
0.1497705578804016,
0.030573638156056404,
0.052341219037771225,
-0.03373581916093826,
0.043998733162879944,
-0.0003009841311722994,
-0.09501088410615921,
0.032110054045915604,
-0.000603223335929215,
0.06362370401620865,
0.01907357946038246,
-0.028629977256059647,
0.044026922434568405,
-0.05044843256473541,
-0.04409933462738991,
0.07521472126245499,
0.3128722608089447,
-0.10562066733837128,
0.008583311922848225,
0.03972017019987106,
-0.037335433065891266,
-0.18131226301193237,
0.03985494375228882,
0.09829901903867722,
0.007914897985756397,
0.06329960376024246,
-0.15425170958042145,
0.0558619499206543,
0.12480537593364716,
-0.04422813653945923,
0.13283036649227142,
-0.3349918723106384,
-0.12679919600486755,
0.0625929981470108,
0.12285853177309036,
0.013062844052910805,
-0.18548813462257385,
-0.046904414892196655,
-0.0019244897412136197,
-0.11053000390529633,
0.051183268427848816,
-0.1144833192229271,
0.08973784744739532,
-0.029973266646265984,
0.012727177701890469,
0.015731532126665115,
-0.05943634733557701,
0.1628127247095108,
-0.03268522024154663,
0.10065677762031555,
-0.02158546634018421,
0.020817525684833527,
0.04801081120967865,
-0.06182203069329262,
0.014971289783716202,
-0.06863372027873993,
0.05303366482257843,
-0.09189525246620178,
-0.00008863770926836878,
-0.10777395218610764,
0.03181052580475807,
-0.05759485810995102,
-0.05644642561674118,
-0.0194336399435997,
0.050546422600746155,
0.014946128241717815,
-0.029869865626096725,
0.13489259779453278,
0.014910644851624966,
0.21085113286972046,
0.14258305728435516,
0.031147673726081848,
-0.030488234013319016,
-0.09481586515903473,
0.0016883796779438853,
-0.01983012817800045,
0.07219069451093674,
-0.16266433894634247,
0.018402833491563797,
0.13149665296077728,
0.04926354065537453,
0.11457724869251251,
0.06863415241241455,
-0.06252975761890411,
0.0044115278869867325,
0.07791148871183395,
-0.14546428620815277,
-0.12663355469703674,
-0.021590502932667732,
0.027920952066779137,
-0.14817653596401215,
0.05998021736741066,
0.11544342339038849,
-0.08159404993057251,
-0.011794651858508587,
0.007618040312081575,
0.03892451524734497,
-0.047135088592767715,
0.24122050404548645,
0.05403945595026016,
0.08515980839729309,
-0.09028840065002441,
0.07086358964443207,
0.03271409869194031,
-0.09361336380243301,
0.006619039922952652,
0.08883289247751236,
-0.07144887000322342,
-0.01843016780912876,
0.07541225850582123,
0.05759235471487045,
-0.020537687465548515,
-0.04373886063694954,
-0.13537070155143738,
-0.12063174694776535,
0.06007475033402443,
0.12105508148670197,
0.06437210738658905,
0.02911977283656597,
-0.002773561980575323,
0.023373287171125412,
-0.12154713273048401,
0.10789734125137329,
0.06449825316667557,
0.09210670739412308,
-0.17791108787059784,
0.14329324662685394,
-0.000026646137484931387,
0.020014557987451553,
-0.012944385409355164,
0.039937302470207214,
-0.12381230294704437,
0.006431960966438055,
-0.11422210931777954,
-0.04494522511959076,
-0.03255479037761688,
-0.007361666765064001,
-0.009980961680412292,
-0.0643891915678978,
-0.05124001204967499,
0.027884718030691147,
-0.11223367601633072,
-0.040941447019577026,
0.010399481281638145,
0.03425126522779465,
-0.13204890489578247,
-0.011247953400015831,
0.038406845182180405,
-0.09955313056707382,
0.08353740721940994,
0.036099184304475784,
0.04595340043306351,
0.04888913780450821,
-0.10143289715051651,
0.022204391658306122,
0.042514074593782425,
-0.02290533483028412,
0.05512794479727745,
-0.1186710000038147,
-0.0060271029360592365,
-0.04335616156458855,
0.03211241587996483,
0.01602429710328579,
0.05882808566093445,
-0.14441385865211487,
0.006702289916574955,
-0.027855144813656807,
-0.05302072688937187,
-0.04541953280568123,
0.028164571151137352,
0.06937438994646072,
0.01567428931593895,
0.1660633385181427,
-0.09462027996778488,
0.02663571573793888,
-0.236319899559021,
-0.01365854125469923,
-0.030506934970617294,
-0.08558493107557297,
-0.11287526041269302,
-0.005485339090228081,
0.08566216379404068,
-0.04958249628543854,
0.049472976475954056,
-0.020433619618415833,
0.089828260242939,
0.046331822872161865,
-0.07871386408805847,
0.009194926358759403,
0.039458997547626495,
0.17433811724185944,
0.018818039447069168,
-0.020250167697668076,
0.049594029784202576,
0.03803717717528343,
0.06677264720201492,
0.10094711184501648,
0.1907331794500351,
0.13873174786567688,
-0.012646976858377457,
0.07575859129428864,
0.03611907735466957,
-0.1096692755818367,
-0.16041943430900574,
0.04750531539320946,
-0.05758780613541603,
0.10900747030973434,
-0.03616278991103172,
0.1621081829071045,
0.13205401599407196,
-0.188242569565773,
0.018245700746774673,
-0.0599275603890419,
-0.07268201559782028,
-0.11428367346525192,
-0.002335387747734785,
-0.06524693965911865,
-0.18229229748249054,
-0.0017326967790722847,
-0.10796421021223068,
0.02781709097325802,
0.10933257639408112,
0.015131479129195213,
0.023052314296364784,
0.16704189777374268,
0.06776265054941177,
0.03681706637144089,
0.0567595474421978,
0.03504118695855141,
-0.00998681504279375,
-0.03374277800321579,
-0.08812326192855835,
0.02226056344807148,
-0.06397177278995514,
0.037612684071063995,
-0.05396418273448944,
-0.09206397831439972,
0.056441936641931534,
0.018389873206615448,
-0.10218389332294464,
0.03324124217033386,
0.0122750299051404,
0.04589997977018356,
0.06592340022325516,
0.02181963436305523,
0.011294381693005562,
-0.03433103486895561,
0.27436622977256775,
-0.1080475002527237,
-0.04817504435777664,
-0.12739387154579163,
0.3255981504917145,
0.021687079221010208,
-0.018299464136362076,
0.024266745895147324,
-0.09850902855396271,
-0.016425559297204018,
0.14255598187446594,
0.1586240530014038,
-0.04725192114710808,
-0.013923163525760174,
0.009271579794585705,
-0.012951810844242573,
-0.058254580944776535,
0.11387727409601212,
0.12047569453716278,
0.06182979792356491,
-0.0825524553656578,
-0.0197201706469059,
-0.04229709506034851,
-0.030509447678923607,
-0.0409100167453289,
0.046718914061784744,
0.007346311118453741,
0.005985603667795658,
-0.048063065856695175,
0.08790122717618942,
-0.0031697184313088655,
-0.127186119556427,
0.0791500136256218,
-0.19278061389923096,
-0.1792127788066864,
-0.01941543072462082,
0.08544009178876877,
0.003250126028433442,
0.07081300765275955,
-0.008964784443378448,
-0.016007035970687866,
0.0943855568766594,
-0.025681357830762863,
-0.038325630128383636,
-0.15324939787387848,
0.08830457180738449,
-0.09337739646434784,
0.22810229659080505,
-0.04951630160212517,
0.02541055530309677,
0.12204191088676453,
0.017218494787812233,
-0.13435621559619904,
0.03879375010728836,
0.07153118401765823,
-0.11876975744962692,
0.004554151091724634,
0.13347238302230835,
-0.03767578676342964,
0.08942866325378418,
0.027541490271687508,
-0.1376032680273056,
-0.017084255814552307,
-0.04962838068604469,
-0.036219093948602676,
-0.0571594163775444,
-0.017130836844444275,
-0.031116126105189323,
0.14313669502735138,
0.21217037737369537,
-0.05898934602737427,
0.009620203636586666,
-0.061665769666433334,
0.03580343350768089,
0.06158244609832764,
0.09237551689147949,
-0.007898185402154922,
-0.26558396220207214,
0.033908210694789886,
0.025394951924681664,
-0.015245246700942516,
-0.24170343577861786,
-0.07262936979532242,
0.03254398703575134,
-0.06004967913031578,
-0.08594813197851181,
0.10239405930042267,
0.05559217929840088,
0.05725894868373871,
-0.051276568323373795,
-0.08533967286348343,
-0.07561778277158737,
0.18237453699111938,
-0.15555039048194885,
-0.0763864666223526
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# 1000_STEPS_5e7
This model is a fine-tuned version of [meta-llama/Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b-hf) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.6782
- Rewards/chosen: -0.0504
- Rewards/rejected: -0.0840
- Rewards/accuracies: 0.5297
- Rewards/margins: 0.0336
- Logps/rejected: -15.9795
- Logps/chosen: -14.6212
- Logits/rejected: -0.0566
- Logits/chosen: -0.0565
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-07
- train_batch_size: 4
- eval_batch_size: 1
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 8
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_steps: 100
- training_steps: 500
### Training results
| Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
|:-------------:|:-----:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
| 0.6932 | 0.1 | 50 | 0.6927 | 0.0015 | 0.0005 | 0.4242 | 0.0010 | -15.1347 | -14.1020 | -0.0215 | -0.0215 |
| 0.6901 | 0.2 | 100 | 0.6901 | -0.0121 | -0.0185 | 0.4835 | 0.0063 | -15.3239 | -14.2383 | -0.0268 | -0.0268 |
| 0.6826 | 0.29 | 150 | 0.6841 | -0.0153 | -0.0346 | 0.5143 | 0.0193 | -15.4851 | -14.2699 | -0.0332 | -0.0331 |
| 0.6776 | 0.39 | 200 | 0.6813 | -0.0237 | -0.0492 | 0.5209 | 0.0255 | -15.6318 | -14.3538 | -0.0437 | -0.0436 |
| 0.678 | 0.49 | 250 | 0.6794 | -0.0441 | -0.0745 | 0.5209 | 0.0304 | -15.8841 | -14.5577 | -0.0475 | -0.0473 |
| 0.6715 | 0.59 | 300 | 0.6786 | -0.0529 | -0.0857 | 0.5253 | 0.0328 | -15.9964 | -14.6462 | -0.0538 | -0.0536 |
| 0.6846 | 0.68 | 350 | 0.6782 | -0.0478 | -0.0812 | 0.5165 | 0.0334 | -15.9511 | -14.5949 | -0.0559 | -0.0557 |
| 0.6846 | 0.78 | 400 | 0.6781 | -0.0501 | -0.0839 | 0.5231 | 0.0338 | -15.9784 | -14.6180 | -0.0567 | -0.0565 |
| 0.6755 | 0.88 | 450 | 0.6783 | -0.0498 | -0.0833 | 0.5209 | 0.0335 | -15.9725 | -14.6153 | -0.0564 | -0.0562 |
| 0.665 | 0.98 | 500 | 0.6782 | -0.0504 | -0.0840 | 0.5297 | 0.0336 | -15.9795 | -14.6212 | -0.0566 | -0.0565 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.0.0+cu117
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"tags": ["trl", "dpo", "generated_from_trainer"], "base_model": "meta-llama/Llama-2-7b-hf", "model-index": [{"name": "1000_STEPS_5e7", "results": []}]} | text-generation | tsavage68/500STEPS_5e7_0.1beta_DPO_zeroshot | [
"transformers",
"safetensors",
"llama",
"text-generation",
"trl",
"dpo",
"generated_from_trainer",
"base_model:meta-llama/Llama-2-7b-hf",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T09:19:30+00:00 | [] | [] | TAGS
#transformers #safetensors #llama #text-generation #trl #dpo #generated_from_trainer #base_model-meta-llama/Llama-2-7b-hf #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| 1000\_STEPS\_5e7
================
This model is a fine-tuned version of meta-llama/Llama-2-7b-hf on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 0.6782
* Rewards/chosen: -0.0504
* Rewards/rejected: -0.0840
* Rewards/accuracies: 0.5297
* Rewards/margins: 0.0336
* Logps/rejected: -15.9795
* Logps/chosen: -14.6212
* Logits/rejected: -0.0566
* Logits/chosen: -0.0565
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-07
* train\_batch\_size: 4
* eval\_batch\_size: 1
* seed: 42
* gradient\_accumulation\_steps: 2
* total\_train\_batch\_size: 8
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: cosine
* lr\_scheduler\_warmup\_steps: 100
* training\_steps: 500
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.0.0+cu117
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-07\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 1\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 8\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 100\n* training\\_steps: 500",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.0.0+cu117\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #trl #dpo #generated_from_trainer #base_model-meta-llama/Llama-2-7b-hf #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-07\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 1\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 8\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 100\n* training\\_steps: 500",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.0.0+cu117\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
78,
145,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #trl #dpo #generated_from_trainer #base_model-meta-llama/Llama-2-7b-hf #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-07\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 1\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 8\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 100\n* training\\_steps: 500### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.0.0+cu117\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.14193975925445557,
0.09611162543296814,
-0.0019720732234418392,
0.07512185722589493,
0.14658428728580475,
0.011171606369316578,
0.10866626352071762,
0.13378599286079407,
-0.09537959843873978,
0.09006843715906143,
0.13701117038726807,
0.12408792227506638,
0.060615647584199905,
0.19523613154888153,
-0.033620428293943405,
-0.30485400557518005,
0.002159582683816552,
-0.016122855246067047,
-0.16598032414913177,
0.1283375471830368,
0.0887812077999115,
-0.12724459171295166,
0.05355114862322807,
-0.039557743817567825,
-0.11999154090881348,
-0.02807861566543579,
-0.018446724861860275,
-0.04014139994978905,
0.12813924252986908,
0.00039565080078318715,
0.10818539559841156,
0.05146857351064682,
0.10584796220064163,
-0.23012223839759827,
0.010249089449644089,
0.06493046134710312,
0.042111992835998535,
0.08586479723453522,
0.06518134474754333,
-0.02590217813849449,
0.07479190081357956,
-0.11034654825925827,
0.06587765365839005,
0.03734516352415085,
-0.11907412856817245,
-0.22864075005054474,
-0.1019957885146141,
0.05379056558012962,
0.1604621410369873,
0.0830584317445755,
-0.023713111877441406,
0.06412151455879211,
-0.07584517449140549,
0.08028963953256607,
0.2444695681333542,
-0.2763167917728424,
-0.08384404331445694,
0.05890754610300064,
0.053573571145534515,
0.0682934820652008,
-0.12756602466106415,
-0.006092457566410303,
0.04486754536628723,
0.00680688489228487,
0.13115747272968292,
0.006604501511901617,
0.08164937049150467,
0.0028073631692677736,
-0.14601244032382965,
-0.03944765031337738,
0.11817032843828201,
0.08219651877880096,
-0.03934166580438614,
-0.0889798104763031,
-0.037351835519075394,
-0.24283204972743988,
-0.04102824628353119,
-0.021512681618332863,
0.03608549386262894,
-0.05324641615152359,
-0.0948348194360733,
0.014148535206913948,
-0.07382192462682724,
-0.11331438273191452,
0.04992031306028366,
0.1402841955423355,
0.036176178604364395,
-0.040432173758745193,
0.02280067838728428,
0.16206230223178864,
0.0639394000172615,
-0.15642784535884857,
-0.0068263583816587925,
0.0237823948264122,
-0.07969123870134354,
-0.04336576536297798,
-0.018743056803941727,
0.0006900654989294708,
0.007560049183666706,
0.15455730259418488,
-0.03654094785451889,
0.04204306751489639,
0.05633080378174782,
0.02818184904754162,
-0.11200892925262451,
0.15183185040950775,
-0.07894650101661682,
-0.09609129279851913,
-0.029143380001187325,
0.14775849878787994,
-0.001533397939056158,
-0.007696449290961027,
-0.08075860142707825,
0.012638350948691368,
0.1130424216389656,
0.07392694801092148,
-0.025230959057807922,
0.03376786410808563,
-0.07719767838716507,
-0.014381323009729385,
0.019343651831150055,
-0.09795940667390823,
0.020304687321186066,
0.003749061608687043,
-0.08553306758403778,
-0.050435371696949005,
-0.004294895566999912,
0.015064530074596405,
0.014480088837444782,
0.13772353529930115,
-0.07971707731485367,
-0.025784287601709366,
-0.1060992106795311,
-0.10450806468725204,
0.004735653754323721,
-0.07800330966711044,
-0.011189867742359638,
-0.08024202287197113,
-0.15549702942371368,
-0.06301245838403702,
0.044486310333013535,
-0.056872833520174026,
-0.0675959661602974,
-0.09349611401557922,
-0.10167363286018372,
0.028227880597114563,
-0.00816099252551794,
0.15564240515232086,
-0.049277398735284805,
0.13446371257305145,
0.019032273441553116,
0.08227790147066116,
0.0665680468082428,
0.043479207903146744,
-0.05351962894201279,
0.0668342262506485,
-0.19338266551494598,
0.06483202427625656,
-0.0654647946357727,
0.09670735150575638,
-0.12554645538330078,
-0.09485278278589249,
-0.03730519860982895,
-0.013316689990460873,
0.09825997054576874,
0.16611483693122864,
-0.18071195483207703,
-0.07693098485469818,
0.197128564119339,
-0.06081858277320862,
-0.13067293167114258,
0.1069340780377388,
-0.02320052497088909,
0.041156575083732605,
0.03990441933274269,
0.13452957570552826,
0.10291416943073273,
-0.07664871215820312,
0.005523914471268654,
-0.04144828021526337,
0.09556667506694794,
0.014000343158841133,
0.09604763984680176,
-0.029256731271743774,
0.017875831574201584,
-0.0015058850403875113,
-0.06430215388536453,
0.04558967798948288,
-0.10766590386629105,
-0.08726353943347931,
-0.004442185629159212,
-0.10453953593969345,
0.0618693009018898,
0.04217809438705444,
0.051201850175857544,
-0.09105858951807022,
-0.10592470318078995,
-0.006108481902629137,
0.10387604683637619,
-0.07010020315647125,
0.010164492763578892,
-0.040785856544971466,
0.07023949176073074,
-0.03337434306740761,
0.004244452342391014,
-0.1382768750190735,
-0.05443188548088074,
0.02785806730389595,
0.027491174638271332,
-0.025652695447206497,
-0.028203429654240608,
0.08405598998069763,
0.06802120804786682,
-0.07686233520507812,
-0.08542002737522125,
-0.06388463079929352,
-0.009591693058609962,
-0.11501345783472061,
-0.23343226313591003,
-0.06365849077701569,
-0.031634341925382614,
0.20570732653141022,
-0.2528097331523895,
0.050353240221738815,
0.005045408848673105,
0.11786577105522156,
0.04031569883227348,
-0.04049469530582428,
-0.004058192018419504,
0.05691719800233841,
-0.029734894633293152,
-0.08901741355657578,
0.03939450904726982,
-0.01701507344841957,
-0.1324252486228943,
-0.027531588450074196,
-0.12858843803405762,
0.13843335211277008,
0.0972895622253418,
0.00736086443066597,
-0.1345687210559845,
-0.08222097158432007,
-0.06910376995801926,
-0.04199934005737305,
-0.036510169506073,
-0.007870522327721119,
0.09103063493967056,
0.038941312581300735,
0.12442954629659653,
-0.07187652587890625,
-0.057208020240068436,
0.030992541462183,
-0.0004666434251703322,
0.019913433119654655,
0.15053798258304596,
0.05210523307323456,
-0.05351429060101509,
0.1292968988418579,
0.13068005442619324,
-0.03779344633221626,
0.14588125050067902,
-0.04201403260231018,
-0.09759873896837234,
-0.0275629423558712,
0.060269054025411606,
0.04299455136060715,
0.1323641985654831,
-0.08643611520528793,
-0.008810277096927166,
0.0037794618401676416,
0.024058189243078232,
-0.003241252154111862,
-0.20801737904548645,
-0.04711323231458664,
0.049527306109666824,
-0.06007126718759537,
-0.009587625041604042,
-0.031031612306833267,
-0.01979580707848072,
0.10805270075798035,
0.04147057980298996,
-0.04992794245481491,
0.006521475967019796,
-0.007985969074070454,
-0.08195959776639938,
0.22923940420150757,
-0.07881570607423782,
-0.12569239735603333,
-0.11099164932966232,
0.032395265996456146,
-0.006435516756027937,
0.017246192321181297,
0.026898810639977455,
-0.1048474907875061,
0.00765231205150485,
-0.07129216194152832,
0.021996350958943367,
-0.018425608053803444,
0.043669313192367554,
-0.015736885368824005,
0.023888105526566505,
0.05125381425023079,
-0.08316124230623245,
0.02004098892211914,
-0.018314961344003677,
-0.06106996163725853,
0.04581667482852936,
0.014833350665867329,
0.11500131338834763,
0.17401130497455597,
0.016433056443929672,
0.023452728986740112,
-0.04453542083501816,
0.1482185274362564,
-0.13319022953510284,
0.001625765929929912,
0.09823745489120483,
0.03378404676914215,
0.049755703657865524,
0.15822692215442657,
0.03857588395476341,
-0.10281109064817429,
0.056608326733112335,
0.042404428124427795,
-0.02320358157157898,
-0.21036282181739807,
0.002244724193587899,
-0.045017920434474945,
0.029901422560214996,
0.11004986613988876,
0.029563287273049355,
0.01336328499019146,
0.062211234122514725,
-0.022890618070960045,
-0.00113504973705858,
0.004907405469566584,
0.07233377546072006,
-0.017653336748480797,
0.030056118965148926,
0.12397616356611252,
-0.015287095680832863,
-0.04342450201511383,
0.008618727326393127,
0.015269899740815163,
0.22508643567562103,
-0.02247585728764534,
0.12591984868049622,
0.039368145167827606,
0.15993300080299377,
-0.013432650826871395,
0.09168799221515656,
0.03340648487210274,
-0.04956818372011185,
0.00774105079472065,
-0.05635824054479599,
-0.03361457586288452,
0.05839671939611435,
0.02072935178875923,
0.07043621689081192,
-0.1521901786327362,
0.03192400932312012,
0.04024084657430649,
0.33431482315063477,
0.09971035271883011,
-0.3189283311367035,
-0.09420498460531235,
0.01619325391948223,
-0.04252094775438309,
-0.034213803708553314,
0.013382314704358578,
0.12862113118171692,
-0.11353294551372528,
0.04034842923283577,
-0.08041573315858841,
0.07546770572662354,
-0.06203131005167961,
0.00015876528050284833,
0.05261068046092987,
0.07250155508518219,
-0.03206790238618851,
0.05945797637104988,
-0.28936561942100525,
0.30432602763175964,
-0.0044040498323738575,
0.07461471110582352,
-0.03822026029229164,
0.008890227414667606,
0.027601540088653564,
0.03478457033634186,
0.11401341110467911,
-0.006916560232639313,
-0.02298494428396225,
-0.21628841757774353,
-0.09778726100921631,
0.002944659674540162,
0.14647041261196136,
-0.1453273743391037,
0.13487577438354492,
-0.023679940029978752,
-0.026730766519904137,
0.04825345054268837,
-0.05530834570527077,
-0.08329782634973526,
-0.0720364972949028,
0.01519810315221548,
-0.03969558700919151,
0.10101742297410965,
-0.11194811761379242,
-0.09843920916318893,
-0.04434855654835701,
0.158989816904068,
-0.11675245314836502,
-0.020499031990766525,
-0.14466632902622223,
0.09054645150899887,
0.1106850653886795,
-0.07286716997623444,
0.045920394361019135,
0.017831921577453613,
0.09985584765672684,
0.009376807138323784,
0.013893039897084236,
0.12469647079706192,
-0.07602114975452423,
-0.24783490598201752,
-0.06980621069669724,
0.167972594499588,
0.04150720685720444,
0.06126847118139267,
-0.018109489232301712,
0.012539851479232311,
0.010600056499242783,
-0.0836348682641983,
0.06786344945430756,
0.010581123642623425,
0.05325433611869812,
0.049671925604343414,
-0.05286717042326927,
0.06815233826637268,
-0.07004944235086441,
-0.062047265470027924,
0.1324007511138916,
0.3355679214000702,
-0.09506475925445557,
0.012120394967496395,
0.03887570649385452,
-0.036458153277635574,
-0.18260280787944794,
0.05136963352560997,
0.10778460651636124,
0.04607197642326355,
0.0015279015060514212,
-0.18439477682113647,
0.04187397658824921,
0.10452140867710114,
-0.030938483774662018,
0.12241018563508987,
-0.32466450333595276,
-0.13539773225784302,
0.0664725974202156,
0.12560921907424927,
0.01585148088634014,
-0.1730673760175705,
-0.05832510069012642,
-0.01751522719860077,
-0.07380971312522888,
0.05130191892385483,
-0.0454782173037529,
0.12343648076057434,
-0.015434904024004936,
0.012109118513762951,
0.02966015227138996,
-0.06325877457857132,
0.1372428983449936,
-0.0012857320252805948,
0.0850125178694725,
-0.021005308255553246,
0.0038854656741023064,
0.02554411254823208,
-0.08925577998161316,
0.005282195750623941,
-0.07606121152639389,
0.032852910459041595,
-0.09741254895925522,
-0.02792515978217125,
-0.0917745903134346,
0.038189731538295746,
-0.06170092895627022,
-0.07393068820238113,
-0.017281729727983475,
0.0595838837325573,
0.05724593251943588,
-0.0013510325225070119,
0.10976922512054443,
-0.03921866789460182,
0.18225590884685516,
0.07982960343360901,
0.1029038280248642,
-0.008355315774679184,
-0.027645377442240715,
0.008673754520714283,
-0.018202025443315506,
0.05259736254811287,
-0.14804351329803467,
0.010556940920650959,
0.14660649001598358,
0.05328255519270897,
0.13901130855083466,
0.07555390149354935,
-0.048087816685438156,
-0.004725028295069933,
0.08785314857959747,
-0.10264267772436142,
-0.11910424381494522,
-0.015084323473274708,
-0.016766825690865517,
-0.15926708281040192,
0.05632311850786209,
0.09967894107103348,
-0.06310882419347763,
-0.005660750437527895,
-0.0008175664115697145,
0.017056426033377647,
-0.039309896528720856,
0.22929337620735168,
0.060547638684511185,
0.10772690176963806,
-0.07214393466711044,
0.07644837349653244,
0.031745120882987976,
-0.11711020022630692,
0.003923884592950344,
0.09152893722057343,
-0.08574406057596207,
-0.018868453800678253,
0.04425093159079552,
0.07217421382665634,
-0.000981608172878623,
-0.007890707813203335,
-0.14121508598327637,
-0.1286787986755371,
0.061816032975912094,
0.11611450463533401,
0.041897065937519073,
0.0382927767932415,
-0.006983769591897726,
0.04580812528729439,
-0.1357830911874771,
0.11491861939430237,
0.0746091678738594,
0.09491629153490067,
-0.15235039591789246,
0.18036499619483948,
-0.018182851374149323,
0.013630270026624203,
-0.007263095583766699,
0.028612008318305016,
-0.12328357249498367,
0.013471216894686222,
-0.060018591582775116,
-0.08023650199174881,
-0.045649170875549316,
-0.02839193120598793,
-0.010479751974344254,
-0.03734835237264633,
-0.019000491127371788,
-0.00452906871214509,
-0.10985293239355087,
-0.059966079890728,
-0.010889803059399128,
0.04000014439225197,
-0.09683898836374283,
-0.03521226346492767,
0.029122555628418922,
-0.118201345205307,
0.09356953203678131,
0.021271690726280212,
0.0534416064620018,
0.007374296896159649,
-0.10952266305685043,
0.04898844659328461,
0.03053857758641243,
-0.03398101031780243,
0.03196370601654053,
-0.12903688848018646,
-0.016519945114850998,
-0.06573431193828583,
0.02304915525019169,
0.02075185254216194,
0.013294477947056293,
-0.14386755228042603,
0.0034485189244151115,
-0.041786495596170425,
-0.039978817105293274,
-0.06911677867174149,
0.05201379954814911,
0.04817340150475502,
0.003200085135176778,
0.14630885422229767,
-0.07506194710731506,
0.05269031599164009,
-0.22328247129917145,
-0.01869121938943863,
-0.017912935465574265,
-0.07668829709291458,
-0.07064944505691528,
-0.033073555678129196,
0.09674135595560074,
-0.06153671070933342,
0.040852054953575134,
-0.05411355569958687,
0.06396740674972534,
0.030514126643538475,
-0.11174720525741577,
0.0844685509800911,
0.05730452015995979,
0.18890538811683655,
0.06209074333310127,
-0.04099464416503906,
0.048575565218925476,
0.05587826296687126,
0.07794009149074554,
0.08962651342153549,
0.1960088610649109,
0.1363075077533722,
-0.005435723811388016,
0.09724443405866623,
0.0240355022251606,
-0.11678548157215118,
-0.154751256108284,
0.0823429599404335,
-0.048575662076473236,
0.08763871341943741,
-0.029772724956274033,
0.18554583191871643,
0.12974026799201965,
-0.19714076817035675,
0.02088545635342598,
-0.03594181686639786,
-0.09262396395206451,
-0.09233143925666809,
-0.04516945034265518,
-0.07124422490596771,
-0.1729286015033722,
0.00708694476634264,
-0.10462687909603119,
0.01439319271594286,
0.08875797688961029,
0.024531356990337372,
0.029620494693517685,
0.17931009829044342,
0.07308264821767807,
0.0347161628305912,
0.1097850427031517,
0.027355968952178955,
0.007272040005773306,
-0.04251077026128769,
-0.11739242821931839,
0.008880861103534698,
-0.06585831195116043,
0.03655213117599487,
-0.07970229536294937,
-0.11217284947633743,
0.05702291429042816,
0.04305896535515785,
-0.11031433939933777,
0.020733056589961052,
0.006615865975618362,
0.06248035281896591,
0.07886362075805664,
0.015997987240552902,
-0.012307325378060341,
-0.02802083268761635,
0.2772132456302643,
-0.11136270314455032,
-0.04598482325673103,
-0.11344564706087112,
0.27199724316596985,
0.03172607347369194,
0.008440335281193256,
0.007786361500620842,
-0.10185393691062927,
0.018842536956071854,
0.16391609609127045,
0.16699260473251343,
-0.052148085087537766,
-0.011888507753610611,
0.012072290293872356,
-0.01831279881298542,
-0.040143806487321854,
0.08433147519826889,
0.11693698167800903,
0.03196170553565025,
-0.07778263092041016,
-0.0081762271001935,
-0.016429441049695015,
-0.06688258051872253,
-0.028373967856168747,
0.08254975825548172,
0.04146059975028038,
0.022931937128305435,
-0.04149489849805832,
0.11070756614208221,
-0.03516754135489464,
-0.12353457510471344,
0.07319691777229309,
-0.19408302009105682,
-0.1749657839536667,
-0.06264633685350418,
0.015250793658196926,
-0.0007033164729364216,
0.0727977454662323,
-0.0028421466704458,
-0.021567588672041893,
0.08989477157592773,
-0.0049685342237353325,
-0.026566877961158752,
-0.1199561059474945,
0.06213963404297829,
-0.07198730856180191,
0.19816182553768158,
-0.06237247586250305,
-0.024683790281414986,
0.13640107214450836,
0.02704271487891674,
-0.08908282220363617,
0.043661750853061676,
0.08770669996738434,
-0.08480359613895416,
0.0481908954679966,
0.16751140356063843,
-0.0369570329785347,
0.1230834499001503,
0.04101640731096268,
-0.1556067168712616,
0.023695239797234535,
-0.08842865377664566,
-0.057569582015275955,
-0.09115821868181229,
0.0075614782981574535,
-0.01672395132482052,
0.1408914476633072,
0.2375085949897766,
-0.06775188446044922,
0.024828819558024406,
-0.060225702822208405,
0.009576220996677876,
0.05646267533302307,
0.10662699490785599,
-0.03139420598745346,
-0.25877484679222107,
0.010245926678180695,
0.05044710263609886,
-0.005290117580443621,
-0.2688823640346527,
-0.09447211772203445,
0.025136690586805344,
-0.05850594490766525,
-0.07870104163885117,
0.10134585946798325,
0.06081615388393402,
0.06147501990199089,
-0.04771803319454193,
-0.11986732482910156,
-0.06203337758779526,
0.20277118682861328,
-0.17598405480384827,
-0.07657849043607712
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.7.1 | {"library_name": "peft", "base_model": "meta-llama/Llama-2-13b-chat-hf"} | null | NikitaZagainov/notebook-segmentation-llama-2-13b-1ep | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:meta-llama/Llama-2-13b-chat-hf",
"region:us"
] | 2024-02-08T09:20:29+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #safetensors #arxiv-1910.09700 #base_model-meta-llama/Llama-2-13b-chat-hf #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.7.1 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.7.1"
] | [
"TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-meta-llama/Llama-2-13b-chat-hf #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.7.1"
] | [
43,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-meta-llama/Llama-2-13b-chat-hf #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.7.1"
] | [
-0.11788157373666763,
0.2025103121995926,
-0.0028418477158993483,
0.02513594552874565,
0.0790175050497055,
0.014644909650087357,
0.05487748235464096,
0.1326894760131836,
0.030531780794262886,
0.11619700491428375,
0.07028989493846893,
0.12141784280538559,
0.1148410513997078,
0.22093915939331055,
0.0027113112155348063,
-0.16705907881259918,
0.018864480778574944,
-0.07344066351652145,
0.01571068912744522,
0.11816342175006866,
0.1429157853126526,
-0.10015536844730377,
0.07803085446357727,
-0.01991301029920578,
0.0036921887658536434,
-0.025843510404229164,
-0.06782865524291992,
-0.011558699421584606,
0.055011067539453506,
0.03277165815234184,
0.05685701593756676,
-0.012581875547766685,
0.08615994453430176,
-0.2717483341693878,
0.01910623349249363,
0.04222232475876808,
0.0004361399041954428,
0.082026906311512,
0.09685737639665604,
-0.04496683180332184,
0.12445446848869324,
-0.022378407418727875,
0.13381695747375488,
0.0902542695403099,
-0.09407521784305573,
-0.23464274406433105,
-0.06301796436309814,
0.07807459682226181,
0.1894497126340866,
0.08497016876935959,
-0.04263485223054886,
0.12393426895141602,
-0.062381669878959656,
0.0231337808072567,
0.0684867575764656,
-0.10580335557460785,
-0.06438134610652924,
0.06463466584682465,
0.13098333775997162,
0.07746705412864685,
-0.12264909595251083,
-0.03583741933107376,
0.037060678005218506,
0.04746377468109131,
0.059295933693647385,
0.005458815023303032,
0.1489870846271515,
0.029969438910484314,
-0.1480439305305481,
-0.049744170159101486,
0.1375211775302887,
0.008236280642449856,
-0.03655404597520828,
-0.21708692610263824,
-0.004266948439180851,
-0.09466829150915146,
-0.03919629752635956,
-0.04703955724835396,
0.03711909055709839,
0.00936040561646223,
0.13455705344676971,
-0.04934484884142876,
-0.0916595533490181,
-0.01584412157535553,
0.11003146320581436,
0.06276513636112213,
0.02184767834842205,
-0.020221911370754242,
0.007613510824739933,
0.12271928042173386,
0.06727160513401031,
-0.13353818655014038,
-0.06328637897968292,
-0.06740202009677887,
-0.03362444415688515,
-0.025729449465870857,
0.0390779934823513,
0.014166852459311485,
0.06217951700091362,
0.27106335759162903,
-0.036810554563999176,
0.06426587700843811,
0.04147655516862869,
0.02288985252380371,
0.03114578314125538,
0.10629495978355408,
-0.03427141532301903,
-0.16344094276428223,
-0.00669145630672574,
0.10196832567453384,
0.0051140859723091125,
-0.03337356075644493,
-0.05716389790177345,
0.03390507772564888,
0.03451387956738472,
0.11938710510730743,
0.10795657336711884,
-0.026408888399600983,
-0.07480237632989883,
-0.05597177892923355,
0.18918707966804504,
-0.1583651900291443,
0.04220565780997276,
0.02860168181359768,
-0.0006102732149884105,
-0.0638180747628212,
0.008215694688260555,
0.017930805683135986,
-0.032395463436841965,
0.07422833889722824,
-0.06666526198387146,
-0.04013180732727051,
-0.1222311481833458,
-0.030767129734158516,
0.03616037964820862,
0.011512357741594315,
-0.04547570273280144,
-0.044152457267045975,
-0.07094040513038635,
-0.10987216234207153,
0.10866580903530121,
-0.05359390377998352,
-0.05877101421356201,
-0.02803731895983219,
-0.08163430541753769,
0.01825849525630474,
0.034869302064180374,
0.07221972942352295,
-0.0267262514680624,
0.04608233645558357,
-0.008477536961436272,
0.06873008608818054,
0.06963228434324265,
0.031963542103767395,
-0.08018555492162704,
0.06634345650672913,
-0.20024055242538452,
0.07262307405471802,
-0.08050692081451416,
0.0455746054649353,
-0.15997160971164703,
-0.003640042617917061,
-0.0014792685396969318,
0.029864763841032982,
0.04312608018517494,
0.15939152240753174,
-0.21266384422779083,
-0.030124397948384285,
0.1682460457086563,
-0.10677676647901535,
-0.13465474545955658,
0.039859261363744736,
-0.037205055356025696,
0.18282857537269592,
0.027857886627316475,
0.030950404703617096,
0.08653967827558517,
-0.16244719922542572,
-0.019717056304216385,
-0.01817350648343563,
0.011286993511021137,
0.06657189875841141,
0.0814373567700386,
-0.09639275819063187,
-0.002265876391902566,
0.009880785830318928,
-0.06378284841775894,
-0.017002616077661514,
-0.040209949016571045,
-0.10548026859760284,
0.004797650501132011,
-0.08805537968873978,
0.0072769722901284695,
0.005076973233371973,
-0.09452961385250092,
-0.00788592267781496,
-0.1486291140317917,
-0.05621597543358803,
0.08575482666492462,
0.00014196978008840233,
-0.013805011287331581,
-0.0946660116314888,
0.06416139751672745,
-0.03400009125471115,
-0.020637493580579758,
-0.14446067810058594,
-0.016076795756816864,
0.017251502722501755,
-0.13779333233833313,
0.0012782185804098845,
-0.12384510040283203,
0.0669412910938263,
0.005050589330494404,
-0.04880156368017197,
-0.04315881058573723,
-0.001010204548947513,
-0.00524371862411499,
-0.061911750584840775,
-0.23732054233551025,
-0.02562497928738594,
-0.052236203104257584,
0.17158274352550507,
-0.23054468631744385,
0.042551323771476746,
0.0013218176318332553,
0.11761953681707382,
0.003304425161331892,
-0.05720871686935425,
0.024332204833626747,
-0.06145508959889412,
-0.024865947663784027,
-0.06902050226926804,
-0.0043387156911194324,
0.003128566313534975,
-0.028387483209371567,
0.017256038263440132,
-0.12189483642578125,
-0.06322982907295227,
0.0958312377333641,
0.05910428985953331,
-0.14457516372203827,
0.007243527565151453,
-0.03951241075992584,
-0.05686984956264496,
-0.06922618299722672,
-0.07263088971376419,
0.08621785789728165,
0.05200279504060745,
0.04850912466645241,
-0.08386287838220596,
-0.06838192790746689,
0.001768954680301249,
-0.0240500308573246,
-0.014463631436228752,
0.12614667415618896,
0.09333369135856628,
-0.09768560528755188,
0.0913463905453682,
0.07177400588989258,
0.02126719243824482,
0.08567578345537186,
-0.022206654772162437,
-0.10630354285240173,
-0.024423716589808464,
0.058100443333387375,
0.010638405568897724,
0.17069987952709198,
-0.07235661894083023,
0.053685713559389114,
0.04647723212838173,
-0.05855085328221321,
0.047724682837724686,
-0.09365744888782501,
0.00628670072183013,
-0.0019485035445541143,
-0.017316928133368492,
0.038373690098524094,
-0.016053196042776108,
0.004685666877776384,
0.08944613486528397,
0.06357455253601074,
0.020298874005675316,
0.011920131742954254,
-0.03656141087412834,
-0.1417326182126999,
0.18056967854499817,
-0.09292636066675186,
-0.23901212215423584,
-0.15010802447795868,
0.05421376973390579,
0.05745307356119156,
-0.013903340324759483,
0.030768904834985733,
-0.053937967866659164,
-0.09592998027801514,
-0.08850184828042984,
0.006417667958885431,
0.032451365143060684,
-0.06015152484178543,
-0.06340000778436661,
0.035531483590602875,
0.03849592059850693,
-0.12112338095903397,
0.02343169040977955,
0.05632079392671585,
-0.0007420660695061088,
-0.004698658362030983,
0.04573182389140129,
0.09440620988607407,
0.2061482071876526,
-0.0025023245252668858,
0.007018395699560642,
0.058498233556747437,
0.276040643453598,
-0.1591096669435501,
0.11200051009654999,
0.13979370892047882,
-0.06493698060512543,
0.07698789983987808,
0.19114595651626587,
0.0302424356341362,
-0.09487387537956238,
0.020369865000247955,
0.03167621046304703,
-0.02390752173960209,
-0.27111610770225525,
-0.051930975168943405,
-0.02317381091415882,
-0.07563389092683792,
0.08103558421134949,
0.08934853971004486,
0.08870835602283478,
0.028369644656777382,
-0.06447386741638184,
-0.09963097423315048,
0.02634870633482933,
0.11165431886911392,
-0.01618480123579502,
0.0005957336979918182,
0.08100581169128418,
-0.04910567030310631,
0.004032977391034365,
0.084804467856884,
-0.019175369292497635,
0.12482133507728577,
0.056135497987270355,
0.10594816505908966,
0.08346930146217346,
0.0840509682893753,
-0.011211014352738857,
0.029751107096672058,
0.001940281130373478,
0.02004975825548172,
0.020541656762361526,
-0.09210331737995148,
0.01743885688483715,
0.11583494395017624,
0.01319670770317316,
0.021101098507642746,
0.013549823313951492,
-0.05889787897467613,
0.0378522053360939,
0.19574348628520966,
0.029605528339743614,
-0.20708759129047394,
-0.07774027436971664,
0.054680973291397095,
-0.07424511015415192,
-0.15421795845031738,
-0.007879722863435745,
0.014505422674119473,
-0.1574283093214035,
0.019816888496279716,
-0.04044210910797119,
0.10735528916120529,
-0.06578231602907181,
-0.03894390910863876,
0.10502928495407104,
0.04858909547328949,
-0.028408242389559746,
0.04954361915588379,
-0.19317233562469482,
0.10876353085041046,
0.02961316891014576,
0.06624200195074081,
-0.08914101123809814,
0.08823274075984955,
-0.0008482593111693859,
-0.008602471090853214,
0.16474327445030212,
-0.0026781773194670677,
-0.060131706297397614,
-0.07745575159788132,
-0.07804002612829208,
-0.004643214866518974,
0.0805710107088089,
-0.13515672087669373,
0.0750945433974266,
-0.03372474014759064,
-0.03128623217344284,
-0.006927921902388334,
-0.0871417224407196,
-0.1181429773569107,
-0.1623523086309433,
0.06011633947491646,
-0.08327510952949524,
0.023717103525996208,
-0.08122113347053528,
-0.052879225462675095,
0.03087249957025051,
0.17839385569095612,
-0.2002856582403183,
-0.10983742028474808,
-0.14319008588790894,
-0.10384400933980942,
0.15116243064403534,
-0.04727339744567871,
0.08746539801359177,
-0.006882337387651205,
0.16186656057834625,
-0.0018413515063002706,
-0.019694453105330467,
0.08511307835578918,
-0.09525609016418457,
-0.18207688629627228,
-0.04612530767917633,
0.18390944600105286,
0.13041752576828003,
0.02810804545879364,
-0.011225296184420586,
0.024338265880942345,
-0.06634529680013657,
-0.10864581912755966,
0.028247011825442314,
0.149430513381958,
0.06784652173519135,
-0.020046968013048172,
-0.04459109902381897,
-0.09517679363489151,
-0.06562554091215134,
-0.043474745005369186,
-0.002455809386447072,
0.20311576128005981,
-0.07044374942779541,
0.15442033112049103,
0.1094876080751419,
-0.059697918593883514,
-0.21334324777126312,
0.0338175892829895,
0.03936067223548889,
0.01768609881401062,
0.03307800367474556,
-0.1929045170545578,
0.08791132271289825,
-0.026313822716474533,
-0.08250562101602554,
0.17991600930690765,
-0.1986837387084961,
-0.1296905279159546,
0.10796400904655457,
0.023953251540660858,
-0.20258675515651703,
-0.15128712356090546,
-0.10375212132930756,
-0.019056186079978943,
-0.1167878732085228,
0.044270843267440796,
0.00699279410764575,
0.012187452986836433,
0.012179792858660221,
0.02266608737409115,
0.041021887212991714,
-0.048118624836206436,
0.2028307318687439,
-0.04459221661090851,
-0.004416223615407944,
-0.05423783138394356,
-0.07714637368917465,
0.01167360320687294,
-0.05537216737866402,
0.1259775459766388,
-0.01797424629330635,
0.032846808433532715,
-0.16335023939609528,
-0.04316803067922592,
-0.06145013868808746,
0.036946866661310196,
-0.09557046741247177,
-0.08004589378833771,
-0.04436483606696129,
0.08199062943458557,
0.09042184799909592,
-0.012540708296000957,
0.013072513975203037,
-0.09800484776496887,
0.09410175681114197,
0.19926108419895172,
0.19393891096115112,
0.05995427817106247,
-0.05162312835454941,
0.03133172169327736,
-0.03741470351815224,
0.044728927314281464,
-0.22015799582004547,
0.04205537587404251,
0.0645650252699852,
0.02615460939705372,
0.06876256316900253,
-0.006028305739164352,
-0.1625821590423584,
-0.09218986332416534,
0.08959945291280746,
-0.06323622167110443,
-0.17259353399276733,
-0.03376561775803566,
0.042873565107584,
-0.2088049352169037,
-0.04544130712747574,
0.037715714424848557,
-0.017989275977015495,
-0.041428472846746445,
0.02545454353094101,
0.08015990257263184,
-0.02190752513706684,
0.08719413727521896,
0.09560935199260712,
0.08916150033473969,
-0.0953352078795433,
0.05223952978849411,
0.07872436940670013,
-0.018873462453484535,
0.03033655695617199,
0.14002232253551483,
-0.03666146099567413,
-0.046344488859176636,
0.07933306694030762,
0.12037548422813416,
-0.003258864628151059,
-0.05549774318933487,
0.0031455522403120995,
-0.049705665558576584,
0.06106950342655182,
0.12411541491746902,
0.02340015582740307,
-0.012639104388654232,
0.07976052910089493,
0.024764331057667732,
-0.09161490201950073,
0.12356899678707123,
0.040597643703222275,
0.021518969908356667,
-0.03645100072026253,
-0.027004897594451904,
-0.013607359491288662,
0.00021324573026504368,
-0.014775843359529972,
0.00006522652256535366,
-0.08998338133096695,
0.0033896011300385,
-0.1141517162322998,
0.016514858230948448,
-0.06856909394264221,
-0.0005768302944488823,
0.02871003746986389,
-0.04715637490153313,
-0.003126622876152396,
-0.004235076252371073,
-0.07826890051364899,
-0.052869509905576706,
-0.023299960419535637,
0.07778995484113693,
-0.1407601535320282,
0.03323814272880554,
0.07304537296295166,
-0.1028575524687767,
0.06794416904449463,
-0.009040433913469315,
0.012576045468449593,
0.006519954185932875,
-0.1437160074710846,
0.05540407821536064,
-0.027348563075065613,
-0.006057131104171276,
0.0018772223265841603,
-0.18099193274974823,
-0.011497852392494678,
-0.042367879301309586,
-0.0702858716249466,
0.013803095556795597,
-0.011336525902152061,
-0.12389353662729263,
0.11192979663610458,
0.008017337881028652,
-0.06569766253232956,
-0.01413482241332531,
0.04526352137327194,
0.06988541036844254,
-0.012181113474071026,
0.10690586268901825,
-0.028097203001379967,
0.08164410293102264,
-0.1796591877937317,
-0.005779837723821402,
-0.017756231129169464,
0.05352712422609329,
-0.01982288621366024,
-0.04605138301849365,
0.055983953177928925,
-0.021016502752900124,
0.16672296822071075,
0.0010196286020800471,
0.07270368188619614,
0.05241123586893082,
0.011164604686200619,
0.04950634762644768,
0.0723496824502945,
0.06387747824192047,
-0.017573459073901176,
-0.0037699334789067507,
0.035219114273786545,
-0.0003119460598099977,
-0.043233949691057205,
-0.1378021389245987,
0.0725247859954834,
0.17799563705921173,
0.07049155980348587,
0.023135408759117126,
0.010230054147541523,
-0.1344994306564331,
-0.0724371075630188,
0.10357820242643356,
-0.016800789162516594,
-0.030321190133690834,
-0.06634638458490372,
0.22777515649795532,
0.15011954307556152,
-0.1912444680929184,
0.0742524191737175,
-0.053979091346263885,
-0.03821665421128273,
-0.14468394219875336,
-0.167638897895813,
-0.05780312418937683,
-0.04853709787130356,
-0.03250361606478691,
-0.05885395035147667,
0.050874046981334686,
0.039369627833366394,
-0.004999021999537945,
-0.02146909572184086,
0.1112421303987503,
0.030625011771917343,
-0.04050149768590927,
0.04534582421183586,
0.06154803931713104,
0.04334854707121849,
-0.10070514678955078,
0.010876684449613094,
0.0014880468370392919,
0.005646043922752142,
0.06036636605858803,
0.02259848453104496,
-0.06996119022369385,
0.030343232676386833,
-0.01802264340221882,
-0.11921427398920059,
0.04814288020133972,
-0.0069939629174768925,
-0.019864631816744804,
0.14960302412509918,
0.03607213869690895,
0.006992223672568798,
-0.010939210653305054,
0.23891356587409973,
-0.07272490113973618,
-0.08256373554468155,
-0.1304820328950882,
0.08596840500831604,
-0.06325113773345947,
0.024178164079785347,
0.014769579283893108,
-0.123263418674469,
0.012303249910473824,
0.1815977543592453,
0.11888858675956726,
-0.01997954212129116,
0.012945982627570629,
0.04363260790705681,
0.009745429269969463,
-0.035624321550130844,
0.012733696028590202,
0.05843547731637955,
0.20640087127685547,
-0.07694563269615173,
0.05828242376446724,
-0.018566392362117767,
-0.06911972910165787,
-0.03318094462156296,
0.10627373307943344,
-0.01030859723687172,
-0.011211195029318333,
-0.056580208241939545,
0.1417265683412552,
-0.07411091029644012,
-0.2113642394542694,
0.049746204167604446,
-0.0821489617228508,
-0.13837212324142456,
-0.04993110150098801,
0.0277590099722147,
-0.026507118716835976,
0.006801436189562082,
0.059315625578165054,
-0.05265038460493088,
0.18020527064800262,
0.030894065275788307,
-0.045119646936655045,
-0.09781262278556824,
0.05726005136966705,
-0.1642291098833084,
0.28240618109703064,
0.021421385928988457,
0.04801954701542854,
0.10929666459560394,
-0.02242736890912056,
-0.13177204132080078,
0.016046447679400444,
0.11444000154733658,
-0.06328914314508438,
0.0635862722992897,
0.15894746780395508,
0.0027237553149461746,
0.1242469772696495,
0.06435003876686096,
-0.05605313554406166,
0.03743471950292587,
-0.07060305029153824,
-0.052798982709646225,
-0.11813118308782578,
0.07744082808494568,
-0.09802453964948654,
0.1527501940727234,
0.12261880934238434,
-0.07232960313558578,
-0.0021184224169701338,
-0.0217424388974905,
0.0801442340016365,
0.018526429310441017,
0.10929414629936218,
0.008120529353618622,
-0.18666334450244904,
0.04575074836611748,
0.005044445861130953,
0.09829195588827133,
-0.2138538807630539,
-0.04899514466524124,
0.04227651283144951,
-0.01758539117872715,
-0.08486651629209518,
0.11394713073968887,
0.03988828510046005,
0.017125489190220833,
-0.034961339086294174,
-0.04907047376036644,
0.01733650080859661,
0.15269286930561066,
-0.1060781478881836,
-0.014094846323132515
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.