sha
null | last_modified
null | library_name
stringclasses 154
values | text
stringlengths 1
900k
| metadata
stringlengths 2
348k
| pipeline_tag
stringclasses 45
values | id
stringlengths 5
122
| tags
sequencelengths 1
1.84k
| created_at
stringlengths 25
25
| arxiv
sequencelengths 0
201
| languages
sequencelengths 0
1.83k
| tags_str
stringlengths 17
9.34k
| text_str
stringlengths 0
389k
| text_lists
sequencelengths 0
722
| processed_texts
sequencelengths 1
723
| tokens_length
sequencelengths 1
723
| input_texts
sequencelengths 1
61
| embeddings
sequencelengths 768
768
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | muzammil-eds/tinyllama-3T-64k-JSONExtractor-v3 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T06:07:10+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
56,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06061961501836777,
0.15481999516487122,
-0.004844071343541145,
0.02074851468205452,
0.0983177199959755,
0.007407687604427338,
0.07119518518447876,
0.11185134947299957,
-0.023851769044995308,
0.1167980208992958,
0.031993988901376724,
0.09781743586063385,
0.11217817664146423,
0.16186554729938507,
0.0015333457849919796,
-0.22897611558437347,
0.049678247421979904,
-0.125278040766716,
-0.0294334813952446,
0.11977242678403854,
0.1422213912010193,
-0.10954539477825165,
0.0752737894654274,
-0.038042325526475906,
-0.005828251596540213,
-0.0323176346719265,
-0.06205610930919647,
-0.05266609415411949,
0.05311284959316254,
0.06794639676809311,
0.07308239489793777,
0.01171939354389906,
0.09106900542974472,
-0.2724283039569855,
0.02348201349377632,
0.0805930644273758,
-0.0006441773730330169,
0.07586129754781723,
0.04993962123990059,
-0.08749990910291672,
0.07524524629116058,
-0.060156844556331635,
0.1498761922121048,
0.07955671846866608,
-0.09018243104219437,
-0.19217631220817566,
-0.07921334356069565,
0.09916994720697403,
0.1890910118818283,
0.05953684076666832,
-0.026427440345287323,
0.11642678081989288,
-0.08593545109033585,
0.013638701289892197,
0.06446459144353867,
-0.06054406240582466,
-0.055855002254247665,
0.06904532760381699,
0.08335285633802414,
0.08567540347576141,
-0.12976622581481934,
-0.010767064057290554,
0.015032444149255753,
0.008952446281909943,
0.08948688954114914,
0.017146794125437737,
0.1335189938545227,
0.040557652711868286,
-0.13501930236816406,
-0.043155476450920105,
0.09761431813240051,
0.03665134683251381,
-0.04888195917010307,
-0.2485782504081726,
-0.023432478308677673,
-0.04339504987001419,
-0.03198111802339554,
-0.03649339824914932,
0.043764639645814896,
-0.014506848528981209,
0.07738617807626724,
-0.004502781666815281,
-0.0837155357003212,
-0.04301247000694275,
0.07241875678300858,
0.06128999963402748,
0.02571401372551918,
-0.015821760520339012,
0.0059297760017216206,
0.12327717989683151,
0.11431120336055756,
-0.126715749502182,
-0.052547648549079895,
-0.06306339055299759,
-0.08449548482894897,
-0.044861067086458206,
0.030838407576084137,
0.037995077669620514,
0.045936476439237595,
0.23867325484752655,
0.007765117567032576,
0.053257301449775696,
0.04455438256263733,
0.014407169073820114,
0.06501194834709167,
0.11008983850479126,
-0.05894824117422104,
-0.09719445556402206,
-0.028582042083144188,
0.10156717151403427,
0.007986726239323616,
-0.04139331728219986,
-0.05712985619902611,
0.07059531658887863,
0.018587570637464523,
0.12360043078660965,
0.08000938594341278,
0.003056557849049568,
-0.0755772516131401,
-0.062465377151966095,
0.17764076590538025,
-0.15825673937797546,
0.04532013460993767,
0.03055616281926632,
-0.0341108962893486,
-0.009745313785970211,
0.012105142697691917,
0.025474950671195984,
-0.021481726318597794,
0.09522198140621185,
-0.05601342022418976,
-0.034448131918907166,
-0.11389608681201935,
-0.03694311901926994,
0.030394554138183594,
0.011153047904372215,
-0.02865210548043251,
-0.03502652049064636,
-0.08865131437778473,
-0.06405586749315262,
0.09101516753435135,
-0.07148737460374832,
-0.04784895107150078,
-0.016645915806293488,
-0.07833752781152725,
0.021804187446832657,
0.01691517047584057,
0.09064167737960815,
-0.0222476739436388,
0.03985358029603958,
-0.0550384595990181,
0.061440225690603256,
0.11723454296588898,
0.027987057343125343,
-0.05787884071469307,
0.061519939452409744,
-0.2424532175064087,
0.10252492874860764,
-0.07715212553739548,
0.04971238598227501,
-0.15203025937080383,
-0.02478341944515705,
0.03986154496669769,
0.01284773275256157,
-0.008251311257481575,
0.14196595549583435,
-0.21994100511074066,
-0.030957341194152832,
0.16964265704154968,
-0.10025953501462936,
-0.08109250664710999,
0.060782887041568756,
-0.05354252830147743,
0.11210215091705322,
0.04557164013385773,
-0.02375967986881733,
0.05775221437215805,
-0.14725260436534882,
-0.011030761525034904,
-0.041942402720451355,
-0.0180682260543108,
0.16207332909107208,
0.0703711211681366,
-0.06047816202044487,
0.07456906884908676,
0.01960151270031929,
-0.014246034435927868,
-0.04887177795171738,
-0.02822130173444748,
-0.1047162413597107,
0.01184528972953558,
-0.06102835759520531,
0.018109694123268127,
-0.021768750622868538,
-0.09445013850927353,
-0.029118487611413002,
-0.17402999103069305,
-0.0031633328180760145,
0.08821269869804382,
-0.011630427092313766,
-0.021509924903512,
-0.11245372891426086,
0.009332616813480854,
0.030967719852924347,
0.0002618339203763753,
-0.13677829504013062,
-0.06033218279480934,
0.026970699429512024,
-0.16097871959209442,
0.029791243374347687,
-0.05741601809859276,
0.04530094936490059,
0.04005871340632439,
-0.03433511033654213,
-0.03489551320672035,
0.010874404571950436,
0.010431389324367046,
-0.01894843392074108,
-0.25422003865242004,
-0.01882786676287651,
-0.0234990194439888,
0.1751047968864441,
-0.22956320643424988,
0.042598169296979904,
0.07489731162786484,
0.1460893303155899,
0.007349682506173849,
-0.03550100699067116,
0.015185600146651268,
-0.07262228429317474,
-0.03268764168024063,
-0.06316669285297394,
-0.01207790058106184,
-0.038400664925575256,
-0.05820201337337494,
0.04906858503818512,
-0.1686294972896576,
-0.030321966856718063,
0.10717973858118057,
0.06342670321464539,
-0.1473218947649002,
-0.02780107781291008,
-0.04056945815682411,
-0.04624456167221069,
-0.06676914542913437,
-0.05461418256163597,
0.11812574416399002,
0.056411582976579666,
0.04860803112387657,
-0.07140495628118515,
-0.07455260306596756,
0.008036690764129162,
-0.01956399530172348,
-0.014917809516191483,
0.09334591031074524,
0.07554110884666443,
-0.12264352291822433,
0.09177418053150177,
0.09668384492397308,
0.08576478064060211,
0.10314212739467621,
-0.014663571491837502,
-0.08914592862129211,
-0.040637146681547165,
0.02245822176337242,
0.016187267377972603,
0.15129362046718597,
-0.012961224652826786,
0.055492039769887924,
0.0358695350587368,
-0.014034898020327091,
0.011105312965810299,
-0.09736533463001251,
0.02655916102230549,
0.030835967510938644,
-0.016302183270454407,
0.03745110332965851,
-0.0447014644742012,
0.019208140671253204,
0.09039704501628876,
0.040895868092775345,
0.040978945791721344,
0.010155045427381992,
-0.04354988783597946,
-0.11037563532590866,
0.1787576973438263,
-0.12389461696147919,
-0.24818050861358643,
-0.13812170922756195,
0.010281167924404144,
0.04737642779946327,
-0.010411068797111511,
0.006690691225230694,
-0.06616118550300598,
-0.1175973042845726,
-0.09878289699554443,
0.018617089837789536,
0.045352302491664886,
-0.07590975612401962,
-0.06842505931854248,
0.06414616107940674,
0.03875524550676346,
-0.13939815759658813,
0.024007495492696762,
0.04662325978279114,
-0.08205481618642807,
-0.0029386086389422417,
0.0791812464594841,
0.06965780258178711,
0.17661017179489136,
0.013885351829230785,
-0.023669935762882233,
0.026634456589818,
0.20819635689258575,
-0.1436755359172821,
0.10975687950849533,
0.13545554876327515,
-0.08767466992139816,
0.08120133727788925,
0.1998777538537979,
0.03777998685836792,
-0.10680917650461197,
0.03608465939760208,
0.028374753892421722,
-0.028325283899903297,
-0.2502254545688629,
-0.06958996504545212,
0.0019060121849179268,
-0.05172049254179001,
0.07064855098724365,
0.08791537582874298,
0.09593888372182846,
0.016860228031873703,
-0.09976044297218323,
-0.07697858661413193,
0.046900223940610886,
0.10824491083621979,
-0.00015424020239152014,
-0.015208319760859013,
0.0904119610786438,
-0.03033481352031231,
0.01743943803012371,
0.09215071052312851,
0.0030607767403125763,
0.17535938322544098,
0.051709048449993134,
0.17189906537532806,
0.07866133749485016,
0.06444311141967773,
0.02004685252904892,
0.007725914940237999,
0.021817529574036598,
0.017227526754140854,
-0.0030957073904573917,
-0.08709781616926193,
-0.0034981227945536375,
0.1202581599354744,
0.049845851957798004,
0.029173865914344788,
0.012042860500514507,
-0.030704669654369354,
0.08337877690792084,
0.1770893782377243,
0.0029054484330117702,
-0.1893385946750641,
-0.07169844210147858,
0.07795937359333038,
-0.08648337423801422,
-0.10729733109474182,
-0.029470939189195633,
0.041069481521844864,
-0.1729043871164322,
0.016882894560694695,
-0.019335895776748657,
0.10788324475288391,
-0.13190391659736633,
-0.01772487722337246,
0.05657728388905525,
0.06932812184095383,
-0.009677323512732983,
0.06694949418306351,
-0.16090403497219086,
0.11770165711641312,
0.01751571334898472,
0.06636732816696167,
-0.09608277678489685,
0.09618937969207764,
-0.007830657996237278,
0.0041499207727611065,
0.1410749852657318,
0.010120149701833725,
-0.05952107161283493,
-0.09608154743909836,
-0.10546442121267319,
-0.009841260500252247,
0.1306990385055542,
-0.14852415025234222,
0.08813067525625229,
-0.02661319263279438,
-0.044553373008966446,
0.003614129964262247,
-0.12497276812791824,
-0.13103094696998596,
-0.18366187810897827,
0.05707118660211563,
-0.12947207689285278,
0.04045100137591362,
-0.10902881622314453,
-0.045833900570869446,
-0.02098964899778366,
0.20040063560009003,
-0.23137451708316803,
-0.06714103370904922,
-0.1551055610179901,
-0.08061286807060242,
0.14446212351322174,
-0.046455029398202896,
0.08550118654966354,
0.0008278203313238919,
0.19068008661270142,
0.021319707855582237,
-0.017237508669495583,
0.1072206199169159,
-0.10052918642759323,
-0.2010865956544876,
-0.09273224323987961,
0.15895552933216095,
0.13766798377037048,
0.03809428587555885,
-0.004381525795906782,
0.03171157464385033,
-0.02098114788532257,
-0.12076930701732635,
0.020226983353495598,
0.17317426204681396,
0.08982043713331223,
0.025265544652938843,
-0.02972041629254818,
-0.11267432570457458,
-0.07061342149972916,
-0.03774050623178482,
0.024755435064435005,
0.18072067201137543,
-0.07222156971693039,
0.18405316770076752,
0.13775517046451569,
-0.05534014105796814,
-0.19904261827468872,
0.021996473893523216,
0.04293542355298996,
0.0070380112156271935,
0.0323902890086174,
-0.20307663083076477,
0.09384101629257202,
0.0008334947633557022,
-0.05131231248378754,
0.1379684954881668,
-0.1823476254940033,
-0.151598259806633,
0.06042521819472313,
0.043563615530729294,
-0.19374065101146698,
-0.12374074012041092,
-0.08848230540752411,
-0.04693066328763962,
-0.15487661957740784,
0.10312657803297043,
0.0020827590487897396,
0.008401188999414444,
0.03778626397252083,
0.02252252586185932,
0.012139533646404743,
-0.04198719933629036,
0.1914343535900116,
-0.025891713798046112,
0.03347287327051163,
-0.0790715217590332,
-0.060851071029901505,
0.062408581376075745,
-0.058187782764434814,
0.0755455270409584,
-0.025226406753063202,
0.015947066247463226,
-0.10598332434892654,
-0.048235729336738586,
-0.02852320298552513,
0.019321219995617867,
-0.09431382268667221,
-0.09348297864198685,
-0.04829427972435951,
0.09367614984512329,
0.09042316675186157,
-0.03652578964829445,
-0.03649144619703293,
-0.078715980052948,
0.038977332413196564,
0.17627815902233124,
0.18159319460391998,
0.04659178853034973,
-0.07959239184856415,
-0.001915142871439457,
-0.014336181804537773,
0.04684065282344818,
-0.22077152132987976,
0.060553863644599915,
0.04557652771472931,
0.016117896884679794,
0.11537692695856094,
-0.0208132341504097,
-0.16198977828025818,
-0.06710557639598846,
0.061360616236925125,
-0.06944561004638672,
-0.17825035750865936,
0.0039279889315366745,
0.07344977557659149,
-0.16578389704227448,
-0.037031736224889755,
0.04200848564505577,
-0.01189455483108759,
-0.0403641052544117,
0.012352054007351398,
0.08063354343175888,
0.007078902795910835,
0.07699975371360779,
0.055281639099121094,
0.09124495089054108,
-0.10227900743484497,
0.07410510629415512,
0.08149529248476028,
-0.08644098788499832,
0.030720343813300133,
0.09573426842689514,
-0.06469762325286865,
-0.0346054881811142,
0.04237886518239975,
0.08354541659355164,
0.024281201884150505,
-0.04682289808988571,
0.0023111123591661453,
-0.09734189510345459,
0.05927345156669617,
0.11483542621135712,
0.03496333956718445,
0.011234734207391739,
0.03813567012548447,
0.04486291855573654,
-0.08093374222517014,
0.11926916986703873,
0.023795632645487785,
0.020354853942990303,
-0.04112942889332771,
-0.040553025901317596,
0.035851649940013885,
-0.026020776480436325,
-0.011440055444836617,
-0.035174157470464706,
-0.0722682997584343,
-0.014069457538425922,
-0.16000694036483765,
-0.0076758842915296555,
-0.03660871088504791,
0.005114538595080376,
0.022510098293423653,
-0.03652830421924591,
0.00792311318218708,
0.012217256240546703,
-0.06868947297334671,
-0.05553458258509636,
-0.023233558982610703,
0.09422210603952408,
-0.16494666039943695,
0.0220257006585598,
0.0823851153254509,
-0.12121747434139252,
0.09289738535881042,
0.016782134771347046,
0.00412249518558383,
0.026962365955114365,
-0.1545863002538681,
0.04763968288898468,
-0.020152103155851364,
0.013473534025251865,
0.04222847521305084,
-0.21637047827243805,
-0.004404853098094463,
-0.04015503451228142,
-0.05566934496164322,
-0.008993052877485752,
-0.0319182425737381,
-0.11338426172733307,
0.09645436704158783,
0.011025024577975273,
-0.08443772792816162,
-0.02965564839541912,
0.03353232145309448,
0.07690354436635971,
-0.027447547763586044,
0.1498211771249771,
-0.004663881380110979,
0.07559948414564133,
-0.17581342160701752,
-0.02282017655670643,
-0.011197620071470737,
0.022367527708411217,
-0.021871577948331833,
-0.01622559316456318,
0.04623444378376007,
-0.02704801969230175,
0.19120801985263824,
-0.024701936170458794,
0.049393873661756516,
0.06364397704601288,
0.009232889860868454,
-0.013832193799316883,
0.11151392012834549,
0.05708572641015053,
0.024334950372576714,
0.022262847051024437,
0.003451440716162324,
-0.04008655622601509,
-0.009981024079024792,
-0.18596695363521576,
0.06803664565086365,
0.14585918188095093,
0.09060460329055786,
-0.012669353745877743,
0.0707244873046875,
-0.10161512345075607,
-0.12005364894866943,
0.10127941519021988,
-0.06415384262800217,
-0.010188822634518147,
-0.06542414426803589,
0.14027701318264008,
0.14953285455703735,
-0.1886233240365982,
0.06583356112241745,
-0.06602055579423904,
-0.0566304549574852,
-0.11457879096269608,
-0.1930263340473175,
-0.057075321674346924,
-0.050602465867996216,
-0.018466074019670486,
-0.05384097993373871,
0.06939727067947388,
0.05750798434019089,
0.01126816775649786,
0.00868057832121849,
0.08568526059389114,
-0.009656033478677273,
0.00248199631460011,
0.030120067298412323,
0.06713981181383133,
0.016768986359238625,
-0.0321255661547184,
0.0179112758487463,
-0.00597198773175478,
0.034156378358602524,
0.059282708913087845,
0.03608176112174988,
-0.028436895459890366,
0.015559280291199684,
-0.034912437200546265,
-0.11309733241796494,
0.042801856994628906,
-0.029640642926096916,
-0.0749855786561966,
0.1347348988056183,
0.026981467381119728,
0.005015076603740454,
-0.023140020668506622,
0.2503887414932251,
-0.07436972856521606,
-0.09334370493888855,
-0.14373961091041565,
0.11701542884111404,
-0.04212593287229538,
0.0635172426700592,
0.03596310690045357,
-0.10810714215040207,
0.017985546961426735,
0.1320217251777649,
0.15442703664302826,
-0.04732590913772583,
0.019251897931098938,
0.028577854856848717,
0.00439635943621397,
-0.04075566306710243,
0.05177190154790878,
0.07100846618413925,
0.14500564336776733,
-0.05157303810119629,
0.08530787378549576,
0.002609728369861841,
-0.1021018698811531,
-0.041973695158958435,
0.11415864527225494,
-0.014296893030405045,
0.017620453611016273,
-0.057136841118335724,
0.124222531914711,
-0.05874236673116684,
-0.23697422444820404,
0.06316976249217987,
-0.0765061303973198,
-0.1432730257511139,
-0.024886758998036385,
0.071670763194561,
-0.016632623970508575,
0.02605951391160488,
0.07167234271764755,
-0.0754380151629448,
0.18880942463874817,
0.03957989811897278,
-0.05233397334814072,
-0.05954399332404137,
0.0744764655828476,
-0.11850855499505997,
0.27879106998443604,
0.010482731275260448,
0.051307905465364456,
0.1042102724313736,
-0.02021743729710579,
-0.13270841538906097,
0.023401619866490364,
0.09579801559448242,
-0.08917027711868286,
0.04087764397263527,
0.21448291838169098,
-0.00629545608535409,
0.11935057491064072,
0.07611140608787537,
-0.07468950748443604,
0.047562725841999054,
-0.11468592286109924,
-0.07639975845813751,
-0.08699081838130951,
0.09244474768638611,
-0.06785612553358078,
0.14258281886577606,
0.12599852681159973,
-0.05530165135860443,
0.011584274470806122,
-0.028389399871230125,
0.045467376708984375,
0.005578654818236828,
0.100032277405262,
0.011115525849163532,
-0.18496567010879517,
0.024811718612909317,
0.016259413212537766,
0.10884406417608261,
-0.18112654983997345,
-0.09105053544044495,
0.046958595514297485,
0.0005061255069449544,
-0.06443515419960022,
0.12483241409063339,
0.057313691824674606,
0.04654949903488159,
-0.0451689288020134,
-0.026830285787582397,
-0.006042256020009518,
0.14264579117298126,
-0.10707559436559677,
-0.005129707511514425
] |
null | null | transformers | Everyone-Coder-33b-v2-Base

EveryoneLLM series of models made by the community, for the community. This is a coding specific model made using fine-tunes of deekseekcoder-33b-base.
This Version 2 of the Everything-Coder-33b model uses the task_arithmetic merging method which has major increases in coding performance as opposed to the ties method. You should find this version having much better coding performance than Version 1, without any of the negative that merging has on the integrity of the model.
Prompt template: Alpaca
```
Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
{prompt}
### Response:
```
The models that were used in this merger were as follow:
- https://huggingface.co/deepseek-ai/deepseek-coder-33b-instruct
- https://huggingface.co/codefuse-ai/CodeFuse-DeepSeek-33B
- https://huggingface.co/WizardLM/WizardCoder-33B-V1.1
Thank you to the creators of the above ai models, they have full credit for the EveryoneLLM series of models. Without their hard work we wouldnt be able to achieve the great success we have in the open source community. 💗
You can find the write up for merging models here:
https://docs.google.com/document/d/1_vOftBnrk9NRk5h10UqrfJ5CDih9KBKL61yvrZtVWPE/edit?usp=sharing
Config for the merger can be found bellow:
```yaml
models:
- model: codefuse-ai_CodeFuse-DeepSeek-33B
parameters:
weight: 1
- model: deepseek-ai_deepseek-coder-33b-instruct
parameters:
weight: 1
- model: WizardLM_WizardCoder-33B-V1.1
parameters:
weight: 1
merge_method: task_arithmetic
base_model: deepseek-ai_deepseek-coder-33b-base
parameters:
normalize: true
int8_mask: true
dtype: float16
```
| {"license": "other", "tags": ["merge"], "license_name": "deepseek", "license_link": "https://github.com/deepseek-ai/DeepSeek-Coder/blob/main/LICENSE-MODEL"} | text-generation | rombodawg/Everyone-Coder-33b-v2-Base | [
"transformers",
"safetensors",
"llama",
"text-generation",
"merge",
"license:other",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T06:08:00+00:00 | [] | [] | TAGS
#transformers #safetensors #llama #text-generation #merge #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| Everyone-Coder-33b-v2-Base
!image/jpeg
EveryoneLLM series of models made by the community, for the community. This is a coding specific model made using fine-tunes of deekseekcoder-33b-base.
This Version 2 of the Everything-Coder-33b model uses the task_arithmetic merging method which has major increases in coding performance as opposed to the ties method. You should find this version having much better coding performance than Version 1, without any of the negative that merging has on the integrity of the model.
Prompt template: Alpaca
The models that were used in this merger were as follow:
- URL
- URL
- URL
Thank you to the creators of the above ai models, they have full credit for the EveryoneLLM series of models. Without their hard work we wouldnt be able to achieve the great success we have in the open source community.
You can find the write up for merging models here:
URL
Config for the merger can be found bellow:
| [] | [
"TAGS\n#transformers #safetensors #llama #text-generation #merge #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
55
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #merge #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
-0.00496361730620265,
0.013435622677206993,
-0.0048941983841359615,
-0.007948110811412334,
0.11774998903274536,
0.010081208311021328,
0.20247454941272736,
0.10724741965532303,
-0.0081727784126997,
-0.02794553153216839,
0.11637203395366669,
0.19246558845043182,
-0.020575184375047684,
0.0786113366484642,
-0.11944705247879028,
-0.19450408220291138,
0.08193118125200272,
0.01286437176167965,
0.019434256479144096,
0.07955759763717651,
0.10736028850078583,
-0.052583519369363785,
0.08838794380426407,
-0.04033555090427399,
-0.10580462217330933,
0.05240209400653839,
0.050350528210401535,
-0.12679429352283478,
0.10726121813058853,
0.09131157398223877,
0.09965533018112183,
0.053541526198387146,
-0.03159579634666443,
-0.21373650431632996,
0.026776643469929695,
-0.007983626797795296,
-0.08475162088871002,
0.023004120215773582,
0.08671828359365463,
-0.07371804863214493,
0.07580311596393585,
0.05730024352669716,
0.00030684471130371094,
0.08479351550340652,
-0.1292436569929123,
-0.030244316905736923,
-0.06460942327976227,
-0.013680852018296719,
0.12005474418401718,
0.0691259503364563,
0.0016497154720127583,
0.09854160249233246,
-0.031248753890395164,
0.09244438260793686,
0.08172664046287537,
-0.36544036865234375,
0.03480618819594383,
0.13793887197971344,
0.10191775113344193,
0.002237213309854269,
-0.04041369631886482,
0.11347141116857529,
0.07477203756570816,
-0.030063968151807785,
0.04138572886586189,
-0.06121840327978134,
-0.04561866819858551,
0.03884280100464821,
-0.05960811302065849,
-0.033487334847450256,
0.19895333051681519,
-0.04296356067061424,
0.0500568225979805,
-0.10024197399616241,
-0.08209642022848129,
0.0004346643399912864,
-0.015289991162717342,
0.03812247887253761,
-0.010191606357693672,
0.09963010251522064,
0.05113557353615761,
-0.035959914326667786,
-0.13030685484409332,
-0.005216804798692465,
-0.2004159688949585,
0.1595696061849594,
0.01788514293730259,
0.03330257534980774,
-0.15076909959316254,
0.0552436038851738,
0.016257518902420998,
-0.10182248055934906,
0.004213124513626099,
-0.07500448077917099,
0.0385168120265007,
-0.015128588303923607,
-0.07416370511054993,
-0.10212622582912445,
0.12873776257038116,
0.1756383180618286,
0.004916585981845856,
0.00740387849509716,
-0.08417452126741409,
0.10456426441669464,
0.011677070520818233,
-0.0003534606075845659,
0.031762804836034775,
-0.04621201008558273,
0.05624647065997124,
-0.10427135974168777,
0.07528972625732422,
-0.039177507162094116,
-0.16654224693775177,
0.010172839276492596,
-0.030613305047154427,
0.12160888314247131,
-0.0015099291922524571,
0.09400424361228943,
-0.04813813418149948,
0.04689636081457138,
0.0641343891620636,
-0.08298645168542862,
0.007679025176912546,
-0.024365952238440514,
0.06925852596759796,
0.0204140767455101,
0.03165958821773529,
0.04615887999534607,
-0.045172665268182755,
0.08644517511129379,
-0.06139305606484413,
-0.04282331094145775,
-0.05274041369557381,
-0.06222843751311302,
0.06560183316469193,
-0.07982368767261505,
0.028125377371907234,
-0.15760309994220734,
-0.2252405285835266,
0.02596941404044628,
0.014920985326170921,
-0.0297084990888834,
-0.01671389676630497,
-0.04438551515340805,
-0.010506823658943176,
0.021031856536865234,
-0.08185165375471115,
-0.049936797469854355,
-0.08322537690401077,
0.05922577157616615,
-0.03619011491537094,
0.034257419407367706,
-0.18331637978553772,
0.0446469821035862,
-0.10228318721055984,
0.027153674513101578,
-0.03186802938580513,
0.04899703338742256,
-0.05404381453990936,
0.1648406684398651,
-0.002505479147657752,
0.04739994555711746,
-0.05548204854130745,
0.08155246824026108,
-0.03349513188004494,
0.1653900295495987,
-0.1219356507062912,
-0.07457999140024185,
0.19437874853610992,
-0.12708114087581635,
-0.2030283808708191,
0.08557765185832977,
-0.011339457705616951,
0.06881066411733627,
0.10909657925367355,
0.20828531682491302,
0.030548078939318657,
-0.050040215253829956,
0.02852419950067997,
0.10709072649478912,
-0.0941527783870697,
-0.1440354734659195,
0.01142947282642126,
-0.016468090936541557,
-0.12143567949533463,
0.047321584075689316,
0.12324801832437515,
0.07120479643344879,
-0.020889470353722572,
-0.05682270973920822,
-0.06243617832660675,
-0.025895575061440468,
-0.001098462613299489,
-0.034107767045497894,
0.06967442482709885,
-0.10257795453071594,
-0.004489772487431765,
0.04407509043812752,
-0.007345555350184441,
-0.008566093631088734,
0.02346433699131012,
-0.07231094688177109,
0.0728871151804924,
-0.02028217539191246,
0.06079154834151268,
-0.11411996930837631,
-0.13449396193027496,
-0.014887942932546139,
0.06686311215162277,
0.007246146909892559,
0.05690915137529373,
0.02400892786681652,
-0.0044649275951087475,
-0.037080373615026474,
0.017351960763335228,
0.22164466977119446,
0.03445685654878616,
-0.05894055962562561,
-0.10250738263130188,
0.07089865207672119,
-0.0733136534690857,
-0.0038810442201793194,
-0.08981648832559586,
0.038887012749910355,
0.04336676374077797,
0.12420043349266052,
-0.007224592845886946,
0.08407974988222122,
-0.022937314584851265,
0.011195441707968712,
-0.11440400779247284,
0.01636681705713272,
0.09614640474319458,
0.007866994477808475,
-0.1337534785270691,
0.23987363278865814,
-0.19760878384113312,
0.27440345287323,
0.20502178370952606,
-0.18877246975898743,
0.03501642867922783,
-0.1149175763130188,
-0.002008592477068305,
0.004755210131406784,
0.0038717782590538263,
-0.04943240433931351,
-0.024568894878029823,
0.0021996109280735254,
0.19166377186775208,
-0.07351944595575333,
-0.044588103890419006,
-0.01404640730470419,
-0.04492517188191414,
-0.04598654806613922,
0.060725100338459015,
0.11691119521856308,
-0.17222446203231812,
0.19953681528568268,
0.26740315556526184,
0.05649516358971596,
0.15636032819747925,
-0.044559624046087265,
0.015058614313602448,
0.03826140612363815,
0.011335236951708794,
0.005399018060415983,
-0.06292850524187088,
-0.04615112766623497,
0.01182850543409586,
0.07680606096982956,
0.036972526460886,
0.07601387798786163,
-0.12983182072639465,
-0.0653335228562355,
0.00989541970193386,
-0.027434425428509712,
0.02707061544060707,
0.09791859239339828,
0.01445250865072012,
0.11266516149044037,
-0.07233676314353943,
-0.0496673583984375,
0.11858335882425308,
-0.010044911876320839,
-0.08638770133256912,
0.1798054426908493,
-0.1244828999042511,
-0.261180579662323,
-0.23111233115196228,
-0.18254457414150238,
-0.06503164023160934,
0.05904011428356171,
0.13374871015548706,
-0.02715570479631424,
-0.06322029232978821,
-0.08102331310510635,
0.03035692125558853,
-0.011322916485369205,
0.006151848006993532,
-0.05763664096593857,
0.08833452314138412,
-0.058100100606679916,
-0.11798392236232758,
-0.04540642723441124,
0.03651372343301773,
-0.028390254825353622,
0.11192913353443146,
-0.08461987972259521,
0.10643972456455231,
0.1271544247865677,
0.007772998884320259,
0.0016197592485696077,
-0.07520508021116257,
0.10091717541217804,
-0.054735612124204636,
-0.015814434736967087,
0.18510334193706512,
-0.06857012957334518,
0.06310728937387466,
0.18042245507240295,
0.015414898283779621,
-0.11694661527872086,
0.05128026008605957,
-0.026471421122550964,
-0.09650994092226028,
-0.24994303286075592,
-0.11167306452989578,
-0.10101259499788284,
0.08947387337684631,
0.0147270318120718,
0.0813816487789154,
0.1235029548406601,
0.0654795840382576,
-0.03870146721601486,
0.0027476151008158922,
0.06207288056612015,
0.09297245740890503,
0.3051879107952118,
-0.027588456869125366,
0.1472405642271042,
-0.1186753362417221,
-0.10952050983905792,
0.08257637172937393,
0.0845542848110199,
0.12338761985301971,
0.13847261667251587,
0.09578624367713928,
0.061665140092372894,
0.017447220161557198,
0.12630344927310944,
0.11974790692329407,
0.03858545050024986,
-0.02270602248609066,
-0.032195474952459335,
-0.03544968366622925,
0.013855401426553726,
0.05927611142396927,
-0.09239695966243744,
-0.11505406349897385,
-0.03534417599439621,
-0.07966310530900955,
0.07341013103723526,
0.11567646265029907,
0.03712856024503708,
-0.22412912547588348,
0.06257343292236328,
0.12581832706928253,
-0.03487724810838699,
-0.07624176144599915,
0.09777287393808365,
0.0010059683118015528,
-0.04102803021669388,
0.1501779407262802,
-0.03483979031443596,
0.12430880963802338,
0.00597972609102726,
0.06967435777187347,
-0.061159081757068634,
-0.09028560668230057,
0.006303017493337393,
0.11862321943044662,
-0.2834312319755554,
0.21129709482192993,
0.009835449047386646,
-0.019402286037802696,
-0.05151335522532463,
0.011698269285261631,
0.01093364879488945,
0.19414639472961426,
0.11842621117830276,
-0.05193806439638138,
-0.1251959204673767,
-0.04806782305240631,
-0.02695751190185547,
0.024282628670334816,
0.09942324459552765,
-0.013378067873418331,
0.02465890161693096,
-0.07476774603128433,
-0.004215550608932972,
0.007410963531583548,
-0.002636012388393283,
-0.03060084767639637,
-0.20665347576141357,
0.03083699382841587,
0.11282823979854584,
0.09943105280399323,
-0.026618046686053276,
0.010056777857244015,
-0.18093986809253693,
0.17984487116336823,
-0.10577816516160965,
-0.0646379142999649,
-0.12176729738712311,
-0.1374046951532364,
0.032386742532253265,
-0.011655131354928017,
0.05578620731830597,
-0.05033711716532707,
0.05199822410941124,
-0.11662887781858444,
-0.20527662336826324,
0.10651233047246933,
-0.10026217252016068,
-0.041269779205322266,
-0.049067843705415726,
0.14421574771404266,
-0.14514556527137756,
-0.01677894778549671,
0.0492711178958416,
0.03700755909085274,
-0.06789186596870422,
-0.09480389207601547,
-0.024957871064543724,
0.031170394271612167,
0.044620662927627563,
0.011302762664854527,
-0.14360836148262024,
-0.09083382785320282,
0.022335760295391083,
-0.08849450945854187,
0.2429637312889099,
0.29421472549438477,
-0.03136972337961197,
0.17118889093399048,
0.1624460071325302,
-0.1346687525510788,
-0.3319091200828552,
-0.13701406121253967,
-0.19598841667175293,
-0.03799448534846306,
0.04184459149837494,
-0.08045858144760132,
0.07849813997745514,
0.07453271746635437,
-0.04596109688282013,
0.0932869017124176,
-0.2135051041841507,
-0.12018171697854996,
0.13025601208209991,
-0.0037368505727499723,
0.3274591863155365,
-0.18035171926021576,
-0.10786277800798416,
-0.147150918841362,
-0.10188359022140503,
0.14011110365390778,
-0.08746092021465302,
0.0968790277838707,
0.010119509883224964,
0.01323111541569233,
0.013967253267765045,
-0.034884002059698105,
0.10382312536239624,
-0.04486394301056862,
0.0827137902379036,
-0.13442641496658325,
0.0311247818171978,
0.0982368141412735,
-0.023089177906513214,
0.06446277350187302,
-0.19565406441688538,
0.0007573386537842453,
-0.04413704201579094,
-0.06428468227386475,
0.003923377953469753,
0.08197945356369019,
0.0119581064209342,
-0.050301458686590195,
-0.03600365296006203,
-0.05647021532058716,
0.004883945919573307,
-0.02668977528810501,
0.297203928232193,
-0.05551900342106819,
0.14349620044231415,
0.14781424403190613,
0.15648621320724487,
-0.13775493204593658,
0.09655493497848511,
-0.04349103942513466,
-0.1007365956902504,
0.05714675411581993,
-0.12492338567972183,
0.04988466575741768,
0.094661645591259,
-0.07334034144878387,
0.07304112613201141,
0.0947517603635788,
0.037457190454006195,
-0.004379583057016134,
0.1802726686000824,
-0.18828874826431274,
-0.09779717028141022,
-0.02748100645840168,
0.04923076182603836,
0.06706935912370682,
0.07469617575407028,
0.16421300172805786,
-0.001201718463562429,
0.015682868659496307,
0.015707042068243027,
0.020014014095067978,
-0.04803552106022835,
0.040308788418769836,
-0.016828909516334534,
0.010336014442145824,
-0.12996217608451843,
0.11176765710115433,
0.022542916238307953,
-0.13203395903110504,
0.0037110543344169855,
0.1117657721042633,
-0.1239880844950676,
-0.12172894924879074,
-0.052547577768564224,
0.17946547269821167,
-0.15897707641124725,
-0.07758819311857224,
-0.06689583510160446,
-0.1884409338235855,
0.06226795166730881,
0.22125835716724396,
0.062441423535346985,
0.10156690329313278,
0.012164506129920483,
-0.048979662358760834,
-0.030817247927188873,
0.04373624548316002,
-0.048446204513311386,
0.06812692433595657,
-0.11580722779035568,
-0.003625471144914627,
-0.051269255578517914,
0.05650920793414116,
-0.08759677410125732,
0.0030319783836603165,
-0.15201345086097717,
0.0209173783659935,
-0.17497412860393524,
-0.0205844696611166,
-0.09589561074972153,
-0.03451237455010414,
0.011887170374393463,
-0.003652766579762101,
-0.04734066128730774,
-0.05812899395823479,
-0.09060658514499664,
0.005249802488833666,
-0.030168622732162476,
0.07605243474245071,
-0.10157889872789383,
-0.03683146834373474,
0.05596957355737686,
-0.025459321215748787,
0.09257930517196655,
0.007215168327093124,
-0.07844144105911255,
0.07323747128248215,
-0.20250831544399261,
-0.07462114840745926,
0.1349623203277588,
0.012877341359853745,
0.01433848962187767,
0.07615745812654495,
-0.00026961302501149476,
0.12541797757148743,
0.0000033646317660895875,
0.053528737276792526,
-0.07837887108325958,
-0.12340746074914932,
0.010294788517057896,
-0.029573259875178337,
-0.1269339770078659,
-0.003393312683328986,
-0.09789466857910156,
0.08011048287153244,
-0.01170063205063343,
0.20980507135391235,
-0.0613788440823555,
0.03947426751255989,
-0.0482708103954792,
0.032428305596113205,
0.004700163844972849,
-0.20592190325260162,
-0.10726362466812134,
-0.10623234510421753,
-0.026636963710188866,
0.006271996535360813,
0.31988534331321716,
0.05655127763748169,
-0.09092064201831818,
0.08231095969676971,
0.0359703004360199,
0.03692442551255226,
0.047804687172174454,
0.32101836800575256,
0.10449174046516418,
-0.015344074927270412,
-0.1473461240530014,
0.02665884979069233,
0.03326180949807167,
-0.09654217213392258,
0.051929157227277756,
0.10810428857803345,
-0.08699777722358704,
0.10889717191457748,
0.07640653103590012,
-0.0012231802102178335,
-0.03349323943257332,
-0.11107002198696136,
-0.02129247784614563,
0.06072830781340599,
-0.005512511823326349,
0.07054830342531204,
0.17545370757579803,
-0.051658809185028076,
0.038718078285455704,
-0.026071172207593918,
-0.02161732129752636,
-0.1754336655139923,
-0.1299535632133484,
-0.10211914032697678,
-0.09993096441030502,
0.025350218638777733,
-0.06776802241802216,
0.046792175620794296,
0.0701855942606926,
0.06920913606882095,
-0.03775874525308609,
0.07067469507455826,
-0.04259522631764412,
-0.03827308118343353,
0.019353482872247696,
-0.04310055449604988,
0.03972002491354942,
-0.043406035751104355,
-0.06280182301998138,
-0.055819831788539886,
-0.04031532257795334,
-0.07156457751989365,
0.06654798984527588,
0.018223078921437263,
0.059505149722099304,
-0.15881334245204926,
-0.06023716554045677,
-0.029088158160448074,
0.06961258500814438,
-0.0010639306856319308,
0.0848631039261818,
0.017299452796578407,
-0.03592558950185776,
0.08237481862306595,
0.1714685559272766,
-0.05526386946439743,
-0.13005387783050537,
-0.01416892558336258,
0.18276989459991455,
0.04249906912446022,
0.13809406757354736,
-0.04787266626954079,
-0.004862284287810326,
-0.02768775075674057,
0.30143216252326965,
0.3054484724998474,
-0.0471254326403141,
0.015752281993627548,
-0.038450077176094055,
0.03920399025082588,
0.08914078027009964,
0.1105283722281456,
0.05389128997921944,
0.22101512551307678,
-0.04384841024875641,
0.005198888946324587,
-0.009552670642733574,
0.0009106346988119185,
-0.139034241437912,
0.11225663125514984,
-0.01098623313009739,
-0.062459368258714676,
-0.0041158562526106834,
0.109149269759655,
-0.15297242999076843,
0.13687030971050262,
-0.04583826661109924,
-0.10477285832166672,
-0.010322065092623234,
-0.004993076901882887,
0.18495936691761017,
-0.038614239543676376,
0.01991380751132965,
-0.03426147624850273,
-0.0722288116812706,
0.02578422613441944,
-0.014197634533047676,
-0.19993196427822113,
0.00019785812764894217,
0.016549699008464813,
-0.0024961435701698065,
0.031246429309248924,
-0.006557455286383629,
0.017212240025401115,
0.07141590118408203,
0.019064750522375107,
-0.051590509712696075,
0.12564879655838013,
-0.0017803513910621405,
-0.0537991039454937,
0.07014787942171097,
-0.06772878021001816,
-0.010519523173570633,
0.005876517854630947,
0.036515455693006516,
-0.09802836924791336,
0.06873522698879242,
-0.034491166472435,
-0.11671977490186691,
-0.019100969657301903,
0.02708382159471512,
-0.0529734268784523,
0.07756980508565903,
0.047662537544965744,
-0.010252631269395351,
0.013023204170167446,
-0.03019876405596733,
0.021127741783857346,
0.011307237669825554,
-0.12513461709022522,
-0.0353403277695179,
-0.10779783129692078,
-0.07460182160139084,
0.16136755049228668,
0.024244174361228943,
-0.273197203874588,
-0.00804264284670353,
-0.12488976866006851,
0.04526819661259651,
-0.19623233377933502,
0.06539338082075119,
0.23118683695793152,
0.007341077551245689,
-0.024383677169680595,
-0.21838563680648804,
0.0626082643866539,
0.04227351397275925,
-0.048489008098840714,
-0.09667101502418518
] |
null | null | transformers | # Sentiment Analysis Model using DistilBERT
This repository hosts a sentiment analysis model fine-tuned on the IMDb movie reviews dataset using DistilBERT architecture. It's designed to classify text inputs into positive or negative sentiment categories.
## Model Description
The model is based on the DistilBERT architecture, a smaller, faster, cheaper, and lighter version of BERT. It has been fine-tuned on the IMDb dataset, which consists of 50,000 movie reviews labeled as positive or negative.
DistilBERT has been proven to retain most of the performance of BERT while being more efficient. This makes it an excellent choice for sentiment analysis tasks where the model's size and speed are essential.
## How to Use
To use the model, you will need to install the `transformers` library from Hugging Face. You can install it using pip:
pip install transformers
Once installed, you can use the following code to classify text using this model:
from transformers import DistilBertTokenizer, DistilBertForSequenceClassification
import torch
# Load the tokenizer and model from the Hugging Face Hub
tokenizer = DistilBertTokenizer.from_pretrained(Pranav-10/Sentimental_Analysis)
model = DistilBertForSequenceClassification.from_pretrained(Pranav-10/Sentimental_Analysis)
# Example text
text = "I loved this movie. The performances were fantastic!"
# Tokenize text and convert to tensor
inputs = tokenizer(text, return_tensors="pt", truncation=True, padding=True, max_length=512)
# Predict sentiment
with torch.no_grad():
logits = model(**inputs).logits
# Convert logits to probabilities using softmax
probabilities = torch.softmax(logits, dim=-1)
# Output the result
print(probabilities)
Evaluation Results
The model achieved the following performance on the IMDb dataset:
Accuracy: 90%
Precision: 89%
Recall: 91%
F1 Score: 90%
These results indicate the model's high efficiency in classifying sentiments as positive or negative.
Training Procedure
The model was trained using the following procedure:
Pre-processing: The dataset was pre-processed by converting all reviews to lowercase and tokenizing using the DistilBERT tokenizer.
Optimization: We used the Adam optimizer with a learning rate of 2e-5, a batch size of 16, and trained the model for 3 epochs.
Hardware: Training was performed on a single NVIDIA GTX 1650 GPU.
| {"license": "apache-2.0"} | text-classification | Pranav-10/Sentiment_analysis | [
"transformers",
"pytorch",
"distilbert",
"text-classification",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T06:08:21+00:00 | [] | [] | TAGS
#transformers #pytorch #distilbert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| # Sentiment Analysis Model using DistilBERT
This repository hosts a sentiment analysis model fine-tuned on the IMDb movie reviews dataset using DistilBERT architecture. It's designed to classify text inputs into positive or negative sentiment categories.
## Model Description
The model is based on the DistilBERT architecture, a smaller, faster, cheaper, and lighter version of BERT. It has been fine-tuned on the IMDb dataset, which consists of 50,000 movie reviews labeled as positive or negative.
DistilBERT has been proven to retain most of the performance of BERT while being more efficient. This makes it an excellent choice for sentiment analysis tasks where the model's size and speed are essential.
## How to Use
To use the model, you will need to install the 'transformers' library from Hugging Face. You can install it using pip:
pip install transformers
Once installed, you can use the following code to classify text using this model:
from transformers import DistilBertTokenizer, DistilBertForSequenceClassification
import torch
# Load the tokenizer and model from the Hugging Face Hub
tokenizer = DistilBertTokenizer.from_pretrained(Pranav-10/Sentimental_Analysis)
model = DistilBertForSequenceClassification.from_pretrained(Pranav-10/Sentimental_Analysis)
# Example text
text = "I loved this movie. The performances were fantastic!"
# Tokenize text and convert to tensor
inputs = tokenizer(text, return_tensors="pt", truncation=True, padding=True, max_length=512)
# Predict sentiment
with torch.no_grad():
logits = model(inputs).logits
# Convert logits to probabilities using softmax
probabilities = torch.softmax(logits, dim=-1)
# Output the result
print(probabilities)
Evaluation Results
The model achieved the following performance on the IMDb dataset:
Accuracy: 90%
Precision: 89%
Recall: 91%
F1 Score: 90%
These results indicate the model's high efficiency in classifying sentiments as positive or negative.
Training Procedure
The model was trained using the following procedure:
Pre-processing: The dataset was pre-processed by converting all reviews to lowercase and tokenizing using the DistilBERT tokenizer.
Optimization: We used the Adam optimizer with a learning rate of 2e-5, a batch size of 16, and trained the model for 3 epochs.
Hardware: Training was performed on a single NVIDIA GTX 1650 GPU.
| [
"# Sentiment Analysis Model using DistilBERT\n\nThis repository hosts a sentiment analysis model fine-tuned on the IMDb movie reviews dataset using DistilBERT architecture. It's designed to classify text inputs into positive or negative sentiment categories.",
"## Model Description\n\nThe model is based on the DistilBERT architecture, a smaller, faster, cheaper, and lighter version of BERT. It has been fine-tuned on the IMDb dataset, which consists of 50,000 movie reviews labeled as positive or negative.\n\nDistilBERT has been proven to retain most of the performance of BERT while being more efficient. This makes it an excellent choice for sentiment analysis tasks where the model's size and speed are essential.",
"## How to Use\n\nTo use the model, you will need to install the 'transformers' library from Hugging Face. You can install it using pip:\n\npip install transformers\n\nOnce installed, you can use the following code to classify text using this model:\n\nfrom transformers import DistilBertTokenizer, DistilBertForSequenceClassification\nimport torch",
"# Load the tokenizer and model from the Hugging Face Hub\ntokenizer = DistilBertTokenizer.from_pretrained(Pranav-10/Sentimental_Analysis)\nmodel = DistilBertForSequenceClassification.from_pretrained(Pranav-10/Sentimental_Analysis)",
"# Example text\ntext = \"I loved this movie. The performances were fantastic!\"",
"# Tokenize text and convert to tensor\ninputs = tokenizer(text, return_tensors=\"pt\", truncation=True, padding=True, max_length=512)",
"# Predict sentiment\nwith torch.no_grad():\n logits = model(inputs).logits",
"# Convert logits to probabilities using softmax\nprobabilities = torch.softmax(logits, dim=-1)",
"# Output the result\nprint(probabilities)\n\nEvaluation Results\nThe model achieved the following performance on the IMDb dataset:\n\nAccuracy: 90%\nPrecision: 89%\nRecall: 91%\nF1 Score: 90%\nThese results indicate the model's high efficiency in classifying sentiments as positive or negative.\n\nTraining Procedure\nThe model was trained using the following procedure:\n\nPre-processing: The dataset was pre-processed by converting all reviews to lowercase and tokenizing using the DistilBERT tokenizer.\nOptimization: We used the Adam optimizer with a learning rate of 2e-5, a batch size of 16, and trained the model for 3 epochs.\nHardware: Training was performed on a single NVIDIA GTX 1650 GPU."
] | [
"TAGS\n#transformers #pytorch #distilbert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Sentiment Analysis Model using DistilBERT\n\nThis repository hosts a sentiment analysis model fine-tuned on the IMDb movie reviews dataset using DistilBERT architecture. It's designed to classify text inputs into positive or negative sentiment categories.",
"## Model Description\n\nThe model is based on the DistilBERT architecture, a smaller, faster, cheaper, and lighter version of BERT. It has been fine-tuned on the IMDb dataset, which consists of 50,000 movie reviews labeled as positive or negative.\n\nDistilBERT has been proven to retain most of the performance of BERT while being more efficient. This makes it an excellent choice for sentiment analysis tasks where the model's size and speed are essential.",
"## How to Use\n\nTo use the model, you will need to install the 'transformers' library from Hugging Face. You can install it using pip:\n\npip install transformers\n\nOnce installed, you can use the following code to classify text using this model:\n\nfrom transformers import DistilBertTokenizer, DistilBertForSequenceClassification\nimport torch",
"# Load the tokenizer and model from the Hugging Face Hub\ntokenizer = DistilBertTokenizer.from_pretrained(Pranav-10/Sentimental_Analysis)\nmodel = DistilBertForSequenceClassification.from_pretrained(Pranav-10/Sentimental_Analysis)",
"# Example text\ntext = \"I loved this movie. The performances were fantastic!\"",
"# Tokenize text and convert to tensor\ninputs = tokenizer(text, return_tensors=\"pt\", truncation=True, padding=True, max_length=512)",
"# Predict sentiment\nwith torch.no_grad():\n logits = model(inputs).logits",
"# Convert logits to probabilities using softmax\nprobabilities = torch.softmax(logits, dim=-1)",
"# Output the result\nprint(probabilities)\n\nEvaluation Results\nThe model achieved the following performance on the IMDb dataset:\n\nAccuracy: 90%\nPrecision: 89%\nRecall: 91%\nF1 Score: 90%\nThese results indicate the model's high efficiency in classifying sentiments as positive or negative.\n\nTraining Procedure\nThe model was trained using the following procedure:\n\nPre-processing: The dataset was pre-processed by converting all reviews to lowercase and tokenizing using the DistilBERT tokenizer.\nOptimization: We used the Adam optimizer with a learning rate of 2e-5, a batch size of 16, and trained the model for 3 epochs.\nHardware: Training was performed on a single NVIDIA GTX 1650 GPU."
] | [
46,
60,
107,
82,
76,
18,
48,
24,
26,
168
] | [
"passage: TAGS\n#transformers #pytorch #distilbert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Sentiment Analysis Model using DistilBERT\n\nThis repository hosts a sentiment analysis model fine-tuned on the IMDb movie reviews dataset using DistilBERT architecture. It's designed to classify text inputs into positive or negative sentiment categories.## Model Description\n\nThe model is based on the DistilBERT architecture, a smaller, faster, cheaper, and lighter version of BERT. It has been fine-tuned on the IMDb dataset, which consists of 50,000 movie reviews labeled as positive or negative.\n\nDistilBERT has been proven to retain most of the performance of BERT while being more efficient. This makes it an excellent choice for sentiment analysis tasks where the model's size and speed are essential.## How to Use\n\nTo use the model, you will need to install the 'transformers' library from Hugging Face. You can install it using pip:\n\npip install transformers\n\nOnce installed, you can use the following code to classify text using this model:\n\nfrom transformers import DistilBertTokenizer, DistilBertForSequenceClassification\nimport torch# Load the tokenizer and model from the Hugging Face Hub\ntokenizer = DistilBertTokenizer.from_pretrained(Pranav-10/Sentimental_Analysis)\nmodel = DistilBertForSequenceClassification.from_pretrained(Pranav-10/Sentimental_Analysis)# Example text\ntext = \"I loved this movie. The performances were fantastic!\"# Tokenize text and convert to tensor\ninputs = tokenizer(text, return_tensors=\"pt\", truncation=True, padding=True, max_length=512)# Predict sentiment\nwith torch.no_grad():\n logits = model(inputs).logits# Convert logits to probabilities using softmax\nprobabilities = torch.softmax(logits, dim=-1)"
] | [
-0.021382782608270645,
0.17072340846061707,
-0.007003192324191332,
0.02860984206199646,
0.08984874933958054,
0.004888338968157768,
0.07578220218420029,
0.09467896819114685,
-0.030645830556750298,
0.0883553996682167,
-0.035233646631240845,
0.007054194342344999,
0.08929986506700516,
0.07483010739088058,
0.05856091156601906,
-0.24132882058620453,
0.014731836505234241,
-0.06576678901910782,
0.07401512563228607,
0.0554586723446846,
0.1048535704612732,
-0.05496898293495178,
0.10190026462078094,
-0.01309223659336567,
-0.06679302453994751,
0.02598765678703785,
-0.0020934673957526684,
-0.015321738086640835,
0.041255202144384384,
0.06928257644176483,
0.06181488558650017,
-0.050807248800992966,
0.08129794150590897,
-0.1720212697982788,
0.005295892246067524,
0.0693480372428894,
0.01945413649082184,
0.04241275414824486,
0.11990221589803696,
-0.019214138388633728,
0.15812641382217407,
-0.12606550753116608,
0.09855885803699493,
0.054491061717271805,
-0.08428190648555756,
-0.01815200410783291,
-0.09305936098098755,
0.06899705529212952,
0.06455203890800476,
0.07473377138376236,
-0.020760508254170418,
0.07155349105596542,
-0.02744319848716259,
0.04348555952310562,
0.1327025145292282,
-0.10472860932350159,
-0.04498361796140671,
0.0012632961152121425,
0.0364207923412323,
0.032562751322984695,
-0.09905210882425308,
-0.011112218722701073,
-0.008256422355771065,
-0.00006446611223509535,
0.10646653175354004,
-0.05007360503077507,
-0.012220212258398533,
-0.044507671147584915,
-0.12056615948677063,
-0.012419757433235645,
0.09589070081710815,
-0.01135249249637127,
-0.07902248203754425,
-0.1474921703338623,
-0.025429101660847664,
0.007833470590412617,
-0.012641738168895245,
-0.03289134055376053,
0.04117174074053764,
0.009271250106394291,
0.05431102216243744,
-0.13297440111637115,
-0.07690241932868958,
0.0003957458247896284,
-0.03642444312572479,
0.031242288649082184,
0.0013435406144708395,
0.005534581374377012,
0.03720813989639282,
0.09166887402534485,
-0.10326419770717621,
-0.06977548450231552,
0.0017308421665802598,
-0.04810131713747978,
-0.06626887619495392,
-0.04712232947349548,
-0.013081539422273636,
-0.0609624981880188,
-0.045126937329769135,
0.12980327010154724,
0.01707576960325241,
0.04871910437941551,
-0.04673168808221817,
-0.013753250241279602,
0.05282304063439369,
0.20095624029636383,
-0.013665640726685524,
-0.12671956419944763,
0.034915853291749954,
0.0167311392724514,
0.051667653024196625,
0.03023611195385456,
0.0007145348936319351,
-0.00985579751431942,
0.02215370163321495,
0.056026458740234375,
0.05222069472074509,
-0.004391566850244999,
-0.09175796061754227,
-0.02681567333638668,
0.16610674560070038,
-0.12881261110305786,
0.07162272185087204,
0.012806299142539501,
-0.10053863376379013,
0.1283235251903534,
0.05660039186477661,
-0.026597216725349426,
-0.09135818481445312,
0.07489757984876633,
-0.04028402641415596,
0.02199428714811802,
-0.09346061944961548,
-0.12006127834320068,
0.04691202566027641,
-0.09278415888547897,
-0.05932808294892311,
-0.06774991005659103,
-0.1446923166513443,
-0.03628939017653465,
0.0203325804322958,
-0.07324067503213882,
0.024895038455724716,
-0.0217621810734272,
0.019473087042570114,
0.0036908916663378477,
0.014882592484354973,
-0.029341448098421097,
0.016748476773500443,
-0.012645510956645012,
-0.038600750267505646,
0.011915860697627068,
0.039788976311683655,
0.018288083374500275,
-0.08044187724590302,
0.014974486082792282,
-0.2493996024131775,
0.14081403613090515,
-0.056950293481349945,
-0.0017859984654933214,
-0.14135773479938507,
-0.026485880836844444,
0.014169439673423767,
0.0019077400211244822,
-0.004414539318531752,
0.1484367549419403,
-0.17392633855342865,
-0.027947161346673965,
0.12405760586261749,
-0.12202887237071991,
0.0061667300760746,
0.1301450878381729,
-0.0300401970744133,
0.05780838802456856,
0.11358970403671265,
0.060165297240018845,
0.11086248606443405,
-0.1718364804983139,
-0.06021334230899811,
0.001512222341261804,
-0.02377507835626602,
0.09930769354104996,
0.06396797299385071,
-0.034378547221422195,
0.0671297237277031,
-0.0049929311498999596,
-0.08889990299940109,
0.03755968064069748,
0.03637583181262016,
-0.014353376813232899,
0.01589379645884037,
-0.023863311856985092,
-0.001705373520962894,
-0.020908690989017487,
-0.016562215983867645,
0.01122707687318325,
-0.11329939216375351,
0.07754379510879517,
0.07470763474702835,
-0.021840423345565796,
0.03593939170241356,
-0.05984009429812431,
0.02182542346417904,
0.0005233981646597385,
-0.020680777728557587,
-0.12394890189170837,
-0.060835909098386765,
0.0642046257853508,
-0.15962402522563934,
0.08349106460809708,
-0.021201254799962044,
0.03687211498618126,
0.08416765183210373,
-0.04302000254392624,
-0.016601160168647766,
0.018150189891457558,
-0.01734554022550583,
0.0006541816983371973,
-0.1038360521197319,
-0.03175727277994156,
-0.03918108344078064,
0.14490410685539246,
-0.07798708230257034,
0.033687982708215714,
0.04877905920147896,
0.10449479520320892,
0.018020788207650185,
-0.0709746852517128,
0.028118597343564034,
-0.03201768174767494,
0.03513073921203613,
-0.07609327882528305,
-0.007249378599226475,
0.011995653621852398,
-0.036703672260046005,
0.06573260575532913,
-0.15050974488258362,
-0.138962060213089,
0.08527722209692001,
0.04789888113737106,
-0.08700211346149445,
-0.060092709958553314,
-0.004475768189877272,
-0.014540579169988632,
-0.07215743511915207,
-0.08903667330741882,
0.13149671256542206,
0.07693400233983994,
0.07655377686023712,
-0.038877081125974655,
-0.033615972846746445,
-0.009173248894512653,
-0.03846846893429756,
-0.03862990811467171,
0.04226796701550484,
-0.03485168516635895,
-0.10209542512893677,
0.048530664294958115,
0.011545190587639809,
-0.010018998757004738,
0.11423103511333466,
0.010753938928246498,
-0.05886662006378174,
-0.044360093772411346,
0.02038346789777279,
0.03004278428852558,
0.00859840027987957,
-0.007866001687943935,
-0.0036458238027989864,
0.03825272247195244,
-0.027428563684225082,
-0.018166128545999527,
-0.06394502520561218,
0.05356800556182861,
0.06217130646109581,
0.010069926269352436,
0.052791982889175415,
-0.03574421629309654,
-0.0020559674594551325,
0.057159971445798874,
0.034874361008405685,
0.09079092741012573,
-0.03519480675458908,
-0.03235497698187828,
-0.1113007441163063,
0.0896490216255188,
-0.12080998718738556,
-0.16640111804008484,
-0.17019475996494293,
-0.047208964824676514,
-0.029445327818393707,
0.010640677995979786,
-0.006520132999867201,
-0.0428016223013401,
-0.06769482791423798,
-0.09541615098714828,
0.14863400161266327,
0.024138491600751877,
-0.08430126309394836,
-0.09432763606309891,
0.003956228494644165,
0.00487254373729229,
-0.1057201400399208,
-0.01515705231577158,
0.04453819990158081,
-0.10185214132070541,
-0.019205812364816666,
-0.00923735462129116,
0.06655965745449066,
0.029461922124028206,
-0.016926029697060585,
-0.05002111196517944,
-0.020533699542284012,
0.1790987253189087,
-0.06743918359279633,
0.1460144817829132,
0.19712775945663452,
0.003584073157981038,
0.08031906187534332,
0.09057479351758957,
0.03152193874120712,
0.024920711293816566,
0.013984058983623981,
0.04395585507154465,
-0.0026475475169718266,
-0.1769220232963562,
-0.07375998795032501,
-0.008589187636971474,
-0.00885664764791727,
0.11031830310821533,
0.02032686024904251,
0.028294669464230537,
0.07009926438331604,
-0.041801657527685165,
0.015616247430443764,
0.07077260315418243,
0.11637084186077118,
0.11101874709129333,
0.0035059186629951,
0.052082426846027374,
-0.04515398293733597,
-0.015281107276678085,
0.12476164847612381,
-0.061251379549503326,
0.10365977883338928,
-0.042947012931108475,
0.1292552649974823,
-0.01989782229065895,
0.05084605515003204,
0.007244409061968327,
0.09535885602235794,
0.004160402342677116,
0.043091826140880585,
-0.02880161814391613,
-0.07964425534009933,
-0.07405709475278854,
0.09292933344841003,
0.044055480509996414,
0.01382234413176775,
0.011919745244085789,
0.002951009664684534,
0.06549231708049774,
0.1742587685585022,
-0.013968462124466896,
-0.1726752072572708,
-0.03224954009056091,
0.03419213742017746,
-0.005049125291407108,
-0.06775827705860138,
-0.03896809369325638,
0.07486819475889206,
-0.14883124828338623,
0.079224593937397,
-0.016451263800263405,
0.07431178539991379,
-0.09367924928665161,
-0.019735056906938553,
-0.007206097710877657,
0.10964533686637878,
0.01101023517549038,
0.06422284245491028,
-0.12913070619106293,
-0.03838589787483215,
0.0230997484177351,
0.11261753737926483,
-0.04933788627386093,
0.04992547258734703,
0.04947257041931152,
-0.017054541036486626,
0.12843286991119385,
0.025768933817744255,
0.05322306603193283,
-0.0927603468298912,
-0.10834305733442307,
-0.03047812730073929,
0.08106549829244614,
-0.06318601965904236,
0.07634248584508896,
-0.03415548428893089,
-0.04417703300714493,
-0.051487162709236145,
-0.06614962965250015,
-0.07997976243495941,
-0.16682371497154236,
0.026517214253544807,
-0.0614466555416584,
0.06004887446761131,
-0.03844727203249931,
-0.006191386375576258,
-0.011397965252399445,
0.12502135336399078,
-0.1096159964799881,
-0.09188855439424515,
-0.11850155144929886,
-0.02056049555540085,
0.1555430144071579,
-0.06490904092788696,
0.056452427059412,
-0.05291935056447983,
0.14653684198856354,
-0.043098896741867065,
-0.06137529015541077,
0.007843077182769775,
-0.0691780224442482,
-0.10733193159103394,
-0.009329103864729404,
0.11152549833059311,
0.06538480520248413,
0.06366360187530518,
-0.005889382213354111,
0.04710599035024643,
-0.012752617709338665,
-0.11350607872009277,
-0.043561261147260666,
0.14414241909980774,
-0.016194932162761688,
0.062493205070495605,
-0.08961347490549088,
-0.13163311779499054,
-0.10910490900278091,
0.021933889016509056,
0.08412396907806396,
0.14382334053516388,
-0.04108310863375664,
0.03534896671772003,
0.09822618216276169,
-0.09852423518896103,
-0.1736912876367569,
0.006267943419516087,
0.07303934544324875,
-0.01679508201777935,
0.04291871190071106,
-0.18087153136730194,
0.11303894221782684,
0.07541672140359879,
0.005710023920983076,
-0.03576620668172836,
-0.1755046248435974,
-0.09377256780862808,
0.008493014611303806,
0.05239715054631233,
-0.103702113032341,
-0.1069151908159256,
-0.033693596720695496,
-0.025220204144716263,
-0.06915018707513809,
0.1230846419930458,
-0.047326985746622086,
0.008374694734811783,
0.03661639243364334,
0.06098804622888565,
0.05637208744883537,
-0.031917378306388855,
0.10391117632389069,
-0.047146428376436234,
0.008631004020571709,
-0.03125869482755661,
-0.027097728103399277,
0.09363255649805069,
-0.0704406127333641,
0.17215858399868011,
-0.02047133445739746,
-0.001902450923807919,
-0.077157162129879,
-0.015084347687661648,
-0.05349959060549736,
0.0504840686917305,
-0.048639927059412,
-0.04507127404212952,
-0.03650619462132454,
0.091614730656147,
0.09124497324228287,
-0.015709688887000084,
-0.04351576045155525,
-0.048085372895002365,
-0.061886634677648544,
0.14308609068393707,
0.060368992388248444,
0.1612086147069931,
-0.10391546040773392,
0.0008478978998027742,
0.0006849392084404826,
0.012870962731540203,
-0.0468304380774498,
0.037287916988134384,
0.08096612244844437,
-0.0037241182290017605,
0.1528298407793045,
-0.04365985095500946,
-0.13307598233222961,
0.010685721412301064,
0.0368337482213974,
-0.06300269812345505,
-0.10727063566446304,
-0.021820297464728355,
0.1094549298286438,
-0.09615849703550339,
-0.10879512131214142,
0.1196809709072113,
-0.06603571772575378,
-0.0349423810839653,
0.0012119244784116745,
0.06280697882175446,
-0.028356822207570076,
0.09907646477222443,
-0.022864727303385735,
0.01519455760717392,
-0.06567925959825516,
0.1254725158214569,
0.07936135679483414,
-0.1311797797679901,
0.006229057442396879,
0.16229598224163055,
-0.04394427686929703,
-0.04337150603532791,
-0.015816962346434593,
0.08328340947628021,
-0.05998068302869797,
-0.013176274485886097,
0.030504141002893448,
-0.07218644767999649,
-0.007481369189918041,
0.04212160035967827,
0.003389104036614299,
0.031968265771865845,
-0.07930899411439896,
0.013540948741137981,
-0.0935148224234581,
0.06606148928403854,
0.10488885641098022,
0.019855190068483353,
-0.021450290456414223,
0.08144202083349228,
0.01562686264514923,
-0.04381364583969116,
-0.010264486074447632,
-0.024677205830812454,
-0.023191306740045547,
-0.05004041641950607,
-0.0880209431052208,
0.009194337762892246,
-0.08224835991859436,
0.0035838482435792685,
-0.025758620351552963,
0.028338665142655373,
0.0471985749900341,
-0.0035060641821473837,
-0.012165612541139126,
-0.07212206721305847,
-0.06957431137561798,
0.06771718710660934,
-0.1404435783624649,
-0.015557152219116688,
0.06034405529499054,
-0.09041773527860641,
0.04664517566561699,
0.0273312758654356,
0.0004210445622447878,
-0.0021400712430477142,
-0.04491373524069786,
-0.046285588294267654,
0.012004823423922062,
-0.005479780957102776,
0.02305843122303486,
-0.09723154455423355,
0.005551034584641457,
-0.018974153324961662,
-0.04971848055720329,
-0.052240531891584396,
0.07985513657331467,
-0.08441869169473648,
0.07429476827383041,
0.045372240245342255,
-0.033734071999788284,
-0.08215198665857315,
0.07380877435207367,
0.08968407660722733,
0.021953970193862915,
0.09073598682880402,
-0.041029561311006546,
0.07512717694044113,
-0.10520150512456894,
-0.0073184482753276825,
0.0021984928753226995,
0.002623288193717599,
-0.046449292451143265,
-0.056010905653238297,
0.03156527504324913,
-0.004273382481187582,
0.0695638358592987,
0.040833890438079834,
0.013284925371408463,
0.044313203543424606,
0.03984270989894867,
-0.03740118816494942,
-0.011607122607529163,
-0.013913899660110474,
0.015784425660967827,
-0.0289603378623724,
-0.011009515263140202,
-0.02371240220963955,
-0.006820301059633493,
-0.010848899371922016,
0.011365992948412895,
0.14406506717205048,
0.030297409743070602,
0.05120960623025894,
0.017314231023192406,
-0.01214810274541378,
-0.08121609687805176,
0.03240521624684334,
-0.016375645995140076,
0.022819925099611282,
-0.08175408840179443,
0.17763839662075043,
0.03120352327823639,
-0.1473751962184906,
0.1008116751909256,
0.061434902250766754,
-0.05397361144423485,
-0.05243457853794098,
-0.1605759710073471,
-0.03353838622570038,
0.047046273946762085,
-0.026511456817388535,
-0.09758193790912628,
0.07121215760707855,
-0.030319849029183388,
0.03503577783703804,
-0.023463906720280647,
0.11796422302722931,
-0.13494201004505157,
-0.14050406217575073,
0.12199810147285461,
0.04659217223525047,
0.04209220036864281,
0.07874605804681778,
0.033429935574531555,
0.01898081973195076,
0.05555612966418266,
0.09614261239767075,
0.08343202620744705,
0.0508057177066803,
0.04037358611822128,
-0.044416334480047226,
-0.09827936440706253,
0.007338901050388813,
-0.012916265986859798,
-0.04881475120782852,
0.17434582114219666,
0.022802533581852913,
-0.009627943858504295,
-0.046674903482198715,
0.14365500211715698,
-0.07478367537260056,
-0.048967961221933365,
-0.10106410086154938,
0.1501626819372177,
-0.0021745762787759304,
0.02590053901076317,
-0.021281233057379723,
-0.10176382958889008,
-0.0013641843106597662,
0.1288057565689087,
0.12531042098999023,
-0.02066359855234623,
-0.020672913640737534,
0.005108607467263937,
0.00944832805544138,
-0.0072877597995102406,
0.04456730931997299,
-0.008891459554433823,
0.16371241211891174,
-0.08432525396347046,
0.21729695796966553,
-0.023532679304480553,
-0.06834068149328232,
-0.04361793026328087,
0.02495654486119747,
-0.04635019600391388,
0.04460293427109718,
-0.04585717245936394,
0.060299258679151535,
-0.013253618963062763,
-0.22023703157901764,
0.05129723995923996,
-0.05695294216275215,
-0.07124622911214828,
0.018398256972432137,
0.01241657231003046,
0.06861893087625504,
0.08883851021528244,
0.046000879257917404,
-0.0038662634324282408,
0.14973579347133636,
0.0045851487666368484,
-0.09182983636856079,
-0.046789344400167465,
0.08277919143438339,
-0.03725675493478775,
0.14328333735466003,
0.009491322562098503,
0.02846737764775753,
0.06528487056493759,
-0.021980270743370056,
-0.13340680301189423,
0.04074842482805252,
0.013450839556753635,
-0.10418468713760376,
-0.0026898893993347883,
0.16520792245864868,
-0.009032031521201134,
0.10982228815555573,
0.04135537147521973,
-0.034268468618392944,
0.03719909489154816,
-0.024123068898916245,
-0.05230274796485901,
-0.09948792308568954,
0.061317816376686096,
-0.08908766508102417,
0.14000916481018066,
0.1915322244167328,
-0.01031279657036066,
0.02130325883626938,
-0.07901465147733688,
-0.009519534185528755,
0.07277557253837585,
0.09880781918764114,
0.0159382876008749,
-0.06853806972503662,
-0.022971654310822487,
0.035328831523656845,
0.0781158059835434,
-0.1497228592634201,
-0.03338303789496422,
-0.06584715098142624,
0.010030794888734818,
-0.01847262866795063,
0.07508019357919693,
0.004906969144940376,
0.04888201132416725,
0.0007578681106679142,
-0.17375245690345764,
0.024491211399435997,
0.08185750991106033,
-0.06869108974933624,
-0.015343504026532173
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# my_mlm_model_masked
This model is a fine-tuned version of [distilroberta-base](https://huggingface.co/distilroberta-base) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.4053
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:-----:|:---------------:|
| 0.4563 | 1.0 | 7300 | 0.4420 |
| 0.434 | 2.0 | 14600 | 0.4119 |
| 0.4114 | 3.0 | 21900 | 0.4039 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "distilroberta-base", "model-index": [{"name": "my_mlm_model_masked", "results": []}]} | fill-mask | ubaskota/my_mlm_model_masked | [
"transformers",
"tensorboard",
"safetensors",
"roberta",
"fill-mask",
"generated_from_trainer",
"base_model:distilroberta-base",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T06:10:23+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #roberta #fill-mask #generated_from_trainer #base_model-distilroberta-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| my\_mlm\_model\_masked
======================
This model is a fine-tuned version of distilroberta-base on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.4053
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #roberta #fill-mask #generated_from_trainer #base_model-distilroberta-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
68,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #roberta #fill-mask #generated_from_trainer #base_model-distilroberta-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.09391044080257416,
0.07210971415042877,
-0.002460902091115713,
0.1092526763677597,
0.13504913449287415,
0.025148652493953705,
0.16647371649742126,
0.10683155804872513,
-0.07183372229337692,
0.04999943822622299,
0.13757100701332092,
0.12186095863580704,
0.014999744482338428,
0.15346604585647583,
-0.04328136518597603,
-0.22304970026016235,
0.01046893559396267,
0.02785126492381096,
-0.08790969103574753,
0.11045260727405548,
0.09516486525535583,
-0.120383121073246,
0.08639081567525864,
0.004777300637215376,
-0.1837863326072693,
0.016302872449159622,
0.026408497244119644,
-0.05516070872545242,
0.1320265680551529,
0.027652133256196976,
0.14696210622787476,
0.009411966428160667,
0.09887604415416718,
-0.1850578933954239,
0.015357950702309608,
0.06093711405992508,
-0.008661320433020592,
0.07559752464294434,
0.01291822362691164,
0.011912074871361256,
0.08568023890256882,
-0.0890500396490097,
0.0660611018538475,
0.01887476071715355,
-0.12530724704265594,
-0.21448664367198944,
-0.08275061845779419,
0.03540593758225441,
0.0876363217830658,
0.08722762763500214,
-0.0036029291804879904,
0.13891996443271637,
-0.04687375947833061,
0.08974240720272064,
0.2278803139925003,
-0.30501052737236023,
-0.07230321317911148,
0.05471920222043991,
0.05207229405641556,
0.044681448489427567,
-0.09813271462917328,
-0.014952917583286762,
0.06458180397748947,
0.029093798249959946,
0.13921833038330078,
-0.030704034492373466,
-0.056200847029685974,
-0.00209024827927351,
-0.14103712141513824,
-0.023632735013961792,
0.1345221996307373,
0.05100131779909134,
-0.04701850190758705,
-0.04200131073594093,
-0.06764388084411621,
-0.14611166715621948,
-0.043612416833639145,
-0.024097753688693047,
0.0459105484187603,
-0.02941848523914814,
-0.07848469167947769,
-0.017067529261112213,
-0.10436448454856873,
-0.08537398278713226,
-0.0657467320561409,
0.16597330570220947,
0.03600774332880974,
0.011267636902630329,
-0.021024296060204506,
0.0856022983789444,
-0.05947311595082283,
-0.13569015264511108,
0.014708307571709156,
0.03034673072397709,
-0.005404547322541475,
-0.04747100546956062,
-0.05674264580011368,
-0.09507111459970474,
0.027914689853787422,
0.1740703135728836,
-0.041803959757089615,
0.04240727424621582,
0.03707709535956383,
0.05241705849766731,
-0.10733136534690857,
0.16343417763710022,
-0.046961747109889984,
-0.029191764071583748,
0.015294112265110016,
0.06423862278461456,
0.05544856935739517,
-0.003678687382489443,
-0.11870282888412476,
0.022260336205363274,
0.07455353438854218,
0.0058155604638159275,
-0.05602561682462692,
0.06699088215827942,
-0.0531407967209816,
0.012186266481876373,
0.008722030557692051,
-0.08636081218719482,
0.022338250651955605,
-0.01833227463066578,
-0.04158856347203255,
-0.06410232186317444,
0.04700164869427681,
0.018499378114938736,
0.026800718158483505,
0.10584130883216858,
-0.09780771285295486,
0.002944675274193287,
-0.09808608144521713,
-0.12008330971002579,
0.016799265518784523,
-0.09332680702209473,
0.02812318503856659,
-0.10958673059940338,
-0.18124569952487946,
0.0023909220471978188,
0.07500658184289932,
-0.03170463442802429,
-0.024152090772986412,
-0.03349152207374573,
-0.07936985790729523,
0.0178051944822073,
-0.0035760074388235807,
0.09134837985038757,
-0.06141821667551994,
0.09142117202281952,
0.05445791408419609,
0.09152306616306305,
-0.05096321180462837,
0.029124336317181587,
-0.08966375142335892,
0.035747066140174866,
-0.19460035860538483,
-0.004671064671128988,
-0.06568534672260284,
0.06331166625022888,
-0.07733616977930069,
-0.0809699073433876,
-0.023082923144102097,
0.014568626880645752,
0.08205310255289078,
0.08353454619646072,
-0.16567237675189972,
-0.06785964965820312,
0.18978933990001678,
-0.09449142217636108,
-0.13158898055553436,
0.13068340718746185,
-0.05138472467660904,
0.057565245777368546,
0.03750499337911606,
0.14821888506412506,
0.05915800854563713,
-0.1342899352312088,
0.002035672077909112,
-0.006354359444230795,
0.03781767934560776,
-0.047472089529037476,
0.06701566278934479,
0.00103561591822654,
0.01799514889717102,
0.011539410799741745,
-0.03296877071261406,
0.05688944458961487,
-0.08200941979885101,
-0.0918632224202156,
-0.04324020445346832,
-0.10471086949110031,
0.05300513654947281,
0.05171854794025421,
0.05690518766641617,
-0.11252503097057343,
-0.1040216013789177,
0.03893457353115082,
0.07659999281167984,
-0.06352432817220688,
0.017376093193888664,
-0.07828288525342941,
0.08113931864500046,
-0.07395468652248383,
-0.026306450366973877,
-0.14604231715202332,
-0.05661734566092491,
0.004008187912404537,
-0.02119610086083412,
0.016737589612603188,
0.004914779681712389,
0.08614364266395569,
0.08401724696159363,
-0.06683031469583511,
-0.019318483769893646,
-0.04482979327440262,
0.01382638979703188,
-0.11053856462240219,
-0.20694658160209656,
-0.028729278594255447,
-0.040740709751844406,
0.13409844040870667,
-0.19867347180843353,
0.04482299089431763,
-0.04197193309664726,
0.09229137748479843,
0.04033257067203522,
-0.012895887717604637,
-0.04774779826402664,
0.06963186711072922,
-0.033306967467069626,
-0.06541375815868378,
0.04297063872218132,
0.006671769078820944,
-0.08553893119096756,
-0.04653629660606384,
-0.13725832104682922,
0.1792699098587036,
0.12582935392856598,
-0.0784367248415947,
-0.0964721143245697,
0.02331063151359558,
-0.054435957223176956,
-0.032331645488739014,
-0.04546225443482399,
0.0008733670692890882,
0.11233600229024887,
-0.01593843474984169,
0.13185366988182068,
-0.07116479426622391,
-0.028060883283615112,
0.04002069681882858,
-0.059742435812950134,
0.0004643604625016451,
0.0774994045495987,
0.12944887578487396,
-0.08110971003770828,
0.1479026824235916,
0.17401443421840668,
-0.11700215935707092,
0.14247114956378937,
-0.033076733350753784,
-0.06654684245586395,
-0.03495514765381813,
-0.006275258958339691,
0.017938967794179916,
0.1442064493894577,
-0.11768322438001633,
0.00476435199379921,
0.009966139681637287,
-0.00276497146114707,
0.012547029182314873,
-0.218326136469841,
-0.03942378982901573,
0.03909139335155487,
-0.04306526482105255,
-0.0032799651380628347,
-0.0033939233981072903,
-0.017227377742528915,
0.08802754431962967,
0.0035249595530331135,
-0.09107521921396255,
0.041070640087127686,
0.00007189290772657841,
-0.07085593789815903,
0.20465177297592163,
-0.07647113502025604,
-0.1303282082080841,
-0.1450480967760086,
-0.074295274913311,
-0.027401121333241463,
0.027442222461104393,
0.05808468908071518,
-0.07194775342941284,
-0.0537603534758091,
-0.09139331430196762,
0.0014161892468109727,
0.04869815707206726,
0.04249851405620575,
0.03792425990104675,
0.0018056457629427314,
0.09037653356790543,
-0.10204794257879257,
-0.01293166633695364,
-0.0437505878508091,
-0.048441603779792786,
0.03827914968132973,
0.039142102003097534,
0.12675516307353973,
0.13245683908462524,
-0.021410997956991196,
-0.002835462801158428,
-0.02421523816883564,
0.22777295112609863,
-0.07030188292264938,
-0.0151444086804986,
0.13545545935630798,
-0.019097361713647842,
0.04854317009449005,
0.14170192182064056,
0.06524867564439774,
-0.09035319834947586,
0.020635921508073807,
0.04032337665557861,
-0.03946599364280701,
-0.19502517580986023,
-0.029914414510130882,
-0.04757509380578995,
-0.011175327934324741,
0.09041347354650497,
0.0296355988830328,
0.04545525833964348,
0.07071797549724579,
0.03684142604470253,
0.08242475986480713,
-0.03207716345787048,
0.0723632201552391,
0.08211838454008102,
0.042592961341142654,
0.12417937815189362,
-0.04695466533303261,
-0.06755781918764114,
0.02429981715977192,
0.0032044346444308758,
0.22327084839344025,
0.02993752807378769,
0.12372038513422012,
0.07279232144355774,
0.15981337428092957,
-0.006798918824642897,
0.07510023564100266,
-0.001465239212848246,
-0.06148083880543709,
-0.00766206718981266,
-0.052602771669626236,
-0.012438038364052773,
0.04548686742782593,
-0.07851341366767883,
0.06354228407144547,
-0.0908292904496193,
0.03836725279688835,
0.04744928702712059,
0.2481933981180191,
0.05831485986709595,
-0.33341410756111145,
-0.09477634727954865,
0.013963475823402405,
-0.022603923454880714,
-0.020604826509952545,
0.016369717195630074,
0.12034254521131516,
-0.05157216638326645,
0.034961603581905365,
-0.07698385417461395,
0.07041530311107635,
0.005219222046434879,
0.03916832059621811,
0.06501666456460953,
0.11051376909017563,
-0.003318189410492778,
0.05753081291913986,
-0.26868271827697754,
0.3003847002983093,
0.007370826788246632,
0.09008505940437317,
-0.05542685464024544,
0.014555134810507298,
0.0365859679877758,
0.0531553216278553,
0.09390514343976974,
-0.023478539660573006,
-0.07555454969406128,
-0.18572594225406647,
-0.047655511647462845,
0.023413361981511116,
0.09331019967794418,
-0.016472918912768364,
0.106987863779068,
-0.03070863150060177,
-0.008017225190997124,
0.0890066847205162,
0.015481313690543175,
-0.08839745819568634,
-0.09099536389112473,
-0.006292812991887331,
0.03568357974290848,
-0.06037410721182823,
-0.0887649729847908,
-0.10465735197067261,
-0.12431932985782623,
0.15459759533405304,
-0.0365651398897171,
-0.014434400014579296,
-0.10128345340490341,
0.0669996589422226,
0.07021380960941315,
-0.0803971067070961,
0.0748463124036789,
0.007697978988289833,
0.09336740523576736,
0.010571797378361225,
-0.054735783487558365,
0.12059769034385681,
-0.08878922462463379,
-0.15557235479354858,
-0.08348949998617172,
0.08681507408618927,
0.018311379477381706,
0.052554067224264145,
-0.0030231932178139687,
0.01936458609998226,
-0.016204535961151123,
-0.06897377222776413,
0.0426059365272522,
-0.022120531648397446,
0.05950088053941727,
0.014001531526446342,
-0.055430371314287186,
-0.010259700007736683,
-0.040182583034038544,
-0.029581746086478233,
0.14510709047317505,
0.28436678647994995,
-0.09511707723140717,
-0.0019535424653440714,
0.047638148069381714,
-0.055345889180898666,
-0.2118454873561859,
0.03244399651885033,
0.038007885217666626,
0.019084248691797256,
0.05776407942175865,
-0.1356431096792221,
0.11155907809734344,
0.09296473860740662,
-0.027196573093533516,
0.11187954246997833,
-0.2831602990627289,
-0.13506247103214264,
0.13817085325717926,
0.15424653887748718,
0.11699520796537399,
-0.14544732868671417,
-0.024155981838703156,
-0.03486773371696472,
-0.12960971891880035,
0.07449252158403397,
-0.11541862785816193,
0.11907552182674408,
-0.014934234321117401,
0.05860438942909241,
-0.0020080977119505405,
-0.07502932101488113,
0.12862025201320648,
-0.024513496086001396,
0.10806500166654587,
-0.06428815424442291,
-0.01205453835427761,
0.06705441325902939,
-0.04785819724202156,
0.020799433812499046,
-0.09908787906169891,
0.02762753516435623,
-0.013773159123957157,
-0.026726551353931427,
-0.04965127632021904,
0.05102228373289108,
-0.033046796917915344,
-0.06658054143190384,
-0.03927503898739815,
0.032826703041791916,
0.028726547956466675,
-0.020820794627070427,
0.13367734849452972,
0.02408546395599842,
0.1633785218000412,
0.1174914762377739,
0.050357524305582047,
-0.05656006932258606,
-0.048144735395908356,
0.0027726886328309774,
-0.037725090980529785,
0.06554438173770905,
-0.14031800627708435,
0.02655143477022648,
0.10853265225887299,
0.015396272763609886,
0.1469321846961975,
0.07502107322216034,
-0.04067452251911163,
0.021461879834532738,
0.08806207031011581,
-0.15426935255527496,
-0.09674573689699173,
0.00839068554341793,
-0.04995928332209587,
-0.12787607312202454,
0.05626966431736946,
0.11130138486623764,
-0.07619272172451019,
0.011684965342283249,
-0.010255718603730202,
0.015568608418107033,
-0.06317470967769623,
0.18579383194446564,
0.06389757245779037,
0.05058746412396431,
-0.0658688172698021,
0.06208473816514015,
0.025970162823796272,
-0.06611255556344986,
0.01458090078085661,
0.03462451323866844,
-0.06884782016277313,
-0.03768137842416763,
0.06444168835878372,
0.1822100132703781,
-0.030445603653788567,
-0.04914141073822975,
-0.15974777936935425,
-0.10847087949514389,
0.04687000811100006,
0.19998444616794586,
0.0911361500620842,
0.0025584325194358826,
-0.02809755504131317,
0.029881777241826057,
-0.11482888460159302,
0.10500433295965195,
0.04451869800686836,
0.0843014195561409,
-0.1469327211380005,
0.12376909703016281,
-0.003077701199799776,
0.0205039381980896,
-0.028043095022439957,
0.05534839630126953,
-0.11551385372877121,
-0.007434104569256306,
-0.11820576339960098,
-0.022199852392077446,
-0.036464132368564606,
-0.008489850908517838,
0.01006210595369339,
-0.055351030081510544,
-0.07274770736694336,
0.021441346034407616,
-0.09815838187932968,
-0.029012568295001984,
0.04885825514793396,
0.04223434254527092,
-0.1177847683429718,
-0.0396590381860733,
0.036391906440258026,
-0.06448213011026382,
0.05313826724886894,
0.02701948769390583,
0.031473901122808456,
0.04789426922798157,
-0.1759011447429657,
0.025713054463267326,
0.0693531334400177,
0.00775432912632823,
0.0487225167453289,
-0.0954616442322731,
-0.01962558925151825,
-0.018427323549985886,
0.04776714742183685,
0.009913351386785507,
0.0725475326180458,
-0.12797336280345917,
0.013030077330768108,
-0.03365355357527733,
-0.06941989809274673,
-0.05282306298613548,
0.00862549152225256,
0.0950465053319931,
-0.005460228770971298,
0.1966373324394226,
-0.10372622311115265,
0.026580344885587692,
-0.19956909120082855,
0.012164541520178318,
-0.00457417219877243,
-0.09236227720975876,
-0.10170310735702515,
-0.048237141221761703,
0.045967139303684235,
-0.06305233389139175,
0.14049048721790314,
0.0018788811285048723,
0.015464424155652523,
0.03165930509567261,
-0.04328536242246628,
0.0421474315226078,
0.020913707092404366,
0.2105739861726761,
0.023603450506925583,
-0.04272926226258278,
0.014731536619365215,
0.04335549473762512,
0.11094480007886887,
0.06459992378950119,
0.17610369622707367,
0.15890610218048096,
-0.017489781603217125,
0.11736936122179031,
0.055162668228149414,
-0.05671205744147301,
-0.13245882093906403,
0.05204485356807709,
-0.04241988807916641,
0.08946487307548523,
-0.013571560382843018,
0.17362627387046814,
0.11412743479013443,
-0.15563127398490906,
0.01818721555173397,
-0.04510512575507164,
-0.07904135435819626,
-0.11118452250957489,
-0.05864588916301727,
-0.09198707342147827,
-0.13509288430213928,
0.014783678576350212,
-0.10030698776245117,
0.0049511282704770565,
0.09747651219367981,
0.00012511806562542915,
-0.010021631605923176,
0.20416224002838135,
-0.0030678659677505493,
0.05048847571015358,
0.03479526937007904,
0.010728304274380207,
-0.035911623388528824,
-0.08194877207279205,
-0.07799416780471802,
-0.008393420837819576,
-0.019145887345075607,
0.01967908814549446,
-0.06040225923061371,
-0.03587152063846588,
0.053615305572748184,
0.002907224465161562,
-0.10418683290481567,
0.005686526652425528,
0.022951599210500717,
0.04518869146704674,
0.029130907729268074,
0.004915625788271427,
0.021661188453435898,
-0.006029201205819845,
0.18543660640716553,
-0.08067134022712708,
-0.08209370821714401,
-0.11549907922744751,
0.19835104048252106,
0.02322378195822239,
0.009530652314424515,
0.0005334829329513013,
-0.08250905573368073,
0.019479254260659218,
0.22931194305419922,
0.19516728818416595,
-0.06240387260913849,
0.010216101072728634,
0.008317595347762108,
-0.01397026889026165,
-0.06006450206041336,
0.08694099634885788,
0.11578245460987091,
0.04520561546087265,
-0.07569553703069687,
-0.06351907551288605,
-0.04562218859791756,
-0.008973068557679653,
-0.04989201948046684,
0.023627497255802155,
0.038474440574645996,
0.014661493711173534,
-0.0389302596449852,
0.04952997714281082,
-0.03178960457444191,
-0.10879628360271454,
0.08651960641145706,
-0.19621984660625458,
-0.14715228974819183,
-0.004385219421237707,
0.11704593896865845,
-0.010118689388036728,
0.06054643169045448,
-0.03863682225346565,
-0.005720995832234621,
0.0770234763622284,
-0.0165326576679945,
-0.06668974459171295,
-0.08208852261304855,
0.060274966061115265,
-0.07864093035459518,
0.2464311122894287,
-0.04252377524971962,
0.05476130172610283,
0.1352820247411728,
0.04775172099471092,
-0.07097026705741882,
0.08577550202608109,
0.048185013234615326,
-0.08070290833711624,
0.022390449419617653,
0.09549110382795334,
-0.0466093011200428,
0.09818175435066223,
0.04904552549123764,
-0.13546496629714966,
0.021495210006833076,
-0.06286372244358063,
-0.07101282477378845,
-0.045581597834825516,
-0.020907985046505928,
-0.07012271881103516,
0.13564155995845795,
0.19010385870933533,
-0.028680307790637016,
0.0024590115062892437,
-0.046951182186603546,
0.039637766778469086,
0.08367584645748138,
0.037406668066978455,
-0.05355844274163246,
-0.23231814801692963,
0.03916998952627182,
0.0400601327419281,
-0.019618745893239975,
-0.2757696509361267,
-0.10430546849966049,
-0.005967473611235619,
-0.06156112253665924,
-0.08559972047805786,
0.06724842637777328,
0.13563649356365204,
0.06314483284950256,
-0.05873878300189972,
-0.09517858922481537,
-0.07246971875429153,
0.14886337518692017,
-0.14838123321533203,
-0.10421202331781387
] |
null | null | diffusers |
# SDXL LoRA DreamBooth - yaneq/jan_zdRM8UdoamtJ6kdZKNKS_SDXL_LoRA_700_9d94_700_1e4
<Gallery />
## Model description
These are yaneq/jan_zdRM8UdoamtJ6kdZKNKS_SDXL_LoRA_700_9d94_700_1e4 LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.
The weights were trained using [DreamBooth](https://dreambooth.github.io/).
LoRA for the text encoder was enabled: False.
Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
## Trigger words
You should use a photo of MDDL man to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
[Download](yaneq/jan_zdRM8UdoamtJ6kdZKNKS_SDXL_LoRA_700_9d94_700_1e4/tree/main) them in the Files & versions tab.
## Training properties
- max_train_steps: 700
- learning_rate: 0.0001
- base_model_name: stabilityai/stable-diffusion-xl-base-1.0
- class_name: man
- training_images_urls: - https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2FWF2NGBPUFgu9eyaCYAwB.jpg?alt=media&token=97c1e215-0a96-4fdf-b292-9ee0e497ba72
- https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2Fcn54hvM4ahi3MzpCQN5D.jpg?alt=media&token=e096f4dc-e7c5-4e14-88fc-a5562d103127
- https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2Fz8D9WdMIx4mXcsDGAZm4.jpg?alt=media&token=fded9422-eb7c-4757-8c1f-cb436a348579
- https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2F6JW19SVZPczh5B2DEqKD.jpg?alt=media&token=0e0dc94f-957d-4b51-8979-0216c0849cf6
- https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2FDAk5k1hGzP9q9y0jpGoO.jpg?alt=media&token=01ed67d1-938a-4f60-bc1a-e1b91412b97e
- https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2F82McawlxnTeA2vBc4bZg.jpg?alt=media&token=f7cfacb2-2186-4005-9211-b7ef762dafad
- https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2FY7nFiafx8co1nK6cnjWJ.jpg?alt=media&token=a1fe8c9a-4d5e-4043-9a82-9304fd430569
- https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2FVYOVRhojKt30NzjWRXL0.jpg?alt=media&token=5a3a2afb-4b83-4488-92e5-6651f5173cc0
- gradient_accumulation_steps: 3
- GPU: T4
- duration: 5284.340887546539
| {"license": "openrail++", "tags": ["stable-diffusion-xl", "stable-diffusion-xl-diffusers", "text-to-image", "diffusers", "lora", "template:sd-lora"], "base_model": "stabilityai/stable-diffusion-xl-base-1.0", "instance_prompt": "a photo of MDDL man"} | text-to-image | yaneq/jan_zdRM8UdoamtJ6kdZKNKS_SDXL_LoRA_700_9d94_700_1e4 | [
"diffusers",
"stable-diffusion-xl",
"stable-diffusion-xl-diffusers",
"text-to-image",
"lora",
"template:sd-lora",
"base_model:stabilityai/stable-diffusion-xl-base-1.0",
"license:openrail++",
"has_space",
"region:us"
] | 2024-02-07T06:10:43+00:00 | [] | [] | TAGS
#diffusers #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us
|
# SDXL LoRA DreamBooth - yaneq/jan_zdRM8UdoamtJ6kdZKNKS_SDXL_LoRA_700_9d94_700_1e4
<Gallery />
## Model description
These are yaneq/jan_zdRM8UdoamtJ6kdZKNKS_SDXL_LoRA_700_9d94_700_1e4 LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.
The weights were trained using DreamBooth.
LoRA for the text encoder was enabled: False.
Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
## Trigger words
You should use a photo of MDDL man to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
Download them in the Files & versions tab.
## Training properties
- max_train_steps: 700
- learning_rate: 0.0001
- base_model_name: stabilityai/stable-diffusion-xl-base-1.0
- class_name: man
- training_images_urls: - URL
- URL
- URL
- URL
- URL
- URL
- URL
- URL
- gradient_accumulation_steps: 3
- GPU: T4
- duration: 5284.340887546539
| [
"# SDXL LoRA DreamBooth - yaneq/jan_zdRM8UdoamtJ6kdZKNKS_SDXL_LoRA_700_9d94_700_1e4\n\n<Gallery />",
"## Model description\n\nThese are yaneq/jan_zdRM8UdoamtJ6kdZKNKS_SDXL_LoRA_700_9d94_700_1e4 LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.\n\nThe weights were trained using DreamBooth.\n\nLoRA for the text encoder was enabled: False.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix.",
"## Trigger words\n\nYou should use a photo of MDDL man to trigger the image generation.",
"## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab.",
"## Training properties\n- max_train_steps: 700\n- learning_rate: 0.0001\n- base_model_name: stabilityai/stable-diffusion-xl-base-1.0\n- class_name: man\n- training_images_urls: - URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- gradient_accumulation_steps: 3\n- GPU: T4\n- duration: 5284.340887546539"
] | [
"TAGS\n#diffusers #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us \n",
"# SDXL LoRA DreamBooth - yaneq/jan_zdRM8UdoamtJ6kdZKNKS_SDXL_LoRA_700_9d94_700_1e4\n\n<Gallery />",
"## Model description\n\nThese are yaneq/jan_zdRM8UdoamtJ6kdZKNKS_SDXL_LoRA_700_9d94_700_1e4 LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.\n\nThe weights were trained using DreamBooth.\n\nLoRA for the text encoder was enabled: False.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix.",
"## Trigger words\n\nYou should use a photo of MDDL man to trigger the image generation.",
"## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab.",
"## Training properties\n- max_train_steps: 700\n- learning_rate: 0.0001\n- base_model_name: stabilityai/stable-diffusion-xl-base-1.0\n- class_name: man\n- training_images_urls: - URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- gradient_accumulation_steps: 3\n- GPU: T4\n- duration: 5284.340887546539"
] | [
82,
51,
116,
19,
28,
101
] | [
"passage: TAGS\n#diffusers #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us \n# SDXL LoRA DreamBooth - yaneq/jan_zdRM8UdoamtJ6kdZKNKS_SDXL_LoRA_700_9d94_700_1e4\n\n<Gallery />## Model description\n\nThese are yaneq/jan_zdRM8UdoamtJ6kdZKNKS_SDXL_LoRA_700_9d94_700_1e4 LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.\n\nThe weights were trained using DreamBooth.\n\nLoRA for the text encoder was enabled: False.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix.## Trigger words\n\nYou should use a photo of MDDL man to trigger the image generation.## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab.## Training properties\n- max_train_steps: 700\n- learning_rate: 0.0001\n- base_model_name: stabilityai/stable-diffusion-xl-base-1.0\n- class_name: man\n- training_images_urls: - URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- gradient_accumulation_steps: 3\n- GPU: T4\n- duration: 5284.340887546539"
] | [
-0.09132502973079681,
0.15559810400009155,
-0.0006037005805410445,
0.07545628398656845,
0.1443294882774353,
0.07180237025022507,
0.14295098185539246,
0.13289864361286163,
0.08236604183912277,
0.12355642765760422,
0.023823589086532593,
0.06814147531986237,
0.09276187419891357,
0.15561127662658691,
-0.003733973251655698,
-0.11661319434642792,
0.02417999878525734,
-0.04632622003555298,
0.011674792505800724,
0.030536679551005363,
0.01744857057929039,
-0.06383923441171646,
0.0668487548828125,
-0.012358568608760834,
-0.03901783376932144,
0.04162485525012016,
-0.009570513851940632,
-0.022884339094161987,
0.027491901069879532,
0.04067675769329071,
0.048000626266002655,
0.03891995549201965,
0.10971355438232422,
-0.24045532941818237,
0.018976543098688126,
0.10265959054231644,
-0.04520983621478081,
0.04601717367768288,
0.01162128709256649,
-0.02238747850060463,
-0.0011555859819054604,
-0.15852470695972443,
0.07440803200006485,
0.05524821951985359,
-0.03191433101892471,
-0.24033963680267334,
-0.12391392886638641,
0.051211435347795486,
0.08189191669225693,
0.1104399785399437,
0.005331376567482948,
0.07912716269493103,
0.05669138953089714,
0.04292202740907669,
0.229820117354393,
-0.1194707453250885,
-0.013558569364249706,
0.14912950992584229,
0.034493423998355865,
0.07675522565841675,
0.0004231760394759476,
0.0020750961266458035,
0.06966768205165863,
-0.01160705927759409,
0.04786286503076553,
-0.035736650228500366,
0.04423920810222626,
-0.07556980848312378,
-0.09587685018777847,
-0.01807512529194355,
0.16219478845596313,
0.06493061035871506,
-0.1043844148516655,
-0.1319831907749176,
-0.013489102944731712,
0.03367949649691582,
-0.0007236131932586432,
0.006497934926301241,
0.007059609983116388,
-0.020468605682253838,
0.006936564110219479,
-0.09644127637147903,
-0.013348315842449665,
-0.05773114040493965,
0.07224690914154053,
0.19044509530067444,
0.03219198063015938,
-0.004716159775853157,
0.022527290508151054,
0.032877933233976364,
-0.061431121081113815,
-0.11863602697849274,
-0.037915296852588654,
-0.02847088687121868,
-0.05489709973335266,
0.016863157972693443,
-0.018757523968815804,
-0.09537413716316223,
0.015128969214856625,
0.09082740545272827,
-0.04596128687262535,
0.04994634911417961,
-0.03560750558972359,
0.030044211074709892,
-0.07750031352043152,
0.11638720333576202,
-0.0498046875,
-0.018743429332971573,
0.03199843317270279,
0.1035970151424408,
0.09885556995868683,
-0.03156008943915367,
-0.1323600560426712,
-0.02731245942413807,
-0.07485301047563553,
0.06734161078929901,
0.0002269126707687974,
-0.009222229942679405,
-0.1063852533698082,
-0.049170345067977905,
0.15673504769802094,
-0.06862098723649979,
0.03146680071949959,
0.006184675730764866,
-0.10378141701221466,
0.027209484949707985,
0.13156351447105408,
-0.0023905052803456783,
-0.050367530435323715,
0.09625259786844254,
-0.03138427808880806,
0.010426382534205914,
-0.04545168578624725,
-0.15880408883094788,
0.03918851539492607,
-0.13848133385181427,
-0.004474548622965813,
-0.12681089341640472,
-0.21528896689414978,
-0.06076015904545784,
0.00488839577883482,
-0.012471217662096024,
0.0436871275305748,
-0.06925840675830841,
-0.09949535131454468,
0.02777104079723358,
0.05249834433197975,
-0.05321161821484566,
-0.011805797927081585,
0.043939199298620224,
-0.06883330643177032,
0.11598540842533112,
0.004685711115598679,
0.017846107482910156,
-0.03995884582400322,
0.011943766847252846,
-0.11501932889223099,
0.10562658309936523,
-0.09749727696180344,
0.03397779539227486,
-0.11675118654966354,
-0.09544618427753448,
0.032034069299697876,
-0.0016660045366734266,
0.06933718174695969,
0.104943186044693,
-0.2050292193889618,
-0.05890677124261856,
0.11340443044900894,
-0.14028075337409973,
-0.07942034304141998,
0.021714452654123306,
-0.032375119626522064,
0.05171975865960121,
0.04982759431004524,
0.11812201887369156,
0.14683401584625244,
-0.17199808359146118,
-0.0623338408768177,
-0.035455893725156784,
0.017840687185525894,
0.08277029544115067,
0.037319671362638474,
-0.029042012989521027,
-0.00944081787019968,
0.00856371596455574,
-0.030315879732370377,
0.025247469544410706,
0.01612202636897564,
-0.054529450833797455,
-0.03421008214354515,
-0.10077446699142456,
0.06583897769451141,
-0.0019494801526889205,
-0.003058034460991621,
-0.012481056153774261,
-0.09077286720275879,
-0.005928349681198597,
0.1385524570941925,
-0.09067416936159134,
0.019584495574235916,
-0.04742266982793808,
-0.037729158997535706,
-0.021161990240216255,
0.003481909167021513,
-0.1011161133646965,
-0.11966998875141144,
0.09036227315664291,
-0.1783192902803421,
-0.011815952137112617,
-0.09420013427734375,
0.05506230145692825,
0.06105814874172211,
-0.05167832598090172,
-0.09012194722890854,
-0.07400750368833542,
-0.009014294482767582,
-0.006580257788300514,
-0.1289030909538269,
-0.051458075642585754,
-0.037662167102098465,
0.0635998547077179,
-0.1885894387960434,
0.05939693748950958,
-0.006634060759097338,
0.09128023684024811,
0.05110102519392967,
-0.05058613419532776,
0.0396876260638237,
0.03611462563276291,
-0.013753116130828857,
-0.10203017294406891,
-0.00478750467300415,
-0.031218180432915688,
-0.16752006113529205,
0.016260987147688866,
-0.1407390981912613,
-0.059329554438591,
0.08902377635240555,
0.10892355442047119,
-0.07070311903953552,
0.009876435622572899,
0.007072319742292166,
-0.04315637797117233,
-0.0568067841231823,
-0.06494928151369095,
0.03712049499154091,
0.004434984643012285,
0.08346084505319595,
-0.0562954805791378,
0.003244325751438737,
0.04302128031849861,
0.011625458486378193,
-0.04903572052717209,
0.09945718199014664,
0.01250565703958273,
-0.018427930772304535,
0.037157606333494186,
0.0034134688321501017,
-0.062113482505083084,
0.1397077441215515,
0.04879404231905937,
-0.04934563487768173,
-0.026919955387711525,
0.016309870406985283,
0.08830759674310684,
0.09496857225894928,
0.08897870779037476,
0.08250366896390915,
0.020134273916482925,
-0.014104708097875118,
-0.033322062343358994,
-0.1317635476589203,
-0.04436799883842468,
0.026938365772366524,
-0.07325450330972672,
0.014055130071938038,
0.04011041671037674,
-0.045488905161619186,
0.049612097442150116,
-0.03602628782391548,
0.10491712391376495,
0.021658478304743767,
-0.05130626633763313,
-0.11623180657625198,
0.1393277943134308,
-0.13779708743095398,
-0.17940522730350494,
-0.1365400105714798,
0.03782461956143379,
-0.051827579736709595,
-0.023068388924002647,
0.061305273324251175,
-0.1389293223619461,
-0.07237044721841812,
-0.09659501165151596,
-0.038228839635849,
0.005793871823698282,
-0.025892911478877068,
0.07656030356884003,
0.0013131970772519708,
0.11706900596618652,
-0.10554002970457077,
0.004116692580282688,
0.016269993036985397,
-0.04508879780769348,
-0.0265969131141901,
0.005548986606299877,
0.10166143625974655,
0.043649837374687195,
0.05122261494398117,
0.026927100494503975,
-0.013636678457260132,
0.18226657807826996,
-0.02092551440000534,
0.10146249085664749,
0.1987817883491516,
0.06221318989992142,
0.110757976770401,
0.1368691772222519,
0.04127691686153412,
-0.04023243114352226,
0.05044054239988327,
0.035628531128168106,
-0.0872519314289093,
-0.22932520508766174,
-0.11592938750982285,
-0.030524590983986855,
-0.06004610285162926,
0.16021224856376648,
0.054774653166532516,
0.10233570635318756,
0.10496571660041809,
-0.08027347177267075,
0.08236509561538696,
0.044002965092659,
0.12368392944335938,
0.028716880828142166,
0.07483156770467758,
0.01130110863596201,
-0.0653267353773117,
0.042975012212991714,
0.07663284242153168,
0.023767735809087753,
0.20517012476921082,
-0.0877203419804573,
0.050583817064762115,
-0.0067031788639724255,
0.03881629556417465,
-0.004730230662971735,
0.015304021537303925,
-0.015909569337964058,
-0.00008639582665637136,
0.019318362697958946,
-0.12871688604354858,
0.07770484685897827,
0.14846859872341156,
-0.0409088134765625,
0.05612112954258919,
-0.05376680940389633,
0.019298162311315536,
0.019550887867808342,
0.04620388522744179,
0.03551504388451576,
-0.24634811282157898,
-0.03414727747440338,
0.04518688842654228,
0.10829611867666245,
0.00148661935236305,
-0.03077988140285015,
0.13298045098781586,
-0.08192610740661621,
0.12544280290603638,
-0.027178069576621056,
0.0802987739443779,
-0.0371781587600708,
-0.05084997043013573,
0.11139475554227829,
0.16774344444274902,
0.007742538582533598,
0.018895544111728668,
-0.1694634109735489,
0.06771949678659439,
0.032927267253398895,
0.12590432167053223,
-0.06716297566890717,
0.03394703194499016,
0.038571640849113464,
-0.006638570688664913,
0.12673965096473694,
0.0310808252543211,
-0.048413246870040894,
-0.19953478872776031,
-0.04482010379433632,
-0.036884650588035583,
0.08630796521902084,
-0.05872824415564537,
0.06746196746826172,
0.04949449002742767,
-0.030698684975504875,
0.004042656626552343,
0.0179549939930439,
-0.18686282634735107,
-0.13956744968891144,
0.028986379504203796,
0.13845007121562958,
0.02461758442223072,
-0.09276233613491058,
-0.0705263763666153,
-0.07436856627464294,
0.08130193501710892,
-0.11817176640033722,
-0.1028144359588623,
-0.09324496984481812,
-0.011202201247215271,
0.10576857626438141,
-0.06337340921163559,
0.0465831533074379,
0.01891389489173889,
0.099847711622715,
-0.10906389355659485,
-0.10569320619106293,
0.01562962494790554,
-0.08601222932338715,
-0.21576157212257385,
-0.0904211550951004,
0.09196390956640244,
0.07313636690378189,
-0.0009582543862052262,
-0.007904842495918274,
0.009169159457087517,
0.04600818455219269,
-0.13149051368236542,
0.060754355043172836,
0.24212880432605743,
0.02718835510313511,
0.15541934967041016,
-0.04180809482932091,
-0.05590497702360153,
-0.06165663152933121,
-0.02125348150730133,
0.0219627283513546,
0.27624252438545227,
-0.12537015974521637,
0.06303102523088455,
0.05978596210479736,
-0.07649824023246765,
-0.13037501275539398,
0.06791598349809647,
0.10324396193027496,
-0.013437560759484768,
0.006351655349135399,
-0.21457253396511078,
0.04761405661702156,
0.06209659203886986,
-0.008198864758014679,
0.07770824432373047,
-0.3640079200267792,
-0.1455264389514923,
-0.0748075619339943,
0.12516425549983978,
0.0384531207382679,
-0.12314227968454361,
-0.05012230947613716,
-0.11895615607500076,
-0.06939525902271271,
0.13167469203472137,
-0.017674846574664116,
0.06484832614660263,
-0.003286034567281604,
0.001524183782748878,
0.026399970054626465,
-0.06294393539428711,
0.1570512056350708,
0.017689762637019157,
0.06041166931390762,
-0.03222765401005745,
0.051453474909067154,
0.09170956164598465,
-0.08628406375646591,
0.08051785081624985,
-0.07460039854049683,
0.06386808305978775,
-0.09191610664129257,
0.002191874198615551,
-0.023715583607554436,
0.0529606007039547,
-0.04775218293070793,
-0.0422009602189064,
-0.07150834798812866,
0.06995031237602234,
0.14059798419475555,
-0.015352715738117695,
-0.03623580187559128,
0.020376909524202347,
0.08231012523174286,
0.1265188604593277,
0.011655041016638279,
0.14904627203941345,
-0.077892005443573,
0.021079355850815773,
-0.0680966004729271,
0.10292328149080276,
-0.14938783645629883,
0.020903317257761955,
0.0752604752779007,
0.047572407871484756,
0.1363997906446457,
0.00605538347736001,
-0.11469728499650955,
0.059800710529088974,
0.019108934327960014,
-0.0846046730875969,
-0.09711239486932755,
-0.023402899503707886,
0.01957031711935997,
-0.04849528521299362,
0.00913102924823761,
0.11206912994384766,
-0.06970420479774475,
0.01347444113343954,
-0.005977202672511339,
0.06460000574588776,
-0.00025353807723149657,
0.12537798285484314,
0.07666265964508057,
0.030964938923716545,
-0.06001075357198715,
0.12555788457393646,
0.0453825406730175,
-0.06515289843082428,
0.09755582362413406,
0.10168168693780899,
-0.09719546139240265,
-0.010541773401200771,
-0.11063980311155319,
0.044042110443115234,
-0.05729443579912186,
-0.04678019881248474,
-0.03569778800010681,
-0.06423553079366684,
-0.001542559708468616,
0.1012079119682312,
0.015171811915934086,
-0.014000963419675827,
0.01819167472422123,
-0.029569216072559357,
-0.11395374685525894,
0.0657079890370369,
0.028480779379606247,
0.06927507370710373,
-0.15524257719516754,
-0.011676397174596786,
0.053013697266578674,
-0.009808060713112354,
0.011027602478861809,
0.007206536363810301,
-0.15453724563121796,
-0.03269327059388161,
-0.08789674937725067,
0.11557628214359283,
-0.0760418251156807,
0.013928570784628391,
-0.04355614259839058,
-0.019193977117538452,
0.004817511886358261,
0.07013440132141113,
-0.022973081097006798,
-0.10649425536394119,
0.016672220081090927,
0.05874556675553322,
-0.10164568573236465,
-0.057035885751247406,
-0.016819030046463013,
-0.07434424757957458,
0.050564415752887726,
0.006452410947531462,
-0.046661898493766785,
-0.0004997883224859834,
-0.14430683851242065,
0.06418853998184204,
0.056266773492097855,
0.02639729343354702,
-0.0049104089848697186,
-0.09765497595071793,
0.0068208640441298485,
-0.050681643187999725,
-0.06059606000781059,
-0.006314197555184364,
0.005001668352633715,
-0.10513844341039658,
0.0029987874440848827,
-0.03629714995622635,
0.03744078427553177,
-0.092313252389431,
0.08631514757871628,
0.14076440036296844,
0.03768313303589821,
0.10690631717443466,
-0.0861685648560524,
0.09304443746805191,
-0.1455908566713333,
-0.042565539479255676,
0.0023449433501809835,
-0.035710860043764114,
-0.015740957111120224,
-0.0036600001621991396,
0.0709514319896698,
-0.013693694025278091,
0.09974731504917145,
-0.07420645654201508,
0.0659617930650711,
-0.011587876826524734,
-0.02212912030518055,
-0.05179724469780922,
-0.0035901523660868406,
0.15886060893535614,
-0.01821075938642025,
0.008829704485833645,
0.012353001162409782,
-0.008611496537923813,
0.0738355815410614,
0.10643840581178665,
0.03332841023802757,
0.09418916702270508,
0.08923308551311493,
0.047521017491817474,
0.03779521584510803,
-0.07057899981737137,
-0.05759673938155174,
0.14326953887939453,
-0.016491061076521873,
0.0825975090265274,
-0.054382290691137314,
0.018760638311505318,
0.06322263181209564,
-0.14679566025733948,
0.000879499944858253,
0.0770966038107872,
-0.06744930893182755,
-0.08735832571983337,
-0.1667889654636383,
-0.07788963615894318,
-0.13271450996398926,
0.012741882354021072,
-0.07817073911428452,
0.046836014837026596,
0.003389874240383506,
-0.032398708164691925,
0.008561419323086739,
0.09175381809473038,
-0.029136577621102333,
-0.01705894060432911,
0.03210067003965378,
-0.019368991255760193,
-0.025068268179893494,
0.08574595302343369,
-0.05332714691758156,
0.12366771697998047,
-0.0034690157044678926,
0.03748009726405144,
0.04483310505747795,
0.04519223794341087,
0.042927131056785583,
-0.0006270386511459947,
-0.0798744261264801,
0.009072591550648212,
0.037093427032232285,
0.010822265408933163,
0.17019294202327728,
0.09530659765005112,
0.011271191760897636,
-0.03699338436126709,
0.14656710624694824,
-0.07225175201892853,
-0.020038370043039322,
-0.09388814121484756,
0.06305298954248428,
0.004456563852727413,
-0.005444562528282404,
0.007982512004673481,
-0.10170222073793411,
-0.033743716776371,
0.07990173250436783,
0.10262145847082138,
0.025889206677675247,
0.02874174155294895,
-0.04145253449678421,
-0.012231791391968727,
-0.05936316028237343,
0.05362273007631302,
0.0515274703502655,
0.07377520203590393,
-0.002943064784631133,
0.04970896244049072,
-0.027281301096081734,
-0.020628999918699265,
-0.03960830345749855,
0.013160575181245804,
-0.04722946137189865,
-0.030212953686714172,
-0.044446058571338654,
0.059017930179834366,
-0.0437772311270237,
-0.14394116401672363,
0.169453427195549,
-0.11386460810899734,
-0.06328179687261581,
-0.01099751889705658,
0.07272066920995712,
0.027145909145474434,
0.003152859630063176,
-0.032732170075178146,
-0.008464362472295761,
0.07713949680328369,
-0.046995941549539566,
-0.04971453920006752,
-0.10945841670036316,
-0.04224197566509247,
-0.1285085678100586,
0.22501172125339508,
-0.006284911651164293,
0.07024088501930237,
0.08312242478132248,
0.03268025070428848,
-0.13707011938095093,
0.07095379382371902,
0.021778428927063942,
-0.1148211658000946,
-0.01993056945502758,
0.18416255712509155,
-0.08059646189212799,
0.1997298300266266,
0.04327107220888138,
-0.09716810286045074,
0.02353893779218197,
0.06464815884828568,
-0.021618502214550972,
-0.08006920665502548,
0.01622522994875908,
-0.10939731448888779,
0.1348738968372345,
0.14378829300403595,
-0.04108945280313492,
-0.024692973122000694,
0.013533556833863258,
0.01793578639626503,
0.013224088586866856,
0.09728231281042099,
0.04658230394124985,
-0.10253717750310898,
0.02524779550731182,
0.006210311781615019,
0.12033049017190933,
-0.15548774600028992,
-0.06465014070272446,
-0.07156258821487427,
-0.06162995845079422,
0.010127202607691288,
0.07951903343200684,
0.08033239096403122,
0.028765788301825523,
-0.04641811177134514,
-0.2056385576725006,
0.00005129708006279543,
0.15709029138088226,
-0.09697046130895615,
-0.08742782473564148
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | octadion/phi-2-jagr-ppg-simpkb | [
"transformers",
"safetensors",
"phi",
"text-generation",
"conversational",
"custom_code",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T06:12:11+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #phi #text-generation #conversational #custom_code #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #phi #text-generation #conversational #custom_code #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
55,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #phi #text-generation #conversational #custom_code #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.05746966227889061,
0.1385236531496048,
-0.004867508076131344,
0.02071472257375717,
0.09852021187543869,
0.00767628476023674,
0.06745423376560211,
0.11078117787837982,
-0.028323214501142502,
0.11590421944856644,
0.029915133491158485,
0.10265152901411057,
0.11196459829807281,
0.16344478726387024,
-0.00402545602992177,
-0.2235129326581955,
0.047733768820762634,
-0.12416195124387741,
-0.02545560710132122,
0.11989402025938034,
0.13988402485847473,
-0.11031996458768845,
0.07310313731431961,
-0.039893634617328644,
-0.005319454241544008,
-0.03720914199948311,
-0.06028740480542183,
-0.049471594393253326,
0.0563783161342144,
0.07132348418235779,
0.07727204263210297,
0.01167363952845335,
0.09068312495946884,
-0.27639180421829224,
0.023529041558504105,
0.08219383656978607,
-0.002843195106834173,
0.074910469353199,
0.04586777091026306,
-0.08579402416944504,
0.07673662900924683,
-0.057560306042432785,
0.15126171708106995,
0.0816478282213211,
-0.0937337577342987,
-0.19621188938617706,
-0.07942938804626465,
0.09498505294322968,
0.19428475201129913,
0.056087613105773926,
-0.028489213436841965,
0.11988316476345062,
-0.08426212519407272,
0.01540372148156166,
0.060048431158065796,
-0.05769569054245949,
-0.05480299890041351,
0.0719008594751358,
0.08189014345407486,
0.09620440006256104,
-0.1283751130104065,
-0.012685850262641907,
0.01773846708238125,
0.0090157650411129,
0.09698512405157089,
0.02159465290606022,
0.12826694548130035,
0.042495064437389374,
-0.14040601253509521,
-0.0429350920021534,
0.08917707949876785,
0.03944757208228111,
-0.048191219568252563,
-0.2459886372089386,
-0.02454492449760437,
-0.043994758278131485,
-0.032698944211006165,
-0.03971369192004204,
0.042680587619543076,
-0.013317018747329712,
0.07610460370779037,
-0.007794694509357214,
-0.08159075677394867,
-0.043010227382183075,
0.07746226340532303,
0.06838548928499222,
0.026839515194296837,
-0.018959667533636093,
0.008074081502854824,
0.12150757014751434,
0.10395351052284241,
-0.12470367550849915,
-0.04892708361148834,
-0.0627409815788269,
-0.07732044905424118,
-0.04740656912326813,
0.029999801889061928,
0.03353775665163994,
0.04419012740254402,
0.23606662452220917,
0.006677375175058842,
0.051339346915483475,
0.040897566825151443,
0.01470019482076168,
0.0598386786878109,
0.10133040696382523,
-0.058189909905195236,
-0.09150446951389313,
-0.024770822376012802,
0.107140451669693,
0.01159056555479765,
-0.041815418750047684,
-0.059919096529483795,
0.07513680309057236,
0.018322143703699112,
0.12354221194982529,
0.07765360921621323,
0.0025391627568751574,
-0.0789392739534378,
-0.0633503645658493,
0.18426553905010223,
-0.15521371364593506,
0.04393265023827553,
0.03151950240135193,
-0.03305595740675926,
-0.016657589003443718,
0.01325246226042509,
0.03188347443938255,
-0.014243376441299915,
0.09450682997703552,
-0.05411379784345627,
-0.02958715707063675,
-0.11437954753637314,
-0.040538184344768524,
0.030216259881854057,
0.01628960482776165,
-0.03070247732102871,
-0.03484250605106354,
-0.07972639799118042,
-0.06037721037864685,
0.09121552109718323,
-0.06952317804098129,
-0.04440006613731384,
-0.022452939301729202,
-0.08314549922943115,
0.017250480130314827,
0.01899232715368271,
0.09558547288179398,
-0.01954829879105091,
0.04530266672372818,
-0.050374262034893036,
0.06292718648910522,
0.11667252331972122,
0.025878911837935448,
-0.05583733320236206,
0.057015560567379,
-0.2518112361431122,
0.0997423455119133,
-0.07118162512779236,
0.04741493612527847,
-0.15088090300559998,
-0.02326461300253868,
0.0342324934899807,
0.013599200174212456,
-0.007171959616243839,
0.13686347007751465,
-0.21894222497940063,
-0.032788969576358795,
0.1728750318288803,
-0.10206284373998642,
-0.0834961086511612,
0.06336672604084015,
-0.05551645904779434,
0.11132661998271942,
0.043416157364845276,
-0.02516593411564827,
0.03925509378314018,
-0.14855700731277466,
-0.008447929285466671,
-0.04647107422351837,
-0.0228714682161808,
0.16645216941833496,
0.07012762874364853,
-0.05589059740304947,
0.07408986240625381,
0.01696590706706047,
-0.011439411900937557,
-0.04799100011587143,
-0.029442038387060165,
-0.10831106454133987,
0.011032086797058582,
-0.05957651510834694,
0.023817073553800583,
-0.022863054648041725,
-0.09880756586790085,
-0.03092763014137745,
-0.17601776123046875,
0.0015118581941351295,
0.08628936856985092,
-0.00961481872946024,
-0.020197568461298943,
-0.11025777459144592,
0.0020687340293079615,
0.02830357663333416,
0.00009873631643131375,
-0.13750053942203522,
-0.05598185211420059,
0.02416086010634899,
-0.16248303651809692,
0.03520655259490013,
-0.053834278136491776,
0.04666110500693321,
0.04319543018937111,
-0.04176093265414238,
-0.03629007562994957,
0.00776281114667654,
0.013402762822806835,
-0.020218633115291595,
-0.263998419046402,
-0.019503481686115265,
-0.026196198537945747,
0.18399229645729065,
-0.24015958607196808,
0.04559517279267311,
0.07493433356285095,
0.14594809710979462,
0.010362114757299423,
-0.032061975449323654,
0.01320319902151823,
-0.07032810896635056,
-0.03549799695611,
-0.060963138937950134,
-0.013593142852187157,
-0.03818603605031967,
-0.06320119649171829,
0.037602249532938004,
-0.17088249325752258,
-0.03702768310904503,
0.10932213813066483,
0.047352101653814316,
-0.1498814821243286,
-0.030633648857474327,
-0.041412971913814545,
-0.04726599529385567,
-0.06418124586343765,
-0.05848364531993866,
0.11017659306526184,
0.05603303015232086,
0.051465123891830444,
-0.0724409893155098,
-0.07407349348068237,
0.008859802968800068,
-0.026088668033480644,
-0.0157513115555048,
0.08809417486190796,
0.0764237716794014,
-0.12180911004543304,
0.09308463335037231,
0.09748471528291702,
0.0825541689991951,
0.10519793629646301,
-0.010203775949776173,
-0.08508020639419556,
-0.039806727319955826,
0.028819195926189423,
0.01749366521835327,
0.15118007361888885,
-0.019246989861130714,
0.05552605167031288,
0.035661302506923676,
-0.015832915902137756,
0.008677844889461994,
-0.10215961188077927,
0.030657034367322922,
0.030794037505984306,
-0.016896063461899757,
0.04252489283680916,
-0.05074205994606018,
0.016952628269791603,
0.09477737545967102,
0.040441352874040604,
0.04382478445768356,
0.00962810404598713,
-0.04335394501686096,
-0.11092008650302887,
0.17614679038524628,
-0.12406712770462036,
-0.2409994751214981,
-0.1334192305803299,
0.00830252468585968,
0.04488078132271767,
-0.008793212473392487,
0.012697882018983364,
-0.07147599756717682,
-0.11402352899312973,
-0.09877244383096695,
0.023178253322839737,
0.04974658042192459,
-0.0792236477136612,
-0.07659852504730225,
0.06992585957050323,
0.03932543471455574,
-0.1392577886581421,
0.023652853444218636,
0.042008016258478165,
-0.08326837420463562,
0.0010953615419566631,
0.07863078266382217,
0.06348864734172821,
0.18350332975387573,
0.01357745099812746,
-0.02588888444006443,
0.02386881783604622,
0.20626242458820343,
-0.14000304043293,
0.11300922930240631,
0.13505563139915466,
-0.0882151797413826,
0.08425850421190262,
0.20395977795124054,
0.04164494574069977,
-0.10641813278198242,
0.03868408501148224,
0.0270888339728117,
-0.02922191470861435,
-0.2477969080209732,
-0.07198986411094666,
0.002102985978126526,
-0.058222394436597824,
0.06480029970407486,
0.08184408396482468,
0.10129287093877792,
0.016439199447631836,
-0.1071258932352066,
-0.07405751198530197,
0.051450569182634354,
0.11125447601079941,
-0.005910532549023628,
-0.016679951921105385,
0.09446603804826736,
-0.025591200217604637,
0.01568126678466797,
0.09125631302595139,
0.0002873673220165074,
0.17465241253376007,
0.05595720559358597,
0.16461676359176636,
0.08274617046117783,
0.06321300566196442,
0.02026492916047573,
0.003029009560123086,
0.016289686784148216,
0.01680661365389824,
-0.008166175335645676,
-0.09074299037456512,
-0.004041763488203287,
0.12580007314682007,
0.04243526980280876,
0.034940507262945175,
0.013306836597621441,
-0.03519197553396225,
0.08732326328754425,
0.1672331541776657,
0.008624950423836708,
-0.1900274157524109,
-0.07184088975191116,
0.07698535174131393,
-0.08199982345104218,
-0.1049899235367775,
-0.03519793227314949,
0.04009527340531349,
-0.16908346116542816,
0.013644976541399956,
-0.020098986104130745,
0.10543514788150787,
-0.12334825843572617,
-0.014638666063547134,
0.05221875011920929,
0.07380601763725281,
-0.011562392115592957,
0.06746756285429001,
-0.16249842941761017,
0.12101735919713974,
0.020899096503853798,
0.06751453131437302,
-0.09755722433328629,
0.09510833024978638,
-0.007531497627496719,
0.0034959791228175163,
0.13583552837371826,
0.008922228589653969,
-0.05192049220204353,
-0.10409228503704071,
-0.09842025488615036,
-0.010315818712115288,
0.13671079277992249,
-0.14452187716960907,
0.09080246835947037,
-0.023235004395246506,
-0.045316025614738464,
0.006856935564428568,
-0.11659596115350723,
-0.1287468671798706,
-0.18072281777858734,
0.054828621447086334,
-0.13112397491931915,
0.04002131149172783,
-0.10866118967533112,
-0.04437215253710747,
-0.015255841426551342,
0.19968634843826294,
-0.22969546914100647,
-0.06825651973485947,
-0.15332789719104767,
-0.07493536174297333,
0.1366935521364212,
-0.0454484298825264,
0.08752236515283585,
-0.00012351905752439052,
0.19069933891296387,
0.018639009445905685,
-0.011175131425261497,
0.11354268342256546,
-0.10132665932178497,
-0.20615456998348236,
-0.09818694740533829,
0.16263408958911896,
0.13958711922168732,
0.04010928422212601,
-0.004340359475463629,
0.03328219801187515,
-0.022340241819620132,
-0.11882122606039047,
0.01661807671189308,
0.1692664623260498,
0.0976729467511177,
0.030817579478025436,
-0.03677118197083473,
-0.11474554985761642,
-0.07340707629919052,
-0.04002682864665985,
0.019006753340363503,
0.18935881555080414,
-0.07132291793823242,
0.17966096103191376,
0.1449604034423828,
-0.051250554621219635,
-0.19625242054462433,
0.020243411883711815,
0.043259236961603165,
0.0051919883117079735,
0.03659573197364807,
-0.20011788606643677,
0.09352371841669083,
-0.0007347991922870278,
-0.05381276085972786,
0.12803494930267334,
-0.17297081649303436,
-0.14956191182136536,
0.055515218526124954,
0.04923489689826965,
-0.18857704102993011,
-0.12271074205636978,
-0.08916878700256348,
-0.04861535131931305,
-0.1435748040676117,
0.10004668682813644,
-0.005967138800770044,
0.008842729963362217,
0.04058210179209709,
0.019071217626333237,
0.012389134615659714,
-0.040754493325948715,
0.19021274149417877,
-0.02488047443330288,
0.03460189700126648,
-0.07648027688264847,
-0.06312233954668045,
0.05980072170495987,
-0.060713592916727066,
0.07130983471870422,
-0.02924666926264763,
0.014641886577010155,
-0.0988130196928978,
-0.050023581832647324,
-0.026615245267748833,
0.01878456026315689,
-0.09396254271268845,
-0.09936850517988205,
-0.04836555942893028,
0.0911177322268486,
0.08790792524814606,
-0.03523729741573334,
-0.036413080990314484,
-0.0784623995423317,
0.04531817510724068,
0.1840154528617859,
0.18332822620868683,
0.03837170451879501,
-0.06480634957551956,
0.00039307636325247586,
-0.014869583770632744,
0.04838959500193596,
-0.22363287210464478,
0.06183580309152603,
0.040492117404937744,
0.019697153940796852,
0.11782129108905792,
-0.021359633654356003,
-0.163929745554924,
-0.06881757825613022,
0.06046493723988533,
-0.06912395358085632,
-0.17333103716373444,
0.005843041930347681,
0.08068350702524185,
-0.16768556833267212,
-0.032293692231178284,
0.042819902300834656,
-0.013471615500748158,
-0.040456850081682205,
0.012034043669700623,
0.08192963898181915,
0.008444741368293762,
0.07340801507234573,
0.058548424392938614,
0.09211587905883789,
-0.10116398334503174,
0.06256914883852005,
0.08027690649032593,
-0.08699701726436615,
0.028788147494196892,
0.08431323617696762,
-0.06640143692493439,
-0.03511952608823776,
0.04261511191725731,
0.08494039624929428,
0.027601271867752075,
-0.044750284403562546,
0.0049463422037661076,
-0.09725427627563477,
0.05377256125211716,
0.1145293116569519,
0.039554793387651443,
0.008861863985657692,
0.03937578946352005,
0.04953936114907265,
-0.08413149416446686,
0.12341699004173279,
0.01909053884446621,
0.02168506197631359,
-0.03827712684869766,
-0.03409188985824585,
0.037789467722177505,
-0.030846722424030304,
-0.011903384700417519,
-0.03247770667076111,
-0.0746593326330185,
-0.014593190513551235,
-0.1536891609430313,
-0.009634746238589287,
-0.03340249881148338,
0.004699328448623419,
0.021953238174319267,
-0.03875463828444481,
0.00921427272260189,
0.012188510037958622,
-0.0689324289560318,
-0.053956206887960434,
-0.021950094029307365,
0.09499738365411758,
-0.1648770421743393,
0.021130844950675964,
0.08321832120418549,
-0.12495473772287369,
0.0901564359664917,
0.019055556505918503,
0.007507782895117998,
0.03336401656270027,
-0.15001636743545532,
0.05081139877438545,
-0.014437399804592133,
0.016506293788552284,
0.0425710491836071,
-0.21409742534160614,
-0.0050269621424376965,
-0.044238410890102386,
-0.05320776626467705,
-0.008403914049267769,
-0.02530374377965927,
-0.11361873894929886,
0.10178086906671524,
0.007690682075917721,
-0.08775308728218079,
-0.02567288652062416,
0.0362861305475235,
0.0736541673541069,
-0.029966186732053757,
0.15486888587474823,
-0.0038720660377293825,
0.07370446622371674,
-0.18215109407901764,
-0.022573357447981834,
-0.01414426602423191,
0.021496392786502838,
-0.025726264342665672,
-0.018330467864871025,
0.04467324912548065,
-0.028773244470357895,
0.19051697850227356,
-0.022742951288819313,
0.05043132230639458,
0.06432893127202988,
0.0073205046355724335,
-0.011716196313500404,
0.1110541895031929,
0.05497567355632782,
0.023867715150117874,
0.023300651460886,
-0.0007449041004292667,
-0.04137673228979111,
-0.007625795900821686,
-0.1867014616727829,
0.06733196973800659,
0.1525408923625946,
0.08798334747552872,
-0.016801979392766953,
0.07248067855834961,
-0.09960407763719559,
-0.11795017123222351,
0.10363016277551651,
-0.0575883612036705,
-0.015193058177828789,
-0.0622229129076004,
0.149260476231575,
0.15269450843334198,
-0.18959222733974457,
0.06297864764928818,
-0.06349852681159973,
-0.054184820502996445,
-0.11384059488773346,
-0.18311649560928345,
-0.057758450508117676,
-0.054391197860240936,
-0.021770363673567772,
-0.05143166705965996,
0.06599576026201248,
0.05969113111495972,
0.009601753205060959,
0.012415576726198196,
0.08726523071527481,
-0.0070823803544044495,
0.006819309201091528,
0.023325880989432335,
0.06842964887619019,
0.013552450574934483,
-0.03182917833328247,
0.014223468489944935,
-0.00021999624732416123,
0.03205692768096924,
0.05638805776834488,
0.03470000624656677,
-0.032354727387428284,
0.013968908227980137,
-0.034115198999643326,
-0.11671015620231628,
0.04292768985033035,
-0.03245917335152626,
-0.07431741803884506,
0.13439905643463135,
0.022394876927137375,
0.002987125189974904,
-0.023156967014074326,
0.25103557109832764,
-0.07349799573421478,
-0.09539806842803955,
-0.13955055177211761,
0.12290097773075104,
-0.03906571865081787,
0.06686554849147797,
0.03291061520576477,
-0.10908804088830948,
0.021335463970899582,
0.1268765926361084,
0.14829246699810028,
-0.04809194430708885,
0.02037576027214527,
0.03027219884097576,
0.0037535117007791996,
-0.04173056408762932,
0.04642251133918762,
0.07435114681720734,
0.14181631803512573,
-0.05114462599158287,
0.08489888161420822,
0.0015237227780744433,
-0.10146874934434891,
-0.040523868054151535,
0.11065921187400818,
-0.01151181384921074,
0.01571481116116047,
-0.055888980627059937,
0.12364766746759415,
-0.04515314847230911,
-0.24378597736358643,
0.06123887747526169,
-0.07869872450828552,
-0.14115175604820251,
-0.02477775514125824,
0.07327479124069214,
-0.018284862861037254,
0.0273441094905138,
0.0708179697394371,
-0.08043792843818665,
0.189077228307724,
0.040144067257642746,
-0.04845357686281204,
-0.05947069823741913,
0.07322884351015091,
-0.12068752944469452,
0.2797746956348419,
0.009804220870137215,
0.056401729583740234,
0.10593809187412262,
-0.022420883178710938,
-0.12631908059120178,
0.031175123527646065,
0.09233500808477402,
-0.08418802171945572,
0.04131191968917847,
0.2145203799009323,
-0.008222459815442562,
0.11580871045589447,
0.07818866521120071,
-0.08888024091720581,
0.048753123730421066,
-0.11053966730833054,
-0.08853860944509506,
-0.081426240503788,
0.0935700312256813,
-0.06710048019886017,
0.14086629450321198,
0.12447112798690796,
-0.0549192950129509,
0.014260023832321167,
-0.027206527069211006,
0.045072708278894424,
0.011562071740627289,
0.10828889906406403,
0.012157557532191277,
-0.1925230622291565,
0.0250005554407835,
0.013929538428783417,
0.10468301922082901,
-0.186534121632576,
-0.09514316916465759,
0.047688741236925125,
0.005185612011700869,
-0.06774647533893585,
0.11928714066743851,
0.060500629246234894,
0.0444079153239727,
-0.04604130610823631,
-0.024775726720690727,
-0.008862584829330444,
0.1463472545146942,
-0.1058395653963089,
-0.004366404842585325
] |
null | null | diffusers |
# SDXL LoRA DreamBooth - yaneq/jan_bYSe9M1l0pUI1xnDnUr2_SDXL_LoRA_700_9d94_700_1e4_2
<Gallery />
## Model description
These are yaneq/jan_bYSe9M1l0pUI1xnDnUr2_SDXL_LoRA_700_9d94_700_1e4_2 LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.
The weights were trained using [DreamBooth](https://dreambooth.github.io/).
LoRA for the text encoder was enabled: False.
Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
## Trigger words
You should use a photo of MDDL man to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
[Download](yaneq/jan_bYSe9M1l0pUI1xnDnUr2_SDXL_LoRA_700_9d94_700_1e4_2/tree/main) them in the Files & versions tab.
## Training properties
- max_train_steps: 700
- learning_rate: 0.0001
- base_model_name: stabilityai/stable-diffusion-xl-base-1.0
- class_name: man
- training_images_urls: - https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2FWF2NGBPUFgu9eyaCYAwB.jpg?alt=media&token=97c1e215-0a96-4fdf-b292-9ee0e497ba72
- https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2FDAk5k1hGzP9q9y0jpGoO.jpg?alt=media&token=01ed67d1-938a-4f60-bc1a-e1b91412b97e
- https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2FY7nFiafx8co1nK6cnjWJ.jpg?alt=media&token=a1fe8c9a-4d5e-4043-9a82-9304fd430569
- https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2F6JW19SVZPczh5B2DEqKD.jpg?alt=media&token=0e0dc94f-957d-4b51-8979-0216c0849cf6
- https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2FVYOVRhojKt30NzjWRXL0.jpg?alt=media&token=5a3a2afb-4b83-4488-92e5-6651f5173cc0
- https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2F82McawlxnTeA2vBc4bZg.jpg?alt=media&token=f7cfacb2-2186-4005-9211-b7ef762dafad
- https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2Fcn54hvM4ahi3MzpCQN5D.jpg?alt=media&token=e096f4dc-e7c5-4e14-88fc-a5562d103127
- https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2Fz8D9WdMIx4mXcsDGAZm4.jpg?alt=media&token=fded9422-eb7c-4757-8c1f-cb436a348579
- gradient_accumulation_steps: 3
- GPU: T4
- duration: 5399.857093095779
| {"license": "openrail++", "tags": ["stable-diffusion-xl", "stable-diffusion-xl-diffusers", "text-to-image", "diffusers", "lora", "template:sd-lora"], "base_model": "stabilityai/stable-diffusion-xl-base-1.0", "instance_prompt": "a photo of MDDL man"} | text-to-image | yaneq/jan_bYSe9M1l0pUI1xnDnUr2_SDXL_LoRA_700_9d94_700_1e4_2 | [
"diffusers",
"stable-diffusion-xl",
"stable-diffusion-xl-diffusers",
"text-to-image",
"lora",
"template:sd-lora",
"base_model:stabilityai/stable-diffusion-xl-base-1.0",
"license:openrail++",
"has_space",
"region:us"
] | 2024-02-07T06:13:16+00:00 | [] | [] | TAGS
#diffusers #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us
|
# SDXL LoRA DreamBooth - yaneq/jan_bYSe9M1l0pUI1xnDnUr2_SDXL_LoRA_700_9d94_700_1e4_2
<Gallery />
## Model description
These are yaneq/jan_bYSe9M1l0pUI1xnDnUr2_SDXL_LoRA_700_9d94_700_1e4_2 LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.
The weights were trained using DreamBooth.
LoRA for the text encoder was enabled: False.
Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
## Trigger words
You should use a photo of MDDL man to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
Download them in the Files & versions tab.
## Training properties
- max_train_steps: 700
- learning_rate: 0.0001
- base_model_name: stabilityai/stable-diffusion-xl-base-1.0
- class_name: man
- training_images_urls: - URL
- URL
- URL
- URL
- URL
- URL
- URL
- URL
- gradient_accumulation_steps: 3
- GPU: T4
- duration: 5399.857093095779
| [
"# SDXL LoRA DreamBooth - yaneq/jan_bYSe9M1l0pUI1xnDnUr2_SDXL_LoRA_700_9d94_700_1e4_2\n\n<Gallery />",
"## Model description\n\nThese are yaneq/jan_bYSe9M1l0pUI1xnDnUr2_SDXL_LoRA_700_9d94_700_1e4_2 LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.\n\nThe weights were trained using DreamBooth.\n\nLoRA for the text encoder was enabled: False.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix.",
"## Trigger words\n\nYou should use a photo of MDDL man to trigger the image generation.",
"## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab.",
"## Training properties\n- max_train_steps: 700\n- learning_rate: 0.0001\n- base_model_name: stabilityai/stable-diffusion-xl-base-1.0\n- class_name: man\n- training_images_urls: - URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- gradient_accumulation_steps: 3\n- GPU: T4\n- duration: 5399.857093095779"
] | [
"TAGS\n#diffusers #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us \n",
"# SDXL LoRA DreamBooth - yaneq/jan_bYSe9M1l0pUI1xnDnUr2_SDXL_LoRA_700_9d94_700_1e4_2\n\n<Gallery />",
"## Model description\n\nThese are yaneq/jan_bYSe9M1l0pUI1xnDnUr2_SDXL_LoRA_700_9d94_700_1e4_2 LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.\n\nThe weights were trained using DreamBooth.\n\nLoRA for the text encoder was enabled: False.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix.",
"## Trigger words\n\nYou should use a photo of MDDL man to trigger the image generation.",
"## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab.",
"## Training properties\n- max_train_steps: 700\n- learning_rate: 0.0001\n- base_model_name: stabilityai/stable-diffusion-xl-base-1.0\n- class_name: man\n- training_images_urls: - URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- gradient_accumulation_steps: 3\n- GPU: T4\n- duration: 5399.857093095779"
] | [
82,
56,
121,
19,
28,
102
] | [
"passage: TAGS\n#diffusers #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us \n# SDXL LoRA DreamBooth - yaneq/jan_bYSe9M1l0pUI1xnDnUr2_SDXL_LoRA_700_9d94_700_1e4_2\n\n<Gallery />## Model description\n\nThese are yaneq/jan_bYSe9M1l0pUI1xnDnUr2_SDXL_LoRA_700_9d94_700_1e4_2 LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.\n\nThe weights were trained using DreamBooth.\n\nLoRA for the text encoder was enabled: False.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix.## Trigger words\n\nYou should use a photo of MDDL man to trigger the image generation.## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab.## Training properties\n- max_train_steps: 700\n- learning_rate: 0.0001\n- base_model_name: stabilityai/stable-diffusion-xl-base-1.0\n- class_name: man\n- training_images_urls: - URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- gradient_accumulation_steps: 3\n- GPU: T4\n- duration: 5399.857093095779"
] | [
-0.0980074405670166,
0.22867363691329956,
-0.000789402169175446,
0.06291297823190689,
0.10987585037946701,
0.04084290936589241,
0.16324663162231445,
0.10596787184476852,
0.004398522898554802,
0.10474595427513123,
0.035356517881155014,
0.04279053956270218,
0.08894488215446472,
0.13172343373298645,
0.03706947714090347,
-0.19874532520771027,
0.010220205411314964,
-0.06730068475008011,
-0.014849124476313591,
0.048373084515333176,
0.030243096873164177,
-0.07129260152578354,
0.086724191904068,
-0.027901116758584976,
-0.03689679875969887,
0.018762338906526566,
-0.028368722647428513,
-0.021619515493512154,
0.036230143159627914,
0.05989646166563034,
0.0479024238884449,
0.06217523291707039,
0.12128115445375443,
-0.2689191997051239,
0.015413422137498856,
0.08643245697021484,
-0.04484066739678383,
0.052575405687093735,
0.0744781419634819,
-0.06730751693248749,
0.029327429831027985,
-0.0986255332827568,
0.10574474930763245,
0.06308102607727051,
-0.06227237358689308,
-0.17458413541316986,
-0.09050439298152924,
0.06892567127943039,
0.09681935608386993,
0.13784760236740112,
0.0003034154651686549,
0.022576358169317245,
0.06484322249889374,
0.03780680522322655,
0.2059476673603058,
-0.1210923045873642,
-0.03308459371328354,
0.12299075722694397,
0.016006184741854668,
0.06654275953769684,
-0.08364509046077728,
0.00014394435856956989,
0.045750562101602554,
-0.008577377535402775,
0.09705282747745514,
-0.0226144976913929,
0.03917638212442398,
-0.07418306171894073,
-0.07466845959424973,
0.008626984432339668,
0.1079024225473404,
0.044228073209524155,
-0.0571540966629982,
-0.1422693431377411,
-0.04791594296693802,
0.0509161613881588,
-0.023695606738328934,
0.01256385538727045,
0.004280085675418377,
-0.030039863660931587,
-0.018964426591992378,
-0.09143007546663284,
-0.029406437650322914,
-0.04883283004164696,
0.08848035335540771,
0.16360972821712494,
0.008772040717303753,
-0.005234174430370331,
0.04147763550281525,
0.07582345604896545,
0.001001875032670796,
-0.13611453771591187,
-0.05533306300640106,
-0.02618447132408619,
-0.08460216969251633,
-0.00044356106081977487,
0.027093736454844475,
-0.10313072800636292,
0.006102694198489189,
0.07925631105899811,
-0.08260217308998108,
0.0478336401283741,
-0.02770184725522995,
0.020975781604647636,
-0.05630744248628616,
0.11030808836221695,
-0.048827603459358215,
-0.08597922325134277,
0.012341242283582687,
0.09003998339176178,
0.05841892585158348,
-0.03621723875403404,
-0.09570623189210892,
-0.004409435670822859,
-0.06603839248418808,
0.0724610909819603,
0.009766066446900368,
-0.0035672129597514868,
-0.06207539513707161,
-0.0563504695892334,
0.16761691868305206,
-0.09627124667167664,
0.015286202542483807,
0.02182207629084587,
-0.06664823740720749,
-0.034175895154476166,
0.10434435307979584,
-0.02146810106933117,
-0.020227501168847084,
0.08757111430168152,
-0.04567543417215347,
0.01937064900994301,
-0.051355794072151184,
-0.13198551535606384,
0.01543411798775196,
-0.09578009694814682,
-0.015906445682048798,
-0.10750271379947662,
-0.21112798154354095,
-0.07089509069919586,
0.012373396195471287,
-0.028604311868548393,
0.04036082699894905,
-0.09134124964475632,
-0.05832461267709732,
0.004436015151441097,
0.03140228986740112,
-0.01293844636529684,
-0.02905895560979843,
0.03438057750463486,
-0.046634748578071594,
0.10423175990581512,
0.0577855184674263,
0.017300572246313095,
-0.030108483508229256,
0.013754044659435749,
-0.0883445292711258,
0.14273756742477417,
-0.10679177194833755,
0.056865040212869644,
-0.09443274140357971,
-0.07825832813978195,
0.041457708925008774,
-0.0029108496382832527,
0.028084229677915573,
0.14415189623832703,
-0.21909159421920776,
-0.05702893063426018,
0.23964980244636536,
-0.14263592660427094,
-0.08561377227306366,
-0.00526337418705225,
-0.049248579889535904,
0.01630222052335739,
0.05030179023742676,
0.08634035289287567,
0.1952117383480072,
-0.1739312708377838,
-0.06527317315340042,
-0.055811807513237,
0.04846632853150368,
0.03838682174682617,
0.0552809052169323,
-0.014447963796555996,
0.016849901527166367,
0.011117017827928066,
-0.05439712479710579,
0.04452482610940933,
-0.014957219362258911,
-0.06676515191793442,
0.0152308139950037,
-0.07932505011558533,
0.011341377161443233,
-0.015283471904695034,
-0.025400180369615555,
0.006362535525113344,
-0.08294336497783661,
0.033933576196432114,
0.13053512573242188,
-0.06931372731924057,
-0.003307270584627986,
-0.05689670518040657,
-0.04234973341226578,
-0.023888451978564262,
-0.003321145661175251,
-0.11272598057985306,
-0.1530066877603531,
0.04059332236647606,
-0.04342728108167648,
0.015879124402999878,
-0.08503004163503647,
0.08344021439552307,
0.12838469445705414,
-0.0713091567158699,
-0.10275787860155106,
-0.06542447209358215,
-0.01273875031620264,
0.0309734083712101,
-0.13575492799282074,
-0.04889640584588051,
-0.09467024356126785,
0.06053189933300018,
-0.18512757122516632,
0.042821865528821945,
0.016112828627228737,
0.12032515555620193,
0.0931181013584137,
-0.06522965431213379,
0.06862366944551468,
0.013577462173998356,
-0.020185543224215508,
-0.11280537396669388,
-0.019296785816550255,
-0.018122274428606033,
-0.12317250669002533,
0.0351482555270195,
-0.16696105897426605,
0.021411795169115067,
0.07585649192333221,
0.1362505704164505,
-0.11173734068870544,
-0.024396052584052086,
-0.03811149299144745,
-0.024786723777651787,
-0.1304996907711029,
-0.035295650362968445,
0.1102999895811081,
0.041833799332380295,
0.10500898957252502,
-0.0426180399954319,
-0.016758350655436516,
0.01881888322532177,
0.02841421216726303,
-0.0440044179558754,
0.10689114034175873,
-0.0269943755120039,
0.0008220930467359722,
0.0761837288737297,
-0.00909108854830265,
-0.02149341255426407,
0.1645984798669815,
0.02772475965321064,
-0.040222130715847015,
-0.01584312878549099,
0.03894728049635887,
0.07978109270334244,
0.11064667999744415,
0.09903939813375473,
0.0483572855591774,
0.007428554818034172,
-0.013264409266412258,
-0.02009943500161171,
-0.12944790720939636,
-0.047492071986198425,
0.036891430616378784,
-0.07954135537147522,
0.0868026539683342,
-0.00102724670432508,
-0.03244886174798012,
0.06975161284208298,
-0.024284614250063896,
0.05902465432882309,
-0.013703004457056522,
-0.054887805134058,
-0.11131066828966141,
0.16064532101154327,
-0.1270996779203415,
-0.20547270774841309,
-0.10229285061359406,
0.02140369825065136,
-0.051335472613573074,
0.002027534181252122,
0.03911533206701279,
-0.12438269704580307,
-0.059276681393384933,
-0.11521603167057037,
0.010790328495204449,
-0.00012323661940172315,
-0.04145625978708267,
0.045101456344127655,
0.018220992758870125,
0.07607852667570114,
-0.08365515619516373,
0.01561189629137516,
0.025896815583109856,
-0.02420046180486679,
0.01862904615700245,
0.04270276054739952,
0.08580084890127182,
0.07147607207298279,
0.04522646218538284,
0.03049875609576702,
-0.02014973945915699,
0.20791974663734436,
-0.021333802491426468,
0.08399268239736557,
0.15419793128967285,
0.10259338468313217,
0.09306450933218002,
0.152849942445755,
0.024409789592027664,
-0.06252976506948471,
0.05062610283493996,
0.045219045132398605,
-0.0744246169924736,
-0.1689000427722931,
-0.09214957058429718,
-0.03698299080133438,
-0.04170789197087288,
0.1309773176908493,
0.06731747090816498,
0.07386398315429688,
0.08287664502859116,
-0.07084039598703384,
0.038130223751068115,
0.07083272933959961,
0.11926702409982681,
-0.006311352364718914,
0.012281421571969986,
0.028232725337147713,
-0.046835191547870636,
0.005333585198968649,
0.07742977887392044,
0.038202185183763504,
0.18629422783851624,
-0.07494708895683289,
0.03230593353509903,
0.010934138670563698,
0.0377754308283329,
0.0154117988422513,
0.008220138028264046,
0.014794543385505676,
0.0032392102293670177,
0.005235428921878338,
-0.1362428367137909,
0.02876678854227066,
0.16257180273532867,
-0.0371139794588089,
0.0025577028281986713,
-0.011228371411561966,
0.037864841520786285,
-0.006194514222443104,
0.07642153650522232,
0.060860056430101395,
-0.30431440472602844,
-0.029797496274113655,
0.035225510597229004,
0.08618616312742233,
-0.029961667954921722,
-0.009332919493317604,
0.1312992423772812,
-0.0670199915766716,
0.12258340418338776,
-0.059563253074884415,
0.08726218342781067,
-0.05225934088230133,
-0.06065630167722702,
0.08365818858146667,
0.14642509818077087,
-0.020046595484018326,
0.023507479578256607,
-0.1370982825756073,
0.10545354336500168,
-0.0014065306168049574,
0.09524856507778168,
0.006237173452973366,
0.04392068833112717,
0.05008354410529137,
0.021332988515496254,
0.16815248131752014,
0.012950177304446697,
-0.10780647397041321,
-0.16873033344745636,
-0.09968043863773346,
-0.02662680298089981,
0.060081660747528076,
-0.11633889377117157,
0.06415604799985886,
0.044164445251226425,
-0.04466859996318817,
-0.007066607009619474,
-0.00570883322507143,
-0.16001516580581665,
-0.13775157928466797,
-0.0015186337986961007,
0.07579827308654785,
0.09443000704050064,
-0.08490054309368134,
-0.05969027429819107,
-0.0032982525881379843,
0.1084408387541771,
-0.07295652478933334,
-0.1212841346859932,
-0.141661137342453,
-0.020336057990789413,
0.08457127213478088,
-0.0773598775267601,
0.001456446130760014,
0.020951220765709877,
0.1715291291475296,
-0.09250659495592117,
-0.09192980080842972,
-0.004097112454473972,
-0.094574935734272,
-0.16725686192512512,
-0.10863501578569412,
0.09232207387685776,
0.04078717529773712,
0.005267453845590353,
-0.008741248399019241,
0.002783097792416811,
0.03644610196352005,
-0.11249268800020218,
0.05974775180220604,
0.22414177656173706,
-0.017729390412569046,
0.09682570397853851,
-0.015825608745217323,
-0.08620379865169525,
-0.06599059700965881,
0.01483639981597662,
0.018419859930872917,
0.2516278624534607,
-0.10444734990596771,
0.09910774230957031,
0.08304957300424576,
-0.07102037221193314,
-0.15148553252220154,
0.014571413397789001,
0.07885468006134033,
-0.01006101630628109,
0.012944751419126987,
-0.25723889470100403,
0.06266273558139801,
0.005880684591829777,
-0.02203448675572872,
0.06944181025028229,
-0.32508230209350586,
-0.12529601156711578,
-0.027346927672624588,
0.16054894030094147,
-0.008784247562289238,
-0.10451926290988922,
-0.06924834847450256,
-0.08452180027961731,
-0.07896328717470169,
0.15984851121902466,
-0.017279978841543198,
0.06457211077213287,
-0.0015097258146852255,
0.04703381285071373,
0.020044483244419098,
-0.05002108961343765,
0.11814422160387039,
0.026613404974341393,
0.05225881561636925,
-0.034855589270591736,
0.07106117904186249,
0.07795350253582001,
-0.07863745093345642,
0.06284452229738235,
-0.06673868000507355,
0.06592023372650146,
-0.10048174858093262,
-0.03755048289895058,
-0.012577162124216557,
0.040398839861154556,
-0.03823867067694664,
-0.08345551788806915,
-0.040955040603876114,
0.09887098520994186,
0.14626170694828033,
-0.01879209838807583,
-0.08379995077848434,
-0.009678727015852928,
0.07439208030700684,
0.12225178629159927,
-0.01158792246133089,
0.1195339635014534,
-0.141649067401886,
-0.027897486463189125,
-0.02739192172884941,
0.10045130550861359,
-0.1168186292052269,
0.00259419740177691,
0.08413410931825638,
0.057679831981658936,
0.15084560215473175,
0.0075190444476902485,
-0.11340038478374481,
0.06138783320784569,
0.03548165783286095,
-0.08409722149372101,
-0.1310107558965683,
-0.04553531855344772,
0.004250032361596823,
-0.06484289467334747,
-0.024709373712539673,
0.13546885550022125,
-0.07771506905555725,
0.03613274544477463,
0.0006063090986572206,
0.04801188036799431,
0.011608261615037918,
0.12642566859722137,
0.04568829387426376,
0.049776773899793625,
-0.041735030710697174,
0.09495782852172852,
0.04228275641798973,
-0.0769498348236084,
0.05239779129624367,
0.1321304440498352,
-0.07952766865491867,
-0.009003780782222748,
-0.09844347089529037,
0.06840808689594269,
-0.11348375678062439,
-0.013599861413240433,
-0.03340890631079674,
-0.0951462909579277,
0.03639064356684685,
0.09281495213508606,
-0.006253615953028202,
-0.0009450280340388417,
0.009308679960668087,
-0.008164994418621063,
-0.07526801526546478,
0.06998255103826523,
0.01406136155128479,
0.05720910802483559,
-0.13238397240638733,
-0.02683485485613346,
0.03832108527421951,
-0.005639240145683289,
-0.0031784565653651953,
-0.023951970040798187,
-0.12334194779396057,
-0.0050606695003807545,
-0.10544241219758987,
0.13185110688209534,
-0.09000074118375778,
0.002791519043967128,
-0.05659036338329315,
0.010482478886842728,
0.019143112003803253,
0.04837380349636078,
-0.0304274745285511,
-0.09263575822114944,
-0.018015706911683083,
0.06987783312797546,
-0.12858864665031433,
-0.04598630219697952,
-0.009486294351518154,
-0.10184872895479202,
0.05497530475258827,
-0.03583088889718056,
-0.04858457297086716,
-0.015316726639866829,
-0.15826274454593658,
0.040299609303474426,
0.04335266724228859,
0.012488403357565403,
0.007609971798956394,
-0.08195029944181442,
-0.004572434816509485,
-0.05252430588006973,
-0.05409913882613182,
-0.007079755887389183,
-0.023396989330649376,
-0.08838415145874023,
0.01986389048397541,
-0.048921264708042145,
0.01983635686337948,
-0.06641669571399689,
0.08414356410503387,
0.11655045300722122,
0.023637671023607254,
0.06420104205608368,
-0.09472377598285675,
0.1278841346502304,
-0.18421350419521332,
-0.03329646214842796,
0.03021918423473835,
-0.01695660874247551,
-0.0606049969792366,
0.0213171374052763,
0.07631183415651321,
-0.021165089681744576,
0.07505031675100327,
-0.040266670286655426,
0.05099201947450638,
-0.003697310108691454,
-0.027254100888967514,
-0.013377657160162926,
0.015594406984746456,
0.16031524538993835,
0.02333802729845047,
-0.01282659824937582,
0.053923167288303375,
0.0011980280978605151,
0.05638696625828743,
0.07971053570508957,
0.08503493666648865,
0.1538434624671936,
0.09173497557640076,
0.03269632160663605,
0.03978826478123665,
-0.07367292046546936,
-0.08489005267620087,
0.10263226926326752,
-0.042648158967494965,
0.06918129324913025,
-0.0704624131321907,
0.03352856636047363,
0.10129842162132263,
-0.15239176154136658,
0.03004494123160839,
0.08546682447195053,
-0.06227711960673332,
-0.06551099568605423,
-0.16620555520057678,
-0.08373339474201202,
-0.09873700141906738,
0.04861283302307129,
-0.10334304720163345,
0.035120539367198944,
0.010819845832884312,
-0.014206147752702236,
0.045808687806129456,
0.11690358072519302,
0.015112594701349735,
-0.039975252002477646,
0.03814456984400749,
0.02306002750992775,
-0.017975617200136185,
0.04290662705898285,
-0.0342472568154335,
0.12485309690237045,
0.015270592644810677,
0.024118918925523758,
0.026806285604834557,
0.03175734356045723,
0.05333800986409187,
0.008808553218841553,
-0.06327399611473083,
0.01114809699356556,
0.006295536644756794,
0.014838595874607563,
0.15645170211791992,
0.0882958173751831,
0.008355752564966679,
-0.01895778998732567,
0.17324256896972656,
-0.08119750022888184,
-0.04338886961340904,
-0.11281611025333405,
0.07969053089618683,
-0.009784875437617302,
0.008534075692296028,
0.02878415770828724,
-0.08403560519218445,
-0.025265904143452644,
0.09017861634492874,
0.1259707361459732,
-0.0004962739767506719,
0.01836756244301796,
-0.049782950431108475,
-0.001673010061495006,
-0.05537962168455124,
0.08995846658945084,
0.024765459820628166,
0.1340513527393341,
-0.03593641519546509,
0.02999204210937023,
-0.02134483866393566,
-0.0642620176076889,
0.004182412289083004,
-0.02800523303449154,
-0.03713085129857063,
0.005263130180537701,
-0.06304233521223068,
0.0810767114162445,
-0.11393752694129944,
-0.17177258431911469,
0.19116197526454926,
-0.10360849648714066,
-0.0780695378780365,
-0.030633993446826935,
0.05548585206270218,
0.03452082350850105,
-0.0034389568027108908,
0.01257381308823824,
-0.029727939516305923,
0.13943730294704437,
-0.029689036309719086,
-0.08298449963331223,
-0.07521943002939224,
-0.025973668321967125,
-0.14373096823692322,
0.2540200650691986,
-0.0392126590013504,
0.013027763925492764,
0.07784071564674377,
0.029204048216342926,
-0.12131685763597488,
0.07296883314847946,
0.016511686146259308,
-0.10565470159053802,
-0.01734989508986473,
0.1882571429014206,
-0.06169172748923302,
0.19644592702388763,
0.023806463927030563,
-0.09845355153083801,
-0.0029213291127234697,
0.0732317864894867,
-0.02316836453974247,
-0.0867353305220604,
0.028912004083395004,
-0.06666365265846252,
0.12694081664085388,
0.12973618507385254,
-0.026924343779683113,
-0.01796543411910534,
-0.0009127118973992765,
0.016831744462251663,
0.012247984297573566,
0.09665712714195251,
0.029494503512978554,
-0.09797443449497223,
-0.02702377922832966,
0.006030981428921223,
0.08086327463388443,
-0.2458517849445343,
-0.11898517608642578,
-0.027294747531414032,
-0.038618847727775574,
0.022672725841403008,
0.12163256108760834,
0.08755642920732498,
0.041980233043432236,
-0.02922956831753254,
-0.2026132494211197,
0.0030061043798923492,
0.13640262186527252,
-0.0897175744175911,
-0.07260693609714508
] |
null | null | transformers |
# Full Parameter Finetuning Qwen1.5 0.5B on Malaysian text
README at https://github.com/huseinzol05/malaya/tree/5.1/session/qwen2
WandB, https://wandb.ai/huseinzol05/finetune-Qwen1.5-0.5B?workspace=user-huseinzol05 | {"language": ["ms"]} | text-generation | mesolitica/Qwen1.5-0.5B-4096-fpf | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"ms",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T06:13:40+00:00 | [] | [
"ms"
] | TAGS
#transformers #safetensors #qwen2 #text-generation #conversational #ms #autotrain_compatible #endpoints_compatible #region-us
|
# Full Parameter Finetuning Qwen1.5 0.5B on Malaysian text
README at URL
WandB, URL | [
"# Full Parameter Finetuning Qwen1.5 0.5B on Malaysian text\n\nREADME at URL\n\nWandB, URL"
] | [
"TAGS\n#transformers #safetensors #qwen2 #text-generation #conversational #ms #autotrain_compatible #endpoints_compatible #region-us \n",
"# Full Parameter Finetuning Qwen1.5 0.5B on Malaysian text\n\nREADME at URL\n\nWandB, URL"
] | [
45,
24
] | [
"passage: TAGS\n#transformers #safetensors #qwen2 #text-generation #conversational #ms #autotrain_compatible #endpoints_compatible #region-us \n# Full Parameter Finetuning Qwen1.5 0.5B on Malaysian text\n\nREADME at URL\n\nWandB, URL"
] | [
-0.05077666416764259,
-0.12399144470691681,
-0.003453192301094532,
0.00850912369787693,
0.08324811607599258,
-0.035280924290418625,
0.11299920082092285,
0.07743149995803833,
-0.10692690312862396,
-0.011682765558362007,
0.06566181033849716,
-0.009335774928331375,
0.03267520293593407,
0.04995323717594147,
-0.11538765579462051,
-0.21191909909248352,
0.08950965851545334,
0.00028403408941812813,
0.05213405191898346,
0.1091923713684082,
0.09561753273010254,
-0.05645259469747543,
0.1170140728354454,
-0.006505968049168587,
-0.0453210324048996,
0.01065157726407051,
0.0077119795605540276,
-0.1370866745710373,
0.07703518122434616,
0.08632609248161316,
0.07766800373792648,
0.1111249104142189,
-0.022536270320415497,
-0.18335817754268646,
0.04102450609207153,
-0.05377192422747612,
-0.03666640818119049,
-0.003601495176553726,
0.044476889073848724,
0.09633023291826248,
0.0036078691482543945,
0.03781761974096298,
-0.03708583861589432,
0.066583551466465,
-0.06080985441803932,
0.09568068385124207,
-0.10925750434398651,
0.12673497200012207,
0.11811820417642593,
0.1313530057668686,
-0.0657484158873558,
0.1773490160703659,
-0.12290295958518982,
0.09688820689916611,
0.10123805701732635,
-0.3846878707408905,
0.011323481798171997,
0.054456744343042374,
0.06431411951780319,
0.10130187124013901,
-0.12059882283210754,
0.01571549102663994,
0.06512570381164551,
-0.012684136629104614,
-0.06432916224002838,
-0.13074254989624023,
-0.06688795983791351,
-0.03796831890940666,
-0.014513228088617325,
0.06344785541296005,
0.1569100320339203,
0.057679835706949234,
-0.04225252568721771,
-0.08992050588130951,
-0.06940441578626633,
-0.008769151754677296,
-0.018431423231959343,
-0.007610405795276165,
-0.0815545991063118,
0.0785168781876564,
-0.05202806740999222,
-0.05878929793834686,
-0.11753203719854355,
0.0030565483029931784,
-0.19656485319137573,
0.1374935656785965,
0.012765396386384964,
0.008295686915516853,
-0.19091922044754028,
-0.0014276480069383979,
0.03386099636554718,
-0.1587810069322586,
-0.055294860154390335,
-0.09138406068086624,
0.050051361322402954,
0.03949220851063728,
-0.0221895482391119,
-0.057196155190467834,
0.14832861721515656,
0.07313582301139832,
-0.09294501692056656,
0.08549327403306961,
-0.006896990817040205,
0.03941110149025917,
-0.04262242838740349,
0.09809710830450058,
-0.054489847272634506,
-0.05455499887466431,
0.14708013832569122,
0.004133199341595173,
0.07759228348731995,
-0.05265176668763161,
-0.05124233663082123,
-0.04207533597946167,
0.04837796464562416,
0.08555803447961807,
-0.05721065402030945,
0.11762875318527222,
0.03579026088118553,
0.02396598644554615,
0.2066044956445694,
-0.08431041240692139,
-0.08944486826658249,
0.032847125083208084,
0.015222647227346897,
0.013306583277881145,
0.0064257229678332806,
0.07151208072900772,
-0.06973521411418915,
0.08462653309106827,
-0.0007223564316518605,
-0.036385346204042435,
0.08453236520290375,
-0.07253527641296387,
-0.029636971652507782,
0.04609723761677742,
0.048161428421735764,
-0.19324377179145813,
-0.04474323242902756,
0.006805538199841976,
-0.09464780241250992,
0.009343436919152737,
-0.05105392634868622,
0.0006966280634514987,
-0.06988541036844254,
0.05094269663095474,
-0.04110544174909592,
-0.015185490250587463,
-0.04878857359290123,
0.11833993345499039,
0.14430361986160278,
0.06342177093029022,
-0.1688695102930069,
0.0018353560008108616,
-0.09615776687860489,
0.016615847125649452,
-0.018734246492385864,
0.1125263050198555,
-0.036653365939855576,
0.12562797963619232,
-0.020148536190390587,
-0.02201998606324196,
-0.09555932134389877,
-0.027513843029737473,
0.0047609261237084866,
0.1767871230840683,
-0.06829329580068588,
-0.06258230656385422,
0.19578607380390167,
-0.10848049074411392,
-0.17148800194263458,
0.13325493037700653,
0.024023666977882385,
-0.00877944566309452,
0.03435777500271797,
0.1956360638141632,
0.1208425909280777,
-0.02401602827012539,
-0.046065885573625565,
0.12049111723899841,
-0.05915047973394394,
0.018098559230566025,
0.032022781670093536,
0.037359464913606644,
-0.01903430186212063,
0.06824784725904465,
0.07362467050552368,
0.07953102886676788,
0.006982012186199427,
-0.09015341103076935,
-0.0368884913623333,
-0.08695463836193085,
0.01929689384996891,
-0.07284620404243469,
0.09792206436395645,
-0.06331767141819,
0.023770593106746674,
-0.13535280525684357,
0.05676645785570145,
0.005508873146027327,
0.02333902381360531,
-0.11957144737243652,
0.04018733277916908,
-0.0007281532161869109,
0.0794588103890419,
-0.11938707530498505,
-0.05234978720545769,
-0.02213423140347004,
0.14571796357631683,
0.07440603524446487,
-0.039621129631996155,
0.02734154276549816,
-0.006957388948649168,
-0.066945880651474,
-0.08226089179515839,
0.08680888265371323,
-0.0174623541533947,
-0.09492380172014236,
-0.005225342232733965,
0.11570490151643753,
-0.015088004991412163,
0.03300837427377701,
-0.054604023694992065,
0.013554622419178486,
0.06994263827800751,
-0.006506793200969696,
0.02430425025522709,
0.06905294209718704,
0.10148736834526062,
0.02763601765036583,
-0.030289512127637863,
0.02109508030116558,
0.02651294507086277,
0.027482977136969566,
-0.15665973722934723,
0.1796753853559494,
-0.18771323561668396,
0.20055682957172394,
0.18315203487873077,
-0.11012037843465805,
0.04055745154619217,
-0.03602916747331619,
-0.018514985218644142,
-0.013775866478681564,
-0.05904935672879219,
0.10227832943201065,
0.08870120346546173,
0.020541660487651825,
0.15881562232971191,
-0.08743049949407578,
-0.020459724590182304,
0.026316212490200996,
-0.0802706703543663,
-0.026975028216838837,
0.055212609469890594,
-0.05213722586631775,
-0.19733785092830658,
0.028934119269251823,
0.11509248614311218,
-0.0038748590741306543,
0.08010128885507584,
0.01714797504246235,
0.011386225931346416,
0.049671720713377,
0.07862989604473114,
-0.013701898977160454,
0.0018654457526281476,
-0.07679297775030136,
-0.014062718488276005,
0.040695223957300186,
0.07082639634609222,
0.06215968355536461,
-0.08472324907779694,
-0.05980871617794037,
-0.01930098980665207,
-0.05298260971903801,
-0.12743425369262695,
0.07890888303518295,
-0.008055497892200947,
0.08971775323152542,
-0.042070768773555756,
0.055139437317848206,
0.027423342689871788,
-0.05577096715569496,
-0.14536000788211823,
0.16204826533794403,
-0.08940145373344421,
-0.28032758831977844,
-0.11921839416027069,
-0.12876050174236298,
-0.12418220192193985,
-0.040808141231536865,
0.1326545625925064,
-0.190651535987854,
-0.04719763249158859,
-0.03220155090093613,
-0.050249792635440826,
0.07092125713825226,
0.07465014606714249,
-0.02593131735920906,
0.03962372988462448,
0.02597588114440441,
-0.09812399744987488,
0.0036244262009859085,
-0.006239838432520628,
-0.047215964645147324,
0.14182962477207184,
-0.11033491790294647,
0.11360419541597366,
0.09209723025560379,
-0.03584353253245354,
0.009492991492152214,
-0.028255900368094444,
0.31672605872154236,
-0.046008624136447906,
-0.010115137323737144,
0.141250878572464,
-0.07555875182151794,
0.025500701740384102,
0.17393431067466736,
-0.04317863658070564,
-0.0779830813407898,
0.09579205513000488,
0.05059221759438515,
-0.007301328703761101,
-0.1900523602962494,
-0.023386100307106972,
-0.05560516193509102,
0.03707810491323471,
-0.11685345321893692,
0.0609176829457283,
0.017659559845924377,
0.0757567286491394,
-0.05132489278912544,
0.004707975313067436,
0.06810817122459412,
0.05024610459804535,
0.15189066529273987,
0.040388964116573334,
0.11103861033916473,
-0.12646064162254333,
-0.0254246536642313,
0.06898128986358643,
0.11349038034677505,
0.0869358628988266,
0.0024684956297278404,
0.046588778495788574,
0.04572072625160217,
0.24455341696739197,
0.15416322648525238,
0.05171927809715271,
-0.11133456230163574,
-0.07554883509874344,
0.009263084270060062,
-0.04086359590291977,
-0.02246619202196598,
-0.03539317101240158,
-0.07749369740486145,
-0.0008825700497254729,
-0.1067325621843338,
0.1955680251121521,
0.10240261256694794,
0.07161534577608109,
-0.008335407823324203,
0.03543027490377426,
-0.07875003665685654,
0.048346761614084244,
-0.04997916519641876,
-0.0146868247538805,
0.09835534542798996,
0.11174891889095306,
-0.07279064506292343,
0.06831219792366028,
0.021127013489603996,
0.14206339418888092,
-0.02414887398481369,
0.057779114693403244,
-0.15049049258232117,
-0.10235687345266342,
-0.030519723892211914,
0.1158231571316719,
-0.3656458258628845,
0.19119219481945038,
0.03615046292543411,
0.0659102126955986,
0.014548305422067642,
-0.04543466866016388,
0.05313585326075554,
0.17440131306648254,
0.07130567729473114,
-0.018041744828224182,
0.03022255003452301,
-0.1263556033372879,
-0.05416812002658844,
0.099620021879673,
0.11832809448242188,
0.19217219948768616,
0.04552476853132248,
-0.027889398857951164,
0.03075379505753517,
-0.013114787638187408,
0.10444610565900803,
-0.13128550350666046,
-0.13122640550136566,
0.0034080713521689177,
0.0993795096874237,
0.028442231938242912,
-0.04107796028256416,
0.042307667434215546,
-0.1172674149274826,
0.1500021070241928,
-0.17319975793361664,
-0.09770891070365906,
-0.06279370188713074,
-0.0612669438123703,
-0.000828389311209321,
-0.14296473562717438,
-0.060944702476263046,
-0.05915714427828789,
0.02041887491941452,
-0.06270577013492584,
-0.07941488176584244,
0.039936695247888565,
-0.07441984117031097,
-0.10887430608272552,
0.03032171167433262,
0.10061219334602356,
-0.102720245718956,
0.025691337883472443,
0.07467757910490036,
-0.11389938741922379,
-0.03364364057779312,
-0.17902128398418427,
-0.0718924030661583,
-0.09661990404129028,
-0.06820518523454666,
-0.0306119192391634,
-0.03322968631982803,
-0.14054960012435913,
-0.13886544108390808,
-0.09920094162225723,
0.22599726915359497,
0.17675358057022095,
-0.01091874111443758,
0.08894740790128708,
0.26790690422058105,
0.015592988580465317,
-0.20846040546894073,
-0.1415051519870758,
-0.08368415385484695,
-0.028327791020274162,
-0.07740413397550583,
0.016894444823265076,
0.0781303271651268,
0.017388567328453064,
0.019552217796444893,
-0.12086353451013565,
-0.13057412207126617,
-0.12608374655246735,
0.12248712033033371,
0.03035096265375614,
0.24865639209747314,
-0.16403383016586304,
-0.12485172599554062,
-0.06697756797075272,
-0.25881922245025635,
0.01522108819335699,
-0.13822202384471893,
0.083316870033741,
0.014140254817903042,
0.10379601269960403,
-0.005628527607768774,
0.017465321347117424,
0.14455801248550415,
-0.03278220817446709,
0.015399323776364326,
-0.07510567456483841,
-0.07829542458057404,
-0.0017570587806403637,
0.020876780152320862,
0.04318026825785637,
-0.1571919173002243,
0.10159717500209808,
-0.08483469486236572,
-0.030331667512655258,
-0.05257485434412956,
0.0067877983674407005,
0.012950635515153408,
-0.04485646262764931,
-0.033194392919540405,
-0.0633009597659111,
0.05025119334459305,
-0.00020893877081107348,
0.11847987025976181,
-0.07800040394067764,
-0.010208383202552795,
0.07198604196310043,
0.17823585867881775,
-0.21803638339042664,
0.1293623298406601,
-0.07743529975414276,
-0.09117407351732254,
0.09359097480773926,
-0.07857517898082733,
0.04598869010806084,
0.039183180779218674,
-0.04730809107422829,
0.07125722616910934,
0.047313570976257324,
0.029493432492017746,
0.1481626033782959,
0.11887892335653305,
-0.13683642446994781,
-0.002725468948483467,
-0.059625398367643356,
0.08402208238840103,
0.015413385815918446,
0.027342749759554863,
0.0878482460975647,
-0.025403916835784912,
-0.016561735421419144,
-0.011023684404790401,
-0.0012379256077110767,
-0.06376931071281433,
0.15172216296195984,
0.03430737927556038,
0.06860048323869705,
-0.1342894732952118,
0.12592798471450806,
0.020825447514653206,
-0.11059366166591644,
0.0107868118211627,
0.19961299002170563,
-0.194970041513443,
-0.07651353627443314,
-0.05982190743088722,
0.12108311802148819,
-0.08388292789459229,
-0.07524118572473526,
-0.04914190620183945,
-0.12109652906656265,
-0.04407788813114166,
0.16213110089302063,
0.026774873957037926,
0.049595680087804794,
-0.0031238843221217394,
-0.036251697689294815,
-0.06989677995443344,
0.032218363136053085,
0.030330948531627655,
0.0443262904882431,
-0.15999650955200195,
0.022180285304784775,
0.05003373324871063,
0.1697552502155304,
-0.05958766117691994,
-0.06589248031377792,
-0.1040550097823143,
0.008902163244783878,
-0.20526960492134094,
0.09624463319778442,
-0.08568506687879562,
0.016308365389704704,
-0.09598835557699203,
-0.062336403876543045,
-0.05142409726977348,
-0.028743088245391846,
-0.007908857427537441,
-0.018235944211483,
-0.05101822316646576,
-0.0023583690635859966,
-0.12829729914665222,
0.03264407813549042,
0.03853674978017807,
0.012394597753882408,
0.1381673514842987,
0.08002365380525589,
-0.12365316599607468,
0.10052486509084702,
-0.05609169229865074,
-0.08390471339225769,
0.09757498651742935,
-0.00047898333286866546,
0.018372582271695137,
0.057387758046388626,
0.017437444999814034,
0.0561840757727623,
0.027367763221263885,
0.08455855399370193,
0.2687167823314667,
-0.08728750050067902,
-0.053089357912540436,
-0.17844697833061218,
-0.041483696550130844,
-0.09138698130846024,
-0.03033990040421486,
0.06975452601909637,
-0.004249743185937405,
0.1895422786474228,
-0.13014647364616394,
0.0017451723106205463,
-0.07937249541282654,
0.012068343348801136,
0.012874377891421318,
-0.13050243258476257,
-0.187033548951149,
-0.058310169726610184,
0.02918202243745327,
-0.06383249163627625,
0.147404745221138,
-0.12554587423801422,
0.00614278856664896,
0.06017984077334404,
-0.010357997380197048,
-0.06728734076023102,
-0.034072861075401306,
0.36908069252967834,
0.15684309601783752,
-0.033311713486909866,
-0.033171601593494415,
0.003968902863562107,
-0.011359656229615211,
-0.038773488253355026,
0.08986453711986542,
0.09497561305761337,
0.002466626465320587,
0.13598689436912537,
0.006197609938681126,
0.027945930138230324,
-0.031614452600479126,
-0.1442534625530243,
-0.1353960782289505,
-0.0488041527569294,
0.009327638894319534,
-0.05705026164650917,
0.3430325388908386,
-0.03084644488990307,
-0.0069969771429896355,
-0.0038997407536953688,
-0.07671553641557693,
-0.16324514150619507,
-0.053035035729408264,
-0.12187723815441132,
-0.08418609201908112,
0.027703184634447098,
-0.1133815199136734,
-0.03199673444032669,
-0.0782565176486969,
0.043997641652822495,
-0.05837481468915939,
0.1804620623588562,
-0.09500238299369812,
-0.06650416553020477,
0.029385054484009743,
-0.05103889852762222,
0.002554124454036355,
0.1234898790717125,
0.0433841347694397,
0.021098416298627853,
-0.034714870154857635,
-0.011951517313718796,
0.08868055790662766,
-0.04774521291255951,
0.06052982062101364,
-0.13062511384487152,
-0.08679713308811188,
-0.03453410416841507,
0.058154381811618805,
0.04996978119015694,
0.21083608269691467,
0.05028079077601433,
-0.020714476704597473,
0.013790611177682877,
0.2544800341129303,
-0.04412470757961273,
-0.19822706282138824,
-0.11041709035634995,
0.03837663680315018,
0.02458825893700123,
0.07812909781932831,
-0.05501357838511467,
-0.07103265821933746,
-0.007804937660694122,
0.3875018060207367,
0.17737796902656555,
-0.14748112857341766,
0.07935941964387894,
-0.03957555443048477,
0.0477588027715683,
0.04479373246431351,
0.12981747090816498,
0.1567174345254898,
0.24334481358528137,
-0.048842478543519974,
-0.059128742665052414,
-0.05025621876120567,
-0.049824293702840805,
-0.0811614990234375,
0.1465277522802353,
0.06241386756300926,
-0.018467720597982407,
-0.10927853733301163,
0.07289714366197586,
-0.18091492354869843,
0.1877679079771042,
-0.14932410418987274,
-0.06778626888990402,
-0.04115128517150879,
-0.016127262264490128,
0.03821113705635071,
0.08435659110546112,
-0.019482389092445374,
0.050974950194358826,
0.047069232910871506,
0.02493254840373993,
0.03121163696050644,
-0.09514125436544418,
0.03411739692091942,
0.051425207406282425,
0.12830060720443726,
-0.026672806590795517,
0.06988455355167389,
0.010212231427431107,
0.06475206464529037,
0.076323501765728,
-0.07098226249217987,
0.23186635971069336,
0.0008730843546800315,
-0.045465003699064255,
0.005159816704690456,
0.07850778847932816,
0.04052295163273811,
-0.030105222016572952,
0.05588752403855324,
-0.06828755140304565,
0.058916185051202774,
-0.003951359074562788,
0.004792851861566305,
-0.07246880233287811,
0.08186695724725723,
-0.0076866233721375465,
0.06234586238861084,
0.07230149954557419,
-0.03018886037170887,
0.05557506904006004,
-0.010067560710012913,
0.03356749564409256,
-0.0058000897988677025,
-0.07662566006183624,
-0.044599924236536026,
-0.1444602757692337,
-0.040386538952589035,
0.04140913113951683,
-0.005101493094116449,
-0.2865702211856842,
0.018545592203736305,
-0.09611119329929352,
0.049873724579811096,
-0.0801682248711586,
0.15177197754383087,
0.16492217779159546,
0.01041433122009039,
-0.014102296903729439,
-0.15125076472759247,
0.03289223462343216,
0.08406585454940796,
-0.12416541576385498,
-0.13141396641731262
] |
null | null | transformers | # RPMerge
A merge of several Yi 34B models with a singular goal: 40K+ context, instruct-enhanced storytelling.
Disappointed with some quirks of my previous kitchen sink merges (like token/instruct formats from various models showing up when they shouldn't), I've gone 'back to the basics' and picked a few Vicuna-format only models:
- [DrNicefellow/ChatAllInOne-Yi-34B-200K-V1](https://huggingface.co/DrNicefellow/ChatAllInOne-Yi-34B-200K-V1) and [migtissera/Tess-34B-v1.5b](https://huggingface.co/migtissera/Tess-34B-v1.5b) both have excellent general instruction-following performance.
- [cgato/Thespis-34b-v0.7](https://huggingface.co/cgato/Thespis-34b-v0.7) is trained on the "Username: {Input} / BotName: {Response}" format, to emphasize it in the merge (but not force it). It also seems to work for multi-character stories.
- [Doctor-Shotgun/limarpv3-yi-llama-34b-lora](https://huggingface.co/Doctor-Shotgun/limarpv3-yi-llama-34b-lora) is trained on roleplaying data, but merged at a modest weight to not over emphasize it. This is the only non-vicuna model (being alpaca format), but it doesn't seem to interefere with the Vicuna format or adversely affect long-context perplexity
- [adamo1139/yi-34b-200k-rawrr-dpo-2](https://huggingface.co/adamo1139/yi-34b-200k-rawrr-dpo-2) the base for the limarp lora, this is base Yi gently finetuned to discourage refusals.
- [migtissera/Tess-M-Creative-v1.0](https://huggingface.co/migtissera/Tess-M-Creative-v1.0) and [NousResearch/Nous-Capybara-34B](https://huggingface.co/NousResearch/Nous-Capybara-34B) are both "undertrained" Yi models. I find they excel at raw completion performance (like long novel continuations) while still retaining some Vicuna instruct ability. This may be why some still prefer the original Tess 1.0/Capybara merge.
I consider this a more "focused" merge that previous ones. I will investigate other models (perhaps chatML models?) for a more "factual assistant" focused merge, as well as a coding-focused merge if I can't find one to suit my needs.
## Prompt template: Orca-Vicuna
```
SYSTEM: {system_message}
USER: {prompt}
ASSISTANT:
```
Raw prompting as described here is also effective: https://old.reddit.com/r/LocalLLaMA/comments/18zqy4s/the_secret_to_writing_quality_stories_with_llms/
As well as a very explicit system prompt like this: https://old.reddit.com/r/LocalLLaMA/comments/1aiz6zu/roleplaying_system_prompts/koygiwa/
## Running
Chinese models with large tokenizer vocabularies like Yi need *careful* parameter tuning due to their huge logit sampling "tails." Yi in particular also runs relatively "hot" even at lower temperatures.
I am a huge fan of Kalomaze's quadratic sampling (shown as "smoothing factor" where available), as described here: https://github.com/oobabooga/text-generation-webui/pull/5403
Otherwise, I recommend a lower temperature with 0.1 or higher MinP, a little repetition penalty, and mirostat with a low tau, and no other samplers. See the explanation here: https://github.com/ggerganov/llama.cpp/pull/3841
24GB GPUs can efficiently run Yi-34B-200K models at **40K-90K context** with exllamav2, and performant UIs like [exui](https://github.com/turboderp/exui). I go into more detail in this [post](https://old.reddit.com/r/LocalLLaMA/comments/1896igc/how_i_run_34b_models_at_75k_context_on_24gb_fast/). Empty 16GB GPUs can still run the high context with aggressive quantization.
To load/train this in full-context backends like transformers, you *must* change `max_position_embeddings` in config.json to a lower value than 200,000, otherwise you will OOM! I do not recommend running high context without context-efficient backends that support flash attention + 8 bit kv cache, like exllamav2, litellm, vllm or unsloth.
## Testing Notes
Thanks to ParasiticRogue for this idea of a Vicuna-only merge, see: https://huggingface.co/brucethemoose/jondurbin_bagel-dpo-34b-v0.2-exl2-4bpw-fiction/discussions
See: https://huggingface.co/brucethemoose/Yi-34B-200K-DARE-megamerge-v8#testing-notes
This is a possible base for a storytelling finetune/LASER in the future, once I can bite the bullet and rent some A100s or a MI300.
I have tested this merge with with novel-style continuation (but not much chat-style roleplay), and some assistant-style responses and long context analysis. I haven't seen any refusals so far.
## Merge Details
### Merge Method
This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using /home/alpha/Models/Raw/chargoddard_Yi-34B-200K-Llama as a base.
### Models Merged
The following models were included in the merge:
* /home/alpha/Models/Raw/migtissera_Tess-34B-v1.5b
* /home/alpha/Models/Raw/migtissera_Tess-M-Creative-v1.0
* /home/alpha/Models/Raw/cgato_Thespis-34b-DPO-v0.7
* /home/alpha/Models/Raw/Nous-Capybara-34B
* /home/alpha/Models/Raw/admo_limarp
* /home/alpha/Models/Raw/DrNicefellow_ChatAllInOne-Yi-34B-200K-V1
### Configuration
The following YAML configuration was used to produce this model:
```yaml
models:
- model: /home/alpha/Models/Raw/chargoddard_Yi-34B-200K-Llama
# No parameters necessary for base model
- model: /home/alpha/Models/Raw/migtissera_Tess-34B-v1.5b
#Emphasize the beginning of Vicuna format models
parameters:
weight: 0.19
density: 0.59
- model: /home/alpha/Models/Raw/Nous-Capybara-34B
parameters:
weight: 0.19
density: 0.55
# Vicuna format
- model: /home/alpha/Models/Raw/migtissera_Tess-M-Creative-v1.0
parameters:
weight: 0.05
density: 0.55
- model: /home/alpha/Models/Raw/DrNicefellow_ChatAllInOne-Yi-34B-200K-V1
parameters:
weight: 0.19
density: 0.55
- model: adamo1139/yi-34b-200k-rawrr-dpo-2+Doctor-Shotgun/limarpv3-yi-llama-34b-lora
parameters:
weight: 0.19
density: 0.48
- model: /home/alpha/Models/Raw/cgato_Thespis-34b-DPO-v0.7
parameters:
weight: 0.19
density: 0.59
merge_method: dare_ties
tokenizer_source: union
base_model: /home/alpha/Models/Raw/chargoddard_Yi-34B-200K-Llama
parameters:
int8_mask: true
dtype: bfloat16
```
## Self Promotion
I'm part of a AI startup called Holocene AI!
We're new, busy, and still setting things up. But if you have any business inquiries, want a job, or just want some consultation, feel free to shoot me an email. We have expertise in RAG applications and llama/embeddings model finetuning, and absolutely *none* of the nonsense of scammy AI startups.
Contact me at: [email protected]
I also set up a Ko-Fi! I want to run some (personal) training/LASERing as well, at 100K context or so. If you'd like to buy me 10 minutes on an A100 (or 5 seconds on an MI300X), I'd appreciate it: https://ko-fi.com/alphaatlas | {"language": ["en"], "license": "other", "library_name": "transformers", "tags": ["mergekit", "merge", "Yi", "exllama", "exllamav2", "exl2"], "license_name": "yi-license", "license_link": "https://huggingface.co/01-ai/Yi-34B/blob/main/LICENSE", "base_model": []} | text-generation | brucethemoose/Yi-34B-200K-RPMerge | [
"transformers",
"safetensors",
"llama",
"text-generation",
"mergekit",
"merge",
"Yi",
"exllama",
"exllamav2",
"exl2",
"en",
"arxiv:2311.03099",
"arxiv:2306.01708",
"license:other",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T06:14:13+00:00 | [
"2311.03099",
"2306.01708"
] | [
"en"
] | TAGS
#transformers #safetensors #llama #text-generation #mergekit #merge #Yi #exllama #exllamav2 #exl2 #en #arxiv-2311.03099 #arxiv-2306.01708 #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # RPMerge
A merge of several Yi 34B models with a singular goal: 40K+ context, instruct-enhanced storytelling.
Disappointed with some quirks of my previous kitchen sink merges (like token/instruct formats from various models showing up when they shouldn't), I've gone 'back to the basics' and picked a few Vicuna-format only models:
- DrNicefellow/ChatAllInOne-Yi-34B-200K-V1 and migtissera/Tess-34B-v1.5b both have excellent general instruction-following performance.
- cgato/Thespis-34b-v0.7 is trained on the "Username: {Input} / BotName: {Response}" format, to emphasize it in the merge (but not force it). It also seems to work for multi-character stories.
- Doctor-Shotgun/limarpv3-yi-llama-34b-lora is trained on roleplaying data, but merged at a modest weight to not over emphasize it. This is the only non-vicuna model (being alpaca format), but it doesn't seem to interefere with the Vicuna format or adversely affect long-context perplexity
- adamo1139/yi-34b-200k-rawrr-dpo-2 the base for the limarp lora, this is base Yi gently finetuned to discourage refusals.
- migtissera/Tess-M-Creative-v1.0 and NousResearch/Nous-Capybara-34B are both "undertrained" Yi models. I find they excel at raw completion performance (like long novel continuations) while still retaining some Vicuna instruct ability. This may be why some still prefer the original Tess 1.0/Capybara merge.
I consider this a more "focused" merge that previous ones. I will investigate other models (perhaps chatML models?) for a more "factual assistant" focused merge, as well as a coding-focused merge if I can't find one to suit my needs.
## Prompt template: Orca-Vicuna
Raw prompting as described here is also effective: URL
As well as a very explicit system prompt like this: URL
## Running
Chinese models with large tokenizer vocabularies like Yi need *careful* parameter tuning due to their huge logit sampling "tails." Yi in particular also runs relatively "hot" even at lower temperatures.
I am a huge fan of Kalomaze's quadratic sampling (shown as "smoothing factor" where available), as described here: URL
Otherwise, I recommend a lower temperature with 0.1 or higher MinP, a little repetition penalty, and mirostat with a low tau, and no other samplers. See the explanation here: URL
24GB GPUs can efficiently run Yi-34B-200K models at 40K-90K context with exllamav2, and performant UIs like exui. I go into more detail in this post. Empty 16GB GPUs can still run the high context with aggressive quantization.
To load/train this in full-context backends like transformers, you *must* change 'max_position_embeddings' in URL to a lower value than 200,000, otherwise you will OOM! I do not recommend running high context without context-efficient backends that support flash attention + 8 bit kv cache, like exllamav2, litellm, vllm or unsloth.
## Testing Notes
Thanks to ParasiticRogue for this idea of a Vicuna-only merge, see: URL
See: URL
This is a possible base for a storytelling finetune/LASER in the future, once I can bite the bullet and rent some A100s or a MI300.
I have tested this merge with with novel-style continuation (but not much chat-style roleplay), and some assistant-style responses and long context analysis. I haven't seen any refusals so far.
## Merge Details
### Merge Method
This model was merged using the DARE TIES merge method using /home/alpha/Models/Raw/chargoddard_Yi-34B-200K-Llama as a base.
### Models Merged
The following models were included in the merge:
* /home/alpha/Models/Raw/migtissera_Tess-34B-v1.5b
* /home/alpha/Models/Raw/migtissera_Tess-M-Creative-v1.0
* /home/alpha/Models/Raw/cgato_Thespis-34b-DPO-v0.7
* /home/alpha/Models/Raw/Nous-Capybara-34B
* /home/alpha/Models/Raw/admo_limarp
* /home/alpha/Models/Raw/DrNicefellow_ChatAllInOne-Yi-34B-200K-V1
### Configuration
The following YAML configuration was used to produce this model:
## Self Promotion
I'm part of a AI startup called Holocene AI!
We're new, busy, and still setting things up. But if you have any business inquiries, want a job, or just want some consultation, feel free to shoot me an email. We have expertise in RAG applications and llama/embeddings model finetuning, and absolutely *none* of the nonsense of scammy AI startups.
Contact me at: URL@URL
I also set up a Ko-Fi! I want to run some (personal) training/LASERing as well, at 100K context or so. If you'd like to buy me 10 minutes on an A100 (or 5 seconds on an MI300X), I'd appreciate it: URL | [
"# RPMerge\nA merge of several Yi 34B models with a singular goal: 40K+ context, instruct-enhanced storytelling.\n\nDisappointed with some quirks of my previous kitchen sink merges (like token/instruct formats from various models showing up when they shouldn't), I've gone 'back to the basics' and picked a few Vicuna-format only models:\n\n- DrNicefellow/ChatAllInOne-Yi-34B-200K-V1 and migtissera/Tess-34B-v1.5b both have excellent general instruction-following performance.\n\n- cgato/Thespis-34b-v0.7 is trained on the \"Username: {Input} / BotName: {Response}\" format, to emphasize it in the merge (but not force it). It also seems to work for multi-character stories.\n\n- Doctor-Shotgun/limarpv3-yi-llama-34b-lora is trained on roleplaying data, but merged at a modest weight to not over emphasize it. This is the only non-vicuna model (being alpaca format), but it doesn't seem to interefere with the Vicuna format or adversely affect long-context perplexity\n\n- adamo1139/yi-34b-200k-rawrr-dpo-2 the base for the limarp lora, this is base Yi gently finetuned to discourage refusals.\n\n- migtissera/Tess-M-Creative-v1.0 and NousResearch/Nous-Capybara-34B are both \"undertrained\" Yi models. I find they excel at raw completion performance (like long novel continuations) while still retaining some Vicuna instruct ability. This may be why some still prefer the original Tess 1.0/Capybara merge.\n\nI consider this a more \"focused\" merge that previous ones. I will investigate other models (perhaps chatML models?) for a more \"factual assistant\" focused merge, as well as a coding-focused merge if I can't find one to suit my needs.",
"## Prompt template: Orca-Vicuna\n\nRaw prompting as described here is also effective: URL\n\nAs well as a very explicit system prompt like this: URL",
"## Running\n\nChinese models with large tokenizer vocabularies like Yi need *careful* parameter tuning due to their huge logit sampling \"tails.\" Yi in particular also runs relatively \"hot\" even at lower temperatures.\n\nI am a huge fan of Kalomaze's quadratic sampling (shown as \"smoothing factor\" where available), as described here: URL\n\nOtherwise, I recommend a lower temperature with 0.1 or higher MinP, a little repetition penalty, and mirostat with a low tau, and no other samplers. See the explanation here: URL\n\n24GB GPUs can efficiently run Yi-34B-200K models at 40K-90K context with exllamav2, and performant UIs like exui. I go into more detail in this post. Empty 16GB GPUs can still run the high context with aggressive quantization.\n\nTo load/train this in full-context backends like transformers, you *must* change 'max_position_embeddings' in URL to a lower value than 200,000, otherwise you will OOM! I do not recommend running high context without context-efficient backends that support flash attention + 8 bit kv cache, like exllamav2, litellm, vllm or unsloth.",
"## Testing Notes\n\nThanks to ParasiticRogue for this idea of a Vicuna-only merge, see: URL\n\nSee: URL\n\nThis is a possible base for a storytelling finetune/LASER in the future, once I can bite the bullet and rent some A100s or a MI300. \n\nI have tested this merge with with novel-style continuation (but not much chat-style roleplay), and some assistant-style responses and long context analysis. I haven't seen any refusals so far.",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the DARE TIES merge method using /home/alpha/Models/Raw/chargoddard_Yi-34B-200K-Llama as a base.",
"### Models Merged\n\nThe following models were included in the merge:\n* /home/alpha/Models/Raw/migtissera_Tess-34B-v1.5b\n* /home/alpha/Models/Raw/migtissera_Tess-M-Creative-v1.0\n* /home/alpha/Models/Raw/cgato_Thespis-34b-DPO-v0.7\n* /home/alpha/Models/Raw/Nous-Capybara-34B\n* /home/alpha/Models/Raw/admo_limarp\n* /home/alpha/Models/Raw/DrNicefellow_ChatAllInOne-Yi-34B-200K-V1",
"### Configuration\n\nThe following YAML configuration was used to produce this model:",
"## Self Promotion\n\nI'm part of a AI startup called Holocene AI!\n\nWe're new, busy, and still setting things up. But if you have any business inquiries, want a job, or just want some consultation, feel free to shoot me an email. We have expertise in RAG applications and llama/embeddings model finetuning, and absolutely *none* of the nonsense of scammy AI startups.\n\nContact me at: URL@URL\n\nI also set up a Ko-Fi! I want to run some (personal) training/LASERing as well, at 100K context or so. If you'd like to buy me 10 minutes on an A100 (or 5 seconds on an MI300X), I'd appreciate it: URL"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #mergekit #merge #Yi #exllama #exllamav2 #exl2 #en #arxiv-2311.03099 #arxiv-2306.01708 #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# RPMerge\nA merge of several Yi 34B models with a singular goal: 40K+ context, instruct-enhanced storytelling.\n\nDisappointed with some quirks of my previous kitchen sink merges (like token/instruct formats from various models showing up when they shouldn't), I've gone 'back to the basics' and picked a few Vicuna-format only models:\n\n- DrNicefellow/ChatAllInOne-Yi-34B-200K-V1 and migtissera/Tess-34B-v1.5b both have excellent general instruction-following performance.\n\n- cgato/Thespis-34b-v0.7 is trained on the \"Username: {Input} / BotName: {Response}\" format, to emphasize it in the merge (but not force it). It also seems to work for multi-character stories.\n\n- Doctor-Shotgun/limarpv3-yi-llama-34b-lora is trained on roleplaying data, but merged at a modest weight to not over emphasize it. This is the only non-vicuna model (being alpaca format), but it doesn't seem to interefere with the Vicuna format or adversely affect long-context perplexity\n\n- adamo1139/yi-34b-200k-rawrr-dpo-2 the base for the limarp lora, this is base Yi gently finetuned to discourage refusals.\n\n- migtissera/Tess-M-Creative-v1.0 and NousResearch/Nous-Capybara-34B are both \"undertrained\" Yi models. I find they excel at raw completion performance (like long novel continuations) while still retaining some Vicuna instruct ability. This may be why some still prefer the original Tess 1.0/Capybara merge.\n\nI consider this a more \"focused\" merge that previous ones. I will investigate other models (perhaps chatML models?) for a more \"factual assistant\" focused merge, as well as a coding-focused merge if I can't find one to suit my needs.",
"## Prompt template: Orca-Vicuna\n\nRaw prompting as described here is also effective: URL\n\nAs well as a very explicit system prompt like this: URL",
"## Running\n\nChinese models with large tokenizer vocabularies like Yi need *careful* parameter tuning due to their huge logit sampling \"tails.\" Yi in particular also runs relatively \"hot\" even at lower temperatures.\n\nI am a huge fan of Kalomaze's quadratic sampling (shown as \"smoothing factor\" where available), as described here: URL\n\nOtherwise, I recommend a lower temperature with 0.1 or higher MinP, a little repetition penalty, and mirostat with a low tau, and no other samplers. See the explanation here: URL\n\n24GB GPUs can efficiently run Yi-34B-200K models at 40K-90K context with exllamav2, and performant UIs like exui. I go into more detail in this post. Empty 16GB GPUs can still run the high context with aggressive quantization.\n\nTo load/train this in full-context backends like transformers, you *must* change 'max_position_embeddings' in URL to a lower value than 200,000, otherwise you will OOM! I do not recommend running high context without context-efficient backends that support flash attention + 8 bit kv cache, like exllamav2, litellm, vllm or unsloth.",
"## Testing Notes\n\nThanks to ParasiticRogue for this idea of a Vicuna-only merge, see: URL\n\nSee: URL\n\nThis is a possible base for a storytelling finetune/LASER in the future, once I can bite the bullet and rent some A100s or a MI300. \n\nI have tested this merge with with novel-style continuation (but not much chat-style roleplay), and some assistant-style responses and long context analysis. I haven't seen any refusals so far.",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the DARE TIES merge method using /home/alpha/Models/Raw/chargoddard_Yi-34B-200K-Llama as a base.",
"### Models Merged\n\nThe following models were included in the merge:\n* /home/alpha/Models/Raw/migtissera_Tess-34B-v1.5b\n* /home/alpha/Models/Raw/migtissera_Tess-M-Creative-v1.0\n* /home/alpha/Models/Raw/cgato_Thespis-34b-DPO-v0.7\n* /home/alpha/Models/Raw/Nous-Capybara-34B\n* /home/alpha/Models/Raw/admo_limarp\n* /home/alpha/Models/Raw/DrNicefellow_ChatAllInOne-Yi-34B-200K-V1",
"### Configuration\n\nThe following YAML configuration was used to produce this model:",
"## Self Promotion\n\nI'm part of a AI startup called Holocene AI!\n\nWe're new, busy, and still setting things up. But if you have any business inquiries, want a job, or just want some consultation, feel free to shoot me an email. We have expertise in RAG applications and llama/embeddings model finetuning, and absolutely *none* of the nonsense of scammy AI startups.\n\nContact me at: URL@URL\n\nI also set up a Ko-Fi! I want to run some (personal) training/LASERing as well, at 100K context or so. If you'd like to buy me 10 minutes on an A100 (or 5 seconds on an MI300X), I'd appreciate it: URL"
] | [
95,
478,
35,
284,
113,
4,
49,
169,
17,
163
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #mergekit #merge #Yi #exllama #exllamav2 #exl2 #en #arxiv-2311.03099 #arxiv-2306.01708 #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"passage: # RPMerge\nA merge of several Yi 34B models with a singular goal: 40K+ context, instruct-enhanced storytelling.\n\nDisappointed with some quirks of my previous kitchen sink merges (like token/instruct formats from various models showing up when they shouldn't), I've gone 'back to the basics' and picked a few Vicuna-format only models:\n\n- DrNicefellow/ChatAllInOne-Yi-34B-200K-V1 and migtissera/Tess-34B-v1.5b both have excellent general instruction-following performance.\n\n- cgato/Thespis-34b-v0.7 is trained on the \"Username: {Input} / BotName: {Response}\" format, to emphasize it in the merge (but not force it). It also seems to work for multi-character stories.\n\n- Doctor-Shotgun/limarpv3-yi-llama-34b-lora is trained on roleplaying data, but merged at a modest weight to not over emphasize it. This is the only non-vicuna model (being alpaca format), but it doesn't seem to interefere with the Vicuna format or adversely affect long-context perplexity\n\n- adamo1139/yi-34b-200k-rawrr-dpo-2 the base for the limarp lora, this is base Yi gently finetuned to discourage refusals.\n\n- migtissera/Tess-M-Creative-v1.0 and NousResearch/Nous-Capybara-34B are both \"undertrained\" Yi models. I find they excel at raw completion performance (like long novel continuations) while still retaining some Vicuna instruct ability. This may be why some still prefer the original Tess 1.0/Capybara merge.\n\nI consider this a more \"focused\" merge that previous ones. I will investigate other models (perhaps chatML models?) for a more \"factual assistant\" focused merge, as well as a coding-focused merge if I can't find one to suit my needs.## Prompt template: Orca-Vicuna\n\nRaw prompting as described here is also effective: URL\n\nAs well as a very explicit system prompt like this: URL## Running\n\nChinese models with large tokenizer vocabularies like Yi need *careful* parameter tuning due to their huge logit sampling \"tails.\" Yi in particular also runs relatively \"hot\" even at lower temperatures.\n\nI am a huge fan of Kalomaze's quadratic sampling (shown as \"smoothing factor\" where available), as described here: URL\n\nOtherwise, I recommend a lower temperature with 0.1 or higher MinP, a little repetition penalty, and mirostat with a low tau, and no other samplers. See the explanation here: URL\n\n24GB GPUs can efficiently run Yi-34B-200K models at 40K-90K context with exllamav2, and performant UIs like exui. I go into more detail in this post. Empty 16GB GPUs can still run the high context with aggressive quantization.\n\nTo load/train this in full-context backends like transformers, you *must* change 'max_position_embeddings' in URL to a lower value than 200,000, otherwise you will OOM! I do not recommend running high context without context-efficient backends that support flash attention + 8 bit kv cache, like exllamav2, litellm, vllm or unsloth.## Testing Notes\n\nThanks to ParasiticRogue for this idea of a Vicuna-only merge, see: URL\n\nSee: URL\n\nThis is a possible base for a storytelling finetune/LASER in the future, once I can bite the bullet and rent some A100s or a MI300. \n\nI have tested this merge with with novel-style continuation (but not much chat-style roleplay), and some assistant-style responses and long context analysis. I haven't seen any refusals so far.## Merge Details### Merge Method\n\nThis model was merged using the DARE TIES merge method using /home/alpha/Models/Raw/chargoddard_Yi-34B-200K-Llama as a base."
] | [
-0.0657280832529068,
-0.03883764147758484,
-0.005031414330005646,
0.03941553458571434,
0.08108654618263245,
0.01211583986878395,
0.1405470073223114,
0.09520211815834045,
-0.011399326846003532,
0.05437276512384415,
0.04094007983803749,
0.07741039991378784,
0.037948913872241974,
0.07974492758512497,
-0.04866745322942734,
-0.13017478585243225,
0.05223392695188522,
-0.018392367288470268,
-0.0001989291631616652,
0.07022418081760406,
0.10999098420143127,
-0.031434014439582825,
0.07722215354442596,
0.0064000519923865795,
-0.09814302623271942,
0.03028661385178566,
0.03687087446451187,
-0.05044526606798172,
0.10735408216714859,
0.11100101470947266,
0.0905713438987732,
0.02822965756058693,
-0.013981180265545845,
-0.15055899322032928,
0.023344827815890312,
0.03601795807480812,
-0.029659118503332138,
0.03170293942093849,
0.07471983134746552,
-0.003429476171731949,
0.06001509726047516,
-0.0813867375254631,
0.000631923321634531,
0.06043754145503044,
-0.11217561364173889,
-0.09124553203582764,
-0.11864979565143585,
0.05379253253340721,
0.09593994170427322,
0.07166111469268799,
-0.014185975305736065,
0.09829781949520111,
-0.0025735031813383102,
0.07770922034978867,
0.07673341035842896,
-0.28871282935142517,
-0.0017346586100757122,
0.06336577981710434,
0.06431692838668823,
-0.015232766047120094,
-0.018899692222476006,
0.03282584622502327,
0.05834983289241791,
-0.019440408796072006,
-0.031616322696208954,
-0.05121326446533203,
0.03579029440879822,
-0.03766779601573944,
-0.06073032319545746,
-0.0050857774913311005,
0.18253330886363983,
0.02199225127696991,
-0.024901170283555984,
-0.08430474251508713,
-0.04446796700358391,
0.005133826285600662,
0.00980633869767189,
-0.0403251051902771,
0.010041328147053719,
0.026971226558089256,
0.08896796405315399,
-0.03390217572450638,
-0.1154056116938591,
0.022528281435370445,
-0.14025329053401947,
0.11859819293022156,
0.014815760776400566,
0.02348189987242222,
-0.061485469341278076,
0.03380409628152847,
-0.037810251116752625,
-0.08981078118085861,
-0.00617143465206027,
-0.030898217111825943,
-0.05352133885025978,
-0.026834644377231598,
-0.06307955086231232,
-0.10260657966136932,
0.04188771918416023,
0.15055236220359802,
-0.047684457153081894,
0.023173881694674492,
-0.03915662690997124,
0.0637236088514328,
0.00929923728108406,
0.0019042976200580597,
-0.02921391651034355,
-0.06958888471126556,
0.06611696630716324,
-0.02303151786327362,
0.06347496807575226,
-0.012448059394955635,
-0.09704916924238205,
-0.02142578735947609,
-0.024604719132184982,
0.059139564633369446,
0.02804766222834587,
0.05573972314596176,
-0.06314535439014435,
0.004633713513612747,
0.10203363746404648,
-0.11167159676551819,
0.046014636754989624,
-0.02713393047451973,
0.02219206839799881,
0.08065091818571091,
0.03941811993718147,
0.028574395924806595,
-0.054419051855802536,
0.065043143928051,
-0.049569278955459595,
-0.010006502270698547,
-0.046763986349105835,
-0.08117979764938354,
0.0675256997346878,
-0.015655145049095154,
-0.01850743032991886,
-0.15776079893112183,
-0.13715451955795288,
-0.038685545325279236,
0.008957135491073132,
-0.04781130701303482,
-0.035338763147592545,
-0.027605805546045303,
-0.032228223979473114,
0.01265384629368782,
-0.018393974751234055,
-0.0017576580867171288,
-0.028948042541742325,
0.01336119044572115,
-0.01967746578156948,
0.055247753858566284,
-0.0829746425151825,
0.02032327651977539,
-0.0707329660654068,
0.039823513478040695,
-0.0743815153837204,
0.08473104238510132,
-0.04316629841923714,
0.041617538779973984,
-0.08055606484413147,
0.02523268386721611,
-0.020383374765515327,
0.05298575013875961,
0.011566353030502796,
0.11714065819978714,
-0.1492558717727661,
-0.060131289064884186,
0.14389604330062866,
-0.1455298364162445,
-0.12742015719413757,
0.1103188768029213,
0.004267839714884758,
0.030449040234088898,
0.08833916485309601,
0.1296926885843277,
0.06646063923835754,
-0.07971236109733582,
-0.06335125863552094,
0.001223372295498848,
-0.08615433424711227,
-0.012586034834384918,
0.053809136152267456,
0.011805062182247639,
-0.015522070229053497,
0.012753700837492943,
0.04241078346967697,
0.040840741246938705,
0.005312885157763958,
-0.05013526603579521,
-0.046981845051050186,
0.0037695933133363724,
-0.014357809908688068,
-0.02479720488190651,
0.0020015714690089226,
-0.0731881707906723,
-0.07266317307949066,
-0.01890711486339569,
0.05987139791250229,
0.0018450133502483368,
0.04860319197177887,
-0.0810045525431633,
0.07580879330635071,
-0.053480975329875946,
0.029483329504728317,
-0.1278400421142578,
-0.08281981945037842,
-0.01352971512824297,
-0.0374082550406456,
0.021694928407669067,
0.1141333281993866,
0.031097371131181717,
0.019829565659165382,
-0.03412093222141266,
-0.0031103226356208324,
0.0919569805264473,
0.011719141155481339,
-0.037552498281002045,
-0.13740158081054688,
0.006156686693429947,
-0.06311877816915512,
0.12366893887519836,
-0.0939408391714096,
0.015294782817363739,
0.04204939305782318,
0.10396246612071991,
-0.008535603992640972,
0.030363380908966064,
-0.02875475212931633,
0.03441561758518219,
-0.05118439346551895,
-0.003483101725578308,
0.05628623068332672,
0.017650699242949486,
-0.11507237702608109,
0.16090981662273407,
-0.13627660274505615,
0.048125021159648895,
0.1134660467505455,
-0.0411580391228199,
-0.06378835439682007,
-0.05231699347496033,
-0.007725239265710115,
-0.03733376786112785,
-0.01214087475091219,
-0.07445776462554932,
0.1207064539194107,
0.044721316546201706,
0.14061316847801208,
-0.04791876673698425,
-0.02019791305065155,
-0.0040846955962479115,
-0.05833081901073456,
-0.04336227476596832,
0.0789361298084259,
0.027516745030879974,
-0.18764980137348175,
0.10679424554109573,
0.13242389261722565,
0.030136337503790855,
0.09338249266147614,
-0.002161476295441389,
-0.03071211464703083,
-0.038854267448186874,
0.01771850883960724,
0.038829367607831955,
-0.046296946704387665,
-0.007829602807760239,
0.030910536646842957,
0.06013588607311249,
0.02263709157705307,
0.02494082599878311,
-0.0999964028596878,
-0.009062595665454865,
0.029500802978873253,
-0.009510510601103306,
0.03353263437747955,
0.062402721494436264,
0.022662142291665077,
0.1080050840973854,
-0.0016961395740509033,
0.005057517439126968,
0.04784578084945679,
-0.027587447315454483,
-0.1093234121799469,
0.13938620686531067,
-0.10654670000076294,
-0.26909613609313965,
-0.185601145029068,
-0.08505947142839432,
-0.048180144280195236,
-0.007993631064891815,
0.0746399462223053,
-0.05848109349608421,
-0.06345906108617783,
-0.06997229158878326,
0.05575196444988251,
-0.010834001004695892,
-0.052163511514663696,
0.003468367038294673,
0.072602778673172,
0.015544583089649677,
-0.09406086802482605,
-0.03407193720340729,
0.01858535036444664,
-0.05503252148628235,
0.07312961667776108,
-0.007379714399576187,
0.09137079864740372,
0.08628709614276886,
0.013123871758580208,
-0.029227014631032944,
-0.04072929173707962,
0.08942665159702301,
-0.06644243001937866,
0.06160276010632515,
0.16159430146217346,
-0.040480948984622955,
0.06916072219610214,
0.1579800695180893,
0.011893810704350471,
-0.0614432655274868,
0.014844642952084541,
0.008633187040686607,
-0.034799445420503616,
-0.2269350290298462,
-0.10152453184127808,
-0.08414267003536224,
0.008608102798461914,
0.016536971554160118,
0.043330300599336624,
0.07278396934270859,
0.0330992117524147,
-0.04249997437000275,
0.04412053897976875,
0.01070931926369667,
0.07258906215429306,
0.244166761636734,
-0.016548264771699905,
0.10595281422138214,
-0.0692313089966774,
-0.03007199801504612,
0.08779381215572357,
0.04356330260634422,
0.17105227708816528,
0.053127966821193695,
0.10980008542537689,
0.07147233933210373,
-0.0002450086176395416,
0.05487074702978134,
0.04210598021745682,
0.02302979677915573,
-0.018233517184853554,
-0.03214295953512192,
-0.04975098371505737,
-0.00974734965711832,
0.0804024338722229,
-0.035794392228126526,
-0.009475989267230034,
-0.03443822264671326,
0.029325082898139954,
0.05771588906645775,
0.12056946754455566,
0.04115460440516472,
-0.17829644680023193,
-0.0222166758030653,
0.05941052734851837,
-0.02254975214600563,
-0.039125002920627594,
0.0010068845003843307,
0.0468110665678978,
-0.07152120769023895,
0.12132088840007782,
-0.03178262338042259,
0.0752815455198288,
-0.054400235414505005,
0.028037432581186295,
-0.030532345175743103,
0.03919224068522453,
0.02187945321202278,
0.07596062123775482,
-0.19628244638442993,
0.19378739595413208,
0.008642751723527908,
0.020173601806163788,
-0.031041111797094345,
0.04095974192023277,
0.006184851285070181,
0.0866684690117836,
0.12100359052419662,
0.005744677037000656,
-0.05385257303714752,
-0.06263434141874313,
-0.05917435511946678,
0.01735786907374859,
0.05682459473609924,
0.015186388045549393,
0.048723869025707245,
-0.05814337357878685,
-0.014166401699185371,
-0.013688569888472557,
0.05666462704539299,
-0.06636025756597519,
-0.18094876408576965,
0.08117005974054337,
0.05203717201948166,
-0.008133228868246078,
-0.04640316218137741,
-0.025232873857021332,
-0.09239903092384338,
0.22749364376068115,
-0.05269606411457062,
-0.05593772232532501,
-0.1077461689710617,
0.009648095816373825,
0.08498600125312805,
-0.07025494426488876,
0.018909797072410583,
-0.029952198266983032,
0.0708247646689415,
-0.0947951227426529,
-0.09568969160318375,
0.054658837616443634,
-0.08118951320648193,
-0.07906282693147659,
-0.0433584600687027,
0.12085403501987457,
-0.06536781042814255,
0.02217540331184864,
0.027056757360696793,
0.04360917583107948,
-0.02287120558321476,
-0.07300559431314468,
0.007690622005611658,
0.06173691526055336,
0.010153927840292454,
0.028257524594664574,
-0.08928604423999786,
-0.09663023054599762,
-0.0382804349064827,
-0.08837516605854034,
0.1702013909816742,
0.3132404685020447,
-0.04544273763895035,
0.10748850554227829,
0.10053829848766327,
-0.09447507560253143,
-0.20322853326797485,
-0.07519013434648514,
-0.028356753289699554,
0.009750918485224247,
-0.0042197974398732185,
-0.07098803669214249,
0.08334368467330933,
0.11546652019023895,
0.0002790875732898712,
0.04947932809591293,
-0.2035357654094696,
-0.11433006078004837,
0.012129522860050201,
-0.004426450468599796,
0.17143988609313965,
-0.13481175899505615,
-0.07242738455533981,
-0.07681001722812653,
-0.09475941956043243,
0.028162313625216484,
0.00009749270975589752,
0.08256494998931885,
-0.0029606204479932785,
0.014319101348519325,
0.024820396676659584,
-0.03594513610005379,
0.1266268640756607,
-0.019459838047623634,
0.040257424116134644,
-0.10071644932031631,
-0.03224731609225273,
0.0757429301738739,
-0.027076777070760727,
0.07807588577270508,
-0.0588790588080883,
-0.0147643331438303,
-0.07716332376003265,
-0.031885311007499695,
-0.038830216974020004,
0.05330120027065277,
-0.022442547604441643,
-0.028773656114935875,
-0.051884956657886505,
0.04810009524226189,
0.03597975894808769,
-0.023731952533125877,
0.11069245636463165,
-0.04521586745977402,
0.0515899732708931,
0.08004868775606155,
0.1286115050315857,
-0.09408014267683029,
-0.02977311611175537,
-0.005204071290791035,
-0.05967184901237488,
0.03385739028453827,
-0.09391520917415619,
0.02618393674492836,
0.08847029507160187,
-0.031126901507377625,
0.09163732826709747,
0.04101702198386192,
-0.04306039214134216,
0.016810962930321693,
0.12569023668766022,
-0.09571833908557892,
-0.08588802814483643,
0.004067686386406422,
0.04861361160874367,
-0.008297178894281387,
-0.024618685245513916,
0.17703309655189514,
-0.03405441343784332,
0.001533351605758071,
0.017706595361232758,
0.03545655682682991,
-0.0511847659945488,
0.0633472353219986,
-0.006050636991858482,
0.027304954826831818,
-0.07668432593345642,
0.06911495327949524,
0.06711360812187195,
-0.13031592965126038,
0.03138890862464905,
0.09494608640670776,
-0.08627772331237793,
-0.09651067852973938,
-0.09330902993679047,
0.12220688164234161,
-0.07118300348520279,
-0.06298600882291794,
-0.029578208923339844,
-0.1187816858291626,
0.05446537584066391,
0.06606360524892807,
0.01786470226943493,
0.03970285505056381,
-0.0038938242942094803,
-0.04486553743481636,
-0.07253476977348328,
0.05445398390293121,
0.016302140429615974,
0.05011041462421417,
-0.07588761299848557,
0.07147583365440369,
-0.007602431811392307,
0.040669478476047516,
-0.024723919108510017,
0.014727730304002762,
-0.09228154271841049,
-0.01287643238902092,
-0.099532350897789,
0.003310559783130884,
-0.08532389998435974,
-0.03085843101143837,
0.013668160885572433,
-0.004751814529299736,
-0.023864924907684326,
0.0025867698714137077,
-0.04348625987768173,
-0.04362434893846512,
-0.020601270720362663,
0.06365004181861877,
-0.09664998203516006,
-0.04308553412556648,
0.05041459947824478,
-0.050062499940395355,
0.06291035562753677,
0.003201656974852085,
-0.024297727271914482,
-0.018750939518213272,
-0.16632132232189178,
-0.0284806489944458,
0.07317766547203064,
0.04787494242191315,
0.02285153977572918,
-0.100013867020607,
0.012889590114355087,
0.055097781121730804,
-0.027065731585025787,
0.02087220549583435,
0.007822365500032902,
-0.08864182978868484,
0.04062385484576225,
-0.03881055489182472,
-0.08156450092792511,
-0.06934794783592224,
-0.013711385428905487,
0.08563999831676483,
0.017630118876695633,
0.14435791969299316,
-0.07137356698513031,
0.06438804417848587,
-0.10845257341861725,
-0.002964867278933525,
0.0019574209582060575,
-0.07187637686729431,
-0.002127423882484436,
-0.09576490521430969,
0.015230609104037285,
-0.0013167544966563582,
0.13867829740047455,
0.0016901481430977583,
-0.053925249725580215,
0.06905888020992279,
-0.0338028147816658,
-0.05296826735138893,
0.008728498592972755,
0.1083044484257698,
0.057108379900455475,
-0.0028216904029250145,
-0.06892446428537369,
0.0024785641580820084,
0.006878012791275978,
-0.045051246881484985,
0.07186750322580338,
0.09244952350854874,
0.027843652293086052,
0.07072296738624573,
0.06569163501262665,
-0.0547233521938324,
-0.039593860507011414,
-0.022033218294382095,
-0.08052010834217072,
0.05015789717435837,
-0.005509848706424236,
0.08118937909603119,
0.12901687622070312,
-0.0845879539847374,
0.07196298241615295,
-0.0244684349745512,
-0.030002731829881668,
-0.07376053929328918,
-0.10548122227191925,
-0.08416025340557098,
-0.04481502249836922,
0.0028209351003170013,
-0.09281622618436813,
0.04092538356781006,
0.022494764998555183,
0.041071467101573944,
-0.010021115653216839,
0.0757955014705658,
-0.05142072215676308,
-0.05178741365671158,
0.044263988733291626,
-0.012002105824649334,
0.03189172223210335,
0.0007363744080066681,
-0.0031508542597293854,
-0.003191964700818062,
0.002954348921775818,
-0.0007225610315799713,
0.07763482630252838,
0.059188373386859894,
0.03573131561279297,
-0.0834631398320198,
-0.07860622555017471,
-0.02141931839287281,
0.0373956561088562,
0.030494384467601776,
0.09456012398004532,
0.04753590002655983,
-0.007341736927628517,
0.014399126172065735,
0.11285136640071869,
-0.011680420488119125,
-0.10551953315734863,
-0.08822522312402725,
0.1045035570859909,
0.011006576009094715,
0.05368790030479431,
-0.033605847507715225,
-0.05680819973349571,
0.011490819975733757,
0.20236489176750183,
0.21078136563301086,
-0.024857042357325554,
0.004568947479128838,
0.01236538402736187,
0.019208485260605812,
0.019741477444767952,
0.09175281226634979,
0.05688662827014923,
0.17234379053115845,
-0.06257987022399902,
0.04585426300764084,
-0.048901960253715515,
-0.00971400085836649,
-0.08482451736927032,
0.08250986784696579,
-0.04413977265357971,
-0.00866202637553215,
-0.0012111756950616837,
0.07304698973894119,
-0.009389229118824005,
-0.07007036358118057,
0.04629310593008995,
-0.11324691027402878,
-0.07670805603265762,
0.0015313923358917236,
0.10390856117010117,
-0.00026108045130968094,
0.06917092204093933,
-0.010289308615028858,
-0.03716389834880829,
0.13385838270187378,
-0.016315503045916557,
-0.10656527429819107,
-0.007231825962662697,
0.028466317802667618,
-0.054986342787742615,
0.03469110652804375,
-0.00012579606845974922,
0.07431881874799728,
0.09958221763372421,
0.026065707206726074,
-0.09610151499509811,
0.07783028483390808,
0.0045998310670256615,
-0.048430632799863815,
0.07149460911750793,
0.03999185934662819,
0.0016619730740785599,
0.049156054854393005,
0.07577541470527649,
-0.09741003811359406,
0.0597427636384964,
0.06995432078838348,
-0.03877744451165199,
-0.0444524809718132,
0.09350777417421341,
-0.06660784780979156,
0.12018497288227081,
0.13278643786907196,
-0.011615656316280365,
-0.007087804842740297,
-0.04651836305856705,
0.0016149003058671951,
0.03558414429426193,
0.03194265812635422,
-0.034115564078092575,
-0.11102356016635895,
0.008872727863490582,
0.038696445524692535,
0.07788583636283875,
-0.20317472517490387,
-0.0828571543097496,
-0.0677414983510971,
-0.031241392716765404,
-0.08241366595029831,
0.08037864416837692,
0.10759099572896957,
0.01854255050420761,
-0.015980437397956848,
-0.16771554946899414,
0.05563141033053398,
0.08061409741640091,
-0.06977464258670807,
-0.05538628250360489
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# mistral-7b-ocn-v1
This model is a fine-tuned version of mistralai/Mistral-7B-v0.1 on the oncc_instruct dataset.
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0005
- train_batch_size: 4
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_steps: 10
- num_epochs: 1.0
- mixed_precision_training: Native AMP
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.1.1+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
### Performance
hf (pretrained=mistralai/Mistral-7B-v0.1,parallelize=True,load_in_4bit=True,peft=chenhugging/mistral-7b-ocn-v1), gen_kwargs: (None), limit: 100.0, num_fewshot: None, batch_size: 1
| Tasks |Version|Filter|n-shot| Metric |Value| |Stderr|
|---------------------|-------|------|-----:|--------|----:|---|-----:|
|pubmedqa | 1|none | 0|acc | 0.98|± |0.0141|
|medmcqa |Yaml |none | 0|acc | 0.36|± |0.0482|
|professional_medicine| 0|none | 0|acc | 0.57|± |0.0498|
|college_medicine | 0|none | 0|acc | 0.54|± |0.0501|
|clinical_knowledge | 0|none | 0|acc | 0.62|± |0.0488|
|aocnp |Yaml |none | 0|acc | 0.44|± |0.0499|
|ocn |Yaml |none | 0|acc | 0.54|± |0.0501|
| {"license": "other", "library_name": "peft", "tags": ["llama-factory", "lora", "generated_from_trainer"], "base_model": "mistralai/Mistral-7B-v0.1", "model-index": [{"name": "mistral-7b-ocn-v1", "results": []}]} | null | chenhugging/mistral-7b-ocn-v1 | [
"peft",
"safetensors",
"llama-factory",
"lora",
"generated_from_trainer",
"base_model:mistralai/Mistral-7B-v0.1",
"license:other",
"region:us"
] | 2024-02-07T06:17:58+00:00 | [] | [] | TAGS
#peft #safetensors #llama-factory #lora #generated_from_trainer #base_model-mistralai/Mistral-7B-v0.1 #license-other #region-us
| mistral-7b-ocn-v1
=================
This model is a fine-tuned version of mistralai/Mistral-7B-v0.1 on the oncc\_instruct dataset.
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0005
* train\_batch\_size: 4
* eval\_batch\_size: 8
* seed: 42
* gradient\_accumulation\_steps: 4
* total\_train\_batch\_size: 16
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: cosine
* lr\_scheduler\_warmup\_steps: 10
* num\_epochs: 1.0
* mixed\_precision\_training: Native AMP
### Framework versions
* PEFT 0.8.2
* Transformers 4.37.2
* Pytorch 2.1.1+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
### Performance
hf (pretrained=mistralai/Mistral-7B-v0.1,parallelize=True,load\_in\_4bit=True,peft=chenhugging/mistral-7b-ocn-v1), gen\_kwargs: (None), limit: 100.0, num\_fewshot: None, batch\_size: 1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 1.0\n* mixed\\_precision\\_training: Native AMP",
"### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1",
"### Performance\n\n\nhf (pretrained=mistralai/Mistral-7B-v0.1,parallelize=True,load\\_in\\_4bit=True,peft=chenhugging/mistral-7b-ocn-v1), gen\\_kwargs: (None), limit: 100.0, num\\_fewshot: None, batch\\_size: 1"
] | [
"TAGS\n#peft #safetensors #llama-factory #lora #generated_from_trainer #base_model-mistralai/Mistral-7B-v0.1 #license-other #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 1.0\n* mixed\\_precision\\_training: Native AMP",
"### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1",
"### Performance\n\n\nhf (pretrained=mistralai/Mistral-7B-v0.1,parallelize=True,load\\_in\\_4bit=True,peft=chenhugging/mistral-7b-ocn-v1), gen\\_kwargs: (None), limit: 100.0, num\\_fewshot: None, batch\\_size: 1"
] | [
51,
159,
39,
94
] | [
"passage: TAGS\n#peft #safetensors #llama-factory #lora #generated_from_trainer #base_model-mistralai/Mistral-7B-v0.1 #license-other #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 1.0\n* mixed\\_precision\\_training: Native AMP### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1### Performance\n\n\nhf (pretrained=mistralai/Mistral-7B-v0.1,parallelize=True,load\\_in\\_4bit=True,peft=chenhugging/mistral-7b-ocn-v1), gen\\_kwargs: (None), limit: 100.0, num\\_fewshot: None, batch\\_size: 1"
] | [
-0.11391337215900421,
0.08055956661701202,
-0.005442911759018898,
0.07870133966207504,
0.08245933800935745,
0.02612409181892872,
0.12946480512619019,
0.13926707208156586,
-0.026486480608582497,
0.12472739815711975,
0.10250131040811539,
0.059227000921964645,
0.06405658274888992,
0.12498190999031067,
-0.007467725779861212,
-0.2672506272792816,
0.020196907222270966,
-0.03376316651701927,
-0.11357303708791733,
0.10353509336709976,
0.06979062408208847,
-0.09297675639390945,
0.049682676792144775,
-0.00936125498265028,
-0.07225380837917328,
-0.023762252181768417,
-0.028774123638868332,
-0.015279402025043964,
0.103571817278862,
0.060707081109285355,
0.08604622632265091,
0.03772377967834473,
0.08544302731752396,
-0.23438774049282074,
0.004325989168137312,
0.06851044297218323,
0.007409828249365091,
0.07299882918596268,
0.1038479134440422,
0.024429133161902428,
0.09583545476198196,
-0.10891230404376984,
0.046620480716228485,
0.03744208440184593,
-0.14006121456623077,
-0.248713880777359,
-0.0960497185587883,
0.07597988843917847,
0.12545651197433472,
0.062496159225702286,
-0.023681513965129852,
0.10363395512104034,
-0.07548755407333374,
0.07849118113517761,
0.27352264523506165,
-0.2789060175418854,
-0.08962634205818176,
0.04011406749486923,
0.06727170199155807,
0.041106048971414566,
-0.1246635764837265,
-0.02904072031378746,
0.035514071583747864,
0.02269042283296585,
0.07062725722789764,
-0.008359633386135101,
0.10700011253356934,
-0.007118352223187685,
-0.13385529816150665,
-0.04421733319759369,
0.11160347610712051,
0.06561578065156937,
-0.02751431055366993,
-0.07271780073642731,
-0.04738876223564148,
-0.18856382369995117,
-0.05054369941353798,
-0.0017744852229952812,
0.02122037671506405,
-0.04050140455365181,
-0.03346395492553711,
0.05786381661891937,
-0.053417373448610306,
-0.061551645398139954,
0.0607788972556591,
0.13073612749576569,
0.06284265220165253,
-0.03597988560795784,
0.016703078523278236,
0.10428374260663986,
0.03167395293712616,
-0.17589746415615082,
-0.014993244782090187,
0.007718878332525492,
-0.08974070847034454,
-0.03652643784880638,
-0.0297517329454422,
0.046452105045318604,
0.04317896068096161,
0.1996179074048996,
-0.09542624652385712,
0.0975327342748642,
0.06328067928552628,
0.021305188536643982,
-0.07869329303503036,
0.12201578915119171,
-0.08268871158361435,
-0.08406717330217361,
-0.04138244315981865,
0.11926717311143875,
0.006125154905021191,
-0.006694318726658821,
-0.05360756069421768,
0.012067291885614395,
0.10351775586605072,
0.0403941310942173,
-0.003961313050240278,
0.028407704085111618,
-0.06410584598779678,
-0.02283080480992794,
0.07975341379642487,
-0.1004970371723175,
0.0392540879547596,
0.042608149349689484,
-0.06535150110721588,
-0.008970833383500576,
-0.01377283874899149,
-0.014998652040958405,
-0.015110963024199009,
0.10472860932350159,
-0.09135877341032028,
-0.01274895016103983,
-0.05548836663365364,
-0.09941378235816956,
0.031093427911400795,
-0.008923838846385479,
-0.000332305149640888,
-0.07683322578668594,
-0.09701944142580032,
-0.06562531739473343,
0.05425850301980972,
-0.07805720716714859,
-0.04314402863383293,
-0.03243936598300934,
-0.09705101698637009,
0.04529636353254318,
0.0011732379207387567,
0.10783226042985916,
-0.05459343641996384,
0.09504300355911255,
-0.0006070167291909456,
0.047402333468198776,
0.07611967623233795,
0.02538733184337616,
-0.05790093541145325,
0.0863189846277237,
-0.1867574006319046,
0.0371658056974411,
-0.08923996239900589,
0.06288035213947296,
-0.1451098769903183,
-0.11376231908798218,
-0.012231475673615932,
-0.01732902228832245,
0.09312418848276138,
0.11728176474571228,
-0.15968067944049835,
-0.06834694743156433,
0.17286555469036102,
-0.08430332690477371,
-0.09226755052804947,
0.11660297960042953,
-0.015109507367014885,
-0.037261322140693665,
0.0164100993424654,
0.15056483447551727,
0.14475645124912262,
-0.11951366811990738,
-0.02419150620698929,
-0.041515253484249115,
0.1159970685839653,
0.07169604301452637,
0.0861978530883789,
-0.009735355153679848,
0.020664773881435394,
0.005591725464910269,
-0.05375019460916519,
0.039563145488500595,
-0.07976279407739639,
-0.07557816803455353,
-0.010541105642914772,
-0.05585150420665741,
0.05488108843564987,
0.05730699747800827,
0.023834621533751488,
-0.07181937992572784,
-0.13162171840667725,
0.011276849545538425,
0.10610523819923401,
-0.07146051526069641,
0.016156664118170738,
-0.03406376764178276,
0.09074139595031738,
0.016346778720617294,
0.0012336346553638577,
-0.15609437227249146,
-0.08234745264053345,
0.03729124367237091,
-0.08855920284986496,
-0.009402674622833729,
-0.002534850500524044,
0.08171682804822922,
0.10360050201416016,
-0.058587539941072464,
-0.05333005264401436,
-0.05249131843447685,
-0.0032020844519138336,
-0.092027448117733,
-0.24238581955432892,
-0.09400384873151779,
-0.022046681493520737,
0.1528882086277008,
-0.22902491688728333,
0.015746746212244034,
0.013385399244725704,
0.1425732970237732,
0.013919590972363949,
-0.03882646560668945,
-0.022800235077738762,
0.07136892527341843,
-0.025327974930405617,
-0.07476019114255905,
0.04031750559806824,
-0.015821406617760658,
-0.07196468114852905,
-0.024219267070293427,
-0.1527910828590393,
0.09730162471532822,
0.0775364339351654,
0.024391423910856247,
-0.1326085925102234,
-0.053266528993844986,
-0.062188521027565,
-0.0580870658159256,
0.0002026808069786057,
0.000012260588846402243,
0.11265942454338074,
0.0374133475124836,
0.10000977665185928,
-0.07224232703447342,
-0.06060631945729256,
0.0353105366230011,
-0.007387703284621239,
0.0005749831325374544,
0.14943845570087433,
0.03579820692539215,
-0.09950035065412521,
0.11084253340959549,
0.10941245406866074,
-0.053264252841472626,
0.12397023290395737,
-0.06821461766958237,
-0.08243665844202042,
-0.07885392755270004,
0.05683087185025215,
0.04102262482047081,
0.11424922943115234,
-0.04564748331904411,
0.01109318807721138,
0.008024594746530056,
0.005545752588659525,
-0.013617700897157192,
-0.16600291430950165,
-0.019270434975624084,
0.024423211812973022,
-0.04418771341443062,
0.0235516969114542,
-0.011077581904828548,
-0.014358434826135635,
0.08674679696559906,
0.016830679029226303,
-0.074521504342556,
-0.03790837153792381,
-0.021666057407855988,
-0.07500075548887253,
0.21307691931724548,
-0.0716511681675911,
-0.0840534120798111,
-0.112911157310009,
-0.010653705336153507,
-0.02608683705329895,
-0.023534204810857773,
0.03481866419315338,
-0.06805356591939926,
-0.03666599839925766,
-0.07322963327169418,
-0.0285163763910532,
-0.007366923149675131,
0.014088373631238937,
-0.0074272374622523785,
0.00045542666339315474,
0.07947596907615662,
-0.09256450086832047,
0.005629588384181261,
0.0034298880491405725,
-0.018789565190672874,
0.027126790955662727,
0.011711015366017818,
0.08437035232782364,
0.14289909601211548,
0.04754679650068283,
0.014305499382317066,
-0.03363025188446045,
0.2322039008140564,
-0.08790937811136246,
0.017987601459026337,
0.08207523077726364,
-0.0011622995370998979,
0.07289829105138779,
0.1548210233449936,
0.06166427209973335,
-0.07521820813417435,
-0.00026600222918204963,
0.03673679009079933,
-0.016021374613046646,
-0.2011527270078659,
-0.03319208323955536,
-0.05656294524669647,
0.009659179486334324,
0.13456840813159943,
0.03440999984741211,
0.02570333704352379,
0.022260254248976707,
-0.03857322409749031,
0.042623620480298996,
0.0353417731821537,
0.08562901616096497,
0.05630103498697281,
0.05411801487207413,
0.10162824392318726,
-0.011112839914858341,
-0.020782098174095154,
0.016890952363610268,
-0.004342189058661461,
0.2374921292066574,
-0.005112217273563147,
0.16115833818912506,
0.05185256153345108,
0.15973153710365295,
-0.006704054772853851,
0.0409877710044384,
0.0256655216217041,
-0.018907370045781136,
0.0028997173067182302,
-0.07663071900606155,
0.001670420984737575,
0.044828955084085464,
0.038627706468105316,
0.026166237890720367,
-0.09453354775905609,
-0.005751343909651041,
0.03489471599459648,
0.26919037103652954,
0.08996909856796265,
-0.288887619972229,
-0.07593341916799545,
0.004270945210009813,
-0.017307715490460396,
-0.009868386201560497,
0.009808332659304142,
0.09321599453687668,
-0.07947363704442978,
0.06987292319536209,
-0.06485630571842194,
0.07565147429704666,
-0.03159038722515106,
0.020806705579161644,
0.10802720487117767,
0.07053086906671524,
0.0005421570385806262,
0.0478377528488636,
-0.19867755472660065,
0.2725054621696472,
-0.0019184426637366414,
0.05785539001226425,
-0.018833978101611137,
0.023992303758859634,
0.030761776491999626,
-0.006310132797807455,
0.08582758158445358,
0.0011540589621290565,
-0.09791265428066254,
-0.23120522499084473,
-0.09100982546806335,
-0.012523968704044819,
0.12149874866008759,
-0.06678588688373566,
0.13737817108631134,
-0.030466753989458084,
-0.034425653517246246,
0.030000703409314156,
0.013534866273403168,
-0.09861141443252563,
-0.08135296404361725,
0.05084004998207092,
-0.043776173144578934,
-0.00047712947707623243,
-0.08918218314647675,
-0.09101127088069916,
-0.08723652362823486,
0.1345185488462448,
-0.10807718336582184,
-0.03141985833644867,
-0.12980739772319794,
0.06870057433843613,
0.16259849071502686,
-0.09703446179628372,
0.029153063893318176,
0.009682665579020977,
0.08061959594488144,
0.0012829393381252885,
-0.031656183302402496,
0.10304959118366241,
-0.054020900279283524,
-0.22575169801712036,
-0.0646284818649292,
0.15306268632411957,
0.035438619554042816,
0.06108348071575165,
-0.026019180193543434,
0.05051492899656296,
0.017118679359555244,
-0.10198225826025009,
0.03703117370605469,
-0.008939523249864578,
0.048678480088710785,
0.010035189799964428,
-0.03615342080593109,
0.058349333703517914,
-0.06987140327692032,
-0.010100317187607288,
0.05298469215631485,
0.3560889661312103,
-0.10337269306182861,
0.002049077767878771,
0.04982916638255119,
-0.054064903408288956,
-0.14749567210674286,
-0.04485681280493736,
0.12153403460979462,
-0.0058111888356506824,
0.03504471853375435,
-0.16557855904102325,
0.03704199194908142,
0.11525934189558029,
-0.014437981881201267,
0.10285339504480362,
-0.34458431601524353,
-0.13606682419776917,
0.06390600651502609,
0.09366559237241745,
-0.013750570826232433,
-0.19445139169692993,
-0.06736432760953903,
0.007646794896572828,
-0.12419898062944412,
0.03117780201137066,
-0.03836723044514656,
0.11540832370519638,
-0.02648034319281578,
-0.03829245641827583,
0.01699872501194477,
-0.062011491507291794,
0.17545387148857117,
0.005817941389977932,
0.07675540447235107,
-0.030572310090065002,
-0.01921227015554905,
0.034940801560878754,
-0.07911961525678635,
0.005703676026314497,
-0.07517295330762863,
0.017754165455698967,
-0.09807264059782028,
-0.005357631482183933,
-0.09340274333953857,
0.019136525690555573,
-0.07918816804885864,
-0.022178800776600838,
-0.04505179077386856,
0.08165469765663147,
0.06328587979078293,
-0.0006225061370059848,
0.12724798917770386,
0.005367811769247055,
0.17694786190986633,
0.16421176493167877,
0.021146481856703758,
0.03903363645076752,
-0.13436409831047058,
-0.0018473375821486115,
-0.005495165474712849,
0.04233657568693161,
-0.11124468594789505,
0.017409734427928925,
0.15131241083145142,
0.04582369327545166,
0.12855751812458038,
0.0566328689455986,
-0.07148187607526779,
-0.011711043305695057,
0.06534935534000397,
-0.09961382299661636,
-0.11880511790513992,
-0.019155466929078102,
0.0012329802848398685,
-0.13035504519939423,
-0.02411951869726181,
0.13843528926372528,
-0.0658101961016655,
-0.00029030925361439586,
0.02842066064476967,
0.041619181632995605,
-0.02751164324581623,
0.23708544671535492,
0.03302839398384094,
0.09349481016397476,
-0.08259101212024689,
0.0544305294752121,
0.08324512094259262,
-0.06814509630203247,
0.025735946372151375,
0.1044539287686348,
-0.06767816841602325,
-0.033929046243429184,
0.030712442472577095,
0.0777924656867981,
0.001445793081074953,
-0.013140900991857052,
-0.11946377903223038,
-0.12582601606845856,
0.08278951048851013,
0.09957445412874222,
0.03852793201804161,
0.03376172482967377,
0.0014305959921330214,
0.02123888023197651,
-0.07552780210971832,
0.11924458295106888,
0.08426041156053543,
0.09002339839935303,
-0.1270757019519806,
0.13243143260478973,
-0.0006223628297448158,
-0.002277931198477745,
0.0029205267783254385,
0.01340226735919714,
-0.16179244220256805,
-0.005484384950250387,
-0.13308438658714294,
0.008005913347005844,
-0.05597558617591858,
-0.007089353632181883,
-0.01691090129315853,
-0.04216349124908447,
-0.03803425654768944,
0.012082240544259548,
-0.09236808866262436,
-0.05766130983829498,
-0.012242477387189865,
0.07157609611749649,
-0.12630529701709747,
-0.029446441680192947,
0.04153410345315933,
-0.11933854967355728,
0.06562455743551254,
0.010326309129595757,
0.03843109682202339,
0.02019384689629078,
-0.04639275744557381,
0.023647595196962357,
0.006959541235119104,
-0.010001958347856998,
0.02684321254491806,
-0.17281708121299744,
-0.012615197338163853,
-0.07052154839038849,
0.01145407184958458,
0.014717469923198223,
0.00035342478076927364,
-0.12920531630516052,
0.012766544707119465,
-0.03986846283078194,
-0.06930989772081375,
-0.0536455474793911,
0.028612570837140083,
0.06805761158466339,
-0.0006434489041566849,
0.1351349949836731,
-0.08337627351284027,
0.06766441464424133,
-0.25264453887939453,
-0.02858128398656845,
0.010375279933214188,
-0.06706863641738892,
-0.030199483036994934,
-0.03986559808254242,
0.08293062448501587,
-0.058701302856206894,
0.052258290350437164,
-0.05616749823093414,
0.07019657641649246,
0.02759694680571556,
-0.06899982690811157,
-0.008151654154062271,
0.05865425616502762,
0.15059582889080048,
0.04178040474653244,
-0.030426178127527237,
0.03480621427297592,
0.017762551084160805,
0.0418785884976387,
0.07619889825582504,
0.1709015816450119,
0.13369908928871155,
0.026832278817892075,
0.05841447040438652,
0.03357482701539993,
-0.13614150881767273,
-0.13281360268592834,
0.11463745683431625,
-0.03889487311244011,
0.12371876835823059,
-0.026076991111040115,
0.13653111457824707,
0.11568793654441833,
-0.2197476178407669,
0.04462454468011856,
-0.05580156669020653,
-0.09537266939878464,
-0.09068039804697037,
-0.08030854165554047,
-0.06938590854406357,
-0.13477906584739685,
0.0015332764014601707,
-0.09795333445072174,
0.02623800002038479,
0.08565051853656769,
0.03284655511379242,
0.0442490354180336,
0.11712201684713364,
0.06581759452819824,
0.012040686793625355,
0.06639853119850159,
0.03901571035385132,
-0.014475593343377113,
-0.029766665771603584,
-0.07247570902109146,
0.022576583549380302,
-0.0635962039232254,
0.027932753786444664,
-0.02835696190595627,
-0.046601802110672,
0.060029905289411545,
0.021462874487042427,
-0.10316122323274612,
0.03571290895342827,
0.004701756406575441,
0.03783300146460533,
0.07672999799251556,
0.027131468057632446,
-0.00046301804832182825,
-0.0261499285697937,
0.18506288528442383,
-0.06859195232391357,
-0.06043018773198128,
-0.10614190250635147,
0.26361003518104553,
0.02166675589978695,
-0.01022487785667181,
0.0185895636677742,
-0.06746063381433487,
0.01939014159142971,
0.13310620188713074,
0.13319562375545502,
-0.04502246528863907,
-0.007217101752758026,
0.03974663093686104,
-0.012437955476343632,
-0.012979645282030106,
0.09935017675161362,
0.089584119617939,
0.07723967730998993,
-0.07576346397399902,
0.005027105566114187,
-0.035274144262075424,
-0.041623249650001526,
-0.066684290766716,
0.062306202948093414,
0.04155117645859718,
0.010001074522733688,
-0.0454062819480896,
0.08942767977714539,
-0.05548522621393204,
-0.12201946973800659,
0.09653171896934509,
-0.19424238801002502,
-0.18044814467430115,
-0.032461151480674744,
0.053340546786785126,
0.02506251446902752,
0.07045814394950867,
-0.01778574101626873,
-0.03347671404480934,
0.10387542098760605,
-0.014086583629250526,
-0.021080249920487404,
-0.11691804230213165,
0.06475906819105148,
-0.05519206449389458,
0.19391334056854248,
-0.03756359592080116,
0.039790328592061996,
0.1313556432723999,
0.024951769039034843,
-0.10372021794319153,
0.010811399668455124,
0.10174329578876495,
-0.12571679055690765,
0.022453386336565018,
0.12561744451522827,
-0.02045503444969654,
0.10998009890317917,
0.054784808307886124,
-0.08372937142848969,
0.015047010034322739,
0.0150813739746809,
-0.014243514277040958,
-0.05575621873140335,
-0.0016793650574982166,
-0.028189361095428467,
0.15235665440559387,
0.22245636582374573,
-0.05867258086800575,
-0.015204270370304585,
-0.03349115327000618,
0.020661836490035057,
0.057866550981998444,
0.10617104917764664,
-0.03775510936975479,
-0.23650498688220978,
0.047409556806087494,
0.009824682027101517,
0.04435738921165466,
-0.18007206916809082,
-0.10243384540081024,
0.039150137454271317,
-0.05287802219390869,
-0.06246200203895569,
0.13509756326675415,
0.037944965064525604,
0.05865303426980972,
-0.0476519837975502,
-0.14342792332172394,
-0.056660957634449005,
0.164807990193367,
-0.15375640988349915,
-0.07526477426290512
] |
null | null | stable-baselines3 |
# **PPO** Agent playing **LunarLander-v2**
This is a trained model of a **PPO** agent playing **LunarLander-v2**
using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3).
## Usage (with Stable-baselines3)
TODO: Add your code
```python
from stable_baselines3 import ...
from huggingface_sb3 import load_from_hub
...
```
| {"library_name": "stable-baselines3", "tags": ["LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "PPO", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "LunarLander-v2", "type": "LunarLander-v2"}, "metrics": [{"type": "mean_reward", "value": "270.77 +/- 22.88", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | EricValen/ppo-LunarLander-v2 | [
"stable-baselines3",
"LunarLander-v2",
"deep-reinforcement-learning",
"reinforcement-learning",
"model-index",
"region:us"
] | 2024-02-07T06:18:24+00:00 | [] | [] | TAGS
#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
|
# PPO Agent playing LunarLander-v2
This is a trained model of a PPO agent playing LunarLander-v2
using the stable-baselines3 library.
## Usage (with Stable-baselines3)
TODO: Add your code
| [
"# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
"TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n",
"# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
39,
41,
17
] | [
"passage: TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
0.03942384943366051,
0.04900386184453964,
-0.005304091144353151,
0.026427261531352997,
0.107408307492733,
-0.026511888951063156,
0.11188238859176636,
0.0814051404595375,
0.10722193866968155,
0.04762078449130058,
0.08338645845651627,
0.06030960753560066,
0.05080918222665787,
0.2571701407432556,
0.04754156619310379,
-0.22987541556358337,
0.036159250885248184,
-0.04869936779141426,
0.12395193427801132,
0.07178173214197159,
-0.0038484656251966953,
-0.06485428661108017,
0.020415637642145157,
-0.013290755450725555,
0.05367108806967735,
0.04282612353563309,
-0.01716216839849949,
-0.08207534998655319,
0.07169748842716217,
-0.06345846503973007,
0.06986866891384125,
0.07677983492612839,
0.13218913972377777,
-0.17832116782665253,
0.029566360637545586,
0.02571309357881546,
-0.07189024239778519,
0.01342033501714468,
0.008019951172173023,
0.05120139941573143,
0.17303818464279175,
0.019879888743162155,
0.07844575494527817,
-0.0025605305563658476,
-0.15412317216396332,
-0.018950799480080605,
0.0436202734708786,
0.12546207010746002,
0.08808347582817078,
0.04605821147561073,
0.01970590092241764,
0.17503218352794647,
-0.054352790117263794,
-0.028833400458097458,
0.21759237349033356,
-0.2881564497947693,
-0.031460098922252655,
0.321048766374588,
0.06997483223676682,
0.09725230932235718,
-0.07540661096572876,
-0.03619609400629997,
0.007783263456076384,
-0.013137873262166977,
-0.028666524216532707,
-0.07447073608636856,
0.17313385009765625,
0.05152064561843872,
-0.05057951435446739,
-0.09541505575180054,
0.16948209702968597,
0.006921638268977404,
0.0018855923553928733,
-0.019282981753349304,
0.009060598909854889,
0.07402525842189789,
-0.016097044572234154,
-0.07255112379789352,
0.057438433170318604,
0.05330665782094002,
0.019649166613817215,
-0.1435653269290924,
-0.10762494057416916,
-0.022740179672837257,
-0.008012006990611553,
0.17786912620067596,
-0.009255532175302505,
0.042902372777462006,
0.003065188182517886,
0.10384012013673782,
-0.12480384111404419,
-0.03354184702038765,
-0.0454259067773819,
-0.07565800100564957,
-0.0223417766392231,
-0.02058211714029312,
-0.03580251708626747,
0.07184842973947525,
0.11971849203109741,
0.027368178591132164,
0.09350208193063736,
0.047715865075588226,
-0.03206788748502731,
0.06343851238489151,
0.05555703118443489,
0.14222665131092072,
0.05807621404528618,
0.012854371219873428,
0.13179877400398254,
0.055213116109371185,
0.033023182302713394,
-0.0613492950797081,
-0.18252409994602203,
0.07489913702011108,
-0.07031869143247604,
0.007941240444779396,
0.12051256000995636,
-0.04480670019984245,
-0.1183447614312172,
-0.037500523030757904,
-0.017392054200172424,
-0.06224250793457031,
-0.025395862758159637,
0.0547584593296051,
-0.02883218228816986,
-0.03973718360066414,
0.0011496668448671699,
0.09384800493717194,
0.00953749567270279,
-0.1752052903175354,
0.03303423151373863,
-0.025042934343218803,
-0.10782608389854431,
0.009975161403417587,
0.0022444494534283876,
0.03394931182265282,
0.04408763721585274,
-0.11822668462991714,
-0.30899152159690857,
-0.07652641832828522,
0.05490870401263237,
-0.06516939401626587,
-0.18425025045871735,
-0.13193942606449127,
0.02454492449760437,
-0.09037084132432938,
-0.044885024428367615,
-0.12759265303611755,
-0.028549788519740105,
0.01743689924478531,
0.011519349180161953,
0.10758619755506516,
-0.0106219332665205,
-0.012188062071800232,
-0.1571401208639145,
0.008273907005786896,
-0.20951123535633087,
0.0890483483672142,
-0.019150104373693466,
0.037884220480918884,
-0.032381169497966766,
-0.07404014468193054,
0.030707746744155884,
0.052499737590551376,
-0.01474119070917368,
0.13510210812091827,
-0.15592676401138306,
-0.03691192343831062,
-0.007996266707777977,
-0.13611900806427002,
-0.04786273464560509,
-0.10358831286430359,
-0.04357128217816353,
0.13354332745075226,
0.018664736300706863,
0.15356586873531342,
-0.08709818124771118,
-0.0722038671374321,
0.20489206910133362,
-0.010411538183689117,
-0.12820468842983246,
-0.076752208173275,
0.10165707021951675,
0.021510310471057892,
-0.056606587022542953,
-0.02523270808160305,
-0.1839766949415207,
-0.0152357779443264,
-0.04550420492887497,
-0.047039128839969635,
0.01796751655638218,
-0.010888241231441498,
0.13837894797325134,
0.08494598418474197,
0.05018039792776108,
-0.06086122244596481,
-0.006730288732796907,
0.10779471695423126,
0.08823856711387634,
0.008680110797286034,
0.023406028747558594,
-0.05774238705635071,
0.09552932530641556,
-0.04003755748271942,
-0.0142367510125041,
-0.08283266425132751,
-0.036246106028556824,
-0.026256313547492027,
0.17507147789001465,
0.09440762549638748,
0.2257927656173706,
0.09567736834287643,
0.039160262793302536,
0.031270865350961685,
-0.13181598484516144,
-0.1425403207540512,
-0.0017254541162401438,
0.09020978957414627,
-0.14270411431789398,
-0.04119925573468208,
-0.08974775671958923,
-0.17768175899982452,
-0.12202505767345428,
0.0006432619411498308,
-0.17960017919540405,
0.06390921026468277,
0.05408334732055664,
-0.035177867859601974,
0.03272094577550888,
0.13032332062721252,
-0.011533179320394993,
-0.03967514634132385,
0.0831870287656784,
0.0379033200442791,
-0.041234664618968964,
-0.021742934361100197,
0.11885567009449005,
0.15673065185546875,
0.13124459981918335,
-0.03511447086930275,
0.004914294462651014,
0.07076404243707657,
-0.02309088408946991,
0.06539414077997208,
0.0558244064450264,
0.20973342657089233,
0.188301220536232,
0.038996949791908264,
0.008822928182780743,
-0.07048165798187256,
0.0855446457862854,
-0.0742373839020729,
-0.14302679896354675,
-0.05579735338687897,
0.08729292452335358,
0.016605578362941742,
0.023469142615795135,
0.08711627870798111,
0.024545932188630104,
0.09132762253284454,
0.15968108177185059,
0.01990218088030815,
-0.09659269452095032,
-0.050218869000673294,
0.01175848301500082,
0.027713103219866753,
0.04794301092624664,
-0.04514073207974434,
-0.00937939714640379,
0.017020760104060173,
-0.10303554683923721,
0.031789086759090424,
-0.1413339376449585,
-0.1358717679977417,
0.044326696544885635,
0.003906996920704842,
0.010907664895057678,
0.02786896750330925,
-0.0038291432429105043,
0.019039705395698547,
0.04351753741502762,
-0.06975466758012772,
0.047416772693395615,
-0.024745507165789604,
-0.020031947642564774,
0.03340689837932587,
-0.057257164269685745,
-0.205775648355484,
-0.17696654796600342,
0.00013708483311347663,
-0.09910997003316879,
0.10194740444421768,
0.018308809027075768,
-0.12373185902833939,
0.047737859189510345,
-0.05822649225592613,
0.027574289590120316,
-0.01875593699514866,
-0.049130141735076904,
0.10507171601057053,
0.1525275856256485,
-0.016146350651979446,
0.018018173053860664,
-0.04865182936191559,
-0.10157987475395203,
-0.19632206857204437,
0.0691583976149559,
0.04680244252085686,
0.014610917307436466,
0.10669491440057755,
0.018072687089443207,
0.02367905154824257,
-0.007674071006476879,
-0.016521066427230835,
-0.011659215204417706,
-0.08781040459871292,
0.31909599900245667,
0.04510033503174782,
-0.025173069909214973,
0.02041010931134224,
-0.0043001663871109486,
-0.028083480894565582,
0.03263787180185318,
-0.0985708013176918,
-0.07548979669809341,
-0.08774089068174362,
-0.04367410019040108,
-0.09784720093011856,
0.053299110382795334,
0.05916472524404526,
0.003188040340319276,
-0.07727594673633575,
0.04221395403146744,
0.11369874328374863,
-0.0923808291554451,
-0.07137343287467957,
0.07477962225675583,
0.0972946360707283,
-0.07331304252147675,
0.00012658814375754446,
0.00874367356300354,
0.023951783776283264,
0.037102166563272476,
0.06778035312891006,
-0.03966575115919113,
0.08589404821395874,
-0.19917890429496765,
0.0372927263379097,
0.106058269739151,
0.023754918947815895,
0.0638108178973198,
0.07643651217222214,
-0.1058402881026268,
-0.008500572293996811,
-0.032518330961465836,
-0.21341575682163239,
0.1668180525302887,
0.1355515867471695,
0.06788124144077301,
-0.025637222453951836,
-0.00461410591378808,
-0.0649740919470787,
0.05773647129535675,
0.02723747305572033,
-0.14758841693401337,
0.004883295856416225,
0.06064270809292793,
0.026899009943008423,
0.01614922471344471,
0.07971042394638062,
0.014697225764393806,
-0.1801026314496994,
-0.014406266622245312,
0.10730406641960144,
0.002390873385593295,
0.0053148469887673855,
-0.03175045922398567,
-0.1755964607000351,
0.0751047357916832,
0.004285442177206278,
0.07233936339616776,
-0.1676585078239441,
0.14297930896282196,
-0.10089799761772156,
0.07726949453353882,
-0.004285062663257122,
-0.021311495453119278,
0.02507244050502777,
-0.0541163794696331,
0.15163759887218475,
0.01058570109307766,
-0.021810131147503853,
-0.1200498715043068,
-0.1717042326927185,
-0.019227758049964905,
-0.11788936704397202,
-0.11679866164922714,
0.050424277782440186,
0.062185097485780716,
0.04923136904835701,
-0.061147067695856094,
0.1518532931804657,
-0.047422297298908234,
0.060713399201631546,
-0.06893875449895859,
-0.06755045056343079,
0.03764858841896057,
-0.12588608264923096,
-0.08176055550575256,
0.05573027580976486,
0.19166934490203857,
0.15833087265491486,
-0.02816431224346161,
-0.03472423925995827,
-0.047419581562280655,
-0.006212298292666674,
-0.007802055217325687,
0.0275666993111372,
0.023223137483000755,
0.07315318286418915,
-0.07681374251842499,
-0.11649256944656372,
0.033787861466407776,
-0.06713802367448807,
-0.055589709430933,
-0.015439179725944996,
0.1513158082962036,
0.04671623185276985,
0.07720734924077988,
-0.018946662545204163,
0.03887668624520302,
-0.001724981120787561,
-0.056474871933460236,
0.16197094321250916,
0.03885216265916824,
-0.05193585529923439,
0.06837689876556396,
0.053174007683992386,
0.043745119124650955,
0.03011113777756691,
-0.026783017441630363,
0.206032395362854,
0.1980147808790207,
0.014206883497536182,
0.2175983190536499,
0.03177616000175476,
-0.03772832080721855,
-0.1300560086965561,
-0.065880686044693,
-0.006372632458806038,
0.03559038043022156,
0.08070417493581772,
-0.18207235634326935,
-0.015011128038167953,
-0.05689644813537598,
-0.034518610686063766,
-0.15059494972229004,
-0.28553900122642517,
-0.05957856774330139,
0.20075850188732147,
0.14706264436244965,
0.27519428730010986,
-0.10432573407888412,
0.035197313874959946,
0.02663275972008705,
-0.04912831634283066,
-0.006501141935586929,
0.00018665487004909664,
0.10268618166446686,
-0.15421873331069946,
0.1176437959074974,
0.08486983180046082,
-0.019002694636583328,
0.01058861706405878,
-0.1619086116552353,
0.00936629343777895,
-0.12191236019134521,
0.05354422330856323,
0.1400289237499237,
-0.048128653317689896,
-0.054873593151569366,
0.14033560454845428,
-0.024562934413552284,
-0.22685599327087402,
-0.04648222774267197,
-0.043600670993328094,
-0.010640020482242107,
0.026607351377606392,
-0.1013401448726654,
0.04101909324526787,
0.1330099105834961,
0.009380043484270573,
0.1147187277674675,
0.11749245226383209,
-0.052566803991794586,
0.10792597383260727,
0.2257719188928604,
-0.018785694614052773,
0.04689010605216026,
-0.12743118405342102,
-0.0012336712097749114,
-0.028270328417420387,
0.013657891191542149,
-0.09504974633455276,
-0.09938385337591171,
0.02366873063147068,
0.02872389927506447,
0.009118586778640747,
0.0921793207526207,
-0.029922157526016235,
0.0759170651435852,
0.06817561388015747,
-0.13014446198940277,
-0.16288450360298157,
0.015828335657715797,
-0.007344507612287998,
0.08354310691356659,
0.00027861111448146403,
0.08878035843372345,
-0.11932205408811569,
-0.018093237653374672,
-0.03153328225016594,
-0.03319635987281799,
-0.130486860871315,
-0.07138993591070175,
0.06156524643301964,
0.028095467016100883,
-0.06602972000837326,
0.1398407518863678,
0.026440169662237167,
0.15942534804344177,
0.049197953194379807,
0.012499804608523846,
0.07227300107479095,
-0.05345509201288223,
0.1283530443906784,
0.13818155229091644,
-0.00868943240493536,
-0.05460423603653908,
-0.1013643890619278,
-0.10236792266368866,
0.08925779908895493,
-0.05773641914129257,
0.07476430386304855,
-0.14885357022285461,
-0.06675903499126434,
0.015772046521306038,
0.016141414642333984,
-0.09562095999717712,
0.02571965754032135,
-0.01625603251159191,
-0.18119946122169495,
0.056570518761873245,
-0.048285093158483505,
0.0440407395362854,
-0.06347788125276566,
-0.1110161691904068,
-0.17226378619670868,
0.06091433763504028,
0.08593481779098511,
-0.053876690566539764,
-0.12229149043560028,
0.011023230850696564,
-0.00012518465518951416,
-0.06341652572154999,
-0.05023367330431938,
0.09722746908664703,
-0.11020902544260025,
0.031452205032110214,
-0.012567701749503613,
0.08853451162576675,
-0.03510405123233795,
-0.011538895778357983,
0.044220831245183945,
-0.08039166033267975,
-0.009481523185968399,
0.03534642979502678,
-0.026372017338871956,
-0.04127239063382149,
-0.2689029574394226,
0.0036654395516961813,
0.0341104120016098,
0.02497158572077751,
0.07856601476669312,
0.011906822212040424,
0.021174922585487366,
0.03993808850646019,
-0.15396519005298615,
-0.013395369984209538,
0.14574195444583893,
-0.07689505815505981,
-0.022186370566487312,
0.05703273415565491,
-0.09054436534643173,
0.013882770203053951,
-0.030287226662039757,
0.1345842480659485,
0.023923413828015327,
0.06404478847980499,
-0.0851147472858429,
0.10106813907623291,
-0.1451139897108078,
-0.04998219385743141,
-0.01244612317532301,
0.09761348366737366,
0.07019034773111343,
-0.10272270441055298,
0.014697125181555748,
0.04210108891129494,
0.19416837394237518,
0.016384804621338844,
-0.0356343574821949,
-0.03396720811724663,
0.004015897400677204,
0.22076453268527985,
0.03044266067445278,
0.10457023978233337,
0.07281364500522614,
-0.026583973318338394,
0.12624378502368927,
0.09929762035608292,
0.11280370503664017,
-0.055645186454057693,
0.13904185593128204,
0.04667386785149574,
0.038641396909952164,
0.0614289753139019,
0.06836545467376709,
0.09098632633686066,
-0.0008288522367365658,
0.1138714924454689,
0.013811973854899406,
-0.02422109805047512,
-0.021335409954190254,
0.17759373784065247,
0.10501719266176224,
-0.14769648015499115,
0.029047364369034767,
-0.01258957851678133,
0.039933037012815475,
-0.014194529503583908,
-0.15634691715240479,
-0.07240267097949982,
-0.3315149247646332,
0.1226184144616127,
-0.07119352370500565,
0.019930170848965645,
0.007913772016763687,
-0.037425633519887924,
-0.03296699747443199,
-0.04477746784687042,
0.13151589035987854,
-0.013641550205647945,
-0.006079165264964104,
-0.04815853759646416,
-0.015360191464424133,
-0.11607866734266281,
-0.11200575530529022,
-0.013207737356424332,
-0.13671602308750153,
-0.010119039565324783,
0.05595948174595833,
0.003977729007601738,
0.01821410097181797,
-0.03142618387937546,
0.0024383175186812878,
0.06541839241981506,
-0.05751744285225868,
0.056182678788900375,
0.12097269296646118,
0.08766137808561325,
-0.1058853268623352,
0.031048951670527458,
0.2011747509241104,
0.04359564557671547,
-0.12483977526426315,
0.01449228823184967,
0.1819491684436798,
0.004885740112513304,
0.017068125307559967,
-0.006097703706473112,
-0.0540788508951664,
-0.07554277032613754,
0.1251034289598465,
0.08296554535627365,
-0.09985227137804031,
0.015833314508199692,
-0.0726347416639328,
-0.01594804972410202,
-0.06374675035476685,
0.10130585730075836,
0.09538925439119339,
0.04440245032310486,
-0.10621760785579681,
-0.08487539738416672,
-0.10891728103160858,
0.040588874369859695,
-0.08629853278398514,
-0.07311757653951645,
0.09629398584365845,
-0.07057105004787445,
-0.07029950618743896,
0.025521177798509598,
-0.17978744208812714,
-0.009467960335314274,
0.1711762249469757,
-0.24654000997543335,
-0.0916430801153183,
-0.10857923328876495,
0.14477859437465668,
0.016497576609253883,
0.1013975441455841,
-0.006207061931490898,
-0.007889035157859325,
-0.20577777922153473,
0.024890204891562462,
-0.05293011665344238,
-0.02073732763528824,
0.07814782857894897,
-0.09476397186517715,
0.22629831731319427,
-0.08276885002851486,
0.020940175279974937,
0.012659613974392414,
0.0870661810040474,
-0.030675338581204414,
0.09283176809549332,
-0.03660329803824425,
-0.12576518952846527,
-0.03620953485369682,
0.03001813031733036,
0.013904244638979435,
0.10071761906147003,
0.09772487729787827,
-0.03414725139737129,
0.03389119729399681,
0.09747414290904999,
0.04172342270612717,
-0.023843804374337196,
0.0360250361263752,
-0.17077107727527618,
0.02182629331946373,
-0.018498148769140244,
-0.06935930997133255,
0.03687669709324837,
-0.06603235751390457,
0.1639697551727295,
0.04022442549467087,
0.0670473501086235,
-0.036152735352516174,
0.0073931049555540085,
-0.014454689808189869,
-0.013775371946394444,
-0.026180334389209747,
-0.17259705066680908,
-0.10422050207853317,
-0.1347656100988388,
-0.012701659463346004,
-0.034971047192811966,
0.04591470584273338,
0.023234914988279343,
-0.0003200018545612693,
-0.014577031135559082,
-0.12090865522623062,
0.04360328987240791,
0.11146783083677292,
-0.04631396010518074,
-0.026193076744675636
] |
null | null | transformers | # RPmerge
See the main model card: https://huggingface.co/brucethemoose/Yi-34B-200K-RPMerge
Quantized with default exl2 quantization, still investigating the benefits/drawbacks of long context (32K) quantization.
## Merge Details
### Merge Method
This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using /home/alpha/Models/Raw/chargoddard_Yi-34B-200K-Llama as a base.
### Models Merged
The following models were included in the merge:
* /home/alpha/Models/Raw/migtissera_Tess-34B-v1.5b
* /home/alpha/Models/Raw/migtissera_Tess-M-Creative-v1.0
* /home/alpha/Models/Raw/cgato_Thespis-34b-DPO-v0.7
* /home/alpha/Models/Raw/Nous-Capybara-34B
* /home/alpha/Models/Raw/admo_limarp
* /home/alpha/Models/Raw/DrNicefellow_ChatAllInOne-Yi-34B-200K-V1
### Configuration
The following YAML configuration was used to produce this model:
```yaml
models:
- model: /home/alpha/Models/Raw/chargoddard_Yi-34B-200K-Llama
# No parameters necessary for base model
- model: /home/alpha/Models/Raw/migtissera_Tess-34B-v1.5b
#Emphasize the beginning of Vicuna format models
parameters:
weight: 0.19
density: 0.59
- model: /home/alpha/Models/Raw/Nous-Capybara-34B
parameters:
weight: 0.19
density: 0.55
# Vicuna format
- model: /home/alpha/Models/Raw/migtissera_Tess-M-Creative-v1.0
parameters:
weight: 0.05
density: 0.55
- model: /home/alpha/Models/Raw/DrNicefellow_ChatAllInOne-Yi-34B-200K-V1
parameters:
weight: 0.19
density: 0.55
- model: /home/alpha/Models/Raw/admo_limarp
parameters:
weight: 0.19
density: 0.48
- model: /home/alpha/Models/Raw/cgato_Thespis-34b-DPO-v0.7
parameters:
weight: 0.19
density: 0.59
merge_method: dare_ties
tokenizer_source: union
base_model: /home/alpha/Models/Raw/chargoddard_Yi-34B-200K-Llama
parameters:
int8_mask: true
dtype: bfloat16
```
| {"language": ["en"], "license": "other", "library_name": "transformers", "tags": ["mergekit", "merge", "Yi", "exllama", "exllamav2", "exl2"], "license_name": "yi-license", "license_link": "https://huggingface.co/01-ai/Yi-34B/blob/main/LICENSE", "base_model": []} | text-generation | brucethemoose/Yi-34B-200K-RPMerge-exl2-31bpw | [
"transformers",
"safetensors",
"llama",
"text-generation",
"mergekit",
"merge",
"Yi",
"exllama",
"exllamav2",
"exl2",
"en",
"arxiv:2311.03099",
"arxiv:2306.01708",
"license:other",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T06:21:14+00:00 | [
"2311.03099",
"2306.01708"
] | [
"en"
] | TAGS
#transformers #safetensors #llama #text-generation #mergekit #merge #Yi #exllama #exllamav2 #exl2 #en #arxiv-2311.03099 #arxiv-2306.01708 #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # RPmerge
See the main model card: URL
Quantized with default exl2 quantization, still investigating the benefits/drawbacks of long context (32K) quantization.
## Merge Details
### Merge Method
This model was merged using the DARE TIES merge method using /home/alpha/Models/Raw/chargoddard_Yi-34B-200K-Llama as a base.
### Models Merged
The following models were included in the merge:
* /home/alpha/Models/Raw/migtissera_Tess-34B-v1.5b
* /home/alpha/Models/Raw/migtissera_Tess-M-Creative-v1.0
* /home/alpha/Models/Raw/cgato_Thespis-34b-DPO-v0.7
* /home/alpha/Models/Raw/Nous-Capybara-34B
* /home/alpha/Models/Raw/admo_limarp
* /home/alpha/Models/Raw/DrNicefellow_ChatAllInOne-Yi-34B-200K-V1
### Configuration
The following YAML configuration was used to produce this model:
| [
"# RPmerge\n\nSee the main model card: URL\n\nQuantized with default exl2 quantization, still investigating the benefits/drawbacks of long context (32K) quantization.",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the DARE TIES merge method using /home/alpha/Models/Raw/chargoddard_Yi-34B-200K-Llama as a base.",
"### Models Merged\n\nThe following models were included in the merge:\n* /home/alpha/Models/Raw/migtissera_Tess-34B-v1.5b\n* /home/alpha/Models/Raw/migtissera_Tess-M-Creative-v1.0\n* /home/alpha/Models/Raw/cgato_Thespis-34b-DPO-v0.7\n* /home/alpha/Models/Raw/Nous-Capybara-34B\n* /home/alpha/Models/Raw/admo_limarp\n* /home/alpha/Models/Raw/DrNicefellow_ChatAllInOne-Yi-34B-200K-V1",
"### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #mergekit #merge #Yi #exllama #exllamav2 #exl2 #en #arxiv-2311.03099 #arxiv-2306.01708 #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# RPmerge\n\nSee the main model card: URL\n\nQuantized with default exl2 quantization, still investigating the benefits/drawbacks of long context (32K) quantization.",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the DARE TIES merge method using /home/alpha/Models/Raw/chargoddard_Yi-34B-200K-Llama as a base.",
"### Models Merged\n\nThe following models were included in the merge:\n* /home/alpha/Models/Raw/migtissera_Tess-34B-v1.5b\n* /home/alpha/Models/Raw/migtissera_Tess-M-Creative-v1.0\n* /home/alpha/Models/Raw/cgato_Thespis-34b-DPO-v0.7\n* /home/alpha/Models/Raw/Nous-Capybara-34B\n* /home/alpha/Models/Raw/admo_limarp\n* /home/alpha/Models/Raw/DrNicefellow_ChatAllInOne-Yi-34B-200K-V1",
"### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
95,
40,
4,
49,
169,
17
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #mergekit #merge #Yi #exllama #exllamav2 #exl2 #en #arxiv-2311.03099 #arxiv-2306.01708 #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# RPmerge\n\nSee the main model card: URL\n\nQuantized with default exl2 quantization, still investigating the benefits/drawbacks of long context (32K) quantization.## Merge Details### Merge Method\n\nThis model was merged using the DARE TIES merge method using /home/alpha/Models/Raw/chargoddard_Yi-34B-200K-Llama as a base.### Models Merged\n\nThe following models were included in the merge:\n* /home/alpha/Models/Raw/migtissera_Tess-34B-v1.5b\n* /home/alpha/Models/Raw/migtissera_Tess-M-Creative-v1.0\n* /home/alpha/Models/Raw/cgato_Thespis-34b-DPO-v0.7\n* /home/alpha/Models/Raw/Nous-Capybara-34B\n* /home/alpha/Models/Raw/admo_limarp\n* /home/alpha/Models/Raw/DrNicefellow_ChatAllInOne-Yi-34B-200K-V1### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
-0.09143144637346268,
0.13088233768939972,
-0.004239476285874844,
0.007002714555710554,
0.07611428201198578,
0.0688571110367775,
0.13210566341876984,
0.13319514691829681,
0.018930861726403236,
0.1146661564707756,
-0.004489467944949865,
0.11506475508213043,
0.11725322157144547,
0.1490548998117447,
-0.009691251441836357,
-0.13874879479408264,
0.043344542384147644,
-0.04330842196941376,
-0.06558141857385635,
0.09411337226629257,
0.09026461094617844,
-0.07027081400156021,
0.10735927522182465,
0.06902097910642624,
-0.09375990182161331,
0.0003966365766245872,
-0.04667649045586586,
0.02636537328362465,
0.07864397019147873,
0.07433581352233887,
0.06466394662857056,
0.05696464702486992,
0.056630536913871765,
-0.14254747331142426,
0.018034514039754868,
0.023142904043197632,
-0.000902830739505589,
0.06992360949516296,
0.10111942887306213,
-0.04483991488814354,
0.10474897176027298,
-0.08026403188705444,
0.03382281959056854,
0.07870031893253326,
-0.09348627179861069,
-0.18507787585258484,
-0.15320031344890594,
0.11755742132663727,
0.09431157261133194,
0.03853112459182739,
-0.022197717800736427,
0.09925086796283722,
0.05696196109056473,
0.05017116665840149,
0.09158790111541748,
-0.23996101319789886,
0.00031927876989357173,
0.1233425959944725,
0.08184559643268585,
-0.10171791166067123,
-0.021577900275588036,
0.03223108500242233,
0.02601911500096321,
0.0010762816527858377,
0.02183089405298233,
-0.010006583295762539,
0.21613149344921112,
-0.05803706869482994,
-0.1130337119102478,
-0.03479117527604103,
0.0580526627600193,
0.06587925553321838,
-0.014316205866634846,
-0.12216382473707199,
-0.07942970097064972,
-0.07798335701227188,
-0.0326395221054554,
-0.02570279873907566,
0.021269816905260086,
-0.05102776363492012,
0.09634225070476532,
-0.03619711101055145,
-0.056511081755161285,
-0.00876806490123272,
-0.06577249616384506,
0.0777861624956131,
0.036285512149333954,
0.01751028373837471,
-0.031084321439266205,
0.06809975951910019,
-0.0890253484249115,
-0.12620246410369873,
-0.006816851440817118,
-0.04581926390528679,
-0.0763077363371849,
-0.04577818140387535,
-0.001394819119013846,
-0.06413495540618896,
0.06751624494791031,
0.1753678172826767,
-0.11126448214054108,
0.03799479082226753,
0.05996380001306534,
0.012898536399006844,
0.06565160304307938,
0.0714573785662651,
-0.13876697421073914,
-0.11260410398244858,
0.008954460732638836,
0.06644444167613983,
0.019819015637040138,
0.03325645998120308,
-0.06100134178996086,
0.024548521265387535,
-0.013515850529074669,
0.03324826806783676,
0.0687195286154747,
0.051566507667303085,
-0.08421696722507477,
-0.0751957893371582,
0.2143598049879074,
-0.10335041582584381,
0.033092107623815536,
0.010988671332597733,
-0.026210401207208633,
-0.004537034314125776,
0.059370871633291245,
-0.005852478090673685,
-0.006877345032989979,
0.0782134160399437,
-0.054428648203611374,
-0.04517027735710144,
-0.06959634274244308,
-0.11619371920824051,
0.01717451587319374,
-0.06540634483098984,
-0.03562006726861,
-0.10333813726902008,
-0.1127791777253151,
-0.08173860609531403,
0.0225505568087101,
-0.08850936591625214,
-0.003282109508290887,
-0.017249062657356262,
-0.014816144481301308,
0.030136849731206894,
0.020879454910755157,
0.010426436550915241,
-0.005551993381232023,
-0.006466520484536886,
-0.025110576301813126,
0.07213345170021057,
-0.016297413036227226,
0.020467659458518028,
-0.05047497898340225,
0.1219911202788353,
-0.1371684968471527,
0.09609949588775635,
-0.03347732126712799,
0.03595913574099541,
-0.175568088889122,
-0.017098214477300644,
-0.0037915981374680996,
-0.004533144645392895,
0.058005981147289276,
0.14057965576648712,
-0.11926346272230148,
-0.06140173226594925,
0.13289830088615417,
-0.09801438450813293,
-0.10742739588022232,
0.053397271782159805,
0.018955418840050697,
0.05846787989139557,
0.03089308924973011,
0.14694742858409882,
0.13983753323554993,
-0.050550125539302826,
-0.08870300650596619,
-0.07149381190538406,
0.010982734151184559,
0.06816384196281433,
0.03878210112452507,
-0.05193159356713295,
0.004449300467967987,
-0.0011385118123143911,
0.04240284115076065,
0.012861059047281742,
-0.05326338857412338,
-0.052353788167238235,
-0.03655942529439926,
-0.0763717070221901,
0.04428849369287491,
-0.043531399220228195,
-0.0027862058486789465,
-0.040726691484451294,
-0.06982230395078659,
-0.03279690816998482,
0.12857794761657715,
-0.019024476408958435,
-0.02646280825138092,
-0.10369996726512909,
0.12207561731338501,
-0.05155794322490692,
0.027999071404337883,
-0.13328449428081512,
-0.07873210310935974,
0.013290425762534142,
-0.14880774915218353,
0.020823977887630463,
-0.048385124653577805,
0.06452542543411255,
0.04076433554291725,
-0.020555315539240837,
-0.060692209750413895,
0.027924058958888054,
0.03729479759931564,
-0.014081302098929882,
-0.17167198657989502,
-0.1198735460639,
-0.03925011679530144,
0.18057720363140106,
-0.06726565212011337,
0.03567621856927872,
0.01323137991130352,
0.20705296099185944,
-0.022352322936058044,
-0.03176422417163849,
0.05123317986726761,
0.00014905599527992308,
-0.011938521638512611,
-0.037665143609046936,
0.04924456775188446,
-0.015136686153709888,
-0.09850086271762848,
0.10377045720815659,
-0.12942194938659668,
-0.059162262827157974,
0.07564891874790192,
0.09444808214902878,
-0.10549592226743698,
0.008000001311302185,
-0.03415174037218094,
-0.032276351004838943,
0.02169433794915676,
-0.09798052161931992,
0.06620953232049942,
0.04097268730401993,
0.08982136100530624,
-0.031179966405034065,
-0.05951347202062607,
0.027041597291827202,
-0.010810350999236107,
-0.0486951619386673,
0.11888406425714493,
0.06309294700622559,
-0.26830580830574036,
0.08573032915592194,
0.1188964769244194,
0.12702177464962006,
0.07135380804538727,
0.01665486767888069,
-0.03582458943128586,
-0.0969366729259491,
-0.002709203865379095,
0.019314058125019073,
0.039211977273225784,
-0.00338642462156713,
0.04723910242319107,
0.06927380710840225,
-0.004969681613147259,
0.019738275557756424,
-0.08276385813951492,
0.05426163971424103,
0.05641549825668335,
0.002137552248314023,
0.044767070561647415,
0.06764590740203857,
0.046978939324617386,
0.07235992699861526,
0.030252300202846527,
0.02796664647758007,
-0.02687324397265911,
-0.02462850697338581,
-0.09307828545570374,
0.16269123554229736,
-0.11206965893507004,
-0.19840660691261292,
-0.15334245562553406,
-0.08850444853305817,
-0.027759697288274765,
-0.03327086940407753,
0.02350410260260105,
-0.045381851494312286,
-0.10191641747951508,
-0.07629477977752686,
0.11286745965480804,
0.021329881623387337,
-0.04006670042872429,
0.019428323954343796,
0.008793728426098824,
0.05060252174735069,
-0.09476018697023392,
-0.03588350862264633,
0.036244116723537445,
-0.0065659708343446255,
0.032573360949754715,
0.02478238381445408,
0.06866764277219772,
0.10987551510334015,
-0.013020339421927929,
-0.013224332593381405,
0.010266613215208054,
0.26109960675239563,
-0.06792005896568298,
0.11300449073314667,
0.1701001673936844,
-0.04405709356069565,
0.062038447707891464,
0.20398791134357452,
0.03727400675415993,
-0.057575833052396774,
-0.02016502246260643,
0.00021238542103674263,
-0.012800226919353008,
-0.22262652218341827,
-0.11957675963640213,
-0.034899547696113586,
-0.01544161420315504,
0.07633636146783829,
0.040822066366672516,
0.000014217610441846773,
0.04809662699699402,
-0.06699933856725693,
-0.0573849156498909,
-0.029192252084612846,
0.05727338418364525,
0.1841459423303604,
0.007880998775362968,
0.07886064052581787,
-0.041331615298986435,
0.007868589833378792,
0.0740891695022583,
0.06255767494440079,
0.11734528094530106,
0.07825468480587006,
0.16789866983890533,
0.10007377713918686,
0.08556236326694489,
0.024385375902056694,
0.014786798506975174,
0.027709078043699265,
0.021460413932800293,
-0.006975389085710049,
-0.07643530517816544,
0.038758520036935806,
0.05182918161153793,
0.031463779509067535,
0.03801362216472626,
-0.0834122896194458,
-0.016985608264803886,
0.03483058139681816,
0.16207928955554962,
0.14048004150390625,
-0.3005506098270416,
-0.08323585987091064,
0.025416415184736252,
0.01383188646286726,
-0.05232508108019829,
-0.06308016926050186,
-0.02035432681441307,
-0.11093750596046448,
0.12947995960712433,
-0.022370731458067894,
0.07623643428087234,
-0.0764886885881424,
-0.01630675606429577,
0.05231771990656853,
0.10738098621368408,
0.005896391812711954,
0.01062813587486744,
-0.02391277439892292,
0.20102538168430328,
0.05562681704759598,
-0.0194980688393116,
0.002547018462792039,
0.06118375062942505,
0.017842551693320274,
0.07136592268943787,
0.10077043622732162,
0.00612291693687439,
-0.03840649500489235,
-0.09589219838380814,
-0.12944872677326202,
-0.03843344375491142,
0.05491048842668533,
-0.10923520475625992,
0.11711625009775162,
-0.014799851924180984,
-0.08156831562519073,
-0.04614942520856857,
0.05058536306023598,
-0.08898196369409561,
-0.11923813074827194,
0.07722939550876617,
0.023040933534502983,
-0.018077459186315536,
-0.06456870585680008,
-0.03413100913167,
-0.1429545134305954,
0.26264381408691406,
0.035666678100824356,
-0.09941735863685608,
-0.11226937919855118,
0.0347491130232811,
0.1675538420677185,
-0.06691380590200424,
0.07872165739536285,
-0.047349750995635986,
0.08966491371393204,
-0.05528796464204788,
-0.12725390493869781,
0.058913927525281906,
-0.10016945749521255,
-0.12759296596050262,
-0.03961240127682686,
0.18509848415851593,
-0.0031884934287518263,
0.029641784727573395,
0.01654040440917015,
0.043276071548461914,
-0.009256722405552864,
-0.043235424906015396,
0.05224913731217384,
0.16106700897216797,
0.03871533274650574,
0.06993894279003143,
-0.047677069902420044,
-0.12623842060565948,
-0.07358108460903168,
-0.045047566294670105,
0.09031318128108978,
0.28172221779823303,
-0.08059272170066833,
0.0703277662396431,
0.04183144122362137,
-0.06414999812841415,
-0.16677998006343842,
-0.03321857750415802,
0.04443689063191414,
0.05941549316048622,
0.08734254539012909,
-0.06102360412478447,
0.0194100309163332,
0.0518232136964798,
-0.024931998923420906,
0.09124879539012909,
-0.31562092900276184,
-0.13086602091789246,
0.012772644869983196,
0.03720805421471596,
-0.03543737530708313,
-0.11141787469387054,
-0.10424143075942993,
-0.029222222045063972,
-0.22971786558628082,
0.05031852051615715,
0.0398586206138134,
0.06705270707607269,
-0.007538771256804466,
0.06800203770399094,
0.040483906865119934,
-0.04256002977490425,
0.15531939268112183,
-0.03938347101211548,
-0.005822100676596165,
-0.05803919956088066,
-0.0028284534346312284,
0.04534327611327171,
-0.06580544263124466,
0.05209860950708389,
0.007137274369597435,
0.05925822630524635,
-0.07757066935300827,
-0.024678707122802734,
-0.06486526131629944,
0.03214328736066818,
-0.05143289268016815,
-0.021723657846450806,
-0.08868027478456497,
0.07818123698234558,
0.06659244745969772,
-0.022818149998784065,
-0.0018113115802407265,
-0.04676906764507294,
0.0822516679763794,
0.20029807090759277,
0.08310268074274063,
0.013505170121788979,
-0.09094647318124771,
0.008243824355304241,
-0.05452888458967209,
0.014529251493513584,
-0.08473841845989227,
0.00394476018846035,
0.1254677027463913,
0.0049360343255102634,
0.08779989182949066,
-0.01823553256690502,
-0.10659253597259521,
-0.03706494718790054,
0.08686766028404236,
-0.07888077199459076,
-0.23521585762500763,
-0.0117771802470088,
0.05145197734236717,
-0.08659669011831284,
0.006848771590739489,
0.18027520179748535,
-0.007439441978931427,
-0.052460767328739166,
0.03836945816874504,
0.030965223908424377,
-0.04342254623770714,
0.14217816293239594,
-0.0177037063986063,
0.06870051473379135,
-0.0893486738204956,
0.07082530111074448,
0.07169808447360992,
-0.07319361716508865,
0.017374640330672264,
0.08868568390607834,
-0.08009906858205795,
-0.05292263254523277,
-0.0856100469827652,
0.07195313274860382,
-0.06030732020735741,
-0.022959688678383827,
-0.048205066472291946,
-0.10104013979434967,
0.02251918986439705,
0.08667980134487152,
0.02457340434193611,
0.02173803374171257,
0.04739495366811752,
-0.06242978200316429,
-0.07013832032680511,
0.08441106975078583,
0.04875132814049721,
0.0726483091711998,
-0.09258172661066055,
0.04837319627404213,
-0.03263453394174576,
0.03473187983036041,
-0.011538718827068806,
0.03207414224743843,
-0.08890228718519211,
-0.04945969954133034,
-0.09284160286188126,
-0.013561494648456573,
-0.1047433614730835,
-0.03561164066195488,
-0.0075886365957558155,
0.00452903239056468,
-0.023436063900589943,
0.006245793309062719,
-0.06833038479089737,
-0.10528631508350372,
-0.03492012992501259,
0.0675821527838707,
-0.07653367519378662,
-0.0202096626162529,
0.025652024894952774,
-0.06017875298857689,
0.05335857346653938,
-0.019051995128393173,
0.018273908644914627,
-0.0521397702395916,
-0.0817580297589302,
-0.05428663268685341,
0.02644684538245201,
0.014924325980246067,
0.054286643862724304,
-0.2144417017698288,
-0.00958106480538845,
-0.023239130154252052,
-0.11878415197134018,
-0.025227610021829605,
-0.006108676083385944,
-0.08585220575332642,
0.004315250087529421,
-0.033441804349422455,
0.03169618919491768,
-0.03416713327169418,
-0.01910240948200226,
0.022437769919633865,
-0.0013304915046319366,
0.08398790657520294,
-0.037253253161907196,
0.06111537665128708,
-0.20021264255046844,
-0.02767246589064598,
-0.043094076216220856,
-0.03537527844309807,
-0.031026748940348625,
-0.039771344512701035,
0.059734586626291275,
-0.028465555980801582,
0.08480904251337051,
-0.029318079352378845,
-0.010752764530479908,
0.040197622030973434,
-0.006120410282164812,
-0.05087416246533394,
0.023314209654927254,
0.10412739962339401,
0.00528274429962039,
-0.00010189160093432292,
-0.004153978079557419,
0.022853892296552658,
-0.03377161920070648,
-0.030008016154170036,
0.1058531403541565,
0.08210869878530502,
0.07707115262746811,
0.03405280411243439,
0.11108583956956863,
-0.05770613253116608,
-0.1038203313946724,
-0.0032150817569345236,
-0.03761368244886398,
0.0912533849477768,
-0.03745587170124054,
0.08848442137241364,
0.08855361491441727,
-0.15082921087741852,
0.10846332460641861,
-0.017351828515529633,
-0.010859545320272446,
-0.07851476222276688,
-0.1239314079284668,
-0.09908819943666458,
-0.07865975052118301,
-0.014563309028744698,
-0.07968620955944061,
0.03894209489226341,
-0.03842732682824135,
0.008507492952048779,
-0.017026571556925774,
0.12609393894672394,
-0.05866537243127823,
-0.02455347590148449,
0.02295229770243168,
0.04421736299991608,
-0.007808969356119633,
-0.0007919921772554517,
-0.0010074416641145945,
0.04876793548464775,
0.06317339092493057,
0.016541825607419014,
0.018818682059645653,
-0.0016199585516005754,
0.04239792004227638,
-0.005740095395594835,
-0.11427447199821472,
-0.009536593221127987,
0.0260044876486063,
0.03104775957763195,
0.03516622632741928,
0.036980461329221725,
0.011867043562233448,
-0.038300104439258575,
0.0863136351108551,
-0.015535220503807068,
-0.040255703032016754,
-0.10217709094285965,
0.1794138103723526,
-0.05375659838318825,
-0.0033138622529804707,
0.046202417463064194,
-0.09528232365846634,
-0.0002491216582711786,
0.14526456594467163,
0.24354273080825806,
-0.002369426190853119,
-0.008058322593569756,
0.019230477511882782,
0.01697108894586563,
-0.013749241828918457,
0.06014157459139824,
0.048449259251356125,
0.11442190408706665,
-0.055674608796834946,
0.04904584586620331,
-0.029404308646917343,
-0.054910171777009964,
-0.023619534447789192,
0.026049857959151268,
0.009405882097780704,
0.007606208324432373,
0.042983558028936386,
0.10001488775014877,
-0.024769606068730354,
-0.14748038351535797,
0.08121319115161896,
-0.16974031925201416,
-0.14832109212875366,
-0.047739796340465546,
0.029321080073714256,
0.020948179066181183,
0.07536580413579941,
-0.03298107534646988,
-0.03171062842011452,
0.20358532667160034,
-0.028074055910110474,
-0.05899084731936455,
-0.09832312911748886,
0.020441092550754547,
-0.09668944776058197,
0.13841167092323303,
-0.008817383088171482,
-0.00542556494474411,
0.12228859215974808,
-0.021948225796222687,
-0.15407922863960266,
0.022388650104403496,
0.05838823318481445,
-0.06662013381719589,
0.07426796853542328,
0.12037602066993713,
-0.0008322051144205034,
0.07671105861663818,
0.020938027650117874,
-0.1675398051738739,
0.03367368504405022,
0.01624918170273304,
0.0028204803820699453,
-0.04721074551343918,
0.09156905859708786,
-0.06716921180486679,
0.1614406257867813,
0.19669775664806366,
-0.05243987590074539,
0.006645268760621548,
-0.005850786808878183,
0.060378286987543106,
0.06362046301364899,
0.1139972135424614,
-0.026050040498375893,
-0.1806458681821823,
0.05474412068724632,
-0.027082225307822227,
0.06820669770240784,
-0.1952790468931198,
-0.09679047763347626,
-0.04316534474492073,
-0.026921940967440605,
-0.059101399034261703,
0.12512192130088806,
0.07807399332523346,
0.039878591895103455,
-0.025211311876773834,
-0.11121244728565216,
0.0005876116920262575,
0.11113984882831573,
-0.12531879544258118,
-0.0628880113363266
] |
null | null | transformers | # RPmerge
See the main model card: https://huggingface.co/brucethemoose/Yi-34B-200K-RPMerge
Quantized with default exl2 quantization, still investigating the benefits/drawbacks of long context (32K) quantization.
## Merge Details
### Merge Method
This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using /home/alpha/Models/Raw/chargoddard_Yi-34B-200K-Llama as a base.
### Models Merged
The following models were included in the merge:
* /home/alpha/Models/Raw/migtissera_Tess-34B-v1.5b
* /home/alpha/Models/Raw/migtissera_Tess-M-Creative-v1.0
* /home/alpha/Models/Raw/cgato_Thespis-34b-DPO-v0.7
* /home/alpha/Models/Raw/Nous-Capybara-34B
* /home/alpha/Models/Raw/admo_limarp
* /home/alpha/Models/Raw/DrNicefellow_ChatAllInOne-Yi-34B-200K-V1
### Configuration
The following YAML configuration was used to produce this model:
```yaml
models:
- model: /home/alpha/Models/Raw/chargoddard_Yi-34B-200K-Llama
# No parameters necessary for base model
- model: /home/alpha/Models/Raw/migtissera_Tess-34B-v1.5b
#Emphasize the beginning of Vicuna format models
parameters:
weight: 0.19
density: 0.59
- model: /home/alpha/Models/Raw/Nous-Capybara-34B
parameters:
weight: 0.19
density: 0.55
# Vicuna format
- model: /home/alpha/Models/Raw/migtissera_Tess-M-Creative-v1.0
parameters:
weight: 0.05
density: 0.55
- model: /home/alpha/Models/Raw/DrNicefellow_ChatAllInOne-Yi-34B-200K-V1
parameters:
weight: 0.19
density: 0.55
- model: /home/alpha/Models/Raw/admo_limarp
parameters:
weight: 0.19
density: 0.48
- model: /home/alpha/Models/Raw/cgato_Thespis-34b-DPO-v0.7
parameters:
weight: 0.19
density: 0.59
merge_method: dare_ties
tokenizer_source: union
base_model: /home/alpha/Models/Raw/chargoddard_Yi-34B-200K-Llama
parameters:
int8_mask: true
dtype: bfloat16
```
| {"language": ["en"], "license": "other", "library_name": "transformers", "tags": ["mergekit", "merge", "Yi", "exllama", "exllamav2", "exl2"], "license_name": "yi-license", "license_link": "https://huggingface.co/01-ai/Yi-34B/blob/main/LICENSE", "base_model": []} | text-generation | brucethemoose/Yi-34B-200K-RPMerge-exl2-40bpw | [
"transformers",
"safetensors",
"llama",
"text-generation",
"mergekit",
"merge",
"Yi",
"exllama",
"exllamav2",
"exl2",
"en",
"arxiv:2311.03099",
"arxiv:2306.01708",
"license:other",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T06:21:51+00:00 | [
"2311.03099",
"2306.01708"
] | [
"en"
] | TAGS
#transformers #safetensors #llama #text-generation #mergekit #merge #Yi #exllama #exllamav2 #exl2 #en #arxiv-2311.03099 #arxiv-2306.01708 #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # RPmerge
See the main model card: URL
Quantized with default exl2 quantization, still investigating the benefits/drawbacks of long context (32K) quantization.
## Merge Details
### Merge Method
This model was merged using the DARE TIES merge method using /home/alpha/Models/Raw/chargoddard_Yi-34B-200K-Llama as a base.
### Models Merged
The following models were included in the merge:
* /home/alpha/Models/Raw/migtissera_Tess-34B-v1.5b
* /home/alpha/Models/Raw/migtissera_Tess-M-Creative-v1.0
* /home/alpha/Models/Raw/cgato_Thespis-34b-DPO-v0.7
* /home/alpha/Models/Raw/Nous-Capybara-34B
* /home/alpha/Models/Raw/admo_limarp
* /home/alpha/Models/Raw/DrNicefellow_ChatAllInOne-Yi-34B-200K-V1
### Configuration
The following YAML configuration was used to produce this model:
| [
"# RPmerge\n\nSee the main model card: URL\n\nQuantized with default exl2 quantization, still investigating the benefits/drawbacks of long context (32K) quantization.",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the DARE TIES merge method using /home/alpha/Models/Raw/chargoddard_Yi-34B-200K-Llama as a base.",
"### Models Merged\n\nThe following models were included in the merge:\n* /home/alpha/Models/Raw/migtissera_Tess-34B-v1.5b\n* /home/alpha/Models/Raw/migtissera_Tess-M-Creative-v1.0\n* /home/alpha/Models/Raw/cgato_Thespis-34b-DPO-v0.7\n* /home/alpha/Models/Raw/Nous-Capybara-34B\n* /home/alpha/Models/Raw/admo_limarp\n* /home/alpha/Models/Raw/DrNicefellow_ChatAllInOne-Yi-34B-200K-V1",
"### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #mergekit #merge #Yi #exllama #exllamav2 #exl2 #en #arxiv-2311.03099 #arxiv-2306.01708 #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# RPmerge\n\nSee the main model card: URL\n\nQuantized with default exl2 quantization, still investigating the benefits/drawbacks of long context (32K) quantization.",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the DARE TIES merge method using /home/alpha/Models/Raw/chargoddard_Yi-34B-200K-Llama as a base.",
"### Models Merged\n\nThe following models were included in the merge:\n* /home/alpha/Models/Raw/migtissera_Tess-34B-v1.5b\n* /home/alpha/Models/Raw/migtissera_Tess-M-Creative-v1.0\n* /home/alpha/Models/Raw/cgato_Thespis-34b-DPO-v0.7\n* /home/alpha/Models/Raw/Nous-Capybara-34B\n* /home/alpha/Models/Raw/admo_limarp\n* /home/alpha/Models/Raw/DrNicefellow_ChatAllInOne-Yi-34B-200K-V1",
"### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
95,
40,
4,
49,
169,
17
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #mergekit #merge #Yi #exllama #exllamav2 #exl2 #en #arxiv-2311.03099 #arxiv-2306.01708 #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# RPmerge\n\nSee the main model card: URL\n\nQuantized with default exl2 quantization, still investigating the benefits/drawbacks of long context (32K) quantization.## Merge Details### Merge Method\n\nThis model was merged using the DARE TIES merge method using /home/alpha/Models/Raw/chargoddard_Yi-34B-200K-Llama as a base.### Models Merged\n\nThe following models were included in the merge:\n* /home/alpha/Models/Raw/migtissera_Tess-34B-v1.5b\n* /home/alpha/Models/Raw/migtissera_Tess-M-Creative-v1.0\n* /home/alpha/Models/Raw/cgato_Thespis-34b-DPO-v0.7\n* /home/alpha/Models/Raw/Nous-Capybara-34B\n* /home/alpha/Models/Raw/admo_limarp\n* /home/alpha/Models/Raw/DrNicefellow_ChatAllInOne-Yi-34B-200K-V1### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
-0.09143144637346268,
0.13088233768939972,
-0.004239476285874844,
0.007002714555710554,
0.07611428201198578,
0.0688571110367775,
0.13210566341876984,
0.13319514691829681,
0.018930861726403236,
0.1146661564707756,
-0.004489467944949865,
0.11506475508213043,
0.11725322157144547,
0.1490548998117447,
-0.009691251441836357,
-0.13874879479408264,
0.043344542384147644,
-0.04330842196941376,
-0.06558141857385635,
0.09411337226629257,
0.09026461094617844,
-0.07027081400156021,
0.10735927522182465,
0.06902097910642624,
-0.09375990182161331,
0.0003966365766245872,
-0.04667649045586586,
0.02636537328362465,
0.07864397019147873,
0.07433581352233887,
0.06466394662857056,
0.05696464702486992,
0.056630536913871765,
-0.14254747331142426,
0.018034514039754868,
0.023142904043197632,
-0.000902830739505589,
0.06992360949516296,
0.10111942887306213,
-0.04483991488814354,
0.10474897176027298,
-0.08026403188705444,
0.03382281959056854,
0.07870031893253326,
-0.09348627179861069,
-0.18507787585258484,
-0.15320031344890594,
0.11755742132663727,
0.09431157261133194,
0.03853112459182739,
-0.022197717800736427,
0.09925086796283722,
0.05696196109056473,
0.05017116665840149,
0.09158790111541748,
-0.23996101319789886,
0.00031927876989357173,
0.1233425959944725,
0.08184559643268585,
-0.10171791166067123,
-0.021577900275588036,
0.03223108500242233,
0.02601911500096321,
0.0010762816527858377,
0.02183089405298233,
-0.010006583295762539,
0.21613149344921112,
-0.05803706869482994,
-0.1130337119102478,
-0.03479117527604103,
0.0580526627600193,
0.06587925553321838,
-0.014316205866634846,
-0.12216382473707199,
-0.07942970097064972,
-0.07798335701227188,
-0.0326395221054554,
-0.02570279873907566,
0.021269816905260086,
-0.05102776363492012,
0.09634225070476532,
-0.03619711101055145,
-0.056511081755161285,
-0.00876806490123272,
-0.06577249616384506,
0.0777861624956131,
0.036285512149333954,
0.01751028373837471,
-0.031084321439266205,
0.06809975951910019,
-0.0890253484249115,
-0.12620246410369873,
-0.006816851440817118,
-0.04581926390528679,
-0.0763077363371849,
-0.04577818140387535,
-0.001394819119013846,
-0.06413495540618896,
0.06751624494791031,
0.1753678172826767,
-0.11126448214054108,
0.03799479082226753,
0.05996380001306534,
0.012898536399006844,
0.06565160304307938,
0.0714573785662651,
-0.13876697421073914,
-0.11260410398244858,
0.008954460732638836,
0.06644444167613983,
0.019819015637040138,
0.03325645998120308,
-0.06100134178996086,
0.024548521265387535,
-0.013515850529074669,
0.03324826806783676,
0.0687195286154747,
0.051566507667303085,
-0.08421696722507477,
-0.0751957893371582,
0.2143598049879074,
-0.10335041582584381,
0.033092107623815536,
0.010988671332597733,
-0.026210401207208633,
-0.004537034314125776,
0.059370871633291245,
-0.005852478090673685,
-0.006877345032989979,
0.0782134160399437,
-0.054428648203611374,
-0.04517027735710144,
-0.06959634274244308,
-0.11619371920824051,
0.01717451587319374,
-0.06540634483098984,
-0.03562006726861,
-0.10333813726902008,
-0.1127791777253151,
-0.08173860609531403,
0.0225505568087101,
-0.08850936591625214,
-0.003282109508290887,
-0.017249062657356262,
-0.014816144481301308,
0.030136849731206894,
0.020879454910755157,
0.010426436550915241,
-0.005551993381232023,
-0.006466520484536886,
-0.025110576301813126,
0.07213345170021057,
-0.016297413036227226,
0.020467659458518028,
-0.05047497898340225,
0.1219911202788353,
-0.1371684968471527,
0.09609949588775635,
-0.03347732126712799,
0.03595913574099541,
-0.175568088889122,
-0.017098214477300644,
-0.0037915981374680996,
-0.004533144645392895,
0.058005981147289276,
0.14057965576648712,
-0.11926346272230148,
-0.06140173226594925,
0.13289830088615417,
-0.09801438450813293,
-0.10742739588022232,
0.053397271782159805,
0.018955418840050697,
0.05846787989139557,
0.03089308924973011,
0.14694742858409882,
0.13983753323554993,
-0.050550125539302826,
-0.08870300650596619,
-0.07149381190538406,
0.010982734151184559,
0.06816384196281433,
0.03878210112452507,
-0.05193159356713295,
0.004449300467967987,
-0.0011385118123143911,
0.04240284115076065,
0.012861059047281742,
-0.05326338857412338,
-0.052353788167238235,
-0.03655942529439926,
-0.0763717070221901,
0.04428849369287491,
-0.043531399220228195,
-0.0027862058486789465,
-0.040726691484451294,
-0.06982230395078659,
-0.03279690816998482,
0.12857794761657715,
-0.019024476408958435,
-0.02646280825138092,
-0.10369996726512909,
0.12207561731338501,
-0.05155794322490692,
0.027999071404337883,
-0.13328449428081512,
-0.07873210310935974,
0.013290425762534142,
-0.14880774915218353,
0.020823977887630463,
-0.048385124653577805,
0.06452542543411255,
0.04076433554291725,
-0.020555315539240837,
-0.060692209750413895,
0.027924058958888054,
0.03729479759931564,
-0.014081302098929882,
-0.17167198657989502,
-0.1198735460639,
-0.03925011679530144,
0.18057720363140106,
-0.06726565212011337,
0.03567621856927872,
0.01323137991130352,
0.20705296099185944,
-0.022352322936058044,
-0.03176422417163849,
0.05123317986726761,
0.00014905599527992308,
-0.011938521638512611,
-0.037665143609046936,
0.04924456775188446,
-0.015136686153709888,
-0.09850086271762848,
0.10377045720815659,
-0.12942194938659668,
-0.059162262827157974,
0.07564891874790192,
0.09444808214902878,
-0.10549592226743698,
0.008000001311302185,
-0.03415174037218094,
-0.032276351004838943,
0.02169433794915676,
-0.09798052161931992,
0.06620953232049942,
0.04097268730401993,
0.08982136100530624,
-0.031179966405034065,
-0.05951347202062607,
0.027041597291827202,
-0.010810350999236107,
-0.0486951619386673,
0.11888406425714493,
0.06309294700622559,
-0.26830580830574036,
0.08573032915592194,
0.1188964769244194,
0.12702177464962006,
0.07135380804538727,
0.01665486767888069,
-0.03582458943128586,
-0.0969366729259491,
-0.002709203865379095,
0.019314058125019073,
0.039211977273225784,
-0.00338642462156713,
0.04723910242319107,
0.06927380710840225,
-0.004969681613147259,
0.019738275557756424,
-0.08276385813951492,
0.05426163971424103,
0.05641549825668335,
0.002137552248314023,
0.044767070561647415,
0.06764590740203857,
0.046978939324617386,
0.07235992699861526,
0.030252300202846527,
0.02796664647758007,
-0.02687324397265911,
-0.02462850697338581,
-0.09307828545570374,
0.16269123554229736,
-0.11206965893507004,
-0.19840660691261292,
-0.15334245562553406,
-0.08850444853305817,
-0.027759697288274765,
-0.03327086940407753,
0.02350410260260105,
-0.045381851494312286,
-0.10191641747951508,
-0.07629477977752686,
0.11286745965480804,
0.021329881623387337,
-0.04006670042872429,
0.019428323954343796,
0.008793728426098824,
0.05060252174735069,
-0.09476018697023392,
-0.03588350862264633,
0.036244116723537445,
-0.0065659708343446255,
0.032573360949754715,
0.02478238381445408,
0.06866764277219772,
0.10987551510334015,
-0.013020339421927929,
-0.013224332593381405,
0.010266613215208054,
0.26109960675239563,
-0.06792005896568298,
0.11300449073314667,
0.1701001673936844,
-0.04405709356069565,
0.062038447707891464,
0.20398791134357452,
0.03727400675415993,
-0.057575833052396774,
-0.02016502246260643,
0.00021238542103674263,
-0.012800226919353008,
-0.22262652218341827,
-0.11957675963640213,
-0.034899547696113586,
-0.01544161420315504,
0.07633636146783829,
0.040822066366672516,
0.000014217610441846773,
0.04809662699699402,
-0.06699933856725693,
-0.0573849156498909,
-0.029192252084612846,
0.05727338418364525,
0.1841459423303604,
0.007880998775362968,
0.07886064052581787,
-0.041331615298986435,
0.007868589833378792,
0.0740891695022583,
0.06255767494440079,
0.11734528094530106,
0.07825468480587006,
0.16789866983890533,
0.10007377713918686,
0.08556236326694489,
0.024385375902056694,
0.014786798506975174,
0.027709078043699265,
0.021460413932800293,
-0.006975389085710049,
-0.07643530517816544,
0.038758520036935806,
0.05182918161153793,
0.031463779509067535,
0.03801362216472626,
-0.0834122896194458,
-0.016985608264803886,
0.03483058139681816,
0.16207928955554962,
0.14048004150390625,
-0.3005506098270416,
-0.08323585987091064,
0.025416415184736252,
0.01383188646286726,
-0.05232508108019829,
-0.06308016926050186,
-0.02035432681441307,
-0.11093750596046448,
0.12947995960712433,
-0.022370731458067894,
0.07623643428087234,
-0.0764886885881424,
-0.01630675606429577,
0.05231771990656853,
0.10738098621368408,
0.005896391812711954,
0.01062813587486744,
-0.02391277439892292,
0.20102538168430328,
0.05562681704759598,
-0.0194980688393116,
0.002547018462792039,
0.06118375062942505,
0.017842551693320274,
0.07136592268943787,
0.10077043622732162,
0.00612291693687439,
-0.03840649500489235,
-0.09589219838380814,
-0.12944872677326202,
-0.03843344375491142,
0.05491048842668533,
-0.10923520475625992,
0.11711625009775162,
-0.014799851924180984,
-0.08156831562519073,
-0.04614942520856857,
0.05058536306023598,
-0.08898196369409561,
-0.11923813074827194,
0.07722939550876617,
0.023040933534502983,
-0.018077459186315536,
-0.06456870585680008,
-0.03413100913167,
-0.1429545134305954,
0.26264381408691406,
0.035666678100824356,
-0.09941735863685608,
-0.11226937919855118,
0.0347491130232811,
0.1675538420677185,
-0.06691380590200424,
0.07872165739536285,
-0.047349750995635986,
0.08966491371393204,
-0.05528796464204788,
-0.12725390493869781,
0.058913927525281906,
-0.10016945749521255,
-0.12759296596050262,
-0.03961240127682686,
0.18509848415851593,
-0.0031884934287518263,
0.029641784727573395,
0.01654040440917015,
0.043276071548461914,
-0.009256722405552864,
-0.043235424906015396,
0.05224913731217384,
0.16106700897216797,
0.03871533274650574,
0.06993894279003143,
-0.047677069902420044,
-0.12623842060565948,
-0.07358108460903168,
-0.045047566294670105,
0.09031318128108978,
0.28172221779823303,
-0.08059272170066833,
0.0703277662396431,
0.04183144122362137,
-0.06414999812841415,
-0.16677998006343842,
-0.03321857750415802,
0.04443689063191414,
0.05941549316048622,
0.08734254539012909,
-0.06102360412478447,
0.0194100309163332,
0.0518232136964798,
-0.024931998923420906,
0.09124879539012909,
-0.31562092900276184,
-0.13086602091789246,
0.012772644869983196,
0.03720805421471596,
-0.03543737530708313,
-0.11141787469387054,
-0.10424143075942993,
-0.029222222045063972,
-0.22971786558628082,
0.05031852051615715,
0.0398586206138134,
0.06705270707607269,
-0.007538771256804466,
0.06800203770399094,
0.040483906865119934,
-0.04256002977490425,
0.15531939268112183,
-0.03938347101211548,
-0.005822100676596165,
-0.05803919956088066,
-0.0028284534346312284,
0.04534327611327171,
-0.06580544263124466,
0.05209860950708389,
0.007137274369597435,
0.05925822630524635,
-0.07757066935300827,
-0.024678707122802734,
-0.06486526131629944,
0.03214328736066818,
-0.05143289268016815,
-0.021723657846450806,
-0.08868027478456497,
0.07818123698234558,
0.06659244745969772,
-0.022818149998784065,
-0.0018113115802407265,
-0.04676906764507294,
0.0822516679763794,
0.20029807090759277,
0.08310268074274063,
0.013505170121788979,
-0.09094647318124771,
0.008243824355304241,
-0.05452888458967209,
0.014529251493513584,
-0.08473841845989227,
0.00394476018846035,
0.1254677027463913,
0.0049360343255102634,
0.08779989182949066,
-0.01823553256690502,
-0.10659253597259521,
-0.03706494718790054,
0.08686766028404236,
-0.07888077199459076,
-0.23521585762500763,
-0.0117771802470088,
0.05145197734236717,
-0.08659669011831284,
0.006848771590739489,
0.18027520179748535,
-0.007439441978931427,
-0.052460767328739166,
0.03836945816874504,
0.030965223908424377,
-0.04342254623770714,
0.14217816293239594,
-0.0177037063986063,
0.06870051473379135,
-0.0893486738204956,
0.07082530111074448,
0.07169808447360992,
-0.07319361716508865,
0.017374640330672264,
0.08868568390607834,
-0.08009906858205795,
-0.05292263254523277,
-0.0856100469827652,
0.07195313274860382,
-0.06030732020735741,
-0.022959688678383827,
-0.048205066472291946,
-0.10104013979434967,
0.02251918986439705,
0.08667980134487152,
0.02457340434193611,
0.02173803374171257,
0.04739495366811752,
-0.06242978200316429,
-0.07013832032680511,
0.08441106975078583,
0.04875132814049721,
0.0726483091711998,
-0.09258172661066055,
0.04837319627404213,
-0.03263453394174576,
0.03473187983036041,
-0.011538718827068806,
0.03207414224743843,
-0.08890228718519211,
-0.04945969954133034,
-0.09284160286188126,
-0.013561494648456573,
-0.1047433614730835,
-0.03561164066195488,
-0.0075886365957558155,
0.00452903239056468,
-0.023436063900589943,
0.006245793309062719,
-0.06833038479089737,
-0.10528631508350372,
-0.03492012992501259,
0.0675821527838707,
-0.07653367519378662,
-0.0202096626162529,
0.025652024894952774,
-0.06017875298857689,
0.05335857346653938,
-0.019051995128393173,
0.018273908644914627,
-0.0521397702395916,
-0.0817580297589302,
-0.05428663268685341,
0.02644684538245201,
0.014924325980246067,
0.054286643862724304,
-0.2144417017698288,
-0.00958106480538845,
-0.023239130154252052,
-0.11878415197134018,
-0.025227610021829605,
-0.006108676083385944,
-0.08585220575332642,
0.004315250087529421,
-0.033441804349422455,
0.03169618919491768,
-0.03416713327169418,
-0.01910240948200226,
0.022437769919633865,
-0.0013304915046319366,
0.08398790657520294,
-0.037253253161907196,
0.06111537665128708,
-0.20021264255046844,
-0.02767246589064598,
-0.043094076216220856,
-0.03537527844309807,
-0.031026748940348625,
-0.039771344512701035,
0.059734586626291275,
-0.028465555980801582,
0.08480904251337051,
-0.029318079352378845,
-0.010752764530479908,
0.040197622030973434,
-0.006120410282164812,
-0.05087416246533394,
0.023314209654927254,
0.10412739962339401,
0.00528274429962039,
-0.00010189160093432292,
-0.004153978079557419,
0.022853892296552658,
-0.03377161920070648,
-0.030008016154170036,
0.1058531403541565,
0.08210869878530502,
0.07707115262746811,
0.03405280411243439,
0.11108583956956863,
-0.05770613253116608,
-0.1038203313946724,
-0.0032150817569345236,
-0.03761368244886398,
0.0912533849477768,
-0.03745587170124054,
0.08848442137241364,
0.08855361491441727,
-0.15082921087741852,
0.10846332460641861,
-0.017351828515529633,
-0.010859545320272446,
-0.07851476222276688,
-0.1239314079284668,
-0.09908819943666458,
-0.07865975052118301,
-0.014563309028744698,
-0.07968620955944061,
0.03894209489226341,
-0.03842732682824135,
0.008507492952048779,
-0.017026571556925774,
0.12609393894672394,
-0.05866537243127823,
-0.02455347590148449,
0.02295229770243168,
0.04421736299991608,
-0.007808969356119633,
-0.0007919921772554517,
-0.0010074416641145945,
0.04876793548464775,
0.06317339092493057,
0.016541825607419014,
0.018818682059645653,
-0.0016199585516005754,
0.04239792004227638,
-0.005740095395594835,
-0.11427447199821472,
-0.009536593221127987,
0.0260044876486063,
0.03104775957763195,
0.03516622632741928,
0.036980461329221725,
0.011867043562233448,
-0.038300104439258575,
0.0863136351108551,
-0.015535220503807068,
-0.040255703032016754,
-0.10217709094285965,
0.1794138103723526,
-0.05375659838318825,
-0.0033138622529804707,
0.046202417463064194,
-0.09528232365846634,
-0.0002491216582711786,
0.14526456594467163,
0.24354273080825806,
-0.002369426190853119,
-0.008058322593569756,
0.019230477511882782,
0.01697108894586563,
-0.013749241828918457,
0.06014157459139824,
0.048449259251356125,
0.11442190408706665,
-0.055674608796834946,
0.04904584586620331,
-0.029404308646917343,
-0.054910171777009964,
-0.023619534447789192,
0.026049857959151268,
0.009405882097780704,
0.007606208324432373,
0.042983558028936386,
0.10001488775014877,
-0.024769606068730354,
-0.14748038351535797,
0.08121319115161896,
-0.16974031925201416,
-0.14832109212875366,
-0.047739796340465546,
0.029321080073714256,
0.020948179066181183,
0.07536580413579941,
-0.03298107534646988,
-0.03171062842011452,
0.20358532667160034,
-0.028074055910110474,
-0.05899084731936455,
-0.09832312911748886,
0.020441092550754547,
-0.09668944776058197,
0.13841167092323303,
-0.008817383088171482,
-0.00542556494474411,
0.12228859215974808,
-0.021948225796222687,
-0.15407922863960266,
0.022388650104403496,
0.05838823318481445,
-0.06662013381719589,
0.07426796853542328,
0.12037602066993713,
-0.0008322051144205034,
0.07671105861663818,
0.020938027650117874,
-0.1675398051738739,
0.03367368504405022,
0.01624918170273304,
0.0028204803820699453,
-0.04721074551343918,
0.09156905859708786,
-0.06716921180486679,
0.1614406257867813,
0.19669775664806366,
-0.05243987590074539,
0.006645268760621548,
-0.005850786808878183,
0.060378286987543106,
0.06362046301364899,
0.1139972135424614,
-0.026050040498375893,
-0.1806458681821823,
0.05474412068724632,
-0.027082225307822227,
0.06820669770240784,
-0.1952790468931198,
-0.09679047763347626,
-0.04316534474492073,
-0.026921940967440605,
-0.059101399034261703,
0.12512192130088806,
0.07807399332523346,
0.039878591895103455,
-0.025211311876773834,
-0.11121244728565216,
0.0005876116920262575,
0.11113984882831573,
-0.12531879544258118,
-0.0628880113363266
] |
null | null | diffusers |
# LoRA DreamBooth - danaleee/CL
These are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using [DreamBooth](https://dreambooth.github.io/). You can find some example images in the following.




LoRA for the text encoder was enabled: False.
| {"license": "creativeml-openrail-m", "tags": ["stable-diffusion", "stable-diffusion-diffusers", "text-to-image", "diffusers", "lora"], "base_model": "CompVis/stable-diffusion-v1-4", "instance_prompt": "a photo of sks teddybear", "inference": true} | text-to-image | danaleee/CL | [
"diffusers",
"tensorboard",
"stable-diffusion",
"stable-diffusion-diffusers",
"text-to-image",
"lora",
"base_model:CompVis/stable-diffusion-v1-4",
"license:creativeml-openrail-m",
"region:us"
] | 2024-02-07T06:24:08+00:00 | [] | [] | TAGS
#diffusers #tensorboard #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us
|
# LoRA DreamBooth - danaleee/CL
These are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using DreamBooth. You can find some example images in the following.
!img_0
!img_1
!img_2
!img_3
LoRA for the text encoder was enabled: False.
| [
"# LoRA DreamBooth - danaleee/CL\n\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
"TAGS\n#diffusers #tensorboard #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us \n",
"# LoRA DreamBooth - danaleee/CL\n\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
70,
98
] | [
"passage: TAGS\n#diffusers #tensorboard #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us \n# LoRA DreamBooth - danaleee/CL\n\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
-0.04772701486945152,
-0.013690099120140076,
-0.002699324395507574,
0.06952100247144699,
0.10886455327272415,
0.010465312749147415,
0.20011374354362488,
0.07357887178659439,
0.052822574973106384,
0.07071219384670258,
0.11647717654705048,
0.06790778785943985,
0.003702660324051976,
0.09241905808448792,
-0.028299707919359207,
-0.14088261127471924,
0.009960704483091831,
-0.0386301651597023,
-0.0810762271285057,
0.023573853075504303,
0.04196403920650482,
-0.041364580392837524,
0.11202314496040344,
-0.03471480309963226,
-0.146956205368042,
0.045379117131233215,
-0.004061137326061726,
-0.06039956584572792,
0.05413028970360756,
0.1013006716966629,
0.046883776783943176,
0.0718577429652214,
0.04726983606815338,
-0.15863919258117676,
0.03603467717766762,
0.02072528563439846,
-0.041677020490169525,
0.06243323162198067,
-0.08367843180894852,
-0.002624831860885024,
0.07120776921510696,
-0.04706888273358345,
0.007748723030090332,
0.017337320372462273,
-0.06217876076698303,
-0.03000943548977375,
-0.02004653587937355,
-0.010808433406054974,
0.026650836691260338,
0.06351792812347412,
0.0052898116409778595,
0.07403484731912613,
-0.015529450960457325,
0.08362326771020889,
0.2606518864631653,
-0.2308049350976944,
-0.01438579335808754,
0.21410350501537323,
-0.016412384808063507,
0.08097325265407562,
-0.026707826182246208,
0.0967051237821579,
0.10384400188922882,
-0.03370849788188934,
0.02356380596756935,
-0.06494735181331635,
-0.016602400690317154,
-0.05942528694868088,
-0.09725475311279297,
0.046263255178928375,
0.16925537586212158,
-0.015014007687568665,
-0.06157928705215454,
-0.12940534949302673,
-0.03269260376691818,
0.04316617548465729,
-0.010759836062788963,
0.028879409655928612,
0.008904486894607544,
0.0063804942183196545,
-0.03507448732852936,
-0.08092934638261795,
-0.07229351252317429,
-0.04764018580317497,
-0.019044453278183937,
0.1201799288392067,
0.006037470884621143,
0.056943461298942566,
-0.01850210689008236,
0.11886085569858551,
-0.11069885641336441,
-0.12371501326560974,
0.025340676307678223,
-0.059221312403678894,
0.027104299515485764,
0.07331908494234085,
-0.038620512932538986,
-0.12049388885498047,
0.07211586833000183,
0.001987893134355545,
0.13716109097003937,
0.020437249913811684,
0.00019604522094596177,
0.10108038783073425,
-0.002137833973392844,
0.05465834215283394,
-0.06673597544431686,
-0.010576003231108189,
0.045634228736162186,
0.011927887797355652,
0.10383769124746323,
-0.07837600260972977,
-0.16445942223072052,
0.00644916296005249,
-0.046881984919309616,
0.056711141020059586,
-0.08414781838655472,
0.01609901152551174,
-0.08317480981349945,
-0.00232956325635314,
0.05365133285522461,
-0.028196560218930244,
0.031223291531205177,
-0.04179829731583595,
-0.03521787002682686,
0.1401311457157135,
0.17488202452659607,
0.017519071698188782,
-0.014060024172067642,
0.050824470818042755,
-0.06504222750663757,
0.05529531091451645,
-0.013609432615339756,
-0.1320790946483612,
0.015597605146467686,
-0.1480133980512619,
0.006481061223894358,
-0.13754941523075104,
-0.039297860115766525,
-0.014601236209273338,
0.01584538072347641,
-0.048199787735939026,
0.042828015983104706,
-0.08108317852020264,
-0.12985612452030182,
-0.009997465647757053,
0.053402118384838104,
-0.0264951903373003,
-0.003636891720816493,
0.04922441020607948,
-0.003444827161729336,
0.12076003104448318,
-0.1335168182849884,
-0.029754268005490303,
-0.06230879947543144,
0.01932070218026638,
-0.12071961909532547,
0.09689406305551529,
-0.05377209931612015,
0.0529094859957695,
-0.056427910923957825,
-0.03644458204507828,
0.007232616655528545,
0.04482332617044449,
0.053788378834724426,
0.12033544480800629,
-0.25011882185935974,
-0.05189071595668793,
0.09323623776435852,
-0.1595487892627716,
-0.10154499858617783,
0.0564577542245388,
-0.023170672357082367,
0.08498209714889526,
0.06904952973127365,
0.0900847315788269,
0.06887326389551163,
-0.3054234981536865,
-0.012790199369192123,
-0.052964888513088226,
-0.01670903153717518,
-0.0322984978556633,
-0.0018346281722187996,
0.07025092095136642,
-0.039318013936281204,
0.04435089975595474,
-0.06567082554101944,
0.07805483788251877,
-0.02902284823358059,
-0.02539302222430706,
-0.03733716532588005,
-0.04366793856024742,
-0.028261400759220123,
-0.011665794998407364,
0.04967682808637619,
-0.030238311737775803,
-0.03001135215163231,
0.03771499544382095,
0.052076518535614014,
-0.06070461496710777,
0.027038119733333588,
-0.04923658072948456,
0.0947851836681366,
-0.10615284740924835,
0.0015892365481704473,
-0.08126421272754669,
-0.03394615277647972,
0.03748369589447975,
0.14010798931121826,
0.11312031000852585,
0.06116370111703873,
0.08867794275283813,
0.07243908941745758,
0.0008803008240647614,
0.00026539989630691707,
0.05347658693790436,
0.0002929343027062714,
-0.05708983168005943,
-0.1861356645822525,
0.034596312791109085,
-0.08183423429727554,
0.05428842455148697,
-0.17118559777736664,
0.02239600382745266,
0.03863650932908058,
0.16075244545936584,
0.0894947499036789,
-0.010520980693399906,
0.044673364609479904,
0.043105777353048325,
-0.05673402547836304,
-0.05453632026910782,
0.01640980690717697,
-0.014539037831127644,
-0.18809905648231506,
0.17642256617546082,
-0.16972781717777252,
0.026157742366194725,
0.12173493206501007,
0.011298201978206635,
-0.05723584070801735,
-0.04905218258500099,
0.013102281838655472,
0.03186330571770668,
-0.03535672277212143,
-0.03393130004405975,
0.1344311684370041,
-0.015423533506691456,
0.15316979587078094,
-0.02595609240233898,
0.028924213722348213,
0.05828564241528511,
-0.06693816930055618,
-0.08941853046417236,
0.09092776477336884,
0.0389103889465332,
0.03171839192509651,
0.010799223557114601,
0.07722964882850647,
-0.031777892261743546,
0.12180623412132263,
-0.0009178211912512779,
-0.045049332082271576,
-0.0780223160982132,
0.021867157891392708,
0.05847705155611038,
0.12103492021560669,
0.02266830950975418,
-0.020790183916687965,
-0.01385205052793026,
-0.06422582268714905,
0.011414160951972008,
-0.15201835334300995,
-0.024612169712781906,
0.022172830998897552,
-0.04404238983988762,
0.17668692767620087,
0.08546862006187439,
-0.08797164261341095,
0.07697296142578125,
-0.10159613192081451,
-0.0804784968495369,
-0.0061297789216041565,
-0.033615756779909134,
-0.053845882415771484,
0.10954184830188751,
-0.07390130311250687,
-0.18028663098812103,
-0.189629927277565,
0.004795357584953308,
-0.021005738526582718,
-0.004215290304273367,
0.04922397807240486,
-0.08549177646636963,
-0.07424535602331161,
-0.13328808546066284,
0.03365626558661461,
0.02469922974705696,
0.041627030819654465,
0.04778566583991051,
-0.027282653376460075,
0.002004383597522974,
-0.08464684337377548,
-0.009203839115798473,
-0.06409899145364761,
0.06619031727313995,
0.06641547381877899,
-0.005657282657921314,
0.10674717277288437,
0.1269015222787857,
0.02961736172437668,
0.035546377301216125,
0.006859153043478727,
0.2212129682302475,
-0.00701368972659111,
0.09331457316875458,
0.14790090918540955,
0.0014649515505880117,
0.06302349269390106,
0.12217087298631668,
0.03191988915205002,
-0.08851377665996552,
0.07074719667434692,
-0.007825286127626896,
-0.14610475301742554,
-0.08706981688737869,
-0.06682711094617844,
-0.03495292365550995,
-0.07270815223455429,
0.05026445910334587,
0.04438494145870209,
0.12068803608417511,
0.08453123271465302,
0.060944292694330215,
0.10679265111684799,
0.06611520051956177,
0.06784187257289886,
0.0863812193274498,
-0.05501343309879303,
0.041674092411994934,
-0.07791031152009964,
-0.11075311154127121,
0.1035139411687851,
-0.05452602356672287,
0.18144354224205017,
-0.08165424317121506,
0.03583480417728424,
0.038571059703826904,
-0.03423029184341431,
0.0829484611749649,
0.013713374733924866,
-0.05051618441939354,
-0.010101541876792908,
-0.04086355119943619,
-0.12125793844461441,
0.09426108002662659,
0.09472436457872391,
0.011740761809051037,
-0.020692724734544754,
-0.018700696527957916,
0.060537636280059814,
0.02552732825279236,
-0.008240430615842342,
0.16074083745479584,
-0.21826858818531036,
0.031485434621572495,
0.025256196036934853,
0.07844684273004532,
0.005782365798950195,
0.005272214766591787,
0.23504753410816193,
-0.004730631597340107,
0.053630974143743515,
-0.041961610317230225,
0.05849510058760643,
-0.012715324759483337,
-0.04520620033144951,
-0.06295555830001831,
0.15765434503555298,
-0.010417955927550793,
-0.02797885239124298,
-0.23945549130439758,
0.014826428145170212,
0.01059003360569477,
0.022246308624744415,
-0.06375398486852646,
-0.01706209033727646,
0.02726718969643116,
0.010115058161318302,
0.09145863354206085,
0.015719538554549217,
0.02411198616027832,
-0.10000088810920715,
-0.17488378286361694,
-0.02557636983692646,
0.08441928774118423,
-0.005934323184192181,
0.06318812072277069,
0.045284420251846313,
-0.018418680876493454,
0.01952318847179413,
0.019474731758236885,
-0.13943932950496674,
-0.10630155354738235,
0.007234492339193821,
0.2045392096042633,
-0.046219322830438614,
-0.03879905864596367,
-0.09013067185878754,
-0.03631002828478813,
0.09726528823375702,
-0.05251668393611908,
-0.06918111443519592,
-0.0761956200003624,
0.008521204814314842,
0.09399077296257019,
-0.036362603306770325,
-0.009273543022572994,
-0.044396210461854935,
0.06284768134355545,
-0.07635119557380676,
-0.13784663379192352,
0.07196469604969025,
-0.01416784804314375,
-0.13207808136940002,
-0.06831185519695282,
0.09866619855165482,
-0.015486509539186954,
-0.007958651520311832,
-0.01732327975332737,
0.027244864031672478,
0.03065708838403225,
-0.11714685708284378,
0.046708233654499054,
0.18415041267871857,
-0.1479247510433197,
0.14085346460342407,
-0.032992832362651825,
-0.000929484551306814,
-0.05961383879184723,
0.0059725637547671795,
0.1367398053407669,
0.268637090921402,
-0.09130032360553741,
0.09747641533613205,
0.04825401306152344,
-0.09934872388839722,
-0.2351149618625641,
-0.051648061722517014,
-0.03593754023313522,
0.041552215814590454,
-0.019566409289836884,
-0.07955325394868851,
0.09207923710346222,
0.03199069947004318,
0.016376184299588203,
0.21348077058792114,
-0.3844316601753235,
-0.1477549970149994,
-0.007520277053117752,
0.15582527220249176,
0.2962019145488739,
-0.13675662875175476,
-0.07633271813392639,
-0.03815867006778717,
-0.03352133929729462,
0.14224474132061005,
-0.06986144185066223,
0.1202114149928093,
-0.018362391740083694,
-0.01065812073647976,
0.02590518817305565,
-0.035769861191511154,
0.14170068502426147,
-0.027097802609205246,
0.05603766068816185,
-0.07559803873300552,
-0.07056697458028793,
0.10324990749359131,
-0.05389432609081268,
0.06302394717931747,
-0.16271433234214783,
0.03377380967140198,
-0.010505271144211292,
-0.014100810512900352,
0.01783696375787258,
0.0032232427038252354,
-0.02410116046667099,
-0.04940176010131836,
-0.11498253047466278,
0.016997257247567177,
0.025118395686149597,
-0.013350676745176315,
0.06678662449121475,
-0.026344526559114456,
-0.031477563083171844,
0.17167098820209503,
-0.057579074054956436,
0.07323646545410156,
-0.016278032213449478,
0.01634852960705757,
-0.04575197771191597,
0.11732082813978195,
-0.13697169721126556,
0.009615194983780384,
0.15586663782596588,
0.05729987472295761,
0.12173517048358917,
0.019246524199843407,
-0.10968120396137238,
0.08486729115247726,
0.08131081610918045,
-0.09687380492687225,
-0.022838285192847252,
-0.02790425904095173,
-0.026960819959640503,
0.11453966796398163,
0.008750783279538155,
0.20241700112819672,
-0.10944218933582306,
0.05045747384428978,
-0.0013740002177655697,
0.009931785054504871,
-0.022219549864530563,
0.09192401170730591,
0.03905188664793968,
-0.008673455566167831,
-0.08204969763755798,
0.08072514086961746,
-0.02014835923910141,
0.00662061246111989,
0.04998154938220978,
0.05766630545258522,
-0.07515507191419601,
-0.013987211510539055,
0.003812001086771488,
0.21802756190299988,
-0.11515101790428162,
-0.01694437488913536,
-0.12161309272050858,
-0.11727695912122726,
0.013256484642624855,
0.15041302144527435,
0.0488910898566246,
0.03794682398438454,
-0.037132713943719864,
-0.05433366075158119,
-0.07026955485343933,
0.050649244338274,
0.042554449290037155,
0.06946627050638199,
-0.23047034442424774,
-0.018620716407895088,
0.022339927032589912,
0.0014852547319605947,
-0.08022525906562805,
-0.052673909813165665,
-0.11996984481811523,
-0.018903890624642372,
-0.07699475437402725,
0.12027004361152649,
-0.038558974862098694,
-0.022239213809370995,
0.005617406219244003,
-0.02432403713464737,
-0.007347386330366135,
0.04423508793115616,
-0.01697111316025257,
-0.02392481081187725,
-0.013302614912390709,
-0.018749823793768883,
-0.05049342289566994,
-0.11430486291646957,
-0.019178664311766624,
-0.08179793506860733,
0.028614383190870285,
-0.0025664863642305136,
-0.08618425577878952,
0.015516886487603188,
-0.2166593074798584,
0.039986029267311096,
0.18195505440235138,
0.0019217870431020856,
0.01193678006529808,
-0.024788720533251762,
-0.001930702361278236,
-0.02155565842986107,
0.06175781786441803,
-0.005201644264161587,
0.08759879320859909,
-0.06981776654720306,
-0.059003956615924835,
-0.06731733679771423,
-0.014411088079214096,
-0.04563508555293083,
0.05960343778133392,
0.12867647409439087,
0.15195044875144958,
0.14186447858810425,
-0.12616510689258575,
0.08592593669891357,
-0.07667958736419678,
-0.001406464260071516,
0.005385212600231171,
-0.043661244213581085,
0.02655956707894802,
-0.03432745859026909,
-0.0165521502494812,
-0.03517040237784386,
0.1417182832956314,
-0.01577390544116497,
-0.14939144253730774,
-0.018218345940113068,
0.014546681195497513,
-0.048308517783880234,
0.024614883586764336,
0.16462503373622894,
0.028280751779675484,
0.034937113523483276,
-0.12292692065238953,
0.04035196825861931,
0.16397418081760406,
0.10463081300258636,
0.04199453443288803,
0.0024780540261417627,
0.049512092024087906,
0.1303616166114807,
0.062273260205984116,
0.030922235921025276,
0.07029002904891968,
0.12814420461654663,
-0.05235698074102402,
0.09441918134689331,
-0.04556066542863846,
0.009876828640699387,
0.10091783851385117,
-0.027818821370601654,
-0.026996301487088203,
0.08020805567502975,
-0.05919817462563515,
-0.07465524226427078,
-0.10031959414482117,
-0.08154650032520294,
-0.12800294160842896,
0.016389280557632446,
-0.054289430379867554,
0.0014623361639678478,
-0.03332353010773659,
0.05383041128516197,
0.04784989356994629,
0.061155036091804504,
-0.030667133629322052,
-0.0486726313829422,
0.10138411074876785,
-0.025467701256275177,
-0.07544740289449692,
0.05092431604862213,
0.021878322586417198,
0.07026545703411102,
0.016283925622701645,
-0.03340228646993637,
0.09618523716926575,
0.08026707172393799,
0.02851685881614685,
-0.03159911930561066,
-0.09229689091444016,
-0.027795247733592987,
0.017293835058808327,
0.017941491678357124,
0.13912872970104218,
0.09845933318138123,
-0.056804899126291275,
-0.050170931965112686,
0.12322357296943665,
-0.06299663335084915,
-0.0312957838177681,
-0.1338912844657898,
0.060481537133455276,
-0.11251752078533173,
0.054208479821681976,
-0.07605099678039551,
-0.08789127320051193,
-0.054408732801675797,
0.1719662994146347,
0.18562012910842896,
-0.10359262675046921,
0.0325346365571022,
-0.05601516366004944,
-0.005631236359477043,
-0.022080598399043083,
0.07032186537981033,
0.030541222542524338,
0.221188485622406,
-0.037332262843847275,
0.004193338565528393,
-0.0910453200340271,
-0.06749469041824341,
-0.05351342260837555,
-0.07442045956850052,
-0.016265353187918663,
-0.04326662793755531,
-0.0847351923584938,
0.03538714349269867,
-0.1592906415462494,
-0.1418677568435669,
0.22124499082565308,
-0.16026665270328522,
-0.025162624195218086,
-0.04694904759526253,
0.061687592417001724,
0.0450003519654274,
0.027161719277501106,
-0.08285463601350784,
0.0029761656187474728,
0.032111089676618576,
-0.01788416877388954,
-0.1361551433801651,
0.0064124781638383865,
-0.07262515276670456,
-0.15935175120830536,
0.04801670461893082,
-0.0027359232772141695,
0.08584744483232498,
0.04758821427822113,
0.021598299965262413,
-0.08748432248830795,
0.08614042401313782,
-0.04502337798476219,
-0.08509556949138641,
-0.08385328203439713,
0.10570521652698517,
-0.03520280122756958,
0.11339162290096283,
0.037914909422397614,
-0.08096589893102646,
0.013157705776393414,
0.08279493451118469,
-0.0830833837389946,
-0.0891086757183075,
-0.03839842230081558,
-0.07882493734359741,
0.09115619212388992,
0.07888247072696686,
-0.03277517110109329,
0.012532473541796207,
0.016202257946133614,
0.0017190580256283283,
0.008193147368729115,
-0.07039826363325119,
0.05332376807928085,
-0.07182261347770691,
-0.011611324734985828,
0.02377287670969963,
0.04544052854180336,
-0.27640390396118164,
-0.06573476642370224,
-0.17081592977046967,
-0.012438450008630753,
-0.01597968302667141,
0.08686709403991699,
0.23341450095176697,
0.03813755139708519,
-0.00186250323895365,
-0.16913507878780365,
0.049388352781534195,
0.08476823568344116,
-0.05493787303566933,
-0.07905025035142899
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# wav2vec_RTSplit0207_4
This model is a fine-tuned version of [jonatasgrosman/wav2vec2-large-xlsr-53-japanese](https://huggingface.co/jonatasgrosman/wav2vec2-large-xlsr-53-japanese) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0532
- Wer: 0.2258
- Cer: 0.1935
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5.1e-05
- train_batch_size: 32
- eval_batch_size: 32
- seed: 4
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 1000
- num_epochs: 10
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|
| 3.8752 | 1.0 | 120 | 3.5938 | 1.0 | 0.9489 |
| 1.6275 | 2.0 | 240 | 1.4828 | 1.0 | 0.7789 |
| 0.8457 | 3.0 | 360 | 0.7008 | 0.8213 | 0.5921 |
| 0.6001 | 4.0 | 480 | 0.5081 | 0.7088 | 0.4447 |
| 0.4644 | 5.0 | 600 | 0.3125 | 0.4872 | 0.2592 |
| 0.3617 | 6.0 | 720 | 0.2162 | 0.3807 | 0.2038 |
| 0.298 | 7.0 | 840 | 0.1448 | 0.3141 | 0.1964 |
| 0.2704 | 8.0 | 960 | 0.1096 | 0.2762 | 0.1944 |
| 0.2499 | 9.0 | 1080 | 0.0630 | 0.2401 | 0.1895 |
| 0.1338 | 10.0 | 1200 | 0.0532 | 0.2258 | 0.1935 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.14.6
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["wer"], "base_model": "jonatasgrosman/wav2vec2-large-xlsr-53-japanese", "model-index": [{"name": "wav2vec_RTSplit0207_4", "results": []}]} | automatic-speech-recognition | tndklab/wav2vec_RTSplit0207_4 | [
"transformers",
"safetensors",
"wav2vec2",
"automatic-speech-recognition",
"generated_from_trainer",
"base_model:jonatasgrosman/wav2vec2-large-xlsr-53-japanese",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-07T06:33:38+00:00 | [] | [] | TAGS
#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-jonatasgrosman/wav2vec2-large-xlsr-53-japanese #license-apache-2.0 #endpoints_compatible #region-us
| wav2vec\_RTSplit0207\_4
=======================
This model is a fine-tuned version of jonatasgrosman/wav2vec2-large-xlsr-53-japanese on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0532
* Wer: 0.2258
* Cer: 0.1935
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5.1e-05
* train\_batch\_size: 32
* eval\_batch\_size: 32
* seed: 4
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 1000
* num\_epochs: 10
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.14.6
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5.1e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-jonatasgrosman/wav2vec2-large-xlsr-53-japanese #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5.1e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
80,
116,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-jonatasgrosman/wav2vec2-large-xlsr-53-japanese #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5.1e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
-0.14140185713768005,
0.1513470560312271,
-0.0005470866453833878,
0.0996072068810463,
0.11818834394216537,
0.008252711035311222,
0.17517869174480438,
0.15004801750183105,
-0.04121510684490204,
0.11099033802747726,
0.1138237863779068,
0.061895813792943954,
0.05501498654484749,
0.19709637761116028,
-0.08205235749483109,
-0.22009775042533875,
0.07736305147409439,
-0.0035447459667921066,
0.010096057318150997,
0.1122141182422638,
0.07091084122657776,
-0.1180148497223854,
0.0901353731751442,
-0.006939710583537817,
-0.14207009971141815,
-0.04258633777499199,
0.016870828345417976,
-0.11024443805217743,
0.10898896306753159,
0.008776349946856499,
0.06617004424333572,
0.03509335592389107,
0.08978565037250519,
-0.18774408102035522,
0.002091117901727557,
0.01791985146701336,
0.014900037087500095,
0.07442665845155716,
0.042306266725063324,
-0.0009124421048909426,
0.0019511813297867775,
-0.1144457459449768,
0.037111811339855194,
0.01577121578156948,
-0.1169668436050415,
-0.1990584135055542,
-0.07826364040374756,
0.016476690769195557,
0.09911130368709564,
0.08347859978675842,
-0.020566459745168686,
0.12296249717473984,
0.0013710098573938012,
0.07954324781894684,
0.19705674052238464,
-0.313181608915329,
-0.0541914701461792,
-0.016705922782421112,
0.039830710738897324,
0.08297010511159897,
-0.1016988456249237,
-0.018417509272694588,
0.05055718868970871,
0.021616870537400246,
0.0928344577550888,
-0.03153528273105621,
-0.0336184948682785,
-0.01177369337528944,
-0.12029176950454712,
-0.03891395404934883,
0.18963198363780975,
0.07309895753860474,
-0.06446082144975662,
-0.08051706850528717,
-0.06428349763154984,
-0.12232383340597153,
-0.05461275577545166,
-0.007258039433509111,
0.026400242000818253,
-0.03910614550113678,
-0.0998644083738327,
-0.005066312849521637,
-0.07962246984243393,
-0.09127762168645859,
-0.017650596797466278,
0.17619140446186066,
0.010853609070181847,
0.014334741979837418,
-0.011432663537561893,
0.054719869047403336,
-0.023717068135738373,
-0.1846146136522293,
-0.022666433826088905,
0.02742605283856392,
-0.033444102853536606,
-0.01444366853684187,
-0.04391278326511383,
-0.034230004996061325,
0.04349740967154503,
0.11757370084524155,
-0.019144944846630096,
0.06593415141105652,
-0.02450522407889366,
0.0017377863405272365,
-0.08481678366661072,
0.18254344165325165,
-0.06381513178348541,
-0.06886767596006393,
0.02049325965344906,
0.12740157544612885,
0.06357104331254959,
-0.02322385273873806,
-0.09904051572084427,
-0.009124419651925564,
0.14740827679634094,
0.03538563847541809,
-0.04271034151315689,
0.05055074393749237,
-0.03877957537770271,
-0.013883186504244804,
0.05678894370794296,
-0.12105908989906311,
0.02533109486103058,
0.02182227000594139,
-0.06302780658006668,
-0.023170191794633865,
-0.011186899617314339,
0.012444955296814442,
0.012831320986151695,
0.05200100317597389,
-0.08241927623748779,
0.004294237587600946,
-0.02347908914089203,
-0.09291993081569672,
0.02673855982720852,
-0.06916772574186325,
-0.0005395279731601477,
-0.10777652263641357,
-0.17966967821121216,
-0.017954057082533836,
0.02419441193342209,
-0.04916108027100563,
-0.01022262405604124,
-0.11255406588315964,
-0.09737443178892136,
0.047409821301698685,
-0.02289048209786415,
0.035422515124082565,
-0.07961729168891907,
0.10830166935920715,
0.079220712184906,
0.08742014318704605,
-0.040515750646591187,
0.02640279196202755,
-0.09470052272081375,
0.03248461335897446,
-0.17534197866916656,
0.07536406815052032,
-0.05433931201696396,
0.034256063401699066,
-0.12034598737955093,
-0.06720847636461258,
0.0199790857732296,
-0.022830627858638763,
0.06995867192745209,
0.14243118464946747,
-0.19172273576259613,
-0.05621590465307236,
0.19613642990589142,
-0.12017670273780823,
-0.14233659207820892,
0.12820763885974884,
-0.036236390471458435,
0.03840848058462143,
0.07050701230764389,
0.22266121208667755,
0.03123670257627964,
-0.10616591572761536,
-0.03942224383354187,
-0.06344962865114212,
0.0835147425532341,
-0.037991005927324295,
0.11146456003189087,
0.004586546681821346,
-0.0014519975520670414,
0.016056343913078308,
-0.08052760362625122,
0.032135628163814545,
-0.07097344845533371,
-0.09986229240894318,
-0.044667378067970276,
-0.10617202520370483,
0.028264937922358513,
0.015629269182682037,
0.0556182935833931,
-0.09861555695533752,
-0.07093813270330429,
0.011688318103551865,
0.10839072614908218,
-0.1172947883605957,
0.013225885108113289,
-0.10380188375711441,
0.09419671446084976,
-0.11421684175729752,
-0.02050180546939373,
-0.1545829027891159,
-0.004180761519819498,
0.05405221879482269,
0.018476149067282677,
0.014612407423555851,
-0.07567939907312393,
0.08224498480558395,
0.077387236058712,
-0.04963484779000282,
-0.07384662330150604,
-0.005162188317626715,
0.01811145804822445,
-0.06272849440574646,
-0.17335008084774017,
-0.02903943508863449,
-0.05393994599580765,
0.1609354168176651,
-0.16454531252384186,
0.0013096441980451345,
0.00870917085558176,
0.09000089019536972,
0.04394276812672615,
-0.0234912671148777,
0.02021574229001999,
0.04835608974099159,
-0.025901198387145996,
-0.07198876887559891,
0.02940317802131176,
0.015535028651356697,
-0.10412604361772537,
0.01979631744325161,
-0.1677049994468689,
0.1497868150472641,
0.13862372934818268,
0.041674740612506866,
-0.052922505885362625,
0.021306047216057777,
-0.014012404717504978,
-0.042311862111091614,
-0.05483417585492134,
-0.015455318614840508,
0.10160964727401733,
0.008567889221012592,
0.12180560827255249,
-0.10307618230581284,
0.01559318695217371,
0.06480160355567932,
-0.027578242123126984,
-0.027729468420147896,
0.08030901104211807,
0.010379251092672348,
-0.13940684497356415,
0.1297876089811325,
0.11201223731040955,
-0.07278957962989807,
0.12587659060955048,
-0.061526842415332794,
-0.08503663539886475,
-0.05001096799969673,
0.03365752473473549,
0.034158892929553986,
0.13780593872070312,
-0.08110926300287247,
-0.022295471280813217,
0.021228276193141937,
0.02263244241476059,
-0.016371827572584152,
-0.19371728599071503,
-0.019441206008195877,
0.014570568688213825,
-0.09517217427492142,
-0.009684785269200802,
0.005465525668114424,
-0.017490917816758156,
0.09453999996185303,
-0.0008943516877479851,
-0.11403311043977737,
0.02380990795791149,
-0.015043700113892555,
-0.08731178194284439,
0.17253625392913818,
-0.09242700040340424,
-0.17422594130039215,
-0.1363985687494278,
-0.07131753116846085,
-0.0563679002225399,
0.03763280063867569,
0.060043156147003174,
-0.06574764847755432,
-0.04060199484229088,
-0.1155543401837349,
-0.04762861132621765,
0.03236564248800278,
0.04535595327615738,
0.05059736967086792,
-0.00873279757797718,
0.06682752072811127,
-0.08195891976356506,
-0.005217843223363161,
-0.014315478503704071,
-0.0076027545146644115,
0.028978558257222176,
0.0006948634400032461,
0.12585537135601044,
0.12219582498073578,
0.005714054219424725,
0.024594463407993317,
-0.03780292719602585,
0.22647689282894135,
-0.06951361149549484,
-0.018394462764263153,
0.12351155281066895,
-0.02730284072458744,
0.04593672975897789,
0.17761284112930298,
0.030772989615797997,
-0.10713522881269455,
0.001591666485182941,
-0.049965858459472656,
-0.015327079221606255,
-0.18867911398410797,
-0.033214159309864044,
-0.04741145297884941,
0.013541110791265965,
0.10095350444316864,
0.029781633988022804,
0.015772372484207153,
0.04822786524891853,
0.021661022678017616,
0.045908328145742416,
0.004295762628316879,
0.08128432929515839,
0.09679067134857178,
0.07629586011171341,
0.10807526856660843,
-0.032302405685186386,
-0.04833135008811951,
0.03228259086608887,
0.021660981699824333,
0.20282496511936188,
0.03015337884426117,
0.19324572384357452,
0.0008202315657399595,
0.15429453551769257,
0.025849437341094017,
0.07994496077299118,
0.019731344655156136,
0.010017940774559975,
-0.02099386602640152,
-0.07817583531141281,
-0.0543784536421299,
0.05450154468417168,
-0.015062488615512848,
0.061184126883745193,
-0.10583194345235825,
0.02077200822532177,
0.049671296030282974,
0.27263399958610535,
0.08806004375219345,
-0.3687765598297119,
-0.08683569729328156,
0.021004049107432365,
-0.03675484284758568,
-0.019846390932798386,
0.016449734568595886,
0.15383316576480865,
-0.06160713732242584,
0.06891895830631256,
-0.07258002460002899,
0.06349083036184311,
-0.06471449881792068,
0.019361155107617378,
0.02476094290614128,
0.047697216272354126,
0.002643988002091646,
0.030670413747429848,
-0.24138197302818298,
0.2870331406593323,
0.03583137318491936,
0.0956231877207756,
-0.05666984245181084,
-0.003703873371705413,
0.039524320513010025,
-0.005946311634033918,
0.11745210736989975,
-0.02467280626296997,
-0.11096319556236267,
-0.17986035346984863,
-0.13503502309322357,
0.04860688000917435,
0.10559181869029999,
-0.006208687089383602,
0.11620091646909714,
-0.014005237258970737,
-0.04423267021775246,
0.04472094029188156,
-0.02351694367825985,
-0.08060543984174728,
-0.07464573532342911,
0.009536982513964176,
0.11420916020870209,
0.044696174561977386,
-0.04999950900673866,
-0.09588051587343216,
-0.08866029977798462,
0.09006531536579132,
0.003077346831560135,
-0.0068000550381839275,
-0.10506249964237213,
0.018146736547350883,
0.15031041204929352,
-0.0914032906293869,
0.053232695907354355,
0.009102074429392815,
0.10987211018800735,
0.027373312041163445,
-0.04930993169546127,
0.0895683616399765,
-0.06205105409026146,
-0.17812830209732056,
-0.05131164565682411,
0.13889925181865692,
-0.008068042807281017,
0.04329867288470268,
0.021038008853793144,
0.05126200616359711,
-0.005076700355857611,
-0.06720000505447388,
0.031923726201057434,
0.0262713935226202,
0.041149307042360306,
0.02022133208811283,
-0.012288776226341724,
-0.09103337675333023,
-0.09267769753932953,
-0.023095399141311646,
0.1508205533027649,
0.29735514521598816,
-0.06623394042253494,
0.018502507358789444,
0.08662042766809464,
-0.017793947830796242,
-0.15151247382164001,
-0.004757965449243784,
0.04431391507387161,
0.04431186616420746,
-0.005465537775307894,
-0.12252914160490036,
0.04505029693245888,
0.06158828362822533,
-0.0449206568300724,
0.07729801535606384,
-0.24901333451271057,
-0.12777471542358398,
0.0886179506778717,
0.13309690356254578,
0.12526464462280273,
-0.15300829708576202,
-0.06722140312194824,
-0.024008916690945625,
-0.107561394572258,
0.10452568531036377,
-0.07231258600950241,
0.13365326821804047,
-0.0022233175113797188,
0.06482834368944168,
0.007398509886115789,
-0.05101599544286728,
0.15042997896671295,
0.022757185623049736,
0.05341789498925209,
-0.02199631929397583,
-0.016445519402623177,
0.04798061028122902,
-0.07572799175977707,
0.06930631399154663,
-0.08586692065000534,
0.04973582178354263,
-0.059643860906362534,
-0.024610988795757294,
-0.06153383478522301,
-0.00620544096454978,
0.0038344664499163628,
-0.03478613495826721,
-0.010418582707643509,
0.03622603043913841,
0.05833699181675911,
0.0030026359017938375,
0.13307882845401764,
0.012143872678279877,
0.08217840641736984,
0.1468023955821991,
0.08832675963640213,
-0.039805274456739426,
0.013990710489451885,
-0.005921985022723675,
-0.05668075382709503,
0.053206101059913635,
-0.13250622153282166,
0.04843001812696457,
0.09629879146814346,
0.018189674243330956,
0.16062672436237335,
0.04690052196383476,
-0.049044668674468994,
0.038352951407432556,
0.0692952498793602,
-0.15845973789691925,
-0.11118871718645096,
0.0038913122843950987,
-0.012815359979867935,
-0.11122631281614304,
0.04957396164536476,
0.13908588886260986,
-0.07029798626899719,
-0.006494930479675531,
-0.017931899055838585,
0.021600743755698204,
-0.039276912808418274,
0.20095659792423248,
0.042017318308353424,
0.05175843834877014,
-0.10942616313695908,
0.08193822205066681,
0.056940533220767975,
-0.0872495248913765,
0.04900488257408142,
0.0369500108063221,
-0.11550870537757874,
-0.022458862513303757,
-0.0003930729872081429,
0.14236915111541748,
0.0053826807998120785,
-0.0761381983757019,
-0.13856051862239838,
-0.08896124362945557,
0.034559931606054306,
0.17758123576641083,
0.06803799420595169,
0.03688894584774971,
-0.018575593829154968,
-0.0021316884085536003,
-0.10354095697402954,
0.09375282377004623,
0.07418568432331085,
0.07488188147544861,
-0.1504599004983902,
0.0830882340669632,
-0.008103721775114536,
0.026941239833831787,
-0.02043088711798191,
0.016762452200055122,
-0.10878368467092514,
0.005377405323088169,
-0.09942983090877533,
0.05725988745689392,
-0.0777355432510376,
-0.016245564445853233,
-0.0013191247126087546,
-0.08153882622718811,
-0.06129556894302368,
0.011646118015050888,
-0.08688964694738388,
-0.026275310665369034,
0.003387901233509183,
0.042967718094587326,
-0.13614435493946075,
-0.03794257342815399,
0.02239089272916317,
-0.09818215668201447,
0.08370166271924973,
0.08636493980884552,
-0.019739601761102676,
0.046102359890937805,
-0.09486469626426697,
-0.021506890654563904,
0.08275867998600006,
0.0022266542073339224,
0.050377439707517624,
-0.14485175907611847,
-0.014273300766944885,
0.0313367061316967,
0.0505339689552784,
0.02149040997028351,
0.14811353385448456,
-0.0964384451508522,
0.005582350771874189,
-0.06706925481557846,
-0.010966062545776367,
-0.05686584860086441,
0.021652808412909508,
0.14174555242061615,
0.0031512153800576925,
0.18426023423671722,
-0.09460330009460449,
0.02249487116932869,
-0.19768789410591125,
0.0014611291699111462,
-0.037174925208091736,
-0.12587666511535645,
-0.14824111759662628,
-0.026885099709033966,
0.07833132892847061,
-0.061970245093107224,
0.09477309137582779,
-0.06163078919053078,
0.07021742314100266,
0.012843438424170017,
-0.05748487263917923,
-0.0012974357232451439,
0.040351368486881256,
0.24937373399734497,
0.0578848160803318,
-0.035781435668468475,
0.07755416631698608,
0.009826667606830597,
0.09480121731758118,
0.12432567030191422,
0.1238613873720169,
0.15756310522556305,
0.03245647996664047,
0.1436290591955185,
0.08281715214252472,
-0.024623874574899673,
-0.11868906021118164,
0.06033610180020332,
-0.06792771816253662,
0.08972831815481186,
0.025742612779140472,
0.21032942831516266,
0.09855382889509201,
-0.1635945737361908,
0.004538338631391525,
-0.03611019253730774,
-0.08512052148580551,
-0.09595491737127304,
-0.06149891018867493,
-0.13092951476573944,
-0.14468815922737122,
0.010765369981527328,
-0.10658024996519089,
0.034474097192287445,
0.0699116662144661,
0.01424720510840416,
0.0003581237397156656,
0.1407388597726822,
0.013667227700352669,
0.028598135337233543,
0.09569618105888367,
0.008547991514205933,
-0.04004355147480965,
0.00004307792551117018,
-0.10350070148706436,
0.024678753688931465,
0.005393704399466515,
0.057250238955020905,
-0.02121918462216854,
-0.02469562739133835,
0.06931012868881226,
-0.025981362909078598,
-0.12543874979019165,
0.011071263812482357,
0.020177770406007767,
0.06038098782300949,
0.044666215777397156,
0.056731726974248886,
-0.017080489546060562,
0.025130413472652435,
0.20746773481369019,
-0.08920416235923767,
-0.07730153203010559,
-0.13300630450248718,
0.14855645596981049,
-0.014304406940937042,
-0.007778032682836056,
0.009742062538862228,
-0.1060720831155777,
0.0023043917026370764,
0.19396871328353882,
0.1490781605243683,
-0.07375820726156235,
-0.0014346797252073884,
-0.027093632146716118,
-0.00682365195825696,
-0.038301561027765274,
0.06577453017234802,
0.07775986939668655,
0.033436257392168045,
-0.05985914170742035,
-0.06112997233867645,
-0.05748649686574936,
-0.04090588912367821,
-0.023135585710406303,
0.039301782846450806,
-0.03290274366736412,
-0.022392241284251213,
-0.05015990138053894,
0.07828210294246674,
-0.08224087208509445,
-0.0979672372341156,
0.007329635787755251,
-0.2170497328042984,
-0.17300307750701904,
-0.001741027575917542,
0.07537645846605301,
0.034900251775979996,
0.025979578495025635,
-0.03377597779035568,
0.026197105646133423,
0.056507695466279984,
-0.014077073894441128,
-0.05688495561480522,
-0.059860311448574066,
0.04226115345954895,
-0.0825837254524231,
0.17499074339866638,
-0.0044199698604643345,
0.06649807840585709,
0.10399214923381805,
0.0813714936375618,
-0.10777547210454941,
0.1042306125164032,
0.06050611287355423,
-0.07391688227653503,
0.05603795871138573,
0.15222999453544617,
-0.056308094412088394,
0.14280712604522705,
0.0513388030230999,
-0.10265466570854187,
0.0005663326592184603,
0.010046265088021755,
-0.028935497626662254,
-0.07468231767416,
-0.06614255905151367,
-0.04580698907375336,
0.14725330471992493,
0.13414359092712402,
-0.0660269483923912,
0.0017900742823258042,
-0.017020488157868385,
0.055780790746212006,
0.06318815052509308,
0.020168932154774666,
-0.061805132776498795,
-0.28363022208213806,
-0.016407405957579613,
0.03769398853182793,
0.02201157622039318,
-0.2418878972530365,
-0.08972179144620895,
-0.009713714942336082,
-0.04622073844075203,
-0.07431346923112869,
0.0936308428645134,
0.0790051743388176,
0.03143533319234848,
-0.05444464832544327,
-0.05092444643378258,
-0.02840057946741581,
0.17316226661205292,
-0.16338981688022614,
-0.11526208370923996
] |
null | null | diffusers |
# SDXL LoRA DreamBooth - yaneq/jan_8gr59VrqueLphjEKA6kl_SDXL_LoRA_900_9d94_900_1e4_2
<Gallery />
## Model description
These are yaneq/jan_8gr59VrqueLphjEKA6kl_SDXL_LoRA_900_9d94_900_1e4_2 LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.
The weights were trained using [DreamBooth](https://dreambooth.github.io/).
LoRA for the text encoder was enabled: False.
Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
## Trigger words
You should use a photo of MDDL man to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
[Download](yaneq/jan_8gr59VrqueLphjEKA6kl_SDXL_LoRA_900_9d94_900_1e4_2/tree/main) them in the Files & versions tab.
## Training properties
- max_train_steps: 900
- learning_rate: 0.0001
- base_model_name: stabilityai/stable-diffusion-xl-base-1.0
- class_name: man
- training_images_urls: - https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2Fcn54hvM4ahi3MzpCQN5D.jpg?alt=media&token=e096f4dc-e7c5-4e14-88fc-a5562d103127
- https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2FY7nFiafx8co1nK6cnjWJ.jpg?alt=media&token=a1fe8c9a-4d5e-4043-9a82-9304fd430569
- https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2Fz8D9WdMIx4mXcsDGAZm4.jpg?alt=media&token=fded9422-eb7c-4757-8c1f-cb436a348579
- https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2FDAk5k1hGzP9q9y0jpGoO.jpg?alt=media&token=01ed67d1-938a-4f60-bc1a-e1b91412b97e
- https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2F82McawlxnTeA2vBc4bZg.jpg?alt=media&token=f7cfacb2-2186-4005-9211-b7ef762dafad
- https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2FWF2NGBPUFgu9eyaCYAwB.jpg?alt=media&token=97c1e215-0a96-4fdf-b292-9ee0e497ba72
- https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2F6JW19SVZPczh5B2DEqKD.jpg?alt=media&token=0e0dc94f-957d-4b51-8979-0216c0849cf6
- https://firebasestorage.googleapis.com/v0/b/axonic-looks.appspot.com/o/models%2FSBGA9KzaKdSZWWzsvHMP%2FSBGA9KzaKdSZWWzsvHMP%2FVYOVRhojKt30NzjWRXL0.jpg?alt=media&token=5a3a2afb-4b83-4488-92e5-6651f5173cc0
- gradient_accumulation_steps: 3
- GPU: T4
- duration: 6676.244818210602
| {"license": "openrail++", "tags": ["stable-diffusion-xl", "stable-diffusion-xl-diffusers", "text-to-image", "diffusers", "lora", "template:sd-lora"], "base_model": "stabilityai/stable-diffusion-xl-base-1.0", "instance_prompt": "a photo of MDDL man"} | text-to-image | yaneq/jan_8gr59VrqueLphjEKA6kl_SDXL_LoRA_900_9d94_900_1e4_2 | [
"diffusers",
"stable-diffusion-xl",
"stable-diffusion-xl-diffusers",
"text-to-image",
"lora",
"template:sd-lora",
"base_model:stabilityai/stable-diffusion-xl-base-1.0",
"license:openrail++",
"has_space",
"region:us"
] | 2024-02-07T06:35:05+00:00 | [] | [] | TAGS
#diffusers #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us
|
# SDXL LoRA DreamBooth - yaneq/jan_8gr59VrqueLphjEKA6kl_SDXL_LoRA_900_9d94_900_1e4_2
<Gallery />
## Model description
These are yaneq/jan_8gr59VrqueLphjEKA6kl_SDXL_LoRA_900_9d94_900_1e4_2 LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.
The weights were trained using DreamBooth.
LoRA for the text encoder was enabled: False.
Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
## Trigger words
You should use a photo of MDDL man to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
Download them in the Files & versions tab.
## Training properties
- max_train_steps: 900
- learning_rate: 0.0001
- base_model_name: stabilityai/stable-diffusion-xl-base-1.0
- class_name: man
- training_images_urls: - URL
- URL
- URL
- URL
- URL
- URL
- URL
- URL
- gradient_accumulation_steps: 3
- GPU: T4
- duration: 6676.244818210602
| [
"# SDXL LoRA DreamBooth - yaneq/jan_8gr59VrqueLphjEKA6kl_SDXL_LoRA_900_9d94_900_1e4_2\n\n<Gallery />",
"## Model description\n\nThese are yaneq/jan_8gr59VrqueLphjEKA6kl_SDXL_LoRA_900_9d94_900_1e4_2 LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.\n\nThe weights were trained using DreamBooth.\n\nLoRA for the text encoder was enabled: False.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix.",
"## Trigger words\n\nYou should use a photo of MDDL man to trigger the image generation.",
"## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab.",
"## Training properties\n- max_train_steps: 900\n- learning_rate: 0.0001\n- base_model_name: stabilityai/stable-diffusion-xl-base-1.0\n- class_name: man\n- training_images_urls: - URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- gradient_accumulation_steps: 3\n- GPU: T4\n- duration: 6676.244818210602"
] | [
"TAGS\n#diffusers #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us \n",
"# SDXL LoRA DreamBooth - yaneq/jan_8gr59VrqueLphjEKA6kl_SDXL_LoRA_900_9d94_900_1e4_2\n\n<Gallery />",
"## Model description\n\nThese are yaneq/jan_8gr59VrqueLphjEKA6kl_SDXL_LoRA_900_9d94_900_1e4_2 LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.\n\nThe weights were trained using DreamBooth.\n\nLoRA for the text encoder was enabled: False.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix.",
"## Trigger words\n\nYou should use a photo of MDDL man to trigger the image generation.",
"## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab.",
"## Training properties\n- max_train_steps: 900\n- learning_rate: 0.0001\n- base_model_name: stabilityai/stable-diffusion-xl-base-1.0\n- class_name: man\n- training_images_urls: - URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- gradient_accumulation_steps: 3\n- GPU: T4\n- duration: 6676.244818210602"
] | [
82,
51,
116,
19,
28,
102
] | [
"passage: TAGS\n#diffusers #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us \n# SDXL LoRA DreamBooth - yaneq/jan_8gr59VrqueLphjEKA6kl_SDXL_LoRA_900_9d94_900_1e4_2\n\n<Gallery />## Model description\n\nThese are yaneq/jan_8gr59VrqueLphjEKA6kl_SDXL_LoRA_900_9d94_900_1e4_2 LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.\n\nThe weights were trained using DreamBooth.\n\nLoRA for the text encoder was enabled: False.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix.## Trigger words\n\nYou should use a photo of MDDL man to trigger the image generation.## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab.## Training properties\n- max_train_steps: 900\n- learning_rate: 0.0001\n- base_model_name: stabilityai/stable-diffusion-xl-base-1.0\n- class_name: man\n- training_images_urls: - URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- URL\n- gradient_accumulation_steps: 3\n- GPU: T4\n- duration: 6676.244818210602"
] | [
-0.08391597121953964,
0.17559656500816345,
-0.0009010335779748857,
0.07892847061157227,
0.14801260828971863,
0.07371649891138077,
0.14368706941604614,
0.13194456696510315,
0.08508437871932983,
0.12607254087924957,
0.030857715755701065,
0.06284894049167633,
0.09124016761779785,
0.15328463912010193,
0.0018332932377234101,
-0.11694500595331192,
0.023884495720267296,
-0.05381598323583603,
0.004501123912632465,
0.030048269778490067,
0.017864562571048737,
-0.05999104678630829,
0.07003498822450638,
-0.009428766556084156,
-0.03822765126824379,
0.045283377170562744,
-0.006916385609656572,
-0.020180128514766693,
0.02584690973162651,
0.04397770017385483,
0.047966133803129196,
0.032483670860528946,
0.110230453312397,
-0.25003695487976074,
0.01723405346274376,
0.1056239902973175,
-0.04653380066156387,
0.04910798743367195,
0.006499258801341057,
-0.03712162375450134,
-0.0111176036298275,
-0.15410861372947693,
0.08117128908634186,
0.05939045920968056,
-0.041057784110307693,
-0.24259193241596222,
-0.12411908805370331,
0.047611303627491,
0.07697867602109909,
0.11055351793766022,
0.004954291507601738,
0.06491655856370926,
0.05074212700128555,
0.03853613883256912,
0.2321757823228836,
-0.11265008896589279,
-0.012036368250846863,
0.1407083123922348,
0.018772248178720474,
0.06906934827566147,
0.0014336364110931754,
-0.0011644434416666627,
0.07400617003440857,
-0.010320354253053665,
0.05042099580168724,
-0.034668609499931335,
0.02506411448121071,
-0.07108967751264572,
-0.0948466956615448,
-0.0211732666939497,
0.1546591967344284,
0.06148311123251915,
-0.10456934571266174,
-0.12540023028850555,
-0.011285445652902126,
0.033238016068935394,
0.003913511522114277,
0.002232825383543968,
0.01686602085828781,
-0.016840027645230293,
0.01377845834940672,
-0.10536061972379684,
-0.013561324216425419,
-0.0582006499171257,
0.07483576983213425,
0.181001216173172,
0.028813837096095085,
-0.0051373448222875595,
0.02756240963935852,
0.03310585394501686,
-0.0750238448381424,
-0.11103874444961548,
-0.03990580514073372,
-0.02249383181333542,
-0.05214696377515793,
0.016610687598586082,
-0.022872604429721832,
-0.08528727293014526,
0.01398442778736353,
0.09720111638307571,
-0.03940005227923393,
0.048287246376276016,
-0.0398656390607357,
0.028899941593408585,
-0.07986661791801453,
0.11947578936815262,
-0.04989923909306526,
-0.013116026297211647,
0.030983073636889458,
0.10370451956987381,
0.10055947303771973,
-0.0299304760992527,
-0.1319674551486969,
-0.024656344205141068,
-0.08667971193790436,
0.06618963181972504,
-0.004054857883602381,
-0.01272543054074049,
-0.11120975762605667,
-0.04099413380026817,
0.15514256060123444,
-0.07442989945411682,
0.03388935327529907,
0.003904688870534301,
-0.1002192571759224,
0.02665185183286667,
0.1405949890613556,
-0.00020243131439201534,
-0.04964758828282356,
0.088863305747509,
-0.03512877970933914,
0.014791974797844887,
-0.04477423056960106,
-0.15648828446865082,
0.038098204880952835,
-0.13690978288650513,
-0.007422741036862135,
-0.1196952611207962,
-0.21922791004180908,
-0.05946054309606552,
0.009827054105699062,
-0.020084280520677567,
0.03815656900405884,
-0.07282119244337082,
-0.09856051206588745,
0.025371914729475975,
0.053636401891708374,
-0.06305491924285889,
-0.014465243555605412,
0.04201779142022133,
-0.06641736626625061,
0.11749135702848434,
0.006556344218552113,
0.01992960460484028,
-0.04274679720401764,
0.013499045744538307,
-0.12973730266094208,
0.11726485937833786,
-0.0997670590877533,
0.03472410887479782,
-0.12077485024929047,
-0.09533410519361496,
0.03713664785027504,
0.0006371294730342925,
0.07027021795511246,
0.10847754776477814,
-0.20493681728839874,
-0.06356227397918701,
0.11136091500520706,
-0.14512576162815094,
-0.07932217419147491,
0.021913548931479454,
-0.03580694645643234,
0.05547022074460983,
0.050198767334222794,
0.12052344530820847,
0.14581866562366486,
-0.1716044396162033,
-0.060475386679172516,
-0.03546387702226639,
0.01730041578412056,
0.08648838847875595,
0.037484604865312576,
-0.03901948034763336,
-0.008175530470907688,
0.008901967667043209,
-0.0342942476272583,
0.02321971394121647,
0.021692069247364998,
-0.05647502467036247,
-0.034527819603681564,
-0.09928734600543976,
0.06451244652271271,
-0.004138631746172905,
-0.009205534122884274,
-0.010932737030088902,
-0.09331072866916656,
0.0077634165063500404,
0.1324639767408371,
-0.09103833138942719,
0.022120270878076553,
-0.04423528537154198,
-0.03218666464090347,
-0.020965587347745895,
0.00016121158841997385,
-0.10014606267213821,
-0.12225482612848282,
0.08467472344636917,
-0.18195803463459015,
-0.004706765059381723,
-0.09358391910791397,
0.05869624763727188,
0.05930262431502342,
-0.0511636957526207,
-0.08608560264110565,
-0.07293471693992615,
-0.01363347191363573,
-0.005149646196514368,
-0.13477647304534912,
-0.049639903008937836,
-0.03825812414288521,
0.07691185176372528,
-0.19105839729309082,
0.058783601969480515,
-0.004516650456935167,
0.09559662640094757,
0.054524000734090805,
-0.052856456488370895,
0.04180411994457245,
0.032685473561286926,
-0.013343945145606995,
-0.10431614518165588,
-0.007612843997776508,
-0.035009633749723434,
-0.1651819497346878,
0.020213622599840164,
-0.13558180630207062,
-0.05694195628166199,
0.09141082316637039,
0.11122655868530273,
-0.07960578054189682,
0.006122123450040817,
0.005743951536715031,
-0.03839418664574623,
-0.05888444557785988,
-0.06065405160188675,
0.05061982199549675,
0.005021647084504366,
0.08711986988782883,
-0.0552540123462677,
0.002450118772685528,
0.039558008313179016,
0.006910574156790972,
-0.05314994230866432,
0.09869466722011566,
0.008001690730452538,
-0.003317680209875107,
0.03981248289346695,
0.00021685520187020302,
-0.06473483890295029,
0.13020449876785278,
0.05629854276776314,
-0.050481848418712616,
-0.028066541999578476,
0.018698519095778465,
0.0890565887093544,
0.08973165601491928,
0.08123616129159927,
0.08176007121801376,
0.018364347517490387,
-0.017151182517409325,
-0.03306201100349426,
-0.1329074203968048,
-0.04465475305914879,
0.02966895140707493,
-0.07480370253324509,
0.010340518318116665,
0.032481513917446136,
-0.04237573221325874,
0.049055177718400955,
-0.03165844827890396,
0.10855186730623245,
0.02122150920331478,
-0.04809393733739853,
-0.11497055739164352,
0.13920858502388,
-0.13759534060955048,
-0.17988520860671997,
-0.14115311205387115,
0.03390282765030861,
-0.047460123896598816,
-0.017937632277607918,
0.0609009750187397,
-0.14635689556598663,
-0.0786178931593895,
-0.09742124378681183,
-0.03753747045993805,
0.007480775937438011,
-0.027622828260064125,
0.07275865226984024,
0.0005325519596226513,
0.12574639916419983,
-0.10478781908750534,
0.005710867699235678,
0.01722765900194645,
-0.047315143048763275,
-0.03467912599444389,
0.011316648684442043,
0.09827558696269989,
0.04249526187777519,
0.049649208784103394,
0.02962338924407959,
-0.012403523549437523,
0.1932699978351593,
-0.021509548649191856,
0.09284734725952148,
0.19853079319000244,
0.06201799586415291,
0.1126953586935997,
0.13911619782447815,
0.04063764587044716,
-0.04638802632689476,
0.04781101644039154,
0.0443405844271183,
-0.08652126789093018,
-0.23424740135669708,
-0.11187294870615005,
-0.027321049943566322,
-0.06805264204740524,
0.1643730252981186,
0.053707554936409,
0.09407281875610352,
0.10618144273757935,
-0.07568585127592087,
0.08950915932655334,
0.04265877231955528,
0.12403033673763275,
0.03855548053979874,
0.07022381573915482,
0.009722600691020489,
-0.06320447474718094,
0.040270332247018814,
0.08019068837165833,
0.023160064592957497,
0.20057977735996246,
-0.09345487505197525,
0.046109430491924286,
-0.006463215220719576,
0.035237863659858704,
-0.011613385751843452,
0.008490964770317078,
-0.01762641780078411,
-0.00035232806112617254,
0.01930486597120762,
-0.12774311006069183,
0.07173504680395126,
0.15605048835277557,
-0.03491207957267761,
0.05525774881243706,
-0.050911176949739456,
0.018613724038004875,
0.019840668886899948,
0.05311741307377815,
0.04489270970225334,
-0.24895842373371124,
-0.035398852080106735,
0.04819166287779808,
0.10787972062826157,
0.00038165997830219567,
-0.028237534686923027,
0.1300685703754425,
-0.0774657279253006,
0.1286228448152542,
-0.024010179564356804,
0.07932055741548538,
-0.038590967655181885,
-0.05237385630607605,
0.11738167703151703,
0.1686115860939026,
0.007226492743939161,
0.0195204745978117,
-0.16776268184185028,
0.07295647263526917,
0.03455440327525139,
0.11946772783994675,
-0.060858216136693954,
0.03972356393933296,
0.03618258237838745,
-0.01273054163902998,
0.13270623981952667,
0.033178843557834625,
-0.039255909621715546,
-0.202878937125206,
-0.04572639241814613,
-0.03715991973876953,
0.08326662331819534,
-0.06315328180789948,
0.06584558635950089,
0.04528096690773964,
-0.03156208246946335,
0.006186881102621555,
0.010634913109242916,
-0.18339918553829193,
-0.1470707654953003,
0.02663181722164154,
0.1345195174217224,
0.029377629980444908,
-0.089724101126194,
-0.07000048458576202,
-0.063212089240551,
0.09031780809164047,
-0.11778054386377335,
-0.10244890302419662,
-0.09394173324108124,
-0.006325879134237766,
0.10823741555213928,
-0.06077811121940613,
0.0489983856678009,
0.01497893501073122,
0.09823442995548248,
-0.10398921370506287,
-0.10192835330963135,
0.015583299100399017,
-0.08018369972705841,
-0.21143393218517303,
-0.08995041251182556,
0.09388992935419083,
0.07143697142601013,
-0.0004044876841362566,
-0.01245963852852583,
0.013712121173739433,
0.04406828060746193,
-0.13127131760120392,
0.07105718553066254,
0.24427242577075958,
0.027823375537991524,
0.16323991119861603,
-0.03758881986141205,
-0.05269143730401993,
-0.05909182131290436,
-0.014834930188953876,
0.020427342504262924,
0.27361416816711426,
-0.12752895057201385,
0.06957724690437317,
0.06179199367761612,
-0.07787363231182098,
-0.1285257637500763,
0.07153298705816269,
0.10082514584064484,
-0.01514314766973257,
0.008381921797990799,
-0.22311623394489288,
0.050504203885793686,
0.06785058975219727,
-0.006062926724553108,
0.07089553028345108,
-0.35586315393447876,
-0.14776663482189178,
-0.0734698623418808,
0.12143237143754959,
0.0371621735394001,
-0.12001169472932816,
-0.05498490482568741,
-0.12151534855365753,
-0.07332588732242584,
0.12386471778154373,
-0.017483431845903397,
0.06419727951288223,
-0.0015445334138348699,
0.0056728823110461235,
0.02645600587129593,
-0.06260836869478226,
0.162850484251976,
0.015043716877698898,
0.06264863908290863,
-0.03253955394029617,
0.047750189900398254,
0.08800870925188065,
-0.08525728434324265,
0.08548945188522339,
-0.07070434093475342,
0.06335746496915817,
-0.08271627128124237,
-0.00079459382686764,
-0.02184421755373478,
0.05075510963797569,
-0.04539790377020836,
-0.0419442318379879,
-0.07216562330722809,
0.06839986890554428,
0.14006784558296204,
-0.014956086874008179,
-0.03857827186584473,
0.015694864094257355,
0.09065134823322296,
0.12573480606079102,
0.012248619459569454,
0.15249599516391754,
-0.08616068959236145,
0.02399544231593609,
-0.06808726489543915,
0.10040357708930969,
-0.15931984782218933,
0.022587260231375694,
0.07429027557373047,
0.04848143830895424,
0.14189481735229492,
0.004502489697188139,
-0.11217654496431351,
0.06370922923088074,
0.017743783071637154,
-0.08486264944076538,
-0.10790220648050308,
-0.018968097865581512,
0.008500419557094574,
-0.05376443639397621,
0.005741443485021591,
0.11956732720136642,
-0.07169894874095917,
0.011599955148994923,
-0.00483773835003376,
0.06592725962400436,
-0.0027648392133414745,
0.1177438423037529,
0.072207972407341,
0.03025505691766739,
-0.06384321302175522,
0.12570081651210785,
0.051881056278944016,
-0.07049956172704697,
0.10187631845474243,
0.09188685566186905,
-0.09066910296678543,
-0.007006607484072447,
-0.11386076360940933,
0.047686636447906494,
-0.061750344932079315,
-0.04178861156105995,
-0.04068397358059883,
-0.0649053156375885,
0.0018497923156246543,
0.10280422121286392,
0.01415610034018755,
-0.008405585773289204,
0.01831570453941822,
-0.027854327112436295,
-0.11174868047237396,
0.0629834458231926,
0.024348527193069458,
0.06350767612457275,
-0.15781813859939575,
-0.007587319705635309,
0.05510186031460762,
-0.013454136438667774,
0.014407297596335411,
0.0049868375062942505,
-0.1561332643032074,
-0.03430718556046486,
-0.08900731801986694,
0.11617910116910934,
-0.07408293336629868,
0.01851031184196472,
-0.040220677852630615,
-0.01977706514298916,
0.002657979028299451,
0.07001321762800217,
-0.023787569254636765,
-0.10690153390169144,
0.01961521990597248,
0.06075188145041466,
-0.11457233130931854,
-0.05408754199743271,
-0.012962119653820992,
-0.07338496297597885,
0.05211269110441208,
0.009046266786754131,
-0.046105869114398956,
-0.0056151943281292915,
-0.14929185807704926,
0.06421762704849243,
0.057840101420879364,
0.02188130095601082,
0.00045920300181023777,
-0.10082592815160751,
0.007560300640761852,
-0.05541539937257767,
-0.06083756312727928,
-0.002111733192577958,
0.011177323758602142,
-0.10265342146158218,
0.009256708435714245,
-0.03277071937918663,
0.031825028359889984,
-0.09241949766874313,
0.0883568823337555,
0.1480056345462799,
0.03654058277606964,
0.10108070075511932,
-0.08981471508741379,
0.096493661403656,
-0.14907988905906677,
-0.044399477541446686,
0.00005322856304701418,
-0.02863943949341774,
-0.030098807066679,
-0.0008494240464642644,
0.06894551217556,
-0.013050827197730541,
0.11114875227212906,
-0.07040688395500183,
0.0683588758111,
-0.009016756899654865,
-0.01883021555840969,
-0.04604557529091835,
-0.003002018202096224,
0.1365870088338852,
-0.01884756051003933,
0.007240384817123413,
0.01307035330682993,
-0.00748720346018672,
0.07503847777843475,
0.10477814078330994,
0.03466523438692093,
0.08837667107582092,
0.09165926277637482,
0.04691735655069351,
0.03550228476524353,
-0.07643063366413116,
-0.054625093936920166,
0.15253080427646637,
-0.015917938202619553,
0.08111230283975601,
-0.05388094484806061,
0.01958196423947811,
0.061885833740234375,
-0.14920645952224731,
0.006582632660865784,
0.07322409749031067,
-0.06755942106246948,
-0.0919940173625946,
-0.16075709462165833,
-0.07842414081096649,
-0.13858389854431152,
0.011223643086850643,
-0.0818624421954155,
0.05254826322197914,
0.006032801698893309,
-0.03603356331586838,
0.014309133403003216,
0.08897948265075684,
-0.03254286199808121,
-0.012684072367846966,
0.027033355087041855,
-0.016849493607878685,
-0.020799262449145317,
0.0826747938990593,
-0.05394672974944115,
0.12741400301456451,
-0.0037267813459038734,
0.033043183386325836,
0.04711037501692772,
0.05623750388622284,
0.043433524668216705,
-0.004383960273116827,
-0.07909369468688965,
0.008416548371315002,
0.04105580225586891,
0.006489052437245846,
0.16857102513313293,
0.09567105770111084,
0.014669603668153286,
-0.037981804460287094,
0.15181151032447815,
-0.07585471868515015,
-0.006309601943939924,
-0.10196178406476974,
0.07034122198820114,
0.0017086694715544581,
-0.00033970799995586276,
0.003580331103876233,
-0.10488463938236237,
-0.0343996100127697,
0.0773310512304306,
0.10131505131721497,
0.022266531363129616,
0.028009744361042976,
-0.04309898987412453,
-0.01368891540914774,
-0.054796002805233,
0.04789293184876442,
0.05421421304345131,
0.07325439900159836,
-0.0069829984568059444,
0.04220074787735939,
-0.021931424736976624,
-0.02143261767923832,
-0.037313349545001984,
0.017896192148327827,
-0.051141899079084396,
-0.02456938661634922,
-0.04359326511621475,
0.05740024894475937,
-0.03883490711450577,
-0.14979299902915955,
0.17368215322494507,
-0.12591825425624847,
-0.06476569175720215,
-0.013433151878416538,
0.06850326061248779,
0.028175240382552147,
0.007237904705107212,
-0.030872296541929245,
-0.016559945419430733,
0.09454813599586487,
-0.04715964198112488,
-0.06328891962766647,
-0.11820686608552933,
-0.04519497975707054,
-0.12106809765100479,
0.23066148161888123,
-0.002150076674297452,
0.06285230815410614,
0.08337387442588806,
0.0323241651058197,
-0.13943444192409515,
0.06253845989704132,
0.022726383060216904,
-0.11397913098335266,
-0.025117065757513046,
0.1819782704114914,
-0.0815015435218811,
0.20115695893764496,
0.04624195769429207,
-0.10105288773775101,
0.02340465784072876,
0.06438717246055603,
-0.029916122555732727,
-0.08158416301012039,
0.007314763031899929,
-0.11277325451374054,
0.12777592241764069,
0.1553797572851181,
-0.039815980941057205,
-0.02474459819495678,
0.014488729648292065,
0.01985710859298706,
0.0146514056250453,
0.09845798462629318,
0.045656926929950714,
-0.09700559079647064,
0.029774243012070656,
0.007709778379648924,
0.11638451367616653,
-0.16679498553276062,
-0.07052110880613327,
-0.06503946334123611,
-0.05952407047152519,
0.0146370530128479,
0.08300046622753143,
0.0767848938703537,
0.03439085930585861,
-0.043949127197265625,
-0.2005295604467392,
0.005200514569878578,
0.16146209836006165,
-0.0880870521068573,
-0.08587422221899033
] |
null | null | diffusers |
# DreamBooth trained by AutoTrain
Text encoder was not trained.
| {"tags": ["text-to-image", "diffusers", "autotrain"], "base_model": "stabilityai/stable-diffusion-xl-base-1.0", "instance_prompt": "photo of a sks business woman hijab person", "inference": true} | text-to-image | nry61/sdxl_businessWoman | [
"diffusers",
"text-to-image",
"autotrain",
"base_model:stabilityai/stable-diffusion-xl-base-1.0",
"region:us"
] | 2024-02-07T06:35:42+00:00 | [] | [] | TAGS
#diffusers #text-to-image #autotrain #base_model-stabilityai/stable-diffusion-xl-base-1.0 #region-us
|
# DreamBooth trained by AutoTrain
Text encoder was not trained.
| [
"# DreamBooth trained by AutoTrain\n\nText encoder was not trained."
] | [
"TAGS\n#diffusers #text-to-image #autotrain #base_model-stabilityai/stable-diffusion-xl-base-1.0 #region-us \n",
"# DreamBooth trained by AutoTrain\n\nText encoder was not trained."
] | [
41,
19
] | [
"passage: TAGS\n#diffusers #text-to-image #autotrain #base_model-stabilityai/stable-diffusion-xl-base-1.0 #region-us \n# DreamBooth trained by AutoTrain\n\nText encoder was not trained."
] | [
-0.036245543509721756,
0.12715020775794983,
-0.001609276863746345,
0.04831791669130325,
0.18532192707061768,
0.04340039938688278,
0.18271183967590332,
0.07303067296743393,
-0.004240434616804123,
0.046504613012075424,
0.19579778611660004,
0.022584218531847,
0.011507879942655563,
0.2310102880001068,
-0.10818178206682205,
-0.149010568857193,
0.05009135231375694,
-0.015542781911790371,
0.11514975130558014,
0.049119893461465836,
0.016962043941020966,
-0.07468947768211365,
0.07071888446807861,
-0.11336331069469452,
-0.21761378645896912,
0.06550896912813187,
0.03588332608342171,
-0.0775405615568161,
0.017985541373491287,
0.07035709917545319,
0.10360643267631531,
0.05894792824983597,
0.07012995332479477,
-0.1112663671374321,
0.031294774264097214,
0.08015334606170654,
-0.032878510653972626,
0.052898384630680084,
0.01470891758799553,
-0.0026460089720785618,
-0.0517716258764267,
0.04980996251106262,
0.04736582189798355,
0.043912142515182495,
-0.07364338636398315,
0.1279863864183426,
0.017672035843133926,
0.05650009959936142,
0.025508619844913483,
0.12453371286392212,
-0.02424733340740204,
0.08564464002847672,
0.020702581852674484,
0.11359728127717972,
0.02213899977505207,
-0.1489952951669693,
-0.05670339614152908,
0.18264123797416687,
0.03225535899400711,
0.18127937614917755,
-0.09674288332462311,
0.07346276938915253,
0.1252599060535431,
0.002823482733219862,
-0.04946395009756088,
-0.014294417575001717,
-0.042641714215278625,
-0.11864033341407776,
-0.024804722517728806,
-0.02916947938501835,
0.18380407989025116,
0.023036038503050804,
-0.03249591588973999,
-0.08120134472846985,
-0.10711746662855148,
-0.0008855643682181835,
0.010800139047205448,
0.013396238908171654,
-0.06835334002971649,
0.06816162914037704,
-0.043746016919612885,
-0.08182105422019958,
-0.03586805239319801,
-0.041306957602500916,
-0.0747617855668068,
0.09806115925312042,
-0.030058175325393677,
0.08185292035341263,
-0.09660601615905762,
0.1333814412355423,
-0.026361621916294098,
-0.12627731263637543,
0.0599898137152195,
-0.09759627282619476,
0.026095055043697357,
0.06731947511434555,
0.0029132110066711903,
-0.18100294470787048,
0.03379857540130615,
0.03375234827399254,
0.0869336947798729,
0.06559991091489792,
-0.09858748316764832,
0.07407204806804657,
0.006027183495461941,
0.0804583728313446,
0.009181375615298748,
-0.022433903068304062,
0.09147199988365173,
0.07103802263736725,
0.025447633117437363,
-0.14230188727378845,
-0.16106386482715607,
0.0827116072177887,
-0.017257221043109894,
0.05629891902208328,
0.020978160202503204,
-0.01937171258032322,
-0.0405338853597641,
-0.019553186371922493,
0.04573526978492737,
-0.05446048080921173,
0.007573939394205809,
-0.07102841138839722,
-0.014178181067109108,
0.05186567083001137,
0.12478453665971756,
0.0021203935611993074,
-0.02912730723619461,
0.004244799725711346,
-0.08898386359214783,
-0.012771227397024632,
-0.05357198417186737,
-0.062320686876773834,
-0.06746604293584824,
-0.11054808646440506,
0.04183540493249893,
-0.15573084354400635,
-0.1716412454843521,
-0.004213833250105381,
0.007832624018192291,
-0.07523111999034882,
0.00938150379806757,
-0.11571332812309265,
-0.10258128494024277,
0.13168372213840485,
-0.011765006929636002,
-0.07873458415269852,
0.008203231729567051,
0.0676451027393341,
0.020236635580658913,
0.08431282639503479,
-0.17113563418388367,
0.008839869871735573,
-0.09166102111339569,
-0.009885294362902641,
-0.08497859537601471,
0.18176661431789398,
-0.026730557903647423,
0.04859798774123192,
-0.03221297636628151,
0.050733115524053574,
0.029140038415789604,
0.003848304972052574,
0.04733651503920555,
0.16666279733181,
-0.17242948710918427,
-0.06218917295336723,
0.11715637892484665,
-0.09599082916975021,
-0.01863897778093815,
0.047500964254140854,
-0.021074924618005753,
0.047988228499889374,
0.041481491178274155,
0.11401445418596268,
-0.07589904963970184,
-0.14255496859550476,
0.015253559686243534,
0.003455876372754574,
-0.05091823637485504,
0.05532316491007805,
-0.04852023348212242,
0.05283749848604202,
-0.1097816452383995,
0.04694690555334091,
-0.008769541047513485,
0.0859716609120369,
-0.07339466363191605,
-0.06417720764875412,
-0.05244520306587219,
-0.0228950884193182,
0.05550098791718483,
0.019141118973493576,
0.07436100393533707,
-0.028765428811311722,
-0.06233592703938484,
0.035375721752643585,
0.04475251957774162,
-0.025758201256394386,
-0.016597965732216835,
-0.033692605793476105,
-0.059320833534002304,
-0.12508338689804077,
0.015558759681880474,
-0.08364531397819519,
-0.08999279886484146,
0.030006708577275276,
0.23967526853084564,
0.10251615196466446,
0.08612723648548126,
0.061787668615579605,
0.05093790963292122,
-0.039444345980882645,
-0.1347617208957672,
0.009005658328533173,
0.007035030052065849,
-0.0708971619606018,
-0.10834814608097076,
0.11448300629854202,
-0.07531768083572388,
0.012892637401819229,
-0.1546088606119156,
0.0028806638438254595,
-0.09619028866291046,
0.1506405621767044,
0.030817557126283646,
-0.03606857731938362,
-0.03634079173207283,
0.04110613092780113,
-0.08826293051242828,
-0.10075326263904572,
0.00478385528549552,
0.020455453544855118,
-0.10564742237329483,
0.05573726445436478,
-0.2142423689365387,
0.037837374955415726,
0.12913189828395844,
-0.0013331864029169083,
-0.06938981264829636,
0.09541511535644531,
0.06282404810190201,
-0.02221987582743168,
-0.023385949432849884,
-0.00370846688747406,
0.14985083043575287,
-0.06380517780780792,
0.21462255716323853,
-0.01891150139272213,
0.09149244427680969,
0.056990042328834534,
-0.07628677785396576,
-0.11459953337907791,
0.002614207100123167,
-0.0183817520737648,
-0.05103599280118942,
0.0967637449502945,
0.032428447157144547,
-0.050694238394498825,
0.27045416831970215,
-0.00465793814510107,
-0.007226756773889065,
-0.02355467714369297,
0.006083228159695864,
-0.013260210864245892,
0.10309924185276031,
0.001494184136390686,
0.02175363525748253,
0.004299492575228214,
-0.026297640055418015,
0.009904916398227215,
-0.08100980520248413,
-0.006161799654364586,
-0.04796920716762543,
-0.03257853537797928,
0.1331891417503357,
-0.006251717917621136,
-0.04229942709207535,
0.06579606980085373,
-0.05657486617565155,
-0.0884828194975853,
0.12328140437602997,
-0.028618741780519485,
-0.018217967823147774,
0.0763513445854187,
-0.16376429796218872,
-0.2900826334953308,
-0.16402651369571686,
-0.013422850519418716,
-0.13986164331436157,
0.03420291095972061,
0.048710525035858154,
-0.11512403935194016,
-0.08244767785072327,
-0.06028057634830475,
-0.08225703239440918,
-0.038554511964321136,
-0.0029783067293465137,
0.1302110105752945,
-0.06166526675224304,
0.04493628442287445,
-0.0501401387155056,
-0.002971479669213295,
-0.02617313340306282,
0.0021297838538885117,
0.10081592947244644,
-0.00761022325605154,
0.03796780854463577,
0.202129065990448,
-0.010098110884428024,
0.044847313314676285,
-0.005482283420860767,
0.23947332799434662,
-0.0626927837729454,
0.045369282364845276,
0.11615905910730362,
0.006890029646456242,
0.0580100379884243,
0.16308575868606567,
-0.011474157683551311,
-0.07871859520673752,
0.07068836688995361,
-0.018553484231233597,
-0.09487346559762955,
-0.13651004433631897,
-0.08400432765483856,
-0.03227277845144272,
-0.05915997549891472,
0.032944828271865845,
0.060708627104759216,
0.2048250138759613,
0.03822575882077217,
-0.014925431460142136,
0.03945036232471466,
-0.021271005272865295,
0.058776192367076874,
0.05568648502230644,
-0.03110971301794052,
0.08776143193244934,
-0.05410825088620186,
-0.08081484586000443,
0.10256079584360123,
0.014764479361474514,
0.10078063607215881,
0.0015681018121540546,
-0.026868712157011032,
-0.06210172548890114,
0.03388858214020729,
0.13113942742347717,
0.03242070600390434,
0.05939389020204544,
-0.03396468609571457,
-0.03142368420958519,
-0.04893866553902626,
-0.028402801603078842,
0.07279334962368011,
-0.0015502171590924263,
0.03518500179052353,
-0.07344528287649155,
0.13397738337516785,
-0.009692627936601639,
0.058724015951156616,
0.08062360435724258,
-0.25072693824768066,
0.03393854945898056,
0.037297505885362625,
0.00938115082681179,
-0.16026899218559265,
-0.0000797836109995842,
0.22579798102378845,
-0.09356516599655151,
0.005204774439334869,
-0.023821650072932243,
0.08528044074773788,
0.05810416862368584,
-0.025039739906787872,
-0.1263386607170105,
0.11799584329128265,
-0.03718066215515137,
-0.011419684626162052,
-0.2191302478313446,
0.026818642392754555,
0.011973787099123001,
0.13413779437541962,
-0.019075613468885422,
0.018596889451146126,
0.02343740314245224,
0.14612221717834473,
0.08113942295312881,
-0.0025373667012900114,
-0.047016777098178864,
-0.1369040161371231,
-0.1009141132235527,
-0.04061482846736908,
0.11433621495962143,
0.10622173547744751,
-0.00976573582738638,
-0.008132239803671837,
0.019755136221647263,
0.021546775475144386,
-0.13333141803741455,
-0.22008948028087616,
-0.12137539684772491,
0.028061518445611,
0.1722775399684906,
0.09369596838951111,
-0.033191703259944916,
-0.06751665472984314,
0.11195974051952362,
0.17584778368473053,
-0.09410375356674194,
-0.05504462495446205,
-0.12103758007287979,
-0.01799076609313488,
0.033899057656526566,
-0.020356787368655205,
0.07893814891576767,
-0.11844030022621155,
0.04901885241270065,
-0.046563804149627686,
-0.16906501352787018,
0.07727976143360138,
-0.09955182671546936,
-0.08727449923753738,
-0.10217057913541794,
-0.005441801622509956,
-0.055659178644418716,
-0.04306938499212265,
0.03298354893922806,
0.00520846713334322,
-0.06480726599693298,
-0.0740930438041687,
0.06889233738183975,
0.06416995823383331,
-0.10268759727478027,
0.11831934750080109,
0.03763728588819504,
-0.06115265190601349,
-0.0276818685233593,
-0.02785094454884529,
0.1583004891872406,
0.27498534321784973,
-0.06539037078619003,
0.12848426401615143,
0.1529870629310608,
-0.0934448093175888,
-0.26671624183654785,
-0.077382393181324,
0.01659390702843666,
0.017424149438738823,
-0.07041475176811218,
-0.1103808730840683,
0.042794182896614075,
-0.02292712777853012,
-0.007760640233755112,
0.11937795579433441,
-0.2564130127429962,
-0.07278794050216675,
0.12020199000835419,
0.03494739532470703,
0.31451332569122314,
-0.11796464771032333,
-0.0594152957201004,
-0.0897473692893982,
0.02157750353217125,
0.08759456127882004,
0.1022130474448204,
0.14900463819503784,
-0.016340946778655052,
0.019258301705121994,
0.01253960095345974,
-0.028015948832035065,
0.13074079155921936,
-0.08528809249401093,
0.06432783603668213,
-0.08787696063518524,
0.04536893218755722,
0.15485939383506775,
-0.05565318465232849,
0.0666361078619957,
-0.08315334469079971,
0.08477341383695602,
-0.13287608325481415,
0.005565148778259754,
-0.016947530210018158,
0.018647341057658195,
0.037531763315200806,
-0.0982481837272644,
-0.029500316828489304,
-0.042952634394168854,
0.0308607816696167,
0.007574446499347687,
-0.008083799853920937,
-0.027663223445415497,
-0.0069111417979002,
0.28180161118507385,
-0.013126120902597904,
-0.09530629962682724,
-0.028030000627040863,
-0.029287368059158325,
-0.08341575413942337,
0.14523804187774658,
-0.09800203889608383,
0.02303970232605934,
0.09674859046936035,
-0.026513192802667618,
0.20614224672317505,
0.037182051688432693,
-0.02977730706334114,
0.06790098547935486,
0.07232028245925903,
-0.16948382556438446,
0.03898109495639801,
-0.09782025963068008,
0.038938265293836594,
0.0875864177942276,
-0.06356405466794968,
0.16797809302806854,
-0.06019335240125656,
0.0390251986682415,
-0.04562242329120636,
0.023244095966219902,
-0.017269715666770935,
0.08887730538845062,
0.03733275458216667,
0.023359721526503563,
-0.09725578129291534,
0.13855582475662231,
0.024098310619592667,
0.0046667009592056274,
0.12440912425518036,
0.1182316318154335,
-0.040450893342494965,
-0.010168218985199928,
0.005212021991610527,
0.27069801092147827,
-0.19435960054397583,
-0.023034915328025818,
-0.0586579255759716,
-0.09062426537275314,
-0.02572054974734783,
0.047252245247364044,
-0.002321677515283227,
0.0055772243067622185,
-0.0626353919506073,
-0.04633498936891556,
-0.12254648655653,
0.03370234742760658,
0.05887467414140701,
0.06559669226408005,
-0.21172763407230377,
-0.019244397059082985,
0.04228638857603073,
0.049211468547582626,
-0.12720735371112823,
-0.1044912338256836,
-0.13925014436244965,
0.01657455414533615,
-0.13711096346378326,
0.08653344959020615,
0.06315533816814423,
-0.04359058290719986,
0.03764145448803902,
-0.0378626249730587,
0.012233911082148552,
0.038636330515146255,
-0.029364967718720436,
-0.0036765364930033684,
0.03113969974219799,
-0.005094574298709631,
-0.0315641388297081,
-0.048442136496305466,
-0.05078553035855293,
-0.030254468321800232,
0.06226435303688049,
0.046048980206251144,
-0.08307506889104843,
-0.0008596591651439667,
-0.21067485213279724,
-0.018223904073238373,
0.14309637248516083,
0.0032463467214256525,
-0.02658955194056034,
0.15721344947814941,
-0.019887013360857964,
0.04572618752717972,
0.03328227251768112,
0.003835827112197876,
0.061016231775283813,
-0.11269015818834305,
-0.12115146219730377,
-0.07395296543836594,
-0.05862290412187576,
-0.09302014857530594,
0.08661885559558868,
0.09580361843109131,
0.07227188348770142,
0.13144655525684357,
-0.16063906252384186,
0.08064062148332596,
-0.06647570431232452,
-0.009136770851910114,
-0.013843890279531479,
-0.06997225433588028,
-0.005904360208660364,
-0.009263744577765465,
0.0424351766705513,
-0.016006316989660263,
0.1162041574716568,
0.05054130405187607,
-0.10427013784646988,
-0.006998004391789436,
-0.019210085272789,
-0.03673923760652542,
-0.020151400938630104,
0.2661339044570923,
0.11022793501615524,
0.0016144451219588518,
-0.09479154646396637,
0.012096131220459938,
0.13538652658462524,
0.10112990438938141,
0.012295308522880077,
-0.010951491072773933,
0.006580990739166737,
0.16312289237976074,
0.020021583884954453,
0.022472593933343887,
-0.04704653471708298,
0.024670634418725967,
-0.11195191740989685,
0.1181398406624794,
-0.07783101499080658,
-0.12228389084339142,
0.0961126908659935,
0.004306115210056305,
-0.05464071035385132,
0.026061825454235077,
-0.08663400262594223,
-0.10763459652662277,
-0.02403993532061577,
-0.08360651880502701,
-0.17054924368858337,
0.04553007706999779,
-0.0644475594162941,
0.10821449756622314,
0.03285038843750954,
0.01913030818104744,
-0.09931070357561111,
0.08828441053628922,
0.049342066049575806,
-0.08149173855781555,
0.12537957727909088,
-0.00206486857496202,
-0.02179819718003273,
-0.0685075893998146,
-0.05024031549692154,
0.07070331275463104,
0.11107023060321808,
0.0027782265096902847,
0.06980131566524506,
0.04400113597512245,
0.08451055735349655,
-0.010072652250528336,
-0.1444101780653,
0.019020602107048035,
0.07610902190208435,
-0.02439282462000847,
0.16260980069637299,
0.05885591357946396,
0.01153038814663887,
-0.030513545498251915,
0.17180931568145752,
-0.08573780953884125,
-0.08169874548912048,
-0.08695080876350403,
0.16584457457065582,
-0.09375490248203278,
0.129506915807724,
-0.07163704186677933,
-0.09326006472110748,
-0.08763561397790909,
0.14160455763339996,
0.10961220413446426,
-0.1615811288356781,
-0.030947357416152954,
-0.07122225314378738,
-0.004950015805661678,
-0.03937298059463501,
0.19569715857505798,
0.015499092638492584,
0.0950700044631958,
-0.0507570318877697,
0.029308956116437912,
-0.053977370262145996,
-0.10537504404783249,
-0.07778677344322205,
-0.09801097959280014,
0.006298981141299009,
-0.03663072735071182,
-0.1205139011144638,
-0.03740362077951431,
-0.1355041116476059,
0.09535817801952362,
0.12423786520957947,
-0.05443605035543442,
-0.032696500420570374,
-0.02045554295182228,
0.14053913950920105,
-0.009911599569022655,
-0.03675956279039383,
-0.06726814061403275,
0.056783780455589294,
0.1183142364025116,
-0.05778191238641739,
-0.037040118128061295,
-0.0331730917096138,
-0.07186004519462585,
-0.2845723628997803,
0.16318918764591217,
0.00482621043920517,
0.05915847420692444,
0.016764864325523376,
0.033719830214977264,
-0.044226400554180145,
0.13406574726104736,
-0.060890257358551025,
-0.015066491439938545,
-0.0018660984933376312,
0.19890965521335602,
-0.02104703150689602,
0.060857538133859634,
0.025196317583322525,
-0.1135951429605484,
-0.03864426165819168,
0.014360340312123299,
-0.08810338377952576,
0.011254840530455112,
-0.021632837131619453,
-0.028297800570726395,
0.10400322079658508,
0.03498094528913498,
-0.015517957508563995,
0.017761504277586937,
-0.02127593383193016,
0.022352565079927444,
-0.013210940174758434,
-0.013586577028036118,
0.03816014528274536,
-0.13842982053756714,
-0.04100436344742775,
0.10049707442522049,
0.04173333942890167,
-0.2507009208202362,
-0.045631106942892075,
-0.23455800116062164,
0.04757317155599594,
-0.0966074988245964,
0.1365291178226471,
0.1447109580039978,
-0.015541939064860344,
-0.003118799766525626,
-0.12650713324546814,
0.016526542603969574,
0.03469578176736832,
0.013082130812108517,
-0.04212687909603119
] |
null | null | transformers |
<div align="center">
<h1>
MiniCPM
</h1>
</div>
<p align="center">
<a href="https://shengdinghu.notion.site/MiniCPM-c805a17c5c8046398914e47f0542095a?pvs=4" target="_blank">MiniCPM 技术报告</a><a href="https://shengdinghu.notion.site/MiniCPM-Unveiling-the-Potential-of-End-side-Large-Language-Models-d4d3a8c426424654a4e80e42a711cb20?pvs=4" target="_blank"> Technical Report</a> |
<a href="https://github.com/OpenBMB/OmniLMM/" target="_blank">OmniLMM 多模态模型 Multi-modal Model</a> |
<a href="https://luca.cn/" target="_blank">CPM-C 千亿模型试用 ~100B Model Trial </a>
</p>
MiniCPM 是面壁与清华大学自然语言处理实验室共同开源的系列端侧语言大模型,主体语言模型 MiniCPM-2B 仅有 24亿(2.4B)的非词嵌入参数量。
- 经过 SFT 后,MiniCPM 在公开综合性评测集上,MiniCPM 与 Mistral-7B相近(中文、数学、代码能力更优),整体性能超越 Llama2-13B、MPT-30B、Falcon-40B 等模型。
- 经过 DPO 后,MiniCPM 在当前最接近用户体感的评测集 MTBench上,MiniCPM-2B 也超越了 Llama2-70B-Chat、Vicuna-33B、Mistral-7B-Instruct-v0.1、Zephyr-7B-alpha 等众多代表性开源大模型。
- 以 MiniCPM-2B 为基础构建端侧多模态大模型 MiniCPM-V,整体性能在同规模模型中实现最佳,超越基于 Phi-2 构建的现有多模态大模型,在部分评测集上达到与 9.6B Qwen-VL-Chat 相当甚至更好的性能。
- 经过 Int4 量化后,MiniCPM 可在手机上进行部署推理,流式输出速度略高于人类说话速度。MiniCPM-V 也首次跑通了多模态大模型在手机上的部署。
- 一张1080/2080可高效参数微调,一张3090/4090可全参数微调,一台机器可持续训练 MiniCPM,二次开发成本较低。
我们将完全开源MiniCPM-2B的模型参数供学术研究和有限商用,以及训练过程中的所有Checkpoint和大部分非专有数据供模型机理研究。
- 基于MiniCPM-2B的指令微调与人类偏好对**MiniCPM-2B-SFT/DPO。**
- 基于MiniCPM-2B的多模态模型**MiniCPM-V**,能力超越基于Phi-2的同参数级别多模态模型**。**
- MiniCPM-2B-SFT/DPO的Int4量化版**MiniCPM-2B-SFT/DPO-Int4。**
- 基于MLC-LLM、LLMFarm开发的MiniCPM手机端程序,**文本及多模态模型均可在手机端进行推理。**
MiniCPM is an End-Size LLM developed by ModelBest Inc. and TsinghuaNLP, with only 2.4B parameters excluding embeddings.
- MiniCPM has very close performance compared with Mistral-7B on open-sourced general benchmarks with better ability on Chinese, Mathmetics and Coding after SFT. The overall performance exceeds Llama2-13B, MPT-30B, Falcon-40B, etc.
- After DPO, MiniCPM outperforms Llama2-70B-Chat, Vicuna-33B, Mistral-7B-Instruct-v0.1, Zephyr-7B-alpha, etc. on MTBench.
- MiniCPM-V, based on MiniCPM-2B, achieves the best overall performance among multimodel models of the same scale, surpassing existing multimodal large models built on Phi-2 and achieving performance comparable to or even better than 9.6B Qwen-VL-Chat on some tasks.
- MiniCPM can be deployed and infer on smartphones, and the speed of streaming output is relatively higher than the verbal speed of human. MiniCPM-V is the first multi-modal models that can be deployed on smartphones.
- The cost of developing based on MiniCPM is low. Parameter efficient finetuning can be conducted with a single 1080/2080 GPU and full parameter finetuning can be conducted with a 3090/4090 GPU.
We release all model parameters for research and limited commercial use. We also release all the checkpoint during training and most public training data for research on model mechanism.
- SFT and DPO version based on MiniCPM-2B and human preference: **MiniCPM-2B-SFT/DPO**
- The multi-modal model **MiniCPM-V** based on MiniCPM-2B, which outperforms models with similar size, i.e., Phi-2
- The INT4 quantized version **MiniCPM-2B-SFT/DPO-Int4** based on MiniCPM-2B-SFT/DPO
- Mobile phone application based on MLC-LLM and LLMFarm. Both language model and multimodel model can conduct inference on smartphones.
### 评测结果 Evaluation Results
详细的评测结果位于[github仓库](https://github.com/OpenBMB/MiniCPM?tab=readme-ov-file#%E8%AF%84%E6%B5%8B%E7%BB%93%E6%9E%9C)
Detailed evaluation results are in [github repo](https://github.com/OpenBMB/MiniCPM/blob/main/README-en.md#evaluation-results)
注意:我们发现使用Huggingface生成质量略差于vLLM,因此推荐使用vLLM进行测试。我们正在排查原因。
Notice: We discovered that the quality of Huggingface generation is slightly lower than vLLM, thus benchmarking using vLLM is recommended.
We are investigating the cause now.
### 局限性 Limitations
- 受限于模型规模,模型可能出现幻觉性问题。其中由于DPO模型生成的回复内容更长,更容易出现幻觉。我们也将持续进行MiniCPM模型的迭代改进;
- 为了保证在学术研究用途上模型的通用性,我们未对模型进行任何身份认同训练。同时由于我们用ShareGPT开源语料作为部分训练数据,模型可能会输出类似GPT系列模型的身份认同信息;
- 受限于模型规模,模型的输出受到提示词(prompt)的影响较大,可能多次尝试产生不一致的结果;
- 受限于模型容量,模型的知识记忆较不准确,后续我们将结合RAG方法来增强模型的知识记忆能力。
- Due to limitations in model size, the model may experience hallucinatory issues. As DPO model tend to generate longer response, hallucinations are more likely to occur. We will also continue to iterate and improve the MiniCPM model.
- To ensure the universality of the model for academic research purposes, we did not conduct any identity training on the model. Meanwhile, as we use ShareGPT open-source corpus as part of the training data, the model may output identity information similar to the GPT series models.
- Due to the limitation of model size, the output of the model is greatly influenced by prompt words, which may result in inconsistent results from multiple attempts.
- Due to limited model capacity, the model's knowledge memory is not accurate. In the future, we will combine the RAG method to enhance the model's knowledge memory ability.
## 模型下载 Download
| HuggingFace | ModelScope | WiseModel |
|-------------|------------|-----------|
|[sft-bf16](https://huggingface.co/openbmb/MiniCPM-2B-sft-bf16)|[sft-bf16](https://modelscope.cn/models/OpenBMB/miniCPM-bf16)|[sft-bf16](https://wisemodel.cn/models/OpenBMB/miniCPM-bf16)
|[sft-fp32](https://huggingface.co/openbmb/MiniCPM-2B-sft-fp32)|[sft-fp32](https://modelscope.cn/models/OpenBMB/MiniCPM-2B-sft-fp32)|[sft-fp32](https://wisemodel.cn/models/OpenBMB/miniCPM-dpo-fp32)
|[dpo-bf16](https://huggingface.co/openbmb/MiniCPM-2B-dpo-bf16)|[dpo-bf16](https://modelscope.cn/models/OpenBMB/MiniCPM-2B-dpo-bf16/summary)|[dpo-bf16](https://wisemodel.cn/models/OpenBMB/MiniCPM-2B-dpo-bf16)
|[dpo-fp16](https://huggingface.co/openbmb/MiniCPM-2B-dpo-fp16)|[dpo-fp16](https://modelscope.cn/models/OpenBMB/MiniCPM-2B-dpo-fp16/)|[dpo-fp16](https://wisemodel.cn/models/OpenBMB/MiniCPM-2B-dpo-fp16)
|[dpo-fp32](https://huggingface.co/openbmb/MiniCPM-2B-dpo-fp32)|[dpo-fp32](https://modelscope.cn/models/OpenBMB/MiniCPM-2B-dpo-fp32)|[dpo-fp32](https://wisemodel.cn/models/OpenBMB/miniCPM-dpo-fp32)
## 模型使用 Usage
* 安装`transformers>=4.36.0`以及`accelerate`后,运行以下代码
* 注意:需要在`from_pretrained`中明确指明模型的数据类型,否则会引起较大计算误差
* Run the following code after install `transformers>=4.36.0` and `accelerate`
* Warning: It is necessary to specify the data type of the model clearly in 'from_pretrained', otherwise large calculation errors will be caused
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
torch.manual_seed(0)
path = 'openbmb/MiniCPM-2B-sft-fp32'
tokenizer = AutoTokenizer.from_pretrained(path)
model = AutoModelForCausalLM.from_pretrained(path, torch_dtype=torch.float32, device_map='cuda', trust_remote_code=True)
responds, history = model.chat(tokenizer, "山东省最高的山是哪座山, 它比黄山高还是矮?差距多少?", temperature=0.8, top_p=0.8)
print(responds)
```
* 期望输出 Expected Output
```shell
山东省最高的山是泰山,海拔1545米。
相对于黄山(海拔1864米),泰山海拔较低,相差约319米。
```
## 开源协议 LICENSE
#### 模型协议 Model LICENSE
* 本仓库中代码依照 [Apache-2.0](https://github.com/OpenBMB/MiniCPM/blob/main/LICENSE) 协议开源
* MiniCPM 模型权重的使用则需要遵循 [“通用模型许可协议-来源说明-宣传限制-商业授权”](https://github.com/OpenBMB/General-Model-License/blob/main/%E9%80%9A%E7%94%A8%E6%A8%A1%E5%9E%8B%E8%AE%B8%E5%8F%AF%E5%8D%8F%E8%AE%AE-%E6%9D%A5%E6%BA%90%E8%AF%B4%E6%98%8E-%E5%AE%A3%E4%BC%A0%E9%99%90%E5%88%B6-%E5%95%86%E4%B8%9A%E6%8E%88%E6%9D%83.md)。
* MiniCPM 模型权重对学术研究完全开放。
* 如需将模型用于商业用途,请联系[email protected]来获取书面授权,在登记后亦允许免费商业使用。
* This repository is released under the [Apache-2.0](https://github.com/OpenBMB/MiniCPM/blob/main/LICENSE) License.
* The usage of MiniCPM model weights must strictly follow [the General Model License (GML)](https://github.com/OpenBMB/General-Model-License/blob/main/%E9%80%9A%E7%94%A8%E6%A8%A1%E5%9E%8B%E8%AE%B8%E5%8F%AF%E5%8D%8F%E8%AE%AE-%E6%9D%A5%E6%BA%90%E8%AF%B4%E6%98%8E-%E5%AE%A3%E4%BC%A0%E9%99%90%E5%88%B6-%E5%95%86%E4%B8%9A%E6%8E%88%E6%9D%83.md).
* The models and weights of MiniCPM are completely free for academic research.
* If you intend to utilize the model for commercial purposes, please reach out to [email protected] to obtain the certificate of authorization.
#### 声明 Statement
* 作为一个语言模型,MiniCPM 通过学习大量的文本来生成内容,但它无法理解、表达个人观点或价值判断,它所输出的任何内容都不代表模型开发者的观点和立场。
* 因此用户在使用 MiniCPM 生成的内容时,应自行负责对其进行评估和验证。
* 如果由于使用 MinCPM 开源模型而导致的任何问题,包括但不限于数据安全问题、公共舆论风险,或模型被误导、滥用、传播或不当利用所带来的任何风险和问题,我们将不承担任何责任。
* As a language model, MiniCPM generates content by learning from a vast amount of text.
* However, it does not possess the ability to comprehend or express personal opinions or value judgments.
* Any content generated by MiniCPM does not represent the viewpoints or positions of the model developers.
* Therefore, when using content generated by MiniCPM, users should take full responsibility for evaluating and verifying it on their own.
<p id="8"></p>
## 工作引用 Citation
* 如果觉得MiniCPM有助于您的工作,请考虑引用下列[技术报告](https://shengdinghu.notion.site/MiniCPM-c805a17c5c8046398914e47f0542095a?pvs=4)
* Please cite our [techinical report](https://shengdinghu.notion.site/MiniCPM-Unveiling-the-Potential-of-End-side-Large-Language-Models-d4d3a8c426424654a4e80e42a711cb20?pvs=4) if you find our work valuable.
```
@inproceedings{minicpm2024,
title={MiniCPM:Unveiling the Potential of End-side Large Language Models},
booktitle={OpenBMB Blog},
year={2024}
}
```
| {"language": ["en", "zh"], "tags": ["MiniCPM", "ModelBest", "THUNLP"]} | text-generation | openbmb/MiniCPM-2B-sft-fp32-llama-format | [
"transformers",
"pytorch",
"text-generation",
"MiniCPM",
"ModelBest",
"THUNLP",
"conversational",
"en",
"zh",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] | 2024-02-07T06:39:35+00:00 | [] | [
"en",
"zh"
] | TAGS
#transformers #pytorch #text-generation #MiniCPM #ModelBest #THUNLP #conversational #en #zh #autotrain_compatible #endpoints_compatible #has_space #region-us
|
MiniCPM
=========
[MiniCPM 技术报告](URL target=) [Technical Report](URL target=) |
[OmniLMM 多模态模型 Multi-modal Model](URL target=) |
[CPM-C 千亿模型试用 ~100B Model Trial](URL target=)
MiniCPM 是面壁与清华大学自然语言处理实验室共同开源的系列端侧语言大模型,主体语言模型 MiniCPM-2B 仅有 24亿(2.4B)的非词嵌入参数量。
* 经过 SFT 后,MiniCPM 在公开综合性评测集上,MiniCPM 与 Mistral-7B相近(中文、数学、代码能力更优),整体性能超越 Llama2-13B、MPT-30B、Falcon-40B 等模型。
* 经过 DPO 后,MiniCPM 在当前最接近用户体感的评测集 MTBench上,MiniCPM-2B 也超越了 Llama2-70B-Chat、Vicuna-33B、Mistral-7B-Instruct-v0.1、Zephyr-7B-alpha 等众多代表性开源大模型。
* 以 MiniCPM-2B 为基础构建端侧多模态大模型 MiniCPM-V,整体性能在同规模模型中实现最佳,超越基于 Phi-2 构建的现有多模态大模型,在部分评测集上达到与 9.6B Qwen-VL-Chat 相当甚至更好的性能。
* 经过 Int4 量化后,MiniCPM 可在手机上进行部署推理,流式输出速度略高于人类说话速度。MiniCPM-V 也首次跑通了多模态大模型在手机上的部署。
* 一张1080/2080可高效参数微调,一张3090/4090可全参数微调,一台机器可持续训练 MiniCPM,二次开发成本较低。
我们将完全开源MiniCPM-2B的模型参数供学术研究和有限商用,以及训练过程中的所有Checkpoint和大部分非专有数据供模型机理研究。
* 基于MiniCPM-2B的指令微调与人类偏好对MiniCPM-2B-SFT/DPO。
* 基于MiniCPM-2B的多模态模型MiniCPM-V,能力超越基于Phi-2的同参数级别多模态模型。
* MiniCPM-2B-SFT/DPO的Int4量化版MiniCPM-2B-SFT/DPO-Int4。
* 基于MLC-LLM、LLMFarm开发的MiniCPM手机端程序,文本及多模态模型均可在手机端进行推理。
MiniCPM is an End-Size LLM developed by ModelBest Inc. and TsinghuaNLP, with only 2.4B parameters excluding embeddings.
* MiniCPM has very close performance compared with Mistral-7B on open-sourced general benchmarks with better ability on Chinese, Mathmetics and Coding after SFT. The overall performance exceeds Llama2-13B, MPT-30B, Falcon-40B, etc.
* After DPO, MiniCPM outperforms Llama2-70B-Chat, Vicuna-33B, Mistral-7B-Instruct-v0.1, Zephyr-7B-alpha, etc. on MTBench.
* MiniCPM-V, based on MiniCPM-2B, achieves the best overall performance among multimodel models of the same scale, surpassing existing multimodal large models built on Phi-2 and achieving performance comparable to or even better than 9.6B Qwen-VL-Chat on some tasks.
* MiniCPM can be deployed and infer on smartphones, and the speed of streaming output is relatively higher than the verbal speed of human. MiniCPM-V is the first multi-modal models that can be deployed on smartphones.
* The cost of developing based on MiniCPM is low. Parameter efficient finetuning can be conducted with a single 1080/2080 GPU and full parameter finetuning can be conducted with a 3090/4090 GPU.
We release all model parameters for research and limited commercial use. We also release all the checkpoint during training and most public training data for research on model mechanism.
* SFT and DPO version based on MiniCPM-2B and human preference: MiniCPM-2B-SFT/DPO
* The multi-modal model MiniCPM-V based on MiniCPM-2B, which outperforms models with similar size, i.e., Phi-2
* The INT4 quantized version MiniCPM-2B-SFT/DPO-Int4 based on MiniCPM-2B-SFT/DPO
* Mobile phone application based on MLC-LLM and LLMFarm. Both language model and multimodel model can conduct inference on smartphones.
### 评测结果 Evaluation Results
详细的评测结果位于github仓库
Detailed evaluation results are in github repo
注意:我们发现使用Huggingface生成质量略差于vLLM,因此推荐使用vLLM进行测试。我们正在排查原因。
Notice: We discovered that the quality of Huggingface generation is slightly lower than vLLM, thus benchmarking using vLLM is recommended.
We are investigating the cause now.
### 局限性 Limitations
* 受限于模型规模,模型可能出现幻觉性问题。其中由于DPO模型生成的回复内容更长,更容易出现幻觉。我们也将持续进行MiniCPM模型的迭代改进;
* 为了保证在学术研究用途上模型的通用性,我们未对模型进行任何身份认同训练。同时由于我们用ShareGPT开源语料作为部分训练数据,模型可能会输出类似GPT系列模型的身份认同信息;
* 受限于模型规模,模型的输出受到提示词(prompt)的影响较大,可能多次尝试产生不一致的结果;
* 受限于模型容量,模型的知识记忆较不准确,后续我们将结合RAG方法来增强模型的知识记忆能力。
* Due to limitations in model size, the model may experience hallucinatory issues. As DPO model tend to generate longer response, hallucinations are more likely to occur. We will also continue to iterate and improve the MiniCPM model.
* To ensure the universality of the model for academic research purposes, we did not conduct any identity training on the model. Meanwhile, as we use ShareGPT open-source corpus as part of the training data, the model may output identity information similar to the GPT series models.
* Due to the limitation of model size, the output of the model is greatly influenced by prompt words, which may result in inconsistent results from multiple attempts.
* Due to limited model capacity, the model's knowledge memory is not accurate. In the future, we will combine the RAG method to enhance the model's knowledge memory ability.
模型下载 Download
-------------
HuggingFace: sft-bf16, ModelScope: sft-bf16, WiseModel: sft-bf16
HuggingFace: sft-fp32, ModelScope: sft-fp32, WiseModel: sft-fp32
HuggingFace: dpo-bf16, ModelScope: dpo-bf16, WiseModel: dpo-bf16
HuggingFace: dpo-fp16, ModelScope: dpo-fp16, WiseModel: dpo-fp16
HuggingFace: dpo-fp32, ModelScope: dpo-fp32, WiseModel: dpo-fp32
模型使用 Usage
----------
* 安装'transformers>=4.36.0'以及'accelerate'后,运行以下代码
* 注意:需要在'from\_pretrained'中明确指明模型的数据类型,否则会引起较大计算误差
* Run the following code after install 'transformers>=4.36.0' and 'accelerate'
* Warning: It is necessary to specify the data type of the model clearly in 'from\_pretrained', otherwise large calculation errors will be caused
* 期望输出 Expected Output
开源协议 LICENSE
------------
#### 模型协议 Model LICENSE
* 本仓库中代码依照 Apache-2.0 协议开源
* MiniCPM 模型权重的使用则需要遵循 “通用模型许可协议-来源说明-宣传限制-商业授权”。
* MiniCPM 模型权重对学术研究完全开放。
* 如需将模型用于商业用途,请联系[email protected]来获取书面授权,在登记后亦允许免费商业使用。
* This repository is released under the Apache-2.0 License.
* The usage of MiniCPM model weights must strictly follow the General Model License (GML).
* The models and weights of MiniCPM are completely free for academic research.
* If you intend to utilize the model for commercial purposes, please reach out to cpm@URL to obtain the certificate of authorization.
#### 声明 Statement
* 作为一个语言模型,MiniCPM 通过学习大量的文本来生成内容,但它无法理解、表达个人观点或价值判断,它所输出的任何内容都不代表模型开发者的观点和立场。
* 因此用户在使用 MiniCPM 生成的内容时,应自行负责对其进行评估和验证。
* 如果由于使用 MinCPM 开源模型而导致的任何问题,包括但不限于数据安全问题、公共舆论风险,或模型被误导、滥用、传播或不当利用所带来的任何风险和问题,我们将不承担任何责任。
* As a language model, MiniCPM generates content by learning from a vast amount of text.
* However, it does not possess the ability to comprehend or express personal opinions or value judgments.
* Any content generated by MiniCPM does not represent the viewpoints or positions of the model developers.
* Therefore, when using content generated by MiniCPM, users should take full responsibility for evaluating and verifying it on their own.
工作引用 Citation
-------------
* 如果觉得MiniCPM有助于您的工作,请考虑引用下列技术报告
* Please cite our techinical report if you find our work valuable.
| [
"### 评测结果 Evaluation Results\n\n\n详细的评测结果位于github仓库\n\n\nDetailed evaluation results are in github repo\n\n\n注意:我们发现使用Huggingface生成质量略差于vLLM,因此推荐使用vLLM进行测试。我们正在排查原因。\n\n\nNotice: We discovered that the quality of Huggingface generation is slightly lower than vLLM, thus benchmarking using vLLM is recommended.\nWe are investigating the cause now.",
"### 局限性 Limitations\n\n\n* 受限于模型规模,模型可能出现幻觉性问题。其中由于DPO模型生成的回复内容更长,更容易出现幻觉。我们也将持续进行MiniCPM模型的迭代改进;\n* 为了保证在学术研究用途上模型的通用性,我们未对模型进行任何身份认同训练。同时由于我们用ShareGPT开源语料作为部分训练数据,模型可能会输出类似GPT系列模型的身份认同信息;\n* 受限于模型规模,模型的输出受到提示词(prompt)的影响较大,可能多次尝试产生不一致的结果;\n* 受限于模型容量,模型的知识记忆较不准确,后续我们将结合RAG方法来增强模型的知识记忆能力。\n* Due to limitations in model size, the model may experience hallucinatory issues. As DPO model tend to generate longer response, hallucinations are more likely to occur. We will also continue to iterate and improve the MiniCPM model.\n* To ensure the universality of the model for academic research purposes, we did not conduct any identity training on the model. Meanwhile, as we use ShareGPT open-source corpus as part of the training data, the model may output identity information similar to the GPT series models.\n* Due to the limitation of model size, the output of the model is greatly influenced by prompt words, which may result in inconsistent results from multiple attempts.\n* Due to limited model capacity, the model's knowledge memory is not accurate. In the future, we will combine the RAG method to enhance the model's knowledge memory ability.\n\n\n模型下载 Download\n-------------\n\n\nHuggingFace: sft-bf16, ModelScope: sft-bf16, WiseModel: sft-bf16\nHuggingFace: sft-fp32, ModelScope: sft-fp32, WiseModel: sft-fp32\nHuggingFace: dpo-bf16, ModelScope: dpo-bf16, WiseModel: dpo-bf16\nHuggingFace: dpo-fp16, ModelScope: dpo-fp16, WiseModel: dpo-fp16\nHuggingFace: dpo-fp32, ModelScope: dpo-fp32, WiseModel: dpo-fp32\n\n\n模型使用 Usage\n----------\n\n\n* 安装'transformers>=4.36.0'以及'accelerate'后,运行以下代码\n* 注意:需要在'from\\_pretrained'中明确指明模型的数据类型,否则会引起较大计算误差\n* Run the following code after install 'transformers>=4.36.0' and 'accelerate'\n* Warning: It is necessary to specify the data type of the model clearly in 'from\\_pretrained', otherwise large calculation errors will be caused\n* 期望输出 Expected Output\n\n\n开源协议 LICENSE\n------------",
"#### 模型协议 Model LICENSE\n\n\n* 本仓库中代码依照 Apache-2.0 协议开源\n* MiniCPM 模型权重的使用则需要遵循 “通用模型许可协议-来源说明-宣传限制-商业授权”。\n* MiniCPM 模型权重对学术研究完全开放。\n* 如需将模型用于商业用途,请联系[email protected]来获取书面授权,在登记后亦允许免费商业使用。\n* This repository is released under the Apache-2.0 License.\n* The usage of MiniCPM model weights must strictly follow the General Model License (GML).\n* The models and weights of MiniCPM are completely free for academic research.\n* If you intend to utilize the model for commercial purposes, please reach out to cpm@URL to obtain the certificate of authorization.",
"#### 声明 Statement\n\n\n* 作为一个语言模型,MiniCPM 通过学习大量的文本来生成内容,但它无法理解、表达个人观点或价值判断,它所输出的任何内容都不代表模型开发者的观点和立场。\n* 因此用户在使用 MiniCPM 生成的内容时,应自行负责对其进行评估和验证。\n* 如果由于使用 MinCPM 开源模型而导致的任何问题,包括但不限于数据安全问题、公共舆论风险,或模型被误导、滥用、传播或不当利用所带来的任何风险和问题,我们将不承担任何责任。\n* As a language model, MiniCPM generates content by learning from a vast amount of text.\n* However, it does not possess the ability to comprehend or express personal opinions or value judgments.\n* Any content generated by MiniCPM does not represent the viewpoints or positions of the model developers.\n* Therefore, when using content generated by MiniCPM, users should take full responsibility for evaluating and verifying it on their own.\n\n\n\n工作引用 Citation\n-------------\n\n\n* 如果觉得MiniCPM有助于您的工作,请考虑引用下列技术报告\n* Please cite our techinical report if you find our work valuable."
] | [
"TAGS\n#transformers #pytorch #text-generation #MiniCPM #ModelBest #THUNLP #conversational #en #zh #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"### 评测结果 Evaluation Results\n\n\n详细的评测结果位于github仓库\n\n\nDetailed evaluation results are in github repo\n\n\n注意:我们发现使用Huggingface生成质量略差于vLLM,因此推荐使用vLLM进行测试。我们正在排查原因。\n\n\nNotice: We discovered that the quality of Huggingface generation is slightly lower than vLLM, thus benchmarking using vLLM is recommended.\nWe are investigating the cause now.",
"### 局限性 Limitations\n\n\n* 受限于模型规模,模型可能出现幻觉性问题。其中由于DPO模型生成的回复内容更长,更容易出现幻觉。我们也将持续进行MiniCPM模型的迭代改进;\n* 为了保证在学术研究用途上模型的通用性,我们未对模型进行任何身份认同训练。同时由于我们用ShareGPT开源语料作为部分训练数据,模型可能会输出类似GPT系列模型的身份认同信息;\n* 受限于模型规模,模型的输出受到提示词(prompt)的影响较大,可能多次尝试产生不一致的结果;\n* 受限于模型容量,模型的知识记忆较不准确,后续我们将结合RAG方法来增强模型的知识记忆能力。\n* Due to limitations in model size, the model may experience hallucinatory issues. As DPO model tend to generate longer response, hallucinations are more likely to occur. We will also continue to iterate and improve the MiniCPM model.\n* To ensure the universality of the model for academic research purposes, we did not conduct any identity training on the model. Meanwhile, as we use ShareGPT open-source corpus as part of the training data, the model may output identity information similar to the GPT series models.\n* Due to the limitation of model size, the output of the model is greatly influenced by prompt words, which may result in inconsistent results from multiple attempts.\n* Due to limited model capacity, the model's knowledge memory is not accurate. In the future, we will combine the RAG method to enhance the model's knowledge memory ability.\n\n\n模型下载 Download\n-------------\n\n\nHuggingFace: sft-bf16, ModelScope: sft-bf16, WiseModel: sft-bf16\nHuggingFace: sft-fp32, ModelScope: sft-fp32, WiseModel: sft-fp32\nHuggingFace: dpo-bf16, ModelScope: dpo-bf16, WiseModel: dpo-bf16\nHuggingFace: dpo-fp16, ModelScope: dpo-fp16, WiseModel: dpo-fp16\nHuggingFace: dpo-fp32, ModelScope: dpo-fp32, WiseModel: dpo-fp32\n\n\n模型使用 Usage\n----------\n\n\n* 安装'transformers>=4.36.0'以及'accelerate'后,运行以下代码\n* 注意:需要在'from\\_pretrained'中明确指明模型的数据类型,否则会引起较大计算误差\n* Run the following code after install 'transformers>=4.36.0' and 'accelerate'\n* Warning: It is necessary to specify the data type of the model clearly in 'from\\_pretrained', otherwise large calculation errors will be caused\n* 期望输出 Expected Output\n\n\n开源协议 LICENSE\n------------",
"#### 模型协议 Model LICENSE\n\n\n* 本仓库中代码依照 Apache-2.0 协议开源\n* MiniCPM 模型权重的使用则需要遵循 “通用模型许可协议-来源说明-宣传限制-商业授权”。\n* MiniCPM 模型权重对学术研究完全开放。\n* 如需将模型用于商业用途,请联系[email protected]来获取书面授权,在登记后亦允许免费商业使用。\n* This repository is released under the Apache-2.0 License.\n* The usage of MiniCPM model weights must strictly follow the General Model License (GML).\n* The models and weights of MiniCPM are completely free for academic research.\n* If you intend to utilize the model for commercial purposes, please reach out to cpm@URL to obtain the certificate of authorization.",
"#### 声明 Statement\n\n\n* 作为一个语言模型,MiniCPM 通过学习大量的文本来生成内容,但它无法理解、表达个人观点或价值判断,它所输出的任何内容都不代表模型开发者的观点和立场。\n* 因此用户在使用 MiniCPM 生成的内容时,应自行负责对其进行评估和验证。\n* 如果由于使用 MinCPM 开源模型而导致的任何问题,包括但不限于数据安全问题、公共舆论风险,或模型被误导、滥用、传播或不当利用所带来的任何风险和问题,我们将不承担任何责任。\n* As a language model, MiniCPM generates content by learning from a vast amount of text.\n* However, it does not possess the ability to comprehend or express personal opinions or value judgments.\n* Any content generated by MiniCPM does not represent the viewpoints or positions of the model developers.\n* Therefore, when using content generated by MiniCPM, users should take full responsibility for evaluating and verifying it on their own.\n\n\n\n工作引用 Citation\n-------------\n\n\n* 如果觉得MiniCPM有助于您的工作,请考虑引用下列技术报告\n* Please cite our techinical report if you find our work valuable."
] | [
57,
101,
655,
180,
262
] | [
"passage: TAGS\n#transformers #pytorch #text-generation #MiniCPM #ModelBest #THUNLP #conversational #en #zh #autotrain_compatible #endpoints_compatible #has_space #region-us \n### 评测结果 Evaluation Results\n\n\n详细的评测结果位于github仓库\n\n\nDetailed evaluation results are in github repo\n\n\n注意:我们发现使用Huggingface生成质量略差于vLLM,因此推荐使用vLLM进行测试。我们正在排查原因。\n\n\nNotice: We discovered that the quality of Huggingface generation is slightly lower than vLLM, thus benchmarking using vLLM is recommended.\nWe are investigating the cause now."
] | [
-0.009963848628103733,
-0.15188519656658173,
-0.004663964733481407,
0.040838971734046936,
0.11296148598194122,
-0.004151855129748583,
0.04998120665550232,
0.12910699844360352,
-0.02972506918013096,
0.01535273902118206,
0.14406965672969818,
0.06965360045433044,
0.013901117257773876,
0.032000426203012466,
-0.014046795666217804,
-0.15578968822956085,
0.06553138792514801,
0.05551905557513237,
-0.020676396787166595,
0.09439079463481903,
0.16194738447666168,
-0.0812826007604599,
0.13612544536590576,
0.04314984753727913,
-0.17512686550617218,
0.009856128133833408,
0.04529460519552231,
0.00197188020683825,
0.08306184411048889,
0.05754504352807999,
-0.02234281785786152,
0.05094146728515625,
-0.024090640246868134,
-0.07107546180486679,
0.050005681812763214,
0.06417179852724075,
0.02152538299560547,
0.03485354781150818,
0.0024379687383770943,
0.019413335248827934,
0.1946471780538559,
-0.02850743941962719,
-0.030730167403817177,
0.055327072739601135,
-0.03788575530052185,
-0.05640626326203346,
-0.06795673817396164,
-0.0019875403959304094,
-0.01639057882130146,
0.03410511091351509,
0.012074085883796215,
0.29111406207084656,
-0.14140450954437256,
0.10867827385663986,
0.2356586456298828,
-0.21030721068382263,
-0.06817213445901871,
0.2313477247953415,
0.09091856330633163,
-0.05366353690624237,
-0.047226689755916595,
0.10838896036148071,
0.08722859621047974,
0.013357307761907578,
0.045655906200408936,
-0.00307396799325943,
-0.04448673501610756,
0.00822632946074009,
-0.05770618095993996,
-0.05444369465112686,
0.21363765001296997,
-0.003263876074925065,
0.018957985565066338,
-0.11875811964273453,
-0.08619099855422974,
-0.007917584851384163,
0.02068524807691574,
-0.03244062513113022,
-0.035888344049453735,
0.013190198689699173,
0.00574712548404932,
-0.05883263424038887,
-0.05092235282063484,
-0.04409996047616005,
-0.05238013342022896,
0.06317425519227982,
0.02778896503150463,
0.05969446152448654,
-0.1079283356666565,
0.0982954204082489,
0.05307566002011299,
-0.10626190155744553,
-0.029404178261756897,
-0.14064525067806244,
0.08544359356164932,
0.0569629967212677,
-0.01715317741036415,
-0.0348917618393898,
0.12155257165431976,
0.05503398925065994,
0.05151749774813652,
-0.002149523701518774,
0.013492915779352188,
0.004082176834344864,
0.030423017218708992,
0.16270262002944946,
-0.0807204470038414,
-0.04579241946339607,
0.11708942800760269,
0.01342683658003807,
0.04790250584483147,
-0.035961806774139404,
0.0013567034620791674,
-0.10785328596830368,
0.12480096518993378,
0.07168927043676376,
0.024156885221600533,
0.12747986614704132,
-0.048523686826229095,
-0.07456890493631363,
-0.09282255172729492,
-0.07381856441497803,
-0.0008883755072019994,
-0.0014976966194808483,
-0.025494802743196487,
0.12369829416275024,
0.053581807762384415,
-0.004466215148568153,
-0.1021031066775322,
-0.013450969941914082,
-0.057195764034986496,
-0.05587713047862053,
-0.09686906635761261,
-0.09749457985162735,
0.034664906561374664,
-0.13242965936660767,
0.01683301478624344,
-0.09892962872982025,
-0.10278419405221939,
-0.056799158453941345,
-0.0036550010554492474,
0.002613748423755169,
-0.06580796092748642,
0.03766433522105217,
-0.06109391897916794,
0.004702324979007244,
-0.024476440623402596,
0.12565796077251434,
-0.0442521907389164,
0.05470563843846321,
-0.0067864577285945415,
0.05335462838411331,
-0.09361041337251663,
0.0030008116737008095,
-0.04243304580450058,
0.03364124894142151,
0.061467304825782776,
-0.007940082810819149,
-0.08252036571502686,
0.10061347484588623,
-0.04629720747470856,
-0.07429531216621399,
-0.03817727044224739,
0.012110764160752296,
0.0294946301728487,
0.13398879766464233,
-0.11478215456008911,
-0.041995029896497726,
0.0647565945982933,
-0.08828625082969666,
-0.10074173659086227,
0.16037124395370483,
-0.019014397636055946,
0.040573880076408386,
0.11098846048116684,
0.06521350145339966,
0.08031408488750458,
-0.1638524830341339,
-0.024770386517047882,
0.0349745936691761,
-0.11249296367168427,
0.06079786270856857,
0.016119958832859993,
0.11210250854492188,
-0.12463803589344025,
0.10161398351192474,
-0.0012497302377596498,
-0.00155666284263134,
-0.07732775807380676,
-0.03788265958428383,
-0.13374386727809906,
-0.008401111699640751,
-0.037229254841804504,
0.03632870316505432,
0.019859254360198975,
-0.05909473076462746,
-0.04766906052827835,
-0.06971175223588943,
0.0598820336163044,
-0.05366324260830879,
0.016148900613188744,
-0.11054450273513794,
0.13435281813144684,
0.020589597523212433,
0.052526917308568954,
-0.061070043593645096,
-0.02086283452808857,
-0.004804848227649927,
0.10296961665153503,
-0.00007636150985490531,
0.16572001576423645,
0.020117109641432762,
-0.060652825981378555,
0.00038681417936459184,
0.029135212302207947,
0.015481819398701191,
-0.039205558598041534,
-0.016027502715587616,
-0.14557918906211853,
0.006099294871091843,
-0.05286874994635582,
-0.0875849798321724,
-0.022143129259347916,
-0.002818894339725375,
0.07531651109457016,
0.17316125333309174,
0.02029617689549923,
0.028132541105151176,
0.03852522000670433,
-0.02643551677465439,
-0.04522136598825455,
-0.05078498274087906,
0.0579528734087944,
-0.02263861708343029,
-0.03458220139145851,
0.14266641438007355,
-0.03786368668079376,
0.20581167936325073,
0.14854344725608826,
-0.07468423992395401,
0.07914749532938004,
0.019191717728972435,
-0.06017221137881279,
0.04059751331806183,
-0.03440424054861069,
-0.07066831737756729,
0.11751186847686768,
-0.008665047585964203,
0.08717501163482666,
-0.09822843968868256,
-0.030138427391648293,
-0.000777266570366919,
-0.047563131898641586,
-0.035248298197984695,
0.09987510740756989,
0.05507684499025345,
0.027442775666713715,
0.03828086331486702,
0.05437777191400528,
0.018055569380521774,
0.2217102348804474,
-0.03222815692424774,
0.04234189912676811,
0.02246813103556633,
-0.04596706107258797,
-0.05174214020371437,
0.013520782813429832,
-0.1866617500782013,
-0.0419137142598629,
0.07376169413328171,
-0.020266063511371613,
0.08636023104190826,
-0.10743571817874908,
0.0011687465012073517,
0.005902913399040699,
-0.01272085402160883,
0.04415442794561386,
0.17054606974124908,
0.03874725103378296,
0.11173390597105026,
-0.05910239368677139,
0.06096939370036125,
-0.061026666313409805,
-0.023633714765310287,
-0.06612072139978409,
0.09836199134588242,
-0.06145980581641197,
-0.2340949922800064,
-0.08506972342729568,
0.0159602090716362,
-0.1627628356218338,
-0.025087624788284302,
0.002109879394993186,
-0.1479577124118805,
-0.07740072160959244,
-0.0471918061375618,
0.21005606651306152,
-0.004354078322649002,
0.052399493753910065,
0.12325792014598846,
0.047970615327358246,
0.009647834114730358,
-0.1885993927717209,
-0.06384705007076263,
-0.028589248657226562,
-0.0758388489484787,
0.08960290253162384,
-0.04645524173974991,
0.07576289027929306,
0.05820562317967415,
-0.009583188220858574,
0.047881364822387695,
-0.019226595759391785,
0.31471145153045654,
-0.035553641617298126,
-0.00533333420753479,
0.3259637951850891,
0.057541560381650925,
-0.02296227216720581,
0.09526355564594269,
-0.019426152110099792,
-0.11046218872070312,
0.006714698858559132,
0.0137480478733778,
-0.08542820811271667,
-0.12188014388084412,
-0.10432010143995285,
-0.10006646811962128,
0.018252413719892502,
-0.007237359415739775,
0.03472958505153656,
-0.05165952071547508,
0.022045481950044632,
-0.06081608682870865,
0.0729842558503151,
0.007330285385251045,
0.05142058804631233,
0.1460603028535843,
-0.06893742829561234,
0.1317494660615921,
-0.03128102794289589,
-0.06270170956850052,
0.1074119508266449,
0.009400169365108013,
0.15202929079532623,
-0.09648088365793228,
-0.05799642205238342,
0.12609103322029114,
0.07756156474351883,
0.05982695519924164,
0.048676904290914536,
-0.10172232985496521,
-0.02687758021056652,
-0.025175301358103752,
-0.03538700193166733,
-0.1050906702876091,
-0.01714756339788437,
-0.0054691811092197895,
-0.02729473076760769,
-0.07963050156831741,
0.11453894525766373,
0.09631408005952835,
0.1296786069869995,
0.022043028846383095,
-0.1542501151561737,
-0.033235929906368256,
0.0383354015648365,
-0.021862171590328217,
-0.0852261483669281,
0.011399954557418823,
0.02884431555867195,
-0.1028372198343277,
0.059421416372060776,
-0.02977173402905464,
0.12298005819320679,
-0.038819245994091034,
0.09354287385940552,
-0.13045786321163177,
0.06127301976084709,
0.021424034610390663,
0.08506978303194046,
-0.23688904941082,
0.13991554081439972,
0.01032151747494936,
0.06803062558174133,
-0.09999697655439377,
-0.01756250485777855,
0.05049979314208031,
0.12621520459651947,
0.03751498833298683,
-0.0017581306165084243,
0.04431116580963135,
0.01654663495719433,
-0.05317969247698784,
0.09479007124900818,
0.019437085837125778,
0.07128685712814331,
0.04661296308040619,
-0.1099143698811531,
0.06197747588157654,
0.017940375953912735,
0.04742567986249924,
-0.15374180674552917,
-0.03743486478924751,
0.023754894733428955,
0.04279281571507454,
-0.011931467801332474,
-0.06670263409614563,
-0.04274728149175644,
-0.031122906133532524,
0.20218594372272491,
0.013009817339479923,
-0.05549423769116402,
-0.09901708364486694,
0.11128373444080353,
0.06029971316456795,
-0.03543815016746521,
-0.04956306517124176,
-0.08541567623615265,
0.11030426621437073,
-0.020894039422273636,
-0.1762648969888687,
-0.04211617261171341,
-0.0951574519276619,
-0.13440901041030884,
0.03815354406833649,
0.14183400571346283,
-0.1303585171699524,
0.030019192025065422,
0.07034850865602493,
-0.02505308948457241,
-0.10968015342950821,
-0.13550199568271637,
0.009742889553308487,
0.18547825515270233,
-0.0842389464378357,
0.011073392815887928,
-0.009521412663161755,
-0.046458449214696884,
0.02931792289018631,
-0.02406197413802147,
0.11245216429233551,
0.21630693972110748,
-0.04430506378412247,
0.13740608096122742,
-0.034626781940460205,
-0.027389081194996834,
-0.30815041065216064,
-0.05139101296663284,
0.02338508330285549,
0.0743558332324028,
-0.0926346629858017,
0.023564178496599197,
0.0676569789648056,
-0.0928911343216896,
0.01399600226432085,
-0.03146687150001526,
-0.24722492694854736,
-0.13510751724243164,
0.05647080019116402,
0.019291043281555176,
0.38187697529792786,
-0.005800116807222366,
-0.0028275251388549805,
-0.08675505965948105,
-0.23328951001167297,
0.08259671181440353,
-0.08128649741411209,
0.0952889546751976,
-0.08773288875818253,
0.03114871121942997,
0.06689145416021347,
-0.06196841970086098,
0.14705424010753632,
-0.1323949247598648,
0.08061888813972473,
-0.07798776775598526,
-0.05378076806664467,
0.030277950689196587,
-0.04154256358742714,
0.12209931015968323,
-0.06934593617916107,
0.08331046253442764,
-0.12698014080524445,
-0.04553611949086189,
-0.12612171471118927,
0.1042281985282898,
-0.007122271694242954,
-0.05714620277285576,
-0.13858848810195923,
0.01962006464600563,
0.033151932060718536,
0.005693190265446901,
0.08062709122896194,
-0.03483660891652107,
0.026492951437830925,
0.15773668885231018,
0.11806971579790115,
-0.2780240476131439,
0.02388387732207775,
0.04517876356840134,
0.0375538170337677,
0.11479126662015915,
-0.12799961864948273,
0.06431050598621368,
0.1083626002073288,
-0.020493876188993454,
0.015316410921514034,
0.044640351086854935,
-0.0243820920586586,
-0.0028237595688551664,
0.12211138755083084,
-0.19227340817451477,
-0.033890023827552795,
-0.047590143978595734,
-0.06605178117752075,
0.03596505895256996,
0.011169283650815487,
0.21548660099506378,
-0.03570568561553955,
0.0028424798510968685,
0.01846596598625183,
0.021426759660243988,
-0.026668604463338852,
0.12632979452610016,
0.035368405282497406,
0.02588573284447193,
-0.1446084827184677,
0.04103390499949455,
-0.059162721037864685,
0.044931408017873764,
0.024975311011075974,
-0.036519814282655716,
-0.1094297468662262,
-0.03689923882484436,
0.008461568504571915,
0.09362412244081497,
-0.02218565344810486,
-0.02586210146546364,
-0.13302430510520935,
-0.14352406561374664,
0.020770419389009476,
-0.010940998792648315,
0.08704866468906403,
0.09805523604154587,
-0.06424272060394287,
-0.03789179399609566,
-0.0443851575255394,
0.03419957682490349,
0.08584510535001755,
-0.007380188442766666,
-0.18932630121707916,
0.13112588226795197,
0.007332389708608389,
0.11594409495592117,
-0.10122021287679672,
0.001954688224941492,
-0.12744253873825073,
-0.012530899606645107,
-0.16673175990581512,
-0.05932587757706642,
0.03281260281801224,
-0.023160288110375404,
0.03921827673912048,
-0.05158775672316551,
-0.12841521203517914,
0.06950194388628006,
-0.0728835016489029,
0.005891803652048111,
-0.04104281961917877,
0.03944343701004982,
-0.025385895743966103,
0.0007178882951848209,
0.11447694897651672,
-0.049809325486421585,
0.08057568222284317,
0.004652941599488258,
-0.06283706426620483,
0.008774153888225555,
-0.16816094517707825,
-0.08980578184127808,
0.1101515144109726,
0.06381706148386002,
-0.008319281041622162,
0.02650187909603119,
0.05772865563631058,
0.046051591634750366,
0.013286971487104893,
0.04211121425032616,
0.09436129033565521,
-0.07549578696489334,
-0.023456688970327377,
-0.024249635636806488,
-0.017127785831689835,
0.015637829899787903,
0.007775090169161558,
0.02024606056511402,
0.12930575013160706,
0.11104942858219147,
-0.003252720693126321,
-0.005167996510863304,
-0.1834798902273178,
0.009027800522744656,
0.004332639742642641,
-0.13733941316604614,
-0.05198432877659798,
-0.002357492921873927,
0.08237162232398987,
-0.013951039873063564,
0.2626967132091522,
0.17083019018173218,
-0.04966462776064873,
0.009036368690431118,
0.185237854719162,
-0.04094427078962326,
-0.009677565656602383,
0.09459075331687927,
-0.0008275508880615234,
0.10812905430793762,
-0.027847466990351677,
-0.03501178324222565,
0.07443016767501831,
-0.05910339578986168,
0.16097435355186462,
0.05772104859352112,
-0.009606534615159035,
0.07273764908313751,
0.04168613627552986,
0.024309605360031128,
-0.011594523675739765,
-0.011031814850866795,
-0.05904736369848251,
0.15188881754875183,
-0.05858268216252327,
-0.09247192740440369,
0.1417888104915619,
-0.07808145880699158,
0.03109031915664673,
-0.03273569047451019,
-0.03036276437342167,
-0.1555784046649933,
-0.11205273121595383,
-0.08796898275613785,
-0.1789209544658661,
0.07044541090726852,
-0.1324620246887207,
0.01269627921283245,
-0.10016889870166779,
0.054952315986156464,
-0.07527055591344833,
0.08428740501403809,
-0.0812181904911995,
-0.11725397408008575,
0.07724128663539886,
-0.021921440958976746,
-0.06624513119459152,
-0.026451576501131058,
0.04098522663116455,
-0.04812319949269295,
0.05449048802256584,
0.00587404053658247,
0.05059178173542023,
-0.005269310902804136,
-0.06425044685602188,
-0.042424578219652176,
-0.03802098333835602,
-0.06550472974777222,
-0.023795176297426224,
-0.03759419918060303,
0.16239631175994873,
-0.007778531406074762,
-0.010661879554390907,
-0.023450244218111038,
0.22331850230693817,
0.017212536185979843,
-0.08924595266580582,
-0.11731912940740585,
0.03265455365180969,
-0.03725731372833252,
0.01228469330817461,
-0.008229444734752178,
0.022487491369247437,
-0.08646856993436813,
0.3139982521533966,
0.22295314073562622,
-0.10870880633592606,
0.005403865594416857,
-0.1040072962641716,
0.029418393969535828,
-0.01845153234899044,
0.1966353803873062,
0.05352490022778511,
0.15518364310264587,
-0.04877052083611488,
0.031942520290613174,
-0.04620177671313286,
-0.017239676788449287,
-0.08565406501293182,
0.0916094183921814,
0.060929156839847565,
-0.07130417227745056,
-0.029773397371172905,
0.013760733418166637,
-0.05542447790503502,
-0.01691482961177826,
-0.05158400163054466,
-0.1761864274740219,
-0.07160589098930359,
-0.005253713112324476,
0.03510793298482895,
-0.07753398269414902,
0.07785817235708237,
-0.06144129857420921,
0.048180509358644485,
0.05500007048249245,
-0.028225751593708992,
-0.12277141958475113,
-0.0002536320243962109,
0.12106063216924667,
0.04896998032927513,
0.06076808646321297,
-0.02277120016515255,
0.011540628969669342,
0.05890783667564392,
-0.07408971339464188,
-0.06837385892868042,
-0.0003167955728713423,
-0.013322574086487293,
-0.11799634248018265,
-0.008204223588109016,
0.06729009747505188,
0.030906997621059418,
-0.06244621053338051,
0.045187823474407196,
-0.011846547946333885,
-0.01924010179936886,
0.06389938294887543,
0.020434485748410225,
-0.04702461510896683,
0.10926884412765503,
-0.10984168946743011,
0.09093936532735825,
0.13239417970180511,
-0.040626611560583115,
-0.017254291102290154,
-0.10518454015254974,
0.09683949500322342,
-0.036645445972681046,
-0.04360179230570793,
-0.01297134067863226,
-0.14086264371871948,
-0.0026363679207861423,
0.12323109805583954,
-0.07062206417322159,
-0.054486632347106934,
-0.11865222454071045,
-0.05708571895956993,
-0.022374669089913368,
-0.03440402075648308,
0.13861094415187836,
0.035979870706796646,
0.015491848811507225,
0.04278513416647911,
-0.14391908049583435,
0.0620354600250721,
0.022109832614660263,
-0.06973956525325775,
-0.08180570602416992
] |
null | null | null | Finetune of miqu-70b-sf dequant of miqudev's leak of Mistral-70B (allegedly an early mistral medium). My diffs are available under CC-0, this is a merge with the leaked model, you can use the other repository to save bandwidth.
EQ-Bench: 84.89
Will run more benches later. | {"license": "cc-by-2.0"} | null | LoneStriker/Senku-70B-Full-GGUF | [
"gguf",
"license:cc-by-2.0",
"region:us"
] | 2024-02-07T06:42:07+00:00 | [] | [] | TAGS
#gguf #license-cc-by-2.0 #region-us
| Finetune of miqu-70b-sf dequant of miqudev's leak of Mistral-70B (allegedly an early mistral medium). My diffs are available under CC-0, this is a merge with the leaked model, you can use the other repository to save bandwidth.
EQ-Bench: 84.89
Will run more benches later. | [] | [
"TAGS\n#gguf #license-cc-by-2.0 #region-us \n"
] | [
18
] | [
"passage: TAGS\n#gguf #license-cc-by-2.0 #region-us \n"
] | [
0.0011595520190894604,
0.1008906364440918,
-0.008277126587927341,
-0.0153642687946558,
-0.03492068871855736,
0.06025968864560127,
0.1851685643196106,
0.03349992260336876,
0.1856045126914978,
-0.05334197357296944,
0.16355285048484802,
0.049112655222415924,
0.02761683613061905,
0.04947838932275772,
0.004204172641038895,
-0.10907959192991257,
0.06244875490665436,
-0.009573734365403652,
-0.028819793835282326,
0.0053512402810156345,
0.02854931354522705,
0.0011576875112950802,
-0.006027223076671362,
-0.02986939437687397,
-0.1685768961906433,
-0.03245604783296585,
0.08751492202281952,
-0.024193275719881058,
0.06312724947929382,
0.053874384611845016,
0.04110157862305641,
0.13036106526851654,
-0.027071915566921234,
-0.1818583607673645,
0.016058344393968582,
-0.07555514574050903,
-0.1735236793756485,
0.04471925273537636,
0.03504736348986626,
0.07272036373615265,
0.11797056347131729,
0.1476445347070694,
-0.1258729100227356,
0.04928118363022804,
-0.19445432722568512,
-0.1651574671268463,
-0.1255561262369156,
0.07546303421258926,
0.01415074709802866,
0.004132893402129412,
0.07642648369073868,
0.02066752314567566,
-0.1976844221353531,
-0.0481024868786335,
0.019797446206212044,
-0.36654016375541687,
0.048862483352422714,
0.2740064859390259,
-0.010004308074712753,
0.06329848617315292,
-0.0680277943611145,
0.13555891811847687,
0.04820406809449196,
-0.031981054693460464,
-0.09860078990459442,
-0.07737178355455399,
0.008312221616506577,
0.1417861431837082,
-0.017208077013492584,
-0.09299086779356003,
0.23472976684570312,
0.02672688476741314,
-0.07084863632917404,
0.10958970338106155,
0.010536650195717812,
0.039409324526786804,
0.005708347540348768,
0.06528723984956741,
0.05813257396221161,
0.19860662519931793,
0.07581155747175217,
-0.08994298428297043,
-0.16485260426998138,
-0.10385251045227051,
-0.21608586609363556,
0.0924827829003334,
-0.037548039108514786,
0.13749542832374573,
-0.10447274893522263,
0.03132317215204239,
-0.22612757980823517,
-0.005731082055717707,
-0.11174938827753067,
-0.051478419452905655,
0.11122458428144455,
0.03165540471673012,
-0.05905789136886597,
0.2158246636390686,
0.14650119841098785,
0.13200172781944275,
-0.11031588912010193,
-0.02857218310236931,
-0.06114452704787254,
0.1764029562473297,
-0.026375548914074898,
-0.01641036756336689,
0.05476361885666847,
0.2188960462808609,
0.07165747135877609,
-0.162868931889534,
0.01418210007250309,
-0.01983271725475788,
-0.1689254641532898,
-0.03025072254240513,
-0.19210316240787506,
0.12934106588363647,
-0.037179235368967056,
-0.1055191233754158,
-0.05420095846056938,
0.07867953926324844,
0.18498492240905762,
0.04822839796543121,
0.012606176547706127,
0.038387421518564224,
0.02951749414205551,
-0.10731499642133713,
-0.04545033350586891,
0.051685918122529984,
0.12180261313915253,
0.05165604129433632,
-0.13913594186306,
0.002520043635740876,
0.04969332367181778,
0.075912244617939,
0.13589423894882202,
-0.04975353553891182,
0.04527550935745239,
-0.10857710242271423,
-0.0865221619606018,
0.054415762424468994,
-0.00015224948583636433,
-0.010297919623553753,
0.02792182005941868,
0.09644226729869843,
0.03419842571020126,
-0.026064401492476463,
-0.06746451556682587,
-0.0806327760219574,
-0.08870553970336914,
0.11862199753522873,
-0.07351814955472946,
-0.013552638702094555,
-0.28421640396118164,
-0.03763074800372124,
-0.07484106719493866,
0.04705019295215607,
0.008275259286165237,
-0.04495861753821373,
-0.11308321356773376,
0.06919229030609131,
-0.02641407959163189,
0.008398644626140594,
-0.10247371345758438,
-0.006398234982043505,
-0.07564003020524979,
0.11529412120580673,
-0.055860742926597595,
-0.06377554684877396,
0.20695243775844574,
-0.14312928915023804,
-0.0856219157576561,
0.033136170357465744,
0.0460883229970932,
-0.057473182678222656,
0.033351972699165344,
0.2864840626716614,
-0.03710563853383064,
-0.10925937443971634,
0.05543743446469307,
0.20663848519325256,
-0.094730444252491,
-0.19516420364379883,
0.1627551019191742,
-0.17130941152572632,
-0.16876088082790375,
-0.004892116412520409,
-0.15383952856063843,
0.113704614341259,
-0.031198807060718536,
-0.06125156581401825,
0.0002903863787651062,
-0.027883905917406082,
-0.028274493291974068,
-0.023593788966536522,
0.07128309458494186,
-0.014494744129478931,
0.05023355782032013,
-0.07908084988594055,
0.011675846762955189,
0.09374605864286423,
0.03009912557899952,
-0.08341597765684128,
0.08898629993200302,
-0.021887850016355515,
0.011341247707605362,
0.015611502341926098,
-0.049292948096990585,
0.036329690366983414,
-0.005462239496409893,
0.11605148762464523,
0.13836906850337982,
0.03273836523294449,
0.001060246373526752,
0.0031342385336756706,
0.06941001862287521,
-0.02835630252957344,
0.00031699848477728665,
0.02964741550385952,
-0.06790052354335785,
0.09893739223480225,
0.02606949396431446,
0.062452755868434906,
-0.06050403043627739,
-0.038328949362039566,
0.2644525170326233,
-0.08032567799091339,
-0.05628533288836479,
0.0076824696734547615,
-0.00813513807952404,
-0.029019735753536224,
0.06691712886095047,
0.02457265369594097,
0.12158234417438507,
0.051102593541145325,
-0.12649741768836975,
0.25464025139808655,
0.021814975887537003,
0.20960621535778046,
0.16022542119026184,
-0.02117924578487873,
0.04937392845749855,
-0.11429550498723984,
-0.004021795932203531,
0.018115581944584846,
0.03620346635580063,
0.025863325223326683,
0.07361716777086258,
-0.06557492166757584,
0.01834634132683277,
-0.04475898668169975,
0.04679056629538536,
0.01387613732367754,
-0.05940280109643936,
-0.07861513644456863,
0.003323150333017111,
0.21123148500919342,
-0.14867042005062103,
0.1565840244293213,
0.36614319682121277,
0.10582873225212097,
0.13521641492843628,
-0.10336575657129288,
-0.02612939290702343,
-0.10750477761030197,
0.05765321105718613,
-0.0036993264220654964,
0.18862055242061615,
-0.05494143068790436,
0.004229951184242964,
0.053485702723264694,
0.02561568096280098,
0.07369932532310486,
-0.22572201490402222,
-0.1541752964258194,
-0.037913087755441666,
-0.11304649710655212,
-0.18841543793678284,
0.0590190514922142,
-0.13209284842014313,
0.02683337777853012,
0.029957830905914307,
-0.06477067619562149,
0.17055083811283112,
-0.008533386513590813,
-0.09367663413286209,
0.08508569002151489,
-0.1732875406742096,
-0.09793684631586075,
-0.14603638648986816,
-0.03374042734503746,
-0.0076118106953799725,
0.06333117932081223,
0.052523668855428696,
-0.033548031002283096,
-0.04136210307478905,
0.050138864666223526,
-0.09441986680030823,
-0.12083601206541061,
-0.004220837727189064,
0.015548563562333584,
0.05213695764541626,
-0.04403722658753395,
-0.07769571989774704,
-0.06085691601037979,
-0.030102429911494255,
-0.13098199665546417,
0.08954382687807083,
-0.05736219882965088,
0.07702329009771347,
0.12555082142353058,
0.05137493461370468,
0.06333070248365402,
-0.049093544483184814,
0.13393935561180115,
-0.05090206488966942,
-0.13872745633125305,
0.11766421049833298,
0.020131098106503487,
0.025496751070022583,
0.09193360805511475,
0.10713472217321396,
-0.100267194211483,
-0.06456533074378967,
-0.11210892349481583,
-0.13933908939361572,
-0.15915995836257935,
-0.03488801047205925,
-0.08391232788562775,
0.10387539863586426,
-0.0014467764412984252,
0.1304897665977478,
0.10017895698547363,
0.06057880073785782,
0.04891480132937431,
0.010672791860997677,
0.04154729098081589,
-0.0064173112623393536,
0.1132214292883873,
-0.03280860558152199,
-0.048583127558231354,
-0.09838015586137772,
0.0750044584274292,
0.19544142484664917,
0.12944787740707397,
0.14806556701660156,
0.25038769841194153,
0.1162492036819458,
0.1500038206577301,
0.12447869032621384,
0.16683925688266754,
-0.014462476596236229,
0.026183968409895897,
-0.020337169989943504,
-0.02514774352312088,
-0.04342241585254669,
0.009672876447439194,
0.04877227172255516,
0.07152079790830612,
-0.2578199505805969,
0.057423386722803116,
-0.26459720730781555,
0.047323718667030334,
-0.0264968890696764,
0.10074887424707413,
0.010484290309250355,
0.07304065674543381,
0.055277325212955475,
0.11413758248090744,
0.006739584729075432,
0.1047661229968071,
0.029175668954849243,
-0.06208187714219093,
0.010828175581991673,
0.02589789777994156,
0.027661502361297607,
0.010889626108109951,
0.026007933542132378,
-0.06981703639030457,
-0.15174192190170288,
0.047919318079948425,
0.10266148298978806,
-0.2341543734073639,
0.22367635369300842,
0.04332678020000458,
-0.07883649319410324,
-0.028190338984131813,
-0.03916743025183678,
0.04511460289359093,
0.15946823358535767,
0.15238574147224426,
0.08564817160367966,
-0.18637973070144653,
-0.08639948815107346,
-0.06085130199790001,
0.024716239422559738,
0.099901482462883,
-0.07874951511621475,
-0.16732285916805267,
-0.019885720685124397,
0.06420081108808517,
0.008748610503971577,
0.0796949565410614,
-0.1043432429432869,
-0.07909058779478073,
0.062365639954805374,
0.08195742964744568,
0.049818217754364014,
-0.10041768103837967,
0.0827634409070015,
-0.04343034327030182,
0.11855664849281311,
-0.24440652132034302,
0.0143445935100317,
-0.06526421755552292,
-0.11118987202644348,
0.006016744766384363,
-0.02954689972102642,
0.026638658717274666,
-0.060394927859306335,
-0.11309129744768143,
-0.13042578101158142,
-0.18037976324558258,
0.1058741882443428,
-0.04664653167128563,
-0.0031292319763451815,
-0.01701299101114273,
0.1249459907412529,
-0.051772020757198334,
0.0416560024023056,
0.005113715305924416,
0.025332754477858543,
0.02089250646531582,
-0.1680154651403427,
0.12925869226455688,
-0.09858588874340057,
0.01665518246591091,
0.0383087582886219,
0.012331852689385414,
0.06972988694906235,
0.07222073525190353,
-0.09775391221046448,
0.1727329045534134,
0.4025157690048218,
-0.07598510384559631,
0.21355897188186646,
0.3101240396499634,
-0.09475591033697128,
-0.23652935028076172,
-0.13385847210884094,
-0.24247144162654877,
-0.08802571892738342,
0.01049027219414711,
-0.21234364807605743,
-0.020915931090712547,
0.23563900589942932,
-0.13717970252037048,
0.3206854462623596,
-0.24011783301830292,
-0.03035585768520832,
0.07553593814373016,
-0.022410085424780846,
0.4501473307609558,
-0.15442438423633575,
-0.14052213728427887,
0.014619076624512672,
-0.19739745557308197,
0.1726800799369812,
-0.02765813283622265,
0.08036572486162186,
0.011479559354484081,
-0.08151566982269287,
-0.036223143339157104,
-0.030575919896364212,
0.23901014029979706,
-0.02394230291247368,
0.07251737266778946,
-0.057778920978307724,
-0.10486498475074768,
0.2418813556432724,
0.03297712653875351,
-0.05480247363448143,
-0.10198857635259628,
-0.03542965650558472,
0.004913857206702232,
0.020295506343245506,
-0.03766421973705292,
0.08225714415311813,
0.0007448230171576142,
-0.08730878680944443,
-0.1066865622997284,
0.013568487949669361,
-0.14038710296154022,
-0.010460998862981796,
0.19220070540905,
-0.05297532305121422,
0.01065878290683031,
0.0864662453532219,
-0.046442899852991104,
-0.13443659245967865,
-0.021216802299022675,
-0.0892062708735466,
-0.07710292935371399,
0.07093127071857452,
-0.19061638414859772,
-0.00037661430542357266,
0.0692189633846283,
0.020628515630960464,
0.08544549345970154,
0.09256470948457718,
-0.05484466254711151,
0.08713635802268982,
0.16987575590610504,
-0.14346818625926971,
-0.10943834483623505,
0.002131785498932004,
-0.015721268951892853,
0.1981460601091385,
0.013927293010056019,
0.06838644295930862,
0.03381659463047981,
0.021103667095303535,
0.01412686426192522,
0.03760719671845436,
-0.17391914129257202,
-0.04344836249947548,
0.021758362650871277,
-0.03347136080265045,
-0.12249195575714111,
0.16211901605129242,
0.0445558987557888,
0.028728395700454712,
-0.06190894544124603,
0.004011211451143026,
-0.06715872883796692,
-0.08004838973283768,
-0.24616838991641998,
-0.07927387207746506,
-0.19086040556430817,
-0.0996348112821579,
0.03957706317305565,
-0.10118209570646286,
-0.04581325128674507,
0.06354472786188126,
0.026706142351031303,
0.18537597358226776,
0.06892016530036926,
0.021364936605095863,
0.0928003117442131,
-0.09637685120105743,
-0.24104343354701996,
-0.005611174274235964,
-0.055640991777181625,
-0.05308784171938896,
0.016761578619480133,
0.07514820247888565,
-0.05256318673491478,
-0.04409831762313843,
-0.1433177888393402,
0.049480143934488297,
-0.05810711532831192,
0.03033287450671196,
-0.07977879792451859,
-0.01970886066555977,
0.058054808527231216,
-0.03677498176693916,
-0.002708164742216468,
0.025268226861953735,
-0.13445106148719788,
0.021365158259868622,
0.0023989423643797636,
0.05454486981034279,
-0.05146732181310654,
-0.0442771390080452,
0.0885404571890831,
0.057835932821035385,
0.14600001275539398,
0.11502771824598312,
0.06255172193050385,
0.13989616930484772,
-0.2353791743516922,
-0.014779730699956417,
0.08913028985261917,
-0.0367823950946331,
-0.03381944075226784,
0.03242078050971031,
0.005717463791370392,
0.046068254858255386,
-0.09858133643865585,
0.062368668615818024,
-0.015455291606485844,
-0.12212461978197098,
-0.13828162848949432,
-0.0370500348508358,
-0.09496131539344788,
-0.0033643355127424,
-0.11221759766340256,
0.20280085504055023,
0.07891929894685745,
0.054196540266275406,
0.04359645023941994,
-0.023567242547869682,
0.026685798540711403,
-0.0025991732254624367,
-0.01844479888677597,
-0.08649298548698425,
-0.13000717759132385,
-0.03876118361949921,
-0.08097053319215775,
-0.0006188988918438554,
0.35593241453170776,
-0.025028323754668236,
-0.17922283709049225,
0.03535255417227745,
0.08669308573007584,
0.11369588971138,
0.017201751470565796,
0.28193867206573486,
0.06560564786195755,
0.0050177136436104774,
-0.1596011072397232,
0.07880295068025589,
-0.04124697297811508,
-0.26902005076408386,
0.06937892735004425,
-0.026316599920392036,
-0.007569037843495607,
0.006292602047324181,
0.10640519112348557,
-0.11218146979808807,
-0.007308312226086855,
0.030519235879182816,
0.02614116109907627,
-0.00870091188699007,
-0.03157037869095802,
0.02083216793835163,
0.18787455558776855,
-0.07524357736110687,
0.005316397175192833,
0.015087205916643143,
-0.012456162832677364,
-0.1474122405052185,
-0.15606985986232758,
0.007377709727734327,
-0.13592395186424255,
0.09237334877252579,
-0.012600034475326538,
0.06192610040307045,
0.21725086867809296,
0.014514447189867496,
-0.04895612969994545,
-0.08309737592935562,
-0.13144530355930328,
-0.06324486434459686,
0.0387842059135437,
-0.02199329249560833,
-0.007349626626819372,
-0.10115515440702438,
-0.12051966041326523,
0.014570314437150955,
-0.1594381183385849,
0.013787773437798023,
0.028851108625531197,
0.08254057914018631,
-0.029874086380004883,
-0.10572104901075363,
-0.021927354857325554,
-0.07995882630348206,
0.0794512927532196,
-0.022885603830218315,
0.17528417706489563,
0.014030463062226772,
0.00012621283531188965,
0.08713336288928986,
0.026602691039443016,
-0.006958980578929186,
-0.04666709899902344,
-0.0013765465700998902,
0.12055453658103943,
-0.04865662753582001,
0.12045476585626602,
-0.02466321736574173,
-0.02633259817957878,
0.037972304970026016,
0.18264660239219666,
0.18724308907985687,
-0.13123492896556854,
0.02243487350642681,
0.0020434707403182983,
0.017737263813614845,
0.12215398997068405,
0.16936393082141876,
-0.024920029565691948,
0.2638060748577118,
-0.0737442895770073,
-0.06217733398079872,
-0.0011923612328246236,
0.05878657102584839,
-0.06870530545711517,
0.05593832954764366,
-0.017074158415198326,
-0.05075569450855255,
-0.10051200538873672,
0.09360083937644958,
-0.12242018431425095,
0.0938929095864296,
0.18321199715137482,
-0.04137864336371422,
0.08520513027906418,
-0.00484077213332057,
-0.002665484556928277,
-0.00784443411976099,
0.07463321089744568,
-0.13740283250808716,
-0.06943826377391815,
-0.09232878684997559,
0.032780054956674576,
-0.33550313115119934,
-0.08123112469911575,
0.0592232383787632,
0.13403183221817017,
0.20259681344032288,
-0.009431025013327599,
0.14047439396381378,
0.009761088527739048,
0.06491469591856003,
-0.08515305817127228,
0.15184764564037323,
-0.005681826267391443,
-0.10309519618749619,
-0.16108819842338562,
-0.14821793138980865,
-0.019807560369372368,
-0.021608194336295128,
0.03813767060637474,
0.07864711433649063,
0.05145956575870514,
0.1440734714269638,
-0.0750885158777237,
-0.02517952397465706,
-0.0535120964050293,
-0.12568412721157074,
0.03698430955410004,
-0.07320766896009445,
-0.0025476450100541115,
-0.09082155674695969,
-0.06315287947654724,
0.004286549054086208,
0.0905798003077507,
-0.12416733056306839,
-0.03217146173119545,
0.15467047691345215,
0.02584569901227951,
0.219248965382576,
0.0017291251569986343,
0.0119365518912673,
0.0010933049488812685,
-0.06383669376373291,
0.13900251686573029,
-0.07753364741802216,
0.06167154014110565,
0.14888395369052887,
-0.003979967441409826,
0.005743835587054491,
-0.24616141617298126,
0.03601286932826042,
-0.06628955155611038,
-0.007753883022814989,
-0.059474650770425797
] |
null | null | diffusers |
# DreamBooth trained by AutoTrain
Text encoder was trained.
| {"tags": ["text-to-image", "diffusers", "autotrain"], "base_model": "stabilityai/stable-diffusion-xl-base-1.0", "instance_prompt": "photo of gbb", "inference": true} | text-to-image | anjith672/gate-boy2 | [
"diffusers",
"text-to-image",
"autotrain",
"base_model:stabilityai/stable-diffusion-xl-base-1.0",
"has_space",
"region:us"
] | 2024-02-07T06:42:19+00:00 | [] | [] | TAGS
#diffusers #text-to-image #autotrain #base_model-stabilityai/stable-diffusion-xl-base-1.0 #has_space #region-us
|
# DreamBooth trained by AutoTrain
Text encoder was trained.
| [
"# DreamBooth trained by AutoTrain\n\nText encoder was trained."
] | [
"TAGS\n#diffusers #text-to-image #autotrain #base_model-stabilityai/stable-diffusion-xl-base-1.0 #has_space #region-us \n",
"# DreamBooth trained by AutoTrain\n\nText encoder was trained."
] | [
45,
18
] | [
"passage: TAGS\n#diffusers #text-to-image #autotrain #base_model-stabilityai/stable-diffusion-xl-base-1.0 #has_space #region-us \n# DreamBooth trained by AutoTrain\n\nText encoder was trained."
] | [
-0.04398449882864952,
0.07840234786272049,
-0.0005392950843088329,
0.02930283546447754,
0.14337566494941711,
-0.006497366353869438,
0.16302914917469025,
0.06228417903184891,
-0.04899086803197861,
0.016302336007356644,
0.17255187034606934,
-0.0001382838818244636,
-0.019391635432839394,
0.2852212190628052,
-0.022004559636116028,
-0.19885662198066711,
0.04498407244682312,
-0.012186543084681034,
0.030130913481116295,
0.06842844933271408,
0.04598861560225487,
-0.10089220106601715,
0.07062596082687378,
-0.10867820680141449,
-0.23793157935142517,
0.061639152467250824,
-0.0024120621383190155,
-0.06989245861768723,
0.07536890357732773,
0.05687239393591881,
0.12212380021810532,
0.08819300681352615,
0.10619136691093445,
-0.10085877031087875,
0.04370025545358658,
0.0618499219417572,
-0.04823364317417145,
0.04618982598185539,
-0.02723781205713749,
0.022576602175831795,
-0.006881612353026867,
0.028912149369716644,
0.053117599338293076,
-0.0006344228750094771,
-0.10675296187400818,
0.11659980565309525,
0.08412143588066101,
0.022547302767634392,
0.12483171373605728,
0.09035079181194305,
-0.02833043783903122,
0.03825448453426361,
-0.023406554013490677,
0.1361367553472519,
0.01914517767727375,
-0.16131064295768738,
-0.07149805128574371,
0.23857855796813965,
0.05967278406023979,
0.1822482943534851,
-0.10062692314386368,
0.1028866246342659,
0.12148696184158325,
-0.01657138578593731,
0.02235104888677597,
-0.03278966248035431,
-0.040145713835954666,
-0.03776431083679199,
-0.07170133292675018,
-0.027627315372228622,
0.20846714079380035,
-0.03431273251771927,
0.0013279474806040525,
-0.0764944851398468,
-0.10208645462989807,
-0.028919914737343788,
-0.01955300383269787,
-0.029654070734977722,
-0.029045764356851578,
0.047539930790662766,
-0.0718991681933403,
-0.08675273507833481,
-0.10041247308254242,
-0.07655340433120728,
-0.0627894327044487,
0.13043764233589172,
-0.0542542040348053,
0.022149493917822838,
-0.06919136643409729,
0.15755492448806763,
0.047388773411512375,
-0.1627960205078125,
0.052154991775751114,
-0.059900205582380295,
-0.00930443312972784,
0.03430337458848953,
-0.0649663582444191,
-0.12501630187034607,
0.043364569544792175,
0.03855999559164047,
0.10298798978328705,
0.02484908327460289,
-0.026976099237799644,
0.09129809588193893,
0.03237375244498253,
0.10407932847738266,
0.026142077520489693,
-0.08161938190460205,
0.02233007736504078,
0.04551607370376587,
-0.012161053717136383,
-0.08278302848339081,
-0.19265371561050415,
0.09385792165994644,
-0.00043560718768276274,
0.03094261884689331,
-0.007527773734182119,
-0.006900179665535688,
-0.015633156523108482,
0.01295509748160839,
0.05424166098237038,
-0.06756512820720673,
0.016468578949570656,
-0.09558184444904327,
-0.019480928778648376,
-0.057944949716329575,
0.10298903286457062,
0.006915216334164143,
0.008255493827164173,
0.01008623093366623,
-0.08427032083272934,
0.01870490238070488,
-0.059023041278123856,
-0.07479500770568848,
-0.05108042433857918,
-0.10132039338350296,
0.06042269244790077,
-0.1875678151845932,
-0.14805662631988525,
-0.04504276439547539,
0.0356040894985199,
-0.051344357430934906,
-0.03839026018977165,
-0.12604625523090363,
-0.0977262333035469,
0.09181345254182816,
-0.028064001351594925,
0.01517318282276392,
-0.011118416674435139,
0.049855079501867294,
-0.007151172496378422,
0.13657556474208832,
-0.1493682861328125,
-0.0002590612566564232,
-0.07700586318969727,
-0.006498237140476704,
-0.05488988757133484,
0.1706833839416504,
-0.041211917996406555,
0.0809163823723793,
0.0008876651991158724,
0.011572323739528656,
-0.01041377056390047,
0.061952777206897736,
0.029133981093764305,
0.15863752365112305,
-0.19792407751083374,
-0.041472580283880234,
0.15125033259391785,
-0.11005294322967529,
-0.05521046742796898,
0.017598092555999756,
-0.03370938077569008,
0.12187504768371582,
0.04926693066954613,
0.15910214185714722,
0.001473760581575334,
-0.13716813921928406,
0.05945860594511032,
0.0362715907394886,
-0.03782058134675026,
-0.0015277958009392023,
-0.01530655100941658,
0.12076370418071747,
-0.07637878507375717,
0.020671680569648743,
-0.04382089897990227,
0.1292186677455902,
-0.0702093318104744,
-0.07982861995697021,
-0.04252699390053749,
-0.05008916184306145,
0.018470339477062225,
-0.0007848223904147744,
0.0934537947177887,
-0.05276156961917877,
-0.055823106318712234,
0.05119020491838455,
0.05955780670046806,
0.0038683379534631968,
-0.02301659807562828,
-0.038894932717084885,
-0.02535870671272278,
-0.11221860349178314,
0.019136687740683556,
-0.12526828050613403,
-0.06441108882427216,
-0.03084789775311947,
0.2759895324707031,
0.06410437822341919,
0.14262592792510986,
0.057107578963041306,
0.06563644856214523,
0.002145481528714299,
-0.10903996229171753,
0.003906522411853075,
0.03306916728615761,
-0.056985098868608475,
-0.09410533308982849,
0.09743699431419373,
-0.13036927580833435,
0.024503711611032486,
-0.15702125430107117,
0.04041634127497673,
-0.08122150599956512,
0.11513203382492065,
0.06794825941324234,
-0.0008592758094891906,
-0.037394165992736816,
0.05535007268190384,
-0.1072564348578453,
-0.07576112449169159,
0.006407795939594507,
0.018456224352121353,
-0.055999305099248886,
0.12139390408992767,
-0.17888638377189636,
0.17045678198337555,
0.14621740579605103,
-0.06571825593709946,
-0.07191716879606247,
0.048308029770851135,
0.0018976664869114757,
0.01003638282418251,
-0.029841840267181396,
-0.00969010591506958,
0.12527291476726532,
-0.06360334157943726,
0.17711034417152405,
-0.013147599995136261,
0.05188120901584625,
0.04656572639942169,
-0.08803264796733856,
-0.06296636164188385,
-0.015434738248586655,
-0.02097306214272976,
-0.08055289089679718,
0.10716642439365387,
0.14278551936149597,
-0.04771167412400246,
0.27152568101882935,
0.006783745251595974,
0.0013470095582306385,
-0.013609991408884525,
-0.03356857970356941,
-0.05762971565127373,
0.13646158576011658,
-0.04819853603839874,
-0.06686806678771973,
-0.03177184984087944,
-0.027023596689105034,
0.03389295935630798,
-0.13654537498950958,
-0.016597507521510124,
0.01478695310652256,
-0.01576651632785797,
0.1418440043926239,
0.012451480142772198,
-0.11274145543575287,
0.061392899602651596,
-0.07733384519815445,
-0.16127309203147888,
0.05899142101407051,
-0.012793055735528469,
-0.012569020502269268,
0.10479377955198288,
-0.10783875733613968,
-0.3281002938747406,
-0.1528884768486023,
-0.03974168747663498,
-0.038401421159505844,
0.06075701117515564,
0.05908828601241112,
-0.1215292364358902,
-0.022533584386110306,
-0.07008418440818787,
-0.07399658113718033,
0.005266725551337004,
0.014565873891115189,
0.07999901473522186,
-0.009142349474132061,
0.029043909162282944,
-0.04068751633167267,
-0.013133842498064041,
-0.021691948175430298,
0.003150958102196455,
0.1055917739868164,
0.010477310046553612,
0.09428497403860092,
0.21048946678638458,
-0.025836963206529617,
0.02378997579216957,
-0.0026474320329725742,
0.24176816642284393,
-0.08467277884483337,
0.043600328266620636,
0.09087527543306351,
0.019019214436411858,
0.0119625860825181,
0.1729641556739807,
-0.005805306136608124,
-0.0862494483590126,
0.09562665969133377,
-0.04648980498313904,
-0.12253358215093613,
-0.027785351499915123,
-0.12838903069496155,
-0.02880263887345791,
-0.02379264496266842,
0.08407612890005112,
0.0792146623134613,
0.12751153111457825,
0.06632808595895767,
0.08829693496227264,
0.010995512828230858,
-0.027369150891900063,
0.05263402312994003,
-0.004726458806544542,
-0.10054438561201096,
0.12419746071100235,
-0.028948385268449783,
-0.12924693524837494,
0.0798472911119461,
0.02917303331196308,
0.03022497147321701,
0.039817824959754944,
-0.09246006608009338,
-0.019542129710316658,
0.030529804527759552,
0.1590501219034195,
0.08309800922870636,
0.08768364787101746,
-0.04060669615864754,
-0.020963242277503014,
-0.00848899781703949,
-0.018347524106502533,
0.10451194643974304,
0.03582065925002098,
-0.06234673038125038,
-0.02469589188694954,
0.08054845780134201,
0.0006566211231984198,
0.024967024102807045,
0.09835029393434525,
-0.31634974479675293,
0.02008422650396824,
0.04085671156644821,
-0.0012458509299904108,
-0.10784264653921127,
0.02262108586728573,
0.2766202688217163,
-0.053005773574113846,
-0.05295083671808243,
-0.0028184684924781322,
0.08323218673467636,
0.06860175728797913,
0.011763076297938824,
-0.13347800076007843,
0.02682235650718212,
-0.012321293354034424,
0.019401103258132935,
-0.26415422558784485,
0.08349677175283432,
-0.04252581298351288,
0.040099579840898514,
-0.0021024071611464024,
0.013640999794006348,
0.05720065161585808,
0.14610251784324646,
0.07030731439590454,
0.009062102995812893,
-0.13646474480628967,
-0.15551920235157013,
-0.15407361090183258,
-0.012840519659221172,
0.07296836376190186,
0.008643276989459991,
0.00720151187852025,
0.037690673023462296,
-0.014467615634202957,
0.01606927439570427,
-0.04130398482084274,
-0.15606921911239624,
-0.1113385260105133,
-0.04238974303007126,
0.13226783275604248,
0.15619324147701263,
-0.04403433948755264,
-0.07892896980047226,
0.046028148382902145,
0.14313365519046783,
0.03001726232469082,
-0.01027599722146988,
-0.11199046671390533,
0.02297889068722725,
-0.023078223690390587,
-0.011686989106237888,
0.07737375050783157,
-0.021515756845474243,
0.05987716093659401,
-0.04767050966620445,
-0.10589359700679779,
0.07817928493022919,
-0.11487112939357758,
-0.0393560528755188,
-0.10965920984745026,
-0.026854852214455605,
-0.060831908136606216,
-0.029004065319895744,
0.04367813467979431,
0.0029224504251033068,
-0.07043464481830597,
-0.07403062283992767,
0.07431022077798843,
0.027637924998998642,
-0.12946130335330963,
0.058321673423051834,
0.005933678243309259,
-0.08159984648227692,
-0.03454512357711792,
0.017108475789427757,
0.18848562240600586,
0.15262630581855774,
-0.08569110929965973,
0.10292760282754898,
0.1318686306476593,
-0.07349705696105957,
-0.32850182056427,
-0.10877734422683716,
0.01566787250339985,
0.07203064858913422,
-0.04180268198251724,
-0.1966724842786789,
0.07605405896902084,
-0.05192657932639122,
-0.048426538705825806,
0.09007546305656433,
-0.2589987516403198,
-0.08489791303873062,
0.1493651568889618,
0.0566369965672493,
0.3180140256881714,
-0.09842395782470703,
-0.01414039172232151,
-0.05555589497089386,
0.04488496482372284,
0.09165439009666443,
-0.01009589433670044,
0.12556472420692444,
0.018507158383727074,
-0.0001558443473186344,
0.006427996326237917,
-0.016958080232143402,
0.1303485780954361,
-0.06565137207508087,
0.06260091811418533,
-0.12285038828849792,
0.036803510040044785,
0.09831491112709045,
-0.05688164010643959,
0.03982168808579445,
-0.10709307342767715,
0.07087772339582443,
-0.11882220208644867,
0.01087918784469366,
0.024886710569262505,
-0.01275100838392973,
0.038805071264505386,
-0.1004834994673729,
0.0027907544281333685,
-0.0017571188509464264,
0.017010176554322243,
0.01606880873441696,
-0.016784148290753365,
-0.04732600226998329,
-0.011930224485695362,
0.2669828534126282,
-0.03779172524809837,
-0.11321185529232025,
-0.00954214297235012,
0.0053439815528690815,
-0.06098959967494011,
0.15120624005794525,
-0.1278066784143448,
0.019782595336437225,
0.09627318382263184,
-0.022020144388079643,
0.1735590398311615,
0.06607235223054886,
-0.007302126847207546,
0.08805494010448456,
0.1653764396905899,
-0.16810080409049988,
0.017034152522683144,
-0.09561080485582352,
-0.013645108789205551,
0.11223644018173218,
-0.05137640982866287,
0.16786274313926697,
-0.0736440047621727,
0.0439506359398365,
-0.03254624083638191,
0.017557300627231598,
-0.07270462065935135,
0.14274218678474426,
0.016568897292017937,
0.03293192759156227,
-0.06983616948127747,
0.07416192442178726,
0.024350028485059738,
-0.033058054745197296,
0.07308115810155869,
0.08398956805467606,
-0.05483098700642586,
-0.07131420075893402,
0.013790532946586609,
0.25049012899398804,
-0.11385786533355713,
-0.03996465727686882,
-0.07691344618797302,
-0.08751638978719711,
0.009756740182638168,
0.04352245107293129,
0.05898971110582352,
0.018403783440589905,
-0.06614949554204941,
-0.016936788335442543,
-0.05309519171714783,
0.043922241777181625,
0.08153904229402542,
0.0859684944152832,
-0.19647157192230225,
-0.011295814998447895,
0.011317769065499306,
0.08331634849309921,
-0.14369851350784302,
-0.10548124462366104,
-0.11811259388923645,
0.015490113757550716,
-0.07760456949472427,
0.06966302543878555,
0.0226095300167799,
-0.04230327531695366,
0.03351517394185066,
-0.002233945531770587,
-0.020256508141756058,
0.011399087496101856,
-0.06992927938699722,
0.006636819336563349,
-0.010699036531150341,
-0.017497427761554718,
-0.012530113570392132,
-0.030591707676649094,
-0.014279176481068134,
-0.041220735758543015,
0.06336469948291779,
-0.0169439148157835,
-0.07126829773187637,
0.019546691328287125,
-0.16222506761550903,
-0.04920502007007599,
0.18533927202224731,
0.002143733436241746,
-0.0010857926681637764,
0.12443237751722336,
-0.007929041981697083,
0.03942568227648735,
0.04640401527285576,
-0.011313479393720627,
0.023228058591485023,
-0.06695953011512756,
-0.040417566895484924,
-0.06301992386579514,
-0.04854709655046463,
-0.06167404726147652,
0.04083915799856186,
0.07926639169454575,
0.08998207747936249,
0.14946192502975464,
-0.16485145688056946,
0.05689458176493645,
-0.08486084640026093,
0.005013452377170324,
-0.003398339729756117,
-0.09475532919168472,
-0.010172471404075623,
-0.07430905103683472,
0.016007618978619576,
-0.02867160737514496,
0.16243410110473633,
0.14979976415634155,
-0.1608181893825531,
-0.021943869069218636,
-0.015085513703525066,
0.030681239441037178,
0.026778919622302055,
0.28097087144851685,
0.11437889188528061,
-0.010923956520855427,
-0.1276623159646988,
0.07557732611894608,
0.1431773453950882,
0.11301200091838837,
-0.037317678332328796,
0.03119102120399475,
0.04769498109817505,
0.15891259908676147,
0.029058467596769333,
0.00782306119799614,
-0.021442264318466187,
0.026038818061351776,
-0.16292111575603485,
0.12082578241825104,
-0.08065149933099747,
-0.11467693746089935,
0.09666301310062408,
-0.022490814328193665,
-0.03920892998576164,
0.0642579197883606,
-0.0634671002626419,
-0.09287332743406296,
-0.03044290281832218,
-0.08640941232442856,
-0.1994573026895523,
0.02100512944161892,
-0.06759560108184814,
0.04678531363606453,
0.06394045799970627,
0.021350771188735962,
-0.023513715714216232,
0.10906615108251572,
0.06311928480863571,
-0.0664956197142601,
0.13000455498695374,
-0.034304264932870865,
-0.0013512027217075229,
-0.10135696828365326,
-0.022439273074269295,
0.020967386662960052,
0.0846933051943779,
-0.012489408254623413,
0.02687808871269226,
-0.00457615265622735,
0.06173684820532799,
-0.024956468492746353,
-0.07466288655996323,
-0.009079880081117153,
0.020233985036611557,
-0.04366102069616318,
0.12752996385097504,
0.06405851989984512,
-0.012991180643439293,
0.0027766325511038303,
0.20657439529895782,
-0.07669415324926376,
-0.1343177855014801,
-0.12075874209403992,
0.15557803213596344,
-0.04695359990000725,
0.15158337354660034,
-0.11404812335968018,
-0.07823263108730316,
-0.0912177637219429,
0.1441309005022049,
0.19782213866710663,
-0.13462242484092712,
0.015302689746022224,
-0.053257331252098083,
-0.0036703660152852535,
-0.05227004736661911,
0.1895093023777008,
0.042662739753723145,
0.11369999498128891,
-0.07013270258903503,
-0.042288873344659805,
-0.07614127546548843,
-0.12223438918590546,
-0.02773137576878071,
-0.14772717654705048,
0.021819084882736206,
-0.03706600144505501,
-0.12499638646841049,
0.005303726065903902,
-0.16816505789756775,
0.06757618486881256,
0.1454460322856903,
-0.11597441881895065,
-0.040313124656677246,
0.003077184548601508,
0.16935569047927856,
0.042297735810279846,
0.017155194655060768,
-0.06543263047933578,
0.037335220724344254,
0.06343232840299606,
-0.06410717964172363,
-0.03337114676833153,
-0.00588108180090785,
-0.032975394278764725,
-0.31616532802581787,
0.14293108880519867,
-0.05051179602742195,
0.021998068317770958,
0.02259230613708496,
0.02136460691690445,
-0.055217403918504715,
0.14757336676120758,
-0.0963498130440712,
-0.006790807005017996,
0.00011305488442303613,
0.18810954689979553,
-0.010394245386123657,
0.030170120298862457,
0.03891069442033768,
-0.13238337635993958,
-0.011865473352372646,
-0.058892469853162766,
-0.06090613827109337,
-0.03851604834198952,
-0.046546224504709244,
-0.0012726830318570137,
0.11132504791021347,
0.027883170172572136,
-0.02131570875644684,
0.05026746541261673,
0.005542791448533535,
0.0306076817214489,
-0.01859206147491932,
-0.004702788311988115,
-0.014624848030507565,
-0.1304505318403244,
-0.07162836194038391,
0.040973417460918427,
-0.004633525386452675,
-0.23884403705596924,
-0.047574233263731,
-0.16205225884914398,
0.04243658483028412,
-0.03479912877082825,
0.1384517401456833,
0.23807911574840546,
0.022481542080640793,
-0.01170669961720705,
-0.1539752334356308,
0.032638370990753174,
0.08036516606807709,
-0.022248782217502594,
-0.0926564484834671
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | rushidesh/mistral_b_finance_finetuned_test | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-07T06:43:41+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
31,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06646376848220825,
0.2168014943599701,
-0.00225935154594481,
0.023818302899599075,
0.1271018385887146,
-0.001635765191167593,
0.04218708351254463,
0.13324736058712006,
-0.020175931975245476,
0.11144465953111649,
0.046588581055402756,
0.09377603232860565,
0.09928803145885468,
0.18404334783554077,
0.04859916493296623,
-0.2059975117444992,
0.007056170143187046,
-0.09090408682823181,
0.014076028019189835,
0.1116579994559288,
0.13719257712364197,
-0.10291384905576706,
0.08272874355316162,
-0.04045208916068077,
-0.02019004337489605,
0.00012576708104461432,
-0.09259183704853058,
-0.07032395154237747,
0.06885425746440887,
0.06264153122901917,
0.051234472543001175,
0.001456156256608665,
0.09140396863222122,
-0.2864592671394348,
0.017265573143959045,
0.08406311273574829,
0.0027674848679453135,
0.06290827691555023,
0.07236549258232117,
-0.07389893382787704,
0.11328595131635666,
-0.08021481335163116,
0.13019037246704102,
0.08625296503305435,
-0.062064990401268005,
-0.23071379959583282,
-0.07525765895843506,
0.0963398814201355,
0.12251301854848862,
0.06215599179267883,
-0.022921854630112648,
0.15455181896686554,
-0.06248689442873001,
0.012971068732440472,
0.1294165402650833,
-0.11526761949062347,
-0.05572471022605896,
0.061741601675748825,
0.11775490641593933,
0.10740239918231964,
-0.14110268652439117,
-0.0017287094378843904,
0.04900608956813812,
0.029121357947587967,
0.08589313924312592,
0.022661056369543076,
0.12003941088914871,
0.04652795568108559,
-0.13695219159126282,
-0.04037507623434067,
0.12011898308992386,
0.038862764835357666,
-0.06446044892072678,
-0.2168138176202774,
-0.006778308190405369,
-0.0601806715130806,
-0.014732478186488152,
-0.07019448280334473,
0.039128515869379044,
-0.02470310963690281,
0.07317749410867691,
-0.04465159401297569,
-0.1063927412033081,
-0.0421026237308979,
0.0892222449183464,
0.07748593389987946,
0.011527054943144321,
-0.02519804798066616,
0.04627908393740654,
0.13455867767333984,
0.05402068421244621,
-0.10399353504180908,
-0.07017925381660461,
-0.06942764669656754,
-0.09420394152402878,
-0.04035796597599983,
0.056760527193546295,
0.031942449510097504,
0.02665667235851288,
0.22703726589679718,
0.016653569415211678,
0.04155244305729866,
0.0224777739495039,
0.01032855175435543,
0.043662428855895996,
0.0955500528216362,
-0.05303520709276199,
-0.15660029649734497,
-0.04072032496333122,
0.09077946096658707,
-0.0027527001220732927,
-0.036689214408397675,
-0.03966725245118141,
0.03849169611930847,
0.06843466311693192,
0.13122352957725525,
0.07552056759595871,
-0.017929591238498688,
-0.04813180863857269,
-0.030096933245658875,
0.23523783683776855,
-0.1493375599384308,
0.04426715523004532,
-0.02271856553852558,
-0.01804111897945404,
-0.03908449783921242,
0.03597262129187584,
0.022118929773569107,
-0.000004518366949923802,
0.09706240892410278,
-0.058981191366910934,
-0.05378659814596176,
-0.10168042778968811,
-0.03272576630115509,
0.04088849574327469,
-0.013975566253066063,
-0.010589460842311382,
-0.09025166928768158,
-0.09490354359149933,
-0.04766594246029854,
0.05537205561995506,
-0.05123869329690933,
-0.03770573064684868,
0.009465423412621021,
-0.08151785284280777,
-0.005444355774670839,
-0.005417742300778627,
0.10699385404586792,
-0.03222226724028587,
0.04445803165435791,
-0.027600755915045738,
0.05225523188710213,
0.09919606149196625,
0.031576547771692276,
-0.0773419588804245,
0.0561848059296608,
-0.22559374570846558,
0.07503069192171097,
-0.11481974273920059,
0.04335082694888115,
-0.1704932004213333,
-0.042439818382263184,
0.005444696638733149,
0.0139949731528759,
0.013206101022660732,
0.12720820307731628,
-0.19255615770816803,
-0.01654396951198578,
0.13260798156261444,
-0.09212633967399597,
-0.118110790848732,
0.07884611934423447,
-0.029701577499508858,
0.1624738723039627,
0.04682036489248276,
-0.027025915682315826,
0.09224298596382141,
-0.16434773802757263,
-0.07092688232660294,
-0.00949116237461567,
-0.01727987825870514,
0.12109188735485077,
0.07512219995260239,
-0.05991523340344429,
0.046571120619773865,
0.02832140028476715,
-0.038078423589468,
-0.04424772411584854,
-0.050857074558734894,
-0.10884185880422592,
-0.01070026308298111,
-0.08987759798765182,
0.04065500199794769,
-0.01250192429870367,
-0.07916021347045898,
-0.029885273426771164,
-0.18612512946128845,
-0.0030564051121473312,
0.10038342326879501,
0.0035033065360039473,
-0.005652366206049919,
-0.08666291832923889,
0.026358824223279953,
-0.03112892620265484,
-0.008404186926782131,
-0.16764774918556213,
-0.04399421438574791,
0.046902090311050415,
-0.16094985604286194,
0.020117372274398804,
-0.06413903087377548,
0.06334125250577927,
0.03641495108604431,
-0.05590536445379257,
-0.0248766727745533,
-0.01730942726135254,
0.011945613659918308,
-0.05083848536014557,
-0.18994836509227753,
-0.056277405470609665,
-0.037882111966609955,
0.149809330701828,
-0.25956398248672485,
0.032966937869787216,
0.051140617579221725,
0.14649195969104767,
0.00406361510977149,
-0.05115427449345589,
0.01429014839231968,
-0.05360214412212372,
-0.054652128368616104,
-0.06746816635131836,
-0.006135428790003061,
-0.027576493099331856,
-0.05147203803062439,
0.019243421033024788,
-0.1755700707435608,
-0.021410830318927765,
0.09424154460430145,
0.12876708805561066,
-0.1486445665359497,
-0.018640631809830666,
-0.048725154250860214,
-0.06339836865663528,
-0.0715010017156601,
-0.07038594037294388,
0.10712739825248718,
0.0513901449739933,
0.04796046018600464,
-0.07435787469148636,
-0.07092321664094925,
0.02726263552904129,
0.006906150374561548,
-0.03382374346256256,
0.08727246522903442,
0.05199531093239784,
-0.09209315478801727,
0.0756213590502739,
0.1092359870672226,
0.07177663594484329,
0.09363535046577454,
0.01574566215276718,
-0.11756632477045059,
-0.028492970392107964,
0.036266472190618515,
0.02740776725113392,
0.1465986967086792,
-0.05952361226081848,
0.04016614332795143,
0.04494241625070572,
-0.04170418903231621,
0.022319864481687546,
-0.08787637203931808,
0.024075502529740334,
0.025203049182891846,
-0.0034381982404738665,
0.06284574419260025,
-0.02525499276816845,
-0.0050758360885083675,
0.07016654312610626,
0.047779910266399384,
0.04621000960469246,
0.009655474685132504,
-0.01720241829752922,
-0.1047825813293457,
0.16950392723083496,
-0.0951867327094078,
-0.269941508769989,
-0.17632324993610382,
0.026197833940386772,
0.04035249724984169,
-0.022378476336598396,
0.031619444489479065,
-0.07056326419115067,
-0.10630585998296738,
-0.1060405746102333,
-0.002429972169920802,
0.01714223250746727,
-0.06364088505506516,
-0.0741225928068161,
0.07348573952913284,
0.04382912442088127,
-0.14902326464653015,
0.038552410900592804,
0.055694397538900375,
-0.057955220341682434,
-0.0233661737293005,
0.09118817001581192,
0.12397737801074982,
0.14583967626094818,
-0.021366750821471214,
-0.028626007959246635,
0.029004426673054695,
0.19620531797409058,
-0.13469526171684265,
0.10371150821447372,
0.13814030587673187,
-0.04545360431075096,
0.08360563963651657,
0.1560150384902954,
0.029186224564909935,
-0.08317049592733383,
0.05044832453131676,
0.04082648828625679,
-0.043159641325473785,
-0.2666129767894745,
-0.0534592866897583,
0.012832709588110447,
-0.06255637854337692,
0.09786593168973923,
0.10183793306350708,
0.11542957276105881,
0.034910861402750015,
-0.07166364789009094,
-0.043925940990448,
-0.0058974819257855415,
0.11737963557243347,
-0.05490213260054588,
-0.012639665976166725,
0.07686592638492584,
-0.05086168646812439,
0.005355054512619972,
0.10266812145709991,
0.02973790094256401,
0.17442677915096283,
0.020399179309606552,
0.11231429129838943,
0.06195578724145889,
0.08633565157651901,
0.0007386076031252742,
0.02951662428677082,
0.05147615820169449,
0.017203815281391144,
-0.002300140680745244,
-0.10421168059110641,
-0.006156572140753269,
0.1449710875749588,
0.028103826567530632,
0.029669636860489845,
-0.0018948549404740334,
-0.005003341939300299,
0.05121048167347908,
0.1746254414319992,
-0.011592294089496136,
-0.22072425484657288,
-0.0845772922039032,
0.06936841458082199,
-0.06218599155545235,
-0.12968985736370087,
-0.026130788028240204,
0.045467354357242584,
-0.17519839107990265,
0.026703642681241035,
-0.027433741837739944,
0.0919293761253357,
-0.09345759451389313,
-0.02221956104040146,
0.03687324374914169,
0.084866963326931,
-0.014529162086546421,
0.08703910559415817,
-0.14498743414878845,
0.11886418610811234,
0.02978132851421833,
0.09024628251791,
-0.11081171780824661,
0.07909037172794342,
-0.007550720125436783,
0.009180475026369095,
0.19379350543022156,
-0.011335089802742004,
-0.03514958545565605,
-0.08774717897176743,
-0.11210042238235474,
-0.013537433929741383,
0.12687496840953827,
-0.1243172138929367,
0.08773399889469147,
-0.015198243781924248,
-0.044079482555389404,
0.00937260314822197,
-0.12100647389888763,
-0.17273177206516266,
-0.19628387689590454,
0.05585884302854538,
-0.09575839340686798,
0.025643249973654747,
-0.11914430558681488,
-0.07089093327522278,
-0.02952558360993862,
0.241120383143425,
-0.1745356321334839,
-0.06510113179683685,
-0.1468164622783661,
-0.046294767409563065,
0.1662203073501587,
-0.04437198117375374,
0.0718095526099205,
-0.0208172257989645,
0.20345525443553925,
0.005988610442727804,
-0.004939318168908358,
0.06724198162555695,
-0.08892562240362167,
-0.16873881220817566,
-0.06771010160446167,
0.1510489284992218,
0.11680185794830322,
0.04907919466495514,
-0.002248800592496991,
0.0011772146681323647,
-0.016943959519267082,
-0.1137804463505745,
-0.0033210667315870523,
0.16037839651107788,
0.03878779336810112,
0.025986969470977783,
-0.05243593826889992,
-0.08797456324100494,
-0.06899320334196091,
-0.06853509694337845,
0.06221301481127739,
0.19590823352336884,
-0.10376439243555069,
0.1700313836336136,
0.147536963224411,
-0.07305635511875153,
-0.23175598680973053,
0.035342130810022354,
0.04983805492520332,
0.0014306638622656465,
0.04886869341135025,
-0.18252557516098022,
0.10521943867206573,
0.019543392583727837,
-0.05505957826972008,
0.13485197722911835,
-0.1557481735944748,
-0.1552847921848297,
0.0722852572798729,
0.03904085233807564,
-0.22423844039440155,
-0.1354004591703415,
-0.09622503817081451,
-0.05825018882751465,
-0.14065024256706238,
0.06054598465561867,
-0.002136280992999673,
0.015948504209518433,
0.03500790148973465,
-0.0015643214574083686,
0.027123261243104935,
-0.058935679495334625,
0.18609118461608887,
-0.004065449349582195,
0.020676052197813988,
-0.060264769941568375,
-0.0478842556476593,
0.09839435666799545,
-0.06130504235625267,
0.12208222597837448,
0.004057085141539574,
0.01594383642077446,
-0.10362856835126877,
-0.048314861953258514,
-0.04328322783112526,
0.05154227837920189,
-0.07548051327466965,
-0.10070807486772537,
-0.043625857681035995,
0.08841723203659058,
0.07005169242620468,
-0.03383097052574158,
0.00549331633374095,
-0.07189501076936722,
0.10019614547491074,
0.17795267701148987,
0.17573626339435577,
0.009926567785441875,
-0.07241068035364151,
0.01677953451871872,
-0.04142116755247116,
0.044231921434402466,
-0.2513144314289093,
0.03756171092391014,
0.06098250672221184,
0.029438555240631104,
0.09217222779989243,
-0.020435843616724014,
-0.1820858269929886,
-0.04050002992153168,
0.08094815909862518,
-0.05452597141265869,
-0.22617179155349731,
-0.019085140898823738,
0.0954197570681572,
-0.2020406424999237,
-0.007372708059847355,
0.03995226323604584,
-0.048725228756666183,
-0.023169852793216705,
0.00010950004070764408,
0.06317184865474701,
0.002471912419423461,
0.09773622453212738,
0.0735151618719101,
0.09715340286493301,
-0.08337292820215225,
0.10562895983457565,
0.10150538384914398,
-0.09572599828243256,
0.03605884686112404,
0.06754924356937408,
-0.05300498008728027,
-0.043293699622154236,
0.03665391728281975,
0.033023297786712646,
0.005234600510448217,
-0.060321882367134094,
0.013913018628954887,
-0.036497246474027634,
0.044923391193151474,
0.08326134830713272,
0.03754979372024536,
-0.013354414142668247,
0.06462216377258301,
0.03401726484298706,
-0.10898099094629288,
0.10366570204496384,
0.01731540448963642,
0.04105307161808014,
-0.08384523540735245,
-0.019968897104263306,
0.035425446927547455,
0.030576206743717194,
-0.01765924133360386,
-0.02306121215224266,
-0.02860277332365513,
-0.01614218018949032,
-0.14299540221691132,
-0.023106401786208153,
-0.07243485748767853,
0.006181265693157911,
0.014656842686235905,
-0.031884219497442245,
-0.011233693920075893,
0.02475680410861969,
-0.06979699432849884,
-0.07426341623067856,
-0.006949664559215307,
0.09833318740129471,
-0.15115703642368317,
0.008848577737808228,
0.06907843053340912,
-0.11088496446609497,
0.08190931379795074,
-0.008411259390413761,
0.016245156526565552,
0.022527478635311127,
-0.15448406338691711,
0.05601610988378525,
0.0008648968650959432,
0.01916889287531376,
0.025886621326208115,
-0.16471809148788452,
0.004104440100491047,
-0.04661374166607857,
-0.02149827405810356,
-0.00004464812809601426,
-0.02647159807384014,
-0.12325995415449142,
0.06858719140291214,
-0.015622655861079693,
-0.035931166261434555,
-0.02701525390148163,
0.0539589487016201,
0.07888586074113846,
-0.027474910020828247,
0.10445091128349304,
-0.008690856397151947,
0.04941811040043831,
-0.16801609098911285,
-0.02470702864229679,
-0.04982255399227142,
0.019377702847123146,
0.009884213097393513,
-0.007693959400057793,
0.04183054715394974,
-0.00976533442735672,
0.21883612871170044,
-0.05075952783226967,
0.1607085019350052,
0.05847611650824547,
-0.017352959141135216,
-0.0007513365126214921,
0.06180921941995621,
0.05997028574347496,
0.04658793285489082,
0.009480604901909828,
0.023740366101264954,
-0.022450892254710197,
-0.006695089396089315,
-0.15932634472846985,
0.01890849508345127,
0.14999441802501678,
0.06301083415746689,
0.024745315313339233,
0.05866100639104843,
-0.12775006890296936,
-0.12135478109121323,
0.09311001747846603,
-0.026755332946777344,
0.00928465835750103,
-0.08245618641376495,
0.1358020007610321,
0.14980104565620422,
-0.14000412821769714,
0.05256148427724838,
-0.06134212389588356,
-0.05217423290014267,
-0.10388828068971634,
-0.12032219022512436,
-0.05887215584516525,
-0.053666237741708755,
0.002330566756427288,
-0.03760887682437897,
0.054546963423490524,
0.03344334661960602,
-0.009351172484457493,
-0.00022941511997487396,
0.13597318530082703,
-0.019751882180571556,
-0.0028988157864660025,
0.048313532024621964,
0.03693558648228645,
0.02373051457107067,
-0.05275435373187065,
0.02940409444272518,
0.02539868652820587,
0.032232340425252914,
0.06546790152788162,
0.033412106335163116,
-0.047448933124542236,
0.03804153576493263,
-0.0025254099164158106,
-0.11207924783229828,
0.019641218706965446,
-0.00460948096588254,
-0.0742158442735672,
0.1268945336341858,
0.0407399944961071,
0.010224059224128723,
-0.03741471841931343,
0.24361543357372284,
-0.06653323769569397,
-0.06378097087144852,
-0.13251738250255585,
0.10491154342889786,
-0.0027236645109951496,
0.06476365029811859,
0.023412218317389488,
-0.1284150779247284,
0.005243356805294752,
0.13858191668987274,
0.12181595712900162,
0.0045748427510261536,
0.009228081442415714,
0.0518609918653965,
0.0025186820421367884,
-0.06998204439878464,
0.054019294679164886,
0.06992026418447495,
0.12919506430625916,
-0.07847554981708527,
0.07680778950452805,
0.0006860480643808842,
-0.08370215445756912,
-0.02947772853076458,
0.11312682181596756,
-0.0409729965031147,
0.03491825982928276,
-0.047444481402635574,
0.10916327685117722,
-0.05787910893559456,
-0.29412412643432617,
0.02350960113108158,
-0.09588567912578583,
-0.15202060341835022,
-0.018367812037467957,
0.05944539234042168,
-0.02624768204987049,
0.018029648810625076,
0.06971040368080139,
-0.06011629104614258,
0.20098382234573364,
0.0335683599114418,
-0.07864278554916382,
-0.0664360448718071,
0.04837050288915634,
-0.06564252078533173,
0.2949807047843933,
0.008418165147304535,
0.02863333560526371,
0.10770907253026962,
-0.03253700211644173,
-0.18271861970424652,
0.010723991319537163,
0.1133992001414299,
-0.08056149631738663,
0.08200647681951523,
0.19000613689422607,
-0.012578671798110008,
0.1209007054567337,
0.05294662341475487,
-0.047376248985528946,
0.04217283055186272,
-0.03389401361346245,
-0.051268599927425385,
-0.10752558708190918,
0.058453381061553955,
-0.05909625440835953,
0.15447644889354706,
0.10152646154165268,
-0.05671518296003342,
-0.004550917539745569,
-0.05555408447980881,
0.04875178262591362,
0.01804669201374054,
0.12263146042823792,
0.02951994352042675,
-0.1865430772304535,
0.032826557755470276,
-0.01144319772720337,
0.10186848044395447,
-0.25588861107826233,
-0.08421015739440918,
0.08833149075508118,
-0.011924264021217823,
-0.05105875805020332,
0.10560628771781921,
0.057650718837976456,
0.04243382066488266,
-0.043439045548439026,
-0.10480839014053345,
-0.02186836116015911,
0.14663739502429962,
-0.1469624787569046,
-0.025013303384184837
] |
null | null | transformers |
# **deplot_kr**
deplot_kr is a Image-to-Data(Text) model based on the google's pix2struct architecture.
It was fine-tuned from [DePlot](https://huggingface.co/google/deplot), using korean chart image-text pairs.
deplot_kr은 google의 pix2struct 구조를 기반으로 한 한국어 image-to-data(텍스트 형태의 데이터 테이블) 모델입니다.
[DePlot](https://huggingface.co/google/deplot) 모델을 한국어 차트 이미지-텍스트 쌍 데이터세트(30만 개)를 이용하여 fine-tuning 했습니다.
## **How to use**
You can run a prediction by input an image.
Model predict the data table of text form in the image.
이미지를 모델에 입력하면 모델은 이미지로부터 표 형태의 데이터 테이블을 예측합니다.
```python
from transformers import Pix2StructForConditionalGeneration, Pix2StructImageProcessor, AutoTokenizer, Pix2StructProcessor
from PIL import Image
image_processor = Pix2StructImageProcessor()
tokenizer = AutoTokenizer.from_pretrained("brainventures/deplot_kr")
processor = Pix2StructProcessor(image_processor=image_processor, tokenizer=tokenizer)
model = Pix2StructForConditionalGeneration.from_pretrained("brainventures/deplot_kr")
image_path = "IMAGE_PATH"
image = Image.open(image_path)
inputs = processor(images=image, return_tensors="pt")
pred = model.generate(flattened_patches=flattened_patches, attention_mask=attention_mask, max_length=1024)
print(processor.batch_decode(deplot_generated_ids, skip_special_token=True)[0])
```
**Model Input Image**

**Model Output - Prediction**
대상:
제목: 2011-2021 보건복지 분야 일자리의 <unk>증
유형: 단일형 일반 세로 <unk>대형
| 보건(천 명) | 복지(천 명)
1분위 | 29.7 | 178.4
2분위 | 70.8 | 97.3
3분위 | 86.4 | 61.3
4분위 | 28.2 | 16.0
5분위 | 52.3 | 0.9
### **Preprocessing**
According to [Liu et al.(2023)](https://arxiv.org/pdf/2212.10505.pdf)...
- markdown format
- | : seperating cells (열 구분)
- \n : seperating rows (행 구분)
### **Train**
The model was trained in a TPU environment.
- num_warmup_steps : 1,000
- num_training_steps : 40,000
## **Evaluation Results**
This model achieves the following results:
|metrics name | % |
|:---|---:|
| RNSS (Relative Number Set Similarity)| 99.5483 |
| RMS F1 (Relative Mapping Similarity)| 16.6401 |
## Contact
For questions and comments, please use the discussion tab or email [email protected] | {"language": ["ko"], "pipeline_tag": "image-to-text"} | image-to-text | brainventures/deplot_kr | [
"transformers",
"pytorch",
"pix2struct",
"text2text-generation",
"image-to-text",
"ko",
"arxiv:2212.10505",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T06:45:49+00:00 | [
"2212.10505"
] | [
"ko"
] | TAGS
#transformers #pytorch #pix2struct #text2text-generation #image-to-text #ko #arxiv-2212.10505 #autotrain_compatible #endpoints_compatible #region-us
| deplot\_kr
==========
deplot\_kr is a Image-to-Data(Text) model based on the google's pix2struct architecture.
It was fine-tuned from DePlot, using korean chart image-text pairs.
deplot\_kr은 google의 pix2struct 구조를 기반으로 한 한국어 image-to-data(텍스트 형태의 데이터 테이블) 모델입니다.
DePlot 모델을 한국어 차트 이미지-텍스트 쌍 데이터세트(30만 개)를 이용하여 fine-tuning 했습니다.
How to use
----------
You can run a prediction by input an image.
Model predict the data table of text form in the image.
이미지를 모델에 입력하면 모델은 이미지로부터 표 형태의 데이터 테이블을 예측합니다.
Model Input Image
!model\_input\_image
Model Output - Prediction
대상:
제목: 2011-2021 보건복지 분야 일자리의 증
유형: 단일형 일반 세로 대형
| 보건(천 명) | 복지(천 명)
1분위 | 29.7 | 178.4
2분위 | 70.8 | 97.3
3분위 | 86.4 | 61.3
4분위 | 28.2 | 16.0
5분위 | 52.3 | 0.9
### Preprocessing
According to Liu et al.(2023)...
* markdown format
* | : seperating cells (열 구분)
* \n : seperating rows (행 구분)
### Train
The model was trained in a TPU environment.
* num\_warmup\_steps : 1,000
* num\_training\_steps : 40,000
Evaluation Results
------------------
This model achieves the following results:
Contact
-------
For questions and comments, please use the discussion tab or email gloria@URL
| [
"### Preprocessing\n\n\nAccording to Liu et al.(2023)...\n\n\n* markdown format\n* | : seperating cells (열 구분)\n* \\n : seperating rows (행 구분)",
"### Train\n\n\nThe model was trained in a TPU environment.\n\n\n* num\\_warmup\\_steps : 1,000\n* num\\_training\\_steps : 40,000\n\n\nEvaluation Results\n------------------\n\n\nThis model achieves the following results:\n\n\n\nContact\n-------\n\n\nFor questions and comments, please use the discussion tab or email gloria@URL"
] | [
"TAGS\n#transformers #pytorch #pix2struct #text2text-generation #image-to-text #ko #arxiv-2212.10505 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Preprocessing\n\n\nAccording to Liu et al.(2023)...\n\n\n* markdown format\n* | : seperating cells (열 구분)\n* \\n : seperating rows (행 구분)",
"### Train\n\n\nThe model was trained in a TPU environment.\n\n\n* num\\_warmup\\_steps : 1,000\n* num\\_training\\_steps : 40,000\n\n\nEvaluation Results\n------------------\n\n\nThis model achieves the following results:\n\n\n\nContact\n-------\n\n\nFor questions and comments, please use the discussion tab or email gloria@URL"
] | [
57,
48,
69
] | [
"passage: TAGS\n#transformers #pytorch #pix2struct #text2text-generation #image-to-text #ko #arxiv-2212.10505 #autotrain_compatible #endpoints_compatible #region-us \n### Preprocessing\n\n\nAccording to Liu et al.(2023)...\n\n\n* markdown format\n* | : seperating cells (열 구분)\n* \\n : seperating rows (행 구분)### Train\n\n\nThe model was trained in a TPU environment.\n\n\n* num\\_warmup\\_steps : 1,000\n* num\\_training\\_steps : 40,000\n\n\nEvaluation Results\n------------------\n\n\nThis model achieves the following results:\n\n\n\nContact\n-------\n\n\nFor questions and comments, please use the discussion tab or email gloria@URL"
] | [
-0.08903109282255173,
0.03867683932185173,
-0.000846041482873261,
0.05244404822587967,
0.12855036556720734,
0.014295900240540504,
0.1558408886194229,
0.0880587175488472,
-0.03728073090314865,
0.035592954605817795,
0.09089095145463943,
0.14053739607334137,
0.0662025511264801,
0.2562636435031891,
-0.025112351402640343,
-0.3282027542591095,
0.02587413787841797,
0.08410262316465378,
-0.03899484872817993,
0.1532602459192276,
0.08845152705907822,
-0.08173644542694092,
0.0841655284166336,
0.015671506524086,
-0.16499435901641846,
-0.047910090535879135,
-0.044133469462394714,
-0.04934290796518326,
0.157749742269516,
0.0511932373046875,
0.06464281678199768,
0.09878268837928772,
0.028157681226730347,
-0.11055216938257217,
0.04040830582380295,
0.0012987196678295732,
-0.015281693078577518,
0.10155629366636276,
0.11449647694826126,
0.03387279063463211,
0.265514612197876,
-0.0015947516076266766,
-0.007100292015820742,
0.015368564054369926,
-0.07358807325363159,
-0.07976842671632767,
-0.06282651424407959,
0.14479659497737885,
0.2031291127204895,
0.04000508412718773,
-0.025977618992328644,
0.15639030933380127,
-0.03983774408698082,
0.09905216842889786,
0.10292056202888489,
-0.31437623500823975,
-0.07128331065177917,
0.06277015060186386,
0.03930741548538208,
0.01980438269674778,
0.0016104134265333414,
0.05317003279924393,
0.07166795432567596,
-0.01919744722545147,
-0.08741595596075058,
-0.02819328010082245,
-0.00025889172684401274,
-0.025966094806790352,
-0.14676429331302643,
-0.038990046828985214,
0.22116254270076752,
0.026699915528297424,
-0.04000508412718773,
-0.05412326008081436,
-0.06812155246734619,
-0.10228212922811508,
-0.03412443771958351,
-0.1108960211277008,
-0.02882741205394268,
-0.04760286957025528,
-0.05493556708097458,
0.0009661411750130355,
-0.10342089086771011,
-0.10202280431985855,
-0.07270660251379013,
0.06009891629219055,
0.04543999955058098,
-0.02145438827574253,
-0.0850006490945816,
0.13977834582328796,
0.058838021010160446,
-0.08185020834207535,
0.05290872976183891,
-0.07644525915384293,
-0.010710673406720161,
-0.025969600304961205,
-0.024812860414385796,
-0.07468022406101227,
-0.01468328945338726,
0.026025807484984398,
0.055045198649168015,
0.031138794496655464,
0.07216981798410416,
0.0436096265912056,
-0.015023892745375633,
0.03328052908182144,
-0.10087773203849792,
-0.04930586367845535,
-0.03588521480560303,
0.0031883385963737965,
-0.05085593834519386,
-0.007641327101737261,
-0.12731993198394775,
-0.06066887453198433,
0.11657169461250305,
0.00888794381171465,
-0.022547047585248947,
0.10690121352672577,
-0.03661579266190529,
-0.0680680125951767,
0.046542249619960785,
-0.04401373118162155,
-0.030907005071640015,
-0.046026892960071564,
-0.06980933994054794,
0.11737857758998871,
0.053126320242881775,
-0.008226449601352215,
-0.07010377198457718,
0.061292730271816254,
-0.09642619639635086,
-0.002455010311678052,
-0.07757015526294708,
-0.04840555042028427,
-0.008202006109058857,
-0.06879077851772308,
-0.023364322260022163,
-0.14451755583286285,
-0.044381264597177505,
-0.023923322558403015,
0.07181314378976822,
-0.05648596212267876,
-0.057338498532772064,
-0.03537946194410324,
-0.004938892554491758,
-0.008465321734547615,
-0.021686652675271034,
0.07839062064886093,
-0.00496641406789422,
0.1074889749288559,
0.0426904670894146,
0.14191769063472748,
0.002875476609915495,
0.03688877075910568,
-0.04211260750889778,
0.037432655692100525,
-0.16326352953910828,
0.03764956444501877,
-0.0016824689228087664,
0.04439053684473038,
-0.11631205677986145,
-0.14457058906555176,
-0.023869508877396584,
-0.018429860472679138,
0.08134786784648895,
0.18873156607151031,
-0.09731710702180862,
-0.04814222455024719,
0.10434602946043015,
-0.11128804832696915,
-0.1477769911289215,
0.1243937760591507,
-0.002398527693003416,
0.10458750277757645,
0.031825751066207886,
0.14889217913150787,
-0.01598542369902134,
-0.031006066128611565,
0.010978671722114086,
0.03376944735646248,
-0.06170723959803581,
-0.08644760400056839,
0.08473419398069382,
0.05939468741416931,
-0.15112845599651337,
0.0327661857008934,
-0.002762257121503353,
0.041468653827905655,
-0.14809629321098328,
-0.06941839307546616,
0.014983958564698696,
-0.05313403531908989,
0.09073714166879654,
0.02462722547352314,
0.1149037703871727,
-0.010577846318483353,
-0.052096664905548096,
0.0012439251877367496,
0.10026521980762482,
-0.0616837739944458,
-0.009809938259422779,
-0.11004114151000977,
0.08752993494272232,
-0.07093970477581024,
0.021676817908883095,
-0.15101595222949982,
0.01981259137392044,
-0.020706484094262123,
0.03363027051091194,
0.031165314838290215,
0.11152574419975281,
0.030251238495111465,
0.05183035507798195,
-0.008787604980170727,
0.006087781395763159,
0.02492203749716282,
-0.016010725870728493,
-0.13429000973701477,
-0.10967656224966049,
-0.05908329784870148,
-0.021252617239952087,
0.145167276263237,
-0.1902662217617035,
0.0025197225622832775,
-0.018580548465251923,
0.0887317955493927,
-0.04934391379356384,
0.014627608470618725,
0.002209074329584837,
0.026300743222236633,
-0.07139867544174194,
-0.010683372616767883,
0.12469886243343353,
-0.0072633675299584866,
-0.04714943841099739,
0.04581385850906372,
-0.10782113671302795,
0.07543666660785675,
0.10465855896472931,
-0.0850507989525795,
-0.13074231147766113,
-0.07161713391542435,
-0.058046258985996246,
-0.02343318611383438,
0.013385401107370853,
0.06195100024342537,
0.08590537309646606,
0.0035959670785814524,
0.14090311527252197,
-0.07114604115486145,
-0.0011793762678280473,
0.04081093147397041,
-0.06267552822828293,
-0.005077903158962727,
0.11209738254547119,
0.1323891580104828,
-0.1722961962223053,
0.07246086746454239,
-0.039771195501089096,
-0.05492827296257019,
0.08936966210603714,
0.07867080718278885,
-0.06652625650167465,
-0.028589393943548203,
0.042590927332639694,
0.00803066324442625,
0.08984831720590591,
-0.16881878674030304,
0.0022216173820197582,
0.08290430158376694,
-0.02773348242044449,
0.07212898135185242,
-0.16878993809223175,
-0.028418026864528656,
0.021949512884020805,
-0.02269141748547554,
-0.03467019274830818,
0.08198452740907669,
-0.05863402038812637,
0.14141350984573364,
0.001465280307456851,
-0.049358922988176346,
0.0024987005162984133,
0.029008328914642334,
-0.11279217898845673,
0.2062295824289322,
0.013956121169030666,
-0.29299262166023254,
-0.1203874722123146,
-0.045432548969984055,
-0.03621964529156685,
0.0431714802980423,
0.03485381230711937,
-0.09721391648054123,
-0.02709950879216194,
-0.020929254591464996,
-0.03180330619215965,
-0.028606094419956207,
0.05027809366583824,
-0.0028006071224808693,
0.024358345195651054,
0.029282081872224808,
-0.01098104938864708,
-0.0067099458537995815,
-0.06481434404850006,
-0.02327442727982998,
0.10694873332977295,
-0.12127912044525146,
0.11288478225469589,
0.18474197387695312,
-0.04485723376274109,
0.07144119590520859,
-0.017734184861183167,
0.13962717354297638,
-0.0795425996184349,
0.020225340500473976,
0.2385944426059723,
0.013618415221571922,
0.03792249783873558,
0.10701458901166916,
-0.02531086653470993,
-0.06835807859897614,
0.07868871092796326,
-0.036119624972343445,
-0.06754487007856369,
-0.20307637751102448,
-0.10558408498764038,
-0.0432576946914196,
0.06715064495801926,
0.08318885415792465,
0.04628964886069298,
0.08087456971406937,
0.15581126511096954,
0.06235251575708389,
0.05229213833808899,
-0.04828120023012161,
0.09641171991825104,
0.0786452665925026,
0.009358768351376057,
0.11666008830070496,
-0.039227403700351715,
-0.09967382252216339,
0.06805623322725296,
-0.014676549471914768,
0.1758105754852295,
-0.008030308410525322,
-0.01989906281232834,
0.05559132993221283,
0.15069977939128876,
0.11800012737512589,
0.15642164647579193,
0.004613957833498716,
-0.04193210229277611,
-0.0466277115046978,
-0.03846890479326248,
-0.09282753616571426,
0.028382018208503723,
0.03936154022812843,
-0.09777683019638062,
-0.12808336317539215,
0.0814131423830986,
0.014812701381742954,
0.08845273405313492,
0.09739164263010025,
-0.3347470462322235,
-0.09046032279729843,
-0.04443836212158203,
-0.010010475292801857,
-0.08882640302181244,
-0.004347512032836676,
0.09365332126617432,
-0.13449829816818237,
-0.0022608365397900343,
-0.07979431748390198,
0.12388131022453308,
-0.0845174491405487,
0.01572735607624054,
-0.017882300540804863,
-0.014432382769882679,
-0.004899362567812204,
0.1203586757183075,
-0.22395232319831848,
0.29666781425476074,
-0.05889074504375458,
-0.016222495585680008,
-0.06446265429258347,
-0.03447782248258591,
-0.010556857101619244,
0.1275399774312973,
0.14130842685699463,
-0.0023729486856609583,
0.030239775776863098,
-0.14149723947048187,
-0.09370870888233185,
0.043968912214040756,
0.08781121671199799,
-0.01992802508175373,
0.07473272830247879,
-0.013501188717782497,
0.011398444883525372,
-0.016398359090089798,
0.02424808032810688,
-0.08579737693071365,
-0.07430195808410645,
0.06923726946115494,
-0.11588697135448456,
0.09886161983013153,
-0.0010831748368218541,
-0.10568113625049591,
-0.06450722366571426,
0.21183758974075317,
0.054843470454216,
-0.040253136307001114,
-0.14158441126346588,
0.04440559819340706,
0.11507266014814377,
-0.09890565276145935,
0.020634034648537636,
-0.03351140767335892,
0.02914595417678356,
0.012727909721434116,
-0.08634420484304428,
0.08202821016311646,
-0.09709324687719345,
-0.0947863832116127,
0.008826230652630329,
0.0834227055311203,
-0.010371783748269081,
0.02942369133234024,
0.039152178913354874,
-0.05639436095952988,
-0.08110606670379639,
-0.11387544870376587,
0.022115156054496765,
-0.023441487923264503,
0.12692730128765106,
-0.039328381419181824,
-0.07233591377735138,
0.06139209866523743,
-0.09959925711154938,
-0.03490874171257019,
0.17253470420837402,
0.23753385245800018,
-0.053671423345804214,
0.025584479793906212,
0.11606817692518234,
-0.024192731827497482,
-0.20923542976379395,
-0.08028129488229752,
0.07207582890987396,
0.048301342874765396,
-0.027517160400748253,
-0.16502828896045685,
0.04401695728302002,
0.05619249865412712,
-0.012256009504199028,
0.016374090686440468,
-0.23942813277244568,
-0.14079827070236206,
0.137544646859169,
0.10799536108970642,
0.21054261922836304,
-0.09889402985572815,
-0.022296668961644173,
-0.07241719961166382,
-0.15360970795154572,
0.10469435900449753,
0.08338414132595062,
0.12175755202770233,
-0.010033862665295601,
0.07531888782978058,
0.018338853493332863,
-0.07389351725578308,
0.14257363975048065,
0.04351334646344185,
0.021441597491502762,
-0.09468182921409607,
-0.1247747465968132,
-0.06705592572689056,
-0.016737401485443115,
0.10501499474048615,
0.04660385847091675,
0.0792149230837822,
-0.17201702296733856,
-0.0792778953909874,
-0.0502755232155323,
0.01599065773189068,
0.015528757125139236,
-0.052776288241147995,
-0.06043101102113724,
0.06820159405469894,
0.026224039494991302,
0.012847336940467358,
0.027648422867059708,
-0.05246131122112274,
0.059724289923906326,
0.12147028744220734,
0.13367822766304016,
-0.030338479205965996,
-0.007499245461076498,
-0.016938023269176483,
-0.005030655302107334,
0.05938596650958061,
-0.06346023827791214,
-0.0007844003848731518,
0.15190677344799042,
-0.01981765776872635,
0.13714250922203064,
0.07344336807727814,
0.008964341133832932,
-0.0019792448729276657,
0.11887354403734207,
-0.18169918656349182,
-0.15056706964969635,
-0.038631316274404526,
-0.001055898261256516,
-0.05968472734093666,
0.0946199968457222,
0.08277592062950134,
-0.08436920493841171,
-0.020821239799261093,
0.01389138214290142,
0.0018542986363172531,
-0.06312637776136398,
0.1685442328453064,
0.03748636320233345,
0.0649927482008934,
-0.07151511311531067,
-0.005767190363258123,
0.021902818232774734,
-0.10072227567434311,
-0.022310929372906685,
0.04226010665297508,
-0.15979474782943726,
-0.07017692923545837,
0.054624076932668686,
0.05788044631481171,
-0.09804052114486694,
-0.025685014203190804,
-0.10192974656820297,
-0.05830098316073418,
0.043052054941654205,
0.046396736055612564,
0.09790582209825516,
0.01972186379134655,
-0.08042203634977341,
-0.014038454741239548,
-0.13934966921806335,
0.03016759268939495,
0.10364391654729843,
0.051723308861255646,
-0.12669062614440918,
0.245949387550354,
-0.0031911544501781464,
0.12006976455450058,
-0.08771207183599472,
-0.028717458248138428,
-0.11786726117134094,
0.01862410642206669,
-0.019412148743867874,
-0.05387958884239197,
-0.13064022362232208,
-0.06285548955202103,
-0.0019330846844241023,
-0.02457953616976738,
-0.03263918682932854,
-0.016443518921732903,
-0.10043373703956604,
0.023015258833765984,
-0.02450893633067608,
0.004238842520862818,
-0.021238340064883232,
0.05198810622096062,
0.06096111983060837,
-0.04405016079545021,
0.1124539002776146,
0.06445643305778503,
-0.025057358667254448,
-0.025210700929164886,
-0.06149527430534363,
0.008655206300318241,
0.03269274905323982,
0.03333442658185959,
0.03905060142278671,
-0.09161420911550522,
0.04430992156267166,
0.01804514229297638,
0.04704194515943527,
0.024851812049746513,
0.07830671221017838,
-0.09847261756658554,
-0.009512178599834442,
-0.12221627682447433,
-0.043777644634246826,
-0.07289015501737595,
0.04656495898962021,
-0.0190894678235054,
0.13480806350708008,
0.11694876104593277,
-0.04768085852265358,
0.03775671869516373,
-0.12217573076486588,
-0.006770072504878044,
0.006953290663659573,
-0.060436952859163284,
0.041980843991041183,
-0.10929817706346512,
0.06785605847835541,
-0.002970830537378788,
0.07330208271741867,
0.052825748920440674,
-0.017459355294704437,
-0.021055128425359726,
0.028450462967157364,
0.030594389885663986,
-0.0257712509483099,
0.2157057225704193,
0.05703524500131607,
0.040610358119010925,
0.0017932781483978033,
0.09633707255125046,
0.0030130059458315372,
0.050444576889276505,
0.24172110855579376,
0.042278438806533813,
-0.09012491255998611,
0.07148327678442001,
-0.04303843155503273,
-0.008940041996538639,
-0.16588683426380157,
-0.057313594967126846,
-0.11153478920459747,
0.05062694847583771,
-0.04650826007127762,
0.1165405660867691,
0.15295925736427307,
-0.11077387630939484,
0.04475748911499977,
0.028237035498023033,
-0.1068151667714119,
-0.11729015409946442,
-0.1967669278383255,
-0.10162894427776337,
-0.18571984767913818,
-0.003385548945516348,
-0.09289256483316422,
-0.0245071891695261,
0.1201530247926712,
0.07376319915056229,
-0.04844904690980911,
0.13832040131092072,
-0.053206343203783035,
-0.04240843653678894,
0.058911796659231186,
-0.026247406378388405,
0.0052270605228841305,
-0.06887722760438919,
-0.023027075454592705,
-0.019502414390444756,
0.02064436487853527,
0.07401535660028458,
-0.04150805249810219,
-0.060677070170640945,
0.030598916113376617,
0.02754197083413601,
-0.050142936408519745,
-0.05080483853816986,
0.030056554824113846,
0.08264986425638199,
0.09629927575588226,
-0.008213914930820465,
0.002216183580458164,
-0.04911353066563606,
0.156930074095726,
-0.03845135495066643,
-0.07894333451986313,
-0.12735934555530548,
0.2830900251865387,
-0.02204371616244316,
0.036972079426050186,
-0.0014046322321519256,
-0.006857461296021938,
-0.02520829811692238,
0.31156647205352783,
0.16356059908866882,
-0.09939707815647125,
-0.045421406626701355,
0.03165151923894882,
0.01664484478533268,
0.04818631708621979,
0.13508693873882294,
0.09334653615951538,
0.20908476412296295,
-0.06316264718770981,
-0.009974135085940361,
-0.1005098819732666,
-0.029995586723089218,
0.0002958191616926342,
0.051069729030132294,
0.0878450870513916,
-0.04891843721270561,
-0.011050512082874775,
0.14453676342964172,
-0.15372246503829956,
0.030625583603978157,
-0.03125656396150589,
-0.06779792159795761,
-0.13785158097743988,
0.03188207745552063,
0.07095348089933395,
-0.0022962153889238834,
0.06944970786571503,
-0.06966035813093185,
0.010795983485877514,
0.08627539128065109,
0.05438163876533508,
-0.09606815129518509,
-0.01699366606771946,
0.09952830523252487,
0.019348399713635445,
0.0618722066283226,
-0.018310723826289177,
0.077669657766819,
0.08056456595659256,
0.012711198069155216,
-0.06101232022047043,
0.0774199366569519,
0.003176923142746091,
-0.05306343734264374,
0.10928645730018616,
0.08138831704854965,
0.01414509117603302,
-0.09513247013092041,
0.008654081262648106,
-0.17875155806541443,
0.004191781394183636,
-0.09098254144191742,
0.06284121423959732,
-0.0794815793633461,
0.06427117437124252,
-0.07350175827741623,
0.13725534081459045,
0.18076622486114502,
-0.05923398211598396,
-0.003723697504028678,
-0.04844889044761658,
0.05903836339712143,
-0.03903832286596298,
0.01522057969123125,
-0.027429470792412758,
-0.20055481791496277,
-0.05752202868461609,
-0.13893839716911316,
-0.020938465371727943,
-0.12259075045585632,
-0.026565009728074074,
-0.062329377979040146,
-0.0616152286529541,
-0.09893079847097397,
0.08136159926652908,
0.09966601431369781,
0.026567790657281876,
-0.047259412705898285,
-0.03009265661239624,
-0.030154507607221603,
0.1263565719127655,
-0.1625586748123169,
-0.09090136736631393
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# videomae-base-finetuned-ucf101-subset
This model is a fine-tuned version of [MCG-NJU/videomae-base](https://huggingface.co/MCG-NJU/videomae-base) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.4111
- Accuracy: 0.8710
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- training_steps: 148
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 2.3295 | 0.13 | 19 | 2.1178 | 0.3 |
| 1.991 | 1.13 | 38 | 1.5624 | 0.5143 |
| 1.227 | 2.13 | 57 | 0.8621 | 0.7714 |
| 0.5665 | 3.13 | 76 | 0.6212 | 0.8 |
| 0.3487 | 4.13 | 95 | 0.3554 | 0.9286 |
| 0.1701 | 5.13 | 114 | 0.3597 | 0.8857 |
| 0.1036 | 6.13 | 133 | 0.2277 | 0.9286 |
| 0.089 | 7.1 | 148 | 0.1989 | 0.9143 |
### Framework versions
- Transformers 4.38.0.dev0
- Pytorch 2.2.0+cu121
- Datasets 2.11.0
- Tokenizers 0.15.1
| {"license": "cc-by-nc-4.0", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "MCG-NJU/videomae-base", "model-index": [{"name": "videomae-base-finetuned-ucf101-subset", "results": []}]} | video-classification | Saini-Manisha/videomae-base-finetuned-ucf101-subset | [
"transformers",
"tensorboard",
"safetensors",
"videomae",
"video-classification",
"generated_from_trainer",
"base_model:MCG-NJU/videomae-base",
"license:cc-by-nc-4.0",
"endpoints_compatible",
"region:us"
] | 2024-02-07T06:59:04+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #videomae #video-classification #generated_from_trainer #base_model-MCG-NJU/videomae-base #license-cc-by-nc-4.0 #endpoints_compatible #region-us
| videomae-base-finetuned-ucf101-subset
=====================================
This model is a fine-tuned version of MCG-NJU/videomae-base on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 0.4111
* Accuracy: 0.8710
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* training\_steps: 148
### Training results
### Framework versions
* Transformers 4.38.0.dev0
* Pytorch 2.2.0+cu121
* Datasets 2.11.0
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 148",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.2.0+cu121\n* Datasets 2.11.0\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #videomae #video-classification #generated_from_trainer #base_model-MCG-NJU/videomae-base #license-cc-by-nc-4.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 148",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.2.0+cu121\n* Datasets 2.11.0\n* Tokenizers 0.15.1"
] | [
69,
115,
4,
38
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #videomae #video-classification #generated_from_trainer #base_model-MCG-NJU/videomae-base #license-cc-by-nc-4.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 148### Training results### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.2.0+cu121\n* Datasets 2.11.0\n* Tokenizers 0.15.1"
] | [
-0.10659623891115189,
0.10453806072473526,
-0.002138470998033881,
0.07598326355218887,
0.1045074611902237,
-0.003080623224377632,
0.14308352768421173,
0.1561223566532135,
-0.12782910466194153,
0.047176338732242584,
0.1385079324245453,
0.1175054982304573,
0.03645259514451027,
0.1809942126274109,
-0.03928786516189575,
-0.2569850981235504,
0.03271397203207016,
0.035292308777570724,
-0.04689547419548035,
0.11613187193870544,
0.09880143404006958,
-0.1558813750743866,
0.08962047845125198,
-0.0029226746410131454,
-0.22355882823467255,
-0.009926815517246723,
0.027797341346740723,
-0.06781097501516342,
0.11280066519975662,
0.01497996598482132,
0.07610711455345154,
0.04688763618469238,
0.11483588069677353,
-0.1990322321653366,
0.012926233001053333,
0.061496980488300323,
-0.0018929133657366037,
0.06852196902036667,
0.06349718570709229,
0.03278119117021561,
0.1058867871761322,
-0.1313723921775818,
0.06998945027589798,
0.027960604056715965,
-0.1281697303056717,
-0.25035789608955383,
-0.09725461900234222,
-0.003012744477018714,
0.09164606779813766,
0.053487036377191544,
-0.008214392699301243,
0.12940388917922974,
-0.040542639791965485,
0.11417101323604584,
0.25101548433303833,
-0.26034438610076904,
-0.07264605164527893,
0.055595699697732925,
0.08617987483739853,
0.07065992802381516,
-0.1248483806848526,
0.022330086678266525,
0.05578337237238884,
0.006826932542026043,
0.13518691062927246,
-0.02090565301477909,
-0.01426601130515337,
-0.026354530826210976,
-0.13083258271217346,
-0.06860015541315079,
0.10821124166250229,
0.06922274827957153,
-0.04070580378174782,
-0.06380521506071091,
-0.029788583517074585,
-0.17520330846309662,
-0.08749116212129593,
0.025325268507003784,
0.057780396193265915,
-0.06553749740123749,
-0.11435065418481827,
0.0030383688863366842,
-0.07884369045495987,
-0.08713909983634949,
-0.002445572754368186,
0.14704398810863495,
0.04148954525589943,
0.037947796285152435,
-0.07433667778968811,
0.08414414525032043,
-0.02178049646317959,
-0.1601027101278305,
-0.01089639961719513,
0.020305251702666283,
-0.005714792292565107,
-0.04322246462106705,
-0.028448205441236496,
-0.07566417753696442,
-0.0034222262911498547,
0.12813794612884521,
-0.11526977270841599,
0.07171458750963211,
-0.010686387307941914,
0.050322163850069046,
-0.08388756960630417,
0.16231919825077057,
-0.051402848213911057,
0.0076899342238903046,
0.012348785065114498,
0.10906367748975754,
0.025464139878749847,
-0.008740793913602829,
-0.10096374899148941,
0.043307553976774216,
0.06643036752939224,
0.032170820981264114,
-0.023337380960583687,
0.06558216363191605,
-0.05207086354494095,
0.001782239880412817,
0.05427417531609535,
-0.09245020896196365,
0.044775839895009995,
-0.0016100216889753938,
-0.06763789057731628,
-0.04603683203458786,
0.004620913416147232,
0.006127037573605776,
0.022059356793761253,
0.07805526256561279,
-0.07296188175678253,
0.03485545516014099,
-0.09493039548397064,
-0.12762512266635895,
0.04874284565448761,
-0.13031472265720367,
0.007465230766683817,
-0.07498116046190262,
-0.09499169886112213,
0.00649592000991106,
0.05013551935553551,
-0.022084852680563927,
-0.00035165148437954485,
-0.06406211107969284,
-0.09400103241205215,
0.03918953984975815,
0.003340622643008828,
0.08740878850221634,
-0.07672083377838135,
0.08753366023302078,
0.03380688279867172,
0.09346561133861542,
-0.025119544938206673,
0.024710452184081078,
-0.045132849365472794,
0.04840078204870224,
-0.23315636813640594,
0.05590327829122543,
-0.08825509250164032,
0.05043448880314827,
-0.09067913889884949,
-0.07740762084722519,
0.06841118633747101,
-0.013019151985645294,
0.04904130473732948,
0.11750738322734833,
-0.23046165704727173,
-0.05932093411684036,
0.1773959994316101,
-0.088056780397892,
-0.14214853942394257,
0.10181652009487152,
-0.04826662689447403,
-0.022415781393647194,
0.03708716481924057,
0.1790039986371994,
0.04895257204771042,
-0.1786939650774002,
0.007153213955461979,
0.0009432249353267252,
0.04119151458144188,
-0.017116248607635498,
0.09007394313812256,
0.043857742100954056,
0.11901914328336716,
-0.022262584418058395,
-0.055611275136470795,
0.03720240667462349,
-0.11573143303394318,
-0.08077014982700348,
-0.03517830744385719,
-0.07612033188343048,
-0.003899846225976944,
0.05018248409032822,
0.04169982299208641,
-0.11496815085411072,
-0.10908655822277069,
0.048020701855421066,
0.08243758976459503,
-0.07590442150831223,
0.07104592770338058,
-0.12896215915679932,
0.081379234790802,
-0.07228641957044601,
-0.02326318249106407,
-0.1445063203573227,
-0.047309473156929016,
0.01607843115925789,
-0.010113606229424477,
-0.014702768996357918,
-0.04639873653650284,
0.06706959009170532,
0.08815736323595047,
-0.05594225227832794,
-0.025359801948070526,
-0.06131429970264435,
0.03408878669142723,
-0.06765580922365189,
-0.24935059249401093,
-0.042875587940216064,
-0.06281458586454391,
0.06548815965652466,
-0.17772792279720306,
0.01452774927020073,
0.10871239006519318,
0.1425345540046692,
0.07025287300348282,
-0.05411909520626068,
0.00505960313603282,
0.04680953919887543,
-0.015902917832136154,
-0.09017964452505112,
0.052494265139102936,
0.008115791715681553,
-0.08898203074932098,
0.0015385838923975825,
-0.134832501411438,
0.11407772451639175,
0.14396582543849945,
-0.08859986066818237,
-0.05222346633672714,
0.02723298780620098,
-0.046472400426864624,
-0.011974607594311237,
-0.010282699018716812,
0.03588835895061493,
0.11149235814809799,
0.014941745437681675,
0.14401419460773468,
-0.09322958439588547,
-0.04899781942367554,
0.07019371539354324,
-0.03275490179657936,
-0.02170424722135067,
0.0784899964928627,
0.06212826073169708,
-0.08983759582042694,
0.12094558775424957,
0.15301984548568726,
-0.03729498013854027,
0.15498510003089905,
-0.08835336565971375,
-0.08198752999305725,
-0.033035650849342346,
0.001917621004395187,
0.034444697201251984,
0.1485266536474228,
-0.08447644114494324,
-0.03059515170753002,
0.00750351045280695,
-0.0057188416831195354,
-0.02833574265241623,
-0.22201192378997803,
-0.028345424681901932,
0.04567909240722656,
-0.0779748484492302,
-0.03890206664800644,
-0.019887100905179977,
-0.009111952036619186,
0.09814446419477463,
0.02731686644256115,
-0.06377384066581726,
0.030023451894521713,
-0.01689804159104824,
-0.06342994421720505,
0.17592589557170868,
-0.09810889512300491,
-0.13548561930656433,
-0.092804454267025,
-0.086464524269104,
-0.029790572822093964,
0.002302071312442422,
0.042908601462841034,
-0.10769529640674591,
-0.042160142213106155,
-0.0899212583899498,
-0.0543351024389267,
-0.005910216365009546,
0.03662335127592087,
0.07529604434967041,
0.022839123383164406,
0.07914320379495621,
-0.09687183797359467,
-0.006871545687317848,
-0.025303829461336136,
-0.05121744051575661,
0.05138498544692993,
0.04505057632923126,
0.12637054920196533,
0.09431934356689453,
-0.048104628920555115,
0.039418697357177734,
-0.04831506311893463,
0.23218262195587158,
-0.11275815218687057,
-0.008982391096651554,
0.13402752578258514,
-0.02304815873503685,
0.051777563989162445,
0.12299450486898422,
0.07083015143871307,
-0.10273256152868271,
-0.018228702247142792,
0.01843358390033245,
-0.04026190936565399,
-0.17984066903591156,
-0.0011022835969924927,
-0.040684331208467484,
-0.0029993297066539526,
0.10727918893098831,
0.025315450504422188,
-0.011037983000278473,
0.033344220370054245,
0.006789744831621647,
0.029222829267382622,
0.020059196278452873,
0.11067153513431549,
0.07611075788736343,
0.05375709757208824,
0.10558183491230011,
-0.06061212345957756,
0.005892458837479353,
0.040712226182222366,
0.018516886979341507,
0.2143566906452179,
0.016813283786177635,
0.17488184571266174,
0.08372155576944351,
0.11070534586906433,
0.027855223044753075,
0.017957553267478943,
-0.006898417137563229,
-0.04335073009133339,
0.003913817461580038,
-0.05707652494311333,
-0.01922641694545746,
0.029255766421556473,
-0.05790599435567856,
-0.011992346495389938,
-0.10222483426332474,
0.07916095107793808,
0.05975573509931564,
0.27436789870262146,
0.0465044267475605,
-0.3625602126121521,
-0.07770220190286636,
0.01062688510864973,
-0.015307381749153137,
-0.02736993134021759,
0.026383180171251297,
0.13960611820220947,
-0.05151739344000816,
0.11214352399110794,
-0.06520913541316986,
0.07912936806678772,
-0.06322050839662552,
0.019755657762289047,
0.10249528288841248,
0.07568547874689102,
-0.005315106827765703,
0.03036939911544323,
-0.2846771478652954,
0.2876308560371399,
0.029402874410152435,
0.08385762572288513,
-0.032288309186697006,
-0.009392482228577137,
0.016102878376841545,
0.09196357429027557,
0.1313110738992691,
-0.008688091300427914,
-0.1276102215051651,
-0.16741594672203064,
-0.05312768369913101,
0.014053735882043839,
0.1273670792579651,
0.028147725388407707,
0.11080589890480042,
-0.01367083378136158,
-0.02062063105404377,
0.06191256642341614,
-0.11070385575294495,
-0.07840344309806824,
-0.0769660547375679,
-0.01119103841483593,
0.02863074094057083,
-0.02240358106791973,
-0.0898359939455986,
-0.0880584567785263,
-0.08900437504053116,
0.15284647047519684,
-0.08539857715368271,
-0.026805486530065536,
-0.11961577832698822,
0.05263392999768257,
0.051946815103292465,
-0.057698238641023636,
0.07726222276687622,
-0.012621650472283363,
0.16495110094547272,
-0.015375406481325626,
-0.05169468745589256,
0.12879662215709686,
-0.07585009187459946,
-0.18467208743095398,
-0.0739884302020073,
0.10267668217420578,
0.014008629135787487,
0.05719335377216339,
-0.01270702201873064,
0.04339565709233284,
0.016639554873108864,
-0.063895583152771,
0.03338852897286415,
-0.002716581104323268,
0.06310012191534042,
-0.06016784533858299,
-0.008633303456008434,
0.01058804988861084,
-0.054521724581718445,
-0.009764472022652626,
0.15012900531291962,
0.3501301109790802,
-0.11695912480354309,
0.0356028787791729,
0.02837901934981346,
-0.04154891148209572,
-0.19494026899337769,
0.051070455461740494,
0.0670929029583931,
-0.03745405003428459,
-0.001965069444850087,
-0.14833983778953552,
0.04571915045380592,
0.07125666737556458,
-0.01743151806294918,
0.08658072352409363,
-0.26903390884399414,
-0.13113589584827423,
0.06374282389879227,
0.17226192355155945,
0.06005851924419403,
-0.13445571064949036,
-0.012009958736598492,
-0.0012539434246718884,
-0.14745566248893738,
0.11377975344657898,
-0.0888514444231987,
0.1024840772151947,
-0.018468080088496208,
0.04071889445185661,
-0.00177458964753896,
-0.06133652850985527,
0.12253278493881226,
-0.018924761563539505,
0.12505853176116943,
-0.05218330770730972,
-0.03730875998735428,
0.13067755103111267,
-0.08773462474346161,
0.008573018945753574,
-0.07617208361625671,
0.016099249944090843,
-0.10305982083082199,
0.007497860584408045,
-0.06520620733499527,
-0.020458143204450607,
-0.032706428319215775,
-0.03800046816468239,
-0.05858733877539635,
0.044673021882772446,
0.04795585572719574,
-0.0133681520819664,
0.2309328019618988,
-0.011138019151985645,
0.10026224702596664,
0.16819407045841217,
0.08051910996437073,
-0.09852737933397293,
-0.04630153626203537,
0.007576589472591877,
-0.02158411405980587,
0.07760750502347946,
-0.17595286667346954,
0.04424753785133362,
0.12206467241048813,
0.026763441041111946,
0.1262584924697876,
0.0561348982155323,
-0.04114804044365883,
0.04733419790863991,
0.08383074402809143,
-0.12955589592456818,
-0.10035743564367294,
0.030566317960619926,
0.013738729059696198,
-0.08838789910078049,
0.02912089042365551,
0.09335008263587952,
-0.07386791706085205,
0.02067822776734829,
-0.01164670754224062,
0.029445892199873924,
-0.04862193018198013,
0.1487128883600235,
0.05295872688293457,
0.0634930282831192,
-0.11722841113805771,
0.11583738774061203,
0.02077253721654415,
-0.10144881159067154,
0.02201826497912407,
0.08928428590297699,
-0.09439418464899063,
-0.012659032829105854,
0.049103349447250366,
0.13687381148338318,
-0.04846048355102539,
-0.05831916630268097,
-0.1535983830690384,
-0.12448764592409134,
0.08465852588415146,
0.19607888162136078,
0.05337660014629364,
0.013891400769352913,
-0.004592704586684704,
0.027683012187480927,
-0.14003276824951172,
0.09883814305067062,
0.005531559232622385,
0.07165487110614777,
-0.16301199793815613,
0.13480380177497864,
0.0073163071647286415,
0.039676733314991,
-0.0304070133715868,
0.03606255352497101,
-0.0742817372083664,
0.02777465060353279,
-0.11056128889322281,
0.019434088841080666,
-0.05208222568035126,
0.022142784669995308,
-0.023152628913521767,
-0.04638684168457985,
-0.06942955404520035,
0.028705613687634468,
-0.0983722060918808,
-0.02561764605343342,
0.028748156502842903,
0.023457270115613937,
-0.14125999808311462,
-0.03848056495189667,
-0.006704468280076981,
-0.08602067083120346,
0.048344846814870834,
0.011136440560221672,
0.013893179595470428,
0.026034235954284668,
-0.12804216146469116,
-0.02579217031598091,
0.08509223163127899,
-0.012542086653411388,
0.056574318557977676,
-0.07797293365001678,
-0.015572674572467804,
-0.009581235237419605,
0.0022741230204701424,
-0.00045793960453011096,
0.07293055951595306,
-0.10313903540372849,
0.014720004051923752,
-0.009115304797887802,
-0.034427493810653687,
-0.055593714118003845,
0.07992690056562424,
0.12229298055171967,
-0.006143598817288876,
0.1682760864496231,
-0.08452585339546204,
-0.00689562875777483,
-0.20221781730651855,
-0.015031976625323296,
0.01023150235414505,
-0.12395023554563522,
-0.09670864045619965,
-0.023125141859054565,
0.0769488736987114,
-0.08319433778524399,
0.1370367705821991,
-0.009467275813221931,
-0.0037158806808292866,
0.07547537237405777,
-0.09061144292354584,
-0.03013489581644535,
0.0431402362883091,
0.17294752597808838,
0.02333156205713749,
-0.028876177966594696,
0.03855982422828674,
0.015845229849219322,
0.11417809128761292,
0.0753978043794632,
0.16298893094062805,
0.15331494808197021,
0.005153084173798561,
0.10381628572940826,
0.06821651756763458,
-0.032514140009880066,
-0.15659601986408234,
0.13321714103221893,
-0.0756223276257515,
0.13661929965019226,
-0.016604747623205185,
0.12705713510513306,
0.17115972936153412,
-0.1768280267715454,
0.02191799134016037,
-0.013324285857379436,
-0.07548119127750397,
-0.09103069454431534,
-0.060059692710638046,
-0.08643710613250732,
-0.16711002588272095,
0.03753148019313812,
-0.12229273468255997,
0.09358461201190948,
0.05492781475186348,
0.04527217522263527,
0.007584554608911276,
0.17913547158241272,
0.020085852593183517,
0.032320186495780945,
0.10070808976888657,
0.023410286754369736,
-0.03446901962161064,
-0.031749263405799866,
-0.07578494399785995,
0.06440013647079468,
-0.047694116830825806,
0.02859068661928177,
-0.017335936427116394,
-0.031283363699913025,
0.07938501983880997,
0.019680365920066833,
-0.1214505210518837,
0.04437464848160744,
0.025785740464925766,
0.061343561857938766,
0.04381568729877472,
0.006663853302598,
0.025130437687039375,
0.006793931592255831,
0.1685948520898819,
-0.070668064057827,
-0.06398812681436539,
-0.0919203981757164,
0.19186529517173767,
0.0007522685918956995,
0.0068016960285604,
-0.0012836054665967822,
-0.07922463864088058,
-0.013953974470496178,
0.13774579763412476,
0.1685316413640976,
-0.05545271560549736,
-0.012897098436951637,
-0.04937412962317467,
-0.008280343376100063,
-0.04995333030819893,
0.10146936029195786,
0.06593988090753555,
-0.03260861709713936,
-0.0741177350282669,
-0.07540587335824966,
-0.0485897995531559,
-0.019623590633273125,
-0.008971335366368294,
0.008621584624052048,
0.03934366628527641,
0.0026903245598077774,
-0.07822690159082413,
0.05203329026699066,
-0.028552807867527008,
-0.114717036485672,
0.1110769584774971,
-0.1841883510351181,
-0.12428782135248184,
0.0036883302964270115,
0.0828961655497551,
-0.003985917661339045,
0.03537943959236145,
-0.002936277538537979,
-0.007526662666350603,
0.052256252616643906,
-0.008451614528894424,
-0.04667797312140465,
-0.10133136063814163,
0.08227432519197464,
-0.1441429704427719,
0.23600158095359802,
-0.03967801854014397,
0.04997151717543602,
0.10286369919776917,
0.021097835153341293,
-0.0998011901974678,
0.06891951709985733,
0.055656980723142624,
-0.08265040814876556,
-0.02336982637643814,
0.1551862210035324,
-0.05378679931163788,
0.1401149183511734,
0.050589513033628464,
-0.09469717741012573,
0.011486401781439781,
-0.11068479716777802,
-0.09344330430030823,
-0.04024229198694229,
-0.048257868736982346,
-0.03071790561079979,
0.1343548446893692,
0.18505123257637024,
-0.03770441934466362,
0.024099715054035187,
-0.06588520109653473,
0.025201858952641487,
0.11384617537260056,
0.017231246456503868,
-0.035664111375808716,
-0.22570286691188812,
0.03833596035838127,
0.0990179032087326,
0.012372237630188465,
-0.1757768988609314,
-0.09478534758090973,
-0.008138330653309822,
-0.045961469411849976,
-0.058415453881025314,
0.08493270725011826,
0.08224477618932724,
0.05655601993203163,
-0.07450738549232483,
-0.09574196487665176,
-0.02088235318660736,
0.159827321767807,
-0.13334909081459045,
-0.07085748016834259
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.7.2.dev0 | {"library_name": "peft", "base_model": "beomi/KoAlpaca-Polyglot-12.8B"} | null | humung/koalpaca-polyglot-12.8B-ia3-vlending-v0.1 | [
"peft",
"arxiv:1910.09700",
"base_model:beomi/KoAlpaca-Polyglot-12.8B",
"region:us"
] | 2024-02-07T06:59:19+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #arxiv-1910.09700 #base_model-beomi/KoAlpaca-Polyglot-12.8B #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.7.2.dev0 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.7.2.dev0"
] | [
"TAGS\n#peft #arxiv-1910.09700 #base_model-beomi/KoAlpaca-Polyglot-12.8B #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.7.2.dev0"
] | [
37,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
14
] | [
"passage: TAGS\n#peft #arxiv-1910.09700 #base_model-beomi/KoAlpaca-Polyglot-12.8B #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.7.2.dev0"
] | [
-0.10540957748889923,
0.20559802651405334,
-0.003004602389410138,
0.03466712683439255,
0.08773048967123032,
0.023959729820489883,
0.053046777844429016,
0.13423356413841248,
-0.026354718953371048,
0.10868921130895615,
0.07335398346185684,
0.09555713832378387,
0.1065594032406807,
0.22315357625484467,
0.007551147136837244,
-0.21237707138061523,
0.027452172711491585,
-0.08461704850196838,
-0.00944582000374794,
0.12461921572685242,
0.14523938298225403,
-0.09219498187303543,
0.07921633869409561,
-0.00964981410652399,
-0.015263928100466728,
-0.036831293255090714,
-0.07659906893968582,
-0.038172997534275055,
0.03601988032460213,
0.04703638702630997,
0.05187511816620827,
-0.00778587069362402,
0.08086880296468735,
-0.2668561339378357,
0.01613675430417061,
0.04639415815472603,
-0.0035148838069289923,
0.08257058262825012,
0.09800540655851364,
-0.037514783442020416,
0.14395184814929962,
-0.03548780083656311,
0.1356346607208252,
0.07572120428085327,
-0.09309115260839462,
-0.21433337032794952,
-0.0732581838965416,
0.09041865915060043,
0.18088139593601227,
0.07243019342422485,
-0.03964708000421524,
0.13519597053527832,
-0.09471599757671356,
0.015010922215878963,
0.04241056740283966,
-0.0877523347735405,
-0.0759870707988739,
0.05766043812036514,
0.09074953943490982,
0.05482340231537819,
-0.12778136134147644,
-0.03361202031373978,
0.025347072631120682,
0.029780209064483643,
0.06822843104600906,
0.010396120138466358,
0.1515258550643921,
0.020694183185696602,
-0.14881348609924316,
-0.04806196689605713,
0.14434587955474854,
0.026913940906524658,
-0.03731377050280571,
-0.22284100949764252,
0.008566704578697681,
-0.08145062625408173,
-0.028312578797340393,
-0.0454905740916729,
0.0370527021586895,
-0.0007401038892567158,
0.09774506837129593,
-0.027208508923649788,
-0.0839231014251709,
-0.009054016321897507,
0.09633250534534454,
0.05235956236720085,
0.02450650930404663,
-0.018564295023679733,
0.0030212849378585815,
0.12530362606048584,
0.04388170689344406,
-0.1294206827878952,
-0.06923345476388931,
-0.07562793791294098,
-0.040979985147714615,
-0.04752693697810173,
0.0435853935778141,
0.04560151323676109,
0.04961647465825081,
0.25797632336616516,
-0.029604587703943253,
0.05874665826559067,
0.07247164100408554,
0.013998884707689285,
0.04994429275393486,
0.10447422415018082,
-0.05734764784574509,
-0.16450390219688416,
-0.018255729228258133,
0.09482507407665253,
-0.007220112718641758,
-0.02308279648423195,
-0.05090372636914253,
0.03542707487940788,
0.03714252635836601,
0.10341520607471466,
0.09943357110023499,
-0.006948645226657391,
-0.07685218006372452,
-0.0603540763258934,
0.19842529296875,
-0.14560015499591827,
0.04970952123403549,
0.019164616242051125,
-0.020797867327928543,
-0.04268357530236244,
0.011600611731410027,
0.016079209744930267,
-0.03291628509759903,
0.10026854276657104,
-0.06706841289997101,
-0.03678459674119949,
-0.11905615031719208,
-0.019217079505324364,
0.038078635931015015,
0.0009777782252058387,
-0.02712208405137062,
-0.028383716940879822,
-0.06745737046003342,
-0.09160862118005753,
0.10519209504127502,
-0.07168067991733551,
-0.06304856389760971,
-0.026171814650297165,
-0.0865248441696167,
0.0203411802649498,
0.029738282784819603,
0.09529437869787216,
-0.027110902592539787,
0.03877667337656021,
-0.008446753025054932,
0.06783518195152283,
0.08302006125450134,
0.037093956023454666,
-0.06604405492544174,
0.061555903404951096,
-0.20855793356895447,
0.08830954134464264,
-0.08274400979280472,
0.021903911605477333,
-0.1620350033044815,
-0.021668478846549988,
0.004943254869431257,
0.017995871603488922,
0.03137851133942604,
0.1654725968837738,
-0.18901321291923523,
-0.03607144579291344,
0.16148042678833008,
-0.10035122185945511,
-0.11384820938110352,
0.03989677503705025,
-0.04917827993631363,
0.16122356057167053,
0.017129600048065186,
-0.012018389999866486,
0.11307433247566223,
-0.1448078602552414,
-0.023520389571785927,
-0.023598995059728622,
-0.0005060981493443251,
0.09559689462184906,
0.088362917304039,
-0.0854744091629982,
0.02989846095442772,
0.014886978082358837,
-0.05555133521556854,
-0.02998429909348488,
-0.04564712196588516,
-0.10921370983123779,
0.006399492733180523,
-0.07876116037368774,
0.02083619311451912,
-0.011231109499931335,
-0.07873944938182831,
-0.006607440300285816,
-0.1585233211517334,
-0.027926629409193993,
0.08556310832500458,
0.013544773682951927,
-0.01887129433453083,
-0.09782706201076508,
0.04953368753194809,
-0.041312918066978455,
-0.023039814084768295,
-0.14905861020088196,
-0.007938822731375694,
0.02067517302930355,
-0.1336175799369812,
0.00863411370664835,
-0.11709819734096527,
0.06830461323261261,
0.0018478805432096124,
-0.05985941365361214,
-0.032764632254838943,
-0.00852511078119278,
0.0034239371307194233,
-0.05219552293419838,
-0.2326623499393463,
-0.0263657346367836,
-0.048467542976140976,
0.16475994884967804,
-0.21808937191963196,
0.034082815051078796,
0.05260705202817917,
0.12895150482654572,
0.00032723211916163564,
-0.06203172728419304,
0.031278472393751144,
-0.06397328525781631,
-0.0214493740350008,
-0.07661542296409607,
0.00008526536112185568,
-0.005617273971438408,
-0.03688585385680199,
0.011217194609344006,
-0.10441116988658905,
-0.04838930815458298,
0.10017045587301254,
0.08055291324853897,
-0.15570560097694397,
-0.013271273113787174,
-0.04083208739757538,
-0.06467369198799133,
-0.0846254900097847,
-0.052871134132146835,
0.11568466573953629,
0.0498887300491333,
0.040877070277929306,
-0.07540175318717957,
-0.07389430701732635,
0.01353177335113287,
-0.024675680324435234,
-0.023185469210147858,
0.11422264575958252,
0.08249276131391525,
-0.1040324792265892,
0.09635740518569946,
0.07085317373275757,
0.026578130200505257,
0.09605441242456436,
-0.027124769985675812,
-0.10774759203195572,
-0.03280458226799965,
0.04691934958100319,
0.014243979938328266,
0.16331171989440918,
-0.08055722713470459,
0.048136383295059204,
0.043023888021707535,
-0.03327047452330589,
0.04607550427317619,
-0.09969968348741531,
0.010920601896941662,
0.015540328808128834,
-0.01461623702198267,
0.013288166373968124,
-0.017842615023255348,
0.00807366892695427,
0.08394591510295868,
0.056021545082330704,
0.036755096167325974,
0.02656877040863037,
-0.030867567285895348,
-0.12846308946609497,
0.1894298940896988,
-0.09403270483016968,
-0.23091864585876465,
-0.15635430812835693,
0.060239486396312714,
0.050140369683504105,
-0.021174944937229156,
0.017909890040755272,
-0.05904591083526611,
-0.10082703828811646,
-0.0797211229801178,
-0.002996908500790596,
0.02637184038758278,
-0.05325498804450035,
-0.06202004477381706,
0.04160214215517044,
0.044552311301231384,
-0.11697765439748764,
0.03576492890715599,
0.05722247436642647,
-0.011896946467459202,
0.0116036431863904,
0.04773005470633507,
0.08603720366954803,
0.18123899400234222,
-0.012317048385739326,
0.0004215012304484844,
0.05266686901450157,
0.27666401863098145,
-0.16215047240257263,
0.11103972792625427,
0.12124417722225189,
-0.05866656452417374,
0.07658068835735321,
0.17997364699840546,
0.027063140645623207,
-0.098046213388443,
0.02566959708929062,
0.02916114777326584,
-0.028762148693203926,
-0.2633727788925171,
-0.049808938056230545,
-0.016081707552075386,
-0.08449802547693253,
0.08712391555309296,
0.08431845158338547,
0.0767303928732872,
0.04020083323121071,
-0.0661969780921936,
-0.07611334323883057,
0.024384990334510803,
0.10425059497356415,
-0.0191093310713768,
0.011448838748037815,
0.081904336810112,
-0.037322912365198135,
0.011437594890594482,
0.09801331162452698,
-0.004299907013773918,
0.16685496270656586,
0.05327363312244415,
0.11320813000202179,
0.07658594846725464,
0.08777978271245956,
-0.002004442038014531,
0.023283634334802628,
0.012051019817590714,
0.02449721097946167,
0.015259896405041218,
-0.079375259578228,
0.020798334851861,
0.11017418652772903,
0.04837658628821373,
0.027168327942490578,
0.013060457073152065,
-0.03896590694785118,
0.04777206853032112,
0.18672892451286316,
0.013180929236114025,
-0.19674982130527496,
-0.07594987004995346,
0.0595211423933506,
-0.07707803696393967,
-0.1365038901567459,
-0.018876932561397552,
0.024392379447817802,
-0.17148363590240479,
0.023588472977280617,
-0.048532236367464066,
0.10074736177921295,
-0.09277769178152084,
-0.0468716137111187,
0.10480275750160217,
0.066237673163414,
-0.022967418655753136,
0.05461890250444412,
-0.18961703777313232,
0.12749436497688293,
0.02305809035897255,
0.07003628462553024,
-0.09290280938148499,
0.10085926949977875,
0.003066322533413768,
-0.008259178139269352,
0.1627831757068634,
0.008199378848075867,
-0.06509269773960114,
-0.0729934498667717,
-0.09288618713617325,
-0.01732928305864334,
0.09777723997831345,
-0.12716518342494965,
0.07387841492891312,
-0.017828097566962242,
-0.027502309530973434,
-0.006057330872863531,
-0.08532759547233582,
-0.11867906898260117,
-0.1710962951183319,
0.06502113491296768,
-0.10218226164579391,
0.028545623645186424,
-0.0869835838675499,
-0.06487492471933365,
0.005896477960050106,
0.17054986953735352,
-0.2083892673254013,
-0.10043428838253021,
-0.14862020313739777,
-0.08871041238307953,
0.16340431571006775,
-0.04621078819036484,
0.08447655290365219,
-0.0008726009982638061,
0.16444042325019836,
0.013363342732191086,
-0.005642919335514307,
0.09241900593042374,
-0.08825422078371048,
-0.19199316203594208,
-0.05295608937740326,
0.16081050038337708,
0.13406766951084137,
0.03866748511791229,
-0.009174387902021408,
0.02346714772284031,
-0.04733990505337715,
-0.12172318994998932,
0.025878695771098137,
0.1349465698003769,
0.07331836223602295,
-0.016982635483145714,
-0.01937372423708439,
-0.08134505152702332,
-0.06843769550323486,
-0.04553724825382233,
-0.001679486595094204,
0.19145622849464417,
-0.07377782464027405,
0.15888291597366333,
0.09683718532323837,
-0.061084065586328506,
-0.20177116990089417,
0.04793189838528633,
0.05791187658905983,
0.018336299806833267,
0.022340089082717896,
-0.20401662588119507,
0.07889525592327118,
-0.000328303431160748,
-0.07100894302129745,
0.16285833716392517,
-0.18476811051368713,
-0.1418207883834839,
0.09888015687465668,
0.030441628769040108,
-0.20764920115470886,
-0.14031247794628143,
-0.09815382212400436,
-0.010438106022775173,
-0.12221009284257889,
0.056168727576732635,
0.005826505366712809,
0.014136971905827522,
0.021295180544257164,
0.028171129524707794,
0.02162000723183155,
-0.04450424015522003,
0.20762448012828827,
-0.02438056282699108,
0.0022295936942100525,
-0.05022121220827103,
-0.0900488868355751,
0.022831372916698456,
-0.053481604903936386,
0.10679526627063751,
0.010150454007089138,
0.023167809471488,
-0.17562416195869446,
-0.03860195353627205,
-0.05738042667508125,
0.02883780561387539,
-0.09595795720815659,
-0.08790213614702225,
-0.039584483951330185,
0.09481174498796463,
0.0935063287615776,
-0.026514306664466858,
0.003174888202920556,
-0.0862383171916008,
0.05826888605952263,
0.19683687388896942,
0.19022683799266815,
0.06592807173728943,
-0.07296352088451385,
0.019150886684656143,
-0.027655238285660744,
0.04857837036252022,
-0.23716144263744354,
0.03750712797045708,
0.05384741723537445,
0.02620970457792282,
0.08409903198480606,
-0.013743074610829353,
-0.15661190450191498,
-0.06760632991790771,
0.0815219059586525,
-0.04874030128121376,
-0.1545884609222412,
-0.020022185519337654,
0.03485345095396042,
-0.20644737780094147,
-0.04766641557216644,
0.024015212431550026,
-0.02572871372103691,
-0.04414347559213638,
0.02390330657362938,
0.07005660235881805,
-0.020936394110322,
0.11210499703884125,
0.08414473384618759,
0.09306694567203522,
-0.10183402895927429,
0.07949715107679367,
0.07005241513252258,
-0.046095892786979675,
0.03028031997382641,
0.1051742285490036,
-0.04836070537567139,
-0.031318049877882004,
0.08050205558538437,
0.08464851975440979,
0.02099420502781868,
-0.052801284939050674,
0.011157471686601639,
-0.05638609081506729,
0.06766611337661743,
0.11316966265439987,
0.03276762738823891,
-0.004189291037619114,
0.05136692523956299,
0.03245152160525322,
-0.09770885109901428,
0.10308299958705902,
0.06513509154319763,
0.015027876012027264,
-0.04574684798717499,
-0.04541699215769768,
-0.01194663718342781,
-0.014374147169291973,
-0.01988505944609642,
-0.011119894683361053,
-0.09364333748817444,
-0.008325990289449692,
-0.09307701885700226,
0.020649690181016922,
-0.07398909330368042,
0.012286537326872349,
0.02572891116142273,
-0.05121826380491257,
0.0026844514068216085,
0.005442524794489145,
-0.0764651671051979,
-0.051145974546670914,
-0.021213872358202934,
0.08480258285999298,
-0.12075527012348175,
0.03950418904423714,
0.08023491501808167,
-0.10450131446123123,
0.06949853152036667,
-0.003907262347638607,
0.013384518213570118,
0.0006123523926362395,
-0.14775694906711578,
0.053266506642103195,
-0.02561413310468197,
-0.017105547711253166,
0.023125868290662766,
-0.2178812175989151,
-0.003649619175121188,
-0.048077210783958435,
-0.04864832013845444,
0.011964813806116581,
-0.02440597675740719,
-0.12254353612661362,
0.09221115708351135,
-0.004571611061692238,
-0.0696164071559906,
-0.01657094806432724,
0.04662122204899788,
0.10417108237743378,
-0.026122163981199265,
0.13508720695972443,
-0.024030504748225212,
0.07298879325389862,
-0.17288702726364136,
-0.008677089586853981,
-0.012730386108160019,
0.038993701338768005,
-0.016805781051516533,
-0.025904789566993713,
0.05921988934278488,
-0.014625982381403446,
0.17589767277240753,
-0.012446380220353603,
0.06950150430202484,
0.05482075363397598,
0.008356618694961071,
0.008194195106625557,
0.08076076954603195,
0.06665685027837753,
0.005010666791349649,
-0.0008032953483052552,
0.044272493571043015,
0.002108176937326789,
-0.03683855012059212,
-0.1519884169101715,
0.06569726765155792,
0.15046218037605286,
0.055599603801965714,
0.030595125630497932,
0.021528474986553192,
-0.11448873579502106,
-0.07594701647758484,
0.12205343693494797,
-0.022100290283560753,
-0.026909690350294113,
-0.06738917529582977,
0.15837456285953522,
0.1377950757741928,
-0.2068730592727661,
0.07074018567800522,
-0.052351925522089005,
-0.05456995964050293,
-0.13457807898521423,
-0.16687943041324615,
-0.0612260103225708,
-0.05464024469256401,
-0.016351282596588135,
-0.06748435646295547,
0.052146222442388535,
0.050902511924505234,
0.008526553399860859,
-0.017359228804707527,
0.10663273930549622,
0.005909161642193794,
-0.03404774144291878,
0.062113918364048004,
0.06192287430167198,
0.03367035463452339,
-0.0905078873038292,
0.008912215009331703,
-0.0068720122799277306,
0.007783703971654177,
0.0713251456618309,
0.01782422699034214,
-0.05224384367465973,
0.018080750480294228,
-0.01283710915595293,
-0.11826425045728683,
0.03787480294704437,
-0.015002252534031868,
-0.029315505176782608,
0.1566014587879181,
0.029751593247056007,
0.013188540935516357,
-0.022483261302113533,
0.2230798304080963,
-0.07382278889417648,
-0.06390204280614853,
-0.1547214835882187,
0.06660162657499313,
-0.06621479988098145,
0.03013068065047264,
0.026573562994599342,
-0.11524501442909241,
0.015479504130780697,
0.17072036862373352,
0.12525522708892822,
-0.0063206497579813,
0.0063994331285357475,
0.04517899081110954,
0.004865913651883602,
-0.028752600774168968,
0.02441132254898548,
0.048747360706329346,
0.13754519820213318,
-0.07439347356557846,
0.06578821688890457,
-0.006502992939203978,
-0.08312483876943588,
-0.01819593459367752,
0.10022545605897903,
0.001182384556159377,
-0.0019503498915582895,
-0.06880789250135422,
0.13801248371601105,
-0.08881956338882446,
-0.21646666526794434,
0.06816039234399796,
-0.07543180882930756,
-0.15402449667453766,
-0.04297800362110138,
0.021336017176508904,
-0.012795302085578442,
0.018365394324064255,
0.07468558102846146,
-0.05250171199440956,
0.1823025941848755,
0.044230978935956955,
-0.05458521842956543,
-0.09213927388191223,
0.05417490378022194,
-0.1392185091972351,
0.2722955644130707,
0.019934138283133507,
0.04954831674695015,
0.10472618043422699,
-0.01806936040520668,
-0.14488475024700165,
0.005098022520542145,
0.10815653204917908,
-0.06299184262752533,
0.054657403379678726,
0.1741686463356018,
0.00031542935175821185,
0.13204926252365112,
0.060122352093458176,
-0.04688252881169319,
0.03457929566502571,
-0.0965464860200882,
-0.04105044901371002,
-0.11572632193565369,
0.08440550416707993,
-0.08579786866903305,
0.16429857909679413,
0.13920848071575165,
-0.07239620387554169,
-0.012172039598226547,
-0.02032407559454441,
0.08366859704256058,
0.005827086511999369,
0.11327318102121353,
0.0020089689642190933,
-0.17611683905124664,
0.031192481517791748,
-0.0014404163230210543,
0.10042709857225418,
-0.19271689653396606,
-0.05438857898116112,
0.04894648864865303,
-0.022859660908579826,
-0.07311827689409256,
0.12349299341440201,
0.032814979553222656,
0.03649212792515755,
-0.038736749440431595,
-0.0547870472073555,
0.005395416636019945,
0.14204010367393494,
-0.11280781030654907,
-0.006454369984567165
] |
null | null | transformers | # maid-yuzu-v5-mix
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
This model was created because I was curious about whether the 8X7B model created randomly by the user would be merged with other existing 8x7b models.
## Merge Details
### Merge Method
This model was merged using the SLERP merge method.
### Models Merged
The following models were included in the merge:
* ../maid-yuzu-v5
* [smelborp/MixtralOrochi8x7B](https://huggingface.co/smelborp/MixtralOrochi8x7B)
### Configuration
The following YAML configuration was used to produce this model:
```yaml
base_model:
model:
path: ../maid-yuzu-v5
dtype: bfloat16
merge_method: slerp
parameters:
t:
- value: 0.5
slices:
- sources:
- layer_range: [0, 32]
model:
model:
path: smelborp/MixtralOrochi8x7B
- layer_range: [0, 32]
model:
model:
path: ../maid-yuzu-v5
```
| {"library_name": "transformers", "tags": ["mergekit", "merge"], "base_model": ["smelborp/MixtralOrochi8x7B"]} | text-generation | rhplus0831/maid-yuzu-v5-mix-exl2-6.0bpw-rpcal | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"mergekit",
"merge",
"conversational",
"base_model:smelborp/MixtralOrochi8x7B",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T07:01:47+00:00 | [] | [] | TAGS
#transformers #safetensors #mixtral #text-generation #mergekit #merge #conversational #base_model-smelborp/MixtralOrochi8x7B #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # maid-yuzu-v5-mix
This is a merge of pre-trained language models created using mergekit.
This model was created because I was curious about whether the 8X7B model created randomly by the user would be merged with other existing 8x7b models.
## Merge Details
### Merge Method
This model was merged using the SLERP merge method.
### Models Merged
The following models were included in the merge:
* ../maid-yuzu-v5
* smelborp/MixtralOrochi8x7B
### Configuration
The following YAML configuration was used to produce this model:
| [
"# maid-yuzu-v5-mix\n\nThis is a merge of pre-trained language models created using mergekit.\n\nThis model was created because I was curious about whether the 8X7B model created randomly by the user would be merged with other existing 8x7b models.",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the SLERP merge method.",
"### Models Merged\n\nThe following models were included in the merge:\n* ../maid-yuzu-v5\n* smelborp/MixtralOrochi8x7B",
"### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
"TAGS\n#transformers #safetensors #mixtral #text-generation #mergekit #merge #conversational #base_model-smelborp/MixtralOrochi8x7B #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# maid-yuzu-v5-mix\n\nThis is a merge of pre-trained language models created using mergekit.\n\nThis model was created because I was curious about whether the 8X7B model created randomly by the user would be merged with other existing 8x7b models.",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the SLERP merge method.",
"### Models Merged\n\nThe following models were included in the merge:\n* ../maid-yuzu-v5\n* smelborp/MixtralOrochi8x7B",
"### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
78,
61,
4,
18,
41,
17
] | [
"passage: TAGS\n#transformers #safetensors #mixtral #text-generation #mergekit #merge #conversational #base_model-smelborp/MixtralOrochi8x7B #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# maid-yuzu-v5-mix\n\nThis is a merge of pre-trained language models created using mergekit.\n\nThis model was created because I was curious about whether the 8X7B model created randomly by the user would be merged with other existing 8x7b models.## Merge Details### Merge Method\n\nThis model was merged using the SLERP merge method.### Models Merged\n\nThe following models were included in the merge:\n* ../maid-yuzu-v5\n* smelborp/MixtralOrochi8x7B### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
-0.06843120604753494,
-0.0639222040772438,
-0.0013013146817684174,
-0.029540063813328743,
0.08714442700147629,
0.0483069084584713,
0.20440888404846191,
-0.011589844711124897,
0.07935865223407745,
0.031766436994075775,
0.03172202780842781,
0.045097604393959045,
0.041795969009399414,
0.20851939916610718,
-0.05296465381979942,
-0.13310706615447998,
0.07394636422395706,
-0.05819404125213623,
-0.14567552506923676,
0.10862846672534943,
0.10552287846803665,
-0.08243416994810104,
0.10070817917585373,
0.03806445747613907,
-0.16787119209766388,
0.032693393528461456,
-0.038313690572977066,
0.03757844492793083,
0.06616710871458054,
0.1408260017633438,
0.0890655517578125,
0.06307199597358704,
0.006891318596899509,
-0.14616958796977997,
0.05050556734204292,
-0.021685952320694923,
-0.020775197073817253,
0.03821534290909767,
0.10106975585222244,
-0.03938208147883415,
0.15368002653121948,
-0.03587837517261505,
0.015648216009140015,
0.08647873252630234,
-0.08509670943021774,
-0.03154776245355606,
-0.13701291382312775,
0.10898718982934952,
0.1109699159860611,
0.008803862147033215,
-0.033845774829387665,
0.058503441512584686,
-0.0036109185311943293,
0.07545683532953262,
-0.03236141428351402,
-0.22564688324928284,
0.004441725090146065,
0.15670117735862732,
0.047048866748809814,
-0.08440861850976944,
0.0466732494533062,
0.05166344717144966,
0.05949703976511955,
-0.024220477789640427,
0.011674253270030022,
-0.047695618122816086,
0.1979600042104721,
-0.09563642740249634,
-0.1449352651834488,
-0.03255826607346535,
0.11577530950307846,
0.05092122405767441,
-0.006258761510252953,
-0.13681697845458984,
-0.10666979849338531,
0.1033659279346466,
-0.020063010975718498,
-0.05134858563542366,
-0.025868836790323257,
0.0018866955069825053,
0.06544151902198792,
-0.06818657368421555,
-0.05431871861219406,
-0.06781480461359024,
-0.0907665267586708,
0.20569109916687012,
0.04548761993646622,
0.050630297511816025,
-0.07318785786628723,
0.044468384236097336,
-0.12843117117881775,
-0.10330376029014587,
0.035723261535167694,
-0.07213455438613892,
-0.08094486594200134,
-0.008117018267512321,
-0.07901088893413544,
-0.24458736181259155,
0.11249540746212006,
0.1591739058494568,
-0.11413689702749252,
0.023858409374952316,
0.11659661680459976,
0.0681755319237709,
0.06769543141126633,
-0.044068317860364914,
-0.13283668458461761,
-0.12147266417741776,
0.021980497986078262,
0.020728150382637978,
0.06628220528364182,
0.019595853984355927,
-0.11531765013933182,
-0.0111357681453228,
-0.018337152898311615,
-0.009052423760294914,
0.0457947701215744,
0.11183793097734451,
-0.014212597161531448,
-0.07490608841180801,
0.12069857865571976,
-0.06401572376489639,
-0.0009044767357409,
0.0015972666442394257,
-0.005769402254372835,
-0.08942517638206482,
0.1363043487071991,
0.05685918405652046,
0.01986008696258068,
0.07784854620695114,
-0.03833550214767456,
-0.022700121626257896,
-0.09149064868688583,
-0.09166353940963745,
-0.015561992302536964,
-0.0035269493237137794,
-0.034474920481443405,
-0.1041935384273529,
-0.23184825479984283,
-0.06673423945903778,
-0.003008090192452073,
-0.0167993251234293,
-0.0057862671092152596,
-0.034993961453437805,
0.06884068995714188,
-0.011400953866541386,
-0.01335702370852232,
-0.040083084255456924,
-0.0030378070659935474,
-0.029023384675383568,
-0.03744013234972954,
0.01690589264035225,
-0.12439455837011337,
0.043461523950099945,
-0.06947531551122665,
0.12019909918308258,
-0.17173951864242554,
0.14761264622211456,
-0.010617638006806374,
0.0570593886077404,
-0.11335103958845139,
-0.014154725708067417,
-0.0074475775472819805,
0.029623208567500114,
0.06394011527299881,
0.18892192840576172,
-0.07565397024154663,
-0.06370799988508224,
0.10898382216691971,
-0.15811783075332642,
-0.10677758604288101,
0.13248348236083984,
0.0005673182313330472,
0.039013754576444626,
0.05550059676170349,
0.2354770302772522,
0.1440514326095581,
0.04166156426072121,
-0.012846787460148335,
-0.02227277308702469,
-0.04535696655511856,
0.01757011003792286,
0.07399287819862366,
-0.060085657984018326,
-0.12495262175798416,
0.02981247380375862,
0.05980315059423447,
0.1421540230512619,
-0.06612318009138107,
-0.039934370666742325,
-0.04066411778330803,
-0.07602005451917648,
0.15498101711273193,
-0.028510762378573418,
0.04034819081425667,
-0.03032013215124607,
0.0008507074089720845,
0.07673333585262299,
0.11908558011054993,
-0.05334082990884781,
-0.023210199549794197,
-0.06055228039622307,
0.15624216198921204,
-0.08581805974245071,
0.03735800459980965,
-0.11090632528066635,
-0.04707346856594086,
-0.03660906106233597,
-0.028143873438239098,
0.0600554421544075,
0.0395779013633728,
0.05712717026472092,
0.0812930092215538,
-0.04392392933368683,
-0.040906548500061035,
0.07622604817152023,
0.031853385269641876,
-0.04695535823702812,
-0.159401074051857,
-0.12689274549484253,
-0.06523746252059937,
0.2642826437950134,
-0.10274860262870789,
0.09030957520008087,
-0.08023979514837265,
0.22306548058986664,
-0.05538932979106903,
0.005113762803375721,
0.0552556999027729,
0.04318397492170334,
-0.030431855469942093,
0.02673659659922123,
0.07703975588083267,
0.014665765687823296,
-0.2017546147108078,
0.138865128159523,
-0.14598605036735535,
-0.03155485913157463,
0.05990435183048248,
0.03078712895512581,
-0.05611727014183998,
-0.12635616958141327,
-0.015236036852002144,
-0.04135280102491379,
0.05407281592488289,
-0.05947459861636162,
0.1204870268702507,
0.026617128401994705,
0.13911697268486023,
-0.031067118048667908,
0.02697308547794819,
-0.0075721037574112415,
-0.03831963613629341,
-0.04433383420109749,
0.05299091339111328,
-0.0027001379057765007,
-0.19107168912887573,
0.10140341520309448,
0.10201133042573929,
0.08355793356895447,
0.08764105290174484,
0.02080153487622738,
0.030412444844841957,
-0.06646081060171127,
0.009621544741094112,
-0.04388677701354027,
-0.06672907620668411,
-0.059814125299453735,
0.03309817612171173,
0.05572317913174629,
-0.013965809717774391,
0.07533956319093704,
-0.07735277712345123,
0.040564894676208496,
0.053608398884534836,
0.011298277415335178,
0.10264929383993149,
0.12920519709587097,
0.023774225264787674,
0.026928994804620743,
0.001405361806973815,
0.02377805858850479,
0.02110757865011692,
-0.025110386312007904,
-0.14003126323223114,
0.17331835627555847,
-0.09697332978248596,
-0.18623889982700348,
-0.21098539233207703,
-0.06057615578174591,
-0.14654959738254547,
0.006514827720820904,
0.05624721944332123,
-0.005526900757104158,
-0.05662225931882858,
-0.09207523614168167,
0.15136820077896118,
0.030032046139240265,
-0.0015023582382127643,
-0.02997763268649578,
-0.02987152338027954,
0.021524833515286446,
-0.06366635113954544,
-0.003786529880017042,
-0.012711330316960812,
0.021234575659036636,
0.051506202667951584,
-0.008109644055366516,
0.13257107138633728,
0.13024963438510895,
-0.027211803942918777,
-0.020717667415738106,
-0.011492032557725906,
0.18200890719890594,
-0.031962793320417404,
0.05111383646726608,
0.23789182305335999,
-0.08544868975877762,
0.054258137941360474,
0.22315268218517303,
0.02804017812013626,
-0.009534792974591255,
0.020052529871463776,
-0.056878410279750824,
-0.06507501006126404,
-0.16685634851455688,
-0.19644702970981598,
-0.07857033610343933,
-0.014846530742943287,
-0.0034583157394081354,
0.021777303889393806,
0.033165328204631805,
0.09578336030244827,
-0.09746982157230377,
-0.05017533898353577,
0.02352118119597435,
0.036099813878536224,
0.18382251262664795,
-0.02154732309281826,
0.09635226428508759,
-0.04138444736599922,
-0.010560935363173485,
0.04225078597664833,
-0.0721960961818695,
0.1021193265914917,
0.04732946678996086,
0.07622075080871582,
0.13246935606002808,
0.04537756368517876,
0.08500906080007553,
0.050415195524692535,
0.010206003673374653,
0.023258144035935402,
-0.021116921678185463,
-0.08371607214212418,
-0.024489780887961388,
0.07147719711065292,
0.0022904532961547375,
0.0670352503657341,
-0.10046742856502533,
0.03849126026034355,
0.05895315483212471,
0.14957763254642487,
0.09230504930019379,
-0.2440607100725174,
-0.10857847332954407,
0.04484729841351509,
0.042968541383743286,
0.006578405387699604,
-0.04573636129498482,
0.04411711543798447,
-0.09495852887630463,
0.18797354400157928,
-0.013604287058115005,
0.10388310998678207,
0.04643859341740608,
-0.009878255426883698,
-0.05888799950480461,
0.13662800192832947,
0.0033531200606375933,
0.05656268447637558,
-0.09442997723817825,
0.1771070659160614,
0.02307979390025139,
-0.02353842556476593,
0.02732444740831852,
0.030906202271580696,
0.008502867072820663,
0.24354548752307892,
0.02178032323718071,
-0.0015200501075014472,
-0.0024081943556666374,
-0.02782076597213745,
-0.09221149981021881,
-0.021869728341698647,
-0.016779636964201927,
-0.007738850079476833,
0.09083613753318787,
-0.044574979692697525,
-0.023444099351763725,
-0.009653910994529724,
0.09667373448610306,
-0.13100691139698029,
-0.1364358514547348,
0.03145826607942581,
0.04912374168634415,
0.08684170991182327,
-0.08435037732124329,
-0.024218063801527023,
-0.08457634598016739,
0.24806424975395203,
0.025502260774374008,
-0.10628342628479004,
-0.10828612744808197,
0.021117301657795906,
0.13267755508422852,
-0.0653638243675232,
0.04773271083831787,
-0.03439606726169586,
0.08114966750144958,
-0.09632830321788788,
-0.1478387713432312,
0.044549498707056046,
-0.0913463681936264,
-0.09610183537006378,
0.008635347709059715,
0.12495633214712143,
-0.00043654791079461575,
0.022259747609496117,
0.008391480892896652,
0.020194966346025467,
-0.05117525905370712,
-0.07859469950199127,
-0.03334537893533707,
0.23390059173107147,
0.04041838273406029,
0.2016979455947876,
0.004988301079720259,
-0.2671132981777191,
-0.01890367455780506,
-0.03931133449077606,
0.15882910788059235,
0.24089168012142181,
-0.0742856115102768,
0.09271016716957092,
0.17187565565109253,
-0.045198120176792145,
-0.21309629082679749,
-0.08431325852870941,
0.008376486599445343,
0.05801926180720329,
0.0024630757980048656,
0.021372303366661072,
0.005760069936513901,
0.07090860605239868,
0.0223477054387331,
-0.019357291981577873,
-0.33111077547073364,
-0.20007042586803436,
0.05570895969867706,
0.024747472256422043,
0.19169025123119354,
-0.09775710105895996,
-0.11197630316019058,
-0.10390988737344742,
-0.1528516709804535,
-0.04293832555413246,
-0.13481849431991577,
0.0492364875972271,
-0.028691070154309273,
0.020763952285051346,
0.0393962524831295,
-0.05597841739654541,
0.13634265959262848,
-0.03251199051737785,
0.04147310554981232,
-0.08551138639450073,
-0.03077055513858795,
0.03788722679018974,
-0.031560298055410385,
0.11537604033946991,
-0.08576514571905136,
0.04238816350698471,
-0.04024871438741684,
-0.04517588019371033,
-0.04635045677423477,
0.07417009025812149,
-0.023048587143421173,
-0.018160084262490273,
-0.08318857848644257,
0.005120753310620785,
0.013729599304497242,
-0.0034072750713676214,
0.10876990109682083,
-0.07078257948160172,
0.13689085841178894,
0.24839186668395996,
0.09304995834827423,
-0.0729423388838768,
-0.024099668487906456,
0.016884101554751396,
-0.05592348426580429,
0.045544613152742386,
-0.03204692527651787,
-0.02161281742155552,
0.10896724462509155,
0.005066742654889822,
0.15614712238311768,
0.013645690865814686,
-0.053354449570178986,
0.02758219465613365,
0.08427751809358597,
-0.160682812333107,
-0.35129037499427795,
-0.035758521407842636,
0.030840419232845306,
-0.07720448076725006,
0.04761664196848869,
0.1550634801387787,
-0.07674756646156311,
-0.003722059540450573,
-0.01304685603827238,
0.021954547613859177,
-0.10819321870803833,
0.11860671639442444,
-0.013681655749678612,
0.03833962604403496,
-0.0873882845044136,
0.05707070976495743,
0.04246305301785469,
-0.06697393953800201,
-0.037975527346134186,
0.06212885305285454,
-0.12234718352556229,
-0.0862002745270729,
-0.07543694227933884,
0.18723788857460022,
-0.06714173406362534,
-0.0886404812335968,
-0.09186343103647232,
-0.13389398157596588,
0.009809162467718124,
0.09460027515888214,
0.057625796645879745,
0.0033381865359842777,
-0.0018094575498253107,
-0.07972849160432816,
-0.03842144086956978,
0.05549291893839836,
0.06602329760789871,
0.049164000898599625,
-0.11887899041175842,
0.16230061650276184,
-0.005176091101020575,
0.06534524261951447,
-0.05852640047669411,
-0.007696274667978287,
-0.0863606184720993,
-0.019062664359807968,
-0.1568649411201477,
-0.01502181589603424,
-0.19992220401763916,
-0.04861157014966011,
-0.010531979613006115,
-0.04456561431288719,
-0.014786415733397007,
0.014676641672849655,
-0.022634314373135567,
-0.030330684036016464,
-0.03950631991028786,
0.04264817759394646,
-0.05754481256008148,
-0.030978143215179443,
-0.0240472499281168,
-0.02426919899880886,
0.04661165922880173,
0.031482744961977005,
-0.036742810159921646,
-0.05538663640618324,
-0.07054294645786285,
-0.08801160752773285,
0.033435940742492676,
0.009101114235818386,
0.002663383958861232,
-0.11606533825397491,
-0.05168473348021507,
0.002162947552278638,
-0.07302268594503403,
-0.03013649396598339,
0.06616384536027908,
-0.054397694766521454,
0.02954094298183918,
0.006883169058710337,
0.023355500772595406,
-0.06894368678331375,
-0.010478587821125984,
-0.00041197799146175385,
0.11370200663805008,
0.13253998756408691,
-0.0742400586605072,
0.0526493601500988,
-0.1615353375673294,
-0.01798221468925476,
-0.018757285550236702,
-0.06455186009407043,
-0.04566846787929535,
-0.09007443487644196,
0.00876540970057249,
0.00022421190806198865,
0.11943325400352478,
0.024187419563531876,
-0.09655886143445969,
0.030179474502801895,
0.0049468777142465115,
0.07233778387308121,
0.046632811427116394,
0.15093374252319336,
-0.007559187710285187,
0.042194414883852005,
-0.06067580729722977,
0.09633226692676544,
-0.013011271134018898,
-0.04433730989694595,
0.1017390787601471,
0.05881595239043236,
0.02009856142103672,
0.042190343141555786,
0.08202511072158813,
-0.0018387535819783807,
-0.0035640050191432238,
-0.17866550385951996,
-0.037982478737831116,
0.01842416077852249,
-0.042152244597673416,
0.13085946440696716,
0.1043601781129837,
-0.15830537676811218,
0.0920833945274353,
0.05433482676744461,
-0.02625524066388607,
-0.06825029850006104,
-0.009410199709236622,
-0.0807776153087616,
-0.14296740293502808,
-0.03315833583474159,
-0.06805068999528885,
-0.029761115089058876,
-0.02877122536301613,
-0.021971261128783226,
-0.006729380693286657,
0.13469713926315308,
-0.06500081717967987,
-0.03252546861767769,
-0.032736390829086304,
-0.03815548121929169,
0.008719287812709808,
0.0030977013520896435,
-0.04440481215715408,
0.07494556903839111,
-0.02284972369670868,
0.010869544930756092,
0.04853672534227371,
0.029079964384436607,
0.0729588195681572,
-0.01610686257481575,
-0.09062428027391434,
-0.0071656364016234875,
0.07866962254047394,
0.10801929235458374,
-0.029885875061154366,
0.04474612697958946,
-0.0430932380259037,
-0.03925381228327751,
0.04018751531839371,
-0.014124483801424503,
-0.05926411971449852,
-0.08674342185258865,
0.21631862223148346,
-0.015389678068459034,
0.04677000269293785,
0.022766191512346268,
-0.09943390637636185,
0.01167245488613844,
0.11646249890327454,
0.31458279490470886,
-0.006817500106990337,
0.009898044168949127,
-0.025610337033867836,
0.018385766074061394,
0.03510136529803276,
0.08510209619998932,
-0.007412495091557503,
0.13292230665683746,
-0.03513246029615402,
0.13508333265781403,
-0.053869444876909256,
-0.0785166546702385,
-0.04232504591345787,
-0.005243903025984764,
-0.034968696534633636,
-0.043462399393320084,
0.06842025369405746,
0.10859037935733795,
-0.07337610423564911,
-0.05257011577486992,
0.01654716581106186,
-0.1003066822886467,
-0.06077892333269119,
-0.0635291337966919,
0.04699564725160599,
0.03216360881924629,
0.05270390212535858,
-0.062145985662937164,
0.024127747863531113,
0.12106502801179886,
-0.0009345266735181212,
-0.08969509601593018,
-0.09212503582239151,
0.06659560650587082,
-0.03250013291835785,
0.003573925234377384,
-0.023334262892603874,
0.03708414360880852,
0.09897339344024658,
-0.009680620394647121,
-0.09147565811872482,
0.020190389826893806,
0.015118655748665333,
0.015114090405404568,
0.0638098493218422,
-0.009387319907546043,
-0.017603537067770958,
0.025901533663272858,
-0.020294584333896637,
-0.2131524682044983,
0.04584284499287605,
-0.035969171673059464,
-0.07941633462905884,
-0.06556347757577896,
0.11676303297281265,
-0.03121768683195114,
0.14073725044727325,
0.1451762616634369,
-0.058565571904182434,
0.01236865296959877,
0.010957813821732998,
0.06164548918604851,
0.049494050443172455,
0.16675898432731628,
-0.03390838950872421,
-0.21221408247947693,
0.027895668521523476,
0.018426816910505295,
0.03779345005750656,
-0.2584504783153534,
-0.07593251764774323,
-0.10832326859235764,
-0.03367210552096367,
-0.0022416403517127037,
0.12176313996315002,
0.16034968197345734,
0.023227864876389503,
-0.006905906833708286,
-0.2322801947593689,
-0.02082138881087303,
0.10201703011989594,
-0.105606310069561,
-0.0947953313589096
] |
null | null | transformers |
# Bangla LLaMA 13B Base v0.1 [pre-trained]
Welcome to the inaugural release of the Bangla LLaMA 13B base model – an important step in advancing LLMs for the Bangla language. This model is ready for immediate inference and is also primed for further fine-tuning to cater to your specific NLP tasks.
> **Please Note:** This model, labeled as a foundational Bangla Language Model (LLM), is designed primarily for Causal Language Modeling (LM) purposes. In other words, if you are looking for an instruction following model in Bangla, you may find [BanglaLLM/bangla-llama-13b-instruct-v0.1](https://huggingface.co/BanglaLLM/bangla-llama-13b-instruct-v0.1) more suitable for your needs.
## Model description
The Bangla LLaMA models have been enhanced and tailored specifically with an extensive Bangla vocabulary of 16,000 tokens, building upon the foundation set by the original LLaMA-2.
- **Model type:** A 13B parameter model for Causal LM pre-trained on [CulturaX](https://huggingface.co/datasets/uonlp/CulturaX) dataset's Bangla subset.
- **Language(s):** Bangla and English
- **License:** GNU General Public License v3.0
- **Source Model:** [meta-llama/Llama-2-13b-hf](https://huggingface.co/meta-llama/Llama-2-13b-hf)
- **Training Precision:** `float16`
- **Code:** [GitHub](https://github.com/abhinand5/bangla-llama)
## Related Models
| Model | Type | Data | Base Model | # Params | Download Links |
|--------------------------|-----------------------------|-------------------|----------------------|------|------------------------------------------------------------------------|
| Bangla LLaMA 7B Base | Base model | 12GB | LLaMA 7B | 7B | [HF Hub](https://huggingface.co/BanglaLLM/bangla-llama-7b-base-v0.1) |
| Bangla LLaMA 13B Base | Base model | 4GB | LLaMA 13B | 13B | [HF Hub](https://huggingface.co/BanglaLLM/bangla-llama-13b-base-v0.1) |
| Bangla LLaMA 7B Instruct | Instruction following model | 145k instructions | Bangla LLaMA 7B Base | 7B | [HF Hub](https://huggingface.co/BanglaLLM/bangla-llama-7b-instruct-v0.1) |
| Bangla LLaMA 13B Instruct | Instruction following model | 145k instructions | Bangla LLaMA 13B Base | 13B | [HF Hub](BanglaLLM/bangla-llama-13b-instruct-v0.1) |
## Usage Note
It's important to note that the models have not undergone detoxification. Therefore, while they possess impressive linguistic capabilities, there is a possibility for them to generate content that could be deemed harmful or offensive. We urge users to exercise discretion and supervise the model's outputs closely, especially in public or sensitive applications.
## Meet the Developers
Get to know the creators behind this innovative model and follow their contributions to the field:
- [Abdullah Khan Zehady](https://www.linkedin.com/in/abdullah-khan-zehady-915ba024/)
## Citation
We hope this model serves as a valuable tool in your NLP toolkit and look forward to seeing the advancements it will enable in the understanding and generation of the Bangla language. | {"language": ["bn", "en"], "license": "llama2"} | text-generation | BanglaLLM/bangla-llama-13b-base-v0.1 | [
"transformers",
"pytorch",
"llama",
"text-generation",
"bn",
"en",
"license:llama2",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T07:04:13+00:00 | [] | [
"bn",
"en"
] | TAGS
#transformers #pytorch #llama #text-generation #bn #en #license-llama2 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| Bangla LLaMA 13B Base v0.1 [pre-trained]
========================================
Welcome to the inaugural release of the Bangla LLaMA 13B base model – an important step in advancing LLMs for the Bangla language. This model is ready for immediate inference and is also primed for further fine-tuning to cater to your specific NLP tasks.
>
> Please Note: This model, labeled as a foundational Bangla Language Model (LLM), is designed primarily for Causal Language Modeling (LM) purposes. In other words, if you are looking for an instruction following model in Bangla, you may find BanglaLLM/bangla-llama-13b-instruct-v0.1 more suitable for your needs.
>
>
>
Model description
-----------------
The Bangla LLaMA models have been enhanced and tailored specifically with an extensive Bangla vocabulary of 16,000 tokens, building upon the foundation set by the original LLaMA-2.
* Model type: A 13B parameter model for Causal LM pre-trained on CulturaX dataset's Bangla subset.
* Language(s): Bangla and English
* License: GNU General Public License v3.0
* Source Model: meta-llama/Llama-2-13b-hf
* Training Precision: 'float16'
* Code: GitHub
Related Models
--------------
Usage Note
----------
It's important to note that the models have not undergone detoxification. Therefore, while they possess impressive linguistic capabilities, there is a possibility for them to generate content that could be deemed harmful or offensive. We urge users to exercise discretion and supervise the model's outputs closely, especially in public or sensitive applications.
Meet the Developers
-------------------
Get to know the creators behind this innovative model and follow their contributions to the field:
* Abdullah Khan Zehady
We hope this model serves as a valuable tool in your NLP toolkit and look forward to seeing the advancements it will enable in the understanding and generation of the Bangla language.
| [] | [
"TAGS\n#transformers #pytorch #llama #text-generation #bn #en #license-llama2 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
57
] | [
"passage: TAGS\n#transformers #pytorch #llama #text-generation #bn #en #license-llama2 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
0.00039963179733604193,
0.03100491687655449,
-0.007689902558922768,
0.0236776452511549,
0.1258000135421753,
0.01841018721461296,
0.1510808914899826,
0.11516890674829483,
0.008001531474292278,
-0.03450804203748703,
0.14374718070030212,
0.21446096897125244,
-0.01299569383263588,
0.005974415689706802,
-0.07826937735080719,
-0.243387833237648,
0.033496931195259094,
0.062039460986852646,
0.028200756758451462,
0.1104864850640297,
0.0843810886144638,
-0.07109877467155457,
0.07808190584182739,
-0.02887132577598095,
-0.13169366121292114,
0.03530477359890938,
0.027938170358538628,
-0.10670400410890579,
0.1096852570772171,
0.07201817631721497,
0.08726633340120316,
0.04146689176559448,
-0.040089502930641174,
-0.19912168383598328,
0.027764474973082542,
-0.03002116084098816,
-0.07767846435308456,
0.05437033623456955,
0.06572148203849792,
-0.06171515956521034,
0.12834538519382477,
0.09635129570960999,
-0.031747713685035706,
0.054592665284872055,
-0.13066156208515167,
-0.029740307480096817,
-0.0346870981156826,
0.04474359378218651,
0.06027894839644432,
0.07389005273580551,
0.007491770666092634,
0.090741828083992,
-0.09369948506355286,
0.08351615071296692,
0.15510603785514832,
-0.3430435359477997,
0.014169045723974705,
0.09794238209724426,
0.08391988277435303,
0.03857037425041199,
-0.044311728328466415,
0.08596581220626831,
0.04933265596628189,
-0.001197698642499745,
0.0338626466691494,
-0.08480820804834366,
-0.059534136205911636,
0.06436925381422043,
-0.07172621786594391,
-0.05111704394221306,
0.2219865471124649,
-0.04604332894086838,
0.05342669039964676,
-0.02978081814944744,
-0.06475381553173065,
-0.04335566982626915,
-0.00979163683950901,
0.03352437540888786,
-0.017983121797442436,
0.07701990008354187,
0.054583579301834106,
-0.0528685599565506,
-0.1293538510799408,
-0.008890986442565918,
-0.20501655340194702,
0.14181284606456757,
0.015383665449917316,
0.043161239475011826,
-0.17773957550525665,
0.09239122271537781,
0.053518809378147125,
-0.10387299209833145,
0.007060840260237455,
-0.05644809082150459,
0.09187544882297516,
0.0212559811770916,
-0.05222187936306,
-0.019018951803445816,
0.09104157239198685,
0.12702420353889465,
0.04100329801440239,
-0.009672271087765694,
-0.08290392905473709,
0.12921437621116638,
0.012458717450499535,
0.05748553201556206,
0.007262629456818104,
-0.036011748015880585,
0.05796250328421593,
-0.09432518482208252,
0.03589311242103577,
-0.03703644499182701,
-0.17928047478199005,
-0.034440070390701294,
0.014073411002755165,
0.10480373352766037,
0.006072054151445627,
0.0699680745601654,
-0.051366399973630905,
-0.008546954020857811,
0.018316250294446945,
-0.07984145730733871,
-0.008999865502119064,
-0.000935689895413816,
0.001041586510837078,
0.08508473634719849,
0.024133743718266487,
0.01347039919346571,
-0.08499184250831604,
0.0809287279844284,
-0.07874024659395218,
-0.013879086822271347,
-0.05372854322195053,
-0.04411569982767105,
0.05348232015967369,
-0.10544142872095108,
0.043818648904561996,
-0.1459655463695526,
-0.1950463503599167,
0.030450081452727318,
0.01354280300438404,
-0.038217708468437195,
-0.04624710604548454,
-0.028230538591742516,
-0.03482498973608017,
0.011380515992641449,
-0.09090231359004974,
0.02202509343624115,
-0.07254661619663239,
0.11759715527296066,
-0.03644835203886032,
0.05348995700478554,
-0.15396854281425476,
0.06720874458551407,
-0.09738647192716599,
0.012428635731339455,
-0.022840695455670357,
0.034559376537799835,
-0.01869348995387554,
0.10808514058589935,
-0.013729255646467209,
-0.01306686457246542,
-0.05574223771691322,
0.055200137197971344,
-0.03884982317686081,
0.1699940711259842,
-0.1131235882639885,
-0.10601221770048141,
0.1754467487335205,
-0.07498863339424133,
-0.17364627122879028,
0.07280892878770828,
-0.005117499735206366,
0.07820719480514526,
0.08614000678062439,
0.18422594666481018,
0.017513474449515343,
-0.03602069616317749,
0.06161961331963539,
0.0902368500828743,
-0.054031144827604294,
-0.14957118034362793,
0.03188066557049751,
-0.005349782761186361,
-0.08195944130420685,
0.05082157254219055,
0.065046526491642,
0.06953375786542892,
-0.0414699949324131,
-0.06104915961623192,
-0.030839843675494194,
-0.02422269806265831,
0.004263648763298988,
0.00394799979403615,
0.09564094245433807,
-0.04644758626818657,
-0.005914347246289253,
0.07772733271121979,
0.0067346165888011456,
0.008949050679802895,
0.043241363018751144,
-0.0529877170920372,
0.1312638223171234,
-0.026734910905361176,
0.036227595061063766,
-0.15185783803462982,
-0.03369409963488579,
-0.024943390861153603,
0.07869657874107361,
0.038206689059734344,
0.07918420433998108,
0.030362272635102272,
-0.023256491869688034,
-0.0336247980594635,
0.02075224742293358,
0.14126934111118317,
0.00003593964720494114,
-0.05773024633526802,
-0.10274713486433029,
0.05236048623919487,
-0.045151639729738235,
-0.022559907287359238,
-0.07138907164335251,
0.015229917131364346,
0.03482264280319214,
0.09614717960357666,
-0.01369383279234171,
0.0682544857263565,
-0.02554209530353546,
0.026217713952064514,
-0.08818408101797104,
0.04260149225592613,
0.11946795135736465,
-0.01610274240374565,
-0.09507033973932266,
0.18824899196624756,
-0.1462714821100235,
0.2621750831604004,
0.2076086550951004,
-0.23637942969799042,
0.05158565193414688,
-0.09068156033754349,
-0.03184454143047333,
0.00011081383127020672,
0.04597749188542366,
-0.03011121228337288,
0.06895357370376587,
0.008235706016421318,
0.18362103402614594,
-0.07548843324184418,
-0.0601915642619133,
-0.015899714082479477,
-0.04350297898054123,
-0.030198469758033752,
0.09503760188817978,
0.1815510094165802,
-0.13465790450572968,
0.18086834251880646,
0.2497747838497162,
0.019733255729079247,
0.20077042281627655,
-0.027697140350937843,
-0.02211826667189598,
0.037999480962753296,
-0.023809025064110756,
-0.03180379420518875,
-0.012925046496093273,
-0.14915169775485992,
-0.004467676859349012,
0.0744965523481369,
0.008714205585420132,
0.08870827406644821,
-0.14132380485534668,
-0.07709675282239914,
-0.02971075475215912,
-0.019772343337535858,
0.0013503945665434003,
0.09013306349515915,
0.027144264429807663,
0.12263942509889603,
-0.03793177381157875,
-0.05973798781633377,
0.09827511012554169,
-0.0037013404071331024,
-0.07750886678695679,
0.17771822214126587,
-0.11437850445508957,
-0.2618994116783142,
-0.17897330224514008,
-0.1673201471567154,
-0.05946248024702072,
0.022734642028808594,
0.10482608526945114,
-0.04227307438850403,
-0.02192295528948307,
-0.009026127867400646,
0.06655540317296982,
-0.0793808326125145,
-0.0013066905084997416,
-0.05718133971095085,
0.059615105390548706,
-0.08615455776453018,
-0.11097824573516846,
-0.06279086321592331,
-0.021967973560094833,
-0.04083240404725075,
0.1102355569601059,
-0.09565897285938263,
0.06649607419967651,
0.16439330577850342,
0.03116457536816597,
0.03239762410521507,
-0.03801896795630455,
0.17329491674900055,
-0.05954023823142052,
-0.006490441504865885,
0.193914532661438,
-0.038687895983457565,
0.06656859070062637,
0.1551823914051056,
0.04660075902938843,
-0.095100998878479,
0.01335546001791954,
-0.017504315823316574,
-0.08511221408843994,
-0.2463100403547287,
-0.10829798132181168,
-0.11480458825826645,
0.06610159575939178,
0.019025769084692,
0.06686916947364807,
0.1354481279850006,
0.06429518759250641,
-0.00926440954208374,
0.0030091649387031794,
0.009165347553789616,
0.08137744665145874,
0.2956153452396393,
-0.031813692301511765,
0.1312772035598755,
-0.06961861252784729,
-0.10502154380083084,
0.07673409581184387,
0.1002812534570694,
0.13174669444561005,
0.11253007501363754,
0.07132120430469513,
0.061690736562013626,
0.0599687285721302,
0.12253932654857635,
0.10508326441049576,
0.003159754676744342,
-0.011339602060616016,
-0.03999870643019676,
-0.043708622455596924,
-0.02142428234219551,
0.047607824206352234,
-0.0050253127701580524,
-0.1028086245059967,
-0.031894609332084656,
-0.13346843421459198,
0.047399844974279404,
0.09125497192144394,
0.03790725767612457,
-0.18087363243103027,
0.04328330233693123,
0.10798877477645874,
-0.03998284786939621,
-0.08407087624073029,
0.09623529762029648,
-0.031238799914717674,
-0.09580818563699722,
0.05858447030186653,
-0.026253830641508102,
0.1276492029428482,
-0.051987141370773315,
0.08364824205636978,
-0.060172196477651596,
-0.10380727797746658,
0.03713517263531685,
0.11134345829486847,
-0.323539137840271,
0.23269937932491302,
0.009172078222036362,
-0.05310484394431114,
-0.07432928681373596,
-0.01306238304823637,
0.00023747466912027448,
0.1315969079732895,
0.11559757590293884,
-0.014860003255307674,
0.00027950279763899744,
-0.07618142664432526,
-0.014188428409397602,
0.025274712592363358,
0.09500951319932938,
-0.04732028767466545,
0.008736209012567997,
-0.04101960361003876,
0.007046594750136137,
-0.011481297202408314,
0.024814195930957794,
0.00499182753264904,
-0.16421374678611755,
0.06389656662940979,
0.0604083351790905,
0.06887499988079071,
-0.00031052669510245323,
-0.03977123647928238,
-0.11082306504249573,
0.17105212807655334,
-0.13845892250537872,
-0.09025376290082932,
-0.0942036584019661,
-0.09851621836423874,
0.059110887348651886,
-0.05723205581307411,
0.03574329987168312,
-0.0750390961766243,
-0.006368307396769524,
-0.08507581800222397,
-0.1988401859998703,
0.10183899104595184,
-0.07478971779346466,
-0.03388943895697594,
-0.04795410856604576,
0.18294352293014526,
-0.0889466181397438,
0.024605775251984596,
0.02199610322713852,
0.01699426956474781,
-0.09087175130844116,
-0.09714306145906448,
-0.005901865195482969,
0.009568894281983376,
0.04527965188026428,
-0.021556010469794273,
-0.15255442261695862,
0.011052114889025688,
-0.01483594998717308,
-0.06511756777763367,
0.2675971984863281,
0.2270585000514984,
-0.046767108142375946,
0.1808474063873291,
0.15788178145885468,
-0.13351255655288696,
-0.3002731502056122,
-0.1326807141304016,
-0.14033588767051697,
-0.030965834856033325,
-0.01453777402639389,
-0.18689578771591187,
0.08598260581493378,
0.04600173979997635,
-0.03159858658909798,
0.14793355762958527,
-0.2820972204208374,
-0.1049380823969841,
0.1565587967634201,
0.007156871259212494,
0.3669073283672333,
-0.16150420904159546,
-0.10656687617301941,
-0.0712006613612175,
-0.1384093016386032,
0.16844354569911957,
0.0010301837464794517,
0.12082559615373611,
-0.05104940012097359,
0.10170739144086838,
0.022365858778357506,
-0.033418554812669754,
0.10652504861354828,
0.01579093374311924,
0.03189196065068245,
-0.10415566712617874,
-0.04816662520170212,
0.05155651643872261,
0.008032895624637604,
0.019926514476537704,
-0.1270199865102768,
0.003662294941022992,
-0.15354059636592865,
-0.04203014075756073,
-0.05461868643760681,
0.07425397634506226,
0.0013892152346670628,
-0.07348562777042389,
-0.022757483646273613,
-0.030256150290369987,
-0.003895790781825781,
-0.005072197411209345,
0.21954123675823212,
-0.06302928924560547,
0.1299687922000885,
0.163323774933815,
0.12175735831260681,
-0.11522255837917328,
0.02555488795042038,
-0.0877765342593193,
-0.0668373554944992,
0.0727396309375763,
-0.10998402535915375,
0.02069256454706192,
0.14341183006763458,
-0.026782136410474777,
0.06981410086154938,
0.08857400715351105,
0.004234542604535818,
-0.009626478888094425,
0.1428283452987671,
-0.20185548067092896,
-0.03375408425927162,
-0.038435522466897964,
-0.0005451614269986749,
0.09166592359542847,
0.03876584768295288,
0.1452634632587433,
-0.005026735365390778,
-0.014114901423454285,
0.024004202336072922,
0.008838124573230743,
-0.05498720332980156,
0.06101951375603676,
0.02293020859360695,
0.019846251234412193,
-0.12817628681659698,
0.0873248279094696,
0.022012460976839066,
-0.092504121363163,
-0.003958105109632015,
0.14066749811172485,
-0.10385801643133163,
-0.11143277585506439,
-0.07067035138607025,
0.10489288717508316,
-0.1946348249912262,
-0.03520084545016289,
-0.04835576191544533,
-0.16183321177959442,
0.0873117744922638,
0.2202673703432083,
0.07331735640764236,
0.08791062235832214,
-0.04319976270198822,
-0.047330982983112335,
-0.005841527134180069,
0.030677368864417076,
-0.04429720714688301,
0.01982804201543331,
-0.07101590931415558,
0.0031806963961571455,
-0.010355575941503048,
0.11982230097055435,
-0.07662578672170639,
-0.05004839971661568,
-0.14686159789562225,
0.05061836540699005,
-0.1810462921857834,
-0.04521346464753151,
-0.06777427345514297,
-0.03953879699110985,
-0.009646476246416569,
-0.03344302624464035,
-0.06548214703798294,
-0.05619901791214943,
-0.11827033758163452,
0.017969006672501564,
-0.025751644745469093,
0.07484845817089081,
-0.08728618174791336,
-0.026478281244635582,
0.08515587449073792,
-0.021757252514362335,
0.10917794704437256,
0.07510827481746674,
-0.0677291676402092,
0.10833549499511719,
-0.12784937024116516,
-0.053884461522102356,
0.10447347164154053,
0.032765574753284454,
0.027659205719828606,
0.05324752256274223,
0.01077474094927311,
0.09132832288742065,
0.01855727843940258,
0.06156758964061737,
-0.015499837696552277,
-0.1436333805322647,
-0.006365438457578421,
-0.02490742690861225,
-0.13857991993427277,
-0.03356311842799187,
-0.05689767003059387,
0.05345777049660683,
0.010894627310335636,
0.14100684225559235,
-0.034903667867183685,
0.0644727274775505,
-0.056309767067432404,
0.03869614377617836,
0.0024324399419128895,
-0.17695015668869019,
-0.07833526283502579,
-0.1029946357011795,
-0.015617404133081436,
-0.00394660048186779,
0.28781893849372864,
0.0353301465511322,
-0.07416320592164993,
0.043365247547626495,
0.0789460837841034,
-0.005395420361310244,
0.0038961386308073997,
0.2684534788131714,
0.1066509336233139,
-0.019473470747470856,
-0.1135689914226532,
0.05608801916241646,
-0.002037764061242342,
0.013041957281529903,
0.11043490469455719,
0.0913643166422844,
-0.01205997820943594,
0.08614636957645416,
0.06347724795341492,
0.023758916184306145,
-0.051706746220588684,
-0.08634679764509201,
0.03522501513361931,
0.06731830537319183,
-0.028662238270044327,
0.1422596126794815,
0.17377819120883942,
-0.040702756494283676,
0.04073568806052208,
-0.017139503732323647,
-0.021096577867865562,
-0.17280647158622742,
-0.15286466479301453,
-0.06190706044435501,
-0.07691023498773575,
0.024390071630477905,
-0.08636157214641571,
0.047756582498550415,
0.07257653772830963,
0.06518800556659698,
-0.06770561635494232,
0.04017225652933121,
0.017153406515717506,
-0.08305252343416214,
0.04920182377099991,
-0.024762101471424103,
0.04117054119706154,
-0.09521535784006119,
0.0036345170810818672,
-0.07936947047710419,
-0.032129742205142975,
-0.05019708722829819,
0.06486977636814117,
-0.022846577689051628,
0.0010149532463401556,
-0.15020166337490082,
-0.08207627385854721,
-0.04885214567184448,
0.0546828992664814,
0.005891055334359407,
0.15040379762649536,
-0.012516397051513195,
-0.022556748241186142,
0.04558714106678963,
0.177068829536438,
-0.04757624492049217,
-0.09768015146255493,
0.021664775907993317,
0.1974608153104782,
0.03539074584841728,
0.09825435280799866,
-0.02498006820678711,
0.016936950385570526,
-0.06865201145410538,
0.3563738465309143,
0.30281028151512146,
-0.08957317471504211,
0.01871534436941147,
0.02771264873445034,
0.04449569061398506,
0.10652880370616913,
0.10991809517145157,
0.09204764664173126,
0.2870418429374695,
-0.07128086686134338,
-0.02773047238588333,
-0.05588487535715103,
0.0006315141217783093,
-0.09961120784282684,
0.1186366006731987,
0.03509176895022392,
-0.08305145800113678,
-0.04625101014971733,
0.0734371766448021,
-0.2277628630399704,
0.10613082349300385,
-0.047711215913295746,
-0.15855270624160767,
-0.04416690394282341,
-0.01734999567270279,
0.13248442113399506,
-0.000570899632293731,
0.043321844190359116,
-0.038000669330358505,
-0.081830233335495,
0.027866769582033157,
0.012120101600885391,
-0.24073727428913116,
-0.00042014484643004835,
0.0712294653058052,
0.002631555777043104,
0.016302291303873062,
-0.018516026437282562,
0.05879293382167816,
0.05746180936694145,
0.048405468463897705,
-0.03885558620095253,
0.08551619201898575,
0.03259602561593056,
-0.056688763201236725,
0.022426128387451172,
-0.0017793134320527315,
0.02331581711769104,
-0.05117538943886757,
0.04963437840342522,
-0.03932167589664459,
0.05729499086737633,
-0.03497404232621193,
-0.08841677010059357,
-0.013087786734104156,
0.023441670462489128,
-0.06566902250051498,
0.05974689871072769,
0.06132481247186661,
-0.0057695587165653706,
-0.029465172439813614,
-0.08071205765008926,
-0.006587418261915445,
0.014502249658107758,
-0.13234500586986542,
-0.06485714018344879,
-0.07300214469432831,
-0.07915813475847244,
0.09436783939599991,
-0.003475506789982319,
-0.24987107515335083,
-0.002151616383343935,
-0.09002591669559479,
0.04427620768547058,
-0.18205176293849945,
0.06975547969341278,
0.1406642198562622,
0.013302133418619633,
-0.00869064312428236,
-0.14345841109752655,
0.053318049758672714,
0.03474859893321991,
-0.11295264959335327,
-0.08103921264410019
] |
null | null | transformers |
### Developed by Jaeyeon Kang ([CCK Solution](https://www.ccksolution.co.kr/))
## Model Description
- This is a model for the korean language.
- We fine-tuned this model based on [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf) with our training dataset(private).
- SFT (QLoRA) training are adjusted .
## Run the model
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "JaeyeonKang/CCK-Asura-v2"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id)
text = "[INST] Put instruction here. [/INST]"
inputs = tokenizer(text, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=20)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
``` | {"language": ["ko"], "license": "cc-by-nc-4.0"} | text-generation | JaeyeonKang/CCK_Asura_v2 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"ko",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T07:05:29+00:00 | [] | [
"ko"
] | TAGS
#transformers #safetensors #llama #text-generation #conversational #ko #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
### Developed by Jaeyeon Kang (CCK Solution)
## Model Description
- This is a model for the korean language.
- We fine-tuned this model based on 152334H/miqu-1-70b-sf with our training dataset(private).
- SFT (QLoRA) training are adjusted .
## Run the model
| [
"### Developed by Jaeyeon Kang (CCK Solution)",
"## Model Description\n- This is a model for the korean language.\n- We fine-tuned this model based on 152334H/miqu-1-70b-sf with our training dataset(private).\n- SFT (QLoRA) training are adjusted .",
"## Run the model"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #conversational #ko #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Developed by Jaeyeon Kang (CCK Solution)",
"## Model Description\n- This is a model for the korean language.\n- We fine-tuned this model based on 152334H/miqu-1-70b-sf with our training dataset(private).\n- SFT (QLoRA) training are adjusted .",
"## Run the model"
] | [
64,
14,
59,
4
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #conversational #ko #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Developed by Jaeyeon Kang (CCK Solution)## Model Description\n- This is a model for the korean language.\n- We fine-tuned this model based on 152334H/miqu-1-70b-sf with our training dataset(private).\n- SFT (QLoRA) training are adjusted .## Run the model"
] | [
-0.02417992800474167,
-0.10156659036874771,
0.0007959955837577581,
0.012417858466506004,
0.14791925251483917,
-0.07627315074205399,
0.15789994597434998,
0.06161500886082649,
-0.004570641089230776,
-0.024070626124739647,
0.10452509671449661,
0.14482615888118744,
0.01662067510187626,
0.14301925897598267,
-0.07653702795505524,
-0.30433276295661926,
0.09204935282468796,
0.010981855913996696,
0.06380762159824371,
0.11077581346035004,
0.12026044726371765,
-0.03768236190080643,
0.12141374498605728,
0.007498616352677345,
-0.09766566008329391,
-0.0021225144155323505,
-0.04421031102538109,
-0.08438463509082794,
0.11406133323907852,
0.005074905697256327,
0.0838238000869751,
0.1028083860874176,
0.030753280967473984,
-0.14990545809268951,
0.0310327410697937,
-0.06438574939966202,
-0.044691432267427444,
0.029985597357153893,
-0.04258424788713455,
0.025523237884044647,
0.216067835688591,
-0.040263064205646515,
-0.05478474870324135,
0.01735602505505085,
-0.06700248271226883,
0.054334431886672974,
-0.013013883493840694,
0.03536156192421913,
0.2622040808200836,
0.10890019685029984,
-0.029538150876760483,
0.08360875397920609,
-0.126357302069664,
0.10164918750524521,
-0.07570002973079681,
-0.3887937366962433,
-0.030661050230264664,
0.23784129321575165,
0.04971190169453621,
0.06454445421695709,
-0.0590074360370636,
0.05323628708720207,
0.0864291712641716,
-0.005096848122775555,
0.05621109530329704,
-0.10282188653945923,
-0.08688712865114212,
0.02251148782670498,
-0.09509912133216858,
0.0749102383852005,
0.2795616090297699,
-0.013626609928905964,
-0.03468071296811104,
-0.08948935568332672,
0.006101381499320269,
-0.09605483710765839,
-0.02223828062415123,
-0.022074587643146515,
-0.04023435339331627,
0.04108208045363426,
-0.008528201840817928,
-0.07952738553285599,
-0.11491642892360687,
-0.14127376675605774,
-0.09541921317577362,
0.17534710466861725,
0.029946230351924896,
-0.0004879003972746432,
-0.14292187988758087,
0.09285867214202881,
-0.10146848857402802,
-0.10870799422264099,
-0.053833018988370895,
-0.08265382796525955,
-0.010546743869781494,
-0.019097428768873215,
-0.031977977603673935,
-0.068741075694561,
0.046609070152044296,
0.01620200090110302,
0.002130471868440509,
0.07426692545413971,
0.001248452696017921,
0.06793957948684692,
-0.01853960007429123,
0.1198672503232956,
0.0009550973190926015,
0.05084550380706787,
0.03257302939891815,
0.03657795861363411,
0.009572142735123634,
-0.0037762403953820467,
-0.17428478598594666,
-0.06449674069881439,
-0.019763998687267303,
0.06939166784286499,
-0.12373372912406921,
0.19178277254104614,
0.05970156937837601,
0.05892983078956604,
0.10348473489284515,
-0.04868369176983833,
-0.07036350667476654,
-0.012919729575514793,
-0.034294985234737396,
-0.03686175122857094,
0.04168322682380676,
0.061066050082445145,
0.0029753490816801786,
0.011035636998713017,
0.0019926794338971376,
-0.026019027456641197,
-0.07023316621780396,
-0.057280778884887695,
-0.012154524214565754,
-0.08101043850183487,
0.054573383182287216,
-0.17982585728168488,
-0.20808780193328857,
-0.027523333206772804,
-0.05420074239373207,
-0.03861001878976822,
-0.027611058205366135,
-0.16882728040218353,
-0.12061148881912231,
0.008137769997119904,
-0.043343424797058105,
0.05349041149020195,
-0.04818490892648697,
0.024933382868766785,
-0.005732635967433453,
0.0730881616473198,
-0.07610144466161728,
0.02115483582019806,
-0.08884681761264801,
0.0353640615940094,
-0.08856016397476196,
0.09575299173593521,
-0.04049772769212723,
0.04575230926275253,
-0.04142564535140991,
-0.023074571043252945,
-0.10755763202905655,
0.09552676975727081,
-0.015123458579182625,
0.20691359043121338,
-0.14808978140354156,
0.018624642863869667,
0.17882412672042847,
-0.09976989775896072,
-0.14730289578437805,
0.09090621024370193,
-0.024677369743585587,
0.17781244218349457,
0.06676792353391647,
0.13692550361156464,
0.08242515474557877,
-0.08556697517633438,
0.01647823676466942,
0.13387005031108856,
-0.09784341603517532,
-0.10786820948123932,
0.060079626739025116,
0.06000254675745964,
-0.17259152233600616,
0.05954087898135185,
-0.10156913846731186,
0.06068141385912895,
-0.04726293310523033,
-0.035260118544101715,
-0.026391757652163506,
-0.08359828591346741,
0.021600494161248207,
-0.05066287890076637,
0.13850148022174835,
-0.025367800146341324,
0.015237731859087944,
0.024190563708543777,
0.10081048309803009,
-0.05031192675232887,
-0.0330132432281971,
-0.12075954675674438,
-0.03984342887997627,
0.01960010826587677,
0.03289403021335602,
-0.04414308816194534,
-0.011683085933327675,
-0.014542612247169018,
0.07294557243585587,
0.03138519451022148,
0.08555804938077927,
0.04734331741929054,
0.028499126434326172,
-0.05703999474644661,
0.012395791709423065,
0.09564081579446793,
0.05654171481728554,
-0.04956221580505371,
-0.10363198816776276,
0.04705429822206497,
-0.04248621314764023,
0.14601784944534302,
-0.16394568979740143,
0.05016310140490532,
-0.012995010241866112,
0.03573121502995491,
-0.03084946610033512,
0.10273191332817078,
0.08139150589704514,
0.09948669373989105,
-0.009290959686040878,
0.03523626923561096,
0.051864512264728546,
0.04669414833188057,
-0.15940912067890167,
0.2873245179653168,
-0.1733347624540329,
-0.008434351533651352,
0.1445460468530655,
-0.029562577605247498,
-0.02583695948123932,
-0.08596035838127136,
-0.00894076656550169,
-0.03659532591700554,
-0.02708391472697258,
-0.008760242722928524,
0.21013571321964264,
-0.038824670016765594,
0.20417520403862,
-0.07273203879594803,
-0.013568511232733727,
-0.011232146993279457,
-0.0883379727602005,
-0.008573313243687153,
0.11526140570640564,
0.10184729099273682,
-0.20096957683563232,
0.10682091116905212,
0.04094662517309189,
-0.06501837819814682,
0.2366943657398224,
0.005662852898240089,
0.0030389726161956787,
-0.013476785272359848,
0.08476054668426514,
-0.039074305444955826,
0.03333137929439545,
-0.12989239394664764,
-0.049112316220998764,
0.024826468899846077,
0.024031903594732285,
0.09952563047409058,
-0.13894957304000854,
-0.06829563528299332,
0.010646716691553593,
-0.04500884190201759,
-0.023414645344018936,
0.11400801688432693,
-0.021871522068977356,
0.08104907721281052,
-0.021259894594550133,
-0.09758585691452026,
0.05461150407791138,
-0.01272839866578579,
-0.12493105977773666,
0.19474786520004272,
-0.08136027306318283,
-0.16073711216449738,
-0.16047614812850952,
-0.01240622065961361,
-0.034612976014614105,
0.008346754126250744,
0.03570648655295372,
-0.14089274406433105,
-0.03698577731847763,
-0.10213842988014221,
-0.018022239208221436,
-0.009483234025537968,
-0.02474493905901909,
-0.04258456453680992,
0.07560114562511444,
-0.09644845873117447,
-0.07377408444881439,
-0.05066698044538498,
-0.02961076609790325,
-0.12456253916025162,
0.07943577319383621,
-0.18262746930122375,
0.04313566908240318,
0.12390875071287155,
0.022069670259952545,
0.030328040942549706,
-0.05464817211031914,
0.08505808562040329,
-0.0911887064576149,
0.01665680482983589,
0.15520313382148743,
-0.026625312864780426,
-0.017580155283212662,
0.15714769065380096,
-0.003088661702349782,
-0.0332459919154644,
0.11103290319442749,
-0.03211377188563347,
-0.11637376248836517,
-0.18825922906398773,
-0.0741976946592331,
-0.06443573534488678,
0.09793635457754135,
-0.021759960800409317,
0.05437200143933296,
0.12575434148311615,
0.12456750869750977,
0.005914285313338041,
0.029434623196721077,
0.045992862433195114,
0.04905139282345772,
0.07186874747276306,
0.01808454841375351,
0.09219513088464737,
-0.08708688616752625,
-0.06755344569683075,
0.02975759655237198,
0.03216206282377243,
0.05048863962292671,
0.034039948135614395,
-0.004188492428511381,
0.06753997504711151,
0.16311129927635193,
0.13281279802322388,
0.09076277911663055,
-0.022954406216740608,
-0.05341581627726555,
0.0007966923294588923,
-0.08505851030349731,
-0.03592684119939804,
0.11383355408906937,
-0.07766050100326538,
-0.07766775786876678,
-0.050152964890003204,
0.1430254578590393,
0.05571107938885689,
0.097465880215168,
0.04972701519727707,
-0.21979819238185883,
-0.06192401796579361,
0.03621475398540497,
0.03631650283932686,
-0.039287954568862915,
0.1257476657629013,
0.20287327468395233,
-0.14030727744102478,
0.04550259932875633,
0.023834381252527237,
0.09581653773784637,
-0.03467395529150963,
0.047755662351846695,
-0.05117446184158325,
-0.010842117480933666,
-0.02349942922592163,
0.0830949917435646,
-0.29295188188552856,
0.2703675627708435,
-0.0017172398511320353,
0.07757318764925003,
-0.07565660029649734,
-0.051325008273124695,
0.04897455498576164,
0.1557551920413971,
0.15221737325191498,
0.0003222957020625472,
-0.12438483536243439,
-0.11494789272546768,
-0.10556082427501678,
0.07548149675130844,
0.06739173084497452,
0.0009522453183308244,
0.04110225662589073,
0.008229147642850876,
-0.034526437520980835,
0.001758829690515995,
-0.055123668164014816,
-0.06881122291088104,
-0.026687249541282654,
-0.046567175537347794,
0.1475333273410797,
0.08861061930656433,
-0.008156507275998592,
-0.01186627522110939,
-0.08757586777210236,
0.10568998754024506,
-0.057066909968853,
-0.07791302353143692,
-0.06956810504198074,
-0.05330875515937805,
0.010568183846771717,
-0.08111268281936646,
0.00030259453342296183,
-0.009426496922969818,
0.02526281028985977,
-0.012426801957190037,
-0.10110088437795639,
0.055376745760440826,
-0.07593026757240295,
-0.06817008554935455,
0.05239346995949745,
0.11819074302911758,
-0.010773898102343082,
-0.016498249024152756,
0.05761439725756645,
-0.03239201381802559,
0.008054861798882484,
-0.15798020362854004,
-0.03442784771323204,
0.03174968808889389,
0.04697072133421898,
0.04173249751329422,
-0.13150611519813538,
-0.06158556416630745,
-0.08802921324968338,
-0.0929880440235138,
0.14178109169006348,
0.19806411862373352,
-0.06813348084688187,
0.028798449784517288,
0.13651065528392792,
-0.02754567563533783,
-0.274172842502594,
-0.09949735552072525,
-0.03429074212908745,
0.08244983106851578,
-0.04417009279131889,
-0.10229364782571793,
0.02376321144402027,
0.03292142227292061,
-0.03206372633576393,
-0.11230673640966415,
-0.2431362420320511,
-0.16987843811511993,
0.1570136994123459,
0.03852343559265137,
0.3095860481262207,
-0.08005165308713913,
-0.06970570236444473,
-0.01922161504626274,
-0.13803745806217194,
0.0657678171992302,
-0.20640438795089722,
0.10229605436325073,
0.007986677810549736,
0.0879107266664505,
0.04014481231570244,
-0.026582282036542892,
0.148848757147789,
0.008190221153199673,
0.05634394660592079,
-0.09881510585546494,
-0.11678536236286163,
-0.0827806293964386,
-0.04945748299360275,
0.16415439546108246,
-0.13529455661773682,
0.06718308478593826,
-0.13214939832687378,
-0.05380741506814957,
-0.07522711157798767,
0.003355270018801093,
-0.0027076334226876497,
-0.07223872095346451,
-0.05427592620253563,
0.03881365805864334,
-0.004011489916592836,
-0.013982751406729221,
0.1190897524356842,
-0.10458728671073914,
0.045901767909526825,
0.1505778580904007,
0.19413135945796967,
0.04223685711622238,
0.19315819442272186,
0.009300434030592442,
-0.054920442402362823,
0.08143644779920578,
-0.18060532212257385,
-0.0027340324595570564,
0.06205248832702637,
-0.02727033942937851,
0.11563935875892639,
0.0576203390955925,
-0.0062448009848594666,
0.05445818230509758,
0.10529134422540665,
-0.11509028822183609,
-0.09744365513324738,
-0.09234623610973358,
0.01970667950809002,
0.12644240260124207,
0.06714551895856857,
0.13779160380363464,
-0.1553899347782135,
0.03302650526165962,
-0.036709778010845184,
-0.00538959726691246,
-0.07276064902544022,
0.003347380319610238,
0.03609038144350052,
0.03408234938979149,
-0.047630660235881805,
0.07547809183597565,
-0.03190091997385025,
0.035899318754673004,
0.0681350901722908,
0.06422065198421478,
-0.17714889347553253,
-0.08238446712493896,
-0.00951445009559393,
0.15981678664684296,
-0.05419522151350975,
-0.12676547467708588,
-0.04876551404595375,
-0.17447926104068756,
-0.03727484494447708,
0.15167950093746185,
0.045766327530145645,
0.05155708268284798,
-0.0016998080536723137,
-0.03139572963118553,
-0.13159416615962982,
0.04008558765053749,
-0.05560164526104927,
-0.021717093884944916,
-0.12496820837259293,
0.12257630378007889,
0.012658904306590557,
0.12729045748710632,
-0.08577588945627213,
-0.026073593646287918,
-0.11848723143339157,
0.04799624904990196,
-0.12138517200946808,
-0.005337011069059372,
-0.15214592218399048,
-0.06046583130955696,
-0.015773996710777283,
-0.05282788723707199,
-0.07328066229820251,
0.015572055242955685,
-0.08489134907722473,
0.07435335218906403,
-0.05099423974752426,
0.05408245697617531,
-0.029975181445479393,
-0.01921766996383667,
0.04953950643539429,
-0.04605500027537346,
0.054344721138477325,
0.10824427753686905,
-0.06577353924512863,
0.13063845038414001,
-0.15722286701202393,
0.09663326293230057,
0.09079130738973618,
-0.00003284236299805343,
0.04327986389398575,
0.024181053042411804,
-0.01400493923574686,
0.0550733245909214,
0.036692384630441666,
0.03968853875994682,
0.0853859931230545,
-0.05949052795767784,
-0.012110747396945953,
0.01471010223031044,
0.006286372896283865,
-0.09370072185993195,
0.07406959682703018,
0.06431299448013306,
0.04708215966820717,
0.08521924167871475,
-0.1328793466091156,
0.006891557015478611,
-0.03667852282524109,
0.016125472262501717,
0.0016422973712906241,
-0.1036730632185936,
-0.022107837721705437,
-0.11761794984340668,
0.05915435403585434,
-0.058237381279468536,
0.1870422065258026,
0.015092222020030022,
-0.12666566669940948,
-0.0065248822793364525,
-0.061344847083091736,
0.05692847818136215,
0.027679793536663055,
0.2732559144496918,
0.060726795345544815,
0.015110582113265991,
-0.07195905596017838,
0.0463407039642334,
-0.0030536013655364513,
-0.004643261898308992,
0.12329556047916412,
0.07462263852357864,
-0.0076798005029559135,
0.13506603240966797,
0.10475925356149673,
-0.018444562330842018,
-0.007085406221449375,
-0.09172406792640686,
-0.17897729575634003,
0.013141842558979988,
-0.08848191052675247,
0.0378374308347702,
0.20371703803539276,
-0.08980369567871094,
0.018992668017745018,
-0.01915716752409935,
-0.09131063520908356,
-0.18280766904354095,
-0.16166502237319946,
-0.12482671439647675,
-0.1206149160861969,
0.03044877015054226,
-0.1117096096277237,
-0.043126530945301056,
0.04331189766526222,
0.06298201531171799,
-0.045221958309412,
0.1799616515636444,
-0.0437498539686203,
-0.033704303205013275,
0.12905530631542206,
-0.06859486550092697,
0.05603351071476936,
-0.10409265011548996,
-0.022548304870724678,
0.0033627403900027275,
-0.05110616981983185,
-0.023910127580165863,
0.018540285527706146,
-0.009305397979915142,
0.023161035031080246,
-0.033212460577487946,
-0.026545580476522446,
-0.03739077225327492,
0.029458561912178993,
0.06217682734131813,
0.040660690516233444,
0.03815777227282524,
-0.06596334278583527,
0.013023938983678818,
0.26773810386657715,
-0.005483982618898153,
-0.11468608677387238,
-0.14108231663703918,
0.20669575035572052,
0.04205290600657463,
0.030786331743001938,
0.008472627028822899,
-0.012072732672095299,
-0.00907763373106718,
0.3207521438598633,
0.2322734147310257,
-0.03627341613173485,
0.037570588290691376,
-0.0896252915263176,
0.02604636549949646,
0.031616926193237305,
0.12486687302589417,
0.07107783108949661,
0.1575823873281479,
-0.04759593680500984,
0.020811952650547028,
-0.029319114983081818,
-0.006677001714706421,
0.014733977615833282,
0.05121226981282234,
0.039109133183956146,
-0.10123606771230698,
-0.0624898225069046,
0.10166001319885254,
-0.17972302436828613,
0.07051762193441391,
-0.06582605093717575,
-0.051744166761636734,
-0.061861801892519,
-0.004165942780673504,
0.03133268281817436,
0.05161992833018303,
0.05463535338640213,
-0.07160627841949463,
0.02607322297990322,
0.011680634692311287,
0.005741420667618513,
-0.02647070400416851,
0.017289290204644203,
0.05429108440876007,
0.10089593380689621,
0.05952643230557442,
0.02702082134783268,
0.06324578076601028,
0.06835068017244339,
0.0014377871993929148,
-0.051945287734270096,
0.2159350961446762,
0.005670381244271994,
-0.039405155926942825,
0.037609510123729706,
0.043484315276145935,
-0.045387085527181625,
0.022124793380498886,
0.0226557869464159,
-0.12773534655570984,
0.09015551954507828,
-0.0030113605316728354,
-0.08857165277004242,
-0.09278810769319534,
0.06166283041238785,
-0.07106748968362808,
0.12586575746536255,
0.18085098266601562,
-0.031130069866776466,
0.019399357959628105,
0.01187924761325121,
0.08191725611686707,
-0.06849110871553421,
-0.08325366675853729,
-0.029431991279125214,
-0.13134704530239105,
-0.02067188173532486,
0.09607602655887604,
0.008906465023756027,
-0.2464817613363266,
0.006885144859552383,
-0.11943984776735306,
-0.007120958063751459,
-0.0596051849424839,
0.025284888222813606,
0.13745324313640594,
0.03504408150911331,
-0.023291369900107384,
-0.17681461572647095,
0.01715843193233013,
0.0971023440361023,
-0.12591281533241272,
-0.1772482544183731
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# mistral-7b-ocn-v2
This model is a fine-tuned version of mistralai/Mistral-7B-v0.1 on the oncc_medqa_instruct dataset.
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0005
- train_batch_size: 4
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_steps: 10
- num_epochs: 1.0
- mixed_precision_training: Native AMP
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.1.1+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
### Performance
hf (pretrained=mistralai/Mistral-7B-v0.1,parallelize=True,load_in_4bit=True,peft=chenhugging/mistral-7b-ocn-v2), gen_kwargs: (None), limit: 100.0, num_fewshot: None, batch_size: 1
| Tasks |Version|Filter|n-shot| Metric |Value| |Stderr|
|---------------------|-------|------|-----:|--------|----:|---|-----:|
|pubmedqa | 1|none | 0|acc | 0.98|± |0.0141|
|medmcqa |Yaml |none | 0|acc | 0.40|± |0.0492|
|professional_medicine| 0|none | 0|acc | 0.69|± |0.0465|
|college_medicine | 0|none | 0|acc | 0.53|± |0.0502|
|clinical_knowledge | 0|none | 0|acc | 0.59|± |0.0494|
|ocn |Yaml |none | 0|acc | 0.80|± |0.0402|
|aocnp |Yaml |none | 0|acc | 0.63|± |0.0485|
| {"license": "other", "library_name": "peft", "tags": ["llama-factory", "lora", "generated_from_trainer"], "base_model": "mistralai/Mistral-7B-v0.1", "model-index": [{"name": "mistral-7b-ocn-v2", "results": []}]} | null | chenhugging/mistral-7b-ocn-v2 | [
"peft",
"safetensors",
"llama-factory",
"lora",
"generated_from_trainer",
"base_model:mistralai/Mistral-7B-v0.1",
"license:other",
"region:us"
] | 2024-02-07T07:07:17+00:00 | [] | [] | TAGS
#peft #safetensors #llama-factory #lora #generated_from_trainer #base_model-mistralai/Mistral-7B-v0.1 #license-other #region-us
| mistral-7b-ocn-v2
=================
This model is a fine-tuned version of mistralai/Mistral-7B-v0.1 on the oncc\_medqa\_instruct dataset.
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0005
* train\_batch\_size: 4
* eval\_batch\_size: 8
* seed: 42
* gradient\_accumulation\_steps: 4
* total\_train\_batch\_size: 16
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: cosine
* lr\_scheduler\_warmup\_steps: 10
* num\_epochs: 1.0
* mixed\_precision\_training: Native AMP
### Framework versions
* PEFT 0.8.2
* Transformers 4.37.2
* Pytorch 2.1.1+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
### Performance
hf (pretrained=mistralai/Mistral-7B-v0.1,parallelize=True,load\_in\_4bit=True,peft=chenhugging/mistral-7b-ocn-v2), gen\_kwargs: (None), limit: 100.0, num\_fewshot: None, batch\_size: 1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 1.0\n* mixed\\_precision\\_training: Native AMP",
"### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1",
"### Performance\n\n\nhf (pretrained=mistralai/Mistral-7B-v0.1,parallelize=True,load\\_in\\_4bit=True,peft=chenhugging/mistral-7b-ocn-v2), gen\\_kwargs: (None), limit: 100.0, num\\_fewshot: None, batch\\_size: 1"
] | [
"TAGS\n#peft #safetensors #llama-factory #lora #generated_from_trainer #base_model-mistralai/Mistral-7B-v0.1 #license-other #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 1.0\n* mixed\\_precision\\_training: Native AMP",
"### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1",
"### Performance\n\n\nhf (pretrained=mistralai/Mistral-7B-v0.1,parallelize=True,load\\_in\\_4bit=True,peft=chenhugging/mistral-7b-ocn-v2), gen\\_kwargs: (None), limit: 100.0, num\\_fewshot: None, batch\\_size: 1"
] | [
51,
159,
39,
94
] | [
"passage: TAGS\n#peft #safetensors #llama-factory #lora #generated_from_trainer #base_model-mistralai/Mistral-7B-v0.1 #license-other #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 1.0\n* mixed\\_precision\\_training: Native AMP### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1### Performance\n\n\nhf (pretrained=mistralai/Mistral-7B-v0.1,parallelize=True,load\\_in\\_4bit=True,peft=chenhugging/mistral-7b-ocn-v2), gen\\_kwargs: (None), limit: 100.0, num\\_fewshot: None, batch\\_size: 1"
] | [
-0.11389196664094925,
0.08067178726196289,
-0.005449953023344278,
0.0786883682012558,
0.08243118226528168,
0.026323290541768074,
0.12948442995548248,
0.1391899138689041,
-0.026751462370157242,
0.12460847198963165,
0.10222358256578445,
0.05935485288500786,
0.06425382941961288,
0.12512601912021637,
-0.007402246817946434,
-0.2675431966781616,
0.020101798698306084,
-0.03387748450040817,
-0.11396182328462601,
0.10353914648294449,
0.0696413516998291,
-0.09309028089046478,
0.04956028237938881,
-0.00951098371297121,
-0.07248617708683014,
-0.023757729679346085,
-0.028880449011921883,
-0.015282725915312767,
0.10363798588514328,
0.06102173402905464,
0.08569139987230301,
0.037820007652044296,
0.08549269288778305,
-0.2342289537191391,
0.004329925402998924,
0.06831774860620499,
0.007264256943017244,
0.07295358926057816,
0.104151152074337,
0.024836961179971695,
0.09606931358575821,
-0.10865248739719391,
0.04643575847148895,
0.03728801757097244,
-0.13984769582748413,
-0.248540997505188,
-0.09609443694353104,
0.07612737268209457,
0.1257929503917694,
0.06228383257985115,
-0.023447299376130104,
0.10376457870006561,
-0.07553131878376007,
0.07837560027837753,
0.27361321449279785,
-0.2791259288787842,
-0.08962708711624146,
0.04010819271206856,
0.06772473454475403,
0.040967028588056564,
-0.12463225424289703,
-0.028895795345306396,
0.03512085601687431,
0.022671889513731003,
0.0705842524766922,
-0.008312057703733444,
0.10700607299804688,
-0.007049419451504946,
-0.13369490206241608,
-0.04411930590867996,
0.11175238341093063,
0.06558153033256531,
-0.027415363118052483,
-0.07277119159698486,
-0.047216400504112244,
-0.18818801641464233,
-0.050516869872808456,
-0.001705256407149136,
0.021248016506433487,
-0.04050089418888092,
-0.03373172879219055,
0.057544562965631485,
-0.053399160504341125,
-0.06168791651725769,
0.06116121634840965,
0.1306367665529251,
0.06277909874916077,
-0.03600397706031799,
0.016850460320711136,
0.10419806838035583,
0.03174398094415665,
-0.17591851949691772,
-0.015094988979399204,
0.007354009430855513,
-0.0897400975227356,
-0.0364639051258564,
-0.0296246949583292,
0.04692734405398369,
0.04321691021323204,
0.19973155856132507,
-0.09515493363142014,
0.09752542525529861,
0.06334645301103592,
0.02145400643348694,
-0.07890380173921585,
0.12187234312295914,
-0.08233527839183807,
-0.08431632071733475,
-0.04155472666025162,
0.11929592490196228,
0.0060208458453416824,
-0.0067588528618216515,
-0.053686998784542084,
0.01173152681440115,
0.10354552417993546,
0.04042396694421768,
-0.003820542013272643,
0.028417449444532394,
-0.06415959447622299,
-0.022971170023083687,
0.07971744239330292,
-0.10051163285970688,
0.03949026018381119,
0.04270954430103302,
-0.0653383657336235,
-0.009267503395676613,
-0.01381538063287735,
-0.014965328387916088,
-0.015062503516674042,
0.10476408153772354,
-0.09099896997213364,
-0.012936954386532307,
-0.05565563589334488,
-0.0993962287902832,
0.03113633580505848,
-0.008944660425186157,
-0.00025898427702486515,
-0.07680410891771317,
-0.09711422771215439,
-0.06536402553319931,
0.054323095828294754,
-0.07808665186166763,
-0.04336772859096527,
-0.03229270502924919,
-0.09709423035383224,
0.04529833793640137,
0.0009250101866200566,
0.10798651725053787,
-0.05460083484649658,
0.0949961319565773,
-0.0008080779807642102,
0.04754440113902092,
0.07601239532232285,
0.02538818120956421,
-0.057522714138031006,
0.08629079908132553,
-0.18662482500076294,
0.03678907826542854,
-0.08905257284641266,
0.06281522661447525,
-0.14498551189899445,
-0.11391749978065491,
-0.012064613401889801,
-0.0172524843364954,
0.09304448962211609,
0.11709074676036835,
-0.15996934473514557,
-0.06822362542152405,
0.1725960671901703,
-0.08426208794116974,
-0.09200804680585861,
0.11632832884788513,
-0.015025732107460499,
-0.037290964275598526,
0.016237691044807434,
0.1504974216222763,
0.1449955552816391,
-0.11927511543035507,
-0.023850245401263237,
-0.041395824402570724,
0.11602193117141724,
0.07134468108415604,
0.08637779951095581,
-0.009852791205048561,
0.02088777907192707,
0.005674221087247133,
-0.0537547692656517,
0.03960952162742615,
-0.07968136668205261,
-0.07552611082792282,
-0.010173308663070202,
-0.055930208414793015,
0.055012378841638565,
0.057278573513031006,
0.023742863908410072,
-0.07181812077760696,
-0.13146507740020752,
0.01148921623826027,
0.10606624186038971,
-0.07165713608264923,
0.01594172976911068,
-0.0341067872941494,
0.09078451246023178,
0.016679923981428146,
0.0012431953800842166,
-0.15607060492038727,
-0.08228758722543716,
0.03742654621601105,
-0.08825894445180893,
-0.009393340907990932,
-0.0023209084756672382,
0.0816929042339325,
0.1038755550980568,
-0.05854436755180359,
-0.053448762744665146,
-0.05226783826947212,
-0.0032217586413025856,
-0.09196745604276657,
-0.2425723522901535,
-0.09429410099983215,
-0.02206343412399292,
0.1523926705121994,
-0.22929395735263824,
0.015685614198446274,
0.01343188900500536,
0.14280137419700623,
0.013967410661280155,
-0.03862336650490761,
-0.02295580692589283,
0.07145416736602783,
-0.025442073121666908,
-0.07470963895320892,
0.040452200919389725,
-0.015689607709646225,
-0.07182314246892929,
-0.024174073711037636,
-0.15282070636749268,
0.09749718010425568,
0.07753787189722061,
0.02450067363679409,
-0.1322939395904541,
-0.053600411862134933,
-0.062188003212213516,
-0.05813342705368996,
0.00031110038980841637,
0.000042670100810937583,
0.11251094937324524,
0.03725891187787056,
0.10012943297624588,
-0.07223070412874222,
-0.060694001615047455,
0.035088978707790375,
-0.00719787273555994,
0.000380634272005409,
0.14924752712249756,
0.03593435883522034,
-0.09948909282684326,
0.110838383436203,
0.10957039892673492,
-0.05309928581118584,
0.12461131811141968,
-0.06819058209657669,
-0.08223050832748413,
-0.07882217317819595,
0.0569944866001606,
0.0412495955824852,
0.11434384435415268,
-0.045183129608631134,
0.011186494491994381,
0.008129574358463287,
0.00544347707182169,
-0.013670001178979874,
-0.1659572422504425,
-0.019245902076363564,
0.024361956864595413,
-0.044073887169361115,
0.02389114536345005,
-0.011040758341550827,
-0.01447354070842266,
0.08665022999048233,
0.016811883077025414,
-0.07474986463785172,
-0.03781145438551903,
-0.02169019542634487,
-0.07505177706480026,
0.2130516767501831,
-0.07175996899604797,
-0.08416932076215744,
-0.11289480328559875,
-0.010706603527069092,
-0.026033077389001846,
-0.023477908223867416,
0.034777767956256866,
-0.06772304326295853,
-0.03677817061543465,
-0.07311858236789703,
-0.028685685247182846,
-0.007570806425064802,
0.014307496137917042,
-0.007471232675015926,
0.0006226288969628513,
0.07949698716402054,
-0.09248168766498566,
0.005747889168560505,
0.0037188639398664236,
-0.01868002489209175,
0.02715073712170124,
0.011701448820531368,
0.08437331020832062,
0.1424836367368698,
0.047572530806064606,
0.01415337435901165,
-0.033622387796640396,
0.23203027248382568,
-0.08775181323289871,
0.017945433035492897,
0.0821688249707222,
-0.0013224442955106497,
0.07286994159221649,
0.15498895943164825,
0.061525460332632065,
-0.07516729831695557,
-0.00037048241938464344,
0.036272063851356506,
-0.015988828614354134,
-0.2009691298007965,
-0.03315064311027527,
-0.056752871721982956,
0.00968174822628498,
0.13450446724891663,
0.03445399925112724,
0.025487305596470833,
0.022001594305038452,
-0.038437504321336746,
0.04290716350078583,
0.03562374785542488,
0.085813507437706,
0.05601615458726883,
0.0543135441839695,
0.101660817861557,
-0.010967418551445007,
-0.020578356459736824,
0.016908766701817513,
-0.004155762027949095,
0.23735657334327698,
-0.005123420152813196,
0.16090920567512512,
0.05175146460533142,
0.1598125696182251,
-0.006449389737099409,
0.04134626314043999,
0.02573816291987896,
-0.01863841526210308,
0.0029645778704434633,
-0.07657656073570251,
0.0016380627639591694,
0.045037638396024704,
0.03878607973456383,
0.02592146024107933,
-0.0946672335267067,
-0.005665894597768784,
0.03492012247443199,
0.26918408274650574,
0.08973036706447601,
-0.2888752520084381,
-0.07598068565130234,
0.004154952242970467,
-0.017288006842136383,
-0.009918934665620327,
0.009600494988262653,
0.09358907490968704,
-0.07919542491436005,
0.06996768712997437,
-0.06497683376073837,
0.0754903182387352,
-0.03158281743526459,
0.020839760079979897,
0.10826300829648972,
0.07043678313493729,
0.000827574054710567,
0.0479561947286129,
-0.19858820736408234,
0.27237018942832947,
-0.0018905879696831107,
0.05780632421374321,
-0.01899152621626854,
0.024090711027383804,
0.030643971636891365,
-0.006632556207478046,
0.08571626991033554,
0.0009385680896230042,
-0.09865502268075943,
-0.23109318315982819,
-0.09087840467691422,
-0.012531797401607037,
0.12181834131479263,
-0.06682141870260239,
0.13741832971572876,
-0.03044954128563404,
-0.03452945500612259,
0.030046645551919937,
0.01307554543018341,
-0.09861122816801071,
-0.08150620758533478,
0.051006559282541275,
-0.044122979044914246,
-0.0003445112088229507,
-0.08910036087036133,
-0.09095947444438934,
-0.08698049932718277,
0.13507674634456635,
-0.10853271186351776,
-0.031373027712106705,
-0.1299026757478714,
0.0685897246003151,
0.16266189515590668,
-0.09691905230283737,
0.029355112463235855,
0.009858696721494198,
0.08111299574375153,
0.001305879675783217,
-0.03164934366941452,
0.10325192660093307,
-0.05426602065563202,
-0.22613191604614258,
-0.06455717980861664,
0.1530441790819168,
0.03538918495178223,
0.06104576960206032,
-0.025764109566807747,
0.05032327026128769,
0.01736171543598175,
-0.1019272431731224,
0.037052493542432785,
-0.008607898838818073,
0.048910610377788544,
0.009949082508683205,
-0.03634624555706978,
0.05795681104063988,
-0.06963452696800232,
-0.009983149357140064,
0.052952300757169724,
0.35670387744903564,
-0.10340434312820435,
0.0022421665489673615,
0.04957885295152664,
-0.05412691459059715,
-0.14754965901374817,
-0.04511202126741409,
0.12152544409036636,
-0.005661847535520792,
0.035340871661901474,
-0.16564762592315674,
0.036545224487781525,
0.11497234553098679,
-0.014428619295358658,
0.10245971381664276,
-0.34452828764915466,
-0.13598820567131042,
0.06350890547037125,
0.0936247706413269,
-0.014527373015880585,
-0.19430139660835266,
-0.06729523837566376,
0.007523959036916494,
-0.12387238442897797,
0.031196406111121178,
-0.03848744556307793,
0.11538185179233551,
-0.026658933609724045,
-0.03869161754846573,
0.017003050073981285,
-0.0620507188141346,
0.17554345726966858,
0.0054880063980817795,
0.0766783207654953,
-0.03068239986896515,
-0.01900036633014679,
0.034824665635824203,
-0.07919814437627792,
0.005316141992807388,
-0.07560079544782639,
0.017698539420962334,
-0.09818106144666672,
-0.00551871582865715,
-0.09332143515348434,
0.01904318481683731,
-0.07921133190393448,
-0.022077003493905067,
-0.04491982236504555,
0.08178629726171494,
0.0633813664317131,
-0.0004019502375740558,
0.12668642401695251,
0.005413373466581106,
0.17689859867095947,
0.16444122791290283,
0.021596014499664307,
0.03957341983914375,
-0.13414528965950012,
-0.0019597280770540237,
-0.00546270189806819,
0.04225482791662216,
-0.11086095124483109,
0.017355868592858315,
0.15140454471111298,
0.04555156081914902,
0.12838803231716156,
0.05673022195696831,
-0.07155550271272659,
-0.011658664792776108,
0.06552045792341232,
-0.0999428778886795,
-0.11853129416704178,
-0.019142750650644302,
0.001617001835256815,
-0.1303415298461914,
-0.024251451715826988,
0.13833926618099213,
-0.0656159296631813,
-0.00025217991787940264,
0.02850193716585636,
0.041604187339544296,
-0.027385761961340904,
0.23704411089420319,
0.032688938081264496,
0.09357202798128128,
-0.0823386013507843,
0.054303400218486786,
0.08298671245574951,
-0.06767084449529648,
0.02582147903740406,
0.10452387481927872,
-0.06769102066755295,
-0.033904630690813065,
0.03102133609354496,
0.0773526206612587,
0.0017149071209132671,
-0.013229615986347198,
-0.11923425644636154,
-0.12553662061691284,
0.08260704576969147,
0.0992971882224083,
0.0383744016289711,
0.03348619118332863,
0.001563508645631373,
0.021193470805883408,
-0.07537636905908585,
0.1191934123635292,
0.08420773595571518,
0.09001296013593674,
-0.12711201608181,
0.13255192339420319,
-0.0006154775619506836,
-0.002347427885979414,
0.0029345008078962564,
0.013714496977627277,
-0.16200301051139832,
-0.005658634006977081,
-0.13277673721313477,
0.007794767152518034,
-0.05598093569278717,
-0.0071749198250472546,
-0.017033932730555534,
-0.0420784056186676,
-0.03797494247555733,
0.012056090869009495,
-0.09226541966199875,
-0.05776184797286987,
-0.012493938207626343,
0.07161842286586761,
-0.1259879320859909,
-0.02960611693561077,
0.04145533964037895,
-0.11943928897380829,
0.06567737460136414,
0.010122928768396378,
0.03848264738917351,
0.0204390287399292,
-0.04600013047456741,
0.02327379211783409,
0.007016187999397516,
-0.009932093322277069,
0.02693352848291397,
-0.17262713611125946,
-0.012768286280333996,
-0.0703473687171936,
0.011443361639976501,
0.014713743701577187,
0.00044896776671521366,
-0.12919721007347107,
0.012799802236258984,
-0.03988642618060112,
-0.06927789747714996,
-0.05358506739139557,
0.02863386459648609,
0.0681874230504036,
-0.0006313629564829171,
0.13520538806915283,
-0.08342944085597992,
0.06752368807792664,
-0.25242239236831665,
-0.02854512259364128,
0.010350826196372509,
-0.06719791889190674,
-0.03017398528754711,
-0.039790160953998566,
0.08290789276361465,
-0.05873791128396988,
0.05209725350141525,
-0.05633990466594696,
0.07038233429193497,
0.027535581961274147,
-0.06950928270816803,
-0.007951105013489723,
0.05880875885486603,
0.15068532526493073,
0.04168160632252693,
-0.030210105702280998,
0.03498000651597977,
0.01783251017332077,
0.04184621572494507,
0.07614793628454208,
0.17110928893089294,
0.13372249901294708,
0.026541322469711304,
0.05833937227725983,
0.034001849591732025,
-0.13649795949459076,
-0.1331232339143753,
0.11436964571475983,
-0.03886128216981888,
0.12415466457605362,
-0.0263799037784338,
0.13655422627925873,
0.11583873629570007,
-0.22001932561397552,
0.04441892355680466,
-0.05594504624605179,
-0.0953427329659462,
-0.09060212224721909,
-0.0803787112236023,
-0.06937780976295471,
-0.13459046185016632,
0.0015093295369297266,
-0.09786705672740936,
0.0260882880538702,
0.08553988486528397,
0.03259391337633133,
0.04430899769067764,
0.11667931824922562,
0.06613612174987793,
0.012197306379675865,
0.06635475903749466,
0.03885863348841667,
-0.014516528695821762,
-0.029604408890008926,
-0.07257682830095291,
0.022226795554161072,
-0.06363100558519363,
0.027869315817952156,
-0.02852349728345871,
-0.04670296236872673,
0.05994512513279915,
0.021558096632361412,
-0.10308144241571426,
0.03577536344528198,
0.004626425914466381,
0.03772401809692383,
0.07643838971853256,
0.027079574763774872,
-0.0007806892972439528,
-0.026001134887337685,
0.18476377427577972,
-0.0687098503112793,
-0.06058061122894287,
-0.10586822032928467,
0.2634528875350952,
0.02178940363228321,
-0.010241547599434853,
0.018663743510842323,
-0.06743123382329941,
0.019432220607995987,
0.13307389616966248,
0.13324999809265137,
-0.04499395191669464,
-0.007168895099312067,
0.03972255438566208,
-0.012367249466478825,
-0.013039389625191689,
0.099636510014534,
0.08995506912469864,
0.07736216485500336,
-0.07568274438381195,
0.004874854348599911,
-0.03521629050374031,
-0.04162042587995529,
-0.0667322650551796,
0.062332846224308014,
0.04164310544729233,
0.009932990185916424,
-0.04515983909368515,
0.08948671817779541,
-0.055650494992733,
-0.12173730880022049,
0.09628342092037201,
-0.19405464828014374,
-0.18073900043964386,
-0.03241397812962532,
0.05325087532401085,
0.024845529347658157,
0.07047177106142044,
-0.018023615702986717,
-0.033497896045446396,
0.10370199382305145,
-0.014184157364070415,
-0.02120964229106903,
-0.11650791019201279,
0.06486198306083679,
-0.05499006807804108,
0.19362370669841766,
-0.03764764964580536,
0.03986886516213417,
0.13136838376522064,
0.02501758001744747,
-0.10367588698863983,
0.010703698731958866,
0.10189057141542435,
-0.12611818313598633,
0.022672239691019058,
0.12577120959758759,
-0.020565373823046684,
0.11005841940641403,
0.05469460040330887,
-0.08347457647323608,
0.015144365839660168,
0.014493116177618504,
-0.014087841846048832,
-0.055653009563684464,
-0.0016727278707548976,
-0.02811826951801777,
0.15233416855335236,
0.22214563190937042,
-0.0587371289730072,
-0.015410025604069233,
-0.03337077051401138,
0.020825974643230438,
0.057730842381715775,
0.10647517442703247,
-0.03791883960366249,
-0.23664763569831848,
0.047270771116018295,
0.009591138921678066,
0.04442792758345604,
-0.17964303493499756,
-0.10245049744844437,
0.03892488777637482,
-0.052904002368450165,
-0.06267140805721283,
0.13528360426425934,
0.037991367280483246,
0.058446601033210754,
-0.047797348350286484,
-0.14342501759529114,
-0.05696254223585129,
0.16475361585617065,
-0.15404729545116425,
-0.07530410587787628
] |
null | null | diffusers |
# DreamBooth - yuandli/dogbooth
This is a dreambooth model derived from stabilityai/stable-diffusion-2-1. The weights were trained on a photo of [v]dog using [DreamBooth](https://dreambooth.github.io/).
You can find some example images in the following.
DreamBooth for the text encoder was enabled: False.
| {"license": "creativeml-openrail-m", "library_name": "diffusers", "tags": ["stable-diffusion", "stable-diffusion-diffusers", "text-to-image", "diffusers", "dreambooth", "text-to-image", "dreambooth", "stable-diffusion", "stable-diffusion-diffusers"], "base_model": "stabilityai/stable-diffusion-2-1", "instance_prompt": "a photo of [v]dog", "inference": true} | text-to-image | yuandli/dogbooth | [
"diffusers",
"tensorboard",
"safetensors",
"stable-diffusion",
"stable-diffusion-diffusers",
"text-to-image",
"dreambooth",
"base_model:stabilityai/stable-diffusion-2-1",
"license:creativeml-openrail-m",
"endpoints_compatible",
"diffusers:StableDiffusionPipeline",
"region:us"
] | 2024-02-07T07:07:47+00:00 | [] | [] | TAGS
#diffusers #tensorboard #safetensors #stable-diffusion #stable-diffusion-diffusers #text-to-image #dreambooth #base_model-stabilityai/stable-diffusion-2-1 #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us
|
# DreamBooth - yuandli/dogbooth
This is a dreambooth model derived from stabilityai/stable-diffusion-2-1. The weights were trained on a photo of [v]dog using DreamBooth.
You can find some example images in the following.
DreamBooth for the text encoder was enabled: False.
| [
"# DreamBooth - yuandli/dogbooth\n\nThis is a dreambooth model derived from stabilityai/stable-diffusion-2-1. The weights were trained on a photo of [v]dog using DreamBooth.\nYou can find some example images in the following. \n\n\n\nDreamBooth for the text encoder was enabled: False."
] | [
"TAGS\n#diffusers #tensorboard #safetensors #stable-diffusion #stable-diffusion-diffusers #text-to-image #dreambooth #base_model-stabilityai/stable-diffusion-2-1 #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us \n",
"# DreamBooth - yuandli/dogbooth\n\nThis is a dreambooth model derived from stabilityai/stable-diffusion-2-1. The weights were trained on a photo of [v]dog using DreamBooth.\nYou can find some example images in the following. \n\n\n\nDreamBooth for the text encoder was enabled: False."
] | [
97,
79
] | [
"passage: TAGS\n#diffusers #tensorboard #safetensors #stable-diffusion #stable-diffusion-diffusers #text-to-image #dreambooth #base_model-stabilityai/stable-diffusion-2-1 #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us \n# DreamBooth - yuandli/dogbooth\n\nThis is a dreambooth model derived from stabilityai/stable-diffusion-2-1. The weights were trained on a photo of [v]dog using DreamBooth.\nYou can find some example images in the following. \n\n\n\nDreamBooth for the text encoder was enabled: False."
] | [
-0.05434555932879448,
0.10205361247062683,
-0.0037138243205845356,
0.017300937324762344,
0.12561379373073578,
-0.004429740831255913,
0.16473163664340973,
0.030074935406446457,
-0.003085080999881029,
0.06343862414360046,
0.10089559853076935,
-0.011445624753832817,
0.002851255005225539,
0.08939076215028763,
0.06156650185585022,
-0.17525452375411987,
0.05255099758505821,
0.003944690804928541,
-0.033001046627759933,
0.019707895815372467,
0.03709039092063904,
-0.09858068078756332,
0.1018982008099556,
0.009228021837770939,
-0.17898504436016083,
0.03458252176642418,
-0.05091124027967453,
0.014594883657991886,
0.08184167742729187,
-0.0032080684322863817,
0.11062434315681458,
0.04249834641814232,
0.023655224591493607,
-0.1480480432510376,
0.02173558436334133,
0.03579750657081604,
-0.03378298133611679,
0.03807848319411278,
0.030685683712363243,
-0.03471533581614494,
0.02612195536494255,
-0.06610558927059174,
0.05246247351169586,
0.025445811450481415,
-0.021229997277259827,
0.00941179133951664,
0.037914667278528214,
0.14623726904392242,
0.08537782728672028,
0.0799267590045929,
-0.0417056679725647,
0.038457948714494705,
0.0382767878472805,
0.09542176127433777,
0.1972428411245346,
-0.23177878558635712,
-0.047521404922008514,
0.2985592186450958,
-0.03436719998717308,
-0.017055701464414597,
-0.026691699400544167,
0.051157739013433456,
0.046622615307569504,
0.006952161435037851,
-0.0044903140515089035,
-0.044751957058906555,
0.04355807229876518,
-0.13311585783958435,
-0.10021836310625076,
0.029522301629185677,
0.03405268117785454,
-0.012621201574802399,
-0.07036422938108444,
-0.16873060166835785,
-0.06860872358083725,
0.1382593810558319,
-0.027921292930841446,
-0.007255225908011198,
-0.017028043046593666,
0.02017851360142231,
-0.0023085151333361864,
-0.019513912498950958,
-0.06384369730949402,
-0.07839293032884598,
-0.007200249005109072,
0.07010874897241592,
0.02175096608698368,
0.010638289153575897,
-0.023385563865303993,
0.14852860569953918,
-0.05574683099985123,
-0.12855447828769684,
0.0680588111281395,
-0.04338730871677399,
-0.02742597833275795,
0.09155207872390747,
-0.034336116164922714,
-0.210780531167984,
0.10859276354312897,
-0.058978524059057236,
0.09473876655101776,
-0.0025555735919624567,
0.0016571362502872944,
0.05348706617951393,
-0.00865318812429905,
0.01773325726389885,
-0.06122202426195145,
-0.02414786070585251,
-0.013805135153234005,
0.002039693295955658,
0.03752947226166725,
-0.019217878580093384,
-0.1082393005490303,
0.01600937359035015,
-0.025165917351841927,
0.02513633295893669,
-0.006841438356786966,
0.023891165852546692,
-0.08941613137722015,
-0.01879679411649704,
0.06421992182731628,
-0.03181791305541992,
-0.011620057746767998,
-0.02261863276362419,
0.01878584362566471,
-0.025646766647696495,
0.12920832633972168,
0.024923689663410187,
-0.03417399153113365,
0.09271812438964844,
-0.08082018792629242,
0.005190630443394184,
-0.01513791922479868,
-0.10858488827943802,
-0.030291594564914703,
-0.1943957805633545,
0.05067262053489685,
-0.14686909317970276,
-0.12474838644266129,
0.01326579786837101,
0.0006193534936755896,
-0.017345063388347626,
-0.007092028856277466,
-0.08569275587797165,
-0.1192047968506813,
-0.045957330614328384,
0.07058065384626389,
0.016138382256031036,
0.009413360618054867,
0.020984334871172905,
-0.023185594007372856,
0.08673279732465744,
-0.05577726289629936,
-0.0451543815433979,
-0.11910495907068253,
0.01315710786730051,
-0.07233448326587677,
0.12291274964809418,
-0.04361400753259659,
0.11342967301607132,
-0.035737838596105576,
-0.006200133357197046,
0.021552281454205513,
0.011659758165478706,
0.03663790971040726,
0.1572021096944809,
-0.20932404696941376,
-0.04133063927292824,
0.21800288558006287,
-0.21218246221542358,
-0.11278356611728668,
0.08350998908281326,
0.008329972624778748,
0.12093660235404968,
0.08048815280199051,
0.11435951292514801,
0.08857367187738419,
-0.3049736022949219,
-0.03996635228395462,
-0.05705006793141365,
-0.07694707065820694,
0.05282466858625412,
-0.03672800958156586,
0.060507725924253464,
-0.02655443176627159,
0.03323997184634209,
-0.08277428150177002,
0.0990588366985321,
-0.05067552626132965,
-0.021906226873397827,
-0.04977276548743248,
-0.08597853779792786,
0.05404156446456909,
0.019236162304878235,
0.04869654029607773,
-0.0041758278384804726,
-0.02452259138226509,
0.0699479877948761,
0.0035129168536514044,
-0.03213626891374588,
-0.012818770483136177,
-0.05304955318570137,
0.02639707922935486,
-0.015575384721159935,
0.0017453266773372889,
-0.10269860178232193,
-0.04259531944990158,
0.0488792359828949,
0.16392041742801666,
0.007639032788574696,
0.06932312250137329,
0.07057145982980728,
0.07891152054071426,
-0.006093194708228111,
-0.07521360367536545,
0.05601011961698532,
0.03073478490114212,
-0.053991734981536865,
-0.13956698775291443,
0.11182830482721329,
-0.0836387425661087,
-0.008889040909707546,
-0.17324386537075043,
0.08808211982250214,
0.07114791870117188,
0.263918399810791,
0.08365695178508759,
-0.04632287099957466,
0.06965219974517822,
0.03939303383231163,
-0.037862829864025116,
-0.07857251912355423,
0.00966076459735632,
0.02471145987510681,
-0.14240573346614838,
0.148367777466774,
-0.12274935096502304,
0.1351877599954605,
0.1253046840429306,
0.11185085773468018,
-0.05300242453813553,
0.00794130191206932,
-0.022303355857729912,
-0.01725168526172638,
-0.08119656145572662,
-0.011549958027899265,
0.1489914208650589,
0.0032908134162425995,
0.15933343768119812,
-0.025067925453186035,
0.05968127399682999,
0.07238273322582245,
-0.04390218108892441,
-0.05854947865009308,
0.10142536461353302,
-0.16308265924453735,
0.046273473650217056,
0.03993533179163933,
0.011140923947095871,
0.023355158045887947,
0.18818633258342743,
-0.031132090836763382,
-0.0012509530643001199,
-0.04910712689161301,
0.030394453555345535,
0.03287895396351814,
0.17040982842445374,
-0.04509067162871361,
-0.006225259508937597,
-0.02227003127336502,
-0.02533849887549877,
0.004457356408238411,
-0.1622992902994156,
-0.0027816586662083864,
0.03376755863428116,
-0.03898485377430916,
0.17984811961650848,
0.030125029385089874,
-0.09301941096782684,
0.022473759949207306,
-0.10688710957765579,
-0.03859947621822357,
0.020563554018735886,
-0.011715229600667953,
-0.08145134896039963,
0.12484846264123917,
-0.11344734579324722,
-0.2768264412879944,
-0.11482349783182144,
0.010869450867176056,
0.0007948229904286563,
0.00955299660563469,
0.057067643851041794,
-0.08919151872396469,
-0.034600235521793365,
-0.09195847064256668,
0.0468793548643589,
0.01601923629641533,
0.09597457945346832,
0.09117189794778824,
0.02323490008711815,
0.04552064090967178,
-0.00783422589302063,
0.0073009212501347065,
-0.06273265182971954,
0.04122374206781387,
0.04109952598810196,
0.026844412088394165,
0.09875726699829102,
0.09220866113901138,
0.0004812979022972286,
-0.023685825988650322,
-0.006831977050751448,
0.2563244104385376,
0.014326417818665504,
0.03116888739168644,
0.1124972477555275,
0.010049916803836823,
0.07266771048307419,
0.1710265427827835,
0.02702181786298752,
-0.05172593146562576,
0.09844890236854553,
0.006136853247880936,
-0.11285509169101715,
-0.03998132050037384,
-0.10321951657533646,
0.01003006100654602,
0.011112247593700886,
0.06088574603199959,
0.06829313933849335,
0.05543355271220207,
0.12292228639125824,
0.06598278135061264,
-0.004903461318463087,
0.009583636187016964,
0.09586238116025925,
0.06504824757575989,
-0.08318775147199631,
0.03325069695711136,
-0.059125546365976334,
-0.06985123455524445,
0.02409859001636505,
-0.03253636881709099,
0.10144875943660736,
-0.08122549951076508,
-0.10071828961372375,
0.059677209705114365,
0.05095699802041054,
0.0878552719950676,
0.02506779320538044,
-0.07137452065944672,
-0.0610104463994503,
-0.011279300786554813,
-0.10678715258836746,
0.05222059413790703,
0.12610799074172974,
-0.029520517215132713,
0.04215633124113083,
-0.022657373920083046,
0.16488969326019287,
0.018685299903154373,
0.038634322583675385,
0.1201331689953804,
-0.19918608665466309,
-0.06330197304487228,
-0.046428531408309937,
0.002849838463589549,
-0.021396396681666374,
-0.011892233043909073,
0.322627454996109,
-0.025240227580070496,
0.03895845264196396,
-0.03386332839727402,
0.028832606971263885,
0.03377208858728409,
0.013207790441811085,
-0.12944000959396362,
0.05418933182954788,
-0.031870074570178986,
-0.01685166172683239,
-0.286385178565979,
0.04936797171831131,
-0.01529592927545309,
0.07652401179075241,
0.005508021451532841,
0.04194163531064987,
-0.007343292236328125,
0.08709633350372314,
0.12639571726322174,
0.002748979488387704,
0.04185101017355919,
-0.052109114825725555,
-0.1506144255399704,
-0.020530881360173225,
0.025470362976193428,
-0.01472426112741232,
0.024750422686338425,
0.10498194396495819,
-0.009238746017217636,
-0.005517279263585806,
0.02428433671593666,
-0.18134643137454987,
-0.06967028230428696,
-0.04506104439496994,
0.1243395283818245,
0.06327683478593826,
-0.08080002665519714,
-0.07347151637077332,
0.06439963728189468,
0.08443721383810043,
-0.24694232642650604,
-0.11419977247714996,
-0.06587439775466919,
-0.013503062538802624,
0.01926305703818798,
-0.02736402302980423,
0.0780276209115982,
-0.0301311407238245,
0.11296343058347702,
-0.11281635612249374,
-0.09766292572021484,
0.010447924956679344,
-0.14278562366962433,
-0.18069368600845337,
-0.14220936596393585,
0.038854118436574936,
0.05647347867488861,
-0.0031307300087064505,
0.009784754365682602,
-0.004032365046441555,
0.000013841895452060271,
-0.0361136794090271,
0.03666874021291733,
0.17093315720558167,
-0.10405860841274261,
0.0759986937046051,
-0.0015034576645120978,
-0.12163170427083969,
-0.08392702788114548,
-0.006243256852030754,
0.061352357268333435,
0.21605220437049866,
-0.05924953520298004,
0.12821146845817566,
0.1300095170736313,
-0.10932659357786179,
-0.20684657990932465,
-0.11051912605762482,
0.07041686773300171,
0.036073148250579834,
0.012614528648555279,
-0.17626668512821198,
0.19062228500843048,
-0.01454654149711132,
-0.0076149809174239635,
0.06102614477276802,
-0.34304285049438477,
-0.14076083898544312,
0.04890097677707672,
0.1810893565416336,
0.2168155461549759,
-0.10834208875894547,
-0.05695496127009392,
0.07656584680080414,
-0.10858771950006485,
0.17970846593379974,
0.024866167455911636,
0.0335363931953907,
-0.0060973563231527805,
-0.013560153543949127,
0.03032909706234932,
-0.02246132865548134,
0.08825413882732391,
0.009962821379303932,
0.006868452299386263,
-0.052054256200790405,
0.02172018028795719,
0.14087039232254028,
-0.024662034586071968,
0.008375182747840881,
-0.101612389087677,
0.051577456295490265,
-0.10547662526369095,
-0.0018293560715392232,
0.021357450634241104,
-0.02508017234504223,
-0.038160331547260284,
-0.14046978950500488,
-0.056055109947919846,
0.035152617841959,
0.07469482719898224,
-0.009039944969117641,
-0.07860296964645386,
0.0013468086253851652,
-0.051869530230760574,
0.1863320916891098,
-0.06944096088409424,
-0.03311065956950188,
-0.10852085798978806,
-0.033837754279375076,
-0.0589669905602932,
0.12342872470617294,
-0.05793972685933113,
-0.02157781459391117,
0.13369041681289673,
0.0724337175488472,
0.08254023641347885,
0.025552092120051384,
-0.11418554186820984,
0.04634954780340195,
0.08694926649332047,
-0.1433376520872116,
-0.07406092435121536,
-0.0039275591261684895,
0.008448929525911808,
0.0784582644701004,
-0.05337563529610634,
0.1587693989276886,
-0.11007858067750931,
0.03187442943453789,
-0.004794721957296133,
0.02465229295194149,
0.00890644546598196,
0.0997590571641922,
-0.03955357149243355,
0.05351511761546135,
-0.05090713128447533,
0.059834957122802734,
-0.026164282113313675,
-0.08453208953142166,
0.056650903075933456,
0.0043128603138029575,
-0.11235427856445312,
0.02891121245920658,
-0.05825294926762581,
0.19279615581035614,
-0.08331088721752167,
-0.06273503601551056,
-0.12451475113630295,
-0.12781699001789093,
0.012773485854268074,
0.09157165884971619,
0.05353332683444023,
0.05339282006025314,
-0.02141699008643627,
-0.041953980922698975,
-0.05927271768450737,
0.0738169252872467,
0.07151415944099426,
0.06564196944236755,
-0.26089605689048767,
-0.0010234107030555606,
0.04868254065513611,
-0.030184375122189522,
-0.06644900143146515,
-0.030820487067103386,
-0.09669893234968185,
-0.007415030617266893,
0.009849547408521175,
0.16041867434978485,
-0.06853730231523514,
-0.059496454894542694,
0.000021345829736674204,
-0.016301129013299942,
0.0358765535056591,
0.03832390159368515,
0.018506033346056938,
0.013408827595412731,
0.0066403732635080814,
-0.01035266276448965,
-0.04012423753738403,
-0.07048112154006958,
-0.024078356102108955,
-0.057817570865154266,
0.05277242138981819,
-0.03228854760527611,
-0.09642840176820755,
-0.010490423068404198,
-0.272365540266037,
0.056445639580488205,
0.19259659945964813,
-0.036631498485803604,
-0.030798524618148804,
0.010337693616747856,
-0.034368474036455154,
-0.021034223958849907,
-0.003598974784836173,
0.008843470364809036,
0.1219087615609169,
-0.09428784996271133,
-0.054500795900821686,
0.005296439863741398,
0.08081487566232681,
-0.07091823220252991,
0.029649240896105766,
0.13037173449993134,
0.10041121393442154,
0.1370924562215805,
-0.19303849339485168,
0.0803687572479248,
-0.062676802277565,
-0.02465437725186348,
0.010251213796436787,
-0.036397747695446014,
0.04622329771518707,
-0.02623864822089672,
-0.030130447819828987,
0.011537681333720684,
0.1333363801240921,
0.02826593443751335,
-0.17429207265377045,
0.012372384779155254,
-0.018562188372015953,
-0.011111228726804256,
0.03232961893081665,
0.26419270038604736,
-0.006305791437625885,
0.024701859802007675,
-0.12319394201040268,
0.05451814830303192,
0.14488321542739868,
0.15658524632453918,
0.056791581213474274,
0.03528905287384987,
0.055245451629161835,
0.09689752012491226,
0.07237932085990906,
0.09304287284612656,
-0.08490324765443802,
0.08682040125131607,
-0.1693151593208313,
0.14735889434814453,
-0.06554339826107025,
0.0176087599247694,
0.08163106441497803,
-0.03787434473633766,
-0.035940323024988174,
0.10674439370632172,
-0.024515075609087944,
-0.000050596230721566826,
-0.039064884185791016,
-0.01881132461130619,
-0.13956093788146973,
0.005513942800462246,
-0.06705734133720398,
-0.06911590695381165,
0.04473710060119629,
0.01572786644101143,
0.02321556583046913,
0.1324044167995453,
0.07259172201156616,
0.008667116053402424,
0.08557499945163727,
-0.012931490316987038,
-0.08322475105524063,
0.06070796772837639,
-0.0015351289184764028,
0.056111596524715424,
0.1099221408367157,
-0.03576826676726341,
0.06559547036886215,
0.05158513784408569,
0.012525158002972603,
0.03679477795958519,
-0.03360236436128616,
0.018595194444060326,
-0.00018392950005363673,
-0.036317165940999985,
0.12964192032814026,
0.1262865513563156,
-0.05137961357831955,
-0.068642258644104,
0.14542756974697113,
-0.022819332778453827,
-0.08738582581281662,
-0.13328798115253448,
0.11937154084444046,
-0.11056319624185562,
0.0707874447107315,
-0.016115780919790268,
-0.11700364202260971,
-0.07252953201532364,
0.12561334669589996,
0.09182272851467133,
-0.0018558199517428875,
0.01325555145740509,
-0.07243038713932037,
-0.009479677304625511,
-0.06399936974048615,
0.06285195052623749,
0.020456325262784958,
0.30524492263793945,
-0.03124091401696205,
0.035471122711896896,
-0.10418223589658737,
-0.15228451788425446,
-0.1284438520669937,
-0.19489845633506775,
0.04028194025158882,
-0.003448279108852148,
-0.07300437241792679,
0.04631310701370239,
-0.19544720649719238,
-0.13318511843681335,
0.22910049557685852,
-0.13051867485046387,
-0.006049108691513538,
-0.06355120241641998,
0.08102425187826157,
0.04136744886636734,
0.042716361582279205,
-0.06360317021608353,
0.06693717837333679,
0.08578619360923767,
-0.05391526222229004,
-0.05810607224702835,
0.03957488015294075,
-0.09704812616109848,
-0.26184648275375366,
0.18960055708885193,
-0.03243452310562134,
0.022890271618962288,
0.033013615757226944,
-0.0041992394253611565,
-0.11723535507917404,
0.012576503679156303,
-0.07184143364429474,
-0.030404606834053993,
-0.08869021385908127,
0.13570116460323334,
0.019121913239359856,
0.04548736661672592,
0.025677694007754326,
-0.09670752286911011,
-0.012598455883562565,
0.021526765078306198,
-0.019707808271050453,
-0.10052625089883804,
0.04746342822909355,
-0.036312419921159744,
0.06158814579248428,
0.06548096239566803,
-0.04471321776509285,
0.06181740760803223,
0.03334301710128784,
0.01047989446669817,
-0.027253633365035057,
-0.011930174194276333,
0.07439155876636505,
-0.11437784880399704,
0.027368219569325447,
0.06209154799580574,
-0.016389088705182076,
-0.2857191860675812,
-0.12683843076229095,
-0.13619506359100342,
0.013528205454349518,
-0.027665212750434875,
0.07227321714162827,
0.2043352574110031,
0.03838711977005005,
0.008567670360207558,
-0.16921767592430115,
-0.010618365369737148,
0.07347100228071213,
0.04797480255365372,
-0.11707714945077896
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.7.1 | {"library_name": "peft", "base_model": "TheBloke/Nous-Hermes-2-SOLAR-10.7B-GPTQ"} | null | PornMixer/LoRA | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:TheBloke/Nous-Hermes-2-SOLAR-10.7B-GPTQ",
"region:us"
] | 2024-02-07T07:09:22+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #safetensors #arxiv-1910.09700 #base_model-TheBloke/Nous-Hermes-2-SOLAR-10.7B-GPTQ #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.7.1 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.7.1"
] | [
"TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-TheBloke/Nous-Hermes-2-SOLAR-10.7B-GPTQ #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.7.1"
] | [
49,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-TheBloke/Nous-Hermes-2-SOLAR-10.7B-GPTQ #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.7.1"
] | [
-0.10011543333530426,
0.19984550774097443,
-0.0034518714528530836,
0.021599847823381424,
0.08469174057245255,
0.02162308432161808,
0.06973709911108017,
0.12053845077753067,
0.01966557838022709,
0.13381998240947723,
0.04983034357428551,
0.10210467875003815,
0.1265787035226822,
0.21692171692848206,
-0.004038906190544367,
-0.19670622050762177,
0.027545232325792313,
-0.07753656804561615,
0.004939262755215168,
0.11741620302200317,
0.13325224816799164,
-0.09465426951646805,
0.07153429836034775,
-0.026208091527223587,
0.000754866108763963,
-0.03674822300672531,
-0.07233580201864243,
-0.02858838438987732,
0.05250626429915428,
0.04811965674161911,
0.03773539885878563,
-0.000007431575795635581,
0.09156256914138794,
-0.27279144525527954,
0.008756092749536037,
0.049986034631729126,
-0.003207660745829344,
0.08768115937709808,
0.10401393473148346,
-0.025355754420161247,
0.12353438884019852,
-0.04439607262611389,
0.13304024934768677,
0.07577473670244217,
-0.09219759702682495,
-0.2284012734889984,
-0.07080194354057312,
0.08408749848604202,
0.1832183450460434,
0.0661805048584938,
-0.0373089425265789,
0.13038356602191925,
-0.0704445019364357,
0.01336069218814373,
0.0733080506324768,
-0.09988495707511902,
-0.0714234784245491,
0.07402066886425018,
0.11392123252153397,
0.08557135611772537,
-0.12009678035974503,
-0.033435478806495667,
0.03123960644006729,
0.037488993257284164,
0.08388682454824448,
0.015478521585464478,
0.16712500154972076,
0.03593394532799721,
-0.1456713080406189,
-0.0563538558781147,
0.14134173095226288,
0.024850042536854744,
-0.04077357426285744,
-0.23132555186748505,
-0.015456721186637878,
-0.06045381352305412,
-0.03387882933020592,
-0.05567502602934837,
0.03836469352245331,
0.008348758332431316,
0.10402250289916992,
-0.033156074583530426,
-0.08095410466194153,
-0.01657908409833908,
0.10910320281982422,
0.08821762353181839,
0.015353989787399769,
-0.014234719797968864,
0.014237852767109871,
0.13152435421943665,
0.05778102949261665,
-0.12391265481710434,
-0.047419436275959015,
-0.06549232453107834,
-0.04010404646396637,
-0.0252122413367033,
0.054758746176958084,
0.04030633717775345,
0.04164772853255272,
0.25864842534065247,
-0.015287918969988823,
0.05633295699954033,
0.05713339522480965,
0.014859852381050587,
0.03132220730185509,
0.09816354513168335,
-0.04791233316063881,
-0.18822675943374634,
-0.02230593003332615,
0.09424133598804474,
0.014290071092545986,
-0.024494187906384468,
-0.039964672178030014,
0.0310208797454834,
0.02865355834364891,
0.11468209326267242,
0.11118196696043015,
-0.0222463458776474,
-0.06552641093730927,
-0.056111082434654236,
0.22487494349479675,
-0.14561279118061066,
0.04748567193746567,
0.01830347627401352,
-0.014806374907493591,
-0.04722954332828522,
0.012777624651789665,
0.011689762584865093,
-0.03424371778964996,
0.10614129155874252,
-0.06127778813242912,
-0.05279256030917168,
-0.11158314347267151,
-0.043979186564683914,
0.03347106650471687,
0.011720947921276093,
-0.03845055401325226,
-0.03222636505961418,
-0.08622606843709946,
-0.08947084844112396,
0.09281636029481888,
-0.06556084752082825,
-0.08008593320846558,
-0.017552224919199944,
-0.07009816914796829,
0.02571456879377365,
0.019422680139541626,
0.08454881608486176,
-0.027540065348148346,
0.040878091007471085,
-0.02966267243027687,
0.06822055578231812,
0.09586042910814285,
0.032897986471652985,
-0.07408582419157028,
0.06825748831033707,
-0.1871829479932785,
0.07703787088394165,
-0.07943654805421829,
0.030931172892451286,
-0.16074785590171814,
-0.01105138286948204,
0.000317205092869699,
0.02052835002541542,
0.03539115563035011,
0.15692433714866638,
-0.19052736461162567,
-0.03313443809747696,
0.16131514310836792,
-0.10742340236902237,
-0.10654831677675247,
0.038282159715890884,
-0.031944312155246735,
0.1568203568458557,
0.026884058490395546,
-0.002598162041977048,
0.0825304463505745,
-0.14142313599586487,
-0.01782327890396118,
-0.024330543354153633,
0.019273141399025917,
0.077140212059021,
0.07334237545728683,
-0.08577253669500351,
0.006607095245271921,
0.01644865982234478,
-0.06530345231294632,
-0.00510529475286603,
-0.040776126086711884,
-0.09610450267791748,
0.00618163775652647,
-0.08160769194364548,
0.01073537953197956,
0.009788998402655125,
-0.08461124449968338,
-0.017117448151111603,
-0.14417128264904022,
-0.038738105446100235,
0.08115411549806595,
0.00808398425579071,
-0.014544068835675716,
-0.06685274839401245,
0.029734959825873375,
-0.03718303143978119,
-0.018052387982606888,
-0.13708224892616272,
-0.018240539357066154,
0.035839855670928955,
-0.15467360615730286,
-0.007199639454483986,
-0.119779072701931,
0.0673796758055687,
0.012544794008135796,
-0.06656042486429214,
-0.038364946842193604,
0.021099349483847618,
-0.001130408956669271,
-0.058692384511232376,
-0.21804502606391907,
-0.0369718000292778,
-0.03927532956004143,
0.14266140758991241,
-0.2236030101776123,
0.0451388917863369,
0.01807735674083233,
0.12160427868366241,
0.007245585788041353,
-0.06839415431022644,
0.02730628289282322,
-0.059574343264102936,
-0.02182965911924839,
-0.07587189227342606,
-0.010573409497737885,
-0.0027459824923425913,
-0.03697500377893448,
0.025284286588430405,
-0.1497042328119278,
-0.061585575342178345,
0.09110169112682343,
0.08377192169427872,
-0.14651161432266235,
0.011293936520814896,
-0.04464523866772652,
-0.06633241474628448,
-0.08328619599342346,
-0.0716790109872818,
0.05562806501984596,
0.04658832773566246,
0.046172816306352615,
-0.07383652031421661,
-0.06950405240058899,
0.006266815587878227,
-0.009876357391476631,
-0.026714257895946503,
0.11932722479104996,
0.08901988714933395,
-0.07789056748151779,
0.0924302265048027,
0.07679087668657303,
0.047110915184020996,
0.08049144595861435,
-0.011162182316184044,
-0.10267847031354904,
-0.03203171119093895,
0.05148942396044731,
0.017985841259360313,
0.15009279549121857,
-0.05520748347043991,
0.0488947257399559,
0.05378099903464317,
-0.0435045063495636,
0.03789263218641281,
-0.10174588859081268,
0.015571072697639465,
0.011069949716329575,
-0.010698487982153893,
0.04275176301598549,
-0.01858624815940857,
0.006216492969542742,
0.08238917589187622,
0.061749327927827835,
0.028303761035203934,
0.02085365727543831,
-0.03752323240041733,
-0.13786093890666962,
0.1725124716758728,
-0.08617261797189713,
-0.22229686379432678,
-0.15386563539505005,
0.03714572265744209,
0.05955427139997482,
-0.02084227278828621,
0.02809171937406063,
-0.033434007316827774,
-0.1108470931649208,
-0.08295315504074097,
0.026753723621368408,
0.03819132596254349,
-0.06433703750371933,
-0.055750004947185516,
0.03512994199991226,
0.03520825505256653,
-0.12707927823066711,
0.03196914494037628,
0.05617063492536545,
-0.00541513180360198,
-0.0005281172343529761,
0.043392568826675415,
0.08656716346740723,
0.18689167499542236,
0.009799916297197342,
-0.0046644024550914764,
0.05074648931622505,
0.2859128713607788,
-0.1565086394548416,
0.12985853850841522,
0.12813088297843933,
-0.0401851087808609,
0.08841293305158615,
0.1943761706352234,
0.043469808995723724,
-0.08061361312866211,
0.02186436578631401,
0.02809552475810051,
-0.03779121860861778,
-0.2590906023979187,
-0.07000578194856644,
-0.023091141134500504,
-0.070566326379776,
0.0872756838798523,
0.0883045345544815,
0.09951240569353104,
0.04468987137079239,
-0.08223925530910492,
-0.0638936311006546,
0.04881645366549492,
0.11833694577217102,
-0.030196571722626686,
0.021870572119951248,
0.08443208038806915,
-0.040706075727939606,
0.006237310357391834,
0.09477083384990692,
-0.013029908761382103,
0.15196755528450012,
0.049533870071172714,
0.10736607015132904,
0.0628935918211937,
0.08043087273836136,
-0.007970614358782768,
0.051655836403369904,
0.015022102743387222,
0.024868547916412354,
0.010502351447939873,
-0.09114103764295578,
0.03241870924830437,
0.12447164952754974,
0.017436305060982704,
0.03187257796525955,
0.025857839733362198,
-0.06727337092161179,
0.03499499335885048,
0.2026103138923645,
0.011157756671309471,
-0.18983444571495056,
-0.07559886574745178,
0.07263948768377304,
-0.0854104608297348,
-0.1450580358505249,
-0.004928472451865673,
0.026062456890940666,
-0.1746186912059784,
0.02631310187280178,
-0.0413331501185894,
0.10520313680171967,
-0.07004868984222412,
-0.04241139814257622,
0.09654097259044647,
0.057929590344429016,
-0.031109655275940895,
0.05105572193861008,
-0.1634128987789154,
0.11349645256996155,
0.027431586757302284,
0.06672204285860062,
-0.09468702226877213,
0.09925594180822372,
0.0012681151274591684,
-0.016837317496538162,
0.16599805653095245,
0.005611192435026169,
-0.0625992938876152,
-0.07338999211788177,
-0.06437905132770538,
-0.021701904013752937,
0.0881941094994545,
-0.1418948471546173,
0.06963834166526794,
-0.016704415902495384,
-0.03952065110206604,
0.002189931459724903,
-0.09976272284984589,
-0.11166784167289734,
-0.17125765979290009,
0.06843320280313492,
-0.07520924508571625,
0.012228247709572315,
-0.09825731813907623,
-0.05545051768422127,
-0.020588628947734833,
0.16821002960205078,
-0.19278044998645782,
-0.11289864033460617,
-0.14448319375514984,
-0.10502643138170242,
0.1739414930343628,
-0.045664217323064804,
0.08813753724098206,
-0.0007365807541646063,
0.1623907834291458,
-0.0065711745992302895,
-0.019189592450857162,
0.08862429857254028,
-0.09653344005346298,
-0.20032361149787903,
-0.046534329652786255,
0.17409661412239075,
0.12691260874271393,
0.035856943577528,
-0.027120986953377724,
0.0327429436147213,
-0.05534672737121582,
-0.11715680360794067,
0.020082583650946617,
0.13325385749340057,
0.04294029623270035,
-0.003752906573936343,
-0.016020342707633972,
-0.11763972789049149,
-0.05197015404701233,
-0.058099307119846344,
-0.014174895361065865,
0.20601952075958252,
-0.08187118172645569,
0.16266004741191864,
0.1228070855140686,
-0.04837575927376747,
-0.2023392915725708,
0.028676943853497505,
0.03215230256319046,
0.011907662265002728,
0.041406407952308655,
-0.19044312834739685,
0.08441182225942612,
-0.008451283909380436,
-0.07925394922494888,
0.15852931141853333,
-0.1908513456583023,
-0.1327408403158188,
0.09165708720684052,
0.0185592882335186,
-0.2308293879032135,
-0.1397762894630432,
-0.11612653732299805,
-0.01595824398100376,
-0.13770876824855804,
0.05122043564915657,
0.03272370994091034,
0.004007798619568348,
0.01506807655096054,
0.016384182497859,
0.04011739417910576,
-0.05099054053425789,
0.2089947611093521,
-0.02500041387975216,
0.0026703248731791973,
-0.0506284162402153,
-0.08928167819976807,
0.026863623410463333,
-0.05361895635724068,
0.10658919811248779,
-0.008810574188828468,
0.0204632468521595,
-0.15733560919761658,
-0.04198085889220238,
-0.06783848255872726,
0.016462218016386032,
-0.09527141600847244,
-0.09116069972515106,
-0.052028462290763855,
0.08024875819683075,
0.11116564273834229,
-0.024985484778881073,
0.013103376142680645,
-0.06996668875217438,
0.08777695894241333,
0.2134270817041397,
0.16374698281288147,
0.05999716371297836,
-0.057268038392066956,
0.010586588643491268,
-0.033842429518699646,
0.03937601298093796,
-0.2192983627319336,
0.036522939801216125,
0.0533607117831707,
0.03790626302361488,
0.0748705044388771,
-0.010533460415899754,
-0.15894447267055511,
-0.07760754972696304,
0.07209770381450653,
-0.06193699687719345,
-0.177755206823349,
-0.034942954778671265,
0.049799516797065735,
-0.1933162659406662,
-0.04311630502343178,
0.025649292394518852,
-0.024195734411478043,
-0.03717893734574318,
0.01780318096280098,
0.08607515692710876,
-0.004012231715023518,
0.10035494714975357,
0.07398845255374908,
0.10083708167076111,
-0.10587744414806366,
0.06955219060182571,
0.08910870552062988,
-0.03494569659233093,
0.007194401230663061,
0.13554798066616058,
-0.048891402781009674,
-0.021590063348412514,
0.0662929117679596,
0.0631977915763855,
0.0011816921178251505,
-0.052265506237745285,
0.013078747317194939,
-0.0709768757224083,
0.05984361469745636,
0.0881018415093422,
0.019543496891856194,
-0.0151747465133667,
0.06579790264368057,
0.02522902376949787,
-0.09068385511636734,
0.11697404086589813,
0.06380623579025269,
0.02549252100288868,
-0.03889762610197067,
-0.027434343472123146,
-0.016808245331048965,
-0.014642560854554176,
-0.015098699368536472,
-0.005685422569513321,
-0.07676553726196289,
-0.004503713455051184,
-0.10249137133359909,
0.013848759233951569,
-0.0874900221824646,
0.0053239064291119576,
0.015556477941572666,
-0.04428046941757202,
-0.001160699874162674,
-0.0013133364263921976,
-0.08085557818412781,
-0.05829818174242973,
-0.025509729981422424,
0.08352628350257874,
-0.12407972663640976,
0.024641629308462143,
0.07469706237316132,
-0.11627605557441711,
0.06368224322795868,
-0.012030119076371193,
0.01555030420422554,
-0.0022677944507449865,
-0.13704846799373627,
0.05670255795121193,
-0.021140338853001595,
-0.006139066070318222,
0.016737615689635277,
-0.16734357178211212,
-0.009363281540572643,
-0.054779794067144394,
-0.07804540544748306,
0.005744592286646366,
-0.041326526552438736,
-0.13220274448394775,
0.11029339581727982,
-0.014529024250805378,
-0.07106789201498032,
-0.024151543155312538,
0.050199829041957855,
0.07898671925067902,
-0.03144354745745659,
0.09049533307552338,
-0.026584196835756302,
0.07448054105043411,
-0.18018051981925964,
-0.009666595607995987,
-0.013425031676888466,
0.03630326688289642,
-0.02306383289396763,
-0.015233989804983139,
0.055429160594940186,
-0.0048938351683318615,
0.17546658217906952,
-0.023451251909136772,
0.07424000650644302,
0.05041109398007393,
0.011855022050440311,
0.033689796924591064,
0.06774409115314484,
0.07527529448270798,
-0.01016127597540617,
-0.0024306371342390776,
0.026926415041089058,
-0.014341405592858791,
-0.048624083399772644,
-0.14778831601142883,
0.0381603017449379,
0.16312290728092194,
0.06382592767477036,
0.03051215223968029,
0.012072240002453327,
-0.14036825299263,
-0.09001783281564713,
0.10902757197618484,
-0.016745341941714287,
-0.0161768589168787,
-0.0706784799695015,
0.19295363128185272,
0.12106192111968994,
-0.19724026322364807,
0.07009638845920563,
-0.04811841994524002,
-0.031017187982797623,
-0.12648892402648926,
-0.1584392935037613,
-0.05463303253054619,
-0.049473367631435394,
-0.02147478237748146,
-0.057317763566970825,
0.057698916643857956,
0.05578102171421051,
-0.0022332260850816965,
-0.0035929037258028984,
0.09350442886352539,
0.002550785895437002,
-0.026455571874976158,
0.054902251809835434,
0.06838791817426682,
0.04442873224616051,
-0.08464677631855011,
0.0030051344074308872,
-0.0034898200538009405,
0.009913599118590355,
0.05990544334053993,
0.02045399136841297,
-0.05767630413174629,
0.02723045088350773,
-0.000012291044185985811,
-0.11828390508890152,
0.0428844653069973,
-0.015610270202159882,
-0.035616111010313034,
0.14467091858386993,
0.020677125081419945,
0.014229864813387394,
-0.02531931921839714,
0.2296341210603714,
-0.09346874803304672,
-0.07215645909309387,
-0.12951606512069702,
0.0748462975025177,
-0.056385718286037445,
0.03425131365656853,
0.04155801981687546,
-0.1271195113658905,
0.00439949706196785,
0.1611509770154953,
0.12640579044818878,
0.014666215516626835,
0.008100620470941067,
0.05995776504278183,
0.002744425553828478,
-0.032190315425395966,
0.0181342214345932,
0.04784432053565979,
0.18599234521389008,
-0.07893554121255875,
0.08084291964769363,
-0.009480226784944534,
-0.07511106133460999,
-0.03082842193543911,
0.13369685411453247,
-0.012725716456770897,
-0.005098688881844282,
-0.05888896808028221,
0.14189082384109497,
-0.05094499886035919,
-0.22213128209114075,
0.06761116534471512,
-0.0855443999171257,
-0.13712133467197418,
-0.03948615491390228,
0.00903940200805664,
-0.019186638295650482,
0.01277577318251133,
0.06777361780405045,
-0.050967223942279816,
0.17815648019313812,
0.03847796842455864,
-0.060378048568964005,
-0.0876951739192009,
0.05489477887749672,
-0.13406549394130707,
0.28202539682388306,
0.023442436009645462,
0.03843826800584793,
0.09967619180679321,
-0.02760070562362671,
-0.14946171641349792,
0.018315806984901428,
0.1177562028169632,
-0.08068549633026123,
0.05843566730618477,
0.17371296882629395,
-0.006197749171406031,
0.1348010152578354,
0.05351943150162697,
-0.06612284481525421,
0.031075187027454376,
-0.04644643887877464,
-0.054170168936252594,
-0.11807858943939209,
0.0672251507639885,
-0.07228586822748184,
0.15162153542041779,
0.11968044191598892,
-0.06592606753110886,
-0.014867263846099377,
-0.018985532224178314,
0.07415646314620972,
0.012263172306120396,
0.13146115839481354,
0.023603320121765137,
-0.17958693206310272,
0.04222927615046501,
-0.013149332255125046,
0.11854638904333115,
-0.19296446442604065,
-0.059448160231113434,
0.03781304508447647,
-0.028584666550159454,
-0.08656022697687149,
0.11344410479068756,
0.05101633444428444,
0.012544621713459492,
-0.03391133248806,
-0.08138445019721985,
0.0014957701787352562,
0.15214985609054565,
-0.10725115984678268,
-0.003998590167611837
] |
null | null | diffusers |
# Gildenface XL – Headshot LoRA
<Gallery />
([CivitAI](https://civitai.com/models/183550))
## Model description
<p><strong>Gildenface XL</strong> – a LoRA focused on <em>exaggerated</em> and <em><span style="color:rgb(189, 193, 198)">less-than-glamorous</span></em> close-ups with very <em>high detailed textures</em>.</p><p>Great for producing <em>unique, grotesque and/or outlandish faces</em>, but it can be used as a <em>enhance details for faces and textures</em>, depending on weight and prompt.</p><p><strong>Trigger word: Gildenface</strong><br /><strong>Useful prompt tips:</strong> Portrait photo, close up on face, detailed skin texture, leathery skin texture, visible skin pores, swollen face, greasy hair, wrinkles, potato nose, addict, blushing, chubby, hard shadows, disgusted, blemish, facial hair, staring <br />+ general enhancers, photography terms, and portrait photographers</p><p><strong>Recommended weights:</strong> between 0.8 – 1.2</p><p>It's a bit rough around the edges, and your milage may vary – but when it hits right it's golden.</p><p><span style="color:rgb(193, 194, 197)">Be sure to check out </span><a target="_blank" rel="ugc" href="https://civitai.com/models/181092?modelVersionId=203235">Caricature XL</a><span style="color:rgb(193, 194, 197)"> LoRA by Blink, if you like creating weird</span></p>
## Trigger words
You should use `Gildenface portrait photo` to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
[Download](/WizWhite/gildenface-xl-headshot-lora/tree/main) them in the Files & versions tab.
## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers)
```py
from diffusers import AutoPipelineForText2Image
import torch
pipeline = AutoPipelineForText2Image.from_pretrained('stabilityai/stable-diffusion-xl-base-1.0', torch_dtype=torch.float16).to('cuda')
pipeline.load_lora_weights('WizWhite/gildenface-xl-headshot-lora', weight_name='GildenfaceXL_Headshot_LoRA_v1.safetensors')
image = pipeline('candid close up photo of a surprised Shrek business man, detailed skin texture, standing outside in a swamp,').images[0]
```
For more details, including weighting, merging and fusing LoRAs, check the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters)
| {"license": "other", "tags": ["text-to-image", "stable-diffusion", "lora", "diffusers", "template:sd-lora", "concept", "portrait", "detailed", "face", "grotesque", "headshot"], "license_name": "bespoke-lora-trained-license", "license_link": "https://multimodal.art/civitai-licenses?allowNoCredit=True&allowCommercialUse=RentCivit&allowDerivatives=True&allowDifferentLicense=True", "base_model": "stabilityai/stable-diffusion-xl-base-1.0", "instance_prompt": "Gildenface portrait photo", "widget": [{"text": "gildenface portrait photography of jabba the hutt, close up photo, professional, breathtaking, close-up on face, highly detailed skin, visible skin pores, dark, gritty", "output": {"url": "3277226.jpeg"}}, {"text": "gildenface portrait photography of a troll from the movie troll hunter", "output": {"url": "3277704.jpeg"}}, {"text": "gildenface portrait of a surprised shocked zombie at a birthday party, highly detailed texture, sharp focus, party hat", "output": {"url": "3277754.jpeg"}}, {"text": "gildenface breathtaking portrait photo inspired by an epic scene from the movie total recall, 1990s, sci-fi, professional, by bruce gilden", "output": {"url": "3277371.jpeg"}}, {"text": "Gildenface close up portrait of real life luigi from (super mario bros:0.5), disgusted", "output": {"url": "3277428.jpeg"}}, {"text": "Gildenface close up portrait of real life super mario, disgusted", "output": {"url": "3277429.jpeg"}}, {"text": "obese (cthulhu:1.4), gildenface style photo, close up on face, detailed skin texture, by bruce gilden", "output": {"url": "3277426.jpeg"}}, {"text": "rusty cast-iron (robot:1.4), gildenface style photo, close up on face, detailed skin texture", "output": {"url": "3277421.jpeg"}}, {"text": "Gildenface close up portrait of real life pikachu by bruce gilden", "output": {"url": "3277419.jpeg"}}, {"text": "candid close up photo of a surprised Shrek business man, detailed skin texture, standing outside in a swamp,", "output": {"url": "3277870.jpeg"}}]} | text-to-image | WizWhite/gildenface-xl-headshot-lora-2 | [
"diffusers",
"text-to-image",
"stable-diffusion",
"lora",
"template:sd-lora",
"concept",
"portrait",
"detailed",
"face",
"grotesque",
"headshot",
"base_model:stabilityai/stable-diffusion-xl-base-1.0",
"license:other",
"region:us"
] | 2024-02-07T07:10:24+00:00 | [] | [] | TAGS
#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #concept #portrait #detailed #face #grotesque #headshot #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-other #region-us
|
# Gildenface XL – Headshot LoRA
<Gallery />
(CivitAI)
## Model description
<p><strong>Gildenface XL</strong> – a LoRA focused on <em>exaggerated</em> and <em><span style="color:rgb(189, 193, 198)">less-than-glamorous</span></em> close-ups with very <em>high detailed textures</em>.</p><p>Great for producing <em>unique, grotesque and/or outlandish faces</em>, but it can be used as a <em>enhance details for faces and textures</em>, depending on weight and prompt.</p><p><strong>Trigger word: Gildenface</strong><br /><strong>Useful prompt tips:</strong> Portrait photo, close up on face, detailed skin texture, leathery skin texture, visible skin pores, swollen face, greasy hair, wrinkles, potato nose, addict, blushing, chubby, hard shadows, disgusted, blemish, facial hair, staring <br />+ general enhancers, photography terms, and portrait photographers</p><p><strong>Recommended weights:</strong> between 0.8 – 1.2</p><p>It's a bit rough around the edges, and your milage may vary – but when it hits right it's golden.</p><p><span style="color:rgb(193, 194, 197)">Be sure to check out </span><a target="_blank" rel="ugc" href="URL XL</a><span style="color:rgb(193, 194, 197)"> LoRA by Blink, if you like creating weird</span></p>
## Trigger words
You should use 'Gildenface portrait photo' to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
Download them in the Files & versions tab.
## Use it with the diffusers library
For more details, including weighting, merging and fusing LoRAs, check the documentation on loading LoRAs in diffusers
| [
"# Gildenface XL – Headshot LoRA \n\n<Gallery />\n\n\n\n(CivitAI)",
"## Model description\n\n<p><strong>Gildenface XL</strong> – a LoRA focused on <em>exaggerated</em> and <em><span style=\"color:rgb(189, 193, 198)\">less-than-glamorous</span></em> close-ups with very <em>high detailed textures</em>.</p><p>Great for producing <em>unique, grotesque and/or outlandish faces</em>, but it can be used as a <em>enhance details for faces and textures</em>, depending on weight and prompt.</p><p><strong>Trigger word: Gildenface</strong><br /><strong>Useful prompt tips:</strong> Portrait photo, close up on face, detailed skin texture, leathery skin texture, visible skin pores, swollen face, greasy hair, wrinkles, potato nose, addict, blushing, chubby, hard shadows, disgusted, blemish, facial hair, staring <br />+ general enhancers, photography terms, and portrait photographers</p><p><strong>Recommended weights:</strong> between 0.8 – 1.2</p><p>It's a bit rough around the edges, and your milage may vary – but when it hits right it's golden.</p><p><span style=\"color:rgb(193, 194, 197)\">Be sure to check out </span><a target=\"_blank\" rel=\"ugc\" href=\"URL XL</a><span style=\"color:rgb(193, 194, 197)\"> LoRA by Blink, if you like creating weird</span></p>",
"## Trigger words\nYou should use 'Gildenface portrait photo' to trigger the image generation.",
"## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab.",
"## Use it with the diffusers library\n\n\n\nFor more details, including weighting, merging and fusing LoRAs, check the documentation on loading LoRAs in diffusers"
] | [
"TAGS\n#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #concept #portrait #detailed #face #grotesque #headshot #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-other #region-us \n",
"# Gildenface XL – Headshot LoRA \n\n<Gallery />\n\n\n\n(CivitAI)",
"## Model description\n\n<p><strong>Gildenface XL</strong> – a LoRA focused on <em>exaggerated</em> and <em><span style=\"color:rgb(189, 193, 198)\">less-than-glamorous</span></em> close-ups with very <em>high detailed textures</em>.</p><p>Great for producing <em>unique, grotesque and/or outlandish faces</em>, but it can be used as a <em>enhance details for faces and textures</em>, depending on weight and prompt.</p><p><strong>Trigger word: Gildenface</strong><br /><strong>Useful prompt tips:</strong> Portrait photo, close up on face, detailed skin texture, leathery skin texture, visible skin pores, swollen face, greasy hair, wrinkles, potato nose, addict, blushing, chubby, hard shadows, disgusted, blemish, facial hair, staring <br />+ general enhancers, photography terms, and portrait photographers</p><p><strong>Recommended weights:</strong> between 0.8 – 1.2</p><p>It's a bit rough around the edges, and your milage may vary – but when it hits right it's golden.</p><p><span style=\"color:rgb(193, 194, 197)\">Be sure to check out </span><a target=\"_blank\" rel=\"ugc\" href=\"URL XL</a><span style=\"color:rgb(193, 194, 197)\"> LoRA by Blink, if you like creating weird</span></p>",
"## Trigger words\nYou should use 'Gildenface portrait photo' to trigger the image generation.",
"## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab.",
"## Use it with the diffusers library\n\n\n\nFor more details, including weighting, merging and fusing LoRAs, check the documentation on loading LoRAs in diffusers"
] | [
79,
20,
397,
22,
28,
38
] | [
"passage: TAGS\n#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #concept #portrait #detailed #face #grotesque #headshot #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-other #region-us \n# Gildenface XL – Headshot LoRA \n\n<Gallery />\n\n\n\n(CivitAI)## Model description\n\n<p><strong>Gildenface XL</strong> – a LoRA focused on <em>exaggerated</em> and <em><span style=\"color:rgb(189, 193, 198)\">less-than-glamorous</span></em> close-ups with very <em>high detailed textures</em>.</p><p>Great for producing <em>unique, grotesque and/or outlandish faces</em>, but it can be used as a <em>enhance details for faces and textures</em>, depending on weight and prompt.</p><p><strong>Trigger word: Gildenface</strong><br /><strong>Useful prompt tips:</strong> Portrait photo, close up on face, detailed skin texture, leathery skin texture, visible skin pores, swollen face, greasy hair, wrinkles, potato nose, addict, blushing, chubby, hard shadows, disgusted, blemish, facial hair, staring <br />+ general enhancers, photography terms, and portrait photographers</p><p><strong>Recommended weights:</strong> between 0.8 – 1.2</p><p>It's a bit rough around the edges, and your milage may vary – but when it hits right it's golden.</p><p><span style=\"color:rgb(193, 194, 197)\">Be sure to check out </span><a target=\"_blank\" rel=\"ugc\" href=\"URL XL</a><span style=\"color:rgb(193, 194, 197)\"> LoRA by Blink, if you like creating weird</span></p>"
] | [
0.024477822706103325,
-0.025832649320364,
-0.007977603003382683,
0.030978973954916,
0.05443288013339043,
0.012751278467476368,
0.02853185124695301,
0.1254209727048874,
0.07602408528327942,
0.14054696261882782,
-0.005379559006541967,
0.05483577027916908,
0.08539647608995438,
0.048581935465335846,
0.06678155809640884,
-0.26244282722473145,
-0.024064159020781517,
-0.0543481819331646,
0.08970627933740616,
0.08000227808952332,
0.09790943562984467,
-0.019399888813495636,
0.06264916062355042,
-0.020395854488015175,
-0.04651745408773422,
0.011873748153448105,
-0.0005521855200640857,
0.0007473547011613846,
0.04031308740377426,
0.12070026993751526,
0.018242236226797104,
-0.005241277161985636,
0.0038566014263778925,
-0.31619638204574585,
0.024221861734986305,
0.06513802707195282,
-0.09408064931631088,
0.019648903980851173,
0.11378970742225647,
-0.10025301575660706,
0.12160985916852951,
-0.2062842845916748,
0.045765768736600876,
0.033598627895116806,
-0.13822239637374878,
-0.16812731325626373,
-0.03979652374982834,
0.06660135835409164,
0.09210105240345001,
0.0076385922729969025,
-0.051300276070833206,
-0.012380761094391346,
-0.09592242538928986,
0.0933884009718895,
0.25693145394325256,
-0.14055849611759186,
-0.06903857737779617,
-0.05992136895656586,
0.074236199259758,
-0.012667334638535976,
-0.15422455966472626,
0.03788631781935692,
-0.04721272736787796,
0.019376548007130623,
0.0602671317756176,
0.00008556133980164304,
0.09155762195587158,
-0.010564710944890976,
-0.07369879633188248,
-0.006668337155133486,
0.03759826719760895,
0.09242519736289978,
-0.05848817899823189,
-0.23981261253356934,
-0.03990533575415611,
0.04784465581178665,
-0.09392734616994858,
-0.05453930422663689,
0.025467399507761,
0.0024931563530117273,
-0.05710030347108841,
-0.06817743927240372,
-0.026687225326895714,
0.005197432357817888,
-0.006308067589998245,
0.11641529947519302,
0.0028136176988482475,
-0.04350490868091583,
0.035518769174814224,
0.03472351282835007,
-0.03296707198023796,
-0.09634895622730255,
-0.012063199654221535,
-0.05054757744073868,
-0.11480129510164261,
0.010662968270480633,
-0.015926942229270935,
0.05838174372911453,
0.10079199820756912,
0.29238882660865784,
0.13703982532024384,
0.0637131929397583,
-0.12467050552368164,
0.021860700100660324,
-0.0016355266561731696,
0.09014544636011124,
-0.032025426626205444,
-0.1689281314611435,
-0.02157188206911087,
-0.01150943897664547,
0.0662410706281662,
-0.04830309748649597,
-0.00942035298794508,
-0.018287096172571182,
-0.008762667886912823,
0.012148626148700714,
0.10121625661849976,
0.03078075870871544,
-0.08255559206008911,
-0.024371011182665825,
0.1861228346824646,
-0.11420901864767075,
0.0438210628926754,
0.04186297580599785,
-0.0262519009411335,
-0.070162832736969,
0.00023517898807767779,
0.03754265978932381,
0.03142548352479935,
0.07734742760658264,
-0.0712563544511795,
-0.04011335223913193,
-0.06553242355585098,
-0.06876594573259354,
0.0138624869287014,
0.013511781580746174,
-0.052113499492406845,
-0.03202572092413902,
-0.028531203046441078,
-0.08088330924510956,
0.012002967298030853,
-0.10739538073539734,
-0.013724718242883682,
-0.03821372613310814,
-0.03784986957907677,
0.06989393383264542,
0.06336360424757004,
0.06391482800245285,
-0.03333892300724983,
0.08409469574689865,
-0.02904975414276123,
0.09758617728948593,
0.10346172749996185,
0.017287375405430794,
-0.03493090718984604,
0.03919948637485504,
-0.2363540381193161,
0.11967209726572037,
-0.08317719399929047,
0.010343346744775772,
-0.06842443346977234,
-0.0440908707678318,
-0.07320040464401245,
0.0294799841940403,
-0.014372806996107101,
0.08324703574180603,
-0.1463019847869873,
-0.10112301260232925,
0.2672567367553711,
-0.07609859108924866,
0.04186573997139931,
0.12079957127571106,
-0.016751553863286972,
-0.01307622343301773,
0.04723221808671951,
0.18906959891319275,
0.049816787242889404,
-0.10378322750329971,
-0.00862488616257906,
-0.014284513890743256,
-0.07629859447479248,
0.23112110793590546,
0.05821400135755539,
-0.11031899601221085,
0.024273797869682312,
0.04183739796280861,
-0.02106463722884655,
0.015400867909193039,
-0.003505406901240349,
-0.003837228985503316,
0.04085754603147507,
0.05424559488892555,
0.12431558966636658,
-0.0020181748550385237,
-0.1326470971107483,
-0.04984363541007042,
-0.17753762006759644,
-0.06447744369506836,
0.028862928971648216,
0.0026827952824532986,
0.02950826659798622,
-0.10455828160047531,
0.0072736735455691814,
0.106313057243824,
-0.014651245437562466,
-0.08390618115663528,
-0.13245254755020142,
0.0066016679629683495,
-0.06966912001371384,
0.08167694509029388,
-0.06396333128213882,
0.07546386122703552,
0.04136984050273895,
-0.06140679493546486,
0.008803757838904858,
-0.02798495814204216,
-0.05284513905644417,
-0.021297989413142204,
-0.16264460980892181,
0.025585127994418144,
-0.04026975482702255,
0.16215205192565918,
-0.06317248195409775,
0.0011606249026954174,
0.10002727061510086,
0.10493394732475281,
0.10109193623065948,
-0.0653834268450737,
0.011563087813556194,
-0.04647544026374817,
0.029260799288749695,
-0.025937428697943687,
0.05167628824710846,
-0.040459226816892624,
-0.05658926069736481,
0.01771998219192028,
-0.16536588966846466,
-0.11999403685331345,
0.07811352610588074,
0.0023644212633371353,
-0.12990893423557281,
-0.11074825376272202,
-0.04356105253100395,
-0.030814679339528084,
-0.029583662748336792,
-0.04279112070798874,
0.08018213510513306,
0.04633590951561928,
-0.0069205062463879585,
-0.01468832977116108,
-0.04136372357606888,
-0.028272757306694984,
-0.026151606813073158,
-0.051996860653162,
0.06217476725578308,
0.0207666028290987,
-0.08594377338886261,
0.006416519172489643,
0.02929023467004299,
0.032003533095121384,
0.16003155708312988,
0.08752590417861938,
-0.04954281449317932,
-0.07767818868160248,
0.051031261682510376,
0.056150052696466446,
0.04107046127319336,
-0.10283216834068298,
0.032603465020656586,
0.03309081867337227,
-0.04769890010356903,
-0.009129960089921951,
-0.07286353409290314,
0.011786267161369324,
-0.028576575219631195,
-0.021319901570677757,
0.06682726740837097,
0.05814748257398605,
0.023028293624520302,
0.05843973904848099,
0.07408545911312103,
0.010921439155936241,
-0.03710160031914711,
-0.05890234559774399,
-0.03370530530810356,
0.08161061257123947,
-0.09441305696964264,
-0.21386587619781494,
-0.10730947554111481,
-0.10384032875299454,
-0.06293010711669922,
0.030257241800427437,
0.039282817393541336,
-0.08161863684654236,
-0.04123193770647049,
-0.015998294577002525,
0.0859755128622055,
0.09417692571878433,
-0.06125788018107414,
-0.09967873245477676,
0.0034997027833014727,
0.02542920596897602,
-0.0999317318201065,
-0.012083352543413639,
-0.0237825158983469,
-0.11651439219713211,
0.01734861172735691,
0.0022128813434392214,
0.13026200234889984,
0.029554763808846474,
0.08780639618635178,
-0.01237211562693119,
-0.04046712443232536,
0.1875983476638794,
-0.13919180631637573,
0.07690232247114182,
0.07891456037759781,
0.0265741478651762,
0.10433951020240784,
0.22503739595413208,
0.03395378962159157,
-0.09132503718137741,
0.03607284650206566,
0.1416323482990265,
-0.019383350387215614,
-0.06826899945735931,
-0.0523674413561821,
-0.02547159418463707,
-0.05100321024656296,
0.056691158562898636,
0.06981347501277924,
0.03079739585518837,
0.016367724165320396,
-0.07608472555875778,
-0.028698479756712914,
0.1484074741601944,
0.10961998254060745,
0.12619775533676147,
-0.0011127303587272763,
0.10102841258049011,
0.024131134152412415,
-0.057681187987327576,
0.07144208997488022,
-0.027385400608181953,
0.11792174726724625,
0.026234665885567665,
0.1519816815853119,
0.1077800765633583,
0.09547140449285507,
0.012376694008708,
-0.11486943066120148,
-0.016933780163526535,
-0.015452876687049866,
-0.022756630554795265,
-0.07810608297586441,
0.012933784164488316,
0.0441063717007637,
0.10581593960523605,
-0.0862768217921257,
0.02781667374074459,
-0.041187938302755356,
0.11509446054697037,
0.07202466577291489,
-0.0036113145761191845,
-0.08323387801647186,
0.056844562292099,
0.06337245553731918,
-0.027784010395407677,
-0.08275353163480759,
-0.004441592376679182,
0.022382529452443123,
-0.13196961581707,
0.11327427625656128,
-0.04233992472290993,
0.05881869047880173,
-0.09375744313001633,
-0.02523413859307766,
0.069567009806633,
0.07336559891700745,
0.023037778213620186,
0.06685274839401245,
-0.08656738698482513,
0.08755875378847122,
0.018915919587016106,
0.020438827574253082,
0.0068990010768175125,
0.04745570942759514,
0.05261510610580444,
-0.04598397761583328,
0.18746860325336456,
0.007737087085843086,
-0.11881888657808304,
-0.0828147605061531,
0.026777584105730057,
-0.015061515383422375,
0.20653094351291656,
-0.117758609354496,
0.07396022230386734,
-0.04587875306606293,
-0.069242462515831,
-0.05471565201878548,
0.08703738451004028,
-0.17165294289588928,
-0.14093820750713348,
0.04023607820272446,
-0.1483038365840912,
0.0721401572227478,
-0.0449230931699276,
0.02187834307551384,
-0.11723355203866959,
0.1847088485956192,
-0.12581847608089447,
-0.04065827280282974,
-0.09226042032241821,
0.009944004006683826,
0.063149593770504,
-0.09535027295351028,
0.02340637519955635,
0.027969790622591972,
0.1711602360010147,
-0.0015700234798714519,
-0.008095747791230679,
0.05611177533864975,
-0.008819079957902431,
-0.09478934854269028,
-0.04220610857009888,
0.14120256900787354,
0.04999922215938568,
0.020391318947076797,
0.01916714385151863,
0.06526824086904526,
-0.00844577420502901,
-0.09487779438495636,
0.12304490059614182,
0.005082282703369856,
-0.08739763498306274,
0.06560081988573074,
-0.050037700682878494,
-0.10430117696523666,
-0.11010462790727615,
0.0435524508357048,
0.09139306843280792,
0.20177853107452393,
-0.07479055970907211,
0.13690343499183655,
0.061157744377851486,
-0.078587107360363,
-0.16878174245357513,
-0.01441111508756876,
0.08697320520877838,
-0.016219425946474075,
0.019515598192811012,
-0.21351778507232666,
0.06443152576684952,
0.04489971697330475,
-0.007124008610844612,
0.02067640982568264,
-0.199379101395607,
-0.1324629932641983,
-0.028950678184628487,
0.03262357413768768,
-0.2087637484073639,
-0.11033008247613907,
-0.11907073855400085,
-0.0761321410536766,
-0.10823175311088562,
0.07337004691362381,
0.019425325095653534,
0.06477214395999908,
0.030820932239294052,
-0.005439466331154108,
0.03504136577248573,
0.011464724317193031,
0.15623517334461212,
-0.03937678411602974,
0.07852792739868164,
-0.11596949398517609,
-0.06839881837368011,
0.010321781039237976,
-0.002345794579014182,
-0.007005574647337198,
-0.004725298378616571,
-0.04406345263123512,
-0.041766878217458725,
-0.0296886395663023,
-0.11256865411996841,
0.0709637850522995,
-0.06372899562120438,
-0.03972483053803444,
-0.029818382114171982,
0.04863636940717697,
0.04237592592835426,
-0.009328316897153854,
-0.021600522100925446,
-0.13665653765201569,
0.14698216319084167,
-0.06345529109239578,
0.032245367765426636,
0.029240552335977554,
-0.19742505252361298,
-0.04488580673933029,
0.009046786464750767,
0.0101535115391016,
-0.023378852754831314,
0.04598892480134964,
0.0332864373922348,
0.028460854664444923,
0.15546339750289917,
-0.02507939748466015,
-0.041329193860292435,
0.01282669510692358,
0.14017502963542938,
-0.04102180153131485,
-0.1508312076330185,
0.005026183091104031,
0.0012408011825755239,
-0.11991886794567108,
-0.17227014899253845,
0.07652059197425842,
0.12637996673583984,
-0.06216306611895561,
0.02343272790312767,
0.08943318575620651,
0.046482279896736145,
-0.0490446574985981,
0.036951154470443726,
0.019689388573169708,
-0.07832470536231995,
-0.00947597250342369,
0.10269232094287872,
-0.12954267859458923,
-0.035352300852537155,
0.14314644038677216,
-0.04080094397068024,
-0.045917809009552,
0.010342707857489586,
0.1257777214050293,
0.0532950833439827,
0.04120946675539017,
0.020833076909184456,
-0.041276250034570694,
-0.005169532261788845,
0.04031167924404144,
0.019807953387498856,
-0.03430025279521942,
0.048730503767728806,
0.038253054022789,
0.02607599087059498,
0.08541641384363174,
0.02156323939561844,
0.02200254425406456,
-0.0005942474817857146,
0.017398705706000328,
0.042901646345853806,
-0.011428853496909142,
-0.004019381478428841,
-0.0275720227509737,
-0.0354117713868618,
-0.037216655910015106,
-0.014780248515307903,
0.04056965187191963,
-0.060369499027729034,
0.02869461104273796,
-0.016068143770098686,
0.054223377257585526,
-0.02194124460220337,
-0.05279981344938278,
-0.07446157932281494,
-0.09265555441379547,
0.016988860443234444,
0.08453051000833511,
-0.18759359419345856,
0.009132107719779015,
0.1633789986371994,
-0.08611436188220978,
0.07530392706394196,
-0.016461651772260666,
-0.007347763981670141,
0.013184778392314911,
-0.11416641622781754,
-0.0619208849966526,
-0.02105797827243805,
-0.031067293137311935,
-0.04525798186659813,
-0.12713560461997986,
0.045260317623615265,
-0.09903648495674133,
-0.0611114539206028,
0.012042472139000893,
0.02807290107011795,
-0.14536459743976593,
0.10592331737279892,
-0.004057521466165781,
-0.0935082659125328,
-0.09628085047006607,
0.04169611260294914,
-0.007861895486712456,
0.002317520324140787,
0.03177208825945854,
-0.07803807407617569,
0.03206130489706993,
-0.12671269476413727,
-0.0019714327063411474,
0.003553648479282856,
-0.004671930801123381,
0.010406900197267532,
-0.030704300850629807,
0.056190162897109985,
0.0226251482963562,
0.028529712930321693,
0.03219180926680565,
0.014631692320108414,
0.016137700527906418,
0.09009810537099838,
-0.05811363458633423,
-0.013901115395128727,
-0.0072845821268856525,
-0.024657487869262695,
-0.02921977825462818,
0.08322833478450775,
-0.03907504305243492,
-0.04638876020908356,
-0.07644475251436234,
0.0716376081109047,
0.1526631861925125,
0.14362576603889465,
-0.04385392367839813,
0.06817208230495453,
-0.09832330793142319,
-0.12042449414730072,
0.11489484459161758,
0.013291265815496445,
0.05789683014154434,
-0.08350745588541031,
0.08087953180074692,
0.1305331587791443,
-0.16324353218078613,
0.13602790236473083,
0.023968107998371124,
-0.045915037393569946,
-0.039023470133543015,
-0.19715850055217743,
-0.06590283662080765,
0.005180463660508394,
0.05905494838953018,
-0.07366067916154861,
0.053140100091695786,
0.07669277489185333,
-0.0345139279961586,
0.045285243541002274,
0.08892343938350677,
-0.13406987488269806,
-0.12520918250083923,
0.043373871594667435,
0.010595698840916157,
0.011923803016543388,
-0.004967494402080774,
0.021823056042194366,
0.009241979569196701,
0.020758602768182755,
0.03597688302397728,
0.037678297609090805,
0.018828436732292175,
-0.028501182794570923,
-0.15545575320720673,
-0.06513874977827072,
0.04187573119997978,
0.00564018776640296,
-0.019999118521809578,
0.15700411796569824,
0.05045652016997337,
-0.00442051887512207,
-0.04119867831468582,
0.27707424759864807,
-0.051053620874881744,
-0.029769692569971085,
-0.09312871098518372,
-0.007306125480681658,
-0.01612784154713154,
0.061354562640190125,
-0.05768008902668953,
-0.08443506807088852,
0.056449949741363525,
0.08793962746858597,
0.08395817875862122,
0.0030685586389154196,
0.03560308367013931,
0.0892428606748581,
0.022410273551940918,
0.046157609671354294,
0.010056293569505215,
0.034955233335494995,
0.23349054157733917,
-0.05940147116780281,
0.060909681022167206,
0.02365214191377163,
-0.05787167325615883,
-0.0496789813041687,
0.04847359284758568,
-0.008320101536810398,
0.022514766082167625,
-0.08753542602062225,
0.12426768988370895,
0.005635613575577736,
-0.15587535500526428,
0.09457392990589142,
-0.08027240633964539,
-0.08479351550340652,
-0.022013910114765167,
0.05038904398679733,
0.03302600979804993,
0.1029486432671547,
0.062256719917058945,
-0.02588604763150215,
0.08878854662179947,
0.02880506031215191,
-0.07186056673526764,
-0.005072820000350475,
0.035472650080919266,
-0.022400854155421257,
0.20462894439697266,
-0.009344225749373436,
0.07813980430364609,
0.11672823131084442,
-0.00871850922703743,
-0.035960279405117035,
0.004817381966859102,
0.04288099706172943,
-0.09390679001808167,
0.024067746475338936,
0.22760726511478424,
-0.006086564622819424,
0.01074038352817297,
0.13562175631523132,
-0.008523890748620033,
0.10506675392389297,
-0.041197966784238815,
-0.039859313517808914,
-0.034046024084091187,
0.15855956077575684,
-0.1634187400341034,
0.059395983815193176,
0.16308803856372833,
0.03555743768811226,
-0.038859475404024124,
-0.05491410568356514,
-0.0753168836236,
0.06298670172691345,
0.1470215916633606,
-0.061635371297597885,
-0.05457483232021332,
0.018763668835163116,
0.06720782816410065,
0.027026616036891937,
-0.15947677195072174,
-0.17828144133090973,
0.0610479936003685,
0.02750750444829464,
-0.044217269867658615,
0.14215195178985596,
0.0860174149274826,
-0.023433977738022804,
-0.020485052838921547,
-0.2032535821199417,
0.030099235475063324,
0.12029887735843658,
-0.11468332260847092,
0.021019916981458664
] |
null | null | null |
# **Q-Learning** Agent playing1 **FrozenLake-v1**
This is a trained model of a **Q-Learning** agent playing **FrozenLake-v1** .
## Usage
```python
model = load_from_hub(repo_id="ergh0/q-FrozenLake-v1-4x4-noSlippery", filename="q-learning.pkl")
# Don't forget to check if you need to add additional attributes (is_slippery=False etc)
env = gym.make(model["env_id"])
```
| {"tags": ["FrozenLake-v1-4x4-no_slippery", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "q-FrozenLake-v1-4x4-noSlippery", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "FrozenLake-v1-4x4-no_slippery", "type": "FrozenLake-v1-4x4-no_slippery"}, "metrics": [{"type": "mean_reward", "value": "1.00 +/- 0.00", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | ergh0/q-FrozenLake-v1-4x4-noSlippery | [
"FrozenLake-v1-4x4-no_slippery",
"q-learning",
"reinforcement-learning",
"custom-implementation",
"model-index",
"region:us"
] | 2024-02-07T07:11:23+00:00 | [] | [] | TAGS
#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us
|
# Q-Learning Agent playing1 FrozenLake-v1
This is a trained model of a Q-Learning agent playing FrozenLake-v1 .
## Usage
| [
"# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage"
] | [
"TAGS\n#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n",
"# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage"
] | [
40,
39
] | [
"passage: TAGS\n#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage"
] | [
0.04578453302383423,
-0.08074592798948288,
-0.00430759321898222,
0.10720831900835037,
0.05034215748310089,
-0.040469273924827576,
0.11997015029191971,
0.018999949097633362,
0.20601962506771088,
-0.010012076236307621,
0.1455274522304535,
0.007022971753031015,
-0.006192410364747047,
0.1867983490228653,
0.04572829231619835,
-0.26324528455734253,
0.01831899583339691,
-0.09495259821414948,
-0.07281816750764847,
0.11870454251766205,
0.05470194295048714,
-0.01901467889547348,
-0.0007633853238075972,
0.056141503155231476,
-0.0673527717590332,
0.0007737681735306978,
0.031996939331293106,
-0.012976245954632759,
0.19804789125919342,
-0.02254498563706875,
0.06641989201307297,
0.054705578833818436,
0.0758768692612648,
-0.1998077929019928,
0.0358855277299881,
-0.04215473681688309,
-0.09439758956432343,
-0.03934839740395546,
-0.018780618906021118,
0.05878105387091637,
0.053356342017650604,
0.03858819976449013,
0.058354366570711136,
0.09384993463754654,
-0.0773480236530304,
0.04328357055783272,
0.04280758649110794,
0.024811049923300743,
0.04589218273758888,
-0.0237203948199749,
-0.027002155780792236,
0.08246652781963348,
-0.22182892262935638,
0.10318073630332947,
-0.010159241035580635,
-0.5270710587501526,
-0.00633762264624238,
0.24088262021541595,
0.11517096310853958,
0.05707438662648201,
-0.06903956830501556,
0.10566288232803345,
0.03913382440805435,
-0.007209456991404295,
0.03210983797907829,
0.02150118350982666,
0.12817370891571045,
0.06009242683649063,
-0.09581366181373596,
0.040699947625398636,
0.13722525537014008,
0.012822695076465607,
0.020306183025240898,
-0.08888901025056839,
0.0410032719373703,
-0.03461858257651329,
-0.007679527159780264,
-0.09758518636226654,
0.05478060990571976,
0.012466507963836193,
-0.0934976264834404,
-0.09247440844774246,
-0.04236573353409767,
-0.06708304584026337,
0.11252415925264359,
0.046419668942689896,
-0.0874939113855362,
0.03884070739150047,
-0.06760413944721222,
0.05918780341744423,
-0.16863860189914703,
0.02074250765144825,
-0.06627868115901947,
-0.09376336634159088,
-0.11799788475036621,
-0.01683047041296959,
-0.07946427166461945,
0.009092256426811218,
0.056664444506168365,
0.1447116881608963,
0.22076484560966492,
0.06690320372581482,
0.09728849679231644,
0.07456006109714508,
0.06531001627445221,
0.1538129299879074,
0.10918238013982773,
0.019075315445661545,
-0.015266558155417442,
0.0948706716299057,
-0.06445580720901489,
-0.1351388692855835,
-0.15579092502593994,
0.005488025024533272,
0.0983937531709671,
0.08871900290250778,
-0.044080477207899094,
-0.006702381651848555,
-0.024641724303364754,
0.08566431701183319,
-0.11314457654953003,
-0.024612564593553543,
-0.002267979085445404,
0.06882024556398392,
-0.024801667779684067,
0.020378148183226585,
-0.06242705136537552,
0.12715265154838562,
0.04222423583269119,
-0.059924717992544174,
-0.055308472365140915,
-0.03053177334368229,
-0.014276440255343914,
-0.027539284899830818,
0.02446848154067993,
-0.07659092545509338,
0.04767750948667526,
-0.16766095161437988,
-0.042871296405792236,
-0.04784649610519409,
0.025697942823171616,
-0.03907240927219391,
-0.13557587563991547,
-0.17699143290519714,
-0.048906855285167694,
-0.022438718006014824,
0.03549358621239662,
-0.038111843168735504,
0.006551501806825399,
-0.006318534724414349,
-0.1583600640296936,
0.09783563017845154,
0.09784027189016342,
-0.03643378987908363,
-0.02749447710812092,
0.056263517588377,
-0.07194498926401138,
0.1561182290315628,
-0.21054518222808838,
-0.054014235734939575,
-0.044764336198568344,
-0.06595750898122787,
0.19673264026641846,
0.012690845876932144,
-0.01202624011784792,
0.19873127341270447,
-0.29073721170425415,
-0.06078760325908661,
0.12533614039421082,
-0.07834373414516449,
-0.0936407670378685,
0.06941844522953033,
-0.04206686094403267,
0.023345354944467545,
0.046047765761613846,
0.36345911026000977,
-0.02069227211177349,
-0.16197136044502258,
-0.021782705560326576,
0.13971707224845886,
-0.1184760183095932,
0.059895481914281845,
0.04240793362259865,
0.12543781101703644,
-0.04250509291887283,
-0.018672896549105644,
-0.09023164212703705,
0.05999075248837471,
-0.05241934582591057,
-0.09016361832618713,
-0.03393383324146271,
-0.07645075023174286,
0.13294468820095062,
-0.0629684180021286,
0.05601520463824272,
-0.03255095332860947,
-0.07133250683546066,
-0.050324998795986176,
-0.016492370516061783,
0.04460815340280533,
0.05951254442334175,
-0.12794871628284454,
0.11029167473316193,
0.13025271892547607,
-0.0006193425506353378,
-0.07498852163553238,
-0.17872096598148346,
0.003240168560296297,
0.009576505981385708,
0.039837226271629333,
0.17141658067703247,
0.12209978699684143,
0.033295199275016785,
0.008770671673119068,
-0.06389404833316803,
-0.18276847898960114,
0.058129217475652695,
-0.056212130934000015,
-0.14230976998806,
-0.052409034222364426,
-0.0728459507226944,
0.017381802201271057,
-0.0859743058681488,
-0.017379917204380035,
0.021926190704107285,
0.006908397190272808,
0.02990424446761608,
-0.026645656675100327,
-0.049561817198991776,
0.021254703402519226,
0.06490101665258408,
-0.0037617047782987356,
0.12023693323135376,
0.008277264423668385,
-0.18308481574058533,
0.07930773496627808,
0.08478537946939468,
0.09196605533361435,
0.013250201940536499,
0.02685922384262085,
-0.021522263064980507,
-0.08061408251523972,
-0.054420311003923416,
0.02957955375313759,
0.11417073011398315,
0.1317172348499298,
0.2361993044614792,
0.08753683418035507,
0.04697408527135849,
-0.02164587564766407,
-0.016415923833847046,
0.002810494042932987,
-0.06318057328462601,
-0.029935607686638832,
0.10614971816539764,
0.05865858122706413,
-0.067733034491539,
-0.04576427489519119,
0.09590928256511688,
0.02732124738395214,
0.21205885708332062,
-0.03342745825648308,
0.01286078616976738,
-0.10957037657499313,
-0.06550975888967514,
-0.031982194632291794,
0.09201868623495102,
0.09498392790555954,
0.009755023755133152,
-0.022056059911847115,
-0.04259001836180687,
0.0012916827108711004,
-0.1334889680147171,
-0.10375088453292847,
0.026475343853235245,
0.013400445692241192,
-0.11206940561532974,
0.11674030870199203,
-0.11352457851171494,
0.039504457265138626,
0.06024791672825813,
-0.13837239146232605,
0.04428480193018913,
-0.029713207855820656,
-0.07886212319135666,
0.16866780817508698,
-0.11075661331415176,
-0.094340018928051,
-0.08831550180912018,
0.004082420375198126,
0.0075836325995624065,
-0.03922267258167267,
-0.009283260442316532,
-0.19952571392059326,
-0.005375816952437162,
-0.03544965013861656,
0.013616434298455715,
-0.06988783925771713,
-0.11287739872932434,
-0.010957922786474228,
0.07084179669618607,
-0.043388739228248596,
-0.07803605496883392,
0.007967432029545307,
-0.08923084288835526,
-0.10623309016227722,
0.028189711272716522,
0.019765101373195648,
-0.022883659228682518,
0.16152891516685486,
0.01816628873348236,
0.05626589432358742,
-0.03298520669341087,
0.30665266513824463,
-0.038163769990205765,
0.08371731638908386,
-0.02993497997522354,
-0.07433546334505081,
0.06130730360746384,
-0.022327827289700508,
0.06086638569831848,
-0.020221687853336334,
-0.02362890914082527,
0.0077952733263373375,
-0.08579335361719131,
-0.18365982174873352,
-0.05417544022202492,
0.03724347800016403,
0.195254847407341,
0.031118987128138542,
0.01910330168902874,
-0.0488768145442009,
-0.010547760874032974,
0.1665220558643341,
-0.10005921125411987,
0.04030545800924301,
-0.05366240441799164,
0.11506262421607971,
-0.08640182018280029,
0.06195629760622978,
0.020486772060394287,
0.04266135022044182,
-0.04877188801765442,
0.09486009180545807,
0.0826394334435463,
0.1121082529425621,
-0.02206910029053688,
0.046257395297288895,
0.019012698903679848,
0.07383184134960175,
0.11073657125234604,
0.0368414968252182,
-0.0729052945971489,
0.001982470043003559,
-0.006313489284366369,
-0.039427030831575394,
0.11933320760726929,
0.17963355779647827,
-0.11991413682699203,
-0.05106910318136215,
0.27167606353759766,
0.0031242913100868464,
0.19481229782104492,
-0.01315275114029646,
0.043591804802417755,
-0.04484925419092178,
0.04572054371237755,
-0.05338600277900696,
-0.04086209088563919,
0.2094656229019165,
0.08045925945043564,
-0.17165091633796692,
-0.08549032360315323,
-0.05912299454212189,
0.07081323862075806,
0.10728751868009567,
0.0013539529172703624,
-0.04156802222132683,
0.0004610282776411623,
0.0014198932331055403,
0.08339415490627289,
-0.14520122110843658,
0.11816094070672989,
-0.03172019124031067,
0.05612684786319733,
0.017555562779307365,
-0.045326150953769684,
0.04264266416430473,
0.07474290579557419,
0.26618310809135437,
0.0904107540845871,
-0.040318213403224945,
-0.0892091691493988,
-0.12260187417268753,
0.010461576282978058,
0.029102616012096405,
-0.03534553572535515,
0.0037547778338193893,
-0.020087555050849915,
0.0318896509706974,
0.008264793083071709,
0.016230624169111252,
-0.08987458795309067,
-0.03175399824976921,
-0.027736429125070572,
-0.023839212954044342,
0.10733365267515182,
-0.09495144337415695,
-0.1444292515516281,
-0.15713949501514435,
0.04191131144762039,
-0.0766405463218689,
-0.056593164801597595,
-0.054507751017808914,
-0.05239389091730118,
-0.0311186034232378,
-0.03773957118391991,
0.09099467098712921,
-0.0021037792321294546,
0.14807306230068207,
-0.1920108050107956,
-0.04220759496092796,
0.051812779158353806,
-0.07607918977737427,
-0.08729588985443115,
0.03410962224006653,
0.12136995792388916,
0.05116051807999611,
0.11504370719194412,
0.013609255664050579,
0.09567681699991226,
0.0045484392903745174,
-0.06713183224201202,
0.15302421152591705,
-0.14069625735282898,
-0.27875974774360657,
-0.03836318850517273,
0.016946332529187202,
0.1615200787782669,
-0.05613167956471443,
0.031766023486852646,
0.3335736393928528,
0.27782970666885376,
-0.1428707242012024,
0.25916144251823425,
0.019178593531250954,
0.004398873541504145,
-0.19130495190620422,
-0.10125631093978882,
0.025324683636426926,
0.04740457236766815,
0.12032642960548401,
-0.14564448595046997,
-0.010732659138739109,
-0.04543145373463631,
-0.025908485054969788,
0.10386138409376144,
-0.12300799041986465,
-0.07263197749853134,
0.07765276730060577,
0.039809420704841614,
0.1808302253484726,
0.03932500258088112,
0.0014799144119024277,
0.13626977801322937,
0.06612244248390198,
0.019124457612633705,
0.05216038227081299,
0.08028066903352737,
-0.018944554030895233,
0.14207926392555237,
0.05448179319500923,
-0.02551644667983055,
0.052681710571050644,
-0.0054580713622272015,
-0.03219012916088104,
0.015605825930833817,
-0.183198019862175,
-0.10147556662559509,
-0.0561356320977211,
-0.10798973590135574,
-0.04978342354297638,
0.056853994727134705,
-0.12395523488521576,
-0.007896827533841133,
-0.03841273859143257,
0.03718273714184761,
-0.07831971347332001,
-0.09360362589359283,
-0.036494381725788116,
0.1351792961359024,
0.07210618257522583,
0.04471297934651375,
0.035655103623867035,
-0.07390819489955902,
0.07097936421632767,
0.21671734750270844,
0.08159157633781433,
0.028919655829668045,
-0.19545674324035645,
-0.024042490869760513,
-0.0803457647562027,
0.06306298077106476,
-0.08856996893882751,
-0.016788700595498085,
0.11923003196716309,
0.08616556972265244,
0.05413002520799637,
0.09640096127986908,
-0.045083072036504745,
0.021686913445591927,
0.02684609219431877,
-0.15131035447120667,
-0.18501274287700653,
-0.08534606546163559,
-0.03519878163933754,
0.11561143398284912,
-0.06398691236972809,
0.10897188633680344,
-0.13615410029888153,
0.010051886551082134,
-0.006060056854039431,
0.02693452313542366,
-0.03596206381917,
-0.11251141875982285,
0.15348562598228455,
0.11999429017305374,
-0.06767056882381439,
0.03127254918217659,
-0.09527092427015305,
-0.04423454403877258,
0.12686803936958313,
-0.013623855076730251,
-0.0371493324637413,
-0.054547641426324844,
-0.03628576174378395,
0.15247689187526703,
-0.03436964750289917,
0.008244883269071579,
-0.041229065507650375,
-0.18217355012893677,
0.0798322781920433,
0.09045056998729706,
0.019827889278531075,
-0.031874191015958786,
-0.09797266125679016,
-0.010231015272438526,
-0.0011165260802954435,
0.11730700731277466,
-0.10696814209222794,
-0.10933240503072739,
-0.15144047141075134,
0.06713984161615372,
-0.0007159380475059152,
0.18502596020698547,
-0.06394898891448975,
-0.08904669433832169,
-0.12429379671812057,
0.02344517596065998,
-0.0027384376153349876,
-0.042264558374881744,
0.01618490368127823,
0.07992301136255264,
-0.04095321521162987,
0.02075677551329136,
-0.06651144474744797,
0.06372585147619247,
-0.11786920577287674,
0.09625071287155151,
0.01063506118953228,
0.016993753612041473,
-0.0417880080640316,
-0.01618220843374729,
0.039470795542001724,
-0.057925306260585785,
0.07921463251113892,
0.011758086271584034,
0.0010938759660348296,
0.10196787863969803,
-0.0034960443153977394,
0.06409632414579391,
-0.05372481048107147,
-0.023290161043405533,
0.06578411161899567,
-0.05874887853860855,
-0.03370826691389084,
-0.1573946475982666,
-0.0709633082151413,
0.020051732659339905,
-0.04775108024477959,
0.002077929675579071,
0.03673801198601723,
0.062159497290849686,
-0.06937079131603241,
-0.12125655263662338,
-0.043812792748212814,
-0.028638383373618126,
0.021301284432411194,
0.10829301923513412,
-0.07526551932096481,
0.1547859013080597,
-0.052787959575653076,
-0.00020603960729204118,
0.07437096536159515,
0.04048224538564682,
0.01393822580575943,
-0.10422444343566895,
-0.04698587954044342,
-0.11035211384296417,
0.1502903699874878,
-0.007902312092483044,
-0.03533121198415756,
0.03719403222203255,
-0.11946307867765427,
-0.1572723090648651,
0.03418220207095146,
0.10199101269245148,
0.0448341928422451,
0.025807438418269157,
0.027079269289970398,
-0.04042419046163559,
-0.021270349621772766,
-0.07034418731927872,
0.0882953479886055,
-0.12085357308387756,
-0.09669415652751923,
0.09555385261774063,
0.12178351730108261,
-0.0036850625183433294,
-0.07441367954015732,
0.11554073542356491,
-0.021787192672491074,
0.05525410920381546,
-0.02971339225769043,
0.10308072715997696,
0.0796005055308342,
-0.12273547053337097,
0.005693064536899328,
-0.036891788244247437,
-0.0741485133767128,
-0.12975730001926422,
0.019545545801520348,
-0.061916105449199677,
-0.13383042812347412,
0.12179028987884521,
-0.09376577287912369,
0.030037038028240204,
-0.10506992787122726,
0.021338803693652153,
0.01864001713693142,
0.061665527522563934,
-0.10988292098045349,
0.08575301617383957,
0.13424484431743622,
-0.043199893087148666,
-0.07184189558029175,
-0.12455986440181732,
-0.05022053420543671,
-0.04231856390833855,
-0.13957437872886658,
-0.11600435525178909,
0.0100301094353199,
-0.023418782278895378,
-0.05818291753530502,
0.0015462689334526658,
-0.03659068048000336,
0.008594646118581295,
0.021907730028033257,
0.04032021388411522,
-0.02693161368370056,
0.05134565755724907,
-0.057569269090890884,
-0.052510857582092285,
0.11489357799291611,
0.04113486409187317,
-0.03561042994260788,
-0.052359987050294876,
0.12997733056545258,
-0.11959461867809296,
0.07662346214056015,
-0.020313527435064316,
0.017129231244325638,
-0.06435854732990265,
0.17131924629211426,
0.11673715710639954,
-0.1367570012807846,
-0.005008010193705559,
-0.08210669457912445,
0.020409544929862022,
0.023555370047688484,
0.13693512976169586,
-0.03411718085408211,
-0.0012358218664303422,
-0.1580323874950409,
0.018575575202703476,
-0.18557456135749817,
-0.03716109320521355,
0.04671547934412956,
0.09917585551738739,
0.15293832123279572,
-0.0034432117827236652,
-0.1263325810432434,
0.10424192249774933,
-0.2118520885705948,
0.0907607227563858,
0.05121984705328941,
-0.11874113976955414,
-0.06765396893024445,
-0.06795281916856766,
0.1198519766330719,
0.009196433238685131,
0.2040700763463974,
-0.013615905307233334,
-0.09132910519838333,
-0.07060808688402176,
-0.01980910450220108,
-0.030524181202054024,
0.09714830666780472,
0.041414931416511536,
0.04653804749250412,
0.12821412086486816,
0.00368314771912992,
0.07533777505159378,
0.060310911387205124,
0.02759413793683052,
-0.012300663627684116,
0.04076618701219559,
0.08261215686798096,
-0.14588621258735657,
-0.1659701019525528,
0.1326720416545868,
0.025149408727884293,
0.11792458593845367,
0.03658788278698921,
-0.1549617499113083,
0.06687124073505402,
0.2523096203804016,
-0.11147607117891312,
0.02505038119852543,
0.12737524509429932,
-0.0366884209215641,
0.0672016367316246,
0.1144871786236763,
-0.02633814327418804,
-0.05217865854501724,
-0.011363590136170387,
0.10233135521411896,
0.028660254552960396,
-0.04646271467208862,
-0.02340836264193058,
-0.03373933956027031,
-0.019070526584982872,
-0.011738128960132599,
-0.0909019410610199,
-0.1543993502855301,
-0.10471053421497345,
-0.16619662940502167,
0.04399140924215317,
-0.04626438021659851,
0.13418889045715332,
0.09469578415155411,
-0.012723101302981377,
0.04568437114357948,
0.028575526550412178,
0.07275456190109253,
0.07916246354579926,
-0.02939477376639843,
-0.036159269511699677
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# finetune_deepseek_6.7b_exp_1_0_yaml
This model is a fine-tuned version of [deepseek-ai/deepseek-coder-6.7b-instruct](https://huggingface.co/deepseek-ai/deepseek-coder-6.7b-instruct) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.4992
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 1
- eval_batch_size: 8
- seed: 42
- distributed_type: multi-GPU
- num_devices: 6
- total_train_batch_size: 6
- total_eval_batch_size: 48
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_ratio: 0.03
- num_epochs: 5
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| No log | 1.0 | 504 | 0.9573 |
| No log | 2.0 | 1008 | 0.6384 |
| No log | 3.0 | 1512 | 0.4705 |
| No log | 4.0 | 2016 | 0.4661 |
| No log | 5.0 | 2520 | 0.4992 |
### Framework versions
- Transformers 4.36.2
- Pytorch 2.1.2
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "other", "tags": ["generated_from_trainer"], "base_model": "deepseek-ai/deepseek-coder-6.7b-instruct", "model-index": [{"name": "finetune_deepseek_6.7b_exp_1_0_yaml", "results": []}]} | text-generation | onur-softtech/finetune_deepseek_6.7b_exp_1_0_yaml | [
"transformers",
"safetensors",
"llama",
"text-generation",
"generated_from_trainer",
"base_model:deepseek-ai/deepseek-coder-6.7b-instruct",
"license:other",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T07:11:44+00:00 | [] | [] | TAGS
#transformers #safetensors #llama #text-generation #generated_from_trainer #base_model-deepseek-ai/deepseek-coder-6.7b-instruct #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| finetune\_deepseek\_6.7b\_exp\_1\_0\_yaml
=========================================
This model is a fine-tuned version of deepseek-ai/deepseek-coder-6.7b-instruct on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 0.4992
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 1
* eval\_batch\_size: 8
* seed: 42
* distributed\_type: multi-GPU
* num\_devices: 6
* total\_train\_batch\_size: 6
* total\_eval\_batch\_size: 48
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: cosine
* lr\_scheduler\_warmup\_ratio: 0.03
* num\_epochs: 5
### Training results
### Framework versions
* Transformers 4.36.2
* Pytorch 2.1.2
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 8\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 6\n* total\\_train\\_batch\\_size: 6\n* total\\_eval\\_batch\\_size: 48\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_ratio: 0.03\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.36.2\n* Pytorch 2.1.2\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #generated_from_trainer #base_model-deepseek-ai/deepseek-coder-6.7b-instruct #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 8\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 6\n* total\\_train\\_batch\\_size: 6\n* total\\_eval\\_batch\\_size: 48\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_ratio: 0.03\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.36.2\n* Pytorch 2.1.2\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
82,
167,
4,
30
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #generated_from_trainer #base_model-deepseek-ai/deepseek-coder-6.7b-instruct #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 8\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 6\n* total\\_train\\_batch\\_size: 6\n* total\\_eval\\_batch\\_size: 48\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_ratio: 0.03\n* num\\_epochs: 5### Training results### Framework versions\n\n\n* Transformers 4.36.2\n* Pytorch 2.1.2\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.0931047573685646,
0.1074134111404419,
-0.003914784174412489,
0.08725771307945251,
0.09090004861354828,
0.047106269747018814,
0.15456458926200867,
0.13639505207538605,
-0.06867090612649918,
0.125626802444458,
0.10067542642354965,
0.058842115104198456,
0.06963453441858292,
0.17908771336078644,
-0.01698909141123295,
-0.23012973368167877,
0.03079228661954403,
-0.04587884992361069,
-0.10103815793991089,
0.10306337475776672,
0.0761142149567604,
-0.11914724856615067,
0.09935928136110306,
-0.04181855916976929,
-0.11377314478158951,
-0.03203625977039337,
-0.04103025421500206,
-0.019744141027331352,
0.09401020407676697,
0.03952089697122574,
0.07923969626426697,
0.03548487275838852,
0.09643247723579407,
-0.23779676854610443,
0.005770414602011442,
0.08343473821878433,
0.004769293591380119,
0.06945238262414932,
0.09583909809589386,
0.014208556152880192,
0.09475744515657425,
-0.10660359263420105,
0.04964669048786163,
0.03955188766121864,
-0.11559073626995087,
-0.17962870001792908,
-0.06301982700824738,
0.05094052478671074,
0.10433195531368256,
0.05372817814350128,
-0.01087403204292059,
0.09190727025270462,
-0.0507509782910347,
0.0849788710474968,
0.22070829570293427,
-0.2782183885574341,
-0.05989638343453407,
0.05153258889913559,
0.028620904311537743,
0.10051940381526947,
-0.1021265909075737,
-0.010080107487738132,
0.03072531893849373,
0.020042255520820618,
0.09066006541252136,
-0.0023282915353775024,
-0.012847624719142914,
0.003903773380443454,
-0.13293786346912384,
-0.07479355484247208,
0.15029489994049072,
0.059032902121543884,
-0.016319289803504944,
-0.10287919640541077,
-0.06058233231306076,
-0.1713537573814392,
-0.03585922345519066,
0.0129671236500144,
0.033818017691373825,
-0.040568429976701736,
-0.04815852642059326,
0.02852359227836132,
-0.08151934295892715,
-0.08928210288286209,
0.005830688867717981,
0.08487354964017868,
0.05572989583015442,
-0.0024109415244311094,
0.01921728067100048,
0.11561581492424011,
0.006964653264731169,
-0.15857458114624023,
-0.020173804834485054,
0.002592305187135935,
-0.06955952197313309,
-0.01534700021147728,
0.00021547926007770002,
0.04800230264663696,
0.0769224613904953,
0.14580442011356354,
-0.06273201107978821,
0.06829491257667542,
0.019745124503970146,
0.018805133178830147,
-0.04356130585074425,
0.11752909421920776,
-0.06782945990562439,
-0.0515214279294014,
-0.014215675182640553,
0.10319020599126816,
0.042808614671230316,
-0.018138820305466652,
-0.08898955583572388,
0.02977289818227291,
0.09866059571504593,
0.06521918624639511,
0.0004967254353687167,
0.043263617902994156,
-0.06736162304878235,
-0.03187581151723862,
0.10034164786338806,
-0.10675571858882904,
0.04335997626185417,
0.05271117761731148,
-0.049775976687669754,
-0.0538654662668705,
-0.0011823097011074424,
-0.004656726494431496,
-0.02525307796895504,
0.05454518273472786,
-0.07881717383861542,
-0.020598692819476128,
-0.07464971393346786,
-0.11230848729610443,
0.038327407091856,
-0.06579210609197617,
-0.010354575701057911,
-0.09016067534685135,
-0.1342018097639084,
-0.03478129953145981,
0.029378684237599373,
-0.06724832952022552,
-0.06268108636140823,
-0.059068016707897186,
-0.09376911073923111,
0.02399534545838833,
-0.006163482554256916,
0.09170737862586975,
-0.07107075303792953,
0.07413328438997269,
0.009243066422641277,
0.04655129462480545,
0.06518851965665817,
0.036782898008823395,
-0.06593497097492218,
0.07751315087080002,
-0.14752709865570068,
0.05686169117689133,
-0.08672083914279938,
0.05285046994686127,
-0.11415380984544754,
-0.10559983551502228,
0.014586269855499268,
-0.019519886001944542,
0.06357887387275696,
0.1282486468553543,
-0.14963789284229279,
-0.048909954726696014,
0.17239369451999664,
-0.09624040126800537,
-0.13606537878513336,
0.12758339941501617,
-0.005835600197315216,
-0.06403891742229462,
0.016952766105532646,
0.1586523950099945,
0.14174798130989075,
-0.10580310225486755,
-0.022319819778203964,
0.00925800297409296,
0.09750731289386749,
0.0068294573575258255,
0.10110552608966827,
0.005978412926197052,
0.054274868220090866,
0.013314702548086643,
-0.0337260365486145,
0.03168397396802902,
-0.08643362671136856,
-0.08707581460475922,
-0.03272494673728943,
-0.08607678860425949,
-0.00648210383951664,
0.03620482236146927,
0.02413284406065941,
-0.09253896772861481,
-0.09684144705533981,
-0.0357033796608448,
0.11065442115068436,
-0.0899338573217392,
0.006001511588692665,
-0.05934828892350197,
0.08212821185588837,
-0.015902314335107803,
0.006039966829121113,
-0.13943356275558472,
-0.1186969131231308,
0.07030720263719559,
-0.04374578222632408,
0.00994036253541708,
0.0010845422511920333,
0.06109630689024925,
0.10635686665773392,
-0.03739415854215622,
-0.059813354164361954,
-0.014180424623191357,
-0.009097376838326454,
-0.07411559671163559,
-0.2383873164653778,
-0.0671813115477562,
-0.030878502875566483,
0.14414219558238983,
-0.20348072052001953,
0.032428842037916183,
0.02197699435055256,
0.1168026328086853,
0.013723582029342651,
-0.03804274648427963,
0.006251683924347162,
0.05217303708195686,
-0.050216592848300934,
-0.08287180960178375,
0.028839552775025368,
-0.006872828118503094,
-0.09525151550769806,
-0.00554416049271822,
-0.1918923556804657,
0.13079552352428436,
0.08206502348184586,
0.0013225581496953964,
-0.08523046225309372,
-0.03083481639623642,
-0.05385551229119301,
-0.05674632638692856,
-0.02104797773063183,
-0.0036627878434956074,
0.10467492789030075,
-0.002834522631019354,
0.10549155622720718,
-0.08342768251895905,
-0.052768126130104065,
0.030878549441695213,
0.002598899183794856,
-0.00986131839454174,
0.14393110573291779,
0.05539459362626076,
-0.09809938073158264,
0.1390392780303955,
0.12157066911458969,
-0.03785353899002075,
0.11798866838216782,
-0.08133949339389801,
-0.07052527368068695,
-0.04480312019586563,
0.05712975561618805,
0.03079826943576336,
0.09707256406545639,
-0.05414685979485512,
0.014295991510152817,
0.029474301263689995,
0.011173367500305176,
0.004380783997476101,
-0.16528797149658203,
0.00019907271780539304,
0.026415152475237846,
-0.08963744342327118,
0.021221982315182686,
-0.03552693873643875,
0.002881848020479083,
0.09372767060995102,
-0.004202614072710276,
-0.03670395165681839,
-0.009943900629878044,
-0.02005835436284542,
-0.08309637010097504,
0.21909360587596893,
-0.10342186689376831,
-0.12253931909799576,
-0.1380317062139511,
0.04337967559695244,
-0.05259205773472786,
0.0038915136829018593,
0.028594374656677246,
-0.0636918768286705,
-0.05172446370124817,
-0.12259470671415329,
-0.007888829335570335,
-0.006302881054580212,
0.025427039712667465,
-0.00107663928065449,
0.017680246382951736,
0.055101748555898666,
-0.10916679352521896,
0.0010396512225270271,
0.016750434413552284,
-0.07087848335504532,
0.0330507792532444,
0.0342601053416729,
0.0936182513833046,
0.14076825976371765,
0.037004243582487106,
0.008847542107105255,
-0.022199193015694618,
0.1781870722770691,
-0.07211199402809143,
0.008710297755897045,
0.10023847967386246,
0.011519106104969978,
0.05746685341000557,
0.15988703072071075,
0.03895982727408409,
-0.07103621959686279,
0.00208285148255527,
0.0269283726811409,
-0.027506854385137558,
-0.20455583930015564,
-0.050434041768312454,
-0.04565907642245293,
0.0659002959728241,
0.10735172033309937,
0.04406965151429176,
-0.012997335754334927,
0.044576309621334076,
-0.045381877571344376,
0.023050924763083458,
0.02333342842757702,
0.07260917872190475,
0.05272989720106125,
0.04870279133319855,
0.10886777937412262,
-0.041351787745952606,
-0.029494037851691246,
0.04830815643072128,
0.01322379894554615,
0.1979856640100479,
-0.03911425918340683,
0.2237827330827713,
0.032438408583402634,
0.15324750542640686,
0.0034693924244493246,
0.06883496046066284,
0.013664663769304752,
0.007827146910130978,
0.009514864534139633,
-0.06402068585157394,
-0.020270589739084244,
0.04551704600453377,
0.01830149069428444,
0.009278752841055393,
-0.08353951573371887,
0.044453490525484085,
0.05381155014038086,
0.2485102117061615,
0.06635696440935135,
-0.3134359121322632,
-0.08444032073020935,
0.04577571153640747,
-0.022615946829319,
-0.03109521046280861,
0.01633376069366932,
0.1701558530330658,
-0.08227551728487015,
0.06742773950099945,
-0.04633176699280739,
0.07688084989786148,
-0.06561188399791718,
0.016363518312573433,
0.0754595547914505,
0.0995970070362091,
0.008436852134764194,
0.08492666482925415,
-0.22198033332824707,
0.24894194304943085,
0.007374035660177469,
0.028781630098819733,
-0.06633224338293076,
0.03570183366537094,
-0.005569190718233585,
0.04254171624779701,
0.08598010987043381,
-0.010202322155237198,
-0.10532073676586151,
-0.19121026992797852,
-0.12715084850788116,
0.018553899601101875,
0.12817688286304474,
-0.08368884772062302,
0.11670750379562378,
-0.013506205752491951,
-0.029257560148835182,
0.034158095717430115,
-0.06131872162222862,
-0.07764620333909988,
-0.10731592029333115,
0.03680526837706566,
-0.01615661196410656,
0.006990110967308283,
-0.07680228352546692,
-0.08123937994241714,
-0.10185296088457108,
0.18116912245750427,
-0.13949118554592133,
-0.041938625276088715,
-0.11637850105762482,
0.0673600286245346,
0.13942909240722656,
-0.09123508632183075,
0.03214017301797867,
-0.017187174409627914,
0.10006800293922424,
0.025913361459970474,
-0.058659784495830536,
0.10118600726127625,
-0.08088207989931107,
-0.23070231080055237,
-0.0376463457942009,
0.13210251927375793,
0.020798707380890846,
0.06097477674484253,
-0.027834033593535423,
0.02074779011309147,
-0.01023991871625185,
-0.10752660781145096,
0.03030313178896904,
0.06492123007774353,
0.06660784780979156,
0.053582336753606796,
-0.05628124624490738,
0.012177064083516598,
-0.02764430083334446,
-0.027807647362351418,
0.10685079544782639,
0.3020934760570526,
-0.09634335339069366,
0.04385776445269585,
0.0508454404771328,
-0.06631869822740555,
-0.18970291316509247,
-0.04986012354493141,
0.06513087451457977,
0.035984911024570465,
0.012809445150196552,
-0.18243804574012756,
0.06293068826198578,
0.08651766180992126,
-0.023918060585856438,
0.08381126075983047,
-0.31079983711242676,
-0.14165765047073364,
0.08382569998502731,
0.08439073711633682,
-0.027677495032548904,
-0.18821284174919128,
-0.05992591008543968,
-0.010854811407625675,
-0.08393873274326324,
0.09695115685462952,
-0.06223750859498978,
0.12101364135742188,
-0.021618805825710297,
0.004805452656000853,
0.025420498102903366,
-0.05974458530545235,
0.1588996946811676,
0.009352440945804119,
0.0853850319981575,
-0.05904698371887207,
0.01769449934363365,
0.07349289953708649,
-0.07647059112787247,
0.044328149408102036,
-0.13392551243305206,
0.046857159584760666,
-0.09340114891529083,
-0.007867995649576187,
-0.0533476285636425,
0.017775828018784523,
-0.04657881706953049,
-0.03099740669131279,
-0.04954474791884422,
0.04919467121362686,
0.06249236315488815,
-0.011852654628455639,
0.14142848551273346,
0.01716964691877365,
0.13363967835903168,
0.1684257686138153,
0.10322429984807968,
0.030299494042992592,
-0.04381635785102844,
-0.016133544966578484,
-0.007575702853500843,
0.03121185675263405,
-0.11823354661464691,
0.027766888961195946,
0.1406746655702591,
0.017104852944612503,
0.11090177297592163,
0.05016801506280899,
-0.06380216032266617,
-0.0037259235978126526,
0.06953778862953186,
-0.14693160355091095,
-0.15492361783981323,
-0.0016221471596509218,
0.00788016989827156,
-0.15659108757972717,
0.025543708354234695,
0.11378581076860428,
-0.03927676007151604,
-0.0014500628458335996,
-0.0017890564631670713,
0.06831425428390503,
-0.017237471416592598,
0.2033742517232895,
0.03513490781188011,
0.09332062304019928,
-0.09561381489038467,
0.08743096143007278,
0.06329238414764404,
-0.09856602549552917,
0.03739457204937935,
0.11157942563295364,
-0.08440349996089935,
-0.044152382761240005,
0.10451172292232513,
0.11521555483341217,
0.0026504427660256624,
-0.041579291224479675,
-0.11879251152276993,
-0.14083334803581238,
0.07393737882375717,
0.11322183161973953,
0.053671106696128845,
0.06245417147874832,
0.00899440050125122,
0.010553739964962006,
-0.08681242913007736,
0.13384245336055756,
0.05530501902103424,
0.08018554002046585,
-0.1470702588558197,
0.11363838613033295,
-0.012764680199325085,
0.02780264802277088,
-0.012697484344244003,
0.03588446229696274,
-0.13397076725959778,
-0.023372124880552292,
-0.1091732457280159,
0.017979219555854797,
-0.05791788548231125,
0.003605748526751995,
0.00800519809126854,
-0.03078327514231205,
-0.0360611267387867,
0.016432475298643112,
-0.08198441565036774,
-0.05444024130702019,
-0.04620231315493584,
0.08208369463682175,
-0.13392460346221924,
-0.03459708392620087,
0.021136917173862457,
-0.10457892715930939,
0.10054061561822891,
0.024502165615558624,
0.03679865226149559,
0.0006592620629817247,
-0.1009388267993927,
0.030374370515346527,
0.027931472286581993,
0.029238596558570862,
0.02055204100906849,
-0.11114658415317535,
-0.007907678373157978,
-0.025728585198521614,
-0.023079123347997665,
0.008939365856349468,
0.05821367725729942,
-0.11004282534122467,
0.033307790756225586,
-0.02268587239086628,
-0.05890725925564766,
-0.06596319377422333,
0.045430902391672134,
0.06704969704151154,
-0.024691317230463028,
0.1368890106678009,
-0.07683336734771729,
0.04990355297923088,
-0.2262069135904312,
-0.015792658552527428,
0.01499534398317337,
-0.08265245705842972,
-0.08447533845901489,
-0.04442683979868889,
0.09853781014680862,
-0.048341505229473114,
0.12259912490844727,
-0.03295919671654701,
0.023332059383392334,
0.013606531545519829,
-0.022259216755628586,
0.0711420550942421,
0.07295537739992142,
0.14579863846302032,
0.02743721939623356,
-0.03797371685504913,
0.045739125460386276,
-0.011669446714222431,
0.06937427073717117,
0.057204924523830414,
0.18264539539813995,
0.12645454704761505,
-0.009285042062401772,
0.06807845830917358,
0.08702945709228516,
-0.14959006011486053,
-0.09594764560461044,
0.08944567292928696,
-0.0814179852604866,
0.12507972121238708,
-0.039648402482271194,
0.1413518488407135,
0.08484365791082382,
-0.19620537757873535,
0.026580892503261566,
-0.04226225987076759,
-0.09238345175981522,
-0.09970398992300034,
-0.10041896998882294,
-0.09647072106599808,
-0.1547461450099945,
0.0008852796745486557,
-0.1257503181695938,
0.03326638415455818,
0.08468785136938095,
0.03770388662815094,
0.02326827123761177,
0.14552059769630432,
0.05201057717204094,
0.03446270525455475,
0.03364323824644089,
0.03417372703552246,
-0.0031896671280264854,
-0.005925710778683424,
-0.09410591423511505,
0.029332812875509262,
-0.025900129228830338,
0.04463278129696846,
-0.023630399256944656,
-0.0016097022453323007,
0.08019611984491348,
-0.014823942445218563,
-0.09756501764059067,
0.014741556718945503,
-0.026665426790714264,
0.013087236322462559,
0.06691569089889526,
0.019129376858472824,
-0.011418307200074196,
-0.011480281129479408,
0.16292394697666168,
-0.07051458209753036,
-0.07935670763254166,
-0.103683702647686,
0.22344401478767395,
-0.021901894360780716,
-0.00906664039939642,
0.03788624703884125,
-0.06257382035255432,
-0.019932985305786133,
0.139170840382576,
0.21965454518795013,
-0.05111798644065857,
-0.005175717640668154,
0.01050588209182024,
-0.010019290260970592,
-0.004529779776930809,
0.09324563294649124,
0.0950862318277359,
0.06637105345726013,
-0.07746759057044983,
-0.02542312629520893,
-0.006646804511547089,
-0.025668339803814888,
-0.07513213902711868,
0.06038985028862953,
0.019902268424630165,
0.007448017597198486,
-0.02005809359252453,
0.05942760407924652,
-0.06098296865820885,
-0.04157059267163277,
0.06547435373067856,
-0.1979244500398636,
-0.15448611974716187,
-0.03148264065384865,
0.07990293204784393,
0.005450621247291565,
0.03902727738022804,
-0.016837377101182938,
-0.01837761141359806,
0.0870051458477974,
-0.0193256214261055,
-0.08972224593162537,
-0.07055530697107315,
0.04181482270359993,
-0.10689593106508255,
0.18392609059810638,
-0.029256416484713554,
0.049789004027843475,
0.12805195152759552,
-0.0020187939517199993,
-0.11271951347589493,
0.04287580028176308,
0.08584198355674744,
-0.0901324599981308,
0.03112233802676201,
0.13129690289497375,
-0.04022400826215744,
0.11074168980121613,
0.04958469420671463,
-0.07649854570627213,
-0.014532438479363918,
-0.03572912514209747,
-0.025725163519382477,
-0.054251715540885925,
-0.034794632345438004,
-0.046641748398542404,
0.16265694797039032,
0.18385960161685944,
-0.056051455438137054,
-0.02292151004076004,
-0.02762744203209877,
0.02566995844244957,
0.03987693786621094,
0.09982344508171082,
-0.007372954860329628,
-0.2588309943675995,
0.029188551008701324,
0.02122361958026886,
0.05741815268993378,
-0.211142435669899,
-0.08591163158416748,
0.02240500971674919,
-0.019026411697268486,
-0.10458067804574966,
0.11404141783714294,
0.08484620600938797,
0.03235548362135887,
-0.06201440095901489,
-0.11752015352249146,
-0.04199101775884628,
0.14804649353027344,
-0.15031151473522186,
-0.09489164501428604
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | areegtarek/patientcommunication-8bit | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"8-bit",
"region:us"
] | 2024-02-07T07:13:36+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #mistral #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #8-bit #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #8-bit #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
64,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #8-bit #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.04128352180123329,
0.19134420156478882,
-0.005646432749927044,
0.016455430537462234,
0.09635572880506516,
0.002479254500940442,
0.056982602924108505,
0.1146617978811264,
-0.0533650778234005,
0.12758329510688782,
0.04321496561169624,
0.10994824767112732,
0.11844903975725174,
0.13765117526054382,
-0.006181856617331505,
-0.21811756491661072,
0.050964366644620895,
-0.10300406813621521,
-0.010971106588840485,
0.12519820034503937,
0.1494244933128357,
-0.09655825048685074,
0.06891559064388275,
-0.037302810698747635,
-0.017924515530467033,
-0.04070461541414261,
-0.05993669107556343,
-0.0399334579706192,
0.04230647534132004,
0.05751945078372955,
0.0668831467628479,
-0.0004572426259983331,
0.0776250883936882,
-0.28147950768470764,
0.0191178098320961,
0.06887131184339523,
-0.0071472292765975,
0.06731373816728592,
0.06974674761295319,
-0.06474964320659637,
0.10905807465314865,
-0.052146319299936295,
0.1333576738834381,
0.08365475386381149,
-0.09233596175909042,
-0.18141965568065643,
-0.08972559869289398,
0.10367386043071747,
0.17589114606380463,
0.04724854975938797,
-0.028098370879888535,
0.10573755949735641,
-0.08052481710910797,
0.0208880677819252,
0.0486016720533371,
-0.09139415621757507,
-0.0580286830663681,
0.06942588835954666,
0.09234505146741867,
0.053686097264289856,
-0.12838797271251678,
-0.03492850437760353,
0.005829438101500273,
0.016645297408103943,
0.07569129765033722,
0.02085166983306408,
0.1499548852443695,
0.03481323644518852,
-0.13231118023395538,
-0.05879415571689606,
0.10943908244371414,
0.038959234952926636,
-0.04054383188486099,
-0.23723065853118896,
-0.032185882329940796,
-0.01542743295431137,
-0.03409914672374725,
-0.043776270002126694,
0.04159987345337868,
-0.0004951093578711152,
0.08771000802516937,
-0.006535344757139683,
-0.07410707324743271,
-0.03462175652384758,
0.0692015141248703,
0.06824567168951035,
0.02959168702363968,
-0.01847192458808422,
0.02023627609014511,
0.10731427371501923,
0.08469881862401962,
-0.11588454991579056,
-0.05925142765045166,
-0.06289152801036835,
-0.0703912228345871,
-0.03641165420413017,
0.03182779252529144,
0.010229445062577724,
0.07366661727428436,
0.2659907042980194,
0.02467755414545536,
0.05592493712902069,
0.029665807262063026,
0.0086274528875947,
0.049165740609169006,
0.10554521530866623,
-0.06093612685799599,
-0.11679372191429138,
-0.01784653216600418,
0.08623374253511429,
0.023653356358408928,
-0.037843264639377594,
-0.042672764509916306,
0.06579094380140305,
0.04653922840952873,
0.10845611244440079,
0.10207968950271606,
0.021128473803400993,
-0.07308052480220795,
-0.05801280587911606,
0.20696866512298584,
-0.1539951115846634,
0.034660134464502335,
0.042221877723932266,
-0.03216218948364258,
-0.027776163071393967,
0.00844907108694315,
0.02751114033162594,
-0.03499545529484749,
0.09118395298719406,
-0.05423082783818245,
-0.046447478234767914,
-0.10708986967802048,
-0.029832858592271805,
0.04285690188407898,
0.009996979497373104,
-0.03377476707100868,
-0.03598395362496376,
-0.07063847035169601,
-0.08379368484020233,
0.08550498634576797,
-0.0685519203543663,
-0.05772706866264343,
-0.02443009614944458,
-0.08077222108840942,
0.023561228066682816,
0.021879766136407852,
0.0703870877623558,
-0.026811322197318077,
0.056948207318782806,
-0.05177130177617073,
0.049297675490379333,
0.09172631800174713,
0.03290620073676109,
-0.062396612018346786,
0.06031118333339691,
-0.2256828397512436,
0.08316178619861603,
-0.07238388806581497,
0.05482398346066475,
-0.15913821756839752,
-0.022542906925082207,
0.03714415431022644,
0.0024592396803200245,
-0.005427270196378231,
0.13136856257915497,
-0.20818060636520386,
-0.020970456302165985,
0.16591490805149078,
-0.09739968180656433,
-0.06593427807092667,
0.056178901344537735,
-0.04690013825893402,
0.09401801228523254,
0.032558146864175797,
0.007820235565304756,
0.060518234968185425,
-0.10578246414661407,
-0.011071745306253433,
-0.05297582969069481,
-0.02497659996151924,
0.13878992199897766,
0.08260059356689453,
-0.08455348759889603,
0.05800924822688103,
0.023025132715702057,
-0.031320903450250626,
-0.06896700710058212,
-0.0159530621021986,
-0.10059244185686111,
0.014233643189072609,
-0.06613922119140625,
0.014380539767444134,
-0.01817280612885952,
-0.09550462663173676,
-0.027999747544527054,
-0.16695383191108704,
-0.03309853747487068,
0.07884068787097931,
-0.0031456074211746454,
-0.011709720827639103,
-0.10673288255929947,
0.02835468389093876,
0.02812434732913971,
0.0028318683616816998,
-0.1288844496011734,
-0.037429507821798325,
0.035795170813798904,
-0.1510521024465561,
0.033771321177482605,
-0.0742814764380455,
0.05265388637781143,
0.01458632294088602,
-0.030257852748036385,
-0.020554034039378166,
0.019695932045578957,
0.008414141833782196,
-0.021155405789613724,
-0.22847439348697662,
-0.0261588953435421,
-0.029949640855193138,
0.15613269805908203,
-0.2023053616285324,
0.0351564846932888,
0.08473288267850876,
0.15631094574928284,
0.003567751031368971,
-0.054057251662015915,
0.020060677081346512,
-0.06852725148200989,
-0.025515809655189514,
-0.053889475762844086,
0.002841021167114377,
-0.01714089699089527,
-0.04609391838312149,
0.028508681803941727,
-0.17861993610858917,
-0.04144664853811264,
0.09436279535293579,
0.052259016782045364,
-0.11904093623161316,
-0.02377285808324814,
-0.035128138959407806,
-0.053802043199539185,
-0.0413530096411705,
-0.06480279564857483,
0.10138877481222153,
0.06199442595243454,
0.03790408745408058,
-0.0645383894443512,
-0.07929814606904984,
-0.004190515726804733,
-0.018222428858280182,
-0.022440388798713684,
0.09439347684383392,
0.07359923422336578,
-0.1254744529724121,
0.09509528428316116,
0.0852997675538063,
0.06737883388996124,
0.08577700704336166,
-0.02042553760111332,
-0.07214456796646118,
-0.03533707559108734,
0.04134025052189827,
0.01990504376590252,
0.12563160061836243,
-0.05059973895549774,
0.04316820576786995,
0.043183207511901855,
-0.029612349346280098,
0.01701227016746998,
-0.07892264425754547,
0.033904582262039185,
0.022262200713157654,
-0.019596343860030174,
0.05044734850525856,
-0.035282403230667114,
0.01822829246520996,
0.08677670359611511,
0.05624454468488693,
0.03339933231472969,
0.017261069267988205,
-0.05203161761164665,
-0.11167068034410477,
0.16080230474472046,
-0.1183275505900383,
-0.21455936133861542,
-0.13397355377674103,
0.023255307227373123,
0.024401554837822914,
-0.013819785788655281,
0.005877576302736998,
-0.05486391857266426,
-0.1059642881155014,
-0.09301234036684036,
0.005422795191407204,
0.056448157876729965,
-0.08602327853441238,
-0.06289160996675491,
0.03975475952029228,
0.0435301810503006,
-0.14439761638641357,
0.021902093663811684,
0.04121612757444382,
-0.09205880016088486,
-0.010940453968942165,
0.07940725982189178,
0.07966019213199615,
0.1842137724161148,
0.01900150626897812,
-0.020863771438598633,
0.02815157361328602,
0.22667044401168823,
-0.1356009542942047,
0.1147211566567421,
0.12862297892570496,
-0.08630791306495667,
0.08614041656255722,
0.2117747962474823,
0.04386679828166962,
-0.09530803561210632,
0.026980465278029442,
0.03248048946261406,
-0.02578095532953739,
-0.23494455218315125,
-0.07109099626541138,
-0.001382913556881249,
-0.06392702460289001,
0.07689134776592255,
0.09707436710596085,
0.07901989668607712,
0.022557096555829048,
-0.0979517325758934,
-0.09100978821516037,
0.05943343788385391,
0.10995136201381683,
0.006441247183829546,
-0.0040697213262319565,
0.08855342864990234,
-0.03425491601228714,
0.014854288659989834,
0.08874541521072388,
0.011025206185877323,
0.15116769075393677,
0.04797039553523064,
0.1799456924200058,
0.08264807611703873,
0.07937188446521759,
0.0043103983625769615,
0.007994822226464748,
0.010194365866482258,
0.04474055394530296,
-0.007578190881758928,
-0.08390802145004272,
-0.028039509430527687,
0.11071117222309113,
0.07189466059207916,
0.014652470126748085,
0.021798308938741684,
-0.05424417927861214,
0.08654633164405823,
0.18378940224647522,
-0.003185821231454611,
-0.17969557642936707,
-0.059234656393527985,
0.07425650209188461,
-0.0982678160071373,
-0.10245729237794876,
-0.005603819154202938,
0.01935960166156292,
-0.16812188923358917,
0.03606696426868439,
-0.025378119200468063,
0.10834769159555435,
-0.12889531254768372,
-0.019157879054546356,
0.0734749361872673,
0.06843684613704681,
-0.004375826567411423,
0.05895286053419113,
-0.18712806701660156,
0.09937801957130432,
0.011319739744067192,
0.06690136343240738,
-0.09462150186300278,
0.08951155096292496,
-0.007744939532130957,
-0.029720822349190712,
0.14525602757930756,
-0.002390886889770627,
-0.06905549764633179,
-0.05996578559279442,
-0.09515837579965591,
-0.008230812847614288,
0.12368303537368774,
-0.13235269486904144,
0.08670174330472946,
-0.03138110786676407,
-0.033665917813777924,
-0.01134360209107399,
-0.08299927413463593,
-0.10598311573266983,
-0.17483621835708618,
0.05878743901848793,
-0.12589916586875916,
0.03890232741832733,
-0.1061476320028305,
-0.02385769784450531,
-0.028394047170877457,
0.170713409781456,
-0.24274958670139313,
-0.07655437290668488,
-0.14117252826690674,
-0.10428055375814438,
0.12920650839805603,
-0.04878317192196846,
0.09075501561164856,
-0.0212484672665596,
0.15581336617469788,
0.01665305159986019,
-0.01994057185947895,
0.08402728289365768,
-0.08377660810947418,
-0.1991237848997116,
-0.06802510470151901,
0.1681114286184311,
0.11482813954353333,
0.03182525187730789,
-0.0019213802879676223,
0.04034852236509323,
-0.02425713650882244,
-0.11918258666992188,
0.02143377810716629,
0.1470739096403122,
0.06553694605827332,
0.011349732056260109,
-0.015210626646876335,
-0.11726926267147064,
-0.07760429382324219,
-0.027843136340379715,
0.02808738872408867,
0.16567941009998322,
-0.07216694951057434,
0.1711350828409195,
0.14943061769008636,
-0.05817114934325218,
-0.20762984454631805,
-0.008944546803832054,
0.025544337928295135,
-0.013771671801805496,
0.009325731545686722,
-0.1831398904323578,
0.08329030126333237,
0.0029838401824235916,
-0.05627384036779404,
0.09755825251340866,
-0.16196595132350922,
-0.13705715537071228,
0.08424854278564453,
0.05096970126032829,
-0.18151970207691193,
-0.13867278397083282,
-0.09906427562236786,
-0.04121728241443634,
-0.163278728723526,
0.09223716706037521,
0.01889505609869957,
0.013068479485809803,
0.029697300866246223,
0.012294040992856026,
0.023423822596669197,
-0.05044947937130928,
0.1777648627758026,
-0.01196333672851324,
0.023793749511241913,
-0.09510384500026703,
-0.08822192251682281,
0.018427608534693718,
-0.048274971544742584,
0.0711488276720047,
-0.027883945032954216,
0.014039237052202225,
-0.1032620519399643,
-0.03400767222046852,
-0.046976327896118164,
0.016659747809171677,
-0.09917358309030533,
-0.08830299228429794,
-0.053166065365076065,
0.08740520477294922,
0.09745458513498306,
-0.02068530023097992,
-0.019553206861019135,
-0.07680746912956238,
0.05917026475071907,
0.2172931730747223,
0.18461643159389496,
0.04595598578453064,
-0.07242146134376526,
-0.003692394122481346,
-0.015667488798499107,
0.046961914747953415,
-0.19701816141605377,
0.05876782909035683,
0.05697258934378624,
0.02131287194788456,
0.10421538352966309,
-0.022381199523806572,
-0.15459972620010376,
-0.07278376072645187,
0.06308430433273315,
-0.06306322664022446,
-0.2119518518447876,
0.010038868524134159,
0.05434399098157883,
-0.1736210584640503,
-0.032768335193395615,
0.04763306677341461,
-0.008661664091050625,
-0.03624432533979416,
0.02099856361746788,
0.09542392939329147,
0.003828928340226412,
0.08080942183732986,
0.0729677751660347,
0.08458419144153595,
-0.09959036111831665,
0.08525462448596954,
0.0973353460431099,
-0.06532556563615799,
0.028015824034810066,
0.09556647390127182,
-0.058179087936878204,
-0.037699561566114426,
0.03495606407523155,
0.07672624289989471,
0.02461671084165573,
-0.044151030480861664,
0.008946296758949757,
-0.0923805981874466,
0.06369789689779282,
0.10878390818834305,
0.03081507422029972,
0.022334326058626175,
0.04528246074914932,
0.04270026832818985,
-0.0696859210729599,
0.12159191817045212,
0.031979672610759735,
0.015123683959245682,
-0.039127860218286514,
-0.043540798127651215,
0.008448269218206406,
-0.03245661407709122,
-0.0051310379058122635,
-0.02233046479523182,
-0.08641903102397919,
-0.015012029558420181,
-0.1271539181470871,
0.0009466670453548431,
-0.06485514342784882,
0.01524441409856081,
0.025784334167838097,
-0.03275036811828613,
0.009026308543980122,
0.007004431914538145,
-0.06976103037595749,
-0.06450826674699783,
-0.011749801225960255,
0.09550291299819946,
-0.16773587465286255,
0.02894597500562668,
0.08614898473024368,
-0.10924353450536728,
0.1020972803235054,
0.007519985549151897,
-0.011424945667386055,
0.020109938457608223,
-0.15724794566631317,
0.0410383902490139,
-0.03755035251379013,
0.006121676880866289,
0.01620732620358467,
-0.19358131289482117,
-0.0012844945304095745,
-0.03225119039416313,
-0.06778176128864288,
-0.00909456703811884,
-0.015164525248110294,
-0.11595740169286728,
0.10752425342798233,
0.0025998002383857965,
-0.08109515905380249,
-0.02987607941031456,
0.030584003776311874,
0.07366982102394104,
-0.030323093757033348,
0.1525343358516693,
-0.011871197260916233,
0.06408598273992538,
-0.16056349873542786,
-0.01094734389334917,
-0.00670700753107667,
0.0098334439098835,
-0.050660640001297,
-0.004018463660031557,
0.048775769770145416,
-0.015142863616347313,
0.1748305708169937,
-0.036267999559640884,
0.007594800554215908,
0.06477022171020508,
0.05298345908522606,
-0.029597442597150803,
0.09591291844844818,
0.051613468676805496,
0.016197748482227325,
0.006904417648911476,
0.008878137916326523,
-0.04523972421884537,
-0.04032763093709946,
-0.19017021358013153,
0.06588494032621384,
0.19552719593048096,
0.10028385370969772,
-0.02118326723575592,
0.07031090557575226,
-0.09745185822248459,
-0.0964348167181015,
0.15125390887260437,
-0.034086938947439194,
-0.004433101043105125,
-0.07185690104961395,
0.12947168946266174,
0.14935120940208435,
-0.18144142627716064,
0.06917683035135269,
-0.06761769950389862,
-0.041149791330099106,
-0.10744229704141617,
-0.19989708065986633,
-0.06394518166780472,
-0.04607472941279411,
-0.01700756512582302,
-0.046806927770376205,
0.06830572336912155,
0.0757194310426712,
-0.0037177233025431633,
-0.008730878122150898,
0.06848712265491486,
-0.035176318138837814,
-0.0042486852034926414,
0.027502894401550293,
0.05977346748113632,
0.010650715790688992,
-0.034018248319625854,
0.014663759618997574,
-0.00959693267941475,
0.05367758870124817,
0.07934602349996567,
0.050391409546136856,
-0.019429398700594902,
0.019031988456845284,
-0.04030049219727516,
-0.10575975477695465,
0.05164264887571335,
-0.025045137852430344,
-0.07230602949857712,
0.1537279337644577,
0.020212531089782715,
0.007143082097172737,
-0.008810870349407196,
0.23593954741954803,
-0.0627041906118393,
-0.10303844511508942,
-0.14499850571155548,
0.07503776997327805,
-0.04297162964940071,
0.04984647408127785,
0.041810184717178345,
-0.11171726882457733,
0.029548052698373795,
0.14282794296741486,
0.15390150249004364,
-0.032964590936899185,
0.023033542558550835,
0.03227212280035019,
0.008083072490990162,
-0.017554832622408867,
0.038818955421447754,
0.056219432502985,
0.15291684865951538,
-0.049605052918195724,
0.07781068980693817,
0.0030934580136090517,
-0.08526960015296936,
-0.034455299377441406,
0.11626563221216202,
-0.018170269206166267,
0.007905791513621807,
-0.06097809970378876,
0.11752495914697647,
-0.0689430758357048,
-0.22074680030345917,
0.03874767944216728,
-0.07022905349731445,
-0.12866510450839996,
-0.021623946726322174,
0.07432129979133606,
-0.009207346476614475,
0.019708409905433655,
0.0795634314417839,
-0.07134775817394257,
0.1859455704689026,
0.03771420568227768,
-0.061885230243206024,
-0.04850851371884346,
0.07243599742650986,
-0.08089011162519455,
0.29468998312950134,
0.017987435683608055,
0.04262674227356911,
0.10908371210098267,
-0.01346576027572155,
-0.13607089221477509,
0.03206384927034378,
0.09783000499010086,
-0.09518790990114212,
0.05292887985706329,
0.17464794218540192,
0.004645544569939375,
0.13460016250610352,
0.07416019588708878,
-0.08443755656480789,
0.046503692865371704,
-0.06481315195560455,
-0.07062353938817978,
-0.10096494853496552,
0.1034814640879631,
-0.09316729009151459,
0.14212666451931,
0.1201847717165947,
-0.05467384681105614,
0.008597280830144882,
-0.03530189394950867,
0.04607437178492546,
-0.00512698432430625,
0.11784914880990982,
0.011154716834425926,
-0.18924684822559357,
0.028253955766558647,
-0.032277483493089676,
0.09944566339254379,
-0.16319908201694489,
-0.08493494987487793,
0.04530351236462593,
0.0115986792370677,
-0.06823078542947769,
0.1211809515953064,
0.06131063029170036,
0.028158580884337425,
-0.048737525939941406,
-0.02928539179265499,
-0.011824419721961021,
0.13896071910858154,
-0.1002424880862236,
-0.0045283581130206585
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# wav2vec_RTSplit0207_5
This model is a fine-tuned version of [jonatasgrosman/wav2vec2-large-xlsr-53-japanese](https://huggingface.co/jonatasgrosman/wav2vec2-large-xlsr-53-japanese) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0318
- Wer: 0.1995
- Cer: 0.0859
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 32
- eval_batch_size: 32
- seed: 4
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 1000
- num_epochs: 10
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|
| 3.8157 | 1.0 | 120 | 3.6297 | 0.9897 | 0.9930 |
| 1.6219 | 2.0 | 240 | 1.4240 | 0.8251 | 0.6728 |
| 0.9197 | 3.0 | 360 | 0.7458 | 0.8210 | 0.5231 |
| 0.6346 | 4.0 | 480 | 0.4789 | 0.6260 | 0.2890 |
| 0.4992 | 5.0 | 600 | 0.3283 | 0.5051 | 0.2244 |
| 0.3764 | 6.0 | 720 | 0.2164 | 0.3602 | 0.1251 |
| 0.3036 | 7.0 | 840 | 0.1210 | 0.2862 | 0.1109 |
| 0.2296 | 8.0 | 960 | 0.0703 | 0.2316 | 0.0890 |
| 0.1974 | 9.0 | 1080 | 0.0453 | 0.2120 | 0.0951 |
| 0.1126 | 10.0 | 1200 | 0.0318 | 0.1995 | 0.0859 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.14.6
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["wer"], "base_model": "jonatasgrosman/wav2vec2-large-xlsr-53-japanese", "model-index": [{"name": "wav2vec_RTSplit0207_5", "results": []}]} | automatic-speech-recognition | tndklab/wav2vec_RTSplit0207_5 | [
"transformers",
"safetensors",
"wav2vec2",
"automatic-speech-recognition",
"generated_from_trainer",
"base_model:jonatasgrosman/wav2vec2-large-xlsr-53-japanese",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-07T07:25:15+00:00 | [] | [] | TAGS
#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-jonatasgrosman/wav2vec2-large-xlsr-53-japanese #license-apache-2.0 #endpoints_compatible #region-us
| wav2vec\_RTSplit0207\_5
=======================
This model is a fine-tuned version of jonatasgrosman/wav2vec2-large-xlsr-53-japanese on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0318
* Wer: 0.1995
* Cer: 0.0859
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 32
* eval\_batch\_size: 32
* seed: 4
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 1000
* num\_epochs: 10
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.14.6
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-jonatasgrosman/wav2vec2-large-xlsr-53-japanese #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
80,
116,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-jonatasgrosman/wav2vec2-large-xlsr-53-japanese #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
-0.14118388295173645,
0.1513376235961914,
-0.0005152393132448196,
0.09957455098628998,
0.11795935779809952,
0.008073536679148674,
0.17546747624874115,
0.14973945915699005,
-0.040967848151922226,
0.11099261045455933,
0.11360207200050354,
0.06234046816825867,
0.05541789159178734,
0.19745732843875885,
-0.08206405490636826,
-0.22053459286689758,
0.07701526582241058,
-0.0033247191458940506,
0.010011174716055393,
0.11214284598827362,
0.07103060185909271,
-0.11864254623651505,
0.09036611765623093,
-0.006963435560464859,
-0.1427508145570755,
-0.04181395843625069,
0.016152987256646156,
-0.11031178385019302,
0.10872969776391983,
0.009288477711379528,
0.06582261621952057,
0.03504345938563347,
0.08960989117622375,
-0.18766173720359802,
0.0022469565737992525,
0.017141321673989296,
0.01435944065451622,
0.07424193620681763,
0.04270157217979431,
-0.0012159834150224924,
0.0027513811364769936,
-0.11490482091903687,
0.03699565678834915,
0.015048346482217312,
-0.11695785820484161,
-0.19873808324337006,
-0.07786379754543304,
0.017535261809825897,
0.09932611882686615,
0.08391966670751572,
-0.02103434130549431,
0.12361589074134827,
0.0008332731667906046,
0.07967887073755264,
0.19855915009975433,
-0.3135106563568115,
-0.05437514930963516,
-0.017246698960661888,
0.03944491222500801,
0.08242533355951309,
-0.10150530934333801,
-0.018765252083539963,
0.05041765794157982,
0.021163977682590485,
0.0925753116607666,
-0.03130659833550453,
-0.0336025208234787,
-0.010997597128152847,
-0.12088784575462341,
-0.03848946467041969,
0.1899806410074234,
0.07315327972173691,
-0.06343334913253784,
-0.0812249630689621,
-0.06397605687379837,
-0.12176059931516647,
-0.054642703384160995,
-0.00814018677920103,
0.026548052206635475,
-0.03949858993291855,
-0.09949330240488052,
-0.00489323865622282,
-0.0798637866973877,
-0.09082527458667755,
-0.017095407471060753,
0.17612586915493011,
0.01114688627421856,
0.013611323200166225,
-0.012470031157135963,
0.054258283227682114,
-0.02387041598558426,
-0.18483866751194,
-0.022159598767757416,
0.026931334286928177,
-0.032688405364751816,
-0.014408618211746216,
-0.04372356832027435,
-0.033641211688518524,
0.04413202032446861,
0.1181967481970787,
-0.018560312688350677,
0.06547611951828003,
-0.024242987856268883,
0.001249976921826601,
-0.08463852852582932,
0.182333305478096,
-0.06463606655597687,
-0.06938374042510986,
0.019906673580408096,
0.1272294521331787,
0.06285319477319717,
-0.02319909818470478,
-0.09877349436283112,
-0.008964418433606625,
0.1465861052274704,
0.03547809645533562,
-0.04232742264866829,
0.05053495988249779,
-0.038675304502248764,
-0.014026766642928123,
0.05737104266881943,
-0.12155470997095108,
0.025580810382962227,
0.022164685651659966,
-0.06321313977241516,
-0.02220059186220169,
-0.012167129665613174,
0.013010796159505844,
0.01259292010217905,
0.05222075432538986,
-0.08240532130002975,
0.0036988360807299614,
-0.02336418628692627,
-0.09296302497386932,
0.02680950053036213,
-0.06885657459497452,
0.0002455086214467883,
-0.10807177424430847,
-0.1792655736207962,
-0.017911894246935844,
0.023634862154722214,
-0.048911672085523605,
-0.009945076890289783,
-0.11301852017641068,
-0.0975419208407402,
0.04686196520924568,
-0.022707432508468628,
0.03525571897625923,
-0.07948608696460724,
0.10767754912376404,
0.07949104905128479,
0.08755897730588913,
-0.04064375162124634,
0.026420699432492256,
-0.09513017535209656,
0.03219624236226082,
-0.17416203022003174,
0.07473883777856827,
-0.054269835352897644,
0.034595683217048645,
-0.12033510208129883,
-0.0673440620303154,
0.019730009138584137,
-0.022371632978320122,
0.06979899108409882,
0.14241422712802887,
-0.1906011998653412,
-0.057025082409381866,
0.19722899794578552,
-0.11986931413412094,
-0.14269396662712097,
0.1278074085712433,
-0.03618486225605011,
0.038308508694171906,
0.07103344798088074,
0.22304697334766388,
0.03084309957921505,
-0.10670869052410126,
-0.04001520201563835,
-0.06360205262899399,
0.08328385651111603,
-0.03686724230647087,
0.1115616038441658,
0.005013670772314072,
-0.0014164310414344072,
0.016217360273003578,
-0.08063652366399765,
0.03222700208425522,
-0.07089976221323013,
-0.10016977041959763,
-0.044187162071466446,
-0.10613103210926056,
0.028292791917920113,
0.015585470013320446,
0.05590164288878441,
-0.09868429601192474,
-0.07060223072767258,
0.011771205812692642,
0.10809079557657242,
-0.11718438565731049,
0.012904723174870014,
-0.10417094081640244,
0.09427313506603241,
-0.11367906630039215,
-0.020158156752586365,
-0.15430019795894623,
-0.00410782964900136,
0.053871724754571915,
0.01823066547513008,
0.014241056516766548,
-0.07536637037992477,
0.08248129487037659,
0.07677027583122253,
-0.048812806606292725,
-0.07412240654230118,
-0.00498126819729805,
0.01805686019361019,
-0.06261217594146729,
-0.17385663092136383,
-0.028737887740135193,
-0.05413864552974701,
0.1608741134405136,
-0.16503508388996124,
0.0014996977988630533,
0.008824673481285572,
0.09039325267076492,
0.044107191264629364,
-0.02336837165057659,
0.019647266715765,
0.04836883023381233,
-0.026001546531915665,
-0.07121788710355759,
0.02903486229479313,
0.015071400441229343,
-0.10376740992069244,
0.019600635394454002,
-0.16807952523231506,
0.15009671449661255,
0.1392962485551834,
0.041799139231443405,
-0.05244087427854538,
0.020045755431056023,
-0.014278299175202847,
-0.04192302003502846,
-0.054265860468149185,
-0.015523474663496017,
0.10089807957410812,
0.007953685708343983,
0.12119784951210022,
-0.10283432900905609,
0.015069880522787571,
0.06469476968050003,
-0.027236688882112503,
-0.028428832069039345,
0.08093540370464325,
0.010932014323771,
-0.13942010700702667,
0.13027305901050568,
0.11321879178285599,
-0.07171899825334549,
0.12632238864898682,
-0.061046771705150604,
-0.085269495844841,
-0.05007721483707428,
0.033539436757564545,
0.03383169323205948,
0.13762885332107544,
-0.08172392845153809,
-0.022573819383978844,
0.02135329321026802,
0.021954631432890892,
-0.01604454591870308,
-0.19322632253170013,
-0.01944556087255478,
0.014740300364792347,
-0.09451686590909958,
-0.009623108431696892,
0.005852391943335533,
-0.017728494480252266,
0.09443813562393188,
-0.0006852737278677523,
-0.11360033601522446,
0.02314845845103264,
-0.015061916783452034,
-0.0876469686627388,
0.17204688489437103,
-0.09222397953271866,
-0.17411498725414276,
-0.13528600335121155,
-0.07141708582639694,
-0.05600440874695778,
0.03755852207541466,
0.06046575680375099,
-0.06563454121351242,
-0.04083354398608208,
-0.11549428105354309,
-0.04752589762210846,
0.03249216452240944,
0.04562458395957947,
0.05096840113401413,
-0.008819636888802052,
0.06599970906972885,
-0.08180946111679077,
-0.005118743982166052,
-0.014260591939091682,
-0.007273803930729628,
0.028529318049550056,
0.0005596327246166766,
0.12637819349765778,
0.12142695486545563,
0.00641365023329854,
0.024710198864340782,
-0.03795645385980606,
0.22719408571720123,
-0.06981682777404785,
-0.019021375104784966,
0.12310279160737991,
-0.027643905952572823,
0.04546043276786804,
0.17790329456329346,
0.03071616031229496,
-0.10723342001438141,
0.001543888240121305,
-0.049799270927906036,
-0.015234281308948994,
-0.188801571726799,
-0.033299293369054794,
-0.04766744375228882,
0.013789367862045765,
0.10153107345104218,
0.029864661395549774,
0.01635241135954857,
0.04831988736987114,
0.02163584530353546,
0.045850567519664764,
0.004909783601760864,
0.08106916397809982,
0.09610755741596222,
0.07607916742563248,
0.10785672068595886,
-0.03219561278820038,
-0.04924085736274719,
0.03241001442074776,
0.02096749283373356,
0.2015942931175232,
0.03050840273499489,
0.19284749031066895,
0.0009453566162846982,
0.15461201965808868,
0.025513680651783943,
0.08029980212450027,
0.0196499302983284,
0.009997223503887653,
-0.020813269540667534,
-0.07838824391365051,
-0.053045131266117096,
0.05521805211901665,
-0.013811358250677586,
0.06097419187426567,
-0.10551154613494873,
0.02013915218412876,
0.050434961915016174,
0.2723030149936676,
0.08769560605287552,
-0.3685109615325928,
-0.08635305613279343,
0.02104947343468666,
-0.036972131580114365,
-0.01903359405696392,
0.016876420006155968,
0.1534101963043213,
-0.06155303493142128,
0.06892576813697815,
-0.07162059843540192,
0.06354092806577682,
-0.06427314132452011,
0.01977919228374958,
0.023796986788511276,
0.047146882861852646,
0.0028003251645714045,
0.030816223472356796,
-0.24248063564300537,
0.28648892045021057,
0.035792022943496704,
0.09525445103645325,
-0.05690842866897583,
-0.0036345715634524822,
0.03943183273077011,
-0.0068437037989497185,
0.11822443455457687,
-0.024794165045022964,
-0.11079125106334686,
-0.1794157326221466,
-0.13511526584625244,
0.048938214778900146,
0.10588322579860687,
-0.006933995056897402,
0.11613228172063828,
-0.013944382779300213,
-0.04456436634063721,
0.04456821084022522,
-0.02340116538107395,
-0.08081210404634476,
-0.0749657079577446,
0.009459982626140118,
0.11387364566326141,
0.044953037053346634,
-0.04946205019950867,
-0.09598170965909958,
-0.09000002592802048,
0.08921240270137787,
0.0022167686838656664,
-0.006378722842782736,
-0.10505329072475433,
0.018605045974254608,
0.1496697962284088,
-0.09152472764253616,
0.053013674914836884,
0.00948717724531889,
0.1103808656334877,
0.027091750875115395,
-0.04968217760324478,
0.09032276272773743,
-0.06209796667098999,
-0.17874783277511597,
-0.050470758229494095,
0.13884679973125458,
-0.007241794373840094,
0.04262242838740349,
0.021225502714514732,
0.05142869055271149,
-0.005594416055828333,
-0.06703758984804153,
0.03248269483447075,
0.0272601880133152,
0.04040838032960892,
0.020026197656989098,
-0.012260128743946552,
-0.09217273443937302,
-0.09288010746240616,
-0.022662481293082237,
0.15083232522010803,
0.2982642650604248,
-0.06597831845283508,
0.01851712167263031,
0.08686181157827377,
-0.017854949459433556,
-0.15149927139282227,
-0.004584764130413532,
0.04337412118911743,
0.044433437287807465,
-0.005071667488664389,
-0.12238527834415436,
0.04546479135751724,
0.061537906527519226,
-0.04462991654872894,
0.07658969610929489,
-0.24918095767498016,
-0.1279260665178299,
0.08928412944078445,
0.1333279311656952,
0.12402694672346115,
-0.15273624658584595,
-0.06697630137205124,
-0.024110818281769753,
-0.10794417560100555,
0.1045057401061058,
-0.07326747477054596,
0.132745623588562,
-0.0021595852449536324,
0.06386832147836685,
0.0073677487671375275,
-0.05131489410996437,
0.15022529661655426,
0.02315061166882515,
0.05349516496062279,
-0.02223295159637928,
-0.01721254177391529,
0.04753045365214348,
-0.07513923943042755,
0.06926781684160233,
-0.0858917385339737,
0.049377862364053726,
-0.05993237346410751,
-0.024863220751285553,
-0.06171049550175667,
-0.006019624415785074,
0.0036315324250608683,
-0.03500806540250778,
-0.010318802669644356,
0.03614702448248863,
0.05882398411631584,
0.003448877949267626,
0.13278301060199738,
0.01197106670588255,
0.08229319751262665,
0.14570775628089905,
0.0886906087398529,
-0.03906968981027603,
0.014958192594349384,
-0.006075490266084671,
-0.05644846335053444,
0.053838834166526794,
-0.13194115459918976,
0.04858151823282242,
0.09669843316078186,
0.018267197534441948,
0.16071775555610657,
0.04687272757291794,
-0.0492333360016346,
0.03808420151472092,
0.06964091956615448,
-0.15898656845092773,
-0.1112910732626915,
0.003452348755672574,
-0.01375506166368723,
-0.11069352924823761,
0.04991121590137482,
0.13865762948989868,
-0.0703156515955925,
-0.007155262865126133,
-0.01823554001748562,
0.021476034075021744,
-0.03947385773062706,
0.20047150552272797,
0.042284511029720306,
0.05118047818541527,
-0.10997112095355988,
0.08147826790809631,
0.05683886259794235,
-0.08807548880577087,
0.049443088471889496,
0.037657495588064194,
-0.11523519456386566,
-0.022784452885389328,
0.00005534075535251759,
0.14294663071632385,
0.005469065625220537,
-0.07627619802951813,
-0.13811787962913513,
-0.08861514180898666,
0.035080134868621826,
0.17778432369232178,
0.06808756291866302,
0.03646450862288475,
-0.01790330372750759,
-0.0017073694616556168,
-0.10371081531047821,
0.09405805170536041,
0.07413934916257858,
0.07491617649793625,
-0.14983569085597992,
0.08254153281450272,
-0.008294850587844849,
0.0264718197286129,
-0.02044260874390602,
0.016870073974132538,
-0.11009109020233154,
0.0050047170370817184,
-0.09842979907989502,
0.05644430220127106,
-0.07891111820936203,
-0.016069279983639717,
-0.0012212666915729642,
-0.08156505972146988,
-0.06149281933903694,
0.011379055678844452,
-0.08712979406118393,
-0.026130566373467445,
0.003458258928731084,
0.0433938093483448,
-0.13683374226093292,
-0.03750019147992134,
0.02246333472430706,
-0.0981111079454422,
0.08394405245780945,
0.0866144597530365,
-0.019790008664131165,
0.04684387892484665,
-0.09545118361711502,
-0.021657824516296387,
0.08283203840255737,
0.0022664524149149656,
0.05060122534632683,
-0.14470165967941284,
-0.013973113149404526,
0.03143203258514404,
0.05041022598743439,
0.021516086533665657,
0.14788220822811127,
-0.09649581462144852,
0.005971268285065889,
-0.06748102605342865,
-0.011066189967095852,
-0.05691782385110855,
0.021555546671152115,
0.14140185713768005,
0.002627637470141053,
0.18363331258296967,
-0.09465426206588745,
0.022600574418902397,
-0.19713501632213593,
0.0013647646410390735,
-0.03708446025848389,
-0.12615914642810822,
-0.14735087752342224,
-0.02729809284210205,
0.07842589169740677,
-0.062242452055215836,
0.09589572995901108,
-0.06168156489729881,
0.06943611055612564,
0.013083240017294884,
-0.05813242867588997,
-0.0013656788505613804,
0.04068639129400253,
0.24912409484386444,
0.057672739028930664,
-0.03611792251467705,
0.07662001997232437,
0.010403653606772423,
0.09453791379928589,
0.12455988675355911,
0.12363877147436142,
0.15776370465755463,
0.03212352469563484,
0.14337152242660522,
0.08356481045484543,
-0.024862807244062424,
-0.11890481412410736,
0.05916835740208626,
-0.06845055520534515,
0.0900888592004776,
0.025596898049116135,
0.2101191282272339,
0.09875553101301193,
-0.1637156754732132,
0.004310329910367727,
-0.03623693808913231,
-0.08481713384389877,
-0.09586827456951141,
-0.06154704466462135,
-0.13099995255470276,
-0.14432699978351593,
0.010924887843430042,
-0.1069321557879448,
0.03407773748040199,
0.06964482367038727,
0.014476785436272621,
0.00010840524919331074,
0.14145208895206451,
0.014773874543607235,
0.029025593772530556,
0.09571029245853424,
0.008392676711082458,
-0.04039377346634865,
-0.0008097448153421283,
-0.10300909727811813,
0.023898394778370857,
0.00501815602183342,
0.056743279099464417,
-0.021962566301226616,
-0.02582181803882122,
0.06900105625391006,
-0.025983864441514015,
-0.1252528429031372,
0.011205081827938557,
0.020153336226940155,
0.06013365089893341,
0.04367395490407944,
0.056713055819272995,
-0.017620805650949478,
0.025014393031597137,
0.20733173191547394,
-0.08926752954721451,
-0.07744365930557251,
-0.13346822559833527,
0.14780977368354797,
-0.0138862831518054,
-0.007439719047397375,
0.009504212997853756,
-0.10589496046304703,
0.00224756239913404,
0.1942233294248581,
0.1499839723110199,
-0.07326409965753555,
-0.0006150471162982285,
-0.02711077593266964,
-0.00640704995021224,
-0.03820818290114403,
0.0650281012058258,
0.07840298116207123,
0.034390028566122055,
-0.059590261429548264,
-0.06017826497554779,
-0.05757676437497139,
-0.041262898594141006,
-0.022971007972955704,
0.03991897031664848,
-0.0317038968205452,
-0.023118194192647934,
-0.04992280900478363,
0.07865259051322937,
-0.08221094310283661,
-0.0975438579916954,
0.00718031357973814,
-0.21697042882442474,
-0.17321449518203735,
-0.0021544182673096657,
0.07518170028924942,
0.03575673699378967,
0.026161469519138336,
-0.033357031643390656,
0.02538921684026718,
0.05649259686470032,
-0.013258103281259537,
-0.057701025158166885,
-0.05936339870095253,
0.04269890859723091,
-0.08285795897245407,
0.17427952587604523,
-0.00461580278351903,
0.06635239720344543,
0.1040571853518486,
0.08133987337350845,
-0.10723624378442764,
0.10374195873737335,
0.06103026121854782,
-0.07403004914522171,
0.055740538984537125,
0.15157853066921234,
-0.056069888174533844,
0.14287318289279938,
0.05193907395005226,
-0.10281984508037567,
0.001149781746789813,
0.010053437203168869,
-0.028709089383482933,
-0.07485763728618622,
-0.06516216695308685,
-0.045741770416498184,
0.14665868878364563,
0.13426938652992249,
-0.06621623039245605,
0.0017491946928203106,
-0.016887987032532692,
0.056069597601890564,
0.06269390136003494,
0.02119249664247036,
-0.06178002804517746,
-0.2836269438266754,
-0.016548514366149902,
0.03701779246330261,
0.022455256432294846,
-0.2418002039194107,
-0.08923479914665222,
-0.010862430557608604,
-0.04599393531680107,
-0.07464209198951721,
0.0931960716843605,
0.08006857335567474,
0.03134746104478836,
-0.054390132427215576,
-0.05154174938797951,
-0.028746988624334335,
0.17312726378440857,
-0.1632525771856308,
-0.11506713181734085
] |
null | null | null |
# **Q-Learning** Agent playing1 **Taxi-v3**
This is a trained model of a **Q-Learning** agent playing **Taxi-v3** .
## Usage
```python
model = load_from_hub(repo_id="bianxg/q-Taxi-v3", filename="q-learning.pkl")
# Don't forget to check if you need to add additional attributes (is_slippery=False etc)
env = gym.make(model["env_id"])
```
| {"tags": ["Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "q-Taxi-v3", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "Taxi-v3", "type": "Taxi-v3"}, "metrics": [{"type": "mean_reward", "value": "7.56 +/- 2.71", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | bianxg/q-Taxi-v3 | [
"Taxi-v3",
"q-learning",
"reinforcement-learning",
"custom-implementation",
"model-index",
"region:us"
] | 2024-02-07T07:26:20+00:00 | [] | [] | TAGS
#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us
|
# Q-Learning Agent playing1 Taxi-v3
This is a trained model of a Q-Learning agent playing Taxi-v3 .
## Usage
| [
"# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage"
] | [
"TAGS\n#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n",
"# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage"
] | [
32,
33
] | [
"passage: TAGS\n#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage"
] | [
0.048862796276807785,
-0.16549694538116455,
-0.005485367961227894,
0.02960980497300625,
0.1345081776380539,
-0.01784728653728962,
0.11895976960659027,
0.07759871333837509,
-0.07461097836494446,
-0.055395450443029404,
0.1418241262435913,
0.09088201075792313,
0.055222880095243454,
0.05699880048632622,
0.09511256217956543,
-0.27440664172172546,
0.048217080533504486,
-0.02918700873851776,
0.05621987581253052,
0.11878681182861328,
0.0670095682144165,
-0.040441032499074936,
0.061956584453582764,
0.11818158626556396,
-0.1018151044845581,
-0.007344264071434736,
0.035402704030275345,
-0.09440053254365921,
0.17413531243801117,
0.07204403728246689,
0.12337774783372879,
0.05132639780640602,
0.179361954331398,
-0.12762396037578583,
0.024310702458024025,
-0.0010275895474478602,
-0.10138072073459625,
-0.03909514099359512,
-0.012415820732712746,
-0.08349097520112991,
0.03230205550789833,
0.23522862792015076,
0.07199250161647797,
0.06632792949676514,
-0.17707863450050354,
-0.06584878265857697,
-0.04375573247671127,
0.069611094892025,
0.14951466023921967,
0.03758616745471954,
-0.033800311386585236,
0.1684885323047638,
-0.2564343810081482,
0.05066783353686333,
0.037275806069374084,
-0.42313119769096375,
0.017119819298386574,
0.1507398933172226,
0.15090937912464142,
0.06909667700529099,
-0.10573802888393402,
0.013512322679162025,
0.051325585693120956,
-0.0005318621988408267,
0.024325110018253326,
0.006554204970598221,
0.15601307153701782,
0.08537693321704865,
-0.1487821787595749,
-0.058576688170433044,
0.17441977560520172,
-0.03788546845316887,
-0.02613203600049019,
-0.039745692163705826,
0.0067160045728087425,
-0.06427708268165588,
-0.004067842848598957,
-0.1777995079755783,
0.00734262028709054,
0.06666424125432968,
-0.014348524622619152,
0.014901017770171165,
-0.035522811114788055,
-0.0966939702630043,
-0.023098144680261612,
-0.08592145889997482,
0.01677769608795643,
-0.006319406442344189,
-0.10187895596027374,
0.05002119392156601,
-0.061138734221458435,
0.0014382408699020743,
-0.05123179033398628,
-0.15047866106033325,
-0.049055423587560654,
-0.03481535613536835,
0.1474713832139969,
-0.0044205985032022,
-0.01873963139951229,
-0.03164304047822952,
0.15474793314933777,
0.049551334232091904,
-0.05370146036148071,
0.05625450983643532,
0.07605006545782089,
0.23867930471897125,
0.10401605814695358,
0.10196955502033234,
-0.06798075139522552,
0.10180158913135529,
-0.12330973148345947,
-0.08915644884109497,
-0.17508824169635773,
0.11820860952138901,
0.00015364694991149008,
0.1317785084247589,
-0.12023144960403442,
0.07898581773042679,
-0.067511186003685,
0.013453764840960503,
0.01636839471757412,
0.0820009782910347,
-0.012399360537528992,
0.10676060616970062,
-0.005061192903667688,
-0.06941985338926315,
0.014177112840116024,
0.05935845896601677,
0.03754841163754463,
-0.038601722568273544,
-0.03192409873008728,
-0.05762290954589844,
-0.05065649375319481,
-0.10128600150346756,
-0.06447898596525192,
0.018573462963104248,
-0.007677143905311823,
-0.1833900660276413,
-0.06407523155212402,
0.00897200871258974,
0.015712225809693336,
-0.03988850116729736,
-0.05148044601082802,
-0.15265507996082306,
-0.042461175471544266,
-0.015450406819581985,
-0.03500641882419586,
-0.06214277446269989,
-0.0383245050907135,
0.046435944736003876,
-0.07560601085424423,
0.013364278711378574,
0.023342855274677277,
0.05405820533633232,
-0.025881100445985794,
0.06068144738674164,
-0.08357544988393784,
0.09493788331747055,
-0.1540430635213852,
-0.03271956741809845,
-0.025445878505706787,
-0.041183918714523315,
0.1752462536096573,
0.06099751964211464,
-0.015994304791092873,
0.15260063111782074,
-0.17141541838645935,
-0.058121129870414734,
0.15596486628055573,
0.008629098534584045,
-0.09967197477817535,
-0.003560945624485612,
-0.09397093951702118,
0.1428760588169098,
0.08571921288967133,
0.2478504776954651,
0.12005335837602615,
-0.22748184204101562,
0.055358242243528366,
0.12515293061733246,
-0.14365963637828827,
0.10365243256092072,
0.07344598323106766,
0.005470725707709789,
-0.18886831402778625,
-0.06843198090791702,
-0.06121627986431122,
0.1053021252155304,
-0.08522345870733261,
-0.0776243582367897,
0.09323626756668091,
-0.05086790770292282,
0.24641476571559906,
-0.028281206265091896,
0.06174173951148987,
-0.026681531220674515,
-0.1389324963092804,
-0.01723906397819519,
0.060955192893743515,
0.05258452147245407,
-0.024835573509335518,
-0.25895482301712036,
0.13646544516086578,
0.048650871962308884,
0.025074828416109085,
0.004106190986931324,
-0.05691491439938545,
0.016934165731072426,
0.1511998474597931,
0.020012924447655678,
0.13717477023601532,
0.027723990380764008,
0.0706823319196701,
-0.006239562761038542,
-0.10560829937458038,
-0.04169593006372452,
0.061916545033454895,
-0.08518962562084198,
-0.06641357392072678,
0.011197872459888458,
-0.06935211271047592,
-0.11783787608146667,
-0.12166737765073776,
-0.026334572583436966,
-0.02980303019285202,
-0.07444227486848831,
0.02368103712797165,
0.06536602973937988,
-0.06702698022127151,
-0.0023908785078674555,
0.007125476840883493,
-0.011537045240402222,
0.16434046626091003,
0.011393417604267597,
-0.007796820718795061,
0.1328643560409546,
-0.11533161997795105,
0.12461213022470474,
0.049438029527664185,
-0.024806302040815353,
-0.04662557691335678,
0.0014137453399598598,
-0.057529181241989136,
0.029044216498732567,
-0.04390640929341316,
0.02774495631456375,
0.20111067593097687,
0.02772962674498558,
0.11389166116714478,
-0.0656520202755928,
0.04385066404938698,
-0.007961965166032314,
-0.009693224914371967,
0.018563594669103622,
0.07608018070459366,
0.07813210040330887,
-0.1324140727519989,
0.02262016013264656,
0.22455167770385742,
0.1385764330625534,
0.18313980102539062,
-0.010877152904868126,
0.06325667351484299,
-0.04875868931412697,
0.027505528181791306,
0.024100203067064285,
0.10314226150512695,
-0.10732068121433258,
-0.0322517491877079,
-0.025407759472727776,
0.023599207401275635,
-0.08197105675935745,
-0.1055799350142479,
-0.090115025639534,
0.01222382951527834,
-0.03125503659248352,
-0.15570329129695892,
0.13300658762454987,
-0.10451057553291321,
0.01802753657102585,
0.04692702740430832,
-0.22163605690002441,
0.11530312895774841,
0.014291439205408096,
-0.10303618758916855,
0.11281087249517441,
-0.12051989883184433,
-0.08699832111597061,
-0.05777236074209213,
-0.18658851087093353,
0.05280197039246559,
0.04673841595649719,
0.05166793242096901,
-0.18521739542484283,
0.024835903197526932,
0.05545609071850777,
0.13426995277404785,
-0.09743253141641617,
-0.07142634689807892,
-0.15038461983203888,
0.016068490222096443,
-0.033661190420389175,
-0.16029728949069977,
-0.005609163548797369,
-0.032781440764665604,
-0.18849676847457886,
-0.04539939761161804,
-0.15086813271045685,
-0.034627582877874374,
0.20464378595352173,
0.026907702907919884,
0.09480511397123337,
-0.07926445454359055,
0.3802889585494995,
-0.042039383202791214,
-0.06146497279405594,
-0.01321389526128769,
-0.07072482258081436,
0.02512686513364315,
0.13271741569042206,
0.0036099457647651434,
-0.017886579036712646,
-0.0037857077550143003,
0.0024592927657067776,
-0.06234965845942497,
-0.13400450348854065,
0.0028710351325571537,
0.03905198723077774,
0.1874423623085022,
0.004639793653041124,
0.06659388542175293,
0.03133883699774742,
0.057546284049749374,
0.07748064398765564,
0.030926106497645378,
0.0011591583024710417,
-0.01591806672513485,
0.06604493409395218,
-0.11684755235910416,
0.042466625571250916,
-0.030429253354668617,
-0.10143838077783585,
-0.013183288276195526,
0.07950251549482346,
0.12755028903484344,
0.17849206924438477,
-0.04790908098220825,
0.17489230632781982,
0.13580141961574554,
0.16576050221920013,
0.049315933138132095,
-0.020801831036806107,
-0.08773037046194077,
-0.06118565797805786,
0.004774159751832485,
-0.031952597200870514,
0.04869702458381653,
0.3231290578842163,
0.037619613111019135,
-0.09036035090684891,
0.11149907857179642,
0.009480619803071022,
0.05359881371259689,
0.022797370329499245,
-0.11162138730287552,
0.11170321702957153,
0.07968773692846298,
-0.06341761350631714,
-0.07602835446596146,
0.16758501529693604,
-0.1109386757016182,
-0.26646625995635986,
-0.11410990357398987,
-0.012305386364459991,
0.07903840392827988,
0.005651174578815699,
0.05498376116156578,
-0.11829282343387604,
-0.16034497320652008,
-0.034191906452178955,
0.1335442066192627,
-0.3077351450920105,
0.2065143585205078,
-0.0198091771453619,
0.06707923114299774,
-0.039657969027757645,
-0.07026876509189606,
0.09694647043943405,
0.13174086809158325,
0.29124146699905396,
0.01396956667304039,
0.04841272905468941,
-0.15176129341125488,
-0.0976925864815712,
0.0018439020495861769,
0.015482662245631218,
-0.02563396655023098,
0.028520405292510986,
-0.0540912002325058,
0.008404579944908619,
-0.018086453899741173,
0.2102297693490982,
-0.11316607892513275,
0.004344627261161804,
-0.06968966871500015,
-0.11707738786935806,
0.19409789144992828,
-0.07178345322608948,
-0.04543264955282211,
-0.14959357678890228,
-0.15512511134147644,
-0.004174166824668646,
-0.02413962036371231,
-0.019664527848362923,
-0.17603960633277893,
-0.18804074823856354,
-0.05204557999968529,
-0.005645004566758871,
-0.003464865731075406,
0.05867868289351463,
-0.07517234236001968,
-0.04805335775017738,
0.1009904220700264,
-0.07743175327777863,
-0.056063808500766754,
-0.1103200614452362,
0.1391381323337555,
0.06248528137803078,
0.16743235290050507,
0.05907081440091133,
0.0006117874872870743,
0.11471151560544968,
-0.02913086675107479,
0.11103474348783493,
-0.11291708797216415,
-0.17145049571990967,
-0.08334989100694656,
-0.018775060772895813,
0.09519003331661224,
-0.04789286106824875,
0.0028788831550627947,
0.2550160884857178,
0.14880181849002838,
-0.0897710770368576,
0.27680760622024536,
0.04414956644177437,
-0.09375058114528656,
-0.18432219326496124,
-0.15961645543575287,
0.03759992495179176,
0.060025621205568314,
0.13095876574516296,
-0.057205069810152054,
-0.08483537286520004,
-0.08492398262023926,
-0.07478608191013336,
-0.13140805065631866,
-0.24232175946235657,
-0.030598774552345276,
0.22874866425991058,
0.08656918257474899,
0.08219650387763977,
-0.012482990510761738,
-0.01186054851859808,
0.00526038184762001,
0.02680150233209133,
0.12018456310033798,
-0.13341329991817474,
0.11107480525970459,
0.022198403254151344,
0.044267985969781876,
0.009712530300021172,
0.07929777354001999,
0.03375575691461563,
-0.003218587953597307,
-0.0006439819699153304,
-0.0988350659608841,
-0.2596651017665863,
0.0816885456442833,
-0.01623627357184887,
-0.09960969537496567,
0.014988959766924381,
0.02061903104186058,
-0.2089255303144455,
0.011128270998597145,
-0.019883770495653152,
-0.03150356933474541,
-0.06483490765094757,
-0.10664787143468857,
-0.056551624089479446,
0.04928823933005333,
0.10853826254606247,
0.011660109274089336,
0.05354316532611847,
-0.0404130220413208,
0.07917837053537369,
0.0826287642121315,
0.15132710337638855,
0.06795957684516907,
-0.190711110830307,
-0.10953907668590546,
-0.0414445661008358,
0.12121522426605225,
-0.12505418062210083,
0.036917757242918015,
0.053161121904850006,
-0.016534561291337013,
0.14621229469776154,
0.1070784479379654,
-0.07452095299959183,
0.11915595084428787,
0.08904775977134705,
-0.04094788804650307,
-0.23367151618003845,
-0.07120766490697861,
0.11133213341236115,
0.07195597887039185,
-0.03961895406246185,
0.018120890483260155,
-0.04960581287741661,
-0.013980977237224579,
0.048759616911411285,
-0.0538676381111145,
-0.07230538129806519,
0.004421027842909098,
0.1247575581073761,
0.1029362753033638,
-0.04655474051833153,
0.01296416949480772,
0.037371400743722916,
0.003788623260334134,
0.04730486497282982,
0.0407949760556221,
-0.08269952982664108,
-0.04124005511403084,
0.02782733179628849,
0.37552911043167114,
-0.010165480896830559,
-0.020456433296203613,
0.018555615097284317,
-0.19949445128440857,
0.09135842323303223,
0.13205479085445404,
0.04697350412607193,
0.004247748292982578,
-0.08139242231845856,
0.026877427473664284,
-0.010625290684401989,
0.09936143457889557,
-0.07806670665740967,
-0.05493134260177612,
-0.21631066501140594,
-0.025010565295815468,
0.017490221187472343,
0.24077683687210083,
-0.08458559215068817,
-0.12801732122898102,
-0.20628872513771057,
0.13128381967544556,
-0.11333390325307846,
-0.03695881739258766,
-0.024473199620842934,
0.03926658630371094,
-0.01989821158349514,
0.06291737407445908,
-0.0710630789399147,
0.006373001262545586,
-0.11024709790945053,
0.055267609655857086,
0.04204455390572548,
0.1229788213968277,
0.014207782223820686,
0.02016810141503811,
0.05822525918483734,
-0.01837925612926483,
0.07173580676317215,
-0.06203491613268852,
-0.04550490900874138,
0.14224006235599518,
-0.020255116745829582,
-0.04152837023139,
-0.0483345128595829,
-0.036874305456876755,
0.11981741338968277,
-0.05059147998690605,
-0.007141099311411381,
-0.054929375648498535,
-0.06906463205814362,
0.03462086617946625,
-0.009175732731819153,
-0.008798843249678612,
0.06801853328943253,
0.04024988040328026,
-0.026994358748197556,
0.005263668950647116,
0.03447828069329262,
-0.10330043733119965,
-0.04955084249377251,
0.16955432295799255,
-0.0749620869755745,
0.10274054110050201,
-0.031069839373230934,
0.018015999346971512,
0.005847334861755371,
-0.022399673238396645,
-0.015360680408775806,
-0.1457086056470871,
-0.06137600541114807,
-0.09489979594945908,
0.11565322428941727,
0.08146517723798752,
0.03358805552124977,
0.04274565726518631,
0.019532648846507072,
-0.04414922371506691,
-0.038583990186452866,
0.12961317598819733,
0.08133101463317871,
0.012996876612305641,
0.01137041300535202,
0.01941833831369877,
-0.020302120596170425,
0.0028480992186814547,
-0.01250747125595808,
-0.07239153981208801,
-0.05874783173203468,
0.09400010108947754,
0.1600283533334732,
-0.06127211079001427,
-0.13325586915016174,
-0.020593497902154922,
0.04988488554954529,
0.0014717020094394684,
-0.08777432143688202,
0.04833676666021347,
0.15805292129516602,
-0.05623878911137581,
0.03216489031910896,
-0.09984751045703888,
-0.07263360917568207,
-0.16060975193977356,
-0.10029061883687973,
-0.06092562898993492,
-0.28350353240966797,
0.09752398729324341,
0.006392303854227066,
-0.014731393195688725,
0.059529416263103485,
0.051305368542671204,
-0.052508849650621414,
0.07068239152431488,
-0.18146829307079315,
-0.007054794579744339,
0.03497592359781265,
-0.13212306797504425,
0.02475893869996071,
-0.2378365397453308,
0.10198072344064713,
-0.04623803123831749,
-0.1519704908132553,
-0.04004510119557381,
0.0641569048166275,
-0.09540136158466339,
-0.01822364516556263,
-0.0475153923034668,
-0.01922670193016529,
0.01624443754553795,
-0.009348669089376926,
-0.031147832050919533,
0.13716529309749603,
0.02827494591474533,
-0.03268734738230705,
0.005254602525383234,
0.0223685409873724,
0.03955082967877388,
-0.0969657450914383,
-0.05986930429935455,
0.08311155438423157,
-0.031056145206093788,
0.14728976786136627,
0.000341245875461027,
0.04181376099586487,
-0.06758682429790497,
0.2593761384487152,
0.2023983597755432,
-0.12479214370250702,
0.008118697442114353,
-0.021801479160785675,
0.012670028023421764,
-0.041751839220523834,
0.13110700249671936,
0.013386172242462635,
0.12186761200428009,
-0.17513342201709747,
-0.01036517322063446,
-0.0818324014544487,
-0.04501292482018471,
0.06702108681201935,
0.14714950323104858,
0.15742522478103638,
0.03436789661645889,
-0.07328428328037262,
0.06722653657197952,
-0.30119743943214417,
0.20540550351142883,
-0.1346001923084259,
-0.01498429011553526,
-0.040251150727272034,
-0.058389630168676376,
0.061147745698690414,
0.11309876292943954,
0.10832664370536804,
-0.021150551736354828,
-0.0905047357082367,
-0.04486766457557678,
-0.039378076791763306,
-0.13019338250160217,
-0.02718670479953289,
0.1654091775417328,
0.06799814850091934,
0.31520840525627136,
-0.017577875405550003,
0.07702425122261047,
0.034410297870635986,
0.06451138854026794,
0.004519328009337187,
0.09537279605865479,
0.07960964739322662,
-0.06345855444669724,
-0.07373003661632538,
-0.001637450186535716,
0.05033271387219429,
0.14567798376083374,
-0.03826142102479935,
-0.18691548705101013,
0.15858715772628784,
0.07192251086235046,
-0.13762691617012024,
-0.05777517706155777,
0.08409425616264343,
-0.0739973932504654,
0.0550808347761631,
0.08115427941083908,
0.015876613557338715,
-0.017793258652091026,
-0.004664506763219833,
0.06074233725667,
0.024694660678505898,
-0.02343848906457424,
0.003570882137864828,
-0.08337053656578064,
-0.04151543974876404,
0.07267895340919495,
-0.0844460055232048,
-0.20546193420886993,
-0.0957019031047821,
-0.07551700621843338,
0.030557552352547646,
-0.0649830624461174,
0.12575586140155792,
0.1717868149280548,
0.0593598335981369,
-0.03307248651981354,
-0.10721943527460098,
-0.035562749952077866,
0.07602505385875702,
-0.044773899018764496,
-0.09409699589014053
] |
null | null | null |
# **Reinforce** Agent playing **CartPole-v1**
This is a trained model of a **Reinforce** agent playing **CartPole-v1** .
To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: https://huggingface.co/deep-rl-course/unit4/introduction
| {"tags": ["CartPole-v1", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class"], "model-index": [{"name": "Reinforce-Cartpole-v1", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "CartPole-v1", "type": "CartPole-v1"}, "metrics": [{"type": "mean_reward", "value": "500.00 +/- 0.00", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | magixn/Reinforce-Cartpole-v1 | [
"CartPole-v1",
"reinforce",
"reinforcement-learning",
"custom-implementation",
"deep-rl-class",
"model-index",
"region:us"
] | 2024-02-07T07:27:02+00:00 | [] | [] | TAGS
#CartPole-v1 #reinforce #reinforcement-learning #custom-implementation #deep-rl-class #model-index #region-us
|
# Reinforce Agent playing CartPole-v1
This is a trained model of a Reinforce agent playing CartPole-v1 .
To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL
| [
"# Reinforce Agent playing CartPole-v1\n This is a trained model of a Reinforce agent playing CartPole-v1 .\n To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL"
] | [
"TAGS\n#CartPole-v1 #reinforce #reinforcement-learning #custom-implementation #deep-rl-class #model-index #region-us \n",
"# Reinforce Agent playing CartPole-v1\n This is a trained model of a Reinforce agent playing CartPole-v1 .\n To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL"
] | [
39,
54
] | [
"passage: TAGS\n#CartPole-v1 #reinforce #reinforcement-learning #custom-implementation #deep-rl-class #model-index #region-us \n# Reinforce Agent playing CartPole-v1\n This is a trained model of a Reinforce agent playing CartPole-v1 .\n To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL"
] | [
0.007526164408773184,
-0.12498430907726288,
-0.0013541718944907188,
0.09601131081581116,
0.11848696321249008,
-0.04186001420021057,
0.11405468732118607,
0.05624859035015106,
0.09539441019296646,
0.04239490255713463,
0.13636724650859833,
0.06906966865062714,
-0.004102868959307671,
0.12412862479686737,
0.09840741008520126,
-0.26058563590049744,
0.07420794665813446,
-0.04403980076313019,
-0.009944677352905273,
0.10139261186122894,
0.07836852967739105,
-0.08325441926717758,
0.051592715084552765,
0.00009572553972247988,
-0.044259943068027496,
0.0321260429918766,
0.013628939166665077,
-0.053157225251197815,
0.1606452465057373,
-0.07313758134841919,
0.10494591295719147,
-0.03843724727630615,
0.14574295282363892,
-0.1126825287938118,
0.04758213832974434,
0.05111503228545189,
-0.04548581689596176,
0.03848232328891754,
-0.12538743019104004,
-0.06033875793218613,
0.026815801858901978,
-0.015865681692957878,
0.12249194830656052,
0.03647647053003311,
-0.1777559220790863,
-0.13461355865001678,
-0.0165896974503994,
0.12325166910886765,
0.1627800315618515,
0.00512364786118269,
0.014270431362092495,
0.16791965067386627,
-0.1761058121919632,
0.025937072932720184,
0.11400806158781052,
-0.37275227904319763,
-0.00034436015994288027,
0.2240462601184845,
0.06164427846670151,
0.1252165287733078,
-0.12646614015102386,
0.010440526530146599,
0.07403992861509323,
0.04368630796670914,
0.049784936010837555,
-0.015430688858032227,
-0.12260042130947113,
0.08455035835504532,
-0.1383819431066513,
-0.058066487312316895,
0.1495426446199417,
-0.019741326570510864,
-0.009476418606936932,
-0.016515808179974556,
-0.009238536469638348,
-0.050979889929294586,
-0.03430935740470886,
-0.11778499186038971,
0.10755524039268494,
0.04975730925798416,
0.0038771627005189657,
-0.04602450504899025,
-0.05612579360604286,
-0.09815777093172073,
-0.03123871050775051,
0.0372777059674263,
-0.013706400990486145,
0.01091629359871149,
0.027692900970578194,
0.09935613721609116,
-0.13446329534053802,
0.01825822703540325,
-0.028096558526158333,
-0.028040969744324684,
-0.1316804438829422,
-0.11984307318925858,
-0.026084421202540398,
0.004223645199090242,
0.03029833547770977,
0.20433813333511353,
0.020139509811997414,
0.059011414647102356,
-0.0022708347532898188,
0.09776382148265839,
0.029780851677060127,
0.13517548143863678,
-0.04466623440384865,
0.19488364458084106,
0.07711011171340942,
0.05364556983113289,
0.03204274922609329,
-0.05344729498028755,
-0.19369827210903168,
0.04861246794462204,
0.06659778952598572,
0.08274952322244644,
-0.1178959533572197,
0.0059632807970047,
-0.10316018015146255,
0.0028950648847967386,
-0.10474003106355667,
-0.0642905905842781,
-0.02892979420721531,
0.031841445714235306,
-0.10535725951194763,
0.028785312548279762,
0.025052599608898163,
0.04140377417206764,
0.0676041767001152,
-0.12253966927528381,
-0.07404746115207672,
-0.021733485162258148,
-0.12817098200321198,
-0.09923440217971802,
0.08802318572998047,
-0.026199497282505035,
-0.005110981408506632,
-0.1253623217344284,
-0.2661486268043518,
-0.05670225992798805,
0.06396034359931946,
-0.03231031447649002,
-0.08589376509189606,
-0.1633463054895401,
0.026403428986668587,
-0.07700273394584656,
0.05221332609653473,
0.04776721075177193,
-0.03665859252214432,
0.02023705095052719,
-0.07958202809095383,
0.12739010155200958,
0.049698662012815475,
0.00541001046076417,
-0.09916839748620987,
0.07882837951183319,
-0.3034103214740753,
-0.02581131085753441,
-0.15228183567523956,
0.0772043839097023,
-0.07893010973930359,
0.01308529730886221,
0.05044940114021301,
0.043790437281131744,
-0.016942394897341728,
0.16269747912883759,
-0.17043575644493103,
-0.05301272124052048,
0.026445282623171806,
-0.09261117875576019,
-0.09916394203901291,
0.07275339215993881,
-0.06339669227600098,
0.21263530850410461,
0.08751397579908371,
0.17006252706050873,
-0.011036526411771774,
-0.16256992518901825,
0.1207515075802803,
0.07522942125797272,
-0.1639646589756012,
0.004287737421691418,
0.061784300953149796,
-0.0016935690073296428,
0.02746843732893467,
-0.01872866041958332,
-0.07289361208677292,
0.06302516162395477,
-0.07825060933828354,
0.022581040859222412,
0.06258945167064667,
-0.09531243145465851,
0.23986859619617462,
-0.005434412509202957,
0.0862451046705246,
-0.025957979261875153,
-0.09802921861410141,
0.00908072479069233,
0.07164718210697174,
-0.0014321404742076993,
0.01703714393079281,
-0.14553219079971313,
0.23044352233409882,
-0.07965081930160522,
0.011176814325153828,
-0.11607582122087479,
-0.1256982982158661,
0.011873425915837288,
0.13336114585399628,
0.059921663254499435,
0.16569606959819794,
0.09518871456384659,
-0.032197169959545135,
0.017584815621376038,
-0.0023385772947221994,
-0.09040450304746628,
0.01580043137073517,
-0.0021571461111307144,
-0.12167251110076904,
-0.07353103160858154,
-0.08134473115205765,
0.12585052847862244,
-0.20988115668296814,
0.015492538921535015,
0.04099845886230469,
0.008103687316179276,
0.04467369243502617,
0.023746047168970108,
-0.013269703835248947,
-0.00007021807687124237,
0.03244573250412941,
-0.10098352283239365,
0.12937165796756744,
0.013381263241171837,
0.014676140621304512,
-0.006365173030644655,
-0.05572463944554329,
0.03720450773835182,
0.040439579635858536,
-0.11237845569849014,
-0.11330515146255493,
-0.009658765979111195,
-0.0015364213613793254,
0.02637762948870659,
-0.022321155294775963,
0.052120618522167206,
0.27587956190109253,
0.05387469753623009,
0.10401033610105515,
-0.05769326910376549,
0.015315087512135506,
-0.015322818420827389,
-0.07135670632123947,
0.06358719617128372,
0.025013601407408714,
0.08050397783517838,
-0.03531401976943016,
0.03759452700614929,
0.1675453782081604,
-0.015888912603259087,
0.11127935349941254,
-0.06545067578554153,
-0.03844274953007698,
-0.043109722435474396,
0.05627678707242012,
0.015021559782326221,
0.04564907029271126,
0.0000015355876712419558,
-0.08444724231958389,
-0.03503387048840523,
-0.03988509997725487,
-0.010637006722390652,
-0.12273643165826797,
-0.00499896751716733,
0.01265440508723259,
-0.021940499544143677,
0.04488934203982353,
0.07375624030828476,
-0.04849626496434212,
0.025821007788181305,
0.06070821359753609,
-0.10193055868148804,
0.08957115560770035,
0.015067169442772865,
-0.06946801394224167,
0.13769419491291046,
-0.07484805583953857,
-0.045293889939785004,
-0.1025395318865776,
-0.1568877100944519,
0.09384927153587341,
0.06704871356487274,
-0.05427970737218857,
-0.1503879576921463,
-0.0016851738328114152,
-0.008973666466772556,
0.09206123650074005,
-0.006399387493729591,
-0.12621140480041504,
0.01989075168967247,
0.08295059949159622,
-0.05633419007062912,
-0.09804849326610565,
-0.0075809285044670105,
-0.05280788615345955,
-0.17707788944244385,
-0.03888550028204918,
-0.06398582458496094,
-0.06734282523393631,
0.23586803674697876,
0.02017230913043022,
0.08274748176336288,
-0.044721852988004684,
0.04250151664018631,
-0.012231717817485332,
0.0006326579605229199,
0.10689259320497513,
-0.09043551236391068,
-0.017900818958878517,
-0.001320177922025323,
-0.024820495396852493,
-0.07327181100845337,
0.029733488336205482,
-0.04272191599011421,
-0.08249637484550476,
-0.1415451467037201,
-0.04993678629398346,
-0.011005163192749023,
0.10754310339689255,
0.07337497919797897,
0.0048001972027122974,
-0.11733713001012802,
0.062058478593826294,
0.13692134618759155,
0.031207585707306862,
0.004062763415277004,
0.028157465159893036,
0.14977529644966125,
-0.10706274956464767,
-0.022463621571660042,
-0.038119975477457047,
-0.054863203316926956,
0.004114252515137196,
0.016883620992302895,
0.08840765058994293,
0.1410384476184845,
0.11468084901571274,
0.047563645988702774,
0.0464191697537899,
0.06561273336410522,
0.1694946140050888,
0.059157438576221466,
-0.10448314249515533,
-0.044678982347249985,
-0.0040070898830890656,
-0.10903503000736237,
0.057307638227939606,
0.16030821204185486,
0.06326017528772354,
-0.14463356137275696,
0.021787412464618683,
-0.038982175290584564,
0.13649246096611023,
0.020638149231672287,
-0.2677258849143982,
-0.008139112964272499,
0.023630544543266296,
-0.0010347915813326836,
-0.012379839085042477,
0.10821118950843811,
-0.040134772658348083,
-0.233198344707489,
-0.12299054861068726,
0.010077533312141895,
0.031144635751843452,
-0.1509784311056137,
0.015542911365628242,
-0.14036494493484497,
0.08027976751327515,
-0.007007129956036806,
0.07418135553598404,
-0.025149788707494736,
0.15060245990753174,
-0.028731435537338257,
0.01628703810274601,
-0.07902143895626068,
-0.047717493027448654,
0.09898673743009567,
-0.0046631391160190105,
0.1931537538766861,
0.005480166990309954,
-0.023713182657957077,
-0.12098433077335358,
-0.05229806900024414,
-0.04967813938856125,
0.010598190128803253,
-0.05373382940888405,
0.0765683576464653,
-0.02441473677754402,
-0.0039579677395522594,
-0.010900177992880344,
0.08942947536706924,
-0.05291692912578583,
0.03636563941836357,
-0.11246588081121445,
-0.05034820735454559,
0.14550213515758514,
-0.09163831174373627,
-0.10174685716629028,
-0.16205860674381256,
0.14137998223304749,
0.15070600807666779,
0.058216437697410583,
-0.04001476243138313,
0.03867831453680992,
-0.019183965399861336,
-0.024241572245955467,
0.07880574464797974,
0.009653856977820396,
0.1324782371520996,
-0.08983246237039566,
0.014327390119433403,
0.14589735865592957,
-0.05275948345661163,
0.016191845759749413,
-0.02304735779762268,
0.12202176451683044,
0.04650457948446274,
0.06189403310418129,
0.018547222018241882,
0.06655703485012054,
0.06466961652040482,
-0.02262885868549347,
0.08456692099571228,
0.030712679028511047,
-0.18644161522388458,
0.058530256152153015,
-0.09805119782686234,
0.22581584751605988,
0.05066308751702309,
0.06047345697879791,
0.2993181645870209,
0.21986234188079834,
-0.05372472479939461,
0.1669820249080658,
0.044286344200372696,
-0.05891284719109535,
-0.21245966851711273,
-0.03684934973716736,
-0.030655447393655777,
0.09436552971601486,
0.15607263147830963,
-0.0981721356511116,
-0.04201313853263855,
-0.00972361396998167,
-0.032264553010463715,
0.020120708271861076,
-0.24663487076759338,
-0.01734781451523304,
0.14379777014255524,
0.10629188269376755,
0.2451348900794983,
-0.006132842972874641,
0.023609744384884834,
0.049030207097530365,
0.018605992197990417,
-0.02483358606696129,
-0.21013511717319489,
0.09079083055257797,
0.006071676965802908,
0.04935038834810257,
0.022885039448738098,
-0.006052911281585693,
0.04500092566013336,
-0.073696069419384,
0.08904470503330231,
-0.08561883866786957,
-0.08341272175312042,
0.2185351401567459,
-0.03945168852806091,
-0.00661163916811347,
0.12917985022068024,
-0.011526807211339474,
-0.1097102016210556,
-0.015364703722298145,
0.027403371408581734,
0.030678823590278625,
-0.030246863141655922,
-0.03609466925263405,
0.024012766778469086,
0.10202405601739883,
-0.04282205551862717,
0.04565315693616867,
0.10240072011947632,
-0.020902957767248154,
0.15945613384246826,
0.13205459713935852,
0.10420060157775879,
0.002927543595433235,
-0.06464727967977524,
0.014349685050547123,
-0.055471502244472504,
0.02962767891585827,
-0.17038846015930176,
-0.0070191239938139915,
0.055695805698633194,
0.04772466421127319,
0.0945243164896965,
0.11333164572715759,
-0.127106174826622,
0.0300484336912632,
0.028996523469686508,
-0.06286120414733887,
-0.06029998138546944,
-0.002275418024510145,
-0.016458535566926003,
-0.008173024281859398,
-0.09947093576192856,
0.07884971052408218,
-0.10555081814527512,
-0.03306307643651962,
0.05025126785039902,
-0.0607193186879158,
-0.12852220237255096,
-0.010904680006206036,
0.1252979338169098,
0.061709314584732056,
-0.05078592896461487,
0.14939077198505402,
0.06109785661101341,
-0.08055379986763,
0.037185851484537125,
0.027442200109362602,
-0.08008874952793121,
-0.10198270529508591,
-0.0004569833690766245,
0.31761088967323303,
0.06076094135642052,
-0.0329466350376606,
-0.11946453154087067,
-0.15002015233039856,
0.04840146750211716,
0.1035679280757904,
0.12359631806612015,
0.011757869273424149,
-0.05322748050093651,
0.02236519381403923,
-0.05275069922208786,
0.03814244270324707,
0.06910209357738495,
-0.03928454965353012,
-0.13761694729328156,
0.0077122850343585014,
0.026647454127669334,
0.10174071043729782,
-0.06771174818277359,
-0.09184598177671432,
-0.18085066974163055,
0.09208621084690094,
-0.03432070091366768,
-0.10890032351016998,
0.027215104550123215,
-0.017406610772013664,
0.014248576015233994,
0.07639352232217789,
-0.047281619161367416,
0.01244808267802,
-0.1517520695924759,
0.07082249224185944,
0.05706808716058731,
0.08926787972450256,
0.000014311663107946515,
-0.054843269288539886,
0.07618319988250732,
-0.05763502046465874,
0.06680037826299667,
-0.053477559238672256,
0.005539732985198498,
0.10781200975179672,
-0.23264040052890778,
-0.021164139732718468,
0.009476077742874622,
-0.04681631922721863,
0.08765807747840881,
-0.19047698378562927,
0.024190550670027733,
-0.08897756040096283,
-0.024605726823210716,
0.01802127994596958,
-0.1086471825838089,
-0.04306677728891373,
0.08475461602210999,
0.037119291722774506,
-0.031288959085941315,
-0.04612116143107414,
-0.019314980134367943,
-0.0914498046040535,
0.053634315729141235,
0.07442525774240494,
-0.0687926784157753,
0.08314394950866699,
-0.05507456883788109,
0.00841207429766655,
-0.052043743431568146,
0.06760627031326294,
-0.012366239912807941,
-0.12672528624534607,
-0.02123171091079712,
-0.044928714632987976,
0.11662110686302185,
-0.023402327671647072,
0.022080281749367714,
0.014599837362766266,
0.0323631577193737,
-0.012065601535141468,
0.05028461292386055,
0.1019197478890419,
0.05136820673942566,
0.014879679307341576,
0.02292765863239765,
0.055746350437402725,
0.0757644772529602,
-0.1134679913520813,
0.06457309424877167,
-0.02098844014108181,
-0.08620109409093857,
0.1013324111700058,
0.06909440457820892,
0.037490107119083405,
0.15593400597572327,
0.22674402594566345,
0.10539932548999786,
-0.03564648702740669,
-0.03126971051096916,
0.12967991828918457,
0.17799612879753113,
-0.07682197540998459,
0.015780627727508545,
-0.0020607721526175737,
-0.017265556380152702,
-0.09849067777395248,
-0.13722245395183563,
-0.060460351407527924,
-0.2453264594078064,
0.1078341007232666,
-0.03288164362311363,
-0.04169659689068794,
0.128489688038826,
0.027952738106250763,
0.03724630922079086,
0.08183616399765015,
-0.12909026443958282,
-0.013460557907819748,
0.07749562710523605,
-0.08914026618003845,
-0.033571500331163406,
-0.17521262168884277,
-0.06771576404571533,
-0.08741120994091034,
-0.15989220142364502,
-0.06844990700483322,
0.029948782175779343,
0.035394806414842606,
0.010386589914560318,
-0.039711855351924896,
-0.01962728053331375,
0.011063394136726856,
-0.0025537724141031504,
-0.04985455423593521,
-0.01753084547817707,
0.021317757666110992,
-0.11333847790956497,
-0.024336790665984154,
0.16320326924324036,
-0.03297848999500275,
-0.18396754562854767,
-0.0405106395483017,
0.2157316505908966,
0.025046708062291145,
0.0590171180665493,
-0.073721744120121,
-0.016323629766702652,
0.021523483097553253,
0.20813441276550293,
0.10171995311975479,
-0.10821312665939331,
0.015457749366760254,
-0.03655189648270607,
0.0013793212128803134,
-0.061893612146377563,
0.10775819420814514,
0.06519263982772827,
-0.07549984753131866,
-0.17567221820354462,
-0.04389495030045509,
-0.08628730475902557,
0.03370477631688118,
-0.14383791387081146,
-0.03786516562104225,
0.1168690100312233,
0.004516853019595146,
-0.053927481174468994,
0.07883694022893906,
-0.17713546752929688,
0.03441957011818886,
-0.04880853369832039,
-0.13215437531471252,
-0.09491758048534393,
-0.10123858600854874,
0.0027463934384286404,
0.08913854509592056,
0.15567956864833832,
-0.06151591241359711,
-0.07471925020217896,
-0.009579092264175415,
-0.028091613203287125,
-0.052700337022542953,
-0.07900123298168182,
0.059512585401535034,
0.0007560851518064737,
0.16147300601005554,
-0.07439453154802322,
0.09558981657028198,
0.09099138528108597,
-0.021246420219540596,
-0.00915549136698246,
0.032866667956113815,
-0.003863809397444129,
-0.07436864078044891,
-0.04970616102218628,
0.02312966249883175,
0.027639856562018394,
0.10846075415611267,
-0.030836544930934906,
-0.1934703141450882,
0.11230092495679855,
0.09140218049287796,
-0.04296138137578964,
-0.046487610787153244,
0.05351927503943443,
-0.07097935676574707,
0.1252279132604599,
0.03444884717464447,
-0.02163051813840866,
0.013762647286057472,
-0.06370721012353897,
0.08370721340179443,
0.11594565212726593,
-0.048265840858221054,
-0.08278503268957138,
-0.06164652109146118,
0.012770666740834713,
0.02961382456123829,
-0.13650155067443848,
-0.21160630881786346,
-0.10802312940359116,
-0.1383298933506012,
0.004740108735859394,
-0.04703504592180252,
0.08498300611972809,
0.12991970777511597,
0.09780163317918777,
-0.011416295543313026,
-0.004867587238550186,
0.018085451796650887,
0.13192623853683472,
-0.11232008039951324,
-0.08192373812198639
] |
null | null | null | # *SeaLLM-7B-v2* - Large Language Models for Southeast Asia
**NOTE: download [seallm.preset.json](https://huggingface.co/SeaLLMs/SeaLLM-7B-v2-gguf/blob/main/seallm.preset.json) for LM-studio to work correctly,**
<p align="center">
<a href="https://huggingface.co/SeaLLMs/SeaLLM-7B-v2" target="_blank" rel="noopener"> 🤗 Tech Memo</a>
<a href="https://huggingface.co/spaces/SeaLLMs/SeaLLM-7B" target="_blank" rel="noopener"> 🤗 DEMO</a>
<a href="https://github.com/DAMO-NLP-SG/SeaLLMs" target="_blank" rel="noopener">Github</a>
<a href="https://arxiv.org/pdf/2312.00738.pdf" target="_blank" rel="noopener">Technical Report</a>
</p>
We introduce [SeaLLM-7B-v2](https://huggingface.co/SeaLLMs/SeaLLM-7B-v2), the state-of-the-art multilingual LLM for Southeast Asian (SEA) languages 🇬🇧 🇨🇳 🇻🇳 🇮🇩 🇹🇭 🇲🇾 🇰🇭 🇱🇦 🇲🇲 🇵🇭. It is the most significant upgrade since [SeaLLM-13B](https://huggingface.co/SeaLLMs/SeaLLM-13B-Chat), with half the size, outperforming performance across diverse multilingual tasks, from world knowledge, math reasoning, instruction following, etc.
### Highlights
* [SeaLLM-7B-v2](https://huggingface.co/SeaLLMs/SeaLLM-7B-v2) achieves the **7B-SOTA** on the **Zero-shot CoT GSM8K** task with **78.2** score and outperforms GPT-3.5 in many GSM8K-translated tasks in SEA languages (🇨🇳 🇻🇳 🇮🇩 🇹🇭) as well as MGSM (🇨🇳 🇹🇭). It also surpasses GPT-3.5 in MATH CoT for Thai 🇹🇭.
* It scores competitively against GPT-3.5 in many zero-shot CoT commonsense benchmark, with **82.5, 68.3, 80.9** scores on Arc-C, Winogrande, and Hellaswag.
* It achieves **7.54** score on the 🇬🇧 **MT-bench**, it ranks 3rd place on the leaderboard for 7B category and is the most outperforming multilingual model.
* It scores **45.74** on the VMLU benchmark for Vietnamese 🇻🇳, and is the only open-source multilingual model that can be competitive to monolingual models ([Vistral-7B](https://huggingface.co/Viet-Mistral/Vistral-7B-Chat)) of similar sizes.
### Release and DEMO
- DEMO: [SeaLLMs/SeaLLM-7B](https://huggingface.co/spaces/SeaLLMs/SeaLLM-7B).
- Technical report: [Arxiv: SeaLLMs - Large Language Models for Southeast Asia](https://arxiv.org/pdf/2312.00738.pdf).
- Model weights:
- [SeaLLM-7B-v2](https://huggingface.co/SeaLLMs/SeaLLM-7B-v2).
- [SeaLLM-7B-v2-gguf](https://huggingface.co/SeaLLMs/SeaLLM-7B-v2-gguf).
- [SeaLLM-7B-v2-q4_0](https://huggingface.co/SeaLLMs/SeaLLM-7B-v2-gguf/blob/main/SeaLLM-7B-v2.q4_0.gguf), [SeaLLM-7B-v2-q8_0](https://huggingface.co/SeaLLMs/SeaLLM-7B-v2-gguf/blob/main/SeaLLM-7B-v2.q8_0.gguf).
- LM-studio requires this [seallm.preset.json](https://huggingface.co/SeaLLMs/SeaLLM-7B-v2-gguf/blob/main/seallm.preset.json) to work properly.
- [SeaLLM-7B-v2-GGUF (thanks Lonestriker)](https://huggingface.co/LoneStriker/SeaLLM-7B-v2-GGUF). NOTE: use [seallm.preset.json](https://huggingface.co/SeaLLMs/SeaLLM-7B-v2-gguf/blob/main/seallm.preset.json) to work properly.
<blockquote style="color:red">
<p><strong style="color: red">Terms of Use and License</strong>:
By using our released weights, codes, and demos, you agree to and comply with the terms and conditions specified in our <a href="https://huggingface.co/SeaLLMs/SeaLLM-Chat-13b/edit/main/LICENSE" target="_blank" rel="noopener">SeaLLMs Terms Of Use</a>.
</blockquote>
> **Disclaimer**:
> We must note that even though the weights, codes, and demos are released in an open manner, similar to other pre-trained language models, and despite our best efforts in red teaming and safety fine-tuning and enforcement, our models come with potential risks, including but not limited to inaccurate, misleading or potentially harmful generation.
> Developers and stakeholders should perform their own red teaming and provide related security measures before deployment, and they must abide by and comply with local governance and regulations.
> In no event shall the authors be held liable for any claim, damages, or other liability arising from the use of the released weights, codes, or demos.
> The logo was generated by DALL-E 3.
### What's new since SeaLLM-13B-v1 and SeaLLM-7B-v1?
* SeaLLM-7B-v2 is continue-pretrained from [Mistral-7B](https://huggingface.co/mistralai/Mistral-7B-v0.1) and underwent carefully designed tuning with focus in reasoning.
## Evaluation
### Zero-shot CoT Multilingual Math Reasoning
[SeaLLM-7B-v2](https://huggingface.co/SeaLLMs/SeaLLM-7B-v2) achieves with **78.2** score on the GSM8K with zero-shot CoT reasoning, making it the **state of the art** in the realm of 7B models. It also outperforms GPT-3.5 in the same GSM8K benchmark as translated into SEA languages (🇨🇳 🇻🇳 🇮🇩 🇹🇭). [SeaLLM-7B-v2](https://huggingface.co/SeaLLMs/SeaLLM-7B-v2) also surpasses GPT-3.5 on the Thai-translated MATH benchmark, with **22.4** vs 18.1 scores.

<details>
<summary>See details on English and translated GSM8K and MATH with zero-shot reasoning</summary>
<br>
| Model | GSM8K<br>en | MATH<br>en | GSM8K<br>zh | MATH<br>zh | GSM8K<br>vi | MATH<br>vi | GSM8K<br>id | MATH<br>id | GSM8K<br>th | MATH<br>th
| --- | --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
| GPT-3.5 | 80.8 | 34.1 | 48.2 | 21.5 | 55 | 26.5 | 64.3 | 26.4 | 35.8 | 18.1
| Qwen-14B-chat | 61.4 | 18.4 | 41.6 | 11.8 | 33.6 | 3.6 | 44.7 | 8.6 | 22 | 6
| Vistral-7b-chat | 48.2 | 12.5 | | | 48.7 | 3.1 | | | |
| SeaLLM-7B-v2 | 78.2 | 27.5 | 53.7 | 17.6 | 69.9 | 23.8 | 71.5 | 24.4 | 59.6 | 22.4
</details>
#### Zero-shot MGSM
[SeaLLM-7B-v2](https://huggingface.co/SeaLLMs/SeaLLM-7B-v2) also outperforms GPT-3.5 and Qwen-14B on the multilingual MGSM for Zh and Th.
| Model | MGSM-Zh | MGSM-Th
|-----| ----- | ---
| ChatGPT (reported) | 61.2* | 47.2*
| Qwen-14B-chat | 59.6 | 28
| SeaLLM-7B-v2 | **64.8** | **62.4**
### Zero-shot Commonsense Reasoning
We compare [SeaLLM-7B-v2](https://huggingface.co/SeaLLMs/SeaLLM-7B-v2) with ChatGPT and Mistral-7B-instruct on various zero-shot commonsense benchmarks (Arc-Challenge, Winogrande and Hellaswag). We use the 2-stage technique in [(Kojima et al., 2023)](https://arxiv.org/pdf/2205.11916.pdf) to grab the answer. Note that we **DID NOT** use "Let's think step-by-step" to invoke explicit CoT.
| Model | Arc-Challenge | Winogrande | Hellaswag
|-----| ----- | --- | -- |
| ChatGPT (reported) | 84.6* | 66.8* | 72.0*
| ChatGPT (reproduced) | 84.1 | 63.1 | 79.5
| Mistral-7B-Instruct | 68.1 | 56.4 | 45.6
| SeaLLM-7B-v2 | 82.5 | 68.3 | 80.9
### Multilingual World Knowledge
We evaluate models on 3 benchmarks following the recommended default setups: 5-shot MMLU for En, 3-shot [M3Exam](https://arxiv.org/pdf/2306.05179.pdf) (M3e) for En, Zh, Vi, Id, Th, and zero-shot [VMLU](https://vmlu.ai/) for Vi.
| Model | Langs | En<br>MMLU | En<br>M3e | Zh<br>M3e | Vi<br>M3e | Vi<br>VMLU | Id<br>M3e | Th<br>M3e
|-----| ----- | --- | -- | ----- | ---- | --- | --- | --- |
| GPT-3.5 | Multi | 68.90 | 75.46 | 60.20 | 58.64 | 46.32 | 49.27 | 37.41
| SeaLLM-13B | Multi | 52.78 | 62.69 | 44.50 | 46.45 | | 39.28 | 36.39
| Vistral-7B-chat | Mono | 56.86 | 67.00 | 44.56 | 54.33 | 50.03 | 36.49 | 25.27
| Qwen1.5-7B-chat | Multi | 61.00 | 52.07 | 81.96 | 43.38 | 45.02 | 24.29 | 20.25
| SeaLLM-7B-v2 | Multi | 61.89 | 70.91 | 55.43 | 51.15 | 45.74 | 42.25 | 35.52
VMLU reproduce script [here](https://github.com/DAMO-NLP-SG/SeaLLMs/blob/main/evaluation/vmlu/vmlu_run.py). Lm-eval was used to evaluate MMLU.
### MT-Bench
On the English [MT-bench](https://arxiv.org/abs/2306.05685) metric, SeaLLM-7B-v2 achieves **7.54** score on the MT-bench (3rd place on the leaderboard for 7B category), outperforms many 70B models and is arguably the only one that handles 10 SEA languages.
Refer to [mt_bench/seallm_7b_v2.jsonl](https://huggingface.co/SeaLLMs/SeaLLM-7B-v2/blob/main/evaluation/mt_bench/seallm_7b_v2.jsonl) for the MT-bench predictions of SeaLLM-7B-v2, and [here](https://github.com/lm-sys/FastChat/issues/3013#issue-2118685341) to reproduce it.
| Model | Access | Langs | MT-Bench
| --- | --- | --- | --- |
| GPT-4-turbo | closed | multi | 9.32
| GPT-4-0613 | closed | multi | 9.18
| Mixtral-8x7b (46B) | open | multi | 8.3
| Starling-LM-7B-alpha | open | mono (en) | 8.0
| OpenChat-3.5-7B | open | mono (en) | 7.81
| **SeaLLM-7B-v2** | **open** | **multi (10+)** | **7.54**
| [Qwen-14B](https://huggingface.co/Qwen/Qwen-14B-Chat) | open | multi | 6.96
| [Llama-2-70B](https://huggingface.co/meta-llama/Llama-2-70b-chat-hf) | open | mono (en) | 6.86
| Mistral-7B-instuct | open | mono (en) | 6.84
### Sea-Bench
Similar to MT-Bench, [Sea-bench](https://huggingface.co/datasets/SeaLLMs/Sea-bench) is a set of categorized instruction test sets to measure models' ability as an assistant that is specifically focused on 9 SEA languages, including non-Latin low-resource languages.
As shown, the huge improvements come from math-reasoning, reaching GPT-3.5 level of performance.

Refer to [sea_bench/seallm_7b_v2.jsonl](https://huggingface.co/SeaLLMs/SeaLLM-7B-v2/blob/main/evaluation/sea_bench/seallm_7b_v2.jsonl) for the Sea-bench predictions of SeaLLM-7B-v2.
### Usage
#### Instruction format
```python
prompt = """<|im_start|>system
You are a helpful assistant.</s><|im_start|>user
Hello world</s><|im_start|>assistant
Hi there, how can I help?</s>"""
# NOTE: previous commit has \n between </s> and <|im_start|>, that was incorrect!
# <|im_start|> is not a special token.
# Transformers chat_template should be consistent with vLLM format below.
# ! ENSURE 1 and only 1 bos `<s>` at the beginning of sequence
print(tokenizer.convert_ids_to_tokens(tokenizer.encode(prompt)))
'<s>', '▁<', '|', 'im', '_', 'start', '|', '>', 'system', '<0x0A>', 'You', '▁are', '▁a', '▁helpful', '▁assistant', '.', '</s>', '▁<', '|', 'im', '_', 'start', '|', '>', 'user', '<0x0A>', 'Hello', '▁world', '</s>', '▁<', '|', 'im', '_', 'start', '|', '>', 'ass', 'istant', '<0x0A>', 'Hi', '▁there', ',', '▁how', '▁can', '▁I', '▁help', '?', '</s>']
"""
```
#### Using transformers's chat_template
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
device = "cuda" # the device to load the model onto
# use bfloat16 to ensure the best performance.
model = AutoModelForCausalLM.from_pretrained("SeaLLMs/SeaLLM-7B-v2", torch_dtype=torch.bfloat16, device_map=device)
tokenizer = AutoTokenizer.from_pretrained("SeaLLMs/SeaLLM-7B-v2")
messages = [
{"role": "system", "content": "You are a helpful assistant."},
{"role": "user", "content": "Hello world"},
{"role": "assistant", "content": "Hi there, how can I help you today?"},
{"role": "user", "content": "Explain general relativity in details."}
]
encodeds = tokenizer.apply_chat_template(messages, return_tensors="pt", add_generation_prompt=True)
print(tokenizer.convert_ids_to_tokens(encodeds[0]))
# ['<s>', '▁<', '|', 'im', '_', 'start', '|', '>', 'system', '<0x0A>', 'You', '▁are', '▁a', '▁helpful', '▁assistant', '.', '</s>', '▁<', '|', 'im', '_', 'start', '|', '>', 'user', '<0x0A>', 'Hello', '▁world', '</s>', '▁<', '|', 'im', '_', 'start', '|', '>', 'ass', 'istant', '<0x0A>', 'Hi', '▁there', ',', '▁how', '▁can', '▁I', '▁help', '▁you', '▁today', '?', '</s>', '▁<', '|', 'im', '_', 'start', '|', '>', 'user', '<0x0A>', 'Ex', 'plain', '▁general', '▁rel', 'ativity', '▁in', '▁details', '.', '</s>', '▁<', '|', 'im', '_', 'start', '|', '>', 'ass', 'istant', '<0x0A>']
model_inputs = encodeds.to(device)
model.to(device)
generated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True, pad_token_id=tokenizer.pad_token_id)
decoded = tokenizer.batch_decode(generated_ids)
print(decoded[0])
```
#### Using vLLM
```python
from vllm import LLM, SamplingParams
TURN_TEMPLATE = "<|im_start|>{role}\n{content}</s>"
TURN_PREFIX = "<|im_start|>{role}\n"
# There is no \n between </s> and <|im_start|>.
def seallm_chat_convo_format(conversations, add_assistant_prefix: bool, system_prompt=None):
# conversations: list of dict with key `role` and `content` (openai format)
if conversations[0]['role'] != 'system' and system_prompt is not None:
conversations = [{"role": "system", "content": system_prompt}] + conversations
text = ''
for turn_id, turn in enumerate(conversations):
prompt = TURN_TEMPLATE.format(role=turn['role'], content=turn['content'])
text += prompt
if add_assistant_prefix:
prompt = TURN_PREFIX.format(role='assistant')
text += prompt
return text
sparams = SamplingParams(temperature=0.1, max_tokens=1024, stop=['</s>', '<|im_start|>'])
llm = LLM("SeaLLMs/SeaLLM-7B-v2", dtype="bfloat16")
message = "Explain general relativity in details."
prompt = seallm_chat_convo_format(message, True)
gen = llm.generate(prompt, sampling_params)
print(gen[0].outputs[0].text)
```
## Acknowledgement to Our Linguists
We would like to express our special thanks to our professional and native linguists, Tantong Champaiboon, Nguyen Ngoc Yen Nhi and Tara Devina Putri, who helped build, evaluate, and fact-check our sampled pretraining and SFT dataset as well as evaluating our models across different aspects, especially safety.
## Citation
If you find our project useful, we hope you would kindly star our repo and cite our work as follows: Corresponding Author: [[email protected]](mailto:[email protected])
**Author list and order will change!**
* `*` and `^` are equal contributions.
```
@article{damonlpsg2023seallm,
author = {Xuan-Phi Nguyen*, Wenxuan Zhang*, Xin Li*, Mahani Aljunied*,
Zhiqiang Hu, Chenhui Shen^, Yew Ken Chia^, Xingxuan Li, Jianyu Wang,
Qingyu Tan, Liying Cheng, Guanzheng Chen, Yue Deng, Sen Yang,
Chaoqun Liu, Hang Zhang, Lidong Bing},
title = {SeaLLMs - Large Language Models for Southeast Asia},
year = 2023,
Eprint = {arXiv:2312.00738},
}
```
| {"license": "other", "license_name": "seallms", "license_link": "https://huggingface.co/SeaLLMs/SeaLLM-13B-Chat/blob/main/LICENSE"} | null | SeaLLMs/SeaLLM-7B-v2-gguf | [
"gguf",
"arxiv:2312.00738",
"arxiv:2205.11916",
"arxiv:2306.05179",
"arxiv:2306.05685",
"license:other",
"region:us"
] | 2024-02-07T07:34:05+00:00 | [
"2312.00738",
"2205.11916",
"2306.05179",
"2306.05685"
] | [] | TAGS
#gguf #arxiv-2312.00738 #arxiv-2205.11916 #arxiv-2306.05179 #arxiv-2306.05685 #license-other #region-us
| *SeaLLM-7B-v2* - Large Language Models for Southeast Asia
=========================================================
NOTE: download URL for LM-studio to work correctly,
[Tech Memo](URL target=)
[DEMO](URL target=)
[Github](URL target=)
[Technical Report](URL target=)
We introduce SeaLLM-7B-v2, the state-of-the-art multilingual LLM for Southeast Asian (SEA) languages 🇬🇧 🇨🇳 🇻🇳 🇮🇩 🇹🇭 🇲🇾 🇰🇭 🇱🇦 🇲🇲 🇵🇭. It is the most significant upgrade since SeaLLM-13B, with half the size, outperforming performance across diverse multilingual tasks, from world knowledge, math reasoning, instruction following, etc.
### Highlights
* SeaLLM-7B-v2 achieves the 7B-SOTA on the Zero-shot CoT GSM8K task with 78.2 score and outperforms GPT-3.5 in many GSM8K-translated tasks in SEA languages (🇨🇳 🇻🇳 🇮🇩 🇹🇭) as well as MGSM (🇨🇳 🇹🇭). It also surpasses GPT-3.5 in MATH CoT for Thai 🇹🇭.
* It scores competitively against GPT-3.5 in many zero-shot CoT commonsense benchmark, with 82.5, 68.3, 80.9 scores on Arc-C, Winogrande, and Hellaswag.
* It achieves 7.54 score on the 🇬🇧 MT-bench, it ranks 3rd place on the leaderboard for 7B category and is the most outperforming multilingual model.
* It scores 45.74 on the VMLU benchmark for Vietnamese 🇻🇳, and is the only open-source multilingual model that can be competitive to monolingual models (Vistral-7B) of similar sizes.
### Release and DEMO
* DEMO: SeaLLMs/SeaLLM-7B.
* Technical report: Arxiv: SeaLLMs - Large Language Models for Southeast Asia.
* Model weights:
+ SeaLLM-7B-v2.
+ SeaLLM-7B-v2-gguf.
- SeaLLM-7B-v2-q4\_0, SeaLLM-7B-v2-q8\_0.
- LM-studio requires this URL to work properly.
+ SeaLLM-7B-v2-GGUF (thanks Lonestriker). NOTE: use URL to work properly.
>
> **Terms of Use and License**:
> By using our released weights, codes, and demos, you agree to and comply with the terms and conditions specified in our [SeaLLMs Terms Of Use](URL target=).
>
>
>
>
> Disclaimer:
> We must note that even though the weights, codes, and demos are released in an open manner, similar to other pre-trained language models, and despite our best efforts in red teaming and safety fine-tuning and enforcement, our models come with potential risks, including but not limited to inaccurate, misleading or potentially harmful generation.
> Developers and stakeholders should perform their own red teaming and provide related security measures before deployment, and they must abide by and comply with local governance and regulations.
> In no event shall the authors be held liable for any claim, damages, or other liability arising from the use of the released weights, codes, or demos.
>
>
>
>
> The logo was generated by DALL-E 3.
>
>
>
### What's new since SeaLLM-13B-v1 and SeaLLM-7B-v1?
* SeaLLM-7B-v2 is continue-pretrained from Mistral-7B and underwent carefully designed tuning with focus in reasoning.
Evaluation
----------
### Zero-shot CoT Multilingual Math Reasoning
SeaLLM-7B-v2 achieves with 78.2 score on the GSM8K with zero-shot CoT reasoning, making it the state of the art in the realm of 7B models. It also outperforms GPT-3.5 in the same GSM8K benchmark as translated into SEA languages (🇨🇳 🇻🇳 🇮🇩 🇹🇭). SeaLLM-7B-v2 also surpasses GPT-3.5 on the Thai-translated MATH benchmark, with 22.4 vs 18.1 scores.
!fig\_sea\_math\_side\_by\_side.png
See details on English and translated GSM8K and MATH with zero-shot reasoning
#### Zero-shot MGSM
SeaLLM-7B-v2 also outperforms GPT-3.5 and Qwen-14B on the multilingual MGSM for Zh and Th.
Model: ChatGPT (reported), MGSM-Zh: 61.2\*, MGSM-Th: 47.2\*
Model: Qwen-14B-chat, MGSM-Zh: 59.6, MGSM-Th: 28
Model: SeaLLM-7B-v2, MGSM-Zh: 64.8, MGSM-Th: 62.4
### Zero-shot Commonsense Reasoning
We compare SeaLLM-7B-v2 with ChatGPT and Mistral-7B-instruct on various zero-shot commonsense benchmarks (Arc-Challenge, Winogrande and Hellaswag). We use the 2-stage technique in (Kojima et al., 2023) to grab the answer. Note that we DID NOT use "Let's think step-by-step" to invoke explicit CoT.
### Multilingual World Knowledge
We evaluate models on 3 benchmarks following the recommended default setups: 5-shot MMLU for En, 3-shot M3Exam (M3e) for En, Zh, Vi, Id, Th, and zero-shot VMLU for Vi.
VMLU reproduce script here. Lm-eval was used to evaluate MMLU.
### MT-Bench
On the English MT-bench metric, SeaLLM-7B-v2 achieves 7.54 score on the MT-bench (3rd place on the leaderboard for 7B category), outperforms many 70B models and is arguably the only one that handles 10 SEA languages.
Refer to mt\_bench/seallm\_7b\_v2.jsonl for the MT-bench predictions of SeaLLM-7B-v2, and here to reproduce it.
### Sea-Bench
Similar to MT-Bench, Sea-bench is a set of categorized instruction test sets to measure models' ability as an assistant that is specifically focused on 9 SEA languages, including non-Latin low-resource languages.
As shown, the huge improvements come from math-reasoning, reaching GPT-3.5 level of performance.
!fig\_sea\_bench\_side\_by\_side.png
Refer to sea\_bench/seallm\_7b\_v2.jsonl for the Sea-bench predictions of SeaLLM-7B-v2.
### Usage
#### Instruction format
#### Using transformers's chat\_template
#### Using vLLM
Acknowledgement to Our Linguists
--------------------------------
We would like to express our special thanks to our professional and native linguists, Tantong Champaiboon, Nguyen Ngoc Yen Nhi and Tara Devina Putri, who helped build, evaluate, and fact-check our sampled pretraining and SFT dataset as well as evaluating our models across different aspects, especially safety.
If you find our project useful, we hope you would kindly star our repo and cite our work as follows: Corresponding Author: l.bing@URL
Author list and order will change!
* '\*' and '^' are equal contributions.
| [
"### Highlights\n\n\n* SeaLLM-7B-v2 achieves the 7B-SOTA on the Zero-shot CoT GSM8K task with 78.2 score and outperforms GPT-3.5 in many GSM8K-translated tasks in SEA languages (🇨🇳 🇻🇳 🇮🇩 🇹🇭) as well as MGSM (🇨🇳 🇹🇭). It also surpasses GPT-3.5 in MATH CoT for Thai 🇹🇭.\n* It scores competitively against GPT-3.5 in many zero-shot CoT commonsense benchmark, with 82.5, 68.3, 80.9 scores on Arc-C, Winogrande, and Hellaswag.\n* It achieves 7.54 score on the 🇬🇧 MT-bench, it ranks 3rd place on the leaderboard for 7B category and is the most outperforming multilingual model.\n* It scores 45.74 on the VMLU benchmark for Vietnamese 🇻🇳, and is the only open-source multilingual model that can be competitive to monolingual models (Vistral-7B) of similar sizes.",
"### Release and DEMO\n\n\n* DEMO: SeaLLMs/SeaLLM-7B.\n* Technical report: Arxiv: SeaLLMs - Large Language Models for Southeast Asia.\n* Model weights:\n\t+ SeaLLM-7B-v2.\n\t+ SeaLLM-7B-v2-gguf.\n\t\t- SeaLLM-7B-v2-q4\\_0, SeaLLM-7B-v2-q8\\_0.\n\t\t- LM-studio requires this URL to work properly.\n\t+ SeaLLM-7B-v2-GGUF (thanks Lonestriker). NOTE: use URL to work properly.\n\n\n\n> \n> **Terms of Use and License**: \n> By using our released weights, codes, and demos, you agree to and comply with the terms and conditions specified in our [SeaLLMs Terms Of Use](URL target=).\n> \n> \n> \n\n\n\n> \n> Disclaimer:\n> We must note that even though the weights, codes, and demos are released in an open manner, similar to other pre-trained language models, and despite our best efforts in red teaming and safety fine-tuning and enforcement, our models come with potential risks, including but not limited to inaccurate, misleading or potentially harmful generation.\n> Developers and stakeholders should perform their own red teaming and provide related security measures before deployment, and they must abide by and comply with local governance and regulations.\n> In no event shall the authors be held liable for any claim, damages, or other liability arising from the use of the released weights, codes, or demos.\n> \n> \n> \n\n\n\n> \n> The logo was generated by DALL-E 3.\n> \n> \n>",
"### What's new since SeaLLM-13B-v1 and SeaLLM-7B-v1?\n\n\n* SeaLLM-7B-v2 is continue-pretrained from Mistral-7B and underwent carefully designed tuning with focus in reasoning.\n\n\nEvaluation\n----------",
"### Zero-shot CoT Multilingual Math Reasoning\n\n\nSeaLLM-7B-v2 achieves with 78.2 score on the GSM8K with zero-shot CoT reasoning, making it the state of the art in the realm of 7B models. It also outperforms GPT-3.5 in the same GSM8K benchmark as translated into SEA languages (🇨🇳 🇻🇳 🇮🇩 🇹🇭). SeaLLM-7B-v2 also surpasses GPT-3.5 on the Thai-translated MATH benchmark, with 22.4 vs 18.1 scores.\n\n\n!fig\\_sea\\_math\\_side\\_by\\_side.png\n\n\n\nSee details on English and translated GSM8K and MATH with zero-shot reasoning",
"#### Zero-shot MGSM\n\n\nSeaLLM-7B-v2 also outperforms GPT-3.5 and Qwen-14B on the multilingual MGSM for Zh and Th.\n\n\nModel: ChatGPT (reported), MGSM-Zh: 61.2\\*, MGSM-Th: 47.2\\*\nModel: Qwen-14B-chat, MGSM-Zh: 59.6, MGSM-Th: 28\nModel: SeaLLM-7B-v2, MGSM-Zh: 64.8, MGSM-Th: 62.4",
"### Zero-shot Commonsense Reasoning\n\n\nWe compare SeaLLM-7B-v2 with ChatGPT and Mistral-7B-instruct on various zero-shot commonsense benchmarks (Arc-Challenge, Winogrande and Hellaswag). We use the 2-stage technique in (Kojima et al., 2023) to grab the answer. Note that we DID NOT use \"Let's think step-by-step\" to invoke explicit CoT.",
"### Multilingual World Knowledge\n\n\nWe evaluate models on 3 benchmarks following the recommended default setups: 5-shot MMLU for En, 3-shot M3Exam (M3e) for En, Zh, Vi, Id, Th, and zero-shot VMLU for Vi.\n\n\n\nVMLU reproduce script here. Lm-eval was used to evaluate MMLU.",
"### MT-Bench\n\n\nOn the English MT-bench metric, SeaLLM-7B-v2 achieves 7.54 score on the MT-bench (3rd place on the leaderboard for 7B category), outperforms many 70B models and is arguably the only one that handles 10 SEA languages.\n\n\nRefer to mt\\_bench/seallm\\_7b\\_v2.jsonl for the MT-bench predictions of SeaLLM-7B-v2, and here to reproduce it.",
"### Sea-Bench\n\n\nSimilar to MT-Bench, Sea-bench is a set of categorized instruction test sets to measure models' ability as an assistant that is specifically focused on 9 SEA languages, including non-Latin low-resource languages.\n\n\nAs shown, the huge improvements come from math-reasoning, reaching GPT-3.5 level of performance.\n\n\n!fig\\_sea\\_bench\\_side\\_by\\_side.png\n\n\nRefer to sea\\_bench/seallm\\_7b\\_v2.jsonl for the Sea-bench predictions of SeaLLM-7B-v2.",
"### Usage",
"#### Instruction format",
"#### Using transformers's chat\\_template",
"#### Using vLLM\n\n\nAcknowledgement to Our Linguists\n--------------------------------\n\n\nWe would like to express our special thanks to our professional and native linguists, Tantong Champaiboon, Nguyen Ngoc Yen Nhi and Tara Devina Putri, who helped build, evaluate, and fact-check our sampled pretraining and SFT dataset as well as evaluating our models across different aspects, especially safety.\n\n\nIf you find our project useful, we hope you would kindly star our repo and cite our work as follows: Corresponding Author: l.bing@URL\n\n\nAuthor list and order will change!\n\n\n* '\\*' and '^' are equal contributions."
] | [
"TAGS\n#gguf #arxiv-2312.00738 #arxiv-2205.11916 #arxiv-2306.05179 #arxiv-2306.05685 #license-other #region-us \n",
"### Highlights\n\n\n* SeaLLM-7B-v2 achieves the 7B-SOTA on the Zero-shot CoT GSM8K task with 78.2 score and outperforms GPT-3.5 in many GSM8K-translated tasks in SEA languages (🇨🇳 🇻🇳 🇮🇩 🇹🇭) as well as MGSM (🇨🇳 🇹🇭). It also surpasses GPT-3.5 in MATH CoT for Thai 🇹🇭.\n* It scores competitively against GPT-3.5 in many zero-shot CoT commonsense benchmark, with 82.5, 68.3, 80.9 scores on Arc-C, Winogrande, and Hellaswag.\n* It achieves 7.54 score on the 🇬🇧 MT-bench, it ranks 3rd place on the leaderboard for 7B category and is the most outperforming multilingual model.\n* It scores 45.74 on the VMLU benchmark for Vietnamese 🇻🇳, and is the only open-source multilingual model that can be competitive to monolingual models (Vistral-7B) of similar sizes.",
"### Release and DEMO\n\n\n* DEMO: SeaLLMs/SeaLLM-7B.\n* Technical report: Arxiv: SeaLLMs - Large Language Models for Southeast Asia.\n* Model weights:\n\t+ SeaLLM-7B-v2.\n\t+ SeaLLM-7B-v2-gguf.\n\t\t- SeaLLM-7B-v2-q4\\_0, SeaLLM-7B-v2-q8\\_0.\n\t\t- LM-studio requires this URL to work properly.\n\t+ SeaLLM-7B-v2-GGUF (thanks Lonestriker). NOTE: use URL to work properly.\n\n\n\n> \n> **Terms of Use and License**: \n> By using our released weights, codes, and demos, you agree to and comply with the terms and conditions specified in our [SeaLLMs Terms Of Use](URL target=).\n> \n> \n> \n\n\n\n> \n> Disclaimer:\n> We must note that even though the weights, codes, and demos are released in an open manner, similar to other pre-trained language models, and despite our best efforts in red teaming and safety fine-tuning and enforcement, our models come with potential risks, including but not limited to inaccurate, misleading or potentially harmful generation.\n> Developers and stakeholders should perform their own red teaming and provide related security measures before deployment, and they must abide by and comply with local governance and regulations.\n> In no event shall the authors be held liable for any claim, damages, or other liability arising from the use of the released weights, codes, or demos.\n> \n> \n> \n\n\n\n> \n> The logo was generated by DALL-E 3.\n> \n> \n>",
"### What's new since SeaLLM-13B-v1 and SeaLLM-7B-v1?\n\n\n* SeaLLM-7B-v2 is continue-pretrained from Mistral-7B and underwent carefully designed tuning with focus in reasoning.\n\n\nEvaluation\n----------",
"### Zero-shot CoT Multilingual Math Reasoning\n\n\nSeaLLM-7B-v2 achieves with 78.2 score on the GSM8K with zero-shot CoT reasoning, making it the state of the art in the realm of 7B models. It also outperforms GPT-3.5 in the same GSM8K benchmark as translated into SEA languages (🇨🇳 🇻🇳 🇮🇩 🇹🇭). SeaLLM-7B-v2 also surpasses GPT-3.5 on the Thai-translated MATH benchmark, with 22.4 vs 18.1 scores.\n\n\n!fig\\_sea\\_math\\_side\\_by\\_side.png\n\n\n\nSee details on English and translated GSM8K and MATH with zero-shot reasoning",
"#### Zero-shot MGSM\n\n\nSeaLLM-7B-v2 also outperforms GPT-3.5 and Qwen-14B on the multilingual MGSM for Zh and Th.\n\n\nModel: ChatGPT (reported), MGSM-Zh: 61.2\\*, MGSM-Th: 47.2\\*\nModel: Qwen-14B-chat, MGSM-Zh: 59.6, MGSM-Th: 28\nModel: SeaLLM-7B-v2, MGSM-Zh: 64.8, MGSM-Th: 62.4",
"### Zero-shot Commonsense Reasoning\n\n\nWe compare SeaLLM-7B-v2 with ChatGPT and Mistral-7B-instruct on various zero-shot commonsense benchmarks (Arc-Challenge, Winogrande and Hellaswag). We use the 2-stage technique in (Kojima et al., 2023) to grab the answer. Note that we DID NOT use \"Let's think step-by-step\" to invoke explicit CoT.",
"### Multilingual World Knowledge\n\n\nWe evaluate models on 3 benchmarks following the recommended default setups: 5-shot MMLU for En, 3-shot M3Exam (M3e) for En, Zh, Vi, Id, Th, and zero-shot VMLU for Vi.\n\n\n\nVMLU reproduce script here. Lm-eval was used to evaluate MMLU.",
"### MT-Bench\n\n\nOn the English MT-bench metric, SeaLLM-7B-v2 achieves 7.54 score on the MT-bench (3rd place on the leaderboard for 7B category), outperforms many 70B models and is arguably the only one that handles 10 SEA languages.\n\n\nRefer to mt\\_bench/seallm\\_7b\\_v2.jsonl for the MT-bench predictions of SeaLLM-7B-v2, and here to reproduce it.",
"### Sea-Bench\n\n\nSimilar to MT-Bench, Sea-bench is a set of categorized instruction test sets to measure models' ability as an assistant that is specifically focused on 9 SEA languages, including non-Latin low-resource languages.\n\n\nAs shown, the huge improvements come from math-reasoning, reaching GPT-3.5 level of performance.\n\n\n!fig\\_sea\\_bench\\_side\\_by\\_side.png\n\n\nRefer to sea\\_bench/seallm\\_7b\\_v2.jsonl for the Sea-bench predictions of SeaLLM-7B-v2.",
"### Usage",
"#### Instruction format",
"#### Using transformers's chat\\_template",
"#### Using vLLM\n\n\nAcknowledgement to Our Linguists\n--------------------------------\n\n\nWe would like to express our special thanks to our professional and native linguists, Tantong Champaiboon, Nguyen Ngoc Yen Nhi and Tara Devina Putri, who helped build, evaluate, and fact-check our sampled pretraining and SFT dataset as well as evaluating our models across different aspects, especially safety.\n\n\nIf you find our project useful, we hope you would kindly star our repo and cite our work as follows: Corresponding Author: l.bing@URL\n\n\nAuthor list and order will change!\n\n\n* '\\*' and '^' are equal contributions."
] | [
47,
247,
387,
62,
171,
123,
104,
81,
118,
143,
4,
5,
13,
145
] | [
"passage: TAGS\n#gguf #arxiv-2312.00738 #arxiv-2205.11916 #arxiv-2306.05179 #arxiv-2306.05685 #license-other #region-us \n### Highlights\n\n\n* SeaLLM-7B-v2 achieves the 7B-SOTA on the Zero-shot CoT GSM8K task with 78.2 score and outperforms GPT-3.5 in many GSM8K-translated tasks in SEA languages (🇨🇳 🇻🇳 🇮🇩 🇹🇭) as well as MGSM (🇨🇳 🇹🇭). It also surpasses GPT-3.5 in MATH CoT for Thai 🇹🇭.\n* It scores competitively against GPT-3.5 in many zero-shot CoT commonsense benchmark, with 82.5, 68.3, 80.9 scores on Arc-C, Winogrande, and Hellaswag.\n* It achieves 7.54 score on the 🇬🇧 MT-bench, it ranks 3rd place on the leaderboard for 7B category and is the most outperforming multilingual model.\n* It scores 45.74 on the VMLU benchmark for Vietnamese 🇻🇳, and is the only open-source multilingual model that can be competitive to monolingual models (Vistral-7B) of similar sizes.",
"passage: ### Release and DEMO\n\n\n* DEMO: SeaLLMs/SeaLLM-7B.\n* Technical report: Arxiv: SeaLLMs - Large Language Models for Southeast Asia.\n* Model weights:\n\t+ SeaLLM-7B-v2.\n\t+ SeaLLM-7B-v2-gguf.\n\t\t- SeaLLM-7B-v2-q4\\_0, SeaLLM-7B-v2-q8\\_0.\n\t\t- LM-studio requires this URL to work properly.\n\t+ SeaLLM-7B-v2-GGUF (thanks Lonestriker). NOTE: use URL to work properly.\n\n\n\n> \n> **Terms of Use and License**: \n> By using our released weights, codes, and demos, you agree to and comply with the terms and conditions specified in our [SeaLLMs Terms Of Use](URL target=).\n> \n> \n> \n\n\n\n> \n> Disclaimer:\n> We must note that even though the weights, codes, and demos are released in an open manner, similar to other pre-trained language models, and despite our best efforts in red teaming and safety fine-tuning and enforcement, our models come with potential risks, including but not limited to inaccurate, misleading or potentially harmful generation.\n> Developers and stakeholders should perform their own red teaming and provide related security measures before deployment, and they must abide by and comply with local governance and regulations.\n> In no event shall the authors be held liable for any claim, damages, or other liability arising from the use of the released weights, codes, or demos.\n> \n> \n> \n\n\n\n> \n> The logo was generated by DALL-E 3.\n> \n> \n>### What's new since SeaLLM-13B-v1 and SeaLLM-7B-v1?\n\n\n* SeaLLM-7B-v2 is continue-pretrained from Mistral-7B and underwent carefully designed tuning with focus in reasoning.\n\n\nEvaluation\n----------### Zero-shot CoT Multilingual Math Reasoning\n\n\nSeaLLM-7B-v2 achieves with 78.2 score on the GSM8K with zero-shot CoT reasoning, making it the state of the art in the realm of 7B models. It also outperforms GPT-3.5 in the same GSM8K benchmark as translated into SEA languages (🇨🇳 🇻🇳 🇮🇩 🇹🇭). SeaLLM-7B-v2 also surpasses GPT-3.5 on the Thai-translated MATH benchmark, with 22.4 vs 18.1 scores.\n\n\n!fig\\_sea\\_math\\_side\\_by\\_side.png\n\n\n\nSee details on English and translated GSM8K and MATH with zero-shot reasoning#### Zero-shot MGSM\n\n\nSeaLLM-7B-v2 also outperforms GPT-3.5 and Qwen-14B on the multilingual MGSM for Zh and Th.\n\n\nModel: ChatGPT (reported), MGSM-Zh: 61.2\\*, MGSM-Th: 47.2\\*\nModel: Qwen-14B-chat, MGSM-Zh: 59.6, MGSM-Th: 28\nModel: SeaLLM-7B-v2, MGSM-Zh: 64.8, MGSM-Th: 62.4### Zero-shot Commonsense Reasoning\n\n\nWe compare SeaLLM-7B-v2 with ChatGPT and Mistral-7B-instruct on various zero-shot commonsense benchmarks (Arc-Challenge, Winogrande and Hellaswag). We use the 2-stage technique in (Kojima et al., 2023) to grab the answer. Note that we DID NOT use \"Let's think step-by-step\" to invoke explicit CoT."
] | [
-0.0454837828874588,
0.0582427978515625,
-0.005183323752135038,
0.05588586628437042,
0.063256174325943,
0.013180009089410305,
0.08342257887125015,
0.06947030127048492,
-0.008440257981419563,
0.08738803118467331,
-0.04695221036672592,
-0.010942909866571426,
0.12331520766019821,
0.0694468691945076,
0.05816822499036789,
-0.20352019369602203,
0.012375932186841965,
-0.06770716607570648,
0.011464880779385567,
0.054482221603393555,
0.06140110641717911,
-0.03654547408223152,
0.06876681745052338,
-0.004022234585136175,
0.05977603420615196,
0.020832180976867676,
-0.01312345452606678,
-0.001562213758006692,
0.05308223143219948,
0.040091853588819504,
0.058419376611709595,
0.013212013989686966,
0.010625745169818401,
-0.14070364832878113,
0.0032472647726535797,
0.0434657484292984,
-0.010885090567171574,
0.03880401328206062,
0.07708220183849335,
-0.005315621383488178,
0.12172679603099823,
-0.04200059175491333,
-0.002460528165102005,
0.05535534769296646,
-0.08658992499113083,
-0.22015872597694397,
-0.10762235522270203,
0.03551476448774338,
0.1450120359659195,
0.009626930579543114,
-0.01713821291923523,
0.08424186706542969,
-0.06624700874090195,
0.03890487179160118,
0.12713979184627533,
-0.307742178440094,
-0.05936051905155182,
0.17416644096374512,
-0.0204862579703331,
0.062366634607315063,
-0.02778663858771324,
0.05423669517040253,
0.02029675617814064,
0.05176911875605583,
-0.03529128432273865,
-0.021577579900622368,
0.058440275490283966,
0.0017015505582094193,
-0.11247697472572327,
-0.01622091606259346,
0.18675366044044495,
0.008299212902784348,
-0.06663835793733597,
-0.12462200224399567,
-0.05872960016131401,
-0.015692085027694702,
-0.00893162377178669,
0.017008991912007332,
0.04877958074212074,
0.08081895112991333,
0.11788716912269592,
0.022401567548513412,
-0.10221529006958008,
-0.05433659255504608,
-0.07349921017885208,
0.14054332673549652,
0.03385648503899574,
0.03523355722427368,
0.018417656421661377,
0.07520908117294312,
-0.15825219452381134,
-0.0717196986079216,
-0.11966016888618469,
-0.05007064342498779,
-0.08779053390026093,
0.03257323056459427,
0.059526730328798294,
-0.024899955838918686,
0.055540405213832855,
0.050969965755939484,
-0.06432992964982986,
0.11084886640310287,
-0.0031767631880939007,
0.03380729258060455,
0.020706763491034508,
0.1201663687825203,
-0.05571344494819641,
0.030333127826452255,
0.03241361677646637,
-0.006796732544898987,
0.07424268126487732,
-0.0013266642345115542,
-0.02190154604613781,
-0.019895458593964577,
0.032864734530448914,
0.0628134235739708,
0.030827246606349945,
-0.0016837352886795998,
-0.10156615078449249,
-0.023052938282489777,
0.2919120788574219,
-0.09197718650102615,
-0.014460407197475433,
0.005279208533465862,
-0.03018888458609581,
0.002593938261270523,
0.04074923321604729,
0.000131240114569664,
-0.011790698394179344,
0.07944495975971222,
-0.002005330752581358,
-0.021455202251672745,
-0.05506046861410141,
-0.04714563116431236,
0.040927331894636154,
0.037519000470638275,
-0.07842932641506195,
-0.05090354382991791,
-0.09931670129299164,
0.00754614919424057,
0.061966490000486374,
-0.07654856890439987,
-0.02968129888176918,
0.03924906253814697,
-0.040368735790252686,
-0.024851197376847267,
-0.017841264605522156,
0.022834639996290207,
-0.019264649599790573,
0.01685977727174759,
0.04635576903820038,
0.055202946066856384,
-0.02964860200881958,
-0.02879190444946289,
-0.005798462312668562,
0.10822651535272598,
-0.16433191299438477,
0.09315799176692963,
-0.11495611071586609,
-0.03885076195001602,
-0.07732374966144562,
-0.04057934880256653,
-0.083281010389328,
0.01767547056078911,
0.02921833097934723,
0.06679420173168182,
-0.13432425260543823,
-0.012160088866949081,
0.13727033138275146,
-0.1295585334300995,
-0.05143972486257553,
0.10373568534851074,
0.032607920467853546,
-0.04354493319988251,
0.043332383036613464,
0.05188780650496483,
0.07311888039112091,
-0.08114005625247955,
-0.05028422549366951,
0.007022494450211525,
-0.05964293330907822,
0.020751720294356346,
0.09204001724720001,
-0.049654580652713776,
-0.001961188390851021,
0.053448501974344254,
-0.010322464630007744,
-0.011881796643137932,
0.00462797237560153,
-0.033496372401714325,
0.014734948053956032,
-0.007042402401566505,
-0.04900279641151428,
0.0381501168012619,
-0.01432315818965435,
-0.025781797245144844,
-0.054533950984478,
-0.0058469343930482864,
0.07397995889186859,
-0.01987418159842491,
-0.014547733590006828,
-0.09192464500665665,
0.08764044940471649,
-0.09421591460704803,
0.012951561249792576,
-0.101294606924057,
-0.10656449943780899,
0.08056407421827316,
-0.09515789896249771,
0.02165394276380539,
0.04196149855852127,
0.06393593549728394,
0.0597008652985096,
-0.06872150301933289,
0.029513418674468994,
0.07788150012493134,
-0.03736060857772827,
-0.025329601019620895,
-0.1529763638973236,
0.02089468576014042,
-0.006106637418270111,
0.0558796189725399,
-0.08531894534826279,
0.008610019460320473,
0.1909267008304596,
0.06665395945310593,
-0.032458461821079254,
-0.06178320199251175,
0.08235685527324677,
-0.0017211390659213066,
-0.007648581638932228,
-0.051991112530231476,
0.017464986070990562,
-0.017586123198270798,
-0.0706292912364006,
0.07561253011226654,
-0.11259836703538895,
0.06893379986286163,
0.07423362135887146,
0.09578998386859894,
-0.06121587008237839,
-0.012266307137906551,
-0.03883245214819908,
-0.019198689609766006,
-0.05906928330659866,
0.043306317180395126,
0.0688437819480896,
0.014394400641322136,
0.09172867238521576,
-0.11812245845794678,
-0.05332262068986893,
0.033797912299633026,
-0.012098556384444237,
-0.017486829310655594,
0.11252893507480621,
0.05963617563247681,
-0.12783515453338623,
0.07083078473806381,
-0.014486271888017654,
0.030040938407182693,
0.12787173688411713,
-0.0008234805427491665,
0.004319450818002224,
-0.06809297204017639,
0.1274150162935257,
0.046133577823638916,
0.13898226618766785,
-0.0241130031645298,
0.047022949904203415,
0.02674875035881996,
0.06820770353078842,
0.0513535812497139,
-0.13408713042736053,
-0.014092615805566311,
0.002011263743042946,
-0.0820382758975029,
-0.032326363027095795,
0.07276596128940582,
-0.03371662646532059,
0.08249376714229584,
0.00925576314330101,
0.0837690681219101,
-0.04298108071088791,
-0.03638788312673569,
-0.08034123480319977,
0.15353861451148987,
-0.07873411476612091,
-0.09983928501605988,
-0.1518605798482895,
0.09344780445098877,
-0.1089458167552948,
0.004147626459598541,
-0.014125239104032516,
-0.1039804220199585,
-0.07653383910655975,
-0.05037909373641014,
-0.032390281558036804,
-0.04885798320174217,
-0.010469801723957062,
-0.013358882628381252,
0.0235771331936121,
0.036664243787527084,
-0.11506374925374985,
-0.007754504680633545,
-0.0110847856849432,
-0.09053830802440643,
0.032480012625455856,
-0.04921790957450867,
0.02599513716995716,
0.1075139045715332,
0.03033892810344696,
0.02942381240427494,
0.0021873759105801582,
0.13834883272647858,
-0.06543625891208649,
0.03626203536987305,
0.10909619927406311,
0.0813317745923996,
0.06564224511384964,
0.0920805037021637,
0.037523262202739716,
-0.10005716979503632,
-0.01915648579597473,
0.04780290648341179,
-0.10200875997543335,
-0.2219213992357254,
-0.03929987549781799,
-0.0669189840555191,
0.04614967107772827,
-0.05067553371191025,
0.08011756092309952,
-0.0088593028485775,
0.028590356931090355,
-0.03845588490366936,
0.040344975888729095,
0.10321361571550369,
0.05785439908504486,
0.08855204284191132,
0.043077945709228516,
0.08142317831516266,
-0.09320737421512604,
0.03035690262913704,
0.1318061351776123,
0.028787020593881607,
0.16263434290885925,
-0.020740877836942673,
0.0657513290643692,
0.09389494359493256,
0.08852046728134155,
0.04521483927965164,
0.04143565893173218,
-0.07014987617731094,
0.014632027596235275,
-0.016010066494345665,
-0.08229875564575195,
-0.002548210322856903,
0.06689649075269699,
0.007152857258915901,
-0.045361220836639404,
0.05549539625644684,
0.012445446103811264,
0.049490369856357574,
0.17824770510196686,
-0.02285619266331196,
-0.047042012214660645,
-0.04294531047344208,
0.08713509142398834,
0.019565468654036522,
-0.024513080716133118,
-0.004652324598282576,
0.1053776890039444,
-0.06340475380420685,
0.16705025732517242,
0.0038276968989521265,
0.057880278676748276,
-0.09921064972877502,
-0.044730789959430695,
0.009631742723286152,
0.08846081793308258,
0.02803981862962246,
0.07137683779001236,
-0.10595771670341492,
0.18070688843727112,
0.05221564322710037,
0.010781215503811836,
-0.01563728041946888,
0.04677855595946312,
-0.010610215365886688,
0.013794615864753723,
0.11430706083774567,
0.030036957934498787,
-0.07544899731874466,
-0.08296320587396622,
-0.06846679747104645,
0.040498871356248856,
0.05435885861515999,
-0.04908706247806549,
0.03766424581408501,
0.02394697070121765,
-0.027727214619517326,
-0.02557969279587269,
-0.030512863770127296,
-0.09836310893297195,
-0.08407637476921082,
0.028889546170830727,
-0.05108867958188057,
-0.11629438400268555,
-0.06378612667322159,
-0.01912868395447731,
-0.10635451972484589,
0.17606478929519653,
-0.23902106285095215,
-0.11170633137226105,
-0.06612899154424667,
-0.06289583444595337,
0.09611594676971436,
-0.06714256852865219,
-0.011201245710253716,
-0.08654665946960449,
0.064975306391716,
0.0023431777954101562,
-0.03925314545631409,
0.05819210410118103,
-0.05576620623469353,
-0.1755256950855255,
-0.018267285078763962,
0.1313723921775818,
0.025213129818439484,
0.01146834809333086,
-0.01564635895192623,
0.01601673848927021,
0.040062639862298965,
-0.16853880882263184,
0.04374402016401291,
0.19587400555610657,
-0.01517007127404213,
0.06317263841629028,
0.03355409950017929,
0.001844768412411213,
-0.01595110446214676,
-0.0908202975988388,
-0.037888605147600174,
0.3065996766090393,
-0.06354121118783951,
0.0896863117814064,
0.07376493513584137,
-0.06158159673213959,
-0.13594013452529907,
-0.10642984509468079,
0.008905338123440742,
0.023083461448550224,
-0.006119675934314728,
-0.11681066453456879,
0.017317403107881546,
0.04282462224364281,
-0.05852224677801132,
-0.027772288769483566,
-0.20354142785072327,
-0.1319267898797989,
-0.048498526215553284,
0.0453060008585453,
0.08602283149957657,
-0.1378946751356125,
-0.0701930969953537,
-0.023315493017435074,
-0.12205629795789719,
0.0381031334400177,
-0.021481899544596672,
0.07025331258773804,
-0.054415833204984665,
-0.0833468958735466,
0.0024088891223073006,
-0.04869839549064636,
0.18406450748443604,
-0.053008072078228,
0.047565899789333344,
-0.020775141194462776,
0.055713728070259094,
0.02064748853445053,
0.008762460201978683,
0.11205143481492996,
0.02822403609752655,
0.0078058564104139805,
-0.09606365859508514,
-0.05904946103692055,
-0.05455104261636734,
-0.0031396523118019104,
-0.02203395590186119,
-0.014338046312332153,
-0.08230942487716675,
0.05154922604560852,
0.004582345485687256,
-0.0022941145580261946,
-0.004970442969352007,
-0.025609824806451797,
-0.021712597459554672,
0.13310053944587708,
0.15495240688323975,
-0.005536071956157684,
0.01797868311405182,
-0.030145611613988876,
0.0008474844507873058,
0.02433125302195549,
-0.18799063563346863,
0.042236290872097015,
0.07024107873439789,
-0.016902629286050797,
0.02338676154613495,
0.020037468522787094,
-0.16297215223312378,
0.03978396952152252,
0.11822189390659332,
-0.06218749284744263,
-0.2193368375301361,
-0.0034885480999946594,
0.005682025104761124,
0.054549440741539,
0.009698159992694855,
0.05829814821481705,
-0.03033865988254547,
0.0033719120547175407,
0.043286968022584915,
0.023683704435825348,
-0.03564225509762764,
0.08261989057064056,
-0.012045072391629219,
0.04456006735563278,
-0.07782772928476334,
0.09619438648223877,
0.0988725796341896,
0.06472641974687576,
-0.0038308370858430862,
0.08983582258224487,
-0.06963305175304413,
-0.008666973561048508,
-0.039471909403800964,
0.0466768816113472,
0.07104310393333435,
-0.0779750794172287,
-0.055303968489170074,
-0.07961679995059967,
-0.016155265271663666,
0.040397532284259796,
0.07231467217206955,
0.05557064712047577,
0.03240598738193512,
0.025451788678765297,
0.014474691823124886,
-0.01379435695707798,
0.03441064432263374,
0.026360156014561653,
-0.15259376168251038,
-0.013971962034702301,
0.02540874294936657,
0.006082044914364815,
-0.01600285805761814,
-0.03842337429523468,
-0.07828192412853241,
-0.004508296959102154,
-0.07020688056945801,
0.020856305956840515,
-0.08801224827766418,
0.004956671968102455,
-0.015013479627668858,
-0.05343012511730194,
-0.049659617245197296,
0.014207734726369381,
-0.06161579489707947,
0.018787886947393417,
-0.06473512202501297,
0.06656467914581299,
-0.10532049834728241,
-0.042471110820770264,
0.06546977162361145,
-0.0433940552175045,
0.040268559008836746,
-0.04605426639318466,
0.029873890802264214,
0.03174365684390068,
-0.07212795317173004,
0.03189869225025177,
0.03431132435798645,
0.01718079298734665,
-0.03214414417743683,
-0.11289623379707336,
0.04377482458949089,
0.01481806393712759,
0.010722827166318893,
0.03055998682975769,
0.07482538372278214,
-0.10984484851360321,
-0.06834279745817184,
-0.048787105828523636,
-0.033765241503715515,
-0.04127292335033417,
0.016429267823696136,
0.12222378700971603,
0.11029858887195587,
0.022189833223819733,
-0.048621684312820435,
-0.013848453760147095,
-0.19913151860237122,
-0.01914210617542267,
0.013641664758324623,
-0.03972819820046425,
0.004094673320651054,
-0.0054565309546887875,
0.03568793460726738,
-0.0028076795861124992,
0.17717602849006653,
0.026250839233398438,
-0.015975555405020714,
0.019036829471588135,
-0.004915941506624222,
0.010331755504012108,
0.018631206825375557,
-0.029386209324002266,
0.03282563015818596,
0.0325479656457901,
0.008700869046151638,
-0.02487868070602417,
-0.024268630892038345,
-0.06919798254966736,
0.04587209224700928,
0.044302064925432205,
-0.0071761831641197205,
0.015797631815075874,
0.060257699340581894,
-0.05526448041200638,
-0.07380270957946777,
0.0579296238720417,
-0.13016726076602936,
0.012794316746294498,
-0.040403515100479126,
-0.04559122771024704,
0.08498785644769669,
-0.14204862713813782,
0.07681472599506378,
-0.024041227996349335,
-0.036105476319789886,
-0.13532838225364685,
-0.17577452957630157,
-0.0522327795624733,
-0.08772002905607224,
0.028162162750959396,
-0.06403040140867233,
0.0194525346159935,
0.12562386691570282,
0.056386519223451614,
0.028873687610030174,
0.04697827994823456,
-0.07452838867902756,
0.02149520441889763,
-0.00788930431008339,
-0.00198329146951437,
0.001734214834868908,
0.13963961601257324,
0.029303602874279022,
0.004617990925908089,
0.0035822587087750435,
0.026771755889058113,
0.05733640491962433,
-0.023163625970482826,
-0.002075674943625927,
0.0029237791895866394,
-0.050824567675590515,
-0.036731161177158356,
0.0016864682547748089,
-0.002472951542586088,
0.14508745074272156,
0.004152788780629635,
-0.03433273732662201,
0.015760235488414764,
0.12694238126277924,
-0.006942713633179665,
-0.09298014640808105,
-0.12666594982147217,
0.22913682460784912,
-0.05777951702475548,
0.012979279272258282,
0.043612945824861526,
-0.044420868158340454,
0.003922400996088982,
0.17556548118591309,
0.05615772306919098,
0.018314391374588013,
0.0006000876892358065,
0.019665850326418877,
-0.004056501667946577,
0.00020203739404678345,
0.04602740705013275,
0.0015151493716984987,
0.23243847489356995,
-0.05187661200761795,
0.06978511065244675,
-0.04087673872709274,
0.02868749015033245,
-0.02900027483701706,
0.08902263641357422,
0.016569847241044044,
0.007394283078610897,
-0.08186028897762299,
0.07579368352890015,
-0.009172435849905014,
-0.05373713746666908,
0.04826204106211662,
-0.11565066128969193,
-0.11098740994930267,
-0.0030908528715372086,
0.012182794511318207,
-0.004815596155822277,
-0.002951459027826786,
-0.04745753854513168,
0.0022249515168368816,
0.06061382591724396,
0.012055178172886372,
-0.1200745701789856,
-0.032305166125297546,
0.0454542376101017,
0.01158760767430067,
0.10909278690814972,
0.02611507847905159,
0.11498735100030899,
0.08876393735408783,
-0.047621551901102066,
-0.056324779987335205,
0.04564369469881058,
0.019366439431905746,
-0.07708477973937988,
-0.028772978112101555,
0.13397449254989624,
-0.022921334952116013,
-0.004492955282330513,
0.041204750537872314,
0.043937940150499344,
0.029017504304647446,
-0.008392563089728355,
-0.0053458488546311855,
-0.08706150203943253,
0.09349077939987183,
-0.09855817258358002,
0.08123236894607544,
0.10159507393836975,
-0.019247569143772125,
-0.0037826411426067352,
-0.0714556947350502,
0.06924303621053696,
-0.04947899281978607,
0.05388082563877106,
-0.029293028637766838,
-0.07463167607784271,
0.016246439889073372,
0.026164941489696503,
0.050698328763246536,
-0.09287838637828827,
-0.06371261179447174,
0.01912063919007778,
-0.02499617636203766,
-0.05251726135611534,
0.10207648575305939,
0.03180861473083496,
0.06381722539663315,
-0.0435204952955246,
-0.15826642513275146,
-0.00917735043913126,
0.008020284585654736,
-0.046723537147045135,
-0.06041189283132553
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->
# pranay-j/distilbert-base-uncased-finetuned-imdb
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 2.8551
- Validation Loss: 2.5646
- Epoch: 0
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'module': 'transformers.optimization_tf', 'class_name': 'WarmUp', 'config': {'initial_learning_rate': 2e-05, 'decay_schedule_fn': {'module': 'keras.optimizers.schedules', 'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': -688, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered_name': None}, 'warmup_steps': 1000, 'power': 1.0, 'name': None}, 'registered_name': 'WarmUp'}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.01}
- training_precision: mixed_float16
### Training results
| Train Loss | Validation Loss | Epoch |
|:----------:|:---------------:|:-----:|
| 2.8551 | 2.5646 | 0 |
### Framework versions
- Transformers 4.35.2
- TensorFlow 2.13.0
- Datasets 2.5.2
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_keras_callback"], "base_model": "distilbert-base-uncased", "model-index": [{"name": "pranay-j/distilbert-base-uncased-finetuned-imdb", "results": []}]} | fill-mask | pranay-j/distilbert-base-uncased-finetuned-imdb | [
"transformers",
"tf",
"distilbert",
"fill-mask",
"generated_from_keras_callback",
"base_model:distilbert-base-uncased",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T07:37:10+00:00 | [] | [] | TAGS
#transformers #tf #distilbert #fill-mask #generated_from_keras_callback #base_model-distilbert-base-uncased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| pranay-j/distilbert-base-uncased-finetuned-imdb
===============================================
This model is a fine-tuned version of distilbert-base-uncased on an unknown dataset.
It achieves the following results on the evaluation set:
* Train Loss: 2.8551
* Validation Loss: 2.5646
* Epoch: 0
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* optimizer: {'name': 'AdamWeightDecay', 'learning\_rate': {'module': 'transformers.optimization\_tf', 'class\_name': 'WarmUp', 'config': {'initial\_learning\_rate': 2e-05, 'decay\_schedule\_fn': {'module': 'keras.optimizers.schedules', 'class\_name': 'PolynomialDecay', 'config': {'initial\_learning\_rate': 2e-05, 'decay\_steps': -688, 'end\_learning\_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered\_name': None}, 'warmup\_steps': 1000, 'power': 1.0, 'name': None}, 'registered\_name': 'WarmUp'}, 'decay': 0.0, 'beta\_1': 0.9, 'beta\_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight\_decay\_rate': 0.01}
* training\_precision: mixed\_float16
### Training results
### Framework versions
* Transformers 4.35.2
* TensorFlow 2.13.0
* Datasets 2.5.2
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* optimizer: {'name': 'AdamWeightDecay', 'learning\\_rate': {'module': 'transformers.optimization\\_tf', 'class\\_name': 'WarmUp', 'config': {'initial\\_learning\\_rate': 2e-05, 'decay\\_schedule\\_fn': {'module': 'keras.optimizers.schedules', 'class\\_name': 'PolynomialDecay', 'config': {'initial\\_learning\\_rate': 2e-05, 'decay\\_steps': -688, 'end\\_learning\\_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered\\_name': None}, 'warmup\\_steps': 1000, 'power': 1.0, 'name': None}, 'registered\\_name': 'WarmUp'}, 'decay': 0.0, 'beta\\_1': 0.9, 'beta\\_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight\\_decay\\_rate': 0.01}\n* training\\_precision: mixed\\_float16",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* TensorFlow 2.13.0\n* Datasets 2.5.2\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tf #distilbert #fill-mask #generated_from_keras_callback #base_model-distilbert-base-uncased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* optimizer: {'name': 'AdamWeightDecay', 'learning\\_rate': {'module': 'transformers.optimization\\_tf', 'class\\_name': 'WarmUp', 'config': {'initial\\_learning\\_rate': 2e-05, 'decay\\_schedule\\_fn': {'module': 'keras.optimizers.schedules', 'class\\_name': 'PolynomialDecay', 'config': {'initial\\_learning\\_rate': 2e-05, 'decay\\_steps': -688, 'end\\_learning\\_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered\\_name': None}, 'warmup\\_steps': 1000, 'power': 1.0, 'name': None}, 'registered\\_name': 'WarmUp'}, 'decay': 0.0, 'beta\\_1': 0.9, 'beta\\_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight\\_decay\\_rate': 0.01}\n* training\\_precision: mixed\\_float16",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* TensorFlow 2.13.0\n* Datasets 2.5.2\n* Tokenizers 0.15.0"
] | [
70,
337,
4,
31
] | [
"passage: TAGS\n#transformers #tf #distilbert #fill-mask #generated_from_keras_callback #base_model-distilbert-base-uncased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* optimizer: {'name': 'AdamWeightDecay', 'learning\\_rate': {'module': 'transformers.optimization\\_tf', 'class\\_name': 'WarmUp', 'config': {'initial\\_learning\\_rate': 2e-05, 'decay\\_schedule\\_fn': {'module': 'keras.optimizers.schedules', 'class\\_name': 'PolynomialDecay', 'config': {'initial\\_learning\\_rate': 2e-05, 'decay\\_steps': -688, 'end\\_learning\\_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered\\_name': None}, 'warmup\\_steps': 1000, 'power': 1.0, 'name': None}, 'registered\\_name': 'WarmUp'}, 'decay': 0.0, 'beta\\_1': 0.9, 'beta\\_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight\\_decay\\_rate': 0.01}\n* training\\_precision: mixed\\_float16### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* TensorFlow 2.13.0\n* Datasets 2.5.2\n* Tokenizers 0.15.0"
] | [
-0.09027089178562164,
0.09097331017255783,
-0.007262998726218939,
0.07662695646286011,
0.09576687216758728,
0.07237124443054199,
0.09313175082206726,
0.11739642918109894,
-0.04821031168103218,
0.15235599875450134,
0.09918253868818283,
0.10013879090547562,
0.05624913051724434,
0.13345445692539215,
-0.05652295798063278,
-0.1616801768541336,
0.04619624838232994,
-0.06281216442584991,
-0.11673121899366379,
0.07637125253677368,
0.06993075460195541,
-0.048549942672252655,
0.06916762888431549,
-0.03849189355969429,
-0.05540524050593376,
-0.04484571889042854,
-0.014560705982148647,
-0.03363855928182602,
0.07579915225505829,
0.062298428267240524,
0.07630600780248642,
0.035112444311380386,
0.005624364595860243,
-0.2245849072933197,
0.011050572618842125,
0.10946193337440491,
0.005491298623383045,
0.07998832315206528,
0.052750878036022186,
-0.05786534771323204,
0.14692972600460052,
-0.10639850795269012,
0.059871286153793335,
0.04821949079632759,
-0.13707056641578674,
-0.24675290286540985,
-0.0787828266620636,
0.06089409068226814,
0.0911962240934372,
0.05373257026076317,
0.002583708381280303,
0.0766761302947998,
-0.08023084700107574,
0.08963502943515778,
0.08776788413524628,
-0.2433914691209793,
-0.05247712507843971,
0.02220243401825428,
-0.006698205601423979,
-0.023159269243478775,
-0.09030596166849136,
-0.011733285151422024,
0.00001705901013338007,
0.014856928028166294,
0.030083399266004562,
0.008718009106814861,
0.03696517273783684,
-0.051988136023283005,
-0.06397105008363724,
-0.0554349347949028,
0.0918082520365715,
0.08441897481679916,
-0.04806015267968178,
-0.07488922774791718,
-0.029513034969568253,
-0.20780439674854279,
-0.022165736183524132,
-0.024675430729985237,
0.003716773819178343,
0.0060614002868533134,
-0.04564407095313072,
0.014268005266785622,
-0.04727866128087044,
-0.058569762855768204,
0.02736620046198368,
0.17067649960517883,
0.03329607471823692,
0.013915691524744034,
0.024364084005355835,
0.07221880555152893,
0.04447569325566292,
-0.14301933348178864,
-0.055468641221523285,
0.0038848971016705036,
-0.09007812291383743,
-0.019434219226241112,
-0.07310515642166138,
0.0744285061955452,
0.10936865955591202,
0.20142680406570435,
-0.08693300187587738,
0.11943399906158447,
0.035875093191862106,
0.0162529144436121,
-0.07758381217718124,
0.04358290508389473,
-0.015352199785411358,
-0.05125126987695694,
-0.035769023001194,
0.05944692716002464,
0.0056230537593364716,
-0.052783720195293427,
-0.021459897980093956,
0.028518591076135635,
0.10765049606561661,
0.03970014303922653,
0.009995406493544579,
0.10156456381082535,
-0.08374262601137161,
-0.015427341684699059,
0.025234950706362724,
-0.11452373117208481,
0.032564375549554825,
0.090106301009655,
-0.08479104936122894,
0.028084177523851395,
0.057800937443971634,
-0.0172318946570158,
-0.055535197257995605,
0.06239715963602066,
-0.058873776346445084,
-0.06587640196084976,
-0.062034983187913895,
-0.08444581925868988,
0.031701479107141495,
-0.065509133040905,
-0.02571795880794525,
-0.07888376712799072,
-0.1574874371290207,
-0.05801070109009743,
0.12264423817396164,
-0.04630960896611214,
-0.036022551357746124,
-0.07683148980140686,
-0.16433939337730408,
0.07570064067840576,
0.004258337430655956,
0.12805666029453278,
-0.0751006156206131,
0.07046415656805038,
-0.03904794901609421,
0.03906114399433136,
0.022071603685617447,
0.007798869162797928,
-0.05935626104474068,
0.0584951750934124,
-0.13332504034042358,
0.13161419332027435,
-0.09095931053161621,
0.027909351512789726,
-0.1542016863822937,
-0.06094963848590851,
0.01079085934907198,
0.009675397537648678,
0.09887724369764328,
0.1546856015920639,
-0.15253716707229614,
-0.07401830703020096,
0.12256859242916107,
-0.10672317445278168,
-0.07384726405143738,
0.07917891442775726,
-0.012021955102682114,
-0.011851084418594837,
0.06954819709062576,
0.07209508121013641,
0.104457788169384,
-0.06935146450996399,
-0.01718594878911972,
-0.0506780743598938,
0.030515795573592186,
0.09199843555688858,
0.04106508567929268,
-0.06792784482240677,
-0.08101152628660202,
0.03724126145243645,
0.002919144928455353,
0.006742360536009073,
-0.06940028816461563,
-0.061783790588378906,
-0.027516648173332214,
-0.08238843083381653,
0.09475986659526825,
0.022449536249041557,
-0.004209074191749096,
-0.07018816471099854,
-0.1933806985616684,
-0.022881008684635162,
0.06661733239889145,
-0.06969315558671951,
0.010490902699530125,
-0.07462435960769653,
0.05120667815208435,
0.09373423457145691,
0.02155151031911373,
-0.14246748387813568,
-0.13069948554039001,
0.015210377052426338,
-0.042149946093559265,
-0.008440246805548668,
-0.06855642795562744,
0.06635591387748718,
0.05417652800679207,
-0.05759124085307121,
-0.027404535561800003,
-0.02800898440182209,
0.02025485597550869,
-0.02721763774752617,
-0.26872459053993225,
-0.058729566633701324,
-0.011923776008188725,
0.18072675168514252,
-0.26490461826324463,
0.005054426845163107,
0.08084863424301147,
0.14465072751045227,
0.030276553705334663,
-0.04329417645931244,
-0.014730988070368767,
0.05289652571082115,
-0.006770472042262554,
-0.06362362951040268,
0.02843535505235195,
0.017292680218815804,
-0.13861031830310822,
-0.04807587340474129,
-0.17159773409366608,
0.06016549840569496,
0.09692694991827011,
0.002966685453429818,
-0.16478325426578522,
-0.00867553148418665,
-0.0331379771232605,
-0.05325181037187576,
0.042135290801525116,
0.043431494385004044,
0.13927626609802246,
0.05711320787668228,
0.09100475162267685,
-0.019104402512311935,
-0.027882270514965057,
0.015471478924155235,
-0.013390800915658474,
-0.0021640383638441563,
0.1343672126531601,
-0.015984635800123215,
-0.10973827540874481,
0.08123769611120224,
0.06032847613096237,
-0.11162757873535156,
0.13693729043006897,
-0.06132357940077782,
-0.06375359743833542,
-0.087253637611866,
0.07113740593194962,
0.052047017961740494,
0.05695854127407074,
-0.08680343627929688,
0.0485379621386528,
0.011476385407149792,
0.0021183653734624386,
-0.000523565337061882,
-0.10693252831697464,
0.03754459321498871,
0.003077182685956359,
-0.060865841805934906,
0.05151806399226189,
0.00038067111745476723,
0.00551837682723999,
0.08887428790330887,
0.03946881741285324,
-0.04492146149277687,
0.041801005601882935,
-0.02420257404446602,
-0.0881066620349884,
0.2293362319469452,
-0.1276639997959137,
-0.0926647037267685,
-0.07843747735023499,
0.0057018655352294445,
-0.05680091679096222,
-0.03018694929778576,
0.0065840259194374084,
-0.06628937274217606,
-0.047692518681287766,
-0.06302697211503983,
-0.028860216960310936,
0.0013835231075063348,
0.019232740625739098,
-0.011993651278316975,
0.014726412482559681,
0.11289837211370468,
-0.08839714527130127,
-0.016271771863102913,
0.006416893098503351,
-0.09445807337760925,
0.011433646082878113,
0.05033447593450546,
0.004934162832796574,
0.13288848102092743,
0.026180267333984375,
0.009252406656742096,
-0.021420439705252647,
0.2105562537908554,
-0.06717391312122345,
0.04773005098104477,
0.07093477994203568,
-0.05087776109576225,
0.06354302167892456,
0.1706678718328476,
0.05379331111907959,
-0.08311811089515686,
0.02660534344613552,
0.07524101436138153,
-0.00040613007149659097,
-0.22398391366004944,
-0.04422977194190025,
-0.041478805243968964,
-0.04513724893331528,
0.07550368458032608,
0.05418035760521889,
0.12401687353849411,
0.016289252787828445,
-0.024998703971505165,
0.0009499165462329984,
0.07212391495704651,
0.07447830587625504,
0.10890462249517441,
0.10694821178913116,
0.08976416289806366,
-0.015299616381525993,
0.036251701414585114,
0.030860772356390953,
-0.006241978611797094,
0.17833079397678375,
-0.02022179216146469,
0.11943928897380829,
0.0993766114115715,
0.0972047746181488,
-0.007520626299083233,
-0.0018846032908186316,
0.03346553072333336,
0.01749936118721962,
0.03364110738039017,
-0.07682014256715775,
-0.06808675080537796,
0.037401165813207626,
0.06545746326446533,
0.03770957142114639,
-0.09048419445753098,
0.026471193879842758,
0.054647620767354965,
0.2607384920120239,
0.11757460236549377,
-0.2927078306674957,
-0.13369449973106384,
-0.010421627201139927,
-0.013984099961817265,
-0.041940826922655106,
-0.010972685180604458,
0.07648120075464249,
-0.07865628600120544,
0.10035847872495651,
-0.026899009943008423,
0.05439528077840805,
-0.08074944466352463,
0.04460825398564339,
0.07831909507513046,
0.11548767238855362,
0.01950855553150177,
0.0024744970723986626,
-0.2970165014266968,
0.26368269324302673,
0.009348372928798199,
0.12759481370449066,
-0.04041832312941551,
0.06025071069598198,
0.04968935623764992,
-0.03575989603996277,
0.05704621970653534,
-0.016153939068317413,
-0.11058373004198074,
-0.16791576147079468,
-0.07651098072528839,
0.022853873670101166,
0.1303364634513855,
-0.08220363408327103,
0.0950273647904396,
-0.02397816814482212,
-0.017716599628329277,
0.036173656582832336,
0.019139019772410393,
-0.18432503938674927,
-0.09308536350727081,
0.07015189528465271,
-0.008567487820982933,
0.044698189944028854,
-0.06371456384658813,
-0.05578009411692619,
-0.06265899538993835,
0.22775186598300934,
-0.17951135337352753,
-0.0651492327451706,
-0.142728790640831,
0.03922715038061142,
0.1350586861371994,
-0.08605322986841202,
0.05050193518400192,
-0.012000405229628086,
0.07246574014425278,
0.06215884909033775,
-0.05459899827837944,
0.115663543343544,
-0.031810514628887177,
-0.22039134800434113,
-0.06055517494678497,
0.09731820970773697,
0.07084718346595764,
0.01937842182815075,
-0.0042501590214669704,
0.0669436827301979,
0.04455922171473503,
-0.10229003429412842,
0.0874878317117691,
0.03205850347876549,
0.018493136391043663,
0.041590552777051926,
-0.02948184125125408,
-0.048431646078825,
-0.030308762565255165,
0.0076976739801466465,
0.05859321728348732,
0.3140515983104706,
-0.09070272743701935,
0.034042537212371826,
0.09521616995334625,
-0.07046692818403244,
-0.17356394231319427,
-0.006680215708911419,
0.13296113908290863,
-0.004646994639188051,
-0.07823114097118378,
-0.21852193772792816,
0.06953698396682739,
0.12080758064985275,
-0.013829289004206657,
0.09590008109807968,
-0.2447979599237442,
-0.14209066331386566,
0.08231668174266815,
0.0815940573811531,
-0.03857085853815079,
-0.16681726276874542,
-0.07056432217359543,
-0.047052543610334396,
-0.08460727334022522,
0.15402837097644806,
-0.04928092285990715,
0.0819680318236351,
0.03724382072687149,
-0.004254519939422607,
0.02519821748137474,
-0.030049530789256096,
0.15750956535339355,
-0.015537654049694538,
0.06770665943622589,
-0.05071232095360756,
0.02815626747906208,
0.018871862441301346,
-0.10594703257083893,
0.019625592976808548,
-0.11501972377300262,
0.04249830171465874,
-0.0826103463768959,
0.00012714287731796503,
-0.07897641509771347,
0.08673817664384842,
-0.08220413327217102,
-0.014596700668334961,
-0.0013707401230931282,
0.07442373037338257,
0.10682099312543869,
0.00533792469650507,
0.053057484328746796,
-0.02807515300810337,
0.2158629596233368,
0.1617993861436844,
0.073032446205616,
0.028062062337994576,
-0.06536154448986053,
0.03881029039621353,
-0.030052099376916885,
0.06252432614564896,
-0.1275167316198349,
0.0387142114341259,
0.123994380235672,
0.02323884516954422,
0.12075649201869965,
0.049789875745773315,
-0.048530302941799164,
-0.012075965292751789,
0.0664636567234993,
-0.09894316643476486,
-0.07814684510231018,
0.01425154972821474,
-0.012370428070425987,
-0.08920513093471527,
0.0005565567407757044,
0.15239904820919037,
-0.033993449062108994,
0.03628222271800041,
0.021939709782600403,
0.05268024653196335,
-0.06364664435386658,
0.12484405189752579,
0.005491953343153,
0.11271296441555023,
-0.07863834500312805,
0.13116511702537537,
0.07640094310045242,
-0.11917561292648315,
0.11707223206758499,
0.05587172135710716,
-0.07148625701665878,
-0.04894670844078064,
0.024542229250073433,
0.08567873388528824,
0.09113216400146484,
-0.05021972209215164,
-0.050449784845113754,
-0.12880057096481323,
0.08871784806251526,
0.17617890238761902,
0.002718881005421281,
0.0863204300403595,
-0.017054861411452293,
0.011620590463280678,
-0.08290526270866394,
0.0745454877614975,
0.058593761175870895,
0.043204423040151596,
-0.08457542955875397,
0.17170800268650055,
0.006209321320056915,
-0.0507316067814827,
0.009813275188207626,
-0.014302941039204597,
-0.18529678881168365,
-0.027823910117149353,
-0.10875258594751358,
0.03620944917201996,
-0.018978621810674667,
0.0094841867685318,
0.03710293397307396,
-0.034935545176267624,
-0.04796107858419418,
0.020561950281262398,
-0.10608059912919998,
-0.07428860664367676,
0.04040435329079628,
0.0926278606057167,
-0.14506544172763824,
-0.042945828288793564,
0.025781573727726936,
-0.1324933022260666,
0.07118277996778488,
0.0373290553689003,
-0.0031229674350470304,
0.03300600126385689,
-0.1545829027891159,
0.02664131298661232,
-0.006980502046644688,
-0.013705425895750523,
0.007755654864013195,
-0.15918941795825958,
0.0023977283854037523,
-0.05322868004441261,
0.009209136478602886,
0.022605564445257187,
0.07240604609251022,
-0.09304455667734146,
-0.037138208746910095,
-0.016654957085847855,
-0.0015765461139380932,
-0.05161917582154274,
0.029863419011235237,
0.12664960324764252,
-0.03783375397324562,
0.16191606223583221,
-0.09339803457260132,
0.0441141314804554,
-0.17415684461593628,
-0.025471407920122147,
0.02559332735836506,
-0.03889256715774536,
-0.10735508054494858,
-0.006795684341341257,
0.11443231254816055,
-0.10092499852180481,
0.031737007200717926,
-0.09222804009914398,
0.037294019013643265,
0.014958903193473816,
-0.10346110165119171,
-0.10478390753269196,
0.08805189281702042,
0.18893401324748993,
0.09286988526582718,
0.0014396719634532928,
0.07267952710390091,
-0.037110112607479095,
0.015873167663812637,
0.033137962222099304,
0.19853626191616058,
0.11399293690919876,
0.021361270919442177,
0.09164057672023773,
0.04319761320948601,
-0.13285666704177856,
-0.06859913468360901,
0.17175379395484924,
-0.0830727145075798,
0.16983410716056824,
-0.056549981236457825,
0.09011861681938171,
0.04004810005426407,
-0.17918148636817932,
0.040777336806058884,
-0.046810492873191833,
-0.09339452534914017,
-0.09630124270915985,
-0.14994333684444427,
-0.07957693934440613,
-0.09248071163892746,
0.002157134935259819,
-0.09980284422636032,
0.015859724953770638,
0.08580194413661957,
0.04086297005414963,
0.034477349370718,
0.044232890009880066,
-0.040774207562208176,
0.030629940330982208,
0.11298445612192154,
0.006529608275741339,
-0.030205674469470978,
-0.03780115023255348,
-0.09297715127468109,
0.020589718595147133,
0.04177331551909447,
0.0343688428401947,
-0.00012334510392975062,
-0.02034362033009529,
0.08259183913469315,
0.01806933805346489,
-0.09557189792394638,
0.08627794682979584,
0.008683317340910435,
0.009265393018722534,
0.08977437764406204,
0.03276010602712631,
-0.035340458154678345,
-0.011576610617339611,
0.15672621130943298,
-0.07640812546014786,
-0.08460778743028641,
-0.1713714599609375,
0.23124192655086517,
-0.03135822340846062,
0.014993483200669289,
0.017140846699476242,
-0.07488761842250824,
-0.034254834055900574,
0.11129958182573318,
0.13536956906318665,
-0.029489878565073013,
0.00008809855353320017,
0.08123847097158432,
-0.011780002154409885,
-0.0319480299949646,
0.12253086268901825,
0.04893754422664642,
0.009892885573208332,
-0.03199268877506256,
-0.006256544031202793,
0.005463073030114174,
-0.05739268288016319,
-0.07061919569969177,
0.07457295060157776,
-0.03491155430674553,
-0.032347533851861954,
-0.0245839674025774,
0.07599522173404694,
-0.14429326355457306,
-0.1520582139492035,
0.1002262532711029,
-0.18077437579631805,
-0.18114034831523895,
-0.04748070240020752,
0.00958032626658678,
0.0331699401140213,
0.044180262833833694,
0.015706012025475502,
-0.02719859778881073,
0.13212773203849792,
-0.0417213998734951,
0.006878875195980072,
-0.10780901461839676,
0.008573179133236408,
0.03787022829055786,
0.2090098112821579,
-0.0038904992397874594,
0.040113549679517746,
0.14979024231433868,
0.047674309462308884,
-0.0986112654209137,
0.027917660772800446,
0.09685686230659485,
-0.14870934188365936,
0.043287791311740875,
0.09395460784435272,
-0.034360066056251526,
0.1481504887342453,
0.09987268596887589,
-0.09877873957157135,
0.004653590731322765,
0.010362150147557259,
-0.04696112126111984,
-0.028050586581230164,
-0.022764889523386955,
-0.04499227553606033,
0.1221206933259964,
0.24752341210842133,
-0.04313002899289131,
-0.023327656090259552,
-0.03431860730051994,
0.052784230560064316,
0.01803453080356121,
0.019214827567338943,
-0.07638544589281082,
-0.22635327279567719,
0.09481967240571976,
0.02305275946855545,
0.0736030861735344,
-0.11576614528894424,
-0.07724875956773758,
0.035714030265808105,
-0.008900117129087448,
-0.07813168317079544,
0.11999951303005219,
0.04976648837327957,
0.03153948485851288,
-0.046411193907260895,
-0.1314178854227066,
-0.026532944291830063,
0.16840007901191711,
-0.13309423625469208,
-0.07632013410329819
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | omartariq612/quran-lora-whisper-medium-epoch-1 | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-07T07:39:58+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
31,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06646376848220825,
0.2168014943599701,
-0.00225935154594481,
0.023818302899599075,
0.1271018385887146,
-0.001635765191167593,
0.04218708351254463,
0.13324736058712006,
-0.020175931975245476,
0.11144465953111649,
0.046588581055402756,
0.09377603232860565,
0.09928803145885468,
0.18404334783554077,
0.04859916493296623,
-0.2059975117444992,
0.007056170143187046,
-0.09090408682823181,
0.014076028019189835,
0.1116579994559288,
0.13719257712364197,
-0.10291384905576706,
0.08272874355316162,
-0.04045208916068077,
-0.02019004337489605,
0.00012576708104461432,
-0.09259183704853058,
-0.07032395154237747,
0.06885425746440887,
0.06264153122901917,
0.051234472543001175,
0.001456156256608665,
0.09140396863222122,
-0.2864592671394348,
0.017265573143959045,
0.08406311273574829,
0.0027674848679453135,
0.06290827691555023,
0.07236549258232117,
-0.07389893382787704,
0.11328595131635666,
-0.08021481335163116,
0.13019037246704102,
0.08625296503305435,
-0.062064990401268005,
-0.23071379959583282,
-0.07525765895843506,
0.0963398814201355,
0.12251301854848862,
0.06215599179267883,
-0.022921854630112648,
0.15455181896686554,
-0.06248689442873001,
0.012971068732440472,
0.1294165402650833,
-0.11526761949062347,
-0.05572471022605896,
0.061741601675748825,
0.11775490641593933,
0.10740239918231964,
-0.14110268652439117,
-0.0017287094378843904,
0.04900608956813812,
0.029121357947587967,
0.08589313924312592,
0.022661056369543076,
0.12003941088914871,
0.04652795568108559,
-0.13695219159126282,
-0.04037507623434067,
0.12011898308992386,
0.038862764835357666,
-0.06446044892072678,
-0.2168138176202774,
-0.006778308190405369,
-0.0601806715130806,
-0.014732478186488152,
-0.07019448280334473,
0.039128515869379044,
-0.02470310963690281,
0.07317749410867691,
-0.04465159401297569,
-0.1063927412033081,
-0.0421026237308979,
0.0892222449183464,
0.07748593389987946,
0.011527054943144321,
-0.02519804798066616,
0.04627908393740654,
0.13455867767333984,
0.05402068421244621,
-0.10399353504180908,
-0.07017925381660461,
-0.06942764669656754,
-0.09420394152402878,
-0.04035796597599983,
0.056760527193546295,
0.031942449510097504,
0.02665667235851288,
0.22703726589679718,
0.016653569415211678,
0.04155244305729866,
0.0224777739495039,
0.01032855175435543,
0.043662428855895996,
0.0955500528216362,
-0.05303520709276199,
-0.15660029649734497,
-0.04072032496333122,
0.09077946096658707,
-0.0027527001220732927,
-0.036689214408397675,
-0.03966725245118141,
0.03849169611930847,
0.06843466311693192,
0.13122352957725525,
0.07552056759595871,
-0.017929591238498688,
-0.04813180863857269,
-0.030096933245658875,
0.23523783683776855,
-0.1493375599384308,
0.04426715523004532,
-0.02271856553852558,
-0.01804111897945404,
-0.03908449783921242,
0.03597262129187584,
0.022118929773569107,
-0.000004518366949923802,
0.09706240892410278,
-0.058981191366910934,
-0.05378659814596176,
-0.10168042778968811,
-0.03272576630115509,
0.04088849574327469,
-0.013975566253066063,
-0.010589460842311382,
-0.09025166928768158,
-0.09490354359149933,
-0.04766594246029854,
0.05537205561995506,
-0.05123869329690933,
-0.03770573064684868,
0.009465423412621021,
-0.08151785284280777,
-0.005444355774670839,
-0.005417742300778627,
0.10699385404586792,
-0.03222226724028587,
0.04445803165435791,
-0.027600755915045738,
0.05225523188710213,
0.09919606149196625,
0.031576547771692276,
-0.0773419588804245,
0.0561848059296608,
-0.22559374570846558,
0.07503069192171097,
-0.11481974273920059,
0.04335082694888115,
-0.1704932004213333,
-0.042439818382263184,
0.005444696638733149,
0.0139949731528759,
0.013206101022660732,
0.12720820307731628,
-0.19255615770816803,
-0.01654396951198578,
0.13260798156261444,
-0.09212633967399597,
-0.118110790848732,
0.07884611934423447,
-0.029701577499508858,
0.1624738723039627,
0.04682036489248276,
-0.027025915682315826,
0.09224298596382141,
-0.16434773802757263,
-0.07092688232660294,
-0.00949116237461567,
-0.01727987825870514,
0.12109188735485077,
0.07512219995260239,
-0.05991523340344429,
0.046571120619773865,
0.02832140028476715,
-0.038078423589468,
-0.04424772411584854,
-0.050857074558734894,
-0.10884185880422592,
-0.01070026308298111,
-0.08987759798765182,
0.04065500199794769,
-0.01250192429870367,
-0.07916021347045898,
-0.029885273426771164,
-0.18612512946128845,
-0.0030564051121473312,
0.10038342326879501,
0.0035033065360039473,
-0.005652366206049919,
-0.08666291832923889,
0.026358824223279953,
-0.03112892620265484,
-0.008404186926782131,
-0.16764774918556213,
-0.04399421438574791,
0.046902090311050415,
-0.16094985604286194,
0.020117372274398804,
-0.06413903087377548,
0.06334125250577927,
0.03641495108604431,
-0.05590536445379257,
-0.0248766727745533,
-0.01730942726135254,
0.011945613659918308,
-0.05083848536014557,
-0.18994836509227753,
-0.056277405470609665,
-0.037882111966609955,
0.149809330701828,
-0.25956398248672485,
0.032966937869787216,
0.051140617579221725,
0.14649195969104767,
0.00406361510977149,
-0.05115427449345589,
0.01429014839231968,
-0.05360214412212372,
-0.054652128368616104,
-0.06746816635131836,
-0.006135428790003061,
-0.027576493099331856,
-0.05147203803062439,
0.019243421033024788,
-0.1755700707435608,
-0.021410830318927765,
0.09424154460430145,
0.12876708805561066,
-0.1486445665359497,
-0.018640631809830666,
-0.048725154250860214,
-0.06339836865663528,
-0.0715010017156601,
-0.07038594037294388,
0.10712739825248718,
0.0513901449739933,
0.04796046018600464,
-0.07435787469148636,
-0.07092321664094925,
0.02726263552904129,
0.006906150374561548,
-0.03382374346256256,
0.08727246522903442,
0.05199531093239784,
-0.09209315478801727,
0.0756213590502739,
0.1092359870672226,
0.07177663594484329,
0.09363535046577454,
0.01574566215276718,
-0.11756632477045059,
-0.028492970392107964,
0.036266472190618515,
0.02740776725113392,
0.1465986967086792,
-0.05952361226081848,
0.04016614332795143,
0.04494241625070572,
-0.04170418903231621,
0.022319864481687546,
-0.08787637203931808,
0.024075502529740334,
0.025203049182891846,
-0.0034381982404738665,
0.06284574419260025,
-0.02525499276816845,
-0.0050758360885083675,
0.07016654312610626,
0.047779910266399384,
0.04621000960469246,
0.009655474685132504,
-0.01720241829752922,
-0.1047825813293457,
0.16950392723083496,
-0.0951867327094078,
-0.269941508769989,
-0.17632324993610382,
0.026197833940386772,
0.04035249724984169,
-0.022378476336598396,
0.031619444489479065,
-0.07056326419115067,
-0.10630585998296738,
-0.1060405746102333,
-0.002429972169920802,
0.01714223250746727,
-0.06364088505506516,
-0.0741225928068161,
0.07348573952913284,
0.04382912442088127,
-0.14902326464653015,
0.038552410900592804,
0.055694397538900375,
-0.057955220341682434,
-0.0233661737293005,
0.09118817001581192,
0.12397737801074982,
0.14583967626094818,
-0.021366750821471214,
-0.028626007959246635,
0.029004426673054695,
0.19620531797409058,
-0.13469526171684265,
0.10371150821447372,
0.13814030587673187,
-0.04545360431075096,
0.08360563963651657,
0.1560150384902954,
0.029186224564909935,
-0.08317049592733383,
0.05044832453131676,
0.04082648828625679,
-0.043159641325473785,
-0.2666129767894745,
-0.0534592866897583,
0.012832709588110447,
-0.06255637854337692,
0.09786593168973923,
0.10183793306350708,
0.11542957276105881,
0.034910861402750015,
-0.07166364789009094,
-0.043925940990448,
-0.0058974819257855415,
0.11737963557243347,
-0.05490213260054588,
-0.012639665976166725,
0.07686592638492584,
-0.05086168646812439,
0.005355054512619972,
0.10266812145709991,
0.02973790094256401,
0.17442677915096283,
0.020399179309606552,
0.11231429129838943,
0.06195578724145889,
0.08633565157651901,
0.0007386076031252742,
0.02951662428677082,
0.05147615820169449,
0.017203815281391144,
-0.002300140680745244,
-0.10421168059110641,
-0.006156572140753269,
0.1449710875749588,
0.028103826567530632,
0.029669636860489845,
-0.0018948549404740334,
-0.005003341939300299,
0.05121048167347908,
0.1746254414319992,
-0.011592294089496136,
-0.22072425484657288,
-0.0845772922039032,
0.06936841458082199,
-0.06218599155545235,
-0.12968985736370087,
-0.026130788028240204,
0.045467354357242584,
-0.17519839107990265,
0.026703642681241035,
-0.027433741837739944,
0.0919293761253357,
-0.09345759451389313,
-0.02221956104040146,
0.03687324374914169,
0.084866963326931,
-0.014529162086546421,
0.08703910559415817,
-0.14498743414878845,
0.11886418610811234,
0.02978132851421833,
0.09024628251791,
-0.11081171780824661,
0.07909037172794342,
-0.007550720125436783,
0.009180475026369095,
0.19379350543022156,
-0.011335089802742004,
-0.03514958545565605,
-0.08774717897176743,
-0.11210042238235474,
-0.013537433929741383,
0.12687496840953827,
-0.1243172138929367,
0.08773399889469147,
-0.015198243781924248,
-0.044079482555389404,
0.00937260314822197,
-0.12100647389888763,
-0.17273177206516266,
-0.19628387689590454,
0.05585884302854538,
-0.09575839340686798,
0.025643249973654747,
-0.11914430558681488,
-0.07089093327522278,
-0.02952558360993862,
0.241120383143425,
-0.1745356321334839,
-0.06510113179683685,
-0.1468164622783661,
-0.046294767409563065,
0.1662203073501587,
-0.04437198117375374,
0.0718095526099205,
-0.0208172257989645,
0.20345525443553925,
0.005988610442727804,
-0.004939318168908358,
0.06724198162555695,
-0.08892562240362167,
-0.16873881220817566,
-0.06771010160446167,
0.1510489284992218,
0.11680185794830322,
0.04907919466495514,
-0.002248800592496991,
0.0011772146681323647,
-0.016943959519267082,
-0.1137804463505745,
-0.0033210667315870523,
0.16037839651107788,
0.03878779336810112,
0.025986969470977783,
-0.05243593826889992,
-0.08797456324100494,
-0.06899320334196091,
-0.06853509694337845,
0.06221301481127739,
0.19590823352336884,
-0.10376439243555069,
0.1700313836336136,
0.147536963224411,
-0.07305635511875153,
-0.23175598680973053,
0.035342130810022354,
0.04983805492520332,
0.0014306638622656465,
0.04886869341135025,
-0.18252557516098022,
0.10521943867206573,
0.019543392583727837,
-0.05505957826972008,
0.13485197722911835,
-0.1557481735944748,
-0.1552847921848297,
0.0722852572798729,
0.03904085233807564,
-0.22423844039440155,
-0.1354004591703415,
-0.09622503817081451,
-0.05825018882751465,
-0.14065024256706238,
0.06054598465561867,
-0.002136280992999673,
0.015948504209518433,
0.03500790148973465,
-0.0015643214574083686,
0.027123261243104935,
-0.058935679495334625,
0.18609118461608887,
-0.004065449349582195,
0.020676052197813988,
-0.060264769941568375,
-0.0478842556476593,
0.09839435666799545,
-0.06130504235625267,
0.12208222597837448,
0.004057085141539574,
0.01594383642077446,
-0.10362856835126877,
-0.048314861953258514,
-0.04328322783112526,
0.05154227837920189,
-0.07548051327466965,
-0.10070807486772537,
-0.043625857681035995,
0.08841723203659058,
0.07005169242620468,
-0.03383097052574158,
0.00549331633374095,
-0.07189501076936722,
0.10019614547491074,
0.17795267701148987,
0.17573626339435577,
0.009926567785441875,
-0.07241068035364151,
0.01677953451871872,
-0.04142116755247116,
0.044231921434402466,
-0.2513144314289093,
0.03756171092391014,
0.06098250672221184,
0.029438555240631104,
0.09217222779989243,
-0.020435843616724014,
-0.1820858269929886,
-0.04050002992153168,
0.08094815909862518,
-0.05452597141265869,
-0.22617179155349731,
-0.019085140898823738,
0.0954197570681572,
-0.2020406424999237,
-0.007372708059847355,
0.03995226323604584,
-0.048725228756666183,
-0.023169852793216705,
0.00010950004070764408,
0.06317184865474701,
0.002471912419423461,
0.09773622453212738,
0.0735151618719101,
0.09715340286493301,
-0.08337292820215225,
0.10562895983457565,
0.10150538384914398,
-0.09572599828243256,
0.03605884686112404,
0.06754924356937408,
-0.05300498008728027,
-0.043293699622154236,
0.03665391728281975,
0.033023297786712646,
0.005234600510448217,
-0.060321882367134094,
0.013913018628954887,
-0.036497246474027634,
0.044923391193151474,
0.08326134830713272,
0.03754979372024536,
-0.013354414142668247,
0.06462216377258301,
0.03401726484298706,
-0.10898099094629288,
0.10366570204496384,
0.01731540448963642,
0.04105307161808014,
-0.08384523540735245,
-0.019968897104263306,
0.035425446927547455,
0.030576206743717194,
-0.01765924133360386,
-0.02306121215224266,
-0.02860277332365513,
-0.01614218018949032,
-0.14299540221691132,
-0.023106401786208153,
-0.07243485748767853,
0.006181265693157911,
0.014656842686235905,
-0.031884219497442245,
-0.011233693920075893,
0.02475680410861969,
-0.06979699432849884,
-0.07426341623067856,
-0.006949664559215307,
0.09833318740129471,
-0.15115703642368317,
0.008848577737808228,
0.06907843053340912,
-0.11088496446609497,
0.08190931379795074,
-0.008411259390413761,
0.016245156526565552,
0.022527478635311127,
-0.15448406338691711,
0.05601610988378525,
0.0008648968650959432,
0.01916889287531376,
0.025886621326208115,
-0.16471809148788452,
0.004104440100491047,
-0.04661374166607857,
-0.02149827405810356,
-0.00004464812809601426,
-0.02647159807384014,
-0.12325995415449142,
0.06858719140291214,
-0.015622655861079693,
-0.035931166261434555,
-0.02701525390148163,
0.0539589487016201,
0.07888586074113846,
-0.027474910020828247,
0.10445091128349304,
-0.008690856397151947,
0.04941811040043831,
-0.16801609098911285,
-0.02470702864229679,
-0.04982255399227142,
0.019377702847123146,
0.009884213097393513,
-0.007693959400057793,
0.04183054715394974,
-0.00976533442735672,
0.21883612871170044,
-0.05075952783226967,
0.1607085019350052,
0.05847611650824547,
-0.017352959141135216,
-0.0007513365126214921,
0.06180921941995621,
0.05997028574347496,
0.04658793285489082,
0.009480604901909828,
0.023740366101264954,
-0.022450892254710197,
-0.006695089396089315,
-0.15932634472846985,
0.01890849508345127,
0.14999441802501678,
0.06301083415746689,
0.024745315313339233,
0.05866100639104843,
-0.12775006890296936,
-0.12135478109121323,
0.09311001747846603,
-0.026755332946777344,
0.00928465835750103,
-0.08245618641376495,
0.1358020007610321,
0.14980104565620422,
-0.14000412821769714,
0.05256148427724838,
-0.06134212389588356,
-0.05217423290014267,
-0.10388828068971634,
-0.12032219022512436,
-0.05887215584516525,
-0.053666237741708755,
0.002330566756427288,
-0.03760887682437897,
0.054546963423490524,
0.03344334661960602,
-0.009351172484457493,
-0.00022941511997487396,
0.13597318530082703,
-0.019751882180571556,
-0.0028988157864660025,
0.048313532024621964,
0.03693558648228645,
0.02373051457107067,
-0.05275435373187065,
0.02940409444272518,
0.02539868652820587,
0.032232340425252914,
0.06546790152788162,
0.033412106335163116,
-0.047448933124542236,
0.03804153576493263,
-0.0025254099164158106,
-0.11207924783229828,
0.019641218706965446,
-0.00460948096588254,
-0.0742158442735672,
0.1268945336341858,
0.0407399944961071,
0.010224059224128723,
-0.03741471841931343,
0.24361543357372284,
-0.06653323769569397,
-0.06378097087144852,
-0.13251738250255585,
0.10491154342889786,
-0.0027236645109951496,
0.06476365029811859,
0.023412218317389488,
-0.1284150779247284,
0.005243356805294752,
0.13858191668987274,
0.12181595712900162,
0.0045748427510261536,
0.009228081442415714,
0.0518609918653965,
0.0025186820421367884,
-0.06998204439878464,
0.054019294679164886,
0.06992026418447495,
0.12919506430625916,
-0.07847554981708527,
0.07680778950452805,
0.0006860480643808842,
-0.08370215445756912,
-0.02947772853076458,
0.11312682181596756,
-0.0409729965031147,
0.03491825982928276,
-0.047444481402635574,
0.10916327685117722,
-0.05787910893559456,
-0.29412412643432617,
0.02350960113108158,
-0.09588567912578583,
-0.15202060341835022,
-0.018367812037467957,
0.05944539234042168,
-0.02624768204987049,
0.018029648810625076,
0.06971040368080139,
-0.06011629104614258,
0.20098382234573364,
0.0335683599114418,
-0.07864278554916382,
-0.0664360448718071,
0.04837050288915634,
-0.06564252078533173,
0.2949807047843933,
0.008418165147304535,
0.02863333560526371,
0.10770907253026962,
-0.03253700211644173,
-0.18271861970424652,
0.010723991319537163,
0.1133992001414299,
-0.08056149631738663,
0.08200647681951523,
0.19000613689422607,
-0.012578671798110008,
0.1209007054567337,
0.05294662341475487,
-0.047376248985528946,
0.04217283055186272,
-0.03389401361346245,
-0.051268599927425385,
-0.10752558708190918,
0.058453381061553955,
-0.05909625440835953,
0.15447644889354706,
0.10152646154165268,
-0.05671518296003342,
-0.004550917539745569,
-0.05555408447980881,
0.04875178262591362,
0.01804669201374054,
0.12263146042823792,
0.02951994352042675,
-0.1865430772304535,
0.032826557755470276,
-0.01144319772720337,
0.10186848044395447,
-0.25588861107826233,
-0.08421015739440918,
0.08833149075508118,
-0.011924264021217823,
-0.05105875805020332,
0.10560628771781921,
0.057650718837976456,
0.04243382066488266,
-0.043439045548439026,
-0.10480839014053345,
-0.02186836116015911,
0.14663739502429962,
-0.1469624787569046,
-0.025013303384184837
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# OndeviceAI-base-v1
This model is a fine-tuned version of [paust/pko-t5-base](https://huggingface.co/paust/pko-t5-base) on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0007
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- gradient_accumulation_steps: 8
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_ratio: 0.06
- num_epochs: 20
### Training results
### Framework versions
- Transformers 4.37.2
- Pytorch 2.2.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "cc-by-4.0", "tags": ["generated_from_trainer"], "base_model": "paust/pko-t5-base", "model-index": [{"name": "OndeviceAI-base-v1", "results": []}]} | text2text-generation | yeye776/OndeviceAI-base-v1 | [
"transformers",
"tensorboard",
"safetensors",
"t5",
"text2text-generation",
"generated_from_trainer",
"base_model:paust/pko-t5-base",
"license:cc-by-4.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T07:40:06+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-paust/pko-t5-base #license-cc-by-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# OndeviceAI-base-v1
This model is a fine-tuned version of paust/pko-t5-base on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0007
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- gradient_accumulation_steps: 8
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_ratio: 0.06
- num_epochs: 20
### Training results
### Framework versions
- Transformers 4.37.2
- Pytorch 2.2.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| [
"# OndeviceAI-base-v1\n\nThis model is a fine-tuned version of paust/pko-t5-base on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0007\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- gradient_accumulation_steps: 8\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_ratio: 0.06\n- num_epochs: 20",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-paust/pko-t5-base #license-cc-by-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# OndeviceAI-base-v1\n\nThis model is a fine-tuned version of paust/pko-t5-base on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0007\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- gradient_accumulation_steps: 8\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_ratio: 0.06\n- num_epochs: 20",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
83,
36,
6,
12,
8,
3,
129,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-paust/pko-t5-base #license-cc-by-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# OndeviceAI-base-v1\n\nThis model is a fine-tuned version of paust/pko-t5-base on the None dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0007\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- gradient_accumulation_steps: 8\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_ratio: 0.06\n- num_epochs: 20### Training results### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.09789319336414337,
0.15943197906017303,
-0.0028766635805368423,
0.07127537578344345,
0.13572563230991364,
0.024393955245614052,
0.11200148612260818,
0.1382688730955124,
-0.12967827916145325,
0.0972563698887825,
0.08155684173107147,
0.0254642553627491,
0.07377476245164871,
0.11654738336801529,
0.0022636547219008207,
-0.2536686062812805,
-0.00408149603754282,
-0.009592100977897644,
-0.08014776557683945,
0.11375895142555237,
0.10719998925924301,
-0.08740157634019852,
0.06363346427679062,
0.012315320782363415,
-0.140313059091568,
0.011337918229401112,
-0.05180791765451431,
-0.06769638508558273,
0.08390724658966064,
0.009217645041644573,
0.08650461584329605,
0.015375218354165554,
0.10686202347278595,
-0.22104771435260773,
-0.005861361976712942,
0.08780699223279953,
0.0187531765550375,
0.09326615184545517,
0.08845113217830658,
-0.0035388099495321512,
0.1255171298980713,
-0.167068213224411,
0.0942421704530716,
0.03902198374271393,
-0.07080760598182678,
-0.1423795074224472,
-0.07974253594875336,
0.10438116639852524,
0.09811804443597794,
0.10546576231718063,
-0.0035313009284436703,
0.15996572375297546,
-0.07596287876367569,
0.07111184298992157,
0.19302639365196228,
-0.28212040662765503,
-0.06907640397548676,
0.033307891339063644,
0.06615542620420456,
0.07146983593702316,
-0.12760576605796814,
-0.012292159721255302,
0.04211704432964325,
0.01663655787706375,
0.08458695560693741,
0.009839198552072048,
0.02502480335533619,
-0.007843540981411934,
-0.11180007457733154,
-0.029096199199557304,
0.14342132210731506,
0.06854557245969772,
-0.044020477682352066,
-0.1336682289838791,
-0.04384828731417656,
-0.15622849762439728,
-0.010451104491949081,
-0.027212943881750107,
0.029279101639986038,
-0.045033328235149384,
-0.09065382182598114,
-0.018481673672795296,
-0.08028330653905869,
-0.05719621852040291,
0.04207056388258934,
0.11270313709974289,
0.03910226747393608,
-0.008048759773373604,
0.0006385837914422154,
0.10034514218568802,
0.040096260607242584,
-0.1394103765487671,
-0.009033811278641224,
-0.0009728369186632335,
-0.09562864899635315,
-0.057341672480106354,
-0.022829927504062653,
-0.0555763840675354,
-0.0029601070564240217,
0.1324797123670578,
-0.04748190939426422,
0.05707107111811638,
0.016518523916602135,
-0.005190867464989424,
-0.02397189848124981,
0.1409701108932495,
-0.06323353946208954,
-0.08530455827713013,
-0.0163392573595047,
0.1068616509437561,
0.023907750844955444,
-0.00775727117434144,
-0.07580266147851944,
-0.035301681607961655,
0.1090736836194992,
0.07659545540809631,
-0.020562944933772087,
0.015690583735704422,
-0.03930845111608505,
-0.03077886626124382,
0.08630228042602539,
-0.11659464240074158,
0.04962025210261345,
-0.016101857647299767,
-0.056333839893341064,
-0.01948937214910984,
0.00577381020411849,
0.0000950446046772413,
-0.04664907604455948,
0.08818676322698593,
-0.09080047160387039,
-0.02267061360180378,
-0.060203272849321365,
-0.04490486904978752,
0.023786751553416252,
-0.047869112342596054,
-0.009390603750944138,
-0.07948488742113113,
-0.14358308911323547,
-0.044919535517692566,
0.02339831180870533,
-0.07735032588243484,
-0.0616399347782135,
-0.012740795500576496,
-0.07627757638692856,
0.025310223922133446,
0.004932533483952284,
0.1290275752544403,
-0.0369110107421875,
0.08057023584842682,
0.009161220863461494,
0.0235750749707222,
0.08015516400337219,
0.03444686532020569,
-0.09116507321596146,
0.03843648359179497,
-0.09565873444080353,
0.06304187327623367,
-0.06410738080739975,
0.025608928874135017,
-0.13029922544956207,
-0.11020004004240036,
-0.04807284474372864,
-0.03509768471121788,
0.060768451541662216,
0.1253393143415451,
-0.13546210527420044,
-0.03866516798734665,
0.17558860778808594,
-0.07182565331459045,
-0.1042453944683075,
0.10832878947257996,
-0.007211374118924141,
-0.015566084533929825,
0.04676676169037819,
0.0890994668006897,
0.09718796610832214,
-0.11709244549274445,
-0.04903562366962433,
0.0056229326874017715,
0.06603864580392838,
0.016841597855091095,
0.10606267303228378,
-0.01832648552954197,
0.04618539288640022,
0.00027905820752494037,
-0.019261395558714867,
0.003899285802617669,
-0.06651525944471359,
-0.071647509932518,
-0.04902581498026848,
-0.08865808695554733,
0.03777434304356575,
0.04217962920665741,
0.031023263931274414,
-0.08240064233541489,
-0.15861637890338898,
0.07421762496232986,
0.12272337079048157,
-0.05104411020874977,
0.014817443676292896,
-0.07063471525907516,
0.023841360583901405,
-0.0654321163892746,
-0.020535266026854515,
-0.18109412491321564,
-0.08885814994573593,
0.04766704514622688,
-0.06782618910074234,
0.011819512583315372,
-0.003939415793865919,
0.06946997344493866,
0.08395326882600784,
-0.07257808744907379,
-0.037093352526426315,
-0.08965451270341873,
-0.0018892985535785556,
-0.09323370456695557,
-0.1716402769088745,
-0.05784084275364876,
-0.022709839046001434,
0.18413212895393372,
-0.22430047392845154,
0.01931462436914444,
0.01611008122563362,
0.16882163286209106,
0.024408496916294098,
-0.06821152567863464,
0.00897327158600092,
0.03771018236875534,
-0.01761951483786106,
-0.10603264719247818,
0.019390901550650597,
0.005685635842382908,
-0.10399734228849411,
-0.032920949161052704,
-0.1659257709980011,
0.0781860500574112,
0.0943002700805664,
0.09745599329471588,
-0.1029156744480133,
-0.03814370930194855,
-0.07192826271057129,
-0.05432991683483124,
-0.06151635944843292,
-0.024334751069545746,
0.15756075084209442,
0.03565144166350365,
0.12166164815425873,
-0.06525657325983047,
-0.08015389740467072,
0.012054392136633396,
0.03263882175087929,
-0.035918544977903366,
0.07918603718280792,
0.036694593727588654,
-0.14100992679595947,
0.10104846209287643,
0.10166462510824203,
-0.03580509498715401,
0.1322897970676422,
-0.04990016669034958,
-0.1095067709684372,
-0.0304923914372921,
0.02354763261973858,
0.023647675290703773,
0.10919477790594101,
-0.08378438651561737,
0.009046663530170918,
0.038321610540151596,
-0.002177897607907653,
0.025756841525435448,
-0.13128380477428436,
-0.006494250614196062,
0.05080290511250496,
-0.017841456457972527,
-0.003098310437053442,
-0.03769238293170929,
0.014923922717571259,
0.07854469120502472,
0.03448391333222389,
0.015675844624638557,
0.021372633054852486,
0.0009478454012423754,
-0.0789295881986618,
0.17076945304870605,
-0.0910600870847702,
-0.15801839530467987,
-0.14391940832138062,
0.07674430310726166,
-0.053340643644332886,
-0.0190492682158947,
0.007097931113094091,
-0.0954519733786583,
-0.05144292116165161,
-0.09337468445301056,
-0.027491247281432152,
-0.06294699758291245,
0.007315516006201506,
0.048051320016384125,
0.038413386791944504,
0.08251849561929703,
-0.12069909274578094,
0.02607887238264084,
0.023757237941026688,
-0.07746589183807373,
-0.014073225669562817,
0.03895040974020958,
0.10438478738069534,
0.11592843383550644,
-0.022754324600100517,
0.004618013743311167,
-0.028602270409464836,
0.17969520390033722,
-0.09649941325187683,
0.008789175190031528,
0.1222282350063324,
0.024326229467988014,
0.05623289942741394,
0.10480481386184692,
0.02841925248503685,
-0.05378711223602295,
0.02595187909901142,
0.05778110772371292,
-0.021053090691566467,
-0.2460048496723175,
-0.028652211651206017,
-0.029192861169576645,
-0.04697028920054436,
0.131143718957901,
0.060885194689035416,
0.046725235879421234,
0.053085021674633026,
-0.04872822389006615,
0.06059197708964348,
-0.0027499236166477203,
0.10578954219818115,
0.028384514153003693,
0.056448258459568024,
0.10126349329948425,
-0.03096957504749298,
-0.02990616485476494,
0.058269888162612915,
0.0038504349067807198,
0.19200411438941956,
-0.037682462483644485,
0.16651760041713715,
-0.004892838653177023,
0.11542801558971405,
-0.012859961949288845,
0.05505865067243576,
0.03434194251894951,
-0.010331070981919765,
0.030929140746593475,
-0.07874124497175217,
-0.03157682716846466,
0.05098094046115875,
0.027317531406879425,
0.05501207709312439,
-0.07087578624486923,
0.03368091583251953,
0.029545417055487633,
0.24650838971138,
0.05215669423341751,
-0.3054507374763489,
-0.09046922624111176,
0.009639294818043709,
-0.026427650824189186,
-0.06974675506353378,
0.006724375765770674,
0.12447631359100342,
-0.13395290076732635,
0.07049538940191269,
-0.06922948360443115,
0.08571866154670715,
-0.05213281139731407,
-0.011244005523622036,
0.07275714725255966,
0.09955441206693649,
-0.012555158697068691,
0.09194426238536835,
-0.16610364615917206,
0.1805528998374939,
0.02044111117720604,
0.10730358213186264,
-0.08054419606924057,
0.049397848546504974,
-0.005134628154337406,
0.06510842591524124,
0.10924379527568817,
-0.010170810855925083,
-0.032812297344207764,
-0.15744896233081818,
-0.11455152928829193,
0.015532298944890499,
0.12471701949834824,
-0.06942080706357956,
0.08226723223924637,
-0.05008788779377937,
-0.017325377091765404,
0.035036422312259674,
-0.05297369137406349,
-0.1587689220905304,
-0.16207802295684814,
0.030557705089449883,
-0.012396935373544693,
0.010627619922161102,
-0.09263498336076736,
-0.10454434901475906,
-0.03273625671863556,
0.23282164335250854,
-0.01421067863702774,
-0.05603373050689697,
-0.15542544424533844,
0.08529870957136154,
0.14551828801631927,
-0.057200752198696136,
0.011170507408678532,
0.00302633224055171,
0.1707766205072403,
0.019266651943325996,
-0.047709282487630844,
0.050711628049612045,
-0.05969865992665291,
-0.18657562136650085,
-0.05669480934739113,
0.1658153235912323,
0.022636068984866142,
0.05691029503941536,
-0.0017354391748085618,
0.02089334838092327,
0.01218002662062645,
-0.08004389703273773,
0.030686670914292336,
0.04168223589658737,
0.10436997562646866,
0.02316664159297943,
-0.044051919132471085,
0.04530402645468712,
-0.030910532921552658,
-0.016923999413847923,
0.11686646938323975,
0.22279492020606995,
-0.07819068431854248,
0.05361026152968407,
0.06310960650444031,
-0.04548392817378044,
-0.1440584659576416,
0.027437999844551086,
0.09940692037343979,
0.024747416377067566,
0.041440561413764954,
-0.16080275177955627,
0.0913686603307724,
0.12373780459165573,
-0.014979171566665173,
0.043035078793764114,
-0.2873922884464264,
-0.1320885419845581,
0.052425336092710495,
0.09617523849010468,
-0.06027565896511078,
-0.15946395695209503,
-0.05617522448301315,
-0.05694520100951195,
-0.12540072202682495,
0.07959236949682236,
-0.06768257915973663,
0.09812356531620026,
-0.017950670793652534,
0.026671532541513443,
0.03701666370034218,
-0.05445244908332825,
0.1516871154308319,
0.0009704928961582482,
0.03889799863100052,
-0.049724530428647995,
0.037950772792100906,
0.11639578640460968,
-0.07516571134328842,
0.04619726166129112,
-0.08682668209075928,
0.06257963180541992,
-0.13149724900722504,
-0.02333524078130722,
-0.057721201330423355,
0.05033458396792412,
-0.06500644981861115,
-0.06604509800672531,
-0.03167342767119408,
0.05475349724292755,
0.06296765059232712,
-0.041579920798540115,
0.08805011212825775,
0.02709093876183033,
0.11727679520845413,
0.12613561749458313,
0.096088707447052,
0.021138986572623253,
-0.11934786289930344,
-0.02130022831261158,
-0.022785061970353127,
0.057730790227651596,
-0.10931195318698883,
0.014987658709287643,
0.10576842725276947,
0.05666730925440788,
0.1130894124507904,
0.029199950397014618,
-0.06961853802204132,
-0.017358137294650078,
0.05494585260748863,
-0.07258552312850952,
-0.12784767150878906,
-0.03445780649781227,
0.009416786022484303,
-0.16624486446380615,
0.0022855005227029324,
0.09689126908779144,
-0.06207229942083359,
-0.017462411895394325,
-0.0007765117334201932,
0.02252311259508133,
-0.015581241808831692,
0.16406986117362976,
0.038574233651161194,
0.08116122335195541,
-0.06304819881916046,
0.10735441744327545,
0.09099661558866501,
-0.09773017466068268,
0.03851556405425072,
0.0627673864364624,
-0.07963323593139648,
-0.014252396300435066,
0.08734643459320068,
0.07325553894042969,
0.006854289676994085,
-0.04007657617330551,
-0.06746362894773483,
-0.08184322714805603,
0.0439329668879509,
0.025589143857359886,
0.024772414937615395,
-0.00211926456540823,
-0.018415383994579315,
0.02778693661093712,
-0.16722126305103302,
0.10500677675008774,
0.03324424475431442,
0.08410405367612839,
-0.1570744812488556,
0.09623777866363525,
0.002876056358218193,
0.013664278201758862,
-0.006969828624278307,
0.034085437655448914,
-0.07670138031244278,
-0.04374641552567482,
-0.11568398028612137,
-0.026124129071831703,
-0.05388278886675835,
-0.004127113614231348,
-0.021090662106871605,
-0.032305993139743805,
-0.03348436579108238,
0.034100066870450974,
-0.0478169247508049,
-0.09548242390155792,
-0.0007421363261528313,
0.05036064609885216,
-0.13518555462360382,
0.016922203823924065,
0.030726678669452667,
-0.11927008628845215,
0.07659344375133514,
0.04441018030047417,
0.04014858230948448,
0.0013839080929756165,
-0.058573413640260696,
0.02783186174929142,
0.01923646219074726,
0.009490315802395344,
0.047567933797836304,
-0.09293850511312485,
-0.0005536414682865143,
-0.03051859140396118,
-0.00228681368753314,
0.021486004814505577,
0.03943921625614166,
-0.12847602367401123,
-0.01705915853381157,
-0.04975692182779312,
-0.04652966186404228,
-0.0621158741414547,
0.05344753339886665,
0.10744686424732208,
-0.005852385889738798,
0.14708523452281952,
-0.06362979859113693,
0.05891269072890282,
-0.21504175662994385,
-0.02606288716197014,
-0.005212216638028622,
-0.01365636382251978,
-0.03409327566623688,
-0.011369665153324604,
0.07205551862716675,
-0.04208897054195404,
0.09780547767877579,
-0.027126476168632507,
0.10691271722316742,
0.05073070153594017,
-0.02576495334506035,
0.021317467093467712,
0.02116004191339016,
0.1940092146396637,
0.06615082919597626,
-0.0266726016998291,
0.0824674442410469,
-0.0337187796831131,
0.06399928033351898,
0.014376415871083736,
0.12904462218284607,
0.15272513031959534,
-0.043070148676633835,
0.06475421786308289,
0.04838152229785919,
-0.11440496146678925,
-0.1892157793045044,
0.0855511873960495,
-0.02977677993476391,
0.10270631313323975,
-0.043855756521224976,
0.15014809370040894,
0.12734100222587585,
-0.1765311360359192,
0.019682791084051132,
-0.05091290920972824,
-0.09933212399482727,
-0.07309309393167496,
-0.07370064407587051,
-0.07525869458913803,
-0.11099068820476532,
0.024306613951921463,
-0.10227477550506592,
0.015781812369823456,
0.06549782305955887,
0.015286828391253948,
0.009346193633973598,
0.16769243776798248,
-0.008880832232534885,
0.010705038905143738,
0.06385349482297897,
0.025586122646927834,
0.0009157349122688174,
-0.034178756177425385,
-0.0492953397333622,
0.03816073387861252,
0.0009432055521756411,
0.09306690096855164,
-0.04288269951939583,
-0.012248399667441845,
0.04755280166864395,
0.017583508044481277,
-0.08191798627376556,
0.03567147254943848,
0.006885915528982878,
0.024526339024305344,
0.06412141770124435,
0.03050442412495613,
0.009418614208698273,
-0.0636163055896759,
0.2597314715385437,
-0.06299465149641037,
-0.053718626499176025,
-0.13181224465370178,
0.15953752398490906,
0.03572148457169533,
-0.015104222111403942,
0.06549655646085739,
-0.1395651400089264,
-0.0053963507525622845,
0.1146751418709755,
0.13381144404411316,
-0.017269067466259003,
-0.016103573143482208,
0.014996308833360672,
-0.022536378353834152,
-0.05319844186306,
0.08785289525985718,
0.08350492268800735,
0.014065518043935299,
-0.053687989711761475,
0.02286337874829769,
0.0022005769424140453,
-0.06204831227660179,
-0.06762562692165375,
0.10241252928972244,
0.01294239517301321,
0.0049048219807446,
-0.028514781966805458,
0.08585942536592484,
0.018390897661447525,
-0.1888134628534317,
0.062206968665122986,
-0.1525067836046219,
-0.1876794397830963,
-0.004849973600357771,
0.08281920850276947,
-0.030701853334903717,
0.061478737741708755,
0.025730859488248825,
-0.014731711708009243,
0.1068737804889679,
-0.0031261893454939127,
-0.05321214348077774,
-0.07604385912418365,
0.06800913065671921,
-0.05679948627948761,
0.22964626550674438,
-0.008867841213941574,
0.046347782015800476,
0.12034085392951965,
0.021022479981184006,
-0.1542600840330124,
0.01241313386708498,
0.08320493996143341,
-0.07186707854270935,
0.05090373381972313,
0.184761643409729,
-0.040344733744859695,
0.07792957127094269,
0.042797207832336426,
-0.12084364145994186,
-0.026484450325369835,
-0.07070191949605942,
0.006931263953447342,
-0.09296785295009613,
0.017404984682798386,
-0.04036698117852211,
0.17144206166267395,
0.2056282013654709,
-0.05294200778007507,
-0.010758241638541222,
-0.06033679097890854,
0.028235342353582382,
0.054588742554187775,
0.1035965085029602,
0.0012649006675928831,
-0.21200834214687347,
0.012996422126889229,
0.01900651305913925,
0.05271465703845024,
-0.24885772168636322,
-0.08648484200239182,
0.06231068819761276,
-0.06409452110528946,
-0.07374738156795502,
0.1024787425994873,
0.04061764478683472,
0.02038969099521637,
-0.03562695160508156,
-0.11453217267990112,
-0.042103320360183716,
0.13460420072078705,
-0.1746665984392166,
-0.04097144305706024
] |
null | null | null |
# **Reinforce** Agent playing **Pixelcopter-PLE-v0**
This is a trained model of a **Reinforce** agent playing **Pixelcopter-PLE-v0** .
To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: https://huggingface.co/deep-rl-course/unit4/introduction
| {"tags": ["Pixelcopter-PLE-v0", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class"], "model-index": [{"name": "Reinforce-Pixelcopter-PLE-v0", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "Pixelcopter-PLE-v0", "type": "Pixelcopter-PLE-v0"}, "metrics": [{"type": "mean_reward", "value": "18.40 +/- 15.38", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | tinywell/Reinforce-Pixelcopter-PLE-v0 | [
"Pixelcopter-PLE-v0",
"reinforce",
"reinforcement-learning",
"custom-implementation",
"deep-rl-class",
"model-index",
"region:us"
] | 2024-02-07T07:42:06+00:00 | [] | [] | TAGS
#Pixelcopter-PLE-v0 #reinforce #reinforcement-learning #custom-implementation #deep-rl-class #model-index #region-us
|
# Reinforce Agent playing Pixelcopter-PLE-v0
This is a trained model of a Reinforce agent playing Pixelcopter-PLE-v0 .
To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL
| [
"# Reinforce Agent playing Pixelcopter-PLE-v0\n This is a trained model of a Reinforce agent playing Pixelcopter-PLE-v0 .\n To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL"
] | [
"TAGS\n#Pixelcopter-PLE-v0 #reinforce #reinforcement-learning #custom-implementation #deep-rl-class #model-index #region-us \n",
"# Reinforce Agent playing Pixelcopter-PLE-v0\n This is a trained model of a Reinforce agent playing Pixelcopter-PLE-v0 .\n To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL"
] | [
41,
58
] | [
"passage: TAGS\n#Pixelcopter-PLE-v0 #reinforce #reinforcement-learning #custom-implementation #deep-rl-class #model-index #region-us \n# Reinforce Agent playing Pixelcopter-PLE-v0\n This is a trained model of a Reinforce agent playing Pixelcopter-PLE-v0 .\n To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL"
] | [
0.0073175891302526,
-0.2259262204170227,
-0.0017347558168694377,
0.05054566636681557,
0.0658537745475769,
-0.055378563702106476,
0.1412602812051773,
0.05916554853320122,
-0.04990595206618309,
0.059261854737997055,
0.14166708290576935,
0.03996060788631439,
0.022112762555480003,
0.1513713151216507,
0.09764605015516281,
-0.2469022423028946,
0.07438477873802185,
0.01641594059765339,
0.008152224123477936,
0.09583204984664917,
0.060265738517045975,
-0.1405058205127716,
0.037032704800367355,
-0.01332044042646885,
-0.13650871813297272,
0.0010478810872882605,
-0.021802188828587532,
-0.03625129908323288,
0.15681709349155426,
0.006844013463705778,
0.09602472931146622,
-0.001560068572871387,
0.06475798785686493,
-0.12438877671957016,
0.05466329678893089,
0.06455880403518677,
-0.06293967366218567,
0.058029334992170334,
-0.057374246418476105,
0.11959903687238693,
0.04641333222389221,
-0.01578129455447197,
0.054811324924230576,
0.010941818356513977,
-0.14131468534469604,
-0.006710252724587917,
0.007013716734945774,
0.15098218619823456,
0.1339312642812729,
0.01409265398979187,
-0.0014771400019526482,
0.1363491266965866,
-0.16774429380893707,
0.045684073120355606,
0.061802688986063004,
-0.2633039951324463,
-0.04168876260519028,
0.12259352207183838,
0.08951573073863983,
0.06848238408565521,
-0.060910262167453766,
0.07636868953704834,
0.049813780933618546,
0.013985024765133858,
0.023094501346349716,
-0.042509064078330994,
-0.040479615330696106,
0.02289252169430256,
-0.0921095609664917,
-0.05999262258410454,
0.11517233401536942,
-0.006806366611272097,
0.03735918551683426,
-0.12476086616516113,
-0.015330453403294086,
-0.07314357161521912,
-0.05917041376233101,
-0.082573801279068,
0.07563583552837372,
0.030191516503691673,
-0.048283837735652924,
-0.08895846456289291,
-0.056533291935920715,
-0.11489585787057877,
-0.023082571104168892,
-0.07226225733757019,
0.005096882116049528,
-0.03157244250178337,
-0.035645097494125366,
0.09446526318788528,
-0.0021088174544274807,
-0.015028090216219425,
-0.03452150896191597,
-0.05930153280496597,
-0.04213470220565796,
-0.02359505370259285,
-0.03510070592164993,
-0.059062156826257706,
0.054655663669109344,
0.0680202916264534,
0.04938843473792076,
0.09133565425872803,
-0.0467856265604496,
0.1667373925447464,
-0.03256719931960106,
0.08078566938638687,
-0.011897698976099491,
0.2012830525636673,
0.11370102316141129,
0.12129533290863037,
0.06716908514499664,
-0.05294690653681755,
-0.16726544499397278,
0.039163749665021896,
0.12641896307468414,
0.07664673775434494,
-0.032492902129888535,
0.018162984400987625,
-0.12440363317728043,
0.05439428985118866,
-0.14826108515262604,
-0.06745084375143051,
0.024251462891697884,
0.01822635903954506,
-0.060682263225317,
0.03656952083110809,
-0.0028792342636734247,
0.003339326474815607,
0.004654870834201574,
-0.16432709991931915,
-0.05568019300699234,
0.028964387252926826,
-0.15712425112724304,
-0.06656725704669952,
0.06277995556592941,
-0.10113482922315598,
-0.012132617644965649,
-0.16982388496398926,
-0.16305199265480042,
-0.03628521412611008,
0.017857929691672325,
-0.040613796561956406,
-0.056917786598205566,
-0.14010562002658844,
-0.019415250048041344,
-0.045320261269807816,
-0.004312154371291399,
0.044072363525629044,
0.0020940210670232773,
0.04635847359895706,
0.0066573889926075935,
0.09289347380399704,
0.010714372619986534,
-0.0014722738415002823,
-0.04595406726002693,
0.0909833237528801,
-0.30731555819511414,
0.07525643706321716,
-0.08645553886890411,
0.05539081245660782,
-0.057316381484270096,
-0.0926317572593689,
-0.007509906310588121,
0.06277763843536377,
0.060464419424533844,
0.20788121223449707,
-0.2800109386444092,
-0.07025618106126785,
0.13655538856983185,
-0.09533236175775528,
-0.13146020472049713,
0.0513952374458313,
-0.050213608890771866,
0.07593657076358795,
0.027370907366275787,
0.140700101852417,
-0.028026295825839043,
-0.15554022789001465,
0.06281048059463501,
0.04586128890514374,
-0.11356306821107864,
0.019295670092105865,
0.03597676753997803,
0.06723599135875702,
0.05744141340255737,
-0.036986757069826126,
-0.04105675220489502,
0.08096802979707718,
-0.07076814025640488,
-0.037564266473054886,
0.04588831216096878,
-0.0579565204679966,
0.1630958467721939,
0.033971156924963,
0.09856503456830978,
-0.04149768501520157,
-0.07435470074415207,
-0.005698562134057283,
0.038746561855077744,
-0.08962973952293396,
0.025353478267788887,
-0.18320298194885254,
0.2423991560935974,
-0.02621818706393242,
0.027546977624297142,
-0.16845986247062683,
-0.0588528998196125,
0.011087946593761444,
0.21568740904331207,
0.030399197712540627,
0.12989304959774017,
0.07485637813806534,
-0.01250512059777975,
0.014156299643218517,
-0.06183977797627449,
-0.1972363442182541,
-0.03247830644249916,
0.008314179256558418,
-0.058311350643634796,
-0.04934588819742203,
-0.0900716632604599,
0.10427892208099365,
-0.19334633648395538,
-0.005319371819496155,
0.08282599598169327,
0.023504555225372314,
0.03946567326784134,
0.0035407328978180885,
-0.03634254261851311,
0.055148303508758545,
0.02030518464744091,
-0.08980578929185867,
0.14668866991996765,
0.0035520538222044706,
-0.03514726087450981,
-0.03927676007151604,
-0.03267495706677437,
0.05703731253743172,
0.08045367896556854,
-0.18214593827724457,
-0.0733821839094162,
-0.0838410034775734,
-0.02458474040031433,
0.050523869693279266,
0.036679428070783615,
0.02738112211227417,
0.44813573360443115,
0.057562243193387985,
0.09003535658121109,
-0.08811535686254501,
0.039806611835956573,
0.012785476632416248,
-0.031281858682632446,
0.013625281862914562,
0.04725322127342224,
0.11279468983411789,
0.028284218162298203,
0.01669839769601822,
0.03680038824677467,
0.01938779093325138,
0.08824212104082108,
-0.10939645022153854,
-0.003965397831052542,
0.002614045049995184,
0.038018375635147095,
0.03672022372484207,
0.07190682739019394,
0.015936892479658127,
-0.09583546966314316,
-0.030848123133182526,
-0.11166880279779434,
0.015594755299389362,
-0.20979784429073334,
-0.025905707851052284,
-0.029619399458169937,
0.0003502996696624905,
0.09109684824943542,
0.04222718998789787,
-0.04444896802306175,
0.035467714071273804,
0.03947039321064949,
-0.0861397460103035,
0.0594942644238472,
-0.014317752793431282,
-0.07008631527423859,
0.13023322820663452,
-0.1002996563911438,
-0.3153233230113983,
-0.08797995746135712,
0.05698639526963234,
0.05295826122164726,
0.06816939264535904,
-0.05876303091645241,
-0.09240786731243134,
0.03294730558991432,
-0.06836386770009995,
-0.0017794050509110093,
0.0037346978206187487,
-0.051060982048511505,
0.07253886014223099,
0.08541567623615265,
-0.014505518600344658,
-0.08911184966564178,
-0.006620637606829405,
-0.041561197489500046,
-0.124965138733387,
0.044060997664928436,
-0.03760828450322151,
0.00007921225915197283,
0.18620672821998596,
0.03724536672234535,
0.06256633251905441,
-0.06291008740663528,
0.07596296072006226,
-0.09150096774101257,
0.0004740063741337508,
0.18428465723991394,
-0.015377625823020935,
-0.004100616089999676,
-0.03996327146887779,
-0.0259257685393095,
-0.10829219967126846,
0.053985193371772766,
-0.07330703735351562,
-0.07349077612161636,
-0.0023273853585124016,
-0.07770214974880219,
-0.0351552739739418,
0.0012160884216427803,
0.07817990332841873,
0.029699061065912247,
-0.09635239094495773,
0.04920589178800583,
0.1298678070306778,
0.0931883230805397,
0.03626195341348648,
0.023981640115380287,
0.13739009201526642,
-0.11230582743883133,
0.019063033163547516,
-0.05148853361606598,
-0.1041760966181755,
-0.042787205427885056,
-0.0714287981390953,
0.07368279993534088,
0.06034531816840172,
-0.09970010071992874,
0.05144011229276657,
0.041872985661029816,
0.0883496031165123,
0.1373600959777832,
-0.04213863983750343,
-0.11244629323482513,
-0.041393622756004333,
-0.022004956379532814,
-0.1777329444885254,
0.0341336652636528,
0.22155584394931793,
0.0073304991237819195,
-0.10497386753559113,
0.07876885682344437,
-0.005956185050308704,
0.11527370661497116,
0.031222699210047722,
-0.278682678937912,
0.016931315883994102,
0.00203216471709311,
0.042359162122011185,
-0.047676295042037964,
0.10937416553497314,
0.11747439950704575,
-0.14421136677265167,
-0.06650938838720322,
-0.03273930773139,
0.044137366116046906,
-0.15618287026882172,
0.036923591047525406,
-0.12602220475673676,
0.06240779533982277,
0.050940994173288345,
0.05090156942605972,
-0.2197665423154831,
0.06881614029407501,
-0.0274215005338192,
0.06763827055692673,
-0.062248338013887405,
-0.01823522336781025,
0.04473711550235748,
0.025079863145947456,
0.14955177903175354,
-0.014347962103784084,
0.14454017579555511,
-0.09031219780445099,
-0.11753576993942261,
0.0027052261866629124,
0.08532248437404633,
0.013173088431358337,
0.013580933213233948,
0.0026939227245748043,
0.041669201105833054,
-0.02811569906771183,
0.17063532769680023,
-0.08147624880075455,
-0.022407781332731247,
-0.06592555344104767,
-0.018158966675400734,
0.2039334923028946,
-0.12064731866121292,
-0.10121093690395355,
-0.11619500070810318,
0.08663272857666016,
-0.04296411573886871,
0.08175522089004517,
-0.020344657823443413,
0.049704354256391525,
-0.02509051002562046,
0.007178863976150751,
0.09594997018575668,
0.01950966566801071,
0.08983828872442245,
-0.09791163355112076,
-0.019585272297263145,
0.13838915526866913,
-0.037155888974666595,
-0.036971647292375565,
-0.019425252452492714,
0.11054370552301407,
-0.0358734093606472,
0.08033111691474915,
0.03929615020751953,
0.03664831817150116,
0.03428546339273453,
-0.039165496826171875,
0.10309428721666336,
0.10041618347167969,
-0.06291446089744568,
0.03864621743559837,
-0.07954532653093338,
0.26597461104393005,
0.040773067623376846,
0.07301845401525497,
0.28390514850616455,
0.19391325116157532,
-0.03036464750766754,
0.10683353990316391,
-0.017607249319553375,
-0.024403288960456848,
-0.2950931787490845,
0.0006976581644266844,
0.027765681967139244,
0.11812873929738998,
0.01744898222386837,
-0.20587195456027985,
-0.1211688369512558,
-0.03560304269194603,
-0.007791717536747456,
0.0310499370098114,
-0.2441052496433258,
-0.06442268192768097,
0.06107868626713753,
0.13779635727405548,
0.15878525376319885,
-0.05917542055249214,
-0.007856467738747597,
0.029358724132180214,
0.07593556493520737,
0.017292039468884468,
-0.11598441749811172,
0.11550791561603546,
0.025637371465563774,
-0.05708931386470795,
0.0267958827316761,
-0.044003549963235855,
0.04214555397629738,
-0.17736166715621948,
0.10933554917573929,
-0.05924695357680321,
-0.08421005308628082,
0.07140472531318665,
-0.02217724733054638,
-0.048552993685007095,
0.0789642184972763,
0.020652711391448975,
-0.13173207640647888,
0.038154006004333496,
0.005618774797767401,
0.04346654564142227,
-0.004941361024975777,
-0.019811764359474182,
-0.029163256287574768,
0.07706235349178314,
-0.03806605935096741,
0.09605937451124191,
0.19590972363948822,
-0.0573095865547657,
0.03974950686097145,
0.085201695561409,
0.09593135863542557,
-0.05523005872964859,
-0.0809539332985878,
-0.03812742978334427,
-0.005277194548398256,
0.0674438327550888,
-0.08598461747169495,
-0.019085103645920753,
0.07938229292631149,
0.015313901007175446,
0.14910826086997986,
0.14389736950397491,
-0.08835655450820923,
0.11321785300970078,
0.10694554448127747,
-0.11366690695285797,
-0.08583837002515793,
-0.02963297814130783,
0.0009990704711526632,
0.04910186678171158,
-0.048617590218782425,
0.05932905897498131,
-0.1035301461815834,
0.012819357216358185,
0.03532040864229202,
0.0038119733799248934,
-0.09975302964448929,
0.009764863178133965,
0.08645275235176086,
0.06119582802057266,
-0.0567571222782135,
0.09250631928443909,
-0.0019178141374140978,
-0.10868195444345474,
0.07241881638765335,
0.009918469935655594,
-0.021528873592615128,
-0.06352251768112183,
0.03211374953389168,
0.2370220273733139,
0.13945111632347107,
-0.04336636886000633,
-0.12396618723869324,
-0.15508891642093658,
0.037849195301532745,
0.024356422945857048,
0.051251959055662155,
0.0062240250408649445,
-0.06906022876501083,
0.01234503649175167,
-0.04392383247613907,
0.005266309250146151,
-0.05930564925074577,
-0.047703344374895096,
-0.12081446498632431,
0.1154373437166214,
0.053290288895368576,
0.11705748736858368,
-0.0842847004532814,
-0.07057584822177887,
-0.1921386867761612,
0.09190598875284195,
0.041707299649715424,
-0.05532265454530716,
0.06002674251794815,
-0.030134430155158043,
0.017344338819384575,
0.11256659775972366,
-0.051967836916446686,
0.008543911390006542,
-0.09269233793020248,
0.03236149623990059,
0.03133073076605797,
0.04903566092252731,
-0.004612727556377649,
-0.017903391271829605,
0.04399999976158142,
-0.05730267986655235,
0.07619527727365494,
-0.07757602632045746,
-0.033709146082401276,
0.0645759105682373,
-0.16051416099071503,
-0.054324716329574585,
0.08708633482456207,
0.013749903067946434,
0.02590017393231392,
-0.05825240537524223,
0.019142305478453636,
-0.05566488951444626,
-0.04483235627412796,
0.01169554702937603,
-0.05552767962217331,
-0.011517677456140518,
0.05293213203549385,
-0.05287189036607742,
-0.040493328124284744,
-0.06794002652168274,
0.061874233186244965,
-0.07247710227966309,
0.09816460311412811,
0.031187955290079117,
-0.10892423242330551,
0.07648903876543045,
-0.037552736699581146,
-0.0049397205002605915,
-0.009439278393983841,
0.039307788014411926,
0.15598824620246887,
-0.1606634259223938,
0.05345672369003296,
-0.0484454482793808,
0.13272921741008759,
0.046888746321201324,
-0.04458791762590408,
-0.020207170397043228,
0.02469455823302269,
-0.05549024045467377,
0.06932897865772247,
0.15877580642700195,
0.09880131483078003,
0.02571805939078331,
0.008134597912430763,
0.10187267512083054,
0.1060529574751854,
0.08136752992868423,
0.08394161611795425,
-0.03428563475608826,
-0.11287897825241089,
0.14338994026184082,
0.09748584777116776,
0.024613093584775925,
0.21077860891819,
0.17944025993347168,
0.03125298395752907,
0.03018142655491829,
-0.06512103229761124,
0.17325744032859802,
0.061261482536792755,
-0.08229418843984604,
0.014424329623579979,
0.03221147879958153,
-0.049809664487838745,
-0.047004032880067825,
-0.09757380187511444,
-0.029556652531027794,
-0.24085633456707,
0.10851483792066574,
-0.057250600308179855,
-0.09750643372535706,
0.022772664204239845,
0.02990041859447956,
-0.018839845433831215,
0.11280566453933716,
-0.07735858112573624,
0.012980576604604721,
0.18577688932418823,
-0.03825045004487038,
-0.022322099655866623,
-0.1633504331111908,
-0.11154003441333771,
-0.014046176336705685,
-0.11750495433807373,
0.025494296103715897,
0.06305963546037674,
0.01117965579032898,
0.04399528726935387,
0.028923438861966133,
-0.020834028720855713,
0.019218796864151955,
-0.05903913825750351,
-0.042673509567976,
-0.01891910657286644,
0.02202831581234932,
-0.09593231230974197,
-0.03627033904194832,
0.12151803076267242,
-0.03246605768799782,
-0.08207374066114426,
-0.006544890813529491,
0.07848484069108963,
-0.042620159685611725,
0.09450104832649231,
-0.07687012106180191,
-0.03479038178920746,
-0.06794454902410507,
0.268902063369751,
0.09388194978237152,
-0.20183001458644867,
0.03341769427061081,
-0.030470456928014755,
0.026735708117485046,
-0.09215684235095978,
0.16250114142894745,
0.0899243950843811,
0.049168527126312256,
-0.12686687707901,
-0.003401300171390176,
-0.09992645680904388,
-0.0028723697178065777,
-0.12552696466445923,
-0.14725084602832794,
0.12093491852283478,
-0.003848524997010827,
-0.06547791510820389,
0.02844911813735962,
-0.15909899771213531,
0.06585367769002914,
0.0978507474064827,
-0.1514272391796112,
-0.038227714598178864,
-0.06086801365017891,
0.06072385236620903,
0.026465637609362602,
0.13005392253398895,
-0.05080926790833473,
0.012067130766808987,
-0.0656723901629448,
-0.011309894733130932,
-0.0000654291216051206,
-0.017478201538324356,
0.001532604917883873,
-0.09828947484493256,
0.05038110539317131,
-0.0835796371102333,
0.12184429168701172,
0.05709611251950264,
0.005326167680323124,
0.008464806713163853,
0.0648408755660057,
-0.02414623089134693,
-0.10202058404684067,
-0.01877439208328724,
0.033475372940301895,
0.03998998552560806,
0.010373802855610847,
0.034506846219301224,
0.0006507808575406671,
0.07714920490980148,
-0.011413984932005405,
-0.027285432443022728,
-0.058209117501974106,
0.03936338797211647,
-0.10441672056913376,
0.10461361706256866,
0.0013552121818065643,
-0.02240127883851528,
-0.010913821868598461,
-0.05532446503639221,
0.045815300196409225,
0.04572062939405441,
0.029743505641818047,
-0.05261747166514397,
-0.09262793511152267,
-0.021781492978334427,
0.023900283500552177,
-0.11539579927921295,
-0.18497975170612335,
-0.0664035826921463,
-0.15038692951202393,
-0.01633414439857006,
-0.0620744526386261,
0.08902198076248169,
0.13558129966259003,
0.030392181128263474,
-0.04822919890284538,
-0.12171997129917145,
0.025026977062225342,
0.13544774055480957,
-0.03851630911231041,
-0.07532322406768799
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# wav2vec2-base-finetuned-ks-ob
This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the audiofolder dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0015
- Accuracy: 1.0000
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 3e-05
- train_batch_size: 32
- eval_batch_size: 32
- seed: 42
- distributed_type: tpu
- gradient_accumulation_steps: 4
- total_train_batch_size: 128
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 5
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.0528 | 1.0 | 256 | 0.0275 | 0.9994 |
| 0.0122 | 2.0 | 512 | 0.0054 | 0.9998 |
| 0.0048 | 3.0 | 768 | 0.0041 | 0.9995 |
| 0.0029 | 4.0 | 1024 | 0.0020 | 0.9999 |
| 0.0019 | 5.0 | 1280 | 0.0015 | 1.0000 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.0.0+cu118
- Datasets 2.17.0
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["audiofolder"], "metrics": ["accuracy"], "base_model": "facebook/wav2vec2-base", "model-index": [{"name": "wav2vec2-base-finetuned-ks-ob", "results": [{"task": {"type": "audio-classification", "name": "Audio Classification"}, "dataset": {"name": "audiofolder", "type": "audiofolder", "config": "default", "split": "train", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.9999694563225412, "name": "Accuracy"}]}]}]} | audio-classification | sravan-gorugantu/wav2vec2-base-finetuned-ks-ob | [
"transformers",
"tensorboard",
"safetensors",
"wav2vec2",
"audio-classification",
"generated_from_trainer",
"dataset:audiofolder",
"base_model:facebook/wav2vec2-base",
"license:apache-2.0",
"model-index",
"endpoints_compatible",
"region:us"
] | 2024-02-07T07:45:13+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #wav2vec2 #audio-classification #generated_from_trainer #dataset-audiofolder #base_model-facebook/wav2vec2-base #license-apache-2.0 #model-index #endpoints_compatible #region-us
| wav2vec2-base-finetuned-ks-ob
=============================
This model is a fine-tuned version of facebook/wav2vec2-base on the audiofolder dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0015
* Accuracy: 1.0000
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 3e-05
* train\_batch\_size: 32
* eval\_batch\_size: 32
* seed: 42
* distributed\_type: tpu
* gradient\_accumulation\_steps: 4
* total\_train\_batch\_size: 128
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* num\_epochs: 5
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.0.0+cu118
* Datasets 2.17.0
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* distributed\\_type: tpu\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.0.0+cu118\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #wav2vec2 #audio-classification #generated_from_trainer #dataset-audiofolder #base_model-facebook/wav2vec2-base #license-apache-2.0 #model-index #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* distributed\\_type: tpu\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.0.0+cu118\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
78,
153,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #wav2vec2 #audio-classification #generated_from_trainer #dataset-audiofolder #base_model-facebook/wav2vec2-base #license-apache-2.0 #model-index #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* distributed\\_type: tpu\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 5### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.0.0+cu118\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
-0.12121367454528809,
0.13954074680805206,
-0.002346578985452652,
0.07390250265598297,
0.1105363741517067,
0.028198910877108574,
0.09423739463090897,
0.1106981709599495,
-0.08634363114833832,
0.11064612865447998,
0.08527939766645432,
0.09703831374645233,
0.06798873096704483,
0.13620199263095856,
-0.016269495710730553,
-0.2697935402393341,
-0.0023534323554486036,
-0.016367793083190918,
-0.1359536051750183,
0.11756301671266556,
0.06442328542470932,
-0.09256460517644882,
0.06635752320289612,
-0.012698670849204063,
-0.11997544765472412,
-0.013258557766675949,
-0.02223590947687626,
-0.06177615374326706,
0.08487173169851303,
0.03244958817958832,
0.06633693724870682,
0.04673555865883827,
0.10648775100708008,
-0.2203703224658966,
0.01837095245718956,
0.09611090272665024,
-0.004760048817843199,
0.08663194626569748,
0.14707064628601074,
-0.005324545316398144,
0.1400076001882553,
-0.09131285548210144,
0.049685943871736526,
0.051168378442525864,
-0.09003095328807831,
-0.24005268514156342,
-0.08156772702932358,
0.09121830016374588,
0.10007857531309128,
0.07018518447875977,
-0.0255054272711277,
0.07476296275854111,
-0.07026797533035278,
0.07727861404418945,
0.23483151197433472,
-0.26527169346809387,
-0.06345000118017197,
0.03504631668329239,
0.04626055061817169,
0.07664090394973755,
-0.11094948649406433,
-0.007239424157887697,
0.049065787345170975,
0.024556774646043777,
0.12425493448972702,
0.009824540466070175,
0.033012859523296356,
0.004034258890897036,
-0.16263996064662933,
-0.03141947090625763,
0.12197320908308029,
0.09819214791059494,
-0.026472683995962143,
-0.06491270661354065,
-0.03806861490011215,
-0.21608339250087738,
-0.024184126406908035,
0.0015347428852692246,
0.024870917201042175,
-0.04577261209487915,
-0.08785060793161392,
0.04537094011902809,
-0.04441770538687706,
-0.09452969580888748,
0.027631226927042007,
0.11832787841558456,
0.0534062422811985,
-0.020143676549196243,
0.031177043914794922,
0.11452510952949524,
0.06332036107778549,
-0.15520241856575012,
0.030170349404215813,
0.020967671647667885,
-0.10876688361167908,
-0.021855004131793976,
0.002127655316144228,
-0.00805111974477768,
0.028255322948098183,
0.11560773104429245,
-0.04668932780623436,
0.05426805466413498,
0.05005623400211334,
0.03516961634159088,
-0.05194941908121109,
0.0898011177778244,
-0.10858356952667236,
-0.12575663626194,
-0.029725851491093636,
0.10435725003480911,
0.01921483688056469,
-0.014264202676713467,
-0.07806724309921265,
0.028799599036574364,
0.11248037219047546,
0.02688608132302761,
-0.0034528065007179976,
0.021206244826316833,
-0.08207416534423828,
-0.05485305190086365,
0.04834919422864914,
-0.0828728899359703,
0.02532058209180832,
0.032619375735521317,
-0.03988788276910782,
0.002762723946943879,
0.0050588431768119335,
0.01027635857462883,
-0.010587132535874844,
0.1407177895307541,
-0.08526541292667389,
-0.049647167325019836,
-0.05484652519226074,
-0.07335805892944336,
0.04932520166039467,
-0.09713998436927795,
0.008773342706263065,
-0.07868245244026184,
-0.08456751704216003,
-0.035891108214855194,
0.07702561467885971,
-0.04993680492043495,
-0.1035836935043335,
-0.03882885351777077,
-0.09233946353197098,
0.047265972942113876,
-0.024561114609241486,
0.14906330406665802,
-0.0765518844127655,
0.1069522574543953,
-0.03633888065814972,
0.06717856973409653,
0.06293144077062607,
0.07097035646438599,
-0.029806310310959816,
0.05172118917107582,
-0.19131653010845184,
0.03878859430551529,
-0.09408364444971085,
0.028224855661392212,
-0.13401490449905396,
-0.11050962656736374,
-0.029731931164860725,
-0.005517411511391401,
0.0703585296869278,
0.09051943570375443,
-0.18474261462688446,
-0.10919356346130371,
0.12684449553489685,
-0.08087488263845444,
-0.10890568792819977,
0.1467493623495102,
0.008536793291568756,
-0.0698988288640976,
0.01681050844490528,
0.17508868873119354,
0.11871787160634995,
-0.12295526266098022,
-0.021690040826797485,
-0.01665714755654335,
0.08211103826761246,
0.014496592804789543,
0.11001741141080856,
-0.024789921939373016,
-0.018555887043476105,
-0.022979330271482468,
-0.012502851895987988,
0.06341294944286346,
-0.09467706829309464,
-0.07407538592815399,
-0.021432790905237198,
-0.10289392620325089,
0.03241295367479324,
0.05139234662055969,
-0.005458418745547533,
-0.06088634207844734,
-0.1333906054496765,
0.0152250025421381,
0.1118135154247284,
-0.06780519336462021,
0.004722948651760817,
-0.05173267796635628,
0.1048559695482254,
-0.057184599339962006,
-0.027337273582816124,
-0.1490320861339569,
-0.014718669466674328,
0.024342911317944527,
-0.06633847951889038,
0.023551534861326218,
0.014892995357513428,
0.05358656495809555,
0.0786970779299736,
-0.07245249301195145,
-0.09056545794010162,
-0.06026914343237877,
0.015336294658482075,
-0.04267236590385437,
-0.2816104292869568,
-0.056324176490306854,
-0.017973875626921654,
0.11126957833766937,
-0.20569515228271484,
-0.01985229179263115,
-0.00278339977376163,
0.12940563261508942,
0.05951378494501114,
-0.061226945370435715,
0.017267970368266106,
0.06569438427686691,
-0.021852582693099976,
-0.07712628692388535,
0.025601355358958244,
0.0077431476674973965,
-0.09706196933984756,
-0.018554501235485077,
-0.11823209375143051,
0.14608165621757507,
0.10188823193311691,
0.0096651716157794,
-0.10428057610988617,
-0.005654633976519108,
-0.07865120470523834,
-0.06339685618877411,
-0.02835864946246147,
0.018923314288258553,
0.13890773057937622,
0.0055801281705498695,
0.09187968820333481,
-0.09859602898359299,
-0.0484485886991024,
0.05792982876300812,
0.012377051636576653,
-0.01065216213464737,
0.13920779526233673,
0.10479237139225006,
-0.08176983147859573,
0.12723621726036072,
0.1444501280784607,
-0.04994174838066101,
0.17900198698043823,
-0.06800520420074463,
-0.12706580758094788,
-0.027128247544169426,
0.01578834466636181,
0.022019285708665848,
0.14332444965839386,
-0.11267882585525513,
0.035940490663051605,
0.024289894849061966,
0.033908016979694366,
0.005938546732068062,
-0.16426779329776764,
-0.019522476941347122,
0.027496660128235817,
-0.06748761236667633,
-0.04404594376683235,
-0.005030326545238495,
-0.005528813693672419,
0.1038932204246521,
0.007410679943859577,
-0.06635569781064987,
-0.0049843983724713326,
-0.00040004830225370824,
-0.08223158121109009,
0.16353759169578552,
-0.09964673221111298,
-0.12418040633201599,
-0.12177231162786484,
0.0214916430413723,
-0.031652119010686874,
-0.012672960758209229,
0.03915994614362717,
-0.09204179793596268,
-0.04288594052195549,
-0.058348048478364944,
0.056917570531368256,
-0.024223027750849724,
0.051565807312726974,
0.01663917489349842,
0.032888274639844894,
0.08353133499622345,
-0.07862371951341629,
0.03601325675845146,
-0.008172906935214996,
-0.0221402645111084,
0.006998395547270775,
0.05538040027022362,
0.09481926262378693,
0.17877034842967987,
0.04689979553222656,
0.011716759763658047,
-0.017049571499228477,
0.1797884851694107,
-0.11604288965463638,
0.039552316069602966,
0.09691887348890305,
-0.02336057648062706,
0.03640199452638626,
0.15794208645820618,
0.05269179120659828,
-0.062423959374427795,
0.020344316959381104,
0.06539913266897202,
-0.02484159730374813,
-0.25004565715789795,
-0.04681703448295593,
-0.0551186203956604,
0.011488604359328747,
0.09314555674791336,
0.03456621617078781,
-0.006421095225960016,
0.045375533401966095,
-0.01113506406545639,
0.012349828146398067,
-0.01467044185847044,
0.03909265622496605,
-0.019596727564930916,
0.0455285869538784,
0.09536396712064743,
-0.033046044409275055,
-0.010031529702246189,
0.06774041801691055,
0.007276370655745268,
0.21905095875263214,
-0.01437083538621664,
0.11991776525974274,
0.0690789744257927,
0.1576615273952484,
0.014025384560227394,
0.05854879692196846,
0.0074646687135100365,
-0.021835535764694214,
0.007441957946866751,
-0.06400461494922638,
-0.0008241864270530641,
0.02992860972881317,
0.05749846249818802,
0.04496387019753456,
-0.11439584940671921,
0.027984607964754105,
0.01732482574880123,
0.28648948669433594,
0.0899350643157959,
-0.2892492413520813,
-0.08342436701059341,
0.002716852817684412,
-0.05013114959001541,
-0.03995963931083679,
0.051838669925928116,
0.15696069598197937,
-0.05496051535010338,
0.09030845761299133,
-0.05196259915828705,
0.07836002111434937,
-0.05540105700492859,
-0.014462033286690712,
0.1225983202457428,
0.13085618615150452,
-0.01677149534225464,
0.05173775181174278,
-0.18963903188705444,
0.2833308279514313,
0.0076772794127464294,
0.05864177271723747,
-0.035638708621263504,
0.01895320788025856,
0.022724471986293793,
0.0158123467117548,
0.11852946877479553,
0.010984831489622593,
-0.08612557500600815,
-0.15092502534389496,
-0.11322261393070221,
0.0016615134663879871,
0.12303581088781357,
-0.05921902135014534,
0.0800854042172432,
-0.032989926636219025,
-0.04081930220127106,
0.04857223108410835,
-0.07382187247276306,
-0.12221711128950119,
-0.08697371929883957,
0.0134552251547575,
-0.004055345430970192,
0.02581709250807762,
-0.08106806874275208,
-0.11240740865468979,
-0.13256819546222687,
0.14351414144039154,
-0.09764011204242706,
-0.012411082163453102,
-0.12919339537620544,
0.04242018982768059,
0.15716250240802765,
-0.05869268625974655,
0.06569699943065643,
0.017032893374562263,
0.098333440721035,
0.04020261764526367,
-0.05647997558116913,
0.11689448356628418,
-0.09946943819522858,
-0.21946774423122406,
-0.06712093204259872,
0.16692841053009033,
0.05439196527004242,
0.06973937898874283,
-0.04447191208600998,
0.017645293846726418,
0.014028249308466911,
-0.0835028886795044,
0.06904754787683487,
0.007936839945614338,
0.05172768980264664,
0.039177607744932175,
-0.0361582450568676,
-0.03080427646636963,
-0.020427262410521507,
-0.06017026677727699,
0.09416422247886658,
0.27391475439071655,
-0.08970589935779572,
0.06817471235990524,
0.06915108859539032,
-0.039442092180252075,
-0.1773531585931778,
0.015808414667844772,
0.12707510590553284,
0.014963395893573761,
0.04579070210456848,
-0.19561444222927094,
0.09315267205238342,
0.06935172528028488,
-0.012651639059185982,
0.10050003975629807,
-0.2978886067867279,
-0.11811315268278122,
0.09299692511558533,
0.10373369604349136,
-0.0358240082859993,
-0.15375876426696777,
-0.051354896277189255,
0.0027353717014193535,
-0.09863964468240738,
0.09729333221912384,
-0.07528505474328995,
0.11624763160943985,
-0.002655802993103862,
0.00780410785228014,
0.02131069451570511,
-0.0503203347325325,
0.11300043016672134,
0.015828223899006844,
0.052883245050907135,
-0.02939656563103199,
0.055449921637773514,
-0.0003983541682828218,
-0.05706977844238281,
-0.008627446368336678,
-0.09952058643102646,
0.04021960496902466,
-0.08464650809764862,
-0.023098615929484367,
-0.0791211873292923,
0.017645373940467834,
-0.04582108184695244,
-0.05481254681944847,
-0.025908907875418663,
0.06587234139442444,
0.07619462162256241,
-0.03667021542787552,
0.11766789853572845,
-0.012011934071779251,
0.1390957087278366,
0.09777748584747314,
0.09054652601480484,
-0.01092552114278078,
-0.1140974909067154,
-0.017757704481482506,
-0.03605413809418678,
0.058370690792798996,
-0.1188245415687561,
0.029056141152977943,
0.12112800031900406,
0.049582310020923615,
0.10846316069364548,
0.05552656948566437,
-0.07864449918270111,
-0.023253709077835083,
0.08823766559362411,
-0.12520289421081543,
-0.11614244431257248,
-0.031644344329833984,
-0.03653612732887268,
-0.1302632838487625,
0.008520007133483887,
0.08007746189832687,
-0.04704716056585312,
0.0065691228955984116,
0.023328015580773354,
0.039681464433670044,
-0.059111591428518295,
0.2049165517091751,
0.06030021235346794,
0.0734735056757927,
-0.09855253249406815,
0.12074180692434311,
0.03806367516517639,
-0.15448938310146332,
0.02626190148293972,
0.06856182217597961,
-0.06980060786008835,
-0.012746245600283146,
0.09614600241184235,
0.09654933214187622,
0.019370557740330696,
-0.052066605538129807,
-0.08841872215270996,
-0.12906599044799805,
0.08606961369514465,
0.12721990048885345,
0.034436192363500595,
0.025704586878418922,
-0.023403331637382507,
0.036015018820762634,
-0.10017671436071396,
0.12861835956573486,
0.09076954424381256,
0.08396196365356445,
-0.20309075713157654,
0.09238169342279434,
0.006026851944625378,
0.014228073880076408,
-0.02705582231283188,
0.046414170414209366,
-0.12546038627624512,
-0.01005737017840147,
-0.0831274762749672,
-0.014196369796991348,
-0.060376547276973724,
0.007402971852570772,
-0.008887574076652527,
-0.0463973805308342,
-0.061042193323373795,
0.01799420639872551,
-0.09031931310892105,
-0.03760562464594841,
0.008575058542191982,
0.07804141938686371,
-0.09653887152671814,
-0.007973941043019295,
0.03199614956974983,
-0.1066896915435791,
0.09864090383052826,
0.019219618290662766,
0.03178975731134415,
0.020545482635498047,
-0.14655549824237823,
0.004043412860482931,
0.044728267937898636,
0.0017933463677763939,
0.043620843440294266,
-0.1799335479736328,
-0.002676802221685648,
-0.05618853494524956,
0.0015225129900500178,
-0.012370980344712734,
0.0418928787112236,
-0.11107030510902405,
-0.0039626313373446465,
-0.056470610201358795,
-0.046753812581300735,
-0.05430954322218895,
0.03490125387907028,
0.09909588098526001,
-0.002588076749816537,
0.1605648696422577,
-0.07312863320112228,
0.030088145285844803,
-0.22570469975471497,
0.005343773402273655,
-0.01226352620869875,
-0.06171485036611557,
-0.08009107410907745,
-0.004595617298036814,
0.07994727045297623,
-0.04887566342949867,
0.07857837527990341,
-0.03885648399591446,
0.005453541409224272,
0.018048856407403946,
-0.06351912766695023,
0.026009460911154747,
0.04587191715836525,
0.18387821316719055,
0.00973803922533989,
-0.025028130039572716,
0.010795295238494873,
-0.008160458877682686,
0.06628745049238205,
0.08217896521091461,
0.17628523707389832,
0.15121330320835114,
-0.028105836361646652,
0.04490355774760246,
0.058428891003131866,
-0.15734267234802246,
-0.16328583657741547,
0.14505654573440552,
-0.07753445953130722,
0.14959806203842163,
0.010901632718741894,
0.1915326714515686,
0.08237984776496887,
-0.19656670093536377,
0.03992364928126335,
-0.0251215398311615,
-0.07592294365167618,
-0.1064484640955925,
-0.07770626991987228,
-0.07716371864080429,
-0.18754032254219055,
0.02068289741873741,
-0.10483946651220322,
0.04939713701605797,
0.042260799556970596,
0.03282589465379715,
0.032404907047748566,
0.14931011199951172,
0.03280926123261452,
0.007580154109746218,
0.07999766618013382,
0.05579029768705368,
-0.05919770896434784,
-0.03586547076702118,
-0.08774857968091965,
0.04528403654694557,
-0.03906300291419029,
0.026791652664542198,
-0.07646774500608444,
-0.10249662399291992,
0.10039462149143219,
0.04075204208493233,
-0.08953606337308884,
0.035328298807144165,
-0.031911153346300125,
0.045975036919116974,
0.10396980494260788,
0.016867531463503838,
0.0054029920138418674,
-0.017913157120347023,
0.20615927875041962,
-0.08554703742265701,
-0.029105035588145256,
-0.12713611125946045,
0.20474448800086975,
-0.03791787475347519,
-0.007961651310324669,
0.04734862968325615,
-0.0791289359331131,
-0.013430857099592686,
0.155223548412323,
0.16647110879421234,
-0.04309780150651932,
-0.022755440324544907,
0.016093185171484947,
-0.008729864843189716,
-0.05183398723602295,
0.09869074076414108,
0.12363072484731674,
0.09060762077569962,
-0.049409378319978714,
-0.053231507539749146,
-0.0682469978928566,
-0.028370238840579987,
-0.002809269353747368,
0.07091372460126877,
0.047917935997247696,
-0.03789903596043587,
-0.012421194463968277,
0.11267231404781342,
-0.10664857923984528,
-0.14490453898906708,
0.046175602823495865,
-0.1605062633752823,
-0.18934179842472076,
-0.06865964084863663,
0.09667466580867767,
0.015156214125454426,
0.028867734596133232,
-0.0031140397768467665,
-0.032765086740255356,
0.08789734542369843,
0.007347944658249617,
-0.06110360845923424,
-0.09165854007005692,
0.03785204887390137,
-0.055136021226644516,
0.20366927981376648,
-0.040369387716054916,
0.02649170160293579,
0.10169144719839096,
0.0701468363404274,
-0.10007276386022568,
0.03530391678214073,
0.07393486797809601,
-0.14900793135166168,
0.031388889998197556,
0.1908065527677536,
-0.05108822137117386,
0.12936781346797943,
0.04315489903092384,
-0.07526242733001709,
-0.011712946929037571,
-0.09146630764007568,
-0.05475602298974991,
-0.051825158298015594,
0.009070063941180706,
-0.056164246052503586,
0.15439535677433014,
0.1749674528837204,
-0.055783264338970184,
-0.013523348607122898,
-0.06738784164190292,
0.03868257999420166,
0.06269171088933945,
0.12210850417613983,
0.018467992544174194,
-0.24309559166431427,
0.034370217472314835,
-0.03212849050760269,
0.02356981858611107,
-0.22730296850204468,
-0.06973212957382202,
0.008035650476813316,
-0.042697444558143616,
-0.12093812227249146,
0.10050279647111893,
0.04837149381637573,
0.045144639909267426,
-0.05571841821074486,
-0.03175071254372597,
-0.04702845960855484,
0.1554143875837326,
-0.185140460729599,
-0.06709416955709457
] |
null | null | null |
# **Q-Learning** Agent playing1 **Taxi-v3**
This is a trained model of a **Q-Learning** agent playing **Taxi-v3** .
## Usage
```python
model = load_from_hub(repo_id="ergh0/Taxi-v3", filename="q-learning.pkl")
# Don't forget to check if you need to add additional attributes (is_slippery=False etc)
env = gym.make(model["env_id"])
```
| {"tags": ["Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "Taxi-v3", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "Taxi-v3", "type": "Taxi-v3"}, "metrics": [{"type": "mean_reward", "value": "7.56 +/- 2.71", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | ergh0/Taxi-v3 | [
"Taxi-v3",
"q-learning",
"reinforcement-learning",
"custom-implementation",
"model-index",
"region:us"
] | 2024-02-07T07:45:38+00:00 | [] | [] | TAGS
#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us
|
# Q-Learning Agent playing1 Taxi-v3
This is a trained model of a Q-Learning agent playing Taxi-v3 .
## Usage
| [
"# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage"
] | [
"TAGS\n#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n",
"# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage"
] | [
32,
33
] | [
"passage: TAGS\n#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage"
] | [
0.048862796276807785,
-0.16549694538116455,
-0.005485367961227894,
0.02960980497300625,
0.1345081776380539,
-0.01784728653728962,
0.11895976960659027,
0.07759871333837509,
-0.07461097836494446,
-0.055395450443029404,
0.1418241262435913,
0.09088201075792313,
0.055222880095243454,
0.05699880048632622,
0.09511256217956543,
-0.27440664172172546,
0.048217080533504486,
-0.02918700873851776,
0.05621987581253052,
0.11878681182861328,
0.0670095682144165,
-0.040441032499074936,
0.061956584453582764,
0.11818158626556396,
-0.1018151044845581,
-0.007344264071434736,
0.035402704030275345,
-0.09440053254365921,
0.17413531243801117,
0.07204403728246689,
0.12337774783372879,
0.05132639780640602,
0.179361954331398,
-0.12762396037578583,
0.024310702458024025,
-0.0010275895474478602,
-0.10138072073459625,
-0.03909514099359512,
-0.012415820732712746,
-0.08349097520112991,
0.03230205550789833,
0.23522862792015076,
0.07199250161647797,
0.06632792949676514,
-0.17707863450050354,
-0.06584878265857697,
-0.04375573247671127,
0.069611094892025,
0.14951466023921967,
0.03758616745471954,
-0.033800311386585236,
0.1684885323047638,
-0.2564343810081482,
0.05066783353686333,
0.037275806069374084,
-0.42313119769096375,
0.017119819298386574,
0.1507398933172226,
0.15090937912464142,
0.06909667700529099,
-0.10573802888393402,
0.013512322679162025,
0.051325585693120956,
-0.0005318621988408267,
0.024325110018253326,
0.006554204970598221,
0.15601307153701782,
0.08537693321704865,
-0.1487821787595749,
-0.058576688170433044,
0.17441977560520172,
-0.03788546845316887,
-0.02613203600049019,
-0.039745692163705826,
0.0067160045728087425,
-0.06427708268165588,
-0.004067842848598957,
-0.1777995079755783,
0.00734262028709054,
0.06666424125432968,
-0.014348524622619152,
0.014901017770171165,
-0.035522811114788055,
-0.0966939702630043,
-0.023098144680261612,
-0.08592145889997482,
0.01677769608795643,
-0.006319406442344189,
-0.10187895596027374,
0.05002119392156601,
-0.061138734221458435,
0.0014382408699020743,
-0.05123179033398628,
-0.15047866106033325,
-0.049055423587560654,
-0.03481535613536835,
0.1474713832139969,
-0.0044205985032022,
-0.01873963139951229,
-0.03164304047822952,
0.15474793314933777,
0.049551334232091904,
-0.05370146036148071,
0.05625450983643532,
0.07605006545782089,
0.23867930471897125,
0.10401605814695358,
0.10196955502033234,
-0.06798075139522552,
0.10180158913135529,
-0.12330973148345947,
-0.08915644884109497,
-0.17508824169635773,
0.11820860952138901,
0.00015364694991149008,
0.1317785084247589,
-0.12023144960403442,
0.07898581773042679,
-0.067511186003685,
0.013453764840960503,
0.01636839471757412,
0.0820009782910347,
-0.012399360537528992,
0.10676060616970062,
-0.005061192903667688,
-0.06941985338926315,
0.014177112840116024,
0.05935845896601677,
0.03754841163754463,
-0.038601722568273544,
-0.03192409873008728,
-0.05762290954589844,
-0.05065649375319481,
-0.10128600150346756,
-0.06447898596525192,
0.018573462963104248,
-0.007677143905311823,
-0.1833900660276413,
-0.06407523155212402,
0.00897200871258974,
0.015712225809693336,
-0.03988850116729736,
-0.05148044601082802,
-0.15265507996082306,
-0.042461175471544266,
-0.015450406819581985,
-0.03500641882419586,
-0.06214277446269989,
-0.0383245050907135,
0.046435944736003876,
-0.07560601085424423,
0.013364278711378574,
0.023342855274677277,
0.05405820533633232,
-0.025881100445985794,
0.06068144738674164,
-0.08357544988393784,
0.09493788331747055,
-0.1540430635213852,
-0.03271956741809845,
-0.025445878505706787,
-0.041183918714523315,
0.1752462536096573,
0.06099751964211464,
-0.015994304791092873,
0.15260063111782074,
-0.17141541838645935,
-0.058121129870414734,
0.15596486628055573,
0.008629098534584045,
-0.09967197477817535,
-0.003560945624485612,
-0.09397093951702118,
0.1428760588169098,
0.08571921288967133,
0.2478504776954651,
0.12005335837602615,
-0.22748184204101562,
0.055358242243528366,
0.12515293061733246,
-0.14365963637828827,
0.10365243256092072,
0.07344598323106766,
0.005470725707709789,
-0.18886831402778625,
-0.06843198090791702,
-0.06121627986431122,
0.1053021252155304,
-0.08522345870733261,
-0.0776243582367897,
0.09323626756668091,
-0.05086790770292282,
0.24641476571559906,
-0.028281206265091896,
0.06174173951148987,
-0.026681531220674515,
-0.1389324963092804,
-0.01723906397819519,
0.060955192893743515,
0.05258452147245407,
-0.024835573509335518,
-0.25895482301712036,
0.13646544516086578,
0.048650871962308884,
0.025074828416109085,
0.004106190986931324,
-0.05691491439938545,
0.016934165731072426,
0.1511998474597931,
0.020012924447655678,
0.13717477023601532,
0.027723990380764008,
0.0706823319196701,
-0.006239562761038542,
-0.10560829937458038,
-0.04169593006372452,
0.061916545033454895,
-0.08518962562084198,
-0.06641357392072678,
0.011197872459888458,
-0.06935211271047592,
-0.11783787608146667,
-0.12166737765073776,
-0.026334572583436966,
-0.02980303019285202,
-0.07444227486848831,
0.02368103712797165,
0.06536602973937988,
-0.06702698022127151,
-0.0023908785078674555,
0.007125476840883493,
-0.011537045240402222,
0.16434046626091003,
0.011393417604267597,
-0.007796820718795061,
0.1328643560409546,
-0.11533161997795105,
0.12461213022470474,
0.049438029527664185,
-0.024806302040815353,
-0.04662557691335678,
0.0014137453399598598,
-0.057529181241989136,
0.029044216498732567,
-0.04390640929341316,
0.02774495631456375,
0.20111067593097687,
0.02772962674498558,
0.11389166116714478,
-0.0656520202755928,
0.04385066404938698,
-0.007961965166032314,
-0.009693224914371967,
0.018563594669103622,
0.07608018070459366,
0.07813210040330887,
-0.1324140727519989,
0.02262016013264656,
0.22455167770385742,
0.1385764330625534,
0.18313980102539062,
-0.010877152904868126,
0.06325667351484299,
-0.04875868931412697,
0.027505528181791306,
0.024100203067064285,
0.10314226150512695,
-0.10732068121433258,
-0.0322517491877079,
-0.025407759472727776,
0.023599207401275635,
-0.08197105675935745,
-0.1055799350142479,
-0.090115025639534,
0.01222382951527834,
-0.03125503659248352,
-0.15570329129695892,
0.13300658762454987,
-0.10451057553291321,
0.01802753657102585,
0.04692702740430832,
-0.22163605690002441,
0.11530312895774841,
0.014291439205408096,
-0.10303618758916855,
0.11281087249517441,
-0.12051989883184433,
-0.08699832111597061,
-0.05777236074209213,
-0.18658851087093353,
0.05280197039246559,
0.04673841595649719,
0.05166793242096901,
-0.18521739542484283,
0.024835903197526932,
0.05545609071850777,
0.13426995277404785,
-0.09743253141641617,
-0.07142634689807892,
-0.15038461983203888,
0.016068490222096443,
-0.033661190420389175,
-0.16029728949069977,
-0.005609163548797369,
-0.032781440764665604,
-0.18849676847457886,
-0.04539939761161804,
-0.15086813271045685,
-0.034627582877874374,
0.20464378595352173,
0.026907702907919884,
0.09480511397123337,
-0.07926445454359055,
0.3802889585494995,
-0.042039383202791214,
-0.06146497279405594,
-0.01321389526128769,
-0.07072482258081436,
0.02512686513364315,
0.13271741569042206,
0.0036099457647651434,
-0.017886579036712646,
-0.0037857077550143003,
0.0024592927657067776,
-0.06234965845942497,
-0.13400450348854065,
0.0028710351325571537,
0.03905198723077774,
0.1874423623085022,
0.004639793653041124,
0.06659388542175293,
0.03133883699774742,
0.057546284049749374,
0.07748064398765564,
0.030926106497645378,
0.0011591583024710417,
-0.01591806672513485,
0.06604493409395218,
-0.11684755235910416,
0.042466625571250916,
-0.030429253354668617,
-0.10143838077783585,
-0.013183288276195526,
0.07950251549482346,
0.12755028903484344,
0.17849206924438477,
-0.04790908098220825,
0.17489230632781982,
0.13580141961574554,
0.16576050221920013,
0.049315933138132095,
-0.020801831036806107,
-0.08773037046194077,
-0.06118565797805786,
0.004774159751832485,
-0.031952597200870514,
0.04869702458381653,
0.3231290578842163,
0.037619613111019135,
-0.09036035090684891,
0.11149907857179642,
0.009480619803071022,
0.05359881371259689,
0.022797370329499245,
-0.11162138730287552,
0.11170321702957153,
0.07968773692846298,
-0.06341761350631714,
-0.07602835446596146,
0.16758501529693604,
-0.1109386757016182,
-0.26646625995635986,
-0.11410990357398987,
-0.012305386364459991,
0.07903840392827988,
0.005651174578815699,
0.05498376116156578,
-0.11829282343387604,
-0.16034497320652008,
-0.034191906452178955,
0.1335442066192627,
-0.3077351450920105,
0.2065143585205078,
-0.0198091771453619,
0.06707923114299774,
-0.039657969027757645,
-0.07026876509189606,
0.09694647043943405,
0.13174086809158325,
0.29124146699905396,
0.01396956667304039,
0.04841272905468941,
-0.15176129341125488,
-0.0976925864815712,
0.0018439020495861769,
0.015482662245631218,
-0.02563396655023098,
0.028520405292510986,
-0.0540912002325058,
0.008404579944908619,
-0.018086453899741173,
0.2102297693490982,
-0.11316607892513275,
0.004344627261161804,
-0.06968966871500015,
-0.11707738786935806,
0.19409789144992828,
-0.07178345322608948,
-0.04543264955282211,
-0.14959357678890228,
-0.15512511134147644,
-0.004174166824668646,
-0.02413962036371231,
-0.019664527848362923,
-0.17603960633277893,
-0.18804074823856354,
-0.05204557999968529,
-0.005645004566758871,
-0.003464865731075406,
0.05867868289351463,
-0.07517234236001968,
-0.04805335775017738,
0.1009904220700264,
-0.07743175327777863,
-0.056063808500766754,
-0.1103200614452362,
0.1391381323337555,
0.06248528137803078,
0.16743235290050507,
0.05907081440091133,
0.0006117874872870743,
0.11471151560544968,
-0.02913086675107479,
0.11103474348783493,
-0.11291708797216415,
-0.17145049571990967,
-0.08334989100694656,
-0.018775060772895813,
0.09519003331661224,
-0.04789286106824875,
0.0028788831550627947,
0.2550160884857178,
0.14880181849002838,
-0.0897710770368576,
0.27680760622024536,
0.04414956644177437,
-0.09375058114528656,
-0.18432219326496124,
-0.15961645543575287,
0.03759992495179176,
0.060025621205568314,
0.13095876574516296,
-0.057205069810152054,
-0.08483537286520004,
-0.08492398262023926,
-0.07478608191013336,
-0.13140805065631866,
-0.24232175946235657,
-0.030598774552345276,
0.22874866425991058,
0.08656918257474899,
0.08219650387763977,
-0.012482990510761738,
-0.01186054851859808,
0.00526038184762001,
0.02680150233209133,
0.12018456310033798,
-0.13341329991817474,
0.11107480525970459,
0.022198403254151344,
0.044267985969781876,
0.009712530300021172,
0.07929777354001999,
0.03375575691461563,
-0.003218587953597307,
-0.0006439819699153304,
-0.0988350659608841,
-0.2596651017665863,
0.0816885456442833,
-0.01623627357184887,
-0.09960969537496567,
0.014988959766924381,
0.02061903104186058,
-0.2089255303144455,
0.011128270998597145,
-0.019883770495653152,
-0.03150356933474541,
-0.06483490765094757,
-0.10664787143468857,
-0.056551624089479446,
0.04928823933005333,
0.10853826254606247,
0.011660109274089336,
0.05354316532611847,
-0.0404130220413208,
0.07917837053537369,
0.0826287642121315,
0.15132710337638855,
0.06795957684516907,
-0.190711110830307,
-0.10953907668590546,
-0.0414445661008358,
0.12121522426605225,
-0.12505418062210083,
0.036917757242918015,
0.053161121904850006,
-0.016534561291337013,
0.14621229469776154,
0.1070784479379654,
-0.07452095299959183,
0.11915595084428787,
0.08904775977134705,
-0.04094788804650307,
-0.23367151618003845,
-0.07120766490697861,
0.11133213341236115,
0.07195597887039185,
-0.03961895406246185,
0.018120890483260155,
-0.04960581287741661,
-0.013980977237224579,
0.048759616911411285,
-0.0538676381111145,
-0.07230538129806519,
0.004421027842909098,
0.1247575581073761,
0.1029362753033638,
-0.04655474051833153,
0.01296416949480772,
0.037371400743722916,
0.003788623260334134,
0.04730486497282982,
0.0407949760556221,
-0.08269952982664108,
-0.04124005511403084,
0.02782733179628849,
0.37552911043167114,
-0.010165480896830559,
-0.020456433296203613,
0.018555615097284317,
-0.19949445128440857,
0.09135842323303223,
0.13205479085445404,
0.04697350412607193,
0.004247748292982578,
-0.08139242231845856,
0.026877427473664284,
-0.010625290684401989,
0.09936143457889557,
-0.07806670665740967,
-0.05493134260177612,
-0.21631066501140594,
-0.025010565295815468,
0.017490221187472343,
0.24077683687210083,
-0.08458559215068817,
-0.12801732122898102,
-0.20628872513771057,
0.13128381967544556,
-0.11333390325307846,
-0.03695881739258766,
-0.024473199620842934,
0.03926658630371094,
-0.01989821158349514,
0.06291737407445908,
-0.0710630789399147,
0.006373001262545586,
-0.11024709790945053,
0.055267609655857086,
0.04204455390572548,
0.1229788213968277,
0.014207782223820686,
0.02016810141503811,
0.05822525918483734,
-0.01837925612926483,
0.07173580676317215,
-0.06203491613268852,
-0.04550490900874138,
0.14224006235599518,
-0.020255116745829582,
-0.04152837023139,
-0.0483345128595829,
-0.036874305456876755,
0.11981741338968277,
-0.05059147998690605,
-0.007141099311411381,
-0.054929375648498535,
-0.06906463205814362,
0.03462086617946625,
-0.009175732731819153,
-0.008798843249678612,
0.06801853328943253,
0.04024988040328026,
-0.026994358748197556,
0.005263668950647116,
0.03447828069329262,
-0.10330043733119965,
-0.04955084249377251,
0.16955432295799255,
-0.0749620869755745,
0.10274054110050201,
-0.031069839373230934,
0.018015999346971512,
0.005847334861755371,
-0.022399673238396645,
-0.015360680408775806,
-0.1457086056470871,
-0.06137600541114807,
-0.09489979594945908,
0.11565322428941727,
0.08146517723798752,
0.03358805552124977,
0.04274565726518631,
0.019532648846507072,
-0.04414922371506691,
-0.038583990186452866,
0.12961317598819733,
0.08133101463317871,
0.012996876612305641,
0.01137041300535202,
0.01941833831369877,
-0.020302120596170425,
0.0028480992186814547,
-0.01250747125595808,
-0.07239153981208801,
-0.05874783173203468,
0.09400010108947754,
0.1600283533334732,
-0.06127211079001427,
-0.13325586915016174,
-0.020593497902154922,
0.04988488554954529,
0.0014717020094394684,
-0.08777432143688202,
0.04833676666021347,
0.15805292129516602,
-0.05623878911137581,
0.03216489031910896,
-0.09984751045703888,
-0.07263360917568207,
-0.16060975193977356,
-0.10029061883687973,
-0.06092562898993492,
-0.28350353240966797,
0.09752398729324341,
0.006392303854227066,
-0.014731393195688725,
0.059529416263103485,
0.051305368542671204,
-0.052508849650621414,
0.07068239152431488,
-0.18146829307079315,
-0.007054794579744339,
0.03497592359781265,
-0.13212306797504425,
0.02475893869996071,
-0.2378365397453308,
0.10198072344064713,
-0.04623803123831749,
-0.1519704908132553,
-0.04004510119557381,
0.0641569048166275,
-0.09540136158466339,
-0.01822364516556263,
-0.0475153923034668,
-0.01922670193016529,
0.01624443754553795,
-0.009348669089376926,
-0.031147832050919533,
0.13716529309749603,
0.02827494591474533,
-0.03268734738230705,
0.005254602525383234,
0.0223685409873724,
0.03955082967877388,
-0.0969657450914383,
-0.05986930429935455,
0.08311155438423157,
-0.031056145206093788,
0.14728976786136627,
0.000341245875461027,
0.04181376099586487,
-0.06758682429790497,
0.2593761384487152,
0.2023983597755432,
-0.12479214370250702,
0.008118697442114353,
-0.021801479160785675,
0.012670028023421764,
-0.041751839220523834,
0.13110700249671936,
0.013386172242462635,
0.12186761200428009,
-0.17513342201709747,
-0.01036517322063446,
-0.0818324014544487,
-0.04501292482018471,
0.06702108681201935,
0.14714950323104858,
0.15742522478103638,
0.03436789661645889,
-0.07328428328037262,
0.06722653657197952,
-0.30119743943214417,
0.20540550351142883,
-0.1346001923084259,
-0.01498429011553526,
-0.040251150727272034,
-0.058389630168676376,
0.061147745698690414,
0.11309876292943954,
0.10832664370536804,
-0.021150551736354828,
-0.0905047357082367,
-0.04486766457557678,
-0.039378076791763306,
-0.13019338250160217,
-0.02718670479953289,
0.1654091775417328,
0.06799814850091934,
0.31520840525627136,
-0.017577875405550003,
0.07702425122261047,
0.034410297870635986,
0.06451138854026794,
0.004519328009337187,
0.09537279605865479,
0.07960964739322662,
-0.06345855444669724,
-0.07373003661632538,
-0.001637450186535716,
0.05033271387219429,
0.14567798376083374,
-0.03826142102479935,
-0.18691548705101013,
0.15858715772628784,
0.07192251086235046,
-0.13762691617012024,
-0.05777517706155777,
0.08409425616264343,
-0.0739973932504654,
0.0550808347761631,
0.08115427941083908,
0.015876613557338715,
-0.017793258652091026,
-0.004664506763219833,
0.06074233725667,
0.024694660678505898,
-0.02343848906457424,
0.003570882137864828,
-0.08337053656578064,
-0.04151543974876404,
0.07267895340919495,
-0.0844460055232048,
-0.20546193420886993,
-0.0957019031047821,
-0.07551700621843338,
0.030557552352547646,
-0.0649830624461174,
0.12575586140155792,
0.1717868149280548,
0.0593598335981369,
-0.03307248651981354,
-0.10721943527460098,
-0.035562749952077866,
0.07602505385875702,
-0.044773899018764496,
-0.09409699589014053
] |
null | null | diffusers | # MedChMTS
<Gallery />
## Model description
Medieval characters bad dataset

## Trigger words
You should use `character concept` to trigger the image generation.
You should use `in the style of medchmts` to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
[Download](/Bukowski89/MedChMTS/tree/main) them in the Files & versions tab.
| {"tags": ["text-to-image", "stable-diffusion", "lora", "diffusers", "template:sd-lora"], "widget": [{"text": "character concept of a medieval soldier", "output": {"url": "images/Capture.JPG"}}], "base_model": "stabilityai/stable-diffusion-xl-base-1.0", "instance_prompt": "character concept, in the style of medchmts"} | text-to-image | Bukowski89/MedChMTS | [
"diffusers",
"text-to-image",
"stable-diffusion",
"lora",
"template:sd-lora",
"base_model:stabilityai/stable-diffusion-xl-base-1.0",
"region:us"
] | 2024-02-07T07:49:11+00:00 | [] | [] | TAGS
#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #region-us
| # MedChMTS
<Gallery />
## Model description
Medieval characters bad dataset
!URL
## Trigger words
You should use 'character concept' to trigger the image generation.
You should use 'in the style of medchmts' to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
Download them in the Files & versions tab.
| [
"# MedChMTS\n\n<Gallery />",
"## Model description \n\nMedieval characters bad dataset\n\n\n!URL",
"## Trigger words\n\nYou should use 'character concept' to trigger the image generation.\n\nYou should use 'in the style of medchmts' to trigger the image generation.",
"## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab."
] | [
"TAGS\n#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #region-us \n",
"# MedChMTS\n\n<Gallery />",
"## Model description \n\nMedieval characters bad dataset\n\n\n!URL",
"## Trigger words\n\nYou should use 'character concept' to trigger the image generation.\n\nYou should use 'in the style of medchmts' to trigger the image generation.",
"## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab."
] | [
56,
10,
12,
38,
28
] | [
"passage: TAGS\n#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #region-us \n# MedChMTS\n\n<Gallery />## Model description \n\nMedieval characters bad dataset\n\n\n!URL## Trigger words\n\nYou should use 'character concept' to trigger the image generation.\n\nYou should use 'in the style of medchmts' to trigger the image generation.## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab."
] | [
-0.11959878355264664,
-0.0135732376947999,
-0.000047560632083332166,
0.006840785499662161,
0.14950983226299286,
0.053602200001478195,
0.14242663979530334,
0.04282201826572418,
0.11018884181976318,
0.08935346454381943,
0.15649153292179108,
-0.042428091168403625,
-0.027040423825383186,
0.13083301484584808,
-0.012036916799843311,
-0.2590581774711609,
0.047434404492378235,
0.004334812518209219,
-0.11057661473751068,
0.039571259170770645,
0.038531478494405746,
-0.07993801683187485,
0.1174149438738823,
-0.05901798978447914,
0.010553312487900257,
0.01464028563350439,
0.0015539717860519886,
0.008024190552532673,
0.00751158595085144,
0.00716740544885397,
-0.06763409823179245,
0.08838941156864166,
0.043404608964920044,
-0.07931031286716461,
0.06785216182470322,
-0.009562012739479542,
-0.0252138189971447,
0.037354644387960434,
-0.0006732271867804229,
-0.06415948271751404,
0.16655229032039642,
-0.16128461062908173,
-0.08106080442667007,
-0.03623427823185921,
-0.03649231418967247,
-0.09133239090442657,
-0.02815857343375683,
-0.013391642831265926,
0.092171810567379,
-0.01871431991457939,
0.02664973959326744,
0.07835293561220169,
-0.06252757459878922,
0.05151687189936638,
0.2656620740890503,
-0.15999427437782288,
-0.026045165956020355,
0.202846959233284,
0.08625269681215286,
0.2269556224346161,
-0.04841449111700058,
0.18087852001190186,
0.009126243181526661,
-0.024237187579274178,
0.013171795755624771,
-0.057661812752485275,
0.13104099035263062,
0.021995922550559044,
-0.10226940363645554,
0.06738600879907608,
0.30860984325408936,
0.0009798106038942933,
0.026257412508130074,
-0.08691678196191788,
-0.014447473920881748,
0.09901171922683716,
-0.08294858038425446,
-0.03616650402545929,
-0.003843311918899417,
0.004877702333033085,
0.02447373978793621,
-0.11975302547216415,
-0.04570094496011734,
-0.037443794310092926,
-0.028662262484431267,
0.09793098270893097,
-0.022501735016703606,
0.05174155533313751,
-0.004122738726437092,
0.028397809714078903,
-0.24127797782421112,
-0.11852778494358063,
-0.0060222870670259,
-0.017153067514300346,
0.0568685419857502,
-0.010167459957301617,
-0.05432828515768051,
-0.10845252871513367,
0.14968155324459076,
-0.04732280597090721,
0.021988272666931152,
0.012600606307387352,
0.01446332037448883,
0.09074194729328156,
0.06338175386190414,
-0.07370781153440475,
-0.15049327909946442,
-0.08811234682798386,
0.13975495100021362,
0.11298397183418274,
0.09916110336780548,
-0.026919668540358543,
-0.13722002506256104,
-0.07399550080299377,
0.012926011346280575,
0.04583822190761566,
-0.001822253572754562,
0.027748335152864456,
-0.009041952900588512,
-0.005209855269640684,
0.17338207364082336,
-0.007384954486042261,
-0.03352717310190201,
0.01675744354724884,
-0.06289775669574738,
0.1675271987915039,
0.06650660187005997,
-0.01197534054517746,
0.016484754160046577,
0.05637100711464882,
-0.04445420578122139,
-0.0034048280213028193,
-0.02621697261929512,
-0.08674798160791397,
0.04124671220779419,
-0.010169253684580326,
0.006345808040350676,
-0.06726963818073273,
-0.28861504793167114,
0.02363188937306404,
0.0501449890434742,
-0.029396438971161842,
0.1380900740623474,
-0.018743524327874184,
-0.06360984593629837,
0.006621480453759432,
0.025976333767175674,
-0.06777223199605942,
-0.07301528751850128,
0.06950397789478302,
-0.11234590411186218,
0.16682811081409454,
-0.21764525771141052,
0.0029670996591448784,
-0.07957948744297028,
0.032401762902736664,
-0.35953694581985474,
0.058399174362421036,
-0.023740066215395927,
-0.004450042732059956,
-0.05628443881869316,
-0.01616855338215828,
-0.0998324528336525,
0.0358155183494091,
0.07651485502719879,
0.17854636907577515,
-0.16828392446041107,
-0.01914687268435955,
-0.01905525103211403,
-0.1486929953098297,
-0.17604056000709534,
0.1241525411605835,
-0.03311421722173691,
0.09302490204572678,
0.11576594412326813,
0.11582666635513306,
0.08095423877239227,
-0.17741262912750244,
0.023440483957529068,
0.010188141837716103,
-0.0818110927939415,
-0.10760164260864258,
0.06590704619884491,
0.021240508183836937,
0.06357212364673615,
0.06268488615751266,
-0.2441146820783615,
0.07301302254199982,
-0.07457812875509262,
0.025790905579924583,
0.0027491615619510412,
-0.12897729873657227,
0.025605063885450363,
0.021027203649282455,
0.06449256092309952,
-0.036136142909526825,
0.04727816581726074,
0.14570505917072296,
0.07206110656261444,
-0.059844009578228,
0.014131993986666203,
0.03650350868701935,
0.17983049154281616,
-0.16945798695087433,
0.04709398373961449,
-0.07937747240066528,
-0.14220720529556274,
-0.012441080994904041,
0.15705816447734833,
0.04590725153684616,
-0.06720098853111267,
0.12605884671211243,
0.10127004981040955,
-0.12917229533195496,
0.021667156368494034,
0.17331358790397644,
0.0012254758039489388,
-0.10399174690246582,
-0.18351967632770538,
-0.0033422857522964478,
-0.040877815335989,
0.09401405602693558,
-0.15127412974834442,
0.004311209078878164,
0.05646393448114395,
0.06762028485536575,
0.04603129252791405,
-0.0064038909040391445,
0.0627155527472496,
-0.0532817542552948,
-0.024216650053858757,
0.01039771270006895,
0.029123792424798012,
-0.03645901009440422,
-0.10920976847410202,
0.15968525409698486,
-0.1694834977388382,
0.12988153100013733,
0.17015233635902405,
-0.035973917692899704,
0.02460794523358345,
-0.2067071795463562,
0.0876956507563591,
0.0023063451517373323,
-0.0011701224138960242,
-0.12182322889566422,
-0.11782842129468918,
-0.005520264618098736,
0.04557298868894577,
-0.0753735676407814,
0.027613239362835884,
-0.006704927422106266,
-0.07731401175260544,
-0.08697903156280518,
0.08040102571249008,
0.10960183292627335,
-0.06809069216251373,
0.032067470252513885,
0.20291465520858765,
-0.13120140135288239,
0.16577742993831635,
0.022626832127571106,
-0.08398118615150452,
-0.03831141069531441,
0.040638502687215805,
0.047415293753147125,
0.11488685756921768,
0.06207265332341194,
0.008168809115886688,
0.03806542232632637,
-0.06111449375748634,
0.037551816552877426,
-0.05653472989797592,
-0.0267635565251112,
-0.017547428607940674,
-0.040483664721250534,
0.05363919585943222,
0.07942695915699005,
-0.08244097977876663,
0.129977285861969,
-0.06277908384799957,
0.027189482003450394,
-0.0007440402405336499,
-0.012879611924290657,
-0.0848865732550621,
0.11299053579568863,
-0.1184152290225029,
-0.08171974122524261,
-0.12387789785861969,
0.024827897548675537,
-0.1334286332130432,
0.02214227244257927,
-0.021206391975283623,
-0.08789186924695969,
-0.04668458178639412,
-0.10489940643310547,
0.0011776566971093416,
0.062064532190561295,
-0.030161935836076736,
-0.02435535378754139,
-0.045754484832286835,
-0.08905433118343353,
-0.050215184688568115,
-0.04904169216752052,
-0.1155422180891037,
0.06098432093858719,
0.08717550337314606,
-0.14483289420604706,
0.10967109352350235,
0.08276830613613129,
-0.01158348098397255,
0.0682985708117485,
-0.01630862057209015,
0.09755486994981766,
-0.01686171628534794,
0.13107749819755554,
0.22133810818195343,
0.08479675650596619,
0.04649186506867409,
0.059298332780599594,
0.05115262046456337,
-0.07674329727888107,
0.021904323250055313,
-0.02466440014541149,
-0.12368602305650711,
0.009077009744942188,
-0.11678043752908707,
-0.10983987152576447,
-0.1328420639038086,
-0.014966464601457119,
0.02653351239860058,
0.018491357564926147,
0.10863907635211945,
-0.03319108113646507,
-0.09653902053833008,
0.11454843729734421,
0.03474987670779228,
0.0012991850962862372,
0.01982519030570984,
0.025694239884614944,
-0.025712590664625168,
0.006383870728313923,
0.1586722433567047,
-0.06164491921663284,
0.21115003526210785,
-0.059388916939496994,
0.10883694887161255,
0.04126931354403496,
-0.013767330907285213,
0.09206247329711914,
0.0653827041387558,
-0.058817218989133835,
-0.035176295787096024,
-0.022523922845721245,
-0.18179260194301605,
0.04435574263334274,
0.059660714119672775,
-0.017429720610380173,
0.001995903206989169,
-0.030531683936715126,
0.10407772660255432,
0.080097496509552,
-0.07517435401678085,
0.008628950454294682,
-0.2981860339641571,
0.02080540359020233,
0.06610254943370819,
0.19572807848453522,
-0.005776606034487486,
0.014635537751019001,
0.07733351737260818,
0.025420671328902245,
0.06362617015838623,
-0.022395161911845207,
0.08000214397907257,
0.05411980301141739,
-0.016327450051903725,
-0.05282111093401909,
0.17290440201759338,
-0.08171959221363068,
-0.03243914246559143,
-0.12990513443946838,
0.07136805355548859,
0.03302265703678131,
0.020188894122838974,
-0.04759446159005165,
-0.08195149153470993,
0.14503808319568634,
0.19747380912303925,
0.05129075050354004,
0.04920722916722298,
-0.01439017802476883,
-0.051820412278175354,
-0.04676702618598938,
-0.02037038281559944,
0.06313281506299973,
-0.03971768915653229,
0.03196073696017265,
0.03579919785261154,
-0.045338667929172516,
0.030324818566441536,
-0.06240447610616684,
-0.1377774477005005,
-0.07654326409101486,
-0.020084302872419357,
0.17567779123783112,
0.013311144895851612,
-0.03427034988999367,
-0.06233253702521324,
-0.09533267468214035,
0.0020966341253370047,
0.21642117202281952,
-0.10090389102697372,
-0.11965855211019516,
0.052059952169656754,
0.13187794387340546,
-0.047651488333940506,
-0.012651833705604076,
-0.03479549661278725,
0.06395891308784485,
-0.0680813416838646,
-0.11865951865911484,
0.07746288180351257,
-0.08915923535823822,
-0.054051607847213745,
-0.041720401495695114,
0.13869449496269226,
0.06275118142366409,
0.003173236735165119,
0.019547246396541595,
0.04506043717265129,
0.03965546190738678,
-0.0720752626657486,
0.0253811776638031,
0.09352416545152664,
0.02702326886355877,
0.1950240433216095,
-0.07983953505754471,
-0.075691357254982,
0.04208097979426384,
0.03708962723612785,
0.10937340557575226,
0.22016757726669312,
-0.030797749757766724,
0.006186732556670904,
0.27393460273742676,
-0.025868380442261696,
-0.2963063418865204,
0.0793694406747818,
-0.02475813776254654,
-0.027748508378863335,
0.0394456647336483,
-0.009127151221036911,
0.17799711227416992,
0.10071830451488495,
-0.01943771354854107,
0.14034993946552277,
-0.31990140676498413,
-0.13388694822788239,
0.0063431779853999615,
0.160608172416687,
0.33424660563468933,
-0.23685869574546814,
-0.035952527076005936,
-0.08270937204360962,
-0.09692632406949997,
0.02738167904317379,
-0.08446168154478073,
0.0175166092813015,
-0.024973643943667412,
-0.09027113020420074,
-0.001234987867064774,
-0.06648513674736023,
0.19683445990085602,
-0.0020798479672521353,
0.08413145691156387,
-0.061456289142370224,
-0.010887211188673973,
0.21663670241832733,
-0.056567635387182236,
0.11609255522489548,
-0.13763704895973206,
-0.03468090668320656,
-0.13932351768016815,
0.017404768615961075,
-0.04513240605592728,
0.02746744640171528,
0.015445043332874775,
-0.04231163486838341,
-0.02938898652791977,
0.022071879357099533,
-0.0033823971170932055,
0.024443412199616432,
-0.023337524384260178,
-0.09443080425262451,
-0.026770323514938354,
0.0969696119427681,
0.032546598464250565,
-0.08776766061782837,
-0.00521597545593977,
0.037703730165958405,
-0.04277754947543144,
0.07810923457145691,
-0.0667012482881546,
-0.027263782918453217,
0.10364293307065964,
0.029588500037789345,
0.08274953067302704,
0.02185114473104477,
0.021457456052303314,
0.074210524559021,
0.1093023419380188,
-0.11757085472345352,
-0.12062212824821472,
-0.05893091484904289,
-0.05407603085041046,
0.023211073130369186,
0.04193829372525215,
0.12205848842859268,
-0.035723309963941574,
0.06089036166667938,
-0.01154135912656784,
0.07187074422836304,
0.0015724593540653586,
0.11576785147190094,
0.07681915909051895,
-0.03715173155069351,
-0.0969669297337532,
0.0623590387403965,
-0.02407168410718441,
-0.040696822106838226,
-0.012462549842894077,
0.08703059703111649,
-0.14040721952915192,
-0.07087939232587814,
0.0683281421661377,
0.11426911503076553,
-0.07460577040910721,
-0.020641183480620384,
-0.06638140231370926,
-0.09462075680494308,
-0.025752892717719078,
0.11357728391885757,
0.07193414866924286,
-0.12524385750293732,
-0.009336472488939762,
-0.005519078113138676,
-0.08323311060667038,
0.1168728768825531,
0.08304913341999054,
0.06497538834810257,
-0.08014200627803802,
-0.07899264246225357,
-0.05030479282140732,
-0.058472152799367905,
-0.09912766516208649,
-0.040750470012426376,
-0.06240241602063179,
0.007223090622574091,
-0.11753135174512863,
0.06907400488853455,
-0.15114541351795197,
-0.0428728424012661,
-0.03680311143398285,
-0.06915899366140366,
-0.03386543691158295,
0.014025244861841202,
-0.05985509604215622,
0.016791705042123795,
0.028507187962532043,
0.020782005041837692,
-0.09060148894786835,
-0.10220450907945633,
-0.0003506603534333408,
-0.02766033448278904,
0.09386661648750305,
0.0916837826371193,
-0.0677344873547554,
0.02786976844072342,
-0.21818986535072327,
-0.06640595942735672,
0.0964064970612526,
0.04873612895607948,
-0.020413465797901154,
0.03236309438943863,
0.019109345972537994,
0.04076887667179108,
0.011458957567811012,
0.028096944093704224,
0.03035874105989933,
-0.01231782604008913,
0.0736592561006546,
-0.12554337084293365,
0.0623440146446228,
-0.0337575301527977,
0.07034897059202194,
0.16433703899383545,
0.060425594449043274,
0.07389680296182632,
-0.08865171670913696,
-0.05194610357284546,
-0.10560929775238037,
0.007132980041205883,
0.005823412444442511,
-0.14355891942977905,
-0.1786017268896103,
0.016603542491793633,
0.06540640443563461,
-0.0004717332776635885,
0.11756207048892975,
-0.06745707988739014,
-0.12263061851263046,
-0.015156100504100323,
0.2331693470478058,
0.1793353408575058,
0.006661885417997837,
0.2005382776260376,
0.012595812790095806,
0.02962740883231163,
-0.11220170557498932,
0.11331692337989807,
0.1289050281047821,
-0.13074153661727905,
0.0972190648317337,
0.05614734813570976,
0.008608202449977398,
0.07572027295827866,
-0.0005698124878108501,
0.10010018944740295,
0.11085669696331024,
0.07773800939321518,
-0.03132829815149307,
0.03370938077569008,
0.007919901050627232,
0.008872252888977528,
0.2433493584394455,
-0.07926267385482788,
-0.030869126319885254,
0.021990807726979256,
-0.027140013873577118,
-0.06315909326076508,
-0.21932955086231232,
-0.03234148398041725,
-0.1576370745897293,
-0.03715630993247032,
-0.12083718925714493,
0.017963383346796036,
0.10869258642196655,
0.04896370694041252,
0.03827960044145584,
0.15982550382614136,
-0.09271818399429321,
-0.025629861280322075,
0.08270511776208878,
-0.09039049595594406,
-0.008532978594303131,
0.06827894598245621,
-0.06696569919586182,
0.1143759936094284,
-0.07989144325256348,
-0.016077006235718727,
0.09808999300003052,
0.0442761667072773,
-0.0025003550108522177,
-0.08158823847770691,
-0.09824905544519424,
-0.04948781058192253,
0.027815887704491615,
-0.027631891891360283,
0.09859366714954376,
0.07757613062858582,
-0.10083391517400742,
-0.00841321051120758,
0.07615645974874496,
0.03648802638053894,
-0.0676109716296196,
-0.01962094195187092,
0.03450142219662666,
-0.0034635644406080246,
0.0065887910313904285,
-0.04132884740829468,
-0.08585750311613083,
0.034064676612615585,
0.13318398594856262,
0.27714329957962036,
-0.1739439070224762,
0.007780042476952076,
-0.10349361598491669,
0.010314545594155788,
0.02167212776839733,
0.09177226573228836,
-0.0030269287526607513,
0.17161080241203308,
-0.03773188963532448,
0.05782974138855934,
-0.1519048511981964,
-0.04545710235834122,
-0.14952802658081055,
-0.1060090884566307,
0.08163084089756012,
-0.09234032779932022,
-0.022175844758749008,
0.15141995251178741,
-0.05880914255976677,
-0.06338349729776382,
-0.0935121402144432,
-0.10805962234735489,
0.014154094271361828,
-0.12768812477588654,
-0.029835151508450508,
0.09454028308391571,
-0.0584370382130146,
-0.11576647311449051,
0.02728305011987686,
-0.09004247933626175,
-0.00193404674064368,
-0.15134297311306,
-0.04159681871533394,
0.026308443397283554,
0.0015372132183983922,
0.029288791120052338,
-0.004966289736330509,
0.032401472330093384,
0.04676095396280289,
-0.0625164806842804,
-0.03596574068069458,
0.059638895094394684,
0.0005743172951042652,
-0.04170951619744301,
0.016085609793663025,
0.03546133264899254,
-0.02365199476480484,
0.052873291075229645,
0.03703876957297325,
-0.03772272169589996,
0.009196741506457329,
0.03594031557440758,
-0.09008213877677917,
-0.040571533143520355,
0.09555937349796295,
-0.07601775228977203,
0.08193063735961914,
0.017227256670594215,
-0.007085885386914015,
0.003000464756041765,
0.001414768397808075,
0.033339399844408035,
0.07197416573762894,
0.01210382953286171,
0.032954249531030655,
-0.1170724406838417,
-0.009862530045211315,
-0.021125759929418564,
0.02371334843337536,
-0.22482989728450775,
-0.03998871520161629,
-0.21090076863765717,
0.03150753676891327,
-0.037569716572761536,
0.0933859571814537,
0.22254875302314758,
-0.0066451518796384335,
-0.01793757639825344,
-0.10990956425666809,
0.02924528531730175,
0.08502333611249924,
-0.11060745269060135,
-0.030938394367694855
] |
null | null | ml-agents |
# **ppo** Agent playing **Huggy**
This is a trained model of a **ppo** agent playing **Huggy**
using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
## Usage (with ML-Agents)
The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/
We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
- A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your
browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction
- A *longer tutorial* to understand how works ML-Agents:
https://huggingface.co/learn/deep-rl-course/unit5/introduction
### Resume the training
```bash
mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume
```
### Watch your Agent play
You can watch your agent **playing directly in your browser**
1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity
2. Step 1: Find your model_id: Atozzio/ppo-Huggy
3. Step 2: Select your *.nn /*.onnx file
4. Click on Watch the agent play 👀
| {"library_name": "ml-agents", "tags": ["Huggy", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Huggy"]} | reinforcement-learning | Atozzio/ppo-Huggy | [
"ml-agents",
"tensorboard",
"onnx",
"Huggy",
"deep-reinforcement-learning",
"reinforcement-learning",
"ML-Agents-Huggy",
"region:us"
] | 2024-02-07T07:50:14+00:00 | [] | [] | TAGS
#ml-agents #tensorboard #onnx #Huggy #deep-reinforcement-learning #reinforcement-learning #ML-Agents-Huggy #region-us
|
# ppo Agent playing Huggy
This is a trained model of a ppo agent playing Huggy
using the Unity ML-Agents Library.
## Usage (with ML-Agents)
The Documentation: URL
We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
- A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your
browser: URL
- A *longer tutorial* to understand how works ML-Agents:
URL
### Resume the training
### Watch your Agent play
You can watch your agent playing directly in your browser
1. If the environment is part of ML-Agents official environments, go to URL
2. Step 1: Find your model_id: Atozzio/ppo-Huggy
3. Step 2: Select your *.nn /*.onnx file
4. Click on Watch the agent play
| [
"# ppo Agent playing Huggy\n This is a trained model of a ppo agent playing Huggy\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: Atozzio/ppo-Huggy\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play"
] | [
"TAGS\n#ml-agents #tensorboard #onnx #Huggy #deep-reinforcement-learning #reinforcement-learning #ML-Agents-Huggy #region-us \n",
"# ppo Agent playing Huggy\n This is a trained model of a ppo agent playing Huggy\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: Atozzio/ppo-Huggy\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play"
] | [
44,
199
] | [
"passage: TAGS\n#ml-agents #tensorboard #onnx #Huggy #deep-reinforcement-learning #reinforcement-learning #ML-Agents-Huggy #region-us \n# ppo Agent playing Huggy\n This is a trained model of a ppo agent playing Huggy\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: Atozzio/ppo-Huggy\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play"
] | [
0.008570084348320961,
0.012797973118722439,
-0.004060710780322552,
0.04699372872710228,
0.14654789865016937,
0.010679289698600769,
0.18720906972885132,
0.12984125316143036,
0.10840514302253723,
0.11266540735960007,
0.05287069454789162,
0.05301020294427872,
0.07244794815778732,
0.17161571979522705,
0.07417457550764084,
-0.2048119306564331,
-0.013516535051167011,
-0.08072344213724136,
0.033370241522789,
0.09146498888731003,
0.04449281096458435,
-0.0378994457423687,
0.08709920197725296,
0.023255012929439545,
-0.03149823099374771,
-0.013553358614444733,
-0.07073161005973816,
-0.03464936092495918,
0.05465712770819664,
0.01062761340290308,
-0.036679867655038834,
-0.038326576352119446,
0.06574205309152603,
-0.2117723524570465,
0.034023433923721313,
0.08636949956417084,
-0.0127339456230402,
0.014533307403326035,
0.11656852066516876,
0.03778518736362457,
0.1092902421951294,
-0.05041981115937233,
0.05575283244252205,
0.05867958813905716,
-0.08171334862709045,
0.02126801200211048,
-0.13746777176856995,
0.024518361315131187,
0.2112603634595871,
0.09120868891477585,
0.004363736603409052,
0.11041499674320221,
-0.08830684423446655,
0.03146621584892273,
0.1646398901939392,
-0.2061523050069809,
-0.07222938537597656,
0.06299372017383575,
0.07297008484601974,
0.009048338048160076,
-0.04147063195705414,
0.04670986533164978,
-0.014171459712088108,
0.028372477740049362,
0.07726152241230011,
-0.03030305914580822,
0.25819242000579834,
-0.035239268094301224,
-0.07251454144716263,
-0.0736246332526207,
0.037425778806209564,
0.08776719868183136,
-0.07765825092792511,
-0.21748735010623932,
0.03711686655879021,
0.1619253158569336,
-0.029216885566711426,
0.0007502386579290032,
0.06896398216485977,
-0.024998748674988747,
-0.03657035157084465,
-0.10979940742254257,
-0.05473872646689415,
-0.06116315722465515,
0.10539431124925613,
0.17534656822681427,
-0.002709694905206561,
-0.042294785380363464,
0.06691252440214157,
0.06335437297821045,
0.016719406470656395,
-0.023603327572345734,
-0.016778305172920227,
-0.029015833511948586,
-0.09891033917665482,
-0.005464350804686546,
-0.007050998508930206,
0.06544511020183563,
0.061445388942956924,
0.14807595312595367,
0.04658099263906479,
0.00950029119849205,
0.023336105048656464,
0.051694173365831375,
-0.00946075189858675,
0.11576347053050995,
0.00679535698145628,
0.01445829775184393,
0.04820414260029793,
0.05376877635717392,
0.05694146454334259,
-0.06977654993534088,
-0.11437687277793884,
0.07015746831893921,
-0.1303476095199585,
0.08646460622549057,
0.09650030732154846,
0.018801983445882797,
-0.0794806256890297,
-0.04718075320124626,
0.003172557568177581,
-0.13215908408164978,
0.08136497437953949,
0.04951775446534157,
-0.04748864844441414,
-0.09551215916872025,
-0.037473153322935104,
0.011458730325102806,
-0.08538803458213806,
0.028309937566518784,
-0.029004398733377457,
0.043685998767614365,
-0.010252779349684715,
-0.04267670214176178,
0.10513269156217575,
-0.0701650008559227,
-0.01905583217740059,
-0.1487879902124405,
-0.08687040954828262,
-0.07937002182006836,
0.044717080891132355,
-0.061481282114982605,
-0.12515361607074738,
-0.0518459677696228,
0.037275832146406174,
-0.1034960001707077,
-0.011050637811422348,
-0.0264996737241745,
-0.06629736721515656,
-0.0008726023370400071,
-0.023517819121479988,
0.07078811526298523,
0.1536737084388733,
0.03957231342792511,
-0.0078110164031386375,
0.07230443507432938,
-0.17121319472789764,
0.09191589802503586,
-0.09734918922185898,
0.17886660993099213,
-0.05536012351512909,
0.020459799095988274,
0.027482150122523308,
0.019027505069971085,
0.03384268283843994,
0.1938091367483139,
-0.053312890231609344,
-0.11829615384340286,
0.13663731515407562,
-0.026963356882333755,
-0.11733400076627731,
0.04956015944480896,
0.023294992744922638,
0.08792342245578766,
0.03792310878634453,
0.23784731328487396,
0.09320104122161865,
-0.2531057596206665,
0.05951499938964844,
0.017775176092982292,
-0.1584598869085312,
-0.0007544442196376622,
0.14120562374591827,
-0.04140659421682358,
0.01229987945407629,
-0.006083074491471052,
-0.16961811482906342,
0.07137709110975266,
-0.013924933038651943,
-0.03488987311720848,
0.0551886186003685,
-0.022973908111453056,
-0.036618683487176895,
-0.00880243256688118,
-0.0009265049011446536,
-0.04864165931940079,
-0.09334117919206619,
-0.0884755402803421,
0.06990565359592438,
-0.003578235162422061,
0.06881902366876602,
-0.06694123148918152,
0.1323026418685913,
0.04624907299876213,
0.05422206595540047,
-0.08093348145484924,
-0.0947645753622055,
0.010299780406057835,
0.03403783217072487,
0.1027444526553154,
-0.11418767273426056,
0.05715811625123024,
0.07169738411903381,
0.00018803926650434732,
-0.07470222562551498,
-0.10939367115497589,
-0.012940971180796623,
-0.058087315410375595,
-0.121157705783844,
-0.06421744078397751,
-0.07111523300409317,
0.1429419219493866,
-0.10155434906482697,
0.06339842826128006,
-0.09603921324014664,
0.03720272332429886,
-0.02073228359222412,
-0.0320831723511219,
0.042748358100652695,
-0.011654573492705822,
0.02779710665345192,
-0.07142829895019531,
0.09744574874639511,
0.035980112850666046,
-0.07536111772060394,
0.08401530981063843,
-0.060404688119888306,
-0.066468246281147,
0.08871060609817505,
0.020109837874770164,
-0.02996441163122654,
-0.08237901329994202,
-0.0780922919511795,
0.013392341323196888,
-0.08743997663259506,
0.00011181759327882901,
0.11807164549827576,
0.09905900806188583,
0.11488252878189087,
-0.08811023086309433,
-0.06186242774128914,
-0.010299762710928917,
-0.11278688907623291,
-0.0643455982208252,
0.15318185091018677,
0.03539852797985077,
0.07232708483934402,
0.05023336037993431,
0.0759958028793335,
0.07427719980478287,
0.07565080374479294,
0.024151740595698357,
-0.12632901966571808,
-0.011778205633163452,
0.08535422384738922,
0.05379963293671608,
-0.006777359638363123,
0.025072652846574783,
0.0026628742925822735,
0.022994419559836388,
-0.03796980157494545,
-0.02343185432255268,
-0.13066914677619934,
-0.07225526124238968,
0.0065861898474395275,
-0.03213377669453621,
0.03721913695335388,
-0.03883086517453194,
-0.048423442989587784,
0.07137785851955414,
0.1168065071105957,
0.01179493498057127,
0.007479189429432154,
-0.05018325522542,
-0.11029298603534698,
0.07028956711292267,
-0.0662166029214859,
-0.3022249639034271,
-0.12229608744382858,
-0.15325090289115906,
-0.06764183938503265,
0.03235001489520073,
0.051558107137680054,
-0.1622489094734192,
-0.007642511278390884,
-0.09801091998815536,
-0.04639348015189171,
0.0674152597784996,
-0.06835734099149704,
0.18841736018657684,
0.1186835989356041,
0.016069022938609123,
-0.05979566648602486,
-0.021808739751577377,
-0.007989542558789253,
-0.048030510544776917,
0.03642963618040085,
0.031801410019397736,
0.0541408509016037,
0.12310851365327835,
0.08359885215759277,
0.0545206293463707,
-0.02584654651582241,
0.07002552598714828,
-0.08787725120782852,
-0.011714653111994267,
0.13098548352718353,
-0.011113462969660759,
0.08285550773143768,
0.031416673213243484,
0.02993200719356537,
-0.022986533120274544,
0.05378824099898338,
0.018690206110477448,
-0.07052639126777649,
-0.19560745358467102,
-0.0991833359003067,
-0.02392307110130787,
0.20304182171821594,
0.0801803469657898,
0.08165598660707474,
-0.040859971195459366,
-0.02534467540681362,
-0.007484420668333769,
-0.05799500644207001,
0.15756219625473022,
0.1104554533958435,
-0.04429168999195099,
-0.08623883873224258,
-0.009771239012479782,
-0.02904101274907589,
0.006079436745494604,
0.11018889397382736,
0.02611655555665493,
0.07360135763883591,
0.03996824845671654,
0.02003495581448078,
0.01816609874367714,
-0.06842658668756485,
-0.10116162896156311,
0.07315269112586975,
0.04910026863217354,
-0.003909206949174404,
-0.040879443287849426,
-0.09149880707263947,
-0.02648892253637314,
0.10243269801139832,
0.13128989934921265,
-0.0874176025390625,
-0.09765741974115372,
0.04822378605604172,
0.08980782330036163,
0.10919412970542908,
0.00654555344954133,
-0.145194873213768,
-0.0313723087310791,
0.00404325919225812,
-0.10034708678722382,
0.0030463915318250656,
-0.025456426665186882,
0.03181827813386917,
-0.1867094188928604,
0.06865561008453369,
0.02254910208284855,
0.1229885146021843,
0.04489481821656227,
0.002277218271046877,
0.031336504966020584,
0.09176045656204224,
-0.022361431270837784,
0.06385725736618042,
-0.19006656110286713,
0.06572485715150833,
-0.0059207733720541,
0.093081995844841,
-0.050222378224134445,
0.015694165602326393,
0.0848165825009346,
-0.018706124275922775,
0.17747047543525696,
0.036042142659425735,
0.08496345579624176,
-0.08467463403940201,
-0.16919998824596405,
-0.06787168979644775,
0.012766272760927677,
-0.08354977518320084,
0.05826043710112572,
0.0038475380279123783,
-0.043124232441186905,
-0.09808138012886047,
0.16244103014469147,
0.02634563483297825,
-0.05089428648352623,
0.020044634118676186,
-0.06255690008401871,
-0.007917948998510838,
-0.04664679616689682,
-0.0383889302611351,
-0.0247971098870039,
0.21172615885734558,
0.14856378734111786,
0.005216390360146761,
-0.09582289308309555,
-0.0542634092271328,
-0.0562024861574173,
-0.014164624735713005,
-0.020043332129716873,
-0.014444582164287567,
0.1321694254875183,
-0.08667748421430588,
-0.027502145618200302,
-0.00945002119988203,
-0.10486922413110733,
-0.11687631905078888,
-0.012119024991989136,
0.2275192141532898,
-0.00744594307616353,
0.08946970850229263,
-0.019245529547333717,
0.007869874127209187,
-0.008485947735607624,
-0.08140330016613007,
0.16004100441932678,
0.16978690028190613,
0.027914835140109062,
0.06565055251121521,
-0.11036700010299683,
0.039009083062410355,
-0.11503171920776367,
-0.014208292588591576,
0.16906093060970306,
0.325131356716156,
-0.008754990063607693,
0.19036993384361267,
0.09003528207540512,
-0.06651697307825089,
-0.21325483918190002,
-0.055338263511657715,
0.038854166865348816,
-0.004876455292105675,
0.14222881197929382,
-0.14352670311927795,
0.01916864886879921,
0.03883304074406624,
-0.012673032470047474,
0.044303297996520996,
-0.14785251021385193,
-0.079727403819561,
-0.013025420717895031,
0.05602509155869484,
0.04450945556163788,
-0.072775699198246,
-0.062517449259758,
-0.0440269373357296,
-0.06490309536457062,
0.06979595869779587,
-0.15705038607120514,
0.09265460073947906,
0.013474139384925365,
0.012708090245723724,
0.04777410253882408,
-0.020058415830135345,
0.1262395828962326,
-0.06195322796702385,
-0.04191100224852562,
-0.0972989872097969,
-0.004924509208649397,
0.021029815077781677,
-0.124601349234581,
0.05604257434606552,
-0.05478343367576599,
-0.07591371983289719,
-0.20599791407585144,
-0.05094285309314728,
-0.03804086148738861,
0.07153740525245667,
-0.011550475843250751,
-0.006378734949976206,
-0.005899116862565279,
0.07713855803012848,
0.07507939636707306,
0.057666368782520294,
0.045834120362997055,
-0.017814233899116516,
-0.020790403708815575,
0.11389756202697754,
0.10214648395776749,
0.012885755859315395,
-0.07878635823726654,
-0.04245300590991974,
-0.026778506115078926,
-0.027018237859010696,
-0.0610780231654644,
-0.005855835974216461,
0.03027193993330002,
0.014916982501745224,
0.044560160487890244,
0.05363672971725464,
-0.10630803555250168,
-0.032339900732040405,
0.08357986062765121,
-0.10088473558425903,
-0.0914420410990715,
-0.05676741525530815,
-0.08099137246608734,
-0.06428321450948715,
-0.06205075979232788,
0.04504017531871796,
-0.019137447699904442,
0.0008988569024950266,
0.047115910798311234,
0.054404497146606445,
-0.07340137660503387,
0.04884413257241249,
-0.004975521005690098,
0.01419366616755724,
-0.05975721776485443,
0.14786183834075928,
0.010279824025928974,
-0.03730865940451622,
0.0197832603007555,
0.19595037400722504,
-0.05118643119931221,
-0.08174612373113632,
-0.042300134897232056,
0.03812992572784424,
0.15458716452121735,
-0.04264259710907936,
-0.0362200066447258,
-0.06197253242135048,
0.0779736191034317,
-0.1301802098751068,
0.0014194807736203074,
-0.09604235738515854,
0.02318214438855648,
0.09213639795780182,
-0.11533543467521667,
0.08302316069602966,
0.010450253263115883,
-0.0498056523501873,
-0.09563639760017395,
0.08607028424739838,
0.050937723368406296,
0.17100121080875397,
-0.0238079521805048,
-0.058921799063682556,
-0.1445530652999878,
0.009715637192130089,
0.012012211605906487,
-0.010758225806057453,
-0.1879505068063736,
-0.01694251224398613,
-0.031559694558382034,
0.0443330742418766,
-0.01103048026561737,
0.031069163233041763,
-0.05532486364245415,
-0.07000190019607544,
-0.06517720222473145,
0.1000455915927887,
-0.03177730739116669,
-0.02485828287899494,
0.01921871304512024,
-0.08215564489364624,
0.11024351418018341,
0.06829046458005905,
-0.029676955193281174,
-0.05646085366606712,
-0.05142936110496521,
-0.016626106575131416,
0.016718104481697083,
-0.03861881420016289,
0.03769807890057564,
-0.17948879301548004,
0.025168854743242264,
-0.04311927407979965,
-0.10785292834043503,
0.0064627788960933685,
0.09108811616897583,
-0.08758582174777985,
0.07253327965736389,
-0.0019070764537900686,
-0.15695102512836456,
-0.08186125755310059,
0.017862340435385704,
0.014490746892988682,
0.05099556967616081,
0.07896294444799423,
-0.05510316416621208,
0.18060551583766937,
-0.12715528905391693,
-0.003024503355845809,
0.01245205570012331,
0.00485001876950264,
-0.005311835091561079,
-0.0915156677365303,
0.040688712149858475,
-0.004498459864407778,
0.12435994297266006,
0.10242483764886856,
-0.050765350461006165,
0.022613095119595528,
0.01715312898159027,
0.11084124445915222,
0.0034441992174834013,
0.0297224260866642,
-0.024614423513412476,
-0.0018208224792033434,
0.043555211275815964,
-0.0031560848001390696,
0.06311070919036865,
-0.13955189287662506,
0.09284385293722153,
0.09574165940284729,
0.13327214121818542,
0.07213144749403,
0.10288678854703903,
-0.1003376767039299,
-0.18428273499011993,
-0.02250850759446621,
0.006545029114931822,
0.030857715755701065,
-0.07889974862337112,
0.22835634648799896,
0.10682310909032822,
-0.20919090509414673,
0.06509923189878464,
-0.005270285066217184,
0.01885811798274517,
-0.08071652054786682,
-0.1376478672027588,
-0.001480186590924859,
-0.21337121725082397,
0.06924184411764145,
-0.057349372655153275,
0.021853186190128326,
-0.03194563835859299,
-0.029009390622377396,
-0.02180892415344715,
0.06671907007694244,
-0.11610245704650879,
-0.06408638507127762,
0.10399086773395538,
-0.041163306683301926,
0.03091319091618061,
-0.008579188957810402,
-0.043834276497364044,
-0.04925166442990303,
-0.05588550865650177,
0.06673935055732727,
0.06273160874843597,
0.03600121662020683,
0.04147947579622269,
-0.06349083036184311,
-0.06431366503238678,
0.03990575671195984,
-0.02096761204302311,
0.02863883785903454,
0.12878873944282532,
0.029879270121455193,
-0.10707379877567291,
0.00401879521086812,
0.20491228997707367,
-0.06045914813876152,
0.023707903921604156,
-0.09229996055364609,
0.1609024703502655,
-0.016915157437324524,
-0.05160403624176979,
-0.06585755944252014,
-0.0999763086438179,
-0.08402998745441437,
0.21341869235038757,
0.11106061190366745,
-0.0441766083240509,
0.0017380908830091357,
-0.025480275973677635,
0.025011496618390083,
0.02191024459898472,
0.12859608232975006,
0.062247999012470245,
0.11970937997102737,
-0.064520925283432,
-0.020483093336224556,
-0.0035647607874125242,
-0.0676618441939354,
-0.16983924806118011,
0.004617467988282442,
0.04369780048727989,
-0.027572516351938248,
-0.042878158390522,
0.08013708889484406,
-0.08108535408973694,
-0.11828260123729706,
0.10782613605260849,
-0.08679697662591934,
-0.08425260335206985,
-0.022365674376487732,
-0.006385850254446268,
0.028621496632695198,
0.14718884229660034,
0.057152267545461655,
0.013370855711400509,
0.0890352725982666,
-0.028896579518914223,
-0.04299250990152359,
0.004393588285893202,
0.0837564766407013,
-0.0848761722445488,
0.22062231600284576,
-0.03775393217802048,
0.04230233281850815,
0.04736042395234108,
0.03016444481909275,
-0.14177733659744263,
0.06392566859722137,
0.009912406094372272,
-0.17358309030532837,
0.01744600385427475,
0.05353960767388344,
-0.07543611526489258,
-0.06076463311910629,
0.06427024304866791,
-0.04370098561048508,
0.009868549183011055,
0.10918335616588593,
0.005294593516737223,
-0.059384386986494064,
0.08542861044406891,
-0.17338718473911285,
0.10054085403680801,
0.13740254938602448,
-0.06153637915849686,
-0.014377258718013763,
-0.05847522243857384,
0.032662104815244675,
0.048301294445991516,
0.06208622828125954,
-0.0194578617811203,
-0.16889597475528717,
0.001532917027361691,
0.015025808475911617,
0.041620053350925446,
-0.2765084505081177,
-0.11616561561822891,
-0.0359724797308445,
-0.03347381204366684,
-0.03198068216443062,
0.10112062841653824,
0.10693880170583725,
-0.01192004419863224,
-0.010397126898169518,
-0.19274921715259552,
0.04218124598264694,
0.17354807257652283,
-0.08912358433008194,
-0.0033991318196058273
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-finetuned-emotion
distilbert is a variant of bert model(one of LLM models). This model with a classification head is used to classify the emotions of the input tweet.
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2195
- Accuracy: 0.9235
- F1: 0.9233
## Emotion Labels
- **label_0:** Sadness
- **label_1:** Joy
- **label_2:** Love
- **label_3:** Anger
- **label_4:** Fear
- **label_5:** Surprise
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
| 0.8537 | 1.0 | 250 | 0.3235 | 0.897 | 0.8958 |
| 0.2506 | 2.0 | 500 | 0.2195 | 0.9235 | 0.9233 |
### Validation metrics
- test_loss : 0.2194512039422989
- test_accuracy : 0.9235
- test_f1 : 0.923296474937779
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["emotion"], "metrics": ["accuracy", "f1"], "base_model": "distilbert-base-uncased", "model-index": [{"name": "distilbert-base-uncased-finetuned-emotion", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "emotion", "type": "emotion", "config": "split", "split": "validation", "args": "split"}, "metrics": [{"type": "accuracy", "value": 0.9235, "name": "Accuracy"}, {"type": "f1", "value": 0.923296474937779, "name": "F1"}]}]}]} | text-classification | pt-sk/distilbert-base-uncased-finetuned-emotion | [
"transformers",
"tensorboard",
"safetensors",
"distilbert",
"text-classification",
"generated_from_trainer",
"dataset:emotion",
"base_model:distilbert-base-uncased",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T07:56:49+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #distilbert #text-classification #generated_from_trainer #dataset-emotion #base_model-distilbert-base-uncased #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| distilbert-base-uncased-finetuned-emotion
=========================================
distilbert is a variant of bert model(one of LLM models). This model with a classification head is used to classify the emotions of the input tweet.
This model is a fine-tuned version of distilbert-base-uncased on the emotion dataset.
It achieves the following results on the evaluation set:
* Loss: 0.2195
* Accuracy: 0.9235
* F1: 0.9233
Emotion Labels
--------------
* label\_0: Sadness
* label\_1: Joy
* label\_2: Love
* label\_3: Anger
* label\_4: Fear
* label\_5: Surprise
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 64
* eval\_batch\_size: 64
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 2
### Training results
### Validation metrics
* test\_loss : 0.2194512039422989
* test\_accuracy : 0.9235
* test\_f1 : 0.923296474937779
### Framework versions
* Transformers 4.37.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Validation metrics\n\n\n* test\\_loss : 0.2194512039422989\n* test\\_accuracy : 0.9235\n* test\\_f1 : 0.923296474937779",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #distilbert #text-classification #generated_from_trainer #dataset-emotion #base_model-distilbert-base-uncased #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Validation metrics\n\n\n* test\\_loss : 0.2194512039422989\n* test\\_accuracy : 0.9235\n* test\\_f1 : 0.923296474937779",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
82,
98,
4,
46,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #distilbert #text-classification #generated_from_trainer #dataset-emotion #base_model-distilbert-base-uncased #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2### Training results### Validation metrics\n\n\n* test\\_loss : 0.2194512039422989\n* test\\_accuracy : 0.9235\n* test\\_f1 : 0.923296474937779### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.1542709469795227,
0.179172083735466,
-0.002975629409775138,
0.11538854241371155,
0.0814908966422081,
0.025383496657013893,
0.12972544133663177,
0.15913385152816772,
-0.03218161687254906,
0.07925713062286377,
0.1573803424835205,
0.10204383730888367,
0.03534587472677231,
0.11918793618679047,
-0.07326939702033997,
-0.1637640744447708,
0.0019779906142503023,
0.03092661127448082,
-0.06349920481443405,
0.1539923995733261,
0.10132023692131042,
-0.1015482023358345,
0.08745655417442322,
-0.001432117191143334,
-0.11724833399057388,
0.01302256528288126,
0.01860240288078785,
-0.048832524567842484,
0.09782889485359192,
0.01745680719614029,
0.09538151323795319,
0.03939313441514969,
0.05907783284783363,
-0.19113226234912872,
0.0031657449435442686,
0.031765181571245193,
0.010191995650529861,
0.07651806622743607,
0.06140683591365814,
0.004556264262646437,
0.08368182927370071,
-0.05197279900312424,
0.06633801013231277,
0.02778535708785057,
-0.11396212130784988,
-0.23424099385738373,
-0.11059549450874329,
0.08382997661828995,
0.08010324835777283,
0.11715387552976608,
-0.03903579339385033,
0.1765247881412506,
-0.06251532584428787,
0.10260205715894699,
0.17719373106956482,
-0.2579871118068695,
-0.08282218873500824,
0.02337721176445484,
0.060090452432632446,
0.04617089033126831,
-0.10911347717046738,
-0.039812762290239334,
0.059224095195531845,
0.046357885003089905,
0.06553252041339874,
-0.030912891030311584,
0.02474156767129898,
0.012369617819786072,
-0.13621777296066284,
-0.07483794540166855,
0.1639067679643631,
0.08160415291786194,
-0.07361046224832535,
-0.07382000237703323,
-0.028019843623042107,
-0.16209886968135834,
-0.004321619868278503,
-0.03407175466418266,
0.026137040928006172,
-0.047138284891843796,
-0.07724430412054062,
0.0249638594686985,
-0.09643005579710007,
-0.08620162308216095,
0.0003980398178100586,
0.10118437558412552,
0.03950429707765579,
-0.0013249319745227695,
0.0038066343404352665,
0.1043931096792221,
0.005585950333625078,
-0.19700759649276733,
-0.022715067490935326,
0.010939782485365868,
-0.022181818261742592,
-0.049426350742578506,
-0.03307560086250305,
-0.06174474209547043,
0.052433669567108154,
0.1584702432155609,
-0.0943160280585289,
0.07108423858880997,
0.016512371599674225,
0.0048837000504136086,
-0.09260950237512589,
0.1976425051689148,
-0.007167099975049496,
-0.03509224206209183,
-0.011694501154124737,
0.16936524212360382,
0.013493488542735577,
0.012416640296578407,
-0.044682059437036514,
0.017535585910081863,
0.1525009274482727,
0.04553358256816864,
-0.03414124250411987,
0.008647841401398182,
-0.08384228497743607,
-0.013101475313305855,
0.05546917766332626,
-0.08691132813692093,
0.03411678597331047,
0.03233025595545769,
-0.066782645881176,
-0.0971827581524849,
0.013629753142595291,
-0.0023355185985565186,
0.013170675374567509,
0.061625126749277115,
-0.08505436778068542,
-0.004488370381295681,
-0.06570637226104736,
-0.07849445194005966,
0.0213213749229908,
-0.04245788976550102,
0.014639290049672127,
-0.08613511174917221,
-0.18916092813014984,
-0.08279170840978622,
0.02439211681485176,
-0.0346590057015419,
-0.0659475103020668,
-0.03975102677941322,
-0.09417641162872314,
0.025236468762159348,
0.004957750905305147,
0.09819739311933517,
-0.05052172392606735,
0.12161578238010406,
0.09458354860544205,
0.0327281653881073,
0.04688403382897377,
0.025035342201590538,
-0.10698358714580536,
0.05964820086956024,
-0.12841182947158813,
0.04176335409283638,
-0.10118561238050461,
0.05686017870903015,
-0.10948652774095535,
-0.1052689328789711,
0.008163564838469028,
-0.022853480651974678,
0.08595854043960571,
0.15290161967277527,
-0.16002947092056274,
-0.052323244512081146,
0.1368907392024994,
-0.09909278899431229,
-0.16733239591121674,
0.12467556446790695,
-0.033088892698287964,
0.01986091211438179,
0.038999855518341064,
0.16725119948387146,
0.11429779976606369,
-0.08748959749937057,
-0.02486986294388771,
-0.03510567173361778,
0.09579028934240341,
0.05197899043560028,
0.12107241153717041,
-0.02291659265756607,
-0.022535890340805054,
-0.004903800785541534,
-0.07235943526029587,
0.01946764439344406,
-0.08469858020544052,
-0.09032715857028961,
-0.055665165185928345,
-0.10037720203399658,
0.06963654607534409,
0.08756674081087112,
0.036382853984832764,
-0.12139791250228882,
-0.117083840072155,
-0.009389643557369709,
0.11873600631952286,
-0.053339313715696335,
-0.003937598317861557,
-0.08602607995271683,
0.11831848323345184,
-0.0646272674202919,
-0.0424937941133976,
-0.1653251200914383,
-0.016037479043006897,
0.05567815154790878,
-0.013822563923895359,
-0.010141649283468723,
-0.021121829748153687,
0.0832834392786026,
0.04266095906496048,
-0.05562157928943634,
-0.08584188669919968,
-0.07449279725551605,
0.00457573588937521,
-0.10739745944738388,
-0.20591387152671814,
-0.022879324853420258,
-0.024833232164382935,
0.18880735337734222,
-0.2652572691440582,
0.026943955570459366,
0.012875041924417019,
0.10345734655857086,
0.023165615275502205,
-0.05008586496114731,
-0.009203287772834301,
0.030961401760578156,
-0.03291648253798485,
-0.06688255071640015,
0.022471334785223007,
0.002855425700545311,
-0.10286902636289597,
-0.04479362070560455,
-0.11067449301481247,
0.15091083943843842,
0.0888611227273941,
0.01852390170097351,
-0.11742682754993439,
0.020990626886487007,
-0.05986044928431511,
-0.035972610116004944,
-0.04159404709935188,
-0.0019173554610460997,
0.13496768474578857,
0.012785215862095356,
0.11597009748220444,
-0.10039068013429642,
-0.05126084014773369,
0.018090922385454178,
-0.028129318729043007,
-0.000011622623787843622,
0.18145622313022614,
-0.00979202426970005,
-0.16576853394508362,
0.138833150267601,
0.15015465021133423,
-0.07905161380767822,
0.1159203052520752,
-0.09303297102451324,
-0.08997385203838348,
-0.0616198368370533,
0.04111620783805847,
0.022833362221717834,
0.14580847322940826,
-0.09267941117286682,
0.03184755519032478,
0.014748739078640938,
-0.002898188540711999,
-0.004533275496214628,
-0.18076853454113007,
-0.0356743223965168,
0.0607437938451767,
-0.05441978946328163,
-0.009332245215773582,
0.007001962978392839,
0.008947152644395828,
0.0866146832704544,
-0.007784461602568626,
-0.07115846872329712,
0.01513091754168272,
-0.012279089540243149,
-0.11746752262115479,
0.22348439693450928,
-0.05812663584947586,
-0.13335749506950378,
-0.13943390548229218,
0.0498170480132103,
-0.046303071081638336,
-0.0027717433404177427,
0.05716688930988312,
-0.07536716014146805,
-0.020126216113567352,
-0.11676620692014694,
-0.05982518196105957,
0.037366826087236404,
0.00993505958467722,
0.02577337436378002,
-0.016020406037569046,
0.10916834324598312,
-0.08610071986913681,
-0.005680889822542667,
-0.0040570637211203575,
-0.005837671924382448,
0.02697524055838585,
0.0024825604632496834,
0.10890091955661774,
0.11310321092605591,
-0.002021188149228692,
0.025099104270339012,
-0.003975950181484222,
0.27493369579315186,
-0.09626861661672592,
-0.04568587988615036,
0.14295892417430878,
-0.021416932344436646,
0.04032343253493309,
0.10620829463005066,
0.026842961087822914,
-0.08741158246994019,
0.018747294321656227,
0.033660028129816055,
0.008635267615318298,
-0.24481044709682465,
0.007817773148417473,
-0.014367884024977684,
-0.011888667941093445,
0.0893068015575409,
0.03685729205608368,
0.03355536609888077,
0.07378789037466049,
0.021168099716305733,
-0.01645197533071041,
-0.033523060381412506,
0.06877659261226654,
0.1021348237991333,
0.03974795341491699,
0.11029190570116043,
-0.045771438628435135,
-0.03469306230545044,
0.030946938320994377,
-0.027816584333777428,
0.18714022636413574,
-0.008020272478461266,
0.12404618412256241,
0.06042462959885597,
0.17202597856521606,
-0.023392267525196075,
0.022873347625136375,
0.008568131364881992,
-0.039103757590055466,
0.030546030029654503,
-0.059939462691545486,
-0.030413663014769554,
0.031242966651916504,
-0.013018546625971794,
0.10012900829315186,
-0.11947208642959595,
-0.00484606996178627,
0.03441851586103439,
0.24208015203475952,
0.08139351010322571,
-0.3213295638561249,
-0.10954979062080383,
0.0052900901064276695,
-0.05427908897399902,
-0.054198529571294785,
-0.0060028815641999245,
0.10851866006851196,
-0.09667579084634781,
0.055911432951688766,
-0.06694772094488144,
0.08020704239606857,
-0.0037554234731942415,
0.025892823934555054,
0.05632674694061279,
0.011567661538720131,
-0.010328011587262154,
0.07618098706007004,
-0.256117582321167,
0.2614983320236206,
0.01001756638288498,
0.0711061954498291,
-0.04132058471441269,
0.0056747086346149445,
0.04539733752608299,
0.04116647318005562,
0.1065206527709961,
-0.019895849749445915,
-0.04071080684661865,
-0.1978159099817276,
-0.0733911544084549,
0.024713197723031044,
0.06877549737691879,
-0.05508895963430405,
0.11916007846593857,
-0.032327815890312195,
0.0025966314133256674,
0.03744838386774063,
-0.01596331223845482,
-0.09982221573591232,
-0.028601594269275665,
0.022498274222016335,
-0.01586247608065605,
-0.020118391141295433,
-0.09749703109264374,
-0.0983535423874855,
-0.07969344407320023,
0.13755296170711517,
-0.07365116477012634,
-0.0306681077927351,
-0.10827483236789703,
0.06830694526433945,
0.14222204685211182,
-0.09065791219472885,
0.010581262409687042,
0.002094422234222293,
0.09103385359048843,
0.03899731859564781,
-0.010989200323820114,
0.062422528862953186,
-0.04906895011663437,
-0.2380148321390152,
-0.048259954899549484,
0.1462153196334839,
0.03208962082862854,
0.06680778414011002,
0.01846020109951496,
0.03888882324099541,
-0.04328937083482742,
-0.07336665689945221,
0.03569900617003441,
-0.0621035173535347,
0.05323072895407677,
0.03381472826004028,
0.006219307892024517,
0.02524731494486332,
-0.08989796787500381,
-0.049568887799978256,
0.1072612777352333,
0.32094132900238037,
-0.06546469777822495,
-0.015024140477180481,
-0.003386772470548749,
-0.06069621816277504,
-0.13513654470443726,
0.04707641527056694,
0.08419395983219147,
0.039528071880340576,
0.07068922370672226,
-0.10903030633926392,
0.049364909529685974,
0.08590183407068253,
-0.04794220253825188,
0.12264543771743774,
-0.27025797963142395,
-0.1345340758562088,
0.127848818898201,
0.13097570836544037,
-0.0008660387829877436,
-0.19313810765743256,
-0.07108035683631897,
0.005952164530754089,
-0.0853082686662674,
0.05032270774245262,
-0.05395302176475525,
0.09307795017957687,
-0.04176914319396019,
0.045276939868927,
0.04193891957402229,
-0.0512847937643528,
0.18387602269649506,
-0.03857263922691345,
0.08479323238134384,
-0.026065213605761528,
0.009050044231116772,
0.04183245077729225,
-0.06999226659536362,
0.05339040607213974,
-0.1150488331913948,
0.07729940116405487,
-0.16562268137931824,
-0.012411633506417274,
-0.09370553493499756,
0.033369697630405426,
-0.07561281323432922,
-0.013251092284917831,
-0.03457208350300789,
0.0584992878139019,
0.035006094723939896,
-0.017897343263030052,
0.10568975657224655,
-0.025958022102713585,
0.17828576266765594,
0.15740233659744263,
0.10607847571372986,
0.0355127789080143,
-0.06913596391677856,
0.015053797513246536,
-0.024465588852763176,
0.030752109363675117,
-0.15362095832824707,
0.024910837411880493,
0.14530256390571594,
0.04010790213942528,
0.14939384162425995,
0.02759145200252533,
-0.07808880507946014,
-0.020868975669145584,
0.06381122767925262,
-0.1416231095790863,
-0.05698646977543831,
-0.03698232024908066,
-0.028864644467830658,
-0.18182748556137085,
-0.020840855315327644,
0.11889521032571793,
-0.04681577533483505,
-0.019713707268238068,
-0.004935280419886112,
0.01994914561510086,
-0.034416843205690384,
0.2259448766708374,
0.06230901926755905,
0.08096425980329514,
-0.06904027611017227,
0.05412496626377106,
0.02629462070763111,
-0.10403987765312195,
0.02390403300523758,
0.06936035305261612,
-0.0473753847181797,
-0.04709808528423309,
0.04163394123315811,
0.1980968564748764,
-0.030756862834095955,
-0.05183747410774231,
-0.17858874797821045,
-0.1093299612402916,
0.05623826012015343,
0.18327946960926056,
0.09617841988801956,
0.03684443607926369,
0.013269023969769478,
-0.014514517970383167,
-0.09600964188575745,
0.1521720439195633,
0.10972333699464798,
0.07996387034654617,
-0.15776222944259644,
0.07506720721721649,
-0.029574355110526085,
0.019250009208917618,
-0.00966227613389492,
0.050066132098436356,
-0.11757748574018478,
-0.005908852443099022,
-0.14983311295509338,
0.01808864437043667,
-0.01420239731669426,
0.0004225534212309867,
0.013695945963263512,
-0.08311846107244492,
-0.053361061960458755,
0.020620547235012054,
-0.09845364838838577,
-0.03413843736052513,
0.03831586614251137,
0.0626932680606842,
-0.1280844807624817,
-0.035674162209033966,
0.058769650757312775,
-0.09667537361383438,
0.07505656033754349,
0.02123388648033142,
0.03104463405907154,
0.013873336836695671,
-0.08083989471197128,
0.0697990134358406,
0.01714300736784935,
0.003750697011128068,
0.007592340465635061,
-0.1835612952709198,
0.007667557802051306,
-0.009976981207728386,
0.018827540799975395,
0.05440676957368851,
0.08219723403453827,
-0.11471092700958252,
-0.009603621438145638,
0.0008417848730459809,
-0.049953676760196686,
-0.05774703994393349,
0.03800339251756668,
0.08973168581724167,
0.021625205874443054,
0.20921212434768677,
-0.09245385229587555,
0.037876375019550323,
-0.24632321298122406,
-0.0007629950996488333,
-0.014182105660438538,
-0.07660485804080963,
-0.10783524066209793,
-0.023128865286707878,
0.07757416367530823,
-0.06560466438531876,
0.07541581988334656,
-0.04357060045003891,
0.07553260028362274,
0.03957272693514824,
-0.025439424440264702,
0.08894460648298264,
0.023514477536082268,
0.17235849797725677,
-0.002691194647923112,
-0.03894644230604172,
0.049343645572662354,
-0.011318252421915531,
0.03707021102309227,
0.024935321882367134,
0.12404729425907135,
0.18887153267860413,
-0.0337674506008625,
0.07144997268915176,
0.0408235527575016,
-0.02736525423824787,
-0.0906405970454216,
0.04124727100133896,
-0.05365797504782677,
0.0882602408528328,
0.01735822483897209,
0.20155541598796844,
0.08116131275892258,
-0.19652824103832245,
0.042653705924749374,
-0.08050092309713364,
-0.0839683786034584,
-0.09727589786052704,
-0.05876470357179642,
-0.0979970321059227,
-0.12245617806911469,
0.02715711109340191,
-0.11265004426240921,
0.043637100607156754,
0.06149303540587425,
0.013214829377830029,
-0.03990286588668823,
0.15108217298984528,
0.013151322491466999,
0.028453122824430466,
0.09378764778375626,
0.008905289694666862,
-0.03782489895820618,
-0.031034566462039948,
-0.053290605545043945,
0.044488728046417236,
-0.02027066797018051,
0.037012048065662384,
-0.06517750024795532,
-0.0587083138525486,
0.016681240871548653,
0.006408688146620989,
-0.11117583513259888,
0.026939401403069496,
0.004769978113472462,
0.0695992112159729,
0.06335190683603287,
0.022192610427737236,
0.050961218774318695,
-0.009817378595471382,
0.21111634373664856,
-0.0660657063126564,
-0.04742201790213585,
-0.12227410823106766,
0.22745372354984283,
-0.01709807850420475,
-0.024174071848392487,
0.03158259764313698,
-0.09114830195903778,
0.05213074013590813,
0.16825427114963531,
0.09660457074642181,
-0.06512413173913956,
-0.016057152301073074,
0.008757931180298328,
-0.0138627290725708,
-0.042380254715681076,
0.035304758697748184,
0.08536409586668015,
0.00527574960142374,
-0.0838548094034195,
-0.02745101787149906,
-0.06243760511279106,
-0.033729732036590576,
-0.015758303925395012,
0.05365689471364021,
0.04633162170648575,
0.005990892648696899,
-0.03880661353468895,
0.0850624069571495,
-0.02711315080523491,
-0.09371998906135559,
0.08183591067790985,
-0.19869758188724518,
-0.18387341499328613,
-0.031123200431466103,
0.07552684843540192,
-0.020360277965664864,
0.02879810519516468,
-0.011300504207611084,
-0.004051053896546364,
0.037008725106716156,
-0.038537126034498215,
-0.025722337886691093,
-0.0980151891708374,
0.08443834632635117,
-0.11841992288827896,
0.2183990627527237,
-0.020702337846159935,
0.09641003608703613,
0.14585183560848236,
0.007396472152322531,
-0.10474923253059387,
0.07115792483091354,
0.09818149358034134,
-0.08881498128175735,
0.05339401960372925,
0.13815400004386902,
0.010588412173092365,
0.13532374799251556,
0.06502953171730042,
-0.09431984275579453,
-0.008908970281481743,
-0.06436847150325775,
-0.04174159839749336,
-0.08584201335906982,
-0.00773859117180109,
-0.05842345952987671,
0.15099439024925232,
0.2125573754310608,
-0.06338012218475342,
0.013551193289458752,
-0.046188801527023315,
0.03409909829497337,
0.07128060609102249,
0.06292515993118286,
-0.020001105964183807,
-0.24305026233196259,
0.046627528965473175,
0.04464460164308548,
-0.005709491670131683,
-0.30332037806510925,
-0.10074547678232193,
0.02328074350953102,
-0.06162932515144348,
-0.04355187714099884,
0.1107015460729599,
0.06234908849000931,
0.05716729164123535,
-0.038583122193813324,
-0.06477699428796768,
-0.05645249783992767,
0.1737036257982254,
-0.1392299383878708,
-0.08397458493709564
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# checkpoints
This model is a fine-tuned version of [nielsr/lilt-xlm-roberta-base](https://huggingface.co/nielsr/lilt-xlm-roberta-base) on the xfun dataset.
It achieves the following results on the evaluation set:
- Precision: 0.4372
- Recall: 0.6574
- F1: 0.5252
- Loss: 0.0001
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 8
- eval_batch_size: 2
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- training_steps: 10000
### Training results
| Training Loss | Epoch | Step | F1 | Validation Loss | Precision | Recall |
|:-------------:|:-----:|:-----:|:------:|:---------------:|:---------:|:------:|
| 0.1954 | 20.0 | 500 | 0 | 0.4094 | 0 | 0 |
| 0.1588 | 40.0 | 1000 | 0.1420 | 0.3055 | 0.3587 | 0.0886 |
| 0.1182 | 60.0 | 1500 | 0.4253 | 0.1384 | 0.3810 | 0.4812 |
| 0.0477 | 80.0 | 2000 | 0.4764 | 0.0216 | 0.3949 | 0.6002 |
| 0.069 | 100.0 | 2500 | 0.5198 | 0.0115 | 0.4564 | 0.6038 |
| 0.0355 | 120.0 | 3000 | 0.5161 | 0.0018 | 0.4271 | 0.6521 |
| 0.0268 | 140.0 | 3500 | 0.5254 | 0.0016 | 0.4395 | 0.6530 |
| 0.0123 | 160.0 | 4000 | 0.5264 | 0.0015 | 0.4382 | 0.6592 |
| 0.0039 | 180.0 | 4500 | 0.5353 | 0.0011 | 0.4510 | 0.6583 |
| 0.0139 | 200.0 | 5000 | 0.5390 | 0.0011 | 0.4533 | 0.6646 |
| 0.001 | 220.0 | 5500 | 0.5430 | 0.0042 | 0.4620 | 0.6583 |
| 0.01 | 240.0 | 6000 | 0.5347 | 0.0013 | 0.4531 | 0.6521 |
| 0.0065 | 260.0 | 6500 | 0.5404 | 0.0001 | 0.4540 | 0.6673 |
| 0.0046 | 280.0 | 7000 | 0.5252 | 0.0001 | 0.4372 | 0.6574 |
| 0.002 | 300.0 | 7500 | 0.5365 | 0.0007 | 0.4474 | 0.6699 |
| 0.0002 | 320.0 | 8000 | 0.5393 | 0.0002 | 0.4546 | 0.6628 |
| 0.0008 | 340.0 | 8500 | 0.5412 | 0.0002 | 0.4569 | 0.6637 |
| 0.0024 | 360.0 | 9000 | 0.4677 | 0.6601 | 0.5475 | 0.0002 |
| 0.0001 | 380.0 | 9500 | 0.4560 | 0.6673 | 0.5418 | 0.0002 |
| 0.002 | 400.0 | 10000 | 0.4594 | 0.6628 | 0.5427 | 0.0003 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "mit", "tags": ["generated_from_trainer"], "datasets": ["xfun"], "metrics": ["precision", "recall", "f1"], "base_model": "nielsr/lilt-xlm-roberta-base", "model-index": [{"name": "checkpoints", "results": []}]} | null | kavg/LiLT-RE-JA | [
"transformers",
"safetensors",
"lilt",
"generated_from_trainer",
"dataset:xfun",
"base_model:nielsr/lilt-xlm-roberta-base",
"license:mit",
"endpoints_compatible",
"region:us"
] | 2024-02-07T07:58:19+00:00 | [] | [] | TAGS
#transformers #safetensors #lilt #generated_from_trainer #dataset-xfun #base_model-nielsr/lilt-xlm-roberta-base #license-mit #endpoints_compatible #region-us
| checkpoints
===========
This model is a fine-tuned version of nielsr/lilt-xlm-roberta-base on the xfun dataset.
It achieves the following results on the evaluation set:
* Precision: 0.4372
* Recall: 0.6574
* F1: 0.5252
* Loss: 0.0001
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 1e-05
* train\_batch\_size: 8
* eval\_batch\_size: 2
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* training\_steps: 10000
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 2\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 10000",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #safetensors #lilt #generated_from_trainer #dataset-xfun #base_model-nielsr/lilt-xlm-roberta-base #license-mit #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 2\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 10000",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
62,
115,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #lilt #generated_from_trainer #dataset-xfun #base_model-nielsr/lilt-xlm-roberta-base #license-mit #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 2\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 10000### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.11091124266386032,
0.06850916892290115,
-0.0023296354338526726,
0.10800515860319138,
0.14272484183311462,
0.011064372025430202,
0.11999479681253433,
0.11600357294082642,
-0.08374476432800293,
0.05140921473503113,
0.1319499909877777,
0.12352370470762253,
0.005310493521392345,
0.17374002933502197,
-0.06204897537827492,
-0.22960329055786133,
0.015373075380921364,
0.006802146788686514,
-0.06994785368442535,
0.1230047196149826,
0.0867888554930687,
-0.1460902839899063,
0.0966864749789238,
-0.020790109410881996,
-0.16934843361377716,
0.0007687713368795812,
0.032481078058481216,
-0.03579093888401985,
0.13908790051937103,
0.023331977427005768,
0.12344641983509064,
0.034073356539011,
0.10318440198898315,
-0.21375662088394165,
0.01278806384652853,
0.06142903119325638,
0.009864133782684803,
0.0739709809422493,
0.03500254824757576,
-0.035148732364177704,
0.09906022995710373,
-0.11316895484924316,
0.0581851527094841,
0.028476044535636902,
-0.15802592039108276,
-0.24174970388412476,
-0.09425810724496841,
0.039056044071912766,
0.08596540987491608,
0.07453063875436783,
-0.010584359057247639,
0.12004561722278595,
-0.06370092183351517,
0.10249217599630356,
0.2677188515663147,
-0.3046685755252838,
-0.07186194509267807,
0.05042284354567528,
0.054702479392290115,
0.09059491753578186,
-0.12329548597335815,
-0.007770544849336147,
0.05281096324324608,
0.016254445537924767,
0.15275873243808746,
-0.04308512806892395,
-0.06092570722103119,
0.021744994446635246,
-0.1405576765537262,
-0.018733074888586998,
0.10363320261240005,
0.05529263988137245,
-0.054680969566106796,
-0.049829140305519104,
-0.05659462884068489,
-0.15665751695632935,
-0.0643305853009224,
-0.021414507180452347,
0.06369718164205551,
-0.04082586243748665,
-0.08936590701341629,
-0.016999492421746254,
-0.09081930667161942,
-0.093085877597332,
-0.022679636254906654,
0.16877000033855438,
0.03328623250126839,
0.0021022262517362833,
-0.009198781102895737,
0.10767022520303726,
-0.03702942654490471,
-0.1335057020187378,
-0.010934073477983475,
0.017424529418349266,
-0.029935745522379875,
-0.05833810195326805,
-0.045755334198474884,
-0.006934540346264839,
0.033086031675338745,
0.14521443843841553,
-0.09829624742269516,
0.041375335305929184,
0.02125525288283825,
0.025963010266423225,
-0.10578275471925735,
0.1568889617919922,
-0.06652272492647171,
-0.06635528057813644,
0.013513315469026566,
0.08928536623716354,
0.02697037160396576,
0.009943069890141487,
-0.07615203410387039,
0.009742991998791695,
0.09059244394302368,
0.03763379529118538,
-0.07143866270780563,
0.07154018431901932,
-0.0527162030339241,
0.01908288337290287,
0.017588946968317032,
-0.09027329087257385,
0.030867349356412888,
0.008157407864928246,
-0.07209328562021255,
-0.07642420381307602,
0.021247588098049164,
0.00869134720414877,
0.025442907586693764,
0.08341223001480103,
-0.10220518708229065,
0.027172334492206573,
-0.08116448670625687,
-0.12765704095363617,
-0.005736074410378933,
-0.08019836992025375,
0.027707679197192192,
-0.087730772793293,
-0.17963066697120667,
-0.021919801831245422,
0.048756759613752365,
-0.040677811950445175,
0.003202010178938508,
-0.06653521955013275,
-0.09869056195020676,
0.0018031122162938118,
-0.0050604394637048244,
0.08497462421655655,
-0.07262725383043289,
0.11364074051380157,
0.047836821526288986,
0.07551374286413193,
-0.020294032990932465,
0.02794768661260605,
-0.11238288134336472,
0.042713675647974014,
-0.22012026607990265,
0.03718909993767738,
-0.062052614986896515,
0.07296259701251984,
-0.09703761339187622,
-0.08751396089792252,
0.011758577078580856,
0.009951705113053322,
0.08341406285762787,
0.11921944469213486,
-0.16762135922908783,
-0.052162352949380875,
0.1958739012479782,
-0.09811758995056152,
-0.1481909155845642,
0.09869173169136047,
-0.06102165952324867,
0.06713885068893433,
0.07292870432138443,
0.2082444429397583,
0.05491584911942482,
-0.12475605309009552,
0.007668833713978529,
-0.014386159367859364,
0.03718801215291023,
-0.023767074570059776,
0.059020623564720154,
0.01157907024025917,
0.022003451362252235,
0.009328681975603104,
-0.07778338342905045,
0.04914334788918495,
-0.10958799719810486,
-0.07678412646055222,
-0.04347272589802742,
-0.11570663750171661,
0.04246535152196884,
0.04550208896398544,
0.03925004228949547,
-0.12523147463798523,
-0.08311811834573746,
0.06932725012302399,
0.09164676070213318,
-0.07352900505065918,
0.02278062142431736,
-0.07725726813077927,
0.05540642514824867,
-0.04668129235506058,
-0.029189351946115494,
-0.1546349823474884,
-0.05810775235295296,
0.0030764031689614058,
0.014849625527858734,
-0.0026265447959303856,
-0.01061892881989479,
0.08681493252515793,
0.08022424578666687,
-0.06850523501634598,
-0.04613653942942619,
-0.023638876155018806,
0.016982505097985268,
-0.11806486546993256,
-0.22240673005580902,
-0.03301287442445755,
-0.03528660908341408,
0.10686822235584259,
-0.25111857056617737,
0.04117092490196228,
-0.00891097728163004,
0.06245404854416847,
0.05062171444296837,
-0.02416347712278366,
-0.022425439208745956,
0.059496164321899414,
-0.023853842169046402,
-0.07488331943750381,
0.053187523037195206,
0.00972640048712492,
-0.07965711504220963,
-0.02637619711458683,
-0.14529062807559967,
0.18064957857131958,
0.11554080992937088,
-0.04154510796070099,
-0.11154916882514954,
-0.00012794247595593333,
-0.05562116578221321,
-0.026308074593544006,
-0.04023601487278938,
0.022303447127342224,
0.10242489725351334,
0.001478404039517045,
0.13885024189949036,
-0.08329793810844421,
-0.03599062189459801,
0.030293019488453865,
-0.0467233881354332,
0.03573571890592575,
0.11898762732744217,
0.04914197698235512,
-0.09189406037330627,
0.13113315403461456,
0.14000461995601654,
-0.04890182986855507,
0.11286744475364685,
-0.04231764376163483,
-0.0653921589255333,
-0.04693159833550453,
0.023133035749197006,
0.005593341309577227,
0.15091191232204437,
-0.07775536924600601,
0.004200808238238096,
-0.0015606743982061744,
0.02192915417253971,
-0.003378194523975253,
-0.22143691778182983,
-0.05354302003979683,
0.039273619651794434,
-0.06795719265937805,
-0.036073558032512665,
-0.024258874356746674,
-0.0075683933682739735,
0.10168924927711487,
0.00023499561939388514,
-0.07164608687162399,
0.010907106101512909,
-0.0012300295056775212,
-0.0738489106297493,
0.2177780121564865,
-0.07445799559354782,
-0.09638175368309021,
-0.08083651959896088,
-0.05211509019136429,
-0.017906326800584793,
0.01610356755554676,
0.06119875982403755,
-0.06633789092302322,
-0.02371036261320114,
-0.10752784460783005,
0.013096682727336884,
0.028802156448364258,
0.02303999289870262,
-0.012512017972767353,
0.0020830919966101646,
0.07581914216279984,
-0.11352528631687164,
-0.009964056313037872,
-0.053690388798713684,
-0.0669960007071495,
0.040065694600343704,
0.04423283785581589,
0.1251552700996399,
0.12454217672348022,
-0.019242266193032265,
0.003404970047995448,
-0.043230652809143066,
0.24240998923778534,
-0.07239282876253128,
-0.006384699605405331,
0.11850845068693161,
0.010298039764165878,
0.04602331295609474,
0.14645519852638245,
0.05800940468907356,
-0.11964225769042969,
0.01431862823665142,
0.042257752269506454,
-0.030576908960938454,
-0.19412356615066528,
-0.022555876523256302,
-0.030526824295520782,
-0.007526764180511236,
0.07629738003015518,
0.017038092017173767,
0.0015076948329806328,
0.0571511909365654,
0.028436338528990746,
0.021465551108121872,
-0.006798330694437027,
0.06847777962684631,
0.060992997139692307,
0.038663703948259354,
0.11813915520906448,
-0.04551559314131737,
-0.04283569008111954,
0.023708214983344078,
-0.003971155267208815,
0.2342909574508667,
-0.006545636337250471,
0.1257757991552353,
0.0688256323337555,
0.16258245706558228,
-0.019153783097863197,
0.059195443987846375,
0.002616442972794175,
-0.06465139240026474,
-0.0063292705453932285,
-0.05192175135016441,
-0.013916390016674995,
0.026734324172139168,
-0.10149960964918137,
0.0732610821723938,
-0.1302664428949356,
0.012223463505506516,
0.0716419667005539,
0.24503442645072937,
0.054659247398376465,
-0.32780593633651733,
-0.09907335788011551,
0.009527526795864105,
-0.018866203725337982,
-0.01864253729581833,
0.019580869004130363,
0.1511472910642624,
-0.062108755111694336,
0.028304405510425568,
-0.07178286463022232,
0.07084602117538452,
-0.008352368138730526,
0.03899955376982689,
0.05333457887172699,
0.1042919009923935,
-0.02985091507434845,
0.04274534434080124,
-0.2666052281856537,
0.30987706780433655,
0.022239062935113907,
0.10302716493606567,
-0.023875568062067032,
-0.01902974769473076,
0.021206725388765335,
0.09492574632167816,
0.11802197247743607,
-0.017141161486506462,
-0.0737774595618248,
-0.23103611171245575,
-0.039019398391246796,
0.03469011187553406,
0.11670517921447754,
-0.031853921711444855,
0.134010910987854,
-0.03454722464084625,
-0.005672285333275795,
0.0813533291220665,
-0.03099741041660309,
-0.08723514527082443,
-0.06105084344744682,
-0.04074858874082565,
0.013626229017972946,
0.0041402303613722324,
-0.08874379843473434,
-0.07867864519357681,
-0.08404526859521866,
0.1254332810640335,
-0.011820406652987003,
-0.025387711822986603,
-0.11140338331460953,
0.06387585401535034,
0.08925723284482956,
-0.08231573551893234,
0.044480036944150925,
0.024361751973628998,
0.06245657429099083,
0.033117156475782394,
-0.029114220291376114,
0.1271509826183319,
-0.07327834516763687,
-0.1848422884941101,
-0.060206089168787,
0.11838560551404953,
0.04251857474446297,
0.04923850670456886,
0.0018504648469388485,
0.02671067975461483,
-0.007312577683478594,
-0.0852898582816124,
0.032376911491155624,
-0.022432532161474228,
0.05776320397853851,
0.009117837063968182,
-0.05700254067778587,
0.036981262266635895,
-0.05652361735701561,
-0.0214137751609087,
0.14302676916122437,
0.3156736195087433,
-0.10344129800796509,
0.02158519998192787,
0.046824876219034195,
-0.062033817172050476,
-0.19471846520900726,
0.06967631727457047,
0.026514045894145966,
0.00040356547106057405,
0.07483882457017899,
-0.14361993968486786,
0.09642304480075836,
0.10722275078296661,
-0.01419061329215765,
0.1014149859547615,
-0.3219514489173889,
-0.1235002800822258,
0.09224962443113327,
0.1754816621541977,
0.10446137934923172,
-0.16959543526172638,
-0.016983406618237495,
0.0009282589890062809,
-0.11316186934709549,
0.09763661026954651,
-0.11949506402015686,
0.11383944004774094,
-0.012983286753296852,
0.03464779630303383,
0.019330855458974838,
-0.057226937264204025,
0.1135164424777031,
0.005009889602661133,
0.14305590093135834,
-0.04248882085084915,
-0.03632392734289169,
0.07314363121986389,
-0.05956840515136719,
0.022949062287807465,
-0.0789095088839531,
0.04445973038673401,
-0.04589514806866646,
-0.020229456946253777,
-0.08348223567008972,
0.027147524058818817,
-0.04022776335477829,
-0.07086179405450821,
-0.04059796407818794,
0.05440054461359978,
0.05900893732905388,
-0.01584290713071823,
0.13783539831638336,
-0.012260206043720245,
0.15262193977832794,
0.09782446920871735,
0.05759386718273163,
-0.05654022842645645,
-0.034806594252586365,
0.023379120975732803,
-0.020582519471645355,
0.03882145136594772,
-0.16576141119003296,
0.043132081627845764,
0.11988434195518494,
0.03596678376197815,
0.13395239412784576,
0.07194579392671585,
-0.02102469466626644,
0.01520474161952734,
0.06400109827518463,
-0.14342467486858368,
-0.12624497711658478,
0.00814066268503666,
-0.0903334692120552,
-0.11942949891090393,
0.06986462324857712,
0.10860977321863174,
-0.07638675719499588,
0.004487472586333752,
-0.02903486229479313,
0.025130731984972954,
-0.04983561858534813,
0.20163525640964508,
0.07547298818826675,
0.0571577213704586,
-0.089650958776474,
0.06666026264429092,
0.020715901628136635,
-0.062160324305295944,
0.004550114274024963,
0.054138753563165665,
-0.08465347439050674,
-0.035575270652770996,
0.06988076120615005,
0.16884851455688477,
-0.04224062338471413,
-0.026528891175985336,
-0.14943702518939972,
-0.11570864170789719,
0.04756505787372589,
0.21687723696231842,
0.0830589309334755,
0.02633271925151348,
-0.00624612532556057,
0.02065572515130043,
-0.13191311061382294,
0.11520599573850632,
0.03697497025132179,
0.07304559648036957,
-0.1452060341835022,
0.17853859066963196,
-0.012712516821920872,
0.014980608597397804,
-0.023436594754457474,
0.046705909073352814,
-0.13603228330612183,
0.005590626038610935,
-0.13128557801246643,
-0.012086641043424606,
-0.05210689827799797,
0.003108836244791746,
-0.005692424718290567,
-0.06625987589359283,
-0.06552331894636154,
0.014348631724715233,
-0.10437563806772232,
-0.01909351348876953,
0.04119280353188515,
0.05380340293049812,
-0.1389332115650177,
-0.042768388986587524,
0.011093330569565296,
-0.06824567168951035,
0.06270929425954819,
0.024700378999114037,
0.02396281063556671,
0.04878449812531471,
-0.18049690127372742,
0.02370947226881981,
0.04813120886683464,
-0.023299463093280792,
0.053398340940475464,
-0.10104575008153915,
-0.013609867542982101,
-0.023266853764653206,
0.043856143951416016,
0.013830083422362804,
0.06081075593829155,
-0.12968909740447998,
0.02150770090520382,
-0.03514576330780983,
-0.04124249890446663,
-0.05370217189192772,
0.017874056473374367,
0.08563162386417389,
0.0030164250638335943,
0.17773477733135223,
-0.09509998559951782,
0.021770210936665535,
-0.22586047649383545,
-0.016165388748049736,
-0.008760582655668259,
-0.10347782075405121,
-0.12252427637577057,
-0.05368814617395401,
0.07204549014568329,
-0.051973626017570496,
0.129298135638237,
-0.002859393833205104,
0.05635492131114006,
0.03489360958337784,
-0.06962666660547256,
0.03651101142168045,
0.03368544206023216,
0.22092430293560028,
0.013549631461501122,
-0.050853047519922256,
0.055471375584602356,
0.053667761385440826,
0.10074598342180252,
0.07259220629930496,
0.21061427891254425,
0.18641120195388794,
-0.010846089571714401,
0.09340488165616989,
0.033669691532850266,
-0.07923606038093567,
-0.11977332830429077,
0.06909535825252533,
-0.03615177422761917,
0.0641869306564331,
-0.03442786633968353,
0.20419538021087646,
0.11362158507108688,
-0.1780390739440918,
0.03647603467106819,
-0.0471116378903389,
-0.08080469816923141,
-0.10343686491250992,
-0.020700745284557343,
-0.0941767543554306,
-0.16349032521247864,
-0.0005123004084452987,
-0.12399931997060776,
0.02776554971933365,
0.09890871495008469,
-0.0015642280923202634,
0.005073768086731434,
0.18351954221725464,
0.031562186777591705,
0.041505713015794754,
0.041980061680078506,
0.011403091251850128,
-0.023923877626657486,
-0.06067851558327675,
-0.08015183359384537,
0.02409016713500023,
-0.05599157512187958,
0.019401762634515762,
-0.04912063851952553,
-0.047234900295734406,
0.04698605090379715,
-0.023521030321717262,
-0.10518106818199158,
0.014694031327962875,
0.040842242538928986,
0.06607324630022049,
0.014196186326444149,
0.0318329744040966,
-0.0076826103031635284,
-0.003393491730093956,
0.26899078488349915,
-0.07213089615106583,
-0.07455974072217941,
-0.09768342226743698,
0.25214073061943054,
0.053326044231653214,
0.01240447722375393,
0.007675265893340111,
-0.0868423581123352,
0.04001669958233833,
0.2137032151222229,
0.1821684092283249,
-0.08408399671316147,
0.014413169585168362,
-0.028126904740929604,
-0.01117358636111021,
-0.032349612563848495,
0.09455512464046478,
0.10109496861696243,
0.02701365016400814,
-0.08404107391834259,
-0.04890086129307747,
-0.047179121524095535,
-0.017935696989297867,
-0.03387090936303139,
0.06030714884400368,
0.03816886246204376,
0.02126738615334034,
-0.046141646802425385,
0.07028932124376297,
-0.038727909326553345,
-0.13361768424510956,
0.08370870351791382,
-0.2039404958486557,
-0.13347503542900085,
-0.03871798887848854,
0.07467317581176758,
0.013289941474795341,
0.06547971814870834,
-0.02781389094889164,
-0.004902031738311052,
0.07311218231916428,
-0.01610635221004486,
-0.05961156636476517,
-0.10586553066968918,
0.06439834833145142,
-0.0809924528002739,
0.22872203588485718,
-0.04332577437162399,
0.010865014046430588,
0.12958933413028717,
0.027939992025494576,
-0.10012894868850708,
0.10289467126131058,
0.050488896667957306,
-0.08472783118486404,
0.02638634853065014,
0.08639806509017944,
-0.04745357111096382,
0.11805219203233719,
0.04820738360285759,
-0.14429689943790436,
0.030389847233891487,
-0.057615045458078384,
-0.08599929511547089,
-0.05454399064183235,
-0.06077701970934868,
-0.03976212441921234,
0.13439710438251495,
0.1946522444486618,
-0.047352105379104614,
0.034707676619291306,
-0.050027016550302505,
0.03713715076446533,
0.0870540514588356,
0.07665467262268066,
-0.034217435866594315,
-0.24617260694503784,
0.03824254497885704,
0.09716778248548508,
-0.017077138647437096,
-0.24484048783779144,
-0.08935652673244476,
0.015055338852107525,
-0.04521496221423149,
-0.07996855676174164,
0.09437962621450424,
0.1139470562338829,
0.06212584674358368,
-0.052593205124139786,
-0.11193755269050598,
-0.06150442734360695,
0.19295255839824677,
-0.12240894883871078,
-0.09847009927034378
] |
null | null | transformers | Finetune of miqu-70b-sf dequant of miqudev's leak of Mistral-70B (allegedly an early mistral medium). My diffs are available under CC-0, this is a merge with the leaked model, you can use the other repository to save bandwidth.
EQ-Bench: 84.89
Will run more benches later. | {"license": "cc-by-2.0"} | text-generation | JackCloudman/Senku-70B-Full-exl2-3.5bpw | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"license:cc-by-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T08:04:36+00:00 | [] | [] | TAGS
#transformers #safetensors #llama #text-generation #conversational #license-cc-by-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| Finetune of miqu-70b-sf dequant of miqudev's leak of Mistral-70B (allegedly an early mistral medium). My diffs are available under CC-0, this is a merge with the leaked model, you can use the other repository to save bandwidth.
EQ-Bench: 84.89
Will run more benches later. | [] | [
"TAGS\n#transformers #safetensors #llama #text-generation #conversational #license-cc-by-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
60
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #conversational #license-cc-by-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
-0.006207775790244341,
0.05186443030834198,
-0.005766681861132383,
0.0034429291263222694,
0.09202545881271362,
-0.026229875162243843,
0.21461424231529236,
0.09413789957761765,
-0.012125303968787193,
-0.027240416035056114,
0.15366454422473907,
0.20107468962669373,
-0.030660197138786316,
0.0625392273068428,
-0.12453678250312805,
-0.15658724308013916,
0.07951316237449646,
-0.0074092973954975605,
0.0468166321516037,
0.08121465146541595,
0.11726208031177521,
-0.05624562129378319,
0.07327274978160858,
-0.05116863548755646,
-0.10777227580547333,
0.01719590090215206,
0.07997997105121613,
-0.1362335979938507,
0.10389529168605804,
0.070436492562294,
0.09596491605043411,
0.09478230774402618,
-0.026663783937692642,
-0.2320748120546341,
0.024640627205371857,
-0.013039902783930302,
-0.08545025438070297,
0.02922174707055092,
0.03788268566131592,
-0.042137615382671356,
0.05930780619382858,
0.04187946766614914,
-0.012704534456133842,
0.0843532532453537,
-0.12006022036075592,
0.03492521494626999,
-0.054476991295814514,
0.008457338437438011,
0.11366608738899231,
0.07924187183380127,
0.002242166781798005,
0.1153382733464241,
-0.05578871816396713,
0.08845382928848267,
0.06245912238955498,
-0.3711334466934204,
0.014689184725284576,
0.13522937893867493,
0.07748636603355408,
0.06413708627223969,
-0.05252519249916077,
0.11249673366546631,
0.06629528105258942,
-0.03810290992259979,
0.04312480613589287,
-0.07572442293167114,
-0.08102772384881973,
0.034424908459186554,
-0.04833368957042694,
-0.030284471809864044,
0.23118719458580017,
-0.03052249550819397,
0.012902313843369484,
-0.08178368955850601,
-0.06274984031915665,
0.003528473898768425,
-0.027013633400201797,
0.033650096505880356,
-0.007114062085747719,
0.09828754514455795,
0.0036289477720856667,
-0.04431737959384918,
-0.15287499129772186,
-0.017103472724556923,
-0.17628958821296692,
0.10724018514156342,
-0.0038528726436197758,
0.03643316775560379,
-0.1433914750814438,
0.03731580823659897,
0.01849939487874508,
-0.09380249679088593,
-0.01922903209924698,
-0.06579022109508514,
0.07399444282054901,
-0.015354927629232407,
-0.04418032616376877,
-0.04420563578605652,
0.146459698677063,
0.11814041435718536,
-0.01364491693675518,
0.003440734464675188,
-0.11379846185445786,
0.0978090688586235,
-0.019500313326716423,
-0.0008985437452793121,
0.047480762004852295,
-0.007553805597126484,
0.1015886589884758,
-0.08041958510875702,
0.08773300796747208,
-0.04154345020651817,
-0.16899511218070984,
0.014984861016273499,
-0.009690387174487114,
0.14272445440292358,
0.0067246644757688046,
0.08934520930051804,
-0.03848949074745178,
0.06400565803050995,
0.07495175302028656,
-0.06882821768522263,
0.002561564091593027,
0.008464166894555092,
0.05782800912857056,
0.020744536072015762,
0.0357230082154274,
0.05749142915010452,
-0.053222328424453735,
0.034963082522153854,
-0.05974286422133446,
-0.030262604355812073,
-0.04071192815899849,
-0.05133391171693802,
0.07166485488414764,
-0.026888888329267502,
0.023412950336933136,
-0.1889147162437439,
-0.17520999908447266,
0.017266716808080673,
0.006569101009517908,
-0.016733024269342422,
-0.0409536249935627,
-0.05526142567396164,
-0.029290098696947098,
0.025005020201206207,
-0.08925633877515793,
-0.08324684202671051,
-0.08618094772100449,
0.08566243946552277,
-0.034463070333004,
0.052008256316185,
-0.17173472046852112,
0.030293233692646027,
-0.09709536284208298,
0.015842683613300323,
-0.024641036987304688,
0.05080806463956833,
-0.03994987532496452,
0.1493525505065918,
-0.03130904585123062,
0.028674336150288582,
-0.049635149538517,
0.06748075783252716,
-0.04132431373000145,
0.20219798386096954,
-0.13369248807430267,
-0.02732842043042183,
0.23248150944709778,
-0.1203412264585495,
-0.2215833067893982,
0.09494956582784653,
-0.007244576700031757,
0.06883425265550613,
0.1192723736166954,
0.16428819298744202,
-0.018312968313694,
-0.084824338555336,
0.043344683945178986,
0.10571722686290741,
-0.05737517401576042,
-0.10720136761665344,
0.012378843501210213,
-0.02602921798825264,
-0.10123564302921295,
0.021436169743537903,
0.054138898849487305,
0.03870760649442673,
-0.010303346440196037,
-0.06431794166564941,
-0.04900350421667099,
-0.04556714743375778,
-0.019114533439278603,
-0.05745845288038254,
0.04891825467348099,
-0.10103435069322586,
-0.004564722999930382,
0.02500876598060131,
-0.0006331629119813442,
-0.017695071175694466,
0.03802654147148132,
-0.1019439622759819,
0.05491343140602112,
-0.01049356535077095,
0.05139922350645065,
-0.08506914973258972,
-0.09540282934904099,
-0.01487639732658863,
0.11419088393449783,
0.04237206280231476,
0.03597130998969078,
0.03009572997689247,
-0.0029274821281433105,
-0.016321944072842598,
0.019669566303491592,
0.1798858940601349,
0.027921929955482483,
-0.06157265976071358,
-0.10305018723011017,
0.10064617544412613,
-0.04469102993607521,
0.06838524341583252,
-0.12369509041309357,
0.021130846813321114,
0.09049873054027557,
0.08164099603891373,
0.011007205583155155,
0.06817352026700974,
-0.016652852296829224,
0.012067300267517567,
-0.08803333342075348,
0.018720898777246475,
0.09845291078090668,
0.023675929754972458,
-0.12399162352085114,
0.24828600883483887,
-0.22121906280517578,
0.23131431639194489,
0.20286236703395844,
-0.2072221040725708,
0.034866541624069214,
-0.11290061473846436,
0.008944046683609486,
0.009122604504227638,
0.00658374885097146,
-0.04158162698149681,
0.02772163227200508,
-0.017428645864129066,
0.17939260601997375,
-0.07259617745876312,
-0.01289988774806261,
-0.015334650874137878,
-0.07770753651857376,
-0.0461907684803009,
0.04808717593550682,
0.10923753678798676,
-0.17211377620697021,
0.18076997995376587,
0.2573873698711395,
0.016237052157521248,
0.145608589053154,
-0.02829170413315296,
0.011588740162551403,
0.031093206256628036,
0.051330193877220154,
0.017952319234609604,
-0.03206970542669296,
-0.09656007587909698,
-0.01615116000175476,
0.06039080768823624,
0.00738911610096693,
0.05942045524716377,
-0.14498206973075867,
-0.06179048866033554,
-0.010635633021593094,
-0.054405391216278076,
0.006760948803275824,
0.05542760342359543,
-0.006918448954820633,
0.12766680121421814,
-0.04561823606491089,
-0.06787645071744919,
0.1212611123919487,
-0.018814176321029663,
-0.11172524094581604,
0.16348372399806976,
-0.1353582739830017,
-0.2530854046344757,
-0.1794736236333847,
-0.16009610891342163,
-0.05283679813146591,
0.06249964237213135,
0.12733663618564606,
-0.021347085013985634,
-0.06735935807228088,
-0.08513124287128448,
-0.025516794994473457,
-0.004286123439669609,
-0.0013990295119583607,
-0.030478790402412415,
0.0740688145160675,
-0.04530498385429382,
-0.11495137214660645,
-0.04566733539104462,
0.02995317615568638,
-0.08298885077238083,
0.11527245491743088,
-0.08257944881916046,
0.09218067675828934,
0.14615941047668457,
0.013489138334989548,
-0.0007384112104773521,
-0.06414268910884857,
0.11027345061302185,
-0.05689837411046028,
-0.021911391988396645,
0.19550743699073792,
-0.06030816212296486,
0.054297950118780136,
0.17741087079048157,
0.019415782764554024,
-0.11608056724071503,
0.05627991259098053,
-0.04948503151535988,
-0.08943437039852142,
-0.2322319746017456,
-0.11231391131877899,
-0.09151040762662888,
0.09372707456350327,
0.02010270394384861,
0.06644773483276367,
0.1561785638332367,
0.07071852684020996,
-0.03820497542619705,
-0.003707042895257473,
0.09500505030155182,
0.10332844406366348,
0.26925957202911377,
-0.03648235276341438,
0.1326010376214981,
-0.09864376485347748,
-0.09166394174098969,
0.08177606761455536,
0.09666159003973007,
0.08463814854621887,
0.1308884173631668,
0.10701965540647507,
0.06300177425146103,
0.06616055965423584,
0.13583797216415405,
0.0930822342634201,
0.04747668653726578,
-0.027060847729444504,
-0.021886298432946205,
-0.06314191222190857,
-0.03154398128390312,
0.06433441489934921,
-0.0616009458899498,
-0.1419346034526825,
-0.01669052243232727,
-0.04656673222780228,
0.08677151799201965,
0.12965995073318481,
0.049826931208372116,
-0.20373210310935974,
0.03404708579182625,
0.12731851637363434,
-0.014891417697072029,
-0.0784522220492363,
0.11266084015369415,
0.041655853390693665,
-0.0492335706949234,
0.0968889519572258,
-0.022453399375081062,
0.10560388118028641,
-0.04346970468759537,
0.06621389091014862,
-0.09852857887744904,
-0.09218709915876389,
0.008176117204129696,
0.10175062716007233,
-0.32708150148391724,
0.19553345441818237,
0.025147762149572372,
0.003030276857316494,
-0.07397007197141647,
-0.00860549882054329,
0.006451844237744808,
0.1687123030424118,
0.14335986971855164,
-0.03559787943959236,
-0.11925935745239258,
-0.036741066724061966,
-0.04582417383790016,
0.021652963012456894,
0.114947110414505,
0.004123690538108349,
0.006702055223286152,
-0.060284681618213654,
-0.005426459014415741,
0.013130738399922848,
-0.03866162151098251,
-0.05004815012216568,
-0.18432049453258514,
0.036161795258522034,
0.15424391627311707,
0.105919748544693,
-0.038923926651477814,
0.025927383452653885,
-0.1276562660932541,
0.19439037144184113,
-0.1566786915063858,
-0.05010409653186798,
-0.10357701778411865,
-0.13866664469242096,
-0.010155638679862022,
-0.015465851873159409,
0.06662318855524063,
-0.05540037900209427,
0.05461542308330536,
-0.0985378548502922,
-0.16780197620391846,
0.11566402018070221,
-0.10529130697250366,
-0.03816051036119461,
-0.04050680622458458,
0.15799902379512787,
-0.10722509026527405,
-0.015996212139725685,
0.0695413127541542,
0.027312718331813812,
-0.04470572620630264,
-0.10069151222705841,
-0.01407475396990776,
0.022582462057471275,
0.03730671480298042,
-0.004337325692176819,
-0.1392161101102829,
-0.12033824622631073,
-0.007708914577960968,
-0.08476556092500687,
0.26667433977127075,
0.2823437750339508,
-0.052704282104969025,
0.1610942929983139,
0.18729859590530396,
-0.10833743214607239,
-0.3613292872905731,
-0.1156543493270874,
-0.19261738657951355,
-0.05812794342637062,
-0.0019929534755647182,
-0.09389892220497131,
0.0683603510260582,
0.03696022555232048,
-0.06046358868479729,
0.10693113505840302,
-0.20837023854255676,
-0.11674803495407104,
0.13628779351711273,
0.022366341203451157,
0.3236163854598999,
-0.1799478828907013,
-0.10869350284337997,
-0.1288607120513916,
-0.09091421216726303,
0.17097102105617523,
-0.11236412078142166,
0.09014905989170074,
0.038183145225048065,
0.05064551532268524,
0.019322404637932777,
-0.028820998966693878,
0.11060543358325958,
-0.05023352429270744,
0.05936712771654129,
-0.125066876411438,
-0.007887703366577625,
0.055251557379961014,
-0.021825825795531273,
0.047600388526916504,
-0.17949266731739044,
0.010565537959337234,
-0.02346070110797882,
-0.03863802179694176,
-0.00010827439837157726,
0.07584916800260544,
-0.0018114936538040638,
-0.05403559282422066,
-0.0338921919465065,
-0.07518398016691208,
0.021392593160271645,
-0.005685644224286079,
0.2621508240699768,
-0.09497393667697906,
0.143461674451828,
0.2097931206226349,
0.18348653614521027,
-0.11326921731233597,
0.10850489139556885,
-0.03179696202278137,
-0.10264644771814346,
0.06764788925647736,
-0.13615255057811737,
0.06894791126251221,
0.07885490357875824,
-0.05221731215715408,
0.09386573731899261,
0.07263122498989105,
0.024896983057260513,
0.011904848739504814,
0.15288308262825012,
-0.19621852040290833,
-0.06335753947496414,
-0.02627270668745041,
0.128641277551651,
0.07603052258491516,
0.07438892126083374,
0.19964289665222168,
-0.02363588474690914,
0.024383772164583206,
0.004766607191413641,
0.0470099002122879,
-0.051470812410116196,
0.03084268979728222,
-0.0036850213073194027,
0.009037865325808525,
-0.11732205748558044,
0.11256258189678192,
0.014299282804131508,
-0.13298094272613525,
0.018583929166197777,
0.10019229352474213,
-0.12479248642921448,
-0.1322273164987564,
-0.06397988647222519,
0.09235106408596039,
-0.1650686413049698,
-0.09092073142528534,
-0.035492803901433945,
-0.18309658765792847,
0.023424223065376282,
0.22705577313899994,
0.03624091297388077,
0.09397074580192566,
0.025309177115559578,
-0.050275810062885284,
-0.03333280235528946,
0.044489964842796326,
-0.0776023343205452,
0.026683444157242775,
-0.09567009657621384,
0.016401100903749466,
-0.0327458456158638,
0.04273902624845505,
-0.08503364771604538,
-0.01675247587263584,
-0.13592562079429626,
0.026447130367159843,
-0.1429377794265747,
0.016726532950997353,
-0.09382633119821548,
-0.022313788533210754,
0.025396376848220825,
-0.011300059035420418,
-0.044356830418109894,
-0.0456419438123703,
-0.08871109038591385,
0.022159568965435028,
-0.04314287006855011,
0.0626419335603714,
-0.09819427877664566,
-0.043954771012067795,
0.04525887221097946,
-0.036994993686676025,
0.11698898673057556,
0.04370922967791557,
-0.0972345620393753,
0.07480837404727936,
-0.24783827364444733,
-0.02655193954706192,
0.13125064969062805,
0.015579255297780037,
0.014812005683779716,
0.06315375864505768,
-0.011836273595690727,
0.1417415887117386,
0.009809926152229309,
0.052264366298913956,
0.013304419815540314,
-0.1011492908000946,
-0.001961590489372611,
-0.04560624808073044,
-0.10192413628101349,
-0.018009837716817856,
-0.0727795958518982,
0.1012018471956253,
-0.024938881397247314,
0.17435556650161743,
-0.092584989964962,
0.03382148593664169,
-0.01980471797287464,
0.03883249685168266,
0.005250109359622002,
-0.15618106722831726,
-0.14053024351596832,
-0.07882187515497208,
-0.001336804125458002,
-0.015131816267967224,
0.27824896574020386,
0.008481241762638092,
-0.06852035224437714,
0.09475342184305191,
0.022017354145646095,
0.034207794815301895,
0.04192714765667915,
0.2883480489253998,
0.08536436408758163,
-0.01155257411301136,
-0.15340177714824677,
0.016927950084209442,
0.0375761017203331,
-0.09943119436502457,
0.06783804297447205,
0.09215667098760605,
-0.07752043753862381,
0.10896968841552734,
0.07114557921886444,
0.0072129033505916595,
-0.018362488597631454,
-0.06760571897029877,
-0.055142998695373535,
0.046072885394096375,
-0.049081407487392426,
0.07889051735401154,
0.20583051443099976,
-0.03462229669094086,
-0.015487121418118477,
-0.04577798396348953,
-0.03612710162997246,
-0.18539372086524963,
-0.13403363525867462,
-0.1109817773103714,
-0.11150851845741272,
0.01701277866959572,
-0.08830517530441284,
0.055540017783641815,
0.05017261207103729,
0.054319605231285095,
-0.04291510581970215,
0.06990589201450348,
-0.05907499045133591,
-0.04540880024433136,
0.0442875400185585,
-0.02594553679227829,
0.06306269764900208,
-0.035188764333724976,
-0.06933852285146713,
-0.053066983819007874,
-0.04323925822973251,
-0.02944331243634224,
0.08734598755836487,
0.050012290477752686,
0.06000626087188721,
-0.13524575531482697,
-0.06855671852827072,
-0.032354552298784256,
0.0697886273264885,
-0.008020002394914627,
0.1599365472793579,
0.022465955466032028,
-0.0408162996172905,
0.08315412700176239,
0.1788318008184433,
-0.06562280654907227,
-0.12326370179653168,
-0.031882673501968384,
0.1764252483844757,
0.0044775111600756645,
0.13507580757141113,
-0.040701255202293396,
-0.013290504924952984,
-0.012223204597830772,
0.3287609815597534,
0.25894030928611755,
-0.09469550848007202,
0.03137436881661415,
-0.07478570938110352,
0.04246385395526886,
0.05169496685266495,
0.11560044437646866,
0.0689227506518364,
0.25144103169441223,
-0.03857225552201271,
-0.03432202711701393,
-0.00518826674669981,
0.017930803820490837,
-0.13905346393585205,
0.0863647311925888,
-0.04497753456234932,
-0.05007867515087128,
-0.02601996622979641,
0.09876758605241776,
-0.15639275312423706,
0.06847500801086426,
-0.04330084100365639,
-0.09984272718429565,
0.007430274039506912,
0.007269448135048151,
0.14150775969028473,
-0.019860094413161278,
0.029579713940620422,
-0.03768281638622284,
-0.07121695578098297,
0.01726885512471199,
-0.005519684869796038,
-0.18614327907562256,
0.03858964890241623,
0.023371171206235886,
-0.005911254324018955,
0.07187901437282562,
0.000053612515330314636,
0.06824749708175659,
0.07142221182584763,
0.030734332278370857,
-0.051403749734163284,
0.15974754095077515,
0.026775244623422623,
-0.0868566483259201,
0.050569936633110046,
-0.040603864938020706,
-0.026959538459777832,
0.04548517242074013,
0.06765088438987732,
-0.09669996052980423,
0.06760630011558533,
0.009751532226800919,
-0.11920495331287384,
-0.03518637642264366,
0.010957258753478527,
-0.07562252134084702,
0.07157810032367706,
0.007729232311248779,
-0.028513247147202492,
0.004583296831697226,
-0.016401249915361404,
0.010413174517452717,
-0.018589364364743233,
-0.15951070189476013,
-0.017387747764587402,
-0.12323169410228729,
-0.05553334206342697,
0.11709286272525787,
0.03462449461221695,
-0.23784784972667694,
0.011705098673701286,
-0.0969860851764679,
0.06600788980722427,
-0.19047285616397858,
0.05791420489549637,
0.2114742547273636,
-0.005059413146227598,
-0.031710248440504074,
-0.17269471287727356,
0.06433253735303879,
0.061398617923259735,
-0.036462731659412384,
-0.10163070261478424
] |
null | null | diffusers |
# LoRA DreamBooth - danaleee/beta
These are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks dog using [DreamBooth](https://dreambooth.github.io/). You can find some example images in the following.




LoRA for the text encoder was enabled: False.
| {"license": "creativeml-openrail-m", "tags": ["stable-diffusion", "stable-diffusion-diffusers", "text-to-image", "diffusers", "lora"], "base_model": "CompVis/stable-diffusion-v1-4", "instance_prompt": "a photo of sks dog", "inference": true} | text-to-image | danaleee/beta | [
"diffusers",
"tensorboard",
"stable-diffusion",
"stable-diffusion-diffusers",
"text-to-image",
"lora",
"base_model:CompVis/stable-diffusion-v1-4",
"license:creativeml-openrail-m",
"region:us"
] | 2024-02-07T08:07:37+00:00 | [] | [] | TAGS
#diffusers #tensorboard #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us
|
# LoRA DreamBooth - danaleee/beta
These are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks dog using DreamBooth. You can find some example images in the following.
!img_0
!img_1
!img_2
!img_3
LoRA for the text encoder was enabled: False.
| [
"# LoRA DreamBooth - danaleee/beta\n\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks dog using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
"TAGS\n#diffusers #tensorboard #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us \n",
"# LoRA DreamBooth - danaleee/beta\n\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks dog using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
70,
95
] | [
"passage: TAGS\n#diffusers #tensorboard #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us \n# LoRA DreamBooth - danaleee/beta\n\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks dog using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
-0.054172154515981674,
-0.01033832784742117,
-0.002750338986515999,
0.03125952556729317,
0.12006928026676178,
0.014381680637598038,
0.20662729442119598,
0.0847255289554596,
0.035181958228349686,
0.06514174491167068,
0.10183650255203247,
0.09683942794799805,
0.0032141092233359814,
0.10633260756731033,
-0.006345484405755997,
-0.12816904485225677,
0.023999659344553947,
-0.043756887316703796,
-0.1023145243525505,
0.010676583275198936,
0.04269254207611084,
-0.03878745809197426,
0.09792973101139069,
-0.019233131781220436,
-0.13073565065860748,
0.059076398611068726,
-0.028020447120070457,
-0.05274296551942825,
0.056479472666978836,
0.07906875014305115,
0.039485156536102295,
0.07632709294557571,
0.05398285761475563,
-0.1579451858997345,
0.03176164627075195,
0.02778986655175686,
-0.04221663251519203,
0.050782863050699234,
-0.05666472762823105,
-0.012776960618793964,
0.03045748546719551,
-0.10826177150011063,
0.00007716544496361166,
0.0024393429048359394,
-0.041698962450027466,
-0.039716992527246475,
0.0035010084975510836,
-0.029369929805397987,
0.03868741914629936,
0.09116216003894806,
-0.013137880712747574,
0.013724524527788162,
0.020035332068800926,
0.08149508386850357,
0.2661606967449188,
-0.25112032890319824,
-0.02136014774441719,
0.1993107795715332,
-0.05254936218261719,
0.10196416825056076,
-0.009974537417292595,
0.09584849327802658,
0.10337737202644348,
-0.04592033103108406,
0.01362625788897276,
-0.05568663403391838,
0.033074818551540375,
-0.08304986357688904,
-0.11344313621520996,
0.03631416708230972,
0.1718919575214386,
-0.023653248324990273,
-0.07649879157543182,
-0.11882010847330093,
-0.04431314766407013,
0.03984576463699341,
-0.023944053798913956,
0.033843327313661575,
0.012651086784899235,
0.018695486709475517,
-0.04280071705579758,
-0.08697943389415741,
-0.06425865739583969,
-0.08217167109251022,
-0.03299884498119354,
0.1743321716785431,
0.034974005073308945,
0.045033764094114304,
-0.011680426076054573,
0.1419675499200821,
-0.083526611328125,
-0.11520861089229584,
0.033181753009557724,
-0.03456516191363335,
-0.0073499539867043495,
0.09416748583316803,
-0.03956103324890137,
-0.1543130725622177,
0.09722968190908432,
-0.010548410005867481,
0.14057938754558563,
0.01643133908510208,
0.008179334923624992,
0.0899791568517685,
-0.018031932413578033,
0.04312853142619133,
-0.062249116599559784,
-0.037478942424058914,
0.03559240326285362,
0.006052736658602953,
0.08681115508079529,
-0.06153898313641548,
-0.1615544706583023,
-0.0038719431031495333,
-0.04745421186089516,
0.050532493740320206,
-0.04707971215248108,
0.02186676114797592,
-0.0695212334394455,
0.009745183400809765,
0.02947961911559105,
-0.04047418385744095,
-0.011257926002144814,
-0.056955937296152115,
-0.011587959714233875,
0.12635594606399536,
0.12897130846977234,
0.017978142946958542,
-0.013167358934879303,
0.04902644827961922,
-0.04200006276369095,
0.034729715436697006,
-0.014059597626328468,
-0.1341528743505478,
-0.017463302239775658,
-0.15652309358119965,
0.014351822435855865,
-0.15867221355438232,
-0.05192156136035919,
-0.005581328645348549,
0.004811443854123354,
-0.058405932039022446,
0.03907839208841324,
-0.08786684274673462,
-0.13050323724746704,
0.0020214563701301813,
0.04319942742586136,
0.025843175128102303,
-0.020840348675847054,
0.045183222740888596,
0.02022363618016243,
0.13211332261562347,
-0.15240365266799927,
-0.021866314113140106,
-0.05746453255414963,
0.014528635889291763,
-0.07909879088401794,
0.12190163880586624,
-0.039318181574344635,
0.037854503840208054,
-0.05103354528546333,
-0.022838493809103966,
-0.021338652819395065,
0.03165588155388832,
0.07076599448919296,
0.09678378701210022,
-0.25184309482574463,
-0.049766991287469864,
0.1223159208893776,
-0.1740199476480484,
-0.0911291316151619,
0.07535011321306229,
0.008912250399589539,
0.09988927096128464,
0.0779830664396286,
0.08514713495969772,
0.09897451847791672,
-0.299857497215271,
-0.020169930532574654,
-0.043694574385881424,
-0.01443848479539156,
-0.03298332914710045,
0.01735749840736389,
0.052028950303792953,
-0.04686250537633896,
0.0459815189242363,
-0.0912734642624855,
0.09661784768104553,
-0.028093896806240082,
-0.02099470980465412,
-0.03757992014288902,
-0.05042925849556923,
-0.011025971733033657,
-0.01576063223183155,
0.06326784193515778,
-0.01792004331946373,
-0.02660067193210125,
0.011892523616552353,
0.05122769996523857,
-0.0526113323867321,
0.028247667476534843,
-0.03426789864897728,
0.11209218204021454,
-0.11693967878818512,
0.020468903705477715,
-0.10218721628189087,
-0.0014047803124412894,
0.020080428570508957,
0.15186966955661774,
0.10071881115436554,
0.08429289609193802,
0.08458813279867172,
0.06185867264866829,
0.0030547522474080324,
-0.021767625585198402,
0.04305112361907959,
0.0053342427127063274,
-0.07089965045452118,
-0.16632826626300812,
0.0760711207985878,
-0.09198850393295288,
0.06723561137914658,
-0.11292428523302078,
0.040681514889001846,
0.026113037019968033,
0.14639587700366974,
0.07130715996026993,
-0.004328817129135132,
0.03222097456455231,
0.048181094229221344,
-0.06200230121612549,
-0.058864399790763855,
0.02385110966861248,
-0.0011846359120681882,
-0.17431093752384186,
0.16982504725456238,
-0.14638261497020721,
0.0025033457204699516,
0.11545119434595108,
0.02906802110373974,
-0.040483251214027405,
-0.040481582283973694,
0.00787185039371252,
0.03934941440820694,
-0.04622066766023636,
-0.054450102150440216,
0.15734684467315674,
-0.013442355208098888,
0.14514459669589996,
-0.02268732711672783,
0.047427013516426086,
0.043753135949373245,
-0.07246817648410797,
-0.06554848700761795,
0.08341792970895767,
0.017545079812407494,
0.03221677988767624,
0.009938645176589489,
0.10279451310634613,
-0.04171359911561012,
0.14757712185382843,
0.0023582701105624437,
-0.0620105005800724,
-0.05385024845600128,
0.041690077632665634,
0.05505813658237457,
0.10580235719680786,
0.01770317740738392,
-0.02045535109937191,
-0.015519768930971622,
-0.031060727313160896,
0.02067018486559391,
-0.15964657068252563,
-0.004933911841362715,
0.014295696280896664,
-0.06585505604743958,
0.1819038987159729,
0.0773361399769783,
-0.08871486783027649,
0.07758942991495132,
-0.08021190017461777,
-0.07203231751918793,
0.007495395373553038,
-0.035639695823192596,
-0.04516546055674553,
0.10099658370018005,
-0.08522311598062515,
-0.17794334888458252,
-0.17500688135623932,
0.014849558472633362,
-0.020078718662261963,
0.0007639157120138407,
0.06736340373754501,
-0.1148858517408371,
-0.042952846735715866,
-0.10762602090835571,
0.04028519615530968,
0.0017180470749735832,
0.07528594881296158,
0.0636826902627945,
-0.0014673101250082254,
0.01790967397391796,
-0.054400984197854996,
-0.018376778811216354,
-0.0683896541595459,
0.053023096174001694,
0.052353791892528534,
0.018007775768637657,
0.12038467824459076,
0.0924333781003952,
0.04021772742271423,
0.01677170768380165,
-0.001201548962853849,
0.220851331949234,
-0.004553430248051882,
0.05261234939098358,
0.14578059315681458,
0.02242334373295307,
0.05877363309264183,
0.12261588126420975,
0.017797213047742844,
-0.09088180214166641,
0.07537670433521271,
0.007240694481879473,
-0.14998984336853027,
-0.08435513824224472,
-0.06325171887874603,
-0.05324850231409073,
-0.06509628891944885,
0.041568323969841,
0.05264390632510185,
0.0674728974699974,
0.09924951940774918,
0.06599614024162292,
0.07450231164693832,
0.056952983140945435,
0.06944330781698227,
0.10140259563922882,
-0.06143604964017868,
0.0593147911131382,
-0.08792371302843094,
-0.11572727560997009,
0.09343095123767853,
-0.02264617569744587,
0.19175870716571808,
-0.07776548713445663,
0.020133264362812042,
0.0555882528424263,
-0.01706787757575512,
0.07887429744005203,
0.025311816483736038,
-0.031422410160303116,
-0.017822816967964172,
-0.042697906494140625,
-0.11278564482927322,
0.07804221659898758,
0.09282790869474411,
0.033769577741622925,
-0.01750960946083069,
-0.014830967411398888,
0.10905255377292633,
0.02432245947420597,
0.040544528514146805,
0.15352170169353485,
-0.21215128898620605,
0.05682188645005226,
0.010346612893044949,
0.06522328406572342,
0.011842848733067513,
-0.008364837616682053,
0.23803828656673431,
-0.010181448422372341,
0.05108176916837692,
-0.03427752107381821,
0.04957402125000954,
-0.03852006793022156,
-0.045973118394613266,
-0.07001940160989761,
0.15953373908996582,
-0.014110604301095009,
-0.021378077566623688,
-0.2765514552593231,
0.030751677230000496,
0.0014496544608846307,
0.03497987240552902,
-0.0451219379901886,
0.013631979003548622,
0.025348613038659096,
0.008708893321454525,
0.08769756555557251,
0.002394049894064665,
0.007794039323925972,
-0.10147135704755783,
-0.18317043781280518,
-0.01616571843624115,
0.0754062756896019,
-0.01860223524272442,
0.06558884680271149,
0.037424586713314056,
-0.0019719426054507494,
-0.006294111721217632,
0.033164165914058685,
-0.13231225311756134,
-0.12892556190490723,
-0.014398418366909027,
0.20651055872440338,
-0.010579746216535568,
-0.04550420865416527,
-0.07653041929006577,
-0.038932669907808304,
0.09578956663608551,
-0.09014048427343369,
-0.09685459733009338,
-0.07754635065793991,
0.026677221059799194,
0.05165000632405281,
-0.033731140196323395,
-0.020352214574813843,
-0.04301181808114052,
0.06203320249915123,
-0.0665869414806366,
-0.1379956305027008,
0.027675922960042953,
-0.05226067453622818,
-0.1248629167675972,
-0.07406647503376007,
0.07014524191617966,
-0.018727902323007584,
0.0036683131475001574,
-0.014594655483961105,
0.010290330275893211,
0.0301810409873724,
-0.09380237013101578,
0.05135785788297653,
0.18473079800605774,
-0.1585153490304947,
0.14017565548419952,
-0.052045535296201706,
-0.025708358734846115,
-0.07746528834104538,
-0.0022110198624432087,
0.14950516819953918,
0.2642119228839874,
-0.07850759476423264,
0.08855579048395157,
0.061396751552820206,
-0.11139043420553207,
-0.21915000677108765,
-0.061773497611284256,
-0.00013488072727341205,
0.037509676069021225,
-0.025165744125843048,
-0.11637526005506516,
0.09528183937072754,
0.04067168012261391,
0.016682105138897896,
0.17293430864810944,
-0.38697919249534607,
-0.14049053192138672,
-0.007555861957371235,
0.16134779155254364,
0.30931881070137024,
-0.13788969814777374,
-0.06861042231321335,
-0.03211921453475952,
-0.030553597956895828,
0.09325528889894485,
-0.06639479845762253,
0.12682634592056274,
-0.004479968920350075,
-0.033103231340646744,
0.016672953963279724,
-0.02275148034095764,
0.1405016928911209,
-0.023731455206871033,
0.027573252096772194,
-0.06783535331487656,
0.005731398705393076,
0.09338881075382233,
-0.04715607315301895,
0.03659556061029434,
-0.18118706345558167,
0.025637326762080193,
-0.01918145827949047,
-0.014517860487103462,
0.04309580475091934,
-0.026263652369379997,
0.0017494340427219868,
-0.07357074320316315,
-0.10335123538970947,
0.00662615243345499,
0.04201361909508705,
-0.02759593166410923,
0.04010189697146416,
-0.024766238406300545,
-0.04713289439678192,
0.13423489034175873,
-0.07554442435503006,
0.07726925611495972,
-0.047643955796957016,
-0.00026431219885125756,
-0.04975516349077225,
0.11366666108369827,
-0.08766064792871475,
-0.0030510739888995886,
0.1423155814409256,
0.03727954253554344,
0.11396811902523041,
0.028319798409938812,
-0.105119988322258,
0.09357118606567383,
0.09432791918516159,
-0.08412271738052368,
-0.010548174381256104,
-0.005427932366728783,
-0.033568743616342545,
0.11060771346092224,
-0.0036655995063483715,
0.2006116360425949,
-0.08325410634279251,
0.06110274791717529,
-0.0033141502644866705,
0.01577182114124298,
-0.030409175902605057,
0.1270712912082672,
-0.0004905261448584497,
-0.0013154579792171717,
-0.06865887343883514,
0.07337026298046112,
-0.03715074807405472,
0.020349418744444847,
0.0597105473279953,
-0.0026744832284748554,
-0.08255315572023392,
0.002732835477218032,
0.01673434115946293,
0.20582474768161774,
-0.09175104647874832,
-0.03938249498605728,
-0.14975038170814514,
-0.11036664247512817,
0.01989823579788208,
0.06984084099531174,
0.05293107405304909,
0.014965342357754707,
-0.03934361785650253,
-0.045814160257577896,
-0.05375919118523598,
0.04846075177192688,
0.014569942839443684,
0.0727226585149765,
-0.24142862856388092,
-0.028425168246030807,
0.019834579899907112,
-0.003417107043787837,
-0.07624939829111099,
-0.05557573586702347,
-0.10790015757083893,
-0.0019987178966403008,
-0.06607288122177124,
0.11630836874246597,
-0.04608852416276932,
-0.03949928283691406,
0.010657562874257565,
-0.013232647441327572,
-0.004613237455487251,
0.03291161358356476,
-0.01020771637558937,
-0.026204368099570274,
-0.008652281016111374,
-0.012783916667103767,
-0.029843397438526154,
-0.09932705014944077,
-0.043074123561382294,
-0.07936932146549225,
0.018255718052387238,
-0.009377941489219666,
-0.08002898097038269,
0.007261726073920727,
-0.20293591916561127,
0.043641429394483566,
0.20616285502910614,
-0.0021309317089617252,
-0.012463884428143501,
-0.02912883646786213,
0.006837605498731136,
-0.009521711617708206,
0.05142536386847496,
0.005971377249807119,
0.08506613224744797,
-0.0628289207816124,
-0.02414761297404766,
-0.08449061214923859,
0.023244261741638184,
-0.039128661155700684,
0.04633418098092079,
0.14007259905338287,
0.16779492795467377,
0.15946777164936066,
-0.15430773794651031,
0.06731399893760681,
-0.0457502044737339,
-0.0009064273326657712,
0.0032938760705292225,
-0.029510777443647385,
0.06713364273309708,
-0.07065870612859726,
-0.009535997174680233,
-0.02131052128970623,
0.15712083876132965,
0.0053427740931510925,
-0.16030895709991455,
-0.005836513824760914,
-0.0015101069584488869,
-0.018021924421191216,
0.007778781466186047,
0.1824970841407776,
0.04359138384461403,
0.04053672403097153,
-0.12818460166454315,
0.05070069432258606,
0.17576372623443604,
0.13323821127414703,
0.06001134589314461,
0.031133538112044334,
0.06216702610254288,
0.13551732897758484,
0.0637163445353508,
0.03497138246893883,
0.03134089335799217,
0.10667392611503601,
-0.0838087722659111,
0.09802467375993729,
-0.036544155329465866,
0.03695995360612869,
0.09249796718358994,
-0.032771043479442596,
-0.03909936547279358,
0.10573693364858627,
-0.030160415917634964,
-0.07629603892564774,
-0.06295756250619888,
-0.06311263144016266,
-0.15300238132476807,
0.009885143488645554,
-0.04647882282733917,
0.005089803598821163,
0.011842360720038414,
0.062141623347997665,
0.05586102232336998,
0.07131556421518326,
0.0008496407535858452,
-0.05899827554821968,
0.09172478318214417,
-0.036045536398887634,
-0.06987860798835754,
0.03079351782798767,
-0.0027809860184788704,
0.08626089990139008,
0.035426873713731766,
-0.02637234516441822,
0.08537698537111282,
0.05542196333408356,
0.03103274293243885,
-0.022090168669819832,
-0.061164528131484985,
-0.020146368071436882,
0.02233528345823288,
0.027413565665483475,
0.17657093703746796,
0.11881376057863235,
-0.07421532273292542,
-0.04904435575008392,
0.1264365166425705,
-0.06194787472486496,
-0.04145525395870209,
-0.14661498367786407,
0.029982391744852066,
-0.10888520628213882,
0.07021680474281311,
-0.06593533605337143,
-0.1066657155752182,
-0.06566891074180603,
0.15566232800483704,
0.15387983620166779,
-0.07641901820898056,
0.02057485468685627,
-0.06427215039730072,
-0.0034160963259637356,
-0.020216211676597595,
0.07284091413021088,
0.04145384579896927,
0.24495381116867065,
-0.028410905972123146,
0.005844183266162872,
-0.10614559054374695,
-0.07409948855638504,
-0.06741627305746078,
-0.09714343398809433,
-0.01838344894349575,
-0.03691950440406799,
-0.07175960391759872,
0.04235868528485298,
-0.14430782198905945,
-0.15959779918193817,
0.22806382179260254,
-0.12516075372695923,
-0.03168332204222679,
-0.05098360776901245,
0.0684470385313034,
0.045571424067020416,
0.022165313363075256,
-0.08261743187904358,
0.020212898030877113,
0.018140122294425964,
-0.03599535673856735,
-0.09031927585601807,
-0.006904130801558495,
-0.07835803180932999,
-0.15983746945858002,
0.04009735584259033,
-0.01807382144033909,
0.07299856841564178,
0.04823331534862518,
0.025526124984025955,
-0.09562382847070694,
0.10017077624797821,
-0.07610142976045609,
-0.08748438209295273,
-0.06344026327133179,
0.10186371207237244,
-0.042732708156108856,
0.07715613394975662,
0.060442954301834106,
-0.03159300237894058,
0.012811592780053616,
0.04494231566786766,
-0.06470823287963867,
-0.10573021322488785,
-0.024390600621700287,
-0.0965503603219986,
0.08119319379329681,
0.07805992662906647,
-0.02141474187374115,
0.022354772314429283,
0.010367368347942829,
-0.004995308816432953,
0.0005580981960520148,
-0.06917006522417068,
0.04543837532401085,
-0.10393901914358139,
-0.02074257843196392,
0.03343547508120537,
0.03537292778491974,
-0.29331865906715393,
-0.08472985774278641,
-0.16273368895053864,
-0.011953757144510746,
-0.0143580362200737,
0.08476707339286804,
0.24579408764839172,
0.033016107976436615,
-0.007523271255195141,
-0.21437034010887146,
0.018765192478895187,
0.09153532236814499,
-0.06267131119966507,
-0.08250268548727036
] |
null | null | null |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# roberta-base-bn-adapter-895K-snli-model2
This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.7648
- Accuracy: 0.7315
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 64
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|
| 0.4332 | 1.0 | 8584 | 0.3469 | 0.8699 |
| 0.4008 | 2.0 | 17168 | 0.3200 | 0.8780 |
| 0.3889 | 3.0 | 25752 | 0.3143 | 0.8805 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| {"license": "mit", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "roberta-base", "model-index": [{"name": "roberta-base-bn-adapter-895K-snli-model2", "results": []}]} | null | varun-v-rao/roberta-base-bn-adapter-895K-snli-model2 | [
"tensorboard",
"generated_from_trainer",
"base_model:roberta-base",
"license:mit",
"region:us"
] | 2024-02-07T08:09:03+00:00 | [] | [] | TAGS
#tensorboard #generated_from_trainer #base_model-roberta-base #license-mit #region-us
| roberta-base-bn-adapter-895K-snli-model2
========================================
This model is a fine-tuned version of roberta-base on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.7648
* Accuracy: 0.7315
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 64
* eval\_batch\_size: 64
* seed: 64
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.1+cu121
* Datasets 2.15.0
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
"TAGS\n#tensorboard #generated_from_trainer #base_model-roberta-base #license-mit #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
31,
98,
4,
33
] | [
"passage: TAGS\n#tensorboard #generated_from_trainer #base_model-roberta-base #license-mit #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
-0.07475393265485764,
-0.017714835703372955,
-0.0011093014618381858,
0.1175146996974945,
0.199581578373909,
0.033305101096630096,
0.1485203504562378,
0.06700192391872406,
-0.1284545660018921,
0.012931725941598415,
0.10589302331209183,
0.13276787102222443,
-0.010016347281634808,
0.05577821284532547,
-0.021496394649147987,
-0.2589932680130005,
-0.030867576599121094,
0.034199438989162445,
-0.14235948026180267,
0.12657976150512695,
0.08383151143789291,
-0.14470088481903076,
0.07032588869333267,
-0.0018644760129973292,
-0.3227788507938385,
0.0532296784222126,
0.03413255885243416,
-0.04161018878221512,
0.14325207471847534,
-0.007914844900369644,
0.19135071337223053,
-0.015160803683102131,
0.13267828524112701,
-0.12823326885700226,
0.013053737580776215,
0.06897308677434921,
0.0017943264683708549,
0.08231586217880249,
0.056769222021102905,
0.027435312047600746,
0.11498178541660309,
-0.08241923153400421,
0.062431663274765015,
0.018346304073929787,
-0.14348003268241882,
-0.24776354432106018,
-0.06946133822202682,
-0.02782427705824375,
0.05361013859510422,
0.08570152521133423,
-0.009992249310016632,
0.15058667957782745,
-0.0923137441277504,
0.09719747304916382,
0.2688377797603607,
-0.24733413755893707,
-0.08268550783395767,
0.08884695172309875,
0.005220783408731222,
0.09686415642499924,
-0.1086171343922615,
-0.027989622205495834,
0.07216697931289673,
0.0430314727127552,
0.12042160332202911,
-0.03158166632056236,
-0.12630656361579895,
0.04484284669160843,
-0.15301978588104248,
0.009162114933133125,
0.06210511550307274,
0.019741728901863098,
-0.01763567514717579,
0.04716676101088524,
-0.08500220626592636,
-0.1384432017803192,
-0.052238333970308304,
-0.021395880728960037,
0.08363467454910278,
-0.05047013610601425,
-0.1300412118434906,
-0.021556653082370758,
-0.10342920571565628,
-0.10336869210004807,
-0.05922522023320198,
0.21434934437274933,
0.05064752697944641,
0.04219092056155205,
-0.06608039140701294,
0.11258300393819809,
-0.05826476588845253,
-0.13001854717731476,
0.05500255897641182,
0.0506310909986496,
-0.02368917129933834,
-0.057791102677583694,
-0.08282174170017242,
-0.09287794679403305,
0.017902921885252,
0.11534342169761658,
-0.093721903860569,
0.052312783896923065,
0.057611554861068726,
0.05404213070869446,
-0.10234405845403671,
0.17427346110343933,
-0.05025380477309227,
0.03339746594429016,
0.009064878337085247,
0.026628121733665466,
-0.005459965672343969,
0.00016136743943206966,
-0.11499515920877457,
0.023156335577368736,
0.08081091940402985,
-0.0021891931537538767,
-0.08867049962282181,
0.03653155267238617,
-0.029007533565163612,
-0.002517419634386897,
-0.031703121960163116,
-0.08794890344142914,
0.04848870262503624,
-0.003793412819504738,
-0.09812897443771362,
-0.02240351215004921,
0.009053951129317284,
0.015105532482266426,
0.026523133739829063,
0.13590876758098602,
-0.10910443961620331,
0.08144775778055191,
-0.14631099998950958,
-0.12735936045646667,
-0.004410628229379654,
-0.04338817670941353,
0.028225958347320557,
-0.07183651626110077,
-0.14536058902740479,
-0.015226799994707108,
0.06742224842309952,
-0.040917009115219116,
-0.008866570889949799,
-0.03476836159825325,
-0.09650605916976929,
-0.014584335498511791,
-0.009945513680577278,
0.18935060501098633,
-0.0451742447912693,
0.10639970749616623,
0.06020219624042511,
0.06926393508911133,
-0.10678685456514359,
0.04169352725148201,
-0.09059275686740875,
0.0037791579961776733,
-0.2654353678226471,
0.013873259536921978,
-0.052873916923999786,
0.06341882795095444,
-0.06508420407772064,
-0.10088736563920975,
0.02093006670475006,
0.0013304909225553274,
0.11042109876871109,
0.06931009888648987,
-0.19004660844802856,
-0.07692719250917435,
0.12319201231002808,
-0.07398407906293869,
-0.06881386041641235,
0.0873371958732605,
-0.07504242658615112,
0.07223214954137802,
0.062358323484659195,
0.1941024661064148,
-0.040140602737665176,
-0.10151940584182739,
0.04610711708664894,
-0.036925096064805984,
0.040352627635002136,
-0.08445976674556732,
0.039315249770879745,
-0.007940866984426975,
0.010666628368198872,
0.027279241010546684,
-0.05941764637827873,
0.06706567853689194,
-0.13013575971126556,
-0.07790613919496536,
-0.049988918006420135,
-0.0946187824010849,
0.020613254979252815,
0.07975970208644867,
0.09930712729692459,
-0.10263907164335251,
-0.055715300142765045,
0.18971209228038788,
0.05886240303516388,
-0.05592130869626999,
0.029504060745239258,
-0.03566564992070198,
0.04614228755235672,
-0.07041586935520172,
-0.044696319848299026,
-0.20008429884910583,
-0.05124339461326599,
0.004241596441715956,
0.006946480832993984,
0.07186046987771988,
0.022618869319558144,
0.09191717207431793,
0.07819274812936783,
-0.07038570940494537,
0.016247980296611786,
-0.08071539551019669,
-0.003306666389107704,
-0.13741962611675262,
-0.22564268112182617,
-0.03651600703597069,
-0.011272558942437172,
0.07179941236972809,
-0.1964970976114273,
0.028477445244789124,
-0.0366688147187233,
0.07279051840305328,
0.0008826391422189772,
-0.024872280657291412,
-0.07309645414352417,
0.10362973809242249,
0.003958879038691521,
-0.0666680559515953,
0.04993414133787155,
-0.01573837548494339,
-0.06682416051626205,
-0.09235772490501404,
-0.08737160265445709,
0.17168691754341125,
0.13765643537044525,
-0.16576111316680908,
-0.10372001677751541,
0.03278122469782829,
-0.0663272961974144,
-0.02667389065027237,
-0.05819068104028702,
0.06024296581745148,
0.1849919557571411,
-0.01751491241157055,
0.1372009515762329,
-0.08131279051303864,
-0.04021052271127701,
0.0015376886585727334,
-0.04459095746278763,
0.07907983660697937,
0.0962468683719635,
0.18345078825950623,
-0.05712341517210007,
0.1127360463142395,
0.13974525034427643,
-0.15563876926898956,
0.1264190971851349,
-0.059313710778951645,
-0.07302986830472946,
-0.041987527161836624,
-0.02648402936756611,
-0.009619939140975475,
0.15616412460803986,
-0.07096295058727264,
0.007621396332979202,
-0.012010314501821995,
0.010820070281624794,
0.04194209724664688,
-0.25343260169029236,
-0.05540056899189949,
0.012922693975269794,
-0.031596552580595016,
-0.017510635778307915,
-0.03824382647871971,
0.00857454352080822,
0.10709058493375778,
-0.03027975931763649,
-0.07016342133283615,
0.01624760776758194,
0.016028832644224167,
-0.057268839329481125,
0.22584983706474304,
-0.0640183612704277,
-0.007865545339882374,
-0.09932583570480347,
-0.038271449506282806,
-0.04785483703017235,
-0.00896566640585661,
0.035262830555438995,
-0.13770060241222382,
-0.017891855910420418,
-0.03293406963348389,
0.04728424549102783,
-0.00033601472387090325,
0.047290559858083725,
-0.01309758611023426,
0.006732778158038855,
0.07837845385074615,
-0.13749851286411285,
0.01817816123366356,
-0.09285251796245575,
-0.1038198247551918,
0.03927066922187805,
0.07736311107873917,
0.12482556700706482,
0.16818830370903015,
-0.03677329421043396,
0.0015368383610621095,
-0.011017702519893646,
0.2803395092487335,
-0.06488595902919769,
-0.04008990526199341,
0.09389489889144897,
0.0007342978497035801,
0.05188651755452156,
0.0901697427034378,
0.11410600692033768,
-0.1147402673959732,
-0.004954434931278229,
0.05302564427256584,
-0.059258442372083664,
-0.22400081157684326,
-0.011986225843429565,
-0.02486732415854931,
-0.10102063417434692,
0.06064709648489952,
0.02945242077112198,
0.02161606214940548,
0.05956052988767624,
0.07426851987838745,
0.11207439005374908,
-0.0715116411447525,
0.04436013102531433,
0.05691972374916077,
0.06056540831923485,
0.11344251036643982,
-0.04857196286320686,
-0.0835137665271759,
0.02000521682202816,
-0.03666288033127785,
0.2722550928592682,
0.004791857209056616,
0.11151908338069916,
0.07511814683675766,
0.17710229754447937,
-0.00441728625446558,
0.09066253155469894,
0.0025761411525309086,
-0.0641959086060524,
-0.0061390213668346405,
-0.049995530396699905,
0.02045358158648014,
0.011897413060069084,
-0.0854366347193718,
0.051196251064538956,
-0.07561656087636948,
-0.01884339191019535,
0.05178448557853699,
0.21077051758766174,
-0.01336502842605114,
-0.3034566044807434,
-0.04394657537341118,
-0.015753470361232758,
0.004282798618078232,
0.0026473836041986942,
-0.0022000432945787907,
0.1375705599784851,
-0.04651518538594246,
-0.004209976643323898,
-0.05867151916027069,
0.08494039624929428,
-0.006153692025691271,
0.03057277202606201,
0.06447528302669525,
0.165020152926445,
-0.017411664128303528,
0.04512619227170944,
-0.2975105941295624,
0.3018229007720947,
0.024604640901088715,
0.10848745703697205,
-0.038883332163095474,
-0.01770537532866001,
0.016760198399424553,
0.02789769321680069,
0.0014502356061711907,
-0.018377121537923813,
-0.05453089252114296,
-0.2146025449037552,
-0.028614576905965805,
0.052944403141736984,
0.13145729899406433,
0.012239460833370686,
0.08638454228639603,
0.005012073554098606,
0.014945306815207005,
0.10343750566244125,
-0.06359148770570755,
-0.10726102441549301,
-0.03291144222021103,
-0.0660511925816536,
0.0022023250348865986,
-0.10666605085134506,
-0.05561085045337677,
-0.10655263811349869,
-0.08546047657728195,
0.10115735977888107,
0.020326783880591393,
-0.03022128902375698,
-0.12451990693807602,
0.13997405767440796,
0.10928459465503693,
-0.06083109974861145,
0.04781898856163025,
0.023947587236762047,
0.03365148976445198,
0.039909038692712784,
-0.05805416405200958,
0.120032399892807,
-0.04326903074979782,
-0.14123669266700745,
-0.051376134157180786,
0.04429418966174126,
0.059027425944805145,
0.06339193880558014,
-0.033860839903354645,
0.023341769352555275,
-0.01778288744390011,
-0.11109589040279388,
0.026515869423747063,
-0.06309778988361359,
0.01995905674993992,
0.008832522667944431,
-0.03575833514332771,
0.057091981172561646,
-0.04059373214840889,
-0.001675014034844935,
0.12161277234554291,
0.2856484055519104,
-0.0800790935754776,
-0.02934160642325878,
0.046779435127973557,
-0.07349865138530731,
-0.1612575352191925,
0.11288801580667496,
0.08126772940158844,
-0.011290083639323711,
0.06943327188491821,
-0.1785479336977005,
0.1565445214509964,
0.13100454211235046,
-0.010218306444585323,
0.11378266662359238,
-0.3223905861377716,
-0.15271008014678955,
0.0757310763001442,
0.2086057960987091,
0.16141444444656372,
-0.1565258651971817,
-0.01734081283211708,
-0.021323636174201965,
-0.09455014765262604,
0.09890241920948029,
-0.15009227395057678,
0.1073819026350975,
-0.012086882255971432,
0.10599377006292343,
-0.008253038860857487,
-0.06590919196605682,
0.11949073523283005,
0.026071971282362938,
0.15058591961860657,
-0.0409066379070282,
-0.03780120611190796,
0.10936670005321503,
-0.00938318483531475,
-0.01104013156145811,
-0.01393925677984953,
0.012540888041257858,
-0.027807138860225677,
-0.00009752086771186441,
-0.08587795495986938,
0.04370386153459549,
-0.03777909651398659,
-0.06980567425489426,
-0.05090554058551788,
-0.0008405999396927655,
-0.012974551878869534,
-0.02806784398853779,
0.12514260411262512,
0.05377015098929405,
0.15139234066009521,
0.09084928035736084,
0.0016517069889232516,
-0.0512702539563179,
-0.06706840544939041,
0.02465876005589962,
-0.0021503958851099014,
0.07514401525259018,
-0.18342944979667664,
-0.011050857603549957,
0.12439505010843277,
0.042387474328279495,
0.09199195355176926,
0.10083737969398499,
-0.05436813086271286,
0.034434471279382706,
0.06739165633916855,
-0.14774778485298157,
-0.08985953032970428,
0.02428152784705162,
-0.10520867258310318,
-0.0652744397521019,
0.060116302222013474,
0.07674793154001236,
-0.07997829467058182,
-0.002550684381276369,
-0.011752395890653133,
-0.03389208763837814,
-0.08400460332632065,
0.2256517857313156,
0.11135032027959824,
0.037366706877946854,
-0.11652616411447525,
0.07407764345407486,
0.05569181963801384,
-0.016596117988228798,
0.0009346549632027745,
0.07391172647476196,
-0.048800546675920486,
0.0132980328053236,
0.1510714739561081,
0.23936761915683746,
-0.10910001397132874,
-0.02936551347374916,
-0.189443901181221,
-0.08038059622049332,
0.04212786629796028,
0.1884782314300537,
0.1147407591342926,
-0.019858920946717262,
-0.028631627559661865,
0.046605415642261505,
-0.15364845097064972,
0.05338914692401886,
0.010879178531467915,
0.09047134965658188,
-0.11368510872125626,
0.18626739084720612,
0.030796127393841743,
0.01839439384639263,
-0.03158188983798027,
0.05061252787709236,
-0.14782902598381042,
0.030198654159903526,
-0.1055854931473732,
-0.06803852319717407,
0.02992464043200016,
-0.017146103084087372,
0.004401281010359526,
-0.07911252975463867,
-0.07808925211429596,
0.03522586077451706,
-0.1446821391582489,
-0.014972825534641743,
0.05801524594426155,
0.012702934443950653,
-0.14669275283813477,
-0.019017741084098816,
0.025810452178120613,
-0.04397733509540558,
0.023243220522999763,
0.05442385375499725,
0.03564239293336868,
0.09325142949819565,
-0.19438046216964722,
-0.0111426692456007,
0.06653879582881927,
-0.013161791488528252,
0.09211081266403198,
-0.020552048459649086,
-0.02697448991239071,
-0.027598824352025986,
0.15139201283454895,
0.009062389843165874,
0.06183305382728577,
-0.134662926197052,
0.00843550730496645,
-0.060163386166095734,
-0.09704585373401642,
-0.04295851290225983,
0.0038129009772092104,
0.08274749666452408,
0.03212431073188782,
0.1570647656917572,
-0.0706486850976944,
0.02624971605837345,
-0.2289324700832367,
-0.01669205166399479,
-0.005091591738164425,
-0.08165226131677628,
-0.1053396537899971,
-0.049344029277563095,
0.06685058027505875,
-0.05818529427051544,
0.14202669262886047,
0.0551544651389122,
0.061387158930301666,
0.03587254881858826,
0.018496448174118996,
0.0105598671361804,
0.029603509232401848,
0.22425638139247894,
0.038315799087285995,
-0.01832764968276024,
0.0697413831949234,
0.09782756119966507,
0.11547084152698517,
0.08346835523843765,
0.20962698757648468,
0.14570488035678864,
-0.058171261101961136,
0.09439346939325333,
0.06460303068161011,
-0.048168450593948364,
-0.11091910302639008,
0.08507418632507324,
-0.022278282791376114,
0.06225833296775818,
-0.033936869353055954,
0.20796671509742737,
0.07659223675727844,
-0.15880735218524933,
0.03877020999789238,
-0.03268436715006828,
-0.10177382081747055,
-0.10245922952890396,
0.04301570728421211,
-0.07914014905691147,
-0.17716224491596222,
0.035248711705207825,
-0.10426777601242065,
0.01913699507713318,
0.157199889421463,
-0.0034432653337717056,
0.008154734037816525,
0.19545809924602509,
0.08775214850902557,
0.0543307401239872,
0.016101188957691193,
-0.005341599229723215,
-0.04086948186159134,
-0.0889374166727066,
-0.07232381403446198,
-0.019491875544190407,
-0.018526431173086166,
0.018520308658480644,
-0.0542416125535965,
-0.09625164419412613,
0.03727414086461067,
-0.003945624455809593,
-0.09825858473777771,
0.02077147737145424,
0.036861203610897064,
0.06726893782615662,
0.00013124800170771778,
-0.0062584890983998775,
0.009733867831528187,
-0.037252962589263916,
0.21817241609096527,
-0.06573435664176941,
-0.09164528548717499,
-0.06735531240701675,
0.23014691472053528,
0.032301656901836395,
-0.002790899248793721,
0.00775654474273324,
-0.08668593317270279,
-0.002038941951468587,
0.19788742065429688,
0.16351957619190216,
-0.11366628110408783,
-0.017749188467860222,
-0.009833009913563728,
-0.02740822732448578,
-0.06558236479759216,
0.1562657654285431,
0.09146503359079361,
0.0213262178003788,
-0.12602870166301727,
-0.05769830569624901,
-0.06878907233476639,
-0.02548471838235855,
-0.05251413956284523,
0.022052815183997154,
0.06907997280359268,
0.006339527200907469,
-0.05708359554409981,
0.07300571352243423,
-0.06542234867811203,
-0.1317337602376938,
0.09377675503492355,
-0.16966968774795532,
-0.1607896387577057,
-0.004568170290440321,
0.13591895997524261,
-0.016043150797486305,
0.07748594880104065,
-0.054651886224746704,
0.0037034074775874615,
0.03522259369492531,
-0.04413066431879997,
-0.10082808136940002,
-0.12409280240535736,
0.11054088175296783,
-0.12651097774505615,
0.20033124089241028,
-0.037229958921670914,
0.11409102380275726,
0.10805658251047134,
0.0468689426779747,
-0.0679241269826889,
0.060261402279138565,
0.033237867057323456,
-0.13597342371940613,
-0.01290198601782322,
0.11421385407447815,
-0.03421597555279732,
0.02231750637292862,
0.01301855593919754,
-0.12435343116521835,
0.03628883138298988,
-0.03998567536473274,
-0.04306679219007492,
-0.05336714908480644,
-0.09049433469772339,
-0.0773928090929985,
0.0878586620092392,
0.20426298677921295,
-0.007984903641045094,
0.04552052915096283,
-0.0874137282371521,
0.03384704142808914,
0.08585871756076813,
0.04565951228141785,
-0.08657516539096832,
-0.2458726167678833,
0.020323947072029114,
0.13361895084381104,
-0.07782333344221115,
-0.2184029072523117,
-0.08471615612506866,
0.025904567912220955,
-0.08172468841075897,
-0.07608304172754288,
0.06640633195638657,
0.10590462386608124,
0.06730035692453384,
-0.04212311655282974,
-0.1393756866455078,
-0.09570799767971039,
0.1443653106689453,
-0.14852774143218994,
-0.10086111724376678
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# decilm-7b-ocn-v1
This model is a fine-tuned version of Deci/DeciLM-7B on the oncc_medqa_instruct dataset.
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0005
- train_batch_size: 4
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_steps: 10
- num_epochs: 1.0
- mixed_precision_training: Native AMP
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.1.1+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
### Performance
hf (pretrained=Deci/DeciLM-7B,trust_remote_code=True,parallelize=True,load_in_4bit=True,peft=chenhugging/decilm-7b-ocn-v1), gen_kwargs: (None), limit: 100.0, num_fewshot: None, batch_size: 1
| Tasks |Version|Filter|n-shot| Metric |Value| |Stderr|
|---------------------|-------|------|-----:|--------|----:|---|-----:|
|pubmedqa | 1|none | 0|acc | 0.95|± |0.0219|
|medmcqa |Yaml |none | 0|acc | 0.34|± |0.0476|
|professional_medicine| 0|none | 0|acc | 0.58|± |0.0496|
|college_medicine | 0|none | 0|acc | 0.54|± |0.0501|
|clinical_knowledge | 0|none | 0|acc | 0.57|± |0.0498|
|ocn |Yaml |none | 0|acc | 0.68|± |0.0469|
|aocnp |Yaml |none | 0|acc | 0.43|± |0.0498|
### Appendix: Original Performance
hf (pretrained=Deci/DeciLM-7B,trust_remote_code=True,parallelize=True,load_in_4bit=True), gen_kwargs: (None), limit: 100.0, num_fewshot: None, batch_size: 1
| Tasks |Version|Filter|n-shot| Metric |Value| |Stderr|
|---------------------|-------|------|-----:|--------|----:|---|-----:|
|pubmedqa | 1|none | 0|acc | 0.89|± |0.0314|
|medmcqa |Yaml |none | 0|acc | 0.37|± |0.0485|
|professional_medicine| 0|none | 0|acc | 0.58|± |0.0496|
|college_medicine | 0|none | 0|acc | 0.54|± |0.0501|
|clinical_knowledge | 0|none | 0|acc | 0.63|± |0.0485|
|ocn |Yaml |none | 0|acc | 0.64|± |0.0482|
|aocnp |Yaml |none | 0|acc | 0.41|± |0.0494|
| {"license": "other", "library_name": "peft", "tags": ["llama-factory", "lora", "generated_from_trainer"], "base_model": "Deci/DeciLM-7B", "model-index": [{"name": "decilm-7b-ocn-v1", "results": []}]} | null | chenhugging/decilm-7b-ocn-v1 | [
"peft",
"safetensors",
"llama-factory",
"lora",
"generated_from_trainer",
"base_model:Deci/DeciLM-7B",
"license:other",
"region:us"
] | 2024-02-07T08:12:11+00:00 | [] | [] | TAGS
#peft #safetensors #llama-factory #lora #generated_from_trainer #base_model-Deci/DeciLM-7B #license-other #region-us
| decilm-7b-ocn-v1
================
This model is a fine-tuned version of Deci/DeciLM-7B on the oncc\_medqa\_instruct dataset.
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0005
* train\_batch\_size: 4
* eval\_batch\_size: 8
* seed: 42
* gradient\_accumulation\_steps: 4
* total\_train\_batch\_size: 16
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: cosine
* lr\_scheduler\_warmup\_steps: 10
* num\_epochs: 1.0
* mixed\_precision\_training: Native AMP
### Framework versions
* PEFT 0.8.2
* Transformers 4.37.2
* Pytorch 2.1.1+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
### Performance
hf (pretrained=Deci/DeciLM-7B,trust\_remote\_code=True,parallelize=True,load\_in\_4bit=True,peft=chenhugging/decilm-7b-ocn-v1), gen\_kwargs: (None), limit: 100.0, num\_fewshot: None, batch\_size: 1
### Appendix: Original Performance
hf (pretrained=Deci/DeciLM-7B,trust\_remote\_code=True,parallelize=True,load\_in\_4bit=True), gen\_kwargs: (None), limit: 100.0, num\_fewshot: None, batch\_size: 1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 1.0\n* mixed\\_precision\\_training: Native AMP",
"### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1",
"### Performance\n\n\nhf (pretrained=Deci/DeciLM-7B,trust\\_remote\\_code=True,parallelize=True,load\\_in\\_4bit=True,peft=chenhugging/decilm-7b-ocn-v1), gen\\_kwargs: (None), limit: 100.0, num\\_fewshot: None, batch\\_size: 1",
"### Appendix: Original Performance\n\n\nhf (pretrained=Deci/DeciLM-7B,trust\\_remote\\_code=True,parallelize=True,load\\_in\\_4bit=True), gen\\_kwargs: (None), limit: 100.0, num\\_fewshot: None, batch\\_size: 1"
] | [
"TAGS\n#peft #safetensors #llama-factory #lora #generated_from_trainer #base_model-Deci/DeciLM-7B #license-other #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 1.0\n* mixed\\_precision\\_training: Native AMP",
"### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1",
"### Performance\n\n\nhf (pretrained=Deci/DeciLM-7B,trust\\_remote\\_code=True,parallelize=True,load\\_in\\_4bit=True,peft=chenhugging/decilm-7b-ocn-v1), gen\\_kwargs: (None), limit: 100.0, num\\_fewshot: None, batch\\_size: 1",
"### Appendix: Original Performance\n\n\nhf (pretrained=Deci/DeciLM-7B,trust\\_remote\\_code=True,parallelize=True,load\\_in\\_4bit=True), gen\\_kwargs: (None), limit: 100.0, num\\_fewshot: None, batch\\_size: 1"
] | [
48,
159,
39,
105,
91
] | [
"passage: TAGS\n#peft #safetensors #llama-factory #lora #generated_from_trainer #base_model-Deci/DeciLM-7B #license-other #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 1.0\n* mixed\\_precision\\_training: Native AMP### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1### Performance\n\n\nhf (pretrained=Deci/DeciLM-7B,trust\\_remote\\_code=True,parallelize=True,load\\_in\\_4bit=True,peft=chenhugging/decilm-7b-ocn-v1), gen\\_kwargs: (None), limit: 100.0, num\\_fewshot: None, batch\\_size: 1### Appendix: Original Performance\n\n\nhf (pretrained=Deci/DeciLM-7B,trust\\_remote\\_code=True,parallelize=True,load\\_in\\_4bit=True), gen\\_kwargs: (None), limit: 100.0, num\\_fewshot: None, batch\\_size: 1"
] | [
-0.10532104969024658,
0.13686434924602509,
-0.0066984714940190315,
0.07450215518474579,
0.07911797612905502,
0.0390290692448616,
0.07124868780374527,
0.16543897986412048,
-0.02276904694736004,
0.14483195543289185,
0.080811507999897,
0.06925346702337265,
0.07614275068044662,
0.137369304895401,
-0.008792483247816563,
-0.23281443119049072,
0.013998176902532578,
-0.041421934962272644,
-0.09322331100702286,
0.10250108689069748,
0.06075635552406311,
-0.09945425391197205,
0.04642562195658684,
-0.037389881908893585,
-0.036841947585344315,
-0.023749100044369698,
-0.054095372557640076,
-0.02703852206468582,
0.06869494915008545,
0.03724740073084831,
0.05773014575242996,
0.046331219375133514,
0.06286665052175522,
-0.24364566802978516,
-0.00283451029099524,
0.09205470979213715,
0.008036602288484573,
0.06811878085136414,
0.10752646625041962,
-0.014759588055312634,
0.07934754341840744,
-0.12342969328165054,
0.02900080196559429,
0.03976471722126007,
-0.14295315742492676,
-0.23006752133369446,
-0.0800226703286171,
0.06536059081554413,
0.13114933669567108,
0.06069646775722504,
-0.03574282303452492,
0.09357888251543045,
-0.0925259068608284,
0.08088506758213043,
0.20797987282276154,
-0.2530263364315033,
-0.06552369892597198,
0.011909542605280876,
0.03995734453201294,
0.01632019132375717,
-0.11065656691789627,
-0.03811008855700493,
0.01803576573729515,
0.025036919862031937,
0.05952942371368408,
-0.0207368154078722,
0.04519481211900711,
-0.012320587411522865,
-0.12518160045146942,
-0.07137394696474075,
0.09299314022064209,
0.05777691304683685,
-0.027188364416360855,
-0.05877363309264183,
-0.04902138188481331,
-0.21085324883460999,
-0.04764566197991371,
0.014958098530769348,
0.015785232186317444,
-0.039768751710653305,
-0.07939009368419647,
0.10123054683208466,
-0.03310532122850418,
-0.051548536866903305,
0.05865482985973358,
0.10255281627178192,
0.07950115948915482,
-0.04145297780632973,
0.017255516722798347,
0.10165999084711075,
0.040000613778829575,
-0.192393496632576,
-0.030894307419657707,
0.014635906554758549,
-0.09157068282365799,
-0.018419532105326653,
-0.028739482164382935,
0.07454320788383484,
0.08337374776601791,
0.22671420872211456,
-0.10223538428544998,
0.11509066820144653,
0.05739164724946022,
0.006090763956308365,
-0.05694423243403435,
0.12034977227449417,
-0.0983302891254425,
-0.10349136590957642,
-0.05405271798372269,
0.10198316723108292,
-0.011969270184636116,
-0.01933097653090954,
-0.03555218502879143,
0.018515143543481827,
0.12288996577262878,
0.05133531987667084,
0.012057052925229073,
0.05873081460595131,
-0.06444252282381058,
-0.017142921686172485,
0.07504665851593018,
-0.12028785049915314,
0.04342572018504143,
0.06832978874444962,
-0.07866550981998444,
-0.024564135819673538,
-0.010498074814677238,
-0.022396322339773178,
-0.06237122043967247,
0.09040015935897827,
-0.07385528087615967,
-0.03192560747265816,
-0.05818302556872368,
-0.08846592158079147,
0.034457579255104065,
-0.0424506701529026,
-0.005638495087623596,
-0.08135471493005753,
-0.07103835791349411,
-0.07310425490140915,
0.055780354887247086,
-0.08434762060642242,
-0.025923429057002068,
-0.043626610189676285,
-0.1164417490363121,
0.050152599811553955,
-0.0035214757081121206,
0.11661185324192047,
-0.05482897907495499,
0.09057388454675674,
0.020320789888501167,
0.05541890859603882,
0.09833857417106628,
0.038228243589401245,
-0.0807432234287262,
0.08319173008203506,
-0.17902198433876038,
0.0657898336648941,
-0.0986468568444252,
0.056585170328617096,
-0.15384046733379364,
-0.1170341894030571,
-0.004451362416148186,
-0.016284748911857605,
0.1079767569899559,
0.11635079234838486,
-0.1596221625804901,
-0.05747233331203461,
0.1811947077512741,
-0.05475660040974617,
-0.08641909807920456,
0.1127176508307457,
-0.01215281430631876,
-0.05085710436105728,
0.01115533709526062,
0.15825165808200836,
0.14768117666244507,
-0.08446783572435379,
-0.06555632501840591,
-0.046145979315042496,
0.09830170124769211,
0.08376048505306244,
0.07346132397651672,
-0.01870209537446499,
0.06170881539583206,
0.015206470154225826,
-0.03645666688680649,
0.027171391993761063,
-0.07170858979225159,
-0.07066291570663452,
0.0014717065496370196,
-0.05069470405578613,
0.041308023035526276,
0.06610498577356339,
0.005455002188682556,
-0.07795130461454391,
-0.14599330723285675,
-0.030247565358877182,
0.09289092570543289,
-0.07713064551353455,
0.013789810240268707,
-0.06617622077465057,
0.07140978425741196,
0.03364177048206329,
0.016813743859529495,
-0.15702779591083527,
-0.0438096709549427,
0.0472341850399971,
-0.09969805181026459,
-0.013337686657905579,
-0.043718788772821426,
0.07177833467721939,
0.06618540734052658,
-0.035192400217056274,
-0.07151398062705994,
-0.016522753983736038,
-0.0008159676799550653,
-0.062463533133268356,
-0.22928588092327118,
-0.09291067719459534,
-0.010820739902555943,
0.14370539784431458,
-0.17948925495147705,
0.007629128638654947,
0.0655052587389946,
0.17019765079021454,
-0.0064945281483232975,
-0.031892966479063034,
-0.001810060115531087,
0.08628810197114944,
-0.026310088112950325,
-0.06956300884485245,
0.03442428633570671,
-0.01686425693333149,
-0.06844327598810196,
-0.02291095070540905,
-0.13969667255878448,
0.03296782076358795,
0.08297314494848251,
-0.0025457378942519426,
-0.12466255575418472,
0.0067707691341638565,
-0.053163912147283554,
-0.06666579842567444,
0.029449492692947388,
-0.006765627767890692,
0.16234710812568665,
0.03843401372432709,
0.08862163126468658,
-0.0644453689455986,
-0.07007116824388504,
0.02469024620950222,
-0.0037172322627156973,
0.003330004634335637,
0.15816576778888702,
0.03261563181877136,
-0.08569303900003433,
0.10033703595399857,
0.09088565409183502,
-0.04716769978404045,
0.11917530745267868,
-0.079447902739048,
-0.08768734335899353,
-0.083001509308815,
0.05599420145153999,
0.05057895928621292,
0.07833979278802872,
-0.08317416161298752,
0.00377267156727612,
0.013232424855232239,
0.002666313434019685,
0.0011665017809718847,
-0.1405765563249588,
0.01674017868936062,
0.003964342642575502,
-0.05027131363749504,
0.028920218348503113,
-0.005406734999269247,
-0.0011866120621562004,
0.07057714462280273,
0.027008237317204475,
-0.0589376762509346,
-0.024793298915028572,
-0.03941496089100838,
-0.05765051394701004,
0.20374833047389984,
-0.08668088912963867,
-0.12031620740890503,
-0.10378840565681458,
-0.010954801924526691,
-0.018960261717438698,
-0.03806368634104729,
0.023746322840452194,
-0.08382319658994675,
-0.045834220945835114,
-0.06909877061843872,
-0.04456136375665665,
-0.011382639408111572,
0.0020599733106791973,
-0.009089138358831406,
0.010552971623837948,
0.1005706787109375,
-0.09663482755422592,
0.000015867213733145036,
0.02188713289797306,
-0.05074493587017059,
0.041143130511045456,
0.04417950659990311,
0.09694793820381165,
0.1448269486427307,
0.03861073777079582,
0.02711343951523304,
-0.017841359600424767,
0.2022496908903122,
-0.08696775883436203,
0.0272380243986845,
0.06756839156150818,
-0.019565369933843613,
0.0816238597035408,
0.17023521661758423,
0.041520845144987106,
-0.07075153291225433,
-0.006426084320992231,
0.04083041846752167,
-0.010177364572882652,
-0.19038380682468414,
-0.025854459032416344,
-0.0445830300450325,
0.03948890045285225,
0.127025306224823,
0.03415418043732643,
0.0361395888030529,
0.01900514028966427,
-0.03509609028697014,
0.04773527383804321,
0.01324858795851469,
0.07196716964244843,
0.03816975653171539,
0.06621425598859787,
0.08800742030143738,
-0.017201464623212814,
-0.021420510485768318,
0.014617147855460644,
-0.0028036420699208975,
0.20277924835681915,
-0.049466028809547424,
0.20197053253650665,
0.07542625814676285,
0.1699511557817459,
-0.01414243783801794,
0.03526794910430908,
0.004256790038198233,
0.01714809611439705,
0.01733626239001751,
-0.07474803179502487,
-0.0027502202428877354,
0.026904093101620674,
0.04772167280316353,
0.010904191993176937,
-0.08630725741386414,
0.005840289872139692,
0.06267475336790085,
0.24275848269462585,
0.11276824027299881,
-0.28515732288360596,
-0.05920400843024254,
0.013927345164120197,
-0.019093738868832588,
-0.023023489862680435,
0.017343979328870773,
0.08725295960903168,
-0.0775860995054245,
0.0723181664943695,
-0.04526159539818764,
0.07756379246711731,
-0.07087590545415878,
0.03424973785877228,
0.11350782215595245,
0.08774256706237793,
0.027243655174970627,
0.053959719836711884,
-0.23761074244976044,
0.2358129769563675,
-0.008275979198515415,
0.06375884264707565,
-0.039973869919776917,
0.04159887507557869,
0.02324976772069931,
-0.055778566747903824,
0.07965385168790817,
0.0014747397508472204,
-0.1025875061750412,
-0.20357483625411987,
-0.09444654732942581,
-0.002762097166851163,
0.134944885969162,
-0.07635173946619034,
0.15013734996318817,
-0.012696923688054085,
-0.05495961755514145,
0.010050861164927483,
-0.015670763328671455,
-0.11206135898828506,
-0.06911836564540863,
0.06941584497690201,
-0.03728695958852768,
0.02741354890167713,
-0.07295555621385574,
-0.05665929242968559,
-0.09652123600244522,
0.14883381128311157,
-0.13596107065677643,
-0.05013756826519966,
-0.11320102214813232,
0.06050323322415352,
0.18113015592098236,
-0.11214655637741089,
0.031920645385980606,
0.0067978654988110065,
0.058403342962265015,
0.03918479382991791,
-0.04405396804213524,
0.10818552225828171,
-0.0471668504178524,
-0.2311749905347824,
-0.05533687770366669,
0.13385413587093353,
0.021024908870458603,
0.04592552036046982,
-0.040080100297927856,
0.054968081414699554,
0.01671099103987217,
-0.10082312673330307,
0.05701916664838791,
-0.03186734393239021,
0.06223156675696373,
0.029829051345586777,
-0.013120315037667751,
0.015538889914751053,
-0.0728931650519371,
0.00993482768535614,
0.037594832479953766,
0.29775068163871765,
-0.09532292932271957,
0.007535990327596664,
0.07727934420108795,
-0.05434621870517731,
-0.17309141159057617,
-0.06428294628858566,
0.10329847037792206,
-0.014922990463674068,
-0.00495339697226882,
-0.17835023999214172,
0.03195052966475487,
0.09075486660003662,
0.0024702600203454494,
0.10407908260822296,
-0.3442223072052002,
-0.12252973765134811,
0.05989501252770424,
0.07496101409196854,
-0.08883972465991974,
-0.2043541818857193,
-0.08018322288990021,
-0.00472624646499753,
-0.14955738186836243,
0.019988304004073143,
-0.04179874062538147,
0.10739952325820923,
-0.02465011551976204,
-0.030133038759231567,
0.016021646559238434,
-0.06611783802509308,
0.1499883383512497,
0.02776726521551609,
0.05919411778450012,
-0.0508192777633667,
-0.0327625572681427,
0.06320509314537048,
-0.08053163439035416,
0.015312878414988518,
-0.08064473420381546,
0.035026367753744125,
-0.09154769033193588,
0.010909765027463436,
-0.09637347608804703,
0.010203099809587002,
-0.09083415567874908,
-0.0229326281696558,
-0.029119564220309258,
0.08723186701536179,
0.08831460773944855,
0.008415931835770607,
0.08590055257081985,
0.007779431529343128,
0.15249879658222198,
0.18652287125587463,
0.03502902761101723,
0.06527399271726608,
-0.11306438595056534,
0.006509723607450724,
-0.002517447806894779,
0.028426289558410645,
-0.12325956672430038,
0.04120380058884621,
0.15259984135627747,
0.047790106385946274,
0.11971426010131836,
0.04895520582795143,
-0.09250137209892273,
-0.01036106888204813,
0.07189257442951202,
-0.1069171130657196,
-0.09259065985679626,
-0.01206126157194376,
-0.042571891099214554,
-0.10868726670742035,
-0.006735837087035179,
0.12705759704113007,
-0.04027136415243149,
0.0014056371292099357,
0.023705651983618736,
0.05514479801058769,
-0.021183567121624947,
0.22132030129432678,
0.035190314054489136,
0.10459601879119873,
-0.09673910588026047,
0.09197326749563217,
0.09792091697454453,
-0.08423413336277008,
0.0562916174530983,
0.11067135632038116,
-0.06780653446912766,
-0.04564826190471649,
0.01951250433921814,
0.0777854397892952,
0.062091559171676636,
-0.03301326930522919,
-0.12261028587818146,
-0.09982756525278091,
0.09620127081871033,
0.11342323571443558,
0.020369501784443855,
0.048595014959573746,
0.022492364048957825,
0.020762905478477478,
-0.08986561000347137,
0.130821093916893,
0.1000867411494255,
0.07781770825386047,
-0.11894398927688599,
0.1606515347957611,
0.004952976480126381,
0.002877848455682397,
0.010121152736246586,
0.006824478507041931,
-0.17004059255123138,
-0.0023457526694983244,
-0.14541275799274445,
0.01047514658421278,
-0.05666372925043106,
-0.019542651250958443,
-0.009757795371115208,
-0.02256840281188488,
-0.04716033488512039,
0.02307642623782158,
-0.09529800713062286,
-0.07602138817310333,
-0.011072621680796146,
0.07894060760736465,
-0.11978451162576675,
-0.028760403394699097,
0.039795782417058945,
-0.12196486443281174,
0.07779757678508759,
0.018146393820643425,
0.020517269149422646,
0.008471641689538956,
-0.022627953439950943,
0.011334611102938652,
-0.017234982922673225,
-0.008732750080525875,
0.03209109604358673,
-0.1479358822107315,
-0.0028217171784490347,
-0.08271332830190659,
-0.0007751929224468768,
0.009443146176636219,
0.014012088999152184,
-0.13176557421684265,
0.04643069580197334,
-0.02419520914554596,
-0.03999555483460426,
-0.06251731514930725,
0.042432915419340134,
0.07753787934780121,
-0.004206882789731026,
0.1215893104672432,
-0.06954384595155716,
0.07455005496740341,
-0.2352745085954666,
-0.03436023369431496,
0.014462037943303585,
-0.058933865278959274,
-0.05016026645898819,
-0.017302557826042175,
0.09794040024280548,
-0.08531662821769714,
0.06031930074095726,
-0.058806177228689194,
0.05788663774728775,
0.011633471585810184,
-0.07048357278108597,
-0.031475309282541275,
0.07917436212301254,
0.13186641037464142,
0.032315660268068314,
-0.04896934702992439,
0.01762407273054123,
0.004153627902269363,
0.013536571525037289,
0.04805417358875275,
0.1358078569173813,
0.14524142444133759,
0.08041635155677795,
0.053160376846790314,
0.038082465529441833,
-0.14307670295238495,
-0.11777685582637787,
0.12362056970596313,
-0.060215577483177185,
0.13621582090854645,
-0.04424196481704712,
0.11088790744543076,
0.0791439563035965,
-0.2025882601737976,
0.06092564016580582,
-0.07080382108688354,
-0.09705166518688202,
-0.0689270943403244,
-0.10798154026269913,
-0.07373618334531784,
-0.10945111513137817,
0.007443446200340986,
-0.09656740725040436,
0.040387578308582306,
0.0989406555891037,
0.03056168742477894,
0.0279388390481472,
0.11735181510448456,
0.054827917367219925,
0.004965836647897959,
0.08708062767982483,
0.031637128442525864,
-0.018026819452643394,
0.010074359364807606,
-0.0469801090657711,
0.01906922087073326,
-0.019007446244359016,
0.04858483001589775,
-0.009575705975294113,
-0.045179370790719986,
0.05690249428153038,
0.015142506919801235,
-0.10221267491579056,
0.03379707783460617,
-0.00204489566385746,
0.025789204984903336,
0.08516708016395569,
0.04331941530108452,
-0.0091093173250556,
-0.01826210506260395,
0.16418525576591492,
-0.0843053013086319,
-0.07616481930017471,
-0.11324580013751984,
0.24999526143074036,
0.0297510027885437,
-0.017778368666768074,
0.03250480443239212,
-0.08927163481712341,
0.0005092322244308889,
0.14320506155490875,
0.11579438298940659,
-0.0306926891207695,
-0.004100461024791002,
0.05184006690979004,
-0.004134908318519592,
-0.006514808163046837,
0.10370584577322006,
0.08888577669858932,
0.08184853196144104,
-0.06909548491239548,
-0.018575090914964676,
-0.01805807650089264,
-0.05868532136082649,
-0.07659177482128143,
0.06616585701704025,
0.031086305156350136,
0.0027098257560282946,
-0.048470187932252884,
0.07718978077173233,
-0.10289303213357925,
-0.14196720719337463,
0.07735931873321533,
-0.2124083787202835,
-0.18090341985225677,
-0.03651938959956169,
0.06281912326812744,
0.04465538263320923,
0.07072082906961441,
0.006455456838011742,
-0.024913765490055084,
0.11918127536773682,
0.009047036990523338,
-0.010816552676260471,
-0.06862545758485794,
0.03133498132228851,
-0.09069392830133438,
0.18863876163959503,
-0.040456052869558334,
0.050668682903051376,
0.15090739727020264,
0.02236046828329563,
-0.10948770493268967,
0.020854098722338676,
0.10355547070503235,
-0.12360504269599915,
0.03231407701969147,
0.14601865410804749,
-0.012111049145460129,
0.0959969162940979,
0.06610246747732162,
-0.06336525827646255,
0.01795933209359646,
0.036720145493745804,
0.004386885091662407,
-0.07481423765420914,
-0.01278655230998993,
-0.026244647800922394,
0.1389477550983429,
0.23222684860229492,
-0.0650138184428215,
-0.01003304310142994,
-0.04772292450070381,
0.008287481032311916,
0.02345329523086548,
0.08692443370819092,
-0.05090619996190071,
-0.2243911325931549,
0.05851515382528305,
0.015378647483885288,
0.07533810287714005,
-0.15943613648414612,
-0.06579636037349701,
0.05578919127583504,
-0.04534696042537689,
-0.0779937133193016,
0.14344899356365204,
0.04172758013010025,
0.06073770299553871,
-0.043004803359508514,
-0.10075004398822784,
-0.05140009894967079,
0.16647829115390778,
-0.1684628129005432,
-0.07872005552053452
] |
null | null | null | **Como traducir con este Modelo**
+ Instalar [Python 3.9](https://www.python.org/downloads/release/python-390/) + ctranslate 2 + subword-nmt
```bash
pip install ctranslate2~=3.20.0
```
```bash
pip install subword-nmt
```
+ procesar texto con BPE:
```bash
subword-nmt apply-bpe -c gl-detok10k.code < input_file.txt > input_file_bpe.txt
```
+ Traducir un input_text utilizando NOS-MT-eu-gl:
```bash
python3 trans_ct2.py eu_gl.ct2_10k input_file_bpe.txt >output_file_bpe.txt
```
+ DeBPEzar output txt:
```bash
cat out_test_bpe.txt | sed "s/@@ //g" > output_file.txt
```
| {"license": "mit"} | null | proxectonos/Nos_MT-OpenNMT-eu-gl | [
"license:mit",
"region:us"
] | 2024-02-07T08:12:49+00:00 | [] | [] | TAGS
#license-mit #region-us
| Como traducir con este Modelo
+ Instalar Python 3.9 + ctranslate 2 + subword-nmt
+ procesar texto con BPE:
+ Traducir un input_text utilizando NOS-MT-eu-gl:
+ DeBPEzar output txt:
| [] | [
"TAGS\n#license-mit #region-us \n"
] | [
11
] | [
"passage: TAGS\n#license-mit #region-us \n"
] | [
0.026221778243780136,
-0.033018264919519424,
-0.008281232789158821,
-0.05295303836464882,
0.052470896393060684,
0.06768012046813965,
0.1598525494337082,
0.04655371606349945,
0.23683255910873413,
-0.05407243221998215,
0.11752297729253769,
0.08923697471618652,
0.004284696187824011,
-0.0009730930323712528,
0.014216204173862934,
-0.17134642601013184,
0.04864625632762909,
-0.02878100797533989,
0.08764812350273132,
0.032233644276857376,
-0.006205103360116482,
-0.03845774009823799,
-0.0022142508532851934,
-0.03178790956735611,
-0.057939812541007996,
0.03869890421628952,
0.045729056000709534,
-0.02754949778318405,
0.14189864695072174,
-0.021783310920000076,
0.13335508108139038,
0.046146418899297714,
-0.011738095432519913,
-0.2486042082309723,
0.008575023151934147,
-0.07252951711416245,
-0.11333522200584412,
0.016201216727495193,
0.035761721432209015,
-0.010069100186228752,
0.032174937427043915,
0.11049123108386993,
-0.011680051684379578,
0.06288356333971024,
-0.2015703022480011,
-0.20486389100551605,
-0.07508610188961029,
-0.07555478066205978,
0.0589042492210865,
0.030872387811541557,
0.05628744140267372,
0.1426718831062317,
-0.18022038042545319,
-0.0018841808196157217,
0.04129622131586075,
-0.3510737717151642,
0.09011197835206985,
0.19666501879692078,
0.06407395005226135,
0.07872317731380463,
-0.04774639382958412,
0.06726468354463577,
0.07745297998189926,
-0.02402484230697155,
-0.10679105669260025,
-0.06142130121588707,
0.040939174592494965,
0.15604156255722046,
-0.03852643445134163,
-0.10356393456459045,
0.2591084837913513,
-0.023262828588485718,
-0.04234466329216957,
0.08201269060373306,
-0.02980397455394268,
-0.040379155427217484,
0.04404358193278313,
0.044016025960445404,
0.036236923187971115,
0.182089164853096,
0.1260262131690979,
-0.03375067934393883,
-0.16269677877426147,
-0.030629513785243034,
-0.2528207004070282,
0.07418664544820786,
-0.003647059667855501,
0.10666298121213913,
-0.20037521421909332,
0.03286786004900932,
-0.15483668446540833,
-0.009493621066212654,
-0.02952384203672409,
-0.059835705906152725,
0.05229754373431206,
-0.0237403754144907,
-0.04600388556718826,
0.07238677144050598,
0.08390641957521439,
0.2046167105436325,
0.023024363443255424,
0.016697337850928307,
-0.10405295342206955,
0.15052515268325806,
0.019140364602208138,
0.024860305711627007,
0.179348424077034,
0.07677878439426422,
-0.04891882464289665,
-0.2251969277858734,
0.027894439175724983,
-0.03671982139348984,
-0.1441805064678192,
0.015881337225437164,
-0.1542915552854538,
0.1736440360546112,
-0.04078168794512749,
-0.06919530034065247,
-0.08578147739171982,
0.09790384024381638,
0.07768166810274124,
-0.021921472623944283,
-0.023105677217245102,
-0.01381723117083311,
0.03522264584898949,
-0.048196230083703995,
-0.11687057465314865,
0.018241960555315018,
0.11869648098945618,
0.12573401629924774,
-0.1483907401561737,
-0.008189842104911804,
-0.017200417816638947,
0.019065292552113533,
0.09696817398071289,
-0.112403005361557,
0.028845038264989853,
-0.09672309458255768,
-0.13033071160316467,
0.036653537303209305,
0.017736904323101044,
-0.019008556380867958,
0.1340927630662918,
0.061849117279052734,
0.056560322642326355,
-0.011025321669876575,
-0.07250872999429703,
-0.14035539329051971,
-0.08679798245429993,
0.1058693379163742,
-0.046787332743406296,
0.010320915840566158,
-0.24556252360343933,
-0.014234079979360104,
-0.14995723962783813,
0.059662189334630966,
-0.0037668521981686354,
-0.08819212019443512,
-0.07740068435668945,
0.21408265829086304,
0.0018596589798107743,
0.04301392287015915,
-0.1078512966632843,
0.054903753101825714,
-0.06764797121286392,
0.10065380483865738,
-0.12895582616329193,
-0.06441528350114822,
0.1613781899213791,
-0.13135331869125366,
-0.14002031087875366,
0.0033312994055449963,
-0.009472889825701714,
0.12053907662630081,
0.0802001804113388,
0.44566696882247925,
-0.058881040662527084,
-0.16201181709766388,
0.1270403116941452,
0.17969723045825958,
-0.13685379922389984,
-0.25928929448127747,
0.12393020838499069,
-0.1636963188648224,
-0.16647985577583313,
0.0040023741312325,
-0.006962866988033056,
0.08049977570772171,
-0.03446655720472336,
-0.056274134665727615,
0.042339932173490524,
0.024350708350539207,
0.029094615951180458,
0.01740112341940403,
0.07037191838026047,
-0.1023021712899208,
0.08444856107234955,
0.058610700070858,
-0.014111426658928394,
0.15077349543571472,
0.011494536884129047,
-0.05393160134553909,
0.014761670492589474,
0.044013332575559616,
-0.015627963468432426,
-0.05899091437458992,
-0.09661509096622467,
0.019826244562864304,
-0.031149597838521004,
0.08229395002126694,
0.1699674129486084,
0.023824702948331833,
-0.02797185815870762,
0.028922779485583305,
0.028606392443180084,
0.1009954959154129,
0.06960704177618027,
0.03099375218153,
-0.04839283227920532,
0.04952205345034599,
-0.0417071171104908,
-0.11430390179157257,
-0.004862460307776928,
-0.011735930107533932,
0.11975742131471634,
-0.08906009048223495,
-0.01223952230066061,
0.05951591953635216,
-0.04513183981180191,
0.0019881438929587603,
0.0428374819457531,
0.0035966038703918457,
0.1388600617647171,
0.004440935328602791,
-0.04352007433772087,
0.17440910637378693,
-0.05288633331656456,
0.15533447265625,
0.1715822070837021,
-0.07049662619829178,
0.015605369582772255,
-0.1273636519908905,
0.003230511210858822,
-0.014480113983154297,
0.05292887985706329,
-0.05400136485695839,
-0.05201306566596031,
-0.01274962443858385,
0.014292534440755844,
-0.03134604170918465,
0.01711403578519821,
-0.06057267636060715,
-0.08167021721601486,
-0.10849859565496445,
0.018649224191904068,
0.20683221518993378,
-0.22544461488723755,
0.1609548032283783,
0.40251004695892334,
0.15190774202346802,
0.21155193448066711,
-0.12478897720575333,
-0.002471078187227249,
-0.06630261242389679,
0.026115071028470993,
-0.024814706295728683,
0.13782677054405212,
-0.13174867630004883,
-0.01413064356893301,
0.03880728408694267,
0.0454997681081295,
0.0661163181066513,
-0.17195898294448853,
-0.15260353684425354,
-0.0034879595041275024,
-0.020591814070940018,
-0.1749730259180069,
0.04874620959162712,
-0.07595308125019073,
0.02181261032819748,
0.018216799944639206,
-0.10832522064447403,
0.16837291419506073,
-0.033566512167453766,
-0.06695768237113953,
0.052613962441682816,
-0.20581911504268646,
-0.07900715619325638,
-0.17772749066352844,
-0.18375012278556824,
0.06050071492791176,
0.05760138854384422,
0.07903145253658295,
-0.05951719731092453,
-0.01922747679054737,
0.061719246208667755,
-0.009363299235701561,
-0.13802112638950348,
-0.04235544428229332,
-0.06993678212165833,
0.08744155615568161,
-0.09474305808544159,
-0.07518411427736282,
-0.07833878695964813,
-0.046996138989925385,
-0.020961694419384003,
0.08125963062047958,
-0.1039251759648323,
0.08903530240058899,
0.1493726521730423,
0.03651920333504677,
0.05440247058868408,
-0.08271230012178421,
0.12693379819393158,
-0.037743739783763885,
-0.09459595382213593,
0.07307634502649307,
0.004350725095719099,
0.04920351505279541,
0.24039287865161896,
0.08962162584066391,
-0.10578162968158722,
-0.01780811697244644,
-0.0968487411737442,
-0.16405464708805084,
-0.2553846538066864,
-0.06823288649320602,
-0.08744750916957855,
0.14417944848537445,
0.014636521227657795,
0.10712126642465591,
0.14313316345214844,
0.01343101728707552,
0.10255914181470871,
-0.08983208239078522,
-0.018939344212412834,
0.031209396198391914,
0.2135104089975357,
-0.05208220332860947,
0.00838248711079359,
-0.13684824109077454,
-0.0256142970174551,
0.14601100981235504,
0.13798639178276062,
0.14503207802772522,
0.31421369314193726,
0.15292863547801971,
0.13410434126853943,
0.13474710285663605,
0.12333164364099503,
0.07403261214494705,
0.03444362059235573,
-0.015304201282560825,
-0.06035377085208893,
-0.003846159903332591,
0.02816268615424633,
0.05421729013323784,
0.06724072247743607,
-0.22906480729579926,
0.041139665991067886,
-0.2661744952201843,
0.03544611483812332,
-0.0854712724685669,
0.1161833181977272,
-0.028890252113342285,
0.11051984131336212,
0.11386284977197647,
0.05553818494081497,
-0.023278791457414627,
0.16036942601203918,
0.032686375081539154,
-0.07703183591365814,
0.020292721688747406,
0.024695809930562973,
0.06633034348487854,
0.08606193959712982,
0.09550496190786362,
-0.020778406411409378,
-0.1831783503293991,
0.025963841006159782,
0.12212833017110825,
-0.20747940242290497,
0.289523184299469,
0.013651901856064796,
-0.0743619054555893,
-0.01690039224922657,
-0.06958060711622238,
0.008433517068624496,
0.12829731404781342,
0.10406835377216339,
0.05508929491043091,
-0.2613787055015564,
-0.13299626111984253,
0.046764206141233444,
-0.00873907096683979,
0.11356569826602936,
-0.0052223424427211285,
-0.14201195538043976,
-0.06640999764204025,
0.05814211815595627,
-0.006591420155018568,
0.13023322820663452,
-0.018290361389517784,
-0.08173255622386932,
-0.010230090469121933,
0.055564697831869125,
-0.001312803477048874,
-0.04580084979534149,
0.07523149996995926,
0.009008137509226799,
0.02259289287030697,
-0.08178020268678665,
0.03887253627181053,
-0.08071476966142654,
-0.25375792384147644,
0.019298138096928596,
-0.04987313598394394,
0.004092312417924404,
-0.04684043675661087,
-0.15448936820030212,
-0.1129264086484909,
-0.15445278584957123,
0.13100723922252655,
-0.03675999864935875,
0.091565802693367,
-0.0817658007144928,
0.13736046850681305,
-0.08521489799022675,
0.05375019088387489,
0.00614814180880785,
0.03918716683983803,
-0.017955513671040535,
-0.1031481996178627,
0.09334362298250198,
-0.1874227225780487,
0.023863423615694046,
0.010427716188132763,
-0.056847453117370605,
-0.01354232057929039,
0.03918023407459259,
-0.08763083070516586,
0.21879427134990692,
0.3331502079963684,
-0.011948764324188232,
0.22546616196632385,
0.35863226652145386,
-0.13763751089572906,
-0.23258967697620392,
-0.1205512136220932,
-0.3263251483440399,
-0.09005610644817352,
0.17321562767028809,
-0.18057219684123993,
0.04850830137729645,
0.16150830686092377,
-0.10868281871080399,
0.22499866783618927,
-0.22723928093910217,
-0.04793389141559601,
0.1823979914188385,
-0.038322996348142624,
0.4527989625930786,
-0.1144307404756546,
-0.1784561723470688,
-0.03637253865599632,
-0.16285361349582672,
0.12426037341356277,
-0.026553882285952568,
0.06700495630502701,
0.02416347898542881,
-0.011372359469532967,
-0.009014161303639412,
-0.04529716446995735,
0.2216065675020218,
0.0522729866206646,
0.10468899458646774,
-0.09159468114376068,
-0.17199653387069702,
0.1907423883676529,
-0.0004908236442133784,
-0.003372655250132084,
-0.05411549657583237,
-0.04850282520055771,
-0.06871756166219711,
0.033092137426137924,
-0.0334564633667469,
0.06195882335305214,
0.03364093229174614,
-0.11903523653745651,
-0.10248823463916779,
0.034111104905605316,
-0.13155671954154968,
-0.054850947111845016,
0.26421889662742615,
-0.02080743946135044,
0.09609334170818329,
0.04959092289209366,
-0.05474294349551201,
-0.13538943231105804,
0.005736751481890678,
-0.07534020394086838,
-0.05711410939693451,
0.06573604047298431,
-0.11453206837177277,
-0.024341827258467674,
0.1293732225894928,
-0.029497180134058,
0.09674722701311111,
0.08061115443706512,
-0.07585363835096359,
0.02032829262316227,
0.15617427229881287,
-0.07247176766395569,
-0.10849180817604065,
0.04999847710132599,
0.04640531167387962,
0.17256882786750793,
0.004101871978491545,
0.02018604800105095,
0.08726977556943893,
0.045959215611219406,
-0.007486662827432156,
0.007311292923986912,
-0.11321697384119034,
-0.04241771996021271,
0.0387241393327713,
-0.005273692775517702,
-0.10946331918239594,
0.16008898615837097,
0.056837860494852066,
0.004653505515307188,
-0.06027700752019882,
0.09720424562692642,
-0.06709636747837067,
-0.07046061009168625,
-0.1753035932779312,
0.018511172384023666,
-0.12734080851078033,
-0.09874535351991653,
0.06846235692501068,
-0.09371624886989594,
-0.04084605351090431,
0.08152704685926437,
0.046927981078624725,
0.14401860535144806,
-0.006597559433430433,
-0.023080874234437943,
0.149825319647789,
-0.0884878933429718,
-0.2241756170988083,
0.01969664730131626,
-0.04083063453435898,
-0.07065816223621368,
-0.0007070365245454013,
0.06069544702768326,
-0.0663156732916832,
-0.11958606541156769,
-0.20477768778800964,
0.10412076860666275,
-0.12043121457099915,
-0.03954985365271568,
-0.1041841059923172,
-0.053260523825883865,
0.07891252636909485,
-0.02613759972155094,
-0.04122013971209526,
-0.047595683485269547,
-0.16630595922470093,
0.054254453629255295,
0.07140932232141495,
0.11125344783067703,
-0.0759999230504036,
-0.018354382365942,
0.1398727148771286,
0.048581548035144806,
0.08479110151529312,
0.07578440010547638,
0.026255371049046516,
0.16728560626506805,
-0.1708206981420517,
-0.0542997270822525,
0.1068294569849968,
-0.026716172695159912,
0.01994573324918747,
0.10631280392408371,
-0.04839588701725006,
0.07042654603719711,
-0.05095988139510155,
0.05859163776040077,
-0.15704534947872162,
-0.13073866069316864,
-0.04184387996792793,
0.023728877305984497,
-0.2260182797908783,
0.015071595087647438,
-0.1769561767578125,
0.19692228734493256,
-0.024228032678365707,
0.11490963399410248,
0.08052190393209457,
0.02052290178835392,
0.03539382666349411,
-0.006019921973347664,
0.00946811307221651,
-0.10524865239858627,
-0.05784677714109421,
-0.07560300827026367,
-0.1168874129652977,
-0.009665017947554588,
0.36614301800727844,
0.02430291846394539,
-0.19682736694812775,
0.051222387701272964,
0.18285293877124786,
0.023639049381017685,
-0.0073763905093073845,
0.26180747151374817,
0.08150359988212585,
-0.023175053298473358,
-0.1782374382019043,
0.0396091528236866,
-0.08699734508991241,
-0.15269799530506134,
0.11385007947683334,
0.09347525984048843,
0.05813581123948097,
0.022930078208446503,
0.10404518246650696,
-0.035940010100603104,
-0.05509711429476738,
-0.13301853835582733,
0.13368983566761017,
-0.001790675800293684,
0.0193882267922163,
0.0897885113954544,
0.19249756634235382,
-0.045275162905454636,
0.05437124893069267,
-0.07336640357971191,
-0.001598604372702539,
-0.15740543603897095,
-0.13358698785305023,
0.06194563955068588,
-0.08269550651311874,
0.06342913210391998,
0.050261519849300385,
0.04341990500688553,
0.31786394119262695,
0.039095040410757065,
-0.046439893543720245,
0.003166865324601531,
-0.14845187962055206,
-0.08075450360774994,
-0.06024569645524025,
-0.03110554814338684,
0.028620192781090736,
-0.13928957283496857,
-0.09898591786623001,
-0.06917677819728851,
-0.130235955119133,
-0.06539803743362427,
0.025270747020840645,
0.014251931570470333,
-0.053083837032318115,
-0.17625881731510162,
-0.04808593541383743,
-0.06644169986248016,
0.10105955600738525,
-0.08462738990783691,
0.1516820639371872,
0.0022449472453445196,
0.030281953513622284,
0.07627002149820328,
0.09585131704807281,
0.018900424242019653,
-0.06975197046995163,
0.05599058046936989,
0.12436293810606003,
0.01323844213038683,
0.1259988248348236,
-0.06034265458583832,
-0.019420607015490532,
-0.014145253226161003,
0.14038437604904175,
0.304447740316391,
-0.01856905221939087,
-0.013814439997076988,
-0.022110093384981155,
0.021388787776231766,
0.10893569141626358,
0.19800719618797302,
-0.03437356278300285,
0.2551359534263611,
-0.058974795043468475,
0.0756678432226181,
-0.013180435635149479,
-0.005362013820558786,
-0.053146667778491974,
0.06074550002813339,
0.06268858164548874,
-0.06877048313617706,
-0.10191375762224197,
0.15178529918193817,
-0.14985080063343048,
0.13306055963039398,
0.14678068459033966,
-0.06057753041386604,
0.03797250986099243,
0.0007459368789568543,
0.19896264374256134,
-0.03570213168859482,
0.0984780564904213,
-0.10653308779001236,
-0.10261140763759613,
-0.14764924347400665,
0.037690844386816025,
-0.36797797679901123,
-0.1756322830915451,
0.11731542646884918,
0.14115898311138153,
0.1759258657693863,
-0.012341637164354324,
0.056479312479496,
0.0033020609989762306,
0.08296097069978714,
-0.04232487455010414,
0.1519634872674942,
0.0612073615193367,
-0.017103128135204315,
-0.15296664834022522,
-0.20328094065189362,
-0.0012039330322295427,
-0.058561209589242935,
0.055583830922842026,
-0.02269243635237217,
0.025347469374537468,
0.07746459543704987,
-0.06768939644098282,
-0.029180381447076797,
-0.02352982573211193,
-0.13262848556041718,
0.052229251712560654,
-0.04354005306959152,
0.0320255309343338,
-0.03958037868142128,
-0.022394726052880287,
-0.039987675845623016,
0.10721533745527267,
-0.22402705252170563,
-0.08517231047153473,
0.1422796994447708,
-0.03421911224722862,
0.1542559564113617,
-0.02848726324737072,
-0.12159585952758789,
-0.024955326691269875,
-0.06977712363004684,
0.10887379199266434,
-0.1419300138950348,
0.038592495024204254,
0.13747453689575195,
0.008710617199540138,
0.031119761988520622,
-0.2533661723136902,
0.050644006580114365,
-0.03556957095861435,
-0.016733208671212196,
-0.057031940668821335
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] | {"language": ["en"], "library_name": "transformers", "tags": ["physics"], "metrics": ["accuracy"], "pipeline_tag": "text-classification"} | text-classification | dylansheils0241/Quantum-Balanced-GPT2-Experimental-Theoretical-Classifier-Arxiv-V3 | [
"transformers",
"safetensors",
"gpt2",
"text-classification",
"physics",
"en",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T08:13:53+00:00 | [
"1910.09700"
] | [
"en"
] | TAGS
#transformers #safetensors #gpt2 #text-classification #physics #en #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #gpt2 #text-classification #physics #en #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
63,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #gpt2 #text-classification #physics #en #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.0454922579228878,
0.1892290860414505,
-0.005267181899398565,
0.023076441138982773,
0.10117173194885254,
0.0077434806153178215,
0.053622350096702576,
0.11688259989023209,
-0.049281761050224304,
0.1273360401391983,
0.04230627045035362,
0.10744371265172958,
0.11827985942363739,
0.14195206761360168,
0.00352189177647233,
-0.22223159670829773,
0.044058289378881454,
-0.1097642034292221,
-0.01773863099515438,
0.12165118753910065,
0.14420785009860992,
-0.10128498077392578,
0.07021533697843552,
-0.03262169286608696,
-0.013463100418448448,
-0.03671392425894737,
-0.0593554861843586,
-0.04263821616768837,
0.040572457015514374,
0.054630160331726074,
0.06357629597187042,
0.0005440081004053354,
0.08636020869016647,
-0.2783202528953552,
0.019143665209412575,
0.06848175823688507,
-0.002689888933673501,
0.06751231104135513,
0.07191142439842224,
-0.06420893967151642,
0.10572454333305359,
-0.05525720492005348,
0.1410498470067978,
0.08407848328351974,
-0.09131639450788498,
-0.18183988332748413,
-0.09535790979862213,
0.10929666459560394,
0.17302729189395905,
0.04880116134881973,
-0.025352975353598595,
0.10172717273235321,
-0.08491719514131546,
0.01533019170165062,
0.05563700571656227,
-0.09077669680118561,
-0.05872148275375366,
0.06778686493635178,
0.0901692658662796,
0.05638265609741211,
-0.1290295124053955,
-0.033455073833465576,
0.006060066167265177,
0.014386994764208794,
0.07489969581365585,
0.023653669282794,
0.14651624858379364,
0.03957021236419678,
-0.13190419971942902,
-0.050189293920993805,
0.10258059203624725,
0.039069969207048416,
-0.043305084109306335,
-0.24386411905288696,
-0.025369178503751755,
-0.029489802196621895,
-0.03168201446533203,
-0.046388521790504456,
0.04390523210167885,
-0.00482727587223053,
0.08556914329528809,
-0.0061739063821733,
-0.0734160766005516,
-0.037340421229600906,
0.07216035574674606,
0.0673036277294159,
0.028943028301000595,
-0.017821909859776497,
0.023992950096726418,
0.11276077479124069,
0.09425251930952072,
-0.11540529131889343,
-0.05989375337958336,
-0.06737136840820312,
-0.07234217971563339,
-0.04444768279790878,
0.03710659220814705,
0.01569688692688942,
0.06569195538759232,
0.2581775188446045,
0.014665181748569012,
0.059788089245557785,
0.0353231355547905,
0.006818118039518595,
0.05111116170883179,
0.10741052776575089,
-0.05909406766295433,
-0.11346599459648132,
-0.023242302238941193,
0.08917469531297684,
0.01908665895462036,
-0.03678778558969498,
-0.0435459166765213,
0.061442095786333084,
0.04259845241904259,
0.1099630668759346,
0.09463895857334137,
0.018771236762404442,
-0.07234785705804825,
-0.061810385435819626,
0.2026224136352539,
-0.15369552373886108,
0.040934253484010696,
0.04223572090268135,
-0.03732837736606598,
-0.026609228923916817,
0.011604798026382923,
0.024884888902306557,
-0.028492165729403496,
0.09093567728996277,
-0.058359935879707336,
-0.043780211359262466,
-0.11050130426883698,
-0.02607148326933384,
0.041717205196619034,
0.013927704654633999,
-0.03360515832901001,
-0.03675388544797897,
-0.06977998465299606,
-0.08461851626634598,
0.0885913148522377,
-0.06979253888130188,
-0.05868910998106003,
-0.02374347299337387,
-0.08104995638132095,
0.02571520209312439,
0.022042836993932724,
0.07542433589696884,
-0.02658507227897644,
0.05562540516257286,
-0.05317474901676178,
0.056244947016239166,
0.10653666406869888,
0.03383060172200203,
-0.059717658907175064,
0.05819431319832802,
-0.2371143102645874,
0.08229552954435349,
-0.07159779965877533,
0.060478564351797104,
-0.15544624626636505,
-0.02904750593006611,
0.0370447039604187,
-0.00012050439545419067,
-0.0036397746298462152,
0.13608399033546448,
-0.20679078996181488,
-0.024356475099921227,
0.16739897429943085,
-0.09645804762840271,
-0.07371331006288528,
0.048024434596300125,
-0.04922673478722572,
0.10172245651483536,
0.030922457575798035,
0.00033895415253937244,
0.06861858069896698,
-0.11190637946128845,
-0.016445549204945564,
-0.05696813389658928,
-0.02647355943918228,
0.1390431970357895,
0.07771971076726913,
-0.0771675780415535,
0.058569807559251785,
0.021676812320947647,
-0.026402225717902184,
-0.06422454863786697,
-0.019683588296175003,
-0.10205245018005371,
0.014703531749546528,
-0.0698927640914917,
0.015332268550992012,
-0.014735503122210503,
-0.09082003682851791,
-0.029301082715392113,
-0.16758349537849426,
-0.031187167391180992,
0.08277160674333572,
-0.0031366399489343166,
-0.014441276900470257,
-0.10866646468639374,
0.022468240931630135,
0.03254542499780655,
0.0007284895400516689,
-0.1306690126657486,
-0.04408187419176102,
0.034751396626234055,
-0.15020820498466492,
0.03625624626874924,
-0.07489804923534393,
0.05473422631621361,
0.01929926685988903,
-0.030721891671419144,
-0.02874412201344967,
0.020097021013498306,
0.00757033983245492,
-0.017352037131786346,
-0.23862533271312714,
-0.026031702756881714,
-0.027882497757673264,
0.16365355253219604,
-0.21298782527446747,
0.03490815684199333,
0.08089017122983932,
0.1532241404056549,
0.006470671389251947,
-0.055513519793748856,
0.017142554745078087,
-0.07452475279569626,
-0.022995643317699432,
-0.05783756822347641,
0.0007704336894676089,
-0.018449876457452774,
-0.041721027344465256,
0.0215263981372118,
-0.1709861308336258,
-0.04279936105012894,
0.0963878259062767,
0.04850459098815918,
-0.1301574409008026,
-0.02655213698744774,
-0.03984624892473221,
-0.05198091268539429,
-0.044672150164842606,
-0.05470692738890648,
0.10276113450527191,
0.06009262427687645,
0.036733414977788925,
-0.05964718386530876,
-0.08222147822380066,
-0.004059364087879658,
-0.014832045882940292,
-0.023698439821600914,
0.09377166628837585,
0.08203411102294922,
-0.12958833575248718,
0.0927903950214386,
0.08085138350725174,
0.07611905038356781,
0.08878163248300552,
-0.017963899299502373,
-0.07635803520679474,
-0.03698829188942909,
0.036995161324739456,
0.020335271954536438,
0.12776018679141998,
-0.03758808225393295,
0.04251377657055855,
0.03941556438803673,
-0.02768935076892376,
0.018462009727954865,
-0.0811423510313034,
0.029668718576431274,
0.021729903295636177,
-0.014222783036530018,
0.0519867017865181,
-0.04036044701933861,
0.0159895159304142,
0.08736182004213333,
0.059590280055999756,
0.046751588582992554,
0.016359178349375725,
-0.05113328620791435,
-0.1142188087105751,
0.15896457433700562,
-0.11944134533405304,
-0.21303002536296844,
-0.13420626521110535,
0.016609150916337967,
0.027823029085993767,
-0.015470955520868301,
0.002491585910320282,
-0.06520005315542221,
-0.10971669852733612,
-0.08959738910198212,
0.003818272612988949,
0.05380045622587204,
-0.0811196118593216,
-0.06572441011667252,
0.048350267112255096,
0.04122333228588104,
-0.14260631799697876,
0.02080303616821766,
0.04214956983923912,
-0.09044978767633438,
-0.01145887840539217,
0.08635754138231277,
0.0769895538687706,
0.18610262870788574,
0.017064902931451797,
-0.018678830936551094,
0.03175484016537666,
0.21870136260986328,
-0.1387357860803604,
0.11223021149635315,
0.13058973848819733,
-0.08224785327911377,
0.08210122585296631,
0.20566321909427643,
0.044577840715646744,
-0.09425488859415054,
0.03326784446835518,
0.032479673624038696,
-0.024472102522850037,
-0.2386406511068344,
-0.06845381110906601,
0.001011490821838379,
-0.06857919692993164,
0.08203253895044327,
0.09691119194030762,
0.07842933386564255,
0.022030703723430634,
-0.0917103961110115,
-0.0919734314084053,
0.05734914541244507,
0.11010482162237167,
0.003062051720917225,
-0.005152191035449505,
0.08908042311668396,
-0.03344396874308586,
0.0154754389077425,
0.0881464034318924,
0.0019678978715091944,
0.15960493683815002,
0.04489469528198242,
0.1688421666622162,
0.08134543895721436,
0.07467558979988098,
0.0013862692285329103,
0.009403538890182972,
0.016823500394821167,
0.040771570056676865,
-0.003894505091011524,
-0.08311600238084793,
-0.025278564542531967,
0.11370491981506348,
0.06369880586862564,
0.0194417554885149,
0.00901445560157299,
-0.047142382711172104,
0.08514023572206497,
0.17779837548732758,
-0.00231233611702919,
-0.18142074346542358,
-0.06185448542237282,
0.07624103128910065,
-0.09856211394071579,
-0.10322526097297668,
-0.0074151307344436646,
0.013185259886085987,
-0.16772492229938507,
0.03721456974744797,
-0.024712707847356796,
0.10490449517965317,
-0.13263392448425293,
-0.020808758214116096,
0.08600576967000961,
0.06833356618881226,
-0.0031068536918610334,
0.05905979871749878,
-0.17244522273540497,
0.09498779475688934,
0.012149130925536156,
0.07152468711137772,
-0.09588398039340973,
0.09354560822248459,
-0.005466605070978403,
-0.02618931420147419,
0.14424623548984528,
-0.002266506664454937,
-0.07356798648834229,
-0.06449858099222183,
-0.09199952334165573,
-0.012544767931103706,
0.1272704303264618,
-0.13194456696510315,
0.0912502184510231,
-0.03380442410707474,
-0.036114346235990524,
-0.011171766556799412,
-0.09125872701406479,
-0.11271911859512329,
-0.17849315702915192,
0.060609254986047745,
-0.1380886286497116,
0.037936948239803314,
-0.1081671193242073,
-0.02932816743850708,
-0.03241882473230362,
0.1853189468383789,
-0.23679997026920319,
-0.07359582930803299,
-0.14321380853652954,
-0.09519625455141068,
0.12957772612571716,
-0.048767514526844025,
0.09029202163219452,
-0.013483068905770779,
0.15911369025707245,
0.019319884479045868,
-0.019532188773155212,
0.08665427565574646,
-0.08434808999300003,
-0.19747509062290192,
-0.06885388493537903,
0.16540810465812683,
0.12358696013689041,
0.03463016077876091,
-0.0038832996506243944,
0.03858266398310661,
-0.020188724622130394,
-0.12180843204259872,
0.02317660115659237,
0.15238697826862335,
0.06875092536211014,
0.007192579563707113,
-0.018620964139699936,
-0.10698071867227554,
-0.07437437772750854,
-0.024687660858035088,
0.029182182624936104,
0.1695101410150528,
-0.07483144104480743,
0.17535264790058136,
0.13886579871177673,
-0.05660459026694298,
-0.2033400684595108,
0.005214088596403599,
0.026235518977046013,
-0.012564047239720821,
0.011621620506048203,
-0.19036664068698883,
0.09390343725681305,
-0.004040845204144716,
-0.05684201791882515,
0.11082622408866882,
-0.15271665155887604,
-0.13844281435012817,
0.07983817160129547,
0.054161109030246735,
-0.18839919567108154,
-0.13631866872310638,
-0.09855357557535172,
-0.03675471246242523,
-0.16187281906604767,
0.09608087688684464,
0.021142609417438507,
0.010354102589190006,
0.027322333306074142,
0.01351181324571371,
0.025093428790569305,
-0.05160268396139145,
0.17773623764514923,
-0.020241370424628258,
0.023158445954322815,
-0.09666916728019714,
-0.08445306122303009,
0.015926335006952286,
-0.0515141598880291,
0.07591541856527328,
-0.012556138448417187,
0.01187922153621912,
-0.10285384207963943,
-0.03707551211118698,
-0.042909085750579834,
0.014773706905543804,
-0.09829625487327576,
-0.08840229362249374,
-0.05165104195475578,
0.09804599732160568,
0.09530165046453476,
-0.022125273942947388,
-0.02740321308374405,
-0.07544717192649841,
0.05670405551791191,
0.2019607424736023,
0.1906883716583252,
0.043060071766376495,
-0.06546362489461899,
0.00039793530595488846,
-0.015163831412792206,
0.04418615624308586,
-0.20139308273792267,
0.0597548671066761,
0.056555576622486115,
0.022511905059218407,
0.10037171095609665,
-0.016469812020659447,
-0.15572866797447205,
-0.0770362988114357,
0.06771523505449295,
-0.06949129700660706,
-0.20054061710834503,
0.004918920807540417,
0.05531409755349159,
-0.17835736274719238,
-0.03684486448764801,
0.0430159829556942,
-0.004673158284276724,
-0.0415334478020668,
0.022451501339673996,
0.09515895694494247,
0.003960844129323959,
0.0767822265625,
0.07493549585342407,
0.0819842666387558,
-0.09586521238088608,
0.08131501078605652,
0.09765590727329254,
-0.07306497544050217,
0.026293009519577026,
0.10408546775579453,
-0.055695097893476486,
-0.03752893954515457,
0.03591118007898331,
0.08107950538396835,
0.029151301831007004,
-0.04298127442598343,
0.009091568179428577,
-0.09248542040586472,
0.06755181401968002,
0.09707692265510559,
0.02941909246146679,
0.01624402031302452,
0.04245876893401146,
0.05362997204065323,
-0.07527279108762741,
0.12204152345657349,
0.030805140733718872,
0.01844998076558113,
-0.04329174384474754,
-0.04292193427681923,
0.009563887491822243,
-0.028133688494563103,
-0.005042788572609425,
-0.02055315673351288,
-0.08608268201351166,
-0.018558956682682037,
-0.12441198527812958,
-0.006824290379881859,
-0.06286269426345825,
0.016643356531858444,
0.024939946830272675,
-0.03083384968340397,
0.006669266149401665,
0.005942019168287516,
-0.0683986023068428,
-0.06568412482738495,
-0.014517217874526978,
0.09688357263803482,
-0.16726930439472198,
0.027356497943401337,
0.08470899611711502,
-0.11206947267055511,
0.10080697387456894,
0.010145304724574089,
-0.0023009981960058212,
0.02145352214574814,
-0.145093634724617,
0.04009302705526352,
-0.03804326802492142,
0.006585637107491493,
0.01963588409125805,
-0.20789407193660736,
0.00012819538824260235,
-0.04070339351892471,
-0.06983331590890884,
-0.008275420404970646,
-0.027385288849473,
-0.1133166179060936,
0.10505366325378418,
0.0018041511066257954,
-0.08291047066450119,
-0.034454453736543655,
0.03384244069457054,
0.07920656353235245,
-0.030854199081659317,
0.14700515568256378,
-0.012957319617271423,
0.06367551535367966,
-0.16380320489406586,
-0.012435955926775932,
-0.008895769715309143,
0.01886560209095478,
-0.03879137337207794,
-0.006686908658593893,
0.05134914815425873,
-0.014987408183515072,
0.17471328377723694,
-0.03153485059738159,
0.015502654016017914,
0.06436023861169815,
0.049617890268564224,
-0.027531292289495468,
0.09797853231430054,
0.049099430441856384,
0.016220733523368835,
0.00771610252559185,
0.015807701274752617,
-0.04161199554800987,
-0.03627050295472145,
-0.1930110603570938,
0.07477480173110962,
0.18303412199020386,
0.09458079934120178,
-0.02132072113454342,
0.0668141096830368,
-0.104291170835495,
-0.09706129878759384,
0.152155339717865,
-0.036607079207897186,
-0.013433953747153282,
-0.07288673520088196,
0.13096018135547638,
0.14686168730258942,
-0.18457777798175812,
0.06662989407777786,
-0.07185389846563339,
-0.04415108263492584,
-0.10994060337543488,
-0.1937890499830246,
-0.06109384074807167,
-0.050476349890232086,
-0.01619703136384487,
-0.04715639352798462,
0.0668591558933258,
0.05701626092195511,
-0.002298380248248577,
-0.006261112634092569,
0.07253046333789825,
-0.03486889600753784,
0.0008380662184208632,
0.0272334236651659,
0.06141670420765877,
0.008357974700629711,
-0.039442744106054306,
0.016402557492256165,
-0.01572669856250286,
0.049971528351306915,
0.07516368478536606,
0.046334292739629745,
-0.027019204571843147,
0.01750812865793705,
-0.03152864798903465,
-0.10467907041311264,
0.049071840941905975,
-0.03048054128885269,
-0.0700785294175148,
0.15455928444862366,
0.01970783993601799,
0.009909378364682198,
-0.013979905284941196,
0.23960374295711517,
-0.06548106670379639,
-0.09619254618883133,
-0.1433766633272171,
0.07538311183452606,
-0.04318699613213539,
0.05031571909785271,
0.04129001498222351,
-0.11302274465560913,
0.02624162845313549,
0.14381884038448334,
0.1547107696533203,
-0.038064345717430115,
0.022280780598521233,
0.036548905074596405,
0.0075757745653390884,
-0.022315723821520805,
0.03925982490181923,
0.06176180765032768,
0.14892682433128357,
-0.050396211445331573,
0.07794349640607834,
0.0004808782250620425,
-0.08597756177186966,
-0.031207749620079994,
0.12030841410160065,
-0.010073723271489143,
0.016193872317671776,
-0.06076192855834961,
0.11857526004314423,
-0.07239425927400589,
-0.21555854380130768,
0.045980796217918396,
-0.0718245655298233,
-0.13439948856830597,
-0.021382097154855728,
0.0745546817779541,
-0.017455222085118294,
0.023925170302391052,
0.08027905970811844,
-0.07115567475557327,
0.18785375356674194,
0.043465130031108856,
-0.06361889094114304,
-0.05098056048154831,
0.07458217442035675,
-0.07300199568271637,
0.3019518256187439,
0.014535597525537014,
0.04014892503619194,
0.11048824340105057,
-0.016689995303750038,
-0.1411888748407364,
0.017861967906355858,
0.09737598896026611,
-0.10161881893873215,
0.049057722091674805,
0.18413949012756348,
0.002281030174344778,
0.1267751008272171,
0.07595405727624893,
-0.08812038600444794,
0.04453841969370842,
-0.07067979127168655,
-0.07145576924085617,
-0.10368072241544724,
0.10495315492153168,
-0.08820059895515442,
0.14607025682926178,
0.12117290496826172,
-0.053774144500494,
0.011705644428730011,
-0.03668057918548584,
0.048356372863054276,
-0.0036693946458399296,
0.12250278145074844,
0.012049626559019089,
-0.18831618130207062,
0.02737637422978878,
-0.026310449466109276,
0.10337604582309723,
-0.17009514570236206,
-0.08743718266487122,
0.04972841218113899,
0.009872285649180412,
-0.07030795514583588,
0.1251453012228012,
0.05528919771313667,
0.029022296890616417,
-0.04736864194273949,
-0.02693687565624714,
-0.012329909019172192,
0.14075808227062225,
-0.10747740417718887,
-0.002252733102068305
] |
null | null | null |
# OBMBX2-7b
OBMBX2-7b is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
* [paulml/OmniBeagleMBX-v3-7B](https://huggingface.co/paulml/OmniBeagleMBX-v3-7B)
* [alnrg2arg/blockchainlabs_joe_bez_seminar](https://huggingface.co/alnrg2arg/blockchainlabs_joe_bez_seminar)
## 🧩 Configuration
```yaml
slices:
models:
- model: flemmingmiguel/MBX-7B-v3
# no parameters necessary for base model
- model: paulml/OmniBeagleMBX-v3-7B
parameters:
density: 0.5
weight: 0.3
- model: alnrg2arg/blockchainlabs_joe_bez_seminar
parameters:
density: 0.5
weight: 0.5
merge_method: dare_ties
base_model: flemmingmiguel/MBX-7B-v3
parameters:
normalize: true
dtype: float16
```
## 💻 Usage
```python
!pip install -qU transformers accelerate
from transformers import AutoTokenizer
import transformers
import torch
model = "gmonsoon/OBMBX2-7b"
messages = [{"role": "user", "content": "What is a large language model?"}]
tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
"text-generation",
model=model,
torch_dtype=torch.float16,
device_map="auto",
)
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
``` | {"tags": ["merge", "mergekit", "lazymergekit", "paulml/OmniBeagleMBX-v3-7B", "alnrg2arg/blockchainlabs_joe_bez_seminar"], "base_model": ["paulml/OmniBeagleMBX-v3-7B", "alnrg2arg/blockchainlabs_joe_bez_seminar"]} | null | gmonsoon/OBMBX2-7b | [
"merge",
"mergekit",
"lazymergekit",
"paulml/OmniBeagleMBX-v3-7B",
"alnrg2arg/blockchainlabs_joe_bez_seminar",
"base_model:paulml/OmniBeagleMBX-v3-7B",
"base_model:alnrg2arg/blockchainlabs_joe_bez_seminar",
"region:us"
] | 2024-02-07T08:16:24+00:00 | [] | [] | TAGS
#merge #mergekit #lazymergekit #paulml/OmniBeagleMBX-v3-7B #alnrg2arg/blockchainlabs_joe_bez_seminar #base_model-paulml/OmniBeagleMBX-v3-7B #base_model-alnrg2arg/blockchainlabs_joe_bez_seminar #region-us
|
# OBMBX2-7b
OBMBX2-7b is a merge of the following models using LazyMergekit:
* paulml/OmniBeagleMBX-v3-7B
* alnrg2arg/blockchainlabs_joe_bez_seminar
## Configuration
## Usage
| [
"# OBMBX2-7b\n\nOBMBX2-7b is a merge of the following models using LazyMergekit:\n* paulml/OmniBeagleMBX-v3-7B\n* alnrg2arg/blockchainlabs_joe_bez_seminar",
"## Configuration",
"## Usage"
] | [
"TAGS\n#merge #mergekit #lazymergekit #paulml/OmniBeagleMBX-v3-7B #alnrg2arg/blockchainlabs_joe_bez_seminar #base_model-paulml/OmniBeagleMBX-v3-7B #base_model-alnrg2arg/blockchainlabs_joe_bez_seminar #region-us \n",
"# OBMBX2-7b\n\nOBMBX2-7b is a merge of the following models using LazyMergekit:\n* paulml/OmniBeagleMBX-v3-7B\n* alnrg2arg/blockchainlabs_joe_bez_seminar",
"## Configuration",
"## Usage"
] | [
95,
61,
4,
3
] | [
"passage: TAGS\n#merge #mergekit #lazymergekit #paulml/OmniBeagleMBX-v3-7B #alnrg2arg/blockchainlabs_joe_bez_seminar #base_model-paulml/OmniBeagleMBX-v3-7B #base_model-alnrg2arg/blockchainlabs_joe_bez_seminar #region-us \n# OBMBX2-7b\n\nOBMBX2-7b is a merge of the following models using LazyMergekit:\n* paulml/OmniBeagleMBX-v3-7B\n* alnrg2arg/blockchainlabs_joe_bez_seminar## Configuration## Usage"
] | [
-0.05019044131040573,
0.2136191427707672,
-0.0011345063103362918,
0.046477820724248886,
-0.013989681378006935,
0.09481295943260193,
0.22896575927734375,
0.09185250103473663,
0.09109412133693695,
0.017817268148064613,
0.1318928599357605,
0.1248677596449852,
0.0009998355526477098,
0.20565998554229736,
-0.06745585799217224,
-0.20941486954689026,
0.025902248919010162,
0.07765557616949081,
-0.10506777465343475,
0.06873491406440735,
0.13746492564678192,
-0.039519116282463074,
0.13100795447826385,
0.018001804128289223,
-0.11379929631948471,
0.023641075938940048,
-0.05319594219326973,
-0.007696256041526794,
0.05582059919834137,
0.1253802627325058,
-0.0009096864378079772,
0.04204043373465538,
-0.030938295647501945,
-0.11121571809053421,
0.04405837133526802,
-0.02618904784321785,
-0.062447480857372284,
0.07790293544530869,
-0.009840318001806736,
-0.060640860348939896,
0.1302449107170105,
-0.01638878509402275,
0.012902380898594856,
0.07319887727499008,
-0.16735988855361938,
-0.07779533416032791,
-0.07197712361812592,
0.18532361090183258,
-0.010778221301734447,
-0.07645434886217117,
-0.0025666633155196905,
0.17503400146961212,
-0.034098803997039795,
0.06065351143479347,
0.1301775723695755,
-0.349128395318985,
-0.04292864724993706,
0.17228952050209045,
0.06063205748796463,
0.052942778915166855,
0.05534368380904198,
0.0372086800634861,
0.05567708611488342,
-0.022044306620955467,
0.01290847733616829,
-0.10159777849912643,
0.16838328540325165,
-0.027171405032277107,
-0.15714550018310547,
-0.005325114354491234,
0.2693125605583191,
0.11187106370925903,
0.031209692358970642,
-0.0844859704375267,
-0.047979604452848434,
0.2164740264415741,
-0.04238085821270943,
-0.055734194815158844,
0.04111732542514801,
-0.03194993734359741,
-0.017719650641083717,
-0.036183200776576996,
0.03653235360980034,
0.030203688889741898,
0.018681127578020096,
0.08414065092802048,
-0.02810916304588318,
0.057248618453741074,
-0.03741166740655899,
0.054077837616205215,
-0.24458910524845123,
-0.08873821794986725,
0.018624980002641678,
-0.09142651408910751,
0.07189982384443283,
-0.04947556182742119,
-0.015831967815756798,
-0.019065745174884796,
0.06195024400949478,
0.3262866139411926,
-0.009118151850998402,
0.027226191014051437,
0.09052152186632156,
0.060273222625255585,
0.07917247712612152,
-0.007935015484690666,
-0.12552598118782043,
-0.30337899923324585,
0.007989684119820595,
0.0778118371963501,
0.005270806606858969,
0.04175275191664696,
-0.12748727202415466,
-0.02632877044379711,
-0.08084689825773239,
-0.07632818818092346,
0.053830791264772415,
0.06517628580331802,
-0.10254873335361481,
-0.13276901841163635,
0.20238253474235535,
0.023550283163785934,
0.07709362357854843,
-0.02443840727210045,
-0.047873105853796005,
-0.04881257936358452,
0.04983852803707123,
0.07106533646583557,
0.049549881368875504,
0.05472787469625473,
-0.11337676644325256,
-0.08230682462453842,
-0.011399297043681145,
-0.04530356079339981,
0.0248013436794281,
-0.0266116913408041,
0.06981828808784485,
-0.05506448820233345,
-0.19020524621009827,
-0.0034698054660111666,
0.12249883264303207,
-0.03762693703174591,
-0.08742328733205795,
-0.0777190774679184,
0.112180694937706,
-0.06424379348754883,
-0.005762285087257624,
0.13278941810131073,
-0.018683718517422676,
-0.0525229349732399,
0.001704239402897656,
0.04527140036225319,
-0.1580900400876999,
-0.027731522917747498,
-0.03622933849692345,
0.08277804404497147,
-0.12489347159862518,
-0.0019134868634864688,
-0.04027731344103813,
0.034863412380218506,
-0.11302539706230164,
0.02989514172077179,
-0.02800988219678402,
0.02728368528187275,
0.036675188690423965,
0.1158636212348938,
0.09821957349777222,
-0.12380488216876984,
0.16184957325458527,
-0.045960720628499985,
-0.1177816316485405,
0.09234538674354553,
-0.06728208065032959,
0.1289234459400177,
0.058132339268922806,
0.12001212686300278,
0.10028374940156937,
-0.026273563504219055,
-0.047804396599531174,
0.019213847815990448,
0.01357482559978962,
-0.0574471578001976,
0.09855750948190689,
0.04099313169717789,
-0.12016180902719498,
0.0221547894179821,
0.03923231363296509,
0.01862511783838272,
-0.06293944269418716,
-0.059689246118068695,
-0.028156038373708725,
-0.14176496863365173,
0.1019795760512352,
-0.04392654821276665,
0.026912249624729156,
-0.08673153072595596,
-0.1303255259990692,
0.06416498124599457,
0.13309425115585327,
0.015237702056765556,
-0.049618303775787354,
-0.11544034630060196,
0.10713588446378708,
-0.1872606873512268,
-0.002532682614400983,
-0.04462236166000366,
0.11660847812891006,
0.008006006479263306,
-0.030393050983548164,
0.11223646253347397,
0.08113095164299011,
0.0991239920258522,
0.0036133532412350178,
-0.09307744354009628,
0.009106426499783993,
0.08548535406589508,
0.013347705826163292,
-0.03515565022826195,
-0.1673772633075714,
-0.09848197549581528,
-0.0538296177983284,
0.18730151653289795,
-0.10339552164077759,
0.042768169194459915,
-0.05464869737625122,
0.16779401898384094,
0.029624633491039276,
0.0034539466723799706,
0.01253480464220047,
0.010883470997214317,
0.0022089460399001837,
0.030064204707741737,
0.06641196459531784,
-0.005365787073969841,
-0.15318070352077484,
0.03736332058906555,
-0.023646311834454536,
0.08642211556434631,
0.08108960092067719,
0.004661920014768839,
0.02424674853682518,
-0.05752544105052948,
-0.038672640919685364,
-0.06668954342603683,
0.14311020076274872,
-0.03649140149354935,
0.07184787839651108,
-0.012229111045598984,
0.0774703249335289,
-0.07594626396894455,
-0.04162583500146866,
0.029205596074461937,
-0.05548682063817978,
-0.08170931786298752,
0.08587043732404709,
0.08510591089725494,
-0.1676868051290512,
0.10534299165010452,
0.18449218571186066,
0.04919269308447838,
0.13925369083881378,
0.04365966096520424,
0.04239025339484215,
-0.10937236994504929,
-0.05642855912446976,
-0.0005624148179776967,
0.08393236249685287,
-0.15322627127170563,
0.00680800573900342,
0.059237249195575714,
-0.011514424346387386,
0.05936799198389053,
-0.030177736654877663,
0.0234641432762146,
0.03974314406514168,
0.03531929478049278,
0.0417136549949646,
0.05936359241604805,
-0.06373417377471924,
0.06567533314228058,
0.05290162190794945,
-0.07270863652229309,
0.005567454732954502,
0.0046075740829110146,
-0.05287551134824753,
0.10026288777589798,
-0.051961276680231094,
-0.14957770705223083,
-0.09896279871463776,
-0.060283806174993515,
-0.0491376668214798,
-0.03451768308877945,
0.014281505718827248,
0.02882539853453636,
-0.036672405898571014,
-0.0901276096701622,
0.009361859411001205,
0.007206897251307964,
-0.0826185867190361,
0.05635830760002136,
-0.035462386906147,
-0.015973050147294998,
-0.12909254431724548,
-0.028965696692466736,
0.0005977742839604616,
0.06579206138849258,
0.02878534235060215,
-0.14901986718177795,
0.04435138776898384,
0.0920545756816864,
0.04040633887052536,
0.013421958312392235,
-0.01813202165067196,
0.24397510290145874,
0.014030521735548973,
0.04515877366065979,
0.15045812726020813,
-0.06997589766979218,
0.04986974596977234,
0.06431739777326584,
0.029054081067442894,
-0.0707901194691658,
-0.036528702825307846,
-0.04954122379422188,
-0.032500170171260834,
-0.09421823918819427,
-0.10309197753667831,
-0.08189204335212708,
0.08238279074430466,
0.08317671716213226,
0.006206071469932795,
0.000327636138536036,
0.10295802354812622,
-0.028970876708626747,
0.06721726059913635,
-0.0264450591057539,
0.0798780769109726,
0.06595147401094437,
-0.03811552748084068,
0.06361804157495499,
0.009429671801626682,
-0.00985694956034422,
0.07295644283294678,
0.10556422173976898,
-0.08264261484146118,
0.13516029715538025,
0.2697998285293579,
0.040829502046108246,
0.008056207560002804,
0.04697263985872269,
0.009365313686430454,
-0.07973799854516983,
-0.00977176334708929,
-0.08042590320110321,
-0.07536085695028305,
-0.0006328831659629941,
0.11524549126625061,
0.011204107664525509,
0.11705607175827026,
-0.018434371799230576,
-0.05535244196653366,
0.07728398591279984,
0.15745201706886292,
0.11139922589063644,
-0.2912435233592987,
-0.06395259499549866,
0.05198822543025017,
0.0008797345799393952,
-0.021562354639172554,
-0.028736408799886703,
-0.05675294250249863,
-0.06980589032173157,
0.07784111052751541,
-0.03456875681877136,
0.12166621536016464,
-0.010495441034436226,
0.020177166908979416,
-0.03972733020782471,
0.07303597033023834,
-0.022481700405478477,
0.0452592708170414,
-0.20784910023212433,
0.052252624183893204,
0.04111913964152336,
-0.0017366998363286257,
-0.049390386790037155,
-0.005528331734240055,
0.000565401918720454,
0.0693226158618927,
0.06163579598069191,
0.021157238632440567,
0.10012690722942352,
0.06529508531093597,
-0.11157629638910294,
0.019718384370207787,
-0.006068942602723837,
-0.0967390388250351,
0.04447875916957855,
-0.0026850830763578415,
-0.049156397581100464,
0.034384582191705704,
0.05147765204310417,
-0.11757539957761765,
-0.09579964727163315,
0.08648068457841873,
0.026860326528549194,
0.03165757283568382,
-0.04180118814110756,
-0.026327483355998993,
-0.035426754504442215,
0.2273474633693695,
-0.0953088030219078,
-0.015013204887509346,
-0.04832790419459343,
0.020035479217767715,
0.11327441036701202,
-0.04094255715608597,
0.015705307945609093,
-0.01597696729004383,
-0.0216576736420393,
-0.08264067769050598,
-0.07668239623308182,
0.06864865869283676,
-0.07432357221841812,
-0.10468046367168427,
-0.10537320375442505,
0.09572295099496841,
-0.0028587474953383207,
0.031675439327955246,
0.015725096687674522,
0.039286572486162186,
-0.04792141914367676,
-0.027428846806287766,
0.017214255407452583,
-0.014352234080433846,
0.11432743817567825,
0.023479677736759186,
0.03573967143893242,
0.08954553306102753,
0.0019953332375735044,
0.004165465943515301,
0.07594924420118332,
0.35116007924079895,
-0.028064830228686333,
0.029999561607837677,
0.1499405801296234,
-0.04608915373682976,
-0.05810077488422394,
-0.05257696285843849,
-0.005588115192949772,
0.056246254593133926,
0.028374528512358665,
-0.10211041569709778,
0.03964526578783989,
0.06314241886138916,
-0.0186386089771986,
0.187516450881958,
-0.32887542247772217,
-0.1276625394821167,
0.10910691320896149,
0.04482661560177803,
0.21602796018123627,
-0.060235053300857544,
-0.017379019409418106,
-0.02037195675075054,
-0.22512370347976685,
0.12914420664310455,
0.012854283675551414,
0.0397881418466568,
-0.014497801661491394,
0.06461896747350693,
0.011203703470528126,
-0.05620803311467171,
0.11399297416210175,
-0.037419773638248444,
0.016900405287742615,
-0.05863824114203453,
-0.21782845258712769,
0.0784640833735466,
-0.048408299684524536,
0.037227630615234375,
-0.16283707320690155,
0.021644804626703262,
-0.1173756867647171,
-0.05976768955588341,
-0.025804271921515465,
0.05483346804976463,
-0.05840067192912102,
-0.09434646368026733,
-0.02252507209777832,
0.11715007573366165,
-0.04475071281194687,
0.08446280658245087,
0.20391710102558136,
-0.04670759290456772,
0.0010245213052257895,
0.30389437079429626,
0.0824889987707138,
-0.21188081800937653,
0.06086432561278343,
-0.00889409240335226,
-0.06687145680189133,
0.07212095707654953,
-0.07295092195272446,
0.001358620123937726,
0.07834310084581375,
0.01828717440366745,
0.14376994967460632,
0.01208745688199997,
-0.09074516594409943,
-0.006200927775353193,
0.09681102633476257,
-0.2230166792869568,
-0.08530011028051376,
-0.015431535430252552,
-0.13095000386238098,
-0.06687894463539124,
0.010778072290122509,
0.21223028004169464,
-0.02901056408882141,
-0.058946095407009125,
0.027626099064946175,
-0.005310394335538149,
-0.15737752616405487,
0.05203517526388168,
0.011930471286177635,
0.0322144441306591,
-0.057901933789253235,
0.02004259265959263,
0.05939492583274841,
0.0020846580155193806,
-0.04523533955216408,
-0.015132764354348183,
-0.019137419760227203,
-0.05385998636484146,
-0.09273694455623627,
0.21557959914207458,
-0.01846238225698471,
-0.005689673125743866,
-0.14097532629966736,
0.0006870750803500414,
0.053756020963191986,
0.11148898303508759,
0.06361664831638336,
0.04006518796086311,
-0.0072665950283408165,
0.001948528690263629,
-0.04220915585756302,
0.040487319231033325,
0.07841897755861282,
0.0748501867055893,
-0.0661020576953888,
-0.04205217584967613,
-0.04468723386526108,
0.0015111271059140563,
-0.03756008669734001,
-0.009891976602375507,
-0.23916052281856537,
-0.05951536074280739,
-0.20633672177791595,
-0.04935656860470772,
-0.13159367442131042,
-0.003742414526641369,
-0.028629714623093605,
-0.05982787907123566,
-0.0492701455950737,
0.00844410341233015,
-0.04674763232469559,
0.01556871086359024,
0.05541121959686279,
0.11138446629047394,
-0.0367678701877594,
0.02393273636698723,
0.05884065106511116,
-0.007960398681461811,
0.03172687068581581,
-0.05410018190741539,
0.0047313678078353405,
-0.03174677863717079,
-0.04392893984913826,
-0.01645834930241108,
-0.041589245200157166,
-0.010133353993296623,
0.050766296684741974,
-0.08496593683958054,
-0.0032062209211289883,
-0.021590957418084145,
-0.0127634322270751,
-0.06013529375195503,
0.12053630501031876,
-0.07362815737724304,
0.028214748948812485,
0.019045881927013397,
-0.09820490330457687,
-0.020420731976628304,
0.005175653379410505,
-0.010033139027655125,
-0.01248372346162796,
0.07132677733898163,
0.0077459197491407394,
-0.019086910411715508,
-0.20528627932071686,
-0.0074644144624471664,
0.04650234803557396,
-0.1074509397149086,
-0.005026942119002342,
-0.05010254308581352,
-0.01546710729598999,
-0.021346643567085266,
0.19679290056228638,
-0.006400327663868666,
-0.18126823008060455,
0.0011132284998893738,
-0.0729975476861,
0.004898337181657553,
0.01361140701919794,
0.1522136777639389,
0.024928467348217964,
0.011785217560827732,
-0.09857077151536942,
0.11684924364089966,
0.05830170586705208,
-0.029656793922185898,
0.015374328009784222,
0.13974319398403168,
0.013988420367240906,
0.04178370162844658,
0.10424870997667313,
0.028243213891983032,
-0.00617844657972455,
0.042180731892585754,
0.06380582600831985,
0.09523913264274597,
-0.06373672932386398,
0.09048766642808914,
0.12126588821411133,
-0.16095750033855438,
0.022265693172812462,
0.04928213730454445,
-0.0048591457307338715,
-0.06632255017757416,
-0.27536866068840027,
-0.12273569405078888,
-0.1283498853445053,
0.013920783996582031,
-0.11781493574380875,
-0.05833969637751579,
0.0024395440705120564,
-0.0029342269990593195,
-0.025490280240774155,
-0.04510700702667236,
-0.03308400884270668,
0.017413116991519928,
0.0356060191988945,
-0.03484621271491051,
-0.0558101050555706,
-0.08678935468196869,
-0.0805462896823883,
-0.0366392619907856,
0.024760788306593895,
-0.008014274761080742,
-0.02920999750494957,
-0.04541094973683357,
-0.03210213780403137,
0.050164029002189636,
-0.09916680306196213,
-0.07774882018566132,
0.04403835907578468,
0.05857539921998978,
-0.031173842027783394,
-0.017724227160215378,
0.021220481023192406,
-0.009362087585031986,
0.05539046600461006,
-0.011212928220629692,
-0.020259330049157143,
-0.09906895458698273,
0.12684087455272675,
-0.08834514766931534,
0.057960331439971924,
0.033544231206178665,
-0.0033638423774391413,
0.034319378435611725,
0.14744143187999725,
0.35406649112701416,
-0.14233095943927765,
-0.01208020281046629,
0.06998341530561447,
0.026331178843975067,
0.01323934830725193,
0.06403732299804688,
-0.03632789105176926,
0.130133718252182,
-0.06310538947582245,
0.06529198586940765,
-0.0024390413891524076,
-0.051876191049814224,
0.03159966692328453,
-0.01178564690053463,
0.03747725486755371,
-0.026889890432357788,
0.04011377692222595,
0.011130592785775661,
-0.08025975525379181,
0.01339031383395195,
0.003319264855235815,
-0.18194924294948578,
-0.133108451962471,
-0.07022427022457123,
-0.09842776507139206,
-0.007627582177519798,
0.12037099152803421,
-0.05317481607198715,
-0.1017008125782013,
0.02909071557223797,
-0.047177378088235855,
-0.06705634295940399,
-0.15004311501979828,
0.07639843970537186,
-0.13587135076522827,
0.07903552055358887,
-0.04076021909713745,
0.03959066793322563,
0.12640850245952606,
-0.029877981171011925,
-0.057700399309396744,
-0.055523183196783066,
0.03190966695547104,
0.0016701313434168696,
-0.030366502702236176,
-0.04228977859020233,
-0.008464036509394646,
0.2516556680202484,
0.027898557484149933,
-0.13780070841312408,
0.03654743358492851,
0.08006620407104492,
-0.07084063440561295,
-0.06895849853754044,
0.0528719536960125,
-0.037205230444669724,
0.14586538076400757,
0.17537543177604675,
-0.06271609663963318,
-0.06536475569009781,
-0.01028021052479744,
0.14450573921203613,
0.10469193756580353,
0.05885422229766846,
-0.11430418491363525,
-0.14308179914951324,
-0.017617736011743546,
0.06508629024028778,
-0.023243948817253113,
-0.15663959085941315,
-0.07245104759931564,
-0.1602208912372589,
0.02328074350953102,
0.026763690635561943,
0.04315638542175293,
0.19113922119140625,
0.04011261835694313,
-0.04989412799477577,
-0.1907459944486618,
-0.02351398766040802,
0.0659966692328453,
-0.11507140100002289,
-0.09325497597455978
] |
null | null | null | **Como traducir con este Modelo**
+ Instalar [Python 3.9](https://www.python.org/downloads/release/python-390/) + ctranslate 2 + subword-nmt
```bash
pip install ctranslate2~=3.20.0
```
```bash
pip install subword-nmt
```
+ procesar texto con BPE:
```bash
subword-nmt apply-bpe -c ca-detok10k.code < input_file.txt > input_file_bpe.txt
```
+ Traducir un input_text utilizando NOS-MT-eu-gl:
```bash
python3 trans_ct2.py ct2_detok-ca-gl_sint_10k input_file_bpe.txt >output_file_bpe.txt
```
+ DeBPEzar output txt:
```bash
cat out_test_bpe.txt | sed "s/@@ //g" > output_file.txt
``` | {"license": "mit"} | null | proxectonos/Nos_MT-OpenNMT-ca-gl | [
"license:mit",
"region:us"
] | 2024-02-07T08:17:58+00:00 | [] | [] | TAGS
#license-mit #region-us
| Como traducir con este Modelo
+ Instalar Python 3.9 + ctranslate 2 + subword-nmt
+ procesar texto con BPE:
+ Traducir un input_text utilizando NOS-MT-eu-gl:
+ DeBPEzar output txt:
| [] | [
"TAGS\n#license-mit #region-us \n"
] | [
11
] | [
"passage: TAGS\n#license-mit #region-us \n"
] | [
0.026221778243780136,
-0.033018264919519424,
-0.008281232789158821,
-0.05295303836464882,
0.052470896393060684,
0.06768012046813965,
0.1598525494337082,
0.04655371606349945,
0.23683255910873413,
-0.05407243221998215,
0.11752297729253769,
0.08923697471618652,
0.004284696187824011,
-0.0009730930323712528,
0.014216204173862934,
-0.17134642601013184,
0.04864625632762909,
-0.02878100797533989,
0.08764812350273132,
0.032233644276857376,
-0.006205103360116482,
-0.03845774009823799,
-0.0022142508532851934,
-0.03178790956735611,
-0.057939812541007996,
0.03869890421628952,
0.045729056000709534,
-0.02754949778318405,
0.14189864695072174,
-0.021783310920000076,
0.13335508108139038,
0.046146418899297714,
-0.011738095432519913,
-0.2486042082309723,
0.008575023151934147,
-0.07252951711416245,
-0.11333522200584412,
0.016201216727495193,
0.035761721432209015,
-0.010069100186228752,
0.032174937427043915,
0.11049123108386993,
-0.011680051684379578,
0.06288356333971024,
-0.2015703022480011,
-0.20486389100551605,
-0.07508610188961029,
-0.07555478066205978,
0.0589042492210865,
0.030872387811541557,
0.05628744140267372,
0.1426718831062317,
-0.18022038042545319,
-0.0018841808196157217,
0.04129622131586075,
-0.3510737717151642,
0.09011197835206985,
0.19666501879692078,
0.06407395005226135,
0.07872317731380463,
-0.04774639382958412,
0.06726468354463577,
0.07745297998189926,
-0.02402484230697155,
-0.10679105669260025,
-0.06142130121588707,
0.040939174592494965,
0.15604156255722046,
-0.03852643445134163,
-0.10356393456459045,
0.2591084837913513,
-0.023262828588485718,
-0.04234466329216957,
0.08201269060373306,
-0.02980397455394268,
-0.040379155427217484,
0.04404358193278313,
0.044016025960445404,
0.036236923187971115,
0.182089164853096,
0.1260262131690979,
-0.03375067934393883,
-0.16269677877426147,
-0.030629513785243034,
-0.2528207004070282,
0.07418664544820786,
-0.003647059667855501,
0.10666298121213913,
-0.20037521421909332,
0.03286786004900932,
-0.15483668446540833,
-0.009493621066212654,
-0.02952384203672409,
-0.059835705906152725,
0.05229754373431206,
-0.0237403754144907,
-0.04600388556718826,
0.07238677144050598,
0.08390641957521439,
0.2046167105436325,
0.023024363443255424,
0.016697337850928307,
-0.10405295342206955,
0.15052515268325806,
0.019140364602208138,
0.024860305711627007,
0.179348424077034,
0.07677878439426422,
-0.04891882464289665,
-0.2251969277858734,
0.027894439175724983,
-0.03671982139348984,
-0.1441805064678192,
0.015881337225437164,
-0.1542915552854538,
0.1736440360546112,
-0.04078168794512749,
-0.06919530034065247,
-0.08578147739171982,
0.09790384024381638,
0.07768166810274124,
-0.021921472623944283,
-0.023105677217245102,
-0.01381723117083311,
0.03522264584898949,
-0.048196230083703995,
-0.11687057465314865,
0.018241960555315018,
0.11869648098945618,
0.12573401629924774,
-0.1483907401561737,
-0.008189842104911804,
-0.017200417816638947,
0.019065292552113533,
0.09696817398071289,
-0.112403005361557,
0.028845038264989853,
-0.09672309458255768,
-0.13033071160316467,
0.036653537303209305,
0.017736904323101044,
-0.019008556380867958,
0.1340927630662918,
0.061849117279052734,
0.056560322642326355,
-0.011025321669876575,
-0.07250872999429703,
-0.14035539329051971,
-0.08679798245429993,
0.1058693379163742,
-0.046787332743406296,
0.010320915840566158,
-0.24556252360343933,
-0.014234079979360104,
-0.14995723962783813,
0.059662189334630966,
-0.0037668521981686354,
-0.08819212019443512,
-0.07740068435668945,
0.21408265829086304,
0.0018596589798107743,
0.04301392287015915,
-0.1078512966632843,
0.054903753101825714,
-0.06764797121286392,
0.10065380483865738,
-0.12895582616329193,
-0.06441528350114822,
0.1613781899213791,
-0.13135331869125366,
-0.14002031087875366,
0.0033312994055449963,
-0.009472889825701714,
0.12053907662630081,
0.0802001804113388,
0.44566696882247925,
-0.058881040662527084,
-0.16201181709766388,
0.1270403116941452,
0.17969723045825958,
-0.13685379922389984,
-0.25928929448127747,
0.12393020838499069,
-0.1636963188648224,
-0.16647985577583313,
0.0040023741312325,
-0.006962866988033056,
0.08049977570772171,
-0.03446655720472336,
-0.056274134665727615,
0.042339932173490524,
0.024350708350539207,
0.029094615951180458,
0.01740112341940403,
0.07037191838026047,
-0.1023021712899208,
0.08444856107234955,
0.058610700070858,
-0.014111426658928394,
0.15077349543571472,
0.011494536884129047,
-0.05393160134553909,
0.014761670492589474,
0.044013332575559616,
-0.015627963468432426,
-0.05899091437458992,
-0.09661509096622467,
0.019826244562864304,
-0.031149597838521004,
0.08229395002126694,
0.1699674129486084,
0.023824702948331833,
-0.02797185815870762,
0.028922779485583305,
0.028606392443180084,
0.1009954959154129,
0.06960704177618027,
0.03099375218153,
-0.04839283227920532,
0.04952205345034599,
-0.0417071171104908,
-0.11430390179157257,
-0.004862460307776928,
-0.011735930107533932,
0.11975742131471634,
-0.08906009048223495,
-0.01223952230066061,
0.05951591953635216,
-0.04513183981180191,
0.0019881438929587603,
0.0428374819457531,
0.0035966038703918457,
0.1388600617647171,
0.004440935328602791,
-0.04352007433772087,
0.17440910637378693,
-0.05288633331656456,
0.15533447265625,
0.1715822070837021,
-0.07049662619829178,
0.015605369582772255,
-0.1273636519908905,
0.003230511210858822,
-0.014480113983154297,
0.05292887985706329,
-0.05400136485695839,
-0.05201306566596031,
-0.01274962443858385,
0.014292534440755844,
-0.03134604170918465,
0.01711403578519821,
-0.06057267636060715,
-0.08167021721601486,
-0.10849859565496445,
0.018649224191904068,
0.20683221518993378,
-0.22544461488723755,
0.1609548032283783,
0.40251004695892334,
0.15190774202346802,
0.21155193448066711,
-0.12478897720575333,
-0.002471078187227249,
-0.06630261242389679,
0.026115071028470993,
-0.024814706295728683,
0.13782677054405212,
-0.13174867630004883,
-0.01413064356893301,
0.03880728408694267,
0.0454997681081295,
0.0661163181066513,
-0.17195898294448853,
-0.15260353684425354,
-0.0034879595041275024,
-0.020591814070940018,
-0.1749730259180069,
0.04874620959162712,
-0.07595308125019073,
0.02181261032819748,
0.018216799944639206,
-0.10832522064447403,
0.16837291419506073,
-0.033566512167453766,
-0.06695768237113953,
0.052613962441682816,
-0.20581911504268646,
-0.07900715619325638,
-0.17772749066352844,
-0.18375012278556824,
0.06050071492791176,
0.05760138854384422,
0.07903145253658295,
-0.05951719731092453,
-0.01922747679054737,
0.061719246208667755,
-0.009363299235701561,
-0.13802112638950348,
-0.04235544428229332,
-0.06993678212165833,
0.08744155615568161,
-0.09474305808544159,
-0.07518411427736282,
-0.07833878695964813,
-0.046996138989925385,
-0.020961694419384003,
0.08125963062047958,
-0.1039251759648323,
0.08903530240058899,
0.1493726521730423,
0.03651920333504677,
0.05440247058868408,
-0.08271230012178421,
0.12693379819393158,
-0.037743739783763885,
-0.09459595382213593,
0.07307634502649307,
0.004350725095719099,
0.04920351505279541,
0.24039287865161896,
0.08962162584066391,
-0.10578162968158722,
-0.01780811697244644,
-0.0968487411737442,
-0.16405464708805084,
-0.2553846538066864,
-0.06823288649320602,
-0.08744750916957855,
0.14417944848537445,
0.014636521227657795,
0.10712126642465591,
0.14313316345214844,
0.01343101728707552,
0.10255914181470871,
-0.08983208239078522,
-0.018939344212412834,
0.031209396198391914,
0.2135104089975357,
-0.05208220332860947,
0.00838248711079359,
-0.13684824109077454,
-0.0256142970174551,
0.14601100981235504,
0.13798639178276062,
0.14503207802772522,
0.31421369314193726,
0.15292863547801971,
0.13410434126853943,
0.13474710285663605,
0.12333164364099503,
0.07403261214494705,
0.03444362059235573,
-0.015304201282560825,
-0.06035377085208893,
-0.003846159903332591,
0.02816268615424633,
0.05421729013323784,
0.06724072247743607,
-0.22906480729579926,
0.041139665991067886,
-0.2661744952201843,
0.03544611483812332,
-0.0854712724685669,
0.1161833181977272,
-0.028890252113342285,
0.11051984131336212,
0.11386284977197647,
0.05553818494081497,
-0.023278791457414627,
0.16036942601203918,
0.032686375081539154,
-0.07703183591365814,
0.020292721688747406,
0.024695809930562973,
0.06633034348487854,
0.08606193959712982,
0.09550496190786362,
-0.020778406411409378,
-0.1831783503293991,
0.025963841006159782,
0.12212833017110825,
-0.20747940242290497,
0.289523184299469,
0.013651901856064796,
-0.0743619054555893,
-0.01690039224922657,
-0.06958060711622238,
0.008433517068624496,
0.12829731404781342,
0.10406835377216339,
0.05508929491043091,
-0.2613787055015564,
-0.13299626111984253,
0.046764206141233444,
-0.00873907096683979,
0.11356569826602936,
-0.0052223424427211285,
-0.14201195538043976,
-0.06640999764204025,
0.05814211815595627,
-0.006591420155018568,
0.13023322820663452,
-0.018290361389517784,
-0.08173255622386932,
-0.010230090469121933,
0.055564697831869125,
-0.001312803477048874,
-0.04580084979534149,
0.07523149996995926,
0.009008137509226799,
0.02259289287030697,
-0.08178020268678665,
0.03887253627181053,
-0.08071476966142654,
-0.25375792384147644,
0.019298138096928596,
-0.04987313598394394,
0.004092312417924404,
-0.04684043675661087,
-0.15448936820030212,
-0.1129264086484909,
-0.15445278584957123,
0.13100723922252655,
-0.03675999864935875,
0.091565802693367,
-0.0817658007144928,
0.13736046850681305,
-0.08521489799022675,
0.05375019088387489,
0.00614814180880785,
0.03918716683983803,
-0.017955513671040535,
-0.1031481996178627,
0.09334362298250198,
-0.1874227225780487,
0.023863423615694046,
0.010427716188132763,
-0.056847453117370605,
-0.01354232057929039,
0.03918023407459259,
-0.08763083070516586,
0.21879427134990692,
0.3331502079963684,
-0.011948764324188232,
0.22546616196632385,
0.35863226652145386,
-0.13763751089572906,
-0.23258967697620392,
-0.1205512136220932,
-0.3263251483440399,
-0.09005610644817352,
0.17321562767028809,
-0.18057219684123993,
0.04850830137729645,
0.16150830686092377,
-0.10868281871080399,
0.22499866783618927,
-0.22723928093910217,
-0.04793389141559601,
0.1823979914188385,
-0.038322996348142624,
0.4527989625930786,
-0.1144307404756546,
-0.1784561723470688,
-0.03637253865599632,
-0.16285361349582672,
0.12426037341356277,
-0.026553882285952568,
0.06700495630502701,
0.02416347898542881,
-0.011372359469532967,
-0.009014161303639412,
-0.04529716446995735,
0.2216065675020218,
0.0522729866206646,
0.10468899458646774,
-0.09159468114376068,
-0.17199653387069702,
0.1907423883676529,
-0.0004908236442133784,
-0.003372655250132084,
-0.05411549657583237,
-0.04850282520055771,
-0.06871756166219711,
0.033092137426137924,
-0.0334564633667469,
0.06195882335305214,
0.03364093229174614,
-0.11903523653745651,
-0.10248823463916779,
0.034111104905605316,
-0.13155671954154968,
-0.054850947111845016,
0.26421889662742615,
-0.02080743946135044,
0.09609334170818329,
0.04959092289209366,
-0.05474294349551201,
-0.13538943231105804,
0.005736751481890678,
-0.07534020394086838,
-0.05711410939693451,
0.06573604047298431,
-0.11453206837177277,
-0.024341827258467674,
0.1293732225894928,
-0.029497180134058,
0.09674722701311111,
0.08061115443706512,
-0.07585363835096359,
0.02032829262316227,
0.15617427229881287,
-0.07247176766395569,
-0.10849180817604065,
0.04999847710132599,
0.04640531167387962,
0.17256882786750793,
0.004101871978491545,
0.02018604800105095,
0.08726977556943893,
0.045959215611219406,
-0.007486662827432156,
0.007311292923986912,
-0.11321697384119034,
-0.04241771996021271,
0.0387241393327713,
-0.005273692775517702,
-0.10946331918239594,
0.16008898615837097,
0.056837860494852066,
0.004653505515307188,
-0.06027700752019882,
0.09720424562692642,
-0.06709636747837067,
-0.07046061009168625,
-0.1753035932779312,
0.018511172384023666,
-0.12734080851078033,
-0.09874535351991653,
0.06846235692501068,
-0.09371624886989594,
-0.04084605351090431,
0.08152704685926437,
0.046927981078624725,
0.14401860535144806,
-0.006597559433430433,
-0.023080874234437943,
0.149825319647789,
-0.0884878933429718,
-0.2241756170988083,
0.01969664730131626,
-0.04083063453435898,
-0.07065816223621368,
-0.0007070365245454013,
0.06069544702768326,
-0.0663156732916832,
-0.11958606541156769,
-0.20477768778800964,
0.10412076860666275,
-0.12043121457099915,
-0.03954985365271568,
-0.1041841059923172,
-0.053260523825883865,
0.07891252636909485,
-0.02613759972155094,
-0.04122013971209526,
-0.047595683485269547,
-0.16630595922470093,
0.054254453629255295,
0.07140932232141495,
0.11125344783067703,
-0.0759999230504036,
-0.018354382365942,
0.1398727148771286,
0.048581548035144806,
0.08479110151529312,
0.07578440010547638,
0.026255371049046516,
0.16728560626506805,
-0.1708206981420517,
-0.0542997270822525,
0.1068294569849968,
-0.026716172695159912,
0.01994573324918747,
0.10631280392408371,
-0.04839588701725006,
0.07042654603719711,
-0.05095988139510155,
0.05859163776040077,
-0.15704534947872162,
-0.13073866069316864,
-0.04184387996792793,
0.023728877305984497,
-0.2260182797908783,
0.015071595087647438,
-0.1769561767578125,
0.19692228734493256,
-0.024228032678365707,
0.11490963399410248,
0.08052190393209457,
0.02052290178835392,
0.03539382666349411,
-0.006019921973347664,
0.00946811307221651,
-0.10524865239858627,
-0.05784677714109421,
-0.07560300827026367,
-0.1168874129652977,
-0.009665017947554588,
0.36614301800727844,
0.02430291846394539,
-0.19682736694812775,
0.051222387701272964,
0.18285293877124786,
0.023639049381017685,
-0.0073763905093073845,
0.26180747151374817,
0.08150359988212585,
-0.023175053298473358,
-0.1782374382019043,
0.0396091528236866,
-0.08699734508991241,
-0.15269799530506134,
0.11385007947683334,
0.09347525984048843,
0.05813581123948097,
0.022930078208446503,
0.10404518246650696,
-0.035940010100603104,
-0.05509711429476738,
-0.13301853835582733,
0.13368983566761017,
-0.001790675800293684,
0.0193882267922163,
0.0897885113954544,
0.19249756634235382,
-0.045275162905454636,
0.05437124893069267,
-0.07336640357971191,
-0.001598604372702539,
-0.15740543603897095,
-0.13358698785305023,
0.06194563955068588,
-0.08269550651311874,
0.06342913210391998,
0.050261519849300385,
0.04341990500688553,
0.31786394119262695,
0.039095040410757065,
-0.046439893543720245,
0.003166865324601531,
-0.14845187962055206,
-0.08075450360774994,
-0.06024569645524025,
-0.03110554814338684,
0.028620192781090736,
-0.13928957283496857,
-0.09898591786623001,
-0.06917677819728851,
-0.130235955119133,
-0.06539803743362427,
0.025270747020840645,
0.014251931570470333,
-0.053083837032318115,
-0.17625881731510162,
-0.04808593541383743,
-0.06644169986248016,
0.10105955600738525,
-0.08462738990783691,
0.1516820639371872,
0.0022449472453445196,
0.030281953513622284,
0.07627002149820328,
0.09585131704807281,
0.018900424242019653,
-0.06975197046995163,
0.05599058046936989,
0.12436293810606003,
0.01323844213038683,
0.1259988248348236,
-0.06034265458583832,
-0.019420607015490532,
-0.014145253226161003,
0.14038437604904175,
0.304447740316391,
-0.01856905221939087,
-0.013814439997076988,
-0.022110093384981155,
0.021388787776231766,
0.10893569141626358,
0.19800719618797302,
-0.03437356278300285,
0.2551359534263611,
-0.058974795043468475,
0.0756678432226181,
-0.013180435635149479,
-0.005362013820558786,
-0.053146667778491974,
0.06074550002813339,
0.06268858164548874,
-0.06877048313617706,
-0.10191375762224197,
0.15178529918193817,
-0.14985080063343048,
0.13306055963039398,
0.14678068459033966,
-0.06057753041386604,
0.03797250986099243,
0.0007459368789568543,
0.19896264374256134,
-0.03570213168859482,
0.0984780564904213,
-0.10653308779001236,
-0.10261140763759613,
-0.14764924347400665,
0.037690844386816025,
-0.36797797679901123,
-0.1756322830915451,
0.11731542646884918,
0.14115898311138153,
0.1759258657693863,
-0.012341637164354324,
0.056479312479496,
0.0033020609989762306,
0.08296097069978714,
-0.04232487455010414,
0.1519634872674942,
0.0612073615193367,
-0.017103128135204315,
-0.15296664834022522,
-0.20328094065189362,
-0.0012039330322295427,
-0.058561209589242935,
0.055583830922842026,
-0.02269243635237217,
0.025347469374537468,
0.07746459543704987,
-0.06768939644098282,
-0.029180381447076797,
-0.02352982573211193,
-0.13262848556041718,
0.052229251712560654,
-0.04354005306959152,
0.0320255309343338,
-0.03958037868142128,
-0.022394726052880287,
-0.039987675845623016,
0.10721533745527267,
-0.22402705252170563,
-0.08517231047153473,
0.1422796994447708,
-0.03421911224722862,
0.1542559564113617,
-0.02848726324737072,
-0.12159585952758789,
-0.024955326691269875,
-0.06977712363004684,
0.10887379199266434,
-0.1419300138950348,
0.038592495024204254,
0.13747453689575195,
0.008710617199540138,
0.031119761988520622,
-0.2533661723136902,
0.050644006580114365,
-0.03556957095861435,
-0.016733208671212196,
-0.057031940668821335
] |
null | null | diffusers |
# LoRA DreamBooth - danaleee/CL_rank4
These are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using [DreamBooth](https://dreambooth.github.io/). You can find some example images in the following.




LoRA for the text encoder was enabled: False.
| {"license": "creativeml-openrail-m", "tags": ["stable-diffusion", "stable-diffusion-diffusers", "text-to-image", "diffusers", "lora"], "base_model": "CompVis/stable-diffusion-v1-4", "instance_prompt": "a photo of sks teddybear", "inference": true} | text-to-image | danaleee/CL_rank4 | [
"diffusers",
"tensorboard",
"stable-diffusion",
"stable-diffusion-diffusers",
"text-to-image",
"lora",
"base_model:CompVis/stable-diffusion-v1-4",
"license:creativeml-openrail-m",
"region:us"
] | 2024-02-07T08:18:39+00:00 | [] | [] | TAGS
#diffusers #tensorboard #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us
|
# LoRA DreamBooth - danaleee/CL_rank4
These are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using DreamBooth. You can find some example images in the following.
!img_0
!img_1
!img_2
!img_3
LoRA for the text encoder was enabled: False.
| [
"# LoRA DreamBooth - danaleee/CL_rank4\n\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
"TAGS\n#diffusers #tensorboard #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us \n",
"# LoRA DreamBooth - danaleee/CL_rank4\n\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
70,
101
] | [
"passage: TAGS\n#diffusers #tensorboard #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us \n# LoRA DreamBooth - danaleee/CL_rank4\n\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
-0.04341907054185867,
-0.015942202880978584,
-0.0028250112663954496,
0.07746413350105286,
0.1169724091887474,
0.025427518412470818,
0.18271614611148834,
0.0738026574254036,
0.049231309443712234,
0.0709879919886589,
0.11736305058002472,
0.07064489275217056,
-0.004964313004165888,
0.09631748497486115,
-0.018667949363589287,
-0.15524275600910187,
0.007252814248204231,
-0.03455926477909088,
-0.05518360808491707,
0.03522060438990593,
0.03546910360455513,
-0.04334699735045433,
0.11702632158994675,
-0.03578236326575279,
-0.16095487773418427,
0.040608540177345276,
0.006910618860274553,
-0.05284915491938591,
0.06334300339221954,
0.07431042939424515,
0.04444604739546776,
0.06587636470794678,
0.049278002232313156,
-0.14593403041362762,
0.031002074480056763,
0.0163436196744442,
-0.052670031785964966,
0.07364684343338013,
-0.08684653788805008,
-0.010339943692088127,
0.0840868204832077,
-0.04559224471449852,
0.011343990452587605,
0.025770828127861023,
-0.08010706305503845,
-0.049801744520664215,
-0.021523071452975273,
-0.0031602235976606607,
0.034805260598659515,
0.055362269282341,
0.006313429679721594,
0.08823953568935394,
-0.029240038245916367,
0.06806129962205887,
0.26588916778564453,
-0.23950862884521484,
-0.01899855211377144,
0.2235211431980133,
-0.008139188401401043,
0.08547473698854446,
-0.04052017629146576,
0.08897750079631805,
0.11105094850063324,
-0.029187314212322235,
0.03702497482299805,
-0.05890991911292076,
-0.004559684544801712,
-0.055189672857522964,
-0.09971760213375092,
0.04510728642344475,
0.1639001965522766,
-0.006552010774612427,
-0.06654545664787292,
-0.12615536153316498,
-0.02784753032028675,
0.02541022188961506,
-0.0031803862657397985,
0.03594784811139107,
0.014548975974321365,
-0.0022297499235719442,
-0.03507556766271591,
-0.05079469829797745,
-0.07389845699071884,
-0.05568958446383476,
-0.007205710280686617,
0.10964683443307877,
-0.005488741211593151,
0.055863216519355774,
-0.004374562297016382,
0.12705139815807343,
-0.12092702835798264,
-0.11754394322633743,
0.033632051199674606,
-0.06341832876205444,
0.01322225946933031,
0.06406484544277191,
-0.025293990969657898,
-0.10047097504138947,
0.05880392715334892,
0.003382892580702901,
0.1380108743906021,
0.016355440020561218,
-0.0033950614742934704,
0.09913190454244614,
-0.007198414299637079,
0.05315367504954338,
-0.053160410374403,
0.002367869019508362,
0.03363179787993431,
0.032208699733018875,
0.1052597388625145,
-0.07429850101470947,
-0.15475058555603027,
-0.0004368349618744105,
-0.055206287652254105,
0.053766872733831406,
-0.087501659989357,
0.014477933757007122,
-0.09005289524793625,
-0.0038406439125537872,
0.07456234097480774,
-0.015409450978040695,
0.03412897512316704,
-0.026990143582224846,
-0.030341247096657753,
0.11830125004053116,
0.16821438074111938,
0.017788158729672432,
0.0031343435402959585,
0.07421474158763885,
-0.07731165736913681,
0.046248506754636765,
-0.014303154312074184,
-0.1297140270471573,
0.01666654273867607,
-0.16157647967338562,
-0.0010764120379462838,
-0.13301143050193787,
-0.03379840403795242,
-0.008777874521911144,
0.02144388109445572,
-0.04588810354471207,
0.02913632057607174,
-0.06573432683944702,
-0.12821175158023834,
-0.0031739489641040564,
0.06216943636536598,
-0.015786021947860718,
-0.013941195793449879,
0.05589903146028519,
-0.006068806163966656,
0.11702877283096313,
-0.12085311859846115,
-0.027835676446557045,
-0.06684071570634842,
0.020470207557082176,
-0.13754083216190338,
0.07727622240781784,
-0.052574124187231064,
0.07806853204965591,
-0.057337477803230286,
-0.046603426337242126,
-0.005025265738368034,
0.03498596325516701,
0.03384779393672943,
0.1205277144908905,
-0.24896466732025146,
-0.05861677601933479,
0.07136847078800201,
-0.1564018577337265,
-0.08534526824951172,
0.047191981226205826,
-0.03015514649450779,
0.08869519084692001,
0.07764715701341629,
0.09606057405471802,
0.08426419645547867,
-0.30525273084640503,
-0.013229794800281525,
-0.057413749396800995,
-0.029245099052786827,
-0.03553345426917076,
-0.0008177062263712287,
0.06475933641195297,
-0.046205006539821625,
0.037332937121391296,
-0.07291214168071747,
0.06947853416204453,
-0.026414241641759872,
-0.016797566786408424,
-0.03371608257293701,
-0.04689457640051842,
-0.002779435832053423,
-0.01032639667391777,
0.04845189303159714,
-0.023798754438757896,
-0.03623393550515175,
0.03102920576930046,
0.04877059534192085,
-0.050546493381261826,
0.0157078318297863,
-0.02974431961774826,
0.09660544991493225,
-0.09479434788227081,
-0.0069621168076992035,
-0.08176911622285843,
-0.014171873219311237,
0.05322252959012985,
0.13276153802871704,
0.11194261908531189,
0.05263838544487953,
0.08583188056945801,
0.07237133383750916,
-0.0018843937432393432,
-0.001962454989552498,
0.048040907829999924,
-0.006709123495966196,
-0.05348801612854004,
-0.19440379738807678,
0.02523115463554859,
-0.08131305128335953,
0.043285779654979706,
-0.19587378203868866,
0.02132508158683777,
0.053477559238672256,
0.17367611825466156,
0.09655749797821045,
-0.013899138197302818,
0.05741235986351967,
0.036425426602363586,
-0.05011986941099167,
-0.045139845460653305,
0.028008436784148216,
-0.01475561410188675,
-0.17448444664478302,
0.15440484881401062,
-0.17193910479545593,
0.035062581300735474,
0.12041056156158447,
-0.0016040849732235074,
-0.06568825989961624,
-0.04588296636939049,
-0.0032709308434277773,
0.03323923796415329,
-0.047381896525621414,
-0.045463677495718,
0.125295951962471,
-0.022256238386034966,
0.15403875708580017,
-0.025979146361351013,
0.008112328127026558,
0.060605812817811966,
-0.03923344239592552,
-0.08103050291538239,
0.09901610016822815,
0.06290307641029358,
0.030513597652316093,
0.010938664898276329,
0.07849103957414627,
-0.013099446892738342,
0.12773942947387695,
-0.001039071474224329,
-0.04228655993938446,
-0.07129907608032227,
0.020174460485577583,
0.05424272269010544,
0.1278262734413147,
0.029769597575068474,
-0.024302449077367783,
-0.012022635899484158,
-0.06184166669845581,
-0.0008379706414416432,
-0.1587388515472412,
-0.028058651834726334,
0.019866183400154114,
-0.041091155260801315,
0.14458370208740234,
0.08805502206087112,
-0.08585330098867416,
0.08541098982095718,
-0.10261666774749756,
-0.06717682629823685,
-0.004778225906193256,
-0.028193825855851173,
-0.06060744822025299,
0.11131694167852402,
-0.06571992486715317,
-0.15074458718299866,
-0.1912163645029068,
0.003397236578166485,
-0.009303453378379345,
-0.00008987563342088833,
0.05273029953241348,
-0.09441886097192764,
-0.06646677106618881,
-0.11243582516908646,
0.029996473342180252,
0.011960997246205807,
0.04876722767949104,
0.0459054559469223,
-0.02460494264960289,
0.00773900980129838,
-0.09628549963235855,
-0.003957229200750589,
-0.07168728858232498,
0.06938584893941879,
0.059952203184366226,
0.0010141970124095678,
0.1007194072008133,
0.12326371669769287,
0.026784677058458328,
0.03313584625720978,
0.0040140156634151936,
0.23651684820652008,
-0.007698030676692724,
0.08224834501743317,
0.1245686337351799,
-0.00448490958660841,
0.061381831765174866,
0.1301926076412201,
0.04333924502134323,
-0.08629945665597916,
0.06321949511766434,
-0.010034860111773014,
-0.1451377123594284,
-0.09342466294765472,
-0.0534212701022625,
-0.033006347715854645,
-0.06310927867889404,
0.07140850275754929,
0.04823211580514908,
0.11957568675279617,
0.07595479488372803,
0.05541970208287239,
0.13485616445541382,
0.06625436246395111,
0.07318418473005295,
0.07824353873729706,
-0.0482291616499424,
0.03650830313563347,
-0.07884727418422699,
-0.11227764934301376,
0.09438467025756836,
-0.06343226879835129,
0.2031732201576233,
-0.06467324495315552,
0.053611837327480316,
0.039042260497808456,
-0.043771810829639435,
0.08480504155158997,
0.029249869287014008,
-0.04769064113497734,
-0.012421797029674053,
-0.03355320543050766,
-0.12436506897211075,
0.10057533532381058,
0.09569450467824936,
0.003375782398506999,
-0.00896451435983181,
-0.036869775503873825,
0.0352763757109642,
0.02298007346689701,
-0.009243073873221874,
0.14936797320842743,
-0.22447653114795685,
0.03024793043732643,
0.021636491641402245,
0.06610236316919327,
0.011202017776668072,
0.00287762307561934,
0.216024711728096,
-0.004215493332594633,
0.06869127601385117,
-0.03940929099917412,
0.05858461186289787,
-0.01287341769784689,
-0.031293123960494995,
-0.043106671422719955,
0.13481405377388,
-0.01601315289735794,
-0.03536630794405937,
-0.22338853776454926,
0.04546704888343811,
0.012524083256721497,
0.014565816149115562,
-0.062143053859472275,
-0.025063209235668182,
0.030040176585316658,
-0.004338495898991823,
0.0890142172574997,
0.01626761443912983,
0.021072275936603546,
-0.11259448528289795,
-0.16940662264823914,
-0.037297990173101425,
0.07631037384271622,
-0.03251313045620918,
0.06547873467206955,
0.0486355759203434,
-0.028622562065720558,
0.014236615970730782,
0.02339186519384384,
-0.1410760134458542,
-0.08791647851467133,
0.00702568469569087,
0.18404577672481537,
-0.0450335294008255,
-0.03229532390832901,
-0.0964898020029068,
-0.02994699403643608,
0.0752733051776886,
-0.05780308321118355,
-0.06531932204961777,
-0.06656667590141296,
0.009424086660146713,
0.09365184605121613,
-0.029468130320310593,
-0.009495943784713745,
-0.03996599093079567,
0.06293158233165741,
-0.0757058635354042,
-0.1436227262020111,
0.06641549617052078,
-0.00830997433513403,
-0.14320321381092072,
-0.07497122883796692,
0.1233392208814621,
-0.0072281318716704845,
0.005523594096302986,
-0.018453087657690048,
0.038608841598033905,
0.03829873725771904,
-0.11585721373558044,
0.06597750633955002,
0.17044483125209808,
-0.14400237798690796,
0.12962660193443298,
-0.03997928276658058,
0.005498290061950684,
-0.06308922171592712,
0.013415500521659851,
0.1249745711684227,
0.2677513360977173,
-0.10195360332727432,
0.10731205344200134,
0.03661147505044937,
-0.09840145707130432,
-0.20925116539001465,
-0.04050299897789955,
-0.016064783558249474,
0.039085373282432556,
-0.016609976068139076,
-0.08601631224155426,
0.10712312161922455,
0.0405130535364151,
0.01219622977077961,
0.23049339652061462,
-0.3584687113761902,
-0.1457986980676651,
-0.02161003090441227,
0.15273447334766388,
0.28606653213500977,
-0.14790372550487518,
-0.08715926110744476,
-0.03193103149533272,
-0.054018907248973846,
0.15796802937984467,
-0.07264256477355957,
0.1155640110373497,
-0.03511131554841995,
-0.02772337570786476,
0.02722114510834217,
-0.03559715673327446,
0.13881368935108185,
-0.0039151934906840324,
0.04988088458776474,
-0.07145698368549347,
-0.09874292463064194,
0.09717526286840439,
-0.055095139890909195,
0.045357201248407364,
-0.1500481367111206,
0.021375473588705063,
-0.04788108915090561,
-0.00894334726035595,
0.006712545175105333,
-0.0023787925019860268,
-0.034200020134449005,
-0.042296912521123886,
-0.11212246119976044,
0.009266091510653496,
0.01802278310060501,
-0.011012709699571133,
0.052590854465961456,
-0.017056431621313095,
0.01603807508945465,
0.15931421518325806,
-0.039860110729932785,
0.07794980704784393,
-0.03389245644211769,
0.016932744532823563,
-0.0487101785838604,
0.1169334426522255,
-0.15169015526771545,
0.00663025863468647,
0.1666049361228943,
0.06880666315555573,
0.11954502761363983,
0.024090943858027458,
-0.10889705270528793,
0.0803801566362381,
0.09752779453992844,
-0.09714475274085999,
-0.010382533073425293,
-0.029842445626854897,
-0.05077598989009857,
0.10155637562274933,
-0.002365934895351529,
0.1972695291042328,
-0.09944852441549301,
0.04274367168545723,
0.0054475730285048485,
0.00012873914965894073,
-0.02746594324707985,
0.08553679287433624,
0.03141888976097107,
0.0072038439102470875,
-0.08020807057619095,
0.07418303936719894,
-0.002940542297437787,
0.022264041006565094,
0.025985099375247955,
0.07080826908349991,
-0.07716336101293564,
-0.010025656782090664,
-0.011744622141122818,
0.20940442383289337,
-0.10714653879404068,
-0.005954008083790541,
-0.11794127523899078,
-0.12231329828500748,
0.018449096009135246,
0.16350921988487244,
0.049284808337688446,
0.03205390274524689,
-0.03870762512087822,
-0.038360677659511566,
-0.0657767727971077,
0.0415608212351799,
0.0455743670463562,
0.06840585172176361,
-0.24581561982631683,
-0.009703245013952255,
0.027007555589079857,
-0.007497615180909634,
-0.07090172916650772,
-0.042931776493787766,
-0.12424822151660919,
-0.02429179847240448,
-0.05454856902360916,
0.11944922059774399,
-0.025623975321650505,
-0.021479638293385506,
0.004044886212795973,
-0.029198044911026955,
-0.013506829738616943,
0.03664661571383476,
-0.032893043011426926,
-0.023084310814738274,
-0.0029477912466973066,
-0.00872444361448288,
-0.0436922125518322,
-0.10598140209913254,
-0.014318812638521194,
-0.08498691767454147,
0.029263371601700783,
-0.0009272302850149572,
-0.07308314740657806,
0.021793996915221214,
-0.20600523054599762,
0.04372995346784592,
0.18137547373771667,
-0.008098607882857323,
0.005238252691924572,
-0.030196962878108025,
0.006245093420147896,
-0.027841169387102127,
0.06456436961889267,
-0.0003592102730181068,
0.06591375917196274,
-0.0804627314209938,
-0.05949690565466881,
-0.05026434734463692,
-0.02139001525938511,
-0.05087758228182793,
0.03658545762300491,
0.12578710913658142,
0.1479327380657196,
0.13985715806484222,
-0.11398666352033615,
0.08274861425161362,
-0.09957148134708405,
-0.005307598039507866,
-0.00316803902387619,
-0.05476774647831917,
0.012079188600182533,
-0.023983296006917953,
-0.017794985324144363,
-0.035985443741083145,
0.13628551363945007,
-0.021533815190196037,
-0.13586582243442535,
-0.01851913146674633,
0.03820221498608589,
-0.06472694128751755,
0.025599969550967216,
0.17170584201812744,
0.02759748511016369,
0.03685925155878067,
-0.1164732277393341,
0.039034709334373474,
0.16031567752361298,
0.08560141175985336,
0.05859524756669998,
0.004794437903910875,
0.044996485114097595,
0.10415461659431458,
0.08182887732982635,
0.02024499513208866,
0.04927076771855354,
0.13218548893928528,
-0.04512936621904373,
0.09530050307512283,
-0.04860490933060646,
0.009282947517931461,
0.10873234272003174,
-0.0372646264731884,
-0.032574865967035294,
0.07393219321966171,
-0.05446444824337959,
-0.07130717486143112,
-0.09393981844186783,
-0.07121126353740692,
-0.14003893733024597,
0.021052874624729156,
-0.050524063408374786,
-0.00530687440186739,
-0.025200514122843742,
0.0443175844848156,
0.05838799104094505,
0.04757940396666527,
-0.024300167337059975,
-0.03986653313040733,
0.11315970122814178,
-0.025783367455005646,
-0.08614547550678253,
0.045022495090961456,
0.026866359636187553,
0.06641175597906113,
-0.0018418427789583802,
-0.04113490879535675,
0.08456914871931076,
0.07918675988912582,
0.015184192918241024,
-0.024443073198199272,
-0.0895819440484047,
-0.020137691870331764,
0.014028104953467846,
0.005418854299932718,
0.14943398535251617,
0.09440691024065018,
-0.057089664041996,
-0.0465204194188118,
0.12006395310163498,
-0.05962083116173744,
-0.015871461480855942,
-0.12768831849098206,
0.07723205536603928,
-0.11577771604061127,
0.0361294262111187,
-0.06352432072162628,
-0.09385763853788376,
-0.04267890751361847,
0.1721668392419815,
0.19318215548992157,
-0.07205932587385178,
0.03672654926776886,
-0.05484435334801674,
-0.006303270347416401,
0.005283446982502937,
0.06800808757543564,
0.03348134830594063,
0.22207047045230865,
-0.04189227148890495,
0.0025824240874499083,
-0.09279385954141617,
-0.05146963521838188,
-0.045852988958358765,
-0.07386578619480133,
-0.0010171178728342056,
-0.047262176871299744,
-0.08262757211923599,
0.03775612264871597,
-0.14955441653728485,
-0.13230860233306885,
0.199258953332901,
-0.17846322059631348,
-0.020945049822330475,
-0.059558842331171036,
0.042158033698797226,
0.04069004952907562,
0.0205673985183239,
-0.08261166512966156,
-0.013337798416614532,
0.028523197397589684,
-0.021592162549495697,
-0.15934202075004578,
-0.009648020379245281,
-0.07408581674098969,
-0.14590181410312653,
0.058233413845300674,
-0.011864174157381058,
0.0915769636631012,
0.04323730990290642,
0.018765060231089592,
-0.08210933208465576,
0.07199398428201675,
-0.038220494985580444,
-0.09182028472423553,
-0.08836804330348969,
0.11282520741224289,
-0.03170490264892578,
0.11498460173606873,
0.026161357760429382,
-0.09033141285181046,
0.013344909064471722,
0.09692980349063873,
-0.0870131403207779,
-0.08466575294733047,
-0.037574514746665955,
-0.0773119330406189,
0.09537110477685928,
0.09085649996995926,
-0.03149992972612381,
0.0069446563720703125,
0.01653466187417507,
0.002026940230280161,
0.021578015759587288,
-0.09083231538534164,
0.050592388957738876,
-0.07775893807411194,
-0.009797858074307442,
0.028886467218399048,
0.037208572030067444,
-0.26692354679107666,
-0.07662384957075119,
-0.1696077436208725,
-0.008669605478644371,
-0.025805475190281868,
0.08991245925426483,
0.21467424929141998,
0.03642063960433006,
0.0008154833340086043,
-0.1791805475950241,
0.04386885091662407,
0.08649415522813797,
-0.05946057289838791,
-0.07728904485702515
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# llama_questioner_DPO_DC_2
This model is a fine-tuned version of [meta-llama/Llama-2-7b-chat-hf](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2169
- Rewards/chosen: -4.2394
- Rewards/rejected: -12.3534
- Rewards/accuracies: 0.9137
- Rewards/margins: 8.1141
- Logps/rejected: -196.4096
- Logps/chosen: -131.2397
- Logits/rejected: 0.1812
- Logits/chosen: 0.1657
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 2
- eval_batch_size: 1
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 4
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 1
### Training results
| Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
|:-------------:|:-----:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
| 0.0194 | 1.0 | 3793 | 0.2169 | -4.2394 | -12.3534 | 0.9137 | 8.1141 | -196.4096 | -131.2397 | 0.1812 | 0.1657 |
### Framework versions
- PEFT 0.7.1
- Transformers 4.36.2
- Pytorch 2.1.2
- Datasets 2.15.0
- Tokenizers 0.15.0 | {"library_name": "peft", "tags": ["trl", "dpo", "generated_from_trainer"], "base_model": "meta-llama/Llama-2-7b-chat-hf", "model-index": [{"name": "llama_questioner_DPO_DC_2", "results": []}]} | null | mazzaqq/llama_questioner_DPO_DC_2 | [
"peft",
"safetensors",
"trl",
"dpo",
"generated_from_trainer",
"base_model:meta-llama/Llama-2-7b-chat-hf",
"region:us"
] | 2024-02-07T08:21:40+00:00 | [] | [] | TAGS
#peft #safetensors #trl #dpo #generated_from_trainer #base_model-meta-llama/Llama-2-7b-chat-hf #region-us
| llama\_questioner\_DPO\_DC\_2
=============================
This model is a fine-tuned version of meta-llama/Llama-2-7b-chat-hf on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 0.2169
* Rewards/chosen: -4.2394
* Rewards/rejected: -12.3534
* Rewards/accuracies: 0.9137
* Rewards/margins: 8.1141
* Logps/rejected: -196.4096
* Logps/chosen: -131.2397
* Logits/rejected: 0.1812
* Logits/chosen: 0.1657
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0002
* train\_batch\_size: 2
* eval\_batch\_size: 1
* seed: 42
* gradient\_accumulation\_steps: 2
* total\_train\_batch\_size: 4
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: cosine
* lr\_scheduler\_warmup\_ratio: 0.1
* num\_epochs: 1
### Training results
### Framework versions
* PEFT 0.7.1
* Transformers 4.36.2
* Pytorch 2.1.2
* Datasets 2.15.0
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 2\n* eval\\_batch\\_size: 1\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 1",
"### Training results",
"### Framework versions\n\n\n* PEFT 0.7.1\n* Transformers 4.36.2\n* Pytorch 2.1.2\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
"TAGS\n#peft #safetensors #trl #dpo #generated_from_trainer #base_model-meta-llama/Llama-2-7b-chat-hf #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 2\n* eval\\_batch\\_size: 1\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 1",
"### Training results",
"### Framework versions\n\n\n* PEFT 0.7.1\n* Transformers 4.36.2\n* Pytorch 2.1.2\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
47,
144,
4,
36
] | [
"passage: TAGS\n#peft #safetensors #trl #dpo #generated_from_trainer #base_model-meta-llama/Llama-2-7b-chat-hf #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 2\n* eval\\_batch\\_size: 1\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 1### Training results### Framework versions\n\n\n* PEFT 0.7.1\n* Transformers 4.36.2\n* Pytorch 2.1.2\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
-0.12815460562705994,
0.07362678647041321,
-0.0030387050937861204,
0.09397165477275848,
0.13555864989757538,
0.008106213994324207,
0.11708234995603561,
0.12782886624336243,
-0.12658336758613586,
0.0822305828332901,
0.12045231461524963,
0.09862366318702698,
0.03968842327594757,
0.21553269028663635,
-0.05761728063225746,
-0.25948667526245117,
0.022048208862543106,
-0.018482090905308723,
-0.08903668075799942,
0.11944059282541275,
0.08411248028278351,
-0.129569873213768,
0.08156486600637436,
-0.02364373952150345,
-0.17302487790584564,
-0.0028785758186131716,
-0.016635721549391747,
-0.036760084331035614,
0.1097821518778801,
-0.006169354543089867,
0.12922990322113037,
0.04526415839791298,
0.10848353058099747,
-0.18451227247714996,
0.007711556740105152,
0.08200474083423615,
0.02123616263270378,
0.08528686314821243,
0.0746433362364769,
-0.01885032281279564,
0.10002977401018143,
-0.10525348782539368,
0.06787969172000885,
0.017859280109405518,
-0.14706026017665863,
-0.2773507535457611,
-0.1126742735505104,
0.04781769961118698,
0.10392667353153229,
0.07026933133602142,
-0.016554437577724457,
0.1615411788225174,
-0.06229695305228233,
0.08477776497602463,
0.30566614866256714,
-0.2768784463405609,
-0.07452896237373352,
0.05150538682937622,
0.0022126506082713604,
0.0921148955821991,
-0.10607907176017761,
-0.02180144190788269,
0.06284728646278381,
0.03495626524090767,
0.11665500700473785,
-0.004495580215007067,
-0.01965779811143875,
-0.0032199432607740164,
-0.14694693684577942,
-0.031375665217638016,
0.13177518546581268,
0.0407976470887661,
-0.03934904560446739,
-0.02005075290799141,
-0.07226730138063431,
-0.21722565591335297,
-0.042063191533088684,
-0.026253847405314445,
0.05780734121799469,
-0.06118124723434448,
-0.07749133557081223,
0.027758704498410225,
-0.07181244343519211,
-0.0886830985546112,
-0.007547182962298393,
0.17631441354751587,
0.05949152261018753,
0.016322962939739227,
-0.023983435705304146,
0.12703554332256317,
-0.02236039564013481,
-0.163008451461792,
-0.013784730806946754,
0.008697688579559326,
-0.03778577595949173,
-0.05532708019018173,
-0.04921555146574974,
-0.01768074370920658,
0.017198756337165833,
0.14920149743556976,
-0.14701946079730988,
0.0730580985546112,
-0.0004137398791499436,
0.017353223636746407,
-0.07049303501844406,
0.12601104378700256,
-0.0764818862080574,
0.007830953225493431,
-0.008835075423121452,
0.0996699184179306,
0.025446465238928795,
-0.0015814426587894559,
-0.07024600356817245,
0.043493255972862244,
0.11381616443395615,
0.035837266594171524,
-0.05900844186544418,
0.03629574179649353,
-0.041540563106536865,
-0.01660413108766079,
0.05454833060503006,
-0.11661184579133987,
0.03598529100418091,
0.021459167823195457,
-0.09227722138166428,
-0.03405699506402016,
0.0026341930497437716,
0.010352976620197296,
0.006624537520110607,
0.12599864602088928,
-0.0888635441660881,
0.027860142290592194,
-0.07448730617761612,
-0.11188007146120071,
0.010479873977601528,
-0.06533174961805344,
-0.0023890468291938305,
-0.0943414717912674,
-0.14688575267791748,
-0.035020969808101654,
0.04058248549699783,
-0.06926972419023514,
-0.02361157350242138,
-0.06240348890423775,
-0.10413254797458649,
0.011004723608493805,
-0.014903126284480095,
0.1060418114066124,
-0.07722140103578568,
0.1097428947687149,
0.031839679926633835,
0.07029137760400772,
-0.022768449038267136,
0.02927134931087494,
-0.08040245622396469,
0.04916730150580406,
-0.270273357629776,
0.04720216989517212,
-0.06263130903244019,
0.06022047996520996,
-0.10419008135795593,
-0.10047989338636398,
-0.015556419268250465,
-0.03069470450282097,
0.11300768703222275,
0.14241379499435425,
-0.1824197620153427,
-0.06024164706468582,
0.21517229080200195,
-0.10305672883987427,
-0.12336238473653793,
0.11052368581295013,
-0.04122283309698105,
0.0017800128553062677,
0.057879578322172165,
0.2028845101594925,
0.04355251416563988,
-0.13892091810703278,
-0.018141580745577812,
-0.062254440039396286,
0.07994980365037918,
-0.009232563897967339,
0.07678795605897903,
-0.009582139551639557,
0.029621116816997528,
0.0079771364107728,
-0.027956360951066017,
0.033886928111314774,
-0.12023397535085678,
-0.07479296624660492,
-0.03418627381324768,
-0.09079082310199738,
0.03611348196864128,
0.05484182760119438,
0.056624092161655426,
-0.12424726784229279,
-0.08512856811285019,
0.04817549139261246,
0.10918882489204407,
-0.05773639306426048,
0.03348131477832794,
-0.056921862065792084,
0.09246339648962021,
-0.05475451052188873,
-0.021429602056741714,
-0.1792852282524109,
-0.07011336088180542,
0.024861305952072144,
0.015804778784513474,
-0.011052154935896397,
-0.045626603066921234,
0.08974310755729675,
0.08939719945192337,
-0.07660000771284103,
-0.035662390291690826,
-0.06498496234416962,
-0.003246423089876771,
-0.1182502806186676,
-0.24159957468509674,
-0.04132038727402687,
-0.041814856231212616,
0.1352064609527588,
-0.21116849780082703,
0.026612496003508568,
0.018731703981757164,
0.13644951581954956,
0.038687873631715775,
-0.04831773042678833,
-0.021129218861460686,
0.06027855724096298,
-0.018056336790323257,
-0.08334333449602127,
0.027914294973015785,
-0.001622346811927855,
-0.05410292372107506,
-0.036345869302749634,
-0.13904449343681335,
0.14976435899734497,
0.12125619500875473,
0.001929301070049405,
-0.12095958739519119,
-0.025295160710811615,
-0.07803943008184433,
-0.03386100381612778,
-0.06692134588956833,
0.029338320717215538,
0.11991428583860397,
0.009690065868198872,
0.12601053714752197,
-0.10093986988067627,
-0.043349891901016235,
0.03855733945965767,
-0.029554393142461777,
0.023165956139564514,
0.11592953652143478,
0.06207852438092232,
-0.09944874048233032,
0.1358334720134735,
0.1304982602596283,
-0.05251271650195122,
0.11702899634838104,
-0.06323662400245667,
-0.09393627941608429,
-0.035723794251680374,
0.025334354490041733,
0.01004878431558609,
0.15680791437625885,
-0.024813560768961906,
0.016271300613880157,
0.006461238954216242,
0.03942129388451576,
0.006841535214334726,
-0.19934433698654175,
-0.031583819538354874,
0.024883560836315155,
-0.05893126502633095,
-0.0342080295085907,
-0.03319765627384186,
0.02528122067451477,
0.1258375346660614,
0.011791164055466652,
-0.07973025739192963,
-0.008969657123088837,
0.010489558801054955,
-0.07015760242938995,
0.20755191147327423,
-0.08896929025650024,
-0.08596935868263245,
-0.046116676181554794,
-0.0017099056858569384,
-0.04548783600330353,
0.002444992773234844,
0.05327117070555687,
-0.10339123010635376,
-0.018679017201066017,
-0.10273021459579468,
-0.010030215606093407,
0.01395424921065569,
0.04528404772281647,
0.02880004420876503,
-0.002601724350824952,
0.0786251425743103,
-0.09363963454961777,
-0.004730171523988247,
-0.05757514387369156,
-0.022427493706345558,
0.04754025861620903,
0.030942430719733238,
0.09619182348251343,
0.15392516553401947,
-0.008661038242280483,
0.0454120896756649,
-0.03301789611577988,
0.2113577425479889,
-0.07959870249032974,
-0.03078470192849636,
0.06400551646947861,
0.00822838582098484,
0.06297493726015091,
0.1447756141424179,
0.05272936448454857,
-0.12036184221506119,
0.0162550937384367,
0.050010547041893005,
-0.04208998382091522,
-0.20404702425003052,
-0.04205765202641487,
-0.02091904915869236,
0.0026556134689599276,
0.09308330714702606,
0.0344749353826046,
-0.005907199811190367,
0.03212855011224747,
0.0021745767444372177,
0.005430043209344149,
-0.008567546494305134,
0.07183487713336945,
-0.004341259133070707,
0.052557170391082764,
0.11070147156715393,
-0.03358035907149315,
-0.04111628234386444,
0.03606261685490608,
-0.016818320378661156,
0.2205241471529007,
-0.03713330999016762,
0.06703437119722366,
0.038300395011901855,
0.1756211817264557,
-0.027401411905884743,
0.07059792429208755,
0.02310744673013687,
-0.04424504190683365,
0.0025211486499756575,
-0.06273103505373001,
-0.009984423406422138,
0.033570583909749985,
-0.03669935092329979,
0.04501033201813698,
-0.13596589863300323,
0.02252715639770031,
0.06647506356239319,
0.2970607280731201,
0.08111522346735,
-0.35316991806030273,
-0.06717296689748764,
-0.0019717898685485125,
0.007061347831040621,
-0.020522553473711014,
0.011685597710311413,
0.1265006959438324,
-0.060448478907346725,
0.07698798179626465,
-0.06222688779234886,
0.07297693192958832,
-0.024876505136489868,
0.020728852599859238,
0.0551469549536705,
0.09852895885705948,
-0.03157133236527443,
0.028491223230957985,
-0.2416187971830368,
0.2969662845134735,
0.023716973140835762,
0.07516492903232574,
-0.03534187376499176,
-0.011764745227992535,
0.016516992822289467,
0.06021052598953247,
0.06861584633588791,
-0.005756249651312828,
-0.0836476981639862,
-0.20504635572433472,
-0.11460769921541214,
0.026673728600144386,
0.13646887242794037,
-0.045434463769197464,
0.11786063015460968,
0.004144873935729265,
-0.004803422372788191,
0.05026169493794441,
-0.05300217866897583,
-0.09712270647287369,
-0.029423531144857407,
0.007308517582714558,
0.0022929315455257893,
-0.017877379432320595,
-0.07577882707118988,
-0.10772490501403809,
-0.08310288190841675,
0.14812806248664856,
0.026623573154211044,
-0.04693519324064255,
-0.13634620606899261,
0.09256470948457718,
0.10419019311666489,
-0.07355309277772903,
0.018275946378707886,
0.03166705742478371,
0.07202830165624619,
0.026637282222509384,
-0.04928148165345192,
0.13472944498062134,
-0.05293550342321396,
-0.21074381470680237,
-0.051014117896556854,
0.13363271951675415,
0.0691148042678833,
0.05008889362215996,
-0.020309844985604286,
0.03112432360649109,
0.029963841661810875,
-0.085725337266922,
0.05170325189828873,
-0.0004115856718271971,
0.08093354105949402,
0.04338978976011276,
-0.017964035272598267,
0.058536361902952194,
-0.06441080570220947,
-0.01369527354836464,
0.10776825249195099,
0.35191088914871216,
-0.08643871545791626,
0.02405436895787716,
0.03269442543387413,
-0.04075344651937485,
-0.1692400723695755,
0.04445899650454521,
0.08584247529506683,
-0.010424620471894741,
0.036294806748628616,
-0.16120392084121704,
0.05763848498463631,
0.11954793334007263,
-0.031708378344774246,
0.10893747955560684,
-0.3269921839237213,
-0.13133926689624786,
0.10111726820468903,
0.13157090544700623,
0.06944505870342255,
-0.17483343183994293,
-0.03105556219816208,
-0.006320709828287363,
-0.10358957946300507,
0.08531705290079117,
-0.152948260307312,
0.10085760802030563,
-0.01650884561240673,
0.038813136518001556,
0.028114542365074158,
-0.05303697660565376,
0.14387701451778412,
0.015934854745864868,
0.11308296024799347,
-0.043208763003349304,
0.02002156525850296,
0.04864125698804855,
-0.07886751741170883,
0.036757685244083405,
-0.036159541457891464,
0.04550966992974281,
-0.07088126987218857,
-0.004681841004639864,
-0.09499665349721909,
0.014393148943781853,
-0.03376278281211853,
-0.057465147227048874,
-0.05384421721100807,
0.035961467772722244,
0.05839407444000244,
-0.01713414490222931,
0.162702277302742,
0.011967498809099197,
0.18008072674274445,
0.14592105150222778,
0.040928762406110764,
-0.05801137909293175,
-0.02226145751774311,
0.022254807874560356,
-0.01538444310426712,
0.047447569668293,
-0.1620238572359085,
0.042768243700265884,
0.13807545602321625,
0.03127158805727959,
0.1066330075263977,
0.0736905038356781,
-0.07266639173030853,
0.004127119667828083,
0.06705775111913681,
-0.16949105262756348,
-0.100782111287117,
0.012248777784407139,
-0.00808768067508936,
-0.12335273623466492,
0.08440060913562775,
0.12282007187604904,
-0.06924551725387573,
-0.01193036325275898,
-0.008237814530730247,
0.03891265019774437,
-0.04840519279241562,
0.23856417834758759,
0.05964931845664978,
0.06491803377866745,
-0.10845424234867096,
0.10216810554265976,
0.04262043535709381,
-0.06997863203287125,
0.04382619634270668,
0.09299077838659286,
-0.08116447180509567,
-0.0100668640807271,
0.09421084821224213,
0.14896337687969208,
-0.023071542382240295,
-0.036038853228092194,
-0.1659618616104126,
-0.12102407217025757,
0.0785081535577774,
0.16386345028877258,
0.07043556123971939,
0.025105619803071022,
0.014056647196412086,
0.018948694691061974,
-0.14484372735023499,
0.09154517203569412,
0.0684390738606453,
0.09156564623117447,
-0.1449739784002304,
0.15543211996555328,
-0.019733870401978493,
0.03957009315490723,
-0.0239461250603199,
0.03077472187578678,
-0.13504792749881744,
0.02073228545486927,
-0.11655038595199585,
-0.0315263532102108,
-0.03757891803979874,
-0.013425040990114212,
-0.0033347143325954676,
-0.05500873178243637,
-0.054496560245752335,
0.015860963612794876,
-0.1252361536026001,
-0.029628999531269073,
0.008457738906145096,
0.015636898577213287,
-0.1449986845254898,
-0.038555510342121124,
0.011418025940656662,
-0.08968399465084076,
0.08509315550327301,
0.06849101185798645,
0.03280850499868393,
0.03696967288851738,
-0.09866730868816376,
0.009557805955410004,
0.06770771741867065,
-0.031676117330789566,
0.07513058185577393,
-0.11819200217723846,
-0.028475970029830933,
-0.028662947937846184,
0.04661983996629715,
0.029181301593780518,
0.09069652110338211,
-0.12622231245040894,
0.013107101432979107,
-0.03656245768070221,
-0.05468517914414406,
-0.036917999386787415,
0.02026619389653206,
0.07088299840688705,
0.027873264625668526,
0.12713198363780975,
-0.08622768521308899,
0.02876957505941391,
-0.2276880145072937,
-0.02901955135166645,
-0.005835325922816992,
-0.08154791593551636,
-0.0492398776113987,
-0.0051063005812466145,
0.09854290634393692,
-0.060456376522779465,
0.08630543947219849,
-0.004568955861032009,
0.05878615379333496,
0.043291933834552765,
-0.07112406194210052,
0.014037096872925758,
0.05036575347185135,
0.1540221869945526,
0.025709278881549835,
-0.04968743026256561,
0.07675717025995255,
0.04353906586766243,
0.11523886770009995,
0.10509803891181946,
0.25196948647499084,
0.13869813084602356,
0.011442589573562145,
0.08427704870700836,
0.036153752356767654,
-0.11623647063970566,
-0.13100402057170868,
0.04825935140252113,
-0.05187327414751053,
0.0917176827788353,
-0.011983255855739117,
0.15192191302776337,
0.11203102022409439,
-0.17799167335033417,
0.016204513609409332,
-0.05278609320521355,
-0.09110420197248459,
-0.12278232723474503,
-0.003140074899420142,
-0.08666501194238663,
-0.1872594654560089,
0.0043540834449231625,
-0.11834164708852768,
0.03118344396352768,
0.11839000880718231,
0.021572742611169815,
0.03418305888772011,
0.17317081987857819,
0.04816411808133125,
0.03477327525615692,
0.05136646330356598,
0.02258443273603916,
-0.021645013242959976,
-0.034270983189344406,
-0.09942011535167694,
0.029359301552176476,
-0.02785157784819603,
0.0350375659763813,
-0.04839786887168884,
-0.05835321918129921,
0.06064651161432266,
0.009766141884028912,
-0.10928505659103394,
0.024540336802601814,
0.023166488856077194,
0.061001043766736984,
0.031085684895515442,
0.025375615805387497,
0.006509930361062288,
-0.025012196972966194,
0.23640796542167664,
-0.08224663883447647,
-0.03464516997337341,
-0.12000130116939545,
0.2850354313850403,
0.04730920121073723,
-0.00098715303465724,
0.031289998441934586,
-0.10772386938333511,
-0.022356020286679268,
0.13448068499565125,
0.17755542695522308,
-0.06366295367479324,
-0.002959158970043063,
-0.034354325383901596,
-0.011339518241584301,
-0.03426269441843033,
0.09860413521528244,
0.11920071393251419,
0.012937894091010094,
-0.09133054316043854,
-0.020140329375863075,
-0.04232707619667053,
-0.02350965142250061,
-0.041769567877054214,
0.042074497789144516,
0.03105233982205391,
0.02049855701625347,
-0.056167446076869965,
0.07927470654249191,
-0.04695768654346466,
-0.11457512527704239,
0.07317181676626205,
-0.20763865113258362,
-0.17215824127197266,
-0.02103712037205696,
0.04049118235707283,
-0.0019640550017356873,
0.07277167588472366,
-0.029221363365650177,
-0.023969219997525215,
0.10306789726018906,
-0.026912668719887733,
-0.059487052261829376,
-0.1524602174758911,
0.07212530821561813,
-0.1389092057943344,
0.21362940967082977,
-0.03854179009795189,
0.03644988313317299,
0.11903737485408783,
0.025389844551682472,
-0.10777517408132553,
0.04074074327945709,
0.061650991439819336,
-0.09115119278430939,
-0.009466529823839664,
0.1465923935174942,
-0.056727681308984756,
0.08162271231412888,
0.049476027488708496,
-0.12992402911186218,
-0.01701592467725277,
-0.039825424551963806,
-0.02819862775504589,
-0.049031730741262436,
-0.04176529496908188,
-0.04924864321947098,
0.12241266667842865,
0.1978878676891327,
-0.04895790293812752,
0.05311164632439613,
-0.05462437868118286,
0.034980323165655136,
0.06190609559416771,
0.0390169695019722,
-0.02757609635591507,
-0.2832527458667755,
0.03600379452109337,
0.10167541354894638,
-0.007862573489546776,
-0.22686976194381714,
-0.07376221567392349,
0.03206394612789154,
-0.05985661968588829,
-0.07733139395713806,
0.10214199125766754,
0.043869633227586746,
0.059335626661777496,
-0.06470576673746109,
-0.08997514098882675,
-0.07297512143850327,
0.1777927130460739,
-0.15683118999004364,
-0.09433213621377945
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | muzammil-eds/tinyllama-3T-64k-JSONExtractor-v4 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T08:21:55+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
56,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06061961501836777,
0.15481999516487122,
-0.004844071343541145,
0.02074851468205452,
0.0983177199959755,
0.007407687604427338,
0.07119518518447876,
0.11185134947299957,
-0.023851769044995308,
0.1167980208992958,
0.031993988901376724,
0.09781743586063385,
0.11217817664146423,
0.16186554729938507,
0.0015333457849919796,
-0.22897611558437347,
0.049678247421979904,
-0.125278040766716,
-0.0294334813952446,
0.11977242678403854,
0.1422213912010193,
-0.10954539477825165,
0.0752737894654274,
-0.038042325526475906,
-0.005828251596540213,
-0.0323176346719265,
-0.06205610930919647,
-0.05266609415411949,
0.05311284959316254,
0.06794639676809311,
0.07308239489793777,
0.01171939354389906,
0.09106900542974472,
-0.2724283039569855,
0.02348201349377632,
0.0805930644273758,
-0.0006441773730330169,
0.07586129754781723,
0.04993962123990059,
-0.08749990910291672,
0.07524524629116058,
-0.060156844556331635,
0.1498761922121048,
0.07955671846866608,
-0.09018243104219437,
-0.19217631220817566,
-0.07921334356069565,
0.09916994720697403,
0.1890910118818283,
0.05953684076666832,
-0.026427440345287323,
0.11642678081989288,
-0.08593545109033585,
0.013638701289892197,
0.06446459144353867,
-0.06054406240582466,
-0.055855002254247665,
0.06904532760381699,
0.08335285633802414,
0.08567540347576141,
-0.12976622581481934,
-0.010767064057290554,
0.015032444149255753,
0.008952446281909943,
0.08948688954114914,
0.017146794125437737,
0.1335189938545227,
0.040557652711868286,
-0.13501930236816406,
-0.043155476450920105,
0.09761431813240051,
0.03665134683251381,
-0.04888195917010307,
-0.2485782504081726,
-0.023432478308677673,
-0.04339504987001419,
-0.03198111802339554,
-0.03649339824914932,
0.043764639645814896,
-0.014506848528981209,
0.07738617807626724,
-0.004502781666815281,
-0.0837155357003212,
-0.04301247000694275,
0.07241875678300858,
0.06128999963402748,
0.02571401372551918,
-0.015821760520339012,
0.0059297760017216206,
0.12327717989683151,
0.11431120336055756,
-0.126715749502182,
-0.052547648549079895,
-0.06306339055299759,
-0.08449548482894897,
-0.044861067086458206,
0.030838407576084137,
0.037995077669620514,
0.045936476439237595,
0.23867325484752655,
0.007765117567032576,
0.053257301449775696,
0.04455438256263733,
0.014407169073820114,
0.06501194834709167,
0.11008983850479126,
-0.05894824117422104,
-0.09719445556402206,
-0.028582042083144188,
0.10156717151403427,
0.007986726239323616,
-0.04139331728219986,
-0.05712985619902611,
0.07059531658887863,
0.018587570637464523,
0.12360043078660965,
0.08000938594341278,
0.003056557849049568,
-0.0755772516131401,
-0.062465377151966095,
0.17764076590538025,
-0.15825673937797546,
0.04532013460993767,
0.03055616281926632,
-0.0341108962893486,
-0.009745313785970211,
0.012105142697691917,
0.025474950671195984,
-0.021481726318597794,
0.09522198140621185,
-0.05601342022418976,
-0.034448131918907166,
-0.11389608681201935,
-0.03694311901926994,
0.030394554138183594,
0.011153047904372215,
-0.02865210548043251,
-0.03502652049064636,
-0.08865131437778473,
-0.06405586749315262,
0.09101516753435135,
-0.07148737460374832,
-0.04784895107150078,
-0.016645915806293488,
-0.07833752781152725,
0.021804187446832657,
0.01691517047584057,
0.09064167737960815,
-0.0222476739436388,
0.03985358029603958,
-0.0550384595990181,
0.061440225690603256,
0.11723454296588898,
0.027987057343125343,
-0.05787884071469307,
0.061519939452409744,
-0.2424532175064087,
0.10252492874860764,
-0.07715212553739548,
0.04971238598227501,
-0.15203025937080383,
-0.02478341944515705,
0.03986154496669769,
0.01284773275256157,
-0.008251311257481575,
0.14196595549583435,
-0.21994100511074066,
-0.030957341194152832,
0.16964265704154968,
-0.10025953501462936,
-0.08109250664710999,
0.060782887041568756,
-0.05354252830147743,
0.11210215091705322,
0.04557164013385773,
-0.02375967986881733,
0.05775221437215805,
-0.14725260436534882,
-0.011030761525034904,
-0.041942402720451355,
-0.0180682260543108,
0.16207332909107208,
0.0703711211681366,
-0.06047816202044487,
0.07456906884908676,
0.01960151270031929,
-0.014246034435927868,
-0.04887177795171738,
-0.02822130173444748,
-0.1047162413597107,
0.01184528972953558,
-0.06102835759520531,
0.018109694123268127,
-0.021768750622868538,
-0.09445013850927353,
-0.029118487611413002,
-0.17402999103069305,
-0.0031633328180760145,
0.08821269869804382,
-0.011630427092313766,
-0.021509924903512,
-0.11245372891426086,
0.009332616813480854,
0.030967719852924347,
0.0002618339203763753,
-0.13677829504013062,
-0.06033218279480934,
0.026970699429512024,
-0.16097871959209442,
0.029791243374347687,
-0.05741601809859276,
0.04530094936490059,
0.04005871340632439,
-0.03433511033654213,
-0.03489551320672035,
0.010874404571950436,
0.010431389324367046,
-0.01894843392074108,
-0.25422003865242004,
-0.01882786676287651,
-0.0234990194439888,
0.1751047968864441,
-0.22956320643424988,
0.042598169296979904,
0.07489731162786484,
0.1460893303155899,
0.007349682506173849,
-0.03550100699067116,
0.015185600146651268,
-0.07262228429317474,
-0.03268764168024063,
-0.06316669285297394,
-0.01207790058106184,
-0.038400664925575256,
-0.05820201337337494,
0.04906858503818512,
-0.1686294972896576,
-0.030321966856718063,
0.10717973858118057,
0.06342670321464539,
-0.1473218947649002,
-0.02780107781291008,
-0.04056945815682411,
-0.04624456167221069,
-0.06676914542913437,
-0.05461418256163597,
0.11812574416399002,
0.056411582976579666,
0.04860803112387657,
-0.07140495628118515,
-0.07455260306596756,
0.008036690764129162,
-0.01956399530172348,
-0.014917809516191483,
0.09334591031074524,
0.07554110884666443,
-0.12264352291822433,
0.09177418053150177,
0.09668384492397308,
0.08576478064060211,
0.10314212739467621,
-0.014663571491837502,
-0.08914592862129211,
-0.040637146681547165,
0.02245822176337242,
0.016187267377972603,
0.15129362046718597,
-0.012961224652826786,
0.055492039769887924,
0.0358695350587368,
-0.014034898020327091,
0.011105312965810299,
-0.09736533463001251,
0.02655916102230549,
0.030835967510938644,
-0.016302183270454407,
0.03745110332965851,
-0.0447014644742012,
0.019208140671253204,
0.09039704501628876,
0.040895868092775345,
0.040978945791721344,
0.010155045427381992,
-0.04354988783597946,
-0.11037563532590866,
0.1787576973438263,
-0.12389461696147919,
-0.24818050861358643,
-0.13812170922756195,
0.010281167924404144,
0.04737642779946327,
-0.010411068797111511,
0.006690691225230694,
-0.06616118550300598,
-0.1175973042845726,
-0.09878289699554443,
0.018617089837789536,
0.045352302491664886,
-0.07590975612401962,
-0.06842505931854248,
0.06414616107940674,
0.03875524550676346,
-0.13939815759658813,
0.024007495492696762,
0.04662325978279114,
-0.08205481618642807,
-0.0029386086389422417,
0.0791812464594841,
0.06965780258178711,
0.17661017179489136,
0.013885351829230785,
-0.023669935762882233,
0.026634456589818,
0.20819635689258575,
-0.1436755359172821,
0.10975687950849533,
0.13545554876327515,
-0.08767466992139816,
0.08120133727788925,
0.1998777538537979,
0.03777998685836792,
-0.10680917650461197,
0.03608465939760208,
0.028374753892421722,
-0.028325283899903297,
-0.2502254545688629,
-0.06958996504545212,
0.0019060121849179268,
-0.05172049254179001,
0.07064855098724365,
0.08791537582874298,
0.09593888372182846,
0.016860228031873703,
-0.09976044297218323,
-0.07697858661413193,
0.046900223940610886,
0.10824491083621979,
-0.00015424020239152014,
-0.015208319760859013,
0.0904119610786438,
-0.03033481352031231,
0.01743943803012371,
0.09215071052312851,
0.0030607767403125763,
0.17535938322544098,
0.051709048449993134,
0.17189906537532806,
0.07866133749485016,
0.06444311141967773,
0.02004685252904892,
0.007725914940237999,
0.021817529574036598,
0.017227526754140854,
-0.0030957073904573917,
-0.08709781616926193,
-0.0034981227945536375,
0.1202581599354744,
0.049845851957798004,
0.029173865914344788,
0.012042860500514507,
-0.030704669654369354,
0.08337877690792084,
0.1770893782377243,
0.0029054484330117702,
-0.1893385946750641,
-0.07169844210147858,
0.07795937359333038,
-0.08648337423801422,
-0.10729733109474182,
-0.029470939189195633,
0.041069481521844864,
-0.1729043871164322,
0.016882894560694695,
-0.019335895776748657,
0.10788324475288391,
-0.13190391659736633,
-0.01772487722337246,
0.05657728388905525,
0.06932812184095383,
-0.009677323512732983,
0.06694949418306351,
-0.16090403497219086,
0.11770165711641312,
0.01751571334898472,
0.06636732816696167,
-0.09608277678489685,
0.09618937969207764,
-0.007830657996237278,
0.0041499207727611065,
0.1410749852657318,
0.010120149701833725,
-0.05952107161283493,
-0.09608154743909836,
-0.10546442121267319,
-0.009841260500252247,
0.1306990385055542,
-0.14852415025234222,
0.08813067525625229,
-0.02661319263279438,
-0.044553373008966446,
0.003614129964262247,
-0.12497276812791824,
-0.13103094696998596,
-0.18366187810897827,
0.05707118660211563,
-0.12947207689285278,
0.04045100137591362,
-0.10902881622314453,
-0.045833900570869446,
-0.02098964899778366,
0.20040063560009003,
-0.23137451708316803,
-0.06714103370904922,
-0.1551055610179901,
-0.08061286807060242,
0.14446212351322174,
-0.046455029398202896,
0.08550118654966354,
0.0008278203313238919,
0.19068008661270142,
0.021319707855582237,
-0.017237508669495583,
0.1072206199169159,
-0.10052918642759323,
-0.2010865956544876,
-0.09273224323987961,
0.15895552933216095,
0.13766798377037048,
0.03809428587555885,
-0.004381525795906782,
0.03171157464385033,
-0.02098114788532257,
-0.12076930701732635,
0.020226983353495598,
0.17317426204681396,
0.08982043713331223,
0.025265544652938843,
-0.02972041629254818,
-0.11267432570457458,
-0.07061342149972916,
-0.03774050623178482,
0.024755435064435005,
0.18072067201137543,
-0.07222156971693039,
0.18405316770076752,
0.13775517046451569,
-0.05534014105796814,
-0.19904261827468872,
0.021996473893523216,
0.04293542355298996,
0.0070380112156271935,
0.0323902890086174,
-0.20307663083076477,
0.09384101629257202,
0.0008334947633557022,
-0.05131231248378754,
0.1379684954881668,
-0.1823476254940033,
-0.151598259806633,
0.06042521819472313,
0.043563615530729294,
-0.19374065101146698,
-0.12374074012041092,
-0.08848230540752411,
-0.04693066328763962,
-0.15487661957740784,
0.10312657803297043,
0.0020827590487897396,
0.008401188999414444,
0.03778626397252083,
0.02252252586185932,
0.012139533646404743,
-0.04198719933629036,
0.1914343535900116,
-0.025891713798046112,
0.03347287327051163,
-0.0790715217590332,
-0.060851071029901505,
0.062408581376075745,
-0.058187782764434814,
0.0755455270409584,
-0.025226406753063202,
0.015947066247463226,
-0.10598332434892654,
-0.048235729336738586,
-0.02852320298552513,
0.019321219995617867,
-0.09431382268667221,
-0.09348297864198685,
-0.04829427972435951,
0.09367614984512329,
0.09042316675186157,
-0.03652578964829445,
-0.03649144619703293,
-0.078715980052948,
0.038977332413196564,
0.17627815902233124,
0.18159319460391998,
0.04659178853034973,
-0.07959239184856415,
-0.001915142871439457,
-0.014336181804537773,
0.04684065282344818,
-0.22077152132987976,
0.060553863644599915,
0.04557652771472931,
0.016117896884679794,
0.11537692695856094,
-0.0208132341504097,
-0.16198977828025818,
-0.06710557639598846,
0.061360616236925125,
-0.06944561004638672,
-0.17825035750865936,
0.0039279889315366745,
0.07344977557659149,
-0.16578389704227448,
-0.037031736224889755,
0.04200848564505577,
-0.01189455483108759,
-0.0403641052544117,
0.012352054007351398,
0.08063354343175888,
0.007078902795910835,
0.07699975371360779,
0.055281639099121094,
0.09124495089054108,
-0.10227900743484497,
0.07410510629415512,
0.08149529248476028,
-0.08644098788499832,
0.030720343813300133,
0.09573426842689514,
-0.06469762325286865,
-0.0346054881811142,
0.04237886518239975,
0.08354541659355164,
0.024281201884150505,
-0.04682289808988571,
0.0023111123591661453,
-0.09734189510345459,
0.05927345156669617,
0.11483542621135712,
0.03496333956718445,
0.011234734207391739,
0.03813567012548447,
0.04486291855573654,
-0.08093374222517014,
0.11926916986703873,
0.023795632645487785,
0.020354853942990303,
-0.04112942889332771,
-0.040553025901317596,
0.035851649940013885,
-0.026020776480436325,
-0.011440055444836617,
-0.035174157470464706,
-0.0722682997584343,
-0.014069457538425922,
-0.16000694036483765,
-0.0076758842915296555,
-0.03660871088504791,
0.005114538595080376,
0.022510098293423653,
-0.03652830421924591,
0.00792311318218708,
0.012217256240546703,
-0.06868947297334671,
-0.05553458258509636,
-0.023233558982610703,
0.09422210603952408,
-0.16494666039943695,
0.0220257006585598,
0.0823851153254509,
-0.12121747434139252,
0.09289738535881042,
0.016782134771347046,
0.00412249518558383,
0.026962365955114365,
-0.1545863002538681,
0.04763968288898468,
-0.020152103155851364,
0.013473534025251865,
0.04222847521305084,
-0.21637047827243805,
-0.004404853098094463,
-0.04015503451228142,
-0.05566934496164322,
-0.008993052877485752,
-0.0319182425737381,
-0.11338426172733307,
0.09645436704158783,
0.011025024577975273,
-0.08443772792816162,
-0.02965564839541912,
0.03353232145309448,
0.07690354436635971,
-0.027447547763586044,
0.1498211771249771,
-0.004663881380110979,
0.07559948414564133,
-0.17581342160701752,
-0.02282017655670643,
-0.011197620071470737,
0.022367527708411217,
-0.021871577948331833,
-0.01622559316456318,
0.04623444378376007,
-0.02704801969230175,
0.19120801985263824,
-0.024701936170458794,
0.049393873661756516,
0.06364397704601288,
0.009232889860868454,
-0.013832193799316883,
0.11151392012834549,
0.05708572641015053,
0.024334950372576714,
0.022262847051024437,
0.003451440716162324,
-0.04008655622601509,
-0.009981024079024792,
-0.18596695363521576,
0.06803664565086365,
0.14585918188095093,
0.09060460329055786,
-0.012669353745877743,
0.0707244873046875,
-0.10161512345075607,
-0.12005364894866943,
0.10127941519021988,
-0.06415384262800217,
-0.010188822634518147,
-0.06542414426803589,
0.14027701318264008,
0.14953285455703735,
-0.1886233240365982,
0.06583356112241745,
-0.06602055579423904,
-0.0566304549574852,
-0.11457879096269608,
-0.1930263340473175,
-0.057075321674346924,
-0.050602465867996216,
-0.018466074019670486,
-0.05384097993373871,
0.06939727067947388,
0.05750798434019089,
0.01126816775649786,
0.00868057832121849,
0.08568526059389114,
-0.009656033478677273,
0.00248199631460011,
0.030120067298412323,
0.06713981181383133,
0.016768986359238625,
-0.0321255661547184,
0.0179112758487463,
-0.00597198773175478,
0.034156378358602524,
0.059282708913087845,
0.03608176112174988,
-0.028436895459890366,
0.015559280291199684,
-0.034912437200546265,
-0.11309733241796494,
0.042801856994628906,
-0.029640642926096916,
-0.0749855786561966,
0.1347348988056183,
0.026981467381119728,
0.005015076603740454,
-0.023140020668506622,
0.2503887414932251,
-0.07436972856521606,
-0.09334370493888855,
-0.14373961091041565,
0.11701542884111404,
-0.04212593287229538,
0.0635172426700592,
0.03596310690045357,
-0.10810714215040207,
0.017985546961426735,
0.1320217251777649,
0.15442703664302826,
-0.04732590913772583,
0.019251897931098938,
0.028577854856848717,
0.00439635943621397,
-0.04075566306710243,
0.05177190154790878,
0.07100846618413925,
0.14500564336776733,
-0.05157303810119629,
0.08530787378549576,
0.002609728369861841,
-0.1021018698811531,
-0.041973695158958435,
0.11415864527225494,
-0.014296893030405045,
0.017620453611016273,
-0.057136841118335724,
0.124222531914711,
-0.05874236673116684,
-0.23697422444820404,
0.06316976249217987,
-0.0765061303973198,
-0.1432730257511139,
-0.024886758998036385,
0.071670763194561,
-0.016632623970508575,
0.02605951391160488,
0.07167234271764755,
-0.0754380151629448,
0.18880942463874817,
0.03957989811897278,
-0.05233397334814072,
-0.05954399332404137,
0.0744764655828476,
-0.11850855499505997,
0.27879106998443604,
0.010482731275260448,
0.051307905465364456,
0.1042102724313736,
-0.02021743729710579,
-0.13270841538906097,
0.023401619866490364,
0.09579801559448242,
-0.08917027711868286,
0.04087764397263527,
0.21448291838169098,
-0.00629545608535409,
0.11935057491064072,
0.07611140608787537,
-0.07468950748443604,
0.047562725841999054,
-0.11468592286109924,
-0.07639975845813751,
-0.08699081838130951,
0.09244474768638611,
-0.06785612553358078,
0.14258281886577606,
0.12599852681159973,
-0.05530165135860443,
0.011584274470806122,
-0.028389399871230125,
0.045467376708984375,
0.005578654818236828,
0.100032277405262,
0.011115525849163532,
-0.18496567010879517,
0.024811718612909317,
0.016259413212537766,
0.10884406417608261,
-0.18112654983997345,
-0.09105053544044495,
0.046958595514297485,
0.0005061255069449544,
-0.06443515419960022,
0.12483241409063339,
0.057313691824674606,
0.04654949903488159,
-0.0451689288020134,
-0.026830285787582397,
-0.006042256020009518,
0.14264579117298126,
-0.10707559436559677,
-0.005129707511514425
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# wav2vec_RTSplit0207_6
This model is a fine-tuned version of [jonatasgrosman/wav2vec2-large-xlsr-53-japanese](https://huggingface.co/jonatasgrosman/wav2vec2-large-xlsr-53-japanese) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0453
- Wer: 0.2160
- Cer: 0.1838
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 32
- eval_batch_size: 32
- seed: 4
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 1000
- num_epochs: 10
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|
| 6.369 | 1.0 | 120 | 5.2384 | 1.0 | 0.8910 |
| 1.1292 | 2.0 | 240 | 0.9018 | 0.9092 | 0.5706 |
| 0.7847 | 3.0 | 360 | 0.6675 | 0.8210 | 0.5470 |
| 0.689 | 4.0 | 480 | 0.6056 | 0.8203 | 0.5506 |
| 0.5764 | 5.0 | 600 | 0.4676 | 0.7040 | 0.4298 |
| 0.4722 | 6.0 | 720 | 0.3901 | 0.5230 | 0.2731 |
| 0.3723 | 7.0 | 840 | 0.2639 | 0.4135 | 0.2273 |
| 0.2916 | 8.0 | 960 | 0.1469 | 0.3003 | 0.1786 |
| 0.2556 | 9.0 | 1080 | 0.0678 | 0.2367 | 0.1940 |
| 0.1352 | 10.0 | 1200 | 0.0453 | 0.2160 | 0.1838 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.14.6
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["wer"], "base_model": "jonatasgrosman/wav2vec2-large-xlsr-53-japanese", "model-index": [{"name": "wav2vec_RTSplit0207_6", "results": []}]} | automatic-speech-recognition | tndklab/wav2vec_RTSplit0207_6 | [
"transformers",
"safetensors",
"wav2vec2",
"automatic-speech-recognition",
"generated_from_trainer",
"base_model:jonatasgrosman/wav2vec2-large-xlsr-53-japanese",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-07T08:23:07+00:00 | [] | [] | TAGS
#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-jonatasgrosman/wav2vec2-large-xlsr-53-japanese #license-apache-2.0 #endpoints_compatible #region-us
| wav2vec\_RTSplit0207\_6
=======================
This model is a fine-tuned version of jonatasgrosman/wav2vec2-large-xlsr-53-japanese on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0453
* Wer: 0.2160
* Cer: 0.1838
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 32
* eval\_batch\_size: 32
* seed: 4
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 1000
* num\_epochs: 10
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.14.6
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-jonatasgrosman/wav2vec2-large-xlsr-53-japanese #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
80,
116,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-jonatasgrosman/wav2vec2-large-xlsr-53-japanese #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
-0.14118388295173645,
0.1513376235961914,
-0.0005152393132448196,
0.09957455098628998,
0.11795935779809952,
0.008073536679148674,
0.17546747624874115,
0.14973945915699005,
-0.040967848151922226,
0.11099261045455933,
0.11360207200050354,
0.06234046816825867,
0.05541789159178734,
0.19745732843875885,
-0.08206405490636826,
-0.22053459286689758,
0.07701526582241058,
-0.0033247191458940506,
0.010011174716055393,
0.11214284598827362,
0.07103060185909271,
-0.11864254623651505,
0.09036611765623093,
-0.006963435560464859,
-0.1427508145570755,
-0.04181395843625069,
0.016152987256646156,
-0.11031178385019302,
0.10872969776391983,
0.009288477711379528,
0.06582261621952057,
0.03504345938563347,
0.08960989117622375,
-0.18766173720359802,
0.0022469565737992525,
0.017141321673989296,
0.01435944065451622,
0.07424193620681763,
0.04270157217979431,
-0.0012159834150224924,
0.0027513811364769936,
-0.11490482091903687,
0.03699565678834915,
0.015048346482217312,
-0.11695785820484161,
-0.19873808324337006,
-0.07786379754543304,
0.017535261809825897,
0.09932611882686615,
0.08391966670751572,
-0.02103434130549431,
0.12361589074134827,
0.0008332731667906046,
0.07967887073755264,
0.19855915009975433,
-0.3135106563568115,
-0.05437514930963516,
-0.017246698960661888,
0.03944491222500801,
0.08242533355951309,
-0.10150530934333801,
-0.018765252083539963,
0.05041765794157982,
0.021163977682590485,
0.0925753116607666,
-0.03130659833550453,
-0.0336025208234787,
-0.010997597128152847,
-0.12088784575462341,
-0.03848946467041969,
0.1899806410074234,
0.07315327972173691,
-0.06343334913253784,
-0.0812249630689621,
-0.06397605687379837,
-0.12176059931516647,
-0.054642703384160995,
-0.00814018677920103,
0.026548052206635475,
-0.03949858993291855,
-0.09949330240488052,
-0.00489323865622282,
-0.0798637866973877,
-0.09082527458667755,
-0.017095407471060753,
0.17612586915493011,
0.01114688627421856,
0.013611323200166225,
-0.012470031157135963,
0.054258283227682114,
-0.02387041598558426,
-0.18483866751194,
-0.022159598767757416,
0.026931334286928177,
-0.032688405364751816,
-0.014408618211746216,
-0.04372356832027435,
-0.033641211688518524,
0.04413202032446861,
0.1181967481970787,
-0.018560312688350677,
0.06547611951828003,
-0.024242987856268883,
0.001249976921826601,
-0.08463852852582932,
0.182333305478096,
-0.06463606655597687,
-0.06938374042510986,
0.019906673580408096,
0.1272294521331787,
0.06285319477319717,
-0.02319909818470478,
-0.09877349436283112,
-0.008964418433606625,
0.1465861052274704,
0.03547809645533562,
-0.04232742264866829,
0.05053495988249779,
-0.038675304502248764,
-0.014026766642928123,
0.05737104266881943,
-0.12155470997095108,
0.025580810382962227,
0.022164685651659966,
-0.06321313977241516,
-0.02220059186220169,
-0.012167129665613174,
0.013010796159505844,
0.01259292010217905,
0.05222075432538986,
-0.08240532130002975,
0.0036988360807299614,
-0.02336418628692627,
-0.09296302497386932,
0.02680950053036213,
-0.06885657459497452,
0.0002455086214467883,
-0.10807177424430847,
-0.1792655736207962,
-0.017911894246935844,
0.023634862154722214,
-0.048911672085523605,
-0.009945076890289783,
-0.11301852017641068,
-0.0975419208407402,
0.04686196520924568,
-0.022707432508468628,
0.03525571897625923,
-0.07948608696460724,
0.10767754912376404,
0.07949104905128479,
0.08755897730588913,
-0.04064375162124634,
0.026420699432492256,
-0.09513017535209656,
0.03219624236226082,
-0.17416203022003174,
0.07473883777856827,
-0.054269835352897644,
0.034595683217048645,
-0.12033510208129883,
-0.0673440620303154,
0.019730009138584137,
-0.022371632978320122,
0.06979899108409882,
0.14241422712802887,
-0.1906011998653412,
-0.057025082409381866,
0.19722899794578552,
-0.11986931413412094,
-0.14269396662712097,
0.1278074085712433,
-0.03618486225605011,
0.038308508694171906,
0.07103344798088074,
0.22304697334766388,
0.03084309957921505,
-0.10670869052410126,
-0.04001520201563835,
-0.06360205262899399,
0.08328385651111603,
-0.03686724230647087,
0.1115616038441658,
0.005013670772314072,
-0.0014164310414344072,
0.016217360273003578,
-0.08063652366399765,
0.03222700208425522,
-0.07089976221323013,
-0.10016977041959763,
-0.044187162071466446,
-0.10613103210926056,
0.028292791917920113,
0.015585470013320446,
0.05590164288878441,
-0.09868429601192474,
-0.07060223072767258,
0.011771205812692642,
0.10809079557657242,
-0.11718438565731049,
0.012904723174870014,
-0.10417094081640244,
0.09427313506603241,
-0.11367906630039215,
-0.020158156752586365,
-0.15430019795894623,
-0.00410782964900136,
0.053871724754571915,
0.01823066547513008,
0.014241056516766548,
-0.07536637037992477,
0.08248129487037659,
0.07677027583122253,
-0.048812806606292725,
-0.07412240654230118,
-0.00498126819729805,
0.01805686019361019,
-0.06261217594146729,
-0.17385663092136383,
-0.028737887740135193,
-0.05413864552974701,
0.1608741134405136,
-0.16503508388996124,
0.0014996977988630533,
0.008824673481285572,
0.09039325267076492,
0.044107191264629364,
-0.02336837165057659,
0.019647266715765,
0.04836883023381233,
-0.026001546531915665,
-0.07121788710355759,
0.02903486229479313,
0.015071400441229343,
-0.10376740992069244,
0.019600635394454002,
-0.16807952523231506,
0.15009671449661255,
0.1392962485551834,
0.041799139231443405,
-0.05244087427854538,
0.020045755431056023,
-0.014278299175202847,
-0.04192302003502846,
-0.054265860468149185,
-0.015523474663496017,
0.10089807957410812,
0.007953685708343983,
0.12119784951210022,
-0.10283432900905609,
0.015069880522787571,
0.06469476968050003,
-0.027236688882112503,
-0.028428832069039345,
0.08093540370464325,
0.010932014323771,
-0.13942010700702667,
0.13027305901050568,
0.11321879178285599,
-0.07171899825334549,
0.12632238864898682,
-0.061046771705150604,
-0.085269495844841,
-0.05007721483707428,
0.033539436757564545,
0.03383169323205948,
0.13762885332107544,
-0.08172392845153809,
-0.022573819383978844,
0.02135329321026802,
0.021954631432890892,
-0.01604454591870308,
-0.19322632253170013,
-0.01944556087255478,
0.014740300364792347,
-0.09451686590909958,
-0.009623108431696892,
0.005852391943335533,
-0.017728494480252266,
0.09443813562393188,
-0.0006852737278677523,
-0.11360033601522446,
0.02314845845103264,
-0.015061916783452034,
-0.0876469686627388,
0.17204688489437103,
-0.09222397953271866,
-0.17411498725414276,
-0.13528600335121155,
-0.07141708582639694,
-0.05600440874695778,
0.03755852207541466,
0.06046575680375099,
-0.06563454121351242,
-0.04083354398608208,
-0.11549428105354309,
-0.04752589762210846,
0.03249216452240944,
0.04562458395957947,
0.05096840113401413,
-0.008819636888802052,
0.06599970906972885,
-0.08180946111679077,
-0.005118743982166052,
-0.014260591939091682,
-0.007273803930729628,
0.028529318049550056,
0.0005596327246166766,
0.12637819349765778,
0.12142695486545563,
0.00641365023329854,
0.024710198864340782,
-0.03795645385980606,
0.22719408571720123,
-0.06981682777404785,
-0.019021375104784966,
0.12310279160737991,
-0.027643905952572823,
0.04546043276786804,
0.17790329456329346,
0.03071616031229496,
-0.10723342001438141,
0.001543888240121305,
-0.049799270927906036,
-0.015234281308948994,
-0.188801571726799,
-0.033299293369054794,
-0.04766744375228882,
0.013789367862045765,
0.10153107345104218,
0.029864661395549774,
0.01635241135954857,
0.04831988736987114,
0.02163584530353546,
0.045850567519664764,
0.004909783601760864,
0.08106916397809982,
0.09610755741596222,
0.07607916742563248,
0.10785672068595886,
-0.03219561278820038,
-0.04924085736274719,
0.03241001442074776,
0.02096749283373356,
0.2015942931175232,
0.03050840273499489,
0.19284749031066895,
0.0009453566162846982,
0.15461201965808868,
0.025513680651783943,
0.08029980212450027,
0.0196499302983284,
0.009997223503887653,
-0.020813269540667534,
-0.07838824391365051,
-0.053045131266117096,
0.05521805211901665,
-0.013811358250677586,
0.06097419187426567,
-0.10551154613494873,
0.02013915218412876,
0.050434961915016174,
0.2723030149936676,
0.08769560605287552,
-0.3685109615325928,
-0.08635305613279343,
0.02104947343468666,
-0.036972131580114365,
-0.01903359405696392,
0.016876420006155968,
0.1534101963043213,
-0.06155303493142128,
0.06892576813697815,
-0.07162059843540192,
0.06354092806577682,
-0.06427314132452011,
0.01977919228374958,
0.023796986788511276,
0.047146882861852646,
0.0028003251645714045,
0.030816223472356796,
-0.24248063564300537,
0.28648892045021057,
0.035792022943496704,
0.09525445103645325,
-0.05690842866897583,
-0.0036345715634524822,
0.03943183273077011,
-0.0068437037989497185,
0.11822443455457687,
-0.024794165045022964,
-0.11079125106334686,
-0.1794157326221466,
-0.13511526584625244,
0.048938214778900146,
0.10588322579860687,
-0.006933995056897402,
0.11613228172063828,
-0.013944382779300213,
-0.04456436634063721,
0.04456821084022522,
-0.02340116538107395,
-0.08081210404634476,
-0.0749657079577446,
0.009459982626140118,
0.11387364566326141,
0.044953037053346634,
-0.04946205019950867,
-0.09598170965909958,
-0.09000002592802048,
0.08921240270137787,
0.0022167686838656664,
-0.006378722842782736,
-0.10505329072475433,
0.018605045974254608,
0.1496697962284088,
-0.09152472764253616,
0.053013674914836884,
0.00948717724531889,
0.1103808656334877,
0.027091750875115395,
-0.04968217760324478,
0.09032276272773743,
-0.06209796667098999,
-0.17874783277511597,
-0.050470758229494095,
0.13884679973125458,
-0.007241794373840094,
0.04262242838740349,
0.021225502714514732,
0.05142869055271149,
-0.005594416055828333,
-0.06703758984804153,
0.03248269483447075,
0.0272601880133152,
0.04040838032960892,
0.020026197656989098,
-0.012260128743946552,
-0.09217273443937302,
-0.09288010746240616,
-0.022662481293082237,
0.15083232522010803,
0.2982642650604248,
-0.06597831845283508,
0.01851712167263031,
0.08686181157827377,
-0.017854949459433556,
-0.15149927139282227,
-0.004584764130413532,
0.04337412118911743,
0.044433437287807465,
-0.005071667488664389,
-0.12238527834415436,
0.04546479135751724,
0.061537906527519226,
-0.04462991654872894,
0.07658969610929489,
-0.24918095767498016,
-0.1279260665178299,
0.08928412944078445,
0.1333279311656952,
0.12402694672346115,
-0.15273624658584595,
-0.06697630137205124,
-0.024110818281769753,
-0.10794417560100555,
0.1045057401061058,
-0.07326747477054596,
0.132745623588562,
-0.0021595852449536324,
0.06386832147836685,
0.0073677487671375275,
-0.05131489410996437,
0.15022529661655426,
0.02315061166882515,
0.05349516496062279,
-0.02223295159637928,
-0.01721254177391529,
0.04753045365214348,
-0.07513923943042755,
0.06926781684160233,
-0.0858917385339737,
0.049377862364053726,
-0.05993237346410751,
-0.024863220751285553,
-0.06171049550175667,
-0.006019624415785074,
0.0036315324250608683,
-0.03500806540250778,
-0.010318802669644356,
0.03614702448248863,
0.05882398411631584,
0.003448877949267626,
0.13278301060199738,
0.01197106670588255,
0.08229319751262665,
0.14570775628089905,
0.0886906087398529,
-0.03906968981027603,
0.014958192594349384,
-0.006075490266084671,
-0.05644846335053444,
0.053838834166526794,
-0.13194115459918976,
0.04858151823282242,
0.09669843316078186,
0.018267197534441948,
0.16071775555610657,
0.04687272757291794,
-0.0492333360016346,
0.03808420151472092,
0.06964091956615448,
-0.15898656845092773,
-0.1112910732626915,
0.003452348755672574,
-0.01375506166368723,
-0.11069352924823761,
0.04991121590137482,
0.13865762948989868,
-0.0703156515955925,
-0.007155262865126133,
-0.01823554001748562,
0.021476034075021744,
-0.03947385773062706,
0.20047150552272797,
0.042284511029720306,
0.05118047818541527,
-0.10997112095355988,
0.08147826790809631,
0.05683886259794235,
-0.08807548880577087,
0.049443088471889496,
0.037657495588064194,
-0.11523519456386566,
-0.022784452885389328,
0.00005534075535251759,
0.14294663071632385,
0.005469065625220537,
-0.07627619802951813,
-0.13811787962913513,
-0.08861514180898666,
0.035080134868621826,
0.17778432369232178,
0.06808756291866302,
0.03646450862288475,
-0.01790330372750759,
-0.0017073694616556168,
-0.10371081531047821,
0.09405805170536041,
0.07413934916257858,
0.07491617649793625,
-0.14983569085597992,
0.08254153281450272,
-0.008294850587844849,
0.0264718197286129,
-0.02044260874390602,
0.016870073974132538,
-0.11009109020233154,
0.0050047170370817184,
-0.09842979907989502,
0.05644430220127106,
-0.07891111820936203,
-0.016069279983639717,
-0.0012212666915729642,
-0.08156505972146988,
-0.06149281933903694,
0.011379055678844452,
-0.08712979406118393,
-0.026130566373467445,
0.003458258928731084,
0.0433938093483448,
-0.13683374226093292,
-0.03750019147992134,
0.02246333472430706,
-0.0981111079454422,
0.08394405245780945,
0.0866144597530365,
-0.019790008664131165,
0.04684387892484665,
-0.09545118361711502,
-0.021657824516296387,
0.08283203840255737,
0.0022664524149149656,
0.05060122534632683,
-0.14470165967941284,
-0.013973113149404526,
0.03143203258514404,
0.05041022598743439,
0.021516086533665657,
0.14788220822811127,
-0.09649581462144852,
0.005971268285065889,
-0.06748102605342865,
-0.011066189967095852,
-0.05691782385110855,
0.021555546671152115,
0.14140185713768005,
0.002627637470141053,
0.18363331258296967,
-0.09465426206588745,
0.022600574418902397,
-0.19713501632213593,
0.0013647646410390735,
-0.03708446025848389,
-0.12615914642810822,
-0.14735087752342224,
-0.02729809284210205,
0.07842589169740677,
-0.062242452055215836,
0.09589572995901108,
-0.06168156489729881,
0.06943611055612564,
0.013083240017294884,
-0.05813242867588997,
-0.0013656788505613804,
0.04068639129400253,
0.24912409484386444,
0.057672739028930664,
-0.03611792251467705,
0.07662001997232437,
0.010403653606772423,
0.09453791379928589,
0.12455988675355911,
0.12363877147436142,
0.15776370465755463,
0.03212352469563484,
0.14337152242660522,
0.08356481045484543,
-0.024862807244062424,
-0.11890481412410736,
0.05916835740208626,
-0.06845055520534515,
0.0900888592004776,
0.025596898049116135,
0.2101191282272339,
0.09875553101301193,
-0.1637156754732132,
0.004310329910367727,
-0.03623693808913231,
-0.08481713384389877,
-0.09586827456951141,
-0.06154704466462135,
-0.13099995255470276,
-0.14432699978351593,
0.010924887843430042,
-0.1069321557879448,
0.03407773748040199,
0.06964482367038727,
0.014476785436272621,
0.00010840524919331074,
0.14145208895206451,
0.014773874543607235,
0.029025593772530556,
0.09571029245853424,
0.008392676711082458,
-0.04039377346634865,
-0.0008097448153421283,
-0.10300909727811813,
0.023898394778370857,
0.00501815602183342,
0.056743279099464417,
-0.021962566301226616,
-0.02582181803882122,
0.06900105625391006,
-0.025983864441514015,
-0.1252528429031372,
0.011205081827938557,
0.020153336226940155,
0.06013365089893341,
0.04367395490407944,
0.056713055819272995,
-0.017620805650949478,
0.025014393031597137,
0.20733173191547394,
-0.08926752954721451,
-0.07744365930557251,
-0.13346822559833527,
0.14780977368354797,
-0.0138862831518054,
-0.007439719047397375,
0.009504212997853756,
-0.10589496046304703,
0.00224756239913404,
0.1942233294248581,
0.1499839723110199,
-0.07326409965753555,
-0.0006150471162982285,
-0.02711077593266964,
-0.00640704995021224,
-0.03820818290114403,
0.0650281012058258,
0.07840298116207123,
0.034390028566122055,
-0.059590261429548264,
-0.06017826497554779,
-0.05757676437497139,
-0.041262898594141006,
-0.022971007972955704,
0.03991897031664848,
-0.0317038968205452,
-0.023118194192647934,
-0.04992280900478363,
0.07865259051322937,
-0.08221094310283661,
-0.0975438579916954,
0.00718031357973814,
-0.21697042882442474,
-0.17321449518203735,
-0.0021544182673096657,
0.07518170028924942,
0.03575673699378967,
0.026161469519138336,
-0.033357031643390656,
0.02538921684026718,
0.05649259686470032,
-0.013258103281259537,
-0.057701025158166885,
-0.05936339870095253,
0.04269890859723091,
-0.08285795897245407,
0.17427952587604523,
-0.00461580278351903,
0.06635239720344543,
0.1040571853518486,
0.08133987337350845,
-0.10723624378442764,
0.10374195873737335,
0.06103026121854782,
-0.07403004914522171,
0.055740538984537125,
0.15157853066921234,
-0.056069888174533844,
0.14287318289279938,
0.05193907395005226,
-0.10281984508037567,
0.001149781746789813,
0.010053437203168869,
-0.028709089383482933,
-0.07485763728618622,
-0.06516216695308685,
-0.045741770416498184,
0.14665868878364563,
0.13426938652992249,
-0.06621623039245605,
0.0017491946928203106,
-0.016887987032532692,
0.056069597601890564,
0.06269390136003494,
0.02119249664247036,
-0.06178002804517746,
-0.2836269438266754,
-0.016548514366149902,
0.03701779246330261,
0.022455256432294846,
-0.2418002039194107,
-0.08923479914665222,
-0.010862430557608604,
-0.04599393531680107,
-0.07464209198951721,
0.0931960716843605,
0.08006857335567474,
0.03134746104478836,
-0.054390132427215576,
-0.05154174938797951,
-0.028746988624334335,
0.17312726378440857,
-0.1632525771856308,
-0.11506713181734085
] |
null | null | diffusers |
# LoRA DreamBooth - danaleee/CL_rank10_iter800
These are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using [DreamBooth](https://dreambooth.github.io/). You can find some example images in the following.




LoRA for the text encoder was enabled: False.
| {"license": "creativeml-openrail-m", "tags": ["stable-diffusion", "stable-diffusion-diffusers", "text-to-image", "diffusers", "lora"], "base_model": "CompVis/stable-diffusion-v1-4", "instance_prompt": "a photo of sks teddybear", "inference": true} | text-to-image | danaleee/CL_rank10_iter800 | [
"diffusers",
"tensorboard",
"stable-diffusion",
"stable-diffusion-diffusers",
"text-to-image",
"lora",
"base_model:CompVis/stable-diffusion-v1-4",
"license:creativeml-openrail-m",
"region:us"
] | 2024-02-07T08:25:51+00:00 | [] | [] | TAGS
#diffusers #tensorboard #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us
|
# LoRA DreamBooth - danaleee/CL_rank10_iter800
These are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using DreamBooth. You can find some example images in the following.
!img_0
!img_1
!img_2
!img_3
LoRA for the text encoder was enabled: False.
| [
"# LoRA DreamBooth - danaleee/CL_rank10_iter800\n\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
"TAGS\n#diffusers #tensorboard #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us \n",
"# LoRA DreamBooth - danaleee/CL_rank10_iter800\n\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
70,
104
] | [
"passage: TAGS\n#diffusers #tensorboard #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us \n# LoRA DreamBooth - danaleee/CL_rank10_iter800\n\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
-0.05028080567717552,
-0.011797548271715641,
-0.0024413312785327435,
0.07628371566534042,
0.10402137041091919,
0.03591272979974747,
0.19729062914848328,
0.0726994127035141,
0.04926155507564545,
0.07333094626665115,
0.11531529575586319,
0.07269470393657684,
-0.001619033981114626,
0.09674135595560074,
-0.021504631265997887,
-0.16172324120998383,
-0.0019511132268235087,
-0.014790990389883518,
-0.06339520961046219,
0.04214738681912422,
0.03581266850233078,
-0.047914136201143265,
0.1267789602279663,
-0.031453508883714676,
-0.15716959536075592,
0.02695307321846485,
0.0008100370177999139,
-0.0550214909017086,
0.07270663976669312,
0.07715949416160583,
0.041827064007520676,
0.061394210904836655,
0.05346483737230301,
-0.14116890728473663,
0.03461005166172981,
0.009203690104186535,
-0.06272967904806137,
0.07817484438419342,
-0.07903348654508591,
-0.0035162782296538353,
0.09527979791164398,
-0.026125473901629448,
0.007147105876356363,
0.014663349837064743,
-0.08258627355098724,
-0.03968124836683273,
-0.023151598870754242,
0.009692165069282055,
0.03698569908738136,
0.0385352298617363,
0.014735916629433632,
0.09459403157234192,
-0.023809917271137238,
0.07420404255390167,
0.2776479423046112,
-0.23208847641944885,
-0.02204415574669838,
0.2349110245704651,
-0.00034296553349122405,
0.07342154532670975,
-0.03536847233772278,
0.09338093549013138,
0.10856090486049652,
-0.03737363591790199,
0.03829101100564003,
-0.06000354140996933,
0.003995109349489212,
-0.05836845189332962,
-0.1017412468791008,
0.041562020778656006,
0.1466391235589981,
0.008008657954633236,
-0.059176184237003326,
-0.13605692982673645,
-0.032812416553497314,
0.0037928589154034853,
-0.014743255451321602,
0.04594723507761955,
0.014888876117765903,
-0.003925745375454426,
-0.02953786589205265,
-0.035860080271959305,
-0.08236448466777802,
-0.0641641765832901,
-0.02181074395775795,
0.09021644294261932,
-0.006285832263529301,
0.05758759751915932,
-0.012378944084048271,
0.1261817365884781,
-0.11278218775987625,
-0.12158762663602829,
0.03849463164806366,
-0.05459170043468475,
0.015807483345270157,
0.07169617712497711,
-0.01926586404442787,
-0.10899198800325394,
0.048057958483695984,
0.020295631140470505,
0.14966648817062378,
0.015499359928071499,
0.0041104573756456375,
0.11869914084672928,
-0.006660458631813526,
0.05749662220478058,
-0.061664555221796036,
-0.00695502944290638,
0.014929343946278095,
0.02064439468085766,
0.10739482194185257,
-0.08231105655431747,
-0.15582765638828278,
-0.006350968033075333,
-0.05701823532581329,
0.04086025804281235,
-0.08913882076740265,
0.02488715387880802,
-0.09217337518930435,
-0.0027198931202292442,
0.08899017423391342,
-0.0015799349639564753,
0.039490848779678345,
-0.03027469478547573,
-0.028390144929289818,
0.11321889609098434,
0.16504748165607452,
0.02246277965605259,
0.017456533387303352,
0.09635588526725769,
-0.07636027038097382,
0.04191894456744194,
-0.011031712405383587,
-0.12333817780017853,
0.021579692140221596,
-0.1789007931947708,
0.007902945391833782,
-0.131269171833992,
-0.03149118646979332,
-0.003063666634261608,
0.02689949795603752,
-0.04270845651626587,
0.051188234239816666,
-0.05987659841775894,
-0.11483065038919449,
-0.004201194271445274,
0.057229068130254745,
-0.013871116563677788,
-0.026394851505756378,
0.05187486857175827,
-0.01988687738776207,
0.1304633915424347,
-0.12472091615200043,
-0.031446728855371475,
-0.07006682455539703,
0.018256573006510735,
-0.12849706411361694,
0.05475984513759613,
-0.041069891303777695,
0.07587821781635284,
-0.05462334677577019,
-0.05449429899454117,
-0.018611779436469078,
0.0498904287815094,
0.03812272474169731,
0.12816385924816132,
-0.2526925802230835,
-0.05487274378538132,
0.07927508652210236,
-0.15070828795433044,
-0.09338653087615967,
0.0465831384062767,
-0.020011689513921738,
0.08908476680517197,
0.06381655484437943,
0.10237003862857819,
0.07236010581254959,
-0.30485767126083374,
-0.014939345419406891,
-0.05071604251861572,
-0.031012700870633125,
-0.05958795174956322,
0.0013311168877407908,
0.06828898191452026,
-0.019378850236535072,
0.031145615503191948,
-0.0688403844833374,
0.0733644962310791,
-0.035630494356155396,
-0.015560277737677097,
-0.027219664305448532,
-0.03750041499733925,
-0.012736666947603226,
-0.0009134786669164896,
0.05083799362182617,
-0.023406194522976875,
-0.03586733341217041,
0.005171850323677063,
0.0471368171274662,
-0.05716005712747574,
0.024121735244989395,
-0.03952747955918312,
0.10146252810955048,
-0.08497539162635803,
-0.012092415243387222,
-0.0860566794872284,
0.009384403936564922,
0.049484360963106155,
0.10919954627752304,
0.10440266132354736,
0.050747789442539215,
0.08688536286354065,
0.0658147931098938,
0.0021876897662878036,
-0.0010147219290956855,
0.05685942992568016,
-0.015628067776560783,
-0.051213670521974564,
-0.19997338950634003,
0.021020332351326942,
-0.07942371815443039,
0.026707623153924942,
-0.18912041187286377,
0.017955312505364418,
0.0582817867398262,
0.1804186999797821,
0.09875927865505219,
-0.013639572076499462,
0.05599997192621231,
0.034279000014066696,
-0.055862754583358765,
-0.044977545738220215,
0.03352579101920128,
-0.014702009037137032,
-0.16669484972953796,
0.15697984397411346,
-0.16696009039878845,
0.057281386107206345,
0.13014227151870728,
-0.0096506979316473,
-0.061450209468603134,
-0.05540476366877556,
0.002653480041772127,
0.032546933740377426,
-0.034625258296728134,
-0.02455390989780426,
0.10589494556188583,
-0.017590085044503212,
0.1549910455942154,
-0.028379475697875023,
0.008504443801939487,
0.06798394024372101,
-0.034969743341207504,
-0.08613581210374832,
0.09626847505569458,
0.09934473782777786,
0.022207332774996758,
0.012034991756081581,
0.08839650452136993,
-0.004507971461862326,
0.13245221972465515,
-0.005835438147187233,
-0.04539274424314499,
-0.06244798004627228,
0.007801850792020559,
0.048613257706165314,
0.1389966756105423,
0.01623259112238884,
-0.02516239695250988,
-0.011378494091331959,
-0.06783406436443329,
0.0019965670071542263,
-0.16112548112869263,
-0.028428610414266586,
0.022830359637737274,
-0.03978472203016281,
0.13009610772132874,
0.098448246717453,
-0.09672357887029648,
0.08819375187158585,
-0.10704336315393448,
-0.08207590878009796,
-0.016072671860456467,
-0.029322881251573563,
-0.053633905947208405,
0.11064661294221878,
-0.06017771363258362,
-0.15650925040245056,
-0.196570485830307,
0.0026739786844700575,
-0.0003947379591409117,
-0.007313541602343321,
0.05368874594569206,
-0.08455752581357956,
-0.0723729208111763,
-0.11837184429168701,
0.02633199281990528,
0.02978408709168434,
0.05258723720908165,
0.060610346496105194,
-0.03201482072472572,
-0.010285570286214352,
-0.10739333927631378,
0.00354552804492414,
-0.08521170914173126,
0.09468308836221695,
0.06680355966091156,
0.0007069047424010932,
0.11659169942140579,
0.13580508530139923,
0.02358207479119301,
0.0391375795006752,
0.0075150178745388985,
0.2336961179971695,
-0.01952388510107994,
0.07894325256347656,
0.12038340419530869,
-0.00529167614877224,
0.059622932225465775,
0.13183844089508057,
0.05256441608071327,
-0.08001229166984558,
0.05561452358961105,
-0.018092580139636993,
-0.14783921837806702,
-0.08062686771154404,
-0.0571013018488884,
-0.0355786457657814,
-0.05312082916498184,
0.07454495877027512,
0.04758160188794136,
0.11840546131134033,
0.08016467094421387,
0.05462179705500603,
0.12163155525922775,
0.051587384194135666,
0.07206692546606064,
0.07216230779886246,
-0.04108631610870361,
0.033635903149843216,
-0.084260493516922,
-0.1121893972158432,
0.10056769847869873,
-0.057649463415145874,
0.18837928771972656,
-0.06548824906349182,
0.05430689826607704,
0.03904002904891968,
-0.04541745409369469,
0.0940607339143753,
0.039250586181879044,
-0.061755403876304626,
-0.015091156587004662,
-0.041503001004457474,
-0.12128560245037079,
0.09792712330818176,
0.09585478156805038,
-0.007354897912591696,
-0.02758045122027397,
-0.01928320899605751,
0.0339682474732399,
0.00620403653010726,
-0.008872803300619125,
0.14811654388904572,
-0.2410709708929062,
0.03235819563269615,
0.02204800397157669,
0.07546635717153549,
-0.000006014794962538872,
0.009692533873021603,
0.2186896950006485,
0.006807912606745958,
0.0818583071231842,
-0.04551195725798607,
0.06085776910185814,
-0.018220296129584312,
-0.027233069762587547,
-0.039021048694849014,
0.12307392805814743,
-0.02115025371313095,
-0.048586003482341766,
-0.2192852795124054,
0.06424496322870255,
0.00869462825357914,
0.0006012200610712171,
-0.0646369606256485,
-0.028818195685744286,
0.022493876516819,
0.00540353637188673,
0.08980569988489151,
0.013177486136555672,
0.006790539715439081,
-0.12246114760637283,
-0.15577802062034607,
-0.039067286998033524,
0.08068419247865677,
-0.026091385632753372,
0.053877320140600204,
0.0526551827788353,
-0.03615414351224899,
0.020107906311750412,
0.004063380416482687,
-0.11510579288005829,
-0.08997223526239395,
0.01042997743934393,
0.17815783619880676,
-0.056361857801675797,
-0.030591586604714394,
-0.1024642363190651,
-0.04323633760213852,
0.05885782092809677,
-0.06155936047434807,
-0.05022382736206055,
-0.05815901234745979,
-0.007413354702293873,
0.09715302288532257,
-0.03522200882434845,
0.011180486530065536,
-0.040690142661333084,
0.058158840984106064,
-0.07143790274858475,
-0.14254160225391388,
0.08039422333240509,
-0.01332706306129694,
-0.138316348195076,
-0.08876928687095642,
0.11563274264335632,
-0.014486722648143768,
0.00560570927336812,
-0.018070854246616364,
0.036955226212739944,
0.0359579399228096,
-0.10946633666753769,
0.07269641011953354,
0.15305989980697632,
-0.13126014173030853,
0.11763346940279007,
-0.026238085702061653,
0.004481542855501175,
-0.05223436653614044,
0.0018697645282372832,
0.11662518233060837,
0.2589791417121887,
-0.10575428605079651,
0.09654492884874344,
0.050745707005262375,
-0.08921655267477036,
-0.2079095095396042,
-0.044483814388513565,
-0.023555217310786247,
0.03724617511034012,
-0.0040595633909106255,
-0.07816151529550552,
0.09736574441194534,
0.03542093187570572,
0.005067048594355583,
0.23200811445713043,
-0.38031214475631714,
-0.14552618563175201,
-0.007928210310637951,
0.15668104588985443,
0.2825857996940613,
-0.15617604553699493,
-0.08304816484451294,
-0.031009754166007042,
-0.09520846605300903,
0.1625533550977707,
-0.04309225082397461,
0.10512508451938629,
-0.04325724020600319,
-0.03956752270460129,
0.023663660511374474,
-0.03862717002630234,
0.12666307389736176,
0.005314597859978676,
0.0577729269862175,
-0.08213398605585098,
-0.0991220474243164,
0.08884307742118835,
-0.04424941539764404,
0.034724146127700806,
-0.14100350439548492,
0.02073495462536812,
-0.04983457177877426,
-0.005885916296392679,
0.015174123458564281,
-0.005876253359019756,
-0.02858152613043785,
-0.04559506103396416,
-0.12057837843894958,
0.015108526684343815,
-0.0010878187604248524,
-0.01120985671877861,
0.07458850741386414,
-0.014052201993763447,
0.01746267080307007,
0.14583468437194824,
-0.04815350100398064,
0.04824553802609444,
-0.04995839670300484,
0.0038398182950913906,
-0.04316438362002373,
0.1192321851849556,
-0.1490287184715271,
0.0016389783704653382,
0.166109099984169,
0.08128432184457779,
0.11125447601079941,
0.02056201733648777,
-0.10338261723518372,
0.08756182342767715,
0.11348801851272583,
-0.09314598143100739,
-0.015003715641796589,
-0.02333313785493374,
-0.06274357438087463,
0.11123571544885635,
0.02066378854215145,
0.1843518614768982,
-0.08816618472337723,
0.045745618641376495,
0.008183696307241917,
0.0014946571318432689,
-0.02104288898408413,
0.08359429240226746,
0.0356239378452301,
0.002730089705437422,
-0.0817968025803566,
0.06573540717363358,
-0.011965052224695683,
0.0236774031072855,
0.02017325721681118,
0.06863812357187271,
-0.07342667877674103,
-0.01027035340666771,
-0.012262595817446709,
0.20276571810245514,
-0.11866427958011627,
-0.007917426526546478,
-0.11477886885404587,
-0.11651476472616196,
0.025125423446297646,
0.15907271206378937,
0.056224334985017776,
0.02343803271651268,
-0.042366500943899155,
-0.04155315086245537,
-0.054421912878751755,
0.03307824954390526,
0.05405519902706146,
0.07772836089134216,
-0.24477483332157135,
-0.03497331961989403,
0.025164587423205376,
-0.012809921987354755,
-0.0693657174706459,
-0.03670283779501915,
-0.11566110700368881,
-0.018911955878138542,
-0.03790481016039848,
0.11511408537626266,
-0.04118110239505768,
-0.01694939285516739,
-0.001790098613128066,
-0.04364541545510292,
-0.02311529964208603,
0.02455328218638897,
-0.033376093953847885,
-0.017405545338988304,
0.0011588030029088259,
-0.009414809755980968,
-0.04776637628674507,
-0.09747077524662018,
-0.01195953693240881,
-0.08186763525009155,
0.03286004438996315,
-0.015314496122300625,
-0.07177690416574478,
0.026697499677538872,
-0.22481507062911987,
0.0382121317088604,
0.1843603551387787,
0.0022838199511170387,
-0.005805046297609806,
0.0007405590731650591,
0.004498075228184462,
-0.026965942233800888,
0.06778524816036224,
-0.004263881593942642,
0.05401778966188431,
-0.08294589817523956,
-0.05129824951291084,
-0.06848222017288208,
-0.015241681598126888,
-0.05401131883263588,
0.05057048425078392,
0.13881194591522217,
0.14108149707317352,
0.13322636485099792,
-0.09686054289340973,
0.07637889683246613,
-0.09642080962657928,
0.002406325889751315,
0.010833526961505413,
-0.06862247735261917,
0.015702979639172554,
-0.02838870882987976,
-0.022866226732730865,
-0.038549717515707016,
0.14917291700839996,
-0.016709979623556137,
-0.1518167406320572,
-0.02232569456100464,
0.05527781322598457,
-0.07058531790971756,
0.025361401960253716,
0.18912667036056519,
0.03625821694731712,
0.043588787317276,
-0.11699364334344864,
0.04673796519637108,
0.1616150438785553,
0.0868055671453476,
0.059919100254774094,
0.009236753918230534,
0.034981075674295425,
0.10602491348981857,
0.09143582731485367,
0.015329285524785519,
0.044940285384655,
0.14109046757221222,
-0.05202440172433853,
0.10610400140285492,
-0.04457937553524971,
0.011164487339556217,
0.112362340092659,
-0.03468303009867668,
-0.03435520827770233,
0.07890858501195908,
-0.04905957356095314,
-0.06320478767156601,
-0.11106663942337036,
-0.07161200046539307,
-0.1463148295879364,
0.026982443407177925,
-0.04296121001243591,
0.0015475817490369081,
-0.001985644455999136,
0.04538577422499657,
0.057904720306396484,
0.042893070727586746,
-0.022833779454231262,
-0.04109810292720795,
0.12011595815420151,
-0.018102971836924553,
-0.0943770483136177,
0.02661413513123989,
0.02417047880589962,
0.06327885389328003,
0.009768228977918625,
-0.0426924005150795,
0.07986718416213989,
0.0767613872885704,
0.00910193845629692,
-0.013721426948904991,
-0.08716652542352676,
-0.022735603153705597,
0.008002989925444126,
0.0033359092194586992,
0.13424375653266907,
0.08819015324115753,
-0.041541971266269684,
-0.039152130484580994,
0.11311078816652298,
-0.05691979452967644,
-0.021093623712658882,
-0.1137707307934761,
0.039473529905080795,
-0.11719316989183426,
0.034063201397657394,
-0.059708960354328156,
-0.09737244248390198,
-0.04925081878900528,
0.1823175996541977,
0.20388133823871613,
-0.06524378806352615,
0.035570353269577026,
-0.034752555191516876,
-0.002861431799829006,
0.0022728480398654938,
0.07994901388883591,
0.038688987493515015,
0.22771237790584564,
-0.03534339368343353,
-0.005916210822761059,
-0.086473248898983,
-0.05051612854003906,
-0.033090248703956604,
-0.07906497269868851,
0.012200726196169853,
-0.057060204446315765,
-0.0891021266579628,
0.043465591967105865,
-0.14455007016658783,
-0.14626698195934296,
0.18550066649913788,
-0.1945117563009262,
-0.024682343006134033,
-0.05371805280447006,
0.0577232800424099,
0.04721716418862343,
0.017938166856765747,
-0.08342348039150238,
-0.016555219888687134,
0.018446138128638268,
-0.011921430006623268,
-0.17204788327217102,
-0.0165278110653162,
-0.0855945497751236,
-0.1614118367433548,
0.0704333558678627,
-0.01312022004276514,
0.08916974067687988,
0.04426039755344391,
0.02417215146124363,
-0.061165809631347656,
0.05992889031767845,
-0.03512541204690933,
-0.07342911511659622,
-0.09544287621974945,
0.09231770038604736,
-0.024963360279798508,
0.0944497212767601,
0.02490954101085663,
-0.09182726591825485,
0.01983695663511753,
0.0754522755742073,
-0.08601792901754379,
-0.0900028869509697,
-0.03926412761211395,
-0.07699103653430939,
0.0988832637667656,
0.07425816357135773,
-0.027079470455646515,
-0.0034265145659446716,
0.02267918735742569,
0.0063239228911697865,
0.019768884405493736,
-0.09658771753311157,
0.05331377685070038,
-0.06255324184894562,
-0.00822108332067728,
-0.0006499484879896045,
0.03906597942113876,
-0.24815276265144348,
-0.07464920729398727,
-0.16821111738681793,
-0.009735208936035633,
-0.02727288380265236,
0.07795692980289459,
0.21861547231674194,
0.030062906444072723,
-0.0005088152247481048,
-0.1745065450668335,
0.04705724120140076,
0.07748077064752579,
-0.061227086931467056,
-0.07859095931053162
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# medicalBERT
This model is a fine-tuned version of [d4data/biomedical-ner-all](https://huggingface.co/d4data/biomedical-ner-all) on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 1
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 30
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "d4data/biomedical-ner-all", "model-index": [{"name": "medicalBERT", "results": []}]} | token-classification | roupenminassian/medicalBERT | [
"transformers",
"tensorboard",
"safetensors",
"distilbert",
"token-classification",
"generated_from_trainer",
"base_model:d4data/biomedical-ner-all",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T08:25:53+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #distilbert #token-classification #generated_from_trainer #base_model-d4data/biomedical-ner-all #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
# medicalBERT
This model is a fine-tuned version of d4data/biomedical-ner-all on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 1
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 30
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| [
"# medicalBERT\n\nThis model is a fine-tuned version of d4data/biomedical-ner-all on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 1\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 30",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #distilbert #token-classification #generated_from_trainer #base_model-d4data/biomedical-ner-all #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# medicalBERT\n\nThis model is a fine-tuned version of d4data/biomedical-ner-all on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 1\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 30",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
75,
32,
6,
12,
8,
3,
90,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #distilbert #token-classification #generated_from_trainer #base_model-d4data/biomedical-ner-all #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# medicalBERT\n\nThis model is a fine-tuned version of d4data/biomedical-ner-all on the None dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 1\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 30### Training results### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.08224538713693619,
0.1450723260641098,
-0.001975311664864421,
0.05178255960345268,
0.14221380650997162,
0.022151120007038116,
0.11371912807226181,
0.1258212774991989,
-0.08535698056221008,
0.07246041297912598,
0.08620651811361313,
0.058279287070035934,
0.03875274211168289,
0.13273978233337402,
-0.016207832843065262,
-0.21654804050922394,
0.00646163010969758,
0.02629321813583374,
-0.08892184495925903,
0.0769641101360321,
0.08283362537622452,
-0.10326030850410461,
0.04711472615599632,
-0.009496540762484074,
-0.12390463799238205,
0.03581894561648369,
-0.007692666258662939,
-0.03255326673388481,
0.10020317882299423,
-0.015433073043823242,
0.1038343533873558,
0.018319938331842422,
0.15221154689788818,
-0.24205729365348816,
-0.00028254915378056467,
0.07532533258199692,
0.04182566702365875,
0.07199647277593613,
0.02328142710030079,
-0.008345850743353367,
0.10334625095129013,
-0.20842605829238892,
0.12051257491111755,
0.010843501426279545,
-0.035768087953329086,
-0.17524640262126923,
-0.0799270048737526,
0.1012382060289383,
0.11510134488344193,
0.10463033616542816,
0.015813229605555534,
0.1368541419506073,
-0.07457584887742996,
0.07683110982179642,
0.1524287313222885,
-0.2266862690448761,
-0.0626215785741806,
0.0621003620326519,
0.0974588617682457,
0.08148372918367386,
-0.10626330226659775,
0.009744051843881607,
0.038932524621486664,
0.02340269647538662,
0.11945303529500961,
-0.020176708698272705,
-0.03373260423541069,
0.011318294331431389,
-0.13487082719802856,
0.009625724516808987,
0.16478563845157623,
0.03908643499016762,
-0.0394996702671051,
-0.07913586497306824,
-0.026161320507526398,
-0.09025031328201294,
-0.01876813732087612,
-0.09804093092679977,
0.05937131121754646,
-0.06428202241659164,
-0.034115344285964966,
-0.053482066839933395,
-0.0675954818725586,
-0.080143041908741,
0.0019005436915904284,
0.077968068420887,
0.049100473523139954,
-0.0067987521179020405,
-0.02058522403240204,
0.10037576407194138,
0.01799009181559086,
-0.08671795576810837,
-0.013754453510046005,
0.017125815153121948,
-0.09726405888795853,
-0.09911775588989258,
-0.03637908771634102,
-0.02811359241604805,
0.042957041412591934,
0.12359536439180374,
-0.06599146872758865,
0.027915867045521736,
0.036124058067798615,
-0.01159865316003561,
-0.035084765404462814,
0.07066681236028671,
-0.0719262883067131,
-0.03736405074596405,
-0.0030826497822999954,
0.1104370728135109,
0.01418163999915123,
0.021348707377910614,
-0.08056071400642395,
0.008890107274055481,
0.0866880938410759,
0.07174764573574066,
-0.06863722950220108,
0.015403389930725098,
-0.027058979496359825,
-0.032724298536777496,
0.014660628512501717,
-0.1210949569940567,
0.026471886783838272,
-0.014653393998742104,
-0.05258147418498993,
-0.01714934967458248,
0.05881502851843834,
0.01002839021384716,
-0.028688913211226463,
0.11657911539077759,
-0.07861477881669998,
-0.010471832938492298,
-0.08210504800081253,
-0.05294758081436157,
0.012415473349392414,
-0.06343857944011688,
0.016467351466417313,
-0.07151306420564651,
-0.18720830976963043,
-0.039684928953647614,
0.028391778469085693,
-0.027785679325461388,
-0.04017159715294838,
-0.05134115368127823,
-0.034360457211732864,
-0.004627063870429993,
0.002768357750028372,
0.09561365842819214,
-0.05203703045845032,
0.06874953955411911,
-0.027946913614869118,
0.03386707976460457,
-0.005784354638308287,
0.029082711786031723,
-0.07915595918893814,
0.03474210575222969,
-0.11292114853858948,
0.023973697796463966,
-0.09595872461795807,
0.0249167513102293,
-0.13043911755084991,
-0.08001583069562912,
-0.01045034360140562,
-0.031237058341503143,
0.046833742409944534,
0.115389384329319,
-0.16682754456996918,
-0.035354018211364746,
0.17505162954330444,
-0.047767557203769684,
-0.10847198218107224,
0.09866949170827866,
-0.04367450997233391,
0.023688532412052155,
0.0765918418765068,
0.1702212542295456,
0.1341819167137146,
-0.16775836050510406,
-0.016151877120137215,
-0.01988103985786438,
0.056787554174661636,
0.003840881399810314,
0.06791076809167862,
-0.03067161701619625,
0.025795919820666313,
0.0019187015714123845,
-0.10055593401193619,
-0.00014573140651918948,
-0.09211471676826477,
-0.08399945497512817,
-0.03378177061676979,
-0.13206179440021515,
0.03185218200087547,
0.024441782385110855,
0.03735343739390373,
-0.058867570012807846,
-0.07239843159914017,
0.13767820596694946,
0.14743340015411377,
-0.06216021999716759,
-0.010863730683922768,
-0.06525198370218277,
-0.023455023765563965,
-0.031472716480493546,
-0.014597930014133453,
-0.1541905701160431,
-0.14173603057861328,
0.021588299423456192,
-0.10012718290090561,
0.021575283259153366,
0.004545186180621386,
0.04588108882308006,
0.08855219930410385,
-0.06760803610086441,
-0.006813500542193651,
-0.08254822343587875,
0.029105238616466522,
-0.08245276659727097,
-0.2240493893623352,
-0.011396576650440693,
-0.04871581122279167,
0.15616142749786377,
-0.26588699221611023,
0.03551405295729637,
-0.08600220829248428,
0.12591703236103058,
0.020018260926008224,
-0.060719460248947144,
0.008442729711532593,
0.057370301336050034,
0.0021704863756895065,
-0.07831823825836182,
0.035568010061979294,
-0.047634322196245193,
-0.09783555567264557,
-0.0810033529996872,
-0.10336590558290482,
0.07100716233253479,
0.052191734313964844,
0.07409293949604034,
-0.1075664535164833,
-0.05314066633582115,
-0.04687534645199776,
-0.05489059165120125,
-0.11299213767051697,
-0.010526967234909534,
0.14673453569412231,
0.0045901197008788586,
0.10951430350542068,
-0.023446273058652878,
-0.04445226863026619,
-0.008226672187447548,
-0.002718766452744603,
0.0006923713954165578,
0.06032155081629753,
0.11163171380758286,
-0.12533771991729736,
0.08782687038183212,
0.10651113092899323,
-0.06864795833826065,
0.15116046369075775,
-0.051536016166210175,
-0.08643541485071182,
-0.031100457534193993,
0.0008437266806140542,
-0.0005099722766317427,
0.13466686010360718,
-0.11998122185468674,
0.005191671662032604,
0.013752369210124016,
-0.006683075334876776,
0.0215331818908453,
-0.13344016671180725,
-0.0204746276140213,
0.030634548515081406,
-0.05037548765540123,
-0.026122642681002617,
-0.04082635045051575,
0.008399376645684242,
0.10188987851142883,
0.02008594572544098,
-0.01298262644559145,
0.02748541161417961,
0.010412750765681267,
-0.10738178342580795,
0.18798907101154327,
-0.10453391075134277,
-0.11700628697872162,
-0.10362076759338379,
0.04873579367995262,
-0.028752367943525314,
-0.015333527699112892,
0.002572328317910433,
-0.08615586906671524,
-0.03961149603128433,
-0.08368316292762756,
-0.03468196094036102,
-0.00006061770181986503,
-0.010739629156887531,
0.03040887787938118,
0.030168287456035614,
0.09681560844182968,
-0.11467816680669785,
0.015658043324947357,
-0.03081454522907734,
-0.06780657172203064,
-0.010352556593716145,
0.036924537271261215,
0.1263194978237152,
0.14266225695610046,
-0.014028459787368774,
-0.00918967928737402,
-0.030657632276415825,
0.15533989667892456,
-0.0755588486790657,
0.00013828543887939304,
0.101657934486866,
0.029136473312973976,
0.02654017135500908,
0.11886345595121384,
0.04991530254483223,
-0.08747049421072006,
0.04551021009683609,
0.0865490734577179,
-0.0003706967690959573,
-0.25621941685676575,
-0.055628225207328796,
-0.024182936176657677,
-0.08903544396162033,
0.07210996001958847,
0.04891805723309517,
0.03536725416779518,
0.04400302842259407,
0.0020706350915133953,
0.05887957662343979,
-0.030809473246335983,
0.07630088925361633,
0.07390633970499039,
0.03376223146915436,
0.1168084517121315,
-0.011064920574426651,
-0.04932026192545891,
0.04659152403473854,
0.02283291332423687,
0.3051290810108185,
-0.0005543130682781339,
0.0863468274474144,
0.06609322130680084,
0.11396845430135727,
-0.030101943761110306,
0.027640657499432564,
0.0695105493068695,
-0.011829052120447159,
0.03219646215438843,
-0.06597217917442322,
-0.011581694707274437,
0.02952309139072895,
-0.041080787777900696,
0.022082773968577385,
-0.09044606983661652,
0.052491918206214905,
0.02113495208323002,
0.23834508657455444,
0.02330387942492962,
-0.27506354451179504,
-0.07510047405958176,
0.014043793082237244,
-0.02686370350420475,
-0.0571177713572979,
0.004114346578717232,
0.12303877621889114,
-0.10565535724163055,
0.06885109841823578,
-0.05704720318317413,
0.0859202891588211,
-0.03972553089261055,
0.028670012950897217,
0.042596250772476196,
0.07216141372919083,
-0.012475819326937199,
0.09819493442773819,
-0.2288649082183838,
0.20572488009929657,
0.007328466512262821,
0.12433075904846191,
-0.045719414949417114,
0.028108183294534683,
0.02478676848113537,
0.16840621829032898,
0.14033812284469604,
0.006400609388947487,
-0.00943304318934679,
-0.17148688435554504,
-0.07003797590732574,
0.04341592639684677,
0.11190018802881241,
-0.04431625083088875,
0.07123542577028275,
-0.027244865894317627,
0.00944464560598135,
0.04772737994790077,
-0.0975431501865387,
-0.21040761470794678,
-0.11025528609752655,
0.025349751114845276,
-0.021516108885407448,
-0.00735803646966815,
-0.07738026231527328,
-0.08469842374324799,
-0.050571855157613754,
0.17703233659267426,
-0.030155682936310768,
-0.03038412146270275,
-0.14547067880630493,
0.08464016765356064,
0.0821714997291565,
-0.06080111488699913,
0.0041307369247078896,
0.04154929891228676,
0.09397363662719727,
0.037036459892988205,
-0.08118487149477005,
0.04166094586253166,
-0.09054592996835709,
-0.1428341567516327,
-0.09081341326236725,
0.1031067967414856,
0.07716796547174454,
0.04600977897644043,
0.019586676731705666,
-0.0007659976254217327,
0.014505501836538315,
-0.0841650441288948,
0.05989905074238777,
0.11481109261512756,
0.08468611538410187,
0.06645757704973221,
-0.14301183819770813,
-0.015584742650389671,
-0.04363590478897095,
-0.04433834180235863,
0.0948677808046341,
0.2023576945066452,
-0.07591580599546432,
0.06665588915348053,
0.10113843530416489,
-0.10848730802536011,
-0.18860088288784027,
0.0713995024561882,
0.10672571510076523,
0.00802744273096323,
0.06251789629459381,
-0.21443210542201996,
0.17146307229995728,
0.14418014883995056,
-0.01675489731132984,
0.05653472617268562,
-0.26142561435699463,
-0.13097204267978668,
0.09981909394264221,
0.12834689021110535,
-0.020966606214642525,
-0.14991462230682373,
-0.04992813989520073,
-0.00823492743074894,
-0.10989779978990555,
0.11182207614183426,
-0.14046825468540192,
0.08947519958019257,
-0.012579972855746746,
0.055847905576229095,
0.023603539913892746,
-0.040497440844774246,
0.13503019511699677,
0.03909346088767052,
0.08594117313623428,
-0.04046134650707245,
0.01976226642727852,
0.08399379253387451,
-0.06077459082007408,
0.07232929766178131,
-0.03606461361050606,
0.053737156093120575,
-0.10763226449489594,
-0.04189496114850044,
-0.06483996659517288,
0.0758945569396019,
-0.04587506130337715,
-0.08466290682554245,
-0.05903790146112442,
0.07747048139572144,
0.06450267136096954,
-0.03322071209549904,
0.058771561831235886,
-0.0031935537699609995,
0.09965234249830246,
0.05590534210205078,
0.11899850517511368,
-0.008640426211059093,
-0.08002432435750961,
0.010857095941901207,
-0.04837710037827492,
0.0619121752679348,
-0.1035229042172432,
0.02455727383494377,
0.12575872242450714,
0.050322528928518295,
0.13370990753173828,
0.025056058540940285,
-0.04595770314335823,
-0.02372697927057743,
0.03395296633243561,
-0.10914620012044907,
-0.17188629508018494,
-0.006212790496647358,
-0.06592795997858047,
-0.15020610392093658,
0.051625706255435944,
0.09916693717241287,
-0.071709044277668,
-0.0001682130096014589,
-0.034715134650468826,
0.02309015393257141,
-0.048062101006507874,
0.22647325694561005,
0.03992640972137451,
0.0673329085111618,
-0.05355150252580643,
0.07101158052682877,
0.08945789933204651,
-0.09893360733985901,
0.051712069660425186,
0.04199475422501564,
-0.10754911601543427,
-0.04563678428530693,
0.07730068266391754,
0.22689679265022278,
-0.029055064544081688,
-0.0621175616979599,
-0.0904732421040535,
-0.10416924953460693,
0.04177561774849892,
0.13125568628311157,
0.02732928656041622,
-0.035161375999450684,
-0.02876630797982216,
0.024627843871712685,
-0.1640850007534027,
0.11988341063261032,
0.004912744276225567,
0.0950741097331047,
-0.11871964484453201,
0.13052500784397125,
-0.005640021990984678,
0.038146574050188065,
-0.030108150094747543,
0.06053481251001358,
-0.09628487378358841,
-0.03184705600142479,
-0.15655502676963806,
-0.03273061290383339,
-0.012301108799874783,
-0.006099454127252102,
-0.001248953863978386,
-0.027975160628557205,
-0.015864605084061623,
0.040340449661016464,
-0.0438760444521904,
-0.06608185917139053,
0.04205949977040291,
0.06865809112787247,
-0.1496177762746811,
-0.020210659131407738,
0.00993945263326168,
-0.07106518000364304,
0.06517373025417328,
0.0255812406539917,
0.04687589034438133,
0.008783011697232723,
-0.08823882043361664,
0.016221443191170692,
0.05170921981334686,
0.06009743735194206,
0.03270433470606804,
-0.13514800369739532,
-0.028294946998357773,
-0.011719398200511932,
0.024945270270109177,
0.04515253007411957,
0.046262528747320175,
-0.10881611704826355,
-0.04009254276752472,
-0.06323285400867462,
-0.025746047496795654,
-0.0637035071849823,
0.027939923107624054,
0.0886329859495163,
0.011842861771583557,
0.17486301064491272,
-0.08182267099618912,
0.026940952986478806,
-0.18505892157554626,
-0.04234466701745987,
0.0032733206171542406,
-0.016764739528298378,
-0.06297506392002106,
-0.014935745857656002,
0.07493247091770172,
-0.05403577536344528,
0.11355570703744888,
-0.045529019087553024,
0.06817986816167831,
0.039942558854818344,
-0.04434105381369591,
0.015672998502850533,
0.013195247389376163,
0.17961661517620087,
0.06224425882101059,
-0.04169551655650139,
0.037298548966646194,
-0.014331069774925709,
0.061770930886268616,
0.029025565832853317,
0.18958699703216553,
0.1500396430492401,
-0.058177631348371506,
0.011416780762374401,
0.04615795239806175,
-0.12006792426109314,
-0.16688059270381927,
0.05663082376122475,
0.014200489968061447,
0.08075687289237976,
-0.013940289616584778,
0.1362571269273758,
0.12049094587564468,
-0.18865202367305756,
0.045245956629514694,
-0.047456588596105576,
-0.13048477470874786,
-0.09844052791595459,
-0.03183567151427269,
-0.06143416836857796,
-0.08752263337373734,
0.0067020682618021965,
-0.12155544012784958,
0.0064301989041268826,
0.06954912841320038,
0.014517649076879025,
0.0030904978048056364,
0.19035303592681885,
-0.038245316594839096,
0.04964076727628708,
0.042066540569067,
0.011784390546381474,
-0.02343117445707321,
-0.08455432206392288,
-0.03765203058719635,
0.0383211188018322,
-0.008674165233969688,
0.05448535457253456,
-0.07232464849948883,
-0.016059748828411102,
0.03475823998451233,
-0.008249979466199875,
-0.05018218606710434,
0.04510122910141945,
-0.003281551878899336,
0.0027646746020764112,
0.06740924715995789,
0.0726422667503357,
-0.02356698550283909,
-0.02245049551129341,
0.2677057385444641,
-0.07199767976999283,
-0.09230107814073563,
-0.1194305568933487,
0.198867067694664,
0.056873492896556854,
0.0005451447796076536,
0.051624398678541183,
-0.10882607847452164,
0.024219300597906113,
0.19094908237457275,
0.13579051196575165,
-0.04255009442567825,
-0.01495310477912426,
-0.01771082915365696,
-0.019740477204322815,
-0.04240071773529053,
0.12093572318553925,
0.11097567528486252,
0.006608869414776564,
-0.035340387374162674,
-0.0034614726901054382,
-0.007473231293261051,
-0.030421694740653038,
-0.12516595423221588,
0.05372922122478485,
0.07147633284330368,
0.0038835019804537296,
-0.003806546563282609,
0.08283586800098419,
-0.0034503911156207323,
-0.1643410325050354,
0.056046005338430405,
-0.12080167233943939,
-0.15172620117664337,
-0.06052052974700928,
0.07810110598802567,
-0.05657133832573891,
-0.005002255085855722,
-0.010311230085790157,
0.005066603887826204,
0.17141564190387726,
-0.011059886775910854,
-0.06800416857004166,
-0.09627121686935425,
0.06909520924091339,
-0.05060786008834839,
0.22194717824459076,
0.008549133315682411,
0.06551346182823181,
0.10549701750278473,
0.0008300678455270827,
-0.11383683234453201,
0.09371782839298248,
0.0421304814517498,
-0.04474039003252983,
0.0692136362195015,
0.1528368443250656,
-0.0418369397521019,
0.10435658693313599,
0.026965368539094925,
-0.10936418175697327,
-0.031301241368055344,
-0.03908552601933479,
-0.0449114553630352,
-0.07710427045822144,
0.05730671435594559,
-0.06223343312740326,
0.13827447593212128,
0.17828071117401123,
-0.024372300133109093,
0.044296663254499435,
-0.08479665219783783,
0.04812343046069145,
0.07774556428194046,
0.10152093321084976,
-0.024504782631993294,
-0.18573467433452606,
0.03653442859649658,
-0.032593246549367905,
0.03923618793487549,
-0.29335644841194153,
-0.10394296050071716,
0.028075184673070908,
-0.03903351351618767,
-0.05130070447921753,
0.08490468561649323,
0.10410253703594208,
0.039789166301488876,
-0.05899525061249733,
-0.1367797702550888,
-0.04028324782848358,
0.1354074627161026,
-0.16821807622909546,
-0.05814087763428688
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# custom_q_and_a
This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 25.0
### Training results
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "mit", "tags": ["generated_from_trainer"], "base_model": "gpt2", "model-index": [{"name": "custom_q_and_a", "results": []}]} | text-generation | smrynrz20/custom_q_and_a | [
"transformers",
"tensorboard",
"safetensors",
"gpt2",
"text-generation",
"generated_from_trainer",
"base_model:gpt2",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T08:26:05+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #gpt2 #text-generation #generated_from_trainer #base_model-gpt2 #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# custom_q_and_a
This model is a fine-tuned version of gpt2 on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 25.0
### Training results
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| [
"# custom_q_and_a\n\nThis model is a fine-tuned version of gpt2 on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 25.0",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #gpt2 #text-generation #generated_from_trainer #base_model-gpt2 #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# custom_q_and_a\n\nThis model is a fine-tuned version of gpt2 on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 25.0",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
72,
29,
6,
12,
8,
3,
91,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #gpt2 #text-generation #generated_from_trainer #base_model-gpt2 #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# custom_q_and_a\n\nThis model is a fine-tuned version of gpt2 on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 25.0### Training results### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.07674609869718552,
0.06359387189149857,
-0.0009764775168150663,
0.08389134705066681,
0.1822912096977234,
0.029226893559098244,
0.1366964727640152,
0.09088768064975739,
-0.12368699163198471,
0.059176087379455566,
0.07532232999801636,
0.10193511098623276,
0.030952464789152145,
0.11716016381978989,
-0.026246102526783943,
-0.23364008963108063,
0.004033609293401241,
-0.00014390079013537616,
-0.10562528669834137,
0.10113803297281265,
0.0853307768702507,
-0.10869024693965912,
0.07676850259304047,
0.006762531585991383,
-0.22243425250053406,
0.024528373032808304,
0.017935743555426598,
-0.05123753473162651,
0.11544322967529297,
0.030171364545822144,
0.09819452464580536,
0.007049850188195705,
0.13713210821151733,
-0.1938001960515976,
0.003230533329769969,
0.10348888486623764,
0.017945431172847748,
0.06462381035089493,
0.04525403305888176,
0.019526569172739983,
0.12385009229183197,
-0.12219863384962082,
0.0797068178653717,
0.021633682772517204,
-0.0938158854842186,
-0.15823231637477875,
-0.07246758043766022,
0.024317560717463493,
0.08364642411470413,
0.1081496849656105,
-0.01209585927426815,
0.14416933059692383,
-0.08563392609357834,
0.06474479287862778,
0.17125804722309113,
-0.30170419812202454,
-0.07621049135923386,
0.11421111971139908,
0.039301931858062744,
0.08720490336418152,
-0.10251890867948532,
0.019035175442695618,
0.06441163271665573,
0.032532233744859695,
0.10018449276685715,
-0.011039412580430508,
-0.060674455016851425,
0.00566965714097023,
-0.15502983331680298,
0.01839226856827736,
0.10684706270694733,
0.03659379482269287,
-0.03731686621904373,
-0.06175198405981064,
-0.0528736375272274,
-0.07030335813760757,
-0.03274272382259369,
-0.06282427161931992,
0.05021892115473747,
-0.021029924973845482,
-0.08903732150793076,
-0.04465954378247261,
-0.07845544070005417,
-0.08810688555240631,
-0.03225478157401085,
0.12198897451162338,
0.04188850522041321,
0.0075774407014250755,
-0.04232450947165489,
0.10313306003808975,
-0.052991483360528946,
-0.09294240176677704,
-0.0011873991461470723,
0.0153592424467206,
-0.05361101031303406,
-0.07420475780963898,
-0.0560867078602314,
-0.0628625676035881,
0.019605737179517746,
0.11941876262426376,
-0.06953172385692596,
0.06857948005199432,
0.01703726127743721,
0.012101395055651665,
-0.0540686771273613,
0.1406097263097763,
-0.04858209565281868,
-0.026698948815464973,
0.01661972887814045,
0.06801331043243408,
0.01341129932552576,
-0.016892090439796448,
-0.1042555421590805,
0.008501778356730938,
0.08339177817106247,
0.03617238625884056,
-0.07384852319955826,
0.06564459204673767,
-0.0202613715082407,
-0.011118829250335693,
-0.013207763433456421,
-0.10384006053209305,
0.048608146607875824,
-0.021472321823239326,
-0.061923444271087646,
-0.054437585175037384,
0.013894387520849705,
0.0220109261572361,
-0.007126225158572197,
0.10671577602624893,
-0.09742370992898941,
0.02012052945792675,
-0.10644803941249847,
-0.0837242379784584,
-0.004275682847946882,
-0.04306056350469589,
-0.013555215671658516,
-0.07767395675182343,
-0.17571702599525452,
-0.0383542999625206,
0.0356864258646965,
-0.032183822244405746,
-0.02060958929359913,
-0.059870656579732895,
-0.06891774386167526,
0.004165660124272108,
-0.004764856304973364,
0.13246232271194458,
-0.05453408509492874,
0.06575999408960342,
0.022752614691853523,
0.03943721577525139,
-0.02150893211364746,
0.023776590824127197,
-0.08869002759456635,
0.008838413283228874,
-0.15047529339790344,
0.062246937304735184,
-0.07807477563619614,
0.03648636117577553,
-0.0801602378487587,
-0.09353567659854889,
-0.008111037313938141,
0.006832680199295282,
0.06444967538118362,
0.08555072546005249,
-0.19318731129169464,
-0.05170443281531334,
0.18898853659629822,
-0.0900234803557396,
-0.0923062264919281,
0.08956317603588104,
-0.045968715101480484,
0.04720626771450043,
0.08141404390335083,
0.1659604161977768,
0.03996383398771286,
-0.11006861180067062,
-0.011314872652292252,
-0.014759133569896221,
0.038348156958818436,
0.0027062632143497467,
0.03553188964724541,
-0.001983549213036895,
0.011629520915448666,
0.012062566354870796,
-0.056349143385887146,
0.01594887487590313,
-0.09728957712650299,
-0.08145952224731445,
-0.06131047382950783,
-0.08933044970035553,
-0.004321717657148838,
0.044053252786397934,
0.060156065970659256,
-0.08169416338205338,
-0.1112009584903717,
0.1420973688364029,
0.08876064419746399,
-0.06859574466943741,
0.012177195399999619,
-0.08399409800767899,
0.045700181275606155,
-0.047140102833509445,
-0.010106037370860577,
-0.17766067385673523,
-0.12426714599132538,
0.005881891120225191,
-0.04583636298775673,
0.05941365286707878,
0.024965940043330193,
0.06782026588916779,
0.08807983249425888,
-0.05084076151251793,
0.01576252654194832,
-0.06075780838727951,
0.005063075106590986,
-0.11289310455322266,
-0.22774744033813477,
-0.001466421177610755,
-0.010626302100718021,
0.16151902079582214,
-0.2584691047668457,
0.030162807554006577,
-0.0242227204144001,
0.10482083261013031,
0.012747063301503658,
-0.056514281779527664,
-0.04114048555493355,
0.06249342858791351,
-0.012160825543105602,
-0.09408010542392731,
0.04737745597958565,
0.002206044038757682,
-0.07750295847654343,
-0.08041124790906906,
-0.1700652688741684,
0.11245941370725632,
0.11268626898527145,
-0.03625643253326416,
-0.09951227903366089,
0.018060971051454544,
-0.06203717365860939,
-0.030380627140402794,
-0.08518802374601364,
0.03687262907624245,
0.1795315146446228,
-0.027001308277249336,
0.12896408140659332,
-0.05953349173069,
-0.053511910140514374,
0.005366768687963486,
-0.01501538511365652,
0.028991153463721275,
0.06165950372815132,
0.12937864661216736,
-0.103965163230896,
0.10523118078708649,
0.09348024427890778,
-0.07854221761226654,
0.143613800406456,
-0.028739308938384056,
-0.06179787218570709,
-0.0167614184319973,
-0.00362161872908473,
0.0018441586289554834,
0.11567919701337814,
-0.1081886738538742,
0.0008521369891241193,
-0.005683973431587219,
0.02820216491818428,
0.04824819788336754,
-0.2115524858236313,
-0.026151837781071663,
0.01687239669263363,
-0.04351693019270897,
-0.014885559678077698,
-0.02430311217904091,
0.002521658316254616,
0.09675527364015579,
0.01950838789343834,
-0.0025414051488041878,
0.01862296275794506,
0.0074597611092031,
-0.09221284836530685,
0.20326004922389984,
-0.10399442911148071,
-0.11829864233732224,
-0.07767827063798904,
0.0283059049397707,
-0.05616717040538788,
-0.0047425138764083385,
0.029254211112856865,
-0.13289284706115723,
-0.05166257917881012,
-0.08237320929765701,
0.014128879643976688,
-0.005208632443100214,
0.01263309083878994,
0.04216136038303375,
0.006432082038372755,
0.08504077047109604,
-0.13027434051036835,
0.002916803350672126,
-0.058214668184518814,
-0.14417600631713867,
0.008193718269467354,
0.07426916062831879,
0.10682183504104614,
0.16503898799419403,
-0.021152904257178307,
0.016276540234684944,
-0.030192872509360313,
0.25444090366363525,
-0.0737607479095459,
-0.024956241250038147,
0.09372280538082123,
0.01959345117211342,
0.036426693201065063,
0.09803137183189392,
0.06065027788281441,
-0.13852722942829132,
0.03146600350737572,
0.09155534952878952,
-0.03275101259350777,
-0.21627196669578552,
-0.04146283119916916,
-0.022715765982866287,
-0.09062158316373825,
0.059011660516262054,
0.052410054951906204,
0.03165748715400696,
0.05521400645375252,
0.022942790761590004,
0.09692776203155518,
-0.02774905413389206,
0.0712277963757515,
0.09799914807081223,
0.05413854122161865,
0.12249985337257385,
-0.04340681806206703,
-0.042052727192640305,
0.05512683838605881,
-0.02805769443511963,
0.26738110184669495,
-0.004369990434497595,
0.037355344742536545,
0.07012627273797989,
0.12126666307449341,
-0.010090013965964317,
0.0027823965065181255,
0.008959265425801277,
-0.04893907159566879,
0.01153481099754572,
-0.06405194103717804,
-0.014680484309792519,
0.026150472462177277,
-0.07816164195537567,
0.0325133353471756,
-0.08522479981184006,
0.0506962426006794,
0.05591016635298729,
0.18284736573696136,
0.013624030165374279,
-0.29514992237091064,
-0.09010584652423859,
0.003913314547389746,
-0.007506858557462692,
-0.04142562299966812,
0.011039048433303833,
0.11315983533859253,
-0.09051431715488434,
0.05070466175675392,
-0.06285455822944641,
0.08536436408758163,
-0.02724445052444935,
0.020309539511799812,
0.07903043180704117,
0.1669589877128601,
-0.01752062514424324,
0.07600483298301697,
-0.2560283839702606,
0.22484061121940613,
0.02420862205326557,
0.14299646019935608,
-0.05248730629682541,
0.01918526366353035,
0.018589666113257408,
0.09998781979084015,
0.0643804669380188,
-0.0012832791544497013,
-0.07282661646604538,
-0.18037045001983643,
-0.02360384352505207,
0.06377770751714706,
0.1339794546365738,
-0.005928443744778633,
0.07066979259252548,
-0.03956065699458122,
0.01763800159096718,
0.07861275970935822,
-0.08404739946126938,
-0.19671575725078583,
-0.10423731058835983,
0.0019741232972592115,
0.011591602116823196,
-0.04637598991394043,
-0.07826890051364899,
-0.09879764169454575,
-0.10366960614919662,
0.20373380184173584,
0.008610776625573635,
-0.019297733902931213,
-0.11787331104278564,
0.11892395466566086,
0.05958808958530426,
-0.0537092499434948,
0.026518208906054497,
0.028564078733325005,
0.09225503355264664,
0.052650853991508484,
-0.08712794631719589,
0.09975633770227432,
-0.07145704329013824,
-0.1506432741880417,
-0.05647185072302818,
0.08391091972589493,
0.06511438637971878,
0.03811407834291458,
0.0005254042916931212,
0.021851642057299614,
0.023367347195744514,
-0.11891990900039673,
0.02888176217675209,
0.0676475539803505,
0.055619679391384125,
0.028652556240558624,
-0.06474073231220245,
0.010077614337205887,
-0.02729858085513115,
-0.035945504903793335,
0.11048644781112671,
0.22098395228385925,
-0.07832644879817963,
0.0337350033223629,
0.04684510454535484,
-0.09829533845186234,
-0.18362365663051605,
0.13246740400791168,
0.078330859541893,
-0.005173014011234045,
0.03521350398659706,
-0.1819707602262497,
0.15974971652030945,
0.1316632330417633,
-0.020180385559797287,
0.08827458322048187,
-0.3294241726398468,
-0.14945268630981445,
0.06102379783987999,
0.15283343195915222,
0.08089511096477509,
-0.1580125242471695,
-0.03193819895386696,
-0.03402275964617729,
-0.11586182564496994,
0.1263628900051117,
-0.1393713504076004,
0.11284086108207703,
0.005972099024802446,
0.07179593294858932,
0.0033694140147417784,
-0.04224938899278641,
0.12721429765224457,
0.013389497995376587,
0.10607641935348511,
-0.0691457986831665,
0.04188867658376694,
0.08968551456928253,
-0.03726000338792801,
-0.0012575002619996667,
-0.0361066497862339,
0.05631810054183006,
-0.04589516669511795,
-0.028379401192069054,
-0.06470835953950882,
0.05776328593492508,
-0.04161084070801735,
-0.07998856902122498,
-0.0546015165746212,
0.041853152215480804,
0.03685156628489494,
-0.03801096975803375,
0.0514412447810173,
0.0002520526759326458,
0.1571923792362213,
0.010163822211325169,
0.0900978371500969,
-0.06956527382135391,
-0.04855676740407944,
0.01413823664188385,
-0.011981481686234474,
0.06873910874128342,
-0.15468117594718933,
0.021513545885682106,
0.11960747838020325,
0.035266563296318054,
0.1264016479253769,
0.07285615801811218,
-0.045309878885746,
0.016261478886008263,
0.06302112340927124,
-0.13010476529598236,
-0.1629713475704193,
0.009106540121138096,
-0.06910925358533859,
-0.0712955892086029,
0.07184401154518127,
0.09432012587785721,
-0.07068082690238953,
0.00643958942964673,
-0.014066515490412712,
0.0034425416961312294,
-0.050561413168907166,
0.18898214399814606,
0.030167775228619576,
0.03899538516998291,
-0.08913738280534744,
0.09292874485254288,
0.04483801871538162,
-0.06674961000680923,
0.04923310875892639,
0.05865286663174629,
-0.0906790941953659,
-0.019245322793722153,
0.0931275486946106,
0.20588544011116028,
-0.05616743117570877,
-0.04137702286243439,
-0.10461156070232391,
-0.09558740258216858,
0.0377664715051651,
0.14190863072872162,
0.0586819089949131,
-0.01806771755218506,
-0.032755497843027115,
0.06742582470178604,
-0.16075778007507324,
0.05951676517724991,
0.011814209632575512,
0.085224948823452,
-0.15273146331310272,
0.1516171097755432,
0.0323232002556324,
0.007715080864727497,
-0.02964932657778263,
0.03701888024806976,
-0.12417293339967728,
-0.020049206912517548,
-0.12294268608093262,
-0.03852790594100952,
-0.025813976302742958,
0.004260458517819643,
-0.0017317644087597728,
-0.04642308130860329,
-0.06249910965561867,
0.04376065358519554,
-0.07473908364772797,
-0.04545526206493378,
0.025780683383345604,
0.03511199355125427,
-0.14254598319530487,
0.013587957248091698,
-0.0009530718089081347,
-0.06075555086135864,
0.06306774169206619,
0.04229649901390076,
0.036095328629016876,
0.0666293352842331,
-0.19784286618232727,
0.012029692530632019,
0.05831700190901756,
0.00791260041296482,
0.06627631187438965,
-0.051722753793001175,
0.00040576353785581887,
-0.01287557277828455,
0.08311386406421661,
0.03276054933667183,
0.06964369863271713,
-0.1091776192188263,
0.009638666175305843,
-0.06707826256752014,
-0.0441594198346138,
-0.04957358166575432,
0.03352558612823486,
0.08463443070650101,
0.013853101991117,
0.162228062748909,
-0.11081130057573318,
0.010299808345735073,
-0.2123415768146515,
-0.03644042834639549,
0.007867998443543911,
-0.03965725004673004,
-0.07923898845911026,
-0.03374691680073738,
0.07932661473751068,
-0.06557310372591019,
0.16216696798801422,
0.021984288468956947,
0.07245294004678726,
0.041959088295698166,
-0.018986942246556282,
0.001005504745990038,
0.008035706356167793,
0.2080141007900238,
0.06936932355165482,
-0.014253823086619377,
0.06422761082649231,
0.03460511565208435,
0.08413390815258026,
0.016808083280920982,
0.21663694083690643,
0.1147146224975586,
-0.0749945268034935,
0.06506061553955078,
0.059245653450489044,
-0.10414499789476395,
-0.1614246368408203,
0.13803483545780182,
-0.07963539659976959,
0.10162995010614395,
-0.05865800008177757,
0.13912875950336456,
0.11092283576726913,
-0.1451403945684433,
0.022289106622338295,
-0.066656194627285,
-0.09957904368638992,
-0.13272938132286072,
-0.04036387428641319,
-0.0886346697807312,
-0.15940938889980316,
0.03820144385099411,
-0.12128244340419769,
0.03572167456150055,
0.09347780793905258,
0.020044146105647087,
0.013755051419138908,
0.1906573474407196,
-0.013201613910496235,
0.03643205761909485,
0.02546338364481926,
-0.005447457544505596,
-0.028819536790251732,
-0.07740198075771332,
-0.060511309653520584,
0.027355097234249115,
-0.008408965542912483,
0.06380008161067963,
-0.05363517627120018,
-0.033142101019620895,
0.048589855432510376,
-0.019208572804927826,
-0.049666598439216614,
0.027722934260964394,
0.038441505283117294,
0.022095607593655586,
0.009603715501725674,
0.03174181282520294,
-0.018691416829824448,
-0.023210912942886353,
0.2771279215812683,
-0.06993517279624939,
-0.10695569217205048,
-0.11233985424041748,
0.25231674313545227,
0.022596918046474457,
0.004702848847955465,
0.035218968987464905,
-0.10261314362287521,
0.019500799477100372,
0.18985863029956818,
0.17783525586128235,
-0.08506050705909729,
-0.01270213257521391,
-0.011871923692524433,
-0.02571958303451538,
-0.0661357045173645,
0.1557387262582779,
0.1056835949420929,
0.056002356112003326,
-0.050733309239149094,
-0.05615602433681488,
-0.01641501672565937,
-0.01697230152785778,
-0.06092814728617668,
0.05290422961115837,
0.04219608008861542,
0.018484625965356827,
-0.021516015753149986,
0.07560916244983673,
-0.01665308140218258,
-0.13539005815982819,
0.037678759545087814,
-0.14982807636260986,
-0.1660965383052826,
-0.008115087635815144,
0.10060260444879532,
-0.04728859290480614,
0.05216822400689125,
-0.015081847086548805,
-0.012041375041007996,
0.07206407189369202,
-0.01895516738295555,
-0.055689599364995956,
-0.09433913975954056,
0.05476406589150429,
-0.07606152445077896,
0.24042053520679474,
-0.020441176369786263,
0.07263779640197754,
0.1223510354757309,
0.033825669437646866,
-0.09751395881175995,
0.08179259300231934,
0.03561058267951012,
-0.09845849126577377,
0.02151435986161232,
0.12175910174846649,
-0.050616275519132614,
0.07229349762201309,
0.05003213882446289,
-0.12997353076934814,
0.01786043681204319,
-0.06657172739505768,
-0.04610763490200043,
-0.07522140443325043,
-0.024341074749827385,
-0.08468938618898392,
0.1453101485967636,
0.19249247014522552,
-0.008508509956300259,
0.037914104759693146,
-0.08893881738185883,
0.03828033432364464,
0.04284820333123207,
0.12251297384500504,
-0.047259602695703506,
-0.24853016436100006,
0.03580097854137421,
0.04520324245095253,
-0.009411470033228397,
-0.24536408483982086,
-0.08461480587720871,
0.03967488557100296,
-0.0489688515663147,
-0.07241341471672058,
0.11064311116933823,
0.12245027720928192,
0.05011357367038727,
-0.05780947208404541,
-0.16574536263942719,
-0.04643798619508743,
0.16550108790397644,
-0.1402004063129425,
-0.0739196389913559
] |
null | null | transformers |
# ALMA-Latxa-7b
ALMA-Latxa-7b is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
* [HiTZ/latxa-7b-v1](https://huggingface.co/HiTZ/latxa-7b-v1)
* [haoranxu/ALMA-7B](https://huggingface.co/haoranxu/ALMA-7B)
## 🧩 Configuration
```yaml
slices:
- sources:
- model: HiTZ/latxa-7b-v1
layer_range: [0, 32]
- model: haoranxu/ALMA-7B
layer_range: [0, 32]
merge_method: slerp
base_model: haoranxu/ALMA-7B
parameters:
t:
- filter: self_attn
value: [0, 0.5, 0.3, 0.7, 1]
- filter: mlp
value: [1, 0.5, 0.7, 0.3, 0]
- value: 0.5
dtype: bfloat16
```
## 💻 Usage
```python
!pip install -qU transformers accelerate
from transformers import AutoTokenizer
import transformers
import torch
model = "airalribalta/ALMA-Latxa-7b"
messages = [{"role": "user", "content": "What is a large language model?"}]
tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
"text-generation",
model=model,
torch_dtype=torch.float16,
device_map="auto",
)
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
``` | {"tags": ["merge", "mergekit", "lazymergekit", "HiTZ/latxa-7b-v1", "haoranxu/ALMA-7B"], "base_model": ["HiTZ/latxa-7b-v1", "haoranxu/ALMA-7B"]} | text-generation | airalribalta/ALMA-Latxa-7b | [
"transformers",
"safetensors",
"llama",
"text-generation",
"merge",
"mergekit",
"lazymergekit",
"HiTZ/latxa-7b-v1",
"haoranxu/ALMA-7B",
"base_model:HiTZ/latxa-7b-v1",
"base_model:haoranxu/ALMA-7B",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T08:30:27+00:00 | [] | [] | TAGS
#transformers #safetensors #llama #text-generation #merge #mergekit #lazymergekit #HiTZ/latxa-7b-v1 #haoranxu/ALMA-7B #base_model-HiTZ/latxa-7b-v1 #base_model-haoranxu/ALMA-7B #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# ALMA-Latxa-7b
ALMA-Latxa-7b is a merge of the following models using LazyMergekit:
* HiTZ/latxa-7b-v1
* haoranxu/ALMA-7B
## Configuration
## Usage
| [
"# ALMA-Latxa-7b\n\nALMA-Latxa-7b is a merge of the following models using LazyMergekit:\n* HiTZ/latxa-7b-v1\n* haoranxu/ALMA-7B",
"## Configuration",
"## Usage"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #merge #mergekit #lazymergekit #HiTZ/latxa-7b-v1 #haoranxu/ALMA-7B #base_model-HiTZ/latxa-7b-v1 #base_model-haoranxu/ALMA-7B #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# ALMA-Latxa-7b\n\nALMA-Latxa-7b is a merge of the following models using LazyMergekit:\n* HiTZ/latxa-7b-v1\n* haoranxu/ALMA-7B",
"## Configuration",
"## Usage"
] | [
106,
48,
4,
3
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #merge #mergekit #lazymergekit #HiTZ/latxa-7b-v1 #haoranxu/ALMA-7B #base_model-HiTZ/latxa-7b-v1 #base_model-haoranxu/ALMA-7B #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# ALMA-Latxa-7b\n\nALMA-Latxa-7b is a merge of the following models using LazyMergekit:\n* HiTZ/latxa-7b-v1\n* haoranxu/ALMA-7B## Configuration## Usage"
] | [
-0.09099698811769485,
-0.10824799537658691,
-0.0024788195732980967,
-0.004307310562580824,
0.03960727900266647,
0.04755571112036705,
0.21851444244384766,
0.08378242701292038,
0.06574037671089172,
0.02981886826455593,
0.08072526752948761,
0.1485772281885147,
0.06479406356811523,
0.13025353848934174,
-0.059962574392557144,
-0.17117300629615784,
0.0340542234480381,
0.019789475947618484,
-0.10256829857826233,
0.044839486479759216,
0.09539151191711426,
-0.04792694374918938,
0.13729475438594818,
-0.007863379083573818,
-0.06674829125404358,
0.08215472847223282,
-0.007283918093889952,
-0.09439422190189362,
0.15065740048885345,
0.16821569204330444,
0.1115088239312172,
0.09577247500419617,
-0.02398826740682125,
-0.10226953029632568,
0.03915007784962654,
-0.006491738371551037,
-0.047593023627996445,
0.09528429806232452,
0.1268402487039566,
0.0045420946553349495,
0.06909334659576416,
-0.07677418738603592,
-0.0483328253030777,
0.02960759587585926,
-0.12085160613059998,
-0.04457956179976463,
-0.06852588057518005,
0.08872286975383759,
0.10684631019830704,
-0.054598618298769,
0.0011533237993717194,
0.0774434432387352,
0.041881076991558075,
0.03289647027850151,
0.16657570004463196,
-0.3487536907196045,
-0.020854922011494637,
0.15193025767803192,
0.0913415253162384,
0.029118157923221588,
0.020226383581757545,
0.13199803233146667,
0.09419843554496765,
-0.0660870298743248,
0.09748263657093048,
-0.07898326963186264,
0.07389294356107712,
-0.009565303102135658,
-0.07662060856819153,
0.013614882715046406,
0.13289350271224976,
-0.006362857297062874,
-0.0005396355409175158,
-0.08425674587488174,
-0.10395806282758713,
0.07965259253978729,
-0.043095774948596954,
-0.06387597322463989,
0.02870829589664936,
-0.044888708740472794,
0.04379125311970711,
-0.07615431398153305,
-0.06625042855739594,
-0.032098617404699326,
-0.047024521976709366,
0.1204872578382492,
0.010218198411166668,
-0.006285023409873247,
-0.03814207389950752,
0.09092500805854797,
-0.16240258514881134,
-0.162292942404747,
0.05021527409553528,
-0.047737617045640945,
0.008535826578736305,
-0.011208388023078442,
-0.03132271766662598,
-0.06797511875629425,
0.10123604536056519,
0.04100741818547249,
0.06130094453692436,
0.06668294221162796,
-0.0034061912447214127,
0.06993191689252853,
0.026242570951581,
0.006283552851527929,
-0.06423491984605789,
-0.23735812306404114,
0.060815148055553436,
0.09695841372013092,
0.09471159428358078,
-0.017242079600691795,
-0.11511684209108353,
-0.09670662879943848,
0.02208441123366356,
0.03827040642499924,
0.05340133607387543,
0.054128702729940414,
-0.03307217359542847,
-0.05365566909313202,
0.1408446580171585,
-0.06824889779090881,
-0.021287821233272552,
-0.05337868630886078,
-0.03352748975157738,
0.029884612187743187,
0.04735425114631653,
0.04944949969649315,
0.01284026075154543,
0.02343321219086647,
-0.10271449387073517,
-0.06727942079305649,
-0.05926739424467087,
-0.05822473391890526,
0.022945404052734375,
-0.05130539461970329,
-0.0021123522892594337,
-0.1367643028497696,
-0.2106340378522873,
-0.023110562935471535,
0.03751828521490097,
-0.0900997519493103,
-0.016111332923173904,
-0.03096909075975418,
0.040905408561229706,
0.022551942616701126,
-0.04311283305287361,
0.02134348265826702,
-0.0041863624937832355,
0.012573922984302044,
0.08317437767982483,
0.10364885628223419,
-0.060335922986269,
0.023643307387828827,
-0.07235979288816452,
0.13443101942539215,
-0.22035285830497742,
0.07093755155801773,
-0.06788330525159836,
0.1347319483757019,
-0.12138824909925461,
0.037997230887413025,
-0.06369096040725708,
0.08534874767065048,
0.07779531925916672,
0.14648686349391937,
-0.10353266447782516,
-0.07893554866313934,
0.0954781100153923,
-0.10222810506820679,
-0.18515951931476593,
0.07830818742513657,
-0.012299800291657448,
0.12201393395662308,
0.05174177139997482,
0.21349883079528809,
0.04998172074556351,
-0.04162658005952835,
0.003176427911967039,
0.01567011885344982,
0.04757217690348625,
-0.09808109700679779,
0.059824664145708084,
0.016355270519852638,
-0.17458537220954895,
0.024590132758021355,
0.02905353158712387,
0.030128516256809235,
-0.027527112513780594,
-0.052821848541498184,
-0.01775539480149746,
-0.04963163658976555,
0.024342423304915428,
-0.0545344203710556,
0.0686071589589119,
-0.05979061871767044,
-0.03494555503129959,
0.02843903936445713,
0.10426493734121323,
-0.003246884560212493,
-0.02875007688999176,
-0.03317714482545853,
0.12174314260482788,
-0.1689787656068802,
0.05522797629237175,
-0.04058113321661949,
-0.037692055106163025,
-0.008703694678843021,
-0.021381430327892303,
0.04328211024403572,
0.028176818042993546,
0.04374614730477333,
0.03222601115703583,
-0.05959341675043106,
-0.03246457129716873,
0.14949356019496918,
0.01615465246140957,
-0.08056492358446121,
-0.22725364565849304,
-0.08030001074075699,
-0.02792220190167427,
0.1345902532339096,
-0.12639130651950836,
0.04175304248929024,
-0.12413299083709717,
0.13774099946022034,
-0.03388223424553871,
0.09844543784856796,
0.015306713059544563,
0.03156205639243126,
-0.07131005078554153,
0.013970974832773209,
0.059661779552698135,
-0.02385713718831539,
-0.11523143947124481,
0.16687695682048798,
-0.138672336935997,
0.1921185404062271,
0.10397820919752121,
-0.048194143921136856,
-0.017421696335077286,
-0.13509781658649445,
0.04350683465600014,
-0.053846582770347595,
0.08814302831888199,
-0.06916266679763794,
0.0257099661976099,
-0.02199867181479931,
0.15369565784931183,
-0.12297247350215912,
-0.004363287705928087,
0.007302260026335716,
-0.11228931695222855,
-0.0744360163807869,
0.06613092124462128,
0.010827829129993916,
-0.2406279444694519,
0.137528195977211,
0.20888106524944305,
-0.014914384111762047,
0.17255091667175293,
0.04622611775994301,
-0.0024429222103208303,
-0.05853913351893425,
-0.022866355255246162,
-0.0051917447708547115,
0.05475935712456703,
0.007581597194075584,
-0.02161569893360138,
0.04127134010195732,
-0.03441675007343292,
0.10565537214279175,
-0.07691136002540588,
0.014146126806735992,
0.024519948288798332,
0.03616448864340782,
0.12529587745666504,
0.08221083879470825,
-0.04212677851319313,
0.1204916313290596,
0.03125748783349991,
-0.06731586158275604,
0.0061064851470291615,
-0.007744978182017803,
-0.07951667159795761,
0.16831733286380768,
-0.05236106738448143,
-0.18549753725528717,
-0.2275429666042328,
-0.04594036936759949,
-0.07951385527849197,
0.038413580507040024,
0.030304256826639175,
-0.03850942477583885,
-0.026364194229245186,
-0.07717027515172958,
0.08827978372573853,
0.004717460833489895,
-0.049235593527555466,
0.026471614837646484,
0.003997277468442917,
0.006486618425697088,
-0.08385270833969116,
-0.02132749930024147,
0.0038866824470460415,
0.027652082964777946,
0.07497484236955643,
-0.06250710040330887,
0.10719361156225204,
0.10189686715602875,
0.02096973918378353,
-0.03904910385608673,
-0.03582003712654114,
0.1100040152668953,
-0.04247278720140457,
0.1076250821352005,
0.1428145468235016,
-0.01674848608672619,
0.07443299144506454,
0.1794753521680832,
0.017297375947237015,
-0.0456533320248127,
-0.030303115025162697,
-0.0535927340388298,
-0.06083904206752777,
-0.1925661712884903,
-0.07857835292816162,
-0.050895776599645615,
0.012123367749154568,
0.042913977056741714,
0.004354733508080244,
0.20052474737167358,
0.11700686067342758,
-0.04702708125114441,
0.08653601258993149,
-0.0265284962952137,
0.05559989809989929,
0.2343696802854538,
-0.0025637750513851643,
0.12367797642946243,
-0.0743282288312912,
-0.05663694068789482,
0.03218929097056389,
0.13281646370887756,
0.05051789805293083,
0.06546427309513092,
0.11469639092683792,
0.04468422010540962,
0.04804505407810211,
0.08482087403535843,
0.06146598979830742,
-0.03637915104627609,
-0.03007274679839611,
-0.08126074075698853,
-0.08954259753227234,
-0.05541185289621353,
0.08562204241752625,
-0.0955655425786972,
0.01736891083419323,
-0.013703458942472935,
-0.003254722338169813,
0.039066608995199203,
0.07737189531326294,
0.03134366124868393,
-0.30997705459594727,
-0.029706258326768875,
0.112772636115551,
0.01941397786140442,
-0.055443283170461655,
-0.030724357813596725,
0.003726045601069927,
-0.033534057438373566,
0.13866086304187775,
-0.007899879477918148,
0.0743156224489212,
0.019880512729287148,
0.05053899809718132,
-0.10310538858175278,
0.09361473470926285,
-0.054358407855033875,
0.08818267285823822,
-0.24936628341674805,
0.19587849080562592,
0.05170449987053871,
-0.046005863696336746,
-0.002949983114376664,
0.03407403081655502,
0.03070196509361267,
0.10726450383663177,
0.057900503277778625,
-0.030872419476509094,
-0.06706113368272781,
-0.10514197498559952,
-0.09926611930131912,
0.02133527584373951,
-0.011879521422088146,
-0.02977857179939747,
0.09692627191543579,
-0.03607219085097313,
-0.03459659591317177,
-0.00099271594081074,
0.13374125957489014,
-0.12229081988334656,
-0.10843859612941742,
0.06755140423774719,
0.02350091002881527,
0.046160124242305756,
-0.07021445780992508,
-0.034547340124845505,
-0.1036846935749054,
0.10714063793420792,
0.015129324980080128,
-0.05609441548585892,
-0.08963285386562347,
-0.06899633258581161,
0.13102324306964874,
-0.07587902247905731,
0.04503031075000763,
-0.019795071333646774,
0.03787025436758995,
-0.07394304126501083,
-0.1537219136953354,
0.12339574843645096,
-0.13743695616722107,
-0.07856394350528717,
-0.063593789935112,
0.14106394350528717,
-0.08577281981706619,
0.02515375055372715,
0.03367543965578079,
0.013876890763640404,
-0.012480274774134159,
-0.005941860843449831,
0.01648714765906334,
0.038243744522333145,
-0.09128043055534363,
0.011189484037458897,
-0.0851626917719841,
-0.16503764688968658,
-0.005370985716581345,
-0.06072955206036568,
0.10049030929803848,
0.29966750741004944,
0.004502535797655582,
0.06437065452337265,
0.2366936206817627,
-0.10759139806032181,
-0.18380017578601837,
-0.08928226679563522,
-0.04828045889735222,
0.05805652216076851,
0.013096051290631294,
-0.12251630425453186,
0.010804662480950356,
0.1123766377568245,
-0.04313216358423233,
0.06977340579032898,
-0.2857990264892578,
-0.124092236161232,
0.13636420667171478,
0.03985196352005005,
0.26925280690193176,
-0.1821133941411972,
-0.11340880393981934,
-0.15149065852165222,
-0.02593526989221573,
0.01801419071853161,
-0.009773940779268742,
0.06895856559276581,
-0.044966068118810654,
-0.004582843277603388,
0.025345662608742714,
-0.0326446071267128,
0.16263063251972198,
-0.024696210399270058,
0.05680136755108833,
-0.0875943973660469,
-0.020660946145653725,
0.04893462359905243,
-0.08951592445373535,
0.04292724281549454,
-0.14072199165821075,
-0.011437475681304932,
-0.12602131068706512,
-0.038832634687423706,
-0.027035001665353775,
0.07902075350284576,
-0.03921911492943764,
-0.031271811574697495,
-0.0025083557702600956,
-0.0005774665623903275,
-0.03141814097762108,
-0.00880915205925703,
0.13927994668483734,
-0.06555450707674026,
0.17055244743824005,
0.20449824631214142,
0.100246362388134,
-0.019255075603723526,
0.09194879233837128,
0.01660945639014244,
-0.042186036705970764,
0.10053472220897675,
-0.15653014183044434,
-0.002463461132720113,
0.09780026972293854,
-0.02964913286268711,
0.05742305889725685,
0.0009226342081092298,
-0.05016656592488289,
-0.00751517154276371,
0.14792561531066895,
-0.138688325881958,
-0.00554738100618124,
-0.06721995770931244,
0.04435620456933975,
0.04233662784099579,
0.08275818824768066,
0.1829218566417694,
-0.03398524969816208,
-0.022541657090187073,
0.01813950017094612,
-0.007224414963275194,
-0.025562629103660583,
0.13836093246936798,
0.033398739993572235,
0.05009216070175171,
-0.05521296709775925,
0.043189164251089096,
0.05715108662843704,
-0.04091034084558487,
0.020541910082101822,
0.0854697972536087,
-0.04514375329017639,
-0.08539241552352905,
-0.09982433170080185,
0.21735861897468567,
-0.15719345211982727,
-0.07145572453737259,
-0.12632310390472412,
-0.09491995722055435,
-0.002210258971899748,
0.1790183037519455,
0.03793052211403847,
-0.022392060607671738,
0.055565182119607925,
-0.05809188261628151,
0.012791906483471394,
0.06947243213653564,
0.045553527772426605,
0.11185254156589508,
-0.10450665652751923,
-0.03165821358561516,
-0.019477449357509613,
0.04347068443894386,
-0.0637945905327797,
0.006163507234305143,
-0.16065451502799988,
-0.02890251949429512,
-0.23411205410957336,
-0.008008638396859169,
-0.11911755055189133,
-0.06581390649080276,
-0.02137886919081211,
0.012058726511895657,
-0.030234524980187416,
-0.027430234476923943,
-0.04623601958155632,
0.01912051998078823,
-0.024870416149497032,
0.11799884587526321,
-0.05085662379860878,
-0.006409418769180775,
0.02740292064845562,
-0.04676646739244461,
0.03536301851272583,
-0.0140827102586627,
0.026041913777589798,
-0.002842130372300744,
-0.07040537893772125,
0.0052866158075630665,
0.033172059804201126,
-0.009323444217443466,
0.03668215498328209,
-0.06279374659061432,
0.00319810607470572,
-0.004124943166971207,
0.018153132870793343,
-0.02674579620361328,
0.02357572317123413,
-0.086324043571949,
0.034598927944898605,
-0.00694398395717144,
-0.0844913199543953,
0.0027815098874270916,
0.019418245181441307,
0.09677337855100632,
-0.038656167685985565,
0.0652269497513771,
-0.08791626244783401,
0.014056033454835415,
-0.16909848153591156,
0.011423854157328606,
-0.02262689173221588,
-0.146665558218956,
-0.024647017940878868,
-0.12421461939811707,
0.0004727649502456188,
-0.029754547402262688,
0.1600804626941681,
0.04253237694501877,
-0.07242340594530106,
-0.0047152903862297535,
-0.027834109961986542,
0.02656789869070053,
0.06505955010652542,
0.23269438743591309,
0.06533578038215637,
0.048585474491119385,
-0.10224930197000504,
0.09122815728187561,
0.05132210999727249,
0.005935986526310444,
-0.003694944316521287,
0.09521042555570602,
-0.022286485880613327,
0.12666258215904236,
0.12731456756591797,
-0.10770092159509659,
0.016396231949329376,
-0.025827130302786827,
-0.05260932818055153,
0.03591989353299141,
0.013380087912082672,
0.08869991451501846,
0.18315619230270386,
-0.09791294485330582,
0.032235898077487946,
-0.02443057671189308,
0.0027021351270377636,
-0.06751736998558044,
-0.11905962228775024,
-0.09869550168514252,
-0.13627350330352783,
-0.04853851720690727,
-0.09044503420591354,
-0.03129321709275246,
0.1276341676712036,
-0.00440851878374815,
-0.00021635468874592334,
0.19331997632980347,
0.06017130985856056,
-0.09577508270740509,
0.030084192752838135,
0.02363571524620056,
-0.0002820150984916836,
-0.09041930735111237,
-0.04886027052998543,
-0.000023433593014487997,
0.029109468683600426,
-0.08395570516586304,
-0.011865111999213696,
0.029837200418114662,
0.061848755925893784,
-0.0015978313749656081,
-0.09885574132204056,
-0.04380110651254654,
0.010259199887514114,
0.06933782994747162,
-0.03748074918985367,
-0.007614876143634319,
-0.043653033673763275,
-0.010945585556328297,
0.12895606458187103,
-0.05385160818696022,
-0.0409851111471653,
0.02227054163813591,
0.2008453905582428,
-0.050344787538051605,
0.06160891428589821,
-0.044173918664455414,
-0.037577662616968155,
0.03496931120753288,
0.15927323698997498,
0.3319663405418396,
-0.07262764126062393,
-0.007973908446729183,
-0.05434046685695648,
0.023856056854128838,
0.04172433912754059,
0.09744922071695328,
0.010740046389400959,
0.15390177071094513,
-0.06885429471731186,
0.07854703068733215,
-0.023248989135026932,
-0.02610059641301632,
-0.026777105405926704,
0.1248249039053917,
0.018599996343255043,
-0.04322824254631996,
0.04204728081822395,
0.0769956186413765,
-0.047161027789115906,
-0.10937133431434631,
-0.0014581349678337574,
-0.16970402002334595,
-0.09206154942512512,
-0.08724605292081833,
0.10422681272029877,
0.01719348505139351,
0.06359653174877167,
-0.043445486575365067,
-0.031899310648441315,
0.05180032178759575,
-0.0080159492790699,
-0.09409540146589279,
-0.09619711339473724,
0.044675763696432114,
-0.07952568680047989,
0.039588652551174164,
-0.05059082806110382,
0.03961896151304245,
0.10513511300086975,
-0.036711499094963074,
-0.06946666538715363,
0.044977471232414246,
0.05568789318203926,
-0.053327012807130814,
0.05712924152612686,
-0.023431671783328056,
0.03857401758432388,
0.12088122218847275,
0.06774412095546722,
-0.15311336517333984,
0.012862387113273144,
0.05658132582902908,
-0.027430828660726547,
-0.0208780188113451,
0.04463784396648407,
-0.0516679510474205,
0.1466076374053955,
0.08152420818805695,
-0.05164617672562599,
-0.02719232626259327,
-0.013003949075937271,
0.04113956168293953,
0.10378215461969376,
0.09898167103528976,
-0.023729581385850906,
-0.2016320377588272,
-0.03716769069433212,
-0.07870142161846161,
0.024260638281702995,
-0.3307252526283264,
-0.03432193771004677,
-0.1363302618265152,
-0.005638754460960627,
-0.0984654650092125,
0.05417736992239952,
0.18993400037288666,
0.04184045270085335,
-0.011164783500134945,
-0.1518934667110443,
-0.0325041227042675,
0.08790484815835953,
-0.08656570315361023,
-0.06904057413339615
] |
null | null | null |
# Detector Guidance for Multi-Object Text-to-Image Generation
by [Luping Liu](https://luping-liu.github.io/)<sup>1</sup>, Zijian Zhang<sup>1</sup>, [Yi Ren](https://rayeren.github.io/)<sup>2</sup>, Rongjie Huang<sup>1</sup>, Zhou Zhao<sup>1</sup>.
<sup>1</sup>Zhejiang University, <sup>2</sup>ByteDance
In this work, we introduce Detector Guidance (DG), which integrates a latent object detection model to separate different objects during the generation process. More precisely, DG first performs latent object detection on cross-attention maps (CAMs) to obtain object information. Based on this information, DG then masks conflicting prompts and enhances related prompts by manipulating the following CAMs. Human evaluations demonstrate that DG provides an 8-22% advantage in preventing the amalgamation of conflicting concepts and ensuring that each object possesses its unique region without any human involvement and additional iterations.
| {"license": "apache-2.0"} | null | luping-liu/Detector_Guidance | [
"license:apache-2.0",
"region:us"
] | 2024-02-07T08:31:10+00:00 | [] | [] | TAGS
#license-apache-2.0 #region-us
|
# Detector Guidance for Multi-Object Text-to-Image Generation
by Luping Liu<sup>1</sup>, Zijian Zhang<sup>1</sup>, Yi Ren<sup>2</sup>, Rongjie Huang<sup>1</sup>, Zhou Zhao<sup>1</sup>.
<sup>1</sup>Zhejiang University, <sup>2</sup>ByteDance
In this work, we introduce Detector Guidance (DG), which integrates a latent object detection model to separate different objects during the generation process. More precisely, DG first performs latent object detection on cross-attention maps (CAMs) to obtain object information. Based on this information, DG then masks conflicting prompts and enhances related prompts by manipulating the following CAMs. Human evaluations demonstrate that DG provides an 8-22% advantage in preventing the amalgamation of conflicting concepts and ensuring that each object possesses its unique region without any human involvement and additional iterations.
| [
"# Detector Guidance for Multi-Object Text-to-Image Generation\n\nby Luping Liu<sup>1</sup>, Zijian Zhang<sup>1</sup>, Yi Ren<sup>2</sup>, Rongjie Huang<sup>1</sup>, Zhou Zhao<sup>1</sup>.\n\n<sup>1</sup>Zhejiang University, <sup>2</sup>ByteDance\n\nIn this work, we introduce Detector Guidance (DG), which integrates a latent object detection model to separate different objects during the generation process. More precisely, DG first performs latent object detection on cross-attention maps (CAMs) to obtain object information. Based on this information, DG then masks conflicting prompts and enhances related prompts by manipulating the following CAMs. Human evaluations demonstrate that DG provides an 8-22% advantage in preventing the amalgamation of conflicting concepts and ensuring that each object possesses its unique region without any human involvement and additional iterations."
] | [
"TAGS\n#license-apache-2.0 #region-us \n",
"# Detector Guidance for Multi-Object Text-to-Image Generation\n\nby Luping Liu<sup>1</sup>, Zijian Zhang<sup>1</sup>, Yi Ren<sup>2</sup>, Rongjie Huang<sup>1</sup>, Zhou Zhao<sup>1</sup>.\n\n<sup>1</sup>Zhejiang University, <sup>2</sup>ByteDance\n\nIn this work, we introduce Detector Guidance (DG), which integrates a latent object detection model to separate different objects during the generation process. More precisely, DG first performs latent object detection on cross-attention maps (CAMs) to obtain object information. Based on this information, DG then masks conflicting prompts and enhances related prompts by manipulating the following CAMs. Human evaluations demonstrate that DG provides an 8-22% advantage in preventing the amalgamation of conflicting concepts and ensuring that each object possesses its unique region without any human involvement and additional iterations."
] | [
14,
260
] | [
"passage: TAGS\n#license-apache-2.0 #region-us \n# Detector Guidance for Multi-Object Text-to-Image Generation\n\nby Luping Liu<sup>1</sup>, Zijian Zhang<sup>1</sup>, Yi Ren<sup>2</sup>, Rongjie Huang<sup>1</sup>, Zhou Zhao<sup>1</sup>.\n\n<sup>1</sup>Zhejiang University, <sup>2</sup>ByteDance\n\nIn this work, we introduce Detector Guidance (DG), which integrates a latent object detection model to separate different objects during the generation process. More precisely, DG first performs latent object detection on cross-attention maps (CAMs) to obtain object information. Based on this information, DG then masks conflicting prompts and enhances related prompts by manipulating the following CAMs. Human evaluations demonstrate that DG provides an 8-22% advantage in preventing the amalgamation of conflicting concepts and ensuring that each object possesses its unique region without any human involvement and additional iterations."
] | [
-0.0680304542183876,
-0.05550987645983696,
-0.004830115009099245,
0.040299903601408005,
0.0057282038033008575,
0.01714492402970791,
0.14682917296886444,
0.08979441970586777,
0.07473848015069962,
-0.009799034334719181,
-0.030522648245096207,
0.06919988244771957,
0.08090291172266006,
-0.039517439901828766,
0.043947480618953705,
-0.24370047450065613,
0.11732348799705505,
-0.035415299236774445,
-0.13173313438892365,
0.04936479777097702,
0.05259519815444946,
-0.04144405946135521,
0.05297888442873955,
-0.0037227680440992117,
-0.03951280564069748,
-0.04818711057305336,
-0.06502984464168549,
0.0073349131271243095,
0.15612100064754486,
-0.0069070723839104176,
0.0021195353474467993,
0.09070582687854767,
0.04836870729923248,
-0.15722425282001495,
0.014391285367310047,
0.01967058889567852,
0.00932113267481327,
-0.019466908648610115,
-0.0005225840723142028,
0.146828293800354,
0.15942399203777313,
-0.004854023922234774,
0.023535948246717453,
-0.015241979621350765,
-0.039576202630996704,
-0.05547785013914108,
-0.1213911846280098,
-0.008607297204434872,
0.13564017415046692,
0.05399682745337486,
-0.003413761267438531,
0.060383301228284836,
-0.14360414445400238,
-0.008750315755605698,
-0.06419114768505096,
-0.27077624201774597,
0.02369205467402935,
0.16431649029254913,
0.04672883450984955,
0.09538447111845016,
-0.04292877018451691,
0.033401548862457275,
-0.022178292274475098,
-0.0459357388317585,
-0.054452307522296906,
-0.06052382290363312,
0.21284149587154388,
-0.030795691534876823,
-0.14152291417121887,
-0.05227717384696007,
0.27294421195983887,
0.05064401030540466,
-0.1035221740603447,
-0.13378272950649261,
-0.05841205269098282,
0.06341962516307831,
-0.04477698728442192,
-0.11260195076465607,
0.06600497663021088,
0.11407450586557388,
0.0955156460404396,
-0.05499576777219772,
-0.08373087644577026,
-0.14638811349868774,
-0.01957569271326065,
0.061168622225522995,
0.06008275970816612,
0.003194943070411682,
-0.08701051026582718,
0.10529055446386337,
0.09386710077524185,
-0.06323256343603134,
-0.15793199837207794,
-0.08209697157144547,
-0.12263008207082748,
-0.0017460015369579196,
-0.016093598678708076,
-0.07451608031988144,
0.11515379697084427,
0.1797642707824707,
0.09725061804056168,
0.04506560042500496,
-0.08836355060338974,
0.013462809845805168,
0.038544751703739166,
0.10799596458673477,
-0.06577557325363159,
0.14531362056732178,
0.03084978647530079,
0.04357830435037613,
0.08365872502326965,
-0.09670905023813248,
-0.05345494672656059,
-0.08055681735277176,
0.03238077834248543,
0.053471896797418594,
-0.02666562609374523,
0.15728004276752472,
-0.012067161500453949,
-0.047212231904268265,
-0.06389646977186203,
-0.09878787398338318,
-0.030894145369529724,
0.0026456189807504416,
-0.02854488044977188,
-0.03528787940740585,
0.0025807907804846764,
0.07270845770835876,
0.006663480307906866,
0.03270541504025459,
-0.030508702620863914,
-0.03821565955877304,
-0.008615159429609776,
-0.03114480711519718,
0.01466185599565506,
0.025432422757148743,
-0.011586803011596203,
-0.10927698016166687,
-0.1712173968553543,
0.0053452677093446255,
0.0010735787218436599,
-0.0012304696720093489,
0.0047624194994568825,
-0.04467564821243286,
0.015513285994529724,
-0.036603350192308426,
0.008710023947060108,
-0.07185191661119461,
-0.05202598124742508,
0.045342426747083664,
0.039852965623140335,
0.05650005862116814,
-0.07727935910224915,
0.009369452483952045,
-0.060050271451473236,
0.11152713745832443,
-0.05602819472551346,
0.1907910257577896,
-0.08522171527147293,
0.04226546734571457,
-0.0034440220333635807,
0.054544657468795776,
-0.06241503357887268,
-0.007803658954799175,
0.00037828207132406533,
0.12384424358606339,
-0.1737428456544876,
-0.03684219345450401,
0.1311071664094925,
-0.17130394279956818,
-0.06560797244310379,
0.04027824103832245,
-0.034350134432315826,
0.08899104595184326,
0.0597325935959816,
0.17616020143032074,
0.06462229043245316,
-0.07554806768894196,
0.037459596991539,
0.06390051543712616,
-0.1008613109588623,
0.16762198507785797,
0.08559916168451309,
0.042747821658849716,
-0.18907243013381958,
0.036372967064380646,
-0.1880159229040146,
0.011981743387877941,
0.00885096937417984,
-0.09231198579072952,
0.021119114011526108,
-0.0023946543224155903,
-0.01954556256532669,
-0.0719110518693924,
0.030208779498934746,
0.048398856073617935,
-0.0003494212869554758,
-0.15572071075439453,
0.12707281112670898,
-0.08427450060844421,
0.021728618070483208,
-0.13909126818180084,
0.033127374947071075,
0.040469132363796234,
0.10620996356010437,
-0.11485151201486588,
-0.06981901079416275,
0.10671865195035934,
0.004375504329800606,
-0.028282830491662025,
0.02896442636847496,
-0.022489670664072037,
0.05311175435781479,
0.08984240144491196,
-0.0065353019163012505,
-0.0864831805229187,
-0.03591194376349449,
0.08562970161437988,
-0.03944623842835426,
0.0066368854604661465,
-0.04772518202662468,
0.2481515258550644,
-0.06211896985769272,
-0.003485976019874215,
0.10992878675460815,
-0.09448381513357162,
0.01943543553352356,
0.009214847348630428,
-0.011811098083853722,
-0.024326598271727562,
-0.04629594832658768,
-0.045598167926073074,
0.10246330499649048,
0.045505721122026443,
-0.1565735936164856,
0.07415471971035004,
-0.08942621201276779,
-0.05862288922071457,
0.0722455158829689,
0.042560648173093796,
-0.07585830241441727,
-0.11137978732585907,
-0.04577343910932541,
-0.041776590049266815,
0.07328026741743088,
-0.04145551100373268,
0.09625086188316345,
0.01754898764193058,
0.04445160925388336,
-0.03947533667087555,
0.03311300650238991,
0.011610928922891617,
0.04059161618351936,
0.0006190672283992171,
-0.013339085504412651,
0.09255320578813553,
0.03381967172026634,
-0.010013658553361893,
-0.018569784238934517,
0.09380555897951126,
0.01258997991681099,
-0.07132703065872192,
-0.03608005493879318,
-0.03320295736193657,
0.14912854135036469,
0.008225291036069393,
0.06111099198460579,
-0.15130794048309326,
-0.01862722635269165,
0.00966683216392994,
0.07078880816698074,
0.004496896639466286,
-0.17906077206134796,
-0.005222070962190628,
0.01586003042757511,
-0.0844801589846611,
-0.08557359129190445,
0.041043348610401154,
-0.07421398907899857,
0.06428244709968567,
0.06470561027526855,
-0.04521747678518295,
0.0754716619849205,
-0.05343899503350258,
-0.08564689010381699,
0.11045581102371216,
-0.05661939084529877,
-0.3207661211490631,
-0.02887639030814171,
-0.09617968648672104,
-0.09664054960012436,
0.10023564100265503,
0.05527061969041824,
-0.02905530110001564,
-0.02240826189517975,
-0.04769646003842354,
0.11334022134542465,
0.004584385547786951,
-0.022405585274100304,
0.09952392429113388,
-0.04453452676534653,
0.0076139685697853565,
-0.14463423192501068,
-0.0153016597032547,
-0.04157670587301254,
-0.11593268066644669,
0.07823479920625687,
-0.06195877864956856,
0.11630382388830185,
0.07500696182250977,
0.07844798266887665,
-0.010779166594147682,
0.0012210191925987601,
0.13814590871334076,
-0.06671492010354996,
-0.04441661387681961,
0.10597605258226395,
-0.06313115358352661,
0.019436854869127274,
0.15531468391418457,
-0.014316613785922527,
-0.17071576416492462,
0.0794418677687645,
-0.016414359211921692,
-0.055311571806669235,
-0.11765532195568085,
-0.03910483419895172,
-0.09382564574480057,
0.030095906928181648,
0.020121948793530464,
0.08169107884168625,
0.09935413300991058,
0.05909807235002518,
0.026369452476501465,
0.12798215448856354,
0.14806489646434784,
0.061579540371894836,
0.11903785914182663,
-0.03209695965051651,
-0.004110350273549557,
-0.03167946636676788,
-0.1178705245256424,
0.04585021734237671,
0.06625573337078094,
0.22115114331245422,
0.017600901424884796,
0.06227109581232071,
0.08889792114496231,
0.11110025644302368,
0.03218310326337814,
0.010627895593643188,
-0.006623005028814077,
0.03714798018336296,
-0.03469427675008774,
-0.0816599652171135,
-0.07601351290941238,
0.012742516584694386,
-0.040221136063337326,
-0.15477855503559113,
-0.05453130230307579,
0.12932522594928741,
0.04025157913565636,
0.27289390563964844,
-0.015891334041953087,
-0.0936952605843544,
0.03943757712841034,
-0.022939419373869896,
0.08831626176834106,
-0.025545764714479446,
0.08006908744573593,
-0.00006228470738278702,
-0.06436467915773392,
0.04393796622753143,
-0.02070632018148899,
0.06095987930893898,
-0.17825941741466522,
0.04482639208436012,
-0.04163988679647446,
-0.009790612384676933,
0.04937667399644852,
0.04765595123171806,
-0.20476019382476807,
0.05785377323627472,
-0.011190081015229225,
0.009280180558562279,
-0.05246315896511078,
0.07263260334730148,
0.0385766476392746,
0.05654534697532654,
0.1346898227930069,
0.01959574781358242,
-0.08491913974285126,
-0.06842447072267532,
-0.07598099857568741,
0.07781066000461578,
0.09692417085170746,
-0.07745812833309174,
0.040516968816518784,
-0.051048193126916885,
0.03983651474118233,
-0.034884072840213776,
0.13104049861431122,
-0.17372886836528778,
-0.14301946759223938,
0.1031254306435585,
-0.03585559502243996,
0.14437329769134521,
-0.058648157864809036,
0.036429163068532944,
-0.02784118615090847,
0.15273678302764893,
-0.04164392501115799,
-0.00014017854118719697,
-0.06069343909621239,
-0.007536869961768389,
-0.015775106847286224,
0.017585350200533867,
-0.011357564479112625,
0.018553389236330986,
0.06937005370855331,
-0.04839060828089714,
-0.11727499216794968,
0.07322126626968384,
0.004670899361371994,
-0.027867980301380157,
-0.0011905243154615164,
0.08207622170448303,
0.03551940992474556,
-0.029923291876912117,
0.046896398067474365,
-0.11429198086261749,
0.11645087599754333,
-0.10448548197746277,
0.04038700461387634,
0.14313620328903198,
-0.005544552579522133,
0.05019408464431763,
-0.06823989003896713,
0.006646177731454372,
-0.10425425320863724,
-0.03002728708088398,
0.09050624817609787,
0.2945263087749481,
-0.018310176208615303,
0.1254364401102066,
0.052875105291604996,
-0.04833528771996498,
-0.22228474915027618,
-0.04722307249903679,
-0.025079306215047836,
-0.02859179861843586,
-0.10388246178627014,
-0.2231770157814026,
-0.03856587037444115,
0.09824111312627792,
-0.05320170894265175,
0.13698343932628632,
-0.2934175431728363,
-0.09290669858455658,
0.05475859344005585,
0.015440675429999828,
0.20467771589756012,
-0.04699782282114029,
-0.03846333175897598,
-0.012926957570016384,
-0.15051844716072083,
0.11236047744750977,
0.1600690335035324,
0.10169930011034012,
0.008634201250970364,
-0.04209980368614197,
0.05011729151010513,
0.013995110988616943,
0.11383350938558578,
0.0645444467663765,
0.13405825197696686,
-0.09258300065994263,
0.0041571930050849915,
-0.007255101576447487,
-0.03504735976457596,
0.13086967170238495,
0.03267524391412735,
0.017723552882671356,
0.0866984874010086,
-0.03553587198257446,
0.048094578087329865,
0.0047563486732542515,
0.016069285571575165,
-0.0853775218129158,
-0.10725748538970947,
0.028509752824902534,
0.0017718476010486484,
0.04198293760418892,
0.12297108769416809,
0.04715527966618538,
-0.09859167784452438,
-0.09919678419828415,
0.020250653848052025,
-0.04858846589922905,
0.002498330781236291,
-0.0764140635728836,
0.034398362040519714,
0.13715964555740356,
-0.11865224689245224,
0.08070165663957596,
0.05716235563158989,
0.0028320292476564646,
0.1318393498659134,
0.10445841401815414,
-0.06606542319059372,
0.047445956617593765,
0.035718031227588654,
-0.12482315301895142,
-0.16361823678016663,
-0.030870681628584862,
-0.038559962064027786,
0.08882445096969604,
0.04037759080529213,
0.07975880801677704,
-0.008566012606024742,
0.017568517476320267,
0.039693910628557205,
0.01956895738840103,
0.0004455708258319646,
0.03138093650341034,
-0.1300089806318283,
-0.010652847588062286,
-0.07323253154754639,
0.15799593925476074,
0.029740765690803528,
-0.10186228901147842,
-0.000055374322982970625,
-0.0869874358177185,
-0.16346700489521027,
-0.033925049006938934,
-0.2178575098514557,
0.11693120002746582,
0.04017551988363266,
-0.04844750091433525,
0.030852969735860825,
-0.10063638538122177,
-0.0006602047360502183,
-0.08093471825122833,
0.045459259301424026,
0.15481141209602356,
0.00019672395137604326,
0.011100267991423607,
-0.029444938525557518,
-0.0011689686216413975,
-0.1000509262084961,
0.06675533950328827,
-0.18388834595680237,
-0.11377367377281189,
0.05088388919830322,
0.0885624960064888,
-0.05343258008360863,
-0.14760787785053253,
-0.1076100692152977,
-0.02276758663356304,
-0.12610282003879547,
0.08932528644800186,
-0.09027088433504105,
0.01598319597542286,
0.029456518590450287,
0.06606326997280121,
0.04041384160518646,
0.044295232743024826,
-0.03650940582156181,
-0.056451089680194855,
-0.050229430198669434,
0.0188197810202837,
-0.010352631099522114,
-0.048589494079351425,
0.04434032738208771,
-0.12228406220674515,
0.16295351088047028,
0.09829382598400116,
-0.01448286883533001,
-0.007615899667143822,
-0.0831679105758667,
-0.017987720668315887,
0.08096162229776382,
0.04303385317325592,
-0.04906942695379257,
-0.11992689222097397,
-0.03494291380047798,
0.030642513185739517,
-0.10891448706388474,
-0.022176196798682213,
0.14635467529296875,
-0.11201068013906479,
0.04789137840270996,
-0.12250310182571411,
-0.05070241913199425,
-0.05229094251990318,
-0.02867022715508938,
0.11200051754713058,
0.14199258387088776,
0.12448133528232574,
-0.04874875769019127,
0.010968833230435848,
-0.07015111297369003,
-0.0554792694747448,
0.05750618129968643,
0.012676926329731941,
0.011829400435090065,
-0.10241231322288513,
0.008608142845332623,
-0.047406259924173355,
0.19175362586975098,
-0.07229046523571014,
0.03331032767891884,
-0.04400256648659706,
0.040128543972969055,
0.055231548845767975,
0.00133641401771456,
0.19341421127319336,
0.09781838953495026,
-0.009964712895452976,
0.011151299811899662,
0.03756784275174141,
0.04805769398808479,
-0.05297417566180229,
-0.006089739501476288,
-0.04392333701252937,
0.03969455510377884,
-0.001826925203204155,
0.0568518303334713,
-0.01881697028875351,
-0.14678865671157837,
0.05301256477832794,
-0.008450526744127274,
0.00800328329205513,
-0.053702160716056824,
0.033080846071243286,
0.14606283605098724,
-0.027776896953582764,
0.1217838004231453,
0.0009684502729214728,
-0.027442442253232002,
-0.07729458808898926,
-0.21404103934764862,
0.0008517460664734244,
-0.18875455856323242,
0.041184697300195694,
-0.014922816306352615,
-0.06457532197237015,
0.04622230678796768,
0.018170863389968872,
-0.06008380651473999,
0.08127977699041367,
-0.13050569593906403,
-0.06275294721126556,
-0.03261435031890869,
-0.052324529737234116,
-0.09650544822216034,
0.0690019354224205,
-0.08448043465614319,
0.10239097476005554,
0.05701137334108353,
0.05564697086811066,
0.012405142188072205,
0.11942300945520401,
0.04832451045513153,
-0.06575184315443039,
0.03257005289196968,
0.0015942768659442663,
-0.0009158271714113653,
-0.04016215726733208,
0.0048911976628005505,
0.09113481640815735,
-0.08245125412940979,
0.04200262203812599,
0.1525431126356125,
0.03145696222782135,
-0.01815124601125717,
-0.10989166051149368,
0.10654747486114502,
-0.06381569802761078,
0.06663985550403595,
0.035002995282411575,
-0.08963914215564728,
-0.038284774869680405,
0.16863314807415009,
0.11645390093326569,
-0.0834316834807396,
-0.03889072686433792,
-0.012714172713458538,
0.014123181812465191,
-0.038969896733760834,
0.12879407405853271,
0.06572423130273819,
0.1704028844833374,
0.06420160084962845,
-0.0038894247263669968,
-0.057053253054618835,
0.007633424364030361,
-0.0003433499950915575,
0.14364266395568848,
0.0440136082470417,
-0.0029987560119479895,
-0.08627410233020782,
0.0418059267103672,
-0.06887883692979813,
-0.21229170262813568,
0.034697920083999634,
0.06952516734600067,
-0.0044992584735155106,
0.025269342586398125,
0.006528396625071764,
-0.03795430436730385,
0.05065863952040672,
-0.07416199892759323,
0.023654909804463387,
-0.02689184434711933,
0.022823922336101532,
-0.09593658894300461,
-0.020260246470570564,
0.057609330862760544,
0.14145433902740479,
0.11780741810798645,
0.006584266200661659,
0.09062225371599197,
-0.006430352106690407,
0.1059618890285492,
-0.11378175020217896,
0.10049550980329514,
-0.02471970207989216,
-0.025656459853053093,
-0.049835722893476486,
0.025228697806596756,
-0.036457888782024384,
0.08623944967985153,
0.08370768278837204,
-0.02672981098294258,
0.03386209160089493,
0.04672880470752716,
-0.03433462604880333,
-0.09332150965929031,
0.011219956912100315,
-0.08059554547071457,
0.054298125207424164,
0.04234005883336067,
0.008613893762230873,
0.10066726803779602,
-0.06880296766757965,
0.012453480623662472,
-0.06377698481082916,
0.03656034544110298,
-0.07916896045207977,
-0.047407716512680054,
-0.015141159296035767,
0.05715208873152733,
-0.01347300037741661,
-0.06437816470861435,
-0.1333644986152649,
0.005091727711260319,
0.008325530216097832,
-0.011402811855077744,
0.07110185921192169,
0.08657783269882202,
0.03727361187338829,
-0.053154513239860535,
-0.21468639373779297,
0.045212339609861374,
0.07572522759437561,
-0.07147444784641266,
-0.07461611181497574
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# phi-2-ft
This model is a fine-tuned version of [microsoft/phi-2](https://huggingface.co/microsoft/phi-2) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 2
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 1
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1 | {"license": "mit", "library_name": "peft", "tags": ["trl", "sft", "generated_from_trainer"], "base_model": "microsoft/phi-2", "model-index": [{"name": "phi-2-ft", "results": []}]} | null | mikeee/phi-2-ft | [
"peft",
"safetensors",
"trl",
"sft",
"generated_from_trainer",
"base_model:microsoft/phi-2",
"license:mit",
"region:us"
] | 2024-02-07T08:32:57+00:00 | [] | [] | TAGS
#peft #safetensors #trl #sft #generated_from_trainer #base_model-microsoft/phi-2 #license-mit #region-us
|
# phi-2-ft
This model is a fine-tuned version of microsoft/phi-2 on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 2
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 1
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1 | [
"# phi-2-ft\n\nThis model is a fine-tuned version of microsoft/phi-2 on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 2\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#peft #safetensors #trl #sft #generated_from_trainer #base_model-microsoft/phi-2 #license-mit #region-us \n",
"# phi-2-ft\n\nThis model is a fine-tuned version of microsoft/phi-2 on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 2\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
41,
28,
6,
12,
8,
3,
89,
4,
39
] | [
"passage: TAGS\n#peft #safetensors #trl #sft #generated_from_trainer #base_model-microsoft/phi-2 #license-mit #region-us \n# phi-2-ft\n\nThis model is a fine-tuned version of microsoft/phi-2 on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 2\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1### Training results### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.12534494698047638,
0.09130492806434631,
-0.0011182319140061736,
0.10207148641347885,
0.125119149684906,
-0.011595148593187332,
0.11447415500879288,
0.08526220917701721,
-0.07133205980062485,
0.12144885957241058,
0.10835271328687668,
0.037459637969732285,
0.05519457161426544,
0.25653648376464844,
-0.04165825992822647,
-0.21809688210487366,
0.0689280554652214,
-0.023147642612457275,
0.06935212016105652,
0.09803628921508789,
0.08332604169845581,
-0.09219539165496826,
0.09228436648845673,
-0.01327939610928297,
-0.14709703624248505,
-0.03067774698138237,
-0.011855307035148144,
-0.04231090843677521,
0.08534927666187286,
0.005363185424357653,
0.08702978491783142,
-0.000005156245606485754,
0.11054041981697083,
-0.17309771478176117,
0.003343293908983469,
0.058953143656253815,
0.02930571511387825,
0.09439331293106079,
0.06998597085475922,
0.04437100887298584,
0.0605781264603138,
-0.053529880940914154,
0.07015712559223175,
0.013377557508647442,
-0.11783692240715027,
-0.136551633477211,
-0.12103836983442307,
0.10375583171844482,
0.06894737482070923,
0.09446141868829727,
-0.0021739755757153034,
0.2295517474412918,
-0.04547223076224327,
0.04937417060136795,
0.22799687087535858,
-0.29730919003486633,
-0.06709539890289307,
0.03699105978012085,
0.05084281414747238,
0.08481688052415848,
-0.0862722173333168,
-0.052190277725458145,
0.07379192858934402,
0.0592615082859993,
0.09141773730516434,
0.013918372802436352,
-0.013604381121695042,
-0.06322258710861206,
-0.15973782539367676,
-0.06202835589647293,
0.15685030817985535,
0.05900172144174576,
-0.07156236469745636,
-0.06471063196659088,
-0.09463859349489212,
-0.1285022497177124,
-0.02047284133732319,
-0.022446706891059875,
0.01524352841079235,
-0.03676094859838486,
-0.04404908046126366,
-0.0475386306643486,
-0.07007773965597153,
-0.060924675315618515,
-0.03378467261791229,
0.12527069449424744,
0.05772895738482475,
0.05418967083096504,
-0.005060540046542883,
0.10074103623628616,
-0.07087066769599915,
-0.08608140796422958,
-0.03833341971039772,
-0.0368066132068634,
0.003947911784052849,
-0.04775870591402054,
-0.03202337026596069,
0.054386790841817856,
0.008304383605718613,
0.1474214494228363,
-0.1280193030834198,
0.03567737340927124,
0.02499394863843918,
0.030542658641934395,
-0.0357174389064312,
0.07853435724973679,
-0.06599787622690201,
0.057528071105480194,
0.03447568789124489,
0.11028733849525452,
0.06733438372612,
-0.0042098634876310825,
-0.09886469691991806,
0.009733013808727264,
0.11938565969467163,
0.03258911147713661,
-0.03433321416378021,
0.011635065078735352,
-0.05949987843632698,
-0.01858230121433735,
0.09844154119491577,
-0.09810478985309601,
0.04603514447808266,
-0.000039233298593899235,
-0.045170143246650696,
-0.05111643299460411,
0.03596470504999161,
0.009158633649349213,
-0.04255165159702301,
0.053362391889095306,
-0.09504436701536179,
0.01829017512500286,
-0.07532073557376862,
-0.06949753314256668,
0.03411742299795151,
-0.08583294600248337,
-0.014774227514863014,
-0.11791007220745087,
-0.18563315272331238,
-0.017190463840961456,
0.00033628116943873465,
-0.06439540535211563,
-0.05447594076395035,
-0.003278605407103896,
-0.08848177641630173,
-0.01792038418352604,
-0.04195239022374153,
0.04926203936338425,
-0.06710069626569748,
0.0900314673781395,
0.024811992421746254,
0.0358208604156971,
-0.09423783421516418,
0.02164938673377037,
-0.08765589445829391,
0.03135334327816963,
-0.15129345655441284,
0.036077916622161865,
-0.07849589735269547,
0.020958267152309418,
-0.08628208935260773,
-0.10202617198228836,
-0.04497208818793297,
-0.02658948488533497,
0.09637857228517532,
0.14399407804012299,
-0.1459452360868454,
0.006930463947355747,
0.18289242684841156,
-0.13556519150733948,
-0.11139634996652603,
0.09747283160686493,
-0.006313533987849951,
0.024519912898540497,
0.05654675140976906,
0.18968425691127777,
0.02146471105515957,
-0.1979200839996338,
-0.007627902086824179,
0.0037863468751311302,
0.06672437489032745,
-0.08201047778129578,
0.0829598531126976,
-0.03845583647489548,
0.013917668722569942,
0.02885206788778305,
-0.0714055746793747,
-0.0020712423138320446,
-0.10993587225675583,
-0.0574231892824173,
-0.067168228328228,
-0.09341656416654587,
0.025701558217406273,
0.028329038992524147,
0.034186217933893204,
-0.07332057505846024,
-0.05625041946768761,
0.12382510304450989,
0.10930654406547546,
-0.028912678360939026,
0.016706913709640503,
-0.08998054265975952,
0.13164009153842926,
-0.05552344396710396,
-0.04456064850091934,
-0.20717746019363403,
-0.08544867485761642,
0.044854357838630676,
-0.03219655156135559,
0.007965384051203728,
-0.027355514466762543,
0.0483783483505249,
0.09935986250638962,
-0.03877436742186546,
-0.011837419122457504,
-0.0983409658074379,
0.002302258973941207,
-0.1427859514951706,
-0.13221557438373566,
-0.06185019016265869,
-0.01817280240356922,
0.14102092385292053,
-0.20992134511470795,
0.021089093759655952,
-0.0014772767899557948,
0.15357795357704163,
0.035434965044260025,
-0.07442397624254227,
-0.007719163782894611,
0.0739765465259552,
0.0030274312011897564,
-0.09473074972629547,
0.049468349665403366,
0.044899262487888336,
-0.031691454350948334,
-0.08559193462133408,
-0.14208628237247467,
0.14641335606575012,
0.11511627584695816,
0.11319281160831451,
-0.07250075042247772,
-0.01779603585600853,
-0.0813654288649559,
-0.017332829535007477,
-0.11681070178747177,
-0.00865939911454916,
0.10303794592618942,
0.00905151292681694,
0.12129057198762894,
-0.10365233570337296,
-0.022124821320176125,
0.03918497636914253,
-0.050791069865226746,
-0.0029024651739746332,
0.06695733219385147,
0.06399089097976685,
-0.10020945966243744,
0.1256626695394516,
0.06739865988492966,
-0.018320897594094276,
0.05517219379544258,
-0.0676707997918129,
-0.09745685756206512,
-0.014076587744057178,
0.03785772621631622,
0.012908336706459522,
0.1656532734632492,
0.016245678067207336,
0.03933725133538246,
0.038535069674253464,
0.03466617688536644,
0.03554418310523033,
-0.19641143083572388,
-0.020888788625597954,
-0.012898300774395466,
-0.04648463428020477,
-0.03804142773151398,
-0.026141123846173286,
0.01958179846405983,
0.09921249002218246,
0.017096299678087234,
-0.034979451447725296,
0.023158162832260132,
0.0193233173340559,
-0.1046995297074318,
0.17348136007785797,
-0.09915182739496231,
-0.1486581712961197,
-0.1313323974609375,
0.07830274105072021,
-0.05545880272984505,
-0.007949496619403362,
0.03687639161944389,
-0.015152035281062126,
-0.0453927144408226,
-0.1400728076696396,
-0.029458140954375267,
0.02162342146039009,
-0.008924839086830616,
0.048893317580223083,
0.0017908724257722497,
0.12228874862194061,
-0.08802095055580139,
-0.012304361909627914,
-0.04470399022102356,
-0.047891516238451004,
0.02834182046353817,
0.013345387764275074,
0.11326996982097626,
0.0817946195602417,
-0.011669342406094074,
0.03217534348368645,
-0.039079729467630386,
0.29038044810295105,
-0.05336805060505867,
-0.02955014631152153,
0.15135742723941803,
0.009228738956153393,
0.0682048499584198,
0.10603686422109604,
0.016175884753465652,
-0.1223205104470253,
0.03439588099718094,
0.017367340624332428,
-0.036676425486803055,
-0.198516845703125,
-0.07764551043510437,
-0.0195308830589056,
-0.019872933626174927,
0.08433721959590912,
0.052456583827733994,
-0.00028507327078841627,
0.04278327152132988,
-0.03209614008665085,
0.019159790128469467,
-0.02742607519030571,
0.0822627991437912,
0.07431958615779877,
0.03832168132066727,
0.08786936104297638,
-0.053938817232847214,
-0.004049140028655529,
0.09169092029333115,
-0.02217663824558258,
0.26645466685295105,
-0.0045495787635445595,
0.008345401845872402,
0.04235751926898956,
0.20897874236106873,
0.0009666351834312081,
0.06789663434028625,
0.03443210572004318,
0.01069257128983736,
-0.0043488964438438416,
-0.06703312695026398,
-0.05445156246423721,
0.022530782967805862,
-0.06571334600448608,
0.06007004529237747,
-0.1069173663854599,
0.03063698299229145,
0.01535086054354906,
0.26357993483543396,
0.06610068678855896,
-0.35485851764678955,
-0.09439397603273392,
-0.009471219033002853,
-0.00735748466104269,
-0.06542439013719559,
0.012325728312134743,
0.13900889456272125,
-0.12126130610704422,
0.05668105185031891,
-0.04531080648303032,
0.08720412105321884,
-0.02459685504436493,
0.020535439252853394,
0.0371200293302536,
0.10032808035612106,
0.001271453918889165,
0.04843907058238983,
-0.1986406296491623,
0.2417442351579666,
0.013427499681711197,
0.11792004108428955,
-0.027884209528565407,
-0.010119251906871796,
-0.006757138762623072,
0.09640564024448395,
0.0942283496260643,
0.010639110580086708,
0.0032445432152599096,
-0.18192650377750397,
-0.16156989336013794,
0.02160857990384102,
0.07369747012853622,
0.003073604078963399,
0.07316772639751434,
-0.023482350632548332,
0.04339863732457161,
0.014589582569897175,
0.0025576967746019363,
-0.15237846970558167,
-0.06063329055905342,
0.0034109053667634726,
-0.0048368205316364765,
-0.02428540214896202,
-0.09692830592393875,
-0.11099303513765335,
-0.02879929728806019,
0.0657728984951973,
0.013407021760940552,
-0.05746873840689659,
-0.13012711703777313,
0.07526934146881104,
0.09003591537475586,
-0.06475654989480972,
0.019879933446645737,
0.0026483896654099226,
0.1123729944229126,
0.022383617237210274,
-0.08099860697984695,
0.08351502567529678,
-0.05126180127263069,
-0.19215883314609528,
-0.02502623200416565,
0.10734774172306061,
0.03500267490744591,
0.04723562300205231,
-0.003298722906038165,
0.028640538454055786,
0.013964097015559673,
-0.09944264590740204,
0.005516984965652227,
0.09425053000450134,
0.07984822243452072,
0.02062499336898327,
-0.04624656215310097,
0.03922715783119202,
-0.029785238206386566,
-0.02849561721086502,
0.09229706972837448,
0.24658040702342987,
-0.08900938928127289,
0.03288885951042175,
0.053366519510746,
-0.038767315447330475,
-0.21582390367984772,
0.019958777353167534,
0.09417851269245148,
0.018376898020505905,
0.050064969807863235,
-0.1183968335390091,
0.05954756215214729,
0.15892408788204193,
-0.05403652414679527,
0.050264786928892136,
-0.29908135533332825,
-0.12207815796136856,
0.0830855593085289,
0.1365196257829666,
0.07065027207136154,
-0.16367119550704956,
-0.0597539097070694,
-0.01411444041877985,
-0.1238522157073021,
0.08421032130718231,
-0.1613425612449646,
0.09246737509965897,
-0.0009958597365766764,
0.07625739276409149,
0.005375928245484829,
-0.02931540459394455,
0.1449497938156128,
0.0013614873168990016,
0.08290088921785355,
-0.05934152007102966,
0.049459878355264664,
0.02564379759132862,
-0.06663993000984192,
0.08880308270454407,
0.03990791365504265,
0.07932135462760925,
-0.09620781242847443,
-0.013200956396758556,
-0.06387188285589218,
0.05423389747738838,
-0.04076332598924637,
-0.03508545830845833,
-0.04854944348335266,
0.03822455555200577,
0.008167303167283535,
-0.027645587921142578,
0.07832649350166321,
0.032830268144607544,
0.11921580135822296,
0.15205712616443634,
0.05691175162792206,
-0.05262264981865883,
-0.09357669204473495,
-0.00035214333911426365,
-0.049153681844472885,
0.04855548217892647,
-0.09607552736997604,
-0.009106962941586971,
0.10828927904367447,
0.020661260932683945,
0.10049648582935333,
0.0336930975317955,
-0.08571969717741013,
0.010240317322313786,
0.04084885120391846,
-0.14449666440486908,
-0.20386658608913422,
0.012431659735739231,
0.052355486899614334,
-0.11516205966472626,
0.06304017454385757,
0.1038888469338417,
-0.09645470976829529,
-0.013297341763973236,
-0.03678971901535988,
0.022681765258312225,
-0.019759399816393852,
0.15197321772575378,
0.059805069118738174,
0.046138226985931396,
-0.1028892770409584,
0.12109313160181046,
0.06478603929281235,
-0.021570853888988495,
0.03653008118271828,
0.03593587130308151,
-0.12063134461641312,
-0.012548438273370266,
0.07588395476341248,
0.11776328086853027,
-0.028772225603461266,
-0.06924193352460861,
-0.07871749252080917,
-0.11178816854953766,
0.05001060664653778,
0.12645423412322998,
0.06893870234489441,
0.0015056735137477517,
0.019978607073426247,
0.011528964154422283,
-0.11610259115695953,
0.05320597440004349,
0.01769786886870861,
0.0688098818063736,
-0.18195772171020508,
0.09106680750846863,
0.0063912877812981606,
0.02594771236181259,
-0.029997291043400764,
0.03889299929141998,
-0.11723922938108444,
-0.0191646758466959,
-0.08452761918306351,
0.01937584951519966,
-0.054554522037506104,
-0.013839803636074066,
-0.009013984352350235,
-0.08178041875362396,
-0.032065000385046005,
0.056316815316677094,
-0.0630025640130043,
-0.04060327261686325,
0.006972708273679018,
0.037107016891241074,
-0.11947229504585266,
0.009841449558734894,
0.005133400205522776,
-0.0774337574839592,
0.08153394609689713,
0.06608455628156662,
0.01885855384171009,
0.01627347618341446,
-0.032252196222543716,
0.013656459748744965,
0.06096893921494484,
0.019710486754775047,
0.07442990690469742,
-0.05034163221716881,
-0.03456718102097511,
0.01944037899374962,
0.046150919049978256,
0.0328323170542717,
0.12555038928985596,
-0.13096527755260468,
-0.0422598198056221,
-0.03891007974743843,
-0.024068348109722137,
-0.03275914117693901,
0.01583409123122692,
0.10693469643592834,
0.03811558708548546,
0.14705424010753632,
-0.09227505326271057,
0.011675489135086536,
-0.19390679895877838,
-0.026342129334807396,
-0.008445116691291332,
-0.02343827299773693,
-0.09776965528726578,
-0.001417858642525971,
0.07843032479286194,
0.0022480397019535303,
0.12081997841596603,
-0.002844329923391342,
0.06133869290351868,
0.03849227726459503,
-0.020732639357447624,
-0.009725436568260193,
0.004045940935611725,
0.14880575239658356,
0.05093618854880333,
-0.012068621814250946,
0.09988810122013092,
-0.006753464229404926,
0.056662287563085556,
0.03841123729944229,
0.21072903275489807,
0.1607600748538971,
-0.015808096155524254,
0.07575210183858871,
0.08421008288860321,
-0.05476163327693939,
-0.12614452838897705,
-0.008219167590141296,
-0.023622428998351097,
0.07944636046886444,
-0.04763392359018326,
0.1326131671667099,
0.0932563915848732,
-0.1578756421804428,
0.001331303850747645,
-0.05431096628308296,
-0.09179346263408661,
-0.11710375547409058,
-0.023026995360851288,
-0.08119795471429825,
-0.1381818950176239,
-0.0019425186328589916,
-0.12142511457204819,
-0.025286031886935234,
0.1536870151758194,
-0.016481582075357437,
-0.004493317566812038,
0.14421024918556213,
-0.020617511123418808,
0.01307652983814478,
0.033477794378995895,
0.011252172291278839,
-0.001995943021029234,
-0.08012328296899796,
-0.08740993589162827,
0.06475265324115753,
0.011371064931154251,
0.09706109017133713,
-0.001759738544933498,
0.0035989885218441486,
0.04865948110818863,
-0.012128680013120174,
-0.06990493088960648,
0.023032842203974724,
0.027640514075756073,
0.004415571223944426,
0.07569754123687744,
0.04821313917636871,
-0.008770893327891827,
-0.03411828726530075,
0.23094163835048676,
-0.07153191417455673,
-0.08082471787929535,
-0.13458670675754547,
0.2554021179676056,
0.013964828103780746,
-0.011858007870614529,
0.058980077505111694,
-0.1204506978392601,
-0.017144065350294113,
0.18323633074760437,
0.117160364985466,
-0.07549241930246353,
-0.018741382285952568,
-0.015886949375271797,
-0.017741773277521133,
-0.04815554618835449,
0.12844786047935486,
0.11500293016433716,
0.07088641077280045,
-0.06156390532851219,
-0.0026769100222736597,
-0.03733956068754196,
-0.014487073756754398,
-0.11306488513946533,
0.007904362864792347,
0.0002649857779033482,
0.0213821642100811,
-0.08018065243959427,
0.07342178374528885,
-0.03613658994436264,
-0.09046810120344162,
0.021879123523831367,
-0.14504830539226532,
-0.16918377578258514,
0.0026057884097099304,
0.04732603207230568,
-0.015671519562602043,
0.06447700411081314,
-0.04108715429902077,
0.04417701065540314,
0.1043497622013092,
-0.03115168958902359,
-0.06355872750282288,
-0.07580192387104034,
0.06815701723098755,
-0.06789262592792511,
0.20534196496009827,
0.002093882067129016,
0.09215904027223587,
0.08772058039903641,
0.015623820945620537,
-0.1670438051223755,
0.08190971612930298,
0.03453727811574936,
-0.024518758058547974,
-0.011774913407862186,
0.05622026324272156,
-0.025949951261281967,
0.029767876490950584,
0.03371382877230644,
-0.148970827460289,
-0.05237087607383728,
0.016995569691061974,
0.002227529650554061,
-0.051041778177022934,
-0.014065239578485489,
-0.07996778935194016,
0.1466948688030243,
0.13671988248825073,
-0.07293000817298889,
0.0018634251318871975,
-0.044764403253793716,
0.050611890852451324,
0.0254513006657362,
0.019096245989203453,
0.019669506698846817,
-0.23310786485671997,
0.008880572393536568,
-0.0004556199419312179,
-0.00237268372438848,
-0.24544310569763184,
-0.04977554827928543,
-0.00019762312876991928,
-0.04817254841327667,
-0.0905769094824791,
0.09509076178073883,
0.04417238384485245,
0.01928340457379818,
-0.05461929738521576,
-0.08688768744468689,
-0.05334585905075073,
0.15114405751228333,
-0.07798688113689423,
-0.07756636291742325
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | mikeee/phi-2-ft-evol-instruct-chinese-gpt4 | [
"transformers",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-07T08:33:07+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
26,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.08389580249786377,
0.19830818474292755,
-0.0013316317927092314,
0.02313883788883686,
0.11396584659814835,
0.01961737498641014,
0.053626976907253265,
0.14538456499576569,
0.0060051376931369305,
0.10656800121068954,
0.066679947078228,
0.09131570905447006,
0.09678101539611816,
0.20042605698108673,
0.04371999576687813,
-0.17659740149974823,
0.010636410675942898,
-0.06930278241634369,
-0.010073255747556686,
0.11651819199323654,
0.141214057803154,
-0.10151198506355286,
0.07627976685762405,
-0.03319970890879631,
-0.02870541252195835,
-0.0070160143077373505,
-0.07769215852022171,
-0.05755697935819626,
0.07573003321886063,
0.054863471537828445,
0.04207949340343475,
-0.0008347301045432687,
0.08447454124689102,
-0.2674994468688965,
0.013753628358244896,
0.07452993094921112,
0.010659529827535152,
0.05990942195057869,
0.07833302766084671,
-0.04036625102162361,
0.12881849706172943,
-0.06320446729660034,
0.13035163283348083,
0.0906217098236084,
-0.0681561604142189,
-0.24378153681755066,
-0.08239314705133438,
0.06505522131919861,
0.12533815205097198,
0.07694927603006363,
-0.02823091857135296,
0.16422191262245178,
-0.07247646898031235,
0.019290022552013397,
0.09481704235076904,
-0.1151006743311882,
-0.060644298791885376,
0.08318385481834412,
0.14101974666118622,
0.10340547561645508,
-0.1255619376897812,
-0.012289565056562424,
0.04275871813297272,
0.045979104936122894,
0.07389909774065018,
0.011339850723743439,
0.1143413558602333,
0.05629947781562805,
-0.13526225090026855,
-0.05700986459851265,
0.14547574520111084,
0.023872992023825645,
-0.057064127177000046,
-0.2138909548521042,
-0.002902575535699725,
-0.07730814069509506,
-0.011685127392411232,
-0.06846728920936584,
0.0291305985301733,
-0.01194276288151741,
0.060226380825042725,
-0.0496203787624836,
-0.09797755628824234,
-0.046314824372529984,
0.1015089675784111,
0.054820988327264786,
0.011354796588420868,
-0.01489334274083376,
0.03576440364122391,
0.13432876765727997,
0.04213530570268631,
-0.10012737661600113,
-0.07065672427415848,
-0.0701170489192009,
-0.09620913118124008,
-0.03947552293539047,
0.04272124543786049,
0.020167991518974304,
0.042202774435281754,
0.2283228635787964,
0.024096308276057243,
0.05459817871451378,
0.029667891561985016,
0.0026177873369306326,
0.03211980313062668,
0.1073630079627037,
-0.041210614144802094,
-0.188126802444458,
-0.03292805701494217,
0.0931866466999054,
-0.009821015410125256,
-0.028658604249358177,
-0.033444397151470184,
0.035014089196920395,
0.08379437029361725,
0.11821532249450684,
0.08875755965709686,
-0.012828069739043713,
-0.037612639367580414,
-0.03493109717965126,
0.2115669697523117,
-0.14141373336315155,
0.045799970626831055,
-0.022097334265708923,
-0.018195297569036484,
-0.06905751675367355,
0.030103791505098343,
0.01831657998263836,
-0.003142025787383318,
0.06966056674718857,
-0.061253178864717484,
-0.05794486775994301,
-0.11518853157758713,
-0.045523155480623245,
0.04711875319480896,
-0.024105608463287354,
-0.024469668045639992,
-0.07765042781829834,
-0.11219723522663116,
-0.06417357176542282,
0.06612563133239746,
-0.04156653955578804,
-0.03974827378988266,
0.005308232270181179,
-0.07131324708461761,
0.008387917652726173,
0.008993842639029026,
0.12122467905282974,
-0.030063031241297722,
0.05833350867033005,
-0.002476902212947607,
0.05916252359747887,
0.10643328726291656,
0.03227818012237549,
-0.08492200076580048,
0.057466037571430206,
-0.20633617043495178,
0.08371785283088684,
-0.11420095711946487,
0.034276340156793594,
-0.17048145830631256,
-0.024183684960007668,
0.008447963744401932,
0.023597201332449913,
0.023726604878902435,
0.1338067352771759,
-0.2097422182559967,
-0.016196569427847862,
0.14133213460445404,
-0.09649793803691864,
-0.12422871589660645,
0.07990546524524689,
-0.03459475561976433,
0.1747698187828064,
0.038475677371025085,
-0.019652999937534332,
0.09909367561340332,
-0.15559963881969452,
-0.05852397903800011,
-0.026064254343509674,
-0.008927824907004833,
0.08823978155851364,
0.07542291283607483,
-0.05844951793551445,
0.02285866066813469,
0.02562655322253704,
-0.04727208614349365,
-0.0268824752420187,
-0.05256075784564018,
-0.10127434879541397,
-0.023140445351600647,
-0.09642518311738968,
0.026515161618590355,
0.000058677000197349116,
-0.07310442626476288,
-0.028560271486639977,
-0.17347893118858337,
-0.02563360333442688,
0.10103316605091095,
0.004820956848561764,
-0.007559072691947222,
-0.08540112525224686,
0.022149885073304176,
-0.05362366884946823,
-0.006164622958749533,
-0.16996455192565918,
-0.03558015450835228,
0.051895126700401306,
-0.14917676150798798,
0.015460150316357613,
-0.07327745854854584,
0.07047311216592789,
0.02098717913031578,
-0.05859505757689476,
-0.03108096309006214,
0.0007694467785768211,
0.004292082041501999,
-0.06229274719953537,
-0.1903683841228485,
-0.058886781334877014,
-0.041500482708215714,
0.15720732510089874,
-0.24841000139713287,
0.0300158578902483,
0.03247617185115814,
0.13185922801494598,
0.007058668415993452,
-0.06344027817249298,
0.02096918225288391,
-0.04676475748419762,
-0.050621338188648224,
-0.06898977607488632,
-0.009901339188218117,
-0.014539826661348343,
-0.031393732875585556,
0.012980648316442966,
-0.14970256388187408,
-0.060514215379953384,
0.09452559798955917,
0.11224991828203201,
-0.14555825293064117,
0.00204002158716321,
-0.0460561066865921,
-0.07002599537372589,
-0.07487804442644119,
-0.0761631652712822,
0.07739497721195221,
0.044650159776210785,
0.049250341951847076,
-0.06317461282014847,
-0.06234706938266754,
0.023210179060697556,
0.005524294450879097,
-0.019023682922124863,
0.0948529988527298,
0.074309803545475,
-0.09122881293296814,
0.07973480224609375,
0.08461450785398483,
0.04414684325456619,
0.086973637342453,
0.005991141777485609,
-0.11396963149309158,
-0.03062884695827961,
0.037754856050014496,
0.024159027263522148,
0.15351562201976776,
-0.08692087233066559,
0.030462130904197693,
0.052177220582962036,
-0.03854219615459442,
0.03157065063714981,
-0.0923321321606636,
0.025362705811858177,
0.021495236083865166,
-0.006555700208991766,
0.05864228308200836,
-0.018769768998026848,
-0.01403577346354723,
0.06336429715156555,
0.05677810311317444,
0.044270504266023636,
0.02595379762351513,
-0.02093072421848774,
-0.1278371512889862,
0.16537296772003174,
-0.09028079360723495,
-0.2540280222892761,
-0.17074446380138397,
0.015454737469553947,
0.03706491366028786,
-0.021728800609707832,
0.039588842540979385,
-0.06286025792360306,
-0.10237989574670792,
-0.09417891502380371,
0.0029635571409016848,
0.023925531655550003,
-0.058347854763269424,
-0.0817074254155159,
0.060779985040426254,
0.04047083482146263,
-0.13689260184764862,
0.0349188968539238,
0.06170675903558731,
-0.03042641654610634,
0.0018567070364952087,
0.07321398705244064,
0.12743599712848663,
0.14838241040706635,
-0.006730219814926386,
-0.012446845881640911,
0.035035960376262665,
0.229813352227211,
-0.1490442156791687,
0.10630457103252411,
0.14053207635879517,
-0.021705523133277893,
0.06635113060474396,
0.1461038440465927,
0.023231739178299904,
-0.07546708732843399,
0.04147516191005707,
0.04027445614337921,
-0.04228919371962547,
-0.2589097023010254,
-0.05694316700100899,
-0.00946022942662239,
-0.07043391466140747,
0.09718906134366989,
0.09238530695438385,
0.11972260475158691,
0.0337289460003376,
-0.05568677559494972,
-0.025771914049983025,
-0.003401360474526882,
0.114128477871418,
-0.027640055865049362,
-0.004564122296869755,
0.07965842634439468,
-0.05878787487745285,
0.011684526689350605,
0.09941446036100388,
0.019347423687577248,
0.17601320147514343,
0.02533329278230667,
0.10681075602769852,
0.06725578010082245,
0.09347675740718842,
-0.0015635732561349869,
0.034774236381053925,
0.05337131395936012,
0.022044572979211807,
0.010453542694449425,
-0.09408048540353775,
-0.012431944720447063,
0.13713060319423676,
0.019816776737570763,
0.009031654335558414,
0.008926562033593655,
-0.01010479498654604,
0.03131420537829399,
0.20501568913459778,
0.0009575071162544191,
-0.22537250816822052,
-0.09500737488269806,
0.059459153562784195,
-0.06931101530790329,
-0.143676295876503,
-0.02094252221286297,
0.030270220711827278,
-0.17292405664920807,
0.016790566965937614,
-0.0316389761865139,
0.09112390875816345,
-0.07145322859287262,
-0.028050832450389862,
0.06891903281211853,
0.07569212466478348,
-0.012108199298381805,
0.07973295450210571,
-0.19069278240203857,
0.12254468351602554,
0.03037673607468605,
0.08605273067951202,
-0.11708726733922958,
0.07849059253931046,
-0.0019813794642686844,
-0.014807495288550854,
0.17999744415283203,
-0.014062200672924519,
-0.0586031936109066,
-0.08878950774669647,
-0.08704045414924622,
-0.011727320961654186,
0.10361312329769135,
-0.09322915226221085,
0.09586969763040543,
-0.02775636687874794,
-0.03705112263560295,
0.012418309226632118,
-0.10469507426023483,
-0.1636953055858612,
-0.18679304420948029,
0.06244563311338425,
-0.07802703976631165,
0.012347841635346413,
-0.11227322369813919,
-0.06334327906370163,
-0.01575082167983055,
0.23160123825073242,
-0.16648635268211365,
-0.07049825042486191,
-0.1498587429523468,
-0.03997112438082695,
0.17463743686676025,
-0.042160745710134506,
0.06849376112222672,
-0.021383514627814293,
0.1873992383480072,
-0.008081548847258091,
-0.013158116489648819,
0.06569221615791321,
-0.09637628495693207,
-0.16879262030124664,
-0.05748843029141426,
0.14160962402820587,
0.10863390564918518,
0.05731578543782234,
-0.0038195757661014795,
0.013171887956559658,
-0.03383830562233925,
-0.09896382689476013,
0.013824623078107834,
0.13817466795444489,
0.0034514935687184334,
0.00682973163202405,
-0.03995988517999649,
-0.07027145475149155,
-0.05825701728463173,
-0.07912654429674149,
0.057147104293107986,
0.187900573015213,
-0.09512355923652649,
0.1602867990732193,
0.12431421875953674,
-0.06468851119279861,
-0.2306901067495346,
0.03996593505144119,
0.04701630026102066,
0.007666614837944508,
0.022401191294193268,
-0.19138796627521515,
0.09788824617862701,
0.0009011493530124426,
-0.06807263940572739,
0.14616990089416504,
-0.16564498841762543,
-0.1461436152458191,
0.08002161979675293,
0.025075770914554596,
-0.22560662031173706,
-0.14821304380893707,
-0.1037549376487732,
-0.03735695406794548,
-0.13707835972309113,
0.048581719398498535,
0.02614329755306244,
0.019834673032164574,
0.025222565978765488,
0.005338077899068594,
0.029657263308763504,
-0.07272187620401382,
0.1870686560869217,
-0.020297454670071602,
0.0072362530045211315,
-0.050640691071748734,
-0.04617878794670105,
0.09227550774812698,
-0.06150037795305252,
0.11741586774587631,
0.018679620698094368,
0.018796883523464203,
-0.1431548148393631,
-0.049209367483854294,
-0.060803934931755066,
0.04456847906112671,
-0.07284719496965408,
-0.09393193572759628,
-0.04137463867664337,
0.08888561278581619,
0.07211937010288239,
-0.032792408019304276,
-0.0027768779546022415,
-0.07569456845521927,
0.09405932575464249,
0.184477761387825,
0.17357055842876434,
0.009977072477340698,
-0.07020942866802216,
0.024555526673793793,
-0.042279548943042755,
0.03349342197179794,
-0.24652716517448425,
0.03456863760948181,
0.066053606569767,
0.03803660348057747,
0.08509242534637451,
-0.016836483031511307,
-0.1781480610370636,
-0.04086102172732353,
0.08498652279376984,
-0.06206206604838371,
-0.19876568019390106,
-0.02703288197517395,
0.08424776047468185,
-0.20383712649345398,
-0.032998621463775635,
0.041543323546648026,
-0.03834589570760727,
-0.02396267279982567,
-0.002415500348433852,
0.06396626681089401,
-0.008327016606926918,
0.12156640738248825,
0.06747189164161682,
0.10266115516424179,
-0.09284433722496033,
0.08920657634735107,
0.10416955500841141,
-0.09140542894601822,
0.03545991703867912,
0.10264154523611069,
-0.05670900270342827,
-0.04460543021559715,
0.033935222774744034,
0.05925208330154419,
-0.028357384726405144,
-0.06409841030836105,
-0.000502707262057811,
-0.0359574519097805,
0.04993389546871185,
0.08058220148086548,
0.036113787442445755,
-0.01202210783958435,
0.06544706225395203,
0.028145326301455498,
-0.11693570017814636,
0.10949387401342392,
0.04405685141682625,
0.04509059712290764,
-0.07182393968105316,
-0.012280966155230999,
0.015999672934412956,
0.032540347427129745,
-0.019734015688300133,
-0.014576527290046215,
-0.03146412968635559,
-0.007561005651950836,
-0.1553635597229004,
-0.02064543403685093,
-0.06516171246767044,
0.006067827809602022,
0.022207623347640038,
-0.03830232471227646,
-0.012014663778245449,
0.01381110493093729,
-0.07979435473680496,
-0.07571027427911758,
-0.01700955256819725,
0.08539021760225296,
-0.1381402313709259,
0.006627439055591822,
0.07182712107896805,
-0.10980239510536194,
0.07347989827394485,
-0.0048679932951927185,
0.017079560086131096,
0.010923396795988083,
-0.11654401570558548,
0.04386281594634056,
-0.005810429807752371,
0.01551580335944891,
0.022556742653250694,
-0.171111062169075,
0.011553828604519367,
-0.038553636521101,
-0.03114982508122921,
0.011926400475203991,
-0.025060230866074562,
-0.11875922232866287,
0.08676479011774063,
-0.028097305446863174,
-0.037512701004743576,
-0.03292486071586609,
0.06296087801456451,
0.08736220002174377,
-0.011740099638700485,
0.09667140990495682,
-0.025766119360923767,
0.04818311333656311,
-0.1756584197282791,
-0.01910574547946453,
-0.050167568027973175,
0.02537350542843342,
-0.01759655587375164,
-0.0070639788173139095,
0.055272240191698074,
-0.004191063344478607,
0.20991376042366028,
-0.03921036794781685,
0.1548677533864975,
0.05199402943253517,
-0.009925156831741333,
0.010884369723498821,
0.05032730847597122,
0.06423956155776978,
0.031145188957452774,
0.00853167474269867,
0.04660189896821976,
-0.004552975296974182,
-0.020357951521873474,
-0.13699717819690704,
0.02791593410074711,
0.16117429733276367,
0.061918217688798904,
0.0392887257039547,
0.03704594820737839,
-0.1422400325536728,
-0.09538721293210983,
0.10306388139724731,
-0.0331864058971405,
0.014331420883536339,
-0.08317886292934418,
0.17621558904647827,
0.12328410148620605,
-0.1574767529964447,
0.0577850341796875,
-0.07234696298837662,
-0.05066767707467079,
-0.1024852767586708,
-0.11832084506750107,
-0.06293155997991562,
-0.06027044355869293,
-0.004747506696730852,
-0.042489297688007355,
0.05734556168317795,
0.026751231402158737,
-0.003270963439717889,
-0.006759525276720524,
0.12665949761867523,
-0.0249644722789526,
-0.004145825747400522,
0.04152364656329155,
0.0326087586581707,
0.019319625571370125,
-0.05872373282909393,
0.017997145652770996,
0.018602589145302773,
0.022180357947945595,
0.06835069507360458,
0.0260987039655447,
-0.059317342936992645,
0.044286735355854034,
0.00319746439345181,
-0.11313364654779434,
0.018146557733416557,
-0.00002245741598017048,
-0.05020225793123245,
0.13557326793670654,
0.04076748713850975,
0.01548024732619524,
-0.029270920902490616,
0.24342355132102966,
-0.07199113070964813,
-0.08681939542293549,
-0.13965600728988647,
0.11511493474245071,
-0.023563209921121597,
0.03755274787545204,
0.016542524099349976,
-0.12659503519535065,
0.011511262506246567,
0.18531471490859985,
0.12824349105358124,
0.012459068559110165,
-0.007656481582671404,
0.05736639350652695,
-0.0007639875984750688,
-0.05985576659440994,
0.05051197111606598,
0.0664999932050705,
0.16097788512706757,
-0.09069112688302994,
0.0652846097946167,
-0.008405503816902637,
-0.0831485390663147,
-0.027498632669448853,
0.11705785244703293,
-0.022675158455967903,
0.02148384228348732,
-0.03778035193681717,
0.11204422265291214,
-0.052532415837049484,
-0.2719486355781555,
0.02952493168413639,
-0.09503202140331268,
-0.13993041217327118,
-0.02591860294342041,
0.041448429226875305,
-0.03349510580301285,
0.01577647216618061,
0.06254769116640091,
-0.045389387756586075,
0.18837277591228485,
0.025987716391682625,
-0.08679025620222092,
-0.07755549252033234,
0.05874146893620491,
-0.08695939928293228,
0.2789687216281891,
0.003863075515255332,
0.04782010242342949,
0.12108923494815826,
-0.03053574077785015,
-0.18664880096912384,
0.014769754372537136,
0.11989909410476685,
-0.09114406257867813,
0.07780203968286514,
0.18139931559562683,
-0.005561648402363062,
0.12649618089199066,
0.04705416411161423,
-0.03877115994691849,
0.03976387158036232,
-0.02721380814909935,
-0.03821522742509842,
-0.12209630757570267,
0.05661242455244064,
-0.0612691193819046,
0.15957388281822205,
0.1158948540687561,
-0.05964287370443344,
0.001120698289014399,
-0.06126941740512848,
0.06300627440214157,
0.014774397015571594,
0.12115653604269028,
0.018452486023306847,
-0.2023056596517563,
0.05087360367178917,
-0.03283824771642685,
0.08166342973709106,
-0.254973828792572,
-0.08186668157577515,
0.07622263580560684,
-0.019022729247808456,
-0.04275642707943916,
0.12311509251594543,
0.06101066991686821,
0.03676839917898178,
-0.03853875398635864,
-0.08537755906581879,
-0.01412904355674982,
0.15376435220241547,
-0.14123432338237762,
-0.029574336484074593
] |
null | null | null |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# CDAgpt-llama-13b-v3
This model is a fine-tuned version of [meta-llama/Llama-2-13b-hf](https://huggingface.co/meta-llama/Llama-2-13b-hf) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- gradient_accumulation_steps: 3
- total_train_batch_size: 12
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: constant
- lr_scheduler_warmup_ratio: 0.03
- num_epochs: 1
- mixed_precision_training: Native AMP
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"tags": ["generated_from_trainer"], "base_model": "meta-llama/Llama-2-13b-hf", "model-index": [{"name": "CDAgpt-llama-13b-v3", "results": []}]} | null | Federic/CDAgpt-llama-13b-v3 | [
"safetensors",
"generated_from_trainer",
"base_model:meta-llama/Llama-2-13b-hf",
"region:us"
] | 2024-02-07T08:37:50+00:00 | [] | [] | TAGS
#safetensors #generated_from_trainer #base_model-meta-llama/Llama-2-13b-hf #region-us
|
# CDAgpt-llama-13b-v3
This model is a fine-tuned version of meta-llama/Llama-2-13b-hf on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- gradient_accumulation_steps: 3
- total_train_batch_size: 12
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: constant
- lr_scheduler_warmup_ratio: 0.03
- num_epochs: 1
- mixed_precision_training: Native AMP
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| [
"# CDAgpt-llama-13b-v3\n\nThis model is a fine-tuned version of meta-llama/Llama-2-13b-hf on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- gradient_accumulation_steps: 3\n- total_train_batch_size: 12\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: constant\n- lr_scheduler_warmup_ratio: 0.03\n- num_epochs: 1\n- mixed_precision_training: Native AMP",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#safetensors #generated_from_trainer #base_model-meta-llama/Llama-2-13b-hf #region-us \n",
"# CDAgpt-llama-13b-v3\n\nThis model is a fine-tuned version of meta-llama/Llama-2-13b-hf on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- gradient_accumulation_steps: 3\n- total_train_batch_size: 12\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: constant\n- lr_scheduler_warmup_ratio: 0.03\n- num_epochs: 1\n- mixed_precision_training: Native AMP",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
36,
43,
6,
12,
8,
3,
141,
33
] | [
"passage: TAGS\n#safetensors #generated_from_trainer #base_model-meta-llama/Llama-2-13b-hf #region-us \n# CDAgpt-llama-13b-v3\n\nThis model is a fine-tuned version of meta-llama/Llama-2-13b-hf on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- gradient_accumulation_steps: 3\n- total_train_batch_size: 12\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: constant\n- lr_scheduler_warmup_ratio: 0.03\n- num_epochs: 1\n- mixed_precision_training: Native AMP### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.1344112604856491,
0.13207533955574036,
-0.003214005148038268,
0.05799411982297897,
0.11716573685407639,
0.027188638225197792,
0.0848921537399292,
0.11359725892543793,
-0.07308313995599747,
0.11785195767879486,
0.10246327519416809,
0.010446201078593731,
0.05598922446370125,
0.17090046405792236,
-0.031695108860731125,
-0.20693884789943695,
0.03565279394388199,
-0.0604541152715683,
-0.05336938798427582,
0.082392118871212,
0.09826060384511948,
-0.09889906644821167,
0.07041513919830322,
0.0005668741068802774,
-0.11673051863908768,
-0.00568764703348279,
-0.048220194876194,
-0.051997728645801544,
0.08595388382673264,
-0.016895629465579987,
0.08012482523918152,
0.028189953416585922,
0.12285102903842926,
-0.20971322059631348,
0.0010990506270900369,
0.06956318020820618,
0.03934435173869133,
0.08914744108915329,
0.061344146728515625,
-0.024917591363191605,
0.08713000267744064,
-0.16408224403858185,
0.08016292750835419,
0.025041161105036736,
-0.059387896209955215,
-0.11734790354967117,
-0.08231113851070404,
0.045282330363988876,
0.09029516577720642,
0.0640900656580925,
0.016522908583283424,
0.13098065555095673,
-0.10144442319869995,
0.04740603640675545,
0.20375943183898926,
-0.27127596735954285,
-0.04888572916388512,
0.06323567032814026,
0.021495269611477852,
0.06192278861999512,
-0.10579446703195572,
-0.012614605017006397,
0.049007564783096313,
0.01031886413693428,
0.10875251889228821,
0.01634978875517845,
-0.050155654549598694,
-0.00019801055896095932,
-0.12380451709032059,
-0.015790574252605438,
0.1495286524295807,
0.047693390399217606,
-0.05412646010518074,
-0.10340997576713562,
-0.053101859986782074,
-0.13268595933914185,
-0.02564200945198536,
0.008148504421114922,
0.02959715574979782,
-0.06395003199577332,
-0.050669994205236435,
-0.013472352176904678,
-0.05130995064973831,
-0.0804053395986557,
0.013050426729023457,
0.14070408046245575,
0.05836959183216095,
0.009655069559812546,
0.011031714268028736,
0.1274172067642212,
-0.03888452798128128,
-0.11153418570756912,
-0.028372518718242645,
-0.014968900009989738,
-0.12212245911359787,
-0.044227633625268936,
-0.043386004865169525,
-0.0008323683287017047,
0.024743564426898956,
0.12880802154541016,
-0.029819436371326447,
0.06054878234863281,
0.04440712183713913,
-0.010482819750905037,
0.006781566422432661,
0.11260809004306793,
-0.0768352746963501,
-0.03438883274793625,
0.025758836418390274,
0.11851738393306732,
0.030564330518245697,
0.00005597388371825218,
-0.0631580725312233,
-0.03963197395205498,
0.10907331109046936,
0.07769154757261276,
-0.05706198513507843,
0.029411720111966133,
-0.03958744928240776,
-0.011537858285009861,
0.027767932042479515,
-0.1287114918231964,
0.011813515797257423,
0.01887037605047226,
-0.08164719492197037,
-0.030300408601760864,
0.01534752082079649,
-0.011263284832239151,
-0.020472168922424316,
0.082016721367836,
-0.09292428195476532,
-0.026675714179873466,
-0.06277989596128464,
-0.04664595052599907,
0.00458933599293232,
-0.0873996913433075,
-0.01531195268034935,
-0.06647777557373047,
-0.16952933371067047,
-0.026861757040023804,
0.01949778012931347,
-0.08422914147377014,
-0.051734477281570435,
-0.04043414443731308,
-0.06837693601846695,
0.030656032264232635,
-0.008272167295217514,
0.12640440464019775,
-0.06360204517841339,
0.07030393183231354,
-0.0004119357618037611,
0.032980792224407196,
0.04732699692249298,
0.030447807163000107,
-0.06876016408205032,
0.06213761121034622,
-0.14286592602729797,
0.07223578542470932,
-0.09146630764007568,
0.022321442142128944,
-0.1228628158569336,
-0.0720478817820549,
-0.00331884459592402,
-0.034789249300956726,
0.08138652890920639,
0.15727455914020538,
-0.15680138766765594,
0.0030292747542262077,
0.1544155478477478,
-0.05455227941274643,
-0.08690602332353592,
0.09108829498291016,
-0.034676603972911835,
-0.015105675905942917,
0.03264966234564781,
0.13757427036762238,
0.09259050339460373,
-0.16016417741775513,
-0.013962469063699245,
0.032855939120054245,
0.06918404996395111,
-0.017642466351389885,
0.07689286023378372,
-0.010505040176212788,
-0.01892497017979622,
0.01878151297569275,
-0.0470198355615139,
-0.00470366794615984,
-0.08407143503427505,
-0.07142254710197449,
-0.06059674546122551,
-0.09669676423072815,
0.005381242837756872,
-0.020572105422616005,
0.009395015425980091,
-0.06657671928405762,
-0.10330750048160553,
0.041015323251485825,
0.16428464651107788,
-0.05075279623270035,
-0.0030744862742722034,
-0.06229829788208008,
0.060682374984025955,
-0.06086909770965576,
-0.03120342828333378,
-0.15260881185531616,
-0.1416456550359726,
0.04495078697800636,
-0.03420542553067207,
0.002664616797119379,
-0.016066960990428925,
0.05387849360704422,
0.08063816279172897,
-0.03226974606513977,
-0.03797314688563347,
-0.05284649506211281,
-0.00903381034731865,
-0.09419377893209457,
-0.18011893332004547,
-0.04297655075788498,
-0.03768562898039818,
0.21209529042243958,
-0.21879027783870697,
-0.0052063013426959515,
-0.024472445249557495,
0.12548828125,
0.017759041860699654,
-0.05629422143101692,
0.03919905051589012,
0.012940246611833572,
-0.013498400337994099,
-0.10644851624965668,
0.01591404341161251,
0.0008756204042583704,
-0.11286456137895584,
-0.007516590878367424,
-0.1484004259109497,
0.04624894633889198,
0.06026804447174072,
0.15930360555648804,
-0.1013282909989357,
-0.10812709480524063,
-0.06560992449522018,
-0.05037442967295647,
-0.07325714081525803,
0.00855676457285881,
0.15089334547519684,
0.029126746580004692,
0.11837809532880783,
-0.08035439252853394,
-0.05458170548081398,
0.011706234887242317,
0.008074034936726093,
-0.017897333949804306,
0.08075357973575592,
0.0728648453950882,
-0.07097107172012329,
0.06035347655415535,
0.07641903311014175,
-0.05134975165128708,
0.1110256239771843,
-0.03955008462071419,
-0.11106892675161362,
-0.027026427909731865,
0.03994978591799736,
-0.003284140955656767,
0.1394931674003601,
-0.03715347871184349,
0.029468536376953125,
0.03819173946976662,
0.019046887755393982,
0.02848953753709793,
-0.14176897704601288,
-0.014483192004263401,
0.02024393156170845,
-0.03952491283416748,
-0.02951831929385662,
-0.00867425836622715,
0.03459910303354263,
0.08869816362857819,
0.021844293922185898,
-0.049294374883174896,
0.0025833139661699533,
-0.010684562847018242,
-0.06427142024040222,
0.20048993825912476,
-0.10919646918773651,
-0.09903982281684875,
-0.09556838870048523,
0.061979956924915314,
-0.056296512484550476,
-0.02835165150463581,
-0.014515519142150879,
-0.05792131647467613,
-0.0661962702870369,
-0.11748063564300537,
-0.04327740520238876,
-0.029152125120162964,
0.015087434090673923,
0.005594911053776741,
0.017272867262363434,
0.08477427065372467,
-0.11390585452318192,
0.008757781237363815,
-0.01959518902003765,
-0.07500488311052322,
0.02603643387556076,
0.0616929717361927,
0.085564523935318,
0.11312548071146011,
-0.001345465425401926,
0.020891563966870308,
-0.016785018146038055,
0.20094488561153412,
-0.049137964844703674,
-0.011684799566864967,
0.0847991555929184,
0.03575705364346504,
0.05441473424434662,
0.12576745450496674,
0.016418583691120148,
-0.09066423773765564,
0.03846156597137451,
0.056162748485803604,
-0.019523926079273224,
-0.24001142382621765,
-0.05592293292284012,
-0.021554190665483475,
-0.07105548679828644,
0.0923047736287117,
0.05071956664323807,
-0.0636724978685379,
0.05733513832092285,
-0.03260986879467964,
-0.0003547716187313199,
-0.037751808762550354,
0.07717598229646683,
0.02059817686676979,
0.047002509236335754,
0.08501759171485901,
-0.010739543475210667,
-0.014905435964465141,
0.05555976182222366,
0.038686592131853104,
0.2355390191078186,
-0.07836508750915527,
0.10943420976400375,
0.00971656572073698,
0.15856365859508514,
-0.040124889463186264,
0.04670464247465134,
0.008646172471344471,
-0.011893000453710556,
0.0017139664851129055,
-0.0772293359041214,
-0.04290924221277237,
0.04394084960222244,
0.023138217628002167,
0.04505935683846474,
-0.11604722589254379,
0.08668986707925797,
0.010348538868129253,
0.26203057169914246,
0.06960514187812805,
-0.31464684009552,
-0.04565928876399994,
0.01347874291241169,
0.004092775750905275,
-0.08378257602453232,
0.014679302461445332,
0.13048018515110016,
-0.1129257008433342,
0.04412895441055298,
-0.04918648302555084,
0.0618310309946537,
-0.051917292177677155,
-0.0186829324811697,
0.00561955152079463,
0.170286625623703,
-0.022225331515073776,
0.090374656021595,
-0.14358766376972198,
0.20223468542099,
0.03258174657821655,
0.09290221333503723,
-0.05674472823739052,
0.01831825263798237,
0.0023043702822178602,
0.06527269631624222,
0.10779384523630142,
0.0060197170823812485,
-0.029319128021597862,
-0.14837951958179474,
-0.1630992293357849,
0.0425153523683548,
0.11903116106987,
-0.024188729003071785,
0.09149474650621414,
-0.03823791444301605,
-0.005215736106038094,
0.030318759381771088,
-0.08056370913982391,
-0.15230847895145416,
-0.11769098788499832,
0.02701922506093979,
0.024027258157730103,
-0.027870170772075653,
-0.09160088747739792,
-0.09117830544710159,
-0.023333096876740456,
0.1878160834312439,
0.014429373666644096,
-0.05000196024775505,
-0.15117168426513672,
0.020868556573987007,
0.1550026535987854,
-0.054889678955078125,
0.02430855669081211,
-0.005103581119328737,
0.14427481591701508,
0.025416214019060135,
-0.06039400398731232,
0.07562946528196335,
-0.07050296664237976,
-0.1589522361755371,
-0.06564568728208542,
0.14687974750995636,
0.04636062681674957,
0.04251707345247269,
-0.008265778422355652,
0.001234513591043651,
0.01098259910941124,
-0.08336953818798065,
0.023575494065880775,
0.09590084105730057,
0.07427934557199478,
0.08615238219499588,
-0.08310864120721817,
0.05991841107606888,
-0.005495638120919466,
-0.0039516924880445,
0.11799771338701248,
0.2773783802986145,
-0.07212784886360168,
0.10094615817070007,
0.0647297203540802,
-0.03355313837528229,
-0.17912213504314423,
0.03982139378786087,
0.10829314589500427,
0.03785267472267151,
0.026219062507152557,
-0.1713894158601761,
0.10354036837816238,
0.11639656126499176,
-0.025902124121785164,
0.05679599568247795,
-0.3148724138736725,
-0.11872890591621399,
0.0935468077659607,
0.09642033278942108,
0.03559967875480652,
-0.10687467455863953,
-0.0401742197573185,
-0.05346907675266266,
-0.08991776406764984,
0.08106434345245361,
-0.10380938649177551,
0.1182045117020607,
-0.03243715688586235,
0.0869031772017479,
0.04882930964231491,
-0.05846049264073372,
0.1578367054462433,
0.029125699773430824,
0.07313060760498047,
-0.03690383955836296,
0.05322955548763275,
0.037423886358737946,
-0.09826897829771042,
0.07927517592906952,
-0.09127268940210342,
0.08623042702674866,
-0.13967905938625336,
-0.012028161436319351,
-0.05472998321056366,
0.08056281507015228,
-0.04396124556660652,
-0.057579390704631805,
-0.04048057645559311,
0.05411694943904877,
0.0720071867108345,
-0.03193425014615059,
0.054264239966869354,
0.022830791771411896,
0.03970540687441826,
0.11835500597953796,
0.08002502471208572,
-0.014650979079306126,
-0.06475748121738434,
0.008503074757754803,
-0.005568403284996748,
0.049681730568408966,
-0.1096397191286087,
0.023056261241436005,
0.10650218278169632,
0.03778880462050438,
0.10100340843200684,
0.01979786530137062,
-0.07554983347654343,
-0.025071464478969574,
0.04784778878092766,
-0.11348852515220642,
-0.11077282577753067,
0.0024852873757481575,
0.028698530048131943,
-0.14755919575691223,
0.02917492389678955,
0.10345731675624847,
-0.05232902243733406,
-0.011277977377176285,
-0.023737184703350067,
0.03539958596229553,
0.005175984464585781,
0.18743711709976196,
0.045971840620040894,
0.0733543336391449,
-0.07856966555118561,
0.13754552602767944,
0.05917738378047943,
-0.05466695502400398,
0.055465422570705414,
0.03457920253276825,
-0.09985014796257019,
-0.007941328920423985,
0.0991227924823761,
0.09694766998291016,
-0.023897569626569748,
-0.028037406504154205,
-0.0703587681055069,
-0.10312359035015106,
0.0408344641327858,
0.09294220060110092,
0.029234370216727257,
-0.005493713542819023,
0.0019338561687618494,
-0.006674083415418863,
-0.1383076310157776,
0.0957787036895752,
0.04933526739478111,
0.08660680800676346,
-0.13102871179580688,
0.09986051917076111,
-0.023543277755379677,
0.043324701488018036,
-0.011157647706568241,
0.014568477869033813,
-0.091846764087677,
-0.0238084327429533,
-0.1436024159193039,
0.006721004843711853,
-0.03123525157570839,
0.0039039875846356153,
-0.008859935216605663,
-0.0188160240650177,
-0.029491089284420013,
0.03378625214099884,
-0.06493216007947922,
-0.05496172234416008,
-0.014750095084309578,
0.05900287628173828,
-0.1318860650062561,
-0.014551792293787003,
0.020692864432930946,
-0.11572624742984772,
0.10307557880878448,
0.07319758087396622,
0.03994082659482956,
0.002343523781746626,
-0.10573981702327728,
-0.014711163006722927,
0.029882052913308144,
0.020280810073018074,
0.0675630047917366,
-0.11754090338945389,
-0.013541188091039658,
-0.035105254501104355,
0.007702062372118235,
0.0007657001260668039,
0.06175024062395096,
-0.13309015333652496,
-0.04884853586554527,
-0.04987272247672081,
-0.024934330955147743,
-0.06134440377354622,
0.028510527685284615,
0.0700615867972374,
0.020900366827845573,
0.13366290926933289,
-0.06668565422296524,
0.017396846786141396,
-0.16902118921279907,
-0.04060514643788338,
-0.016922971233725548,
-0.006847034674137831,
-0.060929689556360245,
0.008382539264857769,
0.09172140806913376,
-0.05466080084443092,
0.13013893365859985,
-0.04269448667764664,
0.03166794031858444,
0.020048700273036957,
-0.03599288687109947,
0.02300652489066124,
0.012487226165831089,
0.16736248135566711,
0.07069598138332367,
-0.0020055079367011786,
0.09473196417093277,
-0.02871895208954811,
0.08728350698947906,
0.06271111965179443,
0.16845716536045074,
0.12376856803894043,
0.008952522650361061,
0.07860048860311508,
0.06664004921913147,
-0.14725548028945923,
-0.08561286330223083,
0.11844853311777115,
-0.046761754900217056,
0.10131364315748215,
-0.028564943000674248,
0.1615964025259018,
0.1151411160826683,
-0.16410930454730988,
0.007780204061418772,
-0.0467967726290226,
-0.12196151167154312,
-0.1111198291182518,
-0.05096305161714554,
-0.08613976836204529,
-0.10132469236850739,
0.030753890052437782,
-0.1097678393125534,
0.014397461898624897,
0.0792812630534172,
0.022860057651996613,
0.03001762181520462,
0.1597343236207962,
-0.008807087317109108,
0.007358151022344828,
0.05043831095099449,
0.0446183867752552,
-0.00045520937419496477,
-0.003212803741917014,
-0.08815448731184006,
0.060000453144311905,
-0.01709439419209957,
0.07624500244855881,
-0.04540734738111496,
0.04354431480169296,
0.06969066709280014,
0.0198746919631958,
-0.07205522060394287,
0.015926359221339226,
-0.007297987584024668,
0.029821906238794327,
0.03965596854686737,
0.06346169114112854,
-0.0034531662240624428,
-0.058458149433135986,
0.26283499598503113,
-0.0699225664138794,
-0.04273257032036781,
-0.11970220506191254,
0.20186203718185425,
0.03293159604072571,
-0.0005230677779763937,
0.059199146926403046,
-0.11774978041648865,
-0.017999479547142982,
0.11607857793569565,
0.14399969577789307,
-0.0929429680109024,
-0.023142362013459206,
-0.04399261623620987,
-0.02514015883207321,
-0.06566151976585388,
0.12128981202840805,
0.0829528272151947,
0.008750745095312595,
-0.053517475724220276,
0.0054081338457763195,
-0.015454686246812344,
-0.029153112322092056,
-0.09554864466190338,
0.0747656300663948,
-0.0076081776060163975,
0.006995225325226784,
-0.0353420227766037,
0.044991668313741684,
-0.013113160617649555,
-0.1651107668876648,
0.020991690456867218,
-0.10032859444618225,
-0.19743365049362183,
-0.045164160430431366,
0.02297753468155861,
-0.007099563721567392,
0.053131766617298126,
-0.04065656289458275,
-0.009658006019890308,
0.12943115830421448,
-0.019489191472530365,
-0.04014059528708458,
-0.10711842775344849,
0.05908474698662758,
-0.018301591277122498,
0.2415282130241394,
0.0194078516215086,
0.05839298665523529,
0.0982171893119812,
0.012807738035917282,
-0.16858267784118652,
0.048052914440631866,
0.0907832533121109,
-0.08821617066860199,
0.0452490895986557,
0.16624298691749573,
-0.05653293430805206,
0.10976902395486832,
0.045034609735012054,
-0.06205594539642334,
-0.027161192148923874,
-0.03559550270438194,
0.00039118321728892624,
-0.08471904695034027,
-0.004745182581245899,
-0.03905383124947548,
0.16200950741767883,
0.18156538903713226,
-0.049848273396492004,
-0.0011215239064767957,
-0.06946511566638947,
0.03455013036727905,
0.045968808233737946,
0.06366852670907974,
0.0014889114536345005,
-0.19289030134677887,
0.04267192631959915,
0.058443110436201096,
0.045831937342882156,
-0.24375173449516296,
-0.07046326994895935,
0.02025958150625229,
-0.053350403904914856,
-0.04854859411716461,
0.12906469404697418,
0.04170047491788864,
0.0358436219394207,
-0.044925104826688766,
-0.06018690764904022,
-0.04977047070860863,
0.1264553666114807,
-0.16381092369556427,
-0.07180782407522202
] |
null | null | null |
# PPO Agent Playing LunarLander-v2
This is a trained model of a PPO agent playing LunarLander-v2.
# Hyperparameters
```python
{'exp_name': 'ppo'
'seed': 1
'torch_deterministic': True
'cuda': True
'track': False
'wandb_project_name': 'cleanRL'
'wandb_entity': None
'capture_video': False
'env_id': 'LunarLander-v2'
'total_timesteps': 500000
'learning_rate': 0.00025
'num_envs': 4
'num_steps': 128
'anneal_lr': True
'gae': True
'gamma': 0.99
'gae_lambda': 0.95
'num_minibatches': 4
'update_epochs': 4
'norm_adv': True
'clip_coef': 0.2
'clip_vloss': True
'ent_coef': 0.01
'vf_coef': 0.5
'max_grad_norm': 0.5
'target_kl': None
'repo_id': 'Wembo/ppo-self-LunarLander-v2'
'batch_size': 512
'minibatch_size': 128}
```
| {"tags": ["LunarLander-v2", "ppo", "deep-reinforcement-learning", "reinforcement-learning", "custom-implementation", "deep-rl-course"], "model-index": [{"name": "PPO", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "LunarLander-v2", "type": "LunarLander-v2"}, "metrics": [{"type": "mean_reward", "value": "20.77 +/- 54.16", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | Wembo/ppo-self-LunarLander-v2 | [
"tensorboard",
"LunarLander-v2",
"ppo",
"deep-reinforcement-learning",
"reinforcement-learning",
"custom-implementation",
"deep-rl-course",
"model-index",
"region:us"
] | 2024-02-07T08:45:25+00:00 | [] | [] | TAGS
#tensorboard #LunarLander-v2 #ppo #deep-reinforcement-learning #reinforcement-learning #custom-implementation #deep-rl-course #model-index #region-us
|
# PPO Agent Playing LunarLander-v2
This is a trained model of a PPO agent playing LunarLander-v2.
# Hyperparameters
| [
"# PPO Agent Playing LunarLander-v2\n\n This is a trained model of a PPO agent playing LunarLander-v2.\n \n # Hyperparameters"
] | [
"TAGS\n#tensorboard #LunarLander-v2 #ppo #deep-reinforcement-learning #reinforcement-learning #custom-implementation #deep-rl-course #model-index #region-us \n",
"# PPO Agent Playing LunarLander-v2\n\n This is a trained model of a PPO agent playing LunarLander-v2.\n \n # Hyperparameters"
] | [
51,
37
] | [
"passage: TAGS\n#tensorboard #LunarLander-v2 #ppo #deep-reinforcement-learning #reinforcement-learning #custom-implementation #deep-rl-course #model-index #region-us \n# PPO Agent Playing LunarLander-v2\n\n This is a trained model of a PPO agent playing LunarLander-v2.\n \n # Hyperparameters"
] | [
0.07948226481676102,
-0.021824665367603302,
-0.005334289278835058,
0.07425090670585632,
0.11451162397861481,
-0.051334477961063385,
0.11827225238084793,
0.05111894756555557,
0.0632978081703186,
0.08233953267335892,
0.09910695254802704,
0.11526558548212051,
0.02103434130549431,
0.12346389144659042,
0.10133372992277145,
-0.26653239130973816,
0.0048308540135622025,
-0.042133692651987076,
0.020121442154049873,
0.07062754780054092,
-0.028985055163502693,
-0.12164036184549332,
0.02042403817176819,
-0.008055811747908592,
0.04164125770330429,
0.03685355558991432,
-0.020250989124178886,
-0.07061084359884262,
0.1035412922501564,
-0.04342407360672951,
0.07646117359399796,
0.04053044691681862,
0.12915800511837006,
-0.11266650259494781,
0.03731851652264595,
0.047094929963350296,
-0.058420803397893906,
0.040810972452163696,
0.023221731185913086,
0.07433853298425674,
0.15582501888275146,
0.0008022422553040087,
0.10807766020298004,
-0.019928930327296257,
-0.15859591960906982,
-0.0564296655356884,
0.04013175517320633,
0.10688508301973343,
0.041339244693517685,
0.05763867497444153,
0.01518392562866211,
0.24210692942142487,
-0.07300914824008942,
0.0014766358071938157,
0.1963091939687729,
-0.2750851511955261,
-0.056198850274086,
0.2650637924671173,
0.08425293117761612,
0.09438422322273254,
-0.09869689494371414,
-0.0236953292042017,
0.007850034162402153,
0.013983802869915962,
-0.038732558488845825,
-0.07621388882398605,
0.1343805193901062,
0.06358266621828079,
-0.07906194031238556,
-0.05448254942893982,
0.09211132675409317,
0.015635671094059944,
0.03398676961660385,
0.0008897133520804346,
-0.015260354615747929,
0.03964465111494064,
-0.008004734292626381,
-0.08323223143815994,
0.067534439265728,
0.017411211505532265,
-0.059903185814619064,
-0.11101946979761124,
-0.11182308942079544,
-0.028280947357416153,
-0.08438915759325027,
0.16840966045856476,
-0.023494480177760124,
0.07285201549530029,
-0.06215810775756836,
0.06860414892435074,
-0.037912189960479736,
0.004227026831358671,
0.006380763836205006,
-0.049948662519454956,
-0.04539962485432625,
-0.025878654792904854,
0.006328459829092026,
0.011017742566764355,
0.11213880032300949,
-0.002449487103149295,
0.0508684441447258,
0.04856472462415695,
0.014653711579740047,
0.0942535474896431,
0.04126615449786186,
0.18958540260791779,
-0.006363034248352051,
0.0650586485862732,
0.062062907963991165,
0.017491057515144348,
0.022076671943068504,
-0.05142693966627121,
-0.1658715307712555,
0.0807771384716034,
-0.08260773122310638,
-0.028765955939888954,
0.09323479980230331,
-0.044928085058927536,
-0.1112084910273552,
-0.01773354969918728,
-0.07590804249048233,
-0.025731517001986504,
-0.01252016518265009,
0.01790926419198513,
-0.035574477165937424,
0.005672375671565533,
0.03449513763189316,
0.08204318583011627,
0.033907562494277954,
-0.08674118667840958,
0.00984077900648117,
0.012360874563455582,
-0.122767873108387,
-0.004771664272993803,
0.010288639925420284,
0.04804306477308273,
0.04491464048624039,
-0.1116413027048111,
-0.2020648866891861,
-0.08828215301036835,
0.053431469947099686,
-0.07537820190191269,
-0.15614600479602814,
-0.11512033641338348,
0.02302604168653488,
-0.10217837989330292,
-0.046169016510248184,
-0.0017400066135451198,
-0.019300667569041252,
0.05366985872387886,
-0.06531468033790588,
0.1828034669160843,
0.0271916463971138,
-0.00020129751646891236,
-0.14947181940078735,
0.019320663064718246,
-0.2362208217382431,
0.07685942947864532,
-0.04987453296780586,
0.07074880599975586,
-0.04584719240665436,
-0.09154892712831497,
-0.01864667609333992,
0.054014526307582855,
0.013841784559190273,
0.10950348526239395,
-0.1638582944869995,
-0.05129624530673027,
0.024843567982316017,
-0.08068934828042984,
-0.0030390452593564987,
-0.04837793856859207,
-0.04604795575141907,
0.1606992781162262,
0.018704978749155998,
0.14688511192798615,
-0.12919624149799347,
-0.09930720180273056,
0.19129104912281036,
0.03531093895435333,
-0.16984215378761292,
-0.036521974951028824,
0.09952033311128616,
0.019277004525065422,
-0.01849931664764881,
-0.05688142776489258,
-0.07599073648452759,
0.015944182872772217,
-0.08702079951763153,
-0.04182637855410576,
0.04013517126441002,
-0.042824242264032364,
0.14606650173664093,
0.10223949700593948,
0.07952884584665298,
-0.07538176327943802,
-0.007020880468189716,
0.08674140274524689,
0.06271850317716599,
0.045035574585199356,
0.03672485426068306,
-0.05614851415157318,
0.03206208720803261,
-0.025039123371243477,
-0.01738123595714569,
-0.13521039485931396,
0.0019960827194154263,
-0.06055765971541405,
0.1118607297539711,
0.13101612031459808,
0.28467631340026855,
0.10075046867132187,
0.02464960888028145,
0.07675616443157196,
-0.07042508572340012,
-0.10758408159017563,
0.002032244112342596,
0.0235405582934618,
-0.1785016655921936,
0.026378504931926727,
-0.07599464803934097,
-0.14044412970542908,
-0.1351996809244156,
-0.025685761123895645,
-0.17195537686347961,
0.02159930020570755,
0.054728612303733826,
-0.018639836460351944,
0.0013907389948144555,
0.12220112234354019,
0.013543038628995419,
-0.053733617067337036,
0.10188740491867065,
0.009542218409478664,
-0.05206648260354996,
-0.045367226004600525,
0.1050298660993576,
0.13431710004806519,
0.1365344226360321,
-0.2098493129014969,
0.008600602857768536,
0.1119711846113205,
-0.04708562791347504,
0.03519878163933754,
0.026510966941714287,
0.21071651577949524,
0.2740876078605652,
0.0374440960586071,
0.008118349127471447,
-0.05789022892713547,
0.0453064851462841,
-0.05260699614882469,
-0.11800429224967957,
-0.05410657823085785,
0.17159637808799744,
0.07862472534179688,
-0.006237224210053682,
0.09871696680784225,
0.07909595966339111,
0.037818074226379395,
0.16045765578746796,
0.03334520757198334,
-0.09544764459133148,
-0.03232238441705704,
-0.026171676814556122,
-0.0047440179623663425,
0.06791821867227554,
-0.0798373743891716,
-0.032012078911066055,
0.021649274975061417,
-0.13788609206676483,
0.018513672053813934,
-0.18612799048423767,
-0.1437452882528305,
0.03805195167660713,
0.043561313301324844,
-0.008401780389249325,
0.04065251722931862,
-0.0160639937967062,
0.05676067993044853,
0.03282754495739937,
-0.08861549198627472,
0.04405612871050835,
-0.005384152289479971,
0.009959283284842968,
0.03441033884882927,
-0.01767686940729618,
-0.21204280853271484,
-0.15340813994407654,
0.013550614938139915,
-0.05142427980899811,
0.05592547729611397,
-0.008550947532057762,
-0.19242143630981445,
0.025911282747983932,
-0.014332908205688,
0.02364996261894703,
-0.03164665028452873,
-0.03833974152803421,
0.1345074623823166,
0.14185978472232819,
-0.026165392249822617,
0.00023905932903289795,
-0.03341824188828468,
-0.14318081736564636,
-0.180479034781456,
0.06557876616716385,
0.0740460753440857,
0.006866236217319965,
0.1220167726278305,
0.004434254486113787,
0.026604121550917625,
-0.00636066310107708,
0.007762894034385681,
-0.07827747613191605,
-0.10268643498420715,
0.2943233549594879,
0.02490289881825447,
-0.022609207779169083,
-0.023361563682556152,
0.022680940106511116,
-0.005913543980568647,
0.020695405080914497,
-0.06731052696704865,
-0.11051533371210098,
-0.10214895755052567,
-0.018064133822917938,
-0.05326148122549057,
0.08696132898330688,
0.05207669362425804,
-0.0023201601579785347,
-0.058658841997385025,
0.0491698756814003,
0.15816207230091095,
0.0022554483730345964,
-0.07889559864997864,
0.00756099633872509,
0.06827649474143982,
-0.10357149690389633,
0.019141824916005135,
-0.011750275269150734,
-0.06115471199154854,
0.01578802429139614,
0.021844392642378807,
0.02698187716305256,
0.10298074781894684,
-0.21004606783390045,
0.04396829754114151,
0.06455216556787491,
0.025463011115789413,
0.08768844604492188,
0.05016043782234192,
-0.11047832667827606,
-0.016628960147500038,
-0.0343489907681942,
-0.16258354485034943,
0.1297316700220108,
0.14130131900310516,
0.06893892586231232,
0.039022352546453476,
0.04288983345031738,
-0.07514789700508118,
0.058336563408374786,
-0.03656633570790291,
-0.1470387876033783,
-0.018523573875427246,
0.03902188688516617,
0.03257647529244423,
0.038807060569524765,
0.10827972739934921,
0.10223158448934555,
-0.14332416653633118,
-0.03201044723391533,
0.06512229144573212,
-0.008886558935046196,
-0.04119880497455597,
0.004403908737003803,
-0.09832779318094254,
0.07498125731945038,
-0.0024919756688177586,
0.04813602566719055,
-0.20199769735336304,
0.16434083878993988,
-0.09330786764621735,
0.034300561994314194,
-0.04896155744791031,
-0.044333528727293015,
0.03555295243859291,
-0.09057865291833878,
0.20472288131713867,
0.0057462104596197605,
0.008313721977174282,
-0.12209630757570267,
-0.17661772668361664,
-0.034985676407814026,
-0.09205599129199982,
-0.07460658252239227,
0.02909865602850914,
0.0682184249162674,
0.029013507068157196,
-0.044006895273923874,
0.1327963024377823,
-0.007539169397205114,
0.08532623946666718,
-0.09495806694030762,
-0.09892267733812332,
-0.06850815564393997,
-0.09003753960132599,
-0.13165755569934845,
-0.069197878241539,
0.05082700401544571,
0.12665395438671112,
0.02109835296869278,
-0.02864154241979122,
0.016000375151634216,
-0.01131656114012003,
0.0060316757299005985,
-0.006539386231452227,
0.0482512004673481,
0.015850301831960678,
-0.05547862499952316,
-0.13189296424388885,
0.08252222090959549,
-0.06544385105371475,
-0.06556238979101181,
-0.023766927421092987,
0.09430349618196487,
0.09706855565309525,
0.1314772367477417,
-0.052682001143693924,
0.028886299580335617,
-0.03723334148526192,
-0.04484548792243004,
0.18565788865089417,
0.0040725888684391975,
-0.07140722125768661,
0.04510314390063286,
0.08041586726903915,
0.05989309027791023,
0.0390491709113121,
-0.031676698476076126,
0.20406655967235565,
0.15550298988819122,
-0.018378838896751404,
0.19636642932891846,
-0.017176153138279915,
-0.0269333329051733,
-0.20952188968658447,
0.006836839485913515,
-0.019357649609446526,
0.029477683827280998,
0.1340312361717224,
-0.1391998678445816,
0.02293945848941803,
-0.004865060094743967,
-0.02284914068877697,
-0.07053285837173462,
-0.3114997148513794,
-0.06468415260314941,
0.20102077722549438,
0.17379379272460938,
0.30399972200393677,
-0.10662104934453964,
0.05403600633144379,
0.02176249772310257,
0.035715505480766296,
0.03934846818447113,
-0.07645441591739655,
0.1000572219491005,
-0.11122481524944305,
0.16528162360191345,
0.08111181855201721,
-0.020749825984239578,
-0.02004031278192997,
-0.13701297342777252,
0.018633954226970673,
-0.12466508150100708,
-0.017992790788412094,
0.08779406547546387,
-0.003319771494716406,
-0.09328535199165344,
0.23242005705833435,
-0.06734555959701538,
-0.127778559923172,
-0.028943995013833046,
-0.057271506637334824,
-0.030531147494912148,
0.012628542259335518,
-0.09404513984918594,
0.005903336685150862,
0.1308545619249344,
-0.011834635399281979,
0.11608193069696426,
0.16071371734142303,
-0.035819161683321,
0.07980551570653915,
0.11671095341444016,
0.041628848761320114,
0.06653126329183578,
-0.16247588396072388,
-0.008802353404462337,
-0.0202709399163723,
0.029673689976334572,
-0.1328430324792862,
-0.08996491879224777,
0.037999510765075684,
0.055287107825279236,
-0.016219541430473328,
0.11157703399658203,
-0.02790040522813797,
0.0671137273311615,
0.05197756364941597,
-0.14911557734012604,
-0.21309031546115875,
0.043088413774967194,
-0.03457297012209892,
0.16741053760051727,
0.032527483999729156,
0.07026690244674683,
-0.1318490356206894,
0.005996404681354761,
-0.008010598830878735,
-0.02555401436984539,
-0.113502137362957,
-0.04016893729567528,
0.10736791044473648,
0.01890859194099903,
-0.05588224157691002,
0.11932288110256195,
0.053731534630060196,
0.07207717001438141,
0.022103527560830116,
0.036430660635232925,
0.10638459026813507,
-0.05759545415639877,
0.08525355905294418,
0.19163745641708374,
0.022084489464759827,
-0.050156377255916595,
-0.1069810688495636,
-0.142279252409935,
0.1059383824467659,
-0.029212607070803642,
0.06867408007383347,
-0.16743674874305725,
-0.09695854038000107,
0.03239866718649864,
-0.006085241679102182,
-0.045712824910879135,
-0.04037291929125786,
-0.029692232608795166,
-0.1638854742050171,
0.07177262753248215,
-0.026750473305583,
0.09733851999044418,
-0.07764898240566254,
-0.08057862520217896,
-0.1878826767206192,
0.0927230566740036,
0.11600489169359207,
-0.09250454604625702,
-0.07816965878009796,
0.0006463889149017632,
0.007188722491264343,
-0.05905555561184883,
-0.05547625944018364,
0.05128099024295807,
-0.1268264353275299,
0.03925716504454613,
0.02211940288543701,
0.07955963909626007,
-0.013168327510356903,
-0.022237133234739304,
0.053730763494968414,
-0.05526714771986008,
-0.004513209220021963,
-0.0007778665167279541,
-0.010598957538604736,
-0.04734821990132332,
-0.2539333701133728,
0.026826584711670876,
0.015074611641466618,
0.023000292479991913,
0.11450504511594772,
0.052672553807497025,
0.002142281737178564,
-0.022901082411408424,
-0.09921795129776001,
0.004082086030393839,
0.0676940307021141,
-0.0444176085293293,
0.02973432093858719,
0.04361078143119812,
-0.10892095416784286,
-0.011856138706207275,
-0.024206269532442093,
0.07134921103715897,
0.010941405780613422,
0.06965811550617218,
-0.07052738219499588,
0.09066002070903778,
-0.1813029795885086,
-0.042003389447927475,
0.02394963428378105,
0.0719861164689064,
0.12007027864456177,
-0.10232933610677719,
0.05554276332259178,
0.007666701916605234,
0.16984406113624573,
0.10653958469629288,
-0.002575549529865384,
-0.03601353242993355,
0.06471540033817291,
0.09858960658311844,
0.034707363694906235,
0.04066390544176102,
0.06345933675765991,
-0.010203788988292217,
0.10382732003927231,
0.10297582298517227,
0.14551296830177307,
0.050692107528448105,
0.15706492960453033,
0.03763074800372124,
0.008729667402803898,
0.07412492483854294,
0.0944521427154541,
0.08652419596910477,
-0.006242257542908192,
0.1731923371553421,
-0.007543493993580341,
-0.01751723699271679,
-0.03595760464668274,
0.16348356008529663,
0.06810002774000168,
-0.10502735525369644,
0.032236937433481216,
-0.05084357038140297,
0.025795334950089455,
-0.021152885630726814,
-0.15513712167739868,
-0.03436838835477829,
-0.2639841139316559,
0.12161721289157867,
-0.04934193193912506,
-0.00526955584064126,
0.0620683990418911,
-0.019800636917352676,
-0.053851764649152756,
-0.00036916558747179806,
0.0654521957039833,
0.026729213073849678,
0.01114212442189455,
-0.028801998123526573,
-0.021474527195096016,
-0.19075548648834229,
-0.11265835911035538,
-0.04041624069213867,
-0.13205185532569885,
-0.026539895683526993,
0.02738100476562977,
-0.05638997629284859,
0.00884995236992836,
-0.0025031883269548416,
-0.01385815255343914,
0.04824291169643402,
-0.052424367517232895,
0.045965224504470825,
0.051154542714357376,
0.06721315532922745,
-0.07684784382581711,
0.00411610584706068,
0.11700203269720078,
0.03185063600540161,
-0.09347992390394211,
0.055158115923404694,
0.12995439767837524,
-0.058530066162347794,
0.026019345968961716,
-0.007744444999843836,
-0.032847896218299866,
-0.09708602726459503,
0.19312189519405365,
0.11783043295145035,
-0.16847896575927734,
0.0006766151054762304,
-0.036616407334804535,
-0.01160040870308876,
-0.09233774989843369,
0.12344596534967422,
0.1592838317155838,
0.055998723953962326,
-0.15062640607357025,
-0.11043619364500046,
-0.10300665348768234,
0.06709197163581848,
-0.07569106668233871,
-0.07460284233093262,
0.15964122116565704,
-0.02457398921251297,
-0.10188330709934235,
0.03819292411208153,
-0.21867942810058594,
-0.01995755359530449,
0.19039398431777954,
-0.29568302631378174,
-0.11494400352239609,
-0.07910088449716568,
0.18586759269237518,
0.025469033047556877,
0.11436232179403305,
-0.023825788870453835,
-0.02012297883629799,
-0.221383735537529,
0.0029703411273658276,
-0.08713068813085556,
0.034245800226926804,
0.0651308074593544,
-0.09516268968582153,
0.24007263779640198,
-0.09044498205184937,
0.05269941687583923,
0.033750344067811966,
0.07691317796707153,
0.01018204540014267,
0.05163824185729027,
-0.048588331788778305,
-0.16688252985477448,
-0.09095858782529831,
0.014404932036995888,
0.03795035555958748,
0.0503084696829319,
0.09903772920370102,
-0.04082057997584343,
0.04713768512010574,
0.0953395888209343,
0.030845828354358673,
-0.004454230889678001,
0.052237071096897125,
-0.15630710124969482,
0.05534590780735016,
0.018921079114079475,
-0.025683825835585594,
0.02539582923054695,
-0.08227502554655075,
0.10333657264709473,
0.03491305932402611,
0.0618959404528141,
-0.0665573701262474,
0.03160114586353302,
-0.009742318652570248,
-0.12334126234054565,
-0.04329211637377739,
-0.18513770401477814,
-0.0893927589058876,
-0.1391412913799286,
-0.03897256776690483,
-0.04044290632009506,
-0.025919048115611076,
0.01644543558359146,
0.00776201207190752,
-0.0044921645894646645,
-0.11029971390962601,
0.07136444747447968,
0.11884529888629913,
-0.030008424073457718,
0.0031494214199483395
] |
null | null | transformers |
## Original model card
Buy me a coffee if you like this project ;)
<a href="https://www.buymeacoffee.com/s3nh"><img src="https://www.buymeacoffee.com/assets/img/guidelines/download-assets-sm-1.svg" alt=""></a>
#### Description
GGUF Format model files for [This project](https://huggingface.co/Test157t/Kunocchini-7b-128k-test).
### GGUF Specs
GGUF is a format based on the existing GGJT, but makes a few changes to the format to make it more extensible and easier to use. The following features are desired:
Single-file deployment: they can be easily distributed and loaded, and do not require any external files for additional information.
Extensible: new features can be added to GGML-based executors/new information can be added to GGUF models without breaking compatibility with existing models.
mmap compatibility: models can be loaded using mmap for fast loading and saving.
Easy to use: models can be easily loaded and saved using a small amount of code, with no need for external libraries, regardless of the language used.
Full information: all information needed to load a model is contained in the model file, and no additional information needs to be provided by the user.
The key difference between GGJT and GGUF is the use of a key-value structure for the hyperparameters (now referred to as metadata), rather than a list of untyped values.
This allows for new metadata to be added without breaking compatibility with existing models, and to annotate the model with additional information that may be useful for
inference or for identifying the model.
### inference
User: Tell me story about what is an quantization and what do we need to build.
Assistant: Quantization is a process used in signal processing, physics, and mathematics to convert continuous signals or data into discrete values. It's often used in digital systems to represent analog signals or phenomena, allowing for efficient storage, transmission, and processing of information. In the context of audio, video, or images, quantization converts the infinite range of possible signal values into a finite number of levels that can be represented using a certain number of bits.
To understand this better, let's consider an example with audio. Sound waves are continuous signals, and when we record them, we need to
# Original model card
| {"language": ["zh", "en"], "license": "openrail", "library_name": "transformers", "pipeline_tag": "text-generation"} | text-generation | s3nh/Kunocchini-7b-128k-test-GGUF | [
"transformers",
"gguf",
"text-generation",
"zh",
"en",
"license:openrail",
"endpoints_compatible",
"region:us"
] | 2024-02-07T08:48:23+00:00 | [] | [
"zh",
"en"
] | TAGS
#transformers #gguf #text-generation #zh #en #license-openrail #endpoints_compatible #region-us
|
## Original model card
Buy me a coffee if you like this project ;)
<a href="URL src="URL alt=""></a>
#### Description
GGUF Format model files for This project.
### GGUF Specs
GGUF is a format based on the existing GGJT, but makes a few changes to the format to make it more extensible and easier to use. The following features are desired:
Single-file deployment: they can be easily distributed and loaded, and do not require any external files for additional information.
Extensible: new features can be added to GGML-based executors/new information can be added to GGUF models without breaking compatibility with existing models.
mmap compatibility: models can be loaded using mmap for fast loading and saving.
Easy to use: models can be easily loaded and saved using a small amount of code, with no need for external libraries, regardless of the language used.
Full information: all information needed to load a model is contained in the model file, and no additional information needs to be provided by the user.
The key difference between GGJT and GGUF is the use of a key-value structure for the hyperparameters (now referred to as metadata), rather than a list of untyped values.
This allows for new metadata to be added without breaking compatibility with existing models, and to annotate the model with additional information that may be useful for
inference or for identifying the model.
### inference
User: Tell me story about what is an quantization and what do we need to build.
Assistant: Quantization is a process used in signal processing, physics, and mathematics to convert continuous signals or data into discrete values. It's often used in digital systems to represent analog signals or phenomena, allowing for efficient storage, transmission, and processing of information. In the context of audio, video, or images, quantization converts the infinite range of possible signal values into a finite number of levels that can be represented using a certain number of bits.
To understand this better, let's consider an example with audio. Sound waves are continuous signals, and when we record them, we need to
# Original model card
| [
"## Original model card \n\nBuy me a coffee if you like this project ;)\n<a href=\"URL src=\"URL alt=\"\"></a>",
"#### Description \n\nGGUF Format model files for This project.",
"### GGUF Specs \n\nGGUF is a format based on the existing GGJT, but makes a few changes to the format to make it more extensible and easier to use. The following features are desired:\n\nSingle-file deployment: they can be easily distributed and loaded, and do not require any external files for additional information.\nExtensible: new features can be added to GGML-based executors/new information can be added to GGUF models without breaking compatibility with existing models.\nmmap compatibility: models can be loaded using mmap for fast loading and saving.\nEasy to use: models can be easily loaded and saved using a small amount of code, with no need for external libraries, regardless of the language used.\nFull information: all information needed to load a model is contained in the model file, and no additional information needs to be provided by the user.\nThe key difference between GGJT and GGUF is the use of a key-value structure for the hyperparameters (now referred to as metadata), rather than a list of untyped values. \nThis allows for new metadata to be added without breaking compatibility with existing models, and to annotate the model with additional information that may be useful for\ninference or for identifying the model.",
"### inference \n\n\n User: Tell me story about what is an quantization and what do we need to build.\n\nAssistant: Quantization is a process used in signal processing, physics, and mathematics to convert continuous signals or data into discrete values. It's often used in digital systems to represent analog signals or phenomena, allowing for efficient storage, transmission, and processing of information. In the context of audio, video, or images, quantization converts the infinite range of possible signal values into a finite number of levels that can be represented using a certain number of bits.\n\nTo understand this better, let's consider an example with audio. Sound waves are continuous signals, and when we record them, we need to",
"# Original model card"
] | [
"TAGS\n#transformers #gguf #text-generation #zh #en #license-openrail #endpoints_compatible #region-us \n",
"## Original model card \n\nBuy me a coffee if you like this project ;)\n<a href=\"URL src=\"URL alt=\"\"></a>",
"#### Description \n\nGGUF Format model files for This project.",
"### GGUF Specs \n\nGGUF is a format based on the existing GGJT, but makes a few changes to the format to make it more extensible and easier to use. The following features are desired:\n\nSingle-file deployment: they can be easily distributed and loaded, and do not require any external files for additional information.\nExtensible: new features can be added to GGML-based executors/new information can be added to GGUF models without breaking compatibility with existing models.\nmmap compatibility: models can be loaded using mmap for fast loading and saving.\nEasy to use: models can be easily loaded and saved using a small amount of code, with no need for external libraries, regardless of the language used.\nFull information: all information needed to load a model is contained in the model file, and no additional information needs to be provided by the user.\nThe key difference between GGJT and GGUF is the use of a key-value structure for the hyperparameters (now referred to as metadata), rather than a list of untyped values. \nThis allows for new metadata to be added without breaking compatibility with existing models, and to annotate the model with additional information that may be useful for\ninference or for identifying the model.",
"### inference \n\n\n User: Tell me story about what is an quantization and what do we need to build.\n\nAssistant: Quantization is a process used in signal processing, physics, and mathematics to convert continuous signals or data into discrete values. It's often used in digital systems to represent analog signals or phenomena, allowing for efficient storage, transmission, and processing of information. In the context of audio, video, or images, quantization converts the infinite range of possible signal values into a finite number of levels that can be represented using a certain number of bits.\n\nTo understand this better, let's consider an example with audio. Sound waves are continuous signals, and when we record them, we need to",
"# Original model card"
] | [
35,
30,
13,
284,
165,
4
] | [
"passage: TAGS\n#transformers #gguf #text-generation #zh #en #license-openrail #endpoints_compatible #region-us \n## Original model card \n\nBuy me a coffee if you like this project ;)\n<a href=\"URL src=\"URL alt=\"\"></a>#### Description \n\nGGUF Format model files for This project.### GGUF Specs \n\nGGUF is a format based on the existing GGJT, but makes a few changes to the format to make it more extensible and easier to use. The following features are desired:\n\nSingle-file deployment: they can be easily distributed and loaded, and do not require any external files for additional information.\nExtensible: new features can be added to GGML-based executors/new information can be added to GGUF models without breaking compatibility with existing models.\nmmap compatibility: models can be loaded using mmap for fast loading and saving.\nEasy to use: models can be easily loaded and saved using a small amount of code, with no need for external libraries, regardless of the language used.\nFull information: all information needed to load a model is contained in the model file, and no additional information needs to be provided by the user.\nThe key difference between GGJT and GGUF is the use of a key-value structure for the hyperparameters (now referred to as metadata), rather than a list of untyped values. \nThis allows for new metadata to be added without breaking compatibility with existing models, and to annotate the model with additional information that may be useful for\ninference or for identifying the model."
] | [
-0.0148029625415802,
0.09197601675987244,
-0.004234130959957838,
0.043110914528369904,
0.05438320338726044,
0.07514534890651703,
0.04146687686443329,
0.10425593703985214,
0.014650163240730762,
0.037674520164728165,
-0.052267879247665405,
-0.0019481116905808449,
0.07060763984918594,
0.09137903153896332,
0.030782945454120636,
-0.22521266341209412,
0.023683352395892143,
-0.006560265552252531,
-0.028240572661161423,
-0.026956811547279358,
0.10527215898036957,
-0.05482003837823868,
0.07565030455589294,
0.052207428961992264,
-0.055180128663778305,
0.016262080520391464,
-0.02084929123520851,
-0.00712186936289072,
0.033579811453819275,
0.10901983827352524,
-0.04072253406047821,
0.07258524745702744,
-0.002151296241208911,
-0.09312070906162262,
0.024879345670342445,
0.05238658934831619,
-0.03063826821744442,
0.010819525457918644,
0.04146100580692291,
-0.01169652957469225,
0.12683649361133575,
-0.041398391127586365,
-0.03536776825785637,
0.06618379801511765,
0.014821314252912998,
-0.2496410459280014,
-0.07568483054637909,
-0.003244952531531453,
-0.07658770680427551,
-0.015829065814614296,
0.023502111434936523,
-0.02375001460313797,
-0.05040609836578369,
-0.03946380689740181,
0.20489402115345,
-0.21734417974948883,
0.02295331284403801,
0.21253164112567902,
-0.00468128826469183,
0.14584723114967346,
-0.04165055230259895,
0.07890539616346359,
-0.03908737376332283,
-0.02645605430006981,
0.1645200550556183,
-0.0012147780507802963,
-0.019270820543169975,
0.0675281435251236,
-0.09024463593959808,
0.03510495647788048,
0.07501108199357986,
-0.050539858639240265,
-0.030882032588124275,
-0.15377868711948395,
-0.043752945959568024,
0.1763928085565567,
-0.05320790037512779,
-0.04080896079540253,
0.048508938401937485,
0.07196901738643646,
0.08426927775144577,
-0.20936553180217743,
-0.06323947757482529,
-0.07219721376895905,
-0.03686573728919029,
0.1079302504658699,
0.06141474470496178,
0.06769857555627823,
0.0019200396491214633,
0.09273772686719894,
-0.28264954686164856,
-0.08921930938959122,
-0.18467073142528534,
-0.03648688271641731,
0.002476438647136092,
0.01943141408264637,
-0.005156117491424084,
-0.07320360839366913,
0.08539215475320816,
0.03413380682468414,
-0.04787705838680267,
0.044615767896175385,
0.03591946139931679,
-0.020042724907398224,
0.02427704632282257,
0.18165095150470734,
-0.051469817757606506,
-0.0002622349711600691,
0.09673429280519485,
-0.10270632058382034,
-0.00807151012122631,
-0.02565731853246689,
-0.11712701618671417,
-0.1142474114894867,
-0.056443389505147934,
0.06956101208925247,
-0.06248188391327858,
0.01671767234802246,
0.08429984003305435,
-0.09691037982702255,
0.2543509304523468,
-0.08042531460523605,
0.012640579603612423,
0.014327357523143291,
-0.057322919368743896,
0.027743814513087273,
-0.010069017298519611,
-0.031523335725069046,
-0.054448485374450684,
-0.024995867162942886,
-0.0641285702586174,
-0.006899749860167503,
-0.04466281086206436,
-0.06580615788698196,
-0.049734026193618774,
-0.03956714645028114,
-0.04125642776489258,
-0.08520030230283737,
-0.28013837337493896,
0.06721289455890656,
0.030235467478632927,
-0.058410100638866425,
0.054631445556879044,
0.04402110353112221,
0.049854204058647156,
-0.024192070588469505,
-0.0004258320841472596,
-0.014999933540821075,
-0.06476770341396332,
-0.020047852769494057,
-0.014186873100697994,
0.030963540077209473,
-0.042911045253276825,
0.02268918789923191,
-0.011082795448601246,
-0.04319373518228531,
-0.17516006529331207,
0.17023363709449768,
-0.13527356088161469,
-0.04877649247646332,
-0.03709572181105614,
-0.0064923521131277084,
-0.00185795733705163,
-0.07556116580963135,
-0.09410107880830765,
0.0158705897629261,
-0.006625530309975147,
-0.08972635865211487,
0.20843631029129028,
-0.12041706591844559,
0.009098769165575504,
0.08425328135490417,
-0.03129435330629349,
-0.024243654683232307,
0.15679562091827393,
0.15591447055339813,
0.2523444592952728,
-0.12113994359970093,
-0.015250160358846188,
0.11862966418266296,
-0.03307455778121948,
0.06351764500141144,
0.08107437938451767,
0.028963308781385422,
-0.013941939920186996,
0.06434166431427002,
-0.03558434918522835,
0.06616167724132538,
0.021328894421458244,
0.0024254745803773403,
-0.0631704181432724,
-0.09018579870462418,
0.0013513342710211873,
-0.03672957792878151,
0.04745275527238846,
0.04133313149213791,
-0.04775498807430267,
0.06622597575187683,
0.1386614590883255,
-0.08445542305707932,
-0.05528418719768524,
-0.06369215995073318,
0.07763306796550751,
-0.12226623296737671,
0.005238848738372326,
0.032268352806568146,
-0.10079829394817352,
0.11198259890079498,
-0.16864238679409027,
0.07018976658582687,
0.11934676766395569,
0.040221404284238815,
0.07955525815486908,
-0.05339035764336586,
0.11405456066131592,
0.05030674859881401,
-0.011907175183296204,
0.03383046016097069,
0.002472157357260585,
-0.0015477249398827553,
0.010057489387691021,
0.028045108541846275,
0.00766347860917449,
0.033141955733299255,
0.10937844961881638,
-0.02360588312149048,
-0.008541104383766651,
-0.047614190727472305,
0.06717684864997864,
-0.1648167073726654,
0.016794392839074135,
-0.10657355189323425,
-0.08468509465456009,
0.058881502598524094,
-0.0573408268392086,
0.1526011824607849,
-0.098458431661129,
0.1680101752281189,
0.01443252433091402,
0.2013317197561264,
0.04429992288351059,
-0.03838447481393814,
-0.012324923649430275,
0.010500174015760422,
-0.10002396255731583,
0.0009103547781705856,
0.19548159837722778,
-0.0507998988032341,
0.12084811180830002,
-0.13766701519489288,
-0.10342829674482346,
0.022943520918488503,
0.01336726639419794,
0.05744268745183945,
-0.024377668276429176,
-0.00024374364875257015,
-0.09343097358942032,
0.07881902158260345,
-0.052177127450704575,
0.06396316736936569,
0.2779701054096222,
0.058816950768232346,
-0.027464082464575768,
-0.03551841154694557,
-0.054301243275403976,
-0.02593158558011055,
0.05575527995824814,
0.011081854812800884,
0.03890885412693024,
0.07159449905157089,
0.07551681995391846,
0.11766620725393295,
-0.06312428414821625,
0.04345468431711197,
0.08404753357172012,
-0.1565222144126892,
0.07459377497434616,
0.03923996165394783,
-0.04784096032381058,
0.04613405093550682,
0.0609043687582016,
0.20698952674865723,
-0.025609923526644707,
0.029080864042043686,
-0.044992923736572266,
0.11582619696855545,
-0.06629101932048798,
-0.150117427110672,
-0.07424093037843704,
0.04288870468735695,
-0.06978470832109451,
0.0552852563560009,
0.05182379484176636,
-0.07682055979967117,
-0.07512616366147995,
-0.06314603239297867,
0.06726919114589691,
-0.06056039780378342,
-0.0639071986079216,
-0.07320567965507507,
-0.010931559838354588,
-0.08259270340204239,
-0.18151339888572693,
-0.03407510370016098,
0.037337157875299454,
-0.17998775839805603,
-0.03136543929576874,
0.006411425769329071,
0.0069209495559334755,
0.11961939930915833,
-0.07088863104581833,
0.03596913442015648,
0.006618928164243698,
0.13686904311180115,
-0.012331800535321236,
0.12884621322155,
0.19315530359745026,
0.10740170627832413,
0.07590875029563904,
-0.055373791605234146,
0.038691505789756775,
-0.01389299426227808,
-0.04714444279670715,
-0.006244257558137178,
-0.0647892951965332,
-0.16268497705459595,
-0.08154039084911346,
-0.06635095179080963,
-0.035289183259010315,
-0.0221770741045475,
0.0943632572889328,
0.12249952554702759,
0.12705382704734802,
-0.12112066894769669,
0.0672779306769371,
0.060800712555646896,
0.05447763577103615,
0.08895272761583328,
0.04139358177781105,
0.05306261032819748,
-0.09734898060560226,
0.03626217693090439,
0.16695111989974976,
0.004587894771248102,
0.09908612072467804,
-0.11933395266532898,
0.17608113586902618,
-0.0018433256773278117,
0.058438610285520554,
0.09817642718553543,
0.006922436412423849,
-0.10484687238931656,
0.0001883206277852878,
-0.04772189259529114,
-0.07804073393344879,
0.0014087066520005465,
0.05869341269135475,
-0.04639194533228874,
-0.11887707561254501,
0.04576956853270531,
-0.07811236381530762,
0.042351122945547104,
-0.007233071606606245,
-0.06467169523239136,
-0.09180092811584473,
-0.12196651101112366,
0.04291791096329689,
0.0014742097118869424,
-0.08146971464157104,
-0.03894428536295891,
0.11828377842903137,
-0.07390891015529633,
0.0676419660449028,
0.028439274057745934,
0.05751262232661247,
-0.10034117847681046,
-0.006596128921955824,
-0.03725668415427208,
0.17185431718826294,
-0.015391496010124683,
0.17729780077934265,
-0.08252207934856415,
-0.1034659817814827,
0.07246450334787369,
0.06127411127090454,
-0.15164366364479065,
0.04442046210169792,
0.06209190562367439,
0.055048421025276184,
0.0101853646337986,
0.03187726065516472,
-0.020882397890090942,
0.031063633039593697,
-0.047046612948179245,
0.12947683036327362,
0.02660417929291725,
-0.12860655784606934,
-0.0229392908513546,
-0.06727144122123718,
-0.007313348352909088,
-0.020393257960677147,
-0.09727198630571365,
-0.16769130527973175,
-0.20561878383159637,
0.11997346580028534,
0.020632456988096237,
-0.03656135872006416,
-0.05353449285030365,
0.058011334389448166,
-0.10631228983402252,
0.18453682959079742,
0.01999487169086933,
-0.15881691873073578,
-0.13810837268829346,
-0.023724112659692764,
0.10978057235479355,
-0.09460192173719406,
0.05427343770861626,
-0.02198653481900692,
0.22822587192058563,
-0.018596507608890533,
-0.18734747171401978,
-0.0435098297894001,
-0.05459628254175186,
-0.054895248264074326,
0.04297473281621933,
-0.028879374265670776,
0.08660124242305756,
0.022515038028359413,
0.05649806931614876,
-0.01667158119380474,
-0.051014624536037445,
-0.18167193233966827,
-0.038484953343868256,
0.20319350063800812,
-0.0674603208899498,
0.07474970072507858,
0.0009453178499825299,
-0.011501198634505272,
0.05167585238814354,
-0.025035597383975983,
0.03430560231208801,
0.07425796985626221,
-0.10317602008581161,
0.1779402196407318,
0.17210035026073456,
-0.051752012223005295,
-0.17836876213550568,
-0.01798187382519245,
-0.03819980099797249,
0.0010077490005642176,
-0.06900961697101593,
-0.23197045922279358,
0.09010354429483414,
0.08883324265480042,
-0.06096335127949715,
0.11729119718074799,
-0.16439656913280487,
-0.04475465416908264,
-0.0182512104511261,
-0.017341414466500282,
0.12672370672225952,
-0.08365978300571442,
-0.05411388352513313,
-0.04248081147670746,
-0.16531500220298767,
0.11528391391038895,
-0.12105181813240051,
0.13618043065071106,
-0.0770038515329361,
-0.00040916245779953897,
0.019575748592615128,
-0.03962848335504532,
0.07912689447402954,
-0.029963187873363495,
0.05482931435108185,
-0.08304261416196823,
0.11120924353599548,
0.11842844635248184,
-0.061302248388528824,
0.16969013214111328,
-0.13047417998313904,
0.11383143067359924,
0.04946231469511986,
-0.10870085656642914,
-0.04552561044692993,
0.08248931169509888,
0.06409218907356262,
-0.11756159365177155,
-0.19009198248386383,
-0.004050345625728369,
0.05284185707569122,
0.0005768910050392151,
-0.09303522109985352,
-0.009570879861712456,
-0.00922592356801033,
0.03508616238832474,
0.005900920368731022,
-0.18826724588871002,
-0.053396955132484436,
-0.03190969303250313,
0.04610881954431534,
0.05712444335222244,
-0.2258535772562027,
0.10215351730585098,
0.015606474131345749,
0.020891238003969193,
0.09190549701452255,
0.03988153487443924,
-0.08891459554433823,
0.02986159734427929,
-0.03156393766403198,
-0.10054024308919907,
-0.17230503261089325,
-0.09043228626251221,
-0.08478228747844696,
0.04052002727985382,
0.022567398846149445,
0.08928454667329788,
-0.025387704372406006,
-0.05864658206701279,
-0.04033420607447624,
-0.019291456788778305,
0.0383356474339962,
0.1241641715168953,
0.02564482018351555,
-0.019662512466311455,
-0.07136540859937668,
0.04697640985250473,
0.04803387448191643,
0.03469250351190567,
0.03630976006388664,
0.03161521255970001,
-0.10075729340314865,
-0.0724359080195427,
-0.009395398199558258,
0.12613730132579803,
-0.09071828424930573,
0.016604898497462273,
-0.03140760213136673,
-0.023935088887810707,
0.00917722750455141,
0.06106339395046234,
-0.01292630285024643,
0.06678856909275055,
-0.03612053021788597,
0.10666448622941971,
-0.06569785624742508,
0.0636918768286705,
-0.05083530396223068,
0.08516556024551392,
-0.11674687266349792,
0.00730510801076889,
-0.011504611000418663,
0.07324716448783875,
-0.0449589341878891,
-0.0034517724998295307,
-0.024578021839261055,
-0.09180212020874023,
-0.17205315828323364,
0.06632912158966064,
0.002657563891261816,
0.011387337930500507,
-0.022154580801725388,
-0.017633378505706787,
0.035558830946683884,
0.10690797865390778,
-0.06598155945539474,
-0.0003923910844605416,
-0.0805053859949112,
0.07641315460205078,
-0.05803035572171211,
0.04064083844423294,
0.027444131672382355,
-0.007128879893571138,
0.16301164031028748,
0.028355903923511505,
-0.028236566111445427,
0.02475597895681858,
-0.21528859436511993,
-0.04258926957845688,
-0.047842249274253845,
0.036110833287239075,
-0.010687212459743023,
0.021360432729125023,
0.10295902192592621,
-0.005911857821047306,
-0.026156794279813766,
-0.0421663373708725,
0.03133561089634895,
-0.09914715588092804,
0.06267432868480682,
-0.09408864378929138,
0.038806211203336716,
-0.01304914616048336,
0.01644941046833992,
0.06611640751361847,
0.012991633266210556,
-0.05953185260295868,
-0.009231150150299072,
-0.006080505438148975,
-0.14795181155204773,
-0.061352021992206573,
0.021509405225515366,
-0.05787117779254913,
-0.003782420651987195,
-0.059957947582006454,
0.028888626024127007,
0.023580007255077362,
0.2898959219455719,
0.07729753851890564,
-0.0454934686422348,
-0.04649794101715088,
0.024441752582788467,
0.17832598090171814,
-0.03108573891222477,
0.17226408421993256,
0.026697352528572083,
-0.005957944318652153,
0.04627556353807449,
0.050162896513938904,
0.028483659029006958,
-0.16182085871696472,
0.04082100838422775,
-0.020657576620578766,
-0.035760294646024704,
-0.04456041753292084,
0.0004391804104670882,
-0.09252358227968216,
-0.08938385546207428,
0.09162823110818863,
-0.03266897425055504,
0.03573327884078026,
-0.049154359847307205,
-0.08467757701873779,
0.1337510049343109,
-0.11817222833633423,
0.12567993998527527,
0.07601795345544815,
-0.024872981011867523,
-0.11827964335680008,
-0.12453451007604599,
-0.0797196552157402,
-0.16760186851024628,
0.02151675894856453,
-0.09732291102409363,
0.09550341963768005,
0.02998759038746357,
0.021919021382927895,
0.00898027140647173,
0.10810766369104385,
0.0029388100374490023,
-0.07441836595535278,
0.013705331832170486,
-0.02462066151201725,
-0.08549414575099945,
0.11992344260215759,
0.01842523366212845,
0.09268759191036224,
0.06882338225841522,
0.09520965814590454,
0.02440427802503109,
0.09404914826154709,
0.0035659626591950655,
-0.049976374953985214,
0.007512437645345926,
-0.05004790797829628,
-0.02824023738503456,
-0.06001001596450806,
-0.0025180717930197716,
-0.02562921680510044,
-0.09601578116416931,
0.02064170315861702,
0.14545120298862457,
-0.016979871317744255,
-0.10153935104608536,
-0.019591504707932472,
0.12457387894392014,
0.038159336894750595,
0.05143444985151291,
0.053770508617162704,
-0.0300933625549078,
0.010053812526166439,
0.18274632096290588,
0.14556989073753357,
0.04082362726330757,
0.015588955953717232,
0.007062407676130533,
-0.017646389082074165,
0.061340879648923874,
0.10114669054746628,
-0.05393661558628082,
0.2615589201450348,
-0.017742959782481194,
0.08670713752508163,
0.05976317450404167,
-0.018402766436338425,
-0.07744009792804718,
0.10675939172506332,
-0.10120204091072083,
0.03298027813434601,
0.010182930156588554,
-0.023644492030143738,
-0.001776991761289537,
-0.09086927771568298,
-0.005754493642598391,
0.025153888389468193,
-0.02911008894443512,
0.02025008760392666,
-0.0025032004341483116,
-0.08414743095636368,
0.005708465818315744,
-0.0069794924929738045,
-0.019739704206585884,
0.06573575735092163,
-0.0022974959574639797,
-0.20477566123008728,
-0.009803768247365952,
0.0531037263572216,
0.02034194953739643,
0.2196461707353592,
-0.0050316001288592815,
-0.027696583420038223,
0.06968850642442703,
-0.0503472201526165,
-0.10115186870098114,
0.06308440864086151,
-0.038913268595933914,
-0.14213871955871582,
0.0026403642259538174,
0.10403532534837723,
-0.04050080105662346,
0.006259065121412277,
0.03175003454089165,
-0.09347830712795258,
0.008184781298041344,
0.04409966617822647,
-0.025983721017837524,
-0.05741210654377937,
-0.023133214563131332,
-0.059568263590335846,
0.10653354227542877,
0.05312012881040573,
0.035469383001327515,
-0.01015428639948368,
-0.10931306332349777,
0.09634536504745483,
0.06286320835351944,
0.1958925575017929,
-0.018883440643548965,
-0.02079462818801403,
-0.03614262491464615,
0.0676172524690628,
0.06840593367815018,
-0.12443508207798004,
0.01047283411026001,
0.03300696983933449,
0.00399126298725605,
0.03142842277884483,
0.11055174469947815,
0.1129893809556961,
0.06441119313240051,
-0.013096584938466549,
-0.2228529155254364,
-0.0598268061876297,
0.009014373645186424,
-0.13074198365211487,
0.004496550187468529
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.8.2 | {"library_name": "peft", "base_model": "ybelkada/blip2-opt-2.7b-fp16-sharded"} | null | satefi/fine-tuned-blip2-opt-2.7b-coco-captions-adapters | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:ybelkada/blip2-opt-2.7b-fp16-sharded",
"region:us"
] | 2024-02-07T08:48:52+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #safetensors #arxiv-1910.09700 #base_model-ybelkada/blip2-opt-2.7b-fp16-sharded #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.8.2 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
"TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-ybelkada/blip2-opt-2.7b-fp16-sharded #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
49,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-ybelkada/blip2-opt-2.7b-fp16-sharded #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2"
] | [
-0.1023959144949913,
0.20364031195640564,
-0.003501187078654766,
0.02728327363729477,
0.08585690706968307,
0.020007897168397903,
0.06716983020305634,
0.12274689972400665,
0.012622345238924026,
0.12875720858573914,
0.05166934058070183,
0.10228630155324936,
0.12245386838912964,
0.22668033838272095,
-0.012838003225624561,
-0.19243070483207703,
0.0242630448192358,
-0.0771036297082901,
0.005593413487076759,
0.12082977592945099,
0.13930879533290863,
-0.09762415289878845,
0.07061418145895004,
-0.022893469780683517,
-0.006428177002817392,
-0.03144235908985138,
-0.06823558360338211,
-0.026020023971796036,
0.05179716274142265,
0.052624523639678955,
0.04507560655474663,
-0.0026365057565271854,
0.08946135640144348,
-0.2719646394252777,
0.01013258844614029,
0.05144821107387543,
-0.0014026554999873042,
0.08499309420585632,
0.0988464504480362,
-0.026928776875138283,
0.1098974272608757,
-0.04064032807946205,
0.13379088044166565,
0.07711464166641235,
-0.09239108860492706,
-0.22410325706005096,
-0.07173959165811539,
0.08625253289937973,
0.1827836036682129,
0.06788639724254608,
-0.03555542975664139,
0.1263314187526703,
-0.07607069611549377,
0.015353861264884472,
0.07337884604930878,
-0.10034869611263275,
-0.07253215461969376,
0.06338474154472351,
0.1148637905716896,
0.08918241411447525,
-0.1183159276843071,
-0.03819655627012253,
0.03314727544784546,
0.04002053290605545,
0.08421246707439423,
0.011725823394954205,
0.16789239645004272,
0.030479831621050835,
-0.14271286129951477,
-0.052292387932538986,
0.13737669587135315,
0.023319685831665993,
-0.040452320128679276,
-0.23331835865974426,
-0.015376968309283257,
-0.06795480847358704,
-0.03502865135669708,
-0.05600921809673309,
0.038739193230867386,
0.003568066982552409,
0.11532311141490936,
-0.035693444311618805,
-0.08089860528707504,
-0.016660982742905617,
0.11346282064914703,
0.07784787565469742,
0.013616815209388733,
-0.015320400707423687,
0.011008908972144127,
0.12671732902526855,
0.05690603330731392,
-0.12490561604499817,
-0.04582716152071953,
-0.06711649894714355,
-0.04033141955733299,
-0.027365777641534805,
0.05342256277799606,
0.0353180356323719,
0.042800284922122955,
0.25008848309516907,
-0.025695038959383965,
0.05624420940876007,
0.05131547898054123,
0.01635528728365898,
0.032757923007011414,
0.09937477856874466,
-0.052993740886449814,
-0.18724198639392853,
-0.014669667929410934,
0.10533948987722397,
0.008535911329090595,
-0.02457994781434536,
-0.04416726529598236,
0.029510188847780228,
0.030378494411706924,
0.11640799790620804,
0.10281682014465332,
-0.024227039888501167,
-0.07068724185228348,
-0.059072207659482956,
0.216552734375,
-0.15430507063865662,
0.04941238462924957,
0.01976020447909832,
-0.02105451002717018,
-0.05080863833427429,
0.015585738234221935,
0.011093972250819206,
-0.03752332180738449,
0.10955710709095001,
-0.06118832156062126,
-0.049624186009168625,
-0.11145076900720596,
-0.04487142711877823,
0.03322191536426544,
0.005177776329219341,
-0.03966242074966431,
-0.031259022653102875,
-0.08961821347475052,
-0.09290562570095062,
0.09356793016195297,
-0.059249572455883026,
-0.07570087909698486,
-0.022671813145279884,
-0.06882917881011963,
0.02134208008646965,
0.017937064170837402,
0.08182405680418015,
-0.025928014889359474,
0.04417520761489868,
-0.031341515481472015,
0.0675363764166832,
0.08862745016813278,
0.03666501119732857,
-0.07080328464508057,
0.06790003925561905,
-0.19235196709632874,
0.08260858803987503,
-0.0774281769990921,
0.027750061824917793,
-0.15746945142745972,
-0.007029620930552483,
0.007698461879044771,
0.02178480476140976,
0.037930868566036224,
0.15137243270874023,
-0.1939772069454193,
-0.03221370279788971,
0.165089413523674,
-0.10368044674396515,
-0.11016760021448135,
0.04177917167544365,
-0.038249846547842026,
0.15739089250564575,
0.030280139297246933,
-0.0012426129542291164,
0.09249915927648544,
-0.14316007494926453,
-0.013938642106950283,
-0.024325892329216003,
0.017603334039449692,
0.0778016746044159,
0.07073507457971573,
-0.08130092173814774,
0.017243625596165657,
0.017793981358408928,
-0.061591293662786484,
-0.0081405621021986,
-0.03970023989677429,
-0.0987352728843689,
0.0057478612288832664,
-0.08744806051254272,
0.016670476645231247,
0.007478628307580948,
-0.0831889808177948,
-0.01675930991768837,
-0.14190812408924103,
-0.03510471433401108,
0.08182565867900848,
0.010915788821876049,
-0.017329296097159386,
-0.06898674368858337,
0.034451693296432495,
-0.03666619583964348,
-0.01648077555000782,
-0.14489136636257172,
-0.021427473053336143,
0.03630919009447098,
-0.15304622054100037,
-0.005145850591361523,
-0.11835219711065292,
0.06576516479253769,
0.01511449459940195,
-0.06323514133691788,
-0.04222502559423447,
0.017977474257349968,
-0.004212197847664356,
-0.057308491319417953,
-0.21832139790058136,
-0.03378254547715187,
-0.04404837638139725,
0.1459498107433319,
-0.22241605818271637,
0.04439277574419975,
0.011004806496202946,
0.12492083758115768,
0.009400355629622936,
-0.06328117102384567,
0.025281473994255066,
-0.060985516756772995,
-0.022151879966259003,
-0.07330615818500519,
-0.010349185205996037,
-0.0023612214718014,
-0.035151150077581406,
0.025524193421006203,
-0.14404910802841187,
-0.04832933098077774,
0.0898718312382698,
0.08889646828174591,
-0.1480346918106079,
0.004582213703542948,
-0.0479704849421978,
-0.06432840973138809,
-0.08677776902914047,
-0.0746101588010788,
0.07303491979837418,
0.050404179841279984,
0.05474574863910675,
-0.08046972006559372,
-0.06728234142065048,
0.006895649246871471,
-0.013016865588724613,
-0.026044650003314018,
0.12051089853048325,
0.08064275234937668,
-0.08795077353715897,
0.09453634172677994,
0.07926716655492783,
0.048489537090063095,
0.09307026863098145,
-0.012605591677129269,
-0.10079944878816605,
-0.029974179342389107,
0.05817700922489166,
0.013242031447589397,
0.15151557326316833,
-0.05331005901098251,
0.04880187660455704,
0.05229213833808899,
-0.046881623566150665,
0.0392768569290638,
-0.1003534123301506,
0.01598189026117325,
0.010466392152011395,
-0.014099279418587685,
0.04221281781792641,
-0.022992460057139397,
0.013768854551017284,
0.08918881416320801,
0.06303547322750092,
0.028075722977519035,
0.02467573806643486,
-0.03831007331609726,
-0.13685983419418335,
0.16914397478103638,
-0.09373056143522263,
-0.2350214421749115,
-0.1508769690990448,
0.033380694687366486,
0.05175327509641647,
-0.02225610613822937,
0.025787750259041786,
-0.03416982665657997,
-0.10775268822908401,
-0.08076061308383942,
0.017707832157611847,
0.04551579803228378,
-0.06924950331449509,
-0.05629436671733856,
0.03717868775129318,
0.03978053480386734,
-0.12887011468410492,
0.03030010685324669,
0.05566076934337616,
0.0030389325693249702,
-0.0023474569898098707,
0.04472624510526657,
0.0838165283203125,
0.19098424911499023,
0.0038206656463444233,
0.0005792943411506712,
0.052969906479120255,
0.2804574966430664,
-0.15404251217842102,
0.1257592886686325,
0.1265062391757965,
-0.05037635564804077,
0.0915827602148056,
0.20057283341884613,
0.04166555032134056,
-0.08155781030654907,
0.024662816897034645,
0.03081902116537094,
-0.033339425921440125,
-0.25858598947525024,
-0.07089923322200775,
-0.02134774811565876,
-0.0718344897031784,
0.08356352150440216,
0.08594240248203278,
0.09141265600919724,
0.034850113093853,
-0.07882706820964813,
-0.06002916395664215,
0.048778217285871506,
0.10919934511184692,
-0.031268373131752014,
0.01373243611305952,
0.08169406652450562,
-0.03843279555439949,
0.00291964178904891,
0.0957426205277443,
-0.016873784363269806,
0.1495479941368103,
0.043862324208021164,
0.10726974159479141,
0.06192391365766525,
0.07807627320289612,
-0.00427240040153265,
0.05275103449821472,
0.014167661778628826,
0.027501339092850685,
0.009811957366764545,
-0.0911397933959961,
0.027544016018509865,
0.12415897846221924,
0.019679436460137367,
0.03910135477781296,
0.02221125178039074,
-0.0599311999976635,
0.03939114511013031,
0.2089223861694336,
0.011348971165716648,
-0.19348034262657166,
-0.07075284421443939,
0.06647099554538727,
-0.08268588781356812,
-0.14843600988388062,
-0.007554139941930771,
0.02860885113477707,
-0.1735040545463562,
0.021763663738965988,
-0.04100511968135834,
0.10741151124238968,
-0.07009892165660858,
-0.04010515660047531,
0.09469404071569443,
0.06040317192673683,
-0.030702605843544006,
0.04793144762516022,
-0.17065903544425964,
0.1120685413479805,
0.030907707288861275,
0.06870585680007935,
-0.09663672745227814,
0.10015694797039032,
0.004520023241639137,
-0.019680628553032875,
0.16754156351089478,
0.004071482457220554,
-0.046421702951192856,
-0.07688678056001663,
-0.068022720515728,
-0.02427726984024048,
0.09113579243421555,
-0.13575509190559387,
0.0659659206867218,
-0.018054954707622528,
-0.038992740213871,
0.0012381378328427672,
-0.10723204910755157,
-0.11051654070615768,
-0.17094869911670685,
0.06820432096719742,
-0.07434363663196564,
0.0032095236238092184,
-0.09686926752328873,
-0.055401481688022614,
-0.009374987334012985,
0.1763092428445816,
-0.17980854213237762,
-0.11631330102682114,
-0.1491946429014206,
-0.10259010642766953,
0.17095473408699036,
-0.04734873026609421,
0.08819764107465744,
-0.0034766437020152807,
0.16787858307361603,
-0.014214687049388885,
-0.015718974173069,
0.087732695043087,
-0.09375686198472977,
-0.1954045295715332,
-0.054642241448163986,
0.1798044741153717,
0.12645776569843292,
0.0340544693171978,
-0.022557005286216736,
0.021083654835820198,
-0.045462194830179214,
-0.1131872907280922,
0.01341725792735815,
0.14102241396903992,
0.044991254806518555,
0.0011787780094891787,
-0.026516659185290337,
-0.11984778940677643,
-0.05476224422454834,
-0.050618626177310944,
-0.007076145615428686,
0.2069881707429886,
-0.08295559138059616,
0.16145899891853333,
0.12309258431196213,
-0.048618387430906296,
-0.20653872191905975,
0.0328410379588604,
0.04221039637923241,
0.015131350606679916,
0.03412385284900665,
-0.18393175303936005,
0.08414941281080246,
-0.011607940308749676,
-0.07911363989114761,
0.1776852309703827,
-0.18600867688655853,
-0.13374063372612,
0.0897565707564354,
0.0228035319596529,
-0.2260301113128662,
-0.13407503068447113,
-0.1137874573469162,
-0.01737418957054615,
-0.12759308516979218,
0.04896645247936249,
0.029215604066848755,
0.0038563767448067665,
0.017197856679558754,
0.017441323027014732,
0.04046647995710373,
-0.05223904177546501,
0.1984073370695114,
-0.026710696518421173,
0.006047636270523071,
-0.050715576857328415,
-0.09581901878118515,
0.02039787359535694,
-0.05753452330827713,
0.110266774892807,
-0.01032858807593584,
0.023703893646597862,
-0.16006609797477722,
-0.04355010390281677,
-0.06762884557247162,
0.020070020109415054,
-0.09522926807403564,
-0.0928168073296547,
-0.05127653107047081,
0.0819212943315506,
0.10820676386356354,
-0.02295168861746788,
0.0005749896517954767,
-0.07361534982919693,
0.07722114026546478,
0.22184693813323975,
0.1649458259344101,
0.048072416335344315,
-0.054855767637491226,
0.010806005448102951,
-0.03211841732263565,
0.0406072698533535,
-0.21899348497390747,
0.0427355132997036,
0.06035930663347244,
0.036828190088272095,
0.08413819968700409,
-0.012402283027768135,
-0.1615157127380371,
-0.07863713055849075,
0.07407596707344055,
-0.06613833457231522,
-0.16754086315631866,
-0.03876882418990135,
0.04338647425174713,
-0.19467578828334808,
-0.0508623942732811,
0.030222972854971886,
-0.021225454285740852,
-0.03482480347156525,
0.015321789309382439,
0.08263038098812103,
-0.006926113273948431,
0.10264649987220764,
0.07923399657011032,
0.09892527759075165,
-0.105184406042099,
0.06824152916669846,
0.08597805351018906,
-0.03393268212676048,
0.009138351306319237,
0.13618622720241547,
-0.05014333128929138,
-0.023080265149474144,
0.06525996327400208,
0.08105384558439255,
0.012546095997095108,
-0.052424877882003784,
0.013543504290282726,
-0.0725960060954094,
0.06149521470069885,
0.10263664275407791,
0.019078141078352928,
-0.016350749880075455,
0.06600207090377808,
0.02000228501856327,
-0.0930938571691513,
0.1252557784318924,
0.06951259076595306,
0.022252975031733513,
-0.038538623601198196,
-0.028527628630399704,
-0.013999884948134422,
-0.007938371039927006,
-0.014650342985987663,
-0.002483789110556245,
-0.07650505006313324,
-0.004464221652597189,
-0.1101321205496788,
0.01693287119269371,
-0.07970034331083298,
0.006229866296052933,
0.013589464128017426,
-0.042377084493637085,
-0.0007004099315963686,
-0.0010941700311377645,
-0.0825127363204956,
-0.060410551726818085,
-0.022047949954867363,
0.08035755902528763,
-0.12847869098186493,
0.01676725037395954,
0.07004719227552414,
-0.11456523090600967,
0.06635120511054993,
-0.01037197932600975,
0.012879660353064537,
0.0023543727584183216,
-0.13481667637825012,
0.05165598541498184,
-0.02099967747926712,
-0.0035665908362716436,
0.02452498860657215,
-0.16873589158058167,
-0.0032408160623162985,
-0.05216484144330025,
-0.0749099999666214,
0.0048598493449389935,
-0.04050567373633385,
-0.13309027254581451,
0.11099666357040405,
-0.012723537161946297,
-0.07056688517332077,
-0.02372054010629654,
0.05077226459980011,
0.08550150692462921,
-0.02510816790163517,
0.0938984677195549,
-0.025579016655683517,
0.08201071619987488,
-0.17812030017375946,
-0.009961896575987339,
-0.01465003564953804,
0.03416622057557106,
-0.018006669357419014,
-0.014842072501778603,
0.05195453017950058,
-0.012286731973290443,
0.17104652523994446,
-0.017782554030418396,
0.07721024751663208,
0.04898650571703911,
-0.002305001951754093,
0.02977832779288292,
0.06967566907405853,
0.06522846221923828,
-0.014649573713541031,
-0.005505430977791548,
0.026262225583195686,
-0.013115668669342995,
-0.043591879308223724,
-0.1496414840221405,
0.03997490927577019,
0.16483180224895477,
0.06666601449251175,
0.031087568029761314,
0.02089330367743969,
-0.137874573469162,
-0.08431115001440048,
0.10268014669418335,
-0.016490649431943893,
-0.017492152750492096,
-0.07017472386360168,
0.1938580423593521,
0.12399069219827652,
-0.19902944564819336,
0.07030948996543884,
-0.047140613198280334,
-0.03475939482450485,
-0.12678919732570648,
-0.1504630148410797,
-0.057876620441675186,
-0.0396706759929657,
-0.023383179679512978,
-0.061611108481884,
0.05963646247982979,
0.045422691851854324,
-0.001120392931625247,
-0.007043042685836554,
0.10047915577888489,
0.002273647114634514,
-0.025097914040088654,
0.06634964048862457,
0.07210102677345276,
0.044173598289489746,
-0.08350478112697601,
0.005473924335092306,
-0.004407464060932398,
0.005279168486595154,
0.05772916600108147,
0.019375650212168694,
-0.059642791748046875,
0.022227197885513306,
-0.001713956706225872,
-0.11705102026462555,
0.04011988639831543,
-0.015719687566161156,
-0.03440998122096062,
0.1458049863576889,
0.023999176919460297,
0.010189180262386799,
-0.02722814306616783,
0.2215377539396286,
-0.08876711875200272,
-0.07316280901432037,
-0.13400240242481232,
0.07494153827428818,
-0.05017445236444473,
0.035624027252197266,
0.03759501129388809,
-0.12485615164041519,
0.006820192094892263,
0.1633101850748062,
0.1287752240896225,
0.0015842759748920798,
0.009181773290038109,
0.05443078279495239,
0.005943849217146635,
-0.03531435877084732,
0.022326532751321793,
0.04972163215279579,
0.18732118606567383,
-0.07171057164669037,
0.07961983233690262,
-0.014104119502007961,
-0.07633235305547714,
-0.02722921036183834,
0.1351630687713623,
-0.0060691554099321365,
-0.0005658621084876359,
-0.061430297791957855,
0.13521265983581543,
-0.05378999561071396,
-0.22143587470054626,
0.061226099729537964,
-0.08844279497861862,
-0.14051449298858643,
-0.03979843854904175,
0.0020295942667871714,
-0.018312251195311546,
0.01529828179627657,
0.06830412894487381,
-0.054241351783275604,
0.1824820339679718,
0.032676003873348236,
-0.06139485538005829,
-0.08912920951843262,
0.050898075103759766,
-0.14380158483982086,
0.2871195673942566,
0.026256408542394638,
0.03674665838479996,
0.10091640055179596,
-0.02551521733403206,
-0.1445848047733307,
0.022934162989258766,
0.11640027910470963,
-0.07552778720855713,
0.05333654582500458,
0.17266467213630676,
-0.007869631983339787,
0.13195273280143738,
0.05329576134681702,
-0.06672891974449158,
0.034466732293367386,
-0.05582581087946892,
-0.06128007173538208,
-0.12193675339221954,
0.07341577857732773,
-0.07046602666378021,
0.14958375692367554,
0.12877047061920166,
-0.06340276449918747,
-0.009213985875248909,
-0.019985152408480644,
0.0771741271018982,
0.016045335680246353,
0.13332360982894897,
0.02260434255003929,
-0.18001601099967957,
0.04458257183432579,
-0.005987263284623623,
0.11051776260137558,
-0.20083686709403992,
-0.06018247827887535,
0.03471638634800911,
-0.02889879234135151,
-0.0835249274969101,
0.11544446647167206,
0.0454484261572361,
0.020198026672005653,
-0.029829418286681175,
-0.07667473703622818,
-0.0012824061559513211,
0.15097537636756897,
-0.10192548483610153,
-0.005492199212312698
] |
null | null | diffusers |
# SDXL LoRA DreamBooth - finalyear2023/virat-kholi
<Gallery />
## Model description
These are finalyear2023/virat-kholi LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.
The weights were trained using [DreamBooth](https://dreambooth.github.io/).
LoRA for the text encoder was enabled: False.
Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
## Trigger words
You should use a photo of virat kholi, to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
[Download](finalyear2023/virat-kholi/tree/main) them in the Files & versions tab.
| {"license": "openrail++", "tags": ["stable-diffusion-xl", "stable-diffusion-xl-diffusers", "text-to-image", "diffusers", "lora", "template:sd-lora"], "base_model": "stabilityai/stable-diffusion-xl-base-1.0", "instance_prompt": "a photo of virat kholi,"} | text-to-image | finalyear2023/virat-kholi | [
"diffusers",
"tensorboard",
"stable-diffusion-xl",
"stable-diffusion-xl-diffusers",
"text-to-image",
"lora",
"template:sd-lora",
"base_model:stabilityai/stable-diffusion-xl-base-1.0",
"license:openrail++",
"has_space",
"region:us"
] | 2024-02-07T08:54:29+00:00 | [] | [] | TAGS
#diffusers #tensorboard #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us
|
# SDXL LoRA DreamBooth - finalyear2023/virat-kholi
<Gallery />
## Model description
These are finalyear2023/virat-kholi LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.
The weights were trained using DreamBooth.
LoRA for the text encoder was enabled: False.
Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
## Trigger words
You should use a photo of virat kholi, to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
Download them in the Files & versions tab.
| [
"# SDXL LoRA DreamBooth - finalyear2023/virat-kholi\n\n<Gallery />",
"## Model description\n\nThese are finalyear2023/virat-kholi LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.\n\nThe weights were trained using DreamBooth.\n\nLoRA for the text encoder was enabled: False.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix.",
"## Trigger words\n\nYou should use a photo of virat kholi, to trigger the image generation.",
"## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab."
] | [
"TAGS\n#diffusers #tensorboard #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us \n",
"# SDXL LoRA DreamBooth - finalyear2023/virat-kholi\n\n<Gallery />",
"## Model description\n\nThese are finalyear2023/virat-kholi LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.\n\nThe weights were trained using DreamBooth.\n\nLoRA for the text encoder was enabled: False.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix.",
"## Trigger words\n\nYou should use a photo of virat kholi, to trigger the image generation.",
"## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab."
] | [
86,
23,
88,
21,
28
] | [
"passage: TAGS\n#diffusers #tensorboard #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us \n# SDXL LoRA DreamBooth - finalyear2023/virat-kholi\n\n<Gallery />## Model description\n\nThese are finalyear2023/virat-kholi LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.\n\nThe weights were trained using DreamBooth.\n\nLoRA for the text encoder was enabled: False.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix.## Trigger words\n\nYou should use a photo of virat kholi, to trigger the image generation.## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab."
] | [
-0.08089561760425568,
0.10146234929561615,
-0.0007670729537494481,
0.024441862478852272,
0.13312192261219025,
0.02954397164285183,
0.13472437858581543,
0.12176594883203506,
0.0981568768620491,
0.08315963298082352,
0.04035024717450142,
0.05871238932013512,
0.08738698810338974,
0.22491255402565002,
-0.01639077439904213,
-0.20869411528110504,
0.016178229823708534,
0.017050471156835556,
-0.035300202667713165,
0.020747072994709015,
0.0308073740452528,
-0.09523496776819229,
0.09033481776714325,
-0.029004421085119247,
-0.10193607956171036,
0.0552823431789875,
0.008124264888465405,
-0.013474701903760433,
0.026474151760339737,
0.05340965837240219,
0.049050599336624146,
0.10018766671419144,
0.08379926532506943,
-0.1458633989095688,
0.025905800983309746,
0.09209869056940079,
-0.012766956351697445,
0.03687377646565437,
-0.03089536726474762,
-0.029393400996923447,
0.10936719924211502,
-0.12972603738307953,
0.026834603399038315,
0.040417805314064026,
-0.0008868347504176199,
-0.1824757158756256,
-0.044758155941963196,
0.010056303814053535,
0.08318297564983368,
0.041310835629701614,
0.0033072649966925383,
0.028327036648988724,
0.12884408235549927,
0.053771648555994034,
0.22646424174308777,
-0.16226153075695038,
-0.0450579971075058,
0.24787724018096924,
0.052743759006261826,
0.08398362249135971,
-0.04172034189105034,
0.08053701370954514,
0.09350284188985825,
-0.041003983467817307,
0.1206974983215332,
-0.058131586760282516,
0.014960932545363903,
-0.09041555970907211,
-0.10774502903223038,
0.008357971906661987,
0.18163521587848663,
-0.013562179170548916,
-0.04990266263484955,
-0.1117626428604126,
-0.060461465269327164,
0.08450619131326675,
-0.021795174106955528,
-0.008334007114171982,
0.031132349744439125,
-0.029368825256824493,
0.022861404344439507,
-0.12967362999916077,
-0.05079222843050957,
-0.0904100239276886,
0.05591406673192978,
0.1928635686635971,
0.03247597813606262,
0.04019042104482651,
0.024129729717969894,
0.11049818247556686,
-0.10042817890644073,
-0.14863863587379456,
-0.022993404418230057,
-0.03718884661793709,
-0.03807573765516281,
0.029409008100628853,
0.018359649926424026,
-0.21794874966144562,
0.035249318927526474,
-0.027069522067904472,
-0.0348251536488533,
0.029326867312192917,
-0.027228228747844696,
0.027329185977578163,
0.015005807392299175,
0.09044936299324036,
-0.018718572333455086,
-0.040998589247465134,
0.05864863470196724,
0.08682659268379211,
0.05749475955963135,
-0.043286386877298355,
-0.08585407584905624,
-0.053162362426519394,
-0.06855392456054688,
0.06813182681798935,
-0.001171226380392909,
0.004239915404468775,
-0.01822623983025551,
-0.06995169818401337,
0.10671089589595795,
-0.07902498543262482,
-0.017654109746217728,
-0.025099271908402443,
-0.0639139860868454,
0.11111876368522644,
0.12974394857883453,
-0.01447712630033493,
-0.02860845811665058,
0.053567517548799515,
-0.046295903623104095,
-0.00869801826775074,
-0.09870053082704544,
-0.10310200601816177,
-0.016086241230368614,
-0.2371486872434616,
-0.020945215597748756,
-0.08090943098068237,
-0.24922975897789001,
-0.047360461205244064,
0.0501820333302021,
-0.014463778585195541,
-0.01629241183400154,
-0.05167325213551521,
-0.05099134519696236,
-0.007163950242102146,
0.03754980117082596,
0.046396803110837936,
-0.004923132248222828,
0.07055758684873581,
-0.025963731110095978,
0.12485277652740479,
-0.020047083497047424,
0.03051094524562359,
-0.054195187985897064,
0.05434979870915413,
-0.15106923878192902,
0.11964735388755798,
-0.10268709063529968,
0.029832229018211365,
-0.09762298315763474,
-0.0344243161380291,
0.0049494169652462006,
-0.017025450244545937,
0.0349603034555912,
0.11943532526493073,
-0.2270164042711258,
-0.05676648020744324,
0.12295947968959808,
-0.19787681102752686,
-0.09008458256721497,
-0.017564842477440834,
-0.016437677666544914,
0.1213725209236145,
0.08198939263820648,
0.11762762069702148,
0.1439676284790039,
-0.25293657183647156,
-0.05125013738870621,
-0.00004050241477671079,
-0.03397146984934807,
-0.059716906398534775,
0.019424932077527046,
0.04849022254347801,
-0.04466618224978447,
0.0629458799958229,
-0.15260136127471924,
0.08477350324392319,
-0.007129716221243143,
-0.03074818290770054,
-0.048722732812166214,
-0.11827266216278076,
0.014532388187944889,
0.03477512300014496,
0.024676013737916946,
0.06947518140077591,
-0.019500741735100746,
0.1137382909655571,
0.1148710697889328,
-0.10667435824871063,
0.004137121606618166,
0.0028435573913156986,
0.08116406947374344,
-0.14223778247833252,
0.016204621642827988,
-0.06861106306314468,
-0.12122281640768051,
0.06622256338596344,
-0.007521778345108032,
0.018143590539693832,
-0.012505258433520794,
0.06722681224346161,
0.05529549717903137,
-0.05433933809399605,
-0.06278350949287415,
0.026884600520133972,
-0.027868468314409256,
-0.031731754541397095,
-0.12991876900196075,
-0.04846309497952461,
-0.07298535108566284,
0.052173808217048645,
-0.1796560436487198,
0.06679555773735046,
0.013582535088062286,
0.09524153918027878,
0.05032725632190704,
-0.0297767985612154,
0.07799060642719269,
-0.03805157169699669,
-0.005570136476308107,
-0.10045794397592545,
0.020100455731153488,
0.0046576643362641335,
-0.10868073254823685,
0.0966554507613182,
-0.14234746992588043,
-0.03899401053786278,
0.09111576527357101,
0.17166785895824432,
0.006486492231488228,
-0.08698383718729019,
-0.018099291250109673,
-0.02597067505121231,
-0.05937913432717323,
0.0024071061052381992,
0.034066226333379745,
0.02249155566096306,
0.07520446926355362,
-0.08160988241434097,
-0.003240960417315364,
0.04522985965013504,
0.0290070753544569,
-0.06061473488807678,
0.025708144530653954,
0.06708502769470215,
0.05370945855975151,
0.02653261087834835,
0.017360417172312737,
-0.07211440056562424,
0.23084653913974762,
0.0015122568001970649,
-0.0688687115907669,
0.0035559250973165035,
0.015047037973999977,
0.028413744643330574,
0.13719114661216736,
0.057979896664619446,
0.03289235755801201,
0.01461449358612299,
-0.07985972613096237,
-0.003984851762652397,
-0.12036025524139404,
-0.017798833549022675,
-0.0010461780475452542,
-0.08449973165988922,
0.12079785764217377,
0.07476067543029785,
-0.05643334612250328,
0.04922768473625183,
-0.04875585809350014,
0.08071114867925644,
0.0019146096892654896,
-0.04054456204175949,
-0.07775596529245377,
0.12905952334403992,
-0.03714752942323685,
-0.11788249015808105,
-0.09649429470300674,
0.04449015110731125,
-0.0884619727730751,
-0.02142355777323246,
0.020041577517986298,
-0.06535427272319794,
-0.04662139341235161,
-0.09433613717556,
0.08118166774511337,
0.03686744347214699,
0.02634526416659355,
-0.016122721135616302,
-0.019474023953080177,
0.023774489760398865,
-0.10128647834062576,
0.0005010917666368186,
-0.03640725463628769,
-0.07592201977968216,
0.008562778122723103,
0.018944835290312767,
0.10665972530841827,
0.12249936163425446,
-0.022419119253754616,
0.050688207149505615,
-0.025828152894973755,
0.12013053148984909,
-0.05492391437292099,
0.1423552930355072,
0.230788916349411,
0.09566271305084229,
0.07812997698783875,
0.13200928270816803,
-0.010369115509092808,
-0.050046827644109726,
0.049650270491838455,
0.06005021184682846,
-0.1251591145992279,
-0.141208216547966,
-0.08460354059934616,
-0.03439391776919365,
-0.030938265845179558,
0.10183032602071762,
0.05941345915198326,
0.10008861869573593,
0.15539759397506714,
-0.013997357338666916,
0.03179201856255531,
0.02994161657989025,
0.08516538143157959,
-0.0036041950806975365,
0.019136982038617134,
0.04212534800171852,
-0.08178605139255524,
0.03330136463046074,
0.08490771055221558,
0.03045131079852581,
0.21337604522705078,
-0.14709584414958954,
-0.010632249526679516,
0.04492960870265961,
0.05616345256567001,
0.037681810557842255,
0.06692066788673401,
-0.051432445645332336,
0.0007601411198265851,
0.0038712495006620884,
-0.1323811411857605,
0.04652133956551552,
0.12709978222846985,
-0.024065876379609108,
0.017047978937625885,
-0.010967599228024483,
0.08646222949028015,
-0.02366519719362259,
0.15038907527923584,
0.02398734726011753,
-0.27031245827674866,
-0.014985929243266582,
0.059076037257909775,
0.11960547417402267,
-0.05437600612640381,
-0.012755835428833961,
0.14800815284252167,
-0.06535419821739197,
0.12021826952695847,
-0.10039845108985901,
0.05126890912652016,
-0.03522635996341705,
-0.08787965774536133,
0.022587113082408905,
0.18991555273532867,
-0.029202543199062347,
0.032419074326753616,
-0.17247959971427917,
0.03995133191347122,
-0.0030541655141860247,
0.11334864795207977,
-0.05675116926431656,
0.02551211602985859,
0.052220411598682404,
0.04817260056734085,
0.1365424543619156,
-0.008758778683841228,
0.04282603785395622,
-0.05761430040001869,
-0.056052401661872864,
0.027719395235180855,
-0.016834791749715805,
-0.07030050456523895,
0.06567300111055374,
0.01325969584286213,
-0.0052398755215108395,
0.002910417504608631,
-0.009368165396153927,
-0.1617453396320343,
-0.08459275960922241,
-0.0018636113964021206,
0.10079294443130493,
0.014415249228477478,
-0.056286610662937164,
-0.08439581096172333,
-0.045578811317682266,
0.10272638499736786,
-0.03364460915327072,
-0.1767098605632782,
-0.11723269522190094,
0.0019125976832583547,
0.1091107651591301,
-0.07050172984600067,
0.014379066415131092,
-0.021945804357528687,
0.16662243008613586,
-0.1278492957353592,
-0.163181334733963,
0.00528927706182003,
-0.054921042174100876,
-0.13005568087100983,
-0.011607619002461433,
0.0783495232462883,
0.017008591443300247,
-0.03189048916101456,
-0.013167609460651875,
-0.006255125626921654,
0.03773726150393486,
-0.0881011113524437,
0.07964204251766205,
0.20577184855937958,
-0.06928537786006927,
0.04143428057432175,
-0.03276825323700905,
-0.08570058643817902,
-0.04848133772611618,
0.04514379799365997,
-0.0011987134348601103,
0.24351298809051514,
-0.07873450219631195,
0.10341314971446991,
0.07680314779281616,
-0.08744017779827118,
-0.21572934091091156,
0.04010464996099472,
0.051814548671245575,
0.015880202874541283,
-0.022057143971323967,
-0.2130403220653534,
0.12829893827438354,
0.015593324787914753,
-0.04225681722164154,
0.16803529858589172,
-0.32791903614997864,
-0.1498252898454666,
-0.026383819058537483,
0.18625634908676147,
0.13181762397289276,
-0.13712038099765778,
-0.05279464274644852,
-0.039558686316013336,
-0.026855695992708206,
0.12644430994987488,
-0.08289188891649246,
0.07050306349992752,
-0.035174410790205,
-0.017825039103627205,
0.02581826038658619,
-0.05371018499135971,
0.08769885450601578,
-0.013881907798349857,
0.07530058175325394,
-0.03960379585623741,
-0.006206386256963015,
0.11579005420207977,
-0.11104172468185425,
0.13155841827392578,
-0.12218441069126129,
0.06365145742893219,
-0.09273866564035416,
-0.037263527512550354,
-0.06981218606233597,
0.12461158633232117,
-0.013764547184109688,
-0.09828802943229675,
-0.05276822671294212,
0.056058645248413086,
0.09807157516479492,
0.007801414467394352,
-0.0261242538690567,
-0.010380901396274567,
0.01114096399396658,
0.1864960640668869,
-0.008455983363091946,
0.06414236128330231,
-0.02967468462884426,
-0.008025010116398335,
-0.023149212822318077,
0.11028653383255005,
-0.16663673520088196,
-0.017725959420204163,
0.09119322896003723,
0.059169866144657135,
0.08871112018823624,
0.0194312185049057,
-0.09386458992958069,
0.05144887417554855,
0.04217218607664108,
-0.09868500381708145,
-0.11481542885303497,
-0.05141398683190346,
-0.04826705530285835,
-0.022890208289027214,
-0.0015785805881023407,
0.12954314053058624,
-0.08609723299741745,
0.023904195055365562,
-0.019505970180034637,
0.022011537104845047,
0.001959846355021,
0.07553520798683167,
0.031664568930864334,
0.015599093399941921,
-0.07918791472911835,
0.1321425586938858,
0.01317578088492155,
-0.06545418500900269,
0.03745895251631737,
0.08933258056640625,
-0.0843159407377243,
-0.005057239904999733,
-0.026571443304419518,
0.12041793763637543,
-0.12016228586435318,
-0.019347481429576874,
-0.05406515300273895,
-0.08768679201602936,
-0.009633730165660381,
0.1442297250032425,
0.03047187067568302,
-0.027197660878300667,
-0.0014255407731980085,
-0.009966936893761158,
-0.11971535533666611,
0.11935364454984665,
0.024046985432505608,
0.08129459619522095,
-0.19509044289588928,
-0.002560853259637952,
0.06411930918693542,
0.02909155562520027,
-0.07491444051265717,
-0.03386801481246948,
-0.05517037585377693,
-0.03344040364027023,
-0.0002997775445692241,
0.08058977127075195,
-0.07621106505393982,
-0.017745982855558395,
-0.05491465702652931,
-0.04225039854645729,
-0.0114170853048563,
0.06850376725196838,
-0.03991227596998215,
-0.021792201325297356,
-0.016373882070183754,
0.044598083943128586,
-0.10883335024118423,
-0.08043152093887329,
-0.035612933337688446,
-0.06813076138496399,
0.039202384650707245,
-0.01921880431473255,
-0.033341359347105026,
0.0022277822718024254,
-0.15055057406425476,
0.05206483602523804,
0.09172762185335159,
-0.006718302611261606,
-0.023478634655475616,
-0.04131089895963669,
0.01179605070501566,
-0.03370188921689987,
-0.02113729529082775,
-0.046336036175489426,
-0.054243139922618866,
-0.10497814416885376,
0.0650208368897438,
-0.014974083751440048,
0.043932244181632996,
-0.03738120198249817,
0.08021508157253265,
0.10270125418901443,
0.10572249442338943,
0.0769025981426239,
-0.10221050679683685,
0.0430733822286129,
-0.1586303859949112,
-0.01933814212679863,
0.026691116392612457,
-0.02401973307132721,
-0.07128141075372696,
-0.019264832139015198,
0.03206723928451538,
0.0003137423482257873,
0.07864130288362503,
0.052420902997255325,
-0.04179641231894493,
-0.006640836596488953,
-0.021254193037748337,
0.03503657132387161,
-0.0005749529809691012,
0.21710723638534546,
0.005879472009837627,
0.04892267659306526,
0.0495016872882843,
0.05518051236867905,
0.07282068580389023,
0.08164379000663757,
0.06658651679754257,
0.04855448007583618,
0.05935860052704811,
0.09652289003133774,
0.03670719638466835,
-0.001707163406535983,
-0.04677603766322136,
0.12735876441001892,
-0.07814721763134003,
0.06673257797956467,
-0.057818423956632614,
0.07268347591161728,
0.11716963350772858,
-0.13240042328834534,
0.006647978909313679,
0.14061440527439117,
-0.05890559032559395,
-0.09314584732055664,
-0.23179668188095093,
-0.07134060561656952,
-0.13376367092132568,
0.041256099939346313,
-0.07587575167417526,
0.025405753403902054,
0.020607387647032738,
-0.0117795430123806,
0.043043456971645355,
0.1630363017320633,
0.024824006482958794,
-0.016108151525259018,
0.07450040429830551,
-0.0032930425368249416,
-0.06389244645833969,
0.0278604906052351,
0.01596684753894806,
0.10041863471269608,
-0.009639858268201351,
-0.01238888781517744,
0.04319554194808006,
0.036288488656282425,
0.07144796848297119,
0.06975287944078445,
-0.0769902914762497,
-0.009968654252588749,
0.03688129037618637,
0.01345175039023161,
0.13098682463169098,
0.07515997439622879,
-0.006999769248068333,
-0.05482374131679535,
0.13058975338935852,
-0.043819963932037354,
0.006570644211024046,
-0.0947699174284935,
0.07480823248624802,
-0.04332035779953003,
0.0012667372357100248,
-0.004400359001010656,
-0.09773319214582443,
-0.042057283222675323,
0.15811677277088165,
0.07600387185811996,
0.04420708492398262,
0.011004932224750519,
-0.08132917433977127,
-0.007041774224489927,
-0.0575440339744091,
0.07642506808042526,
0.009909432381391525,
0.1559745967388153,
-0.07332678884267807,
0.053425904363393784,
-0.060185518115758896,
-0.06625983864068985,
-0.023228034377098083,
-0.033627159893512726,
-0.04998684301972389,
-0.03977213427424431,
-0.028288966044783592,
0.04692510887980461,
-0.1126190721988678,
-0.14780214428901672,
0.1424667090177536,
-0.06491277366876602,
-0.0179216880351305,
-0.08567903190851212,
0.01157731655985117,
0.006810420658439398,
-0.01099473424255848,
-0.06458868086338043,
0.016990352421998978,
0.08906940370798111,
-0.016341274604201317,
-0.12507778406143188,
-0.06536577641963959,
-0.004631379619240761,
-0.1480383276939392,
0.18530616164207458,
-0.04963322728872299,
0.013765105046331882,
0.018361655995249748,
-0.023264583200216293,
-0.12395406514406204,
0.042814720422029495,
-0.015859870240092278,
-0.12482243031263351,
-0.050441332161426544,
0.19452239573001862,
-0.046282075345516205,
0.14653216302394867,
0.000024303912141476758,
-0.06794146448373795,
0.0037379979621618986,
0.07395683228969574,
-0.03996036946773529,
-0.09470177441835403,
0.014563718810677528,
-0.09982084482908249,
0.11769694834947586,
0.11176592856645584,
0.010927227325737476,
-0.00339817488566041,
-0.06668034940958023,
0.03605834022164345,
0.016796434298157692,
0.07964804768562317,
0.052684538066387177,
-0.049973487854003906,
-0.0047730035148561,
0.017140518873929977,
0.03325020521879196,
-0.20223577320575714,
-0.025363778695464134,
-0.1383672058582306,
-0.04921049252152443,
-0.0016864597564563155,
0.08642320334911346,
0.1157284528017044,
0.03731635585427284,
-0.010504716075956821,
-0.22680382430553436,
0.019871622323989868,
0.12434698641300201,
-0.15842242538928986,
-0.05609327182173729
] |
null | null | transformers |
The Moe model was constructed using microsoft/phi-2 as the base, with experts from microsoft/phi-2, g-ronimo/phi-2-OpenHermes-2.5, and mlx-community/phi-2-dpo-7k. Then qlora was applied to all layers of q,v, and gate linear on WizardLM_evol_instruct_70k via mlx.
The model was created using a script from https://github.com/mzbac/mlx-moe
## Evaluation
### hellaswag
| Tasks |Version|Filter|n-shot| Metric |Value | |Stderr|
|---------|------:|------|-----:|--------|-----:|---|-----:|
|hellaswag| 1|none | 0|acc |0.5482|± |0.0050|
| | |none | 0|acc_norm|0.7300|± |0.0044|
### MMLU
| Groups |Version|Filter|n-shot|Metric|Value | |Stderr|
|------------------|-------|------|-----:|------|-----:|---|-----:|
| - humanities |N/A |none | 0|acc |0.5817|± |0.0247|
| - other |N/A |none | 0|acc |0.5795|± |0.0311|
| - social_sciences|N/A |none | 0|acc |0.6347|± |0.0292|
| - stem |N/A |none | 0|acc |0.4486|± |0.0376|
### BBH
| Tasks |Version| Filter |n-shot| Metric |Value | |Stderr|
|----------------------|------:|----------|-----:|-----------|-----:|---|-----:|
|bbh_cot_fewshot_snarks| 2|get-answer| 3|exact_match|0.5281|± |0.0375|
### GSM8k
|Tasks|Version| Filter |n-shot| Metric |Value | |Stderr|
|-----|------:|----------|-----:|-----------|-----:|---|-----:|
|gsm8k| 2|get-answer| 5|exact_match|0.5224|± |0.0138|
## Example
```
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "mzbac/phi-2-2x3-hf"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id, trust_remote_code=True)
text = "Instruct: how backpropagation works.\nOutput:"
inputs = tokenizer(text, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=20)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
``` | {"license": "mit"} | text-generation | mzbac/phi-2-2x3-hf | [
"transformers",
"safetensors",
"phi2moe",
"text-generation",
"custom_code",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T08:55:21+00:00 | [] | [] | TAGS
#transformers #safetensors #phi2moe #text-generation #custom_code #license-mit #autotrain_compatible #endpoints_compatible #region-us
| The Moe model was constructed using microsoft/phi-2 as the base, with experts from microsoft/phi-2, g-ronimo/phi-2-OpenHermes-2.5, and mlx-community/phi-2-dpo-7k. Then qlora was applied to all layers of q,v, and gate linear on WizardLM\_evol\_instruct\_70k via mlx.
The model was created using a script from URL
Evaluation
----------
### hellaswag
### MMLU
### BBH
### GSM8k
Example
-------
| [
"### hellaswag",
"### MMLU",
"### BBH",
"### GSM8k\n\n\n\nExample\n-------"
] | [
"TAGS\n#transformers #safetensors #phi2moe #text-generation #custom_code #license-mit #autotrain_compatible #endpoints_compatible #region-us \n",
"### hellaswag",
"### MMLU",
"### BBH",
"### GSM8k\n\n\n\nExample\n-------"
] | [
50,
5,
4,
4,
9
] | [
"passage: TAGS\n#transformers #safetensors #phi2moe #text-generation #custom_code #license-mit #autotrain_compatible #endpoints_compatible #region-us \n### hellaswag### MMLU### BBH### GSM8k\n\n\n\nExample\n-------"
] | [
-0.018763834610581398,
-0.027750534936785698,
-0.004888508934527636,
0.004249346908181906,
0.08100607246160507,
0.0025800568982958794,
0.17576304078102112,
0.07410439848899841,
0.02963121049106121,
0.002135395770892501,
0.17310649156570435,
0.1854606717824936,
0.001980449305847287,
0.13344621658325195,
-0.07794666290283203,
-0.18238216638565063,
0.06714273989200592,
-0.006061367690563202,
0.04587601125240326,
0.10630793124437332,
0.09124793857336044,
-0.026947680860757828,
0.10231224447488785,
-0.06317305564880371,
-0.09325103461742401,
-0.011414253152906895,
0.03114928863942623,
-0.08493930101394653,
0.10389462113380432,
0.04637649655342102,
0.05529588460922241,
0.0521208830177784,
-0.034473028033971786,
-0.20017041265964508,
0.009262417443096638,
-0.02515977993607521,
-0.0777616873383522,
0.01926236040890217,
0.041213009506464005,
0.003027800703421235,
0.08034054934978485,
0.0730479285120964,
-0.05481346324086189,
0.05550459027290344,
-0.10579799860715866,
-0.08538191765546799,
-0.044449806213378906,
0.05641530454158783,
0.11914719641208649,
0.08313234150409698,
-0.013801963068544865,
0.14136965572834015,
-0.05732237920165062,
0.07580489665269852,
0.14274553954601288,
-0.30847665667533875,
0.022065240889787674,
0.13638973236083984,
0.10281805694103241,
0.03754334896802902,
-0.007642057724297047,
0.06355616450309753,
0.05520440638065338,
0.017377741634845734,
-0.008403083309531212,
-0.05458036810159683,
-0.018535368144512177,
0.06923333555459976,
-0.09358461946249008,
-0.04522440582513809,
0.10842513293027878,
-0.03548617660999298,
0.014900697395205498,
-0.11643857508897781,
-0.07776708155870438,
-0.07060842961072922,
-0.029270587489008904,
-0.00007438251486746594,
-0.06160819157958031,
0.11562830954790115,
-0.024664845317602158,
0.019675936549901962,
-0.09052899479866028,
-0.054654281586408615,
-0.1536022275686264,
0.20974931120872498,
0.003881514072418213,
0.011786078102886677,
-0.08470159024000168,
0.07868102192878723,
-0.059001125395298004,
-0.09581384807825089,
-0.02205459587275982,
-0.08345495164394379,
0.10266879945993423,
0.025089267641305923,
0.019354138523340225,
0.06344935297966003,
0.17181821167469025,
0.1443377435207367,
0.056382402777671814,
0.03621651604771614,
-0.0012890108628198504,
0.06515006721019745,
-0.051287662237882614,
-0.03816164657473564,
0.01939680241048336,
-0.028032969683408737,
0.06311490386724472,
-0.04508931562304497,
0.07278896868228912,
-0.04283538833260536,
-0.15176986157894135,
-0.012371446937322617,
0.11044393479824066,
0.10472427308559418,
0.029358863830566406,
0.05503375455737114,
-0.055998340249061584,
0.04298437759280205,
0.13115200400352478,
-0.03261025622487068,
-0.0044381627812981606,
-0.024560503661632538,
0.09475351125001907,
-0.15028217434883118,
-0.003357888199388981,
0.05171669274568558,
-0.0353308841586113,
0.07966292649507523,
-0.04914816841483116,
-0.046669136732816696,
-0.045892730355262756,
-0.03315671533346176,
0.029319848865270615,
-0.051372673362493515,
0.034495312720537186,
-0.15538355708122253,
-0.18655535578727722,
0.07788547873497009,
0.0035754761192947626,
-0.017271123826503754,
-0.028246216475963593,
0.015124873258173466,
-0.08140154182910919,
0.005705648567527533,
-0.08711271733045578,
-0.03919941931962967,
-0.0896611139178276,
0.0894021987915039,
-0.003986263647675514,
0.005200260318815708,
-0.11024387925863266,
0.003217634279280901,
-0.08721769601106644,
0.03880610689520836,
-0.07083950191736221,
0.025057043880224228,
-0.11994706839323044,
0.10027281939983368,
-0.00980509165674448,
-0.007677755784243345,
-0.07226766645908356,
0.06039191037416458,
-0.01414013933390379,
0.17493167519569397,
-0.0480453185737133,
-0.08988941460847855,
0.1875332146883011,
-0.14225643873214722,
-0.19896098971366882,
0.10978902876377106,
0.028195463120937347,
0.021270262077450752,
0.06532099843025208,
0.21751859784126282,
0.032453592866659164,
-0.15062890946865082,
0.011089405044913292,
0.10069812834262848,
-0.08641430735588074,
-0.08482450991868973,
0.03685098513960838,
-0.04258543625473976,
-0.06756197661161423,
0.0646088644862175,
0.07239366322755814,
0.09696325659751892,
-0.06701009720563889,
-0.04189625009894371,
-0.055078454315662384,
-0.015357526950538158,
-0.01314430683851242,
0.01668705604970455,
0.045380957424640656,
-0.0708116814494133,
-0.0078093744814395905,
-0.0212808046489954,
0.0018557256553322077,
-0.003336203983053565,
-0.014968300238251686,
-0.11835514008998871,
0.07962362468242645,
-0.02058809995651245,
0.03466133028268814,
-0.07884074747562408,
-0.06529844552278519,
0.015079435892403126,
0.02401174232363701,
-0.00026762892957776785,
0.0028316518291831017,
0.09213840961456299,
0.05912403389811516,
-0.018719002604484558,
-0.005498773418366909,
0.18296471238136292,
0.043469950556755066,
-0.08109842240810394,
-0.0979100689291954,
0.04631432518362999,
-0.031009366735816002,
0.058320775628089905,
-0.09298332780599594,
0.04840048775076866,
0.14667968451976776,
0.09166296571493149,
-0.0034435135312378407,
-0.003644738346338272,
-0.024680189788341522,
0.018500883132219315,
-0.05478879436850548,
-0.005755788646638393,
0.04063224047422409,
0.02147221937775612,
-0.087770015001297,
0.14506155252456665,
-0.20840837061405182,
0.379025399684906,
0.19805298745632172,
-0.12812383472919464,
0.014188796281814575,
-0.03576201573014259,
-0.009502826258540154,
0.010053073987364769,
0.0034514893777668476,
-0.025149961933493614,
0.022175822407007217,
-0.019998138770461082,
0.14861150085926056,
-0.05095929279923439,
-0.022495491430163383,
0.03994964435696602,
-0.07150927186012268,
-0.005664851050823927,
0.06583089381456375,
0.06621979176998138,
-0.12940716743469238,
0.18917037546634674,
0.17689348757266998,
0.03109457716345787,
0.16772417724132538,
-0.04279927536845207,
0.035306379199028015,
-0.017656711861491203,
0.03900054469704628,
0.011316045187413692,
0.031821317970752716,
-0.1667506843805313,
0.01545469369739294,
0.0507790744304657,
0.006396425887942314,
0.09264147281646729,
-0.14828650653362274,
-0.07888896018266678,
0.008427167311310768,
-0.044127438217401505,
-0.027424361556768417,
0.07554641366004944,
-0.0263141468167305,
0.06644153594970703,
-0.036906905472278595,
-0.060541413724422455,
0.0922168716788292,
0.012243866920471191,
-0.11912434548139572,
0.20081186294555664,
-0.12028921395540237,
-0.13833661377429962,
-0.18270866572856903,
-0.11746198683977127,
-0.043838877230882645,
0.04887411370873451,
0.12291358411312103,
-0.0818110853433609,
-0.05460377410054207,
-0.06845618784427643,
0.04006676375865936,
0.04381817206740379,
0.04374010115861893,
-0.05928172916173935,
0.03975443169474602,
-0.05612916126847267,
-0.10618440806865692,
-0.08669444918632507,
0.02592533454298973,
-0.03446274623274803,
0.10014776885509491,
-0.0951675996184349,
0.10671675950288773,
0.12629185616970062,
0.060961782932281494,
0.051233477890491486,
-0.04553232342004776,
0.22339768707752228,
-0.05821650102734566,
-0.03794484585523605,
0.17341214418411255,
-0.03352389857172966,
0.0501721128821373,
0.14206287264823914,
0.03522747755050659,
-0.08000482618808746,
0.017496099695563316,
-0.058115363121032715,
-0.09311123192310333,
-0.1362030804157257,
-0.0927121490240097,
-0.08157878369092941,
0.11748014390468597,
-0.02928834781050682,
0.08402005583047867,
0.15872374176979065,
0.09924276918172836,
-0.014463580213487148,
-0.033436186611652374,
0.09769091755151749,
0.1131971925497055,
0.15803058445453644,
-0.0022602861281484365,
0.12566176056861877,
-0.08530005812644958,
-0.10894959419965744,
0.10655400156974792,
0.013397743925452232,
0.1286979615688324,
0.0923735722899437,
0.04162459820508957,
0.08606952428817749,
0.06412015855312347,
0.10652384161949158,
0.08901965618133545,
0.06740357726812363,
-0.05648975446820259,
-0.0002545185561757535,
-0.05635347589850426,
-0.04790813475847244,
0.0741332620382309,
-0.07724406570196152,
-0.15123292803764343,
0.024626752361655235,
-0.03960879519581795,
0.06452158093452454,
0.09813836961984634,
0.04628336802124977,
-0.2246711701154709,
-0.0025254683569073677,
0.08128827065229416,
-0.022796453908085823,
-0.03727129101753235,
0.09194700419902802,
0.04199843853712082,
-0.04854484274983406,
0.08539345860481262,
-0.031241370365023613,
0.06789277493953705,
-0.01377968117594719,
0.06770895421504974,
-0.05662544444203377,
-0.07915302366018295,
-0.011340435594320297,
0.0873631089925766,
-0.2866702079772949,
0.2469393014907837,
0.01675083301961422,
0.03452640771865845,
-0.08402232825756073,
0.02549261599779129,
0.014594465494155884,
0.12974263727664948,
0.12306027859449387,
-0.00742421904578805,
-0.01847493089735508,
-0.13680753111839294,
-0.0015816937666386366,
0.03748176246881485,
0.113643579185009,
-0.019017435610294342,
0.0004957302007824183,
-0.038408875465393066,
-0.0036484934389591217,
0.02659139595925808,
-0.10216561704874039,
-0.11086070537567139,
-0.09806369990110397,
0.05948057025671005,
0.04762519523501396,
0.1242317333817482,
-0.10159903019666672,
0.01990726962685585,
-0.10716038197278976,
0.13728879392147064,
-0.12808626890182495,
-0.05944684520363808,
-0.06891979277133942,
-0.11520390957593918,
0.02181413769721985,
-0.06539838016033173,
0.06884019821882248,
-0.03673223406076431,
-0.016480321064591408,
-0.05377525836229324,
-0.1435961276292801,
0.14614109694957733,
-0.11652424186468124,
-0.10400211811065674,
-0.0239090695977211,
0.09912077337503433,
-0.05821201950311661,
-0.011408334597945213,
0.034374941140413284,
0.03989579528570175,
-0.07461681216955185,
-0.1450498253107071,
0.014452916570007801,
-0.02790459245443344,
0.03221629559993744,
0.03918612003326416,
-0.012853636406362057,
-0.0716143548488617,
0.022865453734993935,
-0.10261980444192886,
0.16930024325847626,
0.35108649730682373,
-0.0146924564614892,
0.0778777077794075,
0.16348548233509064,
-0.04055117815732956,
-0.32045653462409973,
-0.19509904086589813,
-0.11263304948806763,
-0.04425544664263725,
-0.029654324054718018,
-0.07148618251085281,
0.0807926133275032,
0.08528383076190948,
-0.05939389020204544,
0.0582314059138298,
-0.221437007188797,
-0.12838496267795563,
0.16130198538303375,
0.053646937012672424,
0.23284034430980682,
-0.2045404314994812,
-0.10419998317956924,
-0.09195364266633987,
-0.2133997082710266,
0.09532324969768524,
-0.12826962769031525,
0.05887250974774361,
0.010964971967041492,
-0.003814790863543749,
-0.0013758973218500614,
-0.08800821751356125,
0.12686173617839813,
-0.019165560603141785,
0.08138182759284973,
-0.10160710662603378,
-0.04468666389584541,
0.11787848174571991,
0.03249228373169899,
0.008366109803318977,
-0.14383400976657867,
0.017113322392106056,
-0.10359156131744385,
-0.0552607886493206,
-0.04481200501322746,
0.044450655579566956,
0.005458002910017967,
-0.09284411370754242,
-0.032824475318193436,
-0.050237733870744705,
-0.03226061910390854,
-0.03605012968182564,
0.18453529477119446,
-0.0717933177947998,
0.1417682021856308,
0.19875794649124146,
0.1384608894586563,
-0.17353945970535278,
0.09963811933994293,
-0.013604576699435711,
-0.08294294774532318,
0.031017296016216278,
-0.07707998901605606,
0.03309110179543495,
0.05459993705153465,
-0.017151115462183952,
0.09254699945449829,
0.11042260378599167,
0.009595678187906742,
0.028563985601067543,
0.17117901146411896,
-0.18474578857421875,
-0.10719011723995209,
-0.05129478871822357,
0.031399182975292206,
0.09059743583202362,
0.09608583897352219,
0.11593542993068695,
0.027857473120093346,
-0.0067945984192192554,
-0.019544284790754318,
0.05457666143774986,
-0.09447444975376129,
0.10022589564323425,
0.03269146382808685,
0.049408920109272,
-0.11595533788204193,
0.010559520684182644,
-0.07114635407924652,
-0.0634891614317894,
-0.013244238682091236,
0.07624319195747375,
-0.13871067762374878,
-0.1022917702794075,
0.011753909289836884,
0.1715795397758484,
-0.09341391175985336,
-0.0706830695271492,
-0.07999316602945328,
-0.15963827073574066,
0.029788529500365257,
0.16861788928508759,
0.08045387268066406,
0.08781127631664276,
0.024048740044236183,
0.030399955809116364,
-0.015097440220415592,
0.030871860682964325,
-0.009616590104997158,
0.05682634189724922,
-0.16007274389266968,
0.005009881686419249,
-0.048476435244083405,
0.00105723284650594,
-0.09420976042747498,
0.019917132332921028,
-0.15975996851921082,
0.011331414803862572,
-0.08622932434082031,
-0.03334062546491623,
-0.14086033403873444,
-0.029640955850481987,
0.0026299741584807634,
-0.04705313965678215,
-0.03808308392763138,
-0.02456182800233364,
-0.09565465897321701,
0.0046623325906693935,
-0.009771169163286686,
0.05718597024679184,
-0.08526942878961563,
-0.009230907075107098,
0.06876921653747559,
-0.054934222251176834,
0.11540818214416504,
0.08256177604198456,
-0.03368513658642769,
0.03815539553761482,
-0.13928566873073578,
-0.007597463671118021,
0.1367672085762024,
-0.005891208071261644,
-0.023872917518019676,
0.058866310864686966,
0.02809515781700611,
0.08918537944555283,
0.011296953074634075,
0.08030066639184952,
0.1620997041463852,
-0.0961420014500618,
0.033140454441308975,
-0.04655946046113968,
-0.09623072296380997,
-0.04946465790271759,
-0.01534411869943142,
0.12091919779777527,
-0.0030822630506008863,
0.144436314702034,
-0.09179981052875519,
0.002751843072474003,
-0.08786752820014954,
0.013858186081051826,
0.005209358874708414,
-0.15394940972328186,
-0.06401453167200089,
-0.07627080380916595,
-0.00025168684078380466,
0.015581495128571987,
0.28067874908447266,
0.005937475711107254,
-0.1410990208387375,
0.03931817412376404,
0.02936892956495285,
0.13061220943927765,
0.01914326287806034,
0.30164968967437744,
0.07171684503555298,
-0.005618199240416288,
-0.1256617158651352,
0.04663769155740738,
0.01696871593594551,
-0.14658042788505554,
0.039128441363573074,
0.11456988751888275,
-0.10447213053703308,
0.04129587486386299,
0.06989302486181259,
-0.09580468386411667,
-0.10160839557647705,
-0.12716247141361237,
-0.07692620158195496,
0.08973772823810577,
-0.024166172370314598,
0.07617339491844177,
0.16183163225650787,
-0.054773129522800446,
-0.012927068397402763,
-0.007825815118849277,
-0.03325164318084717,
-0.16097265481948853,
-0.0752418041229248,
-0.0823206678032875,
-0.1662834733724594,
0.05843821540474892,
-0.0744270384311676,
0.007036339957267046,
0.13354793190956116,
0.03585973009467125,
-0.012781667523086071,
0.12675173580646515,
-0.011549368500709534,
-0.021080078557133675,
-0.014155583456158638,
-0.053419262170791626,
-0.043467115610837936,
-0.02116106078028679,
-0.07298778742551804,
-0.0776359960436821,
-0.042479582130908966,
-0.002959956182166934,
0.00958860944956541,
-0.051069989800453186,
0.018313691020011902,
-0.13610801100730896,
-0.053938500583171844,
-0.04144447296857834,
0.05037735775113106,
-0.02888222597539425,
0.15861959755420685,
-0.042062561959028244,
-0.041673608124256134,
0.08478944003582001,
0.16983912885189056,
-0.08476489782333374,
-0.22850123047828674,
-0.04431437700986862,
0.21708284318447113,
0.010594593361020088,
0.13710086047649384,
-0.06705665588378906,
0.01309188548475504,
-0.004166803788393736,
0.2628924548625946,
0.2616743743419647,
-0.03376465290784836,
0.03769886493682861,
0.035053499042987823,
0.0220986008644104,
0.06792747229337692,
0.12754330039024353,
0.07169224321842194,
0.29064464569091797,
-0.0658186823129654,
0.013462142087519169,
-0.02672179788351059,
-0.03963703662157059,
-0.14495046436786652,
0.10049690306186676,
0.020816750824451447,
-0.08068002760410309,
-0.015863874927163124,
0.10268399119377136,
-0.12667721509933472,
0.1882927417755127,
-0.06879334151744843,
-0.13894890248775482,
-0.008619480766355991,
0.012279348447918892,
0.12541501224040985,
-0.050560735166072845,
0.04079977050423622,
-0.012973152101039886,
-0.034587759524583817,
-0.008622136898338795,
0.020150568336248398,
-0.22182895243167877,
0.05050681531429291,
-0.0028617808129638433,
-0.06124720349907875,
0.10143059492111206,
-0.036236513406038284,
0.11025447398424149,
0.1327601969242096,
-0.012551337480545044,
-0.043822091072797775,
0.1618357002735138,
0.03554617986083031,
-0.11137443780899048,
-0.01869349554181099,
-0.05768655613064766,
0.0010827649384737015,
-0.06712928414344788,
0.009764399379491806,
-0.15098176896572113,
0.04175106808543205,
-0.008549786172807217,
-0.09288536757230759,
-0.04763704165816307,
-0.01324834581464529,
-0.06776317209005356,
0.07292710989713669,
-0.029496075585484505,
-0.017052272334694862,
0.0012288081925362349,
-0.03337031975388527,
0.06715614348649979,
-0.013892761431634426,
-0.036938052624464035,
-0.016560105606913567,
-0.17841969430446625,
-0.037008192390203476,
0.12578943371772766,
0.02212642878293991,
-0.19091196358203888,
-0.015376339666545391,
-0.09341474622488022,
0.06299912184476852,
-0.16596391797065735,
0.08784022927284241,
0.18727540969848633,
-0.013677123934030533,
-0.011929092928767204,
-0.17665232717990875,
0.015886947512626648,
0.06632731854915619,
-0.04896271973848343,
-0.12185616791248322
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | phamtungthuy/law_model | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-07T08:55:26+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
31,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06646376848220825,
0.2168014943599701,
-0.00225935154594481,
0.023818302899599075,
0.1271018385887146,
-0.001635765191167593,
0.04218708351254463,
0.13324736058712006,
-0.020175931975245476,
0.11144465953111649,
0.046588581055402756,
0.09377603232860565,
0.09928803145885468,
0.18404334783554077,
0.04859916493296623,
-0.2059975117444992,
0.007056170143187046,
-0.09090408682823181,
0.014076028019189835,
0.1116579994559288,
0.13719257712364197,
-0.10291384905576706,
0.08272874355316162,
-0.04045208916068077,
-0.02019004337489605,
0.00012576708104461432,
-0.09259183704853058,
-0.07032395154237747,
0.06885425746440887,
0.06264153122901917,
0.051234472543001175,
0.001456156256608665,
0.09140396863222122,
-0.2864592671394348,
0.017265573143959045,
0.08406311273574829,
0.0027674848679453135,
0.06290827691555023,
0.07236549258232117,
-0.07389893382787704,
0.11328595131635666,
-0.08021481335163116,
0.13019037246704102,
0.08625296503305435,
-0.062064990401268005,
-0.23071379959583282,
-0.07525765895843506,
0.0963398814201355,
0.12251301854848862,
0.06215599179267883,
-0.022921854630112648,
0.15455181896686554,
-0.06248689442873001,
0.012971068732440472,
0.1294165402650833,
-0.11526761949062347,
-0.05572471022605896,
0.061741601675748825,
0.11775490641593933,
0.10740239918231964,
-0.14110268652439117,
-0.0017287094378843904,
0.04900608956813812,
0.029121357947587967,
0.08589313924312592,
0.022661056369543076,
0.12003941088914871,
0.04652795568108559,
-0.13695219159126282,
-0.04037507623434067,
0.12011898308992386,
0.038862764835357666,
-0.06446044892072678,
-0.2168138176202774,
-0.006778308190405369,
-0.0601806715130806,
-0.014732478186488152,
-0.07019448280334473,
0.039128515869379044,
-0.02470310963690281,
0.07317749410867691,
-0.04465159401297569,
-0.1063927412033081,
-0.0421026237308979,
0.0892222449183464,
0.07748593389987946,
0.011527054943144321,
-0.02519804798066616,
0.04627908393740654,
0.13455867767333984,
0.05402068421244621,
-0.10399353504180908,
-0.07017925381660461,
-0.06942764669656754,
-0.09420394152402878,
-0.04035796597599983,
0.056760527193546295,
0.031942449510097504,
0.02665667235851288,
0.22703726589679718,
0.016653569415211678,
0.04155244305729866,
0.0224777739495039,
0.01032855175435543,
0.043662428855895996,
0.0955500528216362,
-0.05303520709276199,
-0.15660029649734497,
-0.04072032496333122,
0.09077946096658707,
-0.0027527001220732927,
-0.036689214408397675,
-0.03966725245118141,
0.03849169611930847,
0.06843466311693192,
0.13122352957725525,
0.07552056759595871,
-0.017929591238498688,
-0.04813180863857269,
-0.030096933245658875,
0.23523783683776855,
-0.1493375599384308,
0.04426715523004532,
-0.02271856553852558,
-0.01804111897945404,
-0.03908449783921242,
0.03597262129187584,
0.022118929773569107,
-0.000004518366949923802,
0.09706240892410278,
-0.058981191366910934,
-0.05378659814596176,
-0.10168042778968811,
-0.03272576630115509,
0.04088849574327469,
-0.013975566253066063,
-0.010589460842311382,
-0.09025166928768158,
-0.09490354359149933,
-0.04766594246029854,
0.05537205561995506,
-0.05123869329690933,
-0.03770573064684868,
0.009465423412621021,
-0.08151785284280777,
-0.005444355774670839,
-0.005417742300778627,
0.10699385404586792,
-0.03222226724028587,
0.04445803165435791,
-0.027600755915045738,
0.05225523188710213,
0.09919606149196625,
0.031576547771692276,
-0.0773419588804245,
0.0561848059296608,
-0.22559374570846558,
0.07503069192171097,
-0.11481974273920059,
0.04335082694888115,
-0.1704932004213333,
-0.042439818382263184,
0.005444696638733149,
0.0139949731528759,
0.013206101022660732,
0.12720820307731628,
-0.19255615770816803,
-0.01654396951198578,
0.13260798156261444,
-0.09212633967399597,
-0.118110790848732,
0.07884611934423447,
-0.029701577499508858,
0.1624738723039627,
0.04682036489248276,
-0.027025915682315826,
0.09224298596382141,
-0.16434773802757263,
-0.07092688232660294,
-0.00949116237461567,
-0.01727987825870514,
0.12109188735485077,
0.07512219995260239,
-0.05991523340344429,
0.046571120619773865,
0.02832140028476715,
-0.038078423589468,
-0.04424772411584854,
-0.050857074558734894,
-0.10884185880422592,
-0.01070026308298111,
-0.08987759798765182,
0.04065500199794769,
-0.01250192429870367,
-0.07916021347045898,
-0.029885273426771164,
-0.18612512946128845,
-0.0030564051121473312,
0.10038342326879501,
0.0035033065360039473,
-0.005652366206049919,
-0.08666291832923889,
0.026358824223279953,
-0.03112892620265484,
-0.008404186926782131,
-0.16764774918556213,
-0.04399421438574791,
0.046902090311050415,
-0.16094985604286194,
0.020117372274398804,
-0.06413903087377548,
0.06334125250577927,
0.03641495108604431,
-0.05590536445379257,
-0.0248766727745533,
-0.01730942726135254,
0.011945613659918308,
-0.05083848536014557,
-0.18994836509227753,
-0.056277405470609665,
-0.037882111966609955,
0.149809330701828,
-0.25956398248672485,
0.032966937869787216,
0.051140617579221725,
0.14649195969104767,
0.00406361510977149,
-0.05115427449345589,
0.01429014839231968,
-0.05360214412212372,
-0.054652128368616104,
-0.06746816635131836,
-0.006135428790003061,
-0.027576493099331856,
-0.05147203803062439,
0.019243421033024788,
-0.1755700707435608,
-0.021410830318927765,
0.09424154460430145,
0.12876708805561066,
-0.1486445665359497,
-0.018640631809830666,
-0.048725154250860214,
-0.06339836865663528,
-0.0715010017156601,
-0.07038594037294388,
0.10712739825248718,
0.0513901449739933,
0.04796046018600464,
-0.07435787469148636,
-0.07092321664094925,
0.02726263552904129,
0.006906150374561548,
-0.03382374346256256,
0.08727246522903442,
0.05199531093239784,
-0.09209315478801727,
0.0756213590502739,
0.1092359870672226,
0.07177663594484329,
0.09363535046577454,
0.01574566215276718,
-0.11756632477045059,
-0.028492970392107964,
0.036266472190618515,
0.02740776725113392,
0.1465986967086792,
-0.05952361226081848,
0.04016614332795143,
0.04494241625070572,
-0.04170418903231621,
0.022319864481687546,
-0.08787637203931808,
0.024075502529740334,
0.025203049182891846,
-0.0034381982404738665,
0.06284574419260025,
-0.02525499276816845,
-0.0050758360885083675,
0.07016654312610626,
0.047779910266399384,
0.04621000960469246,
0.009655474685132504,
-0.01720241829752922,
-0.1047825813293457,
0.16950392723083496,
-0.0951867327094078,
-0.269941508769989,
-0.17632324993610382,
0.026197833940386772,
0.04035249724984169,
-0.022378476336598396,
0.031619444489479065,
-0.07056326419115067,
-0.10630585998296738,
-0.1060405746102333,
-0.002429972169920802,
0.01714223250746727,
-0.06364088505506516,
-0.0741225928068161,
0.07348573952913284,
0.04382912442088127,
-0.14902326464653015,
0.038552410900592804,
0.055694397538900375,
-0.057955220341682434,
-0.0233661737293005,
0.09118817001581192,
0.12397737801074982,
0.14583967626094818,
-0.021366750821471214,
-0.028626007959246635,
0.029004426673054695,
0.19620531797409058,
-0.13469526171684265,
0.10371150821447372,
0.13814030587673187,
-0.04545360431075096,
0.08360563963651657,
0.1560150384902954,
0.029186224564909935,
-0.08317049592733383,
0.05044832453131676,
0.04082648828625679,
-0.043159641325473785,
-0.2666129767894745,
-0.0534592866897583,
0.012832709588110447,
-0.06255637854337692,
0.09786593168973923,
0.10183793306350708,
0.11542957276105881,
0.034910861402750015,
-0.07166364789009094,
-0.043925940990448,
-0.0058974819257855415,
0.11737963557243347,
-0.05490213260054588,
-0.012639665976166725,
0.07686592638492584,
-0.05086168646812439,
0.005355054512619972,
0.10266812145709991,
0.02973790094256401,
0.17442677915096283,
0.020399179309606552,
0.11231429129838943,
0.06195578724145889,
0.08633565157651901,
0.0007386076031252742,
0.02951662428677082,
0.05147615820169449,
0.017203815281391144,
-0.002300140680745244,
-0.10421168059110641,
-0.006156572140753269,
0.1449710875749588,
0.028103826567530632,
0.029669636860489845,
-0.0018948549404740334,
-0.005003341939300299,
0.05121048167347908,
0.1746254414319992,
-0.011592294089496136,
-0.22072425484657288,
-0.0845772922039032,
0.06936841458082199,
-0.06218599155545235,
-0.12968985736370087,
-0.026130788028240204,
0.045467354357242584,
-0.17519839107990265,
0.026703642681241035,
-0.027433741837739944,
0.0919293761253357,
-0.09345759451389313,
-0.02221956104040146,
0.03687324374914169,
0.084866963326931,
-0.014529162086546421,
0.08703910559415817,
-0.14498743414878845,
0.11886418610811234,
0.02978132851421833,
0.09024628251791,
-0.11081171780824661,
0.07909037172794342,
-0.007550720125436783,
0.009180475026369095,
0.19379350543022156,
-0.011335089802742004,
-0.03514958545565605,
-0.08774717897176743,
-0.11210042238235474,
-0.013537433929741383,
0.12687496840953827,
-0.1243172138929367,
0.08773399889469147,
-0.015198243781924248,
-0.044079482555389404,
0.00937260314822197,
-0.12100647389888763,
-0.17273177206516266,
-0.19628387689590454,
0.05585884302854538,
-0.09575839340686798,
0.025643249973654747,
-0.11914430558681488,
-0.07089093327522278,
-0.02952558360993862,
0.241120383143425,
-0.1745356321334839,
-0.06510113179683685,
-0.1468164622783661,
-0.046294767409563065,
0.1662203073501587,
-0.04437198117375374,
0.0718095526099205,
-0.0208172257989645,
0.20345525443553925,
0.005988610442727804,
-0.004939318168908358,
0.06724198162555695,
-0.08892562240362167,
-0.16873881220817566,
-0.06771010160446167,
0.1510489284992218,
0.11680185794830322,
0.04907919466495514,
-0.002248800592496991,
0.0011772146681323647,
-0.016943959519267082,
-0.1137804463505745,
-0.0033210667315870523,
0.16037839651107788,
0.03878779336810112,
0.025986969470977783,
-0.05243593826889992,
-0.08797456324100494,
-0.06899320334196091,
-0.06853509694337845,
0.06221301481127739,
0.19590823352336884,
-0.10376439243555069,
0.1700313836336136,
0.147536963224411,
-0.07305635511875153,
-0.23175598680973053,
0.035342130810022354,
0.04983805492520332,
0.0014306638622656465,
0.04886869341135025,
-0.18252557516098022,
0.10521943867206573,
0.019543392583727837,
-0.05505957826972008,
0.13485197722911835,
-0.1557481735944748,
-0.1552847921848297,
0.0722852572798729,
0.03904085233807564,
-0.22423844039440155,
-0.1354004591703415,
-0.09622503817081451,
-0.05825018882751465,
-0.14065024256706238,
0.06054598465561867,
-0.002136280992999673,
0.015948504209518433,
0.03500790148973465,
-0.0015643214574083686,
0.027123261243104935,
-0.058935679495334625,
0.18609118461608887,
-0.004065449349582195,
0.020676052197813988,
-0.060264769941568375,
-0.0478842556476593,
0.09839435666799545,
-0.06130504235625267,
0.12208222597837448,
0.004057085141539574,
0.01594383642077446,
-0.10362856835126877,
-0.048314861953258514,
-0.04328322783112526,
0.05154227837920189,
-0.07548051327466965,
-0.10070807486772537,
-0.043625857681035995,
0.08841723203659058,
0.07005169242620468,
-0.03383097052574158,
0.00549331633374095,
-0.07189501076936722,
0.10019614547491074,
0.17795267701148987,
0.17573626339435577,
0.009926567785441875,
-0.07241068035364151,
0.01677953451871872,
-0.04142116755247116,
0.044231921434402466,
-0.2513144314289093,
0.03756171092391014,
0.06098250672221184,
0.029438555240631104,
0.09217222779989243,
-0.020435843616724014,
-0.1820858269929886,
-0.04050002992153168,
0.08094815909862518,
-0.05452597141265869,
-0.22617179155349731,
-0.019085140898823738,
0.0954197570681572,
-0.2020406424999237,
-0.007372708059847355,
0.03995226323604584,
-0.048725228756666183,
-0.023169852793216705,
0.00010950004070764408,
0.06317184865474701,
0.002471912419423461,
0.09773622453212738,
0.0735151618719101,
0.09715340286493301,
-0.08337292820215225,
0.10562895983457565,
0.10150538384914398,
-0.09572599828243256,
0.03605884686112404,
0.06754924356937408,
-0.05300498008728027,
-0.043293699622154236,
0.03665391728281975,
0.033023297786712646,
0.005234600510448217,
-0.060321882367134094,
0.013913018628954887,
-0.036497246474027634,
0.044923391193151474,
0.08326134830713272,
0.03754979372024536,
-0.013354414142668247,
0.06462216377258301,
0.03401726484298706,
-0.10898099094629288,
0.10366570204496384,
0.01731540448963642,
0.04105307161808014,
-0.08384523540735245,
-0.019968897104263306,
0.035425446927547455,
0.030576206743717194,
-0.01765924133360386,
-0.02306121215224266,
-0.02860277332365513,
-0.01614218018949032,
-0.14299540221691132,
-0.023106401786208153,
-0.07243485748767853,
0.006181265693157911,
0.014656842686235905,
-0.031884219497442245,
-0.011233693920075893,
0.02475680410861969,
-0.06979699432849884,
-0.07426341623067856,
-0.006949664559215307,
0.09833318740129471,
-0.15115703642368317,
0.008848577737808228,
0.06907843053340912,
-0.11088496446609497,
0.08190931379795074,
-0.008411259390413761,
0.016245156526565552,
0.022527478635311127,
-0.15448406338691711,
0.05601610988378525,
0.0008648968650959432,
0.01916889287531376,
0.025886621326208115,
-0.16471809148788452,
0.004104440100491047,
-0.04661374166607857,
-0.02149827405810356,
-0.00004464812809601426,
-0.02647159807384014,
-0.12325995415449142,
0.06858719140291214,
-0.015622655861079693,
-0.035931166261434555,
-0.02701525390148163,
0.0539589487016201,
0.07888586074113846,
-0.027474910020828247,
0.10445091128349304,
-0.008690856397151947,
0.04941811040043831,
-0.16801609098911285,
-0.02470702864229679,
-0.04982255399227142,
0.019377702847123146,
0.009884213097393513,
-0.007693959400057793,
0.04183054715394974,
-0.00976533442735672,
0.21883612871170044,
-0.05075952783226967,
0.1607085019350052,
0.05847611650824547,
-0.017352959141135216,
-0.0007513365126214921,
0.06180921941995621,
0.05997028574347496,
0.04658793285489082,
0.009480604901909828,
0.023740366101264954,
-0.022450892254710197,
-0.006695089396089315,
-0.15932634472846985,
0.01890849508345127,
0.14999441802501678,
0.06301083415746689,
0.024745315313339233,
0.05866100639104843,
-0.12775006890296936,
-0.12135478109121323,
0.09311001747846603,
-0.026755332946777344,
0.00928465835750103,
-0.08245618641376495,
0.1358020007610321,
0.14980104565620422,
-0.14000412821769714,
0.05256148427724838,
-0.06134212389588356,
-0.05217423290014267,
-0.10388828068971634,
-0.12032219022512436,
-0.05887215584516525,
-0.053666237741708755,
0.002330566756427288,
-0.03760887682437897,
0.054546963423490524,
0.03344334661960602,
-0.009351172484457493,
-0.00022941511997487396,
0.13597318530082703,
-0.019751882180571556,
-0.0028988157864660025,
0.048313532024621964,
0.03693558648228645,
0.02373051457107067,
-0.05275435373187065,
0.02940409444272518,
0.02539868652820587,
0.032232340425252914,
0.06546790152788162,
0.033412106335163116,
-0.047448933124542236,
0.03804153576493263,
-0.0025254099164158106,
-0.11207924783229828,
0.019641218706965446,
-0.00460948096588254,
-0.0742158442735672,
0.1268945336341858,
0.0407399944961071,
0.010224059224128723,
-0.03741471841931343,
0.24361543357372284,
-0.06653323769569397,
-0.06378097087144852,
-0.13251738250255585,
0.10491154342889786,
-0.0027236645109951496,
0.06476365029811859,
0.023412218317389488,
-0.1284150779247284,
0.005243356805294752,
0.13858191668987274,
0.12181595712900162,
0.0045748427510261536,
0.009228081442415714,
0.0518609918653965,
0.0025186820421367884,
-0.06998204439878464,
0.054019294679164886,
0.06992026418447495,
0.12919506430625916,
-0.07847554981708527,
0.07680778950452805,
0.0006860480643808842,
-0.08370215445756912,
-0.02947772853076458,
0.11312682181596756,
-0.0409729965031147,
0.03491825982928276,
-0.047444481402635574,
0.10916327685117722,
-0.05787910893559456,
-0.29412412643432617,
0.02350960113108158,
-0.09588567912578583,
-0.15202060341835022,
-0.018367812037467957,
0.05944539234042168,
-0.02624768204987049,
0.018029648810625076,
0.06971040368080139,
-0.06011629104614258,
0.20098382234573364,
0.0335683599114418,
-0.07864278554916382,
-0.0664360448718071,
0.04837050288915634,
-0.06564252078533173,
0.2949807047843933,
0.008418165147304535,
0.02863333560526371,
0.10770907253026962,
-0.03253700211644173,
-0.18271861970424652,
0.010723991319537163,
0.1133992001414299,
-0.08056149631738663,
0.08200647681951523,
0.19000613689422607,
-0.012578671798110008,
0.1209007054567337,
0.05294662341475487,
-0.047376248985528946,
0.04217283055186272,
-0.03389401361346245,
-0.051268599927425385,
-0.10752558708190918,
0.058453381061553955,
-0.05909625440835953,
0.15447644889354706,
0.10152646154165268,
-0.05671518296003342,
-0.004550917539745569,
-0.05555408447980881,
0.04875178262591362,
0.01804669201374054,
0.12263146042823792,
0.02951994352042675,
-0.1865430772304535,
0.032826557755470276,
-0.01144319772720337,
0.10186848044395447,
-0.25588861107826233,
-0.08421015739440918,
0.08833149075508118,
-0.011924264021217823,
-0.05105875805020332,
0.10560628771781921,
0.057650718837976456,
0.04243382066488266,
-0.043439045548439026,
-0.10480839014053345,
-0.02186836116015911,
0.14663739502429962,
-0.1469624787569046,
-0.025013303384184837
] |
null | null | transformers | <div align="center">
<h1 style="margin-bottom: 0.5em;">WebLINX: Real-World Website Navigation with Multi-Turn Dialogue</h1>
<em>Xing Han Lù*, Zdeněk Kasner*, Siva Reddy</em>
</div>
<div style="margin-bottom: 2em"></div>
<div style="display: flex; justify-content: space-around; align-items: center; font-size: 120%;">
<div><a href="https://mcgill-nlp.github.io/weblinx">🌐Website</a></div>
<div><a href="https://huggingface.co/spaces/McGill-NLP/weblinx-explorer">💻Explorer</a></div>
<div><a href="https://huggingface.co/datasets/McGill-NLP/WebLINX">🤗Dataset</a></div>
<div><a href="https://github.com/McGill-NLP/weblinx">💾Code</a></div>
</div>
## Original Model
This model is finetuned on WebLINX using checkpoints previously published on Huggingface Hub.\
[Click here to access the original model.](https://huggingface.co/google/flan-t5-base) | {"language": ["en"], "license": "apache-2.0", "library_name": "transformers", "tags": ["weblinx", "text-generation-inference", "web-agents", "agents"], "datasets": ["McGill-NLP/WebLINX", "McGill-NLP/WebLINX-full"], "metrics": ["f1", "iou", "chrf"], "pipeline_tag": "text-generation"} | text-generation | McGill-NLP/flan-t5-base-weblinx | [
"transformers",
"safetensors",
"weblinx",
"text-generation-inference",
"web-agents",
"agents",
"text-generation",
"en",
"dataset:McGill-NLP/WebLINX",
"dataset:McGill-NLP/WebLINX-full",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-07T08:55:33+00:00 | [] | [
"en"
] | TAGS
#transformers #safetensors #weblinx #text-generation-inference #web-agents #agents #text-generation #en #dataset-McGill-NLP/WebLINX #dataset-McGill-NLP/WebLINX-full #license-apache-2.0 #endpoints_compatible #region-us
| <div align="center">
<h1 style="margin-bottom: 0.5em;">WebLINX: Real-World Website Navigation with Multi-Turn Dialogue</h1>
<em>Xing Han Lù*, Zdeněk Kasner*, Siva Reddy</em>
</div>
<div style="margin-bottom: 2em"></div>
<div style="display: flex; justify-content: space-around; align-items: center; font-size: 120%;">
<div><a href="URL
<div><a href="URL
<div><a href="URL
<div><a href="URL
</div>
## Original Model
This model is finetuned on WebLINX using checkpoints previously published on Huggingface Hub.\
Click here to access the original model. | [
"## Original Model\n\nThis model is finetuned on WebLINX using checkpoints previously published on Huggingface Hub.\\\nClick here to access the original model."
] | [
"TAGS\n#transformers #safetensors #weblinx #text-generation-inference #web-agents #agents #text-generation #en #dataset-McGill-NLP/WebLINX #dataset-McGill-NLP/WebLINX-full #license-apache-2.0 #endpoints_compatible #region-us \n",
"## Original Model\n\nThis model is finetuned on WebLINX using checkpoints previously published on Huggingface Hub.\\\nClick here to access the original model."
] | [
90,
34
] | [
"passage: TAGS\n#transformers #safetensors #weblinx #text-generation-inference #web-agents #agents #text-generation #en #dataset-McGill-NLP/WebLINX #dataset-McGill-NLP/WebLINX-full #license-apache-2.0 #endpoints_compatible #region-us \n## Original Model\n\nThis model is finetuned on WebLINX using checkpoints previously published on Huggingface Hub.\\\nClick here to access the original model."
] | [
-0.021222325041890144,
0.07018165290355682,
-0.0018198562320321798,
0.011584840714931488,
-0.00456819636747241,
0.0031534789595752954,
0.1586115062236786,
0.04285372793674469,
0.0366867333650589,
-0.10026367008686066,
0.10517100989818573,
0.023754091933369637,
0.009436704218387604,
0.14679713547229767,
0.023823363706469536,
-0.08705909550189972,
0.047060608863830566,
0.024446845054626465,
0.0262104794383049,
0.08939940482378006,
0.09428824484348297,
-0.07575423270463943,
0.10590092837810516,
0.025137441232800484,
-0.044859301298856735,
0.012014728039503098,
0.026307363063097,
-0.015513942576944828,
0.06563590466976166,
0.06281830370426178,
0.07642584294080734,
0.03671393543481827,
0.08641767501831055,
-0.22881801426410675,
0.03984428197145462,
0.039626698940992355,
-0.011783752590417862,
0.0465284138917923,
-0.011927819810807705,
0.018792670220136642,
0.13245056569576263,
-0.02118520438671112,
-0.04019766300916672,
0.050096672028303146,
-0.018715020269155502,
0.02749360166490078,
-0.08490145206451416,
0.11088118702173233,
0.06900694221258163,
0.06804034113883972,
0.013776840642094612,
0.22806023061275482,
-0.033292315900325775,
0.13198988139629364,
0.10730588436126709,
-0.28150469064712524,
-0.02796829864382744,
0.2222190946340561,
0.07488563656806946,
-0.015589375048875809,
-0.04594384878873825,
0.11973663419485092,
0.054866187274456024,
-0.016530262306332588,
0.052385229617357254,
-0.09919256716966629,
-0.129373237490654,
0.032070308923721313,
-0.07702560722827911,
0.01497510727494955,
0.2163708657026291,
0.08343938738107681,
-0.029929954558610916,
-0.11313756555318832,
-0.10875281691551208,
0.11765293776988983,
-0.07557526975870132,
-0.004465811420232058,
0.08772432804107666,
0.04929497092962265,
-0.010114304721355438,
-0.14088551700115204,
-0.07698670029640198,
-0.013745326548814774,
-0.10702376067638397,
0.1190243512392044,
-0.012912278063595295,
0.08465785533189774,
-0.16578027606010437,
0.03947996348142624,
0.05897917598485947,
-0.06625248491764069,
0.021823784336447716,
-0.08956588059663773,
0.11559700965881348,
0.007644249126315117,
-0.04396217316389084,
0.04142085462808609,
0.15981990098953247,
0.10195491462945938,
0.009195408783853054,
-0.015774628147482872,
-0.09348204731941223,
0.0632748156785965,
0.014703458175063133,
0.09117889404296875,
-0.035583145916461945,
-0.14930367469787598,
0.11991146206855774,
-0.011976223438978195,
0.10074775665998459,
0.0020497487857937813,
-0.026411041617393494,
0.0639912337064743,
-0.027416951954364777,
0.03936413675546646,
0.08410236239433289,
0.12314005941152573,
-0.004230388905853033,
-0.004897222388535738,
0.10844048857688904,
-0.09645361453294754,
0.03744376823306084,
0.02106134593486786,
-0.01020008698105812,
0.02446035109460354,
0.1326165646314621,
0.06099997088313103,
-0.0596146397292614,
-0.11885082721710205,
-0.08408939838409424,
-0.061176832765340805,
0.01599116064608097,
-0.06801082193851471,
0.044001027941703796,
-0.02112687937915325,
0.0672977864742279,
-0.10112014412879944,
-0.22565525770187378,
-0.020518753677606583,
0.0821908563375473,
0.05841250717639923,
-0.0485026091337204,
-0.015015936456620693,
-0.0050012716092169285,
-0.011609560810029507,
-0.024520374834537506,
0.01366529893130064,
-0.06032336503267288,
-0.04664697125554085,
-0.06677721440792084,
0.005788411945104599,
-0.07878222316503525,
-0.03592332452535629,
-0.10100644081830978,
-0.02055244892835617,
-0.06730698049068451,
-0.011986675672233105,
-0.0693119540810585,
0.19336189329624176,
-0.05022754147648811,
0.04727478697896004,
0.017664793878793716,
0.06133991852402687,
-0.028018906712532043,
0.16067776083946228,
-0.003275224007666111,
-0.03430666774511337,
0.21620917320251465,
-0.10932547599077225,
-0.22601541876792908,
0.030936099588871002,
-0.02498202584683895,
0.13402588665485382,
0.133483424782753,
0.15982837975025177,
0.10379461944103241,
-0.16877636313438416,
0.036182012408971786,
0.067097008228302,
-0.01976250484585762,
-0.04981627315282822,
-0.07149284332990646,
-0.021241862326860428,
-0.12747034430503845,
0.036388099193573,
-0.13334858417510986,
0.031706903129816055,
-0.016690341755747795,
-0.050568584352731705,
-0.07249800860881805,
-0.09283380210399628,
0.014389454387128353,
-0.04472890496253967,
-0.0017485248390585184,
-0.04355271905660629,
-0.026862308382987976,
-0.0959537923336029,
0.09745708107948303,
-0.04212859645485878,
0.06755275279283524,
-0.09586205333471298,
-0.10331743210554123,
0.025516627356410027,
0.0729532316327095,
-0.07936501502990723,
-0.10062122344970703,
-0.0009191471617668867,
0.05684973672032356,
-0.0007016055751591921,
0.04228636622428894,
0.05182516947388649,
-0.04495089501142502,
-0.009067459963262081,
-0.006481003016233444,
0.04751013591885567,
0.021753260865807533,
-0.03205421194434166,
-0.19062794744968414,
0.016992490738630295,
-0.03937038406729698,
0.022065047174692154,
-0.13701850175857544,
0.030590295791625977,
-0.033762332051992416,
0.08472217619419098,
0.028971778228878975,
0.058795295655727386,
0.034899529069662094,
-0.06274329125881195,
-0.04864140972495079,
-0.03266078978776932,
0.0635925754904747,
0.04255114495754242,
-0.14884540438652039,
0.13075648248195648,
0.0062464457005262375,
0.051126811653375626,
0.15769068896770477,
-0.0351782850921154,
0.09270644187927246,
0.013313611038029194,
-0.06305912882089615,
0.00019239087123423815,
0.04353025183081627,
-0.00021059811115264893,
-0.007918101735413074,
0.03296810761094093,
0.10076507180929184,
-0.06802044063806534,
-0.03117506019771099,
-0.02146143838763237,
-0.07282408326864243,
-0.02314714528620243,
-0.011174840852618217,
0.021499622613191605,
-0.05757316201925278,
0.0008018943481147289,
0.1247420459985733,
0.05104173719882965,
0.047709301114082336,
-0.05348343774676323,
0.029417701065540314,
-0.004684962332248688,
-0.025976937264204025,
-0.0445069819688797,
0.09168009459972382,
-0.07069297134876251,
-0.043796829879283905,
0.03957461193203926,
0.010344958864152431,
0.07508863508701324,
-0.08929906785488129,
-0.00011763651855289936,
0.02274271287024021,
-0.060247402638196945,
0.0062419697642326355,
-0.011824898421764374,
-0.09721781313419342,
0.06804082542657852,
-0.08590628206729889,
-0.057093165814876556,
-0.027055274695158005,
-0.037185076624155045,
-0.13416093587875366,
0.11281641572713852,
-0.08489511162042618,
-0.18041014671325684,
-0.10271430015563965,
-0.19421975314617157,
-0.18108776211738586,
-0.01026380155235529,
0.00925499014556408,
-0.03826535493135452,
-0.0702395960688591,
-0.14804495871067047,
-0.08571991324424744,
0.07777160406112671,
0.0054359715431928635,
0.1893797218799591,
0.0450282096862793,
0.029103733599185944,
-0.16087709367275238,
-0.02560836635529995,
-0.024707093834877014,
-0.02072209119796753,
0.01564621552824974,
-0.06359142065048218,
0.14722177386283875,
0.10005506873130798,
0.03072010539472103,
0.003934965468943119,
0.001979321241378784,
0.1818770170211792,
-0.035957012325525284,
0.10879883170127869,
0.15925812721252441,
-0.012021586298942566,
0.013737115077674389,
0.11198879033327103,
0.02509799599647522,
-0.0725109875202179,
0.05111286789178848,
-0.000617223500739783,
-0.006932103540748358,
-0.20869560539722443,
-0.09724225103855133,
-0.015195057727396488,
0.09856502711772919,
0.023766815662384033,
0.06117073819041252,
0.06241131201386452,
0.06429912149906158,
-0.05388019233942032,
0.01598336361348629,
0.1261700838804245,
0.03538059443235397,
0.03572783246636391,
-0.061635296791791916,
0.05197034403681755,
-0.10403181612491608,
-0.022695468738675117,
0.10684707760810852,
0.10509435087442398,
0.11508216708898544,
0.10000845789909363,
0.07447698712348938,
0.1406693458557129,
0.030831726267933846,
-0.002587369177490473,
0.13470420241355896,
0.0202591884881258,
-0.03626542165875435,
-0.01027034129947424,
-0.10038577020168304,
-0.005433962680399418,
0.026075471192598343,
-0.1648518443107605,
-0.03254007548093796,
-0.03604228049516678,
0.06351303309202194,
0.10315615683794022,
0.12571603059768677,
0.05818505585193634,
-0.1533859372138977,
0.01257017720490694,
0.09130007028579712,
0.030559850856661797,
0.011280844919383526,
0.06489631533622742,
-0.00556802935898304,
-0.05376901477575302,
0.17005997896194458,
0.004884402267634869,
0.14335228502750397,
0.026836954057216644,
0.028540916740894318,
-0.04833287373185158,
-0.02866477519273758,
0.01593007706105709,
0.05851345509290695,
-0.1702987104654312,
0.14926832914352417,
0.05112772807478905,
0.06570970267057419,
-0.050368782132864,
0.02523258700966835,
0.06146357208490372,
0.24696093797683716,
0.138278529047966,
0.038426633924245834,
-0.03957507759332657,
0.10467124730348587,
-0.12302936613559723,
0.0650452971458435,
-0.021542366594076157,
0.03846564143896103,
0.10389277338981628,
-0.056906018406152725,
-0.036995165050029755,
0.036654986441135406,
0.0507313497364521,
-0.25493234395980835,
-0.10709584504365921,
-0.06918162107467651,
0.19051475822925568,
-0.11054514348506927,
-0.08091512322425842,
0.031927868723869324,
-0.0527162067592144,
0.2747766971588135,
-0.03581225126981735,
-0.05068189278244972,
-0.09558023512363434,
-0.047656018286943436,
-0.0010705795139074326,
-0.013308942317962646,
-0.01965280808508396,
-0.023864062502980232,
0.14187933504581451,
-0.0862899199128151,
-0.16465696692466736,
-0.005532992072403431,
-0.16473126411437988,
-0.024282246828079224,
-0.03308350220322609,
0.08017432689666748,
-0.012208091095089912,
-0.03781270608305931,
0.08880128711462021,
-0.03735820949077606,
-0.0980556383728981,
-0.14364877343177795,
-0.01948181726038456,
0.13121548295021057,
-0.02096978947520256,
0.012720013968646526,
-0.14664193987846375,
-0.07468709349632263,
-0.012798362411558628,
0.0014694794081151485,
0.09753498435020447,
0.14239835739135742,
-0.03164052963256836,
0.09343600273132324,
0.21911007165908813,
-0.1046786978840828,
-0.3032979965209961,
-0.09226939082145691,
-0.1707906574010849,
-0.040876902639865875,
0.048727311193943024,
-0.04271368309855461,
0.17153674364089966,
-0.07192384451627731,
-0.059105418622493744,
0.05619104206562042,
-0.21674640476703644,
-0.09094619750976562,
0.11686206609010696,
0.06493991613388062,
0.21786201000213623,
-0.09830866754055023,
-0.05215728282928467,
-0.11257259547710419,
-0.21442162990570068,
0.029232928529381752,
-0.3093061149120331,
-0.004350990056991577,
0.03770551458001137,
0.046685200184583664,
-0.018288735300302505,
-0.04972011595964432,
0.056664228439331055,
-0.012483600527048111,
0.07441719621419907,
-0.08367584645748138,
0.0566365011036396,
0.1850031167268753,
-0.05931615084409714,
0.1424839347600937,
-0.12697073817253113,
0.07234383374452591,
-0.07195829600095749,
-0.024641335010528564,
-0.05058017373085022,
0.08414611965417862,
-0.07608246803283691,
-0.024027325212955475,
-0.08571477234363556,
-0.015754589810967445,
0.08628059178590775,
0.04715639725327492,
0.12759022414684296,
0.0330161452293396,
0.028280138969421387,
0.20718331634998322,
0.07353273779153824,
-0.17821501195430756,
0.05639011412858963,
-0.01951584406197071,
-0.08524004369974136,
0.038989827036857605,
-0.15679100155830383,
0.02707493305206299,
0.047213394194841385,
-0.03291246294975281,
0.020450035110116005,
0.029398053884506226,
-0.02599427103996277,
-0.13382157683372498,
0.08970662951469421,
-0.1747797429561615,
-0.06832166016101837,
-0.03901377320289612,
-0.03201857954263687,
-0.053630977869033813,
0.1284835934638977,
0.2346244901418686,
-0.02621556632220745,
0.004522768780589104,
-0.031481362879276276,
0.07174394279718399,
-0.028820239007472992,
0.013172878883779049,
0.05024237930774689,
-0.0026898602955043316,
-0.11423295736312866,
0.15964296460151672,
0.022389573976397514,
0.03348381444811821,
-0.019494079053401947,
-0.030142448842525482,
-0.1412464678287506,
-0.0705697238445282,
0.004932940471917391,
0.1864844113588333,
-0.03538583219051361,
-0.08880989998579025,
-0.0719086229801178,
-0.12407853454351425,
0.0015264463145285845,
0.0025065140798687935,
0.04013589769601822,
0.08450248837471008,
0.029750466346740723,
-0.06139592081308365,
-0.043277807533741,
0.0705777108669281,
-0.050844185054302216,
-0.01942811906337738,
-0.14937575161457062,
0.003988961223512888,
0.016172926872968674,
0.05235257372260094,
-0.045256108045578,
0.03973129391670227,
-0.04922090843319893,
0.00026686559431254864,
-0.12990231812000275,
0.015771374106407166,
-0.09828590601682663,
0.0028939296025782824,
-0.04298677667975426,
-0.019613152369856834,
-0.10882392525672913,
0.040934156626462936,
-0.03599356859922409,
-0.01644737273454666,
-0.004150992725044489,
0.03850296884775162,
-0.1394466906785965,
0.010179729200899601,
0.03151402249932289,
-0.0450744703412056,
0.05739281699061394,
-0.05193808302283287,
-0.07376417517662048,
0.07081639766693115,
-0.13791243731975555,
-0.02838098630309105,
0.031110772863030434,
0.005903621204197407,
-0.004649123642593622,
-0.03411435708403587,
-0.04364749416708946,
0.0979636013507843,
-0.02518753707408905,
-0.001783302053809166,
0.021325252950191498,
-0.048875629901885986,
-0.07231947034597397,
-0.012374013662338257,
0.011719616129994392,
-0.011605862528085709,
-0.03014637902379036,
0.13525798916816711,
0.03063948079943657,
0.1384371668100357,
-0.04758249595761299,
-0.0224443431943655,
-0.18430857360363007,
0.046028099954128265,
0.004548339638859034,
-0.09202136844396591,
-0.07830039411783218,
-0.05817829445004463,
-0.029256552457809448,
-0.03720324859023094,
0.3151671290397644,
-0.017357656732201576,
-0.023893184959888458,
0.021511942148208618,
-0.01804506592452526,
-0.047220341861248016,
0.003968787379562855,
0.2805250883102417,
-0.014079729095101357,
-0.008546270430088043,
0.014229124411940575,
-0.025703568011522293,
0.11244373768568039,
-0.10052136331796646,
0.11276477575302124,
0.05737374350428581,
0.08472523093223572,
0.15293672680854797,
0.029234062880277634,
0.011493267491459846,
0.04284721612930298,
-0.10917704552412033,
-0.042764633893966675,
0.08347485959529877,
-0.03154876083135605,
0.029606513679027557,
0.17192400991916656,
-0.05654463544487953,
0.012663762085139751,
0.027814166620373726,
-0.03526335582137108,
-0.15126417577266693,
-0.10093432664871216,
-0.09247830510139465,
-0.11214568465948105,
-0.011395438574254513,
-0.15026496350765228,
-0.04601219296455383,
-0.13334836065769196,
0.000010143965482711792,
-0.08159323036670685,
0.02406761609017849,
-0.0905282273888588,
-0.045641280710697174,
0.07043003290891647,
-0.006517861038446426,
-0.08320179581642151,
0.0510505773127079,
-0.04478387162089348,
-0.010366321541368961,
0.05876917392015457,
0.008053512312471867,
0.05491524562239647,
0.052445799112319946,
0.14528487622737885,
-0.03704366087913513,
-0.03537627309560776,
-0.07466281950473785,
-0.02511107735335827,
0.05248693376779556,
0.028590315952897072,
0.06409184634685516,
-0.03513815999031067,
0.04741821438074112,
0.20212966203689575,
-0.03610433638095856,
-0.12098394334316254,
-0.08444462716579437,
-0.032174378633499146,
0.020877333357930183,
0.04930546507239342,
-0.05449466034770012,
-0.03556454926729202,
-0.015534033998847008,
0.3807089626789093,
0.28133973479270935,
-0.09547242522239685,
0.029074009507894516,
-0.08150309324264526,
0.005202847067266703,
0.048969168215990067,
0.10301991552114487,
0.05144472420215607,
0.0965375304222107,
0.03951612859964371,
-0.04297666251659393,
-0.023892246186733246,
-0.06230989098548889,
-0.09634049981832504,
0.06814974546432495,
0.019922303035855293,
-0.06199240684509277,
-0.031735967844724655,
0.07014572620391846,
-0.04402405768632889,
0.010483757592737675,
-0.05891045555472374,
-0.11636233329772949,
0.005337890237569809,
-0.024663031101226807,
0.12695123255252838,
0.046970997005701065,
0.06272334605455399,
-0.018004007637500763,
0.038006022572517395,
0.1519189178943634,
-0.06598782539367676,
-0.10804281383752823,
-0.02192622981965542,
0.03698035702109337,
-0.08098472654819489,
0.2605289816856384,
0.026691382750868797,
0.005204369314014912,
0.06815899908542633,
-0.09357991069555283,
-0.15867705643177032,
0.0820537880063057,
-0.03466859832406044,
-0.00796995498239994,
0.016421107575297356,
-0.09637746214866638,
-0.03417760878801346,
-0.08518568426370621,
0.06972041726112366,
0.0031114043667912483,
0.03102630190551281,
0.12034516036510468,
-0.06934475153684616,
-0.05905064195394516,
0.022118467837572098,
-0.10718359798192978,
0.09798767417669296,
0.048797376453876495,
-0.08364203572273254,
0.061209410429000854,
-0.015863724052906036,
0.0569218173623085,
-0.00448411051183939,
0.002492828294634819,
0.04848877713084221,
-0.07891902327537537,
0.003780954983085394,
0.021189680323004723,
0.026127371937036514,
-0.22183875739574432,
-0.023713208734989166,
-0.07686010003089905,
-0.02341688796877861,
-0.07559123635292053,
0.10296734422445297,
0.20741550624370575,
0.006159932818263769,
-0.011555604636669159,
-0.06918633729219437,
0.06238970533013344,
0.06050347909331322,
-0.021712293848395348,
-0.1280355453491211
] |
null | null | null |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# roberta-base-bn-adapter-895K-snli-model3
This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.7710
- Accuracy: 0.7275
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|
| 0.4273 | 1.0 | 8584 | 0.3416 | 0.8694 |
| 0.4019 | 2.0 | 17168 | 0.3206 | 0.8800 |
| 0.385 | 3.0 | 25752 | 0.3148 | 0.8821 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| {"license": "mit", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "roberta-base", "model-index": [{"name": "roberta-base-bn-adapter-895K-snli-model3", "results": []}]} | null | varun-v-rao/roberta-base-bn-adapter-895K-snli-model3 | [
"tensorboard",
"generated_from_trainer",
"base_model:roberta-base",
"license:mit",
"region:us"
] | 2024-02-07T08:57:02+00:00 | [] | [] | TAGS
#tensorboard #generated_from_trainer #base_model-roberta-base #license-mit #region-us
| roberta-base-bn-adapter-895K-snli-model3
========================================
This model is a fine-tuned version of roberta-base on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.7710
* Accuracy: 0.7275
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 64
* eval\_batch\_size: 64
* seed: 16
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.1+cu121
* Datasets 2.15.0
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
"TAGS\n#tensorboard #generated_from_trainer #base_model-roberta-base #license-mit #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
31,
98,
4,
33
] | [
"passage: TAGS\n#tensorboard #generated_from_trainer #base_model-roberta-base #license-mit #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
-0.07550864666700363,
-0.02109025977551937,
-0.001131002907641232,
0.11642087250947952,
0.19986377656459808,
0.032519303262233734,
0.1489553153514862,
0.06593390554189682,
-0.12863241136074066,
0.013592129573225975,
0.10571130365133286,
0.13225753605365753,
-0.011063135229051113,
0.0559028722345829,
-0.021351991221308708,
-0.2589706778526306,
-0.030275097116827965,
0.03432689607143402,
-0.14379724860191345,
0.12667573988437653,
0.08392859995365143,
-0.14512383937835693,
0.06996306031942368,
-0.00019704063015524298,
-0.3233215808868408,
0.0525614432990551,
0.034350328147411346,
-0.041326943784952164,
0.14337140321731567,
-0.007632463239133358,
0.19256934523582458,
-0.01464041043072939,
0.1317151039838791,
-0.12744224071502686,
0.013083022087812424,
0.06882568448781967,
0.00157072430010885,
0.08348285406827927,
0.05815310776233673,
0.02549658715724945,
0.11559391021728516,
-0.08406563848257065,
0.06291346251964569,
0.01876154914498329,
-0.14418642222881317,
-0.24435856938362122,
-0.06911272555589676,
-0.027874978259205818,
0.054466504603624344,
0.08646325021982193,
-0.010559392161667347,
0.15109072625637054,
-0.09367001801729202,
0.09661748260259628,
0.27046769857406616,
-0.2486611306667328,
-0.08295661211013794,
0.08798392862081528,
0.004633144475519657,
0.09485544264316559,
-0.10890258103609085,
-0.027921518310904503,
0.07303660362958908,
0.04299337789416313,
0.12135104835033417,
-0.03207109495997429,
-0.12717418372631073,
0.04465117305517197,
-0.15398608148097992,
0.010136948898434639,
0.06202498450875282,
0.01977911964058876,
-0.017644230276346207,
0.046597521752119064,
-0.08495069295167923,
-0.1390315741300583,
-0.051450569182634354,
-0.020884055644273758,
0.08361329883337021,
-0.051586441695690155,
-0.1292503923177719,
-0.020815633237361908,
-0.10399257391691208,
-0.10278560221195221,
-0.05850876122713089,
0.2166244238615036,
0.05090004578232765,
0.04132528230547905,
-0.06730999052524567,
0.11228986084461212,
-0.057464469224214554,
-0.12991420924663544,
0.055520132184028625,
0.051684074103832245,
-0.023853400722146034,
-0.058301083743572235,
-0.0822465568780899,
-0.09769855439662933,
0.016940565779805183,
0.11436480283737183,
-0.09314920008182526,
0.05121483653783798,
0.05838734284043312,
0.054602961987257004,
-0.10281053185462952,
0.1760760396718979,
-0.05093870311975479,
0.03548604995012283,
0.00856257788836956,
0.025883512571454048,
-0.005448591895401478,
0.0006493262480944395,
-0.11407793313264847,
0.02257198840379715,
0.07918313145637512,
-0.0012724895495921373,
-0.08747687935829163,
0.03732021525502205,
-0.028683977201581,
-0.0029567989986389875,
-0.03093983791768551,
-0.08754921704530716,
0.04795634001493454,
-0.004087560810148716,
-0.09857125580310822,
-0.02051786333322525,
0.00983775407075882,
0.014469374902546406,
0.026179829612374306,
0.13917991518974304,
-0.11040238291025162,
0.08271413296461105,
-0.14512377977371216,
-0.12798883020877838,
-0.004647684283554554,
-0.04271680861711502,
0.02810935489833355,
-0.07241576910018921,
-0.1429770439863205,
-0.014749432913959026,
0.06704434007406235,
-0.039732903242111206,
-0.008968625217676163,
-0.03561616316437721,
-0.0971234068274498,
-0.014576403424143791,
-0.00978782493621111,
0.18687428534030914,
-0.044896200299263,
0.10582200437784195,
0.05998222529888153,
0.06872709095478058,
-0.10610438883304596,
0.04183264076709747,
-0.09147825837135315,
0.0038459959905594587,
-0.26855674386024475,
0.012316184118390083,
-0.05208326131105423,
0.06633296608924866,
-0.06570014357566833,
-0.10110510140657425,
0.02101411484181881,
0.0013867754023522139,
0.11108902841806412,
0.07041280716657639,
-0.18892501294612885,
-0.07767615467309952,
0.12277442216873169,
-0.0741376280784607,
-0.06978735327720642,
0.08797984570264816,
-0.07468701899051666,
0.07292332500219345,
0.06241711974143982,
0.19417384266853333,
-0.03841821476817131,
-0.10302332788705826,
0.045733459293842316,
-0.03745320439338684,
0.03938659653067589,
-0.08580459654331207,
0.03923599421977997,
-0.007947336882352829,
0.011907441541552544,
0.027324210852384567,
-0.05851885303854942,
0.06829018890857697,
-0.1303066462278366,
-0.07780708372592926,
-0.04980482906103134,
-0.09421304613351822,
0.02210817113518715,
0.07947927713394165,
0.09922768175601959,
-0.10347364842891693,
-0.05577545613050461,
0.18948455154895782,
0.05913086608052254,
-0.05539962276816368,
0.03010958805680275,
-0.035071589052677155,
0.04536290839314461,
-0.07244058698415756,
-0.0447392538189888,
-0.19998502731323242,
-0.05148595944046974,
0.00335444463416934,
0.006074106320738792,
0.07261074334383011,
0.0235604140907526,
0.0912807434797287,
0.07697497308254242,
-0.07017992436885834,
0.01532838772982359,
-0.0801808163523674,
-0.003629057900980115,
-0.1375037282705307,
-0.2262304425239563,
-0.0357191264629364,
-0.011200503446161747,
0.07091870903968811,
-0.19740179181098938,
0.028525032103061676,
-0.03743179515004158,
0.07188045233488083,
0.0015480583533644676,
-0.02407841943204403,
-0.07239106297492981,
0.10274705290794373,
0.0034142271615564823,
-0.06592858582735062,
0.048956505954265594,
-0.016135649755597115,
-0.06682923436164856,
-0.09100838750600815,
-0.08756605535745621,
0.17146696150302887,
0.13840974867343903,
-0.16798058152198792,
-0.1040114015340805,
0.03375178948044777,
-0.06636204570531845,
-0.025873206555843353,
-0.05887632071971893,
0.06160994991660118,
0.18531425297260284,
-0.01694081537425518,
0.13704538345336914,
-0.08096588402986526,
-0.040602922439575195,
0.0017230717930942774,
-0.045180775225162506,
0.07892373949289322,
0.09715266525745392,
0.1830386519432068,
-0.05585618317127228,
0.11234797537326813,
0.1410028636455536,
-0.15472880005836487,
0.12476779520511627,
-0.058977968990802765,
-0.07339634001255035,
-0.04228278994560242,
-0.02538219280540943,
-0.009272892028093338,
0.15566666424274445,
-0.07005835324525833,
0.006908639334142208,
-0.012280694209039211,
0.011151635088026524,
0.04197242856025696,
-0.2537667453289032,
-0.05588109791278839,
0.0135591309517622,
-0.0307995043694973,
-0.019067924469709396,
-0.03805839270353317,
0.00832402054220438,
0.1070316731929779,
-0.030410783365368843,
-0.07071337103843689,
0.01513649895787239,
0.016406225040555,
-0.05824301764369011,
0.22600550949573517,
-0.06315040588378906,
-0.008087812922894955,
-0.0992131382226944,
-0.036052219569683075,
-0.04916299507021904,
-0.010021406225860119,
0.03569858521223068,
-0.13822008669376373,
-0.01852429285645485,
-0.03197009861469269,
0.04900304973125458,
-0.0006061159074306488,
0.04685414955019951,
-0.013910239562392235,
0.007013108115643263,
0.07775688916444778,
-0.13761842250823975,
0.01812126114964485,
-0.09283202886581421,
-0.10282741487026215,
0.03892456367611885,
0.07815247029066086,
0.12542681396007538,
0.1697537899017334,
-0.03703921288251877,
0.0014636920532211661,
-0.011279645375907421,
0.2800275981426239,
-0.0651501789689064,
-0.04108003154397011,
0.0935077890753746,
0.0012395186349749565,
0.051565658301115036,
0.0914212316274643,
0.11409729719161987,
-0.11456598341464996,
-0.0056307935155928135,
0.05230887606739998,
-0.058580223470926285,
-0.22592692077159882,
-0.013285146094858646,
-0.025121476501226425,
-0.1000400185585022,
0.0605548731982708,
0.029185745865106583,
0.022640075534582138,
0.05987611785531044,
0.07355909049510956,
0.11231635510921478,
-0.07145293056964874,
0.04474793002009392,
0.05584476515650749,
0.061446450650691986,
0.11309836804866791,
-0.04984627291560173,
-0.08405034989118576,
0.018838932737708092,
-0.039264462888240814,
0.2717016935348511,
0.005533749237656593,
0.11236810684204102,
0.07535698264837265,
0.1756593883037567,
-0.004780292976647615,
0.09199432283639908,
0.003819235134869814,
-0.0657220110297203,
-0.005626327823847532,
-0.050717275589704514,
0.02146393619477749,
0.011906206607818604,
-0.0850776806473732,
0.052678171545267105,
-0.07786818593740463,
-0.01766118034720421,
0.05314239487051964,
0.21039211750030518,
-0.011656413786113262,
-0.30312132835388184,
-0.04479043185710907,
-0.016887523233890533,
0.004373984877020121,
0.0036168957594782114,
-0.0014206022024154663,
0.1373419314622879,
-0.04568326100707054,
-0.004164909012615681,
-0.05951013043522835,
0.08485835790634155,
-0.005977768450975418,
0.03122648410499096,
0.0636831670999527,
0.16554409265518188,
-0.017838889732956886,
0.045105259865522385,
-0.29897207021713257,
0.3051617741584778,
0.025045698508620262,
0.10859643667936325,
-0.038069747388362885,
-0.01942458376288414,
0.01573902554810047,
0.028783496469259262,
0.002096977550536394,
-0.018081119284033775,
-0.0522296316921711,
-0.21492767333984375,
-0.02787875011563301,
0.05264687538146973,
0.13219349086284637,
0.012311290949583054,
0.08685846626758575,
0.00556850666180253,
0.014383437111973763,
0.10391925275325775,
-0.061493851244449615,
-0.10761425644159317,
-0.031459350138902664,
-0.06531272828578949,
0.0020063994452357292,
-0.10851417481899261,
-0.05457620322704315,
-0.10632651299238205,
-0.08750523626804352,
0.1005142405629158,
0.023098856210708618,
-0.029744992032647133,
-0.1245284304022789,
0.13886696100234985,
0.10871507227420807,
-0.06223147734999657,
0.047791987657547,
0.02439885400235653,
0.03357992693781853,
0.038166582584381104,
-0.05824962258338928,
0.12056294828653336,
-0.04336659982800484,
-0.1398981362581253,
-0.0508989542722702,
0.04542393237352371,
0.059007979929447174,
0.0638081356883049,
-0.034122586250305176,
0.022990986704826355,
-0.017306935042142868,
-0.11066395789384842,
0.025859810411930084,
-0.06455598771572113,
0.018920188769698143,
0.00814618356525898,
-0.03342592343688011,
0.05759702995419502,
-0.04227946326136589,
-0.0006616523023694754,
0.12222117185592651,
0.2861289381980896,
-0.07976929843425751,
-0.03070744499564171,
0.04596384987235069,
-0.07246231287717819,
-0.16037894785404205,
0.11448739469051361,
0.08000379800796509,
-0.012185150757431984,
0.07018603384494781,
-0.1761942207813263,
0.15634940564632416,
0.12950482964515686,
-0.009837312623858452,
0.1124584823846817,
-0.32334575057029724,
-0.1528225839138031,
0.07671817392110825,
0.20860011875629425,
0.16020457446575165,
-0.15608854591846466,
-0.017410937696695328,
-0.022307148203253746,
-0.09726567566394806,
0.09857311844825745,
-0.1502445489168167,
0.10691063851118088,
-0.011442489922046661,
0.10677441209554672,
-0.007696425076574087,
-0.0663372129201889,
0.1189674437046051,
0.027462368831038475,
0.15109477937221527,
-0.04031363129615784,
-0.04056025296449661,
0.10764024406671524,
-0.009527083486318588,
-0.009254918433725834,
-0.013190658763051033,
0.012219099327921867,
-0.026475321501493454,
0.0005941743729636073,
-0.08627329766750336,
0.04414021223783493,
-0.03788895159959793,
-0.0702071413397789,
-0.05172790214419365,
-0.0006335040670819581,
-0.012437880039215088,
-0.027670523151755333,
0.12730368971824646,
0.054277531802654266,
0.152330681681633,
0.08782347291707993,
0.002044210908934474,
-0.05205550789833069,
-0.0684737041592598,
0.02442171424627304,
-0.0016480544582009315,
0.07546316087245941,
-0.18616566061973572,
-0.011071358807384968,
0.12553995847702026,
0.043162163347005844,
0.09249671548604965,
0.10062183439731598,
-0.05536704882979393,
0.03344690054655075,
0.06730186939239502,
-0.14802144467830658,
-0.09077725559473038,
0.023064035922288895,
-0.1054510697722435,
-0.06524056196212769,
0.06190616264939308,
0.07662255316972733,
-0.08175504952669144,
-0.0031208896543830633,
-0.011582564562559128,
-0.03487474471330643,
-0.08492483198642731,
0.22570908069610596,
0.11201847344636917,
0.03791264072060585,
-0.11629263311624527,
0.0740446075797081,
0.05700153484940529,
-0.017699744552373886,
-0.00033888532198034227,
0.07352171093225479,
-0.04792601242661476,
0.012857461348176003,
0.15246431529521942,
0.23906481266021729,
-0.10820063948631287,
-0.028120199218392372,
-0.1905844360589981,
-0.08149953931570053,
0.04207250475883484,
0.1905941516160965,
0.1149011105298996,
-0.021036770194768906,
-0.0280312467366457,
0.046782974153757095,
-0.15396730601787567,
0.05300989747047424,
0.011364448815584183,
0.09041189402341843,
-0.11507397890090942,
0.19036448001861572,
0.030607884749770164,
0.019705891609191895,
-0.030755678191781044,
0.05021960288286209,
-0.14802542328834534,
0.030586469918489456,
-0.10444626212120056,
-0.06871069967746735,
0.028789063915610313,
-0.017145507037639618,
0.004017955623567104,
-0.07905267179012299,
-0.07804034650325775,
0.0341891348361969,
-0.14546343684196472,
-0.014317573979496956,
0.05845281481742859,
0.012435534037649632,
-0.14742572605609894,
-0.019154170528054237,
0.02519853599369526,
-0.04390818998217583,
0.02274506352841854,
0.05413439869880676,
0.03580567240715027,
0.09425567090511322,
-0.19691862165927887,
-0.011683349497616291,
0.06711987406015396,
-0.014050502330064774,
0.09334991872310638,
-0.020273329690098763,
-0.027315817773342133,
-0.028326744213700294,
0.1518830507993698,
0.008648039773106575,
0.059131406247615814,
-0.1347287893295288,
0.007541890721768141,
-0.05966641381382942,
-0.09655211120843887,
-0.04321492090821266,
0.003318303730338812,
0.08146313577890396,
0.031112123280763626,
0.15749716758728027,
-0.07034033536911011,
0.027789749205112457,
-0.22940394282341003,
-0.0173785537481308,
-0.005651835352182388,
-0.08188951015472412,
-0.10494983941316605,
-0.04941686615347862,
0.06675970554351807,
-0.05940667912364006,
0.138103649020195,
0.05657614395022392,
0.06302691996097565,
0.036799319088459015,
0.0211014486849308,
0.006800032686442137,
0.030117686837911606,
0.22187648713588715,
0.03778689354658127,
-0.017923401668667793,
0.06871195882558823,
0.09802889823913574,
0.11715026199817657,
0.08468014001846313,
0.21212255954742432,
0.14536523818969727,
-0.05847734585404396,
0.0961160659790039,
0.06449339538812637,
-0.04772188141942024,
-0.1104942038655281,
0.08270055055618286,
-0.021679287776350975,
0.06267012655735016,
-0.032357390969991684,
0.2063753455877304,
0.07751031219959259,
-0.15839849412441254,
0.03900463879108429,
-0.03360336273908615,
-0.10124436020851135,
-0.10237382352352142,
0.041209980845451355,
-0.07759147882461548,
-0.17633894085884094,
0.03443481773138046,
-0.10364523530006409,
0.01877732388675213,
0.15670935809612274,
-0.003455615835264325,
0.007955928333103657,
0.1977098137140274,
0.08703559637069702,
0.055132560431957245,
0.01685400865972042,
-0.00581470038741827,
-0.040563762187957764,
-0.08998230844736099,
-0.07047480344772339,
-0.019951840862631798,
-0.018591569736599922,
0.017796268686652184,
-0.05436999350786209,
-0.09705153107643127,
0.03804456442594528,
-0.00347969145514071,
-0.09958655387163162,
0.021124955266714096,
0.03666799142956734,
0.06798171252012253,
0.00015520757006015629,
-0.0061292825266718864,
0.010410530492663383,
-0.03757196292281151,
0.21886175870895386,
-0.06533658504486084,
-0.09187676012516022,
-0.06769176572561264,
0.23036164045333862,
0.03295457735657692,
-0.003069742117077112,
0.007284162100404501,
-0.08727272599935532,
-0.002403967548161745,
0.19994013011455536,
0.1651201993227005,
-0.1129685491323471,
-0.016912618651986122,
-0.011078300885856152,
-0.027072446420788765,
-0.06521973758935928,
0.15520121157169342,
0.09028005599975586,
0.021373216062784195,
-0.12615787982940674,
-0.05649424344301224,
-0.06970252096652985,
-0.02549314871430397,
-0.051263801753520966,
0.021897003054618835,
0.06983765959739685,
0.006732372101396322,
-0.05628858134150505,
0.07371435314416885,
-0.06575318425893784,
-0.130406454205513,
0.09200982004404068,
-0.17138557136058807,
-0.15997889637947083,
-0.006055848207324743,
0.13578584790229797,
-0.014369463548064232,
0.07888156920671463,
-0.05527856945991516,
0.0019106845138594508,
0.03554607182741165,
-0.043818261474370956,
-0.10068251937627792,
-0.12491434067487717,
0.11041054129600525,
-0.1267392933368683,
0.2003360092639923,
-0.0367429293692112,
0.1129058301448822,
0.10801129043102264,
0.04799626022577286,
-0.06794044375419617,
0.059138935059309006,
0.03367158770561218,
-0.13556797802448273,
-0.012617677450180054,
0.11374297738075256,
-0.03438149392604828,
0.02105984464287758,
0.012859861366450787,
-0.12624198198318481,
0.03741449490189552,
-0.03947976231575012,
-0.04175402596592903,
-0.0531744621694088,
-0.09006880223751068,
-0.07662985473871231,
0.08738873898983002,
0.2039436399936676,
-0.008297373540699482,
0.04527667537331581,
-0.08675803244113922,
0.032988712191581726,
0.08575893938541412,
0.046451885253190994,
-0.08586663752794266,
-0.24494166672229767,
0.02042527124285698,
0.13210885226726532,
-0.07701734453439713,
-0.21933607757091522,
-0.0841103121638298,
0.02690916880965233,
-0.08306260406970978,
-0.07654892653226852,
0.06617989391088486,
0.10721076279878616,
0.0674024447798729,
-0.04264051467180252,
-0.1382921040058136,
-0.09539025276899338,
0.14478352665901184,
-0.14754818379878998,
-0.10086630284786224
] |
null | null | transformers | <div align="center">
<h1 style="margin-bottom: 0.5em;">WebLINX: Real-World Website Navigation with Multi-Turn Dialogue</h1>
<em>Xing Han Lù*, Zdeněk Kasner*, Siva Reddy</em>
</div>
<div style="margin-bottom: 2em"></div>
<div style="display: flex; justify-content: space-around; align-items: center; font-size: 120%;">
<div><a href="https://arxiv.org/abs/2402.05930">📄Paper</a></div>
<div><a href="https://mcgill-nlp.github.io/weblinx">🌐Website</a></div>
<div><a href="https://huggingface.co/spaces/McGill-NLP/weblinx-explorer">💻Explorer</a></div>
<div><a href="https://huggingface.co/datasets/McGill-NLP/WebLINX">🤗Dataset</a></div>
<div><a href="https://github.com/McGill-NLP/weblinx">💾Code</a></div>
</div>
## Quickstart
```python
from datasets import load_dataset
from huggingface_hub import snapshot_download
from transformers import pipeline
# Load validation split
valid = load_dataset("McGill-NLP/weblinx", split="validation")
# Download and load the templates
snapshot_download(
"McGill-NLP/WebLINX", repo_type="dataset", allow_patterns="templates/*.txt", local_dir="./"
)
with open('templates/llama.txt') as f:
template = f.read()
turn = valid[0]
turn_text = template.format(**turn)
# Load action model and input the text to get prediction
action_model = pipeline(
model="McGill-NLP/Llama-2-13b-chat-weblinx", device=0, torch_dtype='auto'
)
out = action_model(turn_text, return_full_text=False, max_new_tokens=64, truncation=True)
pred = out[0]['generated_text']
print("Ref:", turn["action"])
print("Pred:", pred)
```
## Original Model
This model is finetuned on WebLINX using checkpoints previously published on Huggingface Hub.\
[Click here to access the original model.](https://huggingface.co/meta-llama/Llama-2-13b-chat-hf)
## License
This model is derived from LLaMA-2, which can only be used with the [LLaMA 2 Community License Agreement](https://github.com/facebookresearch/llama/blob/main/LICENSE). By using or distributing any portion or element of this model, you agree to be bound by this Agreement. | {"language": ["en"], "license": "llama2", "library_name": "transformers", "tags": ["weblinx", "text-generation-inference", "web-agents", "agents"], "datasets": ["McGill-NLP/WebLINX", "McGill-NLP/WebLINX-full"], "metrics": ["f1", "iou", "chrf"], "pipeline_tag": "text-generation"} | text-generation | McGill-NLP/Llama-2-13b-chat-weblinx | [
"transformers",
"pytorch",
"weblinx",
"text-generation-inference",
"web-agents",
"agents",
"text-generation",
"en",
"dataset:McGill-NLP/WebLINX",
"dataset:McGill-NLP/WebLINX-full",
"arxiv:2402.05930",
"license:llama2",
"endpoints_compatible",
"region:us"
] | 2024-02-07T08:58:05+00:00 | [
"2402.05930"
] | [
"en"
] | TAGS
#transformers #pytorch #weblinx #text-generation-inference #web-agents #agents #text-generation #en #dataset-McGill-NLP/WebLINX #dataset-McGill-NLP/WebLINX-full #arxiv-2402.05930 #license-llama2 #endpoints_compatible #region-us
| <div align="center">
<h1 style="margin-bottom: 0.5em;">WebLINX: Real-World Website Navigation with Multi-Turn Dialogue</h1>
<em>Xing Han Lù*, Zdeněk Kasner*, Siva Reddy</em>
</div>
<div style="margin-bottom: 2em"></div>
<div style="display: flex; justify-content: space-around; align-items: center; font-size: 120%;">
<div><a href="URL
<div><a href="URL
<div><a href="URL
<div><a href="URL
<div><a href="URL
</div>
## Quickstart
## Original Model
This model is finetuned on WebLINX using checkpoints previously published on Huggingface Hub.\
Click here to access the original model.
## License
This model is derived from LLaMA-2, which can only be used with the LLaMA 2 Community License Agreement. By using or distributing any portion or element of this model, you agree to be bound by this Agreement. | [
"## Quickstart",
"## Original Model\n\nThis model is finetuned on WebLINX using checkpoints previously published on Huggingface Hub.\\\nClick here to access the original model.",
"## License\n\nThis model is derived from LLaMA-2, which can only be used with the LLaMA 2 Community License Agreement. By using or distributing any portion or element of this model, you agree to be bound by this Agreement."
] | [
"TAGS\n#transformers #pytorch #weblinx #text-generation-inference #web-agents #agents #text-generation #en #dataset-McGill-NLP/WebLINX #dataset-McGill-NLP/WebLINX-full #arxiv-2402.05930 #license-llama2 #endpoints_compatible #region-us \n",
"## Quickstart",
"## Original Model\n\nThis model is finetuned on WebLINX using checkpoints previously published on Huggingface Hub.\\\nClick here to access the original model.",
"## License\n\nThis model is derived from LLaMA-2, which can only be used with the LLaMA 2 Community License Agreement. By using or distributing any portion or element of this model, you agree to be bound by this Agreement."
] | [
97,
3,
34,
51
] | [
"passage: TAGS\n#transformers #pytorch #weblinx #text-generation-inference #web-agents #agents #text-generation #en #dataset-McGill-NLP/WebLINX #dataset-McGill-NLP/WebLINX-full #arxiv-2402.05930 #license-llama2 #endpoints_compatible #region-us \n## Quickstart## Original Model\n\nThis model is finetuned on WebLINX using checkpoints previously published on Huggingface Hub.\\\nClick here to access the original model.## License\n\nThis model is derived from LLaMA-2, which can only be used with the LLaMA 2 Community License Agreement. By using or distributing any portion or element of this model, you agree to be bound by this Agreement."
] | [
-0.0296409223228693,
0.05343332886695862,
-0.002485502976924181,
0.051680222153663635,
0.04417257010936737,
0.013712680898606777,
0.2296208143234253,
0.023087117820978165,
0.13175058364868164,
-0.1368471384048462,
0.05337994173169136,
0.10413132607936859,
-0.003542238147929311,
0.06329233944416046,
0.027229076251387596,
-0.14795710146427155,
-0.040844857692718506,
0.05514382943511009,
0.052220240235328674,
0.08444786071777344,
0.10699709504842758,
-0.047850169241428375,
0.10708499699831009,
0.07737737149000168,
-0.07930993288755417,
0.014483344741165638,
0.08389467746019363,
-0.030825840309262276,
0.07119917124509811,
0.0956871286034584,
0.04503067210316658,
0.03149968013167381,
0.08989056199789047,
-0.16889265179634094,
0.01899377442896366,
-0.02771536074578762,
-0.0306047685444355,
0.04581451043486595,
0.048822298645973206,
0.042060259729623795,
0.25183823704719543,
0.04621303081512451,
-0.05176088958978653,
0.04396245256066322,
-0.036876168102025986,
-0.016309335827827454,
-0.047294408082962036,
0.12762965261936188,
0.060992833226919174,
0.0704699233174324,
0.05418843403458595,
0.13403990864753723,
-0.05112022906541824,
0.09810321778059006,
0.14440776407718658,
-0.3096553087234497,
0.021356025710701942,
0.2667551040649414,
0.05164014920592308,
-0.05884915962815285,
-0.006889234762638807,
0.09655949473381042,
0.04002625495195389,
0.014446134679019451,
0.1154959499835968,
-0.10367774218320847,
-0.02925376407802105,
-0.007958393543958664,
-0.07039850205183029,
0.002399106277152896,
0.22113795578479767,
0.0169858206063509,
-0.035341307520866394,
-0.07158268243074417,
-0.0490645207464695,
0.1626574695110321,
-0.05600694566965103,
0.04981519281864166,
0.09432196617126465,
0.07607187330722809,
-0.006885642185807228,
-0.13724017143249512,
-0.09060559421777725,
-0.07435180246829987,
-0.1642199456691742,
0.15459716320037842,
0.013815712183713913,
0.13646557927131653,
-0.17125314474105835,
0.06583050638437271,
-0.08906170725822449,
-0.03560687229037285,
-0.0074757966212928295,
-0.028605500236153603,
0.18130961060523987,
0.022509019821882248,
-0.058544009923934937,
0.03627723827958107,
0.10615255683660507,
0.042291466146707535,
-0.06577084213495255,
-0.09154531359672546,
0.019436432048678398,
0.09084748476743698,
0.036357808858156204,
0.002506040968000889,
-0.032846108078956604,
0.003392173442989588,
0.12294496595859528,
-0.10513916611671448,
0.09792272746562958,
0.0076011656783521175,
-0.09114925563335419,
0.01889542117714882,
-0.10109928250312805,
0.02168193832039833,
0.06709500402212143,
0.09517616033554077,
-0.0058757648803293705,
-0.028946345672011375,
0.11366401612758636,
-0.07400105893611908,
0.040618132799863815,
0.010687326081097126,
-0.07039257138967514,
0.03755034878849983,
0.167292982339859,
0.036736954003572464,
-0.05661310255527496,
-0.06042530760169029,
-0.09250546246767044,
-0.014595020562410355,
-0.046749699860811234,
-0.07675641030073166,
0.06653615087270737,
0.0019143539248034358,
0.04171358421444893,
-0.11349012702703476,
-0.23976773023605347,
-0.006750364787876606,
0.06880087405443192,
0.058974381536245346,
-0.08463744074106216,
0.023784782737493515,
0.03719325736165047,
-0.0468580424785614,
-0.05928516387939453,
-0.038915373384952545,
-0.046471841633319855,
0.028936555609107018,
-0.09988315403461456,
-0.012502691708505154,
-0.1818402111530304,
0.039053287357091904,
-0.0665053278207779,
0.038366079330444336,
-0.02225840464234352,
0.015482742339372635,
0.013464603573083878,
0.10995377600193024,
-0.029480641707777977,
-0.03221777826547623,
0.06104205548763275,
0.04136425256729126,
0.027471138164401054,
0.10732483118772507,
0.012339772656559944,
-0.04044735059142113,
0.1219346821308136,
-0.1419004201889038,
-0.14903299510478973,
0.011691109277307987,
-0.030081279575824738,
0.15032152831554413,
0.12533916532993317,
0.20509988069534302,
0.17078061401844025,
-0.19654680788516998,
0.015115081332623959,
0.09294181317090988,
-0.04273570328950882,
-0.2846440076828003,
-0.013347247615456581,
0.0006378304096870124,
-0.07819517701864243,
0.04388915002346039,
-0.14109787344932556,
0.09648364782333374,
-0.03125271201133728,
-0.05826044827699661,
-0.059627287089824677,
-0.09624665975570679,
-0.0008694712305441499,
-0.022974679246544838,
0.021461451426148415,
0.01618058606982231,
0.04037950560450554,
0.10177178680896759,
0.10164663195610046,
-0.030091378837823868,
0.052263304591178894,
-0.08539185672998428,
-0.0016841108445078135,
0.0028547216206789017,
0.016735736280679703,
-0.07009397447109222,
-0.06811382621526718,
-0.007154102902859449,
0.05028471350669861,
0.04595690965652466,
0.05672946572303772,
0.026693599298596382,
-0.008489026688039303,
0.001203258871100843,
0.057137381285429,
0.027279194444417953,
-0.039506614208221436,
-0.02588379755616188,
-0.13743959367275238,
-0.027814814820885658,
-0.012637527659535408,
-0.012244070880115032,
-0.11476250737905502,
0.0352298766374588,
-0.04823076352477074,
-0.012387726455926895,
-0.023098615929484367,
0.06427907198667526,
0.03713912516832352,
-0.02277226187288761,
-0.015797138214111328,
0.02955901250243187,
0.10951995104551315,
0.05955716595053673,
-0.12054795771837234,
0.18450653553009033,
-0.012256497517228127,
0.0616796500980854,
0.1426030695438385,
-0.07826922088861465,
0.08833979815244675,
-0.036796748638153076,
-0.056814324110746384,
-0.0015061995945870876,
-0.00117584562394768,
0.05058279260993004,
0.05894250422716141,
0.050677135586738586,
0.10167902708053589,
-0.08394249528646469,
0.0010057674953714013,
-0.022706979885697365,
-0.1410665512084961,
-0.008127482607960701,
0.011596359312534332,
0.14084747433662415,
0.010509440675377846,
0.02559407241642475,
0.1278587132692337,
-0.0011408713180571795,
0.07877030968666077,
-0.03267623856663704,
0.011246920563280582,
-0.05113549903035164,
-0.017774594947695732,
-0.03747421130537987,
0.06353124976158142,
-0.010855949483811855,
-0.02867347188293934,
0.03903820738196373,
-0.03127734735608101,
0.059747904539108276,
-0.15778736770153046,
-0.034935954958200455,
0.005586440674960613,
-0.09560400247573853,
-0.05988148972392082,
0.021595720201730728,
-0.0978819951415062,
0.06834924221038818,
-0.05744044855237007,
-0.03999454900622368,
0.01218351535499096,
-0.022947607561945915,
-0.11802057921886444,
0.10926765203475952,
-0.05647342652082443,
-0.08623610436916351,
-0.1569310873746872,
-0.14246836304664612,
-0.20083630084991455,
-0.020058874040842056,
0.009029612876474857,
0.019940124824643135,
-0.034432049840688705,
-0.10913053900003433,
-0.09325440227985382,
0.014703957363963127,
-0.057255469262599945,
0.025150666013360023,
0.027167893946170807,
-0.006050568539649248,
-0.1771286129951477,
-0.0724814385175705,
-0.0548495315015316,
-0.08386638760566711,
0.045183390378952026,
-0.05779443681240082,
0.08730017393827438,
0.1387021392583847,
-0.0021187467500567436,
0.01165302749723196,
-0.021704159677028656,
0.16403687000274658,
0.035026393830776215,
0.037121668457984924,
0.2643772065639496,
0.050013378262519836,
0.06395452469587326,
0.08105865865945816,
0.06949472427368164,
-0.047473758459091187,
0.013913478702306747,
-0.027811694890260696,
-0.08121200650930405,
-0.19963377714157104,
-0.10345137119293213,
-0.03272837772965431,
0.051469139754772186,
0.00845492072403431,
0.06568912416696548,
0.02205798029899597,
0.09307887405157089,
-0.003957103006541729,
0.012884331867098808,
0.06113657355308533,
0.04527485370635986,
0.18988563120365143,
-0.054046835750341415,
0.06849454343318939,
-0.11356790363788605,
0.04307687655091286,
0.10560145229101181,
0.10056627541780472,
0.21091791987419128,
0.04747157171368599,
0.05964682251214981,
0.177734836935997,
0.13753102719783783,
0.0014984625158831477,
0.1268956959247589,
-0.02231523022055626,
0.025931213051080704,
0.00226510944776237,
-0.12470093369483948,
-0.014811457134783268,
0.0649472251534462,
-0.12968815863132477,
-0.048713039606809616,
-0.04734335467219353,
0.010858990252017975,
-0.008179348893463612,
0.1997232437133789,
0.025919629260897636,
-0.175649955868721,
-0.01445000246167183,
0.03719361871480942,
0.046424273401498795,
0.056370705366134644,
0.05627721548080444,
0.005408253986388445,
-0.0728248730301857,
0.13737982511520386,
0.050395213067531586,
0.13734416663646698,
0.003337774658575654,
-0.00864831730723381,
-0.003113707061856985,
0.006011720281094313,
0.028368061408400536,
0.07259523868560791,
-0.21965818107128143,
0.17004740238189697,
0.011927696876227856,
0.02899351343512535,
-0.03421325236558914,
-0.016251064836978912,
0.03536625951528549,
0.24109214544296265,
0.07837440073490143,
0.06534034013748169,
0.08693494647741318,
0.034142445772886276,
-0.052762966603040695,
0.04296443983912468,
-0.0972309410572052,
0.03852931782603264,
0.028233850374817848,
-0.05596400424838066,
0.01856357790529728,
0.03028849884867668,
0.04795415326952934,
-0.14982281625270844,
-0.03849043324589729,
-0.056614626199007034,
0.20782043039798737,
-0.10537190735340118,
-0.07190443575382233,
-0.00016949095879681408,
0.0016718782717362046,
0.2071806788444519,
-0.019410530105233192,
-0.1446511447429657,
-0.06635145843029022,
-0.120254747569561,
-0.013660244643688202,
-0.06857322156429291,
0.012289543636143208,
-0.01642848551273346,
0.055076200515031815,
-0.08505669236183167,
-0.1654210239648819,
-0.013563234359025955,
-0.11363766342401505,
0.003116403240710497,
0.048090241849422455,
0.06488461047410965,
-0.002575291320681572,
-0.016121255233883858,
0.08138550072908401,
-0.04696520417928696,
-0.1101791113615036,
-0.15889954566955566,
-0.07834422588348389,
0.18458807468414307,
-0.017999054864048958,
-0.03983577713370323,
-0.13725309073925018,
-0.04650236666202545,
-0.00008590496145188808,
-0.04122468829154968,
0.09573226422071457,
0.10474736988544464,
-0.06288175284862518,
0.14085203409194946,
0.2398180365562439,
-0.16310124099254608,
-0.25096914172172546,
-0.1319558471441269,
-0.09848345816135406,
-0.058576542884111404,
0.07493297010660172,
-0.09021080285310745,
0.12801480293273926,
-0.005924448370933533,
-0.06691374629735947,
0.04544280096888542,
-0.2899399697780609,
-0.10279250144958496,
0.0608290433883667,
0.13826891779899597,
0.2607450783252716,
-0.12576515972614288,
-0.05426850914955139,
-0.07140879333019257,
-0.2365919053554535,
0.1251966804265976,
-0.15643376111984253,
0.055787816643714905,
-0.0377209298312664,
0.1990390568971634,
0.008926108479499817,
-0.036930471658706665,
0.041017454117536545,
0.04053428769111633,
0.08751165121793747,
-0.11738582700490952,
-0.019798943772912025,
0.204348623752594,
-0.039421770721673965,
0.19315855205059052,
-0.16551832854747772,
0.056711725890636444,
-0.12133936583995819,
-0.05914776399731636,
-0.08696691691875458,
0.11423016339540482,
-0.0718814954161644,
-0.10010028630495071,
-0.0645562931895256,
-0.00549652986228466,
0.04075479507446289,
0.021564779803156853,
0.017633728682994843,
0.007948621176183224,
0.004642283543944359,
0.23354589939117432,
0.057969171553850174,
-0.12478221207857132,
-0.06499668210744858,
-0.044017087668180466,
-0.0783129557967186,
0.05131697654724121,
-0.1911553144454956,
-0.04337352141737938,
0.08182553946971893,
0.026232002303004265,
0.021091748028993607,
0.03156274929642677,
-0.017759177833795547,
-0.03410971909761429,
0.08695707470178604,
-0.12357253581285477,
-0.10250061750411987,
-0.0026323774363845587,
0.010115079581737518,
-0.01265812385827303,
0.09578683972358704,
0.20320922136306763,
-0.09474323689937592,
0.023110508918762207,
-0.022593917325139046,
0.049933239817619324,
-0.09571247547864914,
0.006553749553859234,
0.05931698903441429,
-0.023451875895261765,
-0.10019953548908234,
0.09155162423849106,
0.03676910698413849,
0.13035790622234344,
-0.019788358360528946,
-0.034742891788482666,
-0.09189634025096893,
-0.0738276019692421,
-0.08314850181341171,
0.09516745060682297,
-0.22327686846256256,
-0.12192226201295853,
-0.039150670170784,
-0.11042660474777222,
-0.0013986583799123764,
-0.013463187962770462,
0.030618082731962204,
0.06115268915891647,
-0.008667818270623684,
-0.08167243003845215,
-0.040836431086063385,
0.027128351852297783,
-0.06503809243440628,
-0.037386730313301086,
-0.13241975009441376,
0.09261789172887802,
0.0459497906267643,
0.10276347398757935,
-0.029748819768428802,
-0.0021017128601670265,
-0.04124242812395096,
0.04564850032329559,
-0.1351345032453537,
0.022485116496682167,
-0.1509617567062378,
0.029508724808692932,
-0.04610874503850937,
-0.022198690101504326,
-0.07839134335517883,
0.0779547393321991,
-0.05278138443827629,
-0.00301701040007174,
0.010365519672632217,
0.08092130720615387,
-0.1685735434293747,
-0.05582721158862114,
-0.01919444091618061,
0.023938093334436417,
0.05065903067588806,
-0.03335727006196976,
-0.08940228074789047,
0.008948542177677155,
-0.08623944967985153,
0.025054078549146652,
0.0032882026862353086,
0.012279183603823185,
-0.023777395486831665,
-0.03693051263689995,
-0.0015804936410859227,
0.09994557499885559,
-0.02681250311434269,
0.016563626006245613,
0.040344417095184326,
-0.09205546230077744,
-0.03443417698144913,
0.0004882828216068447,
-0.018803341314196587,
-0.03800297901034355,
-0.014208642765879631,
0.09320824593305588,
0.11484456807374954,
0.1475173532962799,
-0.040376149117946625,
0.010043294169008732,
-0.14693698287010193,
0.02454526722431183,
-0.00028276091325096786,
-0.03010915033519268,
-0.1280723512172699,
-0.11985743045806885,
-0.05697718635201454,
-0.009732152335345745,
0.2650490403175354,
0.1017022579908371,
-0.14540086686611176,
-0.01617739163339138,
0.08026256412267685,
0.014788107946515083,
-0.06035260483622551,
0.2497851848602295,
-0.049268778413534164,
0.0534321554005146,
0.035495445132255554,
0.058255329728126526,
0.043989840894937515,
-0.11365503817796707,
0.11714969575405121,
0.024874966591596603,
0.03625119477510452,
0.05887681618332863,
0.13098785281181335,
-0.006906242109835148,
-0.017906714230775833,
-0.1894065886735916,
0.013519592583179474,
0.06764991581439972,
-0.09694607555866241,
0.09657848626375198,
0.11902067810297012,
-0.10821080952882767,
0.06812600046396255,
0.08212050795555115,
-0.026433110237121582,
-0.16304241120815277,
-0.18522419035434723,
-0.04644053801894188,
-0.10375193506479263,
0.004162898287177086,
-0.09923198074102402,
-0.003284408478066325,
-0.04241622984409332,
-0.023743240162730217,
-0.07925564795732498,
-0.07796194404363632,
-0.1354885846376419,
-0.053214967250823975,
-0.0007884958176873624,
-0.012060025706887245,
-0.04453739896416664,
-0.09825519472360611,
0.011761914007365704,
0.007293719332665205,
0.011823560111224651,
-0.013854686170816422,
0.08756837993860245,
0.08204905688762665,
0.09676803648471832,
-0.01915719360113144,
-0.00883062556385994,
-0.08242011815309525,
0.00895261112600565,
0.11041775345802307,
0.0699157640337944,
0.04190788418054581,
-0.05830233544111252,
0.027655111625790596,
0.10352656245231628,
-0.017918661236763,
-0.14332076907157898,
-0.0031550254207104445,
0.18365471065044403,
-0.036640703678131104,
-0.031358011066913605,
-0.011880096979439259,
0.01593109965324402,
0.0003727012954186648,
0.32516422867774963,
0.28489020466804504,
-0.05422763526439667,
0.024554526433348656,
-0.10750562697649002,
0.009391501545906067,
0.0730963945388794,
0.17988070845603943,
0.018643086776137352,
0.2611449062824249,
0.003380064619705081,
0.024076977744698524,
-0.07213513553142548,
0.006838019005954266,
-0.10102711617946625,
0.019017009064555168,
-0.033059682697057724,
-0.08137805759906769,
-0.027574289590120316,
0.05951618403196335,
-0.08593741804361343,
0.07244797050952911,
0.0423894040286541,
-0.044451598078012466,
0.05651238188147545,
-0.04866095632314682,
0.049188192933797836,
0.03341607749462128,
0.06768310815095901,
-0.0753510370850563,
0.06233787536621094,
0.16880574822425842,
-0.047466736286878586,
-0.22927416861057281,
0.018369046971201897,
0.11708397418260574,
0.031250011175870895,
0.17071329057216644,
0.014098240993916988,
0.07764945924282074,
0.04200688377022743,
-0.02585621364414692,
-0.1329820305109024,
0.11542680114507675,
-0.039418622851371765,
-0.07657233625650406,
-0.014693803153932095,
-0.2185998111963272,
-0.0450856015086174,
-0.07707218080759048,
-0.003776037599891424,
0.04047314450144768,
0.01694202795624733,
0.11869476735591888,
-0.001215149532072246,
-0.10796265304088593,
-0.008565513417124748,
-0.11496415734291077,
0.07434694468975067,
0.045450177043676376,
-0.06033084914088249,
-0.05615730583667755,
-0.07602551579475403,
0.04068280756473541,
0.03521284461021423,
-0.13725949823856354,
0.06639149785041809,
-0.035561494529247284,
0.014008400961756706,
0.0073548126965761185,
0.030753664672374725,
-0.2055412381887436,
-0.007696729619055986,
-0.04431253671646118,
-0.05857739597558975,
-0.09057972580194473,
0.04548297077417374,
0.16499772667884827,
0.004963960498571396,
-0.013817140832543373,
-0.07723918557167053,
0.019789304584264755,
0.03929871320724487,
-0.08047576993703842,
-0.11860509216785431
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.7.2.dev0 | {"library_name": "peft", "base_model": "google/flan-t5-base"} | null | HeydarS/flan-t5-base_peft_v20 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:google/flan-t5-base",
"region:us"
] | 2024-02-07T09:00:20+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #safetensors #arxiv-1910.09700 #base_model-google/flan-t5-base #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.7.2.dev0 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.7.2.dev0"
] | [
"TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-google/flan-t5-base #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.7.2.dev0"
] | [
35,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
14
] | [
"passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-google/flan-t5-base #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.7.2.dev0"
] | [
-0.10626745969057083,
0.1943962424993515,
-0.003387606469914317,
0.03603767976164818,
0.09226002544164658,
0.019435854628682137,
0.05228874087333679,
0.12314661592245102,
-0.03763783350586891,
0.10612741112709045,
0.0677775889635086,
0.1003333255648613,
0.10013683140277863,
0.20788270235061646,
0.01028058398514986,
-0.20106801390647888,
0.025792978703975677,
-0.09577292203903198,
-0.007508585229516029,
0.12317893654108047,
0.15051408112049103,
-0.09860456734895706,
0.08032377064228058,
-0.016225092113018036,
-0.01558137871325016,
-0.03681138902902603,
-0.07320638746023178,
-0.035057611763477325,
0.04912315681576729,
0.049064796417951584,
0.05871843919157982,
-0.00005916723603149876,
0.09118524193763733,
-0.26729705929756165,
0.019125372171401978,
0.04606068879365921,
-0.008893071673810482,
0.08542384952306747,
0.10275928676128387,
-0.03844038024544716,
0.12587106227874756,
-0.03221188858151436,
0.14311404526233673,
0.08030287176370621,
-0.09467478841543198,
-0.21773016452789307,
-0.0719180628657341,
0.0834791362285614,
0.1696113646030426,
0.07952837646007538,
-0.043143972754478455,
0.13347607851028442,
-0.09846170246601105,
0.01899239793419838,
0.04215628653764725,
-0.08037710934877396,
-0.0743836760520935,
0.05872213840484619,
0.10499230772256851,
0.053478822112083435,
-0.1377313882112503,
-0.030999790877103806,
0.022513283416628838,
0.04074760526418686,
0.08075147122144699,
0.018500911071896553,
0.14020903408527374,
0.032465290278196335,
-0.14929986000061035,
-0.04096416383981705,
0.13178586959838867,
0.03145142272114754,
-0.0386669859290123,
-0.22038984298706055,
0.009189976379275322,
-0.08269672095775604,
-0.026602445170283318,
-0.053094200789928436,
0.035357702523469925,
0.0009046635823324323,
0.09034530818462372,
-0.03193694353103638,
-0.09359759837388992,
-0.012948989868164062,
0.0944470688700676,
0.04649518430233002,
0.024375872686505318,
-0.025291787460446358,
0.006814958062022924,
0.1236046850681305,
0.05509087070822716,
-0.12566722929477692,
-0.05955754220485687,
-0.07080139219760895,
-0.04547770693898201,
-0.04806475341320038,
0.033843204379081726,
0.02901734784245491,
0.058218520134687424,
0.2488924115896225,
-0.030058603733778,
0.0549616776406765,
0.05447668582201004,
0.019564809277653694,
0.04267705976963043,
0.09505946934223175,
-0.05727843567728996,
-0.1486940234899521,
-0.01470540463924408,
0.09698048233985901,
-0.00953452754765749,
-0.019430894404649734,
-0.04841039329767227,
0.03599175065755844,
0.046676695346832275,
0.10488031059503555,
0.09370704740285873,
-0.0038008210249245167,
-0.07420773059129715,
-0.04932110756635666,
0.20879322290420532,
-0.14909665286540985,
0.04108875244855881,
0.02244400605559349,
-0.018436485901474953,
-0.06020741909742355,
0.013366634957492352,
0.017413243651390076,
-0.025361547246575356,
0.09793671220541,
-0.06251633912324905,
-0.04163913428783417,
-0.11557204276323318,
-0.021286284551024437,
0.038216814398765564,
0.013806617818772793,
-0.02672383189201355,
-0.035722583532333374,
-0.06411529332399368,
-0.0939604863524437,
0.10544420778751373,
-0.0658579021692276,
-0.06309971213340759,
-0.02974298782646656,
-0.09095096588134766,
0.016579419374465942,
0.028497256338596344,
0.11275363713502884,
-0.024662071838974953,
0.04416226968169212,
-0.014005821198225021,
0.060549791902303696,
0.07976043969392776,
0.037235476076602936,
-0.06926264613866806,
0.061339445412158966,
-0.19918614625930786,
0.09161697328090668,
-0.07697001099586487,
0.028699280694127083,
-0.15864884853363037,
-0.017035553231835365,
0.00033481643185950816,
0.020158562809228897,
0.039182908833026886,
0.15084892511367798,
-0.19582802057266235,
-0.02592594362795353,
0.1559358388185501,
-0.10197359323501587,
-0.12382874637842178,
0.042082399129867554,
-0.05455990135669708,
0.16270385682582855,
0.021871736273169518,
-0.009292042814195156,
0.09196344763040543,
-0.15284870564937592,
-0.02294968068599701,
-0.02839653380215168,
-0.005660604685544968,
0.10367421060800552,
0.08136079460382462,
-0.07848794758319855,
0.02571188658475876,
0.014831089414656162,
-0.041812699288129807,
-0.022545142099261284,
-0.053160153329372406,
-0.11061898618936539,
0.0009782188571989536,
-0.08726845681667328,
0.026618853211402893,
-0.008247734978795052,
-0.07610312849283218,
-0.01254703477025032,
-0.16181626915931702,
-0.02871214598417282,
0.07868514955043793,
0.01724299229681492,
-0.01901300996541977,
-0.09065431356430054,
0.03456268459558487,
-0.02381935343146324,
-0.024170242249965668,
-0.15777996182441711,
-0.0361778698861599,
0.01957101747393608,
-0.13848066329956055,
0.012914540246129036,
-0.12052971124649048,
0.066510409116745,
0.014045109041035175,
-0.07196114957332611,
-0.03254105895757675,
-0.01547172199934721,
0.009743290953338146,
-0.05094056576490402,
-0.23752011358737946,
-0.021084511652588844,
-0.0554598905146122,
0.14978623390197754,
-0.22872553765773773,
0.03972625732421875,
0.04129142314195633,
0.12734706699848175,
0.0021993080154061317,
-0.0596797838807106,
0.027768824249505997,
-0.07154082506895065,
-0.022692902013659477,
-0.07238355278968811,
-0.004847003147006035,
-0.001676840358413756,
-0.04038437828421593,
0.016839073970913887,
-0.11187786608934402,
-0.0447884164750576,
0.10195057839155197,
0.06268775463104248,
-0.17277465760707855,
-0.02041766420006752,
-0.04777924343943596,
-0.06872023642063141,
-0.09030480682849884,
-0.05886688455939293,
0.09846045076847076,
0.04916569963097572,
0.03816893696784973,
-0.07307842373847961,
-0.06809652596712112,
0.010012323036789894,
-0.017745092511177063,
-0.025862621143460274,
0.11348828673362732,
0.06857994943857193,
-0.11471354216337204,
0.09571298211812973,
0.06638070940971375,
0.026577377691864967,
0.07707730680704117,
-0.025765812024474144,
-0.10656370222568512,
-0.0275901909917593,
0.04605529084801674,
0.012250564992427826,
0.16212764382362366,
-0.07695270329713821,
0.0528617724776268,
0.04372003301978111,
-0.034936483949422836,
0.04988601431250572,
-0.09815118461847305,
0.010052197612822056,
0.005698662716895342,
-0.015340845100581646,
0.016503745689988136,
-0.01996779255568981,
0.010818221606314182,
0.08451361209154129,
0.05374802649021149,
0.034732699394226074,
0.03416459634900093,
-0.027584291994571686,
-0.13143029808998108,
0.1801474541425705,
-0.09626327455043793,
-0.23413608968257904,
-0.15353670716285706,
0.05498236417770386,
0.05599696934223175,
-0.018610401079058647,
0.027064653113484383,
-0.05219969525933266,
-0.10190703719854355,
-0.07922445982694626,
-0.004773293621838093,
0.028319185599684715,
-0.060163792222738266,
-0.0763615071773529,
0.05326016992330551,
0.04257733002305031,
-0.11543019115924835,
0.04015215486288071,
0.06237921118736267,
-0.007595696486532688,
0.006950370967388153,
0.052949290722608566,
0.0827041044831276,
0.1778176873922348,
-0.012662996537983418,
-0.001374164829030633,
0.05132483318448067,
0.28036999702453613,
-0.15926094353199005,
0.11483950167894363,
0.11827429383993149,
-0.061905838549137115,
0.07703997194766998,
0.18894267082214355,
0.031426820904016495,
-0.10071038454771042,
0.03919421136379242,
0.0339176207780838,
-0.026562854647636414,
-0.2736181616783142,
-0.04572368040680885,
-0.010000117123126984,
-0.09846661984920502,
0.08148844540119171,
0.09112165868282318,
0.09233558923006058,
0.03964434936642647,
-0.06323430687189102,
-0.09103947132825851,
0.03861045092344284,
0.099139004945755,
-0.016542939469218254,
0.004856142681092024,
0.08154811710119247,
-0.028960175812244415,
0.009979736059904099,
0.09486377984285355,
-0.018665187060832977,
0.1644057333469391,
0.04859054088592529,
0.10418574512004852,
0.07721724361181259,
0.09577159583568573,
-0.00678221695125103,
0.024683605879545212,
0.019262781366705894,
0.023546066135168076,
0.016479844227433205,
-0.0836954340338707,
0.035068731755018234,
0.10865435749292374,
0.03942868858575821,
0.02600918337702751,
0.013191206380724907,
-0.044774990528821945,
0.0474453829228878,
0.18922826647758484,
0.009999711997807026,
-0.20108447968959808,
-0.08201143890619278,
0.05267728865146637,
-0.07563731074333191,
-0.14131365716457367,
-0.017839588224887848,
0.028532851487398148,
-0.16847741603851318,
0.015154414810240269,
-0.044349510222673416,
0.10102768242359161,
-0.06610823422670364,
-0.04040110483765602,
0.10356244444847107,
0.0725359171628952,
-0.02740391157567501,
0.06083453819155693,
-0.20365430414676666,
0.12665313482284546,
0.024027224630117416,
0.07074165344238281,
-0.08825352787971497,
0.0968342199921608,
0.0011091300984844565,
-0.00461592897772789,
0.16639287769794464,
0.002134338952600956,
-0.07097217440605164,
-0.05354686081409454,
-0.0921424999833107,
-0.013771062716841698,
0.10175586491823196,
-0.1249750629067421,
0.06322869658470154,
-0.016876481473445892,
-0.03235099837183952,
0.006789884530007839,
-0.0761876031756401,
-0.13253721594810486,
-0.17009471356868744,
0.05461978167295456,
-0.10156451165676117,
0.03955043852329254,
-0.09254651516675949,
-0.06745260208845139,
0.004009621683508158,
0.17828625440597534,
-0.17609639465808868,
-0.09290038049221039,
-0.14633356034755707,
-0.08385132253170013,
0.16673745214939117,
-0.03696591779589653,
0.08714164793491364,
0.0008085340959951282,
0.1633509248495102,
0.014043261297047138,
-0.0004224503063596785,
0.09846942126750946,
-0.08807885646820068,
-0.19665396213531494,
-0.058461662381887436,
0.1666961908340454,
0.14166374504566193,
0.03869305178523064,
-0.010088405571877956,
0.021535715088248253,
-0.05324366316199303,
-0.11172157526016235,
0.025243619456887245,
0.12963053584098816,
0.08552177995443344,
-0.016120724380016327,
-0.03886369988322258,
-0.09494185447692871,
-0.0650559812784195,
-0.06025339663028717,
0.001782225794158876,
0.18376830220222473,
-0.07549230009317398,
0.16126713156700134,
0.11873925477266312,
-0.054935604333877563,
-0.20435717701911926,
0.051667653024196625,
0.057001613080501556,
0.01535657700151205,
0.03362317755818367,
-0.19608907401561737,
0.08624274283647537,
-0.00034414982656016946,
-0.07186609506607056,
0.16279858350753784,
-0.16888505220413208,
-0.14565519988536835,
0.09902361035346985,
0.0366116538643837,
-0.236861914396286,
-0.13917239010334015,
-0.09933043271303177,
-0.018550176173448563,
-0.11035100370645523,
0.062390539795160294,
0.00021574481797870249,
0.011819778941571712,
0.03125092014670372,
0.01891767606139183,
0.02876076102256775,
-0.04887833073735237,
0.20177625119686127,
-0.02594452165067196,
0.010557923465967178,
-0.05068642273545265,
-0.09157499670982361,
0.033935029059648514,
-0.0506705567240715,
0.1028408408164978,
0.00432779174298048,
0.02938619628548622,
-0.1420229971408844,
-0.04426152631640434,
-0.06219794973731041,
0.03049236349761486,
-0.09847007691860199,
-0.08974216133356094,
-0.04626361280679703,
0.09782198816537857,
0.09823662042617798,
-0.03223685920238495,
0.005748902913182974,
-0.08590634912252426,
0.0708165094256401,
0.20846675336360931,
0.1880311667919159,
0.07307461649179459,
-0.06206341087818146,
0.021826045587658882,
-0.034976597875356674,
0.04174673557281494,
-0.22437147796154022,
0.043028511106967926,
0.056297581642866135,
0.024020612239837646,
0.08573832362890244,
-0.00941250380128622,
-0.1538187861442566,
-0.07839273661375046,
0.0795898586511612,
-0.04730746150016785,
-0.16308970749378204,
-0.027136368677020073,
0.03949432820081711,
-0.21053273975849152,
-0.04798044264316559,
0.015552849508821964,
-0.02477485127747059,
-0.04049832746386528,
0.02438291907310486,
0.07634059339761734,
-0.01943463273346424,
0.11085540056228638,
0.09432967007160187,
0.09200961887836456,
-0.09735818207263947,
0.08124689757823944,
0.07701282203197479,
-0.04806763306260109,
0.021812697872519493,
0.11117840558290482,
-0.050739508122205734,
-0.037840310484170914,
0.09233005344867706,
0.08607760071754456,
0.02519970014691353,
-0.05105016753077507,
0.015625236555933952,
-0.05509917065501213,
0.06538805365562439,
0.12310566008090973,
0.028020475059747696,
-0.00401606410741806,
0.05817234516143799,
0.036317672580480576,
-0.10091570764780045,
0.11022966355085373,
0.060984570533037186,
0.022640034556388855,
-0.04285145550966263,
-0.025664152577519417,
-0.010606732219457626,
-0.00946447066962719,
-0.019042406231164932,
-0.005740757565945387,
-0.09020785242319107,
-0.008166715502738953,
-0.1037309393286705,
0.03222903981804848,
-0.07299622148275375,
0.010693063028156757,
0.026414508000016212,
-0.05268304422497749,
0.009201978333294392,
0.0049271248281002045,
-0.07451415061950684,
-0.051883138716220856,
-0.014390280470252037,
0.08451662957668304,
-0.12997590005397797,
0.03118613176047802,
0.07371551543474197,
-0.1054111048579216,
0.07181970775127411,
0.0018193931318819523,
0.007645344827324152,
0.012531098909676075,
-0.1657782346010208,
0.056539472192525864,
-0.024776315316557884,
-0.014173267409205437,
0.01549680344760418,
-0.2142029106616974,
-0.011480376124382019,
-0.04960622265934944,
-0.05262897163629532,
0.012350749224424362,
-0.02797774039208889,
-0.12506969273090363,
0.09995295107364655,
-0.005099916830658913,
-0.07048572599887848,
-0.017425669357180595,
0.038649704307317734,
0.09460994601249695,
-0.022369757294654846,
0.12504130601882935,
-0.028333604335784912,
0.07363194227218628,
-0.17342032492160797,
-0.0020430213771760464,
-0.016072485595941544,
0.04032069072127342,
-0.022196732461452484,
-0.022031286731362343,
0.05816557630896568,
-0.01751047931611538,
0.17629386484622955,
-0.021942971274256706,
0.07332462072372437,
0.0549980029463768,
0.00522792199626565,
0.005346484947949648,
0.08542075008153915,
0.06342078000307083,
-0.008526364341378212,
-0.004864741116762161,
0.04174121841788292,
-0.006167198531329632,
-0.04359015077352524,
-0.1550811231136322,
0.07007448375225067,
0.16129887104034424,
0.04683932289481163,
0.02103545516729355,
0.03010541759431362,
-0.12103217840194702,
-0.06719450652599335,
0.132915198802948,
-0.007291777525097132,
-0.035461392253637314,
-0.07875702530145645,
0.1820414662361145,
0.12919233739376068,
-0.19846510887145996,
0.0785851925611496,
-0.062020812183618546,
-0.055078864097595215,
-0.12649577856063843,
-0.1544293910264969,
-0.06379027664661407,
-0.04223756864666939,
-0.021467499434947968,
-0.06189240515232086,
0.050444722175598145,
0.04895254969596863,
0.004684585612267256,
-0.019241079688072205,
0.10444451123476028,
0.012044665403664112,
-0.023837493732571602,
0.04863271489739418,
0.06281875818967819,
0.03184809163212776,
-0.09782924503087997,
0.00889611430466175,
-0.003255242481827736,
0.01744062267243862,
0.06000742316246033,
0.016260197386145592,
-0.05058757960796356,
0.013521446846425533,
-0.014316970482468605,
-0.11267974227666855,
0.04368726164102554,
-0.017312193289399147,
-0.034139636904001236,
0.13907207548618317,
0.029180752113461494,
0.008199350908398628,
-0.020409412682056427,
0.2393750250339508,
-0.07374456524848938,
-0.07687947899103165,
-0.15254373848438263,
0.0597771592438221,
-0.06916273385286331,
0.0306389182806015,
0.03442620113492012,
-0.11679217219352722,
0.020971601828932762,
0.16349677741527557,
0.13074807822704315,
-0.010041698813438416,
0.01111083198338747,
0.05241261422634125,
0.0017416616901755333,
-0.030621716752648354,
0.014367358759045601,
0.05457673966884613,
0.13300518691539764,
-0.07221394777297974,
0.0668054074048996,
-0.012110509909689426,
-0.0727691724896431,
-0.017204321920871735,
0.10875110328197479,
-0.00115925632417202,
0.001458446611650288,
-0.07169950008392334,
0.14218659698963165,
-0.0870911106467247,
-0.23450365662574768,
0.05871950089931488,
-0.06713622808456421,
-0.14922666549682617,
-0.0495445616543293,
0.009493282064795494,
-0.01419214066118002,
0.014483078382909298,
0.07484103739261627,
-0.04558661952614784,
0.17140412330627441,
0.04358278587460518,
-0.05350765213370323,
-0.08261451125144958,
0.060234762728214264,
-0.12359639257192612,
0.28447747230529785,
0.018942512571811676,
0.05022856965661049,
0.10444867610931396,
-0.01699936017394066,
-0.14132224023342133,
0.012299301102757454,
0.10720779746770859,
-0.06784209609031677,
0.06262636929750443,
0.17135582864284515,
-0.000842442037537694,
0.1268947869539261,
0.05236613005399704,
-0.05412169173359871,
0.03930345177650452,
-0.08381089568138123,
-0.05141802504658699,
-0.11254512518644333,
0.07955029606819153,
-0.08164326846599579,
0.16071252524852753,
0.13219253718852997,
-0.0665043368935585,
-0.004091443493962288,
-0.0223674438893795,
0.08265941590070724,
0.007208789233118296,
0.11081504821777344,
0.008757385425269604,
-0.1851312816143036,
0.03756987303495407,
0.011224213056266308,
0.09999052435159683,
-0.2103920876979828,
-0.06459777057170868,
0.051976144313812256,
-0.021772118285298347,
-0.07437668740749359,
0.11798278242349625,
0.04241457208991051,
0.03347695991396904,
-0.03955359011888504,
-0.046107664704322815,
0.002645334228873253,
0.15035350620746613,
-0.11573656648397446,
-0.009015773423016071
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# my_ner_model
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2690
- Precision: 0.5545
- Recall: 0.3253
- F1: 0.4100
- Accuracy: 0.9420
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
| No log | 1.0 | 213 | 0.2829 | 0.5103 | 0.2289 | 0.3161 | 0.9377 |
| No log | 2.0 | 426 | 0.2690 | 0.5545 | 0.3253 | 0.4100 | 0.9420 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.2.0
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["precision", "recall", "f1", "accuracy"], "base_model": "distilbert-base-uncased", "model-index": [{"name": "my_ner_model", "results": []}]} | token-classification | veronica1608/my_ner_model | [
"transformers",
"safetensors",
"distilbert",
"token-classification",
"generated_from_trainer",
"base_model:distilbert-base-uncased",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T09:00:58+00:00 | [] | [] | TAGS
#transformers #safetensors #distilbert #token-classification #generated_from_trainer #base_model-distilbert-base-uncased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| my\_ner\_model
==============
This model is a fine-tuned version of distilbert-base-uncased on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 0.2690
* Precision: 0.5545
* Recall: 0.3253
* F1: 0.4100
* Accuracy: 0.9420
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 2
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.2.0
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.2.0\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #safetensors #distilbert #token-classification #generated_from_trainer #base_model-distilbert-base-uncased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.2.0\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
69,
98,
4,
30
] | [
"passage: TAGS\n#transformers #safetensors #distilbert #token-classification #generated_from_trainer #base_model-distilbert-base-uncased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.2.0\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.09100259095430374,
0.10144243389368057,
-0.002475470770150423,
0.11793125420808792,
0.14300183951854706,
0.013321716338396072,
0.13798606395721436,
0.10333547741174698,
-0.08075544238090515,
0.030419252812862396,
0.12973636388778687,
0.14780081808567047,
-0.005629493854939938,
0.144350066781044,
-0.07697660475969315,
-0.2208528369665146,
0.023716187104582787,
0.020043015480041504,
-0.0459410585463047,
0.1137194111943245,
0.10989760607481003,
-0.12915253639221191,
0.08743251115083694,
-0.004197872243821621,
-0.16493737697601318,
0.009085629135370255,
0.022516688331961632,
-0.054311010986566544,
0.12817230820655823,
0.01705576665699482,
0.12202572822570801,
0.01757626421749592,
0.09625823050737381,
-0.19267426431179047,
0.001953189494088292,
0.049804504960775375,
0.00781647115945816,
0.07170090079307556,
0.02684571035206318,
0.0026308398228138685,
0.06881342083215714,
-0.0750543624162674,
0.061759091913700104,
0.016391288489103317,
-0.12402257323265076,
-0.2420976310968399,
-0.08704659342765808,
0.041512176394462585,
0.10782662034034729,
0.06700222194194794,
-0.006896450184285641,
0.13892556726932526,
-0.07981764525175095,
0.08459976315498352,
0.2072015255689621,
-0.31426525115966797,
-0.058341920375823975,
0.04796932637691498,
0.011259264312684536,
0.06401084363460541,
-0.09974464029073715,
-0.03649848699569702,
0.062460850924253464,
0.028589652851223946,
0.14332027733325958,
-0.029064616188406944,
-0.0880604013800621,
0.0037925459910184145,
-0.1455792784690857,
-0.023539487272500992,
0.17346622049808502,
0.05663417652249336,
-0.06290023028850555,
-0.04552239924669266,
-0.06615603715181351,
-0.12375416606664658,
-0.03578045219182968,
-0.030747951939702034,
0.05503097176551819,
-0.017880143597722054,
-0.053590886294841766,
0.004834338556975126,
-0.09325059503316879,
-0.07311946153640747,
-0.046806298196315765,
0.1809282898902893,
0.04394633695483208,
0.007901955395936966,
0.0034786537289619446,
0.10372629761695862,
-0.053903233259916306,
-0.12862686812877655,
0.00311244442127645,
0.009719550609588623,
0.024004150182008743,
-0.05829163268208504,
-0.05733577907085419,
-0.031877994537353516,
0.026511631906032562,
0.18619117140769958,
-0.07197149097919464,
0.03526099771261215,
0.0249521154910326,
0.03624480217695236,
-0.09575407207012177,
0.14673437178134918,
-0.023699697107076645,
-0.024723311886191368,
0.028240542858839035,
0.06771748512983322,
0.05249618738889694,
0.006875194143503904,
-0.11074458062648773,
0.02859567105770111,
0.11588945984840393,
0.01991465501487255,
-0.08445796370506287,
0.07042910903692245,
-0.06290821731090546,
-0.007171615492552519,
0.043516144156455994,
-0.08917319774627686,
0.02555151842534542,
-0.010854472406208515,
-0.05741305276751518,
-0.08241895586252213,
0.02013375051319599,
0.03505709767341614,
0.01677902601659298,
0.10545404255390167,
-0.08411739766597748,
0.005079216789454222,
-0.08238016068935394,
-0.1044001430273056,
0.0034834183752536774,
-0.07056036591529846,
0.04288647696375847,
-0.11415145546197891,
-0.20764540135860443,
0.007587339263409376,
0.06416984647512436,
-0.01853139139711857,
-0.04077097773551941,
-0.04762139916419983,
-0.07244699448347092,
0.004762317053973675,
-0.019348114728927612,
0.06265508383512497,
-0.0697975903749466,
0.09182842820882797,
0.05029675364494324,
0.06032229959964752,
-0.0508536696434021,
0.03449614346027374,
-0.13106824457645416,
0.03680546209216118,
-0.19055290520191193,
0.01620429940521717,
-0.07086510956287384,
0.06322585046291351,
-0.07511381804943085,
-0.08225836604833603,
0.012708496302366257,
-0.0052556623704731464,
0.06413397192955017,
0.09848766028881073,
-0.14492526650428772,
-0.05122390389442444,
0.15617001056671143,
-0.10416006296873093,
-0.14427393674850464,
0.12613102793693542,
-0.06171426922082901,
0.0560101717710495,
0.06292638927698135,
0.1791139543056488,
0.06085597723722458,
-0.08741652965545654,
0.0038433854933828115,
-0.006627559196203947,
0.04948227480053902,
-0.028279289603233337,
0.075567826628685,
0.007008469197899103,
-0.020679647102952003,
0.02460857480764389,
-0.06401578336954117,
0.04470645636320114,
-0.08629574626684189,
-0.08990945667028427,
-0.0524502769112587,
-0.11316055804491043,
0.07095740735530853,
0.0437145009636879,
0.06278958916664124,
-0.12401971966028214,
-0.07983232289552689,
0.07026439905166626,
0.08374929428100586,
-0.058462005108594894,
0.014172783121466637,
-0.07646743953227997,
0.08103906363248825,
-0.051124315708875656,
-0.02507597580552101,
-0.14470379054546356,
-0.046716224402189255,
0.02966930903494358,
-0.00839164573699236,
0.004566361662000418,
-0.020155737176537514,
0.06667378544807434,
0.0961463525891304,
-0.07636364549398422,
-0.04485464468598366,
-0.02720554731786251,
0.02737581357359886,
-0.11642715334892273,
-0.18581892549991608,
-0.025910358875989914,
-0.03158204257488251,
0.12308556586503983,
-0.22107461094856262,
0.05007064342498779,
-0.021093478426337242,
0.09642567485570908,
0.036859575659036636,
-0.00936347246170044,
-0.050499364733695984,
0.06311988830566406,
-0.04975410923361778,
-0.0626838356256485,
0.05211666598916054,
0.014106261543929577,
-0.08496662974357605,
-0.06373561173677444,
-0.11732593923807144,
0.2000179886817932,
0.12210799008607864,
-0.08063925057649612,
-0.08108332753181458,
-0.008494924753904343,
-0.04817047342658043,
-0.027799786999821663,
-0.05662282556295395,
0.015243323519825935,
0.11534248292446136,
-0.02540449984371662,
0.14313635230064392,
-0.08715888112783432,
-0.03243991732597351,
0.017124658450484276,
-0.06335064768791199,
0.019430244341492653,
0.08796637505292892,
0.0782037004828453,
-0.12286189198493958,
0.15592248737812042,
0.19981729984283447,
-0.08801932632923126,
0.11081137508153915,
-0.04960525780916214,
-0.04829023405909538,
-0.02844250574707985,
0.009699699468910694,
0.0010431084083393216,
0.10544252395629883,
-0.1052330955862999,
0.018054837360978127,
0.016345655545592308,
0.03184664249420166,
-0.0022415227722376585,
-0.21195583045482635,
-0.03083869442343712,
0.039593786001205444,
-0.04199492186307907,
-0.0005719709442928433,
-0.018430806696414948,
-0.011321350000798702,
0.08878766000270844,
-0.005864466540515423,
-0.0968066081404686,
0.06248654052615166,
0.008113461546599865,
-0.07644880563020706,
0.20168741047382355,
-0.09802159667015076,
-0.11240458488464355,
-0.1199149340391159,
-0.06968732178211212,
-0.05696345493197441,
0.03951149806380272,
0.07998805493116379,
-0.0610705129802227,
-0.04960748180747032,
-0.10154428333044052,
-0.003484516404569149,
0.0407552532851696,
0.023788558319211006,
0.023343240842223167,
0.0005407214630395174,
0.09285642951726913,
-0.10017557442188263,
-0.01994074322283268,
-0.03110239841043949,
-0.0570540577173233,
0.04183784872293472,
0.04069004952907562,
0.1061188355088234,
0.12105467170476913,
-0.030453352257609367,
-0.00853984896093607,
-0.02445450983941555,
0.25491997599601746,
-0.04503623768687248,
-0.019035296514630318,
0.139680877327919,
-0.017422720789909363,
0.05111532658338547,
0.13847355544567108,
0.05973135679960251,
-0.1071595847606659,
0.022818701341748238,
0.032120924443006516,
-0.030051909387111664,
-0.19584761559963226,
-0.045068319886922836,
-0.031127620488405228,
-0.014414184726774693,
0.09583587199449539,
0.027750860899686813,
0.026328204199671745,
0.07661371678113937,
0.02302101068198681,
0.0717707946896553,
-0.0069164009764790535,
0.08693669736385345,
0.1133236214518547,
0.0504838228225708,
0.12469679862260818,
-0.03352978825569153,
-0.059310007840394974,
0.03670647367835045,
-0.0006604197551496327,
0.20596711337566376,
0.03016677312552929,
0.09904148429632187,
0.06017816439270973,
0.1803111881017685,
-0.0076091173104941845,
0.07263753563165665,
-0.0056687588803470135,
-0.04295380041003227,
-0.020779583603143692,
-0.047553956508636475,
-0.037906844168901443,
0.05087846517562866,
-0.10652033984661102,
0.07285249978303909,
-0.11005225777626038,
0.025552984327077866,
0.06738075613975525,
0.2621721029281616,
0.047966454178094864,
-0.32677850127220154,
-0.10039372742176056,
0.030644334852695465,
-0.030745234340429306,
-0.02554066851735115,
0.04066818580031395,
0.09871044754981995,
-0.05693161487579346,
0.03975886106491089,
-0.05771370977163315,
0.0751088559627533,
-0.01331237517297268,
0.04728398099541664,
0.05828389525413513,
0.07892735302448273,
0.0032150917686522007,
0.07220000773668289,
-0.2694254517555237,
0.26343995332717896,
0.011488499119877815,
0.07933986186981201,
-0.044527698308229446,
0.006763084325939417,
0.031921058893203735,
0.10620134323835373,
0.08555600792169571,
-0.0080759571865201,
-0.07791926711797714,
-0.2017013281583786,
-0.037526220083236694,
0.02359563298523426,
0.08413002640008926,
-0.0347336083650589,
0.09810551255941391,
-0.041153956204652786,
0.00915823969990015,
0.08361631631851196,
-0.010181009769439697,
-0.09910104423761368,
-0.07467804849147797,
-0.02918543666601181,
0.04214604198932648,
0.005185827612876892,
-0.08957214653491974,
-0.08590573072433472,
-0.12310974299907684,
0.15617193281650543,
-0.036811597645282745,
-0.038150034844875336,
-0.09847129136323929,
0.04180131480097771,
0.05383078008890152,
-0.06981948018074036,
0.05645075440406799,
0.009741130284965038,
0.0889197587966919,
0.02971373125910759,
-0.048070505261421204,
0.12320364266633987,
-0.08799895644187927,
-0.1873549073934555,
-0.07208678871393204,
0.09882476180791855,
0.02324294112622738,
0.03909595310688019,
0.005025947000831366,
0.01955607533454895,
-0.012719051912426949,
-0.08093056827783585,
0.02439913898706436,
-0.008353553712368011,
0.06967110931873322,
0.02718210779130459,
-0.06900159269571304,
0.01015947014093399,
-0.04998829960823059,
-0.035240449011325836,
0.1451808363199234,
0.3069882392883301,
-0.09197865426540375,
0.009503208100795746,
0.07194908708333969,
-0.05461099371314049,
-0.18979224562644958,
0.012971635907888412,
0.025735817849636078,
-0.0129573168233037,
0.06089120730757713,
-0.13936927914619446,
0.1309182345867157,
0.1199093908071518,
-0.03376474604010582,
0.0890478566288948,
-0.2689518928527832,
-0.12716154754161835,
0.1421743631362915,
0.15297603607177734,
0.13222968578338623,
-0.14182570576667786,
-0.02611467055976391,
-0.0364019051194191,
-0.12015119194984436,
0.09619622677564621,
-0.107624351978302,
0.10030547529459,
-0.01495044119656086,
0.056432634592056274,
0.0018836382078006864,
-0.04971945658326149,
0.1327899843454361,
0.00825313851237297,
0.12826389074325562,
-0.06133229285478592,
-0.008096963167190552,
0.04838225618004799,
-0.059256989508867264,
0.0336509607732296,
-0.09220980852842331,
0.05378914251923561,
-0.05451986938714981,
-0.024953413754701614,
-0.0444718599319458,
0.0468134805560112,
-0.029858535155653954,
-0.08292223513126373,
-0.043421246111392975,
0.03160868585109711,
0.05380501598119736,
-0.018161412328481674,
0.15436826646327972,
0.03713298588991165,
0.15274086594581604,
0.14560504257678986,
0.06219593808054924,
-0.08719536662101746,
-0.018686313182115555,
-0.014466426335275173,
-0.038598060607910156,
0.06684046238660812,
-0.1278272569179535,
0.052164651453495026,
0.11318832635879517,
0.009459946304559708,
0.14685866236686707,
0.07362381368875504,
-0.011170122772455215,
0.01128233503550291,
0.06031550467014313,
-0.17487263679504395,
-0.0810227319598198,
-0.0012156148441135883,
-0.01746881939470768,
-0.10632334649562836,
0.0777508094906807,
0.11402619630098343,
-0.08130497485399246,
0.0020504710264503956,
-0.017347324639558792,
0.02293272502720356,
-0.048339854925870895,
0.17591683566570282,
0.05783001706004143,
0.04866271838545799,
-0.08266706764698029,
0.0921567752957344,
0.04301045835018158,
-0.04202159494161606,
0.00014751018898095936,
0.01514324452728033,
-0.09446714073419571,
-0.045501623302698135,
0.0516393780708313,
0.17098021507263184,
-0.04606982320547104,
-0.05615374818444252,
-0.1349552720785141,
-0.11718758195638657,
0.053537677973508835,
0.14565302431583405,
0.11101873964071274,
0.01486610434949398,
-0.02587168477475643,
0.016442513093352318,
-0.10764119029045105,
0.11170166730880737,
0.0365653932094574,
0.09208612143993378,
-0.1772576868534088,
0.1014118567109108,
-0.009178759530186653,
0.005879734642803669,
-0.02458517625927925,
0.045409467071294785,
-0.1113060936331749,
-0.0036588453222066164,
-0.12889176607131958,
-0.014368010684847832,
-0.03524814173579216,
0.01933961920440197,
0.009635369293391705,
-0.06334199756383896,
-0.0554632805287838,
0.026184460148215294,
-0.09961079061031342,
-0.01304047554731369,
0.047537870705127716,
0.06507886946201324,
-0.12184294313192368,
-0.04912494868040085,
0.026639005169272423,
-0.06857496500015259,
0.06687584519386292,
0.029206588864326477,
0.020166773349046707,
0.0493672639131546,
-0.1710042953491211,
0.018984273076057434,
0.07665307819843292,
0.007400264497846365,
0.049326855689287186,
-0.10891158133745193,
-0.02115943469107151,
0.006625928450375795,
0.03192480653524399,
0.012650738470256329,
0.0892520323395729,
-0.13332149386405945,
-0.004011222161352634,
-0.01825386844575405,
-0.07121412456035614,
-0.05344090983271599,
-0.0015669480198994279,
0.10809776186943054,
-0.01056461688131094,
0.22367697954177856,
-0.09042926877737045,
0.006140171550214291,
-0.18989986181259155,
-0.0012438681442290545,
-0.003128704847767949,
-0.1133367121219635,
-0.1559244841337204,
-0.04689471796154976,
0.03352901339530945,
-0.05462853983044624,
0.15590086579322815,
0.002955640433356166,
0.025417355820536613,
0.035037070512771606,
-0.05029686167836189,
0.04078175500035286,
0.02409060299396515,
0.23403795063495636,
0.029671918600797653,
-0.04448271542787552,
0.01930161379277706,
0.0282125286757946,
0.1097055971622467,
0.04460493475198746,
0.15789973735809326,
0.17198482155799866,
-0.05655580759048462,
0.09255146235227585,
0.029999958351254463,
-0.05496017634868622,
-0.17159715294837952,
0.03959345817565918,
-0.01896269991993904,
0.08932711184024811,
-0.013091962784528732,
0.20721513032913208,
0.0897650495171547,
-0.1698092222213745,
0.01791069656610489,
-0.056151412427425385,
-0.0701267421245575,
-0.11000338941812515,
-0.05383274331688881,
-0.09823793172836304,
-0.16599813103675842,
0.0022180110681802034,
-0.11303745955228806,
0.0109088234603405,
0.09969376772642136,
-0.007667968049645424,
-0.014740576036274433,
0.16683132946491241,
-0.011036710813641548,
0.049970824271440506,
0.029718581587076187,
-0.01032729260623455,
-0.04914059489965439,
-0.0730094313621521,
-0.1089337170124054,
0.001703928573988378,
-0.021442972123622894,
0.019484907388687134,
-0.05986665561795235,
-0.0345965176820755,
0.026394298300147057,
-0.009305037558078766,
-0.09364556521177292,
0.014991823583841324,
0.024247370660305023,
0.03748083487153053,
0.024189505726099014,
0.007473574485629797,
0.01539329718798399,
0.007018080912530422,
0.22046516835689545,
-0.07959935069084167,
-0.07805922627449036,
-0.1012873724102974,
0.25053396821022034,
0.04597786068916321,
0.02441229857504368,
0.02720210887491703,
-0.0893549695611,
0.026263175532221794,
0.18869759142398834,
0.16530409455299377,
-0.0737047791481018,
-0.00016772552044130862,
-0.02511722780764103,
-0.019392279908061028,
-0.04666602239012718,
0.08993984013795853,
0.11826059222221375,
-0.006980578880757093,
-0.06920389086008072,
-0.030473198741674423,
-0.050579000264406204,
-0.004438641481101513,
-0.06038161367177963,
0.057670556008815765,
0.02113163098692894,
0.010350197553634644,
-0.05180998891592026,
0.04626348242163658,
-0.02632519230246544,
-0.08992914110422134,
0.05905849486589432,
-0.16942524909973145,
-0.14171715080738068,
-0.008897804655134678,
0.10072347521781921,
-0.005478810053318739,
0.04983534663915634,
-0.03602496534585953,
-0.0061430842615664005,
0.06695302575826645,
-0.024130435660481453,
-0.05059386417269707,
-0.08559592813253403,
0.07338928431272507,
-0.07111379504203796,
0.2550458610057831,
-0.0266066025942564,
0.058620501309633255,
0.12292210757732391,
0.03744770959019661,
-0.08665963262319565,
0.09948249161243439,
0.04940016567707062,
-0.06635244190692902,
0.025941388681530952,
0.07494190335273743,
-0.04062628746032715,
0.13898113369941711,
0.050238098949193954,
-0.16191008687019348,
0.002683680271729827,
-0.016643544659018517,
-0.08825495839118958,
-0.05365462973713875,
-0.039167165756225586,
-0.05548419430851936,
0.1300649642944336,
0.17544642090797424,
-0.04247504100203514,
0.004750636871904135,
-0.05398782715201378,
0.05119873955845833,
0.0759635791182518,
0.016714058816432953,
-0.027893343940377235,
-0.2376081496477127,
0.03754596412181854,
0.10090658068656921,
-0.011796272359788418,
-0.23678900301456451,
-0.09794989973306656,
-0.009187194518744946,
-0.050893787294626236,
-0.09158074110746384,
0.08113052695989609,
0.11271882057189941,
0.05037638917565346,
-0.06506321579217911,
-0.10472185164690018,
-0.08860750496387482,
0.15171939134597778,
-0.11172042787075043,
-0.09994826465845108
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | phamtungthuy/quantized_law_model_merged | [
"transformers",
"safetensors",
"mpt",
"text-generation",
"conversational",
"custom_code",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"4-bit",
"region:us"
] | 2024-02-07T09:01:16+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #mpt #text-generation #conversational #custom_code #arxiv-1910.09700 #autotrain_compatible #text-generation-inference #4-bit #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #mpt #text-generation #conversational #custom_code #arxiv-1910.09700 #autotrain_compatible #text-generation-inference #4-bit #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
60,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #mpt #text-generation #conversational #custom_code #arxiv-1910.09700 #autotrain_compatible #text-generation-inference #4-bit #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.046836815774440765,
0.1647104024887085,
-0.00548818614333868,
0.01897842437028885,
0.09896454960107803,
0.012689607217907906,
0.06120520457625389,
0.11300846934318542,
-0.028905285522341728,
0.11334774643182755,
0.03255376219749451,
0.09909575432538986,
0.11503656953573227,
0.1519048511981964,
-0.0030024880543351173,
-0.22314703464508057,
0.051402922719717026,
-0.12195511162281036,
-0.03322732448577881,
0.11792048811912537,
0.14752967655658722,
-0.09863631427288055,
0.07304353266954422,
-0.03148203343153,
-0.013285660184919834,
-0.030106328427791595,
-0.05744732543826103,
-0.044162191450595856,
0.04502193257212639,
0.06684926897287369,
0.06690709292888641,
0.007300524041056633,
0.0916643962264061,
-0.2659870982170105,
0.01935933716595173,
0.07323223352432251,
-0.0024301528465002775,
0.07196123898029327,
0.06251558661460876,
-0.06797171384096146,
0.09986966103315353,
-0.04709329083561897,
0.14694355428218842,
0.08412910252809525,
-0.09047382324934006,
-0.19434987008571625,
-0.08776694536209106,
0.0995001271367073,
0.176052063703537,
0.04804452508687973,
-0.022804055362939835,
0.09717879444360733,
-0.09082359820604324,
0.013806850649416447,
0.047815605998039246,
-0.07298199832439423,
-0.05219341814517975,
0.06204948574304581,
0.07889563590288162,
0.07410836964845657,
-0.12416650354862213,
-0.03171328827738762,
0.006749303545802832,
0.011189286597073078,
0.07576481997966766,
0.020296677947044373,
0.15167999267578125,
0.035466551780700684,
-0.1285610944032669,
-0.051411163061857224,
0.10339448601007462,
0.035137519240379333,
-0.045140668749809265,
-0.24473509192466736,
-0.03564957156777382,
-0.03853682428598404,
-0.027864262461662292,
-0.03630993142724037,
0.041958823800086975,
-0.004923766013234854,
0.0842767283320427,
-0.005395758431404829,
-0.07268313318490982,
-0.04217590391635895,
0.06500858813524246,
0.06705322861671448,
0.03039911389350891,
-0.011621474288403988,
0.011928488500416279,
0.11858756840229034,
0.11163737624883652,
-0.12085368484258652,
-0.05400480702519417,
-0.06527067720890045,
-0.0827568918466568,
-0.03957102820277214,
0.0360938161611557,
0.034597333520650864,
0.05289698392152786,
0.2536163628101349,
0.011734632775187492,
0.062413059175014496,
0.03360540047287941,
0.008391516283154488,
0.05624012649059296,
0.11009805649518967,
-0.061210669577121735,
-0.10028582066297531,
-0.02381979301571846,
0.09070323407649994,
0.009525741450488567,
-0.035405855625867844,
-0.05315864458680153,
0.06331110745668411,
0.019456159323453903,
0.11936915665864944,
0.08937534689903259,
0.009933953173458576,
-0.07105469703674316,
-0.06267762184143066,
0.1954593062400818,
-0.16626228392124176,
0.04384920746088028,
0.03799726068973541,
-0.03902040049433708,
0.00011464420094853267,
0.015001960098743439,
0.015850180760025978,
-0.026051290333271027,
0.08616644144058228,
-0.054621897637844086,
-0.042427536100149155,
-0.1131248027086258,
-0.03744876757264137,
0.03138575330376625,
0.01295038964599371,
-0.03690009564161301,
-0.03263045847415924,
-0.08063790947198868,
-0.07437901198863983,
0.0934433713555336,
-0.07256671786308289,
-0.04976126551628113,
-0.01708609238266945,
-0.07826544344425201,
0.02226773276925087,
0.02073119394481182,
0.07654669135808945,
-0.023182015866041183,
0.04522136598825455,
-0.051821496337652206,
0.05846284702420235,
0.11267654597759247,
0.03567348048090935,
-0.057637013494968414,
0.06343067437410355,
-0.2405025213956833,
0.09473377466201782,
-0.07224822789430618,
0.054827991873025894,
-0.1513880044221878,
-0.025180181488394737,
0.04511411860585213,
0.006036554928869009,
-0.010356215760111809,
0.13423052430152893,
-0.21411681175231934,
-0.02655753120779991,
0.1692247837781906,
-0.10130050033330917,
-0.06695982813835144,
0.057127393782138824,
-0.047052353620529175,
0.10467110574245453,
0.03668500483036041,
-0.028671640902757645,
0.06196305900812149,
-0.12541428208351135,
0.006904494483023882,
-0.0474662147462368,
-0.018319565802812576,
0.15615510940551758,
0.07458817958831787,
-0.07195501774549484,
0.06689044088125229,
0.02665250189602375,
-0.027468763291835785,
-0.046108439564704895,
-0.014795416034758091,
-0.10463820397853851,
0.01510506309568882,
-0.061047863215208054,
0.016948029398918152,
-0.021878130733966827,
-0.09516984224319458,
-0.028752289712429047,
-0.17679062485694885,
-0.03173348307609558,
0.08809247612953186,
-0.00988912396132946,
-0.017979109659790993,
-0.11845765262842178,
0.010264944285154343,
0.037062644958496094,
0.006667076610028744,
-0.13313348591327667,
-0.041100483387708664,
0.029788121581077576,
-0.16209228336811066,
0.034110672771930695,
-0.050082575529813766,
0.05046985298395157,
0.029048535972833633,
-0.028581300750374794,
-0.02944372594356537,
0.018678095191717148,
0.005221392028033733,
-0.009069527499377728,
-0.24494226276874542,
-0.029461096972227097,
-0.024158140644431114,
0.17405138909816742,
-0.21141023933887482,
0.03583250567317009,
0.07916480302810669,
0.15132445096969604,
0.012238884344696999,
-0.042015112936496735,
0.003934313077479601,
-0.0704701766371727,
-0.029022127389907837,
-0.058021429926157,
-0.00698183523491025,
-0.031215690076351166,
-0.045432448387145996,
0.04434436559677124,
-0.16684353351593018,
-0.03446514531970024,
0.09774104505777359,
0.06061384081840515,
-0.13769705593585968,
-0.019792277365922928,
-0.03615017235279083,
-0.04814789816737175,
-0.04649418219923973,
-0.0635431632399559,
0.10675375908613205,
0.05744480714201927,
0.047778159379959106,
-0.05886402353644371,
-0.07373563200235367,
-0.0018592822598293424,
-0.015227645635604858,
-0.017503464594483376,
0.0970890074968338,
0.07871697843074799,
-0.12881506979465485,
0.0925796777009964,
0.09740415215492249,
0.07862301915884018,
0.09160694479942322,
-0.023659368976950645,
-0.08283168822526932,
-0.045791640877723694,
0.03224967047572136,
0.019317863509058952,
0.12954181432724,
-0.022851284593343735,
0.045737724751234055,
0.04265662282705307,
-0.01701509580016136,
0.01527531910687685,
-0.08818325400352478,
0.03562496602535248,
0.030209563672542572,
-0.018250390887260437,
0.04657670110464096,
-0.043136533349752426,
0.022899577394127846,
0.08880213648080826,
0.04652043431997299,
0.03540687635540962,
0.016735758632421494,
-0.051466867327690125,
-0.10955886542797089,
0.16343532502651215,
-0.13016492128372192,
-0.22333437204360962,
-0.14305031299591064,
0.003047437872737646,
0.03404908999800682,
-0.013913608156144619,
0.0010318867862224579,
-0.06298820674419403,
-0.12252139300107956,
-0.08956659585237503,
0.01577884890139103,
0.0467185415327549,
-0.09116512537002563,
-0.06381972134113312,
0.057438019663095474,
0.041743889451026917,
-0.14711973071098328,
0.019101250916719437,
0.049786824733018875,
-0.0982508584856987,
-0.005891401786357164,
0.07870857417583466,
0.06842125952243805,
0.1831832230091095,
0.018143242225050926,
-0.021282153204083443,
0.03438962996006012,
0.2074076384305954,
-0.13331963121891022,
0.11120434105396271,
0.13762928545475006,
-0.09217964857816696,
0.08102559298276901,
0.20039579272270203,
0.04004323109984398,
-0.0997694879770279,
0.03314325585961342,
0.02003711275756359,
-0.030200202018022537,
-0.2437104880809784,
-0.0713467001914978,
-0.0037302353885024786,
-0.056442990899086,
0.07561111450195312,
0.09344951063394547,
0.08132704347372055,
0.011333314701914787,
-0.09457346796989441,
-0.08159025013446808,
0.06176929920911789,
0.10247852653265,
0.020764078944921494,
-0.013180595822632313,
0.08917813003063202,
-0.03367619216442108,
0.019887862727046013,
0.08727820962667465,
0.003822416765615344,
0.1769154965877533,
0.052071988582611084,
0.19217601418495178,
0.0787898451089859,
0.07264366745948792,
0.013349482789635658,
0.013965628109872341,
0.019131729379296303,
0.03359703719615936,
-0.003047162899747491,
-0.08534960448741913,
-0.023618679493665695,
0.11475063860416412,
0.06807617843151093,
0.016076065599918365,
0.01694929040968418,
-0.041291743516922,
0.08016949892044067,
0.17850646376609802,
-0.006623215973377228,
-0.1825191080570221,
-0.05808044224977493,
0.07510574162006378,
-0.09590929001569748,
-0.0984455943107605,
-0.01443559117615223,
0.032915156334638596,
-0.16824102401733398,
0.02115572802722454,
-0.021102940663695335,
0.11152804642915726,
-0.1400838941335678,
-0.021082382649183273,
0.07384687662124634,
0.07785172760486603,
0.004180863033980131,
0.05667582526803017,
-0.16176272928714752,
0.1015118807554245,
0.009188702329993248,
0.06620164960622787,
-0.09429202228784561,
0.10198913514614105,
-0.005289927590638399,
-0.015678511932492256,
0.13372057676315308,
0.004263219889253378,
-0.0682830959558487,
-0.08049130439758301,
-0.08794789016246796,
-0.008455827832221985,
0.12851466238498688,
-0.14218972623348236,
0.08705435693264008,
-0.038971856236457825,
-0.041753776371479034,
-0.0021535323467105627,
-0.10083090513944626,
-0.11627402901649475,
-0.1857561320066452,
0.0577387772500515,
-0.13709378242492676,
0.03257451951503754,
-0.10875391960144043,
-0.030822061002254486,
-0.02859305404126644,
0.18933424353599548,
-0.23575855791568756,
-0.07154747098684311,
-0.14682282507419586,
-0.10252786427736282,
0.14261676371097565,
-0.049167271703481674,
0.08533598482608795,
-0.00865695346146822,
0.17063072323799133,
0.02125515043735504,
-0.025427155196666718,
0.09796741604804993,
-0.09066979587078094,
-0.19584842026233673,
-0.0780235081911087,
0.15831086039543152,
0.13357588648796082,
0.03327145054936409,
-0.004567822441458702,
0.03751498460769653,
-0.0175956841558218,
-0.12013448774814606,
0.016311630606651306,
0.17807579040527344,
0.05787086486816406,
0.026613978669047356,
-0.027188660576939583,
-0.11137328296899796,
-0.06709632277488708,
-0.026733320206403732,
0.02881023660302162,
0.17920275032520294,
-0.06998284161090851,
0.1801978051662445,
0.14515306055545807,
-0.05784882232546806,
-0.2073381245136261,
0.010634344071149826,
0.031087007373571396,
0.0026492057368159294,
0.026879357174038887,
-0.20153191685676575,
0.08359402418136597,
0.003537527984008193,
-0.05130414664745331,
0.13254192471504211,
-0.17374180257320404,
-0.142324298620224,
0.07576381415128708,
0.03726530447602272,
-0.19563426077365875,
-0.12638996541500092,
-0.09056314080953598,
-0.04918690025806427,
-0.18773697316646576,
0.09192302078008652,
0.028096789494156837,
0.01022305991500616,
0.03053116239607334,
0.025923538953065872,
0.018353641033172607,
-0.04306640103459358,
0.1853063702583313,
-0.02672814577817917,
0.02791212871670723,
-0.08583202213048935,
-0.06699703633785248,
0.03955686837434769,
-0.052560221403837204,
0.07334768027067184,
-0.024697571992874146,
0.00856629665941,
-0.10168556869029999,
-0.04047643393278122,
-0.035936903208494186,
0.013404258526861668,
-0.09548021107912064,
-0.08607310801744461,
-0.04715096205472946,
0.09425225108861923,
0.09389013051986694,
-0.0337766595184803,
-0.03096645325422287,
-0.07412707805633545,
0.0430193655192852,
0.2041751891374588,
0.17465740442276,
0.038705646991729736,
-0.07965444028377533,
-0.002935659373179078,
-0.010335927829146385,
0.04545985907316208,
-0.20781102776527405,
0.06431519240140915,
0.05237681791186333,
0.019995132461190224,
0.11615028977394104,
-0.018933145329356194,
-0.15544183552265167,
-0.07166822254657745,
0.06284844875335693,
-0.05964977666735649,
-0.1960112601518631,
0.007406753953546286,
0.05805834010243416,
-0.1648537814617157,
-0.044738590717315674,
0.04522307962179184,
-0.0060577839612960815,
-0.037425119429826736,
0.021866729483008385,
0.09298396110534668,
0.0013969119172543287,
0.07521650940179825,
0.057380616664886475,
0.08283612132072449,
-0.10633254796266556,
0.08069998025894165,
0.08765646070241928,
-0.08192311972379684,
0.02060014009475708,
0.10195332020521164,
-0.06092626228928566,
-0.0331430546939373,
0.02062523178756237,
0.08296512067317963,
0.023787984624505043,
-0.04053160175681114,
0.01279789675027132,
-0.10569491982460022,
0.0591636560857296,
0.07772862911224365,
0.032817769795656204,
0.01771099865436554,
0.03410870581865311,
0.04761794954538345,
-0.07257413119077682,
0.12228159606456757,
0.029085364192724228,
0.016395291313529015,
-0.03863071650266647,
-0.040404897183179855,
0.020513348281383514,
-0.03261692449450493,
-0.005556173622608185,
-0.030997928231954575,
-0.07566902041435242,
-0.014217384159564972,
-0.16211988031864166,
-0.013971314765512943,
-0.05232914909720421,
0.01099184900522232,
0.03223292529582977,
-0.03091696836054325,
0.007455316837877035,
0.011223319917917252,
-0.07765290886163712,
-0.06736263632774353,
-0.020429493859410286,
0.09177552908658981,
-0.1594531089067459,
0.02477598562836647,
0.08419981598854065,
-0.12327192723751068,
0.09286192804574966,
0.017041858285665512,
-0.006877784617245197,
0.027240877971053123,
-0.14548709988594055,
0.033531397581100464,
-0.03081953153014183,
0.011877832002937794,
0.04307198151946068,
-0.2145346701145172,
0.0011035674251616001,
-0.03445719555020332,
-0.0721113309264183,
-0.010611768811941147,
-0.03385230526328087,
-0.11504104733467102,
0.10733716189861298,
0.004277069121599197,
-0.08941003680229187,
-0.03095969185233116,
0.03161302208900452,
0.08224654942750931,
-0.019783973693847656,
0.15552936494350433,
-0.00823113601654768,
0.07357944548130035,
-0.16393354535102844,
-0.02081993967294693,
-0.005984647665172815,
0.021715238690376282,
-0.01979609951376915,
-0.00966197345405817,
0.043640561401844025,
-0.02377203479409218,
0.17872008681297302,
-0.028821680694818497,
0.01852906309068203,
0.0672251284122467,
0.024379827082157135,
-0.03164516016840935,
0.10441958159208298,
0.04121462628245354,
0.020564084872603416,
0.019512219354510307,
0.0012760780518874526,
-0.04783459007740021,
-0.030089499428868294,
-0.19694079458713531,
0.06752467155456543,
0.15307112038135529,
0.09699810296297073,
-0.018511807546019554,
0.08039021492004395,
-0.10178884863853455,
-0.11259745806455612,
0.1282721906900406,
-0.04827006161212921,
-0.0005967059987597167,
-0.07089807838201523,
0.13272404670715332,
0.14681458473205566,
-0.1830657720565796,
0.07114480435848236,
-0.06873263418674469,
-0.04703962802886963,
-0.11192386597394943,
-0.19347000122070312,
-0.05595254525542259,
-0.05544281005859375,
-0.01759531907737255,
-0.04516579583287239,
0.07355371862649918,
0.05612892284989357,
0.006139043252915144,
-0.0019067827379330993,
0.06639719754457474,
-0.030526932328939438,
-0.0062928348779678345,
0.026433905586600304,
0.06542899459600449,
0.010686502791941166,
-0.02786833606660366,
0.014867307618260384,
-0.0063516986556351185,
0.04390629008412361,
0.06825330853462219,
0.04959937557578087,
-0.031064428389072418,
0.016983702778816223,
-0.03989924490451813,
-0.10337188839912415,
0.04382622241973877,
-0.025124363601207733,
-0.08256535232067108,
0.15559422969818115,
0.021912720054388046,
0.006917151156812906,
-0.02032097801566124,
0.23542065918445587,
-0.07307235896587372,
-0.09446928650140762,
-0.14866788685321808,
0.10649941861629486,
-0.0372060127556324,
0.06093543767929077,
0.046831000596284866,
-0.10337284952402115,
0.017050903290510178,
0.12904027104377747,
0.1646832376718521,
-0.039227113127708435,
0.019966665655374527,
0.03195488452911377,
0.005217135418206453,
-0.037870142608881,
0.051399193704128265,
0.06997720152139664,
0.16103850305080414,
-0.04391693323850632,
0.09321633726358414,
0.00042109034257009625,
-0.09568869322538376,
-0.03787053003907204,
0.10844326764345169,
-0.01475430652499199,
0.018231717869639397,
-0.057804450392723083,
0.11598270386457443,
-0.05992904305458069,
-0.23697659373283386,
0.06294666975736618,
-0.0641324520111084,
-0.1347293257713318,
-0.024605415761470795,
0.08455070853233337,
-0.012824739329516888,
0.02644866518676281,
0.07231196016073227,
-0.07240470498800278,
0.20365303754806519,
0.03693865239620209,
-0.05348410829901695,
-0.05722471699118614,
0.08170948922634125,
-0.10032642632722855,
0.27315396070480347,
0.01665438525378704,
0.045854128897190094,
0.10315130650997162,
-0.009667059406638145,
-0.1381330043077469,
0.02517891675233841,
0.09495221078395844,
-0.1030079573392868,
0.039407141506671906,
0.19656206667423248,
-0.0018554746638983488,
0.1250673085451126,
0.07593108713626862,
-0.07669542729854584,
0.04671437293291092,
-0.08359885215759277,
-0.06893200427293777,
-0.09199898689985275,
0.09549407660961151,
-0.07781421393156052,
0.14201928675174713,
0.1377091109752655,
-0.050289053469896317,
0.0130878034979105,
-0.03416995704174042,
0.04503699764609337,
0.00027301054797135293,
0.10270082205533981,
0.007354200817644596,
-0.18405498564243317,
0.023653801530599594,
0.002919376129284501,
0.10706783831119537,
-0.15832123160362244,
-0.09442342072725296,
0.04106706753373146,
0.0052554975263774395,
-0.06749075651168823,
0.12818290293216705,
0.05500555783510208,
0.04493929073214531,
-0.041192516684532166,
-0.029329955577850342,
-0.006561505142599344,
0.1352681815624237,
-0.10633693635463715,
0.0014152000658214092
] |
null | null | transformers | # Model Card
## Summary
This model was trained using [H2O LLM Studio](https://github.com/h2oai/h2o-llmstudio).
- Base model: [h2oai/h2ogpt-4096-llama2-7b](https://huggingface.co/h2oai/h2ogpt-4096-llama2-7b)
## Usage
To use the model with the `transformers` library on a machine with GPUs, first make sure you have the `transformers` library installed.
```bash
pip install transformers==4.36.1
```
Also make sure you are providing your huggingface token to the pipeline if the model is lying in a private repo.
- Either leave `token=True` in the `pipeline` and login to hugginface_hub by running
```python
import huggingface_hub
huggingface_hub.login(<ACCESS_TOKEN>)
```
- Or directly pass your <ACCESS_TOKEN> to `token` in the `pipeline`
```python
from transformers import pipeline
generate_text = pipeline(
model="audreyleteve/curious-muskox",
torch_dtype="auto",
trust_remote_code=True,
use_fast=True,
device_map={"": "cuda:0"},
token=True,
)
res = generate_text(
"Why is drinking water so healthy?",
min_new_tokens=2,
max_new_tokens=256,
do_sample=False,
num_beams=1,
temperature=float(0.0),
repetition_penalty=float(1.2),
renormalize_logits=True
)
print(res[0]["generated_text"])
```
You can print a sample prompt after the preprocessing step to see how it is feed to the tokenizer:
```python
print(generate_text.preprocess("Why is drinking water so healthy?")["prompt_text"])
```
```bash
<|prompt|>Why is drinking water so healthy?</s><|answer|>
```
Alternatively, you can download [h2oai_pipeline.py](h2oai_pipeline.py), store it alongside your notebook, and construct the pipeline yourself from the loaded model and tokenizer. If the model and the tokenizer are fully supported in the `transformers` package, this will allow you to set `trust_remote_code=False`.
```python
from h2oai_pipeline import H2OTextGenerationPipeline
from transformers import AutoModelForCausalLM, AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained(
"audreyleteve/curious-muskox",
use_fast=True,
padding_side="left",
trust_remote_code=True,
)
model = AutoModelForCausalLM.from_pretrained(
"audreyleteve/curious-muskox",
torch_dtype="auto",
device_map={"": "cuda:0"},
trust_remote_code=True,
)
generate_text = H2OTextGenerationPipeline(model=model, tokenizer=tokenizer)
res = generate_text(
"Why is drinking water so healthy?",
min_new_tokens=2,
max_new_tokens=256,
do_sample=False,
num_beams=1,
temperature=float(0.0),
repetition_penalty=float(1.2),
renormalize_logits=True
)
print(res[0]["generated_text"])
```
You may also construct the pipeline from the loaded model and tokenizer yourself and consider the preprocessing steps:
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_name = "audreyleteve/curious-muskox" # either local folder or huggingface model name
# Important: The prompt needs to be in the same format the model was trained with.
# You can find an example prompt in the experiment logs.
prompt = "<|prompt|>How are you?</s><|answer|>"
tokenizer = AutoTokenizer.from_pretrained(
model_name,
use_fast=True,
trust_remote_code=True,
)
model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype="auto",
device_map={"": "cuda:0"},
trust_remote_code=True,
)
model.cuda().eval()
inputs = tokenizer(prompt, return_tensors="pt", add_special_tokens=False).to("cuda")
# generate configuration can be modified to your needs
tokens = model.generate(
input_ids=inputs["input_ids"],
attention_mask=inputs["attention_mask"],
min_new_tokens=2,
max_new_tokens=256,
do_sample=False,
num_beams=1,
temperature=float(0.0),
repetition_penalty=float(1.2),
renormalize_logits=True
)[0]
tokens = tokens[inputs["input_ids"].shape[1]:]
answer = tokenizer.decode(tokens, skip_special_tokens=True)
print(answer)
```
## Quantization and sharding
You can load the models using quantization by specifying ```load_in_8bit=True``` or ```load_in_4bit=True```. Also, sharding on multiple GPUs is possible by setting ```device_map=auto```.
## Model Architecture
```
LlamaForCausalLM(
(model): LlamaModel(
(embed_tokens): Embedding(32000, 4096, padding_idx=0)
(layers): ModuleList(
(0-31): 32 x LlamaDecoderLayer(
(self_attn): LlamaAttention(
(q_proj): Linear(in_features=4096, out_features=4096, bias=False)
(k_proj): Linear(in_features=4096, out_features=4096, bias=False)
(v_proj): Linear(in_features=4096, out_features=4096, bias=False)
(o_proj): Linear(in_features=4096, out_features=4096, bias=False)
(rotary_emb): LlamaRotaryEmbedding()
)
(mlp): LlamaMLP(
(gate_proj): Linear(in_features=4096, out_features=11008, bias=False)
(up_proj): Linear(in_features=4096, out_features=11008, bias=False)
(down_proj): Linear(in_features=11008, out_features=4096, bias=False)
(act_fn): SiLU()
)
(input_layernorm): LlamaRMSNorm()
(post_attention_layernorm): LlamaRMSNorm()
)
)
(norm): LlamaRMSNorm()
)
(lm_head): Linear(in_features=4096, out_features=32000, bias=False)
)
```
## Model Configuration
This model was trained using H2O LLM Studio and with the configuration in [cfg.yaml](cfg.yaml). Visit [H2O LLM Studio](https://github.com/h2oai/h2o-llmstudio) to learn how to train your own large language models.
## Disclaimer
Please read this disclaimer carefully before using the large language model provided in this repository. Your use of the model signifies your agreement to the following terms and conditions.
- Biases and Offensiveness: The large language model is trained on a diverse range of internet text data, which may contain biased, racist, offensive, or otherwise inappropriate content. By using this model, you acknowledge and accept that the generated content may sometimes exhibit biases or produce content that is offensive or inappropriate. The developers of this repository do not endorse, support, or promote any such content or viewpoints.
- Limitations: The large language model is an AI-based tool and not a human. It may produce incorrect, nonsensical, or irrelevant responses. It is the user's responsibility to critically evaluate the generated content and use it at their discretion.
- Use at Your Own Risk: Users of this large language model must assume full responsibility for any consequences that may arise from their use of the tool. The developers and contributors of this repository shall not be held liable for any damages, losses, or harm resulting from the use or misuse of the provided model.
- Ethical Considerations: Users are encouraged to use the large language model responsibly and ethically. By using this model, you agree not to use it for purposes that promote hate speech, discrimination, harassment, or any form of illegal or harmful activities.
- Reporting Issues: If you encounter any biased, offensive, or otherwise inappropriate content generated by the large language model, please report it to the repository maintainers through the provided channels. Your feedback will help improve the model and mitigate potential issues.
- Changes to this Disclaimer: The developers of this repository reserve the right to modify or update this disclaimer at any time without prior notice. It is the user's responsibility to periodically review the disclaimer to stay informed about any changes.
By using the large language model provided in this repository, you agree to accept and comply with the terms and conditions outlined in this disclaimer. If you do not agree with any part of this disclaimer, you should refrain from using the model and any content generated by it. | {"language": ["en"], "library_name": "transformers", "tags": ["gpt", "llm", "large language model", "h2o-llmstudio"], "inference": false, "thumbnail": "https://h2o.ai/etc.clientlibs/h2o/clientlibs/clientlib-site/resources/images/favicon.ico"} | text-generation | audreyleteve/curious-muskox | [
"transformers",
"safetensors",
"llama",
"text-generation",
"gpt",
"llm",
"large language model",
"h2o-llmstudio",
"en",
"autotrain_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T09:02:07+00:00 | [] | [
"en"
] | TAGS
#transformers #safetensors #llama #text-generation #gpt #llm #large language model #h2o-llmstudio #en #autotrain_compatible #text-generation-inference #region-us
| # Model Card
## Summary
This model was trained using H2O LLM Studio.
- Base model: h2oai/h2ogpt-4096-llama2-7b
## Usage
To use the model with the 'transformers' library on a machine with GPUs, first make sure you have the 'transformers' library installed.
Also make sure you are providing your huggingface token to the pipeline if the model is lying in a private repo.
- Either leave 'token=True' in the 'pipeline' and login to hugginface_hub by running
- Or directly pass your <ACCESS_TOKEN> to 'token' in the 'pipeline'
You can print a sample prompt after the preprocessing step to see how it is feed to the tokenizer:
Alternatively, you can download h2oai_pipeline.py, store it alongside your notebook, and construct the pipeline yourself from the loaded model and tokenizer. If the model and the tokenizer are fully supported in the 'transformers' package, this will allow you to set 'trust_remote_code=False'.
You may also construct the pipeline from the loaded model and tokenizer yourself and consider the preprocessing steps:
## Quantization and sharding
You can load the models using quantization by specifying or . Also, sharding on multiple GPUs is possible by setting .
## Model Architecture
## Model Configuration
This model was trained using H2O LLM Studio and with the configuration in URL. Visit H2O LLM Studio to learn how to train your own large language models.
## Disclaimer
Please read this disclaimer carefully before using the large language model provided in this repository. Your use of the model signifies your agreement to the following terms and conditions.
- Biases and Offensiveness: The large language model is trained on a diverse range of internet text data, which may contain biased, racist, offensive, or otherwise inappropriate content. By using this model, you acknowledge and accept that the generated content may sometimes exhibit biases or produce content that is offensive or inappropriate. The developers of this repository do not endorse, support, or promote any such content or viewpoints.
- Limitations: The large language model is an AI-based tool and not a human. It may produce incorrect, nonsensical, or irrelevant responses. It is the user's responsibility to critically evaluate the generated content and use it at their discretion.
- Use at Your Own Risk: Users of this large language model must assume full responsibility for any consequences that may arise from their use of the tool. The developers and contributors of this repository shall not be held liable for any damages, losses, or harm resulting from the use or misuse of the provided model.
- Ethical Considerations: Users are encouraged to use the large language model responsibly and ethically. By using this model, you agree not to use it for purposes that promote hate speech, discrimination, harassment, or any form of illegal or harmful activities.
- Reporting Issues: If you encounter any biased, offensive, or otherwise inappropriate content generated by the large language model, please report it to the repository maintainers through the provided channels. Your feedback will help improve the model and mitigate potential issues.
- Changes to this Disclaimer: The developers of this repository reserve the right to modify or update this disclaimer at any time without prior notice. It is the user's responsibility to periodically review the disclaimer to stay informed about any changes.
By using the large language model provided in this repository, you agree to accept and comply with the terms and conditions outlined in this disclaimer. If you do not agree with any part of this disclaimer, you should refrain from using the model and any content generated by it. | [
"# Model Card",
"## Summary\n\nThis model was trained using H2O LLM Studio.\n- Base model: h2oai/h2ogpt-4096-llama2-7b",
"## Usage\n\nTo use the model with the 'transformers' library on a machine with GPUs, first make sure you have the 'transformers' library installed.\n\n\n\nAlso make sure you are providing your huggingface token to the pipeline if the model is lying in a private repo.\n - Either leave 'token=True' in the 'pipeline' and login to hugginface_hub by running\n \n - Or directly pass your <ACCESS_TOKEN> to 'token' in the 'pipeline'\n\n\n\nYou can print a sample prompt after the preprocessing step to see how it is feed to the tokenizer:\n\n\n\n\n\nAlternatively, you can download h2oai_pipeline.py, store it alongside your notebook, and construct the pipeline yourself from the loaded model and tokenizer. If the model and the tokenizer are fully supported in the 'transformers' package, this will allow you to set 'trust_remote_code=False'.\n\n\n\n\nYou may also construct the pipeline from the loaded model and tokenizer yourself and consider the preprocessing steps:",
"## Quantization and sharding\n\nYou can load the models using quantization by specifying or . Also, sharding on multiple GPUs is possible by setting .",
"## Model Architecture",
"## Model Configuration\n\nThis model was trained using H2O LLM Studio and with the configuration in URL. Visit H2O LLM Studio to learn how to train your own large language models.",
"## Disclaimer\n\nPlease read this disclaimer carefully before using the large language model provided in this repository. Your use of the model signifies your agreement to the following terms and conditions.\n\n- Biases and Offensiveness: The large language model is trained on a diverse range of internet text data, which may contain biased, racist, offensive, or otherwise inappropriate content. By using this model, you acknowledge and accept that the generated content may sometimes exhibit biases or produce content that is offensive or inappropriate. The developers of this repository do not endorse, support, or promote any such content or viewpoints.\n- Limitations: The large language model is an AI-based tool and not a human. It may produce incorrect, nonsensical, or irrelevant responses. It is the user's responsibility to critically evaluate the generated content and use it at their discretion.\n- Use at Your Own Risk: Users of this large language model must assume full responsibility for any consequences that may arise from their use of the tool. The developers and contributors of this repository shall not be held liable for any damages, losses, or harm resulting from the use or misuse of the provided model.\n- Ethical Considerations: Users are encouraged to use the large language model responsibly and ethically. By using this model, you agree not to use it for purposes that promote hate speech, discrimination, harassment, or any form of illegal or harmful activities.\n- Reporting Issues: If you encounter any biased, offensive, or otherwise inappropriate content generated by the large language model, please report it to the repository maintainers through the provided channels. Your feedback will help improve the model and mitigate potential issues.\n- Changes to this Disclaimer: The developers of this repository reserve the right to modify or update this disclaimer at any time without prior notice. It is the user's responsibility to periodically review the disclaimer to stay informed about any changes.\n\nBy using the large language model provided in this repository, you agree to accept and comply with the terms and conditions outlined in this disclaimer. If you do not agree with any part of this disclaimer, you should refrain from using the model and any content generated by it."
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #gpt #llm #large language model #h2o-llmstudio #en #autotrain_compatible #text-generation-inference #region-us \n",
"# Model Card",
"## Summary\n\nThis model was trained using H2O LLM Studio.\n- Base model: h2oai/h2ogpt-4096-llama2-7b",
"## Usage\n\nTo use the model with the 'transformers' library on a machine with GPUs, first make sure you have the 'transformers' library installed.\n\n\n\nAlso make sure you are providing your huggingface token to the pipeline if the model is lying in a private repo.\n - Either leave 'token=True' in the 'pipeline' and login to hugginface_hub by running\n \n - Or directly pass your <ACCESS_TOKEN> to 'token' in the 'pipeline'\n\n\n\nYou can print a sample prompt after the preprocessing step to see how it is feed to the tokenizer:\n\n\n\n\n\nAlternatively, you can download h2oai_pipeline.py, store it alongside your notebook, and construct the pipeline yourself from the loaded model and tokenizer. If the model and the tokenizer are fully supported in the 'transformers' package, this will allow you to set 'trust_remote_code=False'.\n\n\n\n\nYou may also construct the pipeline from the loaded model and tokenizer yourself and consider the preprocessing steps:",
"## Quantization and sharding\n\nYou can load the models using quantization by specifying or . Also, sharding on multiple GPUs is possible by setting .",
"## Model Architecture",
"## Model Configuration\n\nThis model was trained using H2O LLM Studio and with the configuration in URL. Visit H2O LLM Studio to learn how to train your own large language models.",
"## Disclaimer\n\nPlease read this disclaimer carefully before using the large language model provided in this repository. Your use of the model signifies your agreement to the following terms and conditions.\n\n- Biases and Offensiveness: The large language model is trained on a diverse range of internet text data, which may contain biased, racist, offensive, or otherwise inappropriate content. By using this model, you acknowledge and accept that the generated content may sometimes exhibit biases or produce content that is offensive or inappropriate. The developers of this repository do not endorse, support, or promote any such content or viewpoints.\n- Limitations: The large language model is an AI-based tool and not a human. It may produce incorrect, nonsensical, or irrelevant responses. It is the user's responsibility to critically evaluate the generated content and use it at their discretion.\n- Use at Your Own Risk: Users of this large language model must assume full responsibility for any consequences that may arise from their use of the tool. The developers and contributors of this repository shall not be held liable for any damages, losses, or harm resulting from the use or misuse of the provided model.\n- Ethical Considerations: Users are encouraged to use the large language model responsibly and ethically. By using this model, you agree not to use it for purposes that promote hate speech, discrimination, harassment, or any form of illegal or harmful activities.\n- Reporting Issues: If you encounter any biased, offensive, or otherwise inappropriate content generated by the large language model, please report it to the repository maintainers through the provided channels. Your feedback will help improve the model and mitigate potential issues.\n- Changes to this Disclaimer: The developers of this repository reserve the right to modify or update this disclaimer at any time without prior notice. It is the user's responsibility to periodically review the disclaimer to stay informed about any changes.\n\nBy using the large language model provided in this repository, you agree to accept and comply with the terms and conditions outlined in this disclaimer. If you do not agree with any part of this disclaimer, you should refrain from using the model and any content generated by it."
] | [
60,
3,
37,
244,
34,
4,
42,
518
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #gpt #llm #large language model #h2o-llmstudio #en #autotrain_compatible #text-generation-inference #region-us \n# Model Card## Summary\n\nThis model was trained using H2O LLM Studio.\n- Base model: h2oai/h2ogpt-4096-llama2-7b## Usage\n\nTo use the model with the 'transformers' library on a machine with GPUs, first make sure you have the 'transformers' library installed.\n\n\n\nAlso make sure you are providing your huggingface token to the pipeline if the model is lying in a private repo.\n - Either leave 'token=True' in the 'pipeline' and login to hugginface_hub by running\n \n - Or directly pass your <ACCESS_TOKEN> to 'token' in the 'pipeline'\n\n\n\nYou can print a sample prompt after the preprocessing step to see how it is feed to the tokenizer:\n\n\n\n\n\nAlternatively, you can download h2oai_pipeline.py, store it alongside your notebook, and construct the pipeline yourself from the loaded model and tokenizer. If the model and the tokenizer are fully supported in the 'transformers' package, this will allow you to set 'trust_remote_code=False'.\n\n\n\n\nYou may also construct the pipeline from the loaded model and tokenizer yourself and consider the preprocessing steps:## Quantization and sharding\n\nYou can load the models using quantization by specifying or . Also, sharding on multiple GPUs is possible by setting .## Model Architecture## Model Configuration\n\nThis model was trained using H2O LLM Studio and with the configuration in URL. Visit H2O LLM Studio to learn how to train your own large language models."
] | [
-0.11145677417516708,
0.15319277346134186,
-0.004479558672755957,
0.040352970361709595,
0.057110317051410675,
0.011407272890210152,
0.11624860763549805,
0.12530244886875153,
0.10750820487737656,
0.0944189578294754,
-0.005696078296750784,
0.019687650725245476,
0.06053021550178528,
0.17426851391792297,
0.08633320778608322,
-0.23114635050296783,
0.010947249829769135,
-0.08415929973125458,
-0.01139099895954132,
0.06480444222688675,
0.039120323956012726,
-0.053199052810668945,
0.0877903401851654,
0.015427114441990852,
-0.012903944589197636,
0.02010621689260006,
-0.010853469371795654,
0.02893790602684021,
0.05965039134025574,
0.07738528400659561,
-0.003756734309718013,
0.02622949331998825,
0.06072038784623146,
-0.15153974294662476,
0.03379134461283684,
0.1007336899638176,
0.026011615991592407,
0.035575415939092636,
0.012841797433793545,
-0.03369554877281189,
0.1238413155078888,
-0.05513797700405121,
0.044441841542720795,
0.05582619085907936,
-0.07013417035341263,
-0.10770262777805328,
-0.04990456998348236,
0.045828234404325485,
0.1420411616563797,
0.0608811154961586,
0.008411502465605736,
0.09717899560928345,
0.06880462914705276,
0.0717063918709755,
0.1656981110572815,
-0.06019726023077965,
-0.01957298442721367,
0.03843654319643974,
0.0722736045718193,
0.11483681946992874,
-0.04384930059313774,
-0.019511260092258453,
-0.005082618445158005,
0.008928978815674782,
-0.021304238587617874,
-0.07199547439813614,
0.03389766812324524,
-0.05435281619429588,
-0.10842496156692505,
-0.01177296694368124,
0.13243664801120758,
-0.07061423361301422,
-0.06648483127355576,
-0.11258780211210251,
-0.13395428657531738,
-0.04944705590605736,
0.011080941185355186,
0.0036516806576400995,
0.039287019520998,
-0.0012578589376062155,
0.05686313286423683,
-0.15150712430477142,
-0.09609555453062057,
-0.09877774119377136,
0.010440831072628498,
0.11429055035114288,
0.03612194210290909,
0.026758821681141853,
-0.08872006088495255,
0.19743722677230835,
0.0034824609756469727,
-0.09251552820205688,
-0.0970669686794281,
-0.031006965786218643,
-0.12996163964271545,
-0.00961327739059925,
-0.013325587846338749,
-0.11089164763689041,
0.04242148995399475,
0.20630231499671936,
-0.017832402139902115,
0.0963427722454071,
-0.004685491323471069,
0.00602038623765111,
0.058707039803266525,
0.09691468626260757,
-0.01387728936970234,
0.006437751930207014,
0.03653053939342499,
-0.01940631866455078,
0.06598105281591415,
-0.055477045476436615,
-0.04897207394242287,
-0.039876505732536316,
-0.07243700325489044,
0.07970921695232391,
0.03349576145410538,
0.046555086970329285,
0.011254423297941685,
-0.07448527961969376,
0.1526084989309311,
-0.15534809231758118,
0.03744988143444061,
0.03751174733042717,
-0.019564274698495865,
-0.002519048750400543,
0.1118464395403862,
-0.04306725412607193,
-0.08991978317499161,
-0.014162524603307247,
-0.025510519742965698,
0.030593542382121086,
-0.09363982081413269,
-0.034540798515081406,
-0.008772443979978561,
0.027826257050037384,
-0.05219502002000809,
-0.07897923141717911,
-0.1561967134475708,
-0.0246709156781435,
0.07420346140861511,
-0.028604719787836075,
0.040547728538513184,
-0.0022455265279859304,
0.08203623443841934,
-0.0030118378344923258,
0.02649727463722229,
-0.020999277010560036,
-0.03774840757250786,
-0.0076035694219172,
0.018291018903255463,
0.07101599872112274,
-0.03523408994078636,
-0.006416951771825552,
-0.06813164800405502,
0.07136604189872742,
-0.2126355916261673,
0.10349883139133453,
-0.037116844207048416,
0.0439903624355793,
-0.06280192732810974,
-0.007143770344555378,
-0.02348688617348671,
-0.008743628859519958,
0.007398949936032295,
0.08631901443004608,
-0.15106646716594696,
-0.013213472440838814,
0.16974933445453644,
-0.15297627449035645,
-0.012467359192669392,
0.06490898877382278,
-0.002920257393270731,
0.13666729629039764,
0.05171601101756096,
0.10042094439268112,
0.23476003110408783,
-0.2563542425632477,
0.05267183855175972,
0.08452793210744858,
-0.07866313308477402,
0.03750421106815338,
0.02262827754020691,
-0.024827376008033752,
0.05958555266261101,
0.03385971859097481,
-0.07896915078163147,
-0.02365971729159355,
0.06580743938684464,
-0.020184537395834923,
-0.009344258345663548,
-0.01593593694269657,
-0.07371164858341217,
-0.01866483874619007,
-0.028681328520178795,
0.03166902810335159,
-0.060922276228666306,
0.06994076818227768,
0.16556188464164734,
-0.08266504108905792,
0.06544272601604462,
-0.06557033956050873,
0.059354569762945175,
0.021420922130346298,
-0.0276653915643692,
-0.1308359056711197,
-0.056622836738824844,
0.0454796738922596,
-0.15755914151668549,
0.04593629017472267,
0.014321382157504559,
0.02209681272506714,
0.14206847548484802,
0.033907823264598846,
-0.003191788913682103,
0.06799518316984177,
-0.04158956557512283,
-0.03392571583390236,
-0.07321418076753616,
-0.03447093069553375,
-0.04402925446629524,
0.1559130698442459,
-0.023387376219034195,
0.06459828466176987,
0.039383940398693085,
0.03904937580227852,
0.02453630603849888,
-0.047801416367292404,
0.056989893317222595,
-0.09604983031749725,
0.010106485337018967,
-0.06908883154392242,
0.023622648790478706,
0.06821884959936142,
-0.011609951965510845,
0.007775961887091398,
-0.22089934349060059,
-0.20908984541893005,
0.039492394775152206,
0.147694393992424,
-0.04651494696736336,
-0.10206213593482971,
-0.025310995057225227,
-0.05273304134607315,
-0.061573125422000885,
-0.04508151113986969,
0.12979137897491455,
0.051241252571344376,
0.0875122919678688,
-0.09266489744186401,
-0.07442063838243484,
-0.026519840583205223,
-0.06841511279344559,
0.030378922820091248,
0.0683719664812088,
-0.008919858373701572,
-0.03713265806436539,
-0.027100391685962677,
-0.06551765650510788,
-0.039471786469221115,
0.18041329085826874,
0.06540147960186005,
-0.09115010499954224,
-0.05191115662455559,
0.03578221797943115,
0.007496730890125036,
0.0912383422255516,
-0.022032009437680244,
0.0184746403247118,
0.0158622357994318,
-0.010228332132101059,
0.039487987756729126,
-0.1018226146697998,
0.06438704580068588,
0.003204821143299341,
-0.029479341581463814,
-0.00904373824596405,
0.0964939072728157,
-0.05418207496404648,
0.003794832620769739,
-0.024111002683639526,
0.1504124402999878,
-0.03828281909227371,
-0.09307985752820969,
-0.09538555890321732,
0.12171835452318192,
-0.09539631009101868,
-0.22139661014080048,
-0.1528943032026291,
-0.09107368439435959,
-0.03607961907982826,
-0.012427602894604206,
0.05972103402018547,
0.021306147798895836,
-0.03286932036280632,
-0.09928859025239944,
0.007007258478552103,
0.018724987283349037,
-0.06914378702640533,
-0.09710429608821869,
0.02925010770559311,
0.04340182989835739,
-0.12154372781515121,
-0.023347333073616028,
0.025617331266403198,
-0.05067673698067665,
0.017860157415270805,
0.04473227262496948,
0.042177993804216385,
0.1033906564116478,
0.023963844403624535,
0.034245338290929794,
0.052730586379766464,
0.19035552442073822,
-0.048565518110990524,
0.11458902806043625,
0.18452858924865723,
-0.018391886726021767,
0.10651526600122452,
0.1173127144575119,
0.016907967627048492,
-0.06498794257640839,
0.05112266540527344,
-0.001510361093096435,
-0.04889116808772087,
-0.08702309429645538,
-0.046909306198358536,
-0.03994814306497574,
0.008004835806787014,
0.10294737666845322,
0.06288836896419525,
-0.02926754206418991,
-0.00047277825069613755,
-0.039096541702747345,
-0.024357696995139122,
0.004290920682251453,
0.10974228382110596,
-0.010562933050096035,
-0.01625245064496994,
-0.004813366569578648,
-0.027454465627670288,
0.04634961858391762,
0.0793825089931488,
0.10290270298719406,
0.06414400041103363,
-0.09326387196779251,
0.1480717808008194,
0.03683921694755554,
0.0801658034324646,
0.022639388218522072,
0.04179993271827698,
-0.0433523952960968,
0.04849594086408615,
0.020663036033511162,
-0.10703331977128983,
0.010471801273524761,
0.09758714586496353,
-0.0855599045753479,
-0.01512361690402031,
-0.0186432134360075,
0.05463341623544693,
0.016999326646327972,
0.22259119153022766,
0.02000442147254944,
-0.16390767693519592,
-0.031185852363705635,
0.022636305540800095,
-0.019882816821336746,
-0.06453608721494675,
-0.005263238213956356,
0.05391903966665268,
-0.11260996758937836,
0.02668742835521698,
-0.03247600793838501,
0.062030307948589325,
-0.06615475565195084,
-0.007935860194265842,
0.0968092605471611,
0.16709758341312408,
-0.03421623259782791,
0.08257994800806046,
-0.11039762198925018,
0.022593116387724876,
0.014807086437940598,
0.05567336082458496,
-0.038868945091962814,
0.03201393410563469,
0.0353001207113266,
0.06754417717456818,
0.1356879025697708,
0.028407637029886246,
-0.16743804514408112,
-0.07505816966295242,
-0.1083541288971901,
0.02971365861594677,
0.01905873604118824,
-0.08760017901659012,
0.03340170159935951,
-0.02326234057545662,
-0.04089247062802315,
-0.07178030163049698,
-0.045713022351264954,
-0.07954675704240799,
-0.16137930750846863,
0.08176899701356888,
-0.04840409755706787,
0.015924429520964622,
-0.033533524721860886,
0.036282978951931,
0.09262756258249283,
0.096696637570858,
-0.08510138839483261,
-0.1048215925693512,
-0.08190818130970001,
-0.06107267364859581,
0.07629434764385223,
-0.10817908495664597,
-0.004697349388152361,
-0.035671528428792953,
0.15254439413547516,
-0.035466574132442474,
-0.09994252771139145,
0.03877284377813339,
-0.08895976096391678,
-0.08491922914981842,
-0.01987781934440136,
0.12533384561538696,
0.09700848907232285,
-0.0014861254021525383,
-0.032406847923994064,
0.0037231934256851673,
-0.059009090065956116,
-0.08756419271230698,
-0.05168605595827103,
0.20731180906295776,
0.04306800290942192,
-0.03114028088748455,
-0.08600767701864243,
0.05581758916378021,
-0.06955904513597488,
0.04094903543591499,
0.05706886202096939,
0.20423492789268494,
-0.061658360064029694,
0.17223215103149414,
0.14723557233810425,
-0.09778009355068207,
-0.17794951796531677,
-0.034575656056404114,
-0.0023869143333286047,
0.004785811994224787,
-0.017291858792304993,
-0.1638127863407135,
0.10987348109483719,
0.0451480969786644,
-0.015430159866809845,
0.11525557190179825,
-0.29897841811180115,
-0.12340875715017319,
0.041050031781196594,
0.030599037185311317,
-0.0985567569732666,
-0.06715168058872223,
-0.02999451942741871,
-0.0782332643866539,
-0.06868989765644073,
0.057042401283979416,
-0.13327240943908691,
0.07700973749160767,
-0.02924376353621483,
0.03316597267985344,
0.03294858708977699,
-0.06320425122976303,
0.07588215917348862,
-0.015662889927625656,
0.06082187965512276,
-0.059428125619888306,
0.0603201687335968,
0.0040581924840807915,
-0.13374069333076477,
0.13905172049999237,
-0.06464129686355591,
0.06601858139038086,
0.00787746999412775,
-0.035085003823041916,
-0.051495056599378586,
0.08393170684576035,
-0.060723576694726944,
-0.06796705722808838,
-0.030745936557650566,
0.06268048286437988,
0.05349385365843773,
-0.02173415757715702,
-0.1582621932029724,
-0.08525661379098892,
-0.029499640688300133,
0.1985158920288086,
0.06234537437558174,
-0.04245268926024437,
-0.12326159328222275,
0.0009265078115276992,
0.02425234392285347,
0.05627297982573509,
-0.05613534152507782,
0.044884927570819855,
0.05073034018278122,
0.06019618734717369,
0.07570644468069077,
-0.0013805690687149763,
-0.10429014265537262,
-0.011248894967138767,
0.024232754483819008,
-0.14502310752868652,
-0.09151489287614822,
-0.048729658126831055,
0.10619016736745834,
-0.06653761118650436,
-0.02686586230993271,
0.11907976865768433,
0.003698087763041258,
-0.015313842333853245,
0.03765243664383888,
0.04910127818584442,
-0.026611482724547386,
0.08255527168512344,
0.004997944924980402,
0.023300105705857277,
-0.06548731029033661,
0.07750280946493149,
0.017073344439268112,
-0.010229932144284248,
-0.020808082073926926,
0.20799332857131958,
-0.12201420217752457,
-0.07076001167297363,
-0.08358089625835419,
0.03866755589842796,
-0.01813625358045101,
-0.06322064250707626,
-0.003779374063014984,
0.05341450870037079,
-0.05634505674242973,
0.05864622816443443,
0.03733435273170471,
-0.010147512890398502,
-0.021901851519942284,
0.018712621182203293,
-0.05403397977352142,
0.10907808691263199,
0.03321690857410431,
0.04480810463428497,
-0.047437433153390884,
0.12103922665119171,
0.035474058240652084,
0.07296063005924225,
-0.011150498874485493,
-0.08877549320459366,
-0.06760440021753311,
-0.007399145048111677,
-0.083687424659729,
0.04423041641712189,
-0.05977056920528412,
-0.007848354987800121,
-0.00024409443722106516,
0.024472104385495186,
0.03192612901329994,
0.047446053475141525,
-0.03569707274436951,
-0.03767947480082512,
-0.07405383139848709,
0.03884139284491539,
-0.09991118311882019,
0.01713024452328682,
0.07049541920423508,
-0.06853102147579193,
0.09495989978313446,
0.0570962056517601,
-0.08706321567296982,
-0.00411196006461978,
-0.07440769672393799,
0.01949392817914486,
-0.06336396187543869,
0.039188042283058167,
0.02955084666609764,
-0.12652672827243805,
-0.03106589987874031,
-0.03438795357942581,
-0.009045016951858997,
-0.06230321153998375,
0.08567767590284348,
-0.08877534419298172,
0.1569550633430481,
-0.0006484993500635028,
-0.04654954373836517,
-0.09186629951000214,
-0.009367123246192932,
-0.019315510988235474,
0.08204623311758041,
0.02930629812180996,
-0.07272110134363174,
0.08938506245613098,
-0.10009690374135971,
-0.04314398765563965,
0.11730854958295822,
0.04069807752966881,
-0.045336075127124786,
-0.08108934760093689,
0.03669992834329605,
-0.0384625568985939,
0.11075882613658905,
-0.011529684998095036,
0.05768691748380661,
-0.007257582154124975,
-0.015337699092924595,
-0.014043700881302357,
-0.009822942316532135,
0.030519435182213783,
-0.07825072109699249,
0.07380437850952148,
0.05246730148792267,
0.026603952050209045,
-0.019241835922002792,
0.01177167147397995,
0.13203349709510803,
0.02962261624634266,
0.10666108876466751,
0.03082156926393509,
0.004193931818008423,
0.019379088655114174,
-0.1489633470773697,
0.003987194504588842,
-0.04704705625772476,
0.06774333864450455,
-0.11756255477666855,
0.0937078595161438,
0.1004578024148941,
-0.1335824877023697,
0.05847177654504776,
0.06576216220855713,
-0.09771702438592911,
-0.13297323882579803,
-0.26095470786094666,
0.006831461563706398,
-0.10297273844480515,
-0.03179144114255905,
-0.049503766000270844,
0.05107853561639786,
-0.07561878114938736,
0.0245333481580019,
-0.005632129497826099,
0.15626664459705353,
-0.04941701889038086,
-0.07922755181789398,
-0.07256321609020233,
0.016895754262804985,
0.04834209755063057,
0.07566940784454346,
0.028024083003401756,
0.028590785339474678,
0.02088852971792221,
0.018783727660775185,
0.04798280447721481,
0.09750348329544067,
0.011854215525090694,
-0.02981591410934925,
-0.00531784538179636,
-0.008935229852795601,
0.030015720054507256,
-0.027407480403780937,
0.0846775695681572,
0.101722851395607,
-0.05481426417827606,
-0.02392708882689476,
0.15724454820156097,
-0.0775955319404602,
-0.1283494085073471,
-0.10637947916984558,
0.2519557476043701,
-0.0889219120144844,
-0.0341733880341053,
-0.015266534872353077,
-0.09877216815948486,
-0.05363621562719345,
0.21776269376277924,
0.13128897547721863,
-0.032556504011154175,
0.04747045040130615,
-0.006287400610744953,
0.013012861832976341,
-0.037158600986003876,
0.07802706956863403,
0.03698376566171646,
0.2165282815694809,
-0.029270485043525696,
0.06971365958452225,
0.029687052592635155,
-0.05115147680044174,
-0.13971549272537231,
0.058090828359127045,
-0.10008519887924194,
0.008813380263745785,
-0.008711456321179867,
0.057106081396341324,
-0.06352056562900543,
-0.21887069940567017,
-0.04227915406227112,
0.07384327054023743,
-0.04544335976243019,
0.038046710193157196,
0.017653413116931915,
-0.025739384815096855,
0.041594523936510086,
-0.042716134339571,
-0.00996926985681057,
0.2259065955877304,
-0.01593477465212345,
-0.09310025721788406,
-0.06315604597330093,
0.08373694866895676,
-0.10432437807321548,
0.13827021420001984,
0.0014582519652321935,
0.009459874592721462,
0.05038302764296532,
0.0037525754887610674,
-0.08585544675588608,
0.04437702149152756,
0.024292193353176117,
-0.14439573884010315,
0.0013309295754879713,
0.140761137008667,
-0.0016543511301279068,
0.007633795030415058,
0.02118559554219246,
-0.06200085207819939,
0.01988169737160206,
-0.03903118520975113,
0.052197303622961044,
-0.14287324249744415,
0.0467694029211998,
-0.0747203454375267,
0.1588890701532364,
0.13483421504497528,
-0.015038990415632725,
0.0014999606646597385,
-0.042629487812519073,
0.024057673290371895,
-0.0033056375104933977,
-0.03741621598601341,
-0.013222790323197842,
-0.1090642586350441,
0.03808096796274185,
0.035597193986177444,
0.04297751188278198,
-0.14448659121990204,
-0.06067727133631706,
0.011128866113722324,
-0.027708666399121284,
0.053335703909397125,
0.12661437690258026,
0.009677649475634098,
0.05134027078747749,
-0.00867646187543869,
-0.019475119188427925,
0.018742533400654793,
0.08521164953708649,
-0.16785506904125214,
-0.07681126147508621
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | phamtungthuy/law_model_merged | [
"transformers",
"safetensors",
"mpt",
"text-generation",
"conversational",
"custom_code",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T09:05:02+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #mpt #text-generation #conversational #custom_code #arxiv-1910.09700 #autotrain_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #mpt #text-generation #conversational #custom_code #arxiv-1910.09700 #autotrain_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
57,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #mpt #text-generation #conversational #custom_code #arxiv-1910.09700 #autotrain_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.054635148495435715,
0.16052456200122833,
-0.005164579953998327,
0.0220910906791687,
0.09270505607128143,
0.010645394213497639,
0.06736861169338226,
0.1135931983590126,
-0.022680921480059624,
0.11171533912420273,
0.031282417476177216,
0.09966786950826645,
0.11448366194963455,
0.1617225557565689,
-0.00020936582586728036,
-0.23094186186790466,
0.04733486846089363,
-0.1277359277009964,
-0.03125547617673874,
0.11818543821573257,
0.1437927484512329,
-0.10483533143997192,
0.07779167592525482,
-0.03074135258793831,
-0.006630721036344767,
-0.03400503098964691,
-0.06328880041837692,
-0.04955973103642464,
0.0487750843167305,
0.06944658607244492,
0.06819262355566025,
0.007929004728794098,
0.09096074849367142,
-0.26640498638153076,
0.02359817363321781,
0.07491831481456757,
-0.000010058924090117216,
0.07733584940433502,
0.04770263284444809,
-0.08114725351333618,
0.07781454175710678,
-0.05144769325852394,
0.1483687162399292,
0.08374054729938507,
-0.08928101509809494,
-0.19188429415225983,
-0.08276589959859848,
0.09912224113941193,
0.18290962278842926,
0.049899518489837646,
-0.023949703201651573,
0.10654067248106003,
-0.08682400733232498,
0.010167408734560013,
0.05219518393278122,
-0.05916601046919823,
-0.05467120185494423,
0.06819918006658554,
0.08054910600185394,
0.08425354957580566,
-0.129018634557724,
-0.017852487042546272,
0.01335847657173872,
0.006196460220962763,
0.0857679694890976,
0.023681500926613808,
0.1491670459508896,
0.04419431462883949,
-0.13135887682437897,
-0.045980870723724365,
0.09945714473724365,
0.044559575617313385,
-0.052517082542181015,
-0.250519335269928,
-0.028032813221216202,
-0.041343145072460175,
-0.031409576535224915,
-0.037443872541189194,
0.043639495968818665,
-0.009259944781661034,
0.08239757269620895,
-0.00779309868812561,
-0.07888150960206985,
-0.03879553824663162,
0.06594165414571762,
0.05944693833589554,
0.0254695862531662,
-0.012103017419576645,
0.00838468037545681,
0.11912903189659119,
0.11280839890241623,
-0.12819187343120575,
-0.053715288639068604,
-0.06892311573028564,
-0.08053050935268402,
-0.04290087893605232,
0.03527285158634186,
0.045483797788619995,
0.044262275099754333,
0.24731779098510742,
0.015218249522149563,
0.054765526205301285,
0.04213971272110939,
0.011575471609830856,
0.06641978025436401,
0.11105266958475113,
-0.05833715200424194,
-0.09286858141422272,
-0.029584970325231552,
0.09632708132266998,
0.0073746670968830585,
-0.043398164212703705,
-0.06130729988217354,
0.07046598196029663,
0.009513085708022118,
0.12098813056945801,
0.08707986027002335,
0.0007522705709561706,
-0.07445661723613739,
-0.0706939697265625,
0.1917448490858078,
-0.16087934374809265,
0.04857055842876434,
0.03421054407954216,
-0.03818230330944061,
0.002456377260386944,
0.009284881874918938,
0.023113954812288284,
-0.019631540402770042,
0.0855097696185112,
-0.05617918819189072,
-0.03747226670384407,
-0.11360510438680649,
-0.035750407725572586,
0.030426746234297752,
0.008404480293393135,
-0.028276318684220314,
-0.03244270756840706,
-0.08289990574121475,
-0.058483731001615524,
0.09393514692783356,
-0.07486393302679062,
-0.050592172890901566,
-0.015272891148924828,
-0.0791788250207901,
0.023433104157447815,
0.01889639161527157,
0.08151152729988098,
-0.02141479402780533,
0.038186974823474884,
-0.0545668751001358,
0.06371627748012543,
0.11341410130262375,
0.030237682163715363,
-0.05318576470017433,
0.0600082129240036,
-0.24077340960502625,
0.10212498158216476,
-0.07240713387727737,
0.04895240068435669,
-0.149939626455307,
-0.024039480835199356,
0.04290594533085823,
0.00972573272883892,
-0.013753218576312065,
0.14103291928768158,
-0.22007226943969727,
-0.03127304092049599,
0.16302013397216797,
-0.09508489072322845,
-0.0747118815779686,
0.058422572910785675,
-0.054242417216300964,
0.10799778997898102,
0.04084877669811249,
-0.02813737280666828,
0.058678124099969864,
-0.14432768523693085,
-0.007126365322619677,
-0.04207603260874748,
-0.02075313962996006,
0.16577234864234924,
0.07608950883150101,
-0.0627174824476242,
0.07783328741788864,
0.023570606485009193,
-0.01646917127072811,
-0.04628380760550499,
-0.022484013810753822,
-0.11034025251865387,
0.011407471261918545,
-0.05814944952726364,
0.0260792076587677,
-0.026217708364129066,
-0.09773342311382294,
-0.028854506090283394,
-0.17960339784622192,
-0.011270643211901188,
0.08880633860826492,
-0.010637312196195126,
-0.023396793752908707,
-0.12005569040775299,
0.010020323097705841,
0.03054247982800007,
0.0017140507698059082,
-0.13087037205696106,
-0.05012092739343643,
0.027868488803505898,
-0.1645229458808899,
0.034395549446344376,
-0.05180405452847481,
0.04987547546625137,
0.03591126576066017,
-0.029487887397408485,
-0.03373080864548683,
0.009159739129245281,
0.007332916837185621,
-0.010189548134803772,
-0.2527660131454468,
-0.026401663199067116,
-0.023419378325343132,
0.17689085006713867,
-0.22072727978229523,
0.04259980097413063,
0.07535214722156525,
0.14919216930866241,
0.008129291236400604,
-0.03602238744497299,
0.007628660183399916,
-0.08049123734235764,
-0.03433375060558319,
-0.060181934386491776,
-0.01243846956640482,
-0.03811177611351013,
-0.05414266139268875,
0.04925855994224548,
-0.16519419848918915,
-0.03516972064971924,
0.10688495635986328,
0.06264010816812515,
-0.13847990334033966,
-0.022087279707193375,
-0.04074591398239136,
-0.04355667531490326,
-0.05524059757590294,
-0.05962411314249039,
0.11337815225124359,
0.056771762669086456,
0.04850984364748001,
-0.07028688490390778,
-0.07552514225244522,
0.007090869825333357,
-0.019495615735650063,
-0.01900665834546089,
0.0901784747838974,
0.0812983587384224,
-0.11660853773355484,
0.0922618955373764,
0.106283038854599,
0.09227804094552994,
0.10790982842445374,
-0.018234362825751305,
-0.0887838825583458,
-0.045240260660648346,
0.02204916812479496,
0.013851354829967022,
0.1447363644838333,
-0.015503661707043648,
0.05703561380505562,
0.04034547135233879,
-0.01354019995778799,
0.011624177917838097,
-0.09244153648614883,
0.030048804357647896,
0.03612228110432625,
-0.017916856333613396,
0.03961733728647232,
-0.042948927730321884,
0.021726015955209732,
0.08848346769809723,
0.04535902664065361,
0.039198409765958786,
0.010900821536779404,
-0.04336994141340256,
-0.10870755463838577,
0.17140623927116394,
-0.12746594846248627,
-0.24067653715610504,
-0.14221419394016266,
-0.0012358175590634346,
0.04186861217021942,
-0.01059538684785366,
0.0011142924195155501,
-0.07011234015226364,
-0.11840486526489258,
-0.09549368172883987,
0.01936156302690506,
0.0449710451066494,
-0.07725443691015244,
-0.05926412343978882,
0.06101541966199875,
0.03925406560301781,
-0.14526304602622986,
0.022857220843434334,
0.04891752451658249,
-0.08735892176628113,
-0.007337834220379591,
0.0824587494134903,
0.06704602390527725,
0.1819511204957962,
0.010910622775554657,
-0.023387739434838295,
0.03352188318967819,
0.2140638381242752,
-0.1398898959159851,
0.11526372283697128,
0.1414218246936798,
-0.09628640860319138,
0.08139462023973465,
0.19910632073879242,
0.04108594357967377,
-0.10341772437095642,
0.03326005861163139,
0.02280617691576481,
-0.02815237268805504,
-0.24921943247318268,
-0.07115662097930908,
0.00007775821723043919,
-0.04786401987075806,
0.07067081332206726,
0.08930106461048126,
0.09956705570220947,
0.013410673476755619,
-0.09870944917201996,
-0.08179779350757599,
0.04957464337348938,
0.10591937601566315,
0.006951907649636269,
-0.016600094735622406,
0.08980800211429596,
-0.030582370236516,
0.01617647148668766,
0.0952819287776947,
0.004716930445283651,
0.17350728809833527,
0.06039604917168617,
0.1823190301656723,
0.07943164557218552,
0.06654398143291473,
0.021045364439487457,
0.008380318991839886,
0.020960748195648193,
0.02026979625225067,
-0.006007791962474585,
-0.08725067973136902,
-0.005133382044732571,
0.1207573413848877,
0.06482845544815063,
0.028621217235922813,
0.017345979809761047,
-0.037953365594148636,
0.08435174822807312,
0.16896480321884155,
0.0021635573357343674,
-0.1859889179468155,
-0.07005161792039871,
0.08015012741088867,
-0.09120379388332367,
-0.10350073873996735,
-0.03254825249314308,
0.033887993544340134,
-0.17068453133106232,
0.01693805493414402,
-0.016277264803647995,
0.10999585688114166,
-0.1364787369966507,
-0.018435562029480934,
0.06097080558538437,
0.07548227906227112,
-0.002038689563050866,
0.06082221493124962,
-0.1507110297679901,
0.1045399159193039,
0.0174159687012434,
0.06488069891929626,
-0.09838446229696274,
0.10249010473489761,
-0.006359171587973833,
-0.008563103154301643,
0.13387849926948547,
0.009050385095179081,
-0.054088789969682693,
-0.0906732901930809,
-0.10487803816795349,
-0.010847710072994232,
0.1295238435268402,
-0.15052857995033264,
0.08596144616603851,
-0.029183100908994675,
-0.04416894540190697,
0.0022713199723511934,
-0.11619070917367935,
-0.12652844190597534,
-0.18756036460399628,
0.055853214114904404,
-0.13760650157928467,
0.03747143596410751,
-0.10383535921573639,
-0.041909459978342056,
-0.025565309450030327,
0.20074599981307983,
-0.2338542342185974,
-0.06549516320228577,
-0.15260930359363556,
-0.08816627413034439,
0.14466848969459534,
-0.04515206441283226,
0.08901641517877579,
-0.003431829623878002,
0.19132232666015625,
0.022061895579099655,
-0.023516153916716576,
0.10256588459014893,
-0.0946759358048439,
-0.194707989692688,
-0.08654211461544037,
0.15764883160591125,
0.1389331966638565,
0.03798891603946686,
-0.0055480520240962505,
0.030881132930517197,
-0.022598356008529663,
-0.12330599129199982,
0.02234155684709549,
0.17798170447349548,
0.07587705552577972,
0.024513766169548035,
-0.028767045587301254,
-0.11621139198541641,
-0.06686770170927048,
-0.03146045655012131,
0.02390756830573082,
0.1850925236940384,
-0.07339204847812653,
0.18708264827728271,
0.13815747201442719,
-0.05402544513344765,
-0.1952371448278427,
0.017667120322585106,
0.039463941007852554,
0.011406933888792992,
0.03653019294142723,
-0.20459920167922974,
0.08829911798238754,
-0.001095055602490902,
-0.0508810319006443,
0.13820722699165344,
-0.1719542294740677,
-0.14832891523838043,
0.06369562447071075,
0.03551396355032921,
-0.19702890515327454,
-0.11501365154981613,
-0.08721195161342621,
-0.047495681792497635,
-0.17153312265872955,
0.10541050136089325,
0.016818318516016006,
0.006937634665518999,
0.037676215171813965,
0.023313308134675026,
0.01225703302770853,
-0.03891397640109062,
0.19560746848583221,
-0.02633627876639366,
0.03060000389814377,
-0.07944042980670929,
-0.06587760895490646,
0.058183394372463226,
-0.05387004092335701,
0.07589740306138992,
-0.02317744493484497,
0.011171734891831875,
-0.11016803234815598,
-0.04590783640742302,
-0.02669665962457657,
0.015302745625376701,
-0.0945773795247078,
-0.09466630220413208,
-0.05188223719596863,
0.09198068082332611,
0.08989781886339188,
-0.03538026288151741,
-0.03926745429635048,
-0.07457108795642853,
0.03284696862101555,
0.18374256789684296,
0.18051902949810028,
0.03950328007340431,
-0.07674854248762131,
-0.004460221156477928,
-0.011413991451263428,
0.04427469149231911,
-0.2254757583141327,
0.06300739198923111,
0.045056115835905075,
0.014833101071417332,
0.11507917940616608,
-0.020084138959646225,
-0.16268320381641388,
-0.06560098379850388,
0.060816798359155655,
-0.06339775770902634,
-0.18217681348323822,
0.002922016428783536,
0.06140703707933426,
-0.16704855859279633,
-0.04194588586688042,
0.042133230715990067,
-0.005722786299884319,
-0.04149339720606804,
0.018184611573815346,
0.08250795304775238,
0.006119817961007357,
0.06918256729841232,
0.05840451270341873,
0.08533959835767746,
-0.10398739576339722,
0.07208184152841568,
0.07942937314510345,
-0.08207568526268005,
0.027590468525886536,
0.08917015045881271,
-0.06065768748521805,
-0.030156195163726807,
0.0329950712621212,
0.0794283002614975,
0.022911665961146355,
-0.03942554444074631,
0.008366047404706478,
-0.09856892377138138,
0.0626458004117012,
0.09304165095090866,
0.03544410690665245,
0.013699504546821117,
0.033380769193172455,
0.048400793224573135,
-0.07311499118804932,
0.12109450250864029,
0.02168533392250538,
0.015164933167397976,
-0.0412873812019825,
-0.05220061540603638,
0.03134710341691971,
-0.027845747768878937,
-0.009887563064694405,
-0.03943326324224472,
-0.06989603489637375,
-0.017423398792743683,
-0.157952681183815,
-0.014101648703217506,
-0.040278226137161255,
0.0077163297683000565,
0.025573555380105972,
-0.037198007106781006,
0.0065233283676207066,
0.010041977278888226,
-0.07327748090028763,
-0.05960303917527199,
-0.02242644876241684,
0.09257826954126358,
-0.1605987250804901,
0.027755331248044968,
0.08544294536113739,
-0.12294061481952667,
0.09279222786426544,
0.017858223989605904,
0.00018222432117909193,
0.02797253057360649,
-0.15263277292251587,
0.039833471179008484,
-0.023737426847219467,
0.016787255182862282,
0.044893525540828705,
-0.2171315997838974,
0.00009986197255784646,
-0.03639553487300873,
-0.06263463199138641,
-0.01287070382386446,
-0.03437057510018349,
-0.11212986707687378,
0.09930268675088882,
0.009211725555360317,
-0.08627957105636597,
-0.028223544359207153,
0.03517432510852814,
0.08295711874961853,
-0.02777615562081337,
0.1560288816690445,
0.0037433584220707417,
0.07323931157588959,
-0.17735908925533295,
-0.022074343636631966,
-0.011848291382193565,
0.022013992071151733,
-0.01760149747133255,
-0.01711566559970379,
0.04208911582827568,
-0.026420433074235916,
0.18827635049819946,
-0.025936778634786606,
0.03580417111515999,
0.06425658613443375,
0.017719900235533714,
-0.020465362817049026,
0.1115000769495964,
0.05887604504823685,
0.024415457621216774,
0.024907831102609634,
0.004697317257523537,
-0.0426749624311924,
-0.015930678695440292,
-0.19337289035320282,
0.061051346361637115,
0.14045123755931854,
0.08954483270645142,
-0.015985945239663124,
0.08108439296483994,
-0.09789230674505234,
-0.12102881073951721,
0.11737070232629776,
-0.0606815479695797,
-0.00615657726302743,
-0.06569632142782211,
0.13350485265254974,
0.1502852588891983,
-0.1885579228401184,
0.06877799332141876,
-0.062048763036727905,
-0.050999715924263,
-0.1176731213927269,
-0.19573575258255005,
-0.05637215077877045,
-0.05677573010325432,
-0.017962442710995674,
-0.04842783883213997,
0.07284567505121231,
0.056337952613830566,
0.012329373508691788,
0.0035097927320748568,
0.07202450186014175,
-0.017321929335594177,
-0.001713146106339991,
0.027790050953626633,
0.06609232723712921,
0.01376376673579216,
-0.031173428520560265,
0.01871532015502453,
-0.004925333429127932,
0.039521846920251846,
0.06356487423181534,
0.03991309553384781,
-0.02794083207845688,
0.016109539195895195,
-0.04031091928482056,
-0.11215955018997192,
0.042707886546850204,
-0.03184114024043083,
-0.08352098613977432,
0.14015837013721466,
0.02227190136909485,
0.005765206180512905,
-0.02191980369389057,
0.24175137281417847,
-0.07252326607704163,
-0.08980383723974228,
-0.15015986561775208,
0.09920824319124222,
-0.042186010628938675,
0.06808467954397202,
0.041611071676015854,
-0.1055009663105011,
0.01259309146553278,
0.1256643682718277,
0.1585809737443924,
-0.04577846825122833,
0.02053774893283844,
0.03308415785431862,
0.004132973495870829,
-0.04233024641871452,
0.05212663859128952,
0.06862838566303253,
0.14791923761367798,
-0.047211214900016785,
0.08973482996225357,
0.003054187400266528,
-0.10376755148172379,
-0.036125410348176956,
0.11549590528011322,
-0.017118209972977638,
0.016522333025932312,
-0.056225430220365524,
0.11830950528383255,
-0.0581444650888443,
-0.22943419218063354,
0.06189032644033432,
-0.07331333309412003,
-0.14115603268146515,
-0.023689990863204002,
0.07678241282701492,
-0.016217641532421112,
0.024415941908955574,
0.07207741588354111,
-0.07874421030282974,
0.20002593100070953,
0.03732920438051224,
-0.054341766983270645,
-0.057722028344869614,
0.07965662330389023,
-0.11753721535205841,
0.2731863260269165,
0.01380995661020279,
0.04915657266974449,
0.10306801646947861,
-0.016123464331030846,
-0.13350830972194672,
0.018472453579306602,
0.0947251096367836,
-0.08889597654342651,
0.038379255682229996,
0.2097547948360443,
-0.0019173307809978724,
0.12025855481624603,
0.07908619195222855,
-0.07485989481210709,
0.045498862862586975,
-0.11228009313344955,
-0.07637493312358856,
-0.0845867246389389,
0.0944986492395401,
-0.07291062921285629,
0.14091554284095764,
0.12382294237613678,
-0.053785547614097595,
0.009473927319049835,
-0.029916830360889435,
0.047535575926303864,
0.0065670982003211975,
0.10087279975414276,
0.009793313220143318,
-0.18353337049484253,
0.02005654387176037,
0.013835139572620392,
0.10875878483057022,
-0.1678035855293274,
-0.09590578079223633,
0.044682543724775314,
0.002749423496425152,
-0.06389376521110535,
0.1261044591665268,
0.06061521917581558,
0.047957681119441986,
-0.04424063116312027,
-0.01445530541241169,
-0.008313539437949657,
0.13601157069206238,
-0.10486738383769989,
0.0005321411299519241
] |
null | null | transformers | dict(
name="igenius-llama-tinystories-2B", # 2B run on AZURE
hf_config=dict(org="Igenius", name="igenius-llama-tinystories-az"),
block_size=2048,
vocab_size=32_000,
padding_multiple=64,
n_layer=16,
n_head=4 * 16,
n_embd=192 * 16,
rotary_percentage=1.0,
parallel_residual=False,
bias=False,
_norm_class="RMSNorm",
norm_eps=1e-5,
_mlp_class="LLaMAMLP", ##with LLaMAMLP 2B
intermediate_size=8192,
n_query_groups=4 * 16,
)
dummy model trained several epochs (20B token in total) on a small dataset | {} | text-generation | iGenius-AI-Team/Italia-2Bts-ckpt-20B | [
"transformers",
"safetensors",
"llama",
"text-generation",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T09:06:55+00:00 | [] | [] | TAGS
#transformers #safetensors #llama #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| dict(
name="igenius-llama-tinystories-2B", # 2B run on AZURE
hf_config=dict(org="Igenius", name="igenius-llama-tinystories-az"),
block_size=2048,
vocab_size=32_000,
padding_multiple=64,
n_layer=16,
n_head=4 * 16,
n_embd=192 * 16,
rotary_percentage=1.0,
parallel_residual=False,
bias=False,
_norm_class="RMSNorm",
norm_eps=1e-5,
_mlp_class="LLaMAMLP", ##with LLaMAMLP 2B
intermediate_size=8192,
n_query_groups=4 * 16,
)
dummy model trained several epochs (20B token in total) on a small dataset | [
"# 2B run on AZURE\n hf_config=dict(org=\"Igenius\", name=\"igenius-llama-tinystories-az\"),\n block_size=2048,\n vocab_size=32_000,\n padding_multiple=64,\n n_layer=16,\n n_head=4 * 16,\n n_embd=192 * 16,\n rotary_percentage=1.0,\n parallel_residual=False,\n bias=False,\n _norm_class=\"RMSNorm\",\n norm_eps=1e-5,\n _mlp_class=\"LLaMAMLP\", ##with LLaMAMLP 2B\n intermediate_size=8192,\n n_query_groups=4 * 16,\n )\n\n dummy model trained several epochs (20B token in total) on a small dataset"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# 2B run on AZURE\n hf_config=dict(org=\"Igenius\", name=\"igenius-llama-tinystories-az\"),\n block_size=2048,\n vocab_size=32_000,\n padding_multiple=64,\n n_layer=16,\n n_head=4 * 16,\n n_embd=192 * 16,\n rotary_percentage=1.0,\n parallel_residual=False,\n bias=False,\n _norm_class=\"RMSNorm\",\n norm_eps=1e-5,\n _mlp_class=\"LLaMAMLP\", ##with LLaMAMLP 2B\n intermediate_size=8192,\n n_query_groups=4 * 16,\n )\n\n dummy model trained several epochs (20B token in total) on a small dataset"
] | [
47,
184
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# 2B run on AZURE\n hf_config=dict(org=\"Igenius\", name=\"igenius-llama-tinystories-az\"),\n block_size=2048,\n vocab_size=32_000,\n padding_multiple=64,\n n_layer=16,\n n_head=4 * 16,\n n_embd=192 * 16,\n rotary_percentage=1.0,\n parallel_residual=False,\n bias=False,\n _norm_class=\"RMSNorm\",\n norm_eps=1e-5,\n _mlp_class=\"LLaMAMLP\", ##with LLaMAMLP 2B\n intermediate_size=8192,\n n_query_groups=4 * 16,\n )\n\n dummy model trained several epochs (20B token in total) on a small dataset"
] | [
-0.126765176653862,
-0.026687847450375557,
-0.004662765190005302,
0.06378088891506195,
0.048348333686590195,
-0.00008155335672199726,
0.1979302316904068,
0.10813569277524948,
0.0009378024842590094,
0.07792933285236359,
0.1346813142299652,
0.05858774855732918,
0.010449487715959549,
0.094634048640728,
-0.13302923738956451,
-0.1325986534357071,
0.05325626954436302,
0.02879156358540058,
0.13544151186943054,
0.03769855573773384,
0.05586139112710953,
-0.10199109464883804,
0.06396789103746414,
-0.0012527374783530831,
-0.13364547491073608,
0.0039010567124933004,
0.033198099583387375,
-0.0634010061621666,
0.10553666949272156,
0.08362098038196564,
0.028763731941580772,
0.007787052076309919,
0.013172203674912453,
-0.15439583361148834,
0.005646191071718931,
0.049447428435087204,
0.028454119339585304,
0.06255289167165756,
0.09315283596515656,
0.00524070393294096,
0.023234898224473,
-0.07600246369838715,
-0.0194636769592762,
0.0314800888299942,
-0.07248571515083313,
-0.17245033383369446,
-0.07122728228569031,
0.061855390667915344,
0.07882429659366608,
0.027214085683226585,
-0.029068751260638237,
0.18525049090385437,
-0.06834506243467331,
0.11265484988689423,
0.3424246311187744,
-0.3266042470932007,
0.006830153521150351,
0.08215420693159103,
0.008084328845143318,
0.012944313697516918,
-0.008445312269032001,
0.024531392380595207,
0.08501017093658447,
0.03915547579526901,
0.0655590146780014,
-0.07762522250413895,
0.003862130455672741,
-0.006922597531229258,
-0.11089617013931274,
-0.008517393842339516,
0.1763238161802292,
0.0223002340644598,
-0.06688138842582703,
-0.1007755771279335,
-0.09191177785396576,
-0.10510943830013275,
-0.0509197898209095,
0.07711666822433472,
0.011735065840184689,
-0.03173977509140968,
-0.011813893914222717,
0.09400147944688797,
-0.030605921521782875,
-0.05597301200032234,
-0.10093601793050766,
0.15685057640075684,
0.06403811275959015,
0.02702213078737259,
-0.11145661771297455,
0.03735051304101944,
-0.12813830375671387,
-0.09416879713535309,
-0.03232477232813835,
-0.013829933479428291,
-0.06775973737239838,
-0.0009334191563539207,
-0.11055541038513184,
-0.03083164617419243,
0.09415990114212036,
0.17833411693572998,
-0.076056107878685,
0.06230386719107628,
-0.006951815914362669,
0.05045635253190994,
0.03152361884713173,
0.002291920594871044,
-0.10582906007766724,
-0.03465983271598816,
0.0518217571079731,
-0.00661025894805789,
0.11438870429992676,
-0.015978997573256493,
-0.06691833585500717,
-0.029993468895554543,
0.0934242382645607,
0.031309574842453,
-0.08495592325925827,
0.08921441435813904,
-0.014950926415622234,
0.009275298565626144,
-0.031644225120544434,
-0.12587657570838928,
0.0002433312765788287,
-0.005186389200389385,
-0.029103349894285202,
0.06001928076148033,
-0.025046421214938164,
-0.004197185393422842,
-0.07608447223901749,
0.028730733320116997,
-0.09631571918725967,
-0.012066224589943886,
-0.03706634044647217,
-0.1646418571472168,
0.032724034041166306,
-0.08061707019805908,
0.013115203939378262,
-0.13699911534786224,
-0.1719299703836441,
0.026436099782586098,
-0.013451199978590012,
-0.05314105004072189,
0.03246280923485756,
-0.040060222148895264,
-0.0770883858203888,
0.01903962902724743,
-0.013992254622280598,
0.06458062678575516,
-0.06803663820028305,
0.014120267704129219,
0.06982559710741043,
0.11859182268381119,
-0.03751296177506447,
-0.00366334430873394,
-0.059140536934137344,
0.09515046328306198,
-0.11147714406251907,
0.08116340637207031,
0.0037641890812665224,
0.02096005529165268,
-0.0704076737165451,
-0.028954995796084404,
0.00955918151885271,
0.05431683361530304,
0.08664647489786148,
0.08864520490169525,
-0.15601682662963867,
-0.03986600413918495,
0.26568207144737244,
-0.08862470090389252,
-0.1343245655298233,
0.17509707808494568,
-0.03761018440127373,
-0.08686459809541702,
0.06237947940826416,
0.11124837398529053,
-0.06556033343076706,
-0.052291885018348694,
-0.06960966438055038,
0.006721719168126583,
0.05755173787474632,
-0.09171243011951447,
0.04213792830705643,
0.03566230833530426,
-0.020027026534080505,
0.019600478932261467,
0.06355757266283035,
0.017465248703956604,
-0.05984900891780853,
-0.0072070518508553505,
-0.053239911794662476,
-0.035426001995801926,
-0.028861496597528458,
-0.05467134341597557,
0.037488989531993866,
-0.14136222004890442,
-0.06693846732378006,
-0.042025357484817505,
0.09585713595151901,
-0.07507888972759247,
0.017041390761733055,
-0.06780324131250381,
0.09966516494750977,
-0.14572502672672272,
0.03253297507762909,
-0.17093127965927124,
-0.011326407082378864,
0.019562778994441032,
0.02859819307923317,
-0.015912465751171112,
0.13284994661808014,
0.009450728073716164,
0.061779748648405075,
-0.02737230248749256,
0.016045790165662766,
0.10833656787872314,
-0.013396214693784714,
-0.11486738175153732,
-0.12946470081806183,
-0.007747007068246603,
-0.07362983375787735,
0.05833621323108673,
-0.1946365237236023,
0.02692802995443344,
0.0020431370940059423,
0.0709024965763092,
0.04384423792362213,
0.034313105046749115,
0.014916942454874516,
0.05568103492259979,
-0.11138765513896942,
-0.03506404533982277,
0.030065692961215973,
-0.007986584678292274,
-0.07494669407606125,
0.02335762046277523,
-0.238436758518219,
0.06604659557342529,
0.1239723190665245,
-0.0032507628202438354,
-0.04869333282113075,
-0.07638110220432281,
-0.015601971186697483,
-0.008281138725578785,
-0.07388375699520111,
-0.04539426788687706,
0.10417456179857254,
0.054026760160923004,
0.15108045935630798,
-0.10753122717142105,
-0.04915613308548927,
0.04555578902363777,
-0.0960610955953598,
-0.015362902544438839,
0.0976148396730423,
0.03416704759001732,
-0.1502084583044052,
0.06961394101381302,
0.06543906033039093,
-0.10104096680879593,
0.08813656121492386,
-0.02274506539106369,
-0.08327805250883102,
-0.014559818431735039,
0.05298091843724251,
0.0333523191511631,
0.04695965349674225,
-0.06470907479524612,
-0.02557852678000927,
0.057013705372810364,
0.021034566685557365,
0.01424939651042223,
-0.1289823353290558,
0.03310660272836685,
-0.02640666998922825,
-0.022613918408751488,
-0.009488687850534916,
0.04611879214644432,
0.01837293431162834,
0.11012940853834152,
-0.04601512849330902,
-0.11949386447668076,
0.053256142884492874,
-0.0029515086207538843,
-0.08406328409910202,
0.2153594046831131,
-0.04360516741871834,
-0.16405239701271057,
-0.09321951121091843,
-0.15394730865955353,
-0.12557613849639893,
0.01898486167192459,
0.05638248100876808,
-0.036692701280117035,
-0.07980004698038101,
-0.12659332156181335,
0.043810077011585236,
0.03221997618675232,
0.04759812727570534,
0.032047949731349945,
0.0061556026339530945,
0.05665535479784012,
-0.11132188141345978,
-0.047820013016462326,
-0.02235144004225731,
0.059358786791563034,
0.11660570651292801,
-0.030580634251236916,
0.03369912877678871,
0.13242550194263458,
0.007325747050344944,
0.01812913827598095,
-0.016777992248535156,
0.13835978507995605,
0.02538829855620861,
-0.006288487464189529,
0.1559310108423233,
0.011828928254544735,
0.05720818415284157,
0.13212963938713074,
0.006943367887288332,
-0.11658292263746262,
0.021576344966888428,
0.05916377156972885,
-0.06317707896232605,
-0.17931659519672394,
-0.04708598181605339,
-0.06119377166032791,
-0.010568426921963692,
0.08044110238552094,
0.032691244035959244,
-0.009604036808013916,
0.09022750705480576,
-0.02923467569053173,
0.11744720488786697,
-0.046360380947589874,
0.09208913892507553,
0.14620943367481232,
0.016467779874801636,
0.14496953785419464,
-0.03584640845656395,
-0.055432192981243134,
0.0420752614736557,
0.021864328533411026,
0.19847309589385986,
-0.05001486837863922,
0.09872125089168549,
0.06012691557407379,
-0.024898752570152283,
0.049446456134319305,
0.13711529970169067,
-0.012475665658712387,
-0.026970136910676956,
-0.030026722699403763,
-0.05630066990852356,
-0.02193964086472988,
0.043814633041620255,
-0.1058427020907402,
0.013239827007055283,
-0.09401773661375046,
0.09390702098608017,
0.06380258500576019,
0.24428558349609375,
0.15470774471759796,
-0.38971590995788574,
-0.034523963928222656,
0.0519665963947773,
-0.04112737253308296,
-0.041840311139822006,
0.02279987372457981,
0.16039423644542694,
-0.012890532612800598,
0.0919492170214653,
-0.024990014731884003,
0.08455944806337357,
-0.09591671824455261,
0.08411392569541931,
-0.06949596107006073,
0.13888095319271088,
-0.014351807534694672,
0.046987518668174744,
-0.243921160697937,
0.18765611946582794,
0.043095674365758896,
0.015934601426124573,
-0.1060725748538971,
0.04117782041430473,
0.0009948863880708814,
0.0066094123758375645,
0.04372156038880348,
-0.0006356568774208426,
-0.1299951821565628,
-0.12836894392967224,
-0.038216423243284225,
0.0260029099881649,
0.05730012431740761,
0.053288642317056656,
0.12975110113620758,
0.005465197842568159,
-0.01935538835823536,
0.046768005937337875,
0.0076380507089197636,
-0.09809872508049011,
-0.08863472193479538,
0.02368195541203022,
0.18384398519992828,
-0.17886115610599518,
-0.05587523803114891,
-0.06571289151906967,
-0.1358407586812973,
0.22835075855255127,
0.012759245000779629,
-0.08070189505815506,
-0.08090569823980331,
0.0780886560678482,
0.057299867272377014,
-0.09375783801078796,
-0.007113725412636995,
-0.019529342651367188,
0.13765446841716766,
-0.056673917919397354,
-0.09462734311819077,
0.16157084703445435,
-0.09684541821479797,
-0.11659087240695953,
-0.03852948918938637,
0.08469361811876297,
-0.006232390645891428,
-0.00678122416138649,
0.024951748549938202,
-0.00937061756849289,
0.015367455780506134,
-0.08694377541542053,
-0.015672927722334862,
0.09980174154043198,
0.10920412093400955,
0.04971975088119507,
-0.08605152368545532,
-0.042783308774232864,
0.0064500910229980946,
0.04260963201522827,
0.15697522461414337,
0.3337218463420868,
-0.05158801004290581,
-0.032967034727334976,
0.10482871532440186,
-0.01549447886645794,
-0.15879468619823456,
0.003821762977167964,
-0.031088320538401604,
0.03480378910899162,
-0.02330973371863365,
-0.0829768106341362,
0.1293216496706009,
0.13733384013175964,
0.02395652048289776,
0.10494421422481537,
-0.2418006807565689,
-0.0898922011256218,
0.1616828739643097,
0.07695764303207397,
0.226390078663826,
-0.09688088297843933,
-0.012671001255512238,
-0.12820501625537872,
-0.031033428385853767,
0.04677714407444,
-0.15421876311302185,
0.14166992902755737,
-0.0280238576233387,
0.038762785494327545,
0.030038727447390556,
-0.06489133834838867,
0.09397515654563904,
0.07214400172233582,
0.12375238537788391,
-0.062187790870666504,
0.03233812376856804,
0.12421964854001999,
-0.07552601397037506,
0.12172015011310577,
-0.20547735691070557,
0.028054114431142807,
-0.026961537078022957,
-0.04178102687001228,
0.023571038618683815,
0.03792646527290344,
-0.006924719084054232,
-0.053224340081214905,
-0.061626918613910675,
-0.03059104084968567,
0.01796046644449234,
-0.021074745804071426,
0.10203693807125092,
-0.02027503401041031,
0.05525292828679085,
0.19039446115493774,
0.05773130804300308,
-0.18070979416370392,
0.01605962961912155,
0.012348812073469162,
-0.008738983422517776,
0.07113412767648697,
-0.09749951958656311,
0.07622762024402618,
0.08148825913667679,
0.0044942921958863735,
0.0850520059466362,
0.07252969592809677,
-0.06681507080793381,
0.029839973896741867,
0.049486760050058365,
-0.13071362674236298,
-0.03679923340678215,
0.03345613554120064,
0.038375385105609894,
-0.08006482571363449,
0.11059937626123428,
0.18697188794612885,
-0.007275967858731747,
0.007385164499282837,
-0.02495145983994007,
0.05193079635500908,
-0.0037599436473101377,
0.18098324537277222,
-0.020387856289744377,
0.08254574239253998,
-0.11762428283691406,
0.06706500053405762,
-0.014247642830014229,
-0.014681797474622726,
0.05795710161328316,
0.05226687714457512,
-0.09325608611106873,
-0.05763670429587364,
-0.00014375626051332802,
0.13611061871051788,
-0.04494529217481613,
-0.06218542903661728,
-0.14072579145431519,
-0.15603958070278168,
0.07454486936330795,
0.10371926426887512,
0.06364317238330841,
0.028540121391415596,
-0.004515030886977911,
-0.09764640778303146,
-0.08668818324804306,
0.08907326310873032,
0.06805017590522766,
0.10456374287605286,
-0.14961928129196167,
0.09557400643825531,
-0.035465676337480545,
0.024356069043278694,
-0.022185320034623146,
0.011740880087018013,
-0.13531552255153656,
0.005111923906952143,
-0.2744779586791992,
0.05461478978395462,
-0.10227175801992416,
0.0222335122525692,
-0.0028078139293938875,
0.08735120296478271,
-0.04848829284310341,
0.006288753356784582,
-0.050854310393333435,
-0.02188616804778576,
0.0027201054617762566,
0.0322895422577858,
-0.082276351749897,
-0.03504301980137825,
-0.0439506433904171,
-0.09611321985721588,
0.11506885290145874,
0.0279531292617321,
-0.0661730244755745,
-0.022183477878570557,
-0.04183351993560791,
-0.02863805927336216,
0.1450403481721878,
0.04221206530928612,
0.06241731718182564,
-0.10432640463113785,
0.024930043146014214,
0.053281571716070175,
0.049227263778448105,
0.026966826990246773,
0.016649600118398666,
-0.07939327508211136,
0.011879055760800838,
-0.08613505214452744,
-0.03386824205517769,
-0.07665610313415527,
0.026079827919602394,
0.12441346049308777,
0.052968572825193405,
0.10895934700965881,
-0.06233041733503342,
-0.00278872181661427,
-0.11621282994747162,
-0.01567002385854721,
0.007413018029183149,
-0.12206922471523285,
0.040232133120298386,
0.008491378277540207,
0.061354052275419235,
-0.0432308129966259,
0.23805813491344452,
-0.04755182936787605,
-0.03379868343472481,
-0.013668551109731197,
0.0076592471450567245,
0.037084393203258514,
0.02091321162879467,
0.18776321411132812,
0.015397227369248867,
-0.051630716770887375,
-0.06769263744354248,
0.04328492656350136,
0.17486344277858734,
0.13544735312461853,
0.17785541713237762,
0.06334175914525986,
-0.02722761407494545,
0.1207704246044159,
0.018954290077090263,
-0.09588683396577835,
-0.03337850049138069,
0.07601093500852585,
-0.03302754461765289,
0.09510082006454468,
0.0021818389650434256,
0.09700954705476761,
0.19493120908737183,
-0.08180098980665207,
-0.02268490567803383,
-0.027513019740581512,
-0.0752784013748169,
-0.11664441972970963,
-0.1521342694759369,
-0.11431266367435455,
-0.0636037066578865,
-0.02395947463810444,
-0.11510506272315979,
-0.051594607532024384,
0.10779908299446106,
0.0558614656329155,
0.008330444805324078,
0.04756592586636543,
-0.02277456969022751,
-0.011716584675014019,
0.00606370298191905,
-0.00033735367469489574,
-0.021525558084249496,
0.03554973378777504,
-0.046424947679042816,
0.07755172252655029,
-0.02494807355105877,
0.03405357897281647,
0.021056102588772774,
0.0942530483007431,
0.08438976854085922,
-0.08961572498083115,
-0.055735375732183456,
-0.02467929944396019,
0.01188252866268158,
0.05137292295694351,
0.03711773827672005,
0.07958139479160309,
-0.09077410399913788,
-0.024013232439756393,
0.07952263951301575,
-0.09538856893777847,
-0.042183224111795425,
-0.09171479195356369,
0.18639960885047913,
-0.027180496603250504,
0.06874535232782364,
-0.01998739317059517,
-0.07039933651685715,
-0.01604781486093998,
0.14857663214206696,
0.20033147931098938,
-0.02238933928310871,
-0.0019411859102547169,
-0.04167049005627632,
-0.030000105500221252,
-0.08398037403821945,
0.0821700468659401,
0.07691393792629242,
0.12888264656066895,
0.005031662993133068,
-0.011810526251792908,
-0.0034793103113770485,
0.0016033019637688994,
-0.11387208849191666,
0.07084333896636963,
-0.00477257277816534,
-0.00589194567874074,
-0.030145324766635895,
0.011595925316214561,
-0.06677698343992233,
0.007764806970953941,
-0.0025316132232546806,
-0.03759630769491196,
-0.11463071405887604,
0.023081587627530098,
-0.07390288263559341,
0.008807973936200142,
0.10069458931684494,
-0.09477206319570541,
-0.023078661412000656,
0.04983644559979439,
-0.026717456057667732,
-0.09818797558546066,
-0.08426626771688461,
0.02620876580476761,
0.01543025765568018,
0.048697181046009064,
0.002533020917326212,
0.09731101244688034,
0.15121419727802277,
-0.006524829659610987,
-0.11224646866321564,
0.10903527587652206,
0.006506567820906639,
-0.012430638074874878,
0.07030406594276428,
0.07235332578420639,
-0.03948712348937988,
0.05558441951870918,
0.08214092999696732,
-0.06705756485462189,
-0.014590559527277946,
-0.03967280313372612,
-0.011880835518240929,
-0.1136234775185585,
-0.05063289776444435,
-0.039054904133081436,
0.09639782458543777,
0.07994681596755981,
-0.040713708847761154,
-0.0023381507489830256,
0.0038916938938200474,
0.04439133033156395,
0.022442348301410675,
-0.08959244936704636,
0.020127614960074425,
-0.22560831904411316,
0.0819161981344223,
0.0894135907292366,
0.022423161193728447,
-0.2660910189151764,
-0.05131850019097328,
-0.033636923879384995,
-0.03479768708348274,
-0.07633015513420105,
0.023548217490315437,
0.1716759204864502,
0.038317784667015076,
-0.0932241752743721,
-0.18042638897895813,
-0.033510491251945496,
0.12540574371814728,
-0.06212992966175079,
-0.12437327951192856
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bart-summerization-article-title
This model is a fine-tuned version of [facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.001
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.1.2
- Datasets 2.16.1
- Tokenizers 0.15.1 | {"license": "mit", "library_name": "peft", "tags": ["generated_from_trainer"], "base_model": "facebook/bart-large-cnn", "model-index": [{"name": "bart-summerization-article-title", "results": []}]} | null | sanjeev-bhandari01/bart-summerization-article-title | [
"peft",
"tensorboard",
"safetensors",
"generated_from_trainer",
"base_model:facebook/bart-large-cnn",
"license:mit",
"region:us"
] | 2024-02-07T09:08:16+00:00 | [] | [] | TAGS
#peft #tensorboard #safetensors #generated_from_trainer #base_model-facebook/bart-large-cnn #license-mit #region-us
|
# bart-summerization-article-title
This model is a fine-tuned version of facebook/bart-large-cnn on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.001
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.1.2
- Datasets 2.16.1
- Tokenizers 0.15.1 | [
"# bart-summerization-article-title\n\nThis model is a fine-tuned version of facebook/bart-large-cnn on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.001\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 2",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.1.2\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#peft #tensorboard #safetensors #generated_from_trainer #base_model-facebook/bart-large-cnn #license-mit #region-us \n",
"# bart-summerization-article-title\n\nThis model is a fine-tuned version of facebook/bart-large-cnn on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.001\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 2",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.1.2\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
44,
37,
6,
12,
8,
3,
89,
4,
36
] | [
"passage: TAGS\n#peft #tensorboard #safetensors #generated_from_trainer #base_model-facebook/bart-large-cnn #license-mit #region-us \n# bart-summerization-article-title\n\nThis model is a fine-tuned version of facebook/bart-large-cnn on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.001\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 2### Training results### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.1.2\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.11505012959241867,
0.14853249490261078,
-0.002289543393999338,
0.06789897382259369,
0.12549103796482086,
-0.0008737670723348856,
0.1121751070022583,
0.09412989020347595,
-0.05073770880699158,
0.07250529527664185,
0.10286477953195572,
-0.013630606234073639,
0.042245443910360336,
0.15773624181747437,
-0.018620649352669716,
-0.22925087809562683,
0.018818523734807968,
0.012036544270813465,
-0.06155996024608612,
0.11219188570976257,
0.102948397397995,
-0.09809287637472153,
0.07233914732933044,
0.029587747529149055,
-0.20176997780799866,
0.02145584300160408,
-0.01145950611680746,
-0.08889421075582504,
0.09218611568212509,
-0.0006597429164685309,
0.10705491900444031,
0.014681080356240273,
0.10173561424016953,
-0.14636725187301636,
0.008666589856147766,
0.08747771382331848,
0.008103630505502224,
0.10397666692733765,
0.05934859439730644,
0.03092598356306553,
0.159969761967659,
-0.11659588664770126,
0.09286456555128098,
0.03799092024564743,
-0.08500725030899048,
-0.21406863629817963,
-0.06872556358575821,
0.13213485479354858,
0.08598260581493378,
0.09471995383501053,
0.0017782098148018122,
0.13622421026229858,
-0.08354126662015915,
0.05340377986431122,
0.23613616824150085,
-0.2806931734085083,
-0.08230485767126083,
0.01474037952721119,
0.05576615780591965,
0.025247318670153618,
-0.11621779948472977,
-0.022980498149991035,
0.08154551684856415,
0.047144077718257904,
0.10557075589895248,
-0.003228644607588649,
-0.012809262610971928,
-0.03259129077196121,
-0.14268851280212402,
-0.03805406019091606,
0.21967361867427826,
0.06210421398282051,
-0.06480544060468674,
-0.06764665991067886,
-0.053549036383628845,
-0.031544510275125504,
-0.019493019208312035,
-0.052128247916698456,
0.020546551793813705,
-0.051670998334884644,
-0.054855212569236755,
-0.05823340639472008,
-0.09492552280426025,
-0.05373481661081314,
0.02080262452363968,
0.12808535993099213,
0.04824778437614441,
0.03916873782873154,
-0.0658763125538826,
0.0902581736445427,
0.0347377248108387,
-0.12600603699684143,
0.027973152697086334,
-0.02277008816599846,
-0.005988581571727991,
-0.0328587107360363,
-0.02855755016207695,
-0.019079895690083504,
0.0009464616305194795,
0.11363397538661957,
-0.11562684178352356,
0.04066058248281479,
-0.0032257179263979197,
0.03825650364160538,
-0.0233794916421175,
0.09761437773704529,
-0.05519142746925354,
0.012450966984033585,
0.02377377822995186,
0.11173076182603836,
0.012898102402687073,
-0.017020370811223984,
-0.09602949023246765,
-0.015087180770933628,
0.10265252739191055,
0.036760564893484116,
-0.02344691753387451,
-0.007131959777325392,
-0.02348569594323635,
-0.04176259785890579,
0.08247804641723633,
-0.09447146207094193,
0.030718669295310974,
0.010043924674391747,
-0.0848611444234848,
-0.010950340889394283,
0.02956799976527691,
-0.015935901552438736,
-0.03579750284552574,
0.07281395047903061,
-0.10900172591209412,
-0.015127952210605145,
-0.057010360062122345,
-0.04059886932373047,
0.033030204474925995,
-0.0049003856256604195,
-0.005070128012448549,
-0.11011248826980591,
-0.156513512134552,
-0.007730472832918167,
0.020376743748784065,
-0.04559820517897606,
-0.08086270093917847,
0.00039798120269551873,
-0.08716660737991333,
0.006604906637221575,
-0.02384882979094982,
0.0899989977478981,
-0.028145654127001762,
0.09235335886478424,
0.010571245104074478,
0.014848371967673302,
0.00027892994694411755,
0.026816587895154953,
-0.10206467658281326,
0.05079317465424538,
-0.14967092871665955,
0.012230978347361088,
-0.0735504999756813,
0.02251342497766018,
-0.09316855669021606,
-0.11411227285861969,
-0.008233669213950634,
-0.010911758057773113,
0.0771094337105751,
0.12219248712062836,
-0.13303197920322418,
-0.07497583329677582,
0.13655254244804382,
-0.09996550530195236,
-0.1112506240606308,
0.10859930515289307,
-0.030485589057207108,
0.019072657451033592,
0.048413246870040894,
0.18700243532657623,
0.07430415600538254,
-0.14476676285266876,
-0.03305414691567421,
0.02553505077958107,
0.04208119958639145,
-0.04359422251582146,
0.06768820434808731,
0.0024884736631065607,
-0.057934969663619995,
0.032572463154792786,
-0.05301735922694206,
0.014256434515118599,
-0.09578244388103485,
-0.07244506478309631,
-0.058863554149866104,
-0.0872030183672905,
0.114226795732975,
0.03569384664297104,
0.0330730639398098,
-0.07992527633905411,
-0.10371234267950058,
0.11730668693780899,
0.11899401247501373,
-0.023223020136356354,
0.0026151444762945175,
-0.05153855308890343,
0.09866957366466522,
-0.045754820108413696,
-0.027850089594721794,
-0.16301284730434418,
-0.10388738662004471,
0.04266452416777611,
-0.0731506422162056,
0.007301995065063238,
0.01592833735048771,
0.042319681495428085,
0.08659292012453079,
-0.048688918352127075,
-0.0374453030526638,
-0.14834792912006378,
-0.00833223294466734,
-0.09067506343126297,
-0.17344753444194794,
-0.07113493233919144,
-0.020748581737279892,
0.11053933948278427,
-0.2435665726661682,
0.029464149847626686,
-0.005323630291968584,
0.164636492729187,
0.013785464689135551,
-0.03958222642540932,
-0.021497119218111038,
0.03263659402728081,
0.006326906383037567,
-0.06810414046049118,
0.0369865745306015,
0.03125409409403801,
-0.06886351853609085,
-0.058470483869314194,
-0.09387830644845963,
0.13230226933956146,
0.09595046937465668,
0.02998095005750656,
-0.07859022915363312,
-0.014633851125836372,
-0.10715974122285843,
-0.042301055043935776,
-0.05754004418849945,
-0.012579492293298244,
0.13908958435058594,
0.016132112592458725,
0.13932223618030548,
-0.09285194426774979,
-0.060737933963537216,
0.04322817921638489,
-0.024418991059064865,
-0.033185020089149475,
0.06061168015003204,
0.053573861718177795,
-0.06881975382566452,
0.09378978610038757,
0.14750421047210693,
-0.03712424263358116,
0.1555156409740448,
-0.04844819754362106,
-0.0926298052072525,
-0.03678557276725769,
0.008039141073822975,
-0.026209047064185143,
0.16781383752822876,
-0.0848129540681839,
0.014050351455807686,
0.01838085427880287,
0.004743500612676144,
0.05175461992621422,
-0.15738286077976227,
-0.021539421752095222,
-0.0043659270741045475,
-0.019122228026390076,
-0.020364951342344284,
-0.01530543714761734,
0.03856907784938812,
0.09917499870061874,
0.031390972435474396,
-0.03989722579717636,
0.04326224699616432,
0.007162758149206638,
-0.08177868276834488,
0.18402636051177979,
-0.09193111211061478,
-0.21172982454299927,
-0.12805578112602234,
0.13136859238147736,
-0.07970461994409561,
-0.03300260752439499,
0.02503465674817562,
-0.07599283009767532,
-0.023637065663933754,
-0.06636159121990204,
-0.0022270053159445524,
-0.04341212660074234,
0.001861618715338409,
0.038185760378837585,
0.018548978492617607,
0.12285085022449493,
-0.11704013496637344,
0.007945242337882519,
-0.013858254998922348,
-0.06786224246025085,
-0.04371241107583046,
0.06357618421316147,
0.09368542581796646,
0.09631377458572388,
-0.025370750576257706,
0.01948937214910984,
-0.024747787043452263,
0.23086392879486084,
-0.07188373059034348,
-0.011160848662257195,
0.14897498488426208,
0.01580525003373623,
0.05219727382063866,
0.06520875543355942,
0.0393325574696064,
-0.0772266834974289,
0.01527853962033987,
0.054431285709142685,
-0.025896698236465454,
-0.22450265288352966,
-0.07342465221881866,
-0.021261218935251236,
0.012675371021032333,
0.10910724103450775,
0.057466089725494385,
0.02610091120004654,
0.03782857209444046,
-0.034425873309373856,
0.056263551115989685,
-0.03735119104385376,
0.09294440597295761,
0.06557206064462662,
0.02359328418970108,
0.07594095915555954,
-0.018123948946595192,
-0.02942238748073578,
0.0760674849152565,
-0.025480173528194427,
0.21480613946914673,
-0.03404403477907181,
0.04731568694114685,
0.05235280841588974,
0.18734341859817505,
-0.023187832906842232,
0.02294151298701763,
0.019136231392621994,
-0.01518009789288044,
0.012182042934000492,
-0.05356912314891815,
-0.02136204019188881,
0.025069307535886765,
-0.04789033904671669,
0.08409225195646286,
-0.16394266486167908,
0.016047008335590363,
0.024720773100852966,
0.26640602946281433,
0.03153299540281296,
-0.2828451097011566,
-0.10518193989992142,
-0.016993919387459755,
-0.022843345999717712,
-0.07667728513479233,
0.04581350088119507,
0.1579938530921936,
-0.12146104872226715,
0.03894614428281784,
-0.06779434531927109,
0.07628288120031357,
0.0024055493995547295,
0.027216432616114616,
0.03604612499475479,
0.13831138610839844,
0.006209482904523611,
0.10012571513652802,
-0.1937483847141266,
0.22470997273921967,
0.016756776720285416,
0.10071033984422684,
-0.058576539158821106,
0.0033355574123561382,
-0.004444180987775326,
0.07000995427370071,
0.10525868833065033,
0.0028191327583044767,
-0.009628292173147202,
-0.11996348947286606,
-0.09228798002004623,
0.029345281422138214,
0.0696437805891037,
-0.04392460733652115,
0.047185640782117844,
-0.03089437447488308,
0.01629241742193699,
0.04191378504037857,
-0.025371218100190163,
-0.18061861395835876,
-0.08808053284883499,
0.01303108036518097,
0.009287291206419468,
-0.06154603883624077,
-0.11815666407346725,
-0.10773296654224396,
0.02396446280181408,
0.15169338881969452,
0.01232411339879036,
-0.04415435716509819,
-0.12263201922178268,
0.08100108057260513,
0.09947257488965988,
-0.05385471507906914,
0.005937238223850727,
0.02268855832517147,
0.1736343950033188,
0.004034094978123903,
-0.0960172638297081,
0.06147320568561554,
-0.082661472260952,
-0.191665917634964,
-0.034792184829711914,
0.16334691643714905,
0.06390279531478882,
0.034577056765556335,
0.006877837236970663,
-0.006304444745182991,
0.0086655393242836,
-0.09526807814836502,
0.03190337494015694,
0.009184450842440128,
0.054382964968681335,
0.0054457224905490875,
-0.028266077861189842,
0.054114993661642075,
-0.016124241054058075,
0.018535975366830826,
0.08164726942777634,
0.22977930307388306,
-0.08894703537225723,
0.06914149969816208,
0.13562460243701935,
-0.05590781942009926,
-0.1658518761396408,
0.05323263257741928,
0.08466700464487076,
-0.007646552752703428,
0.04376310855150223,
-0.16738517582416534,
0.12585359811782837,
0.12133778631687164,
-0.03992282599210739,
0.0605291873216629,
-0.32637110352516174,
-0.11147955805063248,
0.10507417470216751,
0.09935000538825989,
0.08773355931043625,
-0.11906087398529053,
-0.04467351362109184,
-0.034394778311252594,
-0.10253622382879257,
0.12742304801940918,
-0.2042102813720703,
0.06460557132959366,
-0.023081129416823387,
0.07376226782798767,
0.019219011068344116,
-0.03719383478164673,
0.14295077323913574,
0.009827271103858948,
0.08883409947156906,
-0.041053347289562225,
0.007924840785562992,
0.12624765932559967,
-0.05338715389370918,
0.07496803998947144,
-0.03510763496160507,
0.07299376279115677,
-0.11689195036888123,
-0.016999999061226845,
-0.10885820537805557,
0.10839257389307022,
-0.0758967474102974,
-0.06253580749034882,
-0.045520707964897156,
0.055413637310266495,
0.017748581245541573,
-0.016431255266070366,
0.054631803184747696,
0.07734978199005127,
0.1462302953004837,
0.15913419425487518,
0.07412411272525787,
-0.030973240733146667,
-0.11067159473896027,
-0.009036466479301453,
-0.029175005853176117,
0.09942154586315155,
-0.10421893745660782,
0.0060676163993775845,
0.11169860512018204,
0.045148853212594986,
0.10648272931575775,
0.03572789579629898,
-0.09122612327337265,
-0.007047710008919239,
0.035322487354278564,
-0.16774407029151917,
-0.1598595231771469,
-0.06482627987861633,
0.0379885770380497,
-0.14963141083717346,
0.035531822592020035,
0.10525335371494293,
-0.11009091883897781,
-0.01984008401632309,
-0.009755831211805344,
0.0014185529435053468,
-0.033057697117328644,
0.14220131933689117,
0.08562394976615906,
0.07326069474220276,
-0.08470773696899414,
0.12838751077651978,
0.05692530423402786,
-0.06206134334206581,
0.0670025497674942,
0.0679050013422966,
-0.06684012711048126,
-0.022211546078324318,
0.04052560776472092,
0.16974076628684998,
-0.027154332026839256,
-0.06362022459506989,
-0.10828469693660736,
-0.10594499856233597,
0.04430567845702171,
0.09097433090209961,
0.04417448118329048,
-0.009369635954499245,
-0.01873384229838848,
0.025974011048674583,
-0.16255231201648712,
0.09603313356637955,
0.04477204009890556,
0.05943111330270767,
-0.17250238358974457,
0.12049337476491928,
0.009340637363493443,
0.06601762771606445,
-0.026724813506007195,
0.013669983483850956,
-0.11130325496196747,
-0.01922467350959778,
-0.16458630561828613,
-0.01852242834866047,
-0.02435750886797905,
0.0016633790219202638,
-0.02147107757627964,
-0.03459099680185318,
-0.0449930801987648,
0.04828079417347908,
-0.06798385083675385,
-0.03725709393620491,
0.041332900524139404,
0.05575853958725929,
-0.15117715299129486,
0.019247230142354965,
0.023685548454523087,
-0.08559275418519974,
0.06733651459217072,
0.04914061725139618,
0.04083250090479851,
0.04478522390127182,
-0.0665813684463501,
0.0026202707085758448,
0.014683800749480724,
0.0038285718765109777,
0.06440608948469162,
-0.05680011585354805,
-0.007517378777265549,
-0.015998506918549538,
0.020784107968211174,
0.008684731088578701,
0.03540594130754471,
-0.13915248215198517,
-0.06680437922477722,
-0.013391831889748573,
-0.05696544423699379,
-0.05719693377614021,
0.022337304428219795,
0.09225104004144669,
0.034302305430173874,
0.1408303827047348,
-0.09306646138429642,
0.03345484286546707,
-0.2045888751745224,
-0.03125709295272827,
-0.008549668826162815,
-0.01917295716702938,
-0.01910506747663021,
-0.005052765365689993,
0.06861308217048645,
-0.044267576187849045,
0.09868858009576797,
0.009681696072220802,
0.09664629399776459,
0.0447620265185833,
-0.0649101510643959,
-0.04563704505562782,
0.021092087030410767,
0.13884451985359192,
0.02471124939620495,
-0.009518370032310486,
0.06637562066316605,
-0.018693016842007637,
0.06972789764404297,
0.06662038713693619,
0.1865466684103012,
0.1443815529346466,
-0.038242157548666,
0.05572399124503136,
0.08895787596702576,
-0.07839865982532501,
-0.15996070206165314,
0.07085319608449936,
0.0022437586449086666,
0.11116692423820496,
-0.03761401027441025,
0.1120661050081253,
0.13575895130634308,
-0.15469470620155334,
0.03126960247755051,
-0.03724321350455284,
-0.08938836306333542,
-0.08497551083564758,
-0.05013735964894295,
-0.05571474879980087,
-0.13678862154483795,
0.024259792640805244,
-0.12523038685321808,
-0.01013933215290308,
0.04481649771332741,
-0.015864182263612747,
-0.033188898116350174,
0.15976214408874512,
0.03702837973833084,
0.008043673820793629,
0.0945715606212616,
0.017543131485581398,
-0.021297931671142578,
-0.0756920725107193,
-0.05950606241822243,
0.03376726806163788,
0.015019683167338371,
0.05006035044789314,
-0.05413400009274483,
-0.020962104201316833,
0.0345039963722229,
0.011373619548976421,
-0.08517958968877792,
0.04031507670879364,
-0.0027652166318148375,
0.042704399675130844,
0.026677297428250313,
0.03815242275595665,
-0.0073525612242519855,
-0.0489259734749794,
0.2639272212982178,
-0.05942785367369652,
-0.01752237230539322,
-0.12608669698238373,
0.1597864180803299,
0.018516752868890762,
-0.03453461825847626,
0.05875357985496521,
-0.10628171265125275,
-0.03532082960009575,
0.14329737424850464,
0.13472214341163635,
-0.048415616154670715,
-0.015976954251527786,
-0.02809911035001278,
-0.02170231193304062,
-0.06400810927152634,
0.13375094532966614,
0.09367915242910385,
0.08142933994531631,
-0.05726318806409836,
-0.008758315816521645,
-0.0362224318087101,
-0.026320939883589745,
-0.07834023982286453,
0.09798506647348404,
0.03221521154046059,
-0.004261030349880457,
-0.060383185744285583,
0.05586615204811096,
-0.01222292147576809,
-0.17141304910182953,
0.008271155878901482,
-0.13321544229984283,
-0.1696683168411255,
-0.02370969019830227,
0.0631556436419487,
-0.002250822028145194,
0.07791249454021454,
-0.023769650608301163,
0.003334661712870002,
0.13005493581295013,
-0.02333492785692215,
-0.050805989652872086,
-0.12860263884067535,
0.09273208677768707,
-0.06737220287322998,
0.2826906442642212,
-0.005758506711572409,
0.058213524520397186,
0.08571194112300873,
0.019978895783424377,
-0.15114501118659973,
-0.009001744911074638,
0.06601861119270325,
-0.02819904498755932,
-0.0061340914107859135,
0.15715844929218292,
-0.01976444572210312,
0.07259565591812134,
0.04393716901540756,
-0.1625603437423706,
-0.06734101474285126,
-0.0522395521402359,
-0.005246095359325409,
-0.07042695581912994,
0.020676270127296448,
-0.05068323388695717,
0.1313093602657318,
0.18401308357715607,
-0.06690866500139236,
-0.02726736292243004,
-0.06674163043498993,
0.05625118315219879,
0.05990099534392357,
0.06492741405963898,
0.03245276212692261,
-0.18158011138439178,
0.009096226654946804,
0.04271359369158745,
-0.0034312286879867315,
-0.2690180838108063,
-0.08149385452270508,
0.03955116495490074,
-0.059234924614429474,
-0.05659149959683418,
0.0802576020359993,
0.037550825625658035,
0.0417577363550663,
-0.039357446134090424,
-0.011086758226156235,
-0.08653818815946579,
0.1397024393081665,
-0.13258269429206848,
-0.055437520146369934
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# solar-10b-ocn-v1
This model is a fine-tuned version of upstage/SOLAR-10.7B-v1.0 on the oncc_medqa_instruct dataset.
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0005
- train_batch_size: 4
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_steps: 10
- num_epochs: 1.0
- mixed_precision_training: Native AMP
### Training script
CUDA_VISIBLE_DEVICES=0 python src/train_bash.py --stage sft --do_train True --model_name_or_path upstage/SOLAR-10.7B-v1.0 --template solar --finetuning_type lora --quantization_bit 4 --flash_attn True --dataset_dir data --dataset oncc_medqa_instruct --cutoff_len 1024 --learning_rate 0.0005 --num_train_epochs 1.0 --max_samples 5000 --per_device_train_batch_size 4 --gradient_accumulation_steps 4 --lr_scheduler_type cosine --max_grad_norm 1.0 --logging_steps 10 --save_steps 100 --warmup_steps 10 --neftune_noise_alpha 0.5 --lora_rank 8 --lora_dropout 0.2 --lora_target wqkv --output_dir /workspace/solar-10b-ocn-v1 --fp16 True --plot_loss True
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.1.1+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
### Performance
Test script:
lm_eval --model hf --model_args pretrained=upstage/SOLAR-10.7B-v1.0,peft=chenhugging/solar-10b-ocn-v1,trust_remote_code=True,parallelize=True,load_in_4bit=True --tasks ocn,aocnp,medmcqa,pubmedqa,mmlu_clinical_knowledge,mmlu_college_medicine,mmlu_professional_medicine --device cuda:0 --limit 100
hf (pretrained=upstage/SOLAR-10.7B-v1.0,peft=chenhugging/solar-10b-ocn-v1,trust_remote_code=True,parallelize=True,load_in_4bit=True), gen_kwargs: (None), limit: 100.0, num_fewshot: None, batch_size: 1
| Tasks |Version|Filter|n-shot| Metric |Value| |Stderr|
|---------------------|-------|------|-----:|--------|----:|---|-----:|
|pubmedqa | 1|none | 0|acc | 0.95|± |0.0219|
|medmcqa |Yaml |none | 0|acc | 0.42|± |0.0496|
|professional_medicine| 0|none | 0|acc | 0.72|± |0.0451|
|college_medicine | 0|none | 0|acc | 0.67|± |0.0473|
|clinical_knowledge | 0|none | 0|acc | 0.64|± |0.0482|
|ocn |Yaml |none | 0|acc | 0.83|± |0.0378|
|aocnp |Yaml |none | 0|acc | 0.72|± |0.0451|
| {"license": "other", "library_name": "peft", "tags": ["llama-factory", "lora", "generated_from_trainer"], "base_model": "upstage/SOLAR-10.7B-v1.0", "model-index": [{"name": "solar-10b-ocn-v1", "results": []}]} | null | chenhugging/solar-10b-ocn-v1 | [
"peft",
"safetensors",
"llama-factory",
"lora",
"generated_from_trainer",
"base_model:upstage/SOLAR-10.7B-v1.0",
"license:other",
"region:us"
] | 2024-02-07T09:12:23+00:00 | [] | [] | TAGS
#peft #safetensors #llama-factory #lora #generated_from_trainer #base_model-upstage/SOLAR-10.7B-v1.0 #license-other #region-us
| solar-10b-ocn-v1
================
This model is a fine-tuned version of upstage/SOLAR-10.7B-v1.0 on the oncc\_medqa\_instruct dataset.
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0005
* train\_batch\_size: 4
* eval\_batch\_size: 8
* seed: 42
* gradient\_accumulation\_steps: 4
* total\_train\_batch\_size: 16
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: cosine
* lr\_scheduler\_warmup\_steps: 10
* num\_epochs: 1.0
* mixed\_precision\_training: Native AMP
### Training script
CUDA\_VISIBLE\_DEVICES=0 python src/train\_bash.py --stage sft --do\_train True --model\_name\_or\_path upstage/SOLAR-10.7B-v1.0 --template solar --finetuning\_type lora --quantization\_bit 4 --flash\_attn True --dataset\_dir data --dataset oncc\_medqa\_instruct --cutoff\_len 1024 --learning\_rate 0.0005 --num\_train\_epochs 1.0 --max\_samples 5000 --per\_device\_train\_batch\_size 4 --gradient\_accumulation\_steps 4 --lr\_scheduler\_type cosine --max\_grad\_norm 1.0 --logging\_steps 10 --save\_steps 100 --warmup\_steps 10 --neftune\_noise\_alpha 0.5 --lora\_rank 8 --lora\_dropout 0.2 --lora\_target wqkv --output\_dir /workspace/solar-10b-ocn-v1 --fp16 True --plot\_loss True
### Framework versions
* PEFT 0.8.2
* Transformers 4.37.2
* Pytorch 2.1.1+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
### Performance
Test script:
lm\_eval --model hf --model\_args pretrained=upstage/SOLAR-10.7B-v1.0,peft=chenhugging/solar-10b-ocn-v1,trust\_remote\_code=True,parallelize=True,load\_in\_4bit=True --tasks ocn,aocnp,medmcqa,pubmedqa,mmlu\_clinical\_knowledge,mmlu\_college\_medicine,mmlu\_professional\_medicine --device cuda:0 --limit 100
hf (pretrained=upstage/SOLAR-10.7B-v1.0,peft=chenhugging/solar-10b-ocn-v1,trust\_remote\_code=True,parallelize=True,load\_in\_4bit=True), gen\_kwargs: (None), limit: 100.0, num\_fewshot: None, batch\_size: 1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 1.0\n* mixed\\_precision\\_training: Native AMP",
"### Training script\n\n\nCUDA\\_VISIBLE\\_DEVICES=0 python src/train\\_bash.py --stage sft --do\\_train True --model\\_name\\_or\\_path upstage/SOLAR-10.7B-v1.0 --template solar --finetuning\\_type lora --quantization\\_bit 4 --flash\\_attn True --dataset\\_dir data --dataset oncc\\_medqa\\_instruct --cutoff\\_len 1024 --learning\\_rate 0.0005 --num\\_train\\_epochs 1.0 --max\\_samples 5000 --per\\_device\\_train\\_batch\\_size 4 --gradient\\_accumulation\\_steps 4 --lr\\_scheduler\\_type cosine --max\\_grad\\_norm 1.0 --logging\\_steps 10 --save\\_steps 100 --warmup\\_steps 10 --neftune\\_noise\\_alpha 0.5 --lora\\_rank 8 --lora\\_dropout 0.2 --lora\\_target wqkv --output\\_dir /workspace/solar-10b-ocn-v1 --fp16 True --plot\\_loss True",
"### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1",
"### Performance\n\n\nTest script:\nlm\\_eval --model hf --model\\_args pretrained=upstage/SOLAR-10.7B-v1.0,peft=chenhugging/solar-10b-ocn-v1,trust\\_remote\\_code=True,parallelize=True,load\\_in\\_4bit=True --tasks ocn,aocnp,medmcqa,pubmedqa,mmlu\\_clinical\\_knowledge,mmlu\\_college\\_medicine,mmlu\\_professional\\_medicine --device cuda:0 --limit 100\n\n\nhf (pretrained=upstage/SOLAR-10.7B-v1.0,peft=chenhugging/solar-10b-ocn-v1,trust\\_remote\\_code=True,parallelize=True,load\\_in\\_4bit=True), gen\\_kwargs: (None), limit: 100.0, num\\_fewshot: None, batch\\_size: 1"
] | [
"TAGS\n#peft #safetensors #llama-factory #lora #generated_from_trainer #base_model-upstage/SOLAR-10.7B-v1.0 #license-other #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 1.0\n* mixed\\_precision\\_training: Native AMP",
"### Training script\n\n\nCUDA\\_VISIBLE\\_DEVICES=0 python src/train\\_bash.py --stage sft --do\\_train True --model\\_name\\_or\\_path upstage/SOLAR-10.7B-v1.0 --template solar --finetuning\\_type lora --quantization\\_bit 4 --flash\\_attn True --dataset\\_dir data --dataset oncc\\_medqa\\_instruct --cutoff\\_len 1024 --learning\\_rate 0.0005 --num\\_train\\_epochs 1.0 --max\\_samples 5000 --per\\_device\\_train\\_batch\\_size 4 --gradient\\_accumulation\\_steps 4 --lr\\_scheduler\\_type cosine --max\\_grad\\_norm 1.0 --logging\\_steps 10 --save\\_steps 100 --warmup\\_steps 10 --neftune\\_noise\\_alpha 0.5 --lora\\_rank 8 --lora\\_dropout 0.2 --lora\\_target wqkv --output\\_dir /workspace/solar-10b-ocn-v1 --fp16 True --plot\\_loss True",
"### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1",
"### Performance\n\n\nTest script:\nlm\\_eval --model hf --model\\_args pretrained=upstage/SOLAR-10.7B-v1.0,peft=chenhugging/solar-10b-ocn-v1,trust\\_remote\\_code=True,parallelize=True,load\\_in\\_4bit=True --tasks ocn,aocnp,medmcqa,pubmedqa,mmlu\\_clinical\\_knowledge,mmlu\\_college\\_medicine,mmlu\\_professional\\_medicine --device cuda:0 --limit 100\n\n\nhf (pretrained=upstage/SOLAR-10.7B-v1.0,peft=chenhugging/solar-10b-ocn-v1,trust\\_remote\\_code=True,parallelize=True,load\\_in\\_4bit=True), gen\\_kwargs: (None), limit: 100.0, num\\_fewshot: None, batch\\_size: 1"
] | [
52,
159,
291,
39,
255
] | [
"passage: TAGS\n#peft #safetensors #llama-factory #lora #generated_from_trainer #base_model-upstage/SOLAR-10.7B-v1.0 #license-other #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 1.0\n* mixed\\_precision\\_training: Native AMP### Training script\n\n\nCUDA\\_VISIBLE\\_DEVICES=0 python src/train\\_bash.py --stage sft --do\\_train True --model\\_name\\_or\\_path upstage/SOLAR-10.7B-v1.0 --template solar --finetuning\\_type lora --quantization\\_bit 4 --flash\\_attn True --dataset\\_dir data --dataset oncc\\_medqa\\_instruct --cutoff\\_len 1024 --learning\\_rate 0.0005 --num\\_train\\_epochs 1.0 --max\\_samples 5000 --per\\_device\\_train\\_batch\\_size 4 --gradient\\_accumulation\\_steps 4 --lr\\_scheduler\\_type cosine --max\\_grad\\_norm 1.0 --logging\\_steps 10 --save\\_steps 100 --warmup\\_steps 10 --neftune\\_noise\\_alpha 0.5 --lora\\_rank 8 --lora\\_dropout 0.2 --lora\\_target wqkv --output\\_dir /workspace/solar-10b-ocn-v1 --fp16 True --plot\\_loss True"
] | [
-0.09903895109891891,
0.09014517068862915,
-0.006485908292233944,
0.045815661549568176,
0.07690686732530594,
0.03371012210845947,
0.03729521110653877,
0.16008949279785156,
-0.035615600645542145,
0.1855882704257965,
0.09854533523321152,
0.11807043105363846,
0.08854478597640991,
0.13058185577392578,
0.009540005587041378,
-0.2551251947879791,
0.0036346795968711376,
-0.08758995682001114,
-0.09405792504549026,
0.10081122070550919,
0.052290208637714386,
-0.07495284825563431,
0.03390083834528923,
-0.04566454142332077,
-0.0067045302130281925,
-0.05763209983706474,
-0.0767873004078865,
-0.03358633816242218,
0.07400006800889969,
0.04357726499438286,
0.05137629434466362,
0.05026555433869362,
0.10098245739936829,
-0.31161734461784363,
0.0036392826586961746,
0.0532977357506752,
-0.00711442157626152,
0.07727617025375366,
0.10645953565835953,
-0.03382980823516846,
0.1205061674118042,
-0.1303487867116928,
0.06896486133337021,
0.011481029912829399,
-0.13338124752044678,
-0.22118668258190155,
-0.05943934991955757,
0.02335166744887829,
0.16909056901931763,
0.048322685062885284,
-0.05688749998807907,
0.05108611658215523,
-0.07309824228286743,
0.08018144965171814,
0.1896846890449524,
-0.1799972653388977,
-0.07106310874223709,
0.013446193188428879,
0.036132775247097015,
-0.007063913159072399,
-0.1006963923573494,
-0.03381532430648804,
-0.01016128621995449,
0.005248748231679201,
0.08424809575080872,
0.014135368168354034,
0.055642105638980865,
0.03887910023331642,
-0.11482716351747513,
-0.06181250885128975,
0.08136977255344391,
0.06652440875768661,
-0.008273821324110031,
-0.12779536843299866,
-0.007847407832741737,
-0.21746644377708435,
-0.052843980491161346,
0.01966898702085018,
-0.002220704685896635,
-0.03675860911607742,
-0.04189171642065048,
0.13351984322071075,
-0.03671759366989136,
-0.05537896603345871,
0.11023223400115967,
0.06708967685699463,
0.07442975789308548,
-0.06393089890480042,
0.05575058236718178,
0.11514829099178314,
0.07515094429254532,
-0.19275599718093872,
-0.008889557793736458,
0.02823040634393692,
-0.1131666749715805,
-0.01456861849874258,
0.00998410303145647,
0.1162620559334755,
0.0624658428132534,
0.23580747842788696,
0.008619123138487339,
0.11263296753168106,
0.07439681142568588,
-0.0008884715498425066,
-0.05183509364724159,
0.10431832075119019,
-0.05750161409378052,
-0.11880332976579666,
-0.08812505751848221,
0.11754071712493896,
-0.029639985412359238,
-0.020054304972290993,
-0.017923643812537193,
0.07573708891868591,
0.07428554445505142,
0.05895911529660225,
0.04619929939508438,
0.05215154588222504,
-0.09015017002820969,
0.0026308814994990826,
0.08883458375930786,
-0.1260707974433899,
0.06747545301914215,
0.08693951368331909,
-0.05991765484213829,
-0.07120199501514435,
0.015003551729023457,
-0.0199574064463377,
-0.03518654778599739,
0.11880812048912048,
-0.012846293859183788,
-0.019672807306051254,
-0.08413247764110565,
-0.0771794244647026,
0.02577647566795349,
-0.0139138950034976,
0.013614002615213394,
-0.04739265516400337,
-0.03376109153032303,
-0.09255881607532501,
0.06515570729970932,
-0.12364565581083298,
-0.059742141515016556,
-0.07537078857421875,
-0.12931396067142487,
0.08052236586809158,
0.003373769810423255,
0.14901062846183777,
-0.06387311220169067,
0.0898597463965416,
0.02964983507990837,
0.054070454090833664,
0.17530182003974915,
0.04754620045423508,
-0.026107754558324814,
0.0827116072177887,
-0.15887406468391418,
0.06863702088594437,
-0.0939180999994278,
0.08655235171318054,
-0.19914427399635315,
-0.11897028982639313,
0.007738177198916674,
0.00477151945233345,
0.1178738921880722,
0.13864518702030182,
-0.2010820508003235,
-0.04122482240200043,
0.15812605619430542,
-0.025001363828778267,
-0.08180110156536102,
0.07812391221523285,
-0.014912322163581848,
-0.0045286607928574085,
-0.026914354413747787,
0.11346107721328735,
0.10018415749073029,
-0.10420774668455124,
-0.01930462196469307,
-0.08157224953174591,
0.12720823287963867,
0.13657131791114807,
0.08567477762699127,
-0.03316883742809296,
0.07349137961864471,
0.004335993435233831,
-0.06197470799088478,
0.02644912339746952,
-0.07758887857198715,
-0.09008859843015671,
0.024584423750638962,
-0.01629982516169548,
-0.010769645683467388,
0.05515896528959274,
-0.012756288051605225,
-0.05821438133716583,
-0.1550656110048294,
-0.04258435219526291,
0.08065159618854523,
-0.07697933167219162,
0.008899782784283161,
-0.0793166533112526,
0.057090215384960175,
0.10233157128095627,
0.034954734146595,
-0.1291942298412323,
-0.0035365421790629625,
0.02804652228951454,
-0.03076877072453499,
-0.0019168704748153687,
-0.08754581958055496,
0.0599873848259449,
0.06847503781318665,
-0.03376569226384163,
-0.052969563752412796,
-0.00520310876891017,
-0.0011685590725392103,
-0.03657229617238045,
-0.24683746695518494,
-0.07069046795368195,
-0.023787356913089752,
0.14606145024299622,
-0.2572284936904907,
0.02305016852915287,
0.0657084733247757,
0.11420156806707382,
-0.022635819390416145,
-0.054894860833883286,
0.003463379805907607,
0.05420588329434395,
-0.017134329304099083,
-0.07164261490106583,
0.03243668004870415,
-0.03130011260509491,
-0.09381559491157532,
-0.04374796152114868,
-0.1588173508644104,
0.0035574520006775856,
0.0844498798251152,
-0.023382442072033882,
-0.1727534830570221,
-0.028363239020109177,
-0.030361760407686234,
-0.07237181812524796,
0.01788288913667202,
0.025288552045822144,
0.1560942679643631,
0.03046288900077343,
0.07049322128295898,
-0.04175843298435211,
-0.05250602588057518,
0.024197008460760117,
0.006743847392499447,
0.0022073339205235243,
0.21662858128547668,
0.057128939777612686,
-0.070254847407341,
0.057177767157554626,
0.07540824264287949,
0.011527781374752522,
0.04165765643119812,
-0.07114946842193604,
-0.08733846992254257,
-0.10283785313367844,
0.0701012834906578,
0.07135576009750366,
0.07160984724760056,
-0.07026195526123047,
-0.007230293471366167,
0.02816103771328926,
-0.017417684197425842,
-0.02928701415657997,
-0.1441940814256668,
-0.016092926263809204,
0.04345931112766266,
-0.04064854606986046,
0.021431131288409233,
-0.03246918320655823,
-0.03642312064766884,
0.042299870401620865,
-0.0022700221743434668,
-0.06921714544296265,
-0.06901513785123825,
-0.056482378393411636,
-0.0804659053683281,
0.19719326496124268,
-0.09798476845026016,
-0.08578541874885559,
-0.09034455567598343,
0.0008535573724657297,
0.09592121094465256,
-0.0361335314810276,
0.021166851744055748,
-0.07374465465545654,
-0.056915100663900375,
-0.07764968276023865,
0.017720360308885574,
-0.012647762894630432,
0.0053024897351861,
-0.01517245452851057,
0.018732210621237755,
0.06344112753868103,
-0.06254880130290985,
0.024694979190826416,
0.024048548191785812,
-0.05593331903219223,
0.06053002178668976,
0.040980108082294464,
0.10476695001125336,
0.10675650835037231,
0.08546894043684006,
0.004385138396173716,
-0.012198108248412609,
0.18756204843521118,
-0.1283339411020279,
0.0271738450974226,
0.04922785237431526,
-0.023287391290068626,
0.04967192932963371,
0.18431110680103302,
0.05611790716648102,
-0.0924120768904686,
0.01635296456515789,
0.011694392189383507,
-0.02085048146545887,
-0.1807500422000885,
-0.0049606929533183575,
-0.040576767176389694,
0.023873023688793182,
0.15942256152629852,
0.04171062633395195,
0.005182857625186443,
0.024016816169023514,
-0.01009651180356741,
-0.06743155419826508,
0.06583445519208908,
0.07397331297397614,
0.018095441162586212,
0.05412605032324791,
0.09969273209571838,
0.0017565523739904165,
-0.010407933034002781,
-0.004411166533827782,
-0.021541843190789223,
0.18383724987506866,
-0.019930385053157806,
0.19405722618103027,
0.05984259396791458,
0.1791284829378128,
-0.029597731307148933,
0.006329999305307865,
0.020116962492465973,
-0.003193276934325695,
0.01886313036084175,
-0.06715337932109833,
0.0005673151463270187,
0.04405941441655159,
0.0625971183180809,
-0.03159821778535843,
-0.06193351373076439,
-0.0048639290034770966,
0.05755188316106796,
0.19597771763801575,
0.09648990631103516,
-0.25379088521003723,
-0.050961416214704514,
0.010997000150382519,
-0.05070909112691879,
-0.010331690311431885,
0.008549045771360397,
0.08283395320177078,
-0.0738971158862114,
0.07367893308401108,
-0.051624201238155365,
0.06991828233003616,
-0.06351513415575027,
0.020849129185080528,
0.15411759912967682,
0.05243253707885742,
0.025277400389313698,
0.03554034233093262,
-0.17626738548278809,
0.2639502286911011,
-0.026758253574371338,
0.04133182764053345,
-0.0245275367051363,
0.057824164628982544,
-0.018455760553479195,
-0.08801771700382233,
0.11820988357067108,
-0.026419678702950478,
-0.15744495391845703,
-0.2071157693862915,
-0.09420548379421234,
-0.0031799112912267447,
0.12246013432741165,
-0.1475227326154709,
0.13788637518882751,
-0.0028162458911538124,
-0.05646493658423424,
0.02515893615782261,
-0.015556445345282555,
-0.0453200563788414,
-0.09369362145662308,
0.0435895174741745,
-0.11012834310531616,
0.04704926908016205,
-0.07221017777919769,
-0.045136865228414536,
-0.17321990430355072,
0.11662627011537552,
-0.1796845644712448,
0.005614467430859804,
-0.10509809851646423,
0.07496478408575058,
0.19343717396259308,
-0.09149083495140076,
0.03983178362250328,
0.028863925486803055,
-0.00407664617523551,
0.0714641734957695,
0.01326798927038908,
0.13029484450817108,
-0.05451441928744316,
-0.22732855379581451,
-0.05001353099942207,
0.1613895297050476,
0.0248650424182415,
0.0636235773563385,
-0.06399264186620712,
0.07881283760070801,
-0.001512028626166284,
-0.10462338477373123,
0.09130377322435379,
-0.03832888975739479,
0.0638161301612854,
0.009944383054971695,
-0.02123362571001053,
0.02257097139954567,
-0.09068271517753601,
-0.057004544883966446,
0.0240631103515625,
0.3441091477870941,
-0.10301077365875244,
0.059221088886260986,
0.010891578160226345,
-0.044476840645074844,
-0.15682877600193024,
-0.07871122658252716,
0.09189300984144211,
0.014971667900681496,
0.03799032047390938,
-0.2182706594467163,
0.0239018052816391,
0.0730547234416008,
0.009924286976456642,
0.010029234923422337,
-0.34051454067230225,
-0.14843405783176422,
0.0797424167394638,
0.058537766337394714,
-0.23058657348155975,
-0.2064802348613739,
-0.08125164359807968,
0.03410572558641434,
-0.10739720612764359,
0.029085997492074966,
0.017305534332990646,
0.07295390218496323,
0.0007178995874710381,
-0.03747212141752243,
0.03642510250210762,
-0.05189324542880058,
0.1955413520336151,
-0.014308437705039978,
0.05161594599485397,
-0.05289626866579056,
-0.03995373845100403,
-0.03444496914744377,
-0.05064163729548454,
-0.0044588735327124596,
-0.10286147147417068,
-0.007342284545302391,
-0.09881936758756638,
-0.0007724097813479602,
-0.06009320542216301,
-0.03818110004067421,
-0.09871310740709305,
-0.006594386417418718,
-0.03187904879450798,
0.07628326117992401,
0.11093005537986755,
-0.00494905561208725,
0.09447842091321945,
-0.04101509228348732,
0.110619455575943,
0.1295432150363922,
0.03885457664728165,
0.13957346975803375,
-0.09950793534517288,
0.03009595163166523,
0.03356802091002464,
-0.0072594089433550835,
-0.15984901785850525,
0.007177910767495632,
0.14511096477508545,
0.03470597043633461,
0.13316909968852997,
0.03319702297449112,
-0.061077192425727844,
-0.036013249307870865,
0.08372917771339417,
-0.042871616780757904,
-0.12420400232076645,
0.021027695387601852,
0.022679446265101433,
-0.12781554460525513,
-0.05654317885637283,
0.08405869454145432,
-0.030640916898846626,
-0.0029150324407964945,
0.03479989245533943,
0.09569962322711945,
-0.01927695795893669,
0.21942368149757385,
-0.0015243527013808489,
0.10996250808238983,
-0.083905890583992,
0.04424072057008743,
0.10006652027368546,
-0.06721298396587372,
0.0416988730430603,
0.1481056660413742,
-0.04475918784737587,
-0.039096325635910034,
0.08176273852586746,
0.03559228405356407,
0.07379388809204102,
0.004599387291818857,
-0.06667081266641617,
-0.09450987726449966,
0.09977348893880844,
0.03701457753777504,
0.0194843802601099,
0.014805768616497517,
0.02006518468260765,
0.03551289811730385,
-0.05909218266606331,
0.1165299043059349,
0.08981305360794067,
0.04310397431254387,
-0.05666263774037361,
0.11232678592205048,
-0.012221570126712322,
-0.04472938925027847,
0.01901850290596485,
0.006814104504883289,
-0.1683313250541687,
0.03260642662644386,
-0.007423787843436003,
0.017407208681106567,
-0.09073678404092789,
-0.010745860636234283,
0.010693608783185482,
-0.024596620351076126,
-0.011333855800330639,
-0.025111446157097816,
-0.07863106578588486,
-0.07493928074836731,
-0.032593730837106705,
0.11121819913387299,
-0.10657454282045364,
-0.019336435943841934,
0.07305251806974411,
-0.12777628004550934,
0.07265955954790115,
-0.0005847187130711973,
0.029627613723278046,
-0.01624094881117344,
-0.1000489667057991,
0.0385674349963665,
-0.0011960924603044987,
-0.01242674421519041,
-0.004173900932073593,
-0.21190784871578217,
-0.018778789788484573,
-0.09386973083019257,
-0.05658799037337303,
0.003601223463192582,
-0.013488688506186008,
-0.1225283071398735,
0.028411291539669037,
-0.05411556735634804,
-0.05992971360683441,
-0.07103302329778671,
0.05730108916759491,
0.026878733187913895,
-0.014037810266017914,
0.08176576346158981,
-0.0712587982416153,
0.08404769748449326,
-0.22895504534244537,
-0.037881430238485336,
0.007739521563053131,
-0.031081601977348328,
-0.015659261494874954,
-0.04128134995698929,
0.1175442561507225,
-0.06374320387840271,
0.028080783784389496,
-0.07384549826383591,
-0.0037249268498271704,
0.015373041853308678,
-0.05015140026807785,
-0.01187827531248331,
0.10223814845085144,
0.08997546136379242,
0.03657178208231926,
-0.021680733188986778,
0.009798133745789528,
0.01306409202516079,
-0.008316822350025177,
0.09184050559997559,
0.11773364990949631,
0.18806695938110352,
0.10155607014894485,
0.0036844515707343817,
0.02310897596180439,
-0.22255118191242218,
-0.05591176822781563,
0.22051970660686493,
-0.10515503585338593,
0.12603524327278137,
-0.04920659214258194,
0.11035659164190292,
0.022546349093317986,
-0.21467304229736328,
0.06402892619371414,
-0.08793304860591888,
-0.10354062914848328,
-0.05732943117618561,
-0.1149783730506897,
-0.0661788210272789,
-0.10404900461435318,
0.012858076952397823,
-0.07183539122343063,
0.0506790392100811,
0.047168731689453125,
0.03559102118015289,
0.04170292615890503,
0.09287079423666,
0.07541774213314056,
0.01599031873047352,
0.07274753600358963,
0.051718927919864655,
-0.027255330234766006,
-0.06277652829885483,
-0.04032938554883003,
-0.00044425862142816186,
-0.018015285953879356,
0.06266028434038162,
-0.020094892010092735,
-0.047885920852422714,
0.07132726907730103,
-0.0007304588216356933,
-0.10345254838466644,
0.041311804205179214,
-0.028165416792035103,
0.024831406772136688,
0.10989029705524445,
0.03955421224236488,
0.006009371485561132,
-0.011299737729132175,
0.2064221352338791,
-0.06673695147037506,
-0.06650030612945557,
-0.12529364228248596,
0.13929326832294464,
-0.005275948904454708,
-0.0069086491130292416,
0.02468291111290455,
-0.06069977581501007,
0.00979381613433361,
0.1165151372551918,
0.15372318029403687,
-0.013319509103894234,
0.0026844306848943233,
0.06935294717550278,
0.00026264722691848874,
0.004799751564860344,
0.038671497255563736,
0.11807422339916229,
0.054380595684051514,
-0.056363578885793686,
-0.007624344900250435,
-0.013620540499687195,
-0.09490212798118591,
-0.04409986361861229,
0.03876567259430885,
0.03791448101401329,
0.018309088423848152,
-0.02658027410507202,
0.13109862804412842,
-0.08022136241197586,
-0.19647297263145447,
0.126100093126297,
-0.23216712474822998,
-0.1586146056652069,
-0.04733746126294136,
0.08600368350744247,
0.026128491386771202,
0.04037261754274368,
0.022244984284043312,
-0.052972499281167984,
0.06644009053707123,
0.035746630281209946,
0.01086290180683136,
-0.03689634054899216,
0.03754040226340294,
-0.11434902995824814,
0.179390549659729,
-0.04784604161977768,
0.0498538464307785,
0.13968878984451294,
0.022181415930390358,
-0.08745555579662323,
0.01901153288781643,
0.1347808539867401,
-0.13210712373256683,
0.05410361662507057,
0.15353015065193176,
-0.025212248787283897,
0.12462680041790009,
0.07076288014650345,
-0.04172428324818611,
0.026701033115386963,
-0.044676147401332855,
-0.004277073312550783,
-0.05827288329601288,
-0.003242779988795519,
-0.038092002272605896,
0.1380925476551056,
0.20658206939697266,
-0.07618799060583115,
-0.008311556652188301,
-0.017101213335990906,
0.017810067161917686,
-0.021131964400410652,
0.16004808247089386,
-0.051182910799980164,
-0.2158454805612564,
0.05690015107393265,
-0.03406433388590813,
0.1109074056148529,
-0.1103125661611557,
-0.11004640907049179,
0.09400361776351929,
-0.045064590871334076,
-0.10303651541471481,
0.15213420987129211,
0.07092348486185074,
0.04054810106754303,
-0.05925609543919563,
-0.13536585867404938,
-0.023112736642360687,
0.16632038354873657,
-0.15022577345371246,
-0.07252424210309982
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# classify_food_items
This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 2.5776
- Accuracy: 0.84
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 64
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 1
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 2.5846 | 0.99 | 62 | 2.5776 | 0.84 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "google/vit-base-patch16-224-in21k", "model-index": [{"name": "classify_food_items", "results": []}]} | image-classification | chethanuk/classify_food_items | [
"transformers",
"tensorboard",
"safetensors",
"vit",
"image-classification",
"generated_from_trainer",
"base_model:google/vit-base-patch16-224-in21k",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T09:12:27+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| classify\_food\_items
=====================
This model is a fine-tuned version of google/vit-base-patch16-224-in21k on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 2.5776
* Accuracy: 0.84
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* gradient\_accumulation\_steps: 4
* total\_train\_batch\_size: 64
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* num\_epochs: 1
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 1",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 1",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
75,
144,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 1### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.14522652328014374,
0.12617382407188416,
-0.0015718641225248575,
0.08754861354827881,
0.15124116837978363,
0.013675065711140633,
0.10483653098344803,
0.1344013810157776,
-0.09252185374498367,
0.09950458258390427,
0.13001105189323425,
0.10407785326242447,
0.042161937803030014,
0.1606171727180481,
-0.02536090277135372,
-0.2850876450538635,
-0.007237509824335575,
0.005200713407248259,
-0.16357281804084778,
0.11490899324417114,
0.09476720541715622,
-0.12036359310150146,
0.0818866714835167,
0.02854841575026512,
-0.14496740698814392,
0.0005201386520639062,
-0.020466238260269165,
-0.049481723457574844,
0.10990928113460541,
0.039411913603544235,
0.10141333192586899,
0.027559753507375717,
0.10222900658845901,
-0.20309972763061523,
0.007204372901469469,
0.07703661918640137,
0.010068726725876331,
0.09041100740432739,
0.08620227873325348,
-0.022107573226094246,
0.09827045351266861,
-0.08274886757135391,
0.0673685371875763,
0.04655694589018822,
-0.09780098497867584,
-0.28402218222618103,
-0.1015307679772377,
0.08827988058328629,
0.12498002499341965,
0.07531459629535675,
-0.020450789481401443,
0.09476195275783539,
-0.07914789766073227,
0.08517707139253616,
0.2442970871925354,
-0.2717107832431793,
-0.09011181443929672,
0.03232664242386818,
0.027278805151581764,
0.0162755586206913,
-0.11894040554761887,
-0.009775553829967976,
0.06894122809171677,
0.013676626607775688,
0.09919550269842148,
0.018197258934378624,
0.009397911839187145,
0.009608784690499306,
-0.14548881351947784,
-0.038664497435092926,
0.13060681521892548,
0.0978391021490097,
-0.04994882643222809,
-0.06416817009449005,
-0.034297920763492584,
-0.22293385863304138,
-0.029782230034470558,
0.006554190069437027,
0.04957640543580055,
-0.06151394918560982,
-0.11877363175153732,
0.01371043547987938,
-0.09690608084201813,
-0.09233798831701279,
0.021163510158658028,
0.14809173345565796,
0.05632686987519264,
-0.004189947620034218,
0.014737572520971298,
0.14198154211044312,
0.03372684493660927,
-0.14901523292064667,
-0.00910921674221754,
0.033385612070560455,
-0.06998342275619507,
-0.03421522304415703,
-0.017664993181824684,
0.004154597874730825,
0.008298155851662159,
0.18416525423526764,
-0.04923563078045845,
0.05858105421066284,
0.053383223712444305,
0.03625434637069702,
-0.10279570519924164,
0.17683784663677216,
-0.07310236245393753,
-0.03282950073480606,
-0.028785379603505135,
0.10458208620548248,
0.014183767139911652,
-0.0005685172509402037,
-0.08654222637414932,
0.039124567061662674,
0.08891225606203079,
0.03332549333572388,
-0.02760612964630127,
0.04045799747109413,
-0.06188612058758736,
-0.020820816978812218,
0.06355984508991241,
-0.07379846274852753,
0.03984476998448372,
0.010651876218616962,
-0.10090301930904388,
-0.044174838811159134,
0.03612698242068291,
0.018583714962005615,
0.02526993118226528,
0.16679272055625916,
-0.1061367392539978,
-0.0027110048104077578,
-0.09826980531215668,
-0.08290842175483704,
0.025049155578017235,
-0.07374825328588486,
0.003444439033046365,
-0.09331554174423218,
-0.14185495674610138,
-0.046126462519168854,
0.06703463196754456,
-0.04756467416882515,
-0.07416601479053497,
-0.046010978519916534,
-0.09896262735128403,
0.04226485267281532,
-0.00992924440652132,
0.14898602664470673,
-0.06124398112297058,
0.1084393858909607,
0.044651471078395844,
0.06263669580221176,
0.03536330536007881,
0.04516308009624481,
-0.06954263895750046,
0.06061567738652229,
-0.1841718554496765,
0.0316317118704319,
-0.0966431275010109,
0.09397074580192566,
-0.11713604629039764,
-0.10872241854667664,
-0.015528630465269089,
-0.0010904985247179866,
0.07629762589931488,
0.1165660172700882,
-0.14139120280742645,
-0.09285715967416763,
0.14976665377616882,
-0.08601713925600052,
-0.13936728239059448,
0.11677738279104233,
-0.015067992731928825,
-0.000004675282525568036,
0.028987305238842964,
0.11793825030326843,
0.08009540289640427,
-0.11712327599525452,
-0.03991934284567833,
-0.030668487772345543,
0.09024041146039963,
-0.010638374835252762,
0.10083328932523727,
-0.031963642686605453,
0.027774911373853683,
0.00725618377327919,
-0.07564413547515869,
0.050610706210136414,
-0.11136433482170105,
-0.08434286713600159,
-0.02820713073015213,
-0.09512121230363846,
0.03390110284090042,
0.06490287184715271,
0.0644783228635788,
-0.09659311175346375,
-0.1462012529373169,
0.029692521318793297,
0.12241950631141663,
-0.06353963166475296,
0.00866908859461546,
-0.06543703377246857,
0.09440114349126816,
-0.05294656753540039,
-0.019138971343636513,
-0.14053629338741302,
-0.07066763937473297,
0.01932523027062416,
-0.04839301109313965,
0.0015747689176350832,
-0.03561047464609146,
0.0720781460404396,
0.07729462534189224,
-0.07031764090061188,
-0.06864264607429504,
-0.06911701709032059,
-0.018851226195693016,
-0.09755486994981766,
-0.24226237833499908,
-0.08148568123579025,
-0.01816769503057003,
0.19115570187568665,
-0.26088476181030273,
0.03390834107995033,
0.022782910615205765,
0.15053988993167877,
0.0639447420835495,
-0.0410531722009182,
-0.032953791320323944,
0.044406067579984665,
-0.04536868631839752,
-0.08645313233137131,
0.02956906519830227,
-0.005515079479664564,
-0.09463240206241608,
-0.02908383123576641,
-0.08234808593988419,
0.15294404327869415,
0.12006604671478271,
-0.030611904338002205,
-0.11362364143133163,
-0.0404602475464344,
-0.08484916388988495,
-0.04864739254117012,
-0.03779292851686478,
-0.007024991326034069,
0.09475849568843842,
0.013433664105832577,
0.1361914873123169,
-0.07485278695821762,
-0.06022496521472931,
0.03463727980852127,
-0.011972232721745968,
-0.019561147317290306,
0.1283561885356903,
0.13958682119846344,
-0.06506290286779404,
0.14771759510040283,
0.10981569439172745,
-0.08076275885105133,
0.1403615027666092,
-0.04854581877589226,
-0.09733442962169647,
-0.016698310151696205,
0.03775491192936897,
0.03429438918828964,
0.15707914531230927,
-0.11519376188516617,
-0.01376724150031805,
0.01625298149883747,
0.008921384811401367,
0.023890117183327675,
-0.21108406782150269,
-0.017944779247045517,
0.029705636203289032,
-0.04841373488306999,
-0.0019206303404644132,
-0.02848970890045166,
-0.021647488698363304,
0.09218907356262207,
0.01798267848789692,
-0.04188157990574837,
0.012166742235422134,
0.011348506435751915,
-0.07287881523370743,
0.21017158031463623,
-0.07898492366075516,
-0.12846148014068604,
-0.158858060836792,
0.014159227721393108,
-0.04141365364193916,
0.005608241073787212,
0.040855105966329575,
-0.11313634365797043,
-0.0407530777156353,
-0.062258463352918625,
0.04120006039738655,
-0.0026289334055036306,
0.03484822437167168,
-0.005249921698123217,
0.03248288109898567,
0.08803386986255646,
-0.10405568778514862,
0.023399345576763153,
-0.026972737163305283,
-0.05684598535299301,
0.030499102547764778,
0.04011298343539238,
0.10680755227804184,
0.1430821567773819,
0.018765432760119438,
0.022478874772787094,
-0.028869329020380974,
0.17497245967388153,
-0.1111358255147934,
-0.019801300019025803,
0.10862936079502106,
0.020469820126891136,
0.05123657360672951,
0.12126675248146057,
0.05439896881580353,
-0.0967959612607956,
0.037287387996912,
0.08107324689626694,
-0.0195027943700552,
-0.2140463888645172,
-0.014465969055891037,
-0.034568384289741516,
0.008347718045115471,
0.12220188230276108,
0.055702537298202515,
0.03036959283053875,
0.07403291016817093,
-0.023121582344174385,
0.04460209980607033,
-0.03878402337431908,
0.08002322912216187,
0.0320049487054348,
0.044057492166757584,
0.12752167880535126,
-0.034499432891607285,
-0.036267951130867004,
0.037889596074819565,
-0.006206064950674772,
0.25005054473876953,
-0.021519538015127182,
0.13595907390117645,
0.05933656543493271,
0.17830932140350342,
-0.004649902693927288,
0.062038205564022064,
0.013441350311040878,
-0.05211523547768593,
0.016124900430440903,
-0.058226555585861206,
-0.001140845357440412,
0.06219445541501045,
0.0218890942633152,
0.06800546497106552,
-0.12375745922327042,
0.032604094594717026,
0.0541740320622921,
0.3206157088279724,
0.08031950145959854,
-0.34475356340408325,
-0.10525836795568466,
0.0043173180893063545,
-0.05137115716934204,
-0.03560823202133179,
0.012361230328679085,
0.11597280949354172,
-0.0939144641160965,
0.06309614330530167,
-0.08574756979942322,
0.09331285208463669,
-0.042058419436216354,
-0.009483455680310726,
0.11897988617420197,
0.08809339255094528,
-0.01639948971569538,
0.06523894518613815,
-0.21698206663131714,
0.28849124908447266,
-0.0006852129008620977,
0.07097722589969635,
-0.03932839632034302,
0.03435581550002098,
0.048837512731552124,
0.04376937448978424,
0.08007156103849411,
-0.01571190357208252,
-0.04368017986416817,
-0.20697413384914398,
-0.09321019053459167,
0.009753131307661533,
0.12471301108598709,
-0.12294625490903854,
0.12559382617473602,
-0.02084699086844921,
-0.03580660745501518,
0.04738481342792511,
-0.04179208725690842,
-0.09684938192367554,
-0.08701854944229126,
0.004772581625729799,
-0.03030863031744957,
0.037771593779325485,
-0.11189121752977371,
-0.12703706324100494,
-0.09437832981348038,
0.16343799233436584,
-0.0886252298951149,
-0.032290078699588776,
-0.13688206672668457,
0.11766688525676727,
0.12409235537052155,
-0.07683072984218597,
0.06592772901058197,
0.003316611750051379,
0.11433561146259308,
0.0350368395447731,
-0.01881210319697857,
0.11709997057914734,
-0.08663295954465866,
-0.24229960143566132,
-0.06702174246311188,
0.13694366812705994,
0.02598082274198532,
0.04186505451798439,
-0.0403045155107975,
0.023169994354248047,
-0.01000964641571045,
-0.08576901257038116,
0.06082000210881233,
-0.03267408907413483,
0.05647902935743332,
0.043665751814842224,
-0.027786824852228165,
0.025637028738856316,
-0.04818591848015785,
-0.059056542813777924,
0.09006356447935104,
0.31483885645866394,
-0.09780893474817276,
-0.015251241624355316,
0.0212313961237669,
-0.03012407198548317,
-0.16139012575149536,
0.06686427444219589,
0.12464409321546555,
0.02079891413450241,
0.021582892164587975,
-0.18819494545459747,
0.1115226000547409,
0.11448885500431061,
-0.04638370871543884,
0.17399749159812927,
-0.25634580850601196,
-0.14469067752361298,
0.08584877103567123,
0.1227308064699173,
-0.03522169962525368,
-0.1732255220413208,
-0.06776931881904602,
-0.01002990361303091,
-0.12611033022403717,
0.08622965961694717,
-0.04403003305196762,
0.10848143696784973,
-0.007110068574547768,
0.024032965302467346,
0.014156917110085487,
-0.05891929566860199,
0.15273450314998627,
-0.009288273751735687,
0.09331424534320831,
-0.02335713803768158,
-0.010152122005820274,
0.026690971106290817,
-0.06079662963747978,
0.02061203308403492,
-0.07574407756328583,
0.033864255994558334,
-0.09819526970386505,
-0.02256118133664131,
-0.08282005041837692,
0.052741363644599915,
-0.05883737653493881,
-0.04553870111703873,
-0.0435970313847065,
0.054749250411987305,
0.007706104312092066,
-0.004401977173984051,
0.17166414856910706,
0.006840928923338652,
0.1562190055847168,
0.09819537401199341,
0.039793938398361206,
-0.0295838825404644,
-0.089061439037323,
-0.016530770808458328,
-0.019168352708220482,
0.07156507670879364,
-0.16546989977359772,
0.019152846187353134,
0.1327972263097763,
0.05076620355248451,
0.14493533968925476,
0.0637160986661911,
-0.06225937232375145,
0.01699827052652836,
0.10312416404485703,
-0.10172098875045776,
-0.08704885840415955,
-0.023499371483922005,
-0.03563990816473961,
-0.1331326961517334,
0.0704464241862297,
0.09856309741735458,
-0.06684400886297226,
-0.005323141813278198,
0.0025733637157827616,
0.017481612041592598,
-0.03114287741482258,
0.20833820104599,
0.07938282191753387,
0.07915978878736496,
-0.08725275099277496,
0.08958721160888672,
0.04652116075158119,
-0.13184264302253723,
-0.007877260446548462,
0.05865797400474548,
-0.05908072739839554,
-0.01991390623152256,
0.026895996183156967,
0.1096286028623581,
-0.025746669620275497,
-0.053028423339128494,
-0.15356862545013428,
-0.11253246665000916,
0.07280454784631729,
0.10040508955717087,
0.07440723478794098,
0.019893700256943703,
0.0009046113700605929,
0.06554759293794632,
-0.1116587445139885,
0.11549513041973114,
0.07559346407651901,
0.10283388942480087,
-0.19565175473690033,
0.13675743341445923,
0.01809048280119896,
0.020234741270542145,
-0.0027154902927577496,
0.03323003277182579,
-0.12047309428453445,
-0.0075713638216257095,
-0.08632047474384308,
-0.03233886510133743,
-0.05012365058064461,
-0.0029623648151755333,
0.018179094418883324,
-0.05564497038722038,
-0.06867686659097672,
0.023523647338151932,
-0.12210308015346527,
-0.05405210703611374,
0.024855127558112144,
0.05860923230648041,
-0.11505228281021118,
-0.01395000796765089,
0.04545934870839119,
-0.11166830360889435,
0.09227310866117477,
0.03176369518041611,
0.06029698625206947,
0.03050892800092697,
-0.10391602665185928,
0.03935547545552254,
0.05191883072257042,
-0.02259714901447296,
0.04515581205487251,
-0.11614730954170227,
-0.002800658345222473,
-0.05517025664448738,
0.044431932270526886,
-0.009782007895410061,
0.040281787514686584,
-0.14976981282234192,
-0.022389937192201614,
-0.0326714850962162,
-0.06369079649448395,
-0.05223982036113739,
0.04979296028614044,
0.05240988731384277,
0.00577034056186676,
0.1822797656059265,
-0.07258819043636322,
0.014399007894098759,
-0.23278523981571198,
0.0017568145412951708,
-0.018508732318878174,
-0.07429967075586319,
-0.08209503442049026,
-0.015088244341313839,
0.07179843634366989,
-0.0680566057562828,
0.07916393876075745,
-0.022564690560102463,
0.06279098987579346,
0.0336635485291481,
-0.05867665633559227,
0.04758964851498604,
0.047582581639289856,
0.15341080725193024,
0.024228284135460854,
-0.0016696912935003638,
0.03197108954191208,
0.0319196954369545,
0.09095267951488495,
0.03745405003428459,
0.17358087003231049,
0.12474983930587769,
-0.04946484789252281,
0.11131452769041061,
0.053978320211172104,
-0.12566624581813812,
-0.17943821847438812,
0.08525242656469345,
-0.06269824504852295,
0.12652631103992462,
-0.02054201252758503,
0.15880069136619568,
0.12047543376684189,
-0.18636412918567657,
0.01588951051235199,
-0.03859018534421921,
-0.06988007575273514,
-0.0924990102648735,
-0.05111957713961601,
-0.07547468692064285,
-0.20497164130210876,
0.029554780572652817,
-0.10280710458755493,
0.010690867900848389,
0.08980406075716019,
0.018001437187194824,
0.013086722232401371,
0.1770557016134262,
0.02572275511920452,
0.022391609847545624,
0.07588726282119751,
0.02576722949743271,
-0.03332126885652542,
-0.06490831077098846,
-0.09072718024253845,
0.008447682484984398,
-0.04036109894514084,
0.034090131521224976,
-0.08048917353153229,
-0.1016898974776268,
0.07089916616678238,
0.053720828145742416,
-0.09590506553649902,
0.022210488095879555,
0.0021382186096161604,
0.053279176354408264,
0.05060799792408943,
-0.005600794218480587,
0.03507012501358986,
-0.03051416389644146,
0.24354709684848785,
-0.10288161784410477,
-0.012247144244611263,
-0.14611181616783142,
0.2187098264694214,
0.012489591725170612,
-0.010221160016953945,
0.030105862766504288,
-0.10981393605470657,
-0.003931883722543716,
0.16583861410617828,
0.15250752866268158,
-0.027688374742865562,
-0.02452072501182556,
0.028037941083312035,
-0.02140806056559086,
-0.05518811196088791,
0.07360252737998962,
0.10784656554460526,
0.08001238107681274,
-0.07417119294404984,
-0.05928494408726692,
-0.027213457971811295,
-0.047732990235090256,
-0.011624841019511223,
0.06797446310520172,
0.022323111072182655,
0.010086123831570148,
-0.04950031265616417,
0.08350266516208649,
-0.024082001298666,
-0.14183755218982697,
0.11822318285703659,
-0.19181287288665771,
-0.18818612396717072,
-0.03239361569285393,
0.07885541766881943,
0.005226122681051493,
0.06997416913509369,
-0.005274651572108269,
-0.04678092151880264,
0.08582624047994614,
-0.00829342007637024,
-0.06834834814071655,
-0.1413375735282898,
0.07566316425800323,
-0.08377617597579956,
0.2601117193698883,
-0.05980459228157997,
0.012960643507540226,
0.13074439764022827,
0.033669013530015945,
-0.08958464115858078,
0.015443435870110989,
0.0672626793384552,
-0.12180906534194946,
0.03434606269001961,
0.146161749958992,
-0.03752860054373741,
0.1226252019405365,
0.03248409181833267,
-0.14034128189086914,
0.01145972777158022,
-0.0645560622215271,
-0.04811616241931915,
-0.09005589038133621,
0.0014640215085819364,
-0.061543334275484085,
0.12580940127372742,
0.22646193206310272,
-0.050253693014383316,
-0.012534542009234428,
-0.059854086488485336,
0.04458586871623993,
0.06259529292583466,
0.08308935910463333,
-0.02042110078036785,
-0.26737967133522034,
0.0503995306789875,
0.01577383652329445,
-0.013966238126158714,
-0.2508055865764618,
-0.08383060246706009,
0.03549308329820633,
-0.06081656739115715,
-0.06943875551223755,
0.08763908594846725,
0.0975363701581955,
0.0578521303832531,
-0.0589803121984005,
-0.08460304141044617,
-0.07015988975763321,
0.17148979008197784,
-0.1701369285583496,
-0.08009831607341766
] |
null | null | transformers | <u><b>We are creating a spatial aware vision-language(VL) model.</b></u>
This is a trained model on COCO dataset images including extra information regarding the spatial relationship between the entities of the image.
This is a sequence to sequence model for image-captioning. The architecture is <u><b>ViT encoder and GPT2 decoder.</b></u>
<details>
<summary>Requirements!</summary>
- 4GB GPU RAM.
- CUDA enabled docker
</details>
The way to download and run this:
```python
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
from transformers import pipeline
image_captioner = pipeline("image-to-text", model="VCL3D/rgb-language_cap", max_new_tokens=200, device=device)
filename = 'path/to/file'
generated_captions = image_captioner(filename)
print(generated_captions)
```
The model is trained to produce as many words as possible with a maximum of 200 tokens, which translates to roughly 5 sentences, while the 6th sentence is usually cropped.
<i>The output is always of that form: "Object1" is to the "Left/Right etc." of the "Object2".</i>
## IF YOU WANT TO PRODUCE A SPECIFIC NUMBER OF CAPTIONS UP TO 5.
```python
import os
def print_up_to_n_sentences(captions, n):
for caption in captions:
generated_text = caption.get('generated_text', '')
sentences = generated_text.split('.')
result = '.'.join(sentences[:n])
#print(result)
return result
filename = 'path/to/file'
generated_captions = image_captioner(filename)
captions = print_up_to_n_sentences(generated_captions, 5)
print(captions)
``` | {"language": ["en"], "license": "mit", "library_name": "transformers", "tags": ["text-generation-inference"], "metrics": ["code_eval"], "pipeline_tag": "image-to-text"} | image-to-text | VCL3D/rgb-language_cap | [
"transformers",
"pytorch",
"safetensors",
"vision-encoder-decoder",
"text-generation-inference",
"image-to-text",
"en",
"license:mit",
"endpoints_compatible",
"region:us"
] | 2024-02-07T09:14:11+00:00 | [] | [
"en"
] | TAGS
#transformers #pytorch #safetensors #vision-encoder-decoder #text-generation-inference #image-to-text #en #license-mit #endpoints_compatible #region-us
| <u><b>We are creating a spatial aware vision-language(VL) model.</b></u>
This is a trained model on COCO dataset images including extra information regarding the spatial relationship between the entities of the image.
This is a sequence to sequence model for image-captioning. The architecture is <u><b>ViT encoder and GPT2 decoder.</b></u>
<details>
<summary>Requirements!</summary>
- 4GB GPU RAM.
- CUDA enabled docker
</details>
The way to download and run this:
The model is trained to produce as many words as possible with a maximum of 200 tokens, which translates to roughly 5 sentences, while the 6th sentence is usually cropped.
<i>The output is always of that form: "Object1" is to the "Left/Right etc." of the "Object2".</i>
## IF YOU WANT TO PRODUCE A SPECIFIC NUMBER OF CAPTIONS UP TO 5.
| [
"## IF YOU WANT TO PRODUCE A SPECIFIC NUMBER OF CAPTIONS UP TO 5."
] | [
"TAGS\n#transformers #pytorch #safetensors #vision-encoder-decoder #text-generation-inference #image-to-text #en #license-mit #endpoints_compatible #region-us \n",
"## IF YOU WANT TO PRODUCE A SPECIFIC NUMBER OF CAPTIONS UP TO 5."
] | [
56,
21
] | [
"passage: TAGS\n#transformers #pytorch #safetensors #vision-encoder-decoder #text-generation-inference #image-to-text #en #license-mit #endpoints_compatible #region-us \n## IF YOU WANT TO PRODUCE A SPECIFIC NUMBER OF CAPTIONS UP TO 5."
] | [
-0.07656871527433395,
0.03773869574069977,
-0.00553854089230299,
-0.010143841616809368,
0.1575755774974823,
-0.014238134026527405,
0.1245909333229065,
0.06495143473148346,
0.03643117845058441,
0.043070487678050995,
0.14454689621925354,
0.19630801677703857,
-0.0037797975819557905,
0.13106295466423035,
-0.0819983184337616,
-0.27830517292022705,
0.0399826318025589,
0.10222459584474564,
0.0002967144828289747,
0.09461478143930435,
0.025935476645827293,
-0.13673852384090424,
0.06409755349159241,
-0.08753751963376999,
-0.1465959995985031,
0.029218344017863274,
-0.004642181098461151,
-0.08795276284217834,
0.0896797627210617,
0.03228621557354927,
0.04922127351164818,
0.06292108446359634,
-0.04144807159900665,
-0.1502685248851776,
0.026004446670413017,
0.01940654218196869,
-0.1083860769867897,
0.028209300711750984,
0.09999524801969528,
0.04445609077811241,
0.14287687838077545,
-0.11586137115955353,
-0.016401106491684914,
0.042760349810123444,
-0.08569519966840744,
-0.061525698751211166,
0.03555255010724068,
0.09073232859373093,
0.2552317976951599,
0.07278286665678024,
-0.0046195778995752335,
0.04598436877131462,
-0.08528748899698257,
0.14196617901325226,
0.10871798545122147,
-0.17851406335830688,
-0.058462925255298615,
0.13366568088531494,
0.02490515261888504,
0.008361798711121082,
-0.10957381129264832,
0.06926204264163971,
0.040193017572164536,
-0.004325301386415958,
-0.01930120401084423,
-0.07299622148275375,
-0.09749748557806015,
-0.031744081526994705,
-0.08909012377262115,
-0.06079678609967232,
0.21424955129623413,
0.03964114934206009,
0.04652373120188713,
-0.1375952810049057,
-0.10137438029050827,
-0.030202768743038177,
-0.1108628511428833,
0.06773451715707779,
0.00509608443826437,
-0.02206382341682911,
-0.07285434007644653,
0.03489150106906891,
-0.18589472770690918,
-0.041269998997449875,
-0.08552023768424988,
-0.07870366424322128,
0.02200899086892605,
0.048338375985622406,
-0.1517695039510727,
0.01982557401061058,
-0.00034723422140814364,
-0.10908157378435135,
0.04484214633703232,
-0.10916758328676224,
0.0673588216304779,
0.02580215223133564,
-0.0519028902053833,
-0.06050349026918411,
0.11560807377099991,
0.044459085911512375,
-0.04856446012854576,
0.03786308690905571,
-0.09654458612203598,
0.12558545172214508,
-0.021449733525514603,
-0.014323987066745758,
0.0148130739107728,
0.046524640172719955,
0.0044236211106181145,
-0.07370473444461823,
0.06485618650913239,
-0.03309421241283417,
-0.11934714764356613,
-0.02749510109424591,
0.021941473707556725,
0.09604737907648087,
-0.04952205717563629,
0.0590481236577034,
0.021247349679470062,
0.03761982172727585,
0.13655219972133636,
-0.04088477045297623,
-0.020737117156386375,
0.018508095294237137,
0.07706990838050842,
0.04887276142835617,
-0.01612241566181183,
-0.029473507776856422,
-0.11186603456735611,
0.07357493042945862,
-0.06785987317562103,
0.004235752392560244,
-0.03569939732551575,
-0.014150135219097137,
0.07059555500745773,
-0.04959337040781975,
0.013263200409710407,
-0.09161164611577988,
-0.033684466034173965,
0.0174676813185215,
-0.002395619871094823,
0.0013460059417411685,
-0.009963626973330975,
0.006345272995531559,
-0.054431039839982986,
-0.0022205107379704714,
-0.04448261111974716,
0.025794750079512596,
-0.030646249651908875,
0.1428067833185196,
0.0020668942015618086,
0.12850028276443481,
-0.1477026343345642,
0.018926341086626053,
-0.08271707594394684,
-0.022807933390140533,
-0.19712702929973602,
0.04237594082951546,
-0.023923544213175774,
0.1460365653038025,
0.0020571057684719563,
-0.08995617926120758,
-0.06370890885591507,
0.026274586096405983,
-0.05909954383969307,
0.2322128862142563,
-0.08250708132982254,
-0.053214915096759796,
0.29778945446014404,
-0.15816186368465424,
-0.18009452521800995,
0.07725220918655396,
0.0494932122528553,
-0.10268294811248779,
0.07825929671525955,
0.18188074231147766,
0.05068962648510933,
-0.06495039910078049,
0.022682132199406624,
0.14359360933303833,
-0.06876755505800247,
-0.178352490067482,
0.014686581678688526,
0.03344361111521721,
0.049388375133275986,
0.03669755533337593,
0.012906432151794434,
0.043121226131916046,
-0.034852419048547745,
-0.04166882857680321,
0.013156403787434101,
0.022882331162691116,
-0.016076231375336647,
0.10965996980667114,
0.05613699555397034,
-0.09500022232532501,
0.007447052747011185,
-0.09016608446836472,
0.0048048910684883595,
-0.06464973092079163,
0.03135872632265091,
0.0024384120479226112,
0.0557674840092659,
-0.022585956379771233,
0.03260085731744766,
-0.1665649712085724,
-0.039388228207826614,
-0.05046593025326729,
0.13598164916038513,
-0.005128005985170603,
0.0828142911195755,
0.03604404255747795,
0.017369002103805542,
-0.022633282467722893,
0.016008567065000534,
0.11177770793437958,
-0.04522871971130371,
-0.05640243738889694,
-0.10553674399852753,
0.018027806654572487,
-0.027496734634041786,
0.009421641007065773,
-0.16207994520664215,
0.012286768294870853,
0.1605079025030136,
0.10210596770048141,
-0.003283162834122777,
0.0018342933617532253,
0.031830206513404846,
0.011985830031335354,
-0.030833538621664047,
0.022072510793805122,
0.1386752724647522,
-0.03390984237194061,
-0.02752349153161049,
0.14317750930786133,
-0.16173820197582245,
0.16912555694580078,
0.2079131305217743,
-0.3178832530975342,
0.005077871028333902,
-0.14855702221393585,
-0.04725808650255203,
0.033187415450811386,
-0.016019342467188835,
-0.049604762345552444,
-0.06643512099981308,
0.019275769591331482,
0.13767372071743011,
-0.08097000420093536,
-0.04449241980910301,
0.07154925167560577,
-0.023003462702035904,
-0.0449727326631546,
0.07608756422996521,
0.2085951417684555,
-0.26706042885780334,
0.16760168969631195,
0.1893543004989624,
0.05948503687977791,
0.12154175341129303,
-0.04430506378412247,
-0.05241134390234947,
0.10394946485757828,
0.1041734591126442,
-0.028440317139029503,
0.0888403058052063,
-0.27230122685432434,
-0.04316847771406174,
0.06154712662100792,
-0.089911550283432,
0.03322809562087059,
-0.20269909501075745,
-0.03657834231853485,
-0.05401840806007385,
-0.028722548857331276,
-0.022266393527388573,
0.05778541788458824,
0.018798688426613808,
0.08742569386959076,
-0.09574858099222183,
-0.06967127323150635,
0.06913387775421143,
-0.03585232049226761,
-0.05810164660215378,
0.19872401654720306,
0.007102041970938444,
-0.3830462694168091,
-0.09927358478307724,
-0.09443299472332001,
-0.058857422322034836,
0.0264044888317585,
0.108655646443367,
-0.09711187332868576,
-0.050757043063640594,
-0.002873548772186041,
-0.09837526082992554,
-0.07631109654903412,
0.015648240223526955,
-0.029904304072260857,
0.0823308452963829,
-0.057264771312475204,
-0.04947827011346817,
-0.02485445700585842,
-0.06745728105306625,
-0.03789824619889259,
0.14721305668354034,
-0.09019314497709274,
0.05740351602435112,
0.08826372027397156,
0.0053832982666790485,
0.06556370109319687,
-0.06834446638822556,
0.1599728763103485,
-0.10725993663072586,
-0.014427423477172852,
0.19093602895736694,
0.026392647996544838,
0.10476408153772354,
0.2021808922290802,
0.047214459627866745,
-0.09702306240797043,
0.012745770625770092,
0.03172309696674347,
-0.06635227799415588,
-0.1148686483502388,
-0.048088397830724716,
-0.14938761293888092,
0.09192762523889542,
0.13440224528312683,
0.07196484506130219,
0.09226015210151672,
0.12189316749572754,
-0.09827929735183716,
-0.0008706788066774607,
0.04947308450937271,
0.0874156802892685,
0.255159854888916,
-0.05221996828913689,
0.08851984143257141,
-0.07500610500574112,
-0.11401928961277008,
0.022047467529773712,
0.005293677560985088,
0.14468084275722504,
0.023770097643136978,
-0.02550342120230198,
0.048130884766578674,
0.04950258508324623,
0.09687293320894241,
0.16784417629241943,
0.0006417163531295955,
-0.04151720181107521,
-0.01727244071662426,
-0.08414707332849503,
-0.026846662163734436,
0.008179672993719578,
0.0076560708694159985,
-0.16079121828079224,
-0.09053268283605576,
-0.06117876619100571,
0.04728072136640549,
0.06483867019414902,
0.14452984929084778,
-0.1986512541770935,
0.08383084088563919,
0.033079832792282104,
-0.0436166413128376,
-0.10672090202569962,
0.1130334809422493,
0.13065385818481445,
-0.06049228459596634,
0.1187363937497139,
-0.056961242109537125,
0.12013909965753555,
-0.014259599149227142,
0.07418333739042282,
-0.07366418838500977,
-0.13357508182525635,
0.0405726358294487,
0.11307162046432495,
-0.208957701921463,
0.22430439293384552,
-0.012452874332666397,
-0.05070175975561142,
-0.1320437639951706,
0.008034403435885906,
-0.017535407096147537,
0.10249129682779312,
0.17322799563407898,
-0.024186089634895325,
-0.057045117020606995,
-0.08429612219333649,
-0.011523010209202766,
0.0032776726875454187,
0.14410054683685303,
0.009720170870423317,
-0.018746398389339447,
0.023382486775517464,
-0.05637995898723602,
-0.02798863686621189,
0.05491930618882179,
0.022703254595398903,
-0.17231789231300354,
0.03078545443713665,
0.07209242135286331,
0.004906769376248121,
-0.0010238686809316278,
-0.006937847472727299,
-0.015376349911093712,
-0.007444919086992741,
-0.11336937546730042,
-0.08351670950651169,
-0.05684073641896248,
0.043938592076301575,
0.0390748605132103,
-0.04391901195049286,
0.08310378342866898,
-0.08381477743387222,
0.07895974814891815,
-0.09029243141412735,
-0.09577081352472305,
0.1051478311419487,
-0.09188700467348099,
-0.013739126734435558,
-0.014544632285833359,
0.22031953930854797,
-0.19514501094818115,
-0.04342783987522125,
0.025652803480625153,
0.002487598452717066,
-0.027131574228405952,
-0.11351945996284485,
0.015407494269311428,
-0.2019970715045929,
0.15872961282730103,
0.018932264298200607,
-0.06554507464170456,
-0.09868542104959488,
-0.019627686589956284,
0.023805046454072,
0.21206587553024292,
0.16847166419029236,
-0.00010250913328491151,
0.07020267844200134,
0.1977710723876953,
-0.007146487478166819,
-0.28260013461112976,
-0.12848429381847382,
-0.09049111604690552,
-0.0779833197593689,
-0.020196542143821716,
-0.06851164251565933,
0.1102152094244957,
0.03954146057367325,
-0.013881799764931202,
0.1662236601114273,
-0.23101294040679932,
-0.05681804195046425,
0.04618632793426514,
0.1282605230808258,
0.24233382940292358,
-0.21227507293224335,
-0.043923068791627884,
-0.03866420313715935,
-0.097103051841259,
0.1478937715291977,
-0.00736800255253911,
0.10298295319080353,
-0.015842800959944725,
-0.03453245386481285,
0.02932179719209671,
-0.08942608535289764,
0.05784446746110916,
-0.06981043517589569,
0.0723709985613823,
-0.08386122435331345,
-0.036037296056747437,
0.14717769622802734,
0.002293031197041273,
0.004197486210614443,
-0.1626710146665573,
-0.0160781592130661,
-0.06858573108911514,
-0.02538086660206318,
-0.09407196193933487,
-0.03433941677212715,
-0.01927141658961773,
-0.019589344039559364,
-0.003880303120240569,
-0.02287507802248001,
-0.01123107597231865,
0.01453517284244299,
0.22823487222194672,
-0.10111605376005173,
0.1439560204744339,
0.06881821900606155,
0.07838144153356552,
-0.08364345878362656,
-0.1496279090642929,
-0.11048293113708496,
-0.05476410686969757,
0.06725595891475677,
-0.16484497487545013,
0.04529485106468201,
0.07994180172681808,
0.03506696596741676,
0.03332473710179329,
0.09114498645067215,
0.0698288306593895,
-0.013953312300145626,
0.18394839763641357,
-0.13369055092334747,
-0.06374406069517136,
-0.016740290448069572,
0.07057712972164154,
0.03812688589096069,
0.12173070758581161,
0.05351267382502556,
0.016235798597335815,
0.016017762944102287,
-0.0037812937516719103,
0.03848337382078171,
-0.004543336108326912,
0.033872783184051514,
-0.00822119414806366,
0.08034522086381912,
-0.1588803380727768,
0.04871450364589691,
0.016892459243535995,
-0.2559300363063812,
-0.05761043354868889,
0.17795363068580627,
-0.1074029803276062,
-0.11234108358621597,
-0.038748495280742645,
-0.03103666752576828,
-0.0475226491689682,
-0.038790758699178696,
-0.022474270313978195,
-0.11366454511880875,
0.018478866666555405,
0.21396437287330627,
0.039073407649993896,
0.026081886142492294,
0.05429910868406296,
-0.009133993647992611,
-0.08302334696054459,
0.03653228282928467,
-0.12545299530029297,
0.050931595265865326,
-0.12704023718833923,
0.003924990072846413,
0.021237904205918312,
0.09514199197292328,
-0.06807463616132736,
-0.019797308370471,
-0.14666685461997986,
0.03613343834877014,
0.010222348384559155,
0.027170749381184578,
-0.1423274725675583,
-0.02574744261801243,
-0.03774292394518852,
0.06251856684684753,
-0.0406256802380085,
-0.045191969722509384,
-0.09198825806379318,
0.02421686053276062,
-0.026094354689121246,
0.045749977231025696,
-0.07212217152118683,
-0.015327381901443005,
0.04445889592170715,
-0.05129680037498474,
0.07672169059515,
0.024666020646691322,
-0.1180063858628273,
0.013680901378393173,
-0.12182796001434326,
-0.12008912116289139,
0.13976100087165833,
0.037711963057518005,
-0.02741345949470997,
0.16520415246486664,
0.021581104025244713,
0.07848674058914185,
0.030065394937992096,
0.04346919059753418,
-0.07713019102811813,
-0.12845855951309204,
0.0026702245231717825,
-0.052719783037900925,
-0.16580721735954285,
-0.028124606236815453,
-0.021898746490478516,
0.04450676590204239,
-0.010502536781132221,
0.055167216807603836,
-0.05154789239168167,
0.09984242171049118,
-0.09772203862667084,
0.04245753586292267,
0.008793631568551064,
-0.17114365100860596,
-0.024869786575436592,
-0.08748526126146317,
0.03477088734507561,
-0.019769128412008286,
0.22360387444496155,
0.0015687620034441352,
0.07385841012001038,
0.038366831839084625,
0.07205174118280411,
0.07215959578752518,
0.002399287186563015,
0.3449874818325043,
0.03410275653004646,
-0.019762294366955757,
-0.04376165196299553,
0.0850781798362732,
0.04369688406586647,
0.06161922961473465,
0.11225172877311707,
0.09496709704399109,
-0.0413379892706871,
0.05744524300098419,
0.03982726112008095,
0.09940159320831299,
-0.06582767516374588,
-0.06897023320198059,
0.07321257889270782,
0.07697933912277222,
-0.043010029941797256,
-0.05998389050364494,
0.2369488626718521,
-0.033941905945539474,
0.028683435171842575,
0.022431621327996254,
-0.0518120713531971,
-0.100494384765625,
-0.1278630644083023,
-0.03113006055355072,
-0.12353020906448364,
0.0759332999587059,
-0.1009645015001297,
0.003744015470147133,
0.12768614292144775,
0.08574546128511429,
-0.0547250472009182,
0.1547398716211319,
0.0691235214471817,
-0.051132842898368835,
0.1012360006570816,
0.011721096932888031,
0.058084528893232346,
0.027496661990880966,
0.026014916598796844,
0.006775638088583946,
-0.10231555253267288,
-0.019216086715459824,
0.01642928086221218,
0.026110395789146423,
0.03710580989718437,
-0.09414293617010117,
-0.06276459246873856,
-0.003353757318109274,
0.030302179977297783,
0.042606234550476074,
0.23292593657970428,
-0.02395642176270485,
-0.013622078113257885,
-0.011814005672931671,
0.13738101720809937,
-0.10284914821386337,
-0.08573625981807709,
-0.0011557865655049682,
0.12952406704425812,
0.016928788274526596,
0.047226432710886,
-0.031741585582494736,
-0.039144597947597504,
0.021332569420337677,
0.29983240365982056,
0.1506958305835724,
0.01340058259665966,
0.005199149716645479,
0.02923727035522461,
0.022099994122982025,
0.03615449368953705,
0.10608281940221786,
0.08835054934024811,
0.38374629616737366,
-0.06391012668609619,
-0.04405241087079048,
0.022496771067380905,
-0.008487565442919731,
-0.17998117208480835,
-0.08597566187381744,
0.10166500508785248,
-0.09214041382074356,
-0.027117690071463585,
0.185631662607193,
-0.11901906132698059,
0.1545298844575882,
0.09276952594518661,
-0.10738246887922287,
-0.0014032799517735839,
-0.028882108628749847,
0.1614619344472885,
0.04229738935828209,
0.022691773250699043,
-0.028806908056139946,
-0.012835854664444923,
-0.04887096583843231,
0.031026439741253853,
-0.1864936500787735,
-0.03928283974528313,
-0.012848983518779278,
-0.037212349474430084,
0.0371955931186676,
-0.03089264966547489,
0.10112660378217697,
0.04521075263619423,
0.009180682711303234,
0.021488551050424576,
0.05815751478075981,
0.021178094670176506,
-0.06753210723400116,
0.029496630653738976,
0.09404382109642029,
-0.014591087587177753,
-0.08292260766029358,
0.036696773022413254,
-0.15692415833473206,
0.012695624493062496,
0.07102005928754807,
-0.0016361522721126676,
-0.04343779757618904,
0.11415384709835052,
-0.08359606564044952,
0.04580950736999512,
0.08241156488656998,
0.013483671471476555,
-0.05342496559023857,
-0.056673649698495865,
-0.016415461897850037,
-0.005959778092801571,
-0.14461122453212738,
-0.061856936663389206,
-0.07115393877029419,
-0.06667163968086243,
0.03879217803478241,
-0.009274529293179512,
-0.1737123429775238,
0.02502080611884594,
-0.01964755542576313,
0.03785937651991844,
-0.2098417431116104,
0.025511421263217926,
0.11608007550239563,
-0.02597654238343239,
0.007300082594156265,
-0.17392151057720184,
0.0664677619934082,
0.07670105993747711,
-0.10469570010900497,
-0.06264819204807281
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# roberta-base-detect-cheapfake-co1-co2-v2
This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.3242
- Accuracy: 0.855
- F1: 0.8513
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-06
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 5
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
| No log | 1.0 | 38 | 0.6804 | 0.52 | 0.0204 |
| No log | 2.0 | 76 | 0.5288 | 0.8 | 0.7561 |
| No log | 3.0 | 114 | 0.3494 | 0.865 | 0.8657 |
| No log | 4.0 | 152 | 0.3469 | 0.87 | 0.8725 |
| No log | 5.0 | 190 | 0.3242 | 0.855 | 0.8513 |
### Framework versions
- Transformers 4.36.2
- Pytorch 2.0.0
- Datasets 2.1.0
- Tokenizers 0.15.0
| {"license": "mit", "tags": ["generated_from_trainer"], "metrics": ["accuracy", "f1"], "base_model": "roberta-base", "model-index": [{"name": "roberta-base-detect-cheapfake-co1-co2-v2", "results": []}]} | text-classification | hoanghoavienvo/roberta-base-detect-cheapfake-co1-co2-v2 | [
"transformers",
"tensorboard",
"safetensors",
"roberta",
"text-classification",
"generated_from_trainer",
"base_model:roberta-base",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T09:17:42+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #roberta #text-classification #generated_from_trainer #base_model-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us
| roberta-base-detect-cheapfake-co1-co2-v2
========================================
This model is a fine-tuned version of roberta-base on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.3242
* Accuracy: 0.855
* F1: 0.8513
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-06
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 5
### Training results
### Framework versions
* Transformers 4.36.2
* Pytorch 2.0.0
* Datasets 2.1.0
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.36.2\n* Pytorch 2.0.0\n* Datasets 2.1.0\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #roberta #text-classification #generated_from_trainer #base_model-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.36.2\n* Pytorch 2.0.0\n* Datasets 2.1.0\n* Tokenizers 0.15.0"
] | [
63,
98,
4,
30
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #roberta #text-classification #generated_from_trainer #base_model-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5### Training results### Framework versions\n\n\n* Transformers 4.36.2\n* Pytorch 2.0.0\n* Datasets 2.1.0\n* Tokenizers 0.15.0"
] | [
-0.0872729942202568,
0.06608734279870987,
-0.0018402140121906996,
0.10135751217603683,
0.1675155758857727,
0.01357191614806652,
0.16046130657196045,
0.10870633274316788,
-0.10459641367197037,
0.03928138688206673,
0.12669305503368378,
0.15604309737682343,
-0.003485850291326642,
0.13376066088676453,
-0.07856188714504242,
-0.2434987872838974,
0.002600117353722453,
0.032796766608953476,
-0.08247584849596024,
0.11274389922618866,
0.10568597167730331,
-0.1359047293663025,
0.08753908425569534,
-0.009480123408138752,
-0.20837576687335968,
0.03743283823132515,
0.04446638748049736,
-0.0666339248418808,
0.1381738781929016,
0.04532018303871155,
0.13674187660217285,
0.0321464017033577,
0.08125346899032593,
-0.18660156428813934,
0.019100641831755638,
0.06102104112505913,
-0.015661925077438354,
0.08588720113039017,
0.047358542680740356,
-0.029367763549089432,
0.1105559691786766,
-0.0943804681301117,
0.06365340203046799,
0.022779615595936775,
-0.12993334233760834,
-0.20284292101860046,
-0.0707392543554306,
0.03404094651341438,
0.08907989412546158,
0.08137152343988419,
-0.019688066095113754,
0.15175043046474457,
-0.06305147707462311,
0.10196614265441895,
0.2043527066707611,
-0.30909472703933716,
-0.06914675980806351,
0.05690469965338707,
0.02571505308151245,
0.08627467602491379,
-0.11726520955562592,
0.003577222116291523,
0.08073630183935165,
0.020565267652273178,
0.12236417084932327,
-0.03657818213105202,
-0.06486528366804123,
0.012973707169294357,
-0.14006680250167847,
0.004747745115309954,
0.14671659469604492,
0.044876471161842346,
-0.04360802471637726,
-0.039720021188259125,
-0.05654003471136093,
-0.12649153172969818,
-0.04183842986822128,
-0.030490724369883537,
0.04317954182624817,
-0.028885353356599808,
-0.10950763523578644,
-0.01880469359457493,
-0.11704467982053757,
-0.07468824088573456,
-0.06339125335216522,
0.17693647742271423,
0.03373754024505615,
0.0017746867379173636,
-0.03287970647215843,
0.08995188772678375,
-0.041990894824266434,
-0.12596894800662994,
0.0173062514513731,
0.02494867704808712,
0.006079901475459337,
-0.07307393848896027,
-0.06701092422008514,
-0.09771792590618134,
0.030529692769050598,
0.14856252074241638,
-0.06796810030937195,
0.05014408007264137,
0.00719798170030117,
0.04309980943799019,
-0.09405310451984406,
0.16070455312728882,
-0.04462052509188652,
-0.01590851880609989,
0.02999020181596279,
0.055954620242118835,
0.03782695159316063,
-0.00259994063526392,
-0.12479730695486069,
0.018584761768579483,
0.12026433646678925,
0.013157907873392105,
-0.07674044370651245,
0.08831341564655304,
-0.037040259689092636,
0.004826853051781654,
0.012119737453758717,
-0.0926598310470581,
0.034076280891895294,
0.002792097395285964,
-0.061038196086883545,
-0.06735513359308243,
0.027568645775318146,
0.015938162803649902,
0.010338985361158848,
0.1059475913643837,
-0.09464085847139359,
0.01937321200966835,
-0.09371495246887207,
-0.1320083737373352,
0.007251252420246601,
-0.06348101794719696,
0.03772641345858574,
-0.1180994063615799,
-0.15753714740276337,
-0.016460901126265526,
0.04582826793193817,
-0.027392446994781494,
-0.01917022094130516,
-0.05669393762946129,
-0.08262669295072556,
0.012871086597442627,
-0.014798074960708618,
0.09638716280460358,
-0.055261868983507156,
0.0990075021982193,
0.06297668069601059,
0.06720347702503204,
-0.057597566395998,
0.033339887857437134,
-0.10316026955842972,
0.01398573536425829,
-0.21410338580608368,
0.02292998880147934,
-0.06044796481728554,
0.07861638814210892,
-0.07697700709104538,
-0.08058365434408188,
-0.00814063660800457,
0.025052273645997047,
0.07352124899625778,
0.09091745316982269,
-0.14545080065727234,
-0.0749581977725029,
0.16373999416828156,
-0.10211814939975739,
-0.13635163009166718,
0.11923616379499435,
-0.07076428830623627,
0.0707831084728241,
0.0753115564584732,
0.19543437659740448,
0.05862141773104668,
-0.08901738375425339,
0.004593905061483383,
-0.011637193150818348,
0.03875335305929184,
-0.04887599125504494,
0.056160349398851395,
0.007575794123113155,
-0.003982170484960079,
0.013295128010213375,
-0.020810376852750778,
0.04819609969854355,
-0.09241753816604614,
-0.07803653925657272,
-0.040532227605581284,
-0.10706616938114166,
0.05137501284480095,
0.06208209693431854,
0.08334828913211823,
-0.13340911269187927,
-0.08836691081523895,
0.101030632853508,
0.07163234800100327,
-0.06831535696983337,
0.020421959459781647,
-0.07921966165304184,
0.07088523358106613,
-0.06163021922111511,
-0.029186222702264786,
-0.16390518844127655,
-0.04778002202510834,
-0.003683502320200205,
0.03238211199641228,
0.03382107987999916,
0.013120820745825768,
0.08003579825162888,
0.07187844812870026,
-0.07590454816818237,
-0.025836393237113953,
0.0017989911139011383,
0.015454418025910854,
-0.1279689073562622,
-0.21018916368484497,
0.0002308078546775505,
-0.041370946913957596,
0.13068817555904388,
-0.25381556153297424,
0.05464904010295868,
0.008316610008478165,
0.08639609813690186,
0.04123053699731827,
0.0035423818044364452,
-0.04158870503306389,
0.07146783173084259,
-0.05270799621939659,
-0.0525406114757061,
0.05111796781420708,
0.00639673275873065,
-0.0889396071434021,
-0.04092465341091156,
-0.15823200345039368,
0.2003760188817978,
0.14426562190055847,
-0.11590725183486938,
-0.10827797651290894,
0.0013010026887059212,
-0.03934328258037567,
-0.021488802507519722,
-0.049736957997083664,
0.01555278617888689,
0.12489894777536392,
-0.027551330626010895,
0.154577374458313,
-0.073029525578022,
-0.03547552973031998,
0.021688709035515785,
-0.06440326571464539,
0.007946241647005081,
0.1072741374373436,
0.08657519519329071,
-0.1333802342414856,
0.15223877131938934,
0.14959391951560974,
-0.11022619158029556,
0.15195079147815704,
-0.03198954090476036,
-0.053637150675058365,
-0.01929483748972416,
0.0007391137769445777,
0.015339567326009274,
0.10865878313779831,
-0.11414466053247452,
-0.011621037498116493,
-0.0006649643764831126,
0.009101400151848793,
0.022595210000872612,
-0.21989135444164276,
-0.034414105117321014,
0.028739862143993378,
-0.032747987657785416,
0.015512684360146523,
-0.019877566024661064,
-0.00906267762184143,
0.10469471663236618,
-0.0031259881798177958,
-0.0806354209780693,
0.04326443374156952,
0.0021414505317807198,
-0.09153276681900024,
0.2232935130596161,
-0.0727478563785553,
-0.12019981443881989,
-0.129638209939003,
-0.062360215932130814,
-0.04560421034693718,
0.03414854779839516,
0.06840963661670685,
-0.0870562419295311,
-0.03956850245594978,
-0.09643110632896423,
0.007859847508370876,
0.020154552534222603,
0.03445969521999359,
-0.005109645426273346,
0.015471176244318485,
0.0801493301987648,
-0.10811728984117508,
-0.008116022683680058,
-0.05560486763715744,
-0.07294932752847672,
0.0441703274846077,
0.031075919046998024,
0.11682138592004776,
0.14345350861549377,
-0.04505481198430061,
-0.0011850347509607673,
-0.04478038102388382,
0.2206869125366211,
-0.06959537416696548,
-0.0185348242521286,
0.1205262616276741,
-0.01648399978876114,
0.03943542391061783,
0.13573230803012848,
0.06234784424304962,
-0.09766190499067307,
0.03393910452723503,
0.043939799070358276,
-0.03171302750706673,
-0.21824149787425995,
-0.029734496027231216,
-0.030140580609440804,
-0.012653365731239319,
0.08397942781448364,
0.04136360436677933,
0.05041944980621338,
0.07774405181407928,
0.03395861014723778,
0.06845755130052567,
-0.008551309816539288,
0.075651615858078,
0.10689719766378403,
0.04901786148548126,
0.13647863268852234,
-0.06008579581975937,
-0.07536575198173523,
0.0278773196041584,
-0.017337847501039505,
0.19217246770858765,
0.021955497562885284,
0.10969003289937973,
0.05575046315789223,
0.14343991875648499,
0.012412496842443943,
0.06629960983991623,
-0.0029386382084339857,
-0.06613443791866302,
-0.0023021649103611708,
-0.04609857499599457,
-0.012010710313916206,
0.04037316143512726,
-0.09931240975856781,
0.052487149834632874,
-0.1186591237783432,
0.014166528359055519,
0.06883341073989868,
0.20401787757873535,
0.05591077730059624,
-0.3292448818683624,
-0.09674517810344696,
0.025822315365076065,
-0.022533224895596504,
-0.018792448565363884,
0.024753617122769356,
0.11581360548734665,
-0.04625082388520241,
0.04303446784615517,
-0.06826519966125488,
0.07632776349782944,
-0.023244664072990417,
0.04460268095135689,
0.03938813507556915,
0.09966898709535599,
-0.028653578832745552,
0.06922908872365952,
-0.29875320196151733,
0.2818840444087982,
0.0218399316072464,
0.09036339074373245,
-0.045004621148109436,
-0.010589835233986378,
0.029896503314375877,
0.10234888643026352,
0.06287583708763123,
-0.030833205208182335,
-0.10516851395368576,
-0.1963602602481842,
-0.028252962976694107,
0.031008059158921242,
0.11110185086727142,
-0.016297252848744392,
0.10359761118888855,
-0.035132672637701035,
0.005094053689390421,
0.09670829772949219,
-0.03639379143714905,
-0.09239068627357483,
-0.08732052147388458,
-0.033539559692144394,
0.026768984273076057,
-0.029312903061509132,
-0.07860607653856277,
-0.10084111243486404,
-0.12790824472904205,
0.17554251849651337,
-0.030070176348090172,
-0.01597825065255165,
-0.10685541480779648,
0.08972755819559097,
0.03715252876281738,
-0.08516217768192291,
0.043380286544561386,
0.014188054949045181,
0.0847996175289154,
0.021965356543660164,
-0.06403587758541107,
0.135822132229805,
-0.06647633016109467,
-0.16262401640415192,
-0.06226722151041031,
0.09209457784891129,
0.01873127929866314,
0.0410546138882637,
0.006214599125087261,
0.014503015205264091,
-0.010714574716985226,
-0.07666449248790741,
0.03077417053282261,
-0.03508609160780907,
0.05713530257344246,
0.014647097326815128,
-0.06423678249120712,
-0.027519166469573975,
-0.06183774396777153,
-0.0325285904109478,
0.16661019623279572,
0.2905212938785553,
-0.0873231291770935,
-0.013666264712810516,
0.06713801622390747,
-0.067597396671772,
-0.2199885994195938,
0.07242331653833389,
0.015873469412326813,
0.000044437609176384285,
0.051032036542892456,
-0.1385280191898346,
0.12162170559167862,
0.09741615504026413,
-0.023189550265669823,
0.09330318868160248,
-0.2615320086479187,
-0.14632509648799896,
0.13902096450328827,
0.16806024312973022,
0.13617585599422455,
-0.16751323640346527,
-0.027928046882152557,
-0.04351743683218956,
-0.07248232513666153,
0.09983616322278976,
-0.15078981220722198,
0.11086457222700119,
0.0009276975761167705,
0.060362767428159714,
0.00645743403583765,
-0.057103145867586136,
0.11498405039310455,
-0.017754288390278816,
0.13009601831436157,
-0.06993412226438522,
-0.03729512169957161,
0.06652943044900894,
-0.04756797105073929,
0.019536137580871582,
-0.09769174456596375,
0.03154858201742172,
-0.03981928899884224,
-0.03232339769601822,
-0.05424419790506363,
0.05196463316679001,
-0.03609732910990715,
-0.07104159891605377,
-0.052110832184553146,
0.02983844093978405,
0.019142014905810356,
-0.019201911985874176,
0.17396201193332672,
-0.003315763082355261,
0.18386612832546234,
0.14374534785747528,
0.09199702739715576,
-0.061570554971694946,
0.010656214319169521,
0.013296184130012989,
-0.042659860104322433,
0.05921334773302078,
-0.16055232286453247,
0.039599910378456116,
0.11144288629293442,
0.011234069243073463,
0.15234598517417908,
0.08289815485477448,
-0.035302337259054184,
0.029049204662442207,
0.08850565552711487,
-0.1653183549642563,
-0.11043260991573334,
-0.0083167040720582,
-0.05925484001636505,
-0.10199432820081711,
0.08615361154079437,
0.12101855129003525,
-0.0775105431675911,
0.0008343553636223078,
-0.013820976950228214,
0.0011724523501470685,
-0.0505758672952652,
0.1845986247062683,
0.09260492026805878,
0.04503956437110901,
-0.07386721670627594,
0.07568490505218506,
0.04276810213923454,
-0.06992947310209274,
0.018670709803700447,
0.056341543793678284,
-0.07252013683319092,
-0.05313805863261223,
0.07453782856464386,
0.21586599946022034,
-0.05945160612463951,
-0.0537237823009491,
-0.16226504743099213,
-0.11159386485815048,
0.04425402730703354,
0.21474087238311768,
0.10014452785253525,
0.002039348240941763,
-0.02259734831750393,
0.026590479537844658,
-0.13955098390579224,
0.10729434341192245,
0.033762648701667786,
0.08855681121349335,
-0.15473198890686035,
0.165609672665596,
-0.005095246247947216,
0.009081152267754078,
-0.03446359932422638,
0.0403747633099556,
-0.1347254514694214,
-0.0012021787697449327,
-0.12387602031230927,
-0.015150503255426884,
-0.03458235040307045,
0.006460392381995916,
0.012344620190560818,
-0.062199950218200684,
-0.0753328949213028,
0.010721104219555855,
-0.10760463029146194,
-0.00609162263572216,
0.045288022607564926,
0.050721101462841034,
-0.1295875757932663,
-0.035884179174900055,
0.013588041067123413,
-0.05836851894855499,
0.06122823804616928,
0.01864800415933132,
0.020981483161449432,
0.06851623952388763,
-0.203209787607193,
0.03347880393266678,
0.0750625804066658,
-0.013645204715430737,
0.051553260535001755,
-0.07444515079259872,
-0.004776570945978165,
-0.000050774837291100994,
0.07476620376110077,
0.024769313633441925,
0.07402621954679489,
-0.12295974045991898,
0.02085626870393753,
-0.04257002845406532,
-0.0743708461523056,
-0.06189606338739395,
0.03659456595778465,
0.0803104117512703,
-0.01001213863492012,
0.20015501976013184,
-0.11446156352758408,
0.0045402473770082,
-0.20604464411735535,
0.011455217376351357,
-0.008141632191836834,
-0.12097478657960892,
-0.11917497962713242,
-0.052551522850990295,
0.05748990923166275,
-0.06384813040494919,
0.14290359616279602,
0.031459443271160126,
0.023912299424409866,
0.04692283272743225,
-0.04678844287991524,
0.040723253041505814,
0.039486441761255264,
0.23654697835445404,
0.02277066558599472,
-0.03889310732483864,
0.010829330421984196,
0.055599112063646317,
0.1193910613656044,
0.08069777488708496,
0.18367159366607666,
0.16482187807559967,
-0.08147301524877548,
0.11049287021160126,
0.0435648113489151,
-0.04400802403688431,
-0.1121605634689331,
0.05799002945423126,
-0.04138830304145813,
0.0718410313129425,
-0.02870337665081024,
0.16473762691020966,
0.13899612426757812,
-0.15143676102161407,
0.01114574819803238,
-0.06651344895362854,
-0.08113018423318863,
-0.1252189576625824,
-0.04638320952653885,
-0.11290040612220764,
-0.15452703833580017,
0.008271899074316025,
-0.11619820445775986,
0.005179854109883308,
0.08564002811908722,
0.008730563335120678,
-0.023509321734309196,
0.17945750057697296,
0.016412844881415367,
0.038688093423843384,
0.045998819172382355,
-0.0063506620936095715,
-0.034855376929044724,
-0.08197548985481262,
-0.08002108335494995,
0.0018360961694270372,
-0.032309189438819885,
0.011923772282898426,
-0.05191998556256294,
-0.036109525710344315,
0.04553884267807007,
-0.024049049243330956,
-0.10434871166944504,
0.016677705571055412,
0.04211825504899025,
0.056844595819711685,
0.03699883818626404,
0.01086212508380413,
-0.00400565704330802,
-0.00240694428794086,
0.23883016407489777,
-0.07434157282114029,
-0.07534404844045639,
-0.09767315536737442,
0.2590709626674652,
0.047828491777181625,
0.0312962606549263,
-0.00045436667278409004,
-0.10496097803115845,
0.03741767629981041,
0.224201962351799,
0.20938239991664886,
-0.06882165372371674,
0.014231069013476372,
-0.021472429856657982,
-0.009573426097631454,
-0.030119920149445534,
0.10048364102840424,
0.1027907133102417,
0.01580205373466015,
-0.0804869532585144,
-0.05050399526953697,
-0.031922128051519394,
-0.0024417266249656677,
-0.04772862419486046,
0.06414701789617538,
0.046321187168359756,
0.019289502874016762,
-0.04722324013710022,
0.06444154679775238,
-0.03670699894428253,
-0.11471628397703171,
0.05406592786312103,
-0.19506682455539703,
-0.14013375341892242,
-0.012731866911053658,
0.1118592768907547,
-0.013628995977342129,
0.06617546826601028,
-0.03359965234994888,
-0.006109967362135649,
0.02232666313648224,
-0.01840415969491005,
-0.076612189412117,
-0.06934218853712082,
0.06271838396787643,
-0.10443130880594254,
0.21802997589111328,
-0.045185964554548264,
0.039165448397397995,
0.1445290744304657,
0.03509252890944481,
-0.06933101266622543,
0.1030660942196846,
0.03716365993022919,
-0.07602421939373016,
0.04574407637119293,
0.09153630584478378,
-0.047046348452568054,
0.10535863041877747,
0.054263241589069366,
-0.14843639731407166,
0.03165264427661896,
-0.09596555680036545,
-0.08523643016815186,
-0.05946554243564606,
-0.03543872758746147,
-0.06455115228891373,
0.12446678429841995,
0.18663237988948822,
-0.029476214200258255,
0.022938188165426254,
-0.05215229094028473,
0.024238141253590584,
0.07770445197820663,
0.027831247076392174,
-0.04285553842782974,
-0.2513006329536438,
0.030174626037478447,
0.09305700659751892,
-0.016649393364787102,
-0.3053104877471924,
-0.0707617923617363,
-0.011026682332158089,
-0.04613291099667549,
-0.10008113831281662,
0.08647631853818893,
0.1403307020664215,
0.0571872778236866,
-0.0645022988319397,
-0.11901719868183136,
-0.07755094766616821,
0.1639108955860138,
-0.13391998410224915,
-0.10613751411437988
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# OndeviceAI-base-v2
This model is a fine-tuned version of [paust/pko-t5-base](https://huggingface.co/paust/pko-t5-base) on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0007
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- gradient_accumulation_steps: 8
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_ratio: 0.06
- num_epochs: 20
### Training results
### Framework versions
- Transformers 4.37.2
- Pytorch 2.2.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "cc-by-4.0", "tags": ["generated_from_trainer"], "base_model": "paust/pko-t5-base", "model-index": [{"name": "OndeviceAI-base-v2", "results": []}]} | text2text-generation | yeye776/OndeviceAI-base-v2 | [
"transformers",
"tensorboard",
"safetensors",
"t5",
"text2text-generation",
"generated_from_trainer",
"base_model:paust/pko-t5-base",
"license:cc-by-4.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T09:18:11+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-paust/pko-t5-base #license-cc-by-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# OndeviceAI-base-v2
This model is a fine-tuned version of paust/pko-t5-base on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0007
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- gradient_accumulation_steps: 8
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_ratio: 0.06
- num_epochs: 20
### Training results
### Framework versions
- Transformers 4.37.2
- Pytorch 2.2.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| [
"# OndeviceAI-base-v2\n\nThis model is a fine-tuned version of paust/pko-t5-base on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0007\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- gradient_accumulation_steps: 8\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_ratio: 0.06\n- num_epochs: 20",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-paust/pko-t5-base #license-cc-by-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# OndeviceAI-base-v2\n\nThis model is a fine-tuned version of paust/pko-t5-base on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0007\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- gradient_accumulation_steps: 8\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_ratio: 0.06\n- num_epochs: 20",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
83,
36,
6,
12,
8,
3,
129,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-paust/pko-t5-base #license-cc-by-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# OndeviceAI-base-v2\n\nThis model is a fine-tuned version of paust/pko-t5-base on the None dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0007\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- gradient_accumulation_steps: 8\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_ratio: 0.06\n- num_epochs: 20### Training results### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.09774739295244217,
0.159471333026886,
-0.002910156501457095,
0.07069172710180283,
0.13540774583816528,
0.024444855749607086,
0.1120244562625885,
0.13833943009376526,
-0.13005928695201874,
0.09702067822217941,
0.08143133670091629,
0.02534082531929016,
0.07386012375354767,
0.1164434403181076,
0.0023902603425085545,
-0.2538163363933563,
-0.003843888873234391,
-0.009517994709312916,
-0.0801703929901123,
0.11374800652265549,
0.10727081447839737,
-0.08758463710546494,
0.06389588862657547,
0.012745870277285576,
-0.14038856327533722,
0.011503668501973152,
-0.05123597010970116,
-0.06762293726205826,
0.08419380336999893,
0.009390526451170444,
0.08653785288333893,
0.01562909223139286,
0.10662413388490677,
-0.22120192646980286,
-0.005856629461050034,
0.08803702890872955,
0.018571479246020317,
0.0935649424791336,
0.08842431753873825,
-0.0031093184370547533,
0.12590481340885162,
-0.16655728220939636,
0.09394683688879013,
0.03916265815496445,
-0.07101083546876907,
-0.14274664223194122,
-0.07990586757659912,
0.10431285202503204,
0.09798564016819,
0.10488586872816086,
-0.003321371739730239,
0.16057747602462769,
-0.07592733949422836,
0.07072608917951584,
0.19321970641613007,
-0.2829182744026184,
-0.06894294917583466,
0.033836640417575836,
0.06632733345031738,
0.0712922140955925,
-0.1276370733976364,
-0.012264269404113293,
0.04169296473264694,
0.01656610704958439,
0.08440616726875305,
0.009993503801524639,
0.024238942191004753,
-0.00769051956012845,
-0.11153461039066315,
-0.029186980798840523,
0.1437411606311798,
0.06844966858625412,
-0.043797437101602554,
-0.13357038795948029,
-0.04434159770607948,
-0.15547846257686615,
-0.010332665406167507,
-0.027453048154711723,
0.029430260881781578,
-0.04462255537509918,
-0.09064830094575882,
-0.01870385743677616,
-0.08079259097576141,
-0.057527799159288406,
0.04208396375179291,
0.11229338496923447,
0.03894782438874245,
-0.007681777235120535,
0.00009938803850673139,
0.10077309608459473,
0.03992363438010216,
-0.13939088582992554,
-0.009372659027576447,
-0.0016079703345894814,
-0.09547577053308487,
-0.057330790907144547,
-0.022932305932044983,
-0.05620447173714638,
-0.0030311597511172295,
0.1323290318250656,
-0.04711964353919029,
0.056722406297922134,
0.01621687412261963,
-0.005026258062571287,
-0.024119161069393158,
0.1410425305366516,
-0.06279557943344116,
-0.0850563794374466,
-0.016477666795253754,
0.106824591755867,
0.023914925754070282,
-0.007788300048559904,
-0.07550607621669769,
-0.036335840821266174,
0.10895514488220215,
0.07649951428174973,
-0.02073916420340538,
0.01591663248836994,
-0.03949436917901039,
-0.03105461411178112,
0.08615568280220032,
-0.11658910661935806,
0.049639180302619934,
-0.015586460940539837,
-0.0562201663851738,
-0.0192254688590765,
0.006009863223880529,
0.0002827181888278574,
-0.046456821262836456,
0.08826293796300888,
-0.09065598249435425,
-0.022973177954554558,
-0.06040269508957863,
-0.045028652995824814,
0.02385278232395649,
-0.04781024158000946,
-0.009698101319372654,
-0.07959365844726562,
-0.14326682686805725,
-0.04461251571774483,
0.023235570639371872,
-0.07711546868085861,
-0.06143498793244362,
-0.012597192078828812,
-0.07577870041131973,
0.025443417951464653,
0.004745827056467533,
0.12905076146125793,
-0.03689682111144066,
0.08083610981702805,
0.008762381039559841,
0.02387072704732418,
0.07965930551290512,
0.034219514578580856,
-0.09093651920557022,
0.038328126072883606,
-0.09484058618545532,
0.06256940960884094,
-0.06322493404150009,
0.025032471865415573,
-0.1302107572555542,
-0.10996772348880768,
-0.048324957489967346,
-0.0348823145031929,
0.06054561957716942,
0.1251888871192932,
-0.13614939153194427,
-0.03859692066907883,
0.1762108951807022,
-0.07213512063026428,
-0.10410016775131226,
0.10803545266389847,
-0.007207860704511404,
-0.015135657042264938,
0.047177646309137344,
0.0889018103480339,
0.09731770306825638,
-0.11718981713056564,
-0.04854617640376091,
0.006112582981586456,
0.06558031588792801,
0.016569560393691063,
0.10617469251155853,
-0.01837511919438839,
0.04648340865969658,
0.0005111062782816589,
-0.019596250727772713,
0.003922156989574432,
-0.06622368097305298,
-0.07162626087665558,
-0.04892481118440628,
-0.08844804018735886,
0.038078680634498596,
0.04188481345772743,
0.03112076409161091,
-0.08227331936359406,
-0.1579718440771103,
0.07522860914468765,
0.12309575825929642,
-0.051112137734889984,
0.014749144203960896,
-0.07106784731149673,
0.02386454865336418,
-0.06463298201560974,
-0.020330054685473442,
-0.18079368770122528,
-0.08839629590511322,
0.04745306074619293,
-0.0674419105052948,
0.01157513540238142,
-0.0024609602987766266,
0.06956920772790909,
0.0837073102593422,
-0.07246926426887512,
-0.03718999773263931,
-0.08961046487092972,
-0.002289876341819763,
-0.09335882216691971,
-0.1712733954191208,
-0.05762571468949318,
-0.022456184029579163,
0.1835571676492691,
-0.22417868673801422,
0.01948648691177368,
0.01717081107199192,
0.16913056373596191,
0.02498902566730976,
-0.06808286160230637,
0.008583755232393742,
0.03716828301548958,
-0.017486805096268654,
-0.1057296097278595,
0.019767682999372482,
0.006080586928874254,
-0.10398973524570465,
-0.03242591768503189,
-0.1660187542438507,
0.07854101061820984,
0.09465651959180832,
0.09731876105070114,
-0.10198609530925751,
-0.03839684650301933,
-0.07212553173303604,
-0.05395250767469406,
-0.06154206395149231,
-0.024898169562220573,
0.1567407250404358,
0.035818759351968765,
0.12211164087057114,
-0.06573055684566498,
-0.08014713227748871,
0.011757093481719494,
0.03258245810866356,
-0.03575469180941582,
0.0784529596567154,
0.036439817398786545,
-0.14096863567829132,
0.1006937325000763,
0.10216261446475983,
-0.03553199768066406,
0.13321329653263092,
-0.049552805721759796,
-0.10917497426271439,
-0.030562900006771088,
0.02368750050663948,
0.023887014016509056,
0.10886123776435852,
-0.08368803560733795,
0.009227137081325054,
0.038570698350667953,
-0.0020326655358076096,
0.025489743798971176,
-0.13163137435913086,
-0.006395520642399788,
0.05078866332769394,
-0.017790844663977623,
-0.0029906609561294317,
-0.03670232743024826,
0.014864963479340076,
0.07876035571098328,
0.03426307439804077,
0.016042346134781837,
0.0210084430873394,
0.0007325218175537884,
-0.07913709431886673,
0.17073123157024384,
-0.09121447801589966,
-0.15838509798049927,
-0.14379829168319702,
0.07691330462694168,
-0.05324169248342514,
-0.018466059118509293,
0.007042652927339077,
-0.09524843096733093,
-0.0515202172100544,
-0.0933525487780571,
-0.02707318216562271,
-0.06302302330732346,
0.007815445773303509,
0.047812528908252716,
0.03909672424197197,
0.08311688154935837,
-0.12125058472156525,
0.02596866525709629,
0.023609373718500137,
-0.078118696808815,
-0.014108536764979362,
0.03904222324490547,
0.10347127169370651,
0.11539358645677567,
-0.022654438391327858,
0.004563558846712112,
-0.028463587164878845,
0.17986063659191132,
-0.09618563950061798,
0.008604221977293491,
0.12321612983942032,
0.024006444960832596,
0.056108467280864716,
0.10484730452299118,
0.02826141193509102,
-0.054134469479322433,
0.02593204565346241,
0.05764872208237648,
-0.02093079686164856,
-0.24617727100849152,
-0.028662627562880516,
-0.029444130137562752,
-0.04694359377026558,
0.13112959265708923,
0.06086186319589615,
0.04626850038766861,
0.05284088850021362,
-0.04932129755616188,
0.06117258593440056,
-0.002551737241446972,
0.1057128757238388,
0.028817983344197273,
0.05642003193497658,
0.10127779841423035,
-0.03077930025756359,
-0.02979397028684616,
0.05825016275048256,
0.0038981190882623196,
0.19107282161712646,
-0.03762690722942352,
0.16564849019050598,
-0.004506897646933794,
0.11551782488822937,
-0.012273505330085754,
0.055363740772008896,
0.03401338309049606,
-0.010425603948533535,
0.03060065768659115,
-0.07886170595884323,
-0.03234470635652542,
0.05129370465874672,
0.027489079162478447,
0.05475376173853874,
-0.07102604955434799,
0.033841222524642944,
0.02984587475657463,
0.24670615792274475,
0.05220958963036537,
-0.30527785420417786,
-0.09024996310472488,
0.009744362905621529,
-0.026411086320877075,
-0.07006281614303589,
0.006510124541819096,
0.12503260374069214,
-0.1337452530860901,
0.07046639174222946,
-0.06944656372070312,
0.08525233715772629,
-0.052247241139411926,
-0.010847649537026882,
0.07215788960456848,
0.09940409660339355,
-0.012270125560462475,
0.0919867604970932,
-0.16602358222007751,
0.18020878732204437,
0.020569095388054848,
0.10709499567747116,
-0.08032608777284622,
0.04943620786070824,
-0.005317643750458956,
0.06519286334514618,
0.1086607575416565,
-0.010142323561012745,
-0.0327032096683979,
-0.15687836706638336,
-0.11399715393781662,
0.015650015324354172,
0.12457674741744995,
-0.0695471540093422,
0.08204583078622818,
-0.05040794610977173,
-0.017213789746165276,
0.03499184548854828,
-0.052964530885219574,
-0.1585507094860077,
-0.1619359850883484,
0.03052952140569687,
-0.012501604855060577,
0.011438178829848766,
-0.0928267240524292,
-0.10421887040138245,
-0.03190673142671585,
0.23336724936962128,
-0.015317393466830254,
-0.055953219532966614,
-0.15519344806671143,
0.08538571745157242,
0.14505092799663544,
-0.057001106441020966,
0.011006470769643784,
0.0030742352828383446,
0.17175747454166412,
0.019015861675143242,
-0.047533389180898666,
0.050829000771045685,
-0.05964352563023567,
-0.18708665668964386,
-0.05665476620197296,
0.16616497933864594,
0.02218552678823471,
0.05679560452699661,
-0.0014459172962233424,
0.020694810897111893,
0.01246670726686716,
-0.07999488711357117,
0.030581261962652206,
0.04274890944361687,
0.10425838083028793,
0.022470399737358093,
-0.04392681270837784,
0.044974807649850845,
-0.03047122433781624,
-0.016656745225191116,
0.1169344037771225,
0.2232845574617386,
-0.07804480940103531,
0.054276905953884125,
0.06288430094718933,
-0.045581087470054626,
-0.1445620357990265,
0.027007408440113068,
0.09930175542831421,
0.025056926533579826,
0.041968002915382385,
-0.16069579124450684,
0.0908372551202774,
0.12353907525539398,
-0.015026352368295193,
0.042032673954963684,
-0.2877580523490906,
-0.13234169781208038,
0.051662467420101166,
0.09596512466669083,
-0.059651244431734085,
-0.15853051841259003,
-0.055935971438884735,
-0.05736393481492996,
-0.12495145201683044,
0.07993895560503006,
-0.06803026050329208,
0.09811121970415115,
-0.018420564010739326,
0.026226181536912918,
0.03727247565984726,
-0.05445605888962746,
0.1518842875957489,
0.0006261341040953994,
0.03888057917356491,
-0.0496818870306015,
0.03768913075327873,
0.11643444001674652,
-0.07535967230796814,
0.04551258310675621,
-0.08733824640512466,
0.06262686848640442,
-0.13170525431632996,
-0.023293301463127136,
-0.057828716933727264,
0.05032528191804886,
-0.06493175774812698,
-0.06579194962978363,
-0.031717609614133835,
0.05449535325169563,
0.06326501071453094,
-0.041123393923044205,
0.08797913789749146,
0.026965906843543053,
0.11693592369556427,
0.12643925845623016,
0.09666029363870621,
0.020990921184420586,
-0.11886268854141235,
-0.021387793123722076,
-0.022252511233091354,
0.05827216058969498,
-0.10936970263719559,
0.015382548794150352,
0.105314701795578,
0.056271083652973175,
0.11286554485559464,
0.029336821287870407,
-0.06999189406633377,
-0.01720707304775715,
0.055248260498046875,
-0.07359907031059265,
-0.12725920975208282,
-0.03440822660923004,
0.010057087987661362,
-0.1663120985031128,
0.001870919018983841,
0.09737526625394821,
-0.06157951056957245,
-0.017351556569337845,
-0.00042497587855905294,
0.022430067881941795,
-0.01577458716928959,
0.1637226939201355,
0.03784416615962982,
0.08112452179193497,
-0.0629822313785553,
0.10695559531450272,
0.09043312817811966,
-0.09693432599306107,
0.038409583270549774,
0.06178330257534981,
-0.0796966701745987,
-0.014089119620621204,
0.08719142526388168,
0.07226002961397171,
0.0070696547627449036,
-0.04026160016655922,
-0.06707102060317993,
-0.0818549171090126,
0.04373054578900337,
0.025132915005087852,
0.024904310703277588,
-0.0012589804828166962,
-0.01851065643131733,
0.02757730707526207,
-0.16661570966243744,
0.10451068729162216,
0.03262081742286682,
0.08388286083936691,
-0.15705040097236633,
0.09699860960245132,
0.0030088499188423157,
0.0129412105306983,
-0.006812074687331915,
0.03409554436802864,
-0.07669105380773544,
-0.04400983825325966,
-0.11551617085933685,
-0.026416176930069923,
-0.053266238421201706,
-0.004297303501516581,
-0.02095695212483406,
-0.03231040760874748,
-0.03369104862213135,
0.03410949185490608,
-0.047934893518686295,
-0.0952681303024292,
-0.0010610523167997599,
0.050662729889154434,
-0.13511745631694794,
0.017021114006638527,
0.030938439071178436,
-0.1195266991853714,
0.07671575993299484,
0.044204000383615494,
0.04022042825818062,
0.0021643522195518017,
-0.058610741049051285,
0.026902059093117714,
0.019637808203697205,
0.009953251108527184,
0.04722480848431587,
-0.09234347939491272,
-0.0005673196283169091,
-0.030182363465428352,
-0.002623280044645071,
0.02152489498257637,
0.03950300067663193,
-0.12837377190589905,
-0.016722412779927254,
-0.04947705194354057,
-0.04648168757557869,
-0.06204981729388237,
0.052777208387851715,
0.10725841671228409,
-0.005512712989002466,
0.14733555912971497,
-0.06357831507921219,
0.058207351714372635,
-0.2152944654226303,
-0.02565017342567444,
-0.00512843020260334,
-0.013743996620178223,
-0.03466183692216873,
-0.011392987333238125,
0.0720672607421875,
-0.04199834167957306,
0.09770768135786057,
-0.027248747646808624,
0.10694972425699234,
0.050995636731386185,
-0.02599467523396015,
0.021234778687357903,
0.021164825186133385,
0.1939305067062378,
0.06611200422048569,
-0.02619904838502407,
0.08243619650602341,
-0.03427581489086151,
0.06412532180547714,
0.01421316433697939,
0.129488006234169,
0.15223847329616547,
-0.04318305104970932,
0.06467784196138382,
0.04854446277022362,
-0.11361870169639587,
-0.18945355713367462,
0.08508475124835968,
-0.030592018738389015,
0.10256123542785645,
-0.04402242973446846,
0.15014536678791046,
0.12798333168029785,
-0.176829531788826,
0.01948237232863903,
-0.05076552927494049,
-0.09921381622552872,
-0.07352128624916077,
-0.07388505339622498,
-0.07544921338558197,
-0.11117105185985565,
0.024404454976320267,
-0.1022953987121582,
0.01593935117125511,
0.06478706747293472,
0.015075238421559334,
0.009284146130084991,
0.16705887019634247,
-0.009112724103033543,
0.010315368883311749,
0.06355468183755875,
0.025609707459807396,
0.0009679700597189367,
-0.033849477767944336,
-0.04967917501926422,
0.03819131851196289,
0.0013035207521170378,
0.09285914152860641,
-0.04282480850815773,
-0.011692128144204617,
0.046967215836048126,
0.017468703910708427,
-0.08165156096220016,
0.035537682473659515,
0.006468028295785189,
0.02443060465157032,
0.06333309412002563,
0.030792275443673134,
0.009302855469286442,
-0.06335754692554474,
0.25968900322914124,
-0.06293684989213943,
-0.053756505250930786,
-0.13179640471935272,
0.1588304340839386,
0.03626527637243271,
-0.014704931527376175,
0.06571317464113235,
-0.1392458826303482,
-0.005729457829147577,
0.11473223567008972,
0.13370156288146973,
-0.01740122027695179,
-0.015694350004196167,
0.014679604209959507,
-0.022453203797340393,
-0.053282227367162704,
0.08827468007802963,
0.0835324376821518,
0.0144407469779253,
-0.05347622558474541,
0.022810213267803192,
0.002144873607903719,
-0.06178363785147667,
-0.06719544529914856,
0.10259256511926651,
0.012538665905594826,
0.004877666477113962,
-0.02835199050605297,
0.08558961749076843,
0.018642855808138847,
-0.18869571387767792,
0.06198061630129814,
-0.15252821147441864,
-0.1874830722808838,
-0.004548808094114065,
0.08272005617618561,
-0.031037312000989914,
0.06185625493526459,
0.025192122906446457,
-0.015101241879165173,
0.10718595236539841,
-0.003401846392080188,
-0.05334877967834473,
-0.07619963586330414,
0.0682278499007225,
-0.05684017762541771,
0.22960102558135986,
-0.008846803568303585,
0.0461728535592556,
0.1201167106628418,
0.021077316254377365,
-0.15361542999744415,
0.012401366606354713,
0.08292818069458008,
-0.07232366502285004,
0.050828102976083755,
0.18467648327350616,
-0.04028734192252159,
0.07834542542695999,
0.04320978745818138,
-0.12103943526744843,
-0.02642059326171875,
-0.07074147462844849,
0.006960305385291576,
-0.09270471334457397,
0.01748015359044075,
-0.04032227024435997,
0.17121274769306183,
0.2055787295103073,
-0.052817635238170624,
-0.010933034121990204,
-0.06026912108063698,
0.02842133305966854,
0.0543733686208725,
0.10342614352703094,
0.001326661673374474,
-0.21199668943881989,
0.012996546924114227,
0.01935437135398388,
0.05226440727710724,
-0.24785029888153076,
-0.08679370582103729,
0.061925485730171204,
-0.06412877887487411,
-0.07390634715557098,
0.10272067040205002,
0.04008609056472778,
0.019968340173363686,
-0.035422537475824356,
-0.11461256444454193,
-0.04197439178824425,
0.13459563255310059,
-0.17456746101379395,
-0.041458941996097565
] |
null | null | sentence-transformers |
# e5-base-mlqa-finetuned-arabic-for-rag
This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
<!--- Describe your model here -->
## Usage (Sentence-Transformers)
Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
```
pip install -U sentence-transformers
```
Then you can use the model like this:
```python
from sentence_transformers import SentenceTransformer
sentences = ["This is an example sentence", "Each sentence is converted"]
model = SentenceTransformer('OmarAlsaabi/e5-base-mlqa-finetuned-arabic-for-rag')
embeddings = model.encode(sentences)
print(embeddings)
```
## Evaluation Results
<!--- Describe how your model was evaluated -->
For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME})
## Training
The model was trained with the parameters:
**DataLoader**:
`torch.utils.data.dataloader.DataLoader` of length 2668 with parameters:
```
{'batch_size': 2, 'sampler': 'torch.utils.data.sampler.SequentialSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
```
**Loss**:
`sentence_transformers.losses.MultipleNegativesRankingLoss.MultipleNegativesRankingLoss` with parameters:
```
{'scale': 20.0, 'similarity_fct': 'cos_sim'}
```
Parameters of the fit()-Method:
```
{
"epochs": 2,
"evaluation_steps": 50,
"evaluator": "sentence_transformers.evaluation.InformationRetrievalEvaluator.InformationRetrievalEvaluator",
"max_grad_norm": 1,
"optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
"optimizer_params": {
"lr": 2e-05
},
"scheduler": "WarmupLinear",
"steps_per_epoch": null,
"warmup_steps": 533,
"weight_decay": 0.01
}
```
## Full Model Architecture
```
SentenceTransformer(
(0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: XLMRobertaModel
(1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False})
(2): Normalize()
)
```
## Citing & Authors
<!--- Describe where people can find more information --> | {"library_name": "sentence-transformers", "tags": ["sentence-transformers", "feature-extraction", "sentence-similarity"], "pipeline_tag": "sentence-similarity"} | sentence-similarity | OmarAlsaabi/e5-base-mlqa-finetuned-arabic-for-rag | [
"sentence-transformers",
"safetensors",
"xlm-roberta",
"feature-extraction",
"sentence-similarity",
"endpoints_compatible",
"region:us"
] | 2024-02-07T09:18:22+00:00 | [] | [] | TAGS
#sentence-transformers #safetensors #xlm-roberta #feature-extraction #sentence-similarity #endpoints_compatible #region-us
|
# e5-base-mlqa-finetuned-arabic-for-rag
This is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
## Usage (Sentence-Transformers)
Using this model becomes easy when you have sentence-transformers installed:
Then you can use the model like this:
## Evaluation Results
For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL
## Training
The model was trained with the parameters:
DataLoader:
'URL.dataloader.DataLoader' of length 2668 with parameters:
Loss:
'sentence_transformers.losses.MultipleNegativesRankingLoss.MultipleNegativesRankingLoss' with parameters:
Parameters of the fit()-Method:
## Full Model Architecture
## Citing & Authors
| [
"# e5-base-mlqa-finetuned-arabic-for-rag\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.",
"## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:",
"## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL",
"## Training\nThe model was trained with the parameters:\n\nDataLoader:\n\n'URL.dataloader.DataLoader' of length 2668 with parameters:\n\n\nLoss:\n\n'sentence_transformers.losses.MultipleNegativesRankingLoss.MultipleNegativesRankingLoss' with parameters:\n \n\nParameters of the fit()-Method:",
"## Full Model Architecture",
"## Citing & Authors"
] | [
"TAGS\n#sentence-transformers #safetensors #xlm-roberta #feature-extraction #sentence-similarity #endpoints_compatible #region-us \n",
"# e5-base-mlqa-finetuned-arabic-for-rag\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.",
"## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:",
"## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL",
"## Training\nThe model was trained with the parameters:\n\nDataLoader:\n\n'URL.dataloader.DataLoader' of length 2668 with parameters:\n\n\nLoss:\n\n'sentence_transformers.losses.MultipleNegativesRankingLoss.MultipleNegativesRankingLoss' with parameters:\n \n\nParameters of the fit()-Method:",
"## Full Model Architecture",
"## Citing & Authors"
] | [
44,
61,
38,
29,
86,
5,
6
] | [
"passage: TAGS\n#sentence-transformers #safetensors #xlm-roberta #feature-extraction #sentence-similarity #endpoints_compatible #region-us \n# e5-base-mlqa-finetuned-arabic-for-rag\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL## Training\nThe model was trained with the parameters:\n\nDataLoader:\n\n'URL.dataloader.DataLoader' of length 2668 with parameters:\n\n\nLoss:\n\n'sentence_transformers.losses.MultipleNegativesRankingLoss.MultipleNegativesRankingLoss' with parameters:\n \n\nParameters of the fit()-Method:## Full Model Architecture## Citing & Authors"
] | [
-0.06901340931653976,
0.0946681797504425,
-0.006388369482010603,
0.0654100701212883,
0.09895526617765427,
0.027909578755497932,
0.1597750335931778,
0.07846243679523468,
-0.027231397107243538,
0.08125150948762894,
0.05240620672702789,
0.10371778905391693,
0.001482932479120791,
0.05081447586417198,
-0.027749761939048767,
-0.20956064760684967,
0.04719424247741699,
-0.047402966767549515,
0.03843432664871216,
0.0738433375954628,
0.12282518297433853,
-0.06404361128807068,
0.05903681740164757,
-0.002985726809129119,
-0.0683634877204895,
0.03619344159960747,
-0.005237330216914415,
-0.056767188012599945,
0.0995144173502922,
0.07612059265375137,
0.051484595984220505,
0.036758989095687866,
0.013160533271729946,
-0.23300571739673615,
0.032194774597883224,
0.06151627376675606,
-0.027845459058880806,
0.03902124613523483,
0.02762593887746334,
-0.03497695177793503,
0.14312538504600525,
-0.10910861939191818,
0.02792368084192276,
0.03441285341978073,
-0.08769277483224869,
-0.05189525708556175,
-0.030895452946424484,
-0.013853796757757664,
0.1151466891169548,
0.07301369309425354,
-0.03814117610454559,
0.16985337436199188,
-0.05091492459177971,
0.1160266101360321,
0.14579591155052185,
-0.28569722175598145,
-0.026673194020986557,
0.073898084461689,
0.05297084152698517,
0.06592384725809097,
-0.09153710305690765,
0.03638268634676933,
0.009535126388072968,
0.0250931978225708,
0.08114710450172424,
-0.06075203791260719,
-0.005679543595761061,
-0.006665349006652832,
-0.10005892068147659,
0.015701955184340477,
0.17563404142856598,
0.043047115206718445,
-0.036002885550260544,
-0.1564924120903015,
-0.07948645949363708,
0.12108132988214493,
-0.037201762199401855,
-0.02074337564408779,
0.014125839807093143,
0.021592633798718452,
0.010440591722726822,
-0.08380192518234253,
-0.08634661883115768,
-0.049225375056266785,
-0.08198607712984085,
0.03782135993242264,
0.012956070713698864,
-0.01867608353495598,
-0.01688631996512413,
0.04026417434215546,
-0.11466427147388458,
-0.0997595563530922,
-0.036513447761535645,
-0.03825070336461067,
-0.1186317577958107,
-0.014547372236847878,
-0.04847477748990059,
-0.10418602079153061,
0.07197902351617813,
0.015175174921751022,
0.013902329839766026,
0.021525787189602852,
-0.016635959967970848,
0.0675593763589859,
0.009728280827403069,
0.09621787816286087,
-0.05092492327094078,
-0.05598654970526695,
-0.005155469290912151,
0.054737333208322525,
0.025927376002073288,
-0.0010612104088068008,
-0.04066582769155502,
-0.04453882575035095,
0.0368364192545414,
0.08098199218511581,
0.006252198945730925,
0.07166030257940292,
-0.030032150447368622,
-0.004305445589125156,
0.051401592791080475,
-0.12338408082723618,
0.0006266550626605749,
0.01952163688838482,
-0.041376594454050064,
0.028924571350216866,
0.07266073673963547,
-0.0284422654658556,
-0.1059393435716629,
-0.0015255215112119913,
-0.09638554602861404,
-0.011320863850414753,
-0.04233180731534958,
-0.16183452308177948,
0.005836850497871637,
0.004582347348332405,
-0.0331217385828495,
-0.12393294274806976,
-0.18876633048057556,
-0.058724917471408844,
0.02472819574177265,
-0.021435683593153954,
0.004030328243970871,
-0.11673356592655182,
-0.011492403224110603,
-0.013774902559816837,
-0.024561472237110138,
-0.06775100529193878,
-0.025239167734980583,
0.02764822170138359,
-0.02840648964047432,
0.08588086068630219,
0.0017292201519012451,
0.035181134939193726,
-0.08791068941354752,
0.0376271978020668,
-0.07661283761262894,
0.16307346522808075,
-0.03265703096985817,
0.042326346039772034,
-0.1262032836675644,
0.003115341067314148,
0.010370940901339054,
0.06166524067521095,
0.015282463282346725,
0.1649635136127472,
-0.19247718155384064,
-0.03959684073925018,
0.15207666158676147,
-0.06838096678256989,
-0.15392501652240753,
0.11478358507156372,
-0.04423210024833679,
0.09855061769485474,
0.14281611144542694,
0.13970158994197845,
0.10916055738925934,
-0.04938803240656853,
0.0022238013334572315,
0.08249512314796448,
-0.04016905277967453,
0.057533733546733856,
0.048685070127248764,
-0.026888538151979446,
0.08341371268033981,
-0.010614569298923016,
-0.025531787425279617,
0.020471487194299698,
-0.010279718786478043,
-0.052686337381601334,
0.012299295514822006,
-0.05142846703529358,
0.02081671915948391,
-0.04657639190554619,
0.0276092067360878,
-0.018610406666994095,
-0.07561541348695755,
0.06496062874794006,
0.08410000801086426,
-0.0863804966211319,
0.036317452788352966,
-0.04129934310913086,
-0.0026846781838685274,
-0.07709412276744843,
0.0311889685690403,
-0.1986803561449051,
-0.15692780911922455,
0.003496066201478243,
-0.002715224865823984,
0.033811721950769424,
0.03688404709100723,
0.026505954563617706,
0.0446346290409565,
-0.02499154582619667,
0.005842892453074455,
0.06136424094438553,
0.000497733592055738,
-0.11033440381288528,
-0.1001591831445694,
0.017791373655200005,
-0.0606527253985405,
0.0036619901657104492,
-0.16085298359394073,
0.0288992989808321,
-0.023577727377414703,
0.015346197411417961,
0.037845853716135025,
-0.004314542748034,
0.02192494459450245,
-0.011959780938923359,
-0.018298977985978127,
-0.047529544681310654,
0.05628107115626335,
0.011366226710379124,
-0.1546882838010788,
0.13293713331222534,
-0.2039697915315628,
-0.031119339168071747,
0.06757556647062302,
0.04119224101305008,
-0.056489698588848114,
-0.05590534210205078,
-0.011164408177137375,
-0.002926324959844351,
-0.061440665274858475,
-0.04146905615925789,
0.10050651431083679,
0.06160074844956398,
0.13083507120609283,
-0.08019117265939713,
-0.003960347734391689,
-0.03847615420818329,
-0.04688730835914612,
-0.047777220606803894,
0.11716275662183762,
-0.06441240757703781,
-0.17219170928001404,
0.06125245615839958,
0.10432705283164978,
-0.07509957998991013,
0.09942550957202911,
-0.0007200624677352607,
-0.05722911283373833,
-0.02403089962899685,
0.06193891540169716,
0.0293891578912735,
0.007795162498950958,
-0.025788437575101852,
0.009457249194383621,
0.03914020583033562,
0.03721000999212265,
0.039456356316804886,
-0.08291652053594589,
0.03729074448347092,
0.0441938079893589,
-0.039728034287691116,
0.03743395581841469,
0.04338089004158974,
0.0002580330183263868,
0.08045756071805954,
-0.012706327252089977,
-0.01404341496527195,
-0.007238238118588924,
-0.05227497220039368,
-0.11794404685497284,
0.21428392827510834,
-0.09637780487537384,
-0.16581037640571594,
-0.14453408122062683,
0.006013213191181421,
-0.07484874874353409,
0.013903569430112839,
0.06999582797288895,
-0.03366069495677948,
-0.06640773266553879,
-0.1146448403596878,
0.0629936084151268,
0.038212649524211884,
-0.03774380311369896,
-0.0018678815104067326,
0.025114743039011955,
0.004494071938097477,
-0.12466603517532349,
-0.006766969803720713,
-0.013491190969944,
-0.06195267289876938,
-0.017404155805706978,
-0.07642847299575806,
0.051000453531742096,
0.0707215964794159,
0.021151548251509666,
-0.006542805582284927,
-0.01732354611158371,
0.19270482659339905,
-0.05064386874437332,
0.031216951087117195,
0.1521034985780716,
0.019090676680207253,
0.057413049042224884,
0.10698716342449188,
0.0026009511202573776,
-0.0712272897362709,
0.06092120334506035,
0.05741192027926445,
-0.02604985423386097,
-0.17489901185035706,
-0.09646043181419373,
-0.08847299218177795,
-0.02693403698503971,
0.06717408448457718,
0.05083993077278137,
-0.046255096793174744,
0.07434782385826111,
-0.026191337034106255,
0.006353684701025486,
0.06428228318691254,
0.09691895544528961,
0.13612131774425507,
0.005165200214833021,
0.07748143374919891,
-0.06753546744585037,
-0.054819539189338684,
0.06094105914235115,
0.02681007608771324,
0.181375652551651,
-0.014427471905946732,
0.14452941715717316,
0.08760856837034225,
-0.05008430406451225,
-0.004358686972409487,
0.10552050918340683,
-0.03702593222260475,
0.01769486628472805,
-0.032076478004455566,
-0.08249778300523758,
-0.01372646912932396,
0.042659737169742584,
0.02721463330090046,
-0.005110912956297398,
-0.07131528854370117,
0.08496838063001633,
0.13044962286949158,
0.11041966080665588,
0.08135652542114258,
-0.2709393799304962,
-0.045787226408720016,
0.055590175092220306,
-0.0709272250533104,
-0.06169746071100235,
0.0062517281621694565,
0.09148144721984863,
-0.109605573117733,
0.037601910531520844,
-0.015717817470431328,
0.11303287744522095,
-0.06352590024471283,
0.026359789073467255,
-0.09868495911359787,
0.051235053688287735,
-0.03217652812600136,
0.07744204998016357,
-0.23772019147872925,
0.17318545281887054,
0.048074379563331604,
0.057558998465538025,
-0.03447354957461357,
0.014401916414499283,
0.07764887809753418,
0.08683714270591736,
0.1873277872800827,
-0.03712180256843567,
-0.0027512928936630487,
0.010987276211380959,
-0.046693239361047745,
0.043017156422138214,
0.03658023476600647,
-0.03104550577700138,
0.09354706853628159,
-0.018145650625228882,
0.0006682865205220878,
0.0006496586720459163,
0.10174917429685593,
-0.08288292586803436,
-0.17963309586048126,
-0.034643735736608505,
0.10086280107498169,
-0.05591999366879463,
-0.02681053802371025,
-0.028438612818717957,
-0.00881762895733118,
0.18388961255550385,
-0.0500631183385849,
-0.08654344826936722,
-0.1178336814045906,
0.006351971533149481,
0.08477751165628433,
-0.09197496622800827,
-0.01958058774471283,
-0.028388382866978645,
0.12884843349456787,
-0.06604398041963577,
-0.06365656107664108,
0.05269043520092964,
-0.07287034392356873,
-0.018380550667643547,
-0.02737000770866871,
0.07150061428546906,
0.02048688381910324,
0.012980721890926361,
0.05186326429247856,
0.005061409901827574,
-0.02769523300230503,
-0.10181988775730133,
-0.09428269416093826,
0.07676421105861664,
0.009054009802639484,
0.10396424680948257,
-0.13719290494918823,
-0.07672014087438583,
-0.07007770240306854,
0.03133406862616539,
0.1905442178249359,
0.2068014144897461,
-0.04303096607327461,
0.060356155037879944,
0.18315739929676056,
-0.09659735858440399,
-0.23566074669361115,
-0.07868238538503647,
0.00860140286386013,
0.06845653057098389,
0.0769660621881485,
-0.04294730722904205,
0.07844369858503342,
0.05430472269654274,
0.0029307371005415916,
-0.07508841156959534,
-0.24848602712154388,
-0.11978208273649216,
0.14557933807373047,
0.06201973930001259,
0.047877971082925797,
-0.13415102660655975,
-0.03932128846645355,
-0.1091836467385292,
-0.0228465236723423,
0.06879685819149017,
-0.10922510921955109,
0.1241854652762413,
0.028498338535428047,
0.021279657259583473,
0.0320211797952652,
-0.0031150691211223602,
0.1501767635345459,
0.09363526105880737,
0.07089376449584961,
-0.021731171756982803,
-0.0023210172075778246,
0.08094442635774612,
-0.08552771806716919,
0.14372460544109344,
-0.11105770617723465,
0.05121416971087456,
-0.10700630396604538,
-0.0481782928109169,
-0.03785854950547218,
0.024165647104382515,
-0.017445890232920647,
-0.03931253030896187,
-0.019474655389785767,
0.0372450165450573,
0.11380096524953842,
-0.0042862799018621445,
0.06602826714515686,
-0.07591433078050613,
0.07872677594423294,
0.15368850529193878,
0.11285123229026794,
0.04588997736573219,
-0.1250883787870407,
0.034504134207963943,
0.007001257501542568,
0.09273111820220947,
-0.1668521910905838,
0.07183968275785446,
0.0723264217376709,
-0.02866273745894432,
0.13269641995429993,
0.040982432663440704,
-0.05305405706167221,
0.012427347712218761,
0.051553934812545776,
-0.06993260979652405,
-0.14039810001850128,
-0.01250842772424221,
-0.06284858286380768,
-0.10043171793222427,
-0.004173075780272484,
0.1467868685722351,
-0.0350518599152565,
0.024652749300003052,
0.002189821097999811,
0.03354427218437195,
-0.03537474945187569,
0.1411592960357666,
0.019293317571282387,
0.06087573990225792,
-0.05045458301901817,
0.11454861611127853,
0.0651874989271164,
-0.09018642455339432,
0.012878271751105785,
0.11037182807922363,
-0.1033322662115097,
-0.08257385343313217,
-0.02874084748327732,
0.09891659766435623,
-0.06520357728004456,
-0.029267771169543266,
-0.07230760157108307,
-0.06515313684940338,
0.002524957526475191,
0.09117429703474045,
0.060225240886211395,
0.05249537155032158,
-0.07490696012973785,
-0.053811412304639816,
-0.08103633671998978,
0.07755523175001144,
0.07491989433765411,
0.014833364635705948,
-0.06151164323091507,
0.04441211000084877,
-0.03839901462197304,
0.058550216257572174,
-0.026656705886125565,
-0.0008622908499091864,
-0.0932796373963356,
0.01736750267446041,
-0.06802833825349808,
0.05757267400622368,
-0.12997163832187653,
-0.0008805759716778994,
0.020498255267739296,
0.05116582661867142,
-0.04302147030830383,
0.013235654681921005,
-0.03841916099190712,
-0.06118028610944748,
-0.03855477645993233,
0.11512356996536255,
-0.11755317449569702,
-0.02500164322555065,
-0.007032619323581457,
-0.07904182374477386,
0.07934089004993439,
-0.022360457107424736,
-0.04716264456510544,
0.040343448519706726,
-0.14366485178470612,
-0.03394215926527977,
0.03609804809093475,
0.03525116667151451,
0.039095792919397354,
-0.07336337864398956,
0.023249918594956398,
-0.003930018749088049,
0.039486177265644073,
-0.0001881801726995036,
0.028563082218170166,
-0.08216601610183716,
0.01341842021793127,
-0.07707767188549042,
0.007608307991176844,
-0.0744953602552414,
0.01783062145113945,
0.03631442412734032,
0.054209452122449875,
0.1374770998954773,
-0.08202756941318512,
0.047854356467723846,
-0.10726257413625717,
0.011133513413369656,
0.01979988068342209,
-0.07218018174171448,
0.047357022762298584,
-0.07562795281410217,
0.0708673968911171,
-0.053614091128110886,
0.13185961544513702,
-0.022763986140489578,
0.04338099807500839,
0.06236674636602402,
-0.00020141230197623372,
0.02629494108259678,
-0.010076602920889854,
0.08283551782369614,
0.039116110652685165,
-0.0183218102902174,
-0.025195110589265823,
0.009747098200023174,
0.047786932438611984,
0.04466969519853592,
0.07611262053251266,
0.12346458435058594,
0.011430067010223866,
0.14394980669021606,
0.04975276440382004,
-0.0068761990405619144,
0.02188394032418728,
0.001500522019341588,
-0.006333034951239824,
0.03385576233267784,
-0.015011821873486042,
-0.04287588596343994,
0.20681744813919067,
-0.10588721185922623,
0.08842243254184723,
0.02487623132765293,
-0.08340786397457123,
-0.136520653963089,
-0.08623884618282318,
-0.08953870087862015,
-0.03039601631462574,
-0.013825503177940845,
-0.13060630857944489,
-0.012943889945745468,
0.08801889419555664,
0.02533501386642456,
0.006235551554709673,
0.1558036357164383,
-0.06551147997379303,
-0.1067347377538681,
0.08553334325551987,
-0.04786534979939461,
0.05735268443822861,
0.0454692505300045,
0.03760291635990143,
0.029417486861348152,
0.011543028056621552,
0.047161251306533813,
0.0484827421605587,
0.07465057820081711,
0.06047657132148743,
-0.11071274429559708,
-0.06336798518896103,
-0.02862887643277645,
0.015528707765042782,
-0.01798444613814354,
0.10437629371881485,
0.07737739384174347,
-0.09760788083076477,
-0.0080293919891119,
0.18282002210617065,
-0.05364415794610977,
-0.08649752289056778,
-0.17589271068572998,
0.16757817566394806,
0.0504317507147789,
0.042077045887708664,
-0.0298771895468235,
-0.08423265069723129,
-0.02144913747906685,
0.13335652649402618,
0.20468448102474213,
-0.1240278035402298,
0.01844610646367073,
-0.0022840474266558886,
0.01274255383759737,
0.010459347628057003,
0.05294383317232132,
0.041777852922677994,
0.1770656704902649,
-0.03572547063231468,
0.07050251960754395,
-0.030056096613407135,
-0.021527700126171112,
-0.08326854556798935,
0.09750064462423325,
0.04940328374505043,
0.029308613389730453,
-0.022050198167562485,
0.1255568265914917,
-0.038109008222818375,
-0.06881548464298248,
-0.03728419914841652,
-0.047665074467659,
-0.10429874062538147,
-0.040999673306941986,
-0.027332447469234467,
0.013536734506487846,
0.07310914248228073,
-0.010125004686415195,
-0.017428826540708542,
0.05306989327073097,
-0.017024409025907516,
-0.08889918029308319,
-0.03049001842737198,
0.029096219688653946,
0.014128424227237701,
0.1403864026069641,
0.009405286982655525,
0.012446325272321701,
0.10832901298999786,
-0.04405035451054573,
-0.043447792530059814,
0.15097756683826447,
0.041851140558719635,
-0.027235494926571846,
0.1234789490699768,
0.04489944130182266,
-0.03205489367246628,
0.08960220962762833,
0.06433256715536118,
-0.13111922144889832,
0.04606078192591667,
0.0005681443144567311,
-0.06103179603815079,
-0.08441977202892303,
0.022127462550997734,
-0.07526221126317978,
0.11652791500091553,
0.1609952300786972,
-0.016672270372509956,
0.018627814948558807,
0.009608794003725052,
0.020813796669244766,
0.016362547874450684,
0.04197486862540245,
-0.017734240740537643,
-0.09720425307750702,
0.03149676322937012,
-0.033727191388607025,
0.00903259962797165,
-0.3344166874885559,
-0.07166048884391785,
-0.02169819548726082,
-0.0242316834628582,
-0.05149363353848457,
0.10139346867799759,
0.09955917298793793,
-0.004069928545504808,
-0.045645784586668015,
-0.2824099361896515,
0.018006425350904465,
0.0909096896648407,
-0.08971254527568817,
-0.13309389352798462
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# speecht5_finetuned_voxpopuli_lt_gg
This model is a fine-tuned version of [microsoft/speecht5_tts](https://huggingface.co/microsoft/speecht5_tts) on the facebook/voxpopuli dataset.
It achieves the following results on the evaluation set:
- Loss: 0.4952
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 4
- eval_batch_size: 2
- seed: 42
- gradient_accumulation_steps: 8
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 500
- training_steps: 1500
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:------:|:----:|:---------------:|
| 0.5286 | 51.95 | 500 | 0.5118 |
| 0.4869 | 103.9 | 1000 | 0.4986 |
| 0.481 | 155.84 | 1500 | 0.4952 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1 | {"license": "mit", "tags": ["generated_from_trainer", "text-to-speech"], "datasets": ["facebook/voxpopuli"], "base_model": "microsoft/speecht5_tts", "model-index": [{"name": "speecht5_finetuned_voxpopuli_lt_gg", "results": []}]} | text-to-speech | UnaiGurbindo/speecht5_finetuned_voxpopuli_lt | [
"transformers",
"tensorboard",
"safetensors",
"speecht5",
"text-to-audio",
"generated_from_trainer",
"text-to-speech",
"dataset:facebook/voxpopuli",
"base_model:microsoft/speecht5_tts",
"license:mit",
"endpoints_compatible",
"region:us"
] | 2024-02-07T09:25:12+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #speecht5 #text-to-audio #generated_from_trainer #text-to-speech #dataset-facebook/voxpopuli #base_model-microsoft/speecht5_tts #license-mit #endpoints_compatible #region-us
| speecht5\_finetuned\_voxpopuli\_lt\_gg
======================================
This model is a fine-tuned version of microsoft/speecht5\_tts on the facebook/voxpopuli dataset.
It achieves the following results on the evaluation set:
* Loss: 0.4952
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 1e-05
* train\_batch\_size: 4
* eval\_batch\_size: 2
* seed: 42
* gradient\_accumulation\_steps: 8
* total\_train\_batch\_size: 32
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 500
* training\_steps: 1500
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 2\n* seed: 42\n* gradient\\_accumulation\\_steps: 8\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 1500\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #speecht5 #text-to-audio #generated_from_trainer #text-to-speech #dataset-facebook/voxpopuli #base_model-microsoft/speecht5_tts #license-mit #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 2\n* seed: 42\n* gradient\\_accumulation\\_steps: 8\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 1500\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
81,
158,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #speecht5 #text-to-audio #generated_from_trainer #text-to-speech #dataset-facebook/voxpopuli #base_model-microsoft/speecht5_tts #license-mit #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 2\n* seed: 42\n* gradient\\_accumulation\\_steps: 8\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 1500\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.14155903458595276,
0.11173408478498459,
-0.003154924139380455,
0.05438064783811569,
0.09205067157745361,
0.007710474077612162,
0.11642517894506454,
0.15923161804676056,
-0.05280039831995964,
0.10994135588407516,
0.10602330416440964,
0.08414451032876968,
0.07191348820924759,
0.20538173615932465,
-0.020166730508208275,
-0.29750898480415344,
0.01497836783528328,
-0.002883479231968522,
-0.13483546674251556,
0.11721356958150864,
0.07977326214313507,
-0.10818805545568466,
0.02827444113790989,
-0.009471249766647816,
-0.11115388572216034,
-0.017853790894150734,
-0.022288719192147255,
-0.04484190046787262,
0.10729464888572693,
0.038171928375959396,
0.05497866868972778,
0.05409057438373566,
0.0784318819642067,
-0.2645151913166046,
0.016959073022007942,
0.0602426677942276,
0.021914167329669,
0.07655269652605057,
0.09746693074703217,
-0.025409914553165436,
0.06551870703697205,
-0.08565590530633926,
0.06452671438455582,
0.045288827270269394,
-0.11948342621326447,
-0.30325594544410706,
-0.09473377466201782,
0.054244156926870346,
0.15183718502521515,
0.07070737332105637,
-0.039857495576143265,
0.04890262335538864,
-0.04443875327706337,
0.08996746689081192,
0.25065353512763977,
-0.2586999833583832,
-0.0749128982424736,
-0.005764756351709366,
0.08094161003828049,
0.06912786513566971,
-0.11693284660577774,
-0.010310908779501915,
0.029842359945178032,
0.015183363109827042,
0.12594078481197357,
-0.0023979214020073414,
0.09113045781850815,
-0.02071601338684559,
-0.13731049001216888,
-0.04626114293932915,
0.11950112879276276,
0.09204880148172379,
-0.028302954509854317,
-0.1309202015399933,
-0.015914537012577057,
-0.20912200212478638,
-0.05322158336639404,
-0.0010489667765796185,
0.01802527718245983,
-0.0354413203895092,
-0.09855133295059204,
-0.004217050038278103,
-0.05747926980257034,
-0.0858709067106247,
0.062317293137311935,
0.12034711241722107,
0.03224874287843704,
-0.043289318680763245,
0.018861429765820503,
0.10422109812498093,
0.02394568733870983,
-0.16252006590366364,
0.011280528269708157,
0.028098298236727715,
-0.10860057920217514,
-0.03087376244366169,
-0.006204220931977034,
-0.03750956431031227,
0.013319346122443676,
0.16748790442943573,
-0.03401746228337288,
0.09458644688129425,
0.0302626620978117,
0.034731343388557434,
-0.07094219326972961,
0.12127537280321121,
-0.05069856718182564,
-0.09795582294464111,
-0.04801194369792938,
0.1331070065498352,
0.0006634819437749684,
-0.022272272035479546,
-0.08258578181266785,
0.035730596631765366,
0.07722080498933792,
0.04981021210551262,
0.005161152686923742,
0.008273023180663586,
-0.10160847008228302,
-0.028548866510391235,
0.019546542316675186,
-0.092342808842659,
0.057993579655885696,
0.017295731231570244,
-0.027781911194324493,
-0.06366517394781113,
-0.012840562500059605,
0.025847550481557846,
-0.005580930504947901,
0.13178706169128418,
-0.04890105500817299,
-0.01630464382469654,
-0.0798240602016449,
-0.09761783480644226,
0.031216006726026535,
-0.08115537464618683,
0.0003921102324966341,
-0.03897886350750923,
-0.09323649853467941,
-0.06844718754291534,
0.07980628311634064,
-0.05172187462449074,
-0.07538217306137085,
-0.07393969595432281,
-0.057577285915613174,
0.05193086713552475,
-0.025199340656399727,
0.14674842357635498,
-0.06404778361320496,
0.10891055315732956,
-0.004960821941494942,
0.08339999616146088,
0.06782373785972595,
0.07257430255413055,
-0.022740894928574562,
0.05994625762104988,
-0.231930211186409,
0.09108622372150421,
-0.10349222272634506,
0.06516849249601364,
-0.14163732528686523,
-0.09314943850040436,
-0.02910180762410164,
0.010806750506162643,
0.09661968052387238,
0.12236732244491577,
-0.20594508945941925,
-0.10568594187498093,
0.1836269497871399,
-0.07652296125888824,
-0.09893354028463364,
0.14098192751407623,
-0.02176804468035698,
0.010610039345920086,
0.02203027904033661,
0.21484006941318512,
0.1282573789358139,
-0.11001607030630112,
0.02232174761593342,
-0.05348609760403633,
0.09672117978334427,
0.029146205633878708,
0.08183605968952179,
-0.05762787163257599,
0.06005086749792099,
-0.01711704395711422,
-0.00823312159627676,
0.060459014028310776,
-0.07317835092544556,
-0.07213416695594788,
-0.0013437883462756872,
-0.07738279551267624,
0.03663335740566254,
0.04542246088385582,
0.010026243515312672,
-0.1003744900226593,
-0.12120866775512695,
0.01629476062953472,
0.0848517045378685,
-0.08693862706422806,
0.029048630967736244,
-0.061783090233802795,
0.064989373087883,
-0.013520208187401295,
-0.002724012127146125,
-0.1584644317626953,
0.030595339834690094,
0.033307213336229324,
-0.028726207092404366,
0.0096889054402709,
-0.02850422076880932,
0.07574379444122314,
0.03746434673666954,
-0.07248233258724213,
-0.07866133749485016,
-0.04271344467997551,
0.001129230367951095,
-0.0834820419549942,
-0.24321022629737854,
-0.05706179141998291,
-0.04325385019183159,
0.13001997768878937,
-0.1881498396396637,
0.023586472496390343,
0.07055142521858215,
0.12721683084964752,
0.06961622089147568,
-0.061482012271881104,
0.030866829678416252,
0.07766173034906387,
-0.00005841898382641375,
-0.08287854492664337,
0.03997201472520828,
0.011210828088223934,
-0.12726186215877533,
-0.0064161596819758415,
-0.12573997676372528,
0.11259517073631287,
0.08847779035568237,
0.050737664103507996,
-0.10388562083244324,
-0.0936175212264061,
-0.05796504393219948,
-0.043731022626161575,
-0.03475459665060043,
0.009045064449310303,
0.16035149991512299,
0.042740460485219955,
0.09235768020153046,
-0.08178036659955978,
-0.05079209804534912,
0.046608999371528625,
-0.0020127263851463795,
-0.01381505373865366,
0.14465130865573883,
0.028118275105953217,
-0.06561131030321121,
0.11490961164236069,
0.12076190114021301,
-0.013046185486018658,
0.1750234067440033,
-0.07861834019422531,
-0.09091036021709442,
-0.039112638682127,
0.0484170988202095,
0.0425436832010746,
0.1321457028388977,
-0.13790622353553772,
-0.0078717777505517,
0.011005455628037453,
0.029955178499221802,
-0.004869834519922733,
-0.18517780303955078,
-0.017957016825675964,
0.04750141128897667,
-0.06949277222156525,
-0.021001648157835007,
-0.021450912579894066,
-0.018656214699149132,
0.07466544955968857,
0.014978023245930672,
-0.02645088918507099,
-0.0018314843764528632,
-0.020541297271847725,
-0.0860324278473854,
0.1712833046913147,
-0.1158025860786438,
-0.16634905338287354,
-0.12304238975048065,
-0.01653607189655304,
0.014129307121038437,
-0.009424409829080105,
0.06194911152124405,
-0.10052988678216934,
-0.02347375638782978,
-0.05885559320449829,
0.027901655063033104,
-0.013340691104531288,
0.029055893421173096,
0.014485450461506844,
0.023613320663571358,
0.07130580395460129,
-0.0743638351559639,
0.02207929641008377,
-0.007797107566148043,
0.011048627085983753,
0.03060702048242092,
0.013088605366647243,
0.09374485164880753,
0.14153088629245758,
0.04770645126700401,
0.034176599234342575,
-0.05363491177558899,
0.18326053023338318,
-0.1535242795944214,
0.02603786438703537,
0.10867061465978622,
-0.013931436464190483,
0.057097915560007095,
0.1763305813074112,
0.04623030126094818,
-0.10277998447418213,
0.02675677463412285,
0.03328777104616165,
-0.020459208637475967,
-0.21281081438064575,
-0.008212719112634659,
-0.07184117287397385,
0.007294722367078066,
0.1166997104883194,
0.028721434995532036,
-0.02150685526430607,
0.035564009100198746,
-0.015294339507818222,
-0.024167703464627266,
0.03709053993225098,
0.05974583700299263,
0.02545188181102276,
0.025258751586079597,
0.11037620902061462,
-0.015578326769173145,
-0.012508680112659931,
0.037918996065855026,
0.026407137513160706,
0.24786020815372467,
0.00893721729516983,
0.1802350878715515,
0.05334709212183952,
0.14898216724395752,
0.021945301443338394,
0.053871944546699524,
0.005778031889349222,
-0.032446812838315964,
0.0065048797987401485,
-0.04894868656992912,
-0.018441611900925636,
0.04084925726056099,
0.0721103623509407,
0.013190803118050098,
-0.11890548467636108,
0.02112579718232155,
0.028168385848402977,
0.3510070741176605,
0.08557193726301193,
-0.2579127848148346,
-0.08661675453186035,
0.017826346680521965,
-0.06354820728302002,
-0.04775813966989517,
0.025854211300611496,
0.15524373948574066,
-0.07500612735748291,
0.08686225116252899,
-0.07578139752149582,
0.08894267678260803,
-0.08561339974403381,
0.007781505584716797,
0.073882557451725,
0.07282548397779465,
-0.03330979123711586,
0.03649613633751869,
-0.2750163674354553,
0.3058589994907379,
0.0099369240924716,
0.08003176003694534,
-0.03250843659043312,
0.035587284713983536,
0.021627582609653473,
-0.011910652741789818,
0.1237325444817543,
-0.002024678513407707,
-0.1489536166191101,
-0.16956889629364014,
-0.10220514982938766,
-0.010845374315977097,
0.12998542189598083,
-0.07668455690145493,
0.09162458032369614,
-0.015342885628342628,
-0.03162180632352829,
0.03447675704956055,
-0.08179368823766708,
-0.08776234835386276,
-0.12387624382972717,
0.018519988283514977,
-0.008005277253687382,
0.07877452671527863,
-0.1053691953420639,
-0.09348830580711365,
-0.038371939212083817,
0.14930424094200134,
-0.08564115315675735,
-0.023019829764962196,
-0.15493519604206085,
0.06339159607887268,
0.13467851281166077,
-0.0652657002210617,
0.06392998993396759,
0.022367315366864204,
0.1277087926864624,
-0.020137332379817963,
0.0026430338621139526,
0.13960565626621246,
-0.07729420065879822,
-0.208314910531044,
-0.07609151303768158,
0.18415023386478424,
0.038955215364694595,
0.06761927902698517,
-0.0266131479293108,
0.033859167248010635,
0.015456988476216793,
-0.07275552302598953,
0.0932631716132164,
-0.011993453837931156,
0.01868458464741707,
0.007651841267943382,
-0.024449516087770462,
-0.0065320078283548355,
-0.054139286279678345,
-0.07755346596240997,
0.12395311146974564,
0.3112107515335083,
-0.10349754989147186,
0.07191421836614609,
0.06196097657084465,
-0.04277632758021355,
-0.1754385083913803,
0.03704511746764183,
0.13298675417900085,
0.04692181199789047,
0.030464086681604385,
-0.21269497275352478,
0.0041503035463392735,
0.07712258398532867,
-0.024690726771950722,
0.07340046763420105,
-0.3090954124927521,
-0.13475766777992249,
0.06672607362270355,
0.09006298333406448,
-0.04144526645541191,
-0.14537644386291504,
-0.07040005922317505,
-0.0026736620347946882,
-0.0830678641796112,
0.03008255921304226,
-0.01555339340120554,
0.12230850011110306,
0.01131595391780138,
0.005483689252287149,
0.025846419855952263,
-0.04917815700173378,
0.11652018874883652,
-0.008822579868137836,
0.07490281760692596,
-0.014901609160006046,
0.04487178847193718,
-0.019609158858656883,
-0.07795479893684387,
-0.007447793148458004,
-0.09952205419540405,
0.010160784237086773,
-0.11133541166782379,
-0.025074666365981102,
-0.07136283814907074,
0.01307105552405119,
-0.05442682281136513,
-0.04305649548768997,
-0.020948516204953194,
0.06495664268732071,
0.05649792402982712,
-0.013328458182513714,
0.12901413440704346,
-0.061214201152324677,
0.15221558511257172,
0.10961458086967468,
0.11346456408500671,
-0.04613909125328064,
-0.1081869900226593,
0.008233827538788319,
-0.032188620418310165,
0.041940826922655106,
-0.13854773342609406,
0.03410040959715843,
0.13722926378250122,
0.04337026923894882,
0.15426738560199738,
0.04419998079538345,
-0.07592707127332687,
0.014043348841369152,
0.0830317810177803,
-0.0908876359462738,
-0.17452257871627808,
-0.025720948353409767,
0.019603131338953972,
-0.14225661754608154,
-0.006857036612927914,
0.10115204751491547,
-0.014840264804661274,
-0.011595789343118668,
0.013911768794059753,
0.03253393620252609,
-0.026463128626346588,
0.21047165989875793,
0.018497254699468613,
0.09205006062984467,
-0.09337524324655533,
0.08476246148347855,
0.04070781543850899,
-0.16788533329963684,
0.041524600237607956,
0.09659863263368607,
-0.07411262392997742,
-0.009462668560445309,
0.06207653507590294,
0.09863512963056564,
0.06373338401317596,
-0.021368492394685745,
-0.11261589080095291,
-0.14463099837303162,
0.08024981617927551,
0.10978643596172333,
0.020609451457858086,
0.00015343050472438335,
-0.011979806236922741,
0.03550603985786438,
-0.09027164429426193,
0.12114404141902924,
0.09304486215114594,
0.06981610506772995,
-0.13339675962924957,
0.10778961330652237,
-0.0009823838481679559,
-0.020410774275660515,
-0.003773811971768737,
0.01260310597717762,
-0.13389988243579865,
0.011720577254891396,
-0.06408101320266724,
-0.04589818790555,
-0.06441601365804672,
-0.009145256131887436,
-0.011154784820973873,
-0.0450042262673378,
-0.028175542131066322,
0.009751052595674992,
-0.11505268514156342,
-0.04770394042134285,
-0.015662632882595062,
0.05848760902881622,
-0.09323418885469437,
-0.023638388141989708,
0.040309395641088486,
-0.11070458590984344,
0.09746107459068298,
0.007826515473425388,
0.027792343869805336,
-0.01383398286998272,
-0.0828377828001976,
0.00872552115470171,
0.03205959498882294,
-0.01884128898382187,
0.03387388214468956,
-0.17849400639533997,
-0.0207582488656044,
-0.045535579323768616,
0.0017343248473480344,
0.0036722219083458185,
0.01587730087339878,
-0.12434905767440796,
0.018237223848700523,
-0.039402808994054794,
-0.06268814206123352,
-0.05607438087463379,
0.0690503865480423,
0.09967551380395889,
0.007315809838473797,
0.1463058739900589,
-0.08384907245635986,
0.05183001607656479,
-0.22255490720272064,
-0.00015820658882148564,
-0.002656147815287113,
-0.08377592265605927,
-0.06727839261293411,
-0.035995714366436005,
0.08961698412895203,
-0.054854609072208405,
0.04805533587932587,
-0.045690037310123444,
0.033533740788698196,
0.035152681171894073,
-0.1308789998292923,
0.02428193762898445,
0.05152323842048645,
0.150938019156456,
0.019629601389169693,
-0.03969940170645714,
0.0581914559006691,
0.022842157632112503,
0.0522029846906662,
0.14510953426361084,
0.15501143038272858,
0.14454635977745056,
0.061623334884643555,
0.06361377984285355,
0.0434856079518795,
-0.10198432207107544,
-0.1659361720085144,
0.14671091735363007,
-0.04022965580224991,
0.1375988870859146,
-0.021223751828074455,
0.18340227007865906,
0.09714151918888092,
-0.21410726010799408,
0.06824370473623276,
-0.02590213343501091,
-0.09027335047721863,
-0.09421011060476303,
-0.09230811148881912,
-0.07380703091621399,
-0.19392190873622894,
0.006986051797866821,
-0.10611086338758469,
0.045561980456113815,
0.012718654237687588,
0.03314921632409096,
0.048945557326078415,
0.13666197657585144,
0.04058287292718887,
0.01818055845797062,
0.12747478485107422,
0.02333524264395237,
-0.018620770424604416,
-0.04908200725913048,
-0.0944075658917427,
0.045845143496990204,
-0.050353728234767914,
0.04113991931080818,
-0.04095758870244026,
-0.10985473543405533,
0.05682436004281044,
0.031966593116521835,
-0.10843812674283981,
0.02385248802602291,
-0.011067323386669159,
0.06015227362513542,
0.10936322808265686,
0.02950308471918106,
-0.010324503295123577,
-0.01981787383556366,
0.23377205431461334,
-0.09415382146835327,
-0.03377128764986992,
-0.13710302114486694,
0.19872379302978516,
-0.015125060454010963,
0.0034673367626965046,
0.004575522616505623,
-0.07673730701208115,
0.012733248062431812,
0.15450003743171692,
0.13599993288516998,
-0.008049414493143559,
-0.010702000930905342,
0.01676671765744686,
-0.007721045985817909,
-0.03278272598981857,
0.07323472201824188,
0.11023323237895966,
0.03686637431383133,
-0.05413297936320305,
-0.015995683148503304,
-0.029522141441702843,
-0.07628601044416428,
-0.024422258138656616,
0.07045432925224304,
0.043381836265325546,
0.007936503738164902,
-0.025306090712547302,
0.13673049211502075,
-0.04225797578692436,
-0.1290258914232254,
0.03822239488363266,
-0.20387785136699677,
-0.17606285214424133,
-0.040415700525045395,
0.06714161485433578,
0.02023605816066265,
0.04910862818360329,
0.018625570461153984,
-0.011661258526146412,
0.10228097438812256,
-0.004228969104588032,
-0.019870273768901825,
-0.12006840854883194,
0.06441951543092728,
-0.12603716552257538,
0.15988506376743317,
-0.048097021877765656,
-0.022367404773831367,
0.11958999186754227,
0.05058455467224121,
-0.08200595527887344,
0.02735501155257225,
0.06958483904600143,
-0.10068445652723312,
0.0219524297863245,
0.1846611201763153,
-0.039897821843624115,
0.14598041772842407,
0.05113658681511879,
-0.104805588722229,
0.030896533280611038,
-0.1373448222875595,
-0.05521602928638458,
-0.053945090621709824,
0.019514013081789017,
-0.036005448549985886,
0.13985489308834076,
0.20086203515529633,
-0.060034122318029404,
-0.010290211997926235,
-0.04173946753144264,
0.014966343529522419,
0.0635836198925972,
0.14503799378871918,
-0.026139291003346443,
-0.2657175064086914,
0.024206431582570076,
0.05903831496834755,
0.0019301630090922117,
-0.2281423956155777,
-0.10982964187860489,
0.009355985559523106,
-0.03951611369848251,
-0.07929200679063797,
0.11823548376560211,
0.06627482175827026,
0.056940898299217224,
-0.05547215789556503,
-0.12161098420619965,
-0.029075276106595993,
0.18226739764213562,
-0.1561981439590454,
-0.036802057176828384
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | EnDevSols/tinyllama-3T-64k-JSONExtractor | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T09:26:58+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
56,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06061961501836777,
0.15481999516487122,
-0.004844071343541145,
0.02074851468205452,
0.0983177199959755,
0.007407687604427338,
0.07119518518447876,
0.11185134947299957,
-0.023851769044995308,
0.1167980208992958,
0.031993988901376724,
0.09781743586063385,
0.11217817664146423,
0.16186554729938507,
0.0015333457849919796,
-0.22897611558437347,
0.049678247421979904,
-0.125278040766716,
-0.0294334813952446,
0.11977242678403854,
0.1422213912010193,
-0.10954539477825165,
0.0752737894654274,
-0.038042325526475906,
-0.005828251596540213,
-0.0323176346719265,
-0.06205610930919647,
-0.05266609415411949,
0.05311284959316254,
0.06794639676809311,
0.07308239489793777,
0.01171939354389906,
0.09106900542974472,
-0.2724283039569855,
0.02348201349377632,
0.0805930644273758,
-0.0006441773730330169,
0.07586129754781723,
0.04993962123990059,
-0.08749990910291672,
0.07524524629116058,
-0.060156844556331635,
0.1498761922121048,
0.07955671846866608,
-0.09018243104219437,
-0.19217631220817566,
-0.07921334356069565,
0.09916994720697403,
0.1890910118818283,
0.05953684076666832,
-0.026427440345287323,
0.11642678081989288,
-0.08593545109033585,
0.013638701289892197,
0.06446459144353867,
-0.06054406240582466,
-0.055855002254247665,
0.06904532760381699,
0.08335285633802414,
0.08567540347576141,
-0.12976622581481934,
-0.010767064057290554,
0.015032444149255753,
0.008952446281909943,
0.08948688954114914,
0.017146794125437737,
0.1335189938545227,
0.040557652711868286,
-0.13501930236816406,
-0.043155476450920105,
0.09761431813240051,
0.03665134683251381,
-0.04888195917010307,
-0.2485782504081726,
-0.023432478308677673,
-0.04339504987001419,
-0.03198111802339554,
-0.03649339824914932,
0.043764639645814896,
-0.014506848528981209,
0.07738617807626724,
-0.004502781666815281,
-0.0837155357003212,
-0.04301247000694275,
0.07241875678300858,
0.06128999963402748,
0.02571401372551918,
-0.015821760520339012,
0.0059297760017216206,
0.12327717989683151,
0.11431120336055756,
-0.126715749502182,
-0.052547648549079895,
-0.06306339055299759,
-0.08449548482894897,
-0.044861067086458206,
0.030838407576084137,
0.037995077669620514,
0.045936476439237595,
0.23867325484752655,
0.007765117567032576,
0.053257301449775696,
0.04455438256263733,
0.014407169073820114,
0.06501194834709167,
0.11008983850479126,
-0.05894824117422104,
-0.09719445556402206,
-0.028582042083144188,
0.10156717151403427,
0.007986726239323616,
-0.04139331728219986,
-0.05712985619902611,
0.07059531658887863,
0.018587570637464523,
0.12360043078660965,
0.08000938594341278,
0.003056557849049568,
-0.0755772516131401,
-0.062465377151966095,
0.17764076590538025,
-0.15825673937797546,
0.04532013460993767,
0.03055616281926632,
-0.0341108962893486,
-0.009745313785970211,
0.012105142697691917,
0.025474950671195984,
-0.021481726318597794,
0.09522198140621185,
-0.05601342022418976,
-0.034448131918907166,
-0.11389608681201935,
-0.03694311901926994,
0.030394554138183594,
0.011153047904372215,
-0.02865210548043251,
-0.03502652049064636,
-0.08865131437778473,
-0.06405586749315262,
0.09101516753435135,
-0.07148737460374832,
-0.04784895107150078,
-0.016645915806293488,
-0.07833752781152725,
0.021804187446832657,
0.01691517047584057,
0.09064167737960815,
-0.0222476739436388,
0.03985358029603958,
-0.0550384595990181,
0.061440225690603256,
0.11723454296588898,
0.027987057343125343,
-0.05787884071469307,
0.061519939452409744,
-0.2424532175064087,
0.10252492874860764,
-0.07715212553739548,
0.04971238598227501,
-0.15203025937080383,
-0.02478341944515705,
0.03986154496669769,
0.01284773275256157,
-0.008251311257481575,
0.14196595549583435,
-0.21994100511074066,
-0.030957341194152832,
0.16964265704154968,
-0.10025953501462936,
-0.08109250664710999,
0.060782887041568756,
-0.05354252830147743,
0.11210215091705322,
0.04557164013385773,
-0.02375967986881733,
0.05775221437215805,
-0.14725260436534882,
-0.011030761525034904,
-0.041942402720451355,
-0.0180682260543108,
0.16207332909107208,
0.0703711211681366,
-0.06047816202044487,
0.07456906884908676,
0.01960151270031929,
-0.014246034435927868,
-0.04887177795171738,
-0.02822130173444748,
-0.1047162413597107,
0.01184528972953558,
-0.06102835759520531,
0.018109694123268127,
-0.021768750622868538,
-0.09445013850927353,
-0.029118487611413002,
-0.17402999103069305,
-0.0031633328180760145,
0.08821269869804382,
-0.011630427092313766,
-0.021509924903512,
-0.11245372891426086,
0.009332616813480854,
0.030967719852924347,
0.0002618339203763753,
-0.13677829504013062,
-0.06033218279480934,
0.026970699429512024,
-0.16097871959209442,
0.029791243374347687,
-0.05741601809859276,
0.04530094936490059,
0.04005871340632439,
-0.03433511033654213,
-0.03489551320672035,
0.010874404571950436,
0.010431389324367046,
-0.01894843392074108,
-0.25422003865242004,
-0.01882786676287651,
-0.0234990194439888,
0.1751047968864441,
-0.22956320643424988,
0.042598169296979904,
0.07489731162786484,
0.1460893303155899,
0.007349682506173849,
-0.03550100699067116,
0.015185600146651268,
-0.07262228429317474,
-0.03268764168024063,
-0.06316669285297394,
-0.01207790058106184,
-0.038400664925575256,
-0.05820201337337494,
0.04906858503818512,
-0.1686294972896576,
-0.030321966856718063,
0.10717973858118057,
0.06342670321464539,
-0.1473218947649002,
-0.02780107781291008,
-0.04056945815682411,
-0.04624456167221069,
-0.06676914542913437,
-0.05461418256163597,
0.11812574416399002,
0.056411582976579666,
0.04860803112387657,
-0.07140495628118515,
-0.07455260306596756,
0.008036690764129162,
-0.01956399530172348,
-0.014917809516191483,
0.09334591031074524,
0.07554110884666443,
-0.12264352291822433,
0.09177418053150177,
0.09668384492397308,
0.08576478064060211,
0.10314212739467621,
-0.014663571491837502,
-0.08914592862129211,
-0.040637146681547165,
0.02245822176337242,
0.016187267377972603,
0.15129362046718597,
-0.012961224652826786,
0.055492039769887924,
0.0358695350587368,
-0.014034898020327091,
0.011105312965810299,
-0.09736533463001251,
0.02655916102230549,
0.030835967510938644,
-0.016302183270454407,
0.03745110332965851,
-0.0447014644742012,
0.019208140671253204,
0.09039704501628876,
0.040895868092775345,
0.040978945791721344,
0.010155045427381992,
-0.04354988783597946,
-0.11037563532590866,
0.1787576973438263,
-0.12389461696147919,
-0.24818050861358643,
-0.13812170922756195,
0.010281167924404144,
0.04737642779946327,
-0.010411068797111511,
0.006690691225230694,
-0.06616118550300598,
-0.1175973042845726,
-0.09878289699554443,
0.018617089837789536,
0.045352302491664886,
-0.07590975612401962,
-0.06842505931854248,
0.06414616107940674,
0.03875524550676346,
-0.13939815759658813,
0.024007495492696762,
0.04662325978279114,
-0.08205481618642807,
-0.0029386086389422417,
0.0791812464594841,
0.06965780258178711,
0.17661017179489136,
0.013885351829230785,
-0.023669935762882233,
0.026634456589818,
0.20819635689258575,
-0.1436755359172821,
0.10975687950849533,
0.13545554876327515,
-0.08767466992139816,
0.08120133727788925,
0.1998777538537979,
0.03777998685836792,
-0.10680917650461197,
0.03608465939760208,
0.028374753892421722,
-0.028325283899903297,
-0.2502254545688629,
-0.06958996504545212,
0.0019060121849179268,
-0.05172049254179001,
0.07064855098724365,
0.08791537582874298,
0.09593888372182846,
0.016860228031873703,
-0.09976044297218323,
-0.07697858661413193,
0.046900223940610886,
0.10824491083621979,
-0.00015424020239152014,
-0.015208319760859013,
0.0904119610786438,
-0.03033481352031231,
0.01743943803012371,
0.09215071052312851,
0.0030607767403125763,
0.17535938322544098,
0.051709048449993134,
0.17189906537532806,
0.07866133749485016,
0.06444311141967773,
0.02004685252904892,
0.007725914940237999,
0.021817529574036598,
0.017227526754140854,
-0.0030957073904573917,
-0.08709781616926193,
-0.0034981227945536375,
0.1202581599354744,
0.049845851957798004,
0.029173865914344788,
0.012042860500514507,
-0.030704669654369354,
0.08337877690792084,
0.1770893782377243,
0.0029054484330117702,
-0.1893385946750641,
-0.07169844210147858,
0.07795937359333038,
-0.08648337423801422,
-0.10729733109474182,
-0.029470939189195633,
0.041069481521844864,
-0.1729043871164322,
0.016882894560694695,
-0.019335895776748657,
0.10788324475288391,
-0.13190391659736633,
-0.01772487722337246,
0.05657728388905525,
0.06932812184095383,
-0.009677323512732983,
0.06694949418306351,
-0.16090403497219086,
0.11770165711641312,
0.01751571334898472,
0.06636732816696167,
-0.09608277678489685,
0.09618937969207764,
-0.007830657996237278,
0.0041499207727611065,
0.1410749852657318,
0.010120149701833725,
-0.05952107161283493,
-0.09608154743909836,
-0.10546442121267319,
-0.009841260500252247,
0.1306990385055542,
-0.14852415025234222,
0.08813067525625229,
-0.02661319263279438,
-0.044553373008966446,
0.003614129964262247,
-0.12497276812791824,
-0.13103094696998596,
-0.18366187810897827,
0.05707118660211563,
-0.12947207689285278,
0.04045100137591362,
-0.10902881622314453,
-0.045833900570869446,
-0.02098964899778366,
0.20040063560009003,
-0.23137451708316803,
-0.06714103370904922,
-0.1551055610179901,
-0.08061286807060242,
0.14446212351322174,
-0.046455029398202896,
0.08550118654966354,
0.0008278203313238919,
0.19068008661270142,
0.021319707855582237,
-0.017237508669495583,
0.1072206199169159,
-0.10052918642759323,
-0.2010865956544876,
-0.09273224323987961,
0.15895552933216095,
0.13766798377037048,
0.03809428587555885,
-0.004381525795906782,
0.03171157464385033,
-0.02098114788532257,
-0.12076930701732635,
0.020226983353495598,
0.17317426204681396,
0.08982043713331223,
0.025265544652938843,
-0.02972041629254818,
-0.11267432570457458,
-0.07061342149972916,
-0.03774050623178482,
0.024755435064435005,
0.18072067201137543,
-0.07222156971693039,
0.18405316770076752,
0.13775517046451569,
-0.05534014105796814,
-0.19904261827468872,
0.021996473893523216,
0.04293542355298996,
0.0070380112156271935,
0.0323902890086174,
-0.20307663083076477,
0.09384101629257202,
0.0008334947633557022,
-0.05131231248378754,
0.1379684954881668,
-0.1823476254940033,
-0.151598259806633,
0.06042521819472313,
0.043563615530729294,
-0.19374065101146698,
-0.12374074012041092,
-0.08848230540752411,
-0.04693066328763962,
-0.15487661957740784,
0.10312657803297043,
0.0020827590487897396,
0.008401188999414444,
0.03778626397252083,
0.02252252586185932,
0.012139533646404743,
-0.04198719933629036,
0.1914343535900116,
-0.025891713798046112,
0.03347287327051163,
-0.0790715217590332,
-0.060851071029901505,
0.062408581376075745,
-0.058187782764434814,
0.0755455270409584,
-0.025226406753063202,
0.015947066247463226,
-0.10598332434892654,
-0.048235729336738586,
-0.02852320298552513,
0.019321219995617867,
-0.09431382268667221,
-0.09348297864198685,
-0.04829427972435951,
0.09367614984512329,
0.09042316675186157,
-0.03652578964829445,
-0.03649144619703293,
-0.078715980052948,
0.038977332413196564,
0.17627815902233124,
0.18159319460391998,
0.04659178853034973,
-0.07959239184856415,
-0.001915142871439457,
-0.014336181804537773,
0.04684065282344818,
-0.22077152132987976,
0.060553863644599915,
0.04557652771472931,
0.016117896884679794,
0.11537692695856094,
-0.0208132341504097,
-0.16198977828025818,
-0.06710557639598846,
0.061360616236925125,
-0.06944561004638672,
-0.17825035750865936,
0.0039279889315366745,
0.07344977557659149,
-0.16578389704227448,
-0.037031736224889755,
0.04200848564505577,
-0.01189455483108759,
-0.0403641052544117,
0.012352054007351398,
0.08063354343175888,
0.007078902795910835,
0.07699975371360779,
0.055281639099121094,
0.09124495089054108,
-0.10227900743484497,
0.07410510629415512,
0.08149529248476028,
-0.08644098788499832,
0.030720343813300133,
0.09573426842689514,
-0.06469762325286865,
-0.0346054881811142,
0.04237886518239975,
0.08354541659355164,
0.024281201884150505,
-0.04682289808988571,
0.0023111123591661453,
-0.09734189510345459,
0.05927345156669617,
0.11483542621135712,
0.03496333956718445,
0.011234734207391739,
0.03813567012548447,
0.04486291855573654,
-0.08093374222517014,
0.11926916986703873,
0.023795632645487785,
0.020354853942990303,
-0.04112942889332771,
-0.040553025901317596,
0.035851649940013885,
-0.026020776480436325,
-0.011440055444836617,
-0.035174157470464706,
-0.0722682997584343,
-0.014069457538425922,
-0.16000694036483765,
-0.0076758842915296555,
-0.03660871088504791,
0.005114538595080376,
0.022510098293423653,
-0.03652830421924591,
0.00792311318218708,
0.012217256240546703,
-0.06868947297334671,
-0.05553458258509636,
-0.023233558982610703,
0.09422210603952408,
-0.16494666039943695,
0.0220257006585598,
0.0823851153254509,
-0.12121747434139252,
0.09289738535881042,
0.016782134771347046,
0.00412249518558383,
0.026962365955114365,
-0.1545863002538681,
0.04763968288898468,
-0.020152103155851364,
0.013473534025251865,
0.04222847521305084,
-0.21637047827243805,
-0.004404853098094463,
-0.04015503451228142,
-0.05566934496164322,
-0.008993052877485752,
-0.0319182425737381,
-0.11338426172733307,
0.09645436704158783,
0.011025024577975273,
-0.08443772792816162,
-0.02965564839541912,
0.03353232145309448,
0.07690354436635971,
-0.027447547763586044,
0.1498211771249771,
-0.004663881380110979,
0.07559948414564133,
-0.17581342160701752,
-0.02282017655670643,
-0.011197620071470737,
0.022367527708411217,
-0.021871577948331833,
-0.01622559316456318,
0.04623444378376007,
-0.02704801969230175,
0.19120801985263824,
-0.024701936170458794,
0.049393873661756516,
0.06364397704601288,
0.009232889860868454,
-0.013832193799316883,
0.11151392012834549,
0.05708572641015053,
0.024334950372576714,
0.022262847051024437,
0.003451440716162324,
-0.04008655622601509,
-0.009981024079024792,
-0.18596695363521576,
0.06803664565086365,
0.14585918188095093,
0.09060460329055786,
-0.012669353745877743,
0.0707244873046875,
-0.10161512345075607,
-0.12005364894866943,
0.10127941519021988,
-0.06415384262800217,
-0.010188822634518147,
-0.06542414426803589,
0.14027701318264008,
0.14953285455703735,
-0.1886233240365982,
0.06583356112241745,
-0.06602055579423904,
-0.0566304549574852,
-0.11457879096269608,
-0.1930263340473175,
-0.057075321674346924,
-0.050602465867996216,
-0.018466074019670486,
-0.05384097993373871,
0.06939727067947388,
0.05750798434019089,
0.01126816775649786,
0.00868057832121849,
0.08568526059389114,
-0.009656033478677273,
0.00248199631460011,
0.030120067298412323,
0.06713981181383133,
0.016768986359238625,
-0.0321255661547184,
0.0179112758487463,
-0.00597198773175478,
0.034156378358602524,
0.059282708913087845,
0.03608176112174988,
-0.028436895459890366,
0.015559280291199684,
-0.034912437200546265,
-0.11309733241796494,
0.042801856994628906,
-0.029640642926096916,
-0.0749855786561966,
0.1347348988056183,
0.026981467381119728,
0.005015076603740454,
-0.023140020668506622,
0.2503887414932251,
-0.07436972856521606,
-0.09334370493888855,
-0.14373961091041565,
0.11701542884111404,
-0.04212593287229538,
0.0635172426700592,
0.03596310690045357,
-0.10810714215040207,
0.017985546961426735,
0.1320217251777649,
0.15442703664302826,
-0.04732590913772583,
0.019251897931098938,
0.028577854856848717,
0.00439635943621397,
-0.04075566306710243,
0.05177190154790878,
0.07100846618413925,
0.14500564336776733,
-0.05157303810119629,
0.08530787378549576,
0.002609728369861841,
-0.1021018698811531,
-0.041973695158958435,
0.11415864527225494,
-0.014296893030405045,
0.017620453611016273,
-0.057136841118335724,
0.124222531914711,
-0.05874236673116684,
-0.23697422444820404,
0.06316976249217987,
-0.0765061303973198,
-0.1432730257511139,
-0.024886758998036385,
0.071670763194561,
-0.016632623970508575,
0.02605951391160488,
0.07167234271764755,
-0.0754380151629448,
0.18880942463874817,
0.03957989811897278,
-0.05233397334814072,
-0.05954399332404137,
0.0744764655828476,
-0.11850855499505997,
0.27879106998443604,
0.010482731275260448,
0.051307905465364456,
0.1042102724313736,
-0.02021743729710579,
-0.13270841538906097,
0.023401619866490364,
0.09579801559448242,
-0.08917027711868286,
0.04087764397263527,
0.21448291838169098,
-0.00629545608535409,
0.11935057491064072,
0.07611140608787537,
-0.07468950748443604,
0.047562725841999054,
-0.11468592286109924,
-0.07639975845813751,
-0.08699081838130951,
0.09244474768638611,
-0.06785612553358078,
0.14258281886577606,
0.12599852681159973,
-0.05530165135860443,
0.011584274470806122,
-0.028389399871230125,
0.045467376708984375,
0.005578654818236828,
0.100032277405262,
0.011115525849163532,
-0.18496567010879517,
0.024811718612909317,
0.016259413212537766,
0.10884406417608261,
-0.18112654983997345,
-0.09105053544044495,
0.046958595514297485,
0.0005061255069449544,
-0.06443515419960022,
0.12483241409063339,
0.057313691824674606,
0.04654949903488159,
-0.0451689288020134,
-0.026830285787582397,
-0.006042256020009518,
0.14264579117298126,
-0.10707559436559677,
-0.005129707511514425
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# mixtral-8x7b-v0.1-english-to-hinglish-translation
This model is a fine-tuned version of [mistralai/Mixtral-8x7B-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-v0.1) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 1.0769
- Rouge Scores: {'rouge1': 0.9045408202972536, 'rouge2': 0.795425441228359, 'rougeL': 0.8399846297860634, 'rougeLsum': 0.9043739034131012}
- Bleu Scores: [0.0002881182166187815, 0.0002842750061873772, 0.0002764768847375588, 0.00026750640347869873]
- Gen Len: 2048.0
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 2
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 4
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- num_epochs: 2
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Rouge Scores | Bleu Scores | Gen Len |
|:-------------:|:-----:|:----:|:---------------:|:-----------------------------------------------------------------------------------------------------------------------:|:---------------------------------------------------------------------------------------------:|:--------:|
| 1.1771 | 1.0 | 500 | 1.0579 | {'rouge1': 0.9070255400902434, 'rouge2': 0.7976770190068221, 'rougeL': 0.8400261479965636, 'rougeLsum': 0.9069363147075731} | [0.00028395954091190866, 0.0002796973368739713, 0.0002722057765709132, 0.000263740024418467] | 2047.996 |
| 0.7788 | 2.0 | 1000 | 1.0769 | {'rouge1': 0.90.45408202972536, 'rouge2': 0.795425441228359, 'rougeL': 0.8399846297860634, 'rougeLsum': 0.9043739034131012} | [0.0002881182166187815, 0.0002842750061873772, 0.0002764768847375588, 0.00026750640347869873] | 2048.0 |
### Framework versions
- PEFT 0.8.2
- Transformers 4.38.0.dev0
- Pytorch 2.1.0+cu118
- Datasets 2.16.2.dev0
- Tokenizers 0.15.1 | {"license": "apache-2.0", "library_name": "peft", "tags": ["generated_from_trainer"], "base_model": "mistralai/Mixtral-8x7B-v0.1", "model-index": [{"name": "mixtral-8x7b-v0.1-english-to-hinglish-translation", "results": []}]} | null | DrishtiSharma/mixtral-8x7b-v0.1-english-to-hinglish-translation | [
"peft",
"safetensors",
"generated_from_trainer",
"base_model:mistralai/Mixtral-8x7B-v0.1",
"license:apache-2.0",
"region:us"
] | 2024-02-07T09:27:11+00:00 | [] | [] | TAGS
#peft #safetensors #generated_from_trainer #base_model-mistralai/Mixtral-8x7B-v0.1 #license-apache-2.0 #region-us
| mixtral-8x7b-v0.1-english-to-hinglish-translation
=================================================
This model is a fine-tuned version of mistralai/Mixtral-8x7B-v0.1 on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 1.0769
* Rouge Scores: {'rouge1': 0.9045408202972536, 'rouge2': 0.795425441228359, 'rougeL': 0.8399846297860634, 'rougeLsum': 0.9043739034131012}
* Bleu Scores: [0.0002881182166187815, 0.0002842750061873772, 0.0002764768847375588, 0.00026750640347869873]
* Gen Len: 2048.0
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0001
* train\_batch\_size: 2
* eval\_batch\_size: 8
* seed: 42
* gradient\_accumulation\_steps: 2
* total\_train\_batch\_size: 4
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: cosine
* num\_epochs: 2
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* PEFT 0.8.2
* Transformers 4.38.0.dev0
* Pytorch 2.1.0+cu118
* Datasets 2.16.2.dev0
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 2\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* num\\_epochs: 2\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.0+cu118\n* Datasets 2.16.2.dev0\n* Tokenizers 0.15.1"
] | [
"TAGS\n#peft #safetensors #generated_from_trainer #base_model-mistralai/Mixtral-8x7B-v0.1 #license-apache-2.0 #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 2\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* num\\_epochs: 2\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.0+cu118\n* Datasets 2.16.2.dev0\n* Tokenizers 0.15.1"
] | [
48,
141,
4,
47
] | [
"passage: TAGS\n#peft #safetensors #generated_from_trainer #base_model-mistralai/Mixtral-8x7B-v0.1 #license-apache-2.0 #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 2\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* num\\_epochs: 2\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.0+cu118\n* Datasets 2.16.2.dev0\n* Tokenizers 0.15.1"
] | [
-0.14640560746192932,
0.0949043333530426,
-0.0022138371132314205,
0.0970068871974945,
0.15598048269748688,
0.018265996128320694,
0.1260385364294052,
0.10380406677722931,
-0.1044301837682724,
0.1003396064043045,
0.09516225755214691,
0.06478217244148254,
0.04588183015584946,
0.18417149782180786,
-0.048930875957012177,
-0.22381886839866638,
0.020329153165221214,
-0.013198848813772202,
-0.0700039342045784,
0.12506432831287384,
0.0853566899895668,
-0.1368260532617569,
0.050949953496456146,
-0.020914264023303986,
-0.17555008828639984,
-0.015077201649546623,
-0.0008135774987749755,
-0.03187231346964836,
0.12077738344669342,
0.00787449348717928,
0.15129534900188446,
0.019674276933073997,
0.12004219740629196,
-0.19991706311702728,
0.0075830514542758465,
0.08222773671150208,
0.01841670647263527,
0.08762526512145996,
0.10509158670902252,
0.0036872609052807093,
0.061725009232759476,
-0.09575462341308594,
0.05589897558093071,
0.019824769347906113,
-0.14171110093593597,
-0.2522887885570526,
-0.13585563004016876,
0.056419696658849716,
0.10708194226026535,
0.0875641480088234,
-0.008081763982772827,
0.1197894811630249,
-0.08440963178873062,
0.059690941125154495,
0.2658289074897766,
-0.2635667324066162,
-0.06957884877920151,
0.009885444305837154,
-0.0009996788576245308,
0.09029974788427353,
-0.09722495079040527,
-0.05715815722942352,
0.05654873326420784,
0.039278410375118256,
0.10977347195148468,
0.009684570133686066,
-0.05927809700369835,
-0.005400520283728838,
-0.1615811288356781,
-0.04728570953011513,
0.07683032751083374,
0.039106957614421844,
-0.043738953769207,
-0.018991611897945404,
-0.07860298454761505,
-0.21265336871147156,
-0.04656893014907837,
0.02364559844136238,
0.038019709289073944,
-0.04140007868409157,
-0.040198709815740585,
0.03684745728969574,
-0.08288691937923431,
-0.08950698375701904,
-0.019941911101341248,
0.140784353017807,
0.06139045208692551,
0.010982791893184185,
-0.006894687656313181,
0.12706264853477478,
0.0018892328953370452,
-0.15607218444347382,
0.020821483805775642,
0.020132696256041527,
-0.047173529863357544,
-0.03210809826850891,
-0.04294047877192497,
0.003640729933977127,
0.01291111670434475,
0.16343899071216583,
-0.1410822719335556,
0.04331934079527855,
0.004670781549066305,
0.03465193882584572,
-0.1037691980600357,
0.1318848580121994,
-0.09117485582828522,
0.0008781899814493954,
0.020089734345674515,
0.10393577069044113,
0.04461086913943291,
-0.0013511452125385404,
-0.06939158588647842,
0.008165236562490463,
0.10309961438179016,
0.04833536222577095,
-0.0392504446208477,
0.010205480270087719,
-0.05807731673121452,
-0.0005901326658204198,
0.09520547091960907,
-0.10205360502004623,
0.04453050345182419,
0.034286368638277054,
-0.0725107416510582,
-0.02045944333076477,
0.012275791727006435,
0.014195147901773453,
-0.004967094864696264,
0.14264443516731262,
-0.08581218868494034,
0.025489989668130875,
-0.10065501928329468,
-0.12313704192638397,
0.038347501307725906,
-0.035645078867673874,
-0.012688199058175087,
-0.09586312621831894,
-0.13893474638462067,
-0.04014056175947189,
0.021731141954660416,
-0.04979148879647255,
-0.04125303030014038,
-0.03633689507842064,
-0.09904713928699493,
0.025000646710395813,
-0.029753826558589935,
0.11706487089395523,
-0.07010997086763382,
0.12146713584661484,
0.017469435930252075,
0.048470474779605865,
-0.033865753561258316,
0.03858640789985657,
-0.07970381528139114,
0.049203794449567795,
-0.22624626755714417,
0.026287510991096497,
-0.07164513319730759,
0.053206417709589005,
-0.11028600484132767,
-0.10090553015470505,
0.021386966109275818,
-0.040392667055130005,
0.14162351191043854,
0.12614881992340088,
-0.16985547542572021,
-0.038023196160793304,
0.18542835116386414,
-0.1070709154009819,
-0.1103619635105133,
0.12054172903299332,
-0.03246605396270752,
-0.046367671340703964,
0.034442901611328125,
0.1425393968820572,
0.06116578355431557,
-0.1190655380487442,
-0.010301422327756882,
-0.06341687589883804,
0.11672158539295197,
-0.04421054944396019,
0.09138720482587814,
-0.018448928371071815,
0.011861385777592659,
0.0053605674766004086,
-0.06687971949577332,
0.07834989577531815,
-0.11558070778846741,
-0.0760035440325737,
-0.041269972920417786,
-0.08804500102996826,
0.044668205082416534,
0.056113533675670624,
0.033706434071063995,
-0.08806444704532623,
-0.08996976166963577,
0.050774913281202316,
0.11915984749794006,
-0.05823253467679024,
0.03482460230588913,
-0.03703812137246132,
0.117939293384552,
-0.08279295265674591,
-0.03201018273830414,
-0.19188080728054047,
-0.07984213531017303,
0.03020678274333477,
-0.030651982873678207,
-0.02153867855668068,
-0.07181814312934875,
0.05773966386914253,
0.10689039528369904,
-0.06534454971551895,
-0.05105745792388916,
-0.07410944253206253,
-0.009940562769770622,
-0.11429467052221298,
-0.22755521535873413,
-0.05650068446993828,
-0.022116418927907944,
0.14845941960811615,
-0.2088415026664734,
0.01819704845547676,
-0.010771462693810463,
0.11903396248817444,
0.017239514738321304,
-0.03426909074187279,
-0.03124137967824936,
0.10408259183168411,
-0.015399623662233353,
-0.07275378704071045,
0.04523430019617081,
0.00987403653562069,
-0.050339195877313614,
-0.00679581006988883,
-0.13213437795639038,
0.12911920249462128,
0.11845409125089645,
-0.0038930699229240417,
-0.09486369788646698,
-0.03894888609647751,
-0.07103012502193451,
-0.04006091505289078,
-0.06335346400737762,
0.03822644427418709,
0.11430640518665314,
0.0225739274173975,
0.12791971862316132,
-0.09270412474870682,
-0.034621790051460266,
0.027214353904128075,
-0.0075188023038208485,
0.043796107172966,
0.11789953708648682,
0.12571927905082703,
-0.034433554857969284,
0.1366298943758011,
0.13441993296146393,
-0.08531314879655838,
0.06760851293802261,
-0.07478614151477814,
-0.10614529252052307,
-0.021395782008767128,
0.027746684849262238,
0.02081102877855301,
0.14579027891159058,
-0.010024376213550568,
0.03350664675235748,
0.019527502357959747,
0.03120581805706024,
0.01276959478855133,
-0.223936527967453,
-0.0346142053604126,
0.009619397111237049,
-0.050851840525865555,
-0.049515217542648315,
-0.031200360506772995,
0.027270440012216568,
0.11622195690870285,
-0.0034348813351243734,
-0.06752946972846985,
-0.005863505881279707,
0.0013088744599372149,
-0.08576991409063339,
0.20522081851959229,
-0.10605724155902863,
-0.07638370245695114,
-0.10353247821331024,
-0.010889536701142788,
-0.025792742148041725,
-0.011171880178153515,
0.04821604862809181,
-0.08460057526826859,
-0.029956715181469917,
-0.08284903317689896,
-0.0076535348780453205,
0.034292835742235184,
0.027136551216244698,
-0.012522991746664047,
-0.011083115823566914,
0.08055154234170914,
-0.0930367112159729,
0.01076371781527996,
-0.030039459466934204,
-0.037946898490190506,
0.046903450042009354,
0.06398394703865051,
0.12120697647333145,
0.17254647612571716,
-0.004642007872462273,
-0.001986759016290307,
-0.027518976479768753,
0.2120683789253235,
-0.07005185633897781,
-0.02672528102993965,
0.10966461151838303,
-0.0060749235562980175,
0.07440460473299026,
0.1619066596031189,
0.05885974317789078,
-0.0922505334019661,
0.015361825004220009,
0.0290226973593235,
-0.024730047211050987,
-0.23751556873321533,
-0.06362318247556686,
-0.04427599161863327,
-0.03858146071434021,
0.0944497361779213,
0.028965380042791367,
-0.02627992257475853,
0.04247066751122475,
-0.015238389372825623,
0.022459451109170914,
-0.027570374310016632,
0.07228218764066696,
0.04492979496717453,
0.06365901976823807,
0.11288361251354218,
-0.03225773945450783,
-0.01674531400203705,
0.04038730636239052,
-0.019331449642777443,
0.24213117361068726,
-0.04539690911769867,
0.0992506742477417,
0.05348093807697296,
0.19672344624996185,
0.0041825855150818825,
0.08410915732383728,
0.013840205036103725,
-0.019203627482056618,
0.009809186682105064,
-0.06400056928396225,
-0.021715832874178886,
0.012422718107700348,
-0.06057489290833473,
0.061877958476543427,
-0.13264383375644684,
-0.016325004398822784,
0.04209315404295921,
0.30958953499794006,
0.07665430754423141,
-0.34152984619140625,
-0.08254638314247131,
-0.0028347389306873083,
-0.004233118146657944,
-0.030600599944591522,
0.003474947065114975,
0.13346567749977112,
-0.0630277767777443,
0.03651508688926697,
-0.06716152280569077,
0.08828528970479965,
-0.017427826300263405,
-0.0011469689197838306,
0.07489681243896484,
0.10726141929626465,
-0.011078612878918648,
0.03151366114616394,
-0.2471177875995636,
0.287949800491333,
0.008818118833005428,
0.07353699952363968,
-0.017519250512123108,
0.0002660927420947701,
0.02200806885957718,
0.04728071019053459,
0.03875724598765373,
-0.005930740386247635,
-0.06878013908863068,
-0.2262311577796936,
-0.12226033955812454,
0.027243351563811302,
0.09858562797307968,
-0.01670847274363041,
0.12491755187511444,
-0.01770298182964325,
-0.007265890948474407,
0.05229717120528221,
-0.041998445987701416,
-0.10068504512310028,
-0.05500418692827225,
0.00006825767923146486,
0.00801772065460682,
0.0023668615613132715,
-0.09679742902517319,
-0.10982722043991089,
-0.07639049738645554,
0.09038429707288742,
-0.048076990991830826,
-0.044743429869413376,
-0.13134416937828064,
0.08990121632814407,
0.1396476924419403,
-0.08468132466077805,
0.0587385818362236,
0.013390570878982544,
0.09056355804204941,
0.010453937575221062,
-0.03497816622257233,
0.10335133969783783,
-0.06783188879489899,
-0.2235548347234726,
-0.06738471239805222,
0.11803597211837769,
0.053047046065330505,
0.05816718190908432,
-0.03254913166165352,
0.03924708068370819,
0.013807934708893299,
-0.09808412939310074,
0.002367964945733547,
0.030244402587413788,
0.08549684286117554,
0.048150546848773956,
-0.04151657596230507,
0.007142925634980202,
-0.05275459215044975,
-0.03679803013801575,
0.11358512192964554,
0.3184312582015991,
-0.09193488955497742,
0.013579524122178555,
0.060035932809114456,
-0.04953720420598984,
-0.1771426498889923,
0.03352978080511093,
0.09788811206817627,
0.003706574672833085,
0.01493658684194088,
-0.14693666994571686,
0.07560848444700241,
0.12621374428272247,
-0.0363864004611969,
0.12206154316663742,
-0.334684282541275,
-0.12920978665351868,
0.07328895479440689,
0.1371200531721115,
0.050523243844509125,
-0.17923791706562042,
-0.04395259544253349,
-0.014884069561958313,
-0.13786554336547852,
0.077996164560318,
-0.10167476534843445,
0.08550374954938889,
-0.02473815530538559,
0.04147029295563698,
0.008694744668900967,
-0.06810484826564789,
0.1520131528377533,
0.003699707565829158,
0.10027208179235458,
-0.030788995325565338,
0.01346526574343443,
0.04189818352460861,
-0.06078721955418587,
0.008231984451413155,
-0.03345704823732376,
0.04786330461502075,
-0.08496461808681488,
0.0005464920541271567,
-0.09259392321109772,
0.005488686729222536,
-0.04467669501900673,
-0.03725944086909294,
-0.0405692420899868,
0.04026109352707863,
0.052016451954841614,
-0.012990020215511322,
0.11024760454893112,
0.022404903545975685,
0.17006687819957733,
0.09609874337911606,
0.020526451990008354,
-0.024590739980340004,
-0.04452180489897728,
0.005139636807143688,
-0.035509053617715836,
0.05483386293053627,
-0.15638718008995056,
0.023960549384355545,
0.14691376686096191,
0.035113442689180374,
0.10688168555498123,
0.06898616254329681,
-0.06781566888093948,
0.022913604974746704,
0.06957285106182098,
-0.14013323187828064,
-0.0810093805193901,
0.030661024153232574,
0.038234446197748184,
-0.12150349467992783,
0.026099659502506256,
0.10457810014486313,
-0.0759090855717659,
-0.025991717353463173,
-0.0007750107906758785,
0.034725822508335114,
-0.03855705261230469,
0.23171457648277283,
0.0692497119307518,
0.06545790284872055,
-0.10601788759231567,
0.09132301807403564,
0.07718560099601746,
-0.09955454617738724,
-0.00020752957789227366,
0.09730606526136398,
-0.08695430308580399,
-0.01667947880923748,
0.10304199904203415,
0.11395573616027832,
-0.044851530343294144,
-0.05153551325201988,
-0.1356903612613678,
-0.123543880879879,
0.08148711919784546,
0.1589675396680832,
0.06355654448270798,
0.019572492688894272,
0.015115607529878616,
0.00919468142092228,
-0.1298055201768875,
0.09320537745952606,
0.06768804043531418,
0.09839392453432083,
-0.11900916695594788,
0.15722699463367462,
-0.0020327132660895586,
0.01711803302168846,
-0.00725482078269124,
0.05803662911057472,
-0.1151190921664238,
0.016342276707291603,
-0.1612357646226883,
-0.0031828226055949926,
-0.029432425275444984,
-0.010682722553610802,
-0.017764568328857422,
-0.05984974652528763,
-0.05094067379832268,
0.026116909459233284,
-0.1091768741607666,
-0.04483085870742798,
-0.008391980081796646,
0.02622257173061371,
-0.14530646800994873,
-0.035902239382267,
0.027901697903871536,
-0.09439125657081604,
0.08050365000963211,
0.049852218478918076,
0.051174845546483994,
0.045636024326086044,
-0.13982003927230835,
-0.008802435360848904,
0.05380294471979141,
-0.014280383475124836,
0.06030413508415222,
-0.12806518375873566,
-0.005437965504825115,
-0.022908134385943413,
0.02687859907746315,
0.011885514482855797,
0.06943365186452866,
-0.13297975063323975,
-0.0037933774292469025,
-0.0254488755017519,
-0.0441875234246254,
-0.04294178634881973,
0.016580525785684586,
0.08670788258314133,
0.030220018699765205,
0.13495664298534393,
-0.0906880721449852,
0.049504030495882034,
-0.22559459507465363,
-0.02406885288655758,
-0.03708739951252937,
-0.06856876611709595,
-0.1098412349820137,
-0.01097408402711153,
0.090223528444767,
-0.04142294079065323,
0.07750798761844635,
-0.031939245760440826,
0.08602402359247208,
0.03447055071592331,
-0.04135055094957352,
0.004845249932259321,
0.05529143288731575,
0.18025913834571838,
0.03734496235847473,
-0.038693543523550034,
0.053228314965963364,
0.032912883907556534,
0.0889686718583107,
0.10324250906705856,
0.2116452008485794,
0.1581987589597702,
0.04886293411254883,
0.07390394061803818,
0.043226227164268494,
-0.11370136588811874,
-0.1188710480928421,
0.04357827454805374,
-0.054231442511081696,
0.08382532000541687,
-0.022244051098823547,
0.214538112282753,
0.09783143550157547,
-0.1831723004579544,
0.02576923929154873,
-0.04630438610911369,
-0.0767943263053894,
-0.10127084702253342,
0.021735645830631256,
-0.07298412173986435,
-0.166891410946846,
-0.0052778753452003,
-0.11225634813308716,
0.024257445707917213,
0.13760782778263092,
0.01083324197679758,
0.021915031597018242,
0.14160144329071045,
0.09873900562524796,
0.031898751854896545,
0.04699470475316048,
0.0342303104698658,
0.006053876131772995,
-0.019063834100961685,
-0.10201695561408997,
0.03906271979212761,
-0.0402551144361496,
0.05535975471138954,
-0.030498795211315155,
-0.06071271002292633,
0.07550542801618576,
0.018589969724416733,
-0.10561198741197586,
0.03284674137830734,
0.012290529906749725,
0.06490860134363174,
0.09810072183609009,
0.02577810175716877,
0.041362348943948746,
-0.025007009506225586,
0.21725524961948395,
-0.06829187273979187,
-0.06572375446557999,
-0.11366546899080276,
0.27698400616645813,
0.029860947281122208,
-0.015646588057279587,
0.03445936739444733,
-0.10257834196090698,
-0.01644597202539444,
0.1283629983663559,
0.13758833706378937,
-0.08363451063632965,
-0.01442523580044508,
0.000482631498016417,
-0.02078557386994362,
-0.06039886921644211,
0.12467149645090103,
0.13403566181659698,
0.06291230022907257,
-0.10524236410856247,
-0.029326751828193665,
-0.06572242081165314,
-0.018951665610074997,
-0.032707687467336655,
0.03581508249044418,
-0.004625854082405567,
0.006173963192850351,
-0.04911049082875252,
0.08052129298448563,
-0.035500865429639816,
-0.09903531521558762,
0.07036252319812775,
-0.1930624544620514,
-0.18911586701869965,
-0.015762338414788246,
0.05366590991616249,
0.009017164818942547,
0.05061373859643936,
-0.03329702094197273,
-0.012050766497850418,
0.12464173138141632,
-0.03858862817287445,
-0.03224568068981171,
-0.13386641442775726,
0.06674961745738983,
-0.08406199514865875,
0.22460965812206268,
-0.025015927851200104,
0.05649213120341301,
0.12156196683645248,
0.05242006108164787,
-0.12931400537490845,
0.0516696572303772,
0.07256040722131729,
-0.10403183102607727,
0.012149069458246231,
0.12020670622587204,
-0.0548156276345253,
0.06521465629339218,
0.04206050559878349,
-0.11280994862318039,
0.012115568853914738,
-0.05040190741419792,
-0.04436232149600983,
-0.0537833534181118,
-0.04380316659808159,
-0.04337761551141739,
0.12479574233293533,
0.18094177544116974,
-0.05679793655872345,
0.03044121339917183,
-0.05124770477414131,
0.0183126050978899,
0.03766842558979988,
0.10078079998493195,
-0.016807394102215767,
-0.2507815361022949,
0.05342135205864906,
0.0668652206659317,
0.012538307346403599,
-0.23604899644851685,
-0.059386320412158966,
0.0240347720682621,
-0.07440602779388428,
-0.10452477633953094,
0.1150641068816185,
0.03604064881801605,
0.04834787920117378,
-0.05519169941544533,
-0.12890484929084778,
-0.077102892100811,
0.17093037068843842,
-0.14247262477874756,
-0.09175703674554825
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# daily-mail
This model is a fine-tuned version of [microsoft/phi-2](https://huggingface.co/microsoft/phi-2) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 12
- eval_batch_size: 12
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 48
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 4
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.38.0.dev0
- Pytorch 2.2.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1 | {"license": "mit", "library_name": "peft", "tags": ["summarization", "generated_from_trainer"], "base_model": "microsoft/phi-2", "model-index": [{"name": "daily-mail", "results": []}]} | summarization | RMWeerasinghe/daily-mail | [
"peft",
"tensorboard",
"safetensors",
"summarization",
"generated_from_trainer",
"base_model:microsoft/phi-2",
"license:mit",
"region:us"
] | 2024-02-07T09:27:23+00:00 | [] | [] | TAGS
#peft #tensorboard #safetensors #summarization #generated_from_trainer #base_model-microsoft/phi-2 #license-mit #region-us
|
# daily-mail
This model is a fine-tuned version of microsoft/phi-2 on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 12
- eval_batch_size: 12
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 48
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 4
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.38.0.dev0
- Pytorch 2.2.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1 | [
"# daily-mail\n\nThis model is a fine-tuned version of microsoft/phi-2 on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 12\n- eval_batch_size: 12\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 48\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 4",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.38.0.dev0\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#peft #tensorboard #safetensors #summarization #generated_from_trainer #base_model-microsoft/phi-2 #license-mit #region-us \n",
"# daily-mail\n\nThis model is a fine-tuned version of microsoft/phi-2 on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 12\n- eval_batch_size: 12\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 48\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 4",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.38.0.dev0\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
43,
27,
6,
12,
8,
3,
112,
4,
44
] | [
"passage: TAGS\n#peft #tensorboard #safetensors #summarization #generated_from_trainer #base_model-microsoft/phi-2 #license-mit #region-us \n# daily-mail\n\nThis model is a fine-tuned version of microsoft/phi-2 on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 12\n- eval_batch_size: 12\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 48\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 4### Training results### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.38.0.dev0\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.1333940625190735,
0.10941364616155624,
-0.0018196537857875228,
0.0761205181479454,
0.12676796317100525,
0.0018843988655135036,
0.10394652187824249,
0.11700453609228134,
-0.08630973845720291,
0.11865365505218506,
0.09202639013528824,
0.053347840905189514,
0.05507811903953552,
0.18872582912445068,
-0.00812380388379097,
-0.27205631136894226,
0.019687356427311897,
-0.014896614477038383,
-0.048672616481781006,
0.0927036702632904,
0.10758435726165771,
-0.09086751192808151,
0.06495356559753418,
0.01028546504676342,
-0.12316377460956573,
0.007281976751983166,
-0.031917233020067215,
-0.04190032556653023,
0.0828150138258934,
0.028527075424790382,
0.07287507504224777,
-0.006766681093722582,
0.11664831638336182,
-0.19299636781215668,
0.00313552631996572,
0.08022047579288483,
0.049489397555589676,
0.08940461277961731,
0.08883412927389145,
0.048852548003196716,
0.15380719304084778,
-0.10096298903226852,
0.07332013547420502,
0.034034907817840576,
-0.08796056360006332,
-0.17205595970153809,
-0.12105153501033783,
0.08782710880041122,
0.09944389015436172,
0.09616408497095108,
-0.005764976609498262,
0.16273927688598633,
-0.05502969026565552,
0.05873944237828255,
0.18593019247055054,
-0.27031704783439636,
-0.07902204990386963,
0.09201326221227646,
0.04455938935279846,
0.05882157012820244,
-0.10328864306211472,
-0.036311034113168716,
0.06909365206956863,
0.0408816933631897,
0.06301364302635193,
0.022699344903230667,
0.039490241557359695,
-0.030132519081234932,
-0.14948402345180511,
-0.038921624422073364,
0.15638111531734467,
0.08334998786449432,
-0.05525201931595802,
-0.11475609987974167,
-0.04920687898993492,
-0.13714590668678284,
-0.019430428743362427,
-0.012601821683347225,
0.030777981504797935,
-0.043448567390441895,
-0.041836418211460114,
-0.05907866358757019,
-0.07246346771717072,
-0.07228729128837585,
0.009408934973180294,
0.10009686648845673,
0.058652885258197784,
0.03102683648467064,
-0.019501209259033203,
0.11020743101835251,
-0.02199019119143486,
-0.0993647649884224,
-0.03237491473555565,
-0.007937162183225155,
-0.03133827820420265,
-0.039035867899656296,
-0.03434976935386658,
0.02336924709379673,
0.003964935429394245,
0.12538211047649384,
-0.09235376864671707,
0.06128240376710892,
0.056667037308216095,
0.026712151244282722,
-0.04795144870877266,
0.12093480676412582,
-0.050738625228405,
-0.0017445222474634647,
0.0008176884730346501,
0.12377554178237915,
0.04293370619416237,
0.005653209984302521,
-0.09174764901399612,
0.024298878386616707,
0.08070478588342667,
0.036547306925058365,
-0.041475869715213776,
0.02193920873105526,
-0.038555972278118134,
-0.028167393058538437,
0.06402361392974854,
-0.11373714357614517,
0.036674052476882935,
-0.007168111857026815,
-0.08365891128778458,
-0.00840515922755003,
0.025554930791258812,
-0.0011333145666867495,
-0.03409091383218765,
0.08391004055738449,
-0.09356880933046341,
-0.002585300011560321,
-0.10957653820514679,
-0.05201235041022301,
0.026304708793759346,
-0.05960899591445923,
-0.024520639330148697,
-0.07205216586589813,
-0.1871764361858368,
-0.053356826305389404,
0.015791594982147217,
-0.06290391832590103,
-0.0754828229546547,
-0.021682564169168472,
-0.07851440459489822,
-0.003107953816652298,
-0.02430117130279541,
0.11807243525981903,
-0.05067015439271927,
0.08847551792860031,
0.02962191216647625,
0.027055582031607628,
-0.009141902439296246,
0.036209993064403534,
-0.05541866272687912,
0.03326667845249176,
-0.12716062366962433,
0.052593521773815155,
-0.10664838552474976,
0.027426639571785927,
-0.11049841344356537,
-0.10766144096851349,
-0.04478299990296364,
-0.0020227842032909393,
0.07631257176399231,
0.11916416138410568,
-0.15077455341815948,
-0.026202023029327393,
0.15587559342384338,
-0.10528688877820969,
-0.09794151782989502,
0.07088478654623032,
-0.00425055343657732,
0.026868505403399467,
0.04689399525523186,
0.1476050615310669,
0.05191809684038162,
-0.16629283130168915,
0.0049376701936125755,
0.0281192846596241,
0.08654103428125381,
-0.041164472699165344,
0.06717971712350845,
-0.0568903423845768,
0.030320482328534126,
0.037662409245967865,
-0.06230540573596954,
-0.0032884725369513035,
-0.1026228666305542,
-0.05800911411643028,
-0.054698146879673004,
-0.0972837507724762,
0.0017868157010525465,
0.03815509006381035,
0.055839091539382935,
-0.06523381918668747,
-0.09479434788227081,
0.13041818141937256,
0.1394796371459961,
-0.026386171579360962,
0.0025582073722034693,
-0.08050207793712616,
0.07686550170183182,
-0.04426007345318794,
-0.037082087248563766,
-0.19200360774993896,
-0.09630605578422546,
0.035563625395298004,
-0.05492158979177475,
0.01587788201868534,
0.020818401128053665,
0.0571863055229187,
0.07616260647773743,
-0.052676498889923096,
-0.02568911388516426,
-0.10149889439344406,
-0.01033320277929306,
-0.11549340188503265,
-0.15706637501716614,
-0.06692410260438919,
-0.02155587263405323,
0.13637615740299225,
-0.23832190036773682,
0.0264426339417696,
-0.007573961280286312,
0.15380066633224487,
0.039169877767562866,
-0.08179475367069244,
-0.002755490830168128,
0.05875598266720772,
-0.0013969752471894026,
-0.09526371210813522,
0.05693387985229492,
0.018240952864289284,
-0.07622301578521729,
-0.09384433180093765,
-0.1275843381881714,
0.0888054221868515,
0.09340434521436691,
0.09683030098676682,
-0.07328673452138901,
-0.04109763726592064,
-0.0932442843914032,
-0.036072950810194016,
-0.08580195903778076,
-0.011056283488869667,
0.11329551786184311,
0.02066582255065441,
0.12868636846542358,
-0.08709736168384552,
-0.051991015672683716,
0.01616874523460865,
-0.0212301854044199,
-0.008519920520484447,
0.08751777559518814,
0.08049008250236511,
-0.036305516958236694,
0.09388261288404465,
0.052660778164863586,
-0.04303308576345444,
0.11513502895832062,
-0.07814067602157593,
-0.10982559621334076,
-0.005378415808081627,
0.03724294155836105,
0.006212146021425724,
0.14857040345668793,
-0.05736473947763443,
0.019303075969219208,
0.027564212679862976,
0.038424283266067505,
0.05602412670850754,
-0.19658440351486206,
-0.014485620893537998,
-0.004569095093756914,
-0.04420114681124687,
-0.007666068617254496,
-0.0009160835179500282,
0.007263073697686195,
0.08920209854841232,
0.014072692021727562,
-0.007974781095981598,
0.01215574610978365,
-0.008887657895684242,
-0.08497045189142227,
0.17666040360927582,
-0.08985307812690735,
-0.14594683051109314,
-0.15560415387153625,
0.08584128320217133,
-0.07120910286903381,
-0.02174476906657219,
0.01875157281756401,
-0.05367356538772583,
-0.04535594582557678,
-0.11965254694223404,
0.02117972820997238,
-0.013759653083980083,
-0.009355742484331131,
0.03656160831451416,
0.03153674677014351,
0.13175207376480103,
-0.12594135105609894,
0.014932267367839813,
-0.036480654031038284,
-0.08281604945659637,
-0.008210856467485428,
0.026647310703992844,
0.08901912719011307,
0.10717602074146271,
0.01620909757912159,
0.0368906632065773,
-0.03402169421315193,
0.24745915830135345,
-0.0879426822066307,
-0.014063448645174503,
0.16644124686717987,
0.0477592796087265,
0.051269374787807465,
0.07790818810462952,
0.03753003105521202,
-0.1181671991944313,
0.04400809854269028,
0.08764509111642838,
-0.021821213886141777,
-0.2230015993118286,
-0.05525863543152809,
-0.024368487298488617,
-0.04036528989672661,
0.08053701370954514,
0.07310259342193604,
-0.007239784114062786,
0.028610916808247566,
-0.021185217425227165,
0.004724168684333563,
-0.027430180460214615,
0.07668260484933853,
0.07080084830522537,
0.020259618759155273,
0.0940668061375618,
-0.05000627413392067,
0.003959364257752895,
0.07806100696325302,
0.00036585648194886744,
0.25242993235588074,
-0.009181884117424488,
0.030827045440673828,
0.05135980620980263,
0.16103672981262207,
-0.025076400488615036,
0.03970605880022049,
0.013209527358412743,
-0.0184868723154068,
0.006479938980191946,
-0.06449360400438309,
-0.03375205397605896,
0.03945840150117874,
-0.032179489731788635,
0.0674075037240982,
-0.1112225279211998,
0.030489355325698853,
0.011500190012156963,
0.28073081374168396,
0.06175427883863449,
-0.3008959889411926,
-0.09945966303348541,
0.003026542952284217,
-0.03346020355820656,
-0.06257200986146927,
0.03156988322734833,
0.11682812124490738,
-0.11680489033460617,
0.059152234345674515,
-0.05947718769311905,
0.09860509634017944,
-0.038748599588871,
0.003416069084778428,
0.05856943875551224,
0.12040864676237106,
0.007175874896347523,
0.072458416223526,
-0.19723977148532867,
0.22123213112354279,
0.0030882323626428843,
0.12564167380332947,
-0.024590138345956802,
0.025974348187446594,
0.010892754420638084,
0.08844452351331711,
0.08886412531137466,
-0.0028415226843208075,
-0.028525972738862038,
-0.16409853100776672,
-0.1027008444070816,
0.01690920628607273,
0.09812008589506149,
-0.02588808536529541,
0.07041261345148087,
-0.031013108789920807,
0.03293800354003906,
0.028739670291543007,
-0.07481618970632553,
-0.1585092842578888,
-0.09444445371627808,
0.015056944452226162,
-0.015657292678952217,
-0.025227297097444534,
-0.11686551570892334,
-0.11756084114313126,
-0.03802553564310074,
0.1292002648115158,
-0.023388206958770752,
-0.048357173800468445,
-0.1425132155418396,
0.10319149494171143,
0.11349405348300934,
-0.04562843218445778,
0.020950689911842346,
0.006410672329366207,
0.12673169374465942,
0.0443122498691082,
-0.05272146314382553,
0.068813756108284,
-0.052139583975076675,
-0.20790047943592072,
-0.037025537341833115,
0.13741247355937958,
0.035750359296798706,
0.03228207305073738,
-0.03384953737258911,
0.009061029180884361,
0.018013564869761467,
-0.10450984537601471,
0.030724598094820976,
0.10881071537733078,
0.05807158723473549,
0.020723741501569748,
-0.060597311705350876,
0.07885916531085968,
-0.01876949891448021,
-0.04769188538193703,
0.062023501843214035,
0.24509666860103607,
-0.08159680664539337,
0.029382528737187386,
0.026695340871810913,
-0.05695607513189316,
-0.1925608366727829,
0.049660053104162216,
0.1354835033416748,
0.02627270668745041,
0.04149649292230606,
-0.15921485424041748,
0.09199365228414536,
0.1505960077047348,
-0.0531037263572216,
0.09381725639104843,
-0.339327335357666,
-0.1324971318244934,
0.06101921200752258,
0.11971732974052429,
-0.007484446279704571,
-0.1665990650653839,
-0.069670669734478,
-0.0033131714444607496,
-0.10094006359577179,
0.052608780562877655,
-0.08753480762243271,
0.09985809773206711,
-0.011860388331115246,
0.08480292558670044,
0.015410192310810089,
-0.03172556310892105,
0.15253043174743652,
0.013391928747296333,
0.07985615730285645,
-0.03810897469520569,
0.05672190710902214,
0.03513682261109352,
-0.06912592053413391,
0.06646699458360672,
-0.008848071098327637,
0.08061733841896057,
-0.14906741678714752,
-0.022378329187631607,
-0.07959845662117004,
0.07175492495298386,
-0.06333236396312714,
-0.037610091269016266,
-0.0448460727930069,
0.05614832043647766,
-0.001194746932014823,
-0.02419872395694256,
0.11508775502443314,
-0.01304580643773079,
0.12153352051973343,
0.138810396194458,
0.04331984743475914,
-0.03217214345932007,
-0.1426299661397934,
-0.005260992329567671,
-0.017222201451659203,
0.05138464644551277,
-0.14428700506687164,
-0.017310474067926407,
0.11246512085199356,
0.03688636049628258,
0.11647778004407883,
0.028100719675421715,
-0.07639908045530319,
0.000829230179078877,
0.052147045731544495,
-0.09248816967010498,
-0.20093201100826263,
-0.004938456695526838,
0.049594447016716,
-0.13031090795993805,
0.06473509967327118,
0.08972019702196121,
-0.06185862421989441,
-0.004062621854245663,
-0.019423803314566612,
0.019874563440680504,
-0.020671825855970383,
0.1721196323633194,
0.06801927089691162,
0.06849662214517593,
-0.0747968852519989,
0.1343945413827896,
0.04355758801102638,
-0.09856154024600983,
0.013865575194358826,
0.06729352474212646,
-0.08947867155075073,
-0.009228945709764957,
0.06002639979124069,
0.11232160776853561,
-0.016367090865969658,
-0.06001325324177742,
-0.09315325319766998,
-0.10232294350862503,
0.05825284868478775,
0.05048059672117233,
0.06309912353754044,
-0.003634207881987095,
-0.020040428265929222,
0.03200918063521385,
-0.10831570625305176,
0.07825854420661926,
0.043631408363580704,
0.07154512405395508,
-0.17257103323936462,
0.09449893236160278,
0.014725616201758385,
0.019471479579806328,
-0.015367128886282444,
0.02703048661351204,
-0.08726448565721512,
-0.027634678408503532,
-0.098065085709095,
-0.008262217044830322,
-0.045376863330602646,
-0.00178023730404675,
-0.00507607776671648,
-0.059051353484392166,
-0.035255175083875656,
0.05222347006201744,
-0.07250513881444931,
-0.058661460876464844,
0.005622413009405136,
0.0658242404460907,
-0.09455738961696625,
0.016419216990470886,
0.0038829464465379715,
-0.09295164048671722,
0.07583434134721756,
0.02622324787080288,
0.03823191300034523,
0.028281694278120995,
-0.03215038403868675,
0.049530524760484695,
0.037457406520843506,
0.016328217461705208,
0.052736569195985794,
-0.0537848137319088,
-0.015126422047615051,
-0.020732494071125984,
0.05553127080202103,
0.005932919681072235,
0.04108092561364174,
-0.15494859218597412,
-0.05483793094754219,
-0.04734296351671219,
-0.035031240433454514,
-0.052828364074230194,
0.024016615003347397,
0.0852825939655304,
0.05442086234688759,
0.15038704872131348,
-0.09078691899776459,
0.009300812147557735,
-0.20564773678779602,
-0.01578722894191742,
-0.0006794249638915062,
-0.01177901215851307,
-0.06799151003360748,
-0.010383518412709236,
0.08940303325653076,
-0.031070901080965996,
0.11183101683855057,
-0.009582101367413998,
0.08030081540346146,
0.04775264859199524,
-0.07098866999149323,
0.028510931879281998,
0.005184744019061327,
0.1825186163187027,
0.03790416568517685,
0.022957628592848778,
0.09293901920318604,
-0.020114829763770103,
0.02880639210343361,
0.05560292303562164,
0.19562260806560516,
0.15842878818511963,
-0.04484839737415314,
0.09354456514120102,
0.0780840590596199,
-0.09329525381326675,
-0.14023001492023468,
0.06922443211078644,
-0.0017258194275200367,
0.10567395389080048,
-0.06899895519018173,
0.133492112159729,
0.11020689457654953,
-0.165383979678154,
0.026705611497163773,
-0.05210142582654953,
-0.09668859839439392,
-0.10751170665025711,
-0.036110538989305496,
-0.06906972080469131,
-0.15475773811340332,
0.025221770629286766,
-0.12033925950527191,
0.011587163433432579,
0.10313380509614944,
0.0017093598144128919,
-0.0060829659923911095,
0.14622850716114044,
-0.03161563724279404,
0.004976810421794653,
0.06248695030808449,
0.017539894208312035,
0.0029513004701584578,
-0.08721086382865906,
-0.07709914445877075,
0.03915600851178169,
-0.01216728612780571,
0.07340133190155029,
-0.033618099987506866,
-0.03611621633172035,
0.032635532319545746,
0.01660669595003128,
-0.05914897471666336,
0.029258424416184425,
0.022048065438866615,
0.02999044768512249,
0.08627134561538696,
0.03885490447282791,
-0.009061362594366074,
-0.0610538125038147,
0.27267739176750183,
-0.0910152941942215,
-0.06478430330753326,
-0.13028393685817719,
0.2586393654346466,
0.019808154553174973,
-0.0015246742404997349,
0.03878937289118767,
-0.13168229162693024,
-0.02524172142148018,
0.18182359635829926,
0.12837029993534088,
-0.0405680350959301,
-0.036876190453767776,
-0.007980973459780216,
-0.024534879252314568,
-0.07336151599884033,
0.11949530243873596,
0.10883752256631851,
0.08753181993961334,
-0.033774249255657196,
-0.012646371498703957,
-0.01724391244351864,
-0.018309487029910088,
-0.09052012115716934,
0.07219529151916504,
0.021299896761775017,
0.011363664641976357,
-0.0528528057038784,
0.08877179771661758,
-0.011850368231534958,
-0.15861091017723083,
0.06797818839550018,
-0.12151267379522324,
-0.18591293692588806,
-0.01079198531806469,
0.06588742882013321,
-0.01753147318959236,
0.07303636521100998,
-0.029899321496486664,
0.017461424693465233,
0.12629273533821106,
-0.03464232385158539,
-0.05383925512433052,
-0.11259055137634277,
0.06504621356725693,
-0.07818138599395752,
0.22030672430992126,
-0.007628411985933781,
0.05615084245800972,
0.10151328891515732,
0.0050030178390443325,
-0.15153388679027557,
0.07117520272731781,
0.045345526188611984,
-0.06339036673307419,
0.016574939712882042,
0.1204930767416954,
-0.025462036952376366,
0.08134246617555618,
0.031691793352365494,
-0.15818172693252563,
-0.03553006425499916,
-0.0320126935839653,
-0.039218224585056305,
-0.08435723185539246,
0.02551688626408577,
-0.07412657886743546,
0.14286242425441742,
0.1873387098312378,
-0.05672897771000862,
-0.01154869794845581,
-0.052560530602931976,
0.05606701970100403,
0.02952989749610424,
0.0646117627620697,
0.015227117575705051,
-0.2244129478931427,
0.02381080761551857,
-0.01978795789182186,
-0.013008488342165947,
-0.24992746114730835,
-0.06449048221111298,
0.024308476597070694,
-0.0539468452334404,
-0.05431145429611206,
0.11085532605648041,
0.049459196627140045,
0.01744857430458069,
-0.04685887694358826,
-0.11156603693962097,
-0.05053873360157013,
0.15662303566932678,
-0.1349789947271347,
-0.06682191044092178
] |
null | null | transformers |
# **PPO** Agent playing **LunarLander-v2**
This is a trained model of a **PPO** agent playing **LunarLander-v2**
using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3).
## Usage (with Stable-baselines3)
TODO: Add your code
```python
from stable_baselines3 import ...
from huggingface_sb3 import load_from_hub
...
```
| {"tags": ["Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "c", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "Taxi-v3", "type": "Taxi-v3"}, "metrics": [{"type": "mean_reward", "value": "7.50 +/- 2.72", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | JiajingChen/c | [
"transformers",
"Taxi-v3",
"q-learning",
"reinforcement-learning",
"custom-implementation",
"model-index",
"endpoints_compatible",
"region:us"
] | 2024-02-07T09:28:10+00:00 | [] | [] | TAGS
#transformers #Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #endpoints_compatible #region-us
|
# PPO Agent playing LunarLander-v2
This is a trained model of a PPO agent playing LunarLander-v2
using the stable-baselines3 library.
## Usage (with Stable-baselines3)
TODO: Add your code
| [
"# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
"TAGS\n#transformers #Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #endpoints_compatible #region-us \n",
"# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
43,
41,
17
] | [
"passage: TAGS\n#transformers #Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #endpoints_compatible #region-us \n# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
0.010647284798324108,
0.05591825768351555,
-0.0046826573088765144,
0.039160627871751785,
0.15529821813106537,
-0.02682899683713913,
0.09160283952951431,
0.06924816966056824,
0.06964346766471863,
0.009027299471199512,
0.0883723646402359,
0.13187670707702637,
0.05836737900972366,
0.22727036476135254,
0.06396697461605072,
-0.24795353412628174,
0.04590814933180809,
0.005390627775341272,
0.08898917585611343,
0.08219913393259048,
-0.013840213418006897,
-0.06991897523403168,
0.044786058366298676,
0.03306451067328453,
0.04057827219367027,
0.007438844069838524,
-0.035267822444438934,
-0.10607219487428665,
0.0696987584233284,
-0.03189856559038162,
0.06581401824951172,
0.1112155020236969,
0.11736108362674713,
-0.11452384293079376,
0.024646231904625893,
0.012737803161144257,
-0.0790938064455986,
0.008312288671731949,
0.016975965350866318,
0.06133352220058441,
0.20795203745365143,
0.10411357134580612,
0.06115853041410446,
0.0007176307844929397,
-0.13685089349746704,
-0.03742263838648796,
0.0439700186252594,
0.13477396965026855,
0.09176474809646606,
0.08193127810955048,
0.037638645619153976,
0.23070882260799408,
-0.02457231469452381,
0.013262856751680374,
0.21900737285614014,
-0.2828314006328583,
-0.0394911952316761,
0.26519301533699036,
0.13004806637763977,
0.09168682992458344,
-0.05968813970685005,
-0.07827112823724747,
0.016238823533058167,
-0.0177617110311985,
-0.03870905190706253,
-0.09515080600976944,
0.1946992725133896,
0.04757470637559891,
-0.08908537775278091,
-0.13529157638549805,
0.2035793960094452,
0.0003767554881051183,
-0.021965567022562027,
0.03074795752763748,
-0.010733796283602715,
0.00024969398509711027,
-0.0011879140511155128,
-0.10131251066923141,
0.03524215519428253,
0.04696304351091385,
-0.022483238950371742,
-0.11781688779592514,
-0.09127256274223328,
-0.030165376141667366,
-0.026692936196923256,
0.17069871723651886,
-0.0015929858200252056,
0.044922929257154465,
-0.049541912972927094,
0.09798445552587509,
-0.05672365799546242,
-0.004759443923830986,
-0.07529529929161072,
-0.09587089717388153,
0.009599474258720875,
-0.03857056051492691,
0.016278404742479324,
0.06165105849504471,
0.09520213305950165,
-0.04333045706152916,
0.12774422764778137,
0.03882104530930519,
-0.029234768822789192,
0.05570358783006668,
0.05417723208665848,
0.20908808708190918,
0.04022049903869629,
0.04302575811743736,
0.08491907268762589,
0.016182199120521545,
-0.02770201861858368,
-0.048573583364486694,
-0.1800871640443802,
0.07559529691934586,
-0.07093287259340286,
0.019033240154385567,
0.09380705654621124,
0.0158722884953022,
-0.08823023736476898,
-0.05050503462553024,
0.031153475865721703,
-0.010922384448349476,
-0.02158508263528347,
0.04142273962497711,
-0.011109001003205776,
-0.0375220850110054,
0.038366250693798065,
0.05059873312711716,
-0.04720938578248024,
-0.14566002786159515,
0.019221337512135506,
-0.026322443038225174,
-0.11710379272699356,
-0.006543826311826706,
-0.03290775045752525,
0.029439592733979225,
0.05412356182932854,
-0.14795072376728058,
-0.2794862389564514,
-0.053339965641498566,
0.0512259304523468,
-0.03270750865340233,
-0.17795875668525696,
-0.1364266574382782,
0.013802118599414825,
-0.10756172984838486,
-0.07346183061599731,
-0.1449166089296341,
-0.03172311186790466,
0.04072960838675499,
0.0006291939644142985,
0.0953512042760849,
-0.029116041958332062,
0.02028193138539791,
-0.15725606679916382,
0.021672062575817108,
-0.17373691499233246,
0.08825629949569702,
-0.016491809859871864,
0.03085380420088768,
-0.02600911818444729,
-0.08942197263240814,
0.06754802167415619,
0.04089779779314995,
-0.011137875728309155,
0.1278109848499298,
-0.07392006367444992,
-0.03675844892859459,
0.028277698904275894,
-0.12348049134016037,
-0.08924979716539383,
-0.09467785060405731,
-0.04015415534377098,
0.17784598469734192,
0.021278128027915955,
0.08057721704244614,
-0.02244153805077076,
-0.10666313022375107,
0.18312004208564758,
0.01554274931550026,
-0.18013614416122437,
-0.11668448150157928,
0.08621929585933685,
0.04811639338731766,
-0.12428194284439087,
-0.017266927286982536,
-0.1623666137456894,
0.024469668045639992,
-0.060088809579610825,
-0.030838649719953537,
0.014619441702961922,
-0.024757882580161095,
0.11852094531059265,
0.08734015375375748,
0.055614206939935684,
-0.04231562837958336,
-0.005607897415757179,
0.07611241191625595,
0.08909278362989426,
-0.0014986726455390453,
0.006681247614324093,
-0.10257206112146378,
0.0700591653585434,
-0.07994052022695541,
0.0024976758286356926,
-0.07507266104221344,
0.02607874572277069,
-0.025665542110800743,
0.1582638919353485,
0.047668181359767914,
0.24805226922035217,
0.08952954411506653,
0.05074971169233322,
0.06693936139345169,
-0.15240934491157532,
-0.09140733629465103,
0.0300681721419096,
0.04265368729829788,
-0.07991170138120651,
-0.024021336808800697,
-0.07595556974411011,
-0.22045445442199707,
-0.063082255423069,
0.012250197120010853,
-0.1531268209218979,
0.02449515089392662,
0.048142481595277786,
-0.003279504831880331,
0.006284699309617281,
0.1338403820991516,
-0.03129473328590393,
-0.0330820307135582,
0.09209538996219635,
0.05270373076200485,
-0.024760087952017784,
0.01879495196044445,
0.10817965865135193,
0.21236981451511383,
0.1661815345287323,
-0.06788532435894012,
0.021143194288015366,
0.10299233347177505,
-0.04246918112039566,
0.06631021946668625,
0.06070730835199356,
0.2005346417427063,
0.17523549497127533,
0.027980651706457138,
0.05212448537349701,
-0.0654800534248352,
0.10298378765583038,
-0.029085617512464523,
-0.11104383319616318,
-0.06088893860578537,
0.10297934710979462,
0.04731876403093338,
0.02414126507937908,
0.11098475009202957,
0.06183793023228645,
0.09838372468948364,
0.15126703679561615,
0.050599273294210434,
-0.08263081312179565,
-0.05878102034330368,
0.013103378936648369,
0.029793469235301018,
0.03281376138329506,
-0.14502128958702087,
0.0029670149087905884,
0.02378808706998825,
-0.08370528370141983,
0.022116046398878098,
-0.13854563236236572,
-0.12977540493011475,
0.06063363328576088,
0.02728014439344406,
-0.04079701378941536,
0.0343051478266716,
-0.029259569942951202,
0.014926346018910408,
0.019667591899633408,
-0.10319814085960388,
0.09114885330200195,
0.004913726355880499,
-0.02997569926083088,
0.033152371644973755,
-0.044967059046030045,
-0.2201126217842102,
-0.18726009130477905,
-0.030097972601652145,
-0.08479459583759308,
0.10531604290008545,
0.05659196525812149,
-0.10802599042654037,
0.04359238222241402,
-0.03606465458869934,
0.04601966589689255,
-0.04070669785141945,
-0.023597175255417824,
0.08628398925065994,
0.13684149086475372,
-0.03227521851658821,
-0.006012058816850185,
-0.0567224882543087,
-0.08056720346212387,
-0.20876994729042053,
0.09253337979316711,
-0.0244232639670372,
0.0499388687312603,
0.09185562282800674,
-0.014773430302739143,
0.04052964225411415,
-0.017009317874908447,
0.037261102348566055,
-0.020986419171094894,
-0.12099384516477585,
0.31674373149871826,
0.052270449697971344,
-0.013933219015598297,
0.024204930290579796,
-0.02841733954846859,
-0.004496568348258734,
0.02850200980901718,
-0.14418625831604004,
-0.0958755612373352,
-0.09492833912372589,
-0.03585296496748924,
-0.10057119280099869,
0.10398907214403152,
0.0729277953505516,
0.010605843737721443,
-0.045440420508384705,
0.12204065918922424,
0.10820804536342621,
-0.019174208864569664,
-0.09795982390642166,
0.10012811422348022,
0.08250096440315247,
-0.0734221339225769,
0.0065739392302930355,
-0.00036990095395594835,
-0.02684173546731472,
0.02718646638095379,
0.07512589544057846,
0.024721408262848854,
0.0906963050365448,
-0.16292020678520203,
0.03488987684249878,
0.11809540539979935,
0.04181523993611336,
0.10667180269956589,
0.09474433213472366,
-0.09898936748504639,
-0.01647394709289074,
-0.013986137695610523,
-0.22884415090084076,
0.14280346035957336,
0.14651858806610107,
0.09464110434055328,
-0.04365725815296173,
0.005233872216194868,
-0.07716752588748932,
0.10208811610937119,
-0.050500739365816116,
-0.21222636103630066,
-0.009447703137993813,
0.026418372988700867,
0.026701578870415688,
-0.004739189054816961,
0.09396522492170334,
-0.011989419348537922,
-0.1959003210067749,
-0.05109019577503204,
0.06277002394199371,
0.015130813233554363,
-0.0060146283358335495,
-0.008656795136630535,
-0.14519865810871124,
0.0004791046667378396,
-0.00189327378757298,
0.09880724549293518,
-0.24504920840263367,
0.15712471306324005,
-0.12085813283920288,
0.090500108897686,
-0.03692111745476723,
-0.020244276151061058,
0.02645593136548996,
0.02752482146024704,
0.17936578392982483,
-0.005117486231029034,
0.02031105011701584,
-0.17386992275714874,
-0.12971855700016022,
-0.0005864655831828713,
-0.1480615884065628,
-0.10522768646478653,
0.059657227247953415,
0.051054567098617554,
0.03859470784664154,
-0.07551214098930359,
0.15736886858940125,
-0.007610714063048363,
0.04864040017127991,
-0.048470329493284225,
-0.08273350447416306,
0.1280519664287567,
-0.10929372161626816,
-0.07392164319753647,
0.022464018315076828,
0.11464538425207138,
0.13368380069732666,
-0.04290906339883804,
-0.04230501130223274,
-0.11077317595481873,
-0.0356474332511425,
-0.016711948439478874,
0.01247849315404892,
0.021946726366877556,
0.05157327651977539,
-0.06570819765329361,
-0.09651155024766922,
0.049257513135671616,
-0.10092544555664062,
-0.07836932688951492,
-0.030526813119649887,
0.10828661173582077,
0.05058271810412407,
0.06490276008844376,
-0.011948661878705025,
0.009307135827839375,
-0.0293093454092741,
-0.06860842555761337,
0.12668928503990173,
-0.001004053745418787,
-0.028268957510590553,
0.024960948154330254,
0.09026949107646942,
0.053659021854400635,
0.01804523728787899,
-0.03551432117819786,
0.14342576265335083,
0.13421398401260376,
0.003213057294487953,
0.20312786102294922,
0.02821153961122036,
-0.05183521285653114,
-0.13656708598136902,
-0.09789294749498367,
0.0019813189283013344,
0.02336122840642929,
0.06657157093286514,
-0.11921804398298264,
-0.024277016520500183,
-0.052643030881881714,
-0.026234019547700882,
-0.15535689890384674,
-0.2818443477153778,
-0.042829882353544235,
0.245095357298851,
0.17994263768196106,
0.2527765929698944,
-0.1327243149280548,
0.05441126599907875,
0.005304550752043724,
-0.09499126672744751,
0.0335581973195076,
0.003314525354653597,
0.10501199960708618,
-0.14747551083564758,
0.1250685751438141,
0.07450491935014725,
-0.02283880114555359,
-0.020328914746642113,
-0.1214231550693512,
-0.01090170070528984,
-0.11884025484323502,
0.02239914797246456,
0.16241775453090668,
-0.02951703406870365,
-0.04378203675150871,
0.16558362543582916,
-0.013922485522925854,
-0.21546193957328796,
-0.05773348733782768,
-0.04103512316942215,
-0.009396309964358807,
0.03391183540225029,
-0.11205390840768814,
0.02457444928586483,
0.07816540449857712,
0.034613873809576035,
0.10562039166688919,
0.11730129271745682,
-0.0235004760324955,
0.05691837891936302,
0.2326793670654297,
0.007880778051912785,
0.018537258729338646,
-0.18447016179561615,
-0.009074043482542038,
-0.01836288720369339,
0.06323851644992828,
-0.06384648382663727,
-0.06871052831411362,
0.036990996450185776,
0.004039842169731855,
0.022551363334059715,
0.1097961813211441,
-0.005103940609842539,
0.09572429209947586,
0.034605078399181366,
-0.13101722300052643,
-0.18223898112773895,
-0.0014276646543294191,
0.010660535655915737,
0.10371889919042587,
0.06978340446949005,
0.07731527835130692,
-0.15082892775535583,
-0.011617588810622692,
-0.055527687072753906,
-0.04020017385482788,
-0.10592005401849747,
-0.03344229981303215,
0.0435391403734684,
0.03462565317749977,
-0.06495409458875656,
0.08501116186380386,
0.014927031472325325,
0.05306566506624222,
0.03277252987027168,
-0.033638741821050644,
0.03277148678898811,
-0.07761769741773605,
0.0918535590171814,
0.14246562123298645,
-0.035978544503450394,
-0.07240675389766693,
-0.09530327469110489,
-0.1036783829331398,
0.1284140944480896,
-0.07109623402357101,
0.08009420335292816,
-0.11795760691165924,
-0.08431815356016159,
0.022705620154738426,
-0.046562157571315765,
-0.08316844701766968,
-0.012237336486577988,
-0.00107108848169446,
-0.20584386587142944,
0.0570053830742836,
-0.04102408140897751,
0.10248378664255142,
-0.07692426443099976,
-0.11670845746994019,
-0.18782101571559906,
0.07976038753986359,
0.08463257551193237,
-0.023306291550397873,
-0.16619746387004852,
0.01732823997735977,
0.013083988800644875,
-0.06532903760671616,
-0.06703181564807892,
0.10360711812973022,
-0.1262943297624588,
0.045888613909482956,
0.017442625015974045,
0.10630720108747482,
0.020022472366690636,
0.005188741721212864,
0.033632971346378326,
-0.08204597234725952,
-0.016384363174438477,
0.03214522451162338,
-0.06656890362501144,
-0.06188099831342697,
-0.2197633981704712,
0.027356943115592003,
0.023624416440725327,
0.01920292340219021,
0.09957446902990341,
0.03639257326722145,
0.036462489515542984,
0.0824047327041626,
-0.16432641446590424,
-0.0281576756387949,
0.14066629111766815,
-0.08534271270036697,
-0.030407778918743134,
0.04571075364947319,
-0.0726674422621727,
0.019579846411943436,
-0.011494413949549198,
0.10921904444694519,
-0.002909967675805092,
0.11806085705757141,
-0.0814170092344284,
0.0997634008526802,
-0.14350275695323944,
-0.05444645136594772,
0.014392913319170475,
0.06677838414907455,
0.068927101790905,
-0.14745958149433136,
0.009734352119266987,
0.040032945573329926,
0.18196624517440796,
0.042446982115507126,
0.029254471883177757,
-0.04604627564549446,
-0.006669101770967245,
0.1947937309741974,
-0.0026392790023237467,
0.1311808079481125,
0.053562115877866745,
-0.015460791066288948,
0.09800700098276138,
0.11624976992607117,
0.09903048723936081,
-0.012889127247035503,
0.0688127651810646,
-0.004009180702269077,
0.0035461646039038897,
0.09177166223526001,
0.04425068199634552,
0.09803205728530884,
-0.07000865787267685,
0.023708397522568703,
0.01051484327763319,
0.02246539108455181,
-0.019095558673143387,
0.15566490590572357,
0.09464800357818604,
-0.09451071918010712,
0.02470262162387371,
-0.020145485177636147,
0.04234079644083977,
0.0027999659068882465,
-0.15031374990940094,
-0.07674021273851395,
-0.3746807277202606,
0.11271262913942337,
-0.04379986971616745,
0.0030189661774784327,
0.02338418737053871,
-0.03921474143862724,
-0.061096444725990295,
-0.03540882095694542,
0.12401590496301651,
-0.015259021893143654,
-0.036740392446517944,
-0.07021352648735046,
0.003598464885726571,
-0.17065368592739105,
-0.07408002763986588,
-0.027143921703100204,
-0.14462910592556,
0.0038331467658281326,
0.05723709613084793,
-0.009859615005552769,
0.019808132201433182,
-0.028410116210579872,
0.016416076570749283,
0.05636192858219147,
-0.03607961907982826,
0.029779022559523582,
0.11330032348632812,
0.056339114904403687,
-0.088828906416893,
0.04573938995599747,
0.10672074556350708,
0.03742805868387222,
-0.09923300892114639,
-0.0038765997160226107,
0.1944265216588974,
0.03149283304810524,
0.025310421362519264,
0.01762596145272255,
-0.019724810495972633,
-0.07446398586034775,
0.1724957376718521,
0.0847245529294014,
-0.059195104986429214,
0.00710638053715229,
-0.059128083288669586,
-0.02232721820473671,
-0.05240686610341072,
0.12159226089715958,
0.1103133112192154,
0.06978044658899307,
-0.10929165035486221,
-0.06869625300168991,
-0.1131792813539505,
0.03605156019330025,
-0.04381748288869858,
-0.05376826226711273,
0.08420785516500473,
-0.05683217570185661,
-0.03926567733287811,
0.021155767142772675,
-0.1984243392944336,
0.018296025693416595,
0.1764211654663086,
-0.17612510919570923,
-0.06639091670513153,
-0.08238485455513,
0.1371871680021286,
0.031113281846046448,
0.10920282453298569,
-0.009525357745587826,
-0.018109578639268875,
-0.12288850545883179,
0.023563409224152565,
-0.11131681501865387,
0.009110676124691963,
0.09452588856220245,
-0.11368808150291443,
0.2526414394378662,
-0.08788540959358215,
0.02584596909582615,
0.00981537252664566,
0.0994347333908081,
-0.0407090000808239,
0.10684086382389069,
-0.034308746457099915,
-0.13861458003520966,
0.003457661485299468,
-0.027347424998879433,
0.016104064881801605,
0.08792023360729218,
0.04607471078634262,
-0.07168617099523544,
0.039694417268037796,
0.05636611208319664,
0.07637510448694229,
-0.0311391893774271,
0.024612026289105415,
-0.14713114500045776,
0.035160452127456665,
-0.018204422667622566,
-0.051778316497802734,
0.04406355321407318,
-0.06300771236419678,
0.16269969940185547,
0.005446260329335928,
0.053564101457595825,
-0.014925586991012096,
-0.006913299672305584,
-0.05114142969250679,
-0.02387826330959797,
-0.0262321624904871,
-0.1398717314004898,
-0.0783119946718216,
-0.1376824975013733,
-0.05161554366350174,
-0.02936014160513878,
0.015776759013533592,
0.0064033218659460545,
0.011356758885085583,
-0.02943308651447296,
-0.048532918095588684,
0.005791729316115379,
0.12135656923055649,
-0.05286439508199692,
-0.03476916253566742
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# roberta-base-detect-cheapfake-ca1-ca2
This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.1482
- Accuracy: 0.94
- F1: 0.9450
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-06
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 5
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
| No log | 1.0 | 38 | 0.6724 | 0.705 | 0.7807 |
| No log | 2.0 | 76 | 0.5437 | 0.925 | 0.9309 |
| No log | 3.0 | 114 | 0.1945 | 0.93 | 0.9340 |
| No log | 4.0 | 152 | 0.1559 | 0.94 | 0.9444 |
| No log | 5.0 | 190 | 0.1482 | 0.94 | 0.9450 |
### Framework versions
- Transformers 4.36.2
- Pytorch 2.0.0
- Datasets 2.1.0
- Tokenizers 0.15.0
| {"license": "mit", "tags": ["generated_from_trainer"], "metrics": ["accuracy", "f1"], "base_model": "roberta-base", "model-index": [{"name": "roberta-base-detect-cheapfake-ca1-ca2", "results": []}]} | text-classification | hoanghoavienvo/roberta-base-detect-cheapfake-ca1-ca2 | [
"transformers",
"tensorboard",
"safetensors",
"roberta",
"text-classification",
"generated_from_trainer",
"base_model:roberta-base",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T09:32:30+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #roberta #text-classification #generated_from_trainer #base_model-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us
| roberta-base-detect-cheapfake-ca1-ca2
=====================================
This model is a fine-tuned version of roberta-base on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.1482
* Accuracy: 0.94
* F1: 0.9450
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-06
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 5
### Training results
### Framework versions
* Transformers 4.36.2
* Pytorch 2.0.0
* Datasets 2.1.0
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.36.2\n* Pytorch 2.0.0\n* Datasets 2.1.0\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #roberta #text-classification #generated_from_trainer #base_model-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.36.2\n* Pytorch 2.0.0\n* Datasets 2.1.0\n* Tokenizers 0.15.0"
] | [
63,
98,
4,
30
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #roberta #text-classification #generated_from_trainer #base_model-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5### Training results### Framework versions\n\n\n* Transformers 4.36.2\n* Pytorch 2.0.0\n* Datasets 2.1.0\n* Tokenizers 0.15.0"
] | [
-0.0872729942202568,
0.06608734279870987,
-0.0018402140121906996,
0.10135751217603683,
0.1675155758857727,
0.01357191614806652,
0.16046130657196045,
0.10870633274316788,
-0.10459641367197037,
0.03928138688206673,
0.12669305503368378,
0.15604309737682343,
-0.003485850291326642,
0.13376066088676453,
-0.07856188714504242,
-0.2434987872838974,
0.002600117353722453,
0.032796766608953476,
-0.08247584849596024,
0.11274389922618866,
0.10568597167730331,
-0.1359047293663025,
0.08753908425569534,
-0.009480123408138752,
-0.20837576687335968,
0.03743283823132515,
0.04446638748049736,
-0.0666339248418808,
0.1381738781929016,
0.04532018303871155,
0.13674187660217285,
0.0321464017033577,
0.08125346899032593,
-0.18660156428813934,
0.019100641831755638,
0.06102104112505913,
-0.015661925077438354,
0.08588720113039017,
0.047358542680740356,
-0.029367763549089432,
0.1105559691786766,
-0.0943804681301117,
0.06365340203046799,
0.022779615595936775,
-0.12993334233760834,
-0.20284292101860046,
-0.0707392543554306,
0.03404094651341438,
0.08907989412546158,
0.08137152343988419,
-0.019688066095113754,
0.15175043046474457,
-0.06305147707462311,
0.10196614265441895,
0.2043527066707611,
-0.30909472703933716,
-0.06914675980806351,
0.05690469965338707,
0.02571505308151245,
0.08627467602491379,
-0.11726520955562592,
0.003577222116291523,
0.08073630183935165,
0.020565267652273178,
0.12236417084932327,
-0.03657818213105202,
-0.06486528366804123,
0.012973707169294357,
-0.14006680250167847,
0.004747745115309954,
0.14671659469604492,
0.044876471161842346,
-0.04360802471637726,
-0.039720021188259125,
-0.05654003471136093,
-0.12649153172969818,
-0.04183842986822128,
-0.030490724369883537,
0.04317954182624817,
-0.028885353356599808,
-0.10950763523578644,
-0.01880469359457493,
-0.11704467982053757,
-0.07468824088573456,
-0.06339125335216522,
0.17693647742271423,
0.03373754024505615,
0.0017746867379173636,
-0.03287970647215843,
0.08995188772678375,
-0.041990894824266434,
-0.12596894800662994,
0.0173062514513731,
0.02494867704808712,
0.006079901475459337,
-0.07307393848896027,
-0.06701092422008514,
-0.09771792590618134,
0.030529692769050598,
0.14856252074241638,
-0.06796810030937195,
0.05014408007264137,
0.00719798170030117,
0.04309980943799019,
-0.09405310451984406,
0.16070455312728882,
-0.04462052509188652,
-0.01590851880609989,
0.02999020181596279,
0.055954620242118835,
0.03782695159316063,
-0.00259994063526392,
-0.12479730695486069,
0.018584761768579483,
0.12026433646678925,
0.013157907873392105,
-0.07674044370651245,
0.08831341564655304,
-0.037040259689092636,
0.004826853051781654,
0.012119737453758717,
-0.0926598310470581,
0.034076280891895294,
0.002792097395285964,
-0.061038196086883545,
-0.06735513359308243,
0.027568645775318146,
0.015938162803649902,
0.010338985361158848,
0.1059475913643837,
-0.09464085847139359,
0.01937321200966835,
-0.09371495246887207,
-0.1320083737373352,
0.007251252420246601,
-0.06348101794719696,
0.03772641345858574,
-0.1180994063615799,
-0.15753714740276337,
-0.016460901126265526,
0.04582826793193817,
-0.027392446994781494,
-0.01917022094130516,
-0.05669393762946129,
-0.08262669295072556,
0.012871086597442627,
-0.014798074960708618,
0.09638716280460358,
-0.055261868983507156,
0.0990075021982193,
0.06297668069601059,
0.06720347702503204,
-0.057597566395998,
0.033339887857437134,
-0.10316026955842972,
0.01398573536425829,
-0.21410338580608368,
0.02292998880147934,
-0.06044796481728554,
0.07861638814210892,
-0.07697700709104538,
-0.08058365434408188,
-0.00814063660800457,
0.025052273645997047,
0.07352124899625778,
0.09091745316982269,
-0.14545080065727234,
-0.0749581977725029,
0.16373999416828156,
-0.10211814939975739,
-0.13635163009166718,
0.11923616379499435,
-0.07076428830623627,
0.0707831084728241,
0.0753115564584732,
0.19543437659740448,
0.05862141773104668,
-0.08901738375425339,
0.004593905061483383,
-0.011637193150818348,
0.03875335305929184,
-0.04887599125504494,
0.056160349398851395,
0.007575794123113155,
-0.003982170484960079,
0.013295128010213375,
-0.020810376852750778,
0.04819609969854355,
-0.09241753816604614,
-0.07803653925657272,
-0.040532227605581284,
-0.10706616938114166,
0.05137501284480095,
0.06208209693431854,
0.08334828913211823,
-0.13340911269187927,
-0.08836691081523895,
0.101030632853508,
0.07163234800100327,
-0.06831535696983337,
0.020421959459781647,
-0.07921966165304184,
0.07088523358106613,
-0.06163021922111511,
-0.029186222702264786,
-0.16390518844127655,
-0.04778002202510834,
-0.003683502320200205,
0.03238211199641228,
0.03382107987999916,
0.013120820745825768,
0.08003579825162888,
0.07187844812870026,
-0.07590454816818237,
-0.025836393237113953,
0.0017989911139011383,
0.015454418025910854,
-0.1279689073562622,
-0.21018916368484497,
0.0002308078546775505,
-0.041370946913957596,
0.13068817555904388,
-0.25381556153297424,
0.05464904010295868,
0.008316610008478165,
0.08639609813690186,
0.04123053699731827,
0.0035423818044364452,
-0.04158870503306389,
0.07146783173084259,
-0.05270799621939659,
-0.0525406114757061,
0.05111796781420708,
0.00639673275873065,
-0.0889396071434021,
-0.04092465341091156,
-0.15823200345039368,
0.2003760188817978,
0.14426562190055847,
-0.11590725183486938,
-0.10827797651290894,
0.0013010026887059212,
-0.03934328258037567,
-0.021488802507519722,
-0.049736957997083664,
0.01555278617888689,
0.12489894777536392,
-0.027551330626010895,
0.154577374458313,
-0.073029525578022,
-0.03547552973031998,
0.021688709035515785,
-0.06440326571464539,
0.007946241647005081,
0.1072741374373436,
0.08657519519329071,
-0.1333802342414856,
0.15223877131938934,
0.14959391951560974,
-0.11022619158029556,
0.15195079147815704,
-0.03198954090476036,
-0.053637150675058365,
-0.01929483748972416,
0.0007391137769445777,
0.015339567326009274,
0.10865878313779831,
-0.11414466053247452,
-0.011621037498116493,
-0.0006649643764831126,
0.009101400151848793,
0.022595210000872612,
-0.21989135444164276,
-0.034414105117321014,
0.028739862143993378,
-0.032747987657785416,
0.015512684360146523,
-0.019877566024661064,
-0.00906267762184143,
0.10469471663236618,
-0.0031259881798177958,
-0.0806354209780693,
0.04326443374156952,
0.0021414505317807198,
-0.09153276681900024,
0.2232935130596161,
-0.0727478563785553,
-0.12019981443881989,
-0.129638209939003,
-0.062360215932130814,
-0.04560421034693718,
0.03414854779839516,
0.06840963661670685,
-0.0870562419295311,
-0.03956850245594978,
-0.09643110632896423,
0.007859847508370876,
0.020154552534222603,
0.03445969521999359,
-0.005109645426273346,
0.015471176244318485,
0.0801493301987648,
-0.10811728984117508,
-0.008116022683680058,
-0.05560486763715744,
-0.07294932752847672,
0.0441703274846077,
0.031075919046998024,
0.11682138592004776,
0.14345350861549377,
-0.04505481198430061,
-0.0011850347509607673,
-0.04478038102388382,
0.2206869125366211,
-0.06959537416696548,
-0.0185348242521286,
0.1205262616276741,
-0.01648399978876114,
0.03943542391061783,
0.13573230803012848,
0.06234784424304962,
-0.09766190499067307,
0.03393910452723503,
0.043939799070358276,
-0.03171302750706673,
-0.21824149787425995,
-0.029734496027231216,
-0.030140580609440804,
-0.012653365731239319,
0.08397942781448364,
0.04136360436677933,
0.05041944980621338,
0.07774405181407928,
0.03395861014723778,
0.06845755130052567,
-0.008551309816539288,
0.075651615858078,
0.10689719766378403,
0.04901786148548126,
0.13647863268852234,
-0.06008579581975937,
-0.07536575198173523,
0.0278773196041584,
-0.017337847501039505,
0.19217246770858765,
0.021955497562885284,
0.10969003289937973,
0.05575046315789223,
0.14343991875648499,
0.012412496842443943,
0.06629960983991623,
-0.0029386382084339857,
-0.06613443791866302,
-0.0023021649103611708,
-0.04609857499599457,
-0.012010710313916206,
0.04037316143512726,
-0.09931240975856781,
0.052487149834632874,
-0.1186591237783432,
0.014166528359055519,
0.06883341073989868,
0.20401787757873535,
0.05591077730059624,
-0.3292448818683624,
-0.09674517810344696,
0.025822315365076065,
-0.022533224895596504,
-0.018792448565363884,
0.024753617122769356,
0.11581360548734665,
-0.04625082388520241,
0.04303446784615517,
-0.06826519966125488,
0.07632776349782944,
-0.023244664072990417,
0.04460268095135689,
0.03938813507556915,
0.09966898709535599,
-0.028653578832745552,
0.06922908872365952,
-0.29875320196151733,
0.2818840444087982,
0.0218399316072464,
0.09036339074373245,
-0.045004621148109436,
-0.010589835233986378,
0.029896503314375877,
0.10234888643026352,
0.06287583708763123,
-0.030833205208182335,
-0.10516851395368576,
-0.1963602602481842,
-0.028252962976694107,
0.031008059158921242,
0.11110185086727142,
-0.016297252848744392,
0.10359761118888855,
-0.035132672637701035,
0.005094053689390421,
0.09670829772949219,
-0.03639379143714905,
-0.09239068627357483,
-0.08732052147388458,
-0.033539559692144394,
0.026768984273076057,
-0.029312903061509132,
-0.07860607653856277,
-0.10084111243486404,
-0.12790824472904205,
0.17554251849651337,
-0.030070176348090172,
-0.01597825065255165,
-0.10685541480779648,
0.08972755819559097,
0.03715252876281738,
-0.08516217768192291,
0.043380286544561386,
0.014188054949045181,
0.0847996175289154,
0.021965356543660164,
-0.06403587758541107,
0.135822132229805,
-0.06647633016109467,
-0.16262401640415192,
-0.06226722151041031,
0.09209457784891129,
0.01873127929866314,
0.0410546138882637,
0.006214599125087261,
0.014503015205264091,
-0.010714574716985226,
-0.07666449248790741,
0.03077417053282261,
-0.03508609160780907,
0.05713530257344246,
0.014647097326815128,
-0.06423678249120712,
-0.027519166469573975,
-0.06183774396777153,
-0.0325285904109478,
0.16661019623279572,
0.2905212938785553,
-0.0873231291770935,
-0.013666264712810516,
0.06713801622390747,
-0.067597396671772,
-0.2199885994195938,
0.07242331653833389,
0.015873469412326813,
0.000044437609176384285,
0.051032036542892456,
-0.1385280191898346,
0.12162170559167862,
0.09741615504026413,
-0.023189550265669823,
0.09330318868160248,
-0.2615320086479187,
-0.14632509648799896,
0.13902096450328827,
0.16806024312973022,
0.13617585599422455,
-0.16751323640346527,
-0.027928046882152557,
-0.04351743683218956,
-0.07248232513666153,
0.09983616322278976,
-0.15078981220722198,
0.11086457222700119,
0.0009276975761167705,
0.060362767428159714,
0.00645743403583765,
-0.057103145867586136,
0.11498405039310455,
-0.017754288390278816,
0.13009601831436157,
-0.06993412226438522,
-0.03729512169957161,
0.06652943044900894,
-0.04756797105073929,
0.019536137580871582,
-0.09769174456596375,
0.03154858201742172,
-0.03981928899884224,
-0.03232339769601822,
-0.05424419790506363,
0.05196463316679001,
-0.03609732910990715,
-0.07104159891605377,
-0.052110832184553146,
0.02983844093978405,
0.019142014905810356,
-0.019201911985874176,
0.17396201193332672,
-0.003315763082355261,
0.18386612832546234,
0.14374534785747528,
0.09199702739715576,
-0.061570554971694946,
0.010656214319169521,
0.013296184130012989,
-0.042659860104322433,
0.05921334773302078,
-0.16055232286453247,
0.039599910378456116,
0.11144288629293442,
0.011234069243073463,
0.15234598517417908,
0.08289815485477448,
-0.035302337259054184,
0.029049204662442207,
0.08850565552711487,
-0.1653183549642563,
-0.11043260991573334,
-0.0083167040720582,
-0.05925484001636505,
-0.10199432820081711,
0.08615361154079437,
0.12101855129003525,
-0.0775105431675911,
0.0008343553636223078,
-0.013820976950228214,
0.0011724523501470685,
-0.0505758672952652,
0.1845986247062683,
0.09260492026805878,
0.04503956437110901,
-0.07386721670627594,
0.07568490505218506,
0.04276810213923454,
-0.06992947310209274,
0.018670709803700447,
0.056341543793678284,
-0.07252013683319092,
-0.05313805863261223,
0.07453782856464386,
0.21586599946022034,
-0.05945160612463951,
-0.0537237823009491,
-0.16226504743099213,
-0.11159386485815048,
0.04425402730703354,
0.21474087238311768,
0.10014452785253525,
0.002039348240941763,
-0.02259734831750393,
0.026590479537844658,
-0.13955098390579224,
0.10729434341192245,
0.033762648701667786,
0.08855681121349335,
-0.15473198890686035,
0.165609672665596,
-0.005095246247947216,
0.009081152267754078,
-0.03446359932422638,
0.0403747633099556,
-0.1347254514694214,
-0.0012021787697449327,
-0.12387602031230927,
-0.015150503255426884,
-0.03458235040307045,
0.006460392381995916,
0.012344620190560818,
-0.062199950218200684,
-0.0753328949213028,
0.010721104219555855,
-0.10760463029146194,
-0.00609162263572216,
0.045288022607564926,
0.050721101462841034,
-0.1295875757932663,
-0.035884179174900055,
0.013588041067123413,
-0.05836851894855499,
0.06122823804616928,
0.01864800415933132,
0.020981483161449432,
0.06851623952388763,
-0.203209787607193,
0.03347880393266678,
0.0750625804066658,
-0.013645204715430737,
0.051553260535001755,
-0.07444515079259872,
-0.004776570945978165,
-0.000050774837291100994,
0.07476620376110077,
0.024769313633441925,
0.07402621954679489,
-0.12295974045991898,
0.02085626870393753,
-0.04257002845406532,
-0.0743708461523056,
-0.06189606338739395,
0.03659456595778465,
0.0803104117512703,
-0.01001213863492012,
0.20015501976013184,
-0.11446156352758408,
0.0045402473770082,
-0.20604464411735535,
0.011455217376351357,
-0.008141632191836834,
-0.12097478657960892,
-0.11917497962713242,
-0.052551522850990295,
0.05748990923166275,
-0.06384813040494919,
0.14290359616279602,
0.031459443271160126,
0.023912299424409866,
0.04692283272743225,
-0.04678844287991524,
0.040723253041505814,
0.039486441761255264,
0.23654697835445404,
0.02277066558599472,
-0.03889310732483864,
0.010829330421984196,
0.055599112063646317,
0.1193910613656044,
0.08069777488708496,
0.18367159366607666,
0.16482187807559967,
-0.08147301524877548,
0.11049287021160126,
0.0435648113489151,
-0.04400802403688431,
-0.1121605634689331,
0.05799002945423126,
-0.04138830304145813,
0.0718410313129425,
-0.02870337665081024,
0.16473762691020966,
0.13899612426757812,
-0.15143676102161407,
0.01114574819803238,
-0.06651344895362854,
-0.08113018423318863,
-0.1252189576625824,
-0.04638320952653885,
-0.11290040612220764,
-0.15452703833580017,
0.008271899074316025,
-0.11619820445775986,
0.005179854109883308,
0.08564002811908722,
0.008730563335120678,
-0.023509321734309196,
0.17945750057697296,
0.016412844881415367,
0.038688093423843384,
0.045998819172382355,
-0.0063506620936095715,
-0.034855376929044724,
-0.08197548985481262,
-0.08002108335494995,
0.0018360961694270372,
-0.032309189438819885,
0.011923772282898426,
-0.05191998556256294,
-0.036109525710344315,
0.04553884267807007,
-0.024049049243330956,
-0.10434871166944504,
0.016677705571055412,
0.04211825504899025,
0.056844595819711685,
0.03699883818626404,
0.01086212508380413,
-0.00400565704330802,
-0.00240694428794086,
0.23883016407489777,
-0.07434157282114029,
-0.07534404844045639,
-0.09767315536737442,
0.2590709626674652,
0.047828491777181625,
0.0312962606549263,
-0.00045436667278409004,
-0.10496097803115845,
0.03741767629981041,
0.224201962351799,
0.20938239991664886,
-0.06882165372371674,
0.014231069013476372,
-0.021472429856657982,
-0.009573426097631454,
-0.030119920149445534,
0.10048364102840424,
0.1027907133102417,
0.01580205373466015,
-0.0804869532585144,
-0.05050399526953697,
-0.031922128051519394,
-0.0024417266249656677,
-0.04772862419486046,
0.06414701789617538,
0.046321187168359756,
0.019289502874016762,
-0.04722324013710022,
0.06444154679775238,
-0.03670699894428253,
-0.11471628397703171,
0.05406592786312103,
-0.19506682455539703,
-0.14013375341892242,
-0.012731866911053658,
0.1118592768907547,
-0.013628995977342129,
0.06617546826601028,
-0.03359965234994888,
-0.006109967362135649,
0.02232666313648224,
-0.01840415969491005,
-0.076612189412117,
-0.06934218853712082,
0.06271838396787643,
-0.10443130880594254,
0.21802997589111328,
-0.045185964554548264,
0.039165448397397995,
0.1445290744304657,
0.03509252890944481,
-0.06933101266622543,
0.1030660942196846,
0.03716365993022919,
-0.07602421939373016,
0.04574407637119293,
0.09153630584478378,
-0.047046348452568054,
0.10535863041877747,
0.054263241589069366,
-0.14843639731407166,
0.03165264427661896,
-0.09596555680036545,
-0.08523643016815186,
-0.05946554243564606,
-0.03543872758746147,
-0.06455115228891373,
0.12446678429841995,
0.18663237988948822,
-0.029476214200258255,
0.022938188165426254,
-0.05215229094028473,
0.024238141253590584,
0.07770445197820663,
0.027831247076392174,
-0.04285553842782974,
-0.2513006329536438,
0.030174626037478447,
0.09305700659751892,
-0.016649393364787102,
-0.3053104877471924,
-0.0707617923617363,
-0.011026682332158089,
-0.04613291099667549,
-0.10008113831281662,
0.08647631853818893,
0.1403307020664215,
0.0571872778236866,
-0.0645022988319397,
-0.11901719868183136,
-0.07755094766616821,
0.1639108955860138,
-0.13391998410224915,
-0.10613751411437988
] |
null | null | transformers |
# BagelMIsteryTour-v2-8x7B 3.5bpw
Exllama quant of [ycros/BagelMIsteryTour-v2-8x7B](https://huggingface.co/ycros/BagelMIsteryTour-v2-8x7B)
## Other quants:
EXL2: [8bpw](https://huggingface.co/Kooten/BagelMIsteryTour-v2-8x7B-8bpw-exl2), [6bpw](https://huggingface.co/Kooten/BagelMIsteryTour-v2-8x7B-6bpw-exl2), [5bpw](https://huggingface.co/Kooten/BagelMIsteryTour-v2-8x7B-5bpw-exl2), [4bpw](https://huggingface.co/Kooten/BagelMIsteryTour-v2-8x7B-4bpw-exl2), [3.5bpw](https://huggingface.co/Kooten/BagelMIsteryTour-v2-8x7B-3.5bpw-exl2)
## Prompt format: Alpaca
It is noted to also work with mistral
```
Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
{prompt}
### Input:
{input}
### Response:
```
## Contact
Kooten on discord
[ko-fi.com/kooten](https://ko-fi.com/kooten) if you would like to support me
| {"license": "cc-by-nc-4.0", "tags": ["mergekit", "merge"], "base_model": ["mistralai/Mixtral-8x7B-v0.1", "jondurbin/bagel-dpo-8x7b-v0.2", "Sao10K/Sensualize-Mixtral-bf16", "mistralai/Mixtral-8x7B-v0.1", "Doctor-Shotgun/limarp-zloss-mixtral-8x7b-qlora", "mistralai/Mixtral-8x7B-Instruct-v0.1"]} | text-generation | Kooten/BagelMIsteryTour-v2-8x7B-3.5bpw-exl2 | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"mergekit",
"merge",
"base_model:mistralai/Mixtral-8x7B-v0.1",
"base_model:jondurbin/bagel-dpo-8x7b-v0.2",
"base_model:Sao10K/Sensualize-Mixtral-bf16",
"base_model:Doctor-Shotgun/limarp-zloss-mixtral-8x7b-qlora",
"base_model:mistralai/Mixtral-8x7B-Instruct-v0.1",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T09:32:43+00:00 | [] | [] | TAGS
#transformers #safetensors #mixtral #text-generation #mergekit #merge #base_model-mistralai/Mixtral-8x7B-v0.1 #base_model-jondurbin/bagel-dpo-8x7b-v0.2 #base_model-Sao10K/Sensualize-Mixtral-bf16 #base_model-Doctor-Shotgun/limarp-zloss-mixtral-8x7b-qlora #base_model-mistralai/Mixtral-8x7B-Instruct-v0.1 #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# BagelMIsteryTour-v2-8x7B 3.5bpw
Exllama quant of ycros/BagelMIsteryTour-v2-8x7B
## Other quants:
EXL2: 8bpw, 6bpw, 5bpw, 4bpw, 3.5bpw
## Prompt format: Alpaca
It is noted to also work with mistral
## Contact
Kooten on discord
URL if you would like to support me
| [
"# BagelMIsteryTour-v2-8x7B 3.5bpw\nExllama quant of ycros/BagelMIsteryTour-v2-8x7B",
"## Other quants:\n\nEXL2: 8bpw, 6bpw, 5bpw, 4bpw, 3.5bpw",
"## Prompt format: Alpaca\nIt is noted to also work with mistral",
"## Contact\nKooten on discord\n\nURL if you would like to support me"
] | [
"TAGS\n#transformers #safetensors #mixtral #text-generation #mergekit #merge #base_model-mistralai/Mixtral-8x7B-v0.1 #base_model-jondurbin/bagel-dpo-8x7b-v0.2 #base_model-Sao10K/Sensualize-Mixtral-bf16 #base_model-Doctor-Shotgun/limarp-zloss-mixtral-8x7b-qlora #base_model-mistralai/Mixtral-8x7B-Instruct-v0.1 #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# BagelMIsteryTour-v2-8x7B 3.5bpw\nExllama quant of ycros/BagelMIsteryTour-v2-8x7B",
"## Other quants:\n\nEXL2: 8bpw, 6bpw, 5bpw, 4bpw, 3.5bpw",
"## Prompt format: Alpaca\nIt is noted to also work with mistral",
"## Contact\nKooten on discord\n\nURL if you would like to support me"
] | [
175,
40,
33,
18,
14
] | [
"passage: TAGS\n#transformers #safetensors #mixtral #text-generation #mergekit #merge #base_model-mistralai/Mixtral-8x7B-v0.1 #base_model-jondurbin/bagel-dpo-8x7b-v0.2 #base_model-Sao10K/Sensualize-Mixtral-bf16 #base_model-Doctor-Shotgun/limarp-zloss-mixtral-8x7b-qlora #base_model-mistralai/Mixtral-8x7B-Instruct-v0.1 #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# BagelMIsteryTour-v2-8x7B 3.5bpw\nExllama quant of ycros/BagelMIsteryTour-v2-8x7B## Other quants:\n\nEXL2: 8bpw, 6bpw, 5bpw, 4bpw, 3.5bpw## Prompt format: Alpaca\nIt is noted to also work with mistral## Contact\nKooten on discord\n\nURL if you would like to support me"
] | [
-0.1190851479768753,
0.13947062194347382,
-0.007752615492790937,
0.03501691296696663,
0.03490631654858589,
0.03134389966726303,
0.15435156226158142,
0.145490363240242,
0.16907939314842224,
0.07131358236074448,
0.051510442048311234,
0.15619933605194092,
0.05291961133480072,
0.2056492567062378,
-0.060786809772253036,
-0.10177706927061081,
-0.005316614173352718,
-0.009599603712558746,
0.09977336972951889,
0.06079443171620369,
0.08028469234704971,
-0.01975790411233902,
0.05070866271853447,
-0.02699129842221737,
-0.0575375109910965,
0.03325030207633972,
-0.015494215302169323,
-0.03139106556773186,
0.06189454719424248,
0.08950452506542206,
0.050342705100774765,
0.09772851318120956,
-0.009449271485209465,
-0.21746398508548737,
0.045521330088377,
0.007450601551681757,
-0.06790539622306824,
0.0395759716629982,
0.08713237941265106,
-0.12241916358470917,
-0.008281568996608257,
-0.14164842665195465,
-0.015075878240168095,
0.07059335708618164,
-0.08959458768367767,
-0.042385563254356384,
-0.12589150667190552,
0.14484316110610962,
0.13651248812675476,
0.038601357489824295,
0.004887305200099945,
0.1330670267343521,
0.04703688248991966,
0.11232878267765045,
0.2638123035430908,
-0.2083314061164856,
-0.036783501505851746,
0.11494895070791245,
0.04469021037220955,
0.031389717012643814,
-0.024561380967497826,
-0.0024461825378239155,
0.034570492804050446,
-0.008301610127091408,
-0.028290553018450737,
-0.061542119830846786,
0.013656295835971832,
-0.07121404260396957,
-0.07055527716875076,
-0.018113279715180397,
0.12173096090555191,
0.05591757223010063,
-0.04900694265961647,
-0.020399894565343857,
-0.053165871649980545,
-0.007067680358886719,
-0.02666926197707653,
-0.003154065692797303,
-0.004737348295748234,
-0.02520059235394001,
0.06360836327075958,
-0.05180324614048004,
-0.1097107082605362,
-0.0228448286652565,
-0.030112728476524353,
0.12880301475524902,
0.017570611089468002,
-0.013614099472761154,
-0.011198592372238636,
0.04556238278746605,
-0.1565050184726715,
-0.1374962031841278,
-0.07527308166027069,
-0.06543140113353729,
-0.045938640832901,
-0.06774213165044785,
0.023810921236872673,
-0.04554722458124161,
0.11392442137002945,
0.2533191442489624,
-0.08792272210121155,
0.08430615067481995,
0.033900175243616104,
-0.0030815149657428265,
0.029801970347762108,
-0.01984269917011261,
-0.007593099027872086,
-0.09685937315225601,
0.10485661029815674,
0.07262561470270157,
0.059553563594818115,
0.01726331003010273,
-0.06797651201486588,
-0.08378062397241592,
0.019476819783449173,
0.07366395741701126,
0.042420435696840286,
0.020520202815532684,
-0.09741257131099701,
-0.033527128398418427,
0.31384509801864624,
-0.10099208354949951,
0.046174440532922745,
0.05269468575716019,
0.023452751338481903,
0.014515770599246025,
0.06987113505601883,
0.04029906541109085,
-0.018546050414443016,
0.023560035973787308,
-0.007392123341560364,
0.021772470325231552,
-0.024980714544653893,
-0.059580568224191666,
0.09751684218645096,
0.006878791842609644,
-0.011991254985332489,
-0.1741226613521576,
-0.1343708336353302,
-0.025822607800364494,
0.005942203104496002,
-0.04707321524620056,
-0.021886613219976425,
-0.014921668916940689,
-0.04786718636751175,
0.00677414471283555,
-0.002959706587716937,
-0.006854245904833078,
-0.05228128656744957,
-0.003860148834064603,
0.035651013255119324,
0.09380842000246048,
-0.06951967626810074,
-0.012613733299076557,
-0.004395985044538975,
0.03273467347025871,
-0.18367627263069153,
0.04632730036973953,
-0.16087870299816132,
0.027138663455843925,
-0.15184715390205383,
-0.002800851361826062,
0.019869821146130562,
-0.01915251649916172,
0.06376077979803085,
0.11458861082792282,
-0.11987829953432083,
-0.013674786314368248,
0.1513185352087021,
-0.08583010733127594,
-0.09036874771118164,
0.10086382925510406,
0.031985092908144,
-0.05049669370055199,
0.05969596654176712,
0.04931585118174553,
0.2094743549823761,
-0.13712731003761292,
-0.0988197773694992,
-0.02006506733596325,
0.027294471859931946,
-0.015675200149416924,
0.024473821744322777,
0.006403906736522913,
0.02421230636537075,
0.005044183693826199,
0.0787355974316597,
0.016555726528167725,
-0.009682012721896172,
-0.033563267439603806,
-0.029606971889734268,
-0.07135219871997833,
0.0878349170088768,
-0.03165009245276451,
-0.028744813054800034,
-0.05368926748633385,
-0.04048026353120804,
-0.015256278216838837,
0.10646090656518936,
0.031488578766584396,
-0.0002937323588412255,
-0.07360050827264786,
0.038819070905447006,
-0.044453829526901245,
0.0065826429054141045,
-0.14413626492023468,
-0.09218104928731918,
0.01907695084810257,
-0.08943008631467819,
0.002893761033192277,
-0.043832648545503616,
0.06246047094464302,
0.03402838110923767,
-0.029850680381059647,
-0.02629420906305313,
0.04160390421748161,
0.002057368401437998,
-0.07185456156730652,
-0.07298370450735092,
-0.01403091475367546,
-0.0844547376036644,
0.11393067985773087,
-0.11187591403722763,
0.07213741540908813,
0.10485874861478806,
0.1376824527978897,
0.007770800031721592,
-0.04392548277974129,
0.10263627767562866,
0.03540358319878578,
-0.0032948895823210478,
-0.038423363119363785,
0.05132591351866722,
-0.03557257354259491,
-0.05441340431571007,
0.03957422822713852,
-0.12847039103507996,
-0.024427063763141632,
0.06964103132486343,
0.12739630043506622,
-0.06814081221818924,
-0.07512622326612473,
-0.005602793302386999,
-0.016293413937091827,
0.025396211072802544,
-0.054662998765707016,
0.08284275233745575,
0.04082182049751282,
0.05452028289437294,
-0.04145534709095955,
-0.06806691735982895,
0.015199235640466213,
-0.0630299299955368,
-0.014607138931751251,
0.11598344147205353,
-0.016849713400006294,
-0.154046893119812,
0.05163105949759483,
0.12104885280132294,
-0.055801376700401306,
0.08629526942968369,
0.06973828375339508,
0.0020435957703739405,
-0.034812796860933304,
0.0780741348862648,
0.06165672093629837,
0.023038163781166077,
-0.05400802940130234,
0.027263309806585312,
0.04126685485243797,
-0.03417961671948433,
-0.004838391672819853,
-0.09864560514688492,
0.012683652341365814,
-0.01899822987616062,
-0.05229772627353668,
0.005053096450865269,
0.07026699930429459,
0.015261497348546982,
0.054417770355939865,
-0.025659089908003807,
0.06265994906425476,
-0.011518742889165878,
-0.010236904956400394,
-0.07716278731822968,
0.17767858505249023,
-0.09385330229997635,
-0.1440713256597519,
-0.10653191804885864,
-0.11167754977941513,
-0.07636164873838425,
-0.01832774467766285,
0.09430316090583801,
-0.05516485124826431,
-0.03662611171603203,
-0.06325724720954895,
-0.021721865981817245,
0.10353998094797134,
-0.02781379222869873,
0.03674120828509331,
-0.0020279092714190483,
0.03851492702960968,
-0.0959276407957077,
-0.037080761045217514,
0.05350968614220619,
-0.05539029836654663,
0.12005481123924255,
-0.01444521825760603,
0.06380121409893036,
0.06857100129127502,
0.027159471064805984,
-0.03532503545284271,
-0.015752941370010376,
0.20053987205028534,
-0.016043785959482193,
0.12284428626298904,
0.20715223252773285,
0.0034905243664979935,
0.11095614731311798,
0.15955327451229095,
0.03868190199136734,
-0.05155005678534508,
0.01787155121564865,
0.05844178795814514,
-0.015439698472619057,
-0.23804941773414612,
-0.04528844729065895,
-0.037758924067020416,
0.0201675146818161,
0.06193035468459129,
0.06142440810799599,
0.053396906703710556,
0.10930172353982925,
-0.0622178316116333,
0.005841190926730633,
0.051199160516262054,
0.11074990779161453,
0.0905621349811554,
0.0471690371632576,
0.07531347870826721,
-0.04230450093746185,
0.05666191875934601,
0.07869408279657364,
0.15664935111999512,
0.07001633942127228,
0.003559654112905264,
0.10149198770523071,
0.06425108015537262,
0.07955461740493774,
0.035027261823415756,
-0.0077985613606870174,
0.04449465125799179,
-0.004847305826842785,
-0.0007184291607700288,
-0.0897965207695961,
-0.00882652960717678,
0.035035137087106705,
-0.05164987966418266,
-0.012743319384753704,
-0.022569386288523674,
-0.010572192259132862,
0.04457226023077965,
0.11247031390666962,
0.07360147684812546,
-0.2243669033050537,
-0.045909520238637924,
0.07105901092290878,
0.0025263887364417315,
-0.018617328256368637,
-0.05522284656763077,
-0.0068267034366726875,
-0.08654778450727463,
0.15097130835056305,
-0.010683119297027588,
0.07695089280605316,
-0.059829484671354294,
-0.044369712471961975,
-0.03802577778697014,
0.13045240938663483,
-0.0164813119918108,
0.03248552605509758,
-0.1191922202706337,
0.14646926522254944,
0.039113402366638184,
0.05622156709432602,
-0.01986522600054741,
0.04116568714380264,
0.04752611368894577,
0.10053043067455292,
0.07654407620429993,
0.0071848127990961075,
-0.04865279048681259,
-0.13064604997634888,
-0.10012856870889664,
0.009965959936380386,
0.01827431656420231,
-0.038592904806137085,
0.06535130739212036,
0.06041587144136429,
-0.0910656526684761,
-0.06682880967855453,
0.030686860904097557,
-0.25202813744544983,
-0.09618566185235977,
0.06769689172506332,
0.09178810566663742,
0.0366457961499691,
-0.054112572222948074,
-0.04670417681336403,
-0.06843306124210358,
0.1629396378993988,
-0.04748395457863808,
-0.10633960366249084,
-0.09681916981935501,
-0.05159401521086693,
0.07444734126329422,
-0.11945857107639313,
0.03277973458170891,
-0.04296961426734924,
0.10311221331357956,
-0.0983290821313858,
-0.06940621137619019,
0.030286259949207306,
-0.11813749372959137,
-0.12905536592006683,
-0.028083669021725655,
0.16259515285491943,
-0.01086143497377634,
-0.014064584858715534,
0.0440824031829834,
-0.035679254680871964,
0.03912372887134552,
-0.10289360582828522,
0.03938450291752815,
0.1650070697069168,
0.09176803380250931,
0.01651766337454319,
-0.031000005081295967,
-0.18462297320365906,
-0.09175640344619751,
0.004225586540997028,
0.0893721729516983,
0.40822407603263855,
-0.009249869734048843,
0.07029532641172409,
0.1184024065732956,
-0.06810539960861206,
-0.12631404399871826,
-0.06363508850336075,
0.08175979554653168,
-0.028911277651786804,
0.04565352573990822,
-0.08924425393342972,
0.006060135085135698,
0.12066377699375153,
-0.016309360042214394,
0.04917464405298233,
-0.18275544047355652,
-0.13629984855651855,
-0.0295953918248415,
0.08670442551374435,
0.04897001385688782,
-0.14762108027935028,
-0.07877302169799805,
-0.08402642607688904,
-0.294161856174469,
0.05865649878978729,
0.0037763442378491163,
0.06813958287239075,
-0.033482469618320465,
0.062161024659872055,
0.001730724354274571,
-0.05688653513789177,
0.14110487699508667,
0.020945630967617035,
0.000590726500377059,
-0.06545686721801758,
-0.024417228996753693,
0.034494269639253616,
-0.09531258046627045,
0.06838959455490112,
-0.08880466222763062,
0.021549709141254425,
-0.12354595214128494,
-0.010198074392974377,
-0.07368674874305725,
0.05958930775523186,
-0.0443529449403286,
-0.008460861630737782,
0.01709100417792797,
0.028672311455011368,
0.022027842700481415,
-0.021677883341908455,
0.0748208686709404,
-0.06015373393893242,
0.08380964398384094,
0.21373425424098969,
0.09432922303676605,
-0.06603764742612839,
-0.017398107796907425,
-0.004704915918409824,
-0.09869742393493652,
0.0017961710691452026,
-0.14636342227458954,
0.03363120183348656,
0.05301584675908089,
-0.002160332864150405,
0.07997234165668488,
-0.0181402787566185,
-0.08269183337688446,
0.060775551944971085,
0.08900110423564911,
-0.09101257473230362,
-0.04277852177619934,
0.00648566335439682,
0.11613406240940094,
-0.08570155501365662,
0.034778889268636703,
0.201448455452919,
0.003617453621700406,
-0.07787193357944489,
0.017065754160284996,
0.016582798212766647,
-0.01536513026803732,
0.1199299767613411,
0.029437700286507607,
0.07038172334432602,
-0.03696863353252411,
0.0434933565557003,
-0.000036857061786577106,
-0.055104903876781464,
-0.003055160865187645,
0.11865422129631042,
-0.10797447711229324,
-0.12489762902259827,
-0.020175321027636528,
0.009695978835225105,
-0.061186421662569046,
0.0030594770796597004,
-0.0032128472812473774,
-0.00959816388785839,
0.005970763973891735,
0.18083934485912323,
0.016814256086945534,
-0.06265324354171753,
0.11346794664859772,
-0.013120883144438267,
0.001462024636566639,
0.08930797874927521,
0.05225579813122749,
0.1125272586941719,
-0.101357102394104,
-0.020692650228738785,
-0.06507066637277603,
0.034139230847358704,
0.00947322603315115,
0.021219095215201378,
-0.141306534409523,
-0.06404881924390793,
-0.16667962074279785,
0.06736419349908829,
-0.1727929264307022,
0.011791946366429329,
-0.045733630657196045,
-0.03782228007912636,
-0.0056375134736299515,
0.024943068623542786,
-0.06568022072315216,
-0.05649145320057869,
-0.044370926916599274,
0.0934201255440712,
-0.11029262840747833,
-0.05635194480419159,
0.03640098124742508,
-0.05833571404218674,
0.08764401823282242,
0.016356244683265686,
-0.0183307733386755,
-0.023149946704506874,
-0.2354959100484848,
-0.013239405117928982,
0.07531566917896271,
0.04084651917219162,
0.03221328556537628,
-0.08976274728775024,
-0.004717310890555382,
-0.027930697426199913,
0.003982697147876024,
0.005772626493126154,
0.0825653001666069,
-0.08655782043933868,
-0.0017925339052453637,
-0.04785233363509178,
-0.045606959611177444,
-0.056123752146959305,
0.00008552311192033812,
0.16274069249629974,
-0.009495286270976067,
0.08340559154748917,
-0.0798686221241951,
0.05256979167461395,
-0.1313149780035019,
-0.008829736150801182,
0.0016511721769347787,
-0.07219443470239639,
-0.029048437252640724,
-0.06845339387655258,
0.058175474405288696,
0.01775122433900833,
0.013978447765111923,
-0.10490541160106659,
0.0638447031378746,
0.014434555545449257,
-0.08156383037567139,
0.03335452824831009,
0.014731752686202526,
0.0450281947851181,
-0.008553536608815193,
0.000026302885089535266,
-0.04084122180938721,
0.0463283471763134,
0.017783135175704956,
-0.11472668498754501,
0.05116623267531395,
0.12127363681793213,
0.009535993449389935,
0.06276055425405502,
0.058726850897073746,
-0.017026664689183235,
-0.023589812219142914,
-0.03536336496472359,
-0.07824703305959702,
0.04065033793449402,
-0.08873223513364792,
-0.039584580808877945,
0.13028588891029358,
-0.09166164696216583,
0.010461642406880856,
-0.006741394754499197,
-0.038777802139520645,
-0.08093348890542984,
-0.17669686675071716,
-0.08689405024051666,
-0.08094240725040436,
0.006703532300889492,
-0.0814087763428688,
0.054717227816581726,
-0.01839926838874817,
0.015747010707855225,
0.04632148519158363,
0.10657816380262375,
-0.009315615519881248,
-0.013067835010588169,
0.025368500500917435,
-0.009204969741404057,
-0.0019980433862656355,
0.06083877384662628,
0.014353582635521889,
0.07905486226081848,
-0.022283567115664482,
0.00006458692951127887,
0.06665746867656708,
0.004992975853383541,
0.025385938584804535,
-0.06210072711110115,
-0.11528515070676804,
-0.00921427272260189,
0.06338485330343246,
0.028278112411499023,
0.11225852370262146,
0.07285372167825699,
-0.0068434434942901134,
0.03547145798802376,
0.13603757321834564,
-0.006381679326295853,
-0.09800704568624496,
-0.10844117403030396,
0.07950974255800247,
-0.01540432870388031,
0.050772637128829956,
-0.03336021676659584,
-0.0609499029815197,
0.057834770530462265,
0.07319273799657822,
0.06218266859650612,
0.03866194188594818,
0.0526912584900856,
-0.012094635516405106,
0.017430169507861137,
-0.00610814755782485,
0.05630181357264519,
0.07859960943460464,
0.0887664258480072,
-0.03319748118519783,
0.03171504661440849,
0.024029485881328583,
-0.009548964910209179,
-0.1240968331694603,
-0.026173995807766914,
-0.04650915414094925,
-0.03061109408736229,
0.004802650772035122,
0.12210208922624588,
-0.015549290925264359,
-0.05939660966396332,
0.008253296837210655,
-0.11494598537683487,
-0.10677610337734222,
0.002028942573815584,
0.08376051485538483,
0.027873454615473747,
0.06251518428325653,
-0.024643458425998688,
-0.04355239495635033,
0.15256884694099426,
-0.04850272089242935,
-0.0644543245434761,
-0.01850837655365467,
-0.013854894787073135,
-0.04570400342345238,
0.08849415928125381,
-0.011677567847073078,
0.12125585228204727,
0.13513579964637756,
-0.036585237830877304,
-0.1518787294626236,
0.07227111607789993,
0.05152200162410736,
-0.11875662952661514,
0.07898745685815811,
0.12828023731708527,
-0.031816110014915466,
0.03856617957353592,
0.06046474725008011,
-0.08280228078365326,
0.012370028533041477,
0.06605828553438187,
0.03515943884849548,
-0.07597776502370834,
0.13494855165481567,
-0.08900074660778046,
0.08932466059923172,
0.06318332254886627,
-0.04017160087823868,
-0.007409932091832161,
-0.0026000996585935354,
0.06301441043615341,
0.04298219457268715,
0.03792640566825867,
-0.08533037453889847,
-0.15573129057884216,
0.06413096189498901,
-0.00007507568807341158,
0.11839533597230911,
-0.10857681185007095,
-0.04994645342230797,
-0.07205857336521149,
0.015591910108923912,
-0.08492724597454071,
0.09073789417743683,
0.08838158845901489,
-0.005828432273119688,
-0.015284237451851368,
-0.2658640742301941,
-0.032449644058942795,
0.1264866441488266,
-0.12358944118022919,
-0.08615141361951828
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | DrishtiSharma/mixtral-8x7b-v0.1-english-to-hinglish-translation-merged | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"4-bit",
"region:us"
] | 2024-02-07T09:34:40+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #mixtral #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #mixtral #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
59,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #mixtral #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.0492100790143013,
0.1616126149892807,
-0.00528301764279604,
0.02184453047811985,
0.09754197299480438,
0.01594017818570137,
0.07085862755775452,
0.11018063873052597,
-0.023524558171629906,
0.11237436532974243,
0.03310580179095268,
0.0959877073764801,
0.11302800476551056,
0.1511073112487793,
-0.004449901171028614,
-0.23006680607795715,
0.04969857260584831,
-0.12401237338781357,
-0.039184633642435074,
0.1161055639386177,
0.15064948797225952,
-0.10119860619306564,
0.07616360485553741,
-0.030171431601047516,
-0.00975856464356184,
-0.03212185949087143,
-0.05505085736513138,
-0.04940180853009224,
0.050669461488723755,
0.07503624260425568,
0.06824079900979996,
0.0047172619961202145,
0.09358154237270355,
-0.2653830349445343,
0.019896307960152626,
0.07278038561344147,
-0.0028079862240701914,
0.0752972885966301,
0.05440441146492958,
-0.07567562162876129,
0.0916486456990242,
-0.051031939685344696,
0.14619337022304535,
0.08076590299606323,
-0.09209886193275452,
-0.19179949164390564,
-0.0892757773399353,
0.10176365077495575,
0.18317750096321106,
0.0450831800699234,
-0.02234533429145813,
0.1010935828089714,
-0.08625087887048721,
0.01156448945403099,
0.05020461603999138,
-0.06867926567792892,
-0.052296232432127,
0.06399383395910263,
0.07969621568918228,
0.07660703361034393,
-0.12267867475748062,
-0.02157854288816452,
0.008834225125610828,
0.008101158775389194,
0.08239422738552094,
0.023205358535051346,
0.15234412252902985,
0.03851804882287979,
-0.12687987089157104,
-0.0488545261323452,
0.10819673538208008,
0.04200201481580734,
-0.04793836548924446,
-0.25112876296043396,
-0.030252739787101746,
-0.02640678361058235,
-0.02967509999871254,
-0.03939273953437805,
0.0413043349981308,
-0.00700207008048892,
0.0819111093878746,
-0.007696285843849182,
-0.07608349621295929,
-0.03789140284061432,
0.06117839738726616,
0.062258850783109665,
0.02590334229171276,
-0.011504736728966236,
0.01121334545314312,
0.11612965911626816,
0.10705047845840454,
-0.12408649176359177,
-0.050495926290750504,
-0.06463953107595444,
-0.07856319844722748,
-0.04267695173621178,
0.03387995809316635,
0.036278098821640015,
0.04898930713534355,
0.24677793681621552,
0.014574935659766197,
0.05503120273351669,
0.03896442800760269,
0.010167885571718216,
0.06416473537683487,
0.11215386539697647,
-0.06007103994488716,
-0.09715575724840164,
-0.025286169722676277,
0.08993930369615555,
0.011564950458705425,
-0.040237728506326675,
-0.05735642835497856,
0.06267821043729782,
0.016649754717946053,
0.118519626557827,
0.09037825465202332,
0.004811561666429043,
-0.07140906900167465,
-0.06490059196949005,
0.19362327456474304,
-0.1627010852098465,
0.048158012330532074,
0.03471692278981209,
-0.03803423419594765,
-0.00322781503200531,
0.009035434573888779,
0.024208955466747284,
-0.02166360430419445,
0.09028356522321701,
-0.056031733751297,
-0.04080629721283913,
-0.10969307273626328,
-0.037363357841968536,
0.031433332711458206,
0.010416633449494839,
-0.03290047124028206,
-0.03206469491124153,
-0.08352182805538177,
-0.0684375911951065,
0.09538353979587555,
-0.07277081906795502,
-0.05345873162150383,
-0.016413625329732895,
-0.07309456169605255,
0.025948043912649155,
0.020464899018406868,
0.0749586746096611,
-0.020128712058067322,
0.0424746572971344,
-0.055092595517635345,
0.05906378850340843,
0.10813204944133759,
0.033792946487665176,
-0.05388450250029564,
0.06219671294093132,
-0.2423417717218399,
0.10000433027744293,
-0.06794700026512146,
0.05244508385658264,
-0.1503486931324005,
-0.025011396035552025,
0.049166299402713776,
0.007808702997863293,
-0.010501512326300144,
0.13688668608665466,
-0.21851418912410736,
-0.028239823877811432,
0.16161467134952545,
-0.09608794748783112,
-0.07688594609498978,
0.0607018806040287,
-0.05308416485786438,
0.10767868906259537,
0.04049059376120567,
-0.026768529787659645,
0.06491570919752121,
-0.1332433819770813,
0.0018461219733580947,
-0.046419549733400345,
-0.019821418449282646,
0.15740364789962769,
0.07538620382547379,
-0.06907296180725098,
0.07017243653535843,
0.023979974910616875,
-0.022804807871580124,
-0.045258816331624985,
-0.01728796400129795,
-0.10864134877920151,
0.010852303355932236,
-0.06095534935593605,
0.02172400988638401,
-0.025193115696310997,
-0.09209749847650528,
-0.029975716024637222,
-0.17412637174129486,
-0.018685562536120415,
0.08384992182254791,
-0.009359767660498619,
-0.01940089836716652,
-0.11758141964673996,
0.014672953635454178,
0.03531273826956749,
0.002678109332919121,
-0.131710946559906,
-0.04835474118590355,
0.027603471651673317,
-0.1615411788225174,
0.03370620682835579,
-0.053489942103624344,
0.04953266307711601,
0.03157234191894531,
-0.032672569155693054,
-0.02871638908982277,
0.022493019700050354,
0.005167742725461721,
-0.013918960466980934,
-0.24545986950397491,
-0.02644311636686325,
-0.021760500967502594,
0.16886062920093536,
-0.21524107456207275,
0.037636447697877884,
0.07033006846904755,
0.15214523673057556,
0.00896006915718317,
-0.03784041851758957,
0.0021138733718544245,
-0.07351448386907578,
-0.03246096521615982,
-0.059393927454948425,
-0.008201966993510723,
-0.03462608531117439,
-0.056801993399858475,
0.04834989085793495,
-0.16736289858818054,
-0.028562728315591812,
0.10108717530965805,
0.06430494040250778,
-0.1370258778333664,
-0.02137446030974388,
-0.03499000519514084,
-0.042920805513858795,
-0.05433731526136398,
-0.058740075677633286,
0.10303318500518799,
0.05731228366494179,
0.046605419367551804,
-0.06478330492973328,
-0.07431352138519287,
0.0016415019053965807,
-0.019277552142739296,
-0.02324511855840683,
0.09063004702329636,
0.0750734880566597,
-0.12645095586776733,
0.092108815908432,
0.10475517809391022,
0.08614666759967804,
0.09744621813297272,
-0.023962685838341713,
-0.08241596817970276,
-0.05041659250855446,
0.02309448830783367,
0.01756109483540058,
0.133294478058815,
-0.007595307193696499,
0.05496160313487053,
0.04139304906129837,
-0.013294490985572338,
0.009913320653140545,
-0.09164989739656448,
0.03289409726858139,
0.03452005237340927,
-0.018637165427207947,
0.03988218680024147,
-0.039949383586645126,
0.021179918199777603,
0.08961519598960876,
0.04653824120759964,
0.04183639958500862,
0.014902578666806221,
-0.04666156321763992,
-0.11237750202417374,
0.16579076647758484,
-0.12958046793937683,
-0.23559744656085968,
-0.14693203568458557,
0.002910082694143057,
0.034562259912490845,
-0.009846248663961887,
0.0020482465624809265,
-0.06336327642202377,
-0.11868021637201309,
-0.09202129393815994,
0.011571505106985569,
0.04891825467348099,
-0.0853651687502861,
-0.06010903790593147,
0.05675940960645676,
0.04150831699371338,
-0.14567627012729645,
0.01958550699055195,
0.04915129393339157,
-0.09044831991195679,
-0.008420263417065144,
0.08169661462306976,
0.06742022186517715,
0.17982420325279236,
0.012057109735906124,
-0.022524503991007805,
0.03320202976465225,
0.21853268146514893,
-0.13450473546981812,
0.1139737144112587,
0.14119268953800201,
-0.09259282797574997,
0.08299964666366577,
0.1992565095424652,
0.04091860353946686,
-0.1011200025677681,
0.03255503997206688,
0.017553923651576042,
-0.029480691999197006,
-0.2430734485387802,
-0.07210355252027512,
-0.001007809885777533,
-0.05961545184254646,
0.0748797133564949,
0.08922994881868362,
0.08997989445924759,
0.01629791595041752,
-0.09482269734144211,
-0.07852915674448013,
0.05403036251664162,
0.1029631718993187,
0.021838568150997162,
-0.012486564926803112,
0.09075652807950974,
-0.033757518976926804,
0.017037617042660713,
0.09076420217752457,
0.0003606044629123062,
0.17206108570098877,
0.05808206647634506,
0.1838475614786148,
0.07610926777124405,
0.07362144440412521,
0.014698545448482037,
0.0128402728587389,
0.01911146566271782,
0.027121074497699738,
-0.0047050705179572105,
-0.08515863120555878,
-0.01687251217663288,
0.11894010007381439,
0.07453732192516327,
0.018218141049146652,
0.014295568689703941,
-0.03986431658267975,
0.08226090669631958,
0.17472510039806366,
-0.0028857914730906487,
-0.18157273530960083,
-0.06367392092943192,
0.0815940648317337,
-0.0922710970044136,
-0.10117124021053314,
-0.02492337115108967,
0.03145772963762283,
-0.17042423784732819,
0.024981264024972916,
-0.017380358651280403,
0.11211993545293808,
-0.13744667172431946,
-0.01997678168118,
0.06434136629104614,
0.07104871422052383,
-0.00027726442203857005,
0.0585814043879509,
-0.1615411937236786,
0.10371403396129608,
0.01322200708091259,
0.06754657626152039,
-0.09606750309467316,
0.10096679627895355,
-0.006538857705891132,
-0.011478343978524208,
0.13019709289073944,
0.008899516426026821,
-0.07355503737926483,
-0.08015985041856766,
-0.09150464087724686,
-0.009347239509224892,
0.1269414722919464,
-0.146452397108078,
0.0856066346168518,
-0.03641771897673607,
-0.04169904813170433,
0.002293013734742999,
-0.10595564544200897,
-0.12382242828607559,
-0.18823674321174622,
0.05475292727351189,
-0.13192032277584076,
0.03842078521847725,
-0.10651033371686935,
-0.03452010452747345,
-0.028951672837138176,
0.18889638781547546,
-0.22988250851631165,
-0.06845009326934814,
-0.15192745625972748,
-0.09812023490667343,
0.14535494148731232,
-0.04983055964112282,
0.08612044155597687,
-0.0064140805043280125,
0.17973676323890686,
0.02110406942665577,
-0.02243823930621147,
0.09821193665266037,
-0.09346702694892883,
-0.19534897804260254,
-0.0819322019815445,
0.15730807185173035,
0.13375797867774963,
0.03536279499530792,
-0.0027026666793972254,
0.03800923004746437,
-0.018153997138142586,
-0.12196674942970276,
0.021237321197986603,
0.17568714916706085,
0.06474101543426514,
0.025508904829621315,
-0.02745569683611393,
-0.11437032371759415,
-0.06745380908250809,
-0.0337534137070179,
0.032285626977682114,
0.1848049759864807,
-0.07229581475257874,
0.18418607115745544,
0.1440749317407608,
-0.0586681142449379,
-0.19688928127288818,
0.009774520993232727,
0.03473912551999092,
0.00748026929795742,
0.03411683440208435,
-0.2014101892709732,
0.08306154608726501,
-0.0006713230977766216,
-0.0508250892162323,
0.13138055801391602,
-0.1712750345468521,
-0.15031227469444275,
0.0726422443985939,
0.03674598038196564,
-0.1927690953016281,
-0.11939278244972229,
-0.08918614685535431,
-0.05463467538356781,
-0.18232925236225128,
0.10136403888463974,
0.0357523187994957,
0.008355717174708843,
0.03322768583893776,
0.02632942982017994,
0.016507811844348907,
-0.03891366347670555,
0.1913088858127594,
-0.02729160711169243,
0.031074488535523415,
-0.08418911695480347,
-0.06973641365766525,
0.047041334211826324,
-0.05448983237147331,
0.07591085880994797,
-0.027936331927776337,
0.011565293185412884,
-0.10452011227607727,
-0.04247993230819702,
-0.028639385476708412,
0.014135988429188728,
-0.09575077891349792,
-0.08887862414121628,
-0.0492461696267128,
0.0931631326675415,
0.09460635483264923,
-0.03684133663773537,
-0.036155372858047485,
-0.07004832476377487,
0.04206852987408638,
0.18663634359836578,
0.17629489302635193,
0.04319917783141136,
-0.07922680675983429,
-0.004305878188461065,
-0.011979347094893456,
0.044442515820264816,
-0.21414349973201752,
0.06505125761032104,
0.04941818490624428,
0.017324553802609444,
0.11970899999141693,
-0.019832991063594818,
-0.1546693593263626,
-0.06965133547782898,
0.06436776369810104,
-0.06119891628623009,
-0.19471220672130585,
0.005106652621179819,
0.05302131548523903,
-0.16849607229232788,
-0.04845238849520683,
0.04433823749423027,
-0.00405395170673728,
-0.03977474570274353,
0.018406251445412636,
0.08913438022136688,
0.0035062681417912245,
0.07001189142465591,
0.0575016625225544,
0.08202318102121353,
-0.10269234329462051,
0.07468173652887344,
0.0864284485578537,
-0.0808253362774849,
0.02600708231329918,
0.09163770824670792,
-0.059018705040216446,
-0.029858214780688286,
0.02767745405435562,
0.08199155330657959,
0.013627272099256516,
-0.04234054312109947,
0.011110236868262291,
-0.10469777137041092,
0.061493463814258575,
0.08624710887670517,
0.0332154780626297,
0.014785230159759521,
0.0319269597530365,
0.04486626386642456,
-0.07112183421850204,
0.12188448756933212,
0.028504198417067528,
0.015971386805176735,
-0.04059506952762604,
-0.0482974611222744,
0.02420179732143879,
-0.031129658222198486,
-0.006369255483150482,
-0.034713663160800934,
-0.07348191738128662,
-0.018021773546934128,
-0.16598404943943024,
-0.01642499305307865,
-0.047007784247398376,
0.010076508857309818,
0.031086590141057968,
-0.037944380193948746,
0.00736458133906126,
0.007804956287145615,
-0.07429007440805435,
-0.06553912162780762,
-0.022457081824541092,
0.09357618540525436,
-0.16298171877861023,
0.023078840225934982,
0.0868900716304779,
-0.12058180570602417,
0.09344310313463211,
0.019041698426008224,
-0.004308692645281553,
0.0283166766166687,
-0.14885661005973816,
0.03721340745687485,
-0.030741490423679352,
0.014877090230584145,
0.043578531593084335,
-0.22510670125484467,
0.0016996375052258372,
-0.033317502588033676,
-0.06376626342535019,
-0.008822662755846977,
-0.03464772179722786,
-0.11372490227222443,
0.10571323335170746,
0.006405833642929792,
-0.0893215611577034,
-0.031507380306720734,
0.031805768609046936,
0.08337265253067017,
-0.023627644404768944,
0.15937373042106628,
-0.003927427344024181,
0.0734739750623703,
-0.16826482117176056,
-0.020029379054903984,
-0.010388867929577827,
0.01968802511692047,
-0.021281469613313675,
-0.012884029187262058,
0.040600989013910294,
-0.022154994308948517,
0.1838352531194687,
-0.023590071126818657,
0.021040821447968483,
0.06630481034517288,
0.027481740340590477,
-0.02683531865477562,
0.10472780466079712,
0.0544922761619091,
0.021287893876433372,
0.01933317631483078,
0.002721484052017331,
-0.04222280904650688,
-0.02439267747104168,
-0.1995706409215927,
0.0638027936220169,
0.14158593118190765,
0.09051596373319626,
-0.017823107540607452,
0.08179465681314468,
-0.09895666688680649,
-0.11199329793453217,
0.11719910800457001,
-0.0537027008831501,
-0.0035598385147750378,
-0.0661843940615654,
0.12944552302360535,
0.1438235193490982,
-0.191106915473938,
0.07123035937547684,
-0.06875003129243851,
-0.04897533357143402,
-0.11610150337219238,
-0.19341503083705902,
-0.05786462128162384,
-0.05326216295361519,
-0.01650206930935383,
-0.04736659303307533,
0.0741693526506424,
0.054829590022563934,
0.007494221441447735,
-0.001470523769967258,
0.06295381486415863,
-0.02694425918161869,
-0.0002717367315199226,
0.0268293097615242,
0.06557910144329071,
0.012699071317911148,
-0.029256848618388176,
0.017130669206380844,
-0.010868940502405167,
0.041691455990076065,
0.06406819820404053,
0.04732465744018555,
-0.02999637834727764,
0.01705419085919857,
-0.03900563344359398,
-0.107493095099926,
0.0420064739882946,
-0.02748468704521656,
-0.08087874203920364,
0.1494218111038208,
0.024370772764086723,
0.00888568814843893,
-0.020774761214852333,
0.2394784688949585,
-0.07282441854476929,
-0.09617346525192261,
-0.14871551096439362,
0.10565752536058426,
-0.04225243255496025,
0.06315739452838898,
0.0457107312977314,
-0.10360732674598694,
0.016461793333292007,
0.12599775195121765,
0.16596394777297974,
-0.04424665868282318,
0.020934030413627625,
0.027821561321616173,
0.003930688835680485,
-0.03569735586643219,
0.05258714035153389,
0.06921107321977615,
0.15732763707637787,
-0.04788297414779663,
0.09735256433486938,
-0.002974086906760931,
-0.09634213149547577,
-0.03713185712695122,
0.11821570992469788,
-0.0173256304115057,
0.01732185296714306,
-0.054686542600393295,
0.11868327111005783,
-0.05951813980937004,
-0.22993476688861847,
0.06008617579936981,
-0.06702527403831482,
-0.1364361047744751,
-0.02199983038008213,
0.08053575456142426,
-0.013966009952127934,
0.02633519284427166,
0.0725690945982933,
-0.0749802514910698,
0.19878365099430084,
0.036572057753801346,
-0.05435766279697418,
-0.05330595746636391,
0.08271016925573349,
-0.10416609048843384,
0.27322450280189514,
0.01728520356118679,
0.050410788506269455,
0.10306239128112793,
-0.011722724884748459,
-0.1334574669599533,
0.021906433627009392,
0.09504044055938721,
-0.09268639236688614,
0.04262949526309967,
0.19755758345127106,
0.00013492791913449764,
0.12123726308345795,
0.07934848219156265,
-0.07633684575557709,
0.05047796294093132,
-0.09577839076519012,
-0.07217153161764145,
-0.08968814462423325,
0.09568040817975998,
-0.07660859823226929,
0.1432706117630005,
0.1311088353395462,
-0.052900709211826324,
0.010313699953258038,
-0.029171356931328773,
0.046702686697244644,
0.004070387687534094,
0.101396843791008,
0.009152291342616081,
-0.18525077402591705,
0.0214304830878973,
0.012694042176008224,
0.10595077276229858,
-0.16679531335830688,
-0.0989961102604866,
0.040641963481903076,
0.001496727461926639,
-0.06019711121916771,
0.12934371829032898,
0.06018439307808876,
0.04527871683239937,
-0.04309406131505966,
-0.023274151608347893,
-0.010192750953137875,
0.13592426478862762,
-0.10134663432836533,
0.0017774304142221808
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | unsloth/yi-34b-bnb-4bit | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"4-bit",
"region:us"
] | 2024-02-07T09:34:43+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
59,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.049007222056388855,
0.16460949182510376,
-0.005271392408758402,
0.021910345181822777,
0.09685911983251572,
0.01403510570526123,
0.07018975168466568,
0.11002060770988464,
-0.02425350993871689,
0.11399492621421814,
0.03344893455505371,
0.09780009090900421,
0.11368958652019501,
0.1498505026102066,
-0.002398149576038122,
-0.23227156698703766,
0.04924226179718971,
-0.1249755248427391,
-0.03746527433395386,
0.1159956082701683,
0.15001481771469116,
-0.10170940309762955,
0.07611104100942612,
-0.029819702729582787,
-0.008722295984625816,
-0.032589927315711975,
-0.056551046669483185,
-0.04997202008962631,
0.051094699651002884,
0.07382578402757645,
0.06793182343244553,
0.004094683099538088,
0.09450557827949524,
-0.2669448256492615,
0.0197003111243248,
0.0730973482131958,
-0.002068581758067012,
0.07547242939472198,
0.054895199835300446,
-0.07525460422039032,
0.09282654523849487,
-0.0507965162396431,
0.1469351053237915,
0.08020289987325668,
-0.09152709692716599,
-0.19188682734966278,
-0.0887833908200264,
0.10164182633161545,
0.18469172716140747,
0.045696184039115906,
-0.022488808259367943,
0.09940612316131592,
-0.08621317893266678,
0.011039474047720432,
0.05154034495353699,
-0.06937182694673538,
-0.05223534256219864,
0.06355299055576324,
0.08018788695335388,
0.07678371667861938,
-0.12301702797412872,
-0.02094447799026966,
0.008637533523142338,
0.00831096712499857,
0.08201737701892853,
0.023290244862437248,
0.1510206013917923,
0.03883988782763481,
-0.12744688987731934,
-0.050009194761514664,
0.10665731877088547,
0.041741468012332916,
-0.04784774035215378,
-0.25138479471206665,
-0.030326439067721367,
-0.027732934802770615,
-0.029999805614352226,
-0.03873695060610771,
0.04263332113623619,
-0.0072723389603197575,
0.0826614573597908,
-0.008116158656775951,
-0.07679495960474014,
-0.03798604756593704,
0.06191713735461235,
0.060809630900621414,
0.026244111359119415,
-0.011753023602068424,
0.010934822261333466,
0.1174238994717598,
0.10631082952022552,
-0.12367359548807144,
-0.051516905426979065,
-0.06431761384010315,
-0.07867198437452316,
-0.04216236248612404,
0.03455616533756256,
0.041060756891965866,
0.049376390874385834,
0.2486443817615509,
0.017620395869016647,
0.05382118001580238,
0.03803925961256027,
0.010167144238948822,
0.06406087428331375,
0.11435336619615555,
-0.061582546681165695,
-0.09715550392866135,
-0.025186026468873024,
0.08966731280088425,
0.01176387071609497,
-0.04024789482355118,
-0.05783011019229889,
0.06293477863073349,
0.016524890437722206,
0.1202789843082428,
0.09223750233650208,
0.003793274285271764,
-0.07138240337371826,
-0.06413803994655609,
0.1937950700521469,
-0.1626761257648468,
0.04747059941291809,
0.034180231392383575,
-0.038511235266923904,
-0.0016249394975602627,
0.008853171020746231,
0.024325255304574966,
-0.021725021302700043,
0.08937039971351624,
-0.05618007108569145,
-0.041590798646211624,
-0.10983981937170029,
-0.035744234919548035,
0.03192625194787979,
0.009910091757774353,
-0.03217151761054993,
-0.031847331672906876,
-0.08444786816835403,
-0.06831640005111694,
0.09424425661563873,
-0.07356466352939606,
-0.053753651678562164,
-0.016938211396336555,
-0.07437273859977722,
0.024786023423075676,
0.01960081420838833,
0.07747352123260498,
-0.02004585787653923,
0.042900070548057556,
-0.05549933388829231,
0.06014169380068779,
0.10937028378248215,
0.033117540180683136,
-0.05445994809269905,
0.0621645413339138,
-0.2418462336063385,
0.0997670441865921,
-0.06829129904508591,
0.05325306951999664,
-0.15072302520275116,
-0.02465333603322506,
0.04913770779967308,
0.008168290369212627,
-0.010590006597340107,
0.13754788041114807,
-0.21924975514411926,
-0.027699807658791542,
0.1631394773721695,
-0.09464818984270096,
-0.07676627486944199,
0.05986984074115753,
-0.052457790821790695,
0.10692904144525528,
0.04047565534710884,
-0.026259733363986015,
0.06162377819418907,
-0.13397987186908722,
0.0005626814090646803,
-0.045883387327194214,
-0.01928110048174858,
0.15731419622898102,
0.07587230950593948,
-0.06994020938873291,
0.07348526269197464,
0.023750323802232742,
-0.023168303072452545,
-0.046913031488657,
-0.017583578824996948,
-0.1088033989071846,
0.010729904286563396,
-0.061985816806554794,
0.01937699131667614,
-0.025795195251703262,
-0.09332547336816788,
-0.028493179008364677,
-0.17521639168262482,
-0.020266273990273476,
0.08516935259103775,
-0.009352635592222214,
-0.01925206556916237,
-0.11787936836481094,
0.015734510496258736,
0.03501737862825394,
0.002549536293372512,
-0.1319509893655777,
-0.05043373629450798,
0.02751830592751503,
-0.16075198352336884,
0.033688947558403015,
-0.05403051897883415,
0.0491553395986557,
0.03133281692862511,
-0.031412381678819656,
-0.028679344803094864,
0.022094380110502243,
0.004997676704078913,
-0.014611656777560711,
-0.24550160765647888,
-0.026604164391756058,
-0.02145342156291008,
0.16796952486038208,
-0.21640902757644653,
0.0374150350689888,
0.07194960117340088,
0.15254895389080048,
0.008589224889874458,
-0.038006994873285294,
0.002335198922082782,
-0.075041763484478,
-0.03255171701312065,
-0.06050482019782066,
-0.009038056246936321,
-0.03572068363428116,
-0.05482286959886551,
0.04863523691892624,
-0.16824471950531006,
-0.029467429965734482,
0.1015508770942688,
0.06473538279533386,
-0.13604550063610077,
-0.019663551822304726,
-0.03585261106491089,
-0.042308371514081955,
-0.05517838895320892,
-0.05935737490653992,
0.10260266810655594,
0.05827045813202858,
0.04566904529929161,
-0.06485172361135483,
-0.0747392401099205,
0.0017082487465813756,
-0.019673427566885948,
-0.022536588832736015,
0.09213293343782425,
0.07581926137208939,
-0.12331884354352951,
0.09213830530643463,
0.10402927547693253,
0.08686267584562302,
0.0966128259897232,
-0.023164015263319016,
-0.08361977338790894,
-0.049845483154058456,
0.02228725142776966,
0.017598064616322517,
0.13447505235671997,
-0.007804518099874258,
0.05406574159860611,
0.04160919412970543,
-0.013909573666751385,
0.009752067737281322,
-0.09242741018533707,
0.032518286257982254,
0.03427431732416153,
-0.01857241988182068,
0.041615914553403854,
-0.039849672466516495,
0.019975949078798294,
0.09018522500991821,
0.046917494386434555,
0.04021155461668968,
0.014107138849794865,
-0.04660527780652046,
-0.11187547445297241,
0.16612006723880768,
-0.12780359387397766,
-0.23512837290763855,
-0.1463187336921692,
0.0034277087543159723,
0.03630480915307999,
-0.009390040300786495,
0.0017278295708820224,
-0.06397698074579239,
-0.11876852810382843,
-0.09194197505712509,
0.010153552517294884,
0.04896695911884308,
-0.0851091742515564,
-0.0603698305785656,
0.05686335638165474,
0.04057794436812401,
-0.14546048641204834,
0.019262617453932762,
0.04933769255876541,
-0.09224124997854233,
-0.009894786402583122,
0.08289197087287903,
0.06857553124427795,
0.18091025948524475,
0.013082148507237434,
-0.02271466888487339,
0.03428078070282936,
0.21755947172641754,
-0.13586747646331787,
0.11420658230781555,
0.1426045000553131,
-0.09194567799568176,
0.08309654146432877,
0.19839057326316833,
0.04078111797571182,
-0.10157861560583115,
0.032499175518751144,
0.018653791397809982,
-0.030491048470139503,
-0.24355553090572357,
-0.07171683013439178,
0.00034942623460665345,
-0.057900771498680115,
0.07530075311660767,
0.09018687158823013,
0.09155713021755219,
0.01583298109471798,
-0.0946493074297905,
-0.07830986380577087,
0.05305508151650429,
0.10324970632791519,
0.020061472430825233,
-0.013236436992883682,
0.09051742404699326,
-0.03375976160168648,
0.017617853358387947,
0.09066354483366013,
0.0011531224008649588,
0.17065346240997314,
0.05820678174495697,
0.18275249004364014,
0.07604338973760605,
0.07338658720254898,
0.01378361415117979,
0.01180104911327362,
0.019032908603549004,
0.02708563208580017,
-0.004741039127111435,
-0.08538748323917389,
-0.01599922962486744,
0.12008915096521378,
0.07424698024988174,
0.015674617141485214,
0.014355434104800224,
-0.04089333862066269,
0.08203015476465225,
0.17435193061828613,
-0.001506963511928916,
-0.1824604868888855,
-0.06271602213382721,
0.08220411837100983,
-0.09449198096990585,
-0.10147359222173691,
-0.02445729449391365,
0.03089604340493679,
-0.17088350653648376,
0.023070847615599632,
-0.016430631279945374,
0.11182350665330887,
-0.13931094110012054,
-0.019696295261383057,
0.0640200525522232,
0.07118809968233109,
-0.00031885437783785164,
0.05944213643670082,
-0.16128569841384888,
0.10404066741466522,
0.013166810385882854,
0.06712377816438675,
-0.09715772420167923,
0.10046469420194626,
-0.006883090827614069,
-0.013416164554655552,
0.13275203108787537,
0.008256223052740097,
-0.07161599397659302,
-0.07921489328145981,
-0.09379399567842484,
-0.009093280881643295,
0.12668752670288086,
-0.14835532009601593,
0.08585991710424423,
-0.035368360579013824,
-0.04256736859679222,
0.0022144275717437267,
-0.10755012929439545,
-0.12217973172664642,
-0.1874755620956421,
0.05520224943757057,
-0.1321607530117035,
0.039849888533353806,
-0.10649667680263519,
-0.03462952747941017,
-0.029491933062672615,
0.1882491409778595,
-0.22971367835998535,
-0.06835493445396423,
-0.15157760679721832,
-0.09785088151693344,
0.14553189277648926,
-0.04969761520624161,
0.08694402873516083,
-0.005991519894450903,
0.18016821146011353,
0.022223925217986107,
-0.021585633978247643,
0.09859558939933777,
-0.09382225573062897,
-0.1963716447353363,
-0.08180448412895203,
0.15751656889915466,
0.13459575176239014,
0.03521031513810158,
-0.0027760460507124662,
0.037876322865486145,
-0.01856307126581669,
-0.12259240448474884,
0.021658578887581825,
0.17797763645648956,
0.0652514174580574,
0.02310643345117569,
-0.026529761031270027,
-0.11104881763458252,
-0.06772379577159882,
-0.033685971051454544,
0.03064778819680214,
0.18449479341506958,
-0.0722544714808464,
0.18419069051742554,
0.143813356757164,
-0.05867353826761246,
-0.1976030021905899,
0.008879725821316242,
0.03365374729037285,
0.007196295075118542,
0.03445420414209366,
-0.20255140960216522,
0.0841677114367485,
0.00034181843511760235,
-0.05190233513712883,
0.13343381881713867,
-0.17106693983078003,
-0.15042030811309814,
0.07339101284742355,
0.03619921952486038,
-0.19460853934288025,
-0.11963265389204025,
-0.08913769572973251,
-0.05391303077340126,
-0.18051348626613617,
0.10290905088186264,
0.03496568650007248,
0.008035079576075077,
0.03376363217830658,
0.028494013473391533,
0.01669638603925705,
-0.03928735852241516,
0.1920013129711151,
-0.026591487228870392,
0.029855716973543167,
-0.08456290513277054,
-0.06990274786949158,
0.04655740037560463,
-0.05482156574726105,
0.0760476216673851,
-0.027013001963496208,
0.011612839996814728,
-0.10561433434486389,
-0.042526841163635254,
-0.029051896184682846,
0.013453613966703415,
-0.0963861495256424,
-0.08940120041370392,
-0.0490599125623703,
0.09310506284236908,
0.09519506990909576,
-0.035876575857400894,
-0.03684677556157112,
-0.07069114595651627,
0.039579302072525024,
0.18676936626434326,
0.17657315731048584,
0.04523694887757301,
-0.0789421945810318,
-0.005537794437259436,
-0.011924253776669502,
0.04352729767560959,
-0.21637341380119324,
0.06442029029130936,
0.05013522133231163,
0.017847778275609016,
0.11767403781414032,
-0.02045002020895481,
-0.1556767225265503,
-0.07006701827049255,
0.06328949332237244,
-0.06132598593831062,
-0.1951322853565216,
0.005576360039412975,
0.054395273327827454,
-0.16848263144493103,
-0.048018258064985275,
0.04364382475614548,
-0.004054433200508356,
-0.0402018167078495,
0.01867259293794632,
0.08977478742599487,
0.003425614908337593,
0.0704059898853302,
0.05869606137275696,
0.08224445581436157,
-0.10246741771697998,
0.07471306622028351,
0.08622124791145325,
-0.07954994589090347,
0.026619622483849525,
0.09149482846260071,
-0.05819176882505417,
-0.02969011478126049,
0.02704544924199581,
0.0793747529387474,
0.011502381414175034,
-0.042540501803159714,
0.011518802493810654,
-0.10228829830884933,
0.06203006953001022,
0.08760257810354233,
0.03265642002224922,
0.015443529933691025,
0.03219176456332207,
0.045628782361745834,
-0.07176384329795837,
0.1219232901930809,
0.028246978297829628,
0.015991143882274628,
-0.04067446291446686,
-0.04898078367114067,
0.024271609261631966,
-0.0303955040872097,
-0.006366716232150793,
-0.03475780412554741,
-0.0729878842830658,
-0.0171539094299078,
-0.16714228689670563,
-0.016664555296301842,
-0.04662061110138893,
0.009329318068921566,
0.03086909092962742,
-0.03788549080491066,
0.008464637212455273,
0.007407912518829107,
-0.07459274679422379,
-0.06477426737546921,
-0.022905457764863968,
0.09289900958538055,
-0.16393527388572693,
0.02335011027753353,
0.08690579235553741,
-0.12064014375209808,
0.09392421692609787,
0.01837589405477047,
-0.0037578048650175333,
0.028480252251029015,
-0.14924435317516327,
0.038928523659706116,
-0.03113253228366375,
0.014821149408817291,
0.04454975947737694,
-0.2236335128545761,
0.0009650349384173751,
-0.033828526735305786,
-0.06339430809020996,
-0.009390673600137234,
-0.036760155111551285,
-0.11370383948087692,
0.10629112273454666,
0.007970798760652542,
-0.08916810154914856,
-0.031690530478954315,
0.032128699123859406,
0.08206479996442795,
-0.0239556971937418,
0.15763959288597107,
-0.0023972811177372932,
0.0736590027809143,
-0.1675432026386261,
-0.019303109496831894,
-0.011248460970818996,
0.020926566794514656,
-0.018098697066307068,
-0.01251189224421978,
0.04078914225101471,
-0.02225574664771557,
0.18437865376472473,
-0.023570427671074867,
0.023348741233348846,
0.06592654436826706,
0.027775658294558525,
-0.025002485141158104,
0.10530006885528564,
0.05339968949556351,
0.021854043006896973,
0.02036798559129238,
0.00273964018560946,
-0.04241073876619339,
-0.023610878735780716,
-0.1998770385980606,
0.06446972489356995,
0.14037446677684784,
0.09086652100086212,
-0.017234215512871742,
0.08257289230823517,
-0.1004219725728035,
-0.11521948128938675,
0.11568495631217957,
-0.05446505919098854,
-0.004037478007376194,
-0.0672159418463707,
0.12938179075717926,
0.1446845531463623,
-0.19097456336021423,
0.06995914876461029,
-0.06848131865262985,
-0.049033988267183304,
-0.11654651165008545,
-0.1963350623846054,
-0.05714293569326401,
-0.05161691829562187,
-0.01663723587989807,
-0.046969223767519,
0.07560921460390091,
0.05719533935189247,
0.007424132898449898,
-0.0017566849710419774,
0.06332923471927643,
-0.026077456772327423,
0.00009585227962816134,
0.026813751086592674,
0.06610306352376938,
0.013093758374452591,
-0.02985633723437786,
0.017491595819592476,
-0.012147722765803337,
0.042048826813697815,
0.06357792019844055,
0.04670548066496849,
-0.030032360926270485,
0.016853880137205124,
-0.03863191977143288,
-0.10680584609508514,
0.041318636387586594,
-0.028504958376288414,
-0.08043242245912552,
0.1491626501083374,
0.02454165369272232,
0.008750278502702713,
-0.0205967016518116,
0.2416755110025406,
-0.0737907737493515,
-0.09567341208457947,
-0.1479424238204956,
0.10524045675992966,
-0.04420987144112587,
0.06244929879903793,
0.045180387794971466,
-0.10425344854593277,
0.016717668622732162,
0.12817999720573425,
0.16302813589572906,
-0.044200748205184937,
0.020526019856333733,
0.027614353224635124,
0.004152800887823105,
-0.03678637370467186,
0.0514480359852314,
0.06988705694675446,
0.1595088243484497,
-0.048713311553001404,
0.09546878933906555,
-0.0016016386216506362,
-0.09618084132671356,
-0.03802286460995674,
0.11709540337324142,
-0.018092934042215347,
0.017691975459456444,
-0.055210161954164505,
0.11857418715953827,
-0.06138255074620247,
-0.2316483110189438,
0.06108921393752098,
-0.06591550260782242,
-0.13765475153923035,
-0.02143050730228424,
0.08041442185640335,
-0.013238796964287758,
0.02708347514271736,
0.07207029312849045,
-0.07533451914787292,
0.20003929734230042,
0.037636954337358475,
-0.05420409142971039,
-0.05360380560159683,
0.08255447447299957,
-0.10376271605491638,
0.27565470337867737,
0.016520937904715538,
0.04948882386088371,
0.10317612439393997,
-0.012690499424934387,
-0.13475549221038818,
0.02108365297317505,
0.09600389003753662,
-0.0946137085556984,
0.04216265305876732,
0.19903649389743805,
0.0003853837260976434,
0.1207512691617012,
0.0790785402059555,
-0.07618726044893265,
0.049590613692998886,
-0.0941753089427948,
-0.07070460170507431,
-0.09001081436872482,
0.09455035626888275,
-0.07685617357492447,
0.14261877536773682,
0.1292559802532196,
-0.053739987313747406,
0.010677514597773552,
-0.028576120734214783,
0.04638256877660751,
0.0034859003499150276,
0.1005801111459732,
0.010024284943938255,
-0.18460705876350403,
0.02157641015946865,
0.01203901320695877,
0.1056026741862297,
-0.16518552601337433,
-0.09804878383874893,
0.042120642960071564,
0.0014211505185812712,
-0.060778699815273285,
0.12909291684627533,
0.06027422100305557,
0.04478219151496887,
-0.04292554408311844,
-0.020403601229190826,
-0.009860116057097912,
0.13677826523780823,
-0.10241927951574326,
0.0014122816501185298
] |
null | null | transformers |
# Midjourney v6 Image Classification | {"pipeline_tag": "image-classification"} | image-classification | ideepankarsharma2003/Midjourney_v6_ImageClassifier | [
"transformers",
"safetensors",
"swin",
"image-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T09:35:14+00:00 | [] | [] | TAGS
#transformers #safetensors #swin #image-classification #autotrain_compatible #endpoints_compatible #region-us
|
# Midjourney v6 Image Classification | [
"# Midjourney v6 Image Classification"
] | [
"TAGS\n#transformers #safetensors #swin #image-classification #autotrain_compatible #endpoints_compatible #region-us \n",
"# Midjourney v6 Image Classification"
] | [
38,
9
] | [
"passage: TAGS\n#transformers #safetensors #swin #image-classification #autotrain_compatible #endpoints_compatible #region-us \n# Midjourney v6 Image Classification"
] | [
-0.05098317936062813,
0.05216138809919357,
-0.003817607183009386,
0.0512714758515358,
0.18920056521892548,
0.025373613461852074,
0.03656882792711258,
0.04512016102671623,
-0.05225476622581482,
-0.027700349688529968,
0.10219785571098328,
0.14220058917999268,
-0.01383967138826847,
0.07106702029705048,
-0.10416664928197861,
-0.3276037275791168,
0.06711508333683014,
-0.009458671323955059,
0.0032341654878109694,
0.061215732246637344,
0.0539303794503212,
-0.16787448525428772,
0.11854287981987,
-0.04578652232885361,
-0.2646918296813965,
0.02666923590004444,
0.03388627991080284,
-0.0807749330997467,
0.07598554342985153,
0.03331948444247246,
0.13547992706298828,
0.07544193416833878,
0.08990103006362915,
-0.047893226146698,
0.033198997378349304,
0.017842212691903114,
-0.07894404977560043,
-0.011817793361842632,
0.13523218035697937,
-0.047729503363370895,
0.020923733711242676,
0.08484205603599548,
0.007318609859794378,
0.010225672274827957,
-0.04994628578424454,
-0.040051527321338654,
0.016463110223412514,
0.022157108411192894,
0.21817751228809357,
-0.0019043105421587825,
0.05691412463784218,
0.0414109081029892,
-0.095647431910038,
0.13337816298007965,
0.06458472460508347,
-0.1902133673429489,
-0.06860947608947754,
0.1671273410320282,
-0.058605559170246124,
-0.037640344351530075,
-0.12268257141113281,
0.10880564898252487,
0.02235383167862892,
-0.011248079128563404,
-0.022702280431985855,
-0.04373760148882866,
-0.097007617354393,
-0.04280730336904526,
-0.08255089819431305,
-0.017150895670056343,
0.0586027055978775,
0.042357008904218674,
0.032796964049339294,
-0.0552690327167511,
-0.09988057613372803,
-0.017866939306259155,
-0.11495605111122131,
0.054033223539590836,
0.025899475440382957,
0.01195924449712038,
-0.06611011177301407,
0.09326259791851044,
-0.12938131392002106,
-0.026154475286602974,
-0.16928313672542572,
0.0937923938035965,
0.009440667927265167,
0.09918469190597534,
-0.15607671439647675,
0.03419024124741554,
0.0049996888265013695,
-0.08667795360088348,
0.051072947680950165,
-0.11265140771865845,
-0.023449869826436043,
-0.009738507680594921,
-0.004663476720452309,
-0.12143608182668686,
0.07426492869853973,
0.048151902854442596,
0.062324341386556625,
0.06635688245296478,
-0.054983630776405334,
0.09707889705896378,
0.014825296588242054,
0.05079139024019241,
-0.011028922162950039,
-0.027425561100244522,
0.023587867617607117,
-0.061422012746334076,
0.01799580454826355,
-0.029881423339247704,
-0.13584551215171814,
-0.021679280325770378,
0.04465382173657417,
-0.010525580495595932,
-0.04723427817225456,
0.06765620410442352,
-0.03404918685555458,
-0.02770288474857807,
0.17038360238075256,
-0.005981782451272011,
0.003776494413614273,
-0.024265656247735023,
0.04119125381112099,
0.02099069394171238,
0.024829385802149773,
0.003075616667047143,
0.05390835925936699,
0.18364201486110687,
-0.028265628963708878,
0.022265104576945305,
-0.006867637392133474,
-0.01366608589887619,
-0.011922777630388737,
-0.08789114654064178,
0.02247171476483345,
-0.21013100445270538,
-0.11665130406618118,
0.013854398392140865,
0.024925652891397476,
-0.008202928118407726,
0.04413960501551628,
0.011086240410804749,
-0.023580297827720642,
0.061411429196596146,
-0.0009030118235386908,
-0.009212610311806202,
-0.06166590750217438,
0.04440153390169144,
-0.06370113044977188,
0.13568420708179474,
-0.05951477959752083,
0.01973152346909046,
-0.10372935980558395,
0.013553038239479065,
-0.1455869823694229,
0.038968272507190704,
-0.07520349323749542,
0.2293609231710434,
0.01683901809155941,
-0.02949857898056507,
-0.06459672003984451,
0.030328387394547462,
-0.10400035232305527,
0.15236710011959076,
-0.13832174241542816,
-0.055749788880348206,
0.13202467560768127,
-0.1677583009004593,
-0.12329155206680298,
0.04706526920199394,
-0.0033199950121343136,
-0.09457013010978699,
0.05567965656518936,
0.11066707223653793,
0.08305703103542328,
-0.04076066613197327,
0.0067604780197143555,
0.014437716454267502,
-0.1643877774477005,
-0.08723852783441544,
-0.005569105036556721,
0.13093793392181396,
-0.10928528755903244,
0.03768935799598694,
0.010024122893810272,
0.13181284070014954,
-0.07066424190998077,
-0.0485040545463562,
-0.011662162840366364,
-0.031994421035051346,
0.03615782782435417,
0.10141002386808395,
0.08917625993490219,
-0.11627388745546341,
0.03463886305689812,
-0.03026166558265686,
0.009350736625492573,
-0.0004424213257152587,
-0.026717763394117355,
-0.06221742555499077,
0.09542974829673767,
-0.10395992547273636,
0.005059948191046715,
-0.11877825111150742,
-0.11719328165054321,
-0.021183127537369728,
0.17713165283203125,
-0.051651425659656525,
0.028139373287558556,
0.0972164198756218,
0.009406455792486668,
-0.05043940618634224,
-0.10512661188840866,
0.15710234642028809,
0.03339603543281555,
-0.059660691767930984,
-0.1268622726202011,
0.04677669331431389,
-0.0657619759440422,
-0.02541518770158291,
-0.15343086421489716,
-0.04845702275633812,
0.17438049614429474,
0.19133049249649048,
0.0907362774014473,
0.05387157201766968,
-0.003737298771739006,
-0.039042990654706955,
-0.08144935220479965,
-0.039298880845308304,
0.11982668191194534,
-0.009571697562932968,
0.0031549257691949606,
0.1623261719942093,
-0.1216605007648468,
0.2114126980304718,
0.18879111111164093,
-0.3162611126899719,
-0.07055522501468658,
-0.01563357189297676,
0.009592763148248196,
0.01658065803349018,
-0.04333314672112465,
0.04966594651341438,
-0.1470196545124054,
-0.055774375796318054,
0.17026497423648834,
0.004066196735948324,
0.010270064696669579,
0.052426498383283615,
0.0013760667061433196,
-0.12356775254011154,
0.008856951259076595,
0.041219595819711685,
-0.28117939829826355,
0.1472582370042801,
0.2691199779510498,
-0.0009684322867542505,
0.12755946815013885,
0.04572216048836708,
-0.006882919929921627,
0.10437137633562088,
-0.006025131791830063,
0.003196439938619733,
0.0582406111061573,
-0.03870633989572525,
-0.08664562553167343,
0.05100713670253754,
-0.014880554750561714,
-0.017419958487153053,
-0.1361723095178604,
0.009032858535647392,
0.019921189174056053,
0.06429249793291092,
0.12293490022420883,
0.03732375428080559,
0.0022329515777528286,
0.12705391645431519,
-0.04746229574084282,
-0.1379057615995407,
0.07603874802589417,
-0.04026643931865692,
-0.02949380688369274,
0.16998298466205597,
-0.050635356456041336,
-0.33163338899612427,
-0.07753226161003113,
-0.16333578526973724,
-0.08619142323732376,
0.09822253882884979,
0.054166387766599655,
-0.15073639154434204,
-0.1304442435503006,
-0.014963626861572266,
-0.10862510651350021,
0.001902171759866178,
0.08705389499664307,
-0.02737937495112419,
0.05307653918862343,
-0.01907791756093502,
-0.004256758373230696,
-0.05582477152347565,
-0.057510413229465485,
-0.033970341086387634,
0.21042479574680328,
-0.07317600399255753,
0.11325620859861374,
0.07929325103759766,
-0.06520941853523254,
0.059367209672927856,
-0.009628579951822758,
0.1317114382982254,
-0.09774293005466461,
0.014450939372181892,
0.1683254987001419,
-0.04478519409894943,
0.031401604413986206,
0.16638624668121338,
0.0255923792719841,
-0.11647579818964005,
0.042506638914346695,
-0.0053227124735713005,
-0.13598337769508362,
-0.018002301454544067,
-0.10170504450798035,
-0.06164601817727089,
0.005195058882236481,
0.1364133656024933,
0.10823051631450653,
0.033723488450050354,
0.15198133885860443,
0.003569172229617834,
0.00021841246052645147,
0.07577645778656006,
0.09010543674230576,
0.07434456795454025,
0.021385198459029198,
0.14554369449615479,
-0.0714162290096283,
-0.10797329246997833,
0.07583818584680557,
0.03426583483815193,
0.15611152350902557,
0.10219815373420715,
-0.1098964661359787,
0.009293659590184689,
0.022289274260401726,
0.20249351859092712,
0.07716483622789383,
0.010431473143398762,
-0.06984637677669525,
0.011916602030396461,
-0.006398916710168123,
0.003959180321544409,
0.02737375721335411,
0.029782453551888466,
-0.05618768557906151,
-0.04909927770495415,
0.03485262393951416,
0.07359091937541962,
0.14753417670726776,
0.059567905962467194,
-0.30702823400497437,
0.06502245366573334,
0.0895375981926918,
0.036762308329343796,
-0.09444169700145721,
0.09080610424280167,
0.06468819826841354,
-0.06824751943349838,
0.16995157301425934,
-0.0748588964343071,
0.08390790224075317,
-0.06937500089406967,
0.0078009748831391335,
0.028277505189180374,
-0.16691575944423676,
0.03407275304198265,
0.04159601405262947,
-0.10877170413732529,
0.17892798781394958,
-0.005048607941716909,
-0.057094376534223557,
-0.03998720273375511,
-0.023191366344690323,
0.10660839825868607,
0.29398900270462036,
0.2068663388490677,
-0.0021833786740899086,
-0.19822734594345093,
-0.12626779079437256,
0.008704455569386482,
-0.04895556718111038,
0.10012464970350266,
0.0036643065977841616,
-0.013431441970169544,
-0.033292412757873535,
-0.0861043706536293,
-0.034436460584402084,
-0.02957618050277233,
-0.005981672555208206,
-0.1204894557595253,
-0.011789998970925808,
0.03443022444844246,
0.006114250514656305,
0.0127346096560359,
-0.0570933036506176,
-0.17653310298919678,
0.13589948415756226,
-0.06251662224531174,
0.0013713198713958263,
-0.16664241254329681,
-0.000017142763681476936,
-0.060756176710128784,
-0.03329376503825188,
0.09542597085237503,
-0.08779198676347733,
0.128880575299263,
-0.035848360508680344,
-0.21295398473739624,
0.16591666638851166,
-0.13787153363227844,
-0.07303764671087265,
-0.08478284627199173,
0.05586857348680496,
-0.09976710379123688,
-0.03256913647055626,
0.035008613020181656,
0.09885696321725845,
-0.020561696961522102,
-0.01845669560134411,
0.132639542222023,
0.0007882083300501108,
0.12133773416280746,
0.05375561863183975,
-0.004866128321737051,
-0.16628751158714294,
0.033919237554073334,
0.05149427056312561,
0.15786327421665192,
0.17924335598945618,
-0.09495127201080322,
0.07109366357326508,
0.03896554931998253,
0.01652265340089798,
-0.4162823557853699,
0.03420571610331535,
-0.053632814437150955,
-0.055615440011024475,
0.009691664017736912,
0.005338732153177261,
0.18839111924171448,
0.0680406391620636,
-0.058466341346502304,
0.07539767026901245,
-0.11476486921310425,
-0.10575385391712189,
0.13445855677127838,
0.1514720767736435,
0.3159759044647217,
-0.16846992075443268,
-0.01696452684700489,
-0.13450750708580017,
-0.15346935391426086,
0.11288689821958542,
0.056108344346284866,
0.0361168310046196,
-0.016866454854607582,
-0.08036930114030838,
0.010650989599525928,
-0.0449090339243412,
0.07708625495433807,
-0.04922394081950188,
0.16816595196723938,
-0.16144926846027374,
-0.06659071147441864,
0.010410084389150143,
-0.050446342676877975,
-0.007925176061689854,
0.07393378019332886,
0.04500265792012215,
-0.09736718237400055,
0.014299392700195312,
-0.03307412564754486,
0.05874057486653328,
0.07639066129922867,
-0.0014146411558613181,
-0.01583784446120262,
0.002302114386111498,
0.0013692593201994896,
0.024655397981405258,
0.3694572448730469,
-0.0202174074947834,
0.07457569241523743,
0.05997183918952942,
0.08817189931869507,
-0.14633050560951233,
-0.009545822627842426,
-0.16010938584804535,
-0.049340471625328064,
0.13713020086288452,
-0.10989617556333542,
0.11290045082569122,
0.08830109983682632,
-0.008094623684883118,
0.04683172330260277,
0.1095714345574379,
0.12746451795101166,
0.006222342606633902,
0.22948558628559113,
-0.14469587802886963,
-0.06487210094928741,
-0.04554285854101181,
0.0956217423081398,
0.156520277261734,
0.10393719375133514,
0.0795164555311203,
0.00040612646262161434,
0.011281711980700493,
0.01746363937854767,
0.014739763922989368,
0.03107667900621891,
0.061393242329359055,
0.06683794409036636,
-0.0029912670142948627,
-0.13423530757427216,
0.13464267551898956,
0.010791598819196224,
-0.18093881011009216,
-0.08124207705259323,
0.053248900920152664,
-0.12834900617599487,
-0.10587236285209656,
0.05396272614598274,
0.14341439306735992,
-0.12727417051792145,
-0.021986236795783043,
-0.0018049338832497597,
-0.12215012311935425,
0.03488217294216156,
0.14904607832431793,
0.11965236812829971,
0.02269412763416767,
0.041390642523765564,
-0.010694939643144608,
-0.04705282300710678,
0.0575900673866272,
-0.05596357211470604,
0.14310994744300842,
-0.27426138520240784,
-0.08813907951116562,
-0.02395613305270672,
0.1121990978717804,
-0.14224931597709656,
-0.036330997943878174,
-0.1103150025010109,
0.048072829842567444,
-0.08881386369466782,
0.05496162921190262,
-0.11394122242927551,
-0.009011815302073956,
0.0036227540113031864,
0.0010127349523827434,
-0.044032223522663116,
-0.03368191048502922,
-0.11682663857936859,
0.0022949213162064552,
0.015611007809638977,
-0.0020034138578921556,
-0.012324435636401176,
-0.06696880608797073,
0.014945338480174541,
-0.04240834340453148,
0.07647676020860672,
0.07073404639959335,
-0.0543515719473362,
0.02443739026784897,
-0.1462935358285904,
-0.22509823739528656,
0.1843857318162918,
-0.02556663565337658,
0.03821023181080818,
0.14153122901916504,
0.031053708866238594,
0.06866943091154099,
-0.05982388183474541,
0.024316763505339622,
0.07680584490299225,
-0.0958554670214653,
-0.0016616197535768151,
-0.12065642327070236,
-0.1197519302368164,
-0.033145200461149216,
0.010239509865641594,
0.13632144033908844,
-0.032459985464811325,
0.16839160025119781,
-0.09987575560808182,
0.015580868348479271,
-0.07144094258546829,
-0.01095577422529459,
-0.024906402453780174,
-0.21981187164783478,
-0.16360683739185333,
-0.010426834225654602,
0.024239148944616318,
-0.09106742590665817,
0.12825973331928253,
0.07978319376707077,
0.016156429424881935,
0.050135280936956406,
0.14035531878471375,
0.041439902037382126,
0.0741349458694458,
0.17637108266353607,
0.06797391921281815,
-0.0254123006016016,
-0.014263337478041649,
0.09546748548746109,
0.09643781185150146,
-0.014736450277268887,
0.02714533917605877,
0.10779691487550735,
-0.1572103500366211,
0.0892808735370636,
0.05972614511847496,
0.0580887608230114,
-0.16180692613124847,
-0.05656154081225395,
-0.1520928144454956,
0.12002265453338623,
-0.010960935615003109,
-0.08954694867134094,
0.23177632689476013,
-0.04742438346147537,
0.0023793356958776712,
-0.07253663241863251,
0.011083264835178852,
-0.11253409087657928,
-0.19063910841941833,
-0.16309496760368347,
-0.16157718002796173,
0.08555708825588226,
-0.014802679419517517,
-0.02937699854373932,
0.17560985684394836,
0.0802973285317421,
-0.00004151857865508646,
0.13360385596752167,
-0.09102853387594223,
0.03160160034894943,
0.10951709002256393,
-0.014716935344040394,
-0.05557027831673622,
0.07389607280492783,
-0.019100124016404152,
0.009942932054400444,
-0.04759260639548302,
-0.09094821661710739,
-0.023548675701022148,
-0.0009319659438915551,
0.02534862980246544,
-0.039483606815338135,
-0.11333911865949631,
0.005712772719562054,
0.009953057393431664,
-0.05913354828953743,
-0.02186122164130211,
-0.021549740806221962,
0.08498521894216537,
0.02872699871659279,
0.1632102131843567,
-0.0908195748925209,
0.003622280666604638,
-0.053789593279361725,
0.04178878292441368,
-0.008205389603972435,
0.1520290970802307,
-0.041393883526325226,
-0.05739251896739006,
-0.010827355086803436,
0.1863587647676468,
0.15464620292186737,
-0.11698516458272934,
0.02565336972475052,
0.0014597229892387986,
-0.00021766184363514185,
0.03290431946516037,
0.1462290734052658,
-0.0016576671041548252,
0.13629567623138428,
-0.04158775880932808,
-0.11982826888561249,
-0.03700665384531021,
-0.06772056967020035,
-0.04585842043161392,
0.051733240485191345,
0.05654412508010864,
-0.006522502284497023,
-0.13767006993293762,
0.07235065847635269,
0.012593536637723446,
0.20160475373268127,
0.15354818105697632,
-0.16701030731201172,
-0.06290493905544281,
-0.014174090698361397,
0.13680702447891235,
0.03936843201518059,
-0.0411820262670517,
-0.07629794627428055,
-0.03685915842652321,
-0.0867755189538002,
0.017561152577400208,
-0.17594130337238312,
-0.09506864845752716,
-0.06251293420791626,
-0.023875409737229347,
0.16815969347953796,
-0.02985801361501217,
-0.02992996573448181,
0.023965194821357727,
0.022074833512306213,
-0.07866574078798294,
0.028356481343507767,
-0.05738025903701782,
-0.114836685359478,
0.024548538029193878,
0.19983220100402832,
-0.023837115615606308,
0.001856421702541411,
0.0508204810321331,
-0.1433607041835785,
0.021977830678224564,
-0.04538090154528618,
-0.07874581217765808,
-0.02374928630888462,
0.09917638450860977,
-0.0318363718688488,
0.09050247073173523,
0.04984551668167114,
0.05187264457345009,
-0.011057390831410885,
-0.04230450093746185,
0.008373795077204704,
0.06091725453734398,
-0.030217856168746948,
-0.03962817043066025,
-0.06605855375528336,
-0.05049958825111389,
0.04053719341754913,
0.01172164361923933,
-0.03211916238069534,
-0.07882515341043472,
-0.13037723302841187,
0.017971258610486984,
-0.11540958285331726,
0.10750184208154678,
0.1484668105840683,
0.015447511337697506,
-0.03333177790045738,
-0.07717633992433548,
0.08499789983034134,
0.07591315358877182,
-0.04153817147016525,
-0.11150839924812317
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.