sha
null | last_modified
null | library_name
stringclasses 154
values | text
stringlengths 1
900k
| metadata
stringlengths 2
348k
| pipeline_tag
stringclasses 45
values | id
stringlengths 5
122
| tags
sequencelengths 1
1.84k
| created_at
stringlengths 25
25
| arxiv
sequencelengths 0
201
| languages
sequencelengths 0
1.83k
| tags_str
stringlengths 17
9.34k
| text_str
stringlengths 0
389k
| text_lists
sequencelengths 0
722
| processed_texts
sequencelengths 1
723
| tokens_length
sequencelengths 1
723
| input_texts
sequencelengths 1
61
| embeddings
sequencelengths 768
768
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | Enginable/phi2_DPO-13B | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-07T15:28:36+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
31,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06646376848220825,
0.2168014943599701,
-0.00225935154594481,
0.023818302899599075,
0.1271018385887146,
-0.001635765191167593,
0.04218708351254463,
0.13324736058712006,
-0.020175931975245476,
0.11144465953111649,
0.046588581055402756,
0.09377603232860565,
0.09928803145885468,
0.18404334783554077,
0.04859916493296623,
-0.2059975117444992,
0.007056170143187046,
-0.09090408682823181,
0.014076028019189835,
0.1116579994559288,
0.13719257712364197,
-0.10291384905576706,
0.08272874355316162,
-0.04045208916068077,
-0.02019004337489605,
0.00012576708104461432,
-0.09259183704853058,
-0.07032395154237747,
0.06885425746440887,
0.06264153122901917,
0.051234472543001175,
0.001456156256608665,
0.09140396863222122,
-0.2864592671394348,
0.017265573143959045,
0.08406311273574829,
0.0027674848679453135,
0.06290827691555023,
0.07236549258232117,
-0.07389893382787704,
0.11328595131635666,
-0.08021481335163116,
0.13019037246704102,
0.08625296503305435,
-0.062064990401268005,
-0.23071379959583282,
-0.07525765895843506,
0.0963398814201355,
0.12251301854848862,
0.06215599179267883,
-0.022921854630112648,
0.15455181896686554,
-0.06248689442873001,
0.012971068732440472,
0.1294165402650833,
-0.11526761949062347,
-0.05572471022605896,
0.061741601675748825,
0.11775490641593933,
0.10740239918231964,
-0.14110268652439117,
-0.0017287094378843904,
0.04900608956813812,
0.029121357947587967,
0.08589313924312592,
0.022661056369543076,
0.12003941088914871,
0.04652795568108559,
-0.13695219159126282,
-0.04037507623434067,
0.12011898308992386,
0.038862764835357666,
-0.06446044892072678,
-0.2168138176202774,
-0.006778308190405369,
-0.0601806715130806,
-0.014732478186488152,
-0.07019448280334473,
0.039128515869379044,
-0.02470310963690281,
0.07317749410867691,
-0.04465159401297569,
-0.1063927412033081,
-0.0421026237308979,
0.0892222449183464,
0.07748593389987946,
0.011527054943144321,
-0.02519804798066616,
0.04627908393740654,
0.13455867767333984,
0.05402068421244621,
-0.10399353504180908,
-0.07017925381660461,
-0.06942764669656754,
-0.09420394152402878,
-0.04035796597599983,
0.056760527193546295,
0.031942449510097504,
0.02665667235851288,
0.22703726589679718,
0.016653569415211678,
0.04155244305729866,
0.0224777739495039,
0.01032855175435543,
0.043662428855895996,
0.0955500528216362,
-0.05303520709276199,
-0.15660029649734497,
-0.04072032496333122,
0.09077946096658707,
-0.0027527001220732927,
-0.036689214408397675,
-0.03966725245118141,
0.03849169611930847,
0.06843466311693192,
0.13122352957725525,
0.07552056759595871,
-0.017929591238498688,
-0.04813180863857269,
-0.030096933245658875,
0.23523783683776855,
-0.1493375599384308,
0.04426715523004532,
-0.02271856553852558,
-0.01804111897945404,
-0.03908449783921242,
0.03597262129187584,
0.022118929773569107,
-0.000004518366949923802,
0.09706240892410278,
-0.058981191366910934,
-0.05378659814596176,
-0.10168042778968811,
-0.03272576630115509,
0.04088849574327469,
-0.013975566253066063,
-0.010589460842311382,
-0.09025166928768158,
-0.09490354359149933,
-0.04766594246029854,
0.05537205561995506,
-0.05123869329690933,
-0.03770573064684868,
0.009465423412621021,
-0.08151785284280777,
-0.005444355774670839,
-0.005417742300778627,
0.10699385404586792,
-0.03222226724028587,
0.04445803165435791,
-0.027600755915045738,
0.05225523188710213,
0.09919606149196625,
0.031576547771692276,
-0.0773419588804245,
0.0561848059296608,
-0.22559374570846558,
0.07503069192171097,
-0.11481974273920059,
0.04335082694888115,
-0.1704932004213333,
-0.042439818382263184,
0.005444696638733149,
0.0139949731528759,
0.013206101022660732,
0.12720820307731628,
-0.19255615770816803,
-0.01654396951198578,
0.13260798156261444,
-0.09212633967399597,
-0.118110790848732,
0.07884611934423447,
-0.029701577499508858,
0.1624738723039627,
0.04682036489248276,
-0.027025915682315826,
0.09224298596382141,
-0.16434773802757263,
-0.07092688232660294,
-0.00949116237461567,
-0.01727987825870514,
0.12109188735485077,
0.07512219995260239,
-0.05991523340344429,
0.046571120619773865,
0.02832140028476715,
-0.038078423589468,
-0.04424772411584854,
-0.050857074558734894,
-0.10884185880422592,
-0.01070026308298111,
-0.08987759798765182,
0.04065500199794769,
-0.01250192429870367,
-0.07916021347045898,
-0.029885273426771164,
-0.18612512946128845,
-0.0030564051121473312,
0.10038342326879501,
0.0035033065360039473,
-0.005652366206049919,
-0.08666291832923889,
0.026358824223279953,
-0.03112892620265484,
-0.008404186926782131,
-0.16764774918556213,
-0.04399421438574791,
0.046902090311050415,
-0.16094985604286194,
0.020117372274398804,
-0.06413903087377548,
0.06334125250577927,
0.03641495108604431,
-0.05590536445379257,
-0.0248766727745533,
-0.01730942726135254,
0.011945613659918308,
-0.05083848536014557,
-0.18994836509227753,
-0.056277405470609665,
-0.037882111966609955,
0.149809330701828,
-0.25956398248672485,
0.032966937869787216,
0.051140617579221725,
0.14649195969104767,
0.00406361510977149,
-0.05115427449345589,
0.01429014839231968,
-0.05360214412212372,
-0.054652128368616104,
-0.06746816635131836,
-0.006135428790003061,
-0.027576493099331856,
-0.05147203803062439,
0.019243421033024788,
-0.1755700707435608,
-0.021410830318927765,
0.09424154460430145,
0.12876708805561066,
-0.1486445665359497,
-0.018640631809830666,
-0.048725154250860214,
-0.06339836865663528,
-0.0715010017156601,
-0.07038594037294388,
0.10712739825248718,
0.0513901449739933,
0.04796046018600464,
-0.07435787469148636,
-0.07092321664094925,
0.02726263552904129,
0.006906150374561548,
-0.03382374346256256,
0.08727246522903442,
0.05199531093239784,
-0.09209315478801727,
0.0756213590502739,
0.1092359870672226,
0.07177663594484329,
0.09363535046577454,
0.01574566215276718,
-0.11756632477045059,
-0.028492970392107964,
0.036266472190618515,
0.02740776725113392,
0.1465986967086792,
-0.05952361226081848,
0.04016614332795143,
0.04494241625070572,
-0.04170418903231621,
0.022319864481687546,
-0.08787637203931808,
0.024075502529740334,
0.025203049182891846,
-0.0034381982404738665,
0.06284574419260025,
-0.02525499276816845,
-0.0050758360885083675,
0.07016654312610626,
0.047779910266399384,
0.04621000960469246,
0.009655474685132504,
-0.01720241829752922,
-0.1047825813293457,
0.16950392723083496,
-0.0951867327094078,
-0.269941508769989,
-0.17632324993610382,
0.026197833940386772,
0.04035249724984169,
-0.022378476336598396,
0.031619444489479065,
-0.07056326419115067,
-0.10630585998296738,
-0.1060405746102333,
-0.002429972169920802,
0.01714223250746727,
-0.06364088505506516,
-0.0741225928068161,
0.07348573952913284,
0.04382912442088127,
-0.14902326464653015,
0.038552410900592804,
0.055694397538900375,
-0.057955220341682434,
-0.0233661737293005,
0.09118817001581192,
0.12397737801074982,
0.14583967626094818,
-0.021366750821471214,
-0.028626007959246635,
0.029004426673054695,
0.19620531797409058,
-0.13469526171684265,
0.10371150821447372,
0.13814030587673187,
-0.04545360431075096,
0.08360563963651657,
0.1560150384902954,
0.029186224564909935,
-0.08317049592733383,
0.05044832453131676,
0.04082648828625679,
-0.043159641325473785,
-0.2666129767894745,
-0.0534592866897583,
0.012832709588110447,
-0.06255637854337692,
0.09786593168973923,
0.10183793306350708,
0.11542957276105881,
0.034910861402750015,
-0.07166364789009094,
-0.043925940990448,
-0.0058974819257855415,
0.11737963557243347,
-0.05490213260054588,
-0.012639665976166725,
0.07686592638492584,
-0.05086168646812439,
0.005355054512619972,
0.10266812145709991,
0.02973790094256401,
0.17442677915096283,
0.020399179309606552,
0.11231429129838943,
0.06195578724145889,
0.08633565157651901,
0.0007386076031252742,
0.02951662428677082,
0.05147615820169449,
0.017203815281391144,
-0.002300140680745244,
-0.10421168059110641,
-0.006156572140753269,
0.1449710875749588,
0.028103826567530632,
0.029669636860489845,
-0.0018948549404740334,
-0.005003341939300299,
0.05121048167347908,
0.1746254414319992,
-0.011592294089496136,
-0.22072425484657288,
-0.0845772922039032,
0.06936841458082199,
-0.06218599155545235,
-0.12968985736370087,
-0.026130788028240204,
0.045467354357242584,
-0.17519839107990265,
0.026703642681241035,
-0.027433741837739944,
0.0919293761253357,
-0.09345759451389313,
-0.02221956104040146,
0.03687324374914169,
0.084866963326931,
-0.014529162086546421,
0.08703910559415817,
-0.14498743414878845,
0.11886418610811234,
0.02978132851421833,
0.09024628251791,
-0.11081171780824661,
0.07909037172794342,
-0.007550720125436783,
0.009180475026369095,
0.19379350543022156,
-0.011335089802742004,
-0.03514958545565605,
-0.08774717897176743,
-0.11210042238235474,
-0.013537433929741383,
0.12687496840953827,
-0.1243172138929367,
0.08773399889469147,
-0.015198243781924248,
-0.044079482555389404,
0.00937260314822197,
-0.12100647389888763,
-0.17273177206516266,
-0.19628387689590454,
0.05585884302854538,
-0.09575839340686798,
0.025643249973654747,
-0.11914430558681488,
-0.07089093327522278,
-0.02952558360993862,
0.241120383143425,
-0.1745356321334839,
-0.06510113179683685,
-0.1468164622783661,
-0.046294767409563065,
0.1662203073501587,
-0.04437198117375374,
0.0718095526099205,
-0.0208172257989645,
0.20345525443553925,
0.005988610442727804,
-0.004939318168908358,
0.06724198162555695,
-0.08892562240362167,
-0.16873881220817566,
-0.06771010160446167,
0.1510489284992218,
0.11680185794830322,
0.04907919466495514,
-0.002248800592496991,
0.0011772146681323647,
-0.016943959519267082,
-0.1137804463505745,
-0.0033210667315870523,
0.16037839651107788,
0.03878779336810112,
0.025986969470977783,
-0.05243593826889992,
-0.08797456324100494,
-0.06899320334196091,
-0.06853509694337845,
0.06221301481127739,
0.19590823352336884,
-0.10376439243555069,
0.1700313836336136,
0.147536963224411,
-0.07305635511875153,
-0.23175598680973053,
0.035342130810022354,
0.04983805492520332,
0.0014306638622656465,
0.04886869341135025,
-0.18252557516098022,
0.10521943867206573,
0.019543392583727837,
-0.05505957826972008,
0.13485197722911835,
-0.1557481735944748,
-0.1552847921848297,
0.0722852572798729,
0.03904085233807564,
-0.22423844039440155,
-0.1354004591703415,
-0.09622503817081451,
-0.05825018882751465,
-0.14065024256706238,
0.06054598465561867,
-0.002136280992999673,
0.015948504209518433,
0.03500790148973465,
-0.0015643214574083686,
0.027123261243104935,
-0.058935679495334625,
0.18609118461608887,
-0.004065449349582195,
0.020676052197813988,
-0.060264769941568375,
-0.0478842556476593,
0.09839435666799545,
-0.06130504235625267,
0.12208222597837448,
0.004057085141539574,
0.01594383642077446,
-0.10362856835126877,
-0.048314861953258514,
-0.04328322783112526,
0.05154227837920189,
-0.07548051327466965,
-0.10070807486772537,
-0.043625857681035995,
0.08841723203659058,
0.07005169242620468,
-0.03383097052574158,
0.00549331633374095,
-0.07189501076936722,
0.10019614547491074,
0.17795267701148987,
0.17573626339435577,
0.009926567785441875,
-0.07241068035364151,
0.01677953451871872,
-0.04142116755247116,
0.044231921434402466,
-0.2513144314289093,
0.03756171092391014,
0.06098250672221184,
0.029438555240631104,
0.09217222779989243,
-0.020435843616724014,
-0.1820858269929886,
-0.04050002992153168,
0.08094815909862518,
-0.05452597141265869,
-0.22617179155349731,
-0.019085140898823738,
0.0954197570681572,
-0.2020406424999237,
-0.007372708059847355,
0.03995226323604584,
-0.048725228756666183,
-0.023169852793216705,
0.00010950004070764408,
0.06317184865474701,
0.002471912419423461,
0.09773622453212738,
0.0735151618719101,
0.09715340286493301,
-0.08337292820215225,
0.10562895983457565,
0.10150538384914398,
-0.09572599828243256,
0.03605884686112404,
0.06754924356937408,
-0.05300498008728027,
-0.043293699622154236,
0.03665391728281975,
0.033023297786712646,
0.005234600510448217,
-0.060321882367134094,
0.013913018628954887,
-0.036497246474027634,
0.044923391193151474,
0.08326134830713272,
0.03754979372024536,
-0.013354414142668247,
0.06462216377258301,
0.03401726484298706,
-0.10898099094629288,
0.10366570204496384,
0.01731540448963642,
0.04105307161808014,
-0.08384523540735245,
-0.019968897104263306,
0.035425446927547455,
0.030576206743717194,
-0.01765924133360386,
-0.02306121215224266,
-0.02860277332365513,
-0.01614218018949032,
-0.14299540221691132,
-0.023106401786208153,
-0.07243485748767853,
0.006181265693157911,
0.014656842686235905,
-0.031884219497442245,
-0.011233693920075893,
0.02475680410861969,
-0.06979699432849884,
-0.07426341623067856,
-0.006949664559215307,
0.09833318740129471,
-0.15115703642368317,
0.008848577737808228,
0.06907843053340912,
-0.11088496446609497,
0.08190931379795074,
-0.008411259390413761,
0.016245156526565552,
0.022527478635311127,
-0.15448406338691711,
0.05601610988378525,
0.0008648968650959432,
0.01916889287531376,
0.025886621326208115,
-0.16471809148788452,
0.004104440100491047,
-0.04661374166607857,
-0.02149827405810356,
-0.00004464812809601426,
-0.02647159807384014,
-0.12325995415449142,
0.06858719140291214,
-0.015622655861079693,
-0.035931166261434555,
-0.02701525390148163,
0.0539589487016201,
0.07888586074113846,
-0.027474910020828247,
0.10445091128349304,
-0.008690856397151947,
0.04941811040043831,
-0.16801609098911285,
-0.02470702864229679,
-0.04982255399227142,
0.019377702847123146,
0.009884213097393513,
-0.007693959400057793,
0.04183054715394974,
-0.00976533442735672,
0.21883612871170044,
-0.05075952783226967,
0.1607085019350052,
0.05847611650824547,
-0.017352959141135216,
-0.0007513365126214921,
0.06180921941995621,
0.05997028574347496,
0.04658793285489082,
0.009480604901909828,
0.023740366101264954,
-0.022450892254710197,
-0.006695089396089315,
-0.15932634472846985,
0.01890849508345127,
0.14999441802501678,
0.06301083415746689,
0.024745315313339233,
0.05866100639104843,
-0.12775006890296936,
-0.12135478109121323,
0.09311001747846603,
-0.026755332946777344,
0.00928465835750103,
-0.08245618641376495,
0.1358020007610321,
0.14980104565620422,
-0.14000412821769714,
0.05256148427724838,
-0.06134212389588356,
-0.05217423290014267,
-0.10388828068971634,
-0.12032219022512436,
-0.05887215584516525,
-0.053666237741708755,
0.002330566756427288,
-0.03760887682437897,
0.054546963423490524,
0.03344334661960602,
-0.009351172484457493,
-0.00022941511997487396,
0.13597318530082703,
-0.019751882180571556,
-0.0028988157864660025,
0.048313532024621964,
0.03693558648228645,
0.02373051457107067,
-0.05275435373187065,
0.02940409444272518,
0.02539868652820587,
0.032232340425252914,
0.06546790152788162,
0.033412106335163116,
-0.047448933124542236,
0.03804153576493263,
-0.0025254099164158106,
-0.11207924783229828,
0.019641218706965446,
-0.00460948096588254,
-0.0742158442735672,
0.1268945336341858,
0.0407399944961071,
0.010224059224128723,
-0.03741471841931343,
0.24361543357372284,
-0.06653323769569397,
-0.06378097087144852,
-0.13251738250255585,
0.10491154342889786,
-0.0027236645109951496,
0.06476365029811859,
0.023412218317389488,
-0.1284150779247284,
0.005243356805294752,
0.13858191668987274,
0.12181595712900162,
0.0045748427510261536,
0.009228081442415714,
0.0518609918653965,
0.0025186820421367884,
-0.06998204439878464,
0.054019294679164886,
0.06992026418447495,
0.12919506430625916,
-0.07847554981708527,
0.07680778950452805,
0.0006860480643808842,
-0.08370215445756912,
-0.02947772853076458,
0.11312682181596756,
-0.0409729965031147,
0.03491825982928276,
-0.047444481402635574,
0.10916327685117722,
-0.05787910893559456,
-0.29412412643432617,
0.02350960113108158,
-0.09588567912578583,
-0.15202060341835022,
-0.018367812037467957,
0.05944539234042168,
-0.02624768204987049,
0.018029648810625076,
0.06971040368080139,
-0.06011629104614258,
0.20098382234573364,
0.0335683599114418,
-0.07864278554916382,
-0.0664360448718071,
0.04837050288915634,
-0.06564252078533173,
0.2949807047843933,
0.008418165147304535,
0.02863333560526371,
0.10770907253026962,
-0.03253700211644173,
-0.18271861970424652,
0.010723991319537163,
0.1133992001414299,
-0.08056149631738663,
0.08200647681951523,
0.19000613689422607,
-0.012578671798110008,
0.1209007054567337,
0.05294662341475487,
-0.047376248985528946,
0.04217283055186272,
-0.03389401361346245,
-0.051268599927425385,
-0.10752558708190918,
0.058453381061553955,
-0.05909625440835953,
0.15447644889354706,
0.10152646154165268,
-0.05671518296003342,
-0.004550917539745569,
-0.05555408447980881,
0.04875178262591362,
0.01804669201374054,
0.12263146042823792,
0.02951994352042675,
-0.1865430772304535,
0.032826557755470276,
-0.01144319772720337,
0.10186848044395447,
-0.25588861107826233,
-0.08421015739440918,
0.08833149075508118,
-0.011924264021217823,
-0.05105875805020332,
0.10560628771781921,
0.057650718837976456,
0.04243382066488266,
-0.043439045548439026,
-0.10480839014053345,
-0.02186836116015911,
0.14663739502429962,
-0.1469624787569046,
-0.025013303384184837
] |
null | null | null |
!pip install -q -U bitsandbytes
| {"license": "apache-2.0"} | null | breebrenda/fineTuneMistralAI | [
"license:apache-2.0",
"region:us"
] | 2024-02-07T15:29:41+00:00 | [] | [] | TAGS
#license-apache-2.0 #region-us
|
!pip install -q -U bitsandbytes
| [] | [
"TAGS\n#license-apache-2.0 #region-us \n"
] | [
14
] | [
"passage: TAGS\n#license-apache-2.0 #region-us \n"
] | [
-0.014972950331866741,
0.1377405971288681,
-0.008658665232360363,
-0.026387644931674004,
-0.06495039910078049,
0.02879125252366066,
0.153837651014328,
0.10483942180871964,
0.1256849616765976,
-0.0962781235575676,
0.15542295575141907,
0.06252618134021759,
0.01139750611037016,
0.01549961045384407,
0.010183668695390224,
-0.10781864821910858,
0.10650338232517242,
-0.03490327671170235,
-0.0707244873046875,
0.013394840992987156,
0.033960238099098206,
0.01981612667441368,
-0.023830559104681015,
-0.012257595546543598,
0.01112176850438118,
0.0037393702659755945,
0.07398658245801926,
-0.03678586333990097,
0.0712570771574974,
-0.0307270847260952,
0.04963022843003273,
0.037079621106386185,
-0.007749613840132952,
-0.26644203066825867,
0.0024822098203003407,
-0.024022065103054047,
-0.07085459679365158,
0.029720399528741837,
0.013389789499342442,
0.01578286662697792,
0.00231059524230659,
0.07178761065006256,
-0.05940864980220795,
0.031044602394104004,
-0.10492347925901413,
-0.29222142696380615,
-0.17006294429302216,
0.01744023710489273,
0.06068360060453415,
0.05476497858762741,
0.09300920367240906,
0.11550862342119217,
-0.14644485712051392,
-0.044547248631715775,
0.07099119573831558,
-0.3628372251987457,
0.050203535705804825,
0.08580845594406128,
-0.01402231678366661,
0.057996317744255066,
0.02550811693072319,
0.06553854793310165,
0.10383764654397964,
-0.038684334605932236,
-0.08282918483018875,
-0.048850782215595245,
-0.07356837391853333,
0.10531846433877945,
-0.00149822561070323,
-0.12009333074092865,
0.345758318901062,
0.07138761878013611,
-0.013982822187244892,
0.1401686668395996,
-0.03401775285601616,
0.08242715895175934,
0.00856063887476921,
0.07128778845071793,
0.11027930676937103,
0.21097713708877563,
0.1913771778345108,
-0.10426254570484161,
-0.16983820497989655,
-0.05970091372728348,
-0.17783339321613312,
0.02624516934156418,
0.006967680528759956,
0.1441962569952011,
-0.15566711127758026,
-0.007344068959355354,
-0.08945375680923462,
-0.05843057110905647,
-0.052450742572546005,
-0.06543421745300293,
0.15477518737316132,
0.08793632686138153,
-0.09496267139911652,
0.12857593595981598,
0.16295284032821655,
0.27066588401794434,
0.0397113561630249,
0.0019443186465650797,
-0.1110435351729393,
0.16749770939350128,
-0.07042741030454636,
0.013758119195699692,
0.11881717294454575,
0.07653947174549103,
0.10706596076488495,
-0.16501086950302124,
0.12452162802219391,
-0.016609763726592064,
-0.11789992451667786,
-0.018261663615703583,
-0.1484517753124237,
0.16352272033691406,
0.06240275129675865,
-0.10497567802667618,
-0.06494749337434769,
0.07674259692430496,
0.11038123816251755,
-0.04079411178827286,
-0.009143678471446037,
-0.015199431218206882,
0.01051195990294218,
-0.009426284581422806,
0.027398476377129555,
0.04705727845430374,
0.058609262108802795,
0.0033366302959620953,
-0.07381071150302887,
-0.020023247227072716,
0.003280751407146454,
0.1267043501138687,
0.1569293588399887,
-0.05185375362634659,
0.04033694788813591,
-0.06643003970384598,
-0.15634682774543762,
0.033434003591537476,
0.08608060330152512,
0.0321158766746521,
-0.007054667454212904,
0.10480355471372604,
0.0473899245262146,
-0.004594990983605385,
-0.090055912733078,
-0.06355541944503784,
-0.08356712758541107,
0.04190095514059067,
-0.11996523290872574,
-0.0058386498130857944,
-0.25945019721984863,
0.00316249905154109,
-0.1388063281774521,
0.0704830139875412,
0.04737875238060951,
-0.11123143136501312,
-0.11911641061306,
0.18649786710739136,
-0.07714612036943436,
0.055908672511577606,
-0.05615519732236862,
-0.008492839522659779,
-0.03699778765439987,
0.06682316213846207,
-0.14747841656208038,
-0.0005399030051194131,
0.18717823922634125,
-0.14897091686725616,
-0.183698832988739,
0.007495424710214138,
0.03250853717327118,
0.014232120476663113,
0.028690440580248833,
0.30018556118011475,
-0.04365801066160202,
-0.022711357101798058,
0.14066553115844727,
0.16189178824424744,
-0.0973799079656601,
-0.2718583643436432,
0.14789041876792908,
-0.18816958367824554,
-0.19481199979782104,
0.02703963965177536,
-0.10175284743309021,
0.06735242903232574,
0.0434052050113678,
-0.12173084169626236,
-0.040761545300483704,
-0.06889492273330688,
-0.03939614072442055,
-0.04728172346949577,
0.01916317269206047,
-0.06263907253742218,
0.06502450257539749,
-0.0881565511226654,
0.0690668523311615,
0.1269703209400177,
0.08447451889514923,
-0.026536764577031136,
0.009126527234911919,
0.025589320808649063,
0.017330490052700043,
-0.03833708539605141,
0.01929517462849617,
0.01688491925597191,
-0.09582766890525818,
0.07343678176403046,
0.10080436617136002,
0.0517999529838562,
-0.10638647526502609,
0.023197486996650696,
0.03179466351866722,
0.0024750155862420797,
0.0694531574845314,
0.06032438576221466,
-0.10287857055664062,
0.06489933282136917,
-0.0037468012887984514,
0.0518820621073246,
0.07478269934654236,
-0.022493045777082443,
0.020172767341136932,
-0.044141460210084915,
-0.04319985210895538,
0.08567604422569275,
-0.019361646845936775,
-0.08695904165506363,
0.02959357015788555,
0.005028232932090759,
0.10603106021881104,
0.04762826859951019,
-0.10406646132469177,
0.16851206123828888,
0.03276379778981209,
0.14172857999801636,
0.16983027756214142,
-0.05003548413515091,
0.13265447318553925,
-0.018061332404613495,
0.011779023334383965,
-0.027860842645168304,
0.08203933387994766,
0.013741283677518368,
-0.0884268656373024,
0.010677173733711243,
-0.0012559969909489155,
-0.04808245599269867,
0.026862075552344322,
-0.05627221614122391,
-0.11836695671081543,
-0.059314463287591934,
-0.028883814811706543,
0.22526615858078003,
-0.10537033528089523,
0.12464691698551178,
0.5217018723487854,
0.024548746645450592,
0.047969620674848557,
-0.16087572276592255,
-0.06556912511587143,
-0.03724734112620354,
0.01234061736613512,
-0.03181084617972374,
0.13119523227214813,
-0.06553105264902115,
0.03862816467881203,
0.0817440003156662,
0.07771708816289902,
0.04734378680586815,
-0.17659273743629456,
-0.12391189485788345,
0.0027862393762916327,
-0.06264690309762955,
-0.1301562488079071,
-0.015394099988043308,
-0.11034313589334488,
0.036011673510074615,
0.015018938109278679,
-0.09842728823423386,
0.16451223194599152,
-0.03630390390753746,
-0.045461565256118774,
0.04944632574915886,
-0.2314302921295166,
-0.0839247852563858,
-0.12943525612354279,
-0.03678453713655472,
-0.023447291925549507,
0.016634400933980942,
0.09078312665224075,
-0.055623311549425125,
-0.0538143664598465,
0.03897318243980408,
-0.08703092485666275,
-0.05625670403242111,
-0.015357088297605515,
0.04862850904464722,
0.08293753117322922,
0.04238975793123245,
-0.10195945203304291,
-0.04045995697379112,
-0.002702324651181698,
-0.01255359873175621,
0.033146876841783524,
-0.07412629574537277,
0.0853082686662674,
0.11064175516366959,
0.049554772675037384,
0.03361808881163597,
-0.0033064892049878836,
0.07316921651363373,
-0.013150024227797985,
-0.06427362561225891,
0.1754506230354309,
-0.014537261798977852,
0.052527204155921936,
0.1614663302898407,
0.06942721456289291,
-0.08693967014551163,
-0.016334567219018936,
-0.05160483345389366,
-0.11219155788421631,
-0.313568115234375,
-0.0514010526239872,
-0.0697566568851471,
0.10370723158121109,
0.016293581575155258,
0.11431120336055756,
0.11805130541324615,
0.05553880333900452,
0.02477053552865982,
-0.030986998230218887,
-0.013935171999037266,
-0.008552669547498226,
0.14244908094406128,
-0.040016934275627136,
-0.025634407997131348,
-0.16582822799682617,
0.05352972075343132,
0.19532184302806854,
0.17346875369548798,
0.16141186654567719,
0.2949795722961426,
0.1219889223575592,
0.1499485820531845,
0.20596453547477722,
0.03277221694588661,
0.08221390843391418,
0.06475520879030228,
0.0147244893014431,
-0.07361224293708801,
-0.02132425457239151,
-0.0358458049595356,
0.09254280477762222,
-0.020926684141159058,
-0.1894368976354599,
0.04358299449086189,
-0.20253854990005493,
0.04529424384236336,
0.1325015276670456,
0.10060963034629822,
0.0372280478477478,
0.12041204422712326,
0.10496384650468826,
0.07236151397228241,
0.02265322208404541,
0.155935138463974,
-0.10047086328268051,
-0.04170655459165573,
0.10314960777759552,
0.03088374249637127,
0.08923831582069397,
0.05249778553843498,
0.024546442553400993,
-0.09486964344978333,
-0.18898503482341766,
0.08735544979572296,
0.15138362348079681,
-0.18271131813526154,
0.25446999073028564,
0.007140269037336111,
-0.10152944922447205,
-0.04616072401404381,
-0.034160908311605453,
0.07485716044902802,
0.1744716465473175,
0.09998820722103119,
0.07455950975418091,
-0.2311946600675583,
0.06371928006410599,
-0.0805746391415596,
0.03850121796131134,
0.009798256680369377,
-0.0012644194066524506,
-0.1523386389017105,
-0.06277230381965637,
0.03955406695604324,
0.028464701026678085,
0.16074317693710327,
-0.09644434601068497,
-0.07261957228183746,
0.0019746189936995506,
0.14707164466381073,
-0.027826759964227676,
-0.12161600589752197,
0.07778380811214447,
0.026536058634519577,
0.10129949450492859,
-0.046719279140233994,
0.01687805913388729,
-0.0395737923681736,
-0.22956150770187378,
0.06549007445573807,
-0.02001434937119484,
0.016561396420001984,
-0.057295773178339005,
-0.095908522605896,
-0.09420177340507507,
-0.1935003399848938,
0.09841206669807434,
-0.0820716917514801,
0.02724429965019226,
-0.03322573006153107,
0.12508079409599304,
-0.09227079153060913,
0.022051848471164703,
0.002473256317898631,
-0.0009367846651002765,
-0.0559639148414135,
-0.12471262365579605,
0.08924731612205505,
-0.02971985563635826,
-0.0013593619223684072,
-0.004879474639892578,
-0.037194594740867615,
0.05458180233836174,
0.06104811653494835,
-0.09029456228017807,
0.17284651100635529,
0.304045170545578,
-0.06797412037849426,
0.21419626474380493,
0.3437173664569855,
-0.12620802223682404,
-0.22579985857009888,
-0.2027367204427719,
-0.287395715713501,
-0.1499757170677185,
0.09507355093955994,
-0.18301111459732056,
0.10197417438030243,
0.19583699107170105,
-0.1618303805589676,
0.17492994666099548,
-0.18311083316802979,
-0.021827004849910736,
0.20329692959785461,
-0.06383024156093597,
0.3788199722766876,
-0.11906247586011887,
-0.10778811573982239,
-0.09945111721754074,
-0.1618940234184265,
0.10558515042066574,
-0.18797975778579712,
0.02240607887506485,
0.03505697101354599,
-0.06375153362751007,
-0.049790963530540466,
-0.018343381583690643,
0.24272295832633972,
-0.001803387189283967,
0.0726020559668541,
-0.07976466417312622,
0.016474680975079536,
0.1855529397726059,
-0.05465783178806305,
0.036813490092754364,
-0.15827761590480804,
-0.028589751571416855,
-0.009538229554891586,
0.037033237516880035,
-0.03485805541276932,
0.07510014623403549,
0.0011767554096877575,
-0.07188761979341507,
-0.0968698114156723,
-0.021874895319342613,
-0.04847799241542816,
-0.005289438646286726,
0.26872649788856506,
0.076040118932724,
-0.053988635540008545,
0.10166085511445999,
-0.0624149851500988,
-0.1741536259651184,
0.016763299703598022,
-0.10044022649526596,
-0.07138258218765259,
0.0554080568253994,
-0.24670164287090302,
0.034909263253211975,
0.05687393248081207,
-0.06160387769341469,
0.03683772310614586,
0.05927642062306404,
-0.09550898522138596,
-0.021464845165610313,
0.12709221243858337,
-0.0498422347009182,
-0.0725388154387474,
0.06216653808951378,
0.1259109526872635,
0.11850999295711517,
0.0343933068215847,
0.07898429036140442,
0.04615308716893196,
0.00829920545220375,
0.021227914839982986,
0.07433275133371353,
-0.17039799690246582,
-0.05188438296318054,
0.05058757960796356,
-0.02595781534910202,
-0.1195429265499115,
0.25599008798599243,
0.024486735463142395,
-0.03208741545677185,
-0.03641199693083763,
0.034413471817970276,
-0.05278664082288742,
-0.09139014035463333,
-0.055758245289325714,
-0.012565652839839458,
-0.09509796649217606,
-0.18549469113349915,
0.037074074149131775,
-0.09632396697998047,
-0.031133420765399933,
-0.03407083451747894,
0.1050528809428215,
0.10973811894655228,
0.0608430951833725,
-0.035467516630887985,
0.16503585875034332,
-0.08250437676906586,
-0.18473856151103973,
-0.014520380645990372,
-0.04597662389278412,
-0.2016236037015915,
0.02965214103460312,
0.06940829008817673,
-0.013816002756357193,
-0.046515315771102905,
-0.05482683703303337,
0.08873733133077621,
-0.2119728922843933,
0.016329854726791382,
-0.08455836027860641,
0.001034914399497211,
0.0714198648929596,
-0.0625232607126236,
-0.02158266305923462,
0.019058139994740486,
-0.16247235238552094,
-0.057285383343696594,
-0.010741667822003365,
0.05505165457725525,
-0.10866090655326843,
-0.059575166553258896,
0.13569991290569305,
0.05747954174876213,
0.09156093001365662,
0.08844062685966492,
0.012601320631802082,
0.14171354472637177,
-0.1320880800485611,
-0.07235664874315262,
0.07277880609035492,
0.03337378427386284,
-0.02606162242591381,
0.016614673659205437,
-0.08515988290309906,
0.08969198167324066,
-0.07962571084499359,
0.01820359006524086,
-0.060848966240882874,
-0.13395224511623383,
-0.14613622426986694,
-0.008169831708073616,
-0.17318809032440186,
0.04386909306049347,
-0.1962924599647522,
0.20117418467998505,
0.049670200794935226,
0.11739125847816467,
0.09586193412542343,
0.0011940872063860297,
0.02116451971232891,
0.03084888681769371,
-0.038015320897102356,
-0.06214485689997673,
-0.1336703896522522,
-0.034515380859375,
-0.14342233538627625,
-0.05508563295006752,
0.3311520218849182,
-0.04312245175242424,
-0.13223427534103394,
0.05347722768783569,
0.08347365260124207,
0.016498390585184097,
0.023317286744713783,
0.24889890849590302,
0.04722274839878082,
0.01573793776333332,
-0.13385194540023804,
-0.026428358629345894,
0.01794368028640747,
-0.16727277636528015,
0.06783340871334076,
0.09281940758228302,
0.16844220459461212,
0.05829422175884247,
0.05544407665729523,
-0.019741542637348175,
-0.05980636551976204,
-0.07777530699968338,
0.14353948831558228,
0.032101042568683624,
0.07295279204845428,
0.09418365359306335,
0.159424290060997,
-0.01132612582296133,
0.011219488456845284,
-0.048546407371759415,
0.018409814685583115,
-0.15742093324661255,
-0.13183769583702087,
0.0072431364096701145,
-0.15321512520313263,
0.010697826743125916,
0.011604762636125088,
0.031117310747504234,
0.2556179463863373,
0.040744051337242126,
-0.06757266819477081,
-0.061626169830560684,
-0.15643614530563354,
-0.04558565840125084,
-0.04036465659737587,
-0.007026704493910074,
-0.038584496825933456,
-0.04671114683151245,
-0.11182798445224762,
-0.0216030515730381,
-0.08770470321178436,
-0.06560606509447098,
0.03681713715195656,
0.030047502368688583,
0.026680879294872284,
-0.10008109360933304,
-0.026802673935890198,
-0.08784958720207214,
0.040374286472797394,
0.0004315301775932312,
0.18315750360488892,
0.03631989657878876,
0.0276914332062006,
0.1346033811569214,
0.07949449867010117,
-0.04593978822231293,
-0.1404203623533249,
-0.04276864603161812,
0.050976864993572235,
-0.04557491093873978,
0.06821722537279129,
-0.047834381461143494,
-0.008090421557426453,
-0.03426219895482063,
0.22107382118701935,
0.2087077796459198,
-0.07518038153648376,
-0.0023365700617432594,
-0.04285150393843651,
0.01813213713467121,
0.007908736355602741,
0.15072014927864075,
0.0591626912355423,
0.10565771162509918,
-0.0679246187210083,
-0.012176652438938618,
-0.016756610944867134,
0.013343945145606995,
-0.18723566830158234,
0.07582846283912659,
-0.030456462875008583,
-0.11726406216621399,
-0.038775019347667694,
0.12641341984272003,
-0.06347712874412537,
0.09165021777153015,
0.09027697890996933,
-0.029973473399877548,
0.027466343715786934,
0.0041959225200116634,
0.20493634045124054,
0.02617124654352665,
0.048617783933877945,
-0.1167941614985466,
-0.09022688120603561,
0.04032917320728302,
0.00874406099319458,
-0.3118123412132263,
-0.1954881250858307,
0.10563000291585922,
0.06715425848960876,
0.2813838720321655,
0.0376100093126297,
0.04810243472456932,
0.015465166419744492,
0.07121489197015762,
-0.12309351563453674,
0.13339413702487946,
0.046601615846157074,
-0.01129146758466959,
-0.11312273144721985,
-0.22528712451457977,
-0.08437834680080414,
-0.05826718360185623,
0.07650711387395859,
0.06533077359199524,
0.0024006948806345463,
0.1862800121307373,
-0.0665736049413681,
-0.02629813738167286,
-0.025829114019870758,
-0.15687870979309082,
0.03671088442206383,
-0.06590475142002106,
-0.042691074311733246,
-0.0751107782125473,
-0.03413062542676926,
-0.023224810138344765,
0.055103402584791183,
-0.24346224963665009,
-0.05677183344960213,
0.2217833697795868,
0.021673565730452538,
0.14538446068763733,
0.02648567035794258,
0.03698272630572319,
-0.035830456763505936,
-0.04817749559879303,
0.054671525955200195,
-0.08730608224868774,
0.004225281998515129,
0.10913265496492386,
-0.03198149800300598,
0.021041326224803925,
-0.1597670465707779,
0.046472933143377304,
-0.043475229293107986,
-0.005544837564229965,
-0.0814170315861702
] |
null | null | generic |
# Fork of [Salesforce/blip-image-captioning-base](https://huggingface.co/Salesforce/blip-image-captioning-base) for a `image-to-text` Inference endpoint.
> Inspired by https://huggingface.co/sergeipetrov/blip_captioning
This repository implements a `custom` task for `image-to-text` for 🤗 Inference Endpoints to allow image capturing.
The code for the customized pipeline is in the handler.py.
To use deploy this model an Inference Endpoint you have to select `Custom` as task to use the `handler.py` file.
### expected Request payload
Image to be labeled as binary.
#### CURL
```
curl URL \
-X POST \
--data-binary @car.png \
-H "Content-Type: image/png"
```
#### Python
```python
requests.post(ENDPOINT_URL, headers={"Content-Type": "image/png"}, data=open("car.png", 'rb').read()).json()
``` | {"library_name": "generic", "tags": ["vision", "image-to-text", "endpoints-template"], "inference": false, "pipeline_tag": "image-to-text", "base_model": "Salesforce/blip-image-captioning-base"} | image-to-text | pimcore/IEP__image-capturing-base | [
"generic",
"vision",
"image-to-text",
"endpoints-template",
"base_model:Salesforce/blip-image-captioning-base",
"endpoints_compatible",
"region:us"
] | 2024-02-07T15:30:01+00:00 | [] | [] | TAGS
#generic #vision #image-to-text #endpoints-template #base_model-Salesforce/blip-image-captioning-base #endpoints_compatible #region-us
|
# Fork of Salesforce/blip-image-captioning-base for a 'image-to-text' Inference endpoint.
> Inspired by URL
This repository implements a 'custom' task for 'image-to-text' for Inference Endpoints to allow image capturing.
The code for the customized pipeline is in the URL.
To use deploy this model an Inference Endpoint you have to select 'Custom' as task to use the 'URL' file.
### expected Request payload
Image to be labeled as binary.
#### CURL
#### Python
| [
"# Fork of Salesforce/blip-image-captioning-base for a 'image-to-text' Inference endpoint.\n\n> Inspired by URL\n\nThis repository implements a 'custom' task for 'image-to-text' for Inference Endpoints to allow image capturing. \nThe code for the customized pipeline is in the URL.\n\nTo use deploy this model an Inference Endpoint you have to select 'Custom' as task to use the 'URL' file.",
"### expected Request payload\n\nImage to be labeled as binary.",
"#### CURL",
"#### Python"
] | [
"TAGS\n#generic #vision #image-to-text #endpoints-template #base_model-Salesforce/blip-image-captioning-base #endpoints_compatible #region-us \n",
"# Fork of Salesforce/blip-image-captioning-base for a 'image-to-text' Inference endpoint.\n\n> Inspired by URL\n\nThis repository implements a 'custom' task for 'image-to-text' for Inference Endpoints to allow image capturing. \nThe code for the customized pipeline is in the URL.\n\nTo use deploy this model an Inference Endpoint you have to select 'Custom' as task to use the 'URL' file.",
"### expected Request payload\n\nImage to be labeled as binary.",
"#### CURL",
"#### Python"
] | [
51,
114,
16,
4,
3
] | [
"passage: TAGS\n#generic #vision #image-to-text #endpoints-template #base_model-Salesforce/blip-image-captioning-base #endpoints_compatible #region-us \n# Fork of Salesforce/blip-image-captioning-base for a 'image-to-text' Inference endpoint.\n\n> Inspired by URL\n\nThis repository implements a 'custom' task for 'image-to-text' for Inference Endpoints to allow image capturing. \nThe code for the customized pipeline is in the URL.\n\nTo use deploy this model an Inference Endpoint you have to select 'Custom' as task to use the 'URL' file.### expected Request payload\n\nImage to be labeled as binary.#### CURL#### Python"
] | [
-0.021503081545233727,
-0.04389744997024536,
-0.0018433758523315191,
0.03415074944496155,
0.11996952444314957,
0.04011815786361694,
-0.006862712558358908,
0.034829117357730865,
0.08094219118356705,
-0.02605469897389412,
0.12322847545146942,
0.15208914875984192,
0.0204868596047163,
0.09145643562078476,
-0.011768252588808537,
-0.26651209592819214,
-0.013413612730801105,
0.09524278342723846,
0.12220582365989685,
0.06662532687187195,
0.03499610722064972,
-0.058931346982717514,
0.15626150369644165,
-0.04564923793077469,
-0.16009128093719482,
0.06357908248901367,
0.006692121736705303,
0.04007294401526451,
0.030593646690249443,
0.06701017171144485,
0.06477458029985428,
-0.02527417242527008,
0.07074269652366638,
-0.13174380362033844,
0.027087751775979996,
0.0317542627453804,
-0.08221958577632904,
-0.0023744383361190557,
-0.014704283326864243,
-0.06887441128492355,
0.1425328105688095,
-0.045773301273584366,
0.04924384877085686,
0.024063555523753166,
-0.06590507924556732,
0.06094904616475105,
0.11646943539381027,
0.20785129070281982,
0.10781147330999374,
0.01009036973118782,
0.06109561398625374,
-0.07632564008235931,
-0.005075028631836176,
0.15129759907722473,
0.032965559512376785,
-0.11825628578662872,
-0.00849708542227745,
0.1769006997346878,
0.01626131869852543,
0.04494425281882286,
-0.03707713633775711,
-0.0019441103795543313,
-0.011084116995334625,
0.0681007131934166,
0.021777279675006866,
-0.07400334626436234,
-0.10165957361459732,
-0.011114728637039661,
-0.04619657248258591,
-0.13851791620254517,
0.12474982440471649,
0.016918664798140526,
-0.04968498274683952,
-0.0007552293827757239,
-0.08419220894575119,
-0.054340314120054245,
-0.12239397317171097,
0.12332730740308762,
0.03716569021344185,
0.06453161686658859,
-0.03366051986813545,
-0.06669935584068298,
-0.06783315539360046,
-0.06977982074022293,
-0.019901633262634277,
-0.014746428467333317,
-0.034197334200143814,
0.12265684455633163,
-0.12423412501811981,
0.1626981645822525,
-0.0998847559094429,
-0.08969051390886307,
-0.04021294787526131,
-0.10529539734125137,
0.024498717859387398,
0.034475162625312805,
0.010553887113928795,
-0.08636463433504105,
-0.01744367554783821,
0.032521139830350876,
0.14963726699352264,
0.03382396325469017,
-0.07427108287811279,
0.12113410234451294,
0.05624401196837425,
0.056693535298109055,
-0.02442942187190056,
-0.03695201873779297,
0.11357064545154572,
-0.062387675046920776,
0.02575368992984295,
0.024001043289899826,
-0.14603158831596375,
-0.06708681583404541,
-0.03268410637974739,
-0.03569909185171127,
0.001997205661609769,
0.10760299116373062,
-0.058189477771520615,
-0.08510618656873703,
0.12732382118701935,
-0.015731412917375565,
-0.030403491109609604,
-0.024059811607003212,
-0.049177467823028564,
0.07788796722888947,
0.19157880544662476,
-0.07975068688392639,
-0.0502128042280674,
0.09673118591308594,
-0.05223369225859642,
-0.007092425134032965,
-0.12167734652757645,
-0.03219850733876228,
0.0018352476181462407,
-0.07714110612869263,
0.06558551639318466,
-0.11025748401880264,
-0.3180733621120453,
0.012500620447099209,
0.13507595658302307,
-0.003982876893132925,
-0.020926235243678093,
-0.005491221323609352,
0.04397105053067207,
-0.02110511250793934,
-0.054070476442575455,
0.0709192082285881,
-0.08691968768835068,
0.07448375970125198,
-0.1032373383641243,
0.10412702709436417,
-0.1623128354549408,
-0.0016846244689077139,
-0.07771078497171402,
0.06274875998497009,
-0.1997348666191101,
0.1143999770283699,
0.01997729204595089,
0.15377196669578552,
-0.020022882148623466,
-0.04352714866399765,
-0.022046931087970734,
0.014222859404981136,
-0.002611920703202486,
0.11607113480567932,
-0.11844857037067413,
0.06025487557053566,
0.11137431859970093,
-0.06525588780641556,
-0.09321561455726624,
-0.01692969724535942,
-0.073109470307827,
0.10182827711105347,
0.07018116116523743,
0.07931074500083923,
0.002751487772911787,
-0.16664020717144012,
0.04846173897385597,
0.10805509239435196,
-0.18647488951683044,
-0.11592231690883636,
-0.03755764290690422,
0.09140787273645401,
-0.018401673063635826,
0.07218924164772034,
-0.0435054749250412,
0.06345295161008835,
0.012221056967973709,
0.004222316201776266,
-0.06646116077899933,
-0.032924458384513855,
-0.10972601175308228,
0.022536756470799446,
-0.006308458745479584,
0.05139961466193199,
-0.03897359222173691,
0.0856989324092865,
0.02770773321390152,
-0.07844474166631699,
0.023579230532050133,
0.009189200587570667,
0.10522029548883438,
-0.1866413950920105,
-0.020657820627093315,
-0.11876927316188812,
-0.049466922879219055,
-0.04247879236936569,
0.18584735691547394,
0.08375295251607895,
-0.07345562428236008,
0.022750334814190865,
-0.03778965026140213,
0.023772817105054855,
-0.016368364915251732,
0.07269470393657684,
-0.054410308599472046,
-0.04325907677412033,
-0.09215301275253296,
-0.13773931562900543,
-0.01875240169465542,
-0.08941427618265152,
-0.11371596902608871,
0.0003628453123383224,
0.1152934804558754,
0.0888722836971283,
0.020387869328260422,
0.016510579735040665,
0.02011471427977085,
-0.0914047583937645,
-0.021779872477054596,
-0.049688179045915604,
0.10691533982753754,
0.03614867851138115,
0.04453137889504433,
0.01486269198358059,
0.04715608060359955,
0.042419757694005966,
0.12640121579170227,
-0.2317027896642685,
-0.011214439757168293,
-0.10879044979810715,
-0.042146384716033936,
-0.0012560384348034859,
0.028496110811829567,
-0.03723881393671036,
0.13847123086452484,
0.05724252760410309,
0.11736354231834412,
-0.09034198522567749,
0.0062654162757098675,
0.028268735855817795,
-0.03525625914335251,
0.012336988933384418,
-0.0882214680314064,
0.08586381375789642,
-0.23492442071437836,
0.07154926657676697,
-0.08618871867656708,
-0.007444242015480995,
0.16507558524608612,
0.05614820867776871,
-0.029642295092344284,
0.08058363944292068,
0.07589880377054214,
-0.037783488631248474,
0.20785456895828247,
-0.2115929126739502,
-0.06998051702976227,
0.0527508407831192,
-0.1083429828286171,
0.07913167774677277,
-0.08008065819740295,
0.031069515272974968,
0.004710198845714331,
-0.006457386072725058,
0.007244420237839222,
-0.0383169949054718,
-0.0035397044848650694,
0.04994494095444679,
0.06062932685017586,
0.009856652468442917,
0.06467174738645554,
-0.053660303354263306,
-0.08826734125614166,
0.1399967521429062,
0.0207405723631382,
-0.17520691454410553,
-0.16249755024909973,
-0.21502012014389038,
-0.11242745816707611,
0.040973398834466934,
-0.0009642126387916505,
0.0363970585167408,
-0.057245735079050064,
-0.03206051513552666,
-0.038381606340408325,
0.046810783445835114,
-0.04949437826871872,
-0.2342570424079895,
0.05418366938829422,
-0.027019551023840904,
-0.08908532559871674,
-0.02365550771355629,
-0.027186933904886246,
0.08025163412094116,
0.0466708280146122,
-0.11406879872083664,
0.08910385519266129,
0.09963147342205048,
-0.028959790244698524,
0.04247858002781868,
0.02559473179280758,
0.2105673998594284,
-0.03358198702335358,
0.06005523353815079,
0.10389289259910583,
-0.0484711155295372,
0.039460547268390656,
0.21871456503868103,
-0.003845663508400321,
-0.12222110480070114,
0.03353538736701012,
-0.04480607062578201,
-0.09553369879722595,
-0.0072856186889112,
-0.10568634420633316,
-0.12300512939691544,
-0.07779771834611893,
0.20533354580402374,
0.041105639189481735,
-0.02724236622452736,
0.16668805480003357,
-0.01806337758898735,
0.1684056669473648,
-0.043600283563137054,
0.20426678657531738,
0.052076034247875214,
-0.03622150793671608,
0.03342096880078316,
-0.03546226769685745,
0.009765240363776684,
0.044630225747823715,
0.19145353138446808,
0.12718138098716736,
-0.11343477666378021,
-0.023283351212739944,
0.06952109187841415,
0.05406992509961128,
0.12162768095731735,
0.15532046556472778,
-0.12090078741312027,
0.023987198248505592,
0.006961738225072622,
-0.03409701585769653,
-0.13636811077594757,
0.08697187900543213,
-0.06265180557966232,
0.027480710297822952,
0.01194210909307003,
-0.14284655451774597,
-0.018786128610372543,
0.24301765859127045,
0.009728061966598034,
-0.36480528116226196,
-0.05153223127126694,
0.07660230994224548,
0.033121075481176376,
-0.17227520048618317,
0.006437270436435938,
0.018428213894367218,
-0.09157652407884598,
0.03581065312027931,
-0.014645176008343697,
0.14794456958770752,
-0.04385875165462494,
-0.0038185424637049437,
0.03309082239866257,
-0.05678306147456169,
0.00551055371761322,
0.07411754876375198,
-0.0801728367805481,
0.0469924621284008,
0.029449691995978355,
0.015008872374892235,
-0.10858386009931564,
0.0035357805900275707,
-0.036300186067819595,
0.18804220855236053,
0.13130469620227814,
-0.008160458877682686,
0.10287144035100937,
-0.01986866071820259,
0.049098335206508636,
0.023744819685816765,
-0.021392177790403366,
0.04622998088598251,
0.06038111448287964,
0.05724038556218147,
-0.04095647484064102,
-0.05563326179981232,
0.012172691524028778,
0.036991801112890244,
-0.17916157841682434,
-0.08382469415664673,
-0.08221238106489182,
0.12714672088623047,
0.045836254954338074,
0.12170232087373734,
0.24917806684970856,
0.1840343326330185,
-0.07532120496034622,
-0.14719168841838837,
-0.10706339031457901,
0.08912893384695053,
-0.00721032964065671,
0.004386797547340393,
0.10198865830898285,
-0.0944317951798439,
0.09424520283937454,
-0.02703886665403843,
-0.18005986511707306,
0.07505838572978973,
-0.08705279231071472,
0.08503159880638123,
-0.06032726913690567,
0.06355742365121841,
-0.07148648053407669,
-0.06780724227428436,
0.04733849689364433,
-0.03896291181445122,
-0.08040894567966461,
-0.09119538962841034,
-0.06692000478506088,
0.06497788429260254,
0.18626390397548676,
0.06388729810714722,
-0.12418626993894577,
-0.053456950932741165,
0.04439187049865723,
0.1510360836982727,
0.08544736355543137,
-0.010354908183217049,
-0.04133138433098793,
0.023749543353915215,
0.0289745032787323,
-0.04727252572774887,
-0.18582670390605927,
0.016649464145302773,
0.007988804951310158,
0.01123803947120905,
-0.13023199141025543,
-0.04362025484442711,
0.14744365215301514,
-0.08650176227092743,
-0.003633772488683462,
0.1933993548154831,
-0.13191676139831543,
-0.0883256271481514,
0.0452139750123024,
0.17195169627666473,
0.15396951138973236,
-0.09443829208612442,
-0.008099475875496864,
-0.13589023053646088,
-0.28326109051704407,
0.13333627581596375,
0.015702933073043823,
0.05015214532613754,
-0.0324212871491909,
0.15217988193035126,
0.05295368656516075,
-0.04816032946109772,
0.0011884297709912062,
-0.003004615195095539,
0.11080283671617508,
-0.0733913779258728,
0.015060101635754108,
0.0680205300450325,
-0.04045036435127258,
0.12381017208099365,
-0.012918202206492424,
0.048673730343580246,
-0.1855502724647522,
-0.044570792466402054,
-0.06630782037973404,
0.028944160789251328,
0.015306653454899788,
-0.10583627223968506,
-0.06244058161973953,
-0.034937784075737,
0.10543807595968246,
0.09536492824554443,
0.11211048811674118,
-0.12254205346107483,
-0.08857567608356476,
0.17607873678207397,
0.003382408292964101,
-0.12456545233726501,
-0.1938641369342804,
-0.042512714862823486,
-0.03576961159706116,
0.10853822529315948,
-0.24568785727024078,
0.014251493848860264,
0.054214976727962494,
0.04501355439424515,
0.05541060492396355,
0.08454694598913193,
0.013580926693975925,
0.0019921697676181793,
0.0675012543797493,
-0.08978858590126038,
0.10002855956554413,
-0.007510564289987087,
0.06376180797815323,
-0.017752060666680336,
-0.05400899425148964,
0.08858044445514679,
-0.03924285247921944,
-0.0014488701708614826,
-0.007951048202812672,
-0.006327996030449867,
-0.13284151256084442,
-0.059821028262376785,
0.06942471116781235,
0.013459748588502407,
-0.07937583327293396,
0.0037417125422507524,
0.005941685754805803,
-0.10461942851543427,
-0.05997776612639427,
0.0663963109254837,
-0.1315920501947403,
-0.0784233883023262,
-0.040731798857450485,
0.12311964482069016,
-0.15757425129413605,
0.030423859134316444,
0.029374979436397552,
-0.018162278458476067,
0.0469542033970356,
0.15842607617378235,
0.04647153988480568,
-0.021352354437112808,
-0.029519276693463326,
-0.03646547347307205,
-0.08536122739315033,
0.0024154013954102993,
-0.048717088997364044,
0.07703190296888351,
-0.11728421598672867,
-0.031579356640577316,
0.02946288138628006,
0.07764166593551636,
-0.06777041405439377,
-0.029366308823227882,
-0.15282104909420013,
0.05547838658094406,
-0.15650400519371033,
0.12227871268987656,
-0.0875365138053894,
0.017379039898514748,
-0.02141553908586502,
0.07108990848064423,
-0.05175192654132843,
0.030192993581295013,
-0.0647827759385109,
0.020187977701425552,
0.048103392124176025,
-0.0018120782915502787,
-0.020445849746465683,
-0.010436145588755608,
0.0034483992494642735,
-0.06833513081073761,
0.02381535805761814,
0.03678971529006958,
-0.0766742080450058,
-0.0366375632584095,
-0.10451191663742065,
-0.06646224856376648,
0.07388988137245178,
0.018284885212779045,
0.005689146462827921,
0.19495880603790283,
0.07419245690107346,
0.040894556790590286,
-0.024977464228868484,
-0.06562250107526779,
0.07955454289913177,
-0.09361635893583298,
0.06974434852600098,
0.016740312799811363,
0.033536896109580994,
-0.07925129681825638,
0.0635046437382698,
0.016904080286622047,
0.10038721561431885,
0.035126447677612305,
-0.04810497537255287,
0.028074536472558975,
-0.09027449041604996,
-0.03943576291203499,
0.02247723937034607,
-0.00043140281923115253,
0.07653896510601044,
-0.09850727021694183,
-0.02340516448020935,
-0.0269511379301548,
0.2323211282491684,
0.08564811199903488,
0.047665901482105255,
-0.019428549334406853,
0.13289158046245575,
0.145407035946846,
0.01952332817018032,
0.21906757354736328,
-0.06016908958554268,
0.0370122529566288,
0.014370397664606571,
0.10729886591434479,
0.057536885142326355,
0.144913911819458,
-0.01862371899187565,
0.10008355230093002,
-0.0019021416082978249,
0.07443711161613464,
-0.007754830177873373,
0.07996959984302521,
0.002821049652993679,
-0.09864087402820587,
0.08166442811489105,
0.08765210956335068,
-0.11646842956542969,
0.052465759217739105,
0.14563043415546417,
-0.0913572683930397,
0.04196302220225334,
0.12868589162826538,
-0.06491167098283768,
0.01593606546521187,
-0.28285616636276245,
-0.006667500361800194,
-0.1476619988679886,
0.050632987171411514,
-0.061830684542655945,
-0.007767158094793558,
0.1063869372010231,
-0.0224405936896801,
-0.02198215015232563,
0.07169029861688614,
-0.061379574239254,
-0.07409995794296265,
0.05507275089621544,
0.04466159641742706,
-0.010280655696988106,
0.03736187517642975,
0.10328838974237442,
0.025668315589427948,
-0.11911839246749878,
-0.0394282229244709,
0.026176398620009422,
0.13934533298015594,
-0.012819594703614712,
0.0427568145096302,
-0.0023535373620688915,
-0.034069307148456573,
0.06351560354232788,
-0.05768303945660591,
0.2280074954032898,
-0.011441449634730816,
-0.001054501859471202,
-0.030293481424450874,
0.1318543553352356,
-0.04479812830686569,
-0.10446509718894958,
-0.056239716708660126,
0.1091061607003212,
-0.03176872059702873,
-0.015855269506573677,
-0.046084269881248474,
-0.07027299702167511,
0.0187839362770319,
0.2871204614639282,
0.0539223738014698,
-0.11427117139101028,
0.002449166728183627,
0.01737164705991745,
0.02166791260242462,
0.02420433983206749,
0.12478283047676086,
-0.00003769147224375047,
0.3579959273338318,
-0.06151801347732544,
-0.040804509073495865,
-0.008118144236505032,
-0.05874905735254288,
-0.045499831438064575,
-0.0716773122549057,
-0.01770544983446598,
-0.11475741863250732,
-0.05520298704504967,
0.04781680554151535,
-0.2782486081123352,
0.06862454116344452,
0.14479613304138184,
-0.06675297766923904,
0.02277381718158722,
-0.07624335587024689,
-0.11763692647218704,
-0.004716651048511267,
0.02829209715127945,
-0.08537538349628448,
-0.05089329928159714,
0.1407926380634308,
-0.02798517607152462,
-0.24284963309764862,
-0.07538514584302902,
0.00034307059831917286,
-0.10403354465961456,
0.15980032086372375,
-0.06340960413217545,
0.05333954468369484,
0.0334131233394146,
0.01847025565803051,
-0.07069801539182663,
-0.048245176672935486,
-0.042626868933439255,
-0.08436589688062668,
0.01862270198762417,
0.08523433655500412,
-0.046771224588155746,
-0.037943389266729355,
-0.07785143703222275,
-0.14890515804290771,
0.013395056128501892,
0.05756869912147522,
0.054132137447595596,
-0.11161786317825317,
-0.07962004095315933,
-0.10902953147888184,
0.08655108511447906,
0.0360206700861454,
0.0004887569812126458,
-0.05559688061475754,
-0.09089983254671097,
0.10841181874275208,
0.08145514875650406,
0.0010702927829697728,
-0.003145382972434163,
0.008811608888208866,
-0.059821318835020065,
0.0023925332352519035,
-0.07435590028762817,
-0.16504442691802979,
0.0035085175186395645,
-0.021265992894768715,
-0.003893821267411113,
-0.04154835268855095,
0.07934194803237915,
0.1214897483587265,
0.06492958217859268,
-0.051433973014354706,
-0.08248630911111832,
0.04057253897190094,
0.06992188096046448,
-0.05689972639083862,
-0.14816558361053467
] |
null | null | transformers | (Note: From short testing, this Alt version generated much better code)
Alternate version of DeepMagic-Coder-7b which can be found bellow.
- https://huggingface.co/rombodawg/DeepMagic-Coder-7b

This version uses a diffrent config setup, with the actual base model of the two merges as the "base_model". Test both for yourself and see which is better at coding. Benchmarks coming soon.
Config can be found bellow:
```yaml
models:
- model: deepseek-ai_deepseek-coder-6.7b-instruct
parameters:
weight: 1
- model: ise-uiuc_Magicoder-S-DS-6.7B
parameters:
weight: 1
merge_method: task_arithmetic
base_model: deepseek-ai_deepseek-coder-6.7b-base
parameters:
normalize: true
int8_mask: true
dtype: float16
```
| {"license": "other", "license_name": "deepseek", "license_link": "https://github.com/deepseek-ai/DeepSeek-Coder/blob/main/LICENSE-MODEL"} | text-generation | LoneStriker/DeepMagic-Coder-7b-Alt-8.0bpw-h8-exl2 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"license:other",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T15:31:58+00:00 | [] | [] | TAGS
#transformers #safetensors #llama #text-generation #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| (Note: From short testing, this Alt version generated much better code)
Alternate version of DeepMagic-Coder-7b which can be found bellow.
- URL
!image/jpeg
This version uses a diffrent config setup, with the actual base model of the two merges as the "base_model". Test both for yourself and see which is better at coding. Benchmarks coming soon.
Config can be found bellow:
| [] | [
"TAGS\n#transformers #safetensors #llama #text-generation #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
52
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
-0.003849297296255827,
0.04157582297921181,
-0.005358298774808645,
-0.002006892580538988,
0.12121443450450897,
-0.0023140516132116318,
0.21995262801647186,
0.0976245328783989,
-0.002134085865691304,
-0.028051059693098068,
0.14098598062992096,
0.20353935658931732,
-0.027530256658792496,
0.06547725945711136,
-0.10989534854888916,
-0.19001781940460205,
0.07975570857524872,
-0.005405021365731955,
0.055104125291109085,
0.07527986913919449,
0.09911483526229858,
-0.05641859024763107,
0.08538983762264252,
-0.043939460068941116,
-0.09532630443572998,
0.04172266274690628,
0.06300311535596848,
-0.14478042721748352,
0.10952402651309967,
0.07509230077266693,
0.10293574631214142,
0.06388837099075317,
-0.01375104021281004,
-0.2282211035490036,
0.019493047147989273,
-0.011446521617472172,
-0.07956469058990479,
0.01857081986963749,
0.0724555253982544,
-0.06962201744318008,
0.06803163141012192,
0.07595796138048172,
-0.0069822589866817,
0.07794570922851562,
-0.12478917837142944,
-0.03773628547787666,
-0.0527474582195282,
-0.01737235300242901,
0.10033683478832245,
0.08288641273975372,
0.008553036488592625,
0.12861965596675873,
-0.0460137315094471,
0.08864674717187881,
0.10105884075164795,
-0.36422982811927795,
0.03252127021551132,
0.13258597254753113,
0.08639060705900192,
0.02636159583926201,
-0.04908933863043785,
0.11520615965127945,
0.0778331384062767,
-0.025230219587683678,
0.04683642461895943,
-0.06403718888759613,
-0.06023075804114342,
0.0474824495613575,
-0.04937823489308357,
-0.035021159797906876,
0.2255091667175293,
-0.048003487288951874,
0.02201833389699459,
-0.08495626598596573,
-0.063540980219841,
0.004415482748299837,
-0.015621200203895569,
0.04226779192686081,
-0.004650085233151913,
0.10234159231185913,
0.047058794647455215,
-0.02541493996977806,
-0.1463351547718048,
-0.010123618878424168,
-0.19214987754821777,
0.15616413950920105,
0.009995682165026665,
0.03602566570043564,
-0.14876310527324677,
0.052005585283041,
0.020091798156499863,
-0.1022171825170517,
-0.009024777449667454,
-0.07613623142242432,
0.07293427735567093,
-0.021311713382601738,
-0.05060402303934097,
-0.08264583349227905,
0.14676575362682343,
0.16682443022727966,
0.0238858200609684,
0.0036225037183612585,
-0.11841477453708649,
0.09255154430866241,
0.005842770449817181,
0.033214397728443146,
0.048933349549770355,
-0.020335614681243896,
0.0804562196135521,
-0.10641513764858246,
0.0777621641755104,
-0.045458484441041946,
-0.16236527264118195,
0.01291638519614935,
0.00952336098998785,
0.1481045037508011,
-0.004036768805235624,
0.089394710958004,
-0.04640040919184685,
0.05940914899110794,
0.07988601177930832,
-0.08486296236515045,
-0.0037709602620452642,
-0.015596665441989899,
0.05994424223899841,
0.014862819574773312,
0.014005301520228386,
0.03976818919181824,
-0.046839725226163864,
0.07250332832336426,
-0.06159265711903572,
-0.04051261022686958,
-0.049518883228302,
-0.06224117800593376,
0.06905776262283325,
-0.061363302171230316,
0.04468187317252159,
-0.1777738332748413,
-0.22070357203483582,
0.02464439533650875,
0.013951590284705162,
-0.018896007910370827,
-0.0012315894709900022,
-0.03315489739179611,
-0.03501160815358162,
0.022209595888853073,
-0.08758897334337234,
-0.07052310556173325,
-0.09185780584812164,
0.07831136137247086,
-0.027050333097577095,
0.030020903795957565,
-0.1604417860507965,
0.038697756826877594,
-0.11656199395656586,
0.01296173594892025,
-0.028051257133483887,
0.023482143878936768,
-0.06817398965358734,
0.1713334172964096,
-0.009197886101901531,
0.0350252240896225,
-0.053295914083719254,
0.08215028047561646,
-0.04779068008065224,
0.16681306064128876,
-0.13494452834129333,
-0.05061241611838341,
0.20699985325336456,
-0.13453488051891327,
-0.2181762009859085,
0.07937384396791458,
-0.012063896283507347,
0.07402712851762772,
0.10656873881816864,
0.20221243798732758,
0.012460126541554928,
-0.08428488671779633,
0.04308132454752922,
0.09250641614198685,
-0.08357719331979752,
-0.13048051297664642,
-0.0002972103247884661,
-0.01989022269845009,
-0.1173701286315918,
0.04816463589668274,
0.07768440246582031,
0.05200492963194847,
-0.016429955139756203,
-0.057088013738393784,
-0.05853211134672165,
-0.038411833345890045,
-0.016033632680773735,
-0.03397468850016594,
0.06806734949350357,
-0.09934092313051224,
0.0019110505236312747,
0.03648265823721886,
-0.015667276456952095,
-0.003068840829655528,
0.014169437810778618,
-0.08264804631471634,
0.06614229828119278,
-0.005409609992057085,
0.0475180447101593,
-0.11178915202617645,
-0.12964870035648346,
-0.0007647328893654048,
0.07509008795022964,
0.016813769936561584,
0.03709910437464714,
0.029702192172408104,
0.010672100819647312,
-0.01521940529346466,
0.007333923596888781,
0.20648764073848724,
0.038250911980867386,
-0.04768528416752815,
-0.08994092047214508,
0.08715750277042389,
-0.06458824127912521,
-0.015310880728065968,
-0.11385233700275421,
0.023347778245806694,
0.044380247592926025,
0.09104777127504349,
0.00921452883630991,
0.07495976239442825,
-0.020303400233387947,
0.013424382545053959,
-0.11474888026714325,
0.006550895981490612,
0.08273644745349884,
0.0022236108779907227,
-0.10209061205387115,
0.22667057812213898,
-0.2210896760225296,
0.3080860674381256,
0.21094214916229248,
-0.1825488954782486,
0.026943033561110497,
-0.0945882648229599,
0.011032522656023502,
0.014552914537489414,
0.005726390518248081,
-0.025031547993421555,
-0.03026294708251953,
-0.008293206803500652,
0.1792396456003189,
-0.0730908140540123,
-0.025023970752954483,
-0.006241809111088514,
-0.06615357100963593,
-0.04816410690546036,
0.05591263249516487,
0.1242951825261116,
-0.17330072820186615,
0.18749713897705078,
0.26384732127189636,
0.046951763331890106,
0.1530172973871231,
-0.05455179885029793,
0.013769618235528469,
0.04229217767715454,
0.02884567342698574,
0.016804341226816177,
-0.04273542761802673,
-0.04881971701979637,
-0.003401821246370673,
0.06570019572973251,
0.026174848899245262,
0.05847599357366562,
-0.1476474553346634,
-0.07770358771085739,
-0.00043314375216141343,
-0.03968307748436928,
0.006200890988111496,
0.08856471627950668,
0.00619640713557601,
0.10802275687456131,
-0.07824309915304184,
-0.04909283667802811,
0.11786139756441116,
-0.008468481712043285,
-0.08553964644670486,
0.18271438777446747,
-0.12980367243289948,
-0.24880115687847137,
-0.20721422135829926,
-0.17201396822929382,
-0.054968323558568954,
0.06625471264123917,
0.13368694484233856,
-0.029557034373283386,
-0.075899139046669,
-0.07619262486696243,
0.002861606189981103,
-0.0057288226671516895,
0.004272588063031435,
-0.05925639718770981,
0.08340255171060562,
-0.054597072303295135,
-0.11993831396102905,
-0.05887284502387047,
0.031332191079854965,
-0.04659777134656906,
0.12162791192531586,
-0.07940787822008133,
0.10349738597869873,
0.11328767985105515,
0.012108376249670982,
0.002099009230732918,
-0.06695496290922165,
0.12976272404193878,
-0.05371017009019852,
-0.013419275172054768,
0.20312604308128357,
-0.051885515451431274,
0.05902451276779175,
0.17539414763450623,
0.023335132747888565,
-0.1270580142736435,
0.05697086080908775,
-0.03616220876574516,
-0.09935794770717621,
-0.243087038397789,
-0.10316934436559677,
-0.08488531410694122,
0.08033953607082367,
0.024067701771855354,
0.08842440694570541,
0.14749181270599365,
0.0748928040266037,
-0.02551400661468506,
0.005888370331376791,
0.06971859931945801,
0.09860600531101227,
0.2775948643684387,
-0.014959827996790409,
0.1402406245470047,
-0.11787562817335129,
-0.1081637367606163,
0.08385855704545975,
0.08064430952072144,
0.1274876594543457,
0.14238470792770386,
0.07077115029096603,
0.06407109647989273,
0.049006592482328415,
0.12872321903705597,
0.12011539936065674,
0.051090482622385025,
-0.033788446336984634,
-0.024373143911361694,
-0.04308266192674637,
-0.009716058149933815,
0.06506809592247009,
-0.10034585744142532,
-0.1401059776544571,
-0.02602313458919525,
-0.08214320242404938,
0.0613565668463707,
0.09771771728992462,
0.023694660514593124,
-0.21600355207920074,
0.06241931393742561,
0.12548938393592834,
-0.027630848810076714,
-0.07887393981218338,
0.11614454537630081,
-0.006654286291450262,
-0.046861473470926285,
0.12004013359546661,
-0.029947400093078613,
0.11082738637924194,
-0.010047024115920067,
0.07613486796617508,
-0.08575134724378586,
-0.09170379489660263,
0.00663441326469183,
0.11974678933620453,
-0.302707701921463,
0.21962690353393555,
0.012638079933822155,
0.004754889290779829,
-0.07237697392702103,
-0.0002597958955448121,
0.01731303706765175,
0.19032515585422516,
0.13372834026813507,
-0.04475002735853195,
-0.16154153645038605,
-0.07699186354875565,
-0.013818249106407166,
0.027343200519680977,
0.10924059897661209,
-0.005134155508130789,
0.023065051063895226,
-0.06813602894544601,
-0.001331697916612029,
0.008712774142622948,
-0.01654650829732418,
-0.05658411979675293,
-0.18648874759674072,
0.022294819355010986,
0.11972842365503311,
0.10918665677309036,
-0.04220011457800865,
0.012974285520613194,
-0.1632457822561264,
0.153584286570549,
-0.10504484176635742,
-0.057048499584198,
-0.11677833646535873,
-0.14972363412380219,
0.026337679475545883,
-0.022613978013396263,
0.053408823907375336,
-0.06106256693601608,
0.046389348804950714,
-0.0939030796289444,
-0.19687974452972412,
0.10409051179885864,
-0.10229508578777313,
-0.042705755680799484,
-0.048289064317941666,
0.13027821481227875,
-0.12902523577213287,
-0.03166760131716728,
0.061804402619600296,
0.03401830047369003,
-0.0572851225733757,
-0.10815569013357162,
-0.02057097665965557,
0.027827328070998192,
0.033202916383743286,
-0.017214270308613777,
-0.15045465528964996,
-0.08616852760314941,
0.013691711239516735,
-0.08751906454563141,
0.2202940136194229,
0.29374775290489197,
-0.039665885269641876,
0.1598970890045166,
0.17970344424247742,
-0.1352005898952484,
-0.34689784049987793,
-0.13214148581027985,
-0.20071037113666534,
-0.05137982591986656,
0.055562909692525864,
-0.10422831773757935,
0.09389669448137283,
0.0568624809384346,
-0.05788401886820793,
0.1129763126373291,
-0.1865699291229248,
-0.11794386804103851,
0.15749907493591309,
0.02788655273616314,
0.33786776661872864,
-0.19532208144664764,
-0.10973913222551346,
-0.1346137374639511,
-0.09878473728895187,
0.14908848702907562,
-0.07051460444927216,
0.09469954669475555,
0.013681183569133282,
0.029611049219965935,
0.01568800024688244,
-0.041363585740327835,
0.10135187208652496,
-0.05262637138366699,
0.09835208207368851,
-0.13683371245861053,
0.03563297167420387,
0.0805124044418335,
-0.02058546058833599,
0.07582128047943115,
-0.1828145682811737,
0.016589630395174026,
-0.030878504738211632,
-0.058801911771297455,
0.0028995408210903406,
0.09404236078262329,
0.0157973300665617,
-0.05681571736931801,
-0.024002641439437866,
-0.061485033482313156,
-0.0036316998302936554,
-0.029109850525856018,
0.26888102293014526,
-0.05835554376244545,
0.1487407386302948,
0.16098777949810028,
0.1460089534521103,
-0.12262018024921417,
0.12195087224245071,
-0.04960666224360466,
-0.10001746565103531,
0.0586608462035656,
-0.12557490170001984,
0.05669065937399864,
0.08642797917127609,
-0.062301959842443466,
0.07540272176265717,
0.08279389888048172,
0.02674143947660923,
-0.0030215971637517214,
0.17022624611854553,
-0.17954237759113312,
-0.07991205155849457,
-0.0313558466732502,
0.06571418046951294,
0.08947695046663284,
0.07575149089097977,
0.16392293572425842,
-0.010046266950666904,
0.01506747305393219,
0.00009138336463365704,
0.022046448662877083,
-0.023308994248509407,
0.03990858048200607,
-0.00040541309863328934,
0.007664828095585108,
-0.12263605743646622,
0.10800116509199142,
0.012595436535775661,
-0.09544645994901657,
0.023100027814507484,
0.09208671003580093,
-0.1304319202899933,
-0.12365005165338516,
-0.039133913815021515,
0.14645399153232574,
-0.1784994900226593,
-0.08615804463624954,
-0.057105544954538345,
-0.17860136926174164,
0.03919219970703125,
0.22984002530574799,
0.056675586849451065,
0.10148686170578003,
0.014567098580300808,
-0.04951230064034462,
-0.027642622590065002,
0.03900628536939621,
-0.060044918209314346,
0.0603107213973999,
-0.1290147751569748,
-0.022691836580634117,
-0.03921836242079735,
0.03983237221837044,
-0.08672855794429779,
-0.010996577329933643,
-0.154677152633667,
0.025562196969985962,
-0.1745959371328354,
-0.003359198337420821,
-0.08859938383102417,
-0.019591230899095535,
0.021335775032639503,
-0.022274639457464218,
-0.04903541877865791,
-0.045714445412158966,
-0.10395531356334686,
0.011114559136331081,
-0.03104853630065918,
0.08278977125883102,
-0.11301403492689133,
-0.046055469661951065,
0.0567261166870594,
-0.02520120143890381,
0.09041257202625275,
0.029151948168873787,
-0.08768359571695328,
0.10448309034109116,
-0.23303459584712982,
-0.04687062278389931,
0.13416028022766113,
0.017081361263990402,
0.015140390954911709,
0.08166962116956711,
-0.001032560714520514,
0.14004859328269958,
0.008282414637506008,
0.05614238604903221,
-0.04957761615514755,
-0.13021619617938995,
-0.010631983168423176,
-0.02670212648808956,
-0.12377263605594635,
-0.00032062403624877334,
-0.08736969530582428,
0.10374079644680023,
-0.03747629374265671,
0.2043464183807373,
-0.07429692894220352,
0.029200153425335884,
-0.030702412128448486,
0.027821263298392296,
-0.002358036581426859,
-0.19099481403827667,
-0.1315377652645111,
-0.09178810566663742,
-0.025928905233740807,
-0.001532270573079586,
0.32111430168151855,
0.04376078397035599,
-0.05720257759094238,
0.07128024846315384,
0.049720410257577896,
0.04592670127749443,
0.03617539256811142,
0.31660541892051697,
0.09487446397542953,
-0.013420739211142063,
-0.14970850944519043,
0.026235517114400864,
0.02326009050011635,
-0.11589925736188889,
0.04454168304800987,
0.10622580349445343,
-0.11225781589746475,
0.09958390146493912,
0.07320310920476913,
0.007487726863473654,
-0.03509373590350151,
-0.10228554159402847,
-0.02670189179480076,
0.06416782736778259,
-0.015020408667623997,
0.05821194127202034,
0.1985262930393219,
-0.021395107731223106,
0.006586323957890272,
-0.038559723645448685,
-0.021712077781558037,
-0.1926947385072708,
-0.15573188662528992,
-0.10866611450910568,
-0.09517345577478409,
0.039143793284893036,
-0.06912236660718918,
0.05956843122839928,
0.07048705965280533,
0.06756238639354706,
-0.04013029858469963,
0.08135770261287689,
-0.02699987404048443,
-0.04420517012476921,
0.014290579594671726,
-0.041168082505464554,
0.028459150344133377,
-0.03677079454064369,
-0.05804334208369255,
-0.06928186863660812,
-0.0491744764149189,
-0.062447406351566315,
0.06913433223962784,
0.011559789068996906,
0.05507809296250343,
-0.16446392238140106,
-0.06341475993394852,
-0.029220281168818474,
0.07237408310174942,
-0.02244047448039055,
0.11889401823282242,
0.014598897658288479,
-0.04319441691040993,
0.08626241236925125,
0.19991669058799744,
-0.04993418604135513,
-0.1447323113679886,
-0.013448497280478477,
0.2002435177564621,
0.035074882209300995,
0.13375608623027802,
-0.05605895072221756,
0.008848054334521294,
-0.018097292631864548,
0.3221629559993744,
0.2753550112247467,
-0.04600027948617935,
0.02637862227857113,
-0.05468853935599327,
0.035121895372867584,
0.06386703997850418,
0.13155029714107513,
0.057988543063402176,
0.22688718140125275,
-0.034239523112773895,
-0.004356452729552984,
-0.017280329018831253,
0.01592135615646839,
-0.15614387392997742,
0.11534588783979416,
-0.015370954759418964,
-0.06237677484750748,
-0.024119241163134575,
0.1149347648024559,
-0.15318956971168518,
0.1179792732000351,
-0.052648093551397324,
-0.08660911023616791,
-0.00515630142763257,
0.003058725269511342,
0.1731954663991928,
-0.03028198704123497,
0.01488418411463499,
-0.03697666898369789,
-0.06821247935295105,
0.005594014190137386,
-0.008830344304442406,
-0.20003125071525574,
0.02245105430483818,
0.0068825832568109035,
0.012758585624396801,
0.07741796970367432,
0.006922852713614702,
0.03329480439424515,
0.06797651201486588,
0.02341562509536743,
-0.059864841401576996,
0.15672925114631653,
0.017178725451231003,
-0.06865647435188293,
0.0689980611205101,
-0.06539874523878098,
-0.006503082811832428,
0.004121152218431234,
0.04110436886548996,
-0.07741575688123703,
0.06828004866838455,
-0.020052211359143257,
-0.11550668627023697,
-0.02141040377318859,
0.021401267498731613,
-0.07135076820850372,
0.07339077442884445,
0.014393203891813755,
-0.007730762008577585,
0.0074863978661596775,
-0.03275369480252266,
0.04045829549431801,
0.005226201377809048,
-0.16471591591835022,
-0.029747474938631058,
-0.10361180454492569,
-0.07068151980638504,
0.166717991232872,
0.02268947847187519,
-0.2698356509208679,
0.004231145139783621,
-0.11386311054229736,
0.04944133013486862,
-0.1964423656463623,
0.06345852464437485,
0.2272513210773468,
0.00894148088991642,
-0.02555731311440468,
-0.19690142571926117,
0.057916779071092606,
0.04835796728730202,
-0.048489492386579514,
-0.10247565805912018
] |
null | null | diffusers |
# LoRA DreamBooth - danaleee/CL_rank10_iter800_valprompt
These are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks duck using [DreamBooth](https://dreambooth.github.io/). You can find some example images in the following.




LoRA for the text encoder was enabled: False.
| {"license": "creativeml-openrail-m", "tags": ["stable-diffusion", "stable-diffusion-diffusers", "text-to-image", "diffusers", "lora"], "base_model": "CompVis/stable-diffusion-v1-4", "instance_prompt": "a photo of sks duck", "inference": true} | text-to-image | danaleee/CL_rank10_iter800_valprompt | [
"diffusers",
"tensorboard",
"stable-diffusion",
"stable-diffusion-diffusers",
"text-to-image",
"lora",
"base_model:CompVis/stable-diffusion-v1-4",
"license:creativeml-openrail-m",
"region:us"
] | 2024-02-07T15:35:01+00:00 | [] | [] | TAGS
#diffusers #tensorboard #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us
|
# LoRA DreamBooth - danaleee/CL_rank10_iter800_valprompt
These are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks duck using DreamBooth. You can find some example images in the following.
!img_0
!img_1
!img_2
!img_3
LoRA for the text encoder was enabled: False.
| [
"# LoRA DreamBooth - danaleee/CL_rank10_iter800_valprompt\n\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks duck using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
"TAGS\n#diffusers #tensorboard #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us \n",
"# LoRA DreamBooth - danaleee/CL_rank10_iter800_valprompt\n\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks duck using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
70,
107
] | [
"passage: TAGS\n#diffusers #tensorboard #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us \n# LoRA DreamBooth - danaleee/CL_rank10_iter800_valprompt\n\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks duck using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
-0.07934951037168503,
-0.003535887924954295,
-0.0023701880127191544,
0.07230953872203827,
0.0848681852221489,
0.014420729130506516,
0.1609642207622528,
0.07304100692272186,
0.03519773110747337,
0.05502532795071602,
0.10429251939058304,
0.10181208699941635,
0.020037852227687836,
0.06093335524201393,
-0.008816270157694817,
-0.15248744189739227,
-0.004746070131659508,
-0.018193518742918968,
-0.10643191635608673,
0.053695641458034515,
0.03153160959482193,
-0.02331419847905636,
0.11743499338626862,
-0.0028342732694000006,
-0.11974266916513443,
0.012735876254737377,
-0.018723415210843086,
-0.03668356314301491,
0.08759114146232605,
0.0378478579223156,
0.03287378326058388,
0.06567445397377014,
0.046267908066511154,
-0.1554860770702362,
0.038007982075214386,
0.024351082742214203,
-0.042747557163238525,
0.07084835320711136,
-0.07366135716438293,
0.0020744933281093836,
0.05561099573969841,
-0.03279285132884979,
-0.0030106790363788605,
0.016432499513030052,
-0.06287678331136703,
-0.07011695951223373,
-0.02934282273054123,
-0.02953116036951542,
0.06415305286645889,
0.04563593864440918,
0.019238371402025223,
0.0688786506652832,
-0.0056807235814630985,
0.09288907051086426,
0.27412864565849304,
-0.2293311059474945,
-0.03820507228374481,
0.20510442554950714,
0.0055316961370408535,
0.09518444538116455,
-0.04139236733317375,
0.08535923808813095,
0.09708081185817719,
-0.03132037818431854,
0.006521920673549175,
-0.04647887125611305,
0.03587787225842476,
-0.06865262985229492,
-0.099862240254879,
0.0726105198264122,
0.14610733091831207,
0.003566527273505926,
-0.05606098100543022,
-0.13305284082889557,
-0.030724095180630684,
-0.013787003234028816,
-0.02649681828916073,
0.023914610967040062,
0.01761072315275669,
0.008250046521425247,
-0.02713075838983059,
-0.04177919030189514,
-0.08439190685749054,
-0.06727482378482819,
-0.02702510356903076,
0.055500373244285583,
-0.004356942605227232,
0.05494433268904686,
-0.04111088812351227,
0.11931262165307999,
-0.07619375735521317,
-0.12021378427743912,
0.02567421644926071,
-0.0516178160905838,
0.01531843189150095,
0.08420761674642563,
-0.004766588564962149,
-0.1620713174343109,
0.06968311965465546,
0.013912544585764408,
0.15979556739330292,
0.02816639468073845,
-0.008154675364494324,
0.1177946999669075,
-0.022229475900530815,
0.019628245383501053,
-0.0543246865272522,
-0.056331854313611984,
0.017908673733472824,
0.02122451551258564,
0.09418194741010666,
-0.08786647021770477,
-0.12939903140068054,
-0.004274553153663874,
-0.03003661334514618,
0.02231275476515293,
-0.050489362329244614,
0.008365154266357422,
-0.10354695469141006,
0.0023862214293330908,
0.08242392539978027,
0.007593720220029354,
0.024060683324933052,
-0.03645285964012146,
0.020748460665345192,
0.10559588670730591,
0.15953774750232697,
0.033628761768341064,
0.02196929045021534,
0.05359385535120964,
-0.05772523581981659,
0.026715388521552086,
0.013843145221471786,
-0.11279233545064926,
-0.0039972891099750996,
-0.16671055555343628,
-0.018424436450004578,
-0.13245894014835358,
-0.03173496574163437,
-0.00541836628690362,
0.030826086178421974,
-0.034830112010240555,
0.030059929937124252,
-0.06390424817800522,
-0.11842480301856995,
-0.011254891753196716,
0.06263041496276855,
0.019727399572730064,
-0.03205904737114906,
0.0458492636680603,
-0.043927550315856934,
0.15046773850917816,
-0.12866990268230438,
-0.0270601287484169,
-0.05498310923576355,
0.007054036017507315,
-0.11194551736116409,
0.09693728387355804,
-0.05836131051182747,
0.0693947970867157,
-0.056976258754730225,
-0.055506959557533264,
-0.06396610289812088,
0.036827556788921356,
0.055146779865026474,
0.13360704481601715,
-0.27965283393859863,
-0.04890454187989235,
0.08008050918579102,
-0.1644320785999298,
-0.07375026494264603,
0.05994488671422005,
-0.0072801318019628525,
0.11628709733486176,
0.06946633011102676,
0.11791379004716873,
0.10466320812702179,
-0.26677781343460083,
0.011175552383065224,
-0.03224416822195053,
-0.04526464268565178,
-0.06354670971632004,
-0.021065503358840942,
0.05134933814406395,
-0.03817499056458473,
0.0364467017352581,
-0.07580337673425674,
0.06776763498783112,
-0.02285643294453621,
-0.022187447175383568,
-0.022765081375837326,
-0.03481446951627731,
-0.017516029998660088,
-0.008103475905954838,
0.0452125146985054,
-0.02965022251009941,
-0.028026210144162178,
-0.031846530735492706,
0.026760319247841835,
-0.05408531427383423,
0.03833981603384018,
-0.011576656252145767,
0.10833925008773804,
-0.0617976151406765,
-0.014722700230777264,
-0.1086517870426178,
0.011319834738969803,
0.022079620510339737,
0.10162865370512009,
0.08323479443788528,
0.06018262356519699,
0.0952618420124054,
0.0613027922809124,
0.026656946167349815,
0.015833573415875435,
0.07328468561172485,
-0.003934844397008419,
-0.07810041308403015,
-0.17978069186210632,
0.04790716618299484,
-0.07260022312402725,
0.03467966988682747,
-0.17887069284915924,
0.03209368512034416,
0.0975898802280426,
0.1702105849981308,
0.09708128124475479,
-0.0324329249560833,
0.04956107959151268,
0.022949157282710075,
-0.058464545756578445,
-0.0459001325070858,
0.039186764508485794,
0.012925694696605206,
-0.1451149582862854,
0.15917949378490448,
-0.12389747053384781,
0.032335568219423294,
0.11992821097373962,
-0.0048417882062494755,
-0.07085259258747101,
-0.030631186440587044,
0.0019055871525779366,
0.02220819890499115,
-0.04956963658332825,
-0.022213611751794815,
0.13420473039150238,
0.016633331775665283,
0.14617758989334106,
-0.023718426004052162,
-0.003351920982822776,
0.06810309737920761,
-0.016394490376114845,
-0.05267629027366638,
0.11411720514297485,
0.07507692277431488,
0.045850783586502075,
0.006298159249126911,
0.09850400686264038,
0.016913384199142456,
0.1357104480266571,
0.004170678090304136,
-0.030279317870736122,
-0.04807606711983681,
0.03168873488903046,
0.0460825189948082,
0.1522519737482071,
-0.0033711870200932026,
-0.02535906992852688,
-0.007992413826286793,
-0.05243242532014847,
0.002544826129451394,
-0.18760253489017487,
-0.03558719903230667,
0.033838752657175064,
-0.049285829067230225,
0.12444201111793518,
0.09180305898189545,
-0.09395718574523926,
0.08807994425296783,
-0.11511821299791336,
-0.1041707843542099,
-0.02153666876256466,
-0.03990795090794563,
-0.044786736369132996,
0.11173690855503082,
-0.0714995339512825,
-0.16819655895233154,
-0.17421428859233856,
0.04528474062681198,
-0.004611548967659473,
-0.038419079035520554,
0.0635513961315155,
-0.10390113294124603,
-0.051824431866407394,
-0.1118420660495758,
0.005673820618540049,
0.029923858121037483,
0.05726901814341545,
0.060025256127119064,
-0.03191668167710304,
-0.01221941877156496,
-0.12017323076725006,
-0.004474703688174486,
-0.09013819694519043,
0.0877893790602684,
0.06738360226154327,
0.005066817160695791,
0.11279302090406418,
0.11458354443311691,
0.028086327016353607,
0.03174334391951561,
0.018310125917196274,
0.21252426505088806,
-0.021794518455863,
0.0920904353260994,
0.09967470914125443,
0.013604411855340004,
0.08436048775911331,
0.11640764772891998,
0.039804715663194656,
-0.07613462954759598,
0.04219930246472359,
0.0026140790432691574,
-0.15623317658901215,
-0.06883635371923447,
-0.03588726744055748,
-0.039902228862047195,
-0.029105370864272118,
0.05949030816555023,
0.04949763044714928,
0.0766381174325943,
0.08662846684455872,
0.09612823277711868,
0.0797288566827774,
0.07312987744808197,
0.0660722404718399,
0.08777271211147308,
-0.04246828332543373,
0.05580948293209076,
-0.08623964339494705,
-0.11519881337881088,
0.11310374736785889,
-0.03991539031267166,
0.19077935814857483,
-0.06349537521600723,
0.04394100606441498,
0.04657349735498428,
-0.023940784856677055,
0.08141449838876724,
0.05554954707622528,
-0.05514156073331833,
-0.0056969947181642056,
-0.04530227556824684,
-0.11596725136041641,
0.09177521616220474,
0.08075284957885742,
-0.02321513369679451,
-0.03605811297893524,
0.008598732762038708,
0.09188073873519897,
-0.000577326980419457,
0.0621669627726078,
0.15757058560848236,
-0.20829975605010986,
0.047451529651880264,
0.0032957124058157206,
0.06320808827877045,
-0.000513922015670687,
0.03488023579120636,
0.22351984679698944,
-0.003177134320139885,
0.07875481247901917,
-0.06778132170438766,
0.06322626769542694,
-0.03345440700650215,
-0.02111879177391529,
-0.013475554995238781,
0.12153096497058868,
-0.02461138926446438,
-0.05714753642678261,
-0.16630138456821442,
0.06461489945650101,
-0.014826350845396519,
-0.014922892674803734,
-0.061314668506383896,
-0.022214025259017944,
0.03374014422297478,
-0.012404565699398518,
0.10797145217657089,
0.021871933713555336,
-0.034008417278528214,
-0.09833521395921707,
-0.17884507775306702,
-0.02499420754611492,
0.059950679540634155,
-0.043629515916109085,
0.02779972553253174,
0.058758631348609924,
-0.02523750811815262,
0.0065716649405658245,
-0.019597411155700684,
-0.10797559469938278,
-0.1002429947257042,
0.00761494180187583,
0.15909956395626068,
-0.0748579278588295,
-0.028966033831238747,
-0.10941805690526962,
-0.03606279939413071,
0.08366884291172028,
-0.10722803324460983,
-0.07377836108207703,
-0.04696901887655258,
-0.040572673082351685,
0.06162848323583603,
-0.021853826940059662,
-0.005971888080239296,
-0.044952619820833206,
0.056574977934360504,
-0.08630169928073883,
-0.13242220878601074,
0.08168548345565796,
-0.019158896058797836,
-0.12442649900913239,
-0.08127779513597488,
0.09541183710098267,
0.00044844436342827976,
-0.012921304441988468,
-0.008474117144942284,
0.0376221239566803,
0.011823313310742378,
-0.11520080268383026,
0.06248968839645386,
0.10639139264822006,
-0.13393454253673553,
0.1080697551369667,
-0.01710212230682373,
0.001023743418045342,
-0.0780511423945427,
0.020956171676516533,
0.14376036822795868,
0.255912721157074,
-0.1112532764673233,
0.10810170322656631,
0.033399540930986404,
-0.09431549161672592,
-0.20936229825019836,
-0.05194348469376564,
-0.014507623389363289,
0.034449875354766846,
0.025096192955970764,
-0.08178222179412842,
0.14332309365272522,
0.06512928009033203,
0.0061301118694245815,
0.2651495337486267,
-0.37732404470443726,
-0.14429236948490143,
-0.008728858083486557,
0.1521431803703308,
0.2833644449710846,
-0.15977610647678375,
-0.06473408639431,
-0.026883309707045555,
-0.08976875245571136,
0.1543763428926468,
-0.03310806676745415,
0.09364496171474457,
-0.034629084169864655,
-0.0209352970123291,
0.013888584449887276,
-0.02830924466252327,
0.11934860050678253,
0.04225057736039162,
0.057104673236608505,
-0.0841703861951828,
-0.0610678493976593,
0.07377328723669052,
-0.042835015803575516,
0.025411905720829964,
-0.15382157266139984,
0.016107942909002304,
-0.07601891458034515,
0.012714138254523277,
0.010337701998651028,
-0.010126892477273941,
-0.016822574660182,
-0.03151043877005577,
-0.1260198950767517,
0.019668634980916977,
0.007298169657588005,
-0.007676307111978531,
0.07787474989891052,
-0.002965275663882494,
0.03923385962843895,
0.13254253566265106,
-0.03234148770570755,
0.03027176670730114,
-0.10129499435424805,
-0.017255250364542007,
-0.03498328849673271,
0.09909963607788086,
-0.11124099791049957,
-0.003619537455961108,
0.16053634881973267,
0.09305808693170547,
0.08035267889499664,
0.035299353301525116,
-0.08698973804712296,
0.09235204011201859,
0.11020690202713013,
-0.08655696362257004,
-0.024347370490431786,
-0.011052708141505718,
-0.05142001807689667,
0.11349919438362122,
0.01920885220170021,
0.18747028708457947,
-0.06192709505558014,
0.04820215702056885,
0.00873687956482172,
0.03189985081553459,
-0.01899491436779499,
0.0790877416729927,
0.03415914252400398,
0.010285167023539543,
-0.07436265051364899,
0.04685882478952408,
-0.041899777948856354,
0.05581381916999817,
0.02495674602687359,
0.05354040488600731,
-0.07074963301420212,
-0.006558479741215706,
0.00023768718529026955,
0.19630038738250732,
-0.08381898701190948,
0.014816840179264545,
-0.10832278430461884,
-0.12049578130245209,
0.020214173942804337,
0.12408287823200226,
0.06072678044438362,
0.016930032521486282,
-0.04118271544575691,
-0.05640061944723129,
-0.04690214619040489,
0.03896626830101013,
0.057088352739810944,
0.06728368997573853,
-0.22755466401576996,
-0.07087317109107971,
0.023940283805131912,
-0.020446762442588806,
-0.061402272433042526,
-0.04713587090373039,
-0.09851796180009842,
-0.029260272160172462,
-0.044653039425611496,
0.1214706227183342,
-0.06477449834346771,
-0.0008935217629186809,
-0.007903271354734898,
-0.03854377940297127,
-0.035404279828071594,
0.01978945918381214,
-0.033094681799411774,
-0.029194343835115433,
-0.00270967953838408,
-0.025294063612818718,
-0.045332081615924835,
-0.095279261469841,
-0.014277513138949871,
-0.08910419046878815,
0.01849225163459778,
-0.046157363802194595,
-0.07307963073253632,
0.007183407433331013,
-0.19084970653057098,
0.0288395993411541,
0.16962531208992004,
-0.0025018034502863884,
-0.01631348766386509,
-0.013826828449964523,
-0.009569554589688778,
-0.021440576761960983,
0.07314451783895493,
0.007440177723765373,
0.06561008095741272,
-0.07767489552497864,
-0.02716864086687565,
-0.08058374375104904,
-0.006231367122381926,
-0.049122877418994904,
0.036915238946676254,
0.1512407809495926,
0.16245251893997192,
0.16026653349399567,
-0.10313235223293304,
0.08841695636510849,
-0.10757803171873093,
0.0025890495162457228,
0.020593339577317238,
-0.06917861104011536,
0.06331954151391983,
-0.03831591457128525,
-0.01561181340366602,
-0.03613661602139473,
0.13871324062347412,
-0.016655605286359787,
-0.1401759833097458,
-0.02174750529229641,
0.0524725578725338,
-0.09008678048849106,
0.010500794276595116,
0.14869914948940277,
0.03879579156637192,
0.04777621105313301,
-0.09199003130197525,
0.050495896488428116,
0.1493990123271942,
0.09923405200242996,
0.09846789389848709,
0.012990407645702362,
0.01884191669523716,
0.105982206761837,
0.09693972021341324,
0.03827814385294914,
0.02018088847398758,
0.1631513386964798,
-0.055024757981300354,
0.12304412573575974,
-0.03296086937189102,
0.0009619725169613957,
0.09182476997375488,
-0.04027154669165611,
-0.025186626240611076,
0.08990693837404251,
-0.05163160711526871,
-0.08636712282896042,
-0.11801038682460785,
-0.05882475897669792,
-0.1391059309244156,
0.009074747562408447,
-0.04120205342769623,
-0.01969265379011631,
0.02148251235485077,
0.06372981518507004,
0.07718673348426819,
0.023793358355760574,
-0.03150186687707901,
-0.046276699751615524,
0.1440598964691162,
-0.029298903420567513,
-0.09393336623907089,
0.024225369095802307,
0.011036185547709465,
0.075925812125206,
0.034454356878995895,
-0.044146835803985596,
0.07775569707155228,
0.06476251780986786,
-0.0036450729239732027,
-0.026483722031116486,
-0.08178271353244781,
-0.027776135131716728,
0.003273703856393695,
-0.0010013504652306437,
0.1317920833826065,
0.12445857375860214,
-0.06105012446641922,
-0.0289506446570158,
0.0762379989027977,
-0.03353940322995186,
-0.028782567009329796,
-0.13104940950870514,
0.015934865921735764,
-0.1541983187198639,
0.03685036674141884,
-0.06580980122089386,
-0.09340901672840118,
-0.029808443039655685,
0.1717093288898468,
0.19582143425941467,
-0.08705282211303711,
0.033753275871276855,
-0.041764289140701294,
-0.004363609943538904,
0.010504531674087048,
0.05512956157326698,
0.04724040627479553,
0.22490088641643524,
-0.03681692108511925,
-0.02135797031223774,
-0.10298021882772446,
-0.06751831620931625,
-0.07200244069099426,
-0.11447921395301819,
0.02080957219004631,
-0.03542766720056534,
-0.09144233912229538,
0.08347045630216599,
-0.13264992833137512,
-0.13570137321949005,
0.16414499282836914,
-0.17461225390434265,
-0.05589672178030014,
-0.04532632231712341,
0.05411875620484352,
0.026699550449848175,
0.011020099744200706,
-0.06786558032035828,
-0.01164780929684639,
0.026150664314627647,
-0.018845506012439728,
-0.14924265444278717,
-0.023132042959332466,
-0.09346165508031845,
-0.16869693994522095,
0.05931047722697258,
-0.018007943406701088,
0.06224459037184715,
0.043549731373786926,
0.021999211981892586,
-0.053781818598508835,
0.04273322969675064,
-0.04922281578183174,
-0.02839423343539238,
-0.10074673593044281,
0.07555408775806427,
-0.024958692491054535,
0.06917744129896164,
0.0546269565820694,
-0.05780499801039696,
0.007342635653913021,
0.054381344467401505,
-0.08853152394294739,
-0.08889257162809372,
-0.03565602749586105,
-0.07453850656747818,
0.09504814445972443,
0.07758201658725739,
-0.017545899376273155,
0.0151499779894948,
0.01656445302069187,
-0.030970273539423943,
-0.01927553117275238,
-0.11492729187011719,
0.03447720780968666,
-0.06821170449256897,
-0.028488362208008766,
-0.0014124615117907524,
0.04116791859269142,
-0.2075652927160263,
-0.09348497539758682,
-0.15489093959331512,
-0.02207650989294052,
0.0005192434182390571,
0.07520324736833572,
0.2086511254310608,
0.012959737330675125,
0.0015245669055730104,
-0.11902838945388794,
0.04405065253376961,
0.09616470336914062,
-0.04378679022192955,
-0.06500529497861862
] |
null | null | diffusers |
# LoRA DreamBooth - danaleee/CL_rank4_iter500_valprompt
These are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using [DreamBooth](https://dreambooth.github.io/). You can find some example images in the following.




LoRA for the text encoder was enabled: False.
| {"license": "creativeml-openrail-m", "tags": ["stable-diffusion", "stable-diffusion-diffusers", "text-to-image", "diffusers", "lora"], "base_model": "CompVis/stable-diffusion-v1-4", "instance_prompt": "a photo of sks teddybear", "inference": true} | text-to-image | danaleee/CL_rank4_iter500_valprompt | [
"diffusers",
"tensorboard",
"stable-diffusion",
"stable-diffusion-diffusers",
"text-to-image",
"lora",
"base_model:CompVis/stable-diffusion-v1-4",
"license:creativeml-openrail-m",
"region:us"
] | 2024-02-07T15:38:10+00:00 | [] | [] | TAGS
#diffusers #tensorboard #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us
|
# LoRA DreamBooth - danaleee/CL_rank4_iter500_valprompt
These are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using DreamBooth. You can find some example images in the following.
!img_0
!img_1
!img_2
!img_3
LoRA for the text encoder was enabled: False.
| [
"# LoRA DreamBooth - danaleee/CL_rank4_iter500_valprompt\n\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
"TAGS\n#diffusers #tensorboard #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us \n",
"# LoRA DreamBooth - danaleee/CL_rank4_iter500_valprompt\n\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
70,
109
] | [
"passage: TAGS\n#diffusers #tensorboard #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us \n# LoRA DreamBooth - danaleee/CL_rank4_iter500_valprompt\n\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
-0.06418299674987793,
-0.018912747502326965,
-0.0027262975927442312,
0.08229614794254303,
0.10149101912975311,
0.028092797845602036,
0.2142983376979828,
0.08223975449800491,
0.03945368528366089,
0.07039938867092133,
0.11435072124004364,
0.07524985074996948,
-0.004199519753456116,
0.09547212719917297,
-0.01573575660586357,
-0.19504620134830475,
-0.013157961890101433,
-0.01925741136074066,
-0.06336776912212372,
0.04583438113331795,
0.04577314481139183,
-0.04636969789862633,
0.12092532217502594,
-0.019518636167049408,
-0.1653718203306198,
0.030590012669563293,
-0.003577628405764699,
-0.057035181671381,
0.07986747473478317,
0.06811925023794174,
0.033372070640325546,
0.06165529042482376,
0.04540754109621048,
-0.13661551475524902,
0.03741907328367233,
0.0214117132127285,
-0.057370875030756,
0.08327311277389526,
-0.04634563624858856,
-0.025751426815986633,
0.10904999822378159,
-0.02058963105082512,
0.0031554640736430883,
0.017855674028396606,
-0.0821266919374466,
-0.050687629729509354,
-0.021519280970096588,
0.007891660556197166,
0.035894691944122314,
0.036791346967220306,
0.010679805651307106,
0.08802878856658936,
-0.024516530334949493,
0.08295455574989319,
0.2638590633869171,
-0.2178805023431778,
-0.03983232378959656,
0.21264106035232544,
0.01126022171229124,
0.08533849567174911,
-0.046776313334703445,
0.09862919896841049,
0.10026884078979492,
-0.03205820173025131,
0.050036247819662094,
-0.04575381800532341,
-0.005896759685128927,
-0.049453672021627426,
-0.10963941365480423,
0.0643748939037323,
0.14287912845611572,
0.00806680228561163,
-0.04854487255215645,
-0.1521986424922943,
-0.03915920853614807,
0.0024792945478111506,
-0.025849007070064545,
0.045873239636421204,
0.007054173853248358,
-0.008698996156454086,
-0.03865257650613785,
-0.039899032562971115,
-0.09310562163591385,
-0.05993153154850006,
-0.017562994733452797,
0.07222195714712143,
-0.00674669723957777,
0.059819769114255905,
-0.015178336761891842,
0.11205703765153885,
-0.09805933386087418,
-0.12787772715091705,
0.0486496202647686,
-0.06781159341335297,
0.013497332111001015,
0.06519239395856857,
-0.01232688408344984,
-0.15576398372650146,
0.04882984608411789,
0.009736795909702778,
0.14874927699565887,
0.018451154232025146,
-0.007753320038318634,
0.12787489593029022,
-0.004319640342146158,
0.04534268006682396,
-0.05662492290139198,
-0.004481386858969927,
0.01541709341108799,
0.028249409049749374,
0.0898570716381073,
-0.07914416491985321,
-0.1386418342590332,
-0.013875186443328857,
-0.04593800753355026,
0.02215789072215557,
-0.08345227688550949,
0.028793497011065483,
-0.09615430235862732,
-0.012646431103348732,
0.0796346664428711,
-0.00002647977089509368,
0.04345474764704704,
-0.02293120138347149,
-0.023804081603884697,
0.12179939448833466,
0.1615922451019287,
0.018632616847753525,
0.017964746803045273,
0.09068770706653595,
-0.06607595831155777,
0.04962111636996269,
-0.02466740645468235,
-0.11617621779441833,
0.017638830468058586,
-0.16308484971523285,
-0.009228448383510113,
-0.13539597392082214,
-0.028218654915690422,
-0.0037412235978990793,
0.04089301452040672,
-0.04712569713592529,
0.05919678509235382,
-0.058769527822732925,
-0.0895925760269165,
-0.007940621115267277,
0.06204827502369881,
0.006275657098740339,
-0.024193037301301956,
0.0538598969578743,
-0.026133596897125244,
0.1434963047504425,
-0.10315122455358505,
-0.03129434585571289,
-0.06551443040370941,
0.007670012768357992,
-0.14267580211162567,
0.06733082234859467,
-0.0564904548227787,
0.05929706618189812,
-0.06762213259935379,
-0.04573891684412956,
-0.053048815578222275,
0.05376031622290611,
0.03532619774341583,
0.13737702369689941,
-0.2583237588405609,
-0.06637634336948395,
0.0994638130068779,
-0.15313133597373962,
-0.08481644093990326,
0.05437285080552101,
-0.020304638892412186,
0.10454267263412476,
0.06025119870901108,
0.10788421332836151,
0.07356289029121399,
-0.2981001138687134,
-0.0018920204602181911,
-0.04695946350693703,
-0.03234155476093292,
-0.05116299167275429,
0.0004546486015897244,
0.06443236023187637,
0.0058749462477862835,
0.03389078751206398,
-0.04734363034367561,
0.07686816155910492,
-0.0420999750494957,
-0.011391944251954556,
-0.021614937111735344,
-0.03435469791293144,
-0.007907038554549217,
0.0030873222276568413,
0.06446218490600586,
-0.01828949525952339,
-0.04592176899313927,
0.0028552187141031027,
0.03504567965865135,
-0.06352318078279495,
0.02428700029850006,
-0.026736317202448845,
0.09744129329919815,
-0.08019720762968063,
-0.015170658938586712,
-0.10088687390089035,
0.018860438838601112,
0.0404965914785862,
0.12282110750675201,
0.0919545590877533,
0.05620947480201721,
0.10066108405590057,
0.06333731859922409,
0.0004777531430590898,
0.0149540388956666,
0.06863486766815186,
-0.024696214124560356,
-0.05555563047528267,
-0.2027713656425476,
0.026103857904672623,
-0.08682089298963547,
0.045405469834804535,
-0.20798920094966888,
0.022663423791527748,
0.07233396917581558,
0.1803480088710785,
0.11090733855962753,
-0.029039442539215088,
0.06563389301300049,
0.022416643798351288,
-0.0625271201133728,
-0.04814508929848671,
0.039592672139406204,
-0.008508970029652119,
-0.13504213094711304,
0.16368542611598969,
-0.1560274362564087,
0.06851736456155777,
0.12929487228393555,
-0.03649670258164406,
-0.07392759621143341,
-0.07473240047693253,
-0.005547874141484499,
0.03936390206217766,
-0.06389831751585007,
-0.018877560272812843,
0.08980810642242432,
-0.0029679371509701014,
0.15673953294754028,
-0.02241353504359722,
0.001536842086352408,
0.07032348960638046,
-0.03103133849799633,
-0.07968855649232864,
0.10076771676540375,
0.114347904920578,
0.010833165608346462,
0.014701553620398045,
0.06911300867795944,
-0.016215166077017784,
0.1453261822462082,
-0.0011236510472372174,
-0.04821697995066643,
-0.04495001211762428,
0.017267540097236633,
0.056186508387327194,
0.14903457462787628,
-0.013463693670928478,
-0.03197360783815384,
-0.007363441400229931,
-0.07849038392305374,
0.007583041209727526,
-0.15707476437091827,
-0.02620118297636509,
0.030069734901189804,
-0.03377244994044304,
0.12391023337841034,
0.10341400653123856,
-0.08899880200624466,
0.09639044106006622,
-0.1102340966463089,
-0.07694737613201141,
-0.030761756002902985,
-0.021310558542609215,
-0.0643637403845787,
0.11669499427080154,
-0.05177381634712219,
-0.15599282085895538,
-0.17590995132923126,
-0.0033547047059983015,
0.007945813238620758,
-0.010779913514852524,
0.047617170959711075,
-0.09045766294002533,
-0.06644513458013535,
-0.1156124547123909,
0.042128369212150574,
0.037665318697690964,
0.04925888404250145,
0.04851960763335228,
-0.021906904876232147,
-0.007163176313042641,
-0.11097445338964462,
0.006530500948429108,
-0.08332798629999161,
0.09842030704021454,
0.07704228907823563,
-0.0035689626820385456,
0.11332497000694275,
0.12774693965911865,
0.02736900933086872,
0.050412822514772415,
0.009075192734599113,
0.2239321768283844,
-0.033322010189294815,
0.08541291952133179,
0.11314002424478531,
-0.010009157471358776,
0.06088259443640709,
0.10830061882734299,
0.056307174265384674,
-0.07249888777732849,
0.048726070672273636,
-0.0022157595958560705,
-0.14814406633377075,
-0.07249896228313446,
-0.0464768223464489,
-0.0412897951900959,
-0.039127085357904434,
0.08533217012882233,
0.048816002905368805,
0.10204526782035828,
0.09487661719322205,
0.0628654882311821,
0.11528392136096954,
0.06026574596762657,
0.07945925742387772,
0.05738219618797302,
-0.04233522713184357,
0.04548025503754616,
-0.08358309417963028,
-0.11313305050134659,
0.1067851260304451,
-0.07181516289710999,
0.20256352424621582,
-0.06561342626810074,
0.04786397144198418,
0.05324472486972809,
-0.050830986350774765,
0.09725233167409897,
0.04236939549446106,
-0.06838183850049973,
-0.011160791851580143,
-0.04207644984126091,
-0.12868286669254303,
0.09581393003463745,
0.08684778213500977,
-0.02505795657634735,
-0.04326456040143967,
-0.009829756803810596,
0.035732023417949677,
0.01581287384033203,
0.01510569453239441,
0.14451587200164795,
-0.26124581694602966,
0.038218677043914795,
0.013036970049142838,
0.06473032385110855,
0.0006797550013288856,
0.0065226065926253796,
0.21941041946411133,
0.003469473449513316,
0.08424799889326096,
-0.0700354054570198,
0.058595605194568634,
-0.027147026732563972,
-0.02528681792318821,
-0.012525627389550209,
0.1245952770113945,
-0.029643259942531586,
-0.05136352777481079,
-0.20565329492092133,
0.0849960446357727,
0.008792269043624401,
-0.0092240609228611,
-0.07020686566829681,
-0.02674381621181965,
0.03235439211130142,
-0.007164948154240847,
0.08573758602142334,
0.006760905496776104,
0.00998890120536089,
-0.10887441039085388,
-0.14783889055252075,
-0.041071124374866486,
0.08852431178092957,
-0.03549104556441307,
0.04656554013490677,
0.05683058500289917,
-0.04089975729584694,
0.015249114483594894,
0.011805362068116665,
-0.13045944273471832,
-0.10874779522418976,
0.010989522561430931,
0.15458311140537262,
-0.07909547537565231,
-0.025835070759058,
-0.1036653071641922,
-0.04688705503940582,
0.05916960909962654,
-0.04689452797174454,
-0.04845086485147476,
-0.06529805064201355,
0.015913978219032288,
0.08518365025520325,
-0.02813866175711155,
0.005188861396163702,
-0.038892313838005066,
0.07894497364759445,
-0.08180247992277145,
-0.15058307349681854,
0.07720887660980225,
-0.019096799194812775,
-0.11984112113714218,
-0.09659121185541153,
0.1064775288105011,
-0.00850335881114006,
-0.00336822890676558,
-0.015049038454890251,
0.03584633395075798,
0.024472184479236603,
-0.0949539765715599,
0.08284656703472137,
0.14138762652873993,
-0.11951662600040436,
0.08751257508993149,
-0.03609180450439453,
-0.008353895507752895,
-0.051783159375190735,
0.008006689138710499,
0.11435526609420776,
0.23708970844745636,
-0.10293562710285187,
0.09419462084770203,
0.022321602329611778,
-0.08667124807834625,
-0.1952994465827942,
-0.014769538305699825,
-0.0021213763393461704,
0.03829343616962433,
0.01652071811258793,
-0.09201062470674515,
0.10008629411458969,
0.05155745893716812,
0.0029268977232277393,
0.2542369067668915,
-0.3790925145149231,
-0.14453691244125366,
-0.019906627014279366,
0.15968157351016998,
0.24026772379875183,
-0.16969752311706543,
-0.07967353612184525,
-0.02607867680490017,
-0.07353286445140839,
0.15346787869930267,
-0.045787811279296875,
0.0985286682844162,
-0.0322524793446064,
-0.03415520861744881,
0.027305446565151215,
-0.03809938579797745,
0.11587652564048767,
-0.0070856837555766106,
0.04776732623577118,
-0.08498542010784149,
-0.07991649210453033,
0.0827866941690445,
-0.041713763028383255,
0.020432734861969948,
-0.14520210027694702,
0.010113086551427841,
-0.06771478056907654,
-0.01193829532712698,
0.012363247573375702,
-0.0013986711855977774,
-0.021818790584802628,
-0.038545090705156326,
-0.11062075942754745,
0.018612416461110115,
-0.009190519340336323,
-0.012345358729362488,
0.0768439844250679,
-0.02003147266805172,
0.04261121153831482,
0.1435038447380066,
-0.03052370436489582,
0.06049887835979462,
-0.06982533633708954,
-0.015033663250505924,
-0.03915785998106003,
0.11067766696214676,
-0.16621935367584229,
-0.001920429291203618,
0.15103386342525482,
0.08426020294427872,
0.10053417831659317,
0.022479703649878502,
-0.09515807032585144,
0.09171885251998901,
0.13182635605335236,
-0.08787329494953156,
-0.011774555779993534,
-0.027886154130101204,
-0.03895745426416397,
0.1094108521938324,
0.01906515285372734,
0.18613214790821075,
-0.07754961401224136,
0.0405406653881073,
0.016877194866538048,
0.00751029746606946,
-0.017208239063620567,
0.08661099523305893,
0.04017716646194458,
0.005723201669752598,
-0.06811929494142532,
0.0558452308177948,
-0.01248290203511715,
0.016655657440423965,
0.010945789515972137,
0.06970028579235077,
-0.06392413377761841,
-0.011781214736402035,
0.008447197265923023,
0.20203185081481934,
-0.10351057350635529,
0.009993305429816246,
-0.10702608525753021,
-0.10653061419725418,
0.023334961384534836,
0.1592043936252594,
0.05443944036960602,
0.006893273908644915,
-0.04025273397564888,
-0.024759892374277115,
-0.05429800972342491,
0.03898369520902634,
0.059652406722307205,
0.07637076824903488,
-0.23308521509170532,
-0.028016014024615288,
0.01498111616820097,
-0.032238081097602844,
-0.07133933901786804,
-0.038721635937690735,
-0.11363272368907928,
-0.022022148594260216,
-0.03643966466188431,
0.10991808027029037,
-0.05700239539146423,
-0.020528091117739677,
-0.013454734347760677,
-0.039845045655965805,
-0.02019309811294079,
0.013778693042695522,
-0.03979776054620743,
-0.016050158068537712,
0.0013366822386160493,
-0.01107795350253582,
-0.053478315472602844,
-0.08516447991132736,
-0.014902912080287933,
-0.09314320981502533,
0.0422285795211792,
-0.01747868023812771,
-0.07660457491874695,
0.014480250887572765,
-0.22524647414684296,
0.034930210560560226,
0.1821153610944748,
-0.004053331445902586,
-0.01108773797750473,
0.015767082571983337,
0.018086755648255348,
-0.033949658274650574,
0.07735703140497208,
0.001190165407024324,
0.058055367320775986,
-0.07741210609674454,
-0.011815165169537067,
-0.07350485026836395,
-0.006107613909989595,
-0.055290337651968,
0.0711207315325737,
0.12678216397762299,
0.14138171076774597,
0.1272771805524826,
-0.10297728329896927,
0.0798184871673584,
-0.11208395659923553,
0.014515606686472893,
0.020970283076167107,
-0.06997498869895935,
0.025694992393255234,
-0.025029879063367844,
-0.012171813286840916,
-0.04987766221165657,
0.14694081246852875,
0.011714240536093712,
-0.14285235106945038,
-0.020898794755339622,
0.05307836830615997,
-0.06882023066282272,
0.024102739989757538,
0.17778030037879944,
0.035099804401397705,
0.054697636514902115,
-0.1130315363407135,
0.05743202194571495,
0.16680355370044708,
0.08059245347976685,
0.07213893532752991,
0.034875668585300446,
-0.0039161089807748795,
0.09504704922437668,
0.09331244230270386,
0.012849390506744385,
0.027224527671933174,
0.14041221141815186,
-0.046797752380371094,
0.12710773944854736,
-0.06306220591068268,
-0.009246826171875,
0.11080794036388397,
-0.028964495286345482,
-0.028982238844037056,
0.07898515462875366,
-0.04808035492897034,
-0.07057083398103714,
-0.13866637647151947,
-0.06756004691123962,
-0.15011711418628693,
0.03476601466536522,
-0.04494220390915871,
-0.0029460578225553036,
-0.00862309243530035,
0.05448642745614052,
0.061279598623514175,
0.03670012205839157,
-0.016324348747730255,
-0.04786834120750427,
0.14299121499061584,
-0.017961928620934486,
-0.10476014763116837,
0.014549206010997295,
0.025163548067212105,
0.04722118377685547,
0.02126329205930233,
-0.041314367204904556,
0.07115349173545837,
0.06726939976215363,
-0.004118652548640966,
-0.020570596680045128,
-0.0826282724738121,
-0.020444480702280998,
-0.0007801253232173622,
0.00378950173035264,
0.13386061787605286,
0.08258241415023804,
-0.046381089836359024,
-0.04312576353549957,
0.12494907528162003,
-0.04604757949709892,
-0.01654989644885063,
-0.12741665542125702,
0.013055805116891861,
-0.13396745920181274,
0.028979523107409477,
-0.0726531445980072,
-0.09909548610448837,
-0.04503459110856056,
0.20093299448490143,
0.21518562734127045,
-0.07466297596693039,
0.03571831434965134,
-0.025606989860534668,
-0.005865235812962055,
0.011621295474469662,
0.08178704977035522,
0.025334810838103294,
0.23699824512004852,
-0.042567674070596695,
-0.015511632896959782,
-0.08188240975141525,
-0.05238429829478264,
-0.04101480543613434,
-0.0677781030535698,
0.02445944771170616,
-0.045903269201517105,
-0.09124352782964706,
0.06739615648984909,
-0.14292645454406738,
-0.13578572869300842,
0.1469876617193222,
-0.19399957358837128,
-0.04436710104346275,
-0.05973048135638237,
0.05110796540975571,
0.040536776185035706,
0.025204580277204514,
-0.07401380687952042,
-0.021916083991527557,
0.003665306605398655,
-0.008297248743474483,
-0.16026224195957184,
-0.018606144934892654,
-0.08421655744314194,
-0.1707712858915329,
0.07204151898622513,
-0.022090284153819084,
0.0695299282670021,
0.049273908138275146,
0.018450450152158737,
-0.052820149809122086,
0.033286985009908676,
-0.04045113921165466,
-0.0689159706234932,
-0.09235561639070511,
0.09607861936092377,
-0.02399926632642746,
0.08546322584152222,
0.029929574579000473,
-0.0904746949672699,
0.024320058524608612,
0.06929364055395126,
-0.0920771062374115,
-0.09393647313117981,
-0.02280461974442005,
-0.07781423628330231,
0.09361373633146286,
0.08818564563989639,
-0.017490334808826447,
-0.00442441226914525,
0.007104700431227684,
0.013883836567401886,
0.01279937382787466,
-0.09604765474796295,
0.033282577991485596,
-0.07627931982278824,
-0.02096392586827278,
-0.0063968850299716,
0.02650025114417076,
-0.23762933909893036,
-0.07757503539323807,
-0.1556054949760437,
-0.01214381493628025,
-0.0185533594340086,
0.07773834466934204,
0.2324351966381073,
0.03069939650595188,
0.0026952270418405533,
-0.16656982898712158,
0.055784646421670914,
0.08317137509584427,
-0.05393415316939354,
-0.05615796148777008
] |
null | null | ml-agents |
# **ppo** Agent playing **SnowballTarget**
This is a trained model of a **ppo** agent playing **SnowballTarget**
using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
## Usage (with ML-Agents)
The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/
We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
- A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your
browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction
- A *longer tutorial* to understand how works ML-Agents:
https://huggingface.co/learn/deep-rl-course/unit5/introduction
### Resume the training
```bash
mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume
```
### Watch your Agent play
You can watch your agent **playing directly in your browser**
1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity
2. Step 1: Find your model_id: tizayi/ppo-SnowballTarget
3. Step 2: Select your *.nn /*.onnx file
4. Click on Watch the agent play 👀
| {"library_name": "ml-agents", "tags": ["SnowballTarget", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-SnowballTarget"]} | reinforcement-learning | tizayi/ppo-SnowballTarget | [
"ml-agents",
"tensorboard",
"onnx",
"SnowballTarget",
"deep-reinforcement-learning",
"reinforcement-learning",
"ML-Agents-SnowballTarget",
"region:us"
] | 2024-02-07T15:38:12+00:00 | [] | [] | TAGS
#ml-agents #tensorboard #onnx #SnowballTarget #deep-reinforcement-learning #reinforcement-learning #ML-Agents-SnowballTarget #region-us
|
# ppo Agent playing SnowballTarget
This is a trained model of a ppo agent playing SnowballTarget
using the Unity ML-Agents Library.
## Usage (with ML-Agents)
The Documentation: URL
We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
- A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your
browser: URL
- A *longer tutorial* to understand how works ML-Agents:
URL
### Resume the training
### Watch your Agent play
You can watch your agent playing directly in your browser
1. If the environment is part of ML-Agents official environments, go to URL
2. Step 1: Find your model_id: tizayi/ppo-SnowballTarget
3. Step 2: Select your *.nn /*.onnx file
4. Click on Watch the agent play
| [
"# ppo Agent playing SnowballTarget\n This is a trained model of a ppo agent playing SnowballTarget\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: tizayi/ppo-SnowballTarget\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play"
] | [
"TAGS\n#ml-agents #tensorboard #onnx #SnowballTarget #deep-reinforcement-learning #reinforcement-learning #ML-Agents-SnowballTarget #region-us \n",
"# ppo Agent playing SnowballTarget\n This is a trained model of a ppo agent playing SnowballTarget\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: tizayi/ppo-SnowballTarget\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play"
] | [
50,
206
] | [
"passage: TAGS\n#ml-agents #tensorboard #onnx #SnowballTarget #deep-reinforcement-learning #reinforcement-learning #ML-Agents-SnowballTarget #region-us \n# ppo Agent playing SnowballTarget\n This is a trained model of a ppo agent playing SnowballTarget\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: tizayi/ppo-SnowballTarget\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play"
] | [
-0.04098378121852875,
0.0529199056327343,
-0.003549708053469658,
0.10989358276128769,
0.182369664311409,
-0.008339453488588333,
0.16344261169433594,
0.10107908397912979,
0.12922435998916626,
0.05771038681268692,
0.08012408018112183,
0.08950730413198471,
0.06882720440626144,
0.1351473182439804,
0.0794931948184967,
-0.23413294553756714,
-0.048011843115091324,
-0.10449908673763275,
-0.0018571363762021065,
0.08271785825490952,
0.04468458518385887,
-0.03373988717794418,
0.03420308604836464,
0.054149042814970016,
-0.013200533576309681,
-0.0021219193004071712,
-0.06870964169502258,
-0.044829726219177246,
0.06800080090761185,
-0.0272233709692955,
0.0161039587110281,
-0.053344883024692535,
0.09825397282838821,
-0.15128284692764282,
0.02825840190052986,
0.050211191177368164,
-0.01252512726932764,
-0.027781883254647255,
0.14913398027420044,
0.021974077448248863,
0.11340059340000153,
-0.11480125784873962,
0.09507449716329575,
0.07622415572404861,
-0.05976797267794609,
0.004587485454976559,
-0.07227209955453873,
0.06150665879249573,
0.20108813047409058,
0.14492012560367584,
-0.00511375954374671,
0.08886195719242096,
-0.029426563531160355,
0.0647302120923996,
0.173537015914917,
-0.28197959065437317,
-0.07274053990840912,
0.16157184541225433,
-0.03830846771597862,
0.032431576400995255,
-0.01700725592672825,
0.03988894820213318,
-0.020000385120511055,
0.022813888266682625,
-0.02082325518131256,
0.030785931274294853,
0.2675333619117737,
0.02055923268198967,
-0.100214384496212,
-0.0824086144566536,
0.004630753304809332,
0.027874434366822243,
-0.054017722606658936,
-0.17895154654979706,
0.00552399642765522,
0.11186398565769196,
0.010367024689912796,
0.028273746371269226,
0.054212599992752075,
0.019534412771463394,
-0.08308723568916321,
-0.14887776970863342,
-0.041613031178712845,
-0.05392478406429291,
0.09620925784111023,
0.10088382661342621,
-0.02287670411169529,
-0.00985200796276331,
0.03908514231443405,
0.07816862314939499,
0.11934881657361984,
-0.03944442421197891,
-0.03567606210708618,
-0.021447235718369484,
-0.15477532148361206,
-0.023447897285223007,
-0.03258021920919418,
-0.021612176671624184,
0.0358271524310112,
0.14682091772556305,
0.17139588296413422,
0.029796432703733444,
0.02625354193150997,
0.028123529627919197,
0.005864174570888281,
0.11615810543298721,
0.03744589909911156,
-0.019981669262051582,
-0.0019492002902552485,
0.011789297685027122,
0.050143562257289886,
-0.09253565222024918,
-0.09391118586063385,
0.05430150404572487,
-0.039227139204740524,
0.12933531403541565,
0.16247205436229706,
-0.018924815580248833,
-0.015735993161797523,
-0.04037395119667053,
0.022768516093492508,
-0.1449621468782425,
0.07361099123954773,
0.05369172617793083,
-0.04395667091012001,
-0.07198955118656158,
-0.060804352164268494,
0.05503930523991585,
-0.08684615045785904,
0.03416658192873001,
0.0031746188178658485,
0.07361357659101486,
0.00562661699950695,
-0.039612092077732086,
0.0474630668759346,
-0.1338263303041458,
-0.011642860248684883,
-0.17178532481193542,
-0.115993931889534,
-0.07950884103775024,
0.03891054540872574,
-0.04847198352217674,
-0.10767044126987457,
-0.09711021929979324,
0.031614307314157486,
-0.0708157941699028,
0.03272192180156708,
-0.037277236580848694,
-0.06380008161067963,
-0.026633944362401962,
-0.1218406930565834,
0.05737626925110817,
0.165813609957695,
0.006489320192486048,
-0.031023725867271423,
0.024677693843841553,
-0.14962482452392578,
0.15709297358989716,
-0.12849430739879608,
0.15462730824947357,
-0.08079733699560165,
0.03905471786856651,
0.1292700320482254,
-0.023967692628502846,
0.05051327124238014,
0.19915105402469635,
-0.11122147738933563,
-0.06642105430364609,
0.05281524732708931,
-0.09002136439085007,
-0.11080270260572433,
0.06570585817098618,
0.023760898038744926,
0.06481507420539856,
0.06607804447412491,
0.20425748825073242,
0.09689196199178696,
-0.24670380353927612,
0.04493148997426033,
0.006199343595653772,
-0.1374945491552353,
0.009624255821108818,
0.12840886414051056,
-0.07198239117860794,
-0.003205646062269807,
-0.04442322626709938,
-0.12342221289873123,
0.10113406181335449,
-0.014323818497359753,
-0.06112702563405037,
0.03013695776462555,
-0.053861308842897415,
-0.03743404895067215,
-0.00431743124499917,
0.04046843945980072,
-0.040764398872852325,
-0.04933340847492218,
-0.022954534739255905,
0.03604642674326897,
0.0029638810083270073,
0.07594636082649231,
-0.04218924045562744,
0.11589580774307251,
-0.00806925818324089,
0.008964641019701958,
-0.10415187478065491,
-0.13978052139282227,
-0.014807204715907574,
0.039243850857019424,
0.08433466404676437,
-0.08207552134990692,
0.09946483373641968,
0.08184745162725449,
0.03919168561697006,
-0.07373923808336258,
-0.06056208536028862,
0.019914336502552032,
-0.10581814497709274,
-0.10809377580881119,
-0.06906905770301819,
-0.05947890877723694,
0.11028288304805756,
-0.09085224568843842,
0.06596285104751587,
-0.0577579140663147,
0.08906977623701096,
-0.013549931347370148,
-0.07485060393810272,
0.03790700063109398,
-0.01854090206325054,
0.03361358493566513,
-0.1068611592054367,
0.10067728161811829,
0.07300616055727005,
-0.125491663813591,
0.036353323608636856,
0.04969164356589317,
-0.0781359076499939,
0.12885965406894684,
0.046310629695653915,
-0.012887101620435715,
-0.043387383222579956,
-0.06028680130839348,
0.0037583978846669197,
-0.06863429397344589,
0.02384813316166401,
0.2055853009223938,
0.12759093940258026,
0.07760853320360184,
-0.03620769456028938,
-0.054980844259262085,
-0.027286481112241745,
-0.05628794804215431,
-0.06348063051700592,
0.13000820577144623,
0.04217938333749771,
-0.04052901268005371,
0.03546704724431038,
0.0013320902362465858,
0.10189229249954224,
0.12133844196796417,
-0.005130067002028227,
-0.12240750342607498,
0.017917199060320854,
0.06283443421125412,
0.06595512479543686,
0.006589919328689575,
0.04562569782137871,
-0.023731021210551262,
-0.012475389987230301,
-0.06276204437017441,
-0.0208546444773674,
-0.10733989626169205,
-0.06420833617448807,
0.06508959829807281,
-0.009438042528927326,
-0.006434346083551645,
-0.06961895525455475,
-0.044936224818229675,
0.033226024359464645,
0.10084925591945648,
-0.003798249876126647,
0.0329916849732399,
-0.03088686242699623,
-0.1294674128293991,
0.04041530191898346,
-0.0894462838768959,
-0.23845547437667847,
-0.10242882370948792,
-0.056789856404066086,
-0.0630212053656578,
0.02538285404443741,
0.07570388913154602,
-0.1911509931087494,
-0.0008297935128211975,
-0.09626291692256927,
0.005728547461330891,
-0.009324065409600735,
-0.027396146208047867,
0.13604237139225006,
0.09620498865842819,
-0.030323157086968422,
-0.06664206832647324,
0.012255567125976086,
0.01579935848712921,
-0.07138976454734802,
-0.004269793629646301,
0.06230081990361214,
0.09430377185344696,
0.06814081221818924,
0.0622490718960762,
0.06017772853374481,
-0.033957332372665405,
0.15253770351409912,
-0.05997501686215401,
0.036763157695531845,
0.076695017516613,
-0.011441041715443134,
0.07667173445224762,
0.01665789820253849,
0.024466006085276604,
-0.003134249011054635,
0.015294126234948635,
0.006290608085691929,
-0.07949943095445633,
-0.21455761790275574,
-0.07382936775684357,
-0.006992490030825138,
0.18034577369689941,
0.16712896525859833,
0.09635119885206223,
-0.10476502031087875,
0.029447777196764946,
0.0003757561498787254,
-0.10593870282173157,
0.11285550892353058,
0.13088764250278473,
-0.08051309734582901,
-0.015704648569226265,
0.03346430882811546,
-0.04271337389945984,
0.05449610576033592,
0.06539816409349442,
-0.041257694363594055,
0.09736967831850052,
0.02528228797018528,
-0.005307361483573914,
-0.023376334458589554,
-0.0463639460504055,
-0.05875924974679947,
0.13625065982341766,
0.06723589450120926,
0.02545555680990219,
0.011573567055165768,
-0.06245598569512367,
-0.07167071104049683,
0.13265979290008545,
0.16241782903671265,
-0.07184037566184998,
-0.04475639387965202,
0.10736007988452911,
0.057448506355285645,
0.20028088986873627,
-0.000418262294260785,
-0.12171950191259384,
-0.06789018958806992,
-0.010765139013528824,
-0.11544956266880035,
-0.00003228462810511701,
0.030497848987579346,
-0.016559522598981857,
-0.16793116927146912,
0.05304029583930969,
-0.0011710247490555048,
0.11186084896326065,
0.023505037650465965,
-0.038036707788705826,
0.07145557552576065,
0.006709098815917969,
-0.03693971410393715,
0.04183237627148628,
-0.16243626177310944,
0.024546068161725998,
-0.008329181000590324,
0.09778161346912384,
-0.06828241050243378,
0.02054610475897789,
0.08767081052064896,
-0.03720792755484581,
0.1611662060022354,
0.047171588987112045,
-0.00006823080911999568,
-0.13103270530700684,
-0.1679753214120865,
-0.06063307076692581,
-0.014066667295992374,
-0.11346833407878876,
0.0751575380563736,
0.02830325812101364,
-0.015582310035824776,
-0.10009351372718811,
0.02613743580877781,
-0.036952562630176544,
-0.12257689237594604,
-0.03908638283610344,
-0.09073127061128616,
0.07615601271390915,
-0.05476415157318115,
-0.07411961257457733,
-0.09470679610967636,
0.1776716113090515,
0.09314217418432236,
-0.10637100040912628,
-0.12333746999502182,
0.003860590048134327,
-0.05538829043507576,
-0.02675521932542324,
0.0726916715502739,
0.017282584682106972,
0.10320502519607544,
-0.10639213770627975,
-0.057899635285139084,
-0.024791177362203598,
-0.10941940546035767,
-0.09747610241174698,
0.029629483819007874,
0.1698271632194519,
0.04384578764438629,
0.08917541801929474,
-0.009674817323684692,
0.09496860951185226,
-0.007854023016989231,
-0.06834184378385544,
0.11049678176641464,
0.08961082249879837,
-0.022387325763702393,
0.05761708691716194,
0.038153450936079025,
0.05888863280415535,
-0.13361139595508575,
-0.02621307782828808,
0.20781491696834564,
0.26472008228302,
-0.06367078423500061,
0.1991555094718933,
-0.007024930790066719,
-0.04344520717859268,
-0.1532459408044815,
-0.052700627595186234,
0.01673869602382183,
-0.050537895411252975,
0.10263630747795105,
-0.18782684206962585,
0.08497937023639679,
-0.006532377563416958,
-0.004863146226853132,
0.05165965110063553,
-0.15187136828899384,
-0.08463441580533981,
0.02394740842282772,
0.10007314383983612,
-0.05620739981532097,
-0.10274745523929596,
-0.0692826360464096,
0.022953925654292107,
-0.06416765600442886,
0.028095945715904236,
-0.10488197207450867,
0.05891454592347145,
0.023563414812088013,
0.027999674901366234,
0.060603611171245575,
-0.051750071346759796,
0.12299831956624985,
-0.038988932967185974,
-0.062033023685216904,
-0.074203722178936,
0.032660674303770065,
-0.006521870847791433,
-0.08562731742858887,
0.052219342440366745,
-0.002434292109683156,
-0.020653149113059044,
-0.18471768498420715,
-0.05465877428650856,
0.022856412455439568,
0.03643636777997017,
-0.029973700642585754,
-0.08569569140672684,
-0.028931336477398872,
0.06259029358625412,
0.08713486790657043,
0.017723163589835167,
0.10930006951093674,
0.001192907802760601,
-0.011129175312817097,
0.04222127050161362,
0.0310501828789711,
0.037075672298669815,
-0.1325163096189499,
-0.07111836969852448,
-0.06776129454374313,
-0.0006896533886902034,
-0.044814012944698334,
-0.022198613733053207,
0.04685387760400772,
0.05934849753975868,
-0.006959015969187021,
0.05962630733847618,
-0.08197775483131409,
-0.013341798447072506,
0.024468421936035156,
-0.09079927206039429,
-0.11531445384025574,
-0.08583271503448486,
-0.10917658358812332,
0.021336935460567474,
-0.07487473636865616,
0.0856672003865242,
-0.05265228822827339,
-0.0002655634598340839,
0.013331345282495022,
0.033426683396101,
-0.0036436435766518116,
0.03249802067875862,
0.019214164465665817,
0.027740320190787315,
-0.07345886528491974,
0.1239372119307518,
0.013188697397708893,
-0.04662786424160004,
0.05215933918952942,
0.189858078956604,
-0.06342898309230804,
-0.06667077541351318,
-0.04515092074871063,
0.06531614065170288,
0.04408220946788788,
-0.021405426785349846,
-0.035508036613464355,
-0.04979728162288666,
0.12192501872777939,
-0.15009033679962158,
0.010588173754513264,
-0.11245087534189224,
0.003139214124530554,
0.05775299295783043,
-0.054886095225811005,
0.06277627497911453,
-0.015598032623529434,
-0.05988084524869919,
-0.1495409458875656,
0.07729826867580414,
0.02686776965856552,
0.09308472275733948,
-0.01311352476477623,
-0.02997097745537758,
-0.13766837120056152,
0.03005053848028183,
0.011607196182012558,
0.010924785397946835,
-0.169045552611351,
0.02085859701037407,
-0.004023473244160414,
0.018491171300411224,
0.03427879884839058,
0.061175886541604996,
-0.03803861886262894,
-0.09178780019283295,
-0.05070106312632561,
0.061063796281814575,
-0.0840589851140976,
-0.023145440965890884,
-0.026467368006706238,
-0.08518654108047485,
0.060855764895677567,
0.08051826804876328,
-0.031659796833992004,
-0.04757829010486603,
-0.061810001730918884,
0.018396340310573578,
-0.023122312501072884,
-0.04697102680802345,
0.04558492824435234,
-0.127726748585701,
0.02251792512834072,
-0.06382514536380768,
-0.11664263904094696,
0.0360114760696888,
0.12829530239105225,
-0.06363853812217712,
0.04619182273745537,
0.044716790318489075,
-0.08108962327241898,
-0.06899017095565796,
-0.008936959318816662,
0.07002992182970047,
0.04742471128702164,
0.11024948209524155,
-0.07536568492650986,
0.1986466348171234,
-0.10611319541931152,
-0.032475389540195465,
0.010789576917886734,
0.07239305973052979,
0.03385153040289879,
-0.09288091212511063,
0.04158269613981247,
-0.01315687783062458,
0.04815317317843437,
0.07473523169755936,
0.015389840118587017,
0.04881086200475693,
0.051684461534023285,
0.1290411502122879,
0.01243575755506754,
0.09520698338747025,
-0.008036455139517784,
0.017129996791481972,
0.10040397197008133,
-0.007381570991128683,
0.06918653845787048,
-0.07066506892442703,
0.06761480867862701,
0.06277016550302505,
0.07265552878379822,
0.06780099868774414,
0.04709988459944725,
-0.0969739556312561,
-0.18666045367717743,
-0.057585593312978745,
0.0320846252143383,
0.02992001734673977,
-0.04036248102784157,
0.1826094388961792,
0.13986286520957947,
-0.19815145432949066,
0.013509061187505722,
0.002263470785692334,
0.04164675250649452,
-0.06561879068613052,
-0.09108549356460571,
0.007701550144702196,
-0.13661056756973267,
0.09699695557355881,
-0.013723512180149555,
0.00938367284834385,
-0.027335871011018753,
0.014562109485268593,
0.030491597950458527,
0.052522264420986176,
-0.049070410430431366,
-0.010166038759052753,
0.052068326622247696,
-0.03190019354224205,
0.006981526035815477,
-0.002153052482753992,
-0.09039614349603653,
-0.044448480010032654,
-0.0600823312997818,
-0.014615992084145546,
0.027369914576411247,
0.005656045861542225,
0.06088044121861458,
0.002088156994432211,
-0.061928726732730865,
0.07614651322364807,
0.004647333174943924,
0.01393088884651661,
0.19388121366500854,
0.0937906876206398,
-0.04468357935547829,
-0.04463585093617439,
0.20890755951404572,
-0.03384842351078987,
-0.07303118705749512,
-0.09097466617822647,
0.11311762779951096,
-0.04896659031510353,
-0.050488606095314026,
-0.0394461564719677,
-0.16609002649784088,
-0.05898142606019974,
0.17456208169460297,
0.12370084971189499,
-0.019923241809010506,
0.005798882804811001,
-0.06439116597175598,
0.005692438222467899,
0.028925567865371704,
0.1043548732995987,
0.06365318596363068,
0.06117650121450424,
-0.10105711966753006,
-0.002694229129701853,
-0.07452672719955444,
-0.09772974252700806,
-0.19333066046237946,
0.05550137162208557,
0.0303680170327425,
-0.025847913697361946,
-0.018936866894364357,
0.12911085784435272,
-0.10297761112451553,
-0.09720737487077713,
0.11406700313091278,
-0.032472867518663406,
-0.07459414750337601,
0.0008576067048124969,
0.02202761359512806,
0.004843822680413723,
0.1133904755115509,
0.08899756520986557,
0.042774323374032974,
0.015304460190236568,
-0.00886387750506401,
-0.0976245105266571,
0.03333142399787903,
0.05021558701992035,
-0.13223566114902496,
0.22518566250801086,
-0.02516271360218525,
0.00085493887308985,
0.09049301594495773,
0.0714409202337265,
-0.18268121778964996,
0.0034652368631213903,
0.05156494304537773,
-0.165214404463768,
0.024001633748412132,
0.07457488775253296,
-0.041195448487997055,
-0.0030552376993000507,
0.055432017892599106,
-0.04526424780488014,
0.01002510916441679,
0.185214102268219,
0.04524118825793266,
-0.042377352714538574,
0.08282893896102905,
-0.15402548015117645,
0.10899679362773895,
0.09371355175971985,
-0.059855345636606216,
0.008760214783251286,
-0.038066599518060684,
0.0047675869427621365,
-0.002753907348960638,
-0.019418559968471527,
-0.017698395997285843,
-0.1215875893831253,
-0.033026669174432755,
-0.060504186898469925,
0.03272201493382454,
-0.2089284062385559,
-0.12753762304782867,
-0.0580277293920517,
-0.07796721160411835,
-0.05617211386561394,
0.08387314528226852,
0.07416725903749466,
-0.052682992070913315,
0.013679921627044678,
-0.11236847937107086,
0.027562826871871948,
0.15471597015857697,
-0.07627357542514801,
0.0008323897491209209
] |
null | null | transformers | Finetune of miqu-70b-sf dequant of miqudev's leak of Mistral-70B (allegedly an early mistral medium). My diffs are available under CC-0, this is a merge with the leaked model, you can use the other repository to save bandwidth.
EQ-Bench: 84.89
Will run more benches later. | {"license": "cc-by-2.0"} | text-generation | LoneStriker/Senku-70B-Full-5.0bpw-h6-exl2 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"license:cc-by-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T15:38:23+00:00 | [] | [] | TAGS
#transformers #safetensors #llama #text-generation #conversational #license-cc-by-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| Finetune of miqu-70b-sf dequant of miqudev's leak of Mistral-70B (allegedly an early mistral medium). My diffs are available under CC-0, this is a merge with the leaked model, you can use the other repository to save bandwidth.
EQ-Bench: 84.89
Will run more benches later. | [] | [
"TAGS\n#transformers #safetensors #llama #text-generation #conversational #license-cc-by-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
60
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #conversational #license-cc-by-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
-0.006207775790244341,
0.05186443030834198,
-0.005766681861132383,
0.0034429291263222694,
0.09202545881271362,
-0.026229875162243843,
0.21461424231529236,
0.09413789957761765,
-0.012125303968787193,
-0.027240416035056114,
0.15366454422473907,
0.20107468962669373,
-0.030660197138786316,
0.0625392273068428,
-0.12453678250312805,
-0.15658724308013916,
0.07951316237449646,
-0.0074092973954975605,
0.0468166321516037,
0.08121465146541595,
0.11726208031177521,
-0.05624562129378319,
0.07327274978160858,
-0.05116863548755646,
-0.10777227580547333,
0.01719590090215206,
0.07997997105121613,
-0.1362335979938507,
0.10389529168605804,
0.070436492562294,
0.09596491605043411,
0.09478230774402618,
-0.026663783937692642,
-0.2320748120546341,
0.024640627205371857,
-0.013039902783930302,
-0.08545025438070297,
0.02922174707055092,
0.03788268566131592,
-0.042137615382671356,
0.05930780619382858,
0.04187946766614914,
-0.012704534456133842,
0.0843532532453537,
-0.12006022036075592,
0.03492521494626999,
-0.054476991295814514,
0.008457338437438011,
0.11366608738899231,
0.07924187183380127,
0.002242166781798005,
0.1153382733464241,
-0.05578871816396713,
0.08845382928848267,
0.06245912238955498,
-0.3711334466934204,
0.014689184725284576,
0.13522937893867493,
0.07748636603355408,
0.06413708627223969,
-0.05252519249916077,
0.11249673366546631,
0.06629528105258942,
-0.03810290992259979,
0.04312480613589287,
-0.07572442293167114,
-0.08102772384881973,
0.034424908459186554,
-0.04833368957042694,
-0.030284471809864044,
0.23118719458580017,
-0.03052249550819397,
0.012902313843369484,
-0.08178368955850601,
-0.06274984031915665,
0.003528473898768425,
-0.027013633400201797,
0.033650096505880356,
-0.007114062085747719,
0.09828754514455795,
0.0036289477720856667,
-0.04431737959384918,
-0.15287499129772186,
-0.017103472724556923,
-0.17628958821296692,
0.10724018514156342,
-0.0038528726436197758,
0.03643316775560379,
-0.1433914750814438,
0.03731580823659897,
0.01849939487874508,
-0.09380249679088593,
-0.01922903209924698,
-0.06579022109508514,
0.07399444282054901,
-0.015354927629232407,
-0.04418032616376877,
-0.04420563578605652,
0.146459698677063,
0.11814041435718536,
-0.01364491693675518,
0.003440734464675188,
-0.11379846185445786,
0.0978090688586235,
-0.019500313326716423,
-0.0008985437452793121,
0.047480762004852295,
-0.007553805597126484,
0.1015886589884758,
-0.08041958510875702,
0.08773300796747208,
-0.04154345020651817,
-0.16899511218070984,
0.014984861016273499,
-0.009690387174487114,
0.14272445440292358,
0.0067246644757688046,
0.08934520930051804,
-0.03848949074745178,
0.06400565803050995,
0.07495175302028656,
-0.06882821768522263,
0.002561564091593027,
0.008464166894555092,
0.05782800912857056,
0.020744536072015762,
0.0357230082154274,
0.05749142915010452,
-0.053222328424453735,
0.034963082522153854,
-0.05974286422133446,
-0.030262604355812073,
-0.04071192815899849,
-0.05133391171693802,
0.07166485488414764,
-0.026888888329267502,
0.023412950336933136,
-0.1889147162437439,
-0.17520999908447266,
0.017266716808080673,
0.006569101009517908,
-0.016733024269342422,
-0.0409536249935627,
-0.05526142567396164,
-0.029290098696947098,
0.025005020201206207,
-0.08925633877515793,
-0.08324684202671051,
-0.08618094772100449,
0.08566243946552277,
-0.034463070333004,
0.052008256316185,
-0.17173472046852112,
0.030293233692646027,
-0.09709536284208298,
0.015842683613300323,
-0.024641036987304688,
0.05080806463956833,
-0.03994987532496452,
0.1493525505065918,
-0.03130904585123062,
0.028674336150288582,
-0.049635149538517,
0.06748075783252716,
-0.04132431373000145,
0.20219798386096954,
-0.13369248807430267,
-0.02732842043042183,
0.23248150944709778,
-0.1203412264585495,
-0.2215833067893982,
0.09494956582784653,
-0.007244576700031757,
0.06883425265550613,
0.1192723736166954,
0.16428819298744202,
-0.018312968313694,
-0.084824338555336,
0.043344683945178986,
0.10571722686290741,
-0.05737517401576042,
-0.10720136761665344,
0.012378843501210213,
-0.02602921798825264,
-0.10123564302921295,
0.021436169743537903,
0.054138898849487305,
0.03870760649442673,
-0.010303346440196037,
-0.06431794166564941,
-0.04900350421667099,
-0.04556714743375778,
-0.019114533439278603,
-0.05745845288038254,
0.04891825467348099,
-0.10103435069322586,
-0.004564722999930382,
0.02500876598060131,
-0.0006331629119813442,
-0.017695071175694466,
0.03802654147148132,
-0.1019439622759819,
0.05491343140602112,
-0.01049356535077095,
0.05139922350645065,
-0.08506914973258972,
-0.09540282934904099,
-0.01487639732658863,
0.11419088393449783,
0.04237206280231476,
0.03597130998969078,
0.03009572997689247,
-0.0029274821281433105,
-0.016321944072842598,
0.019669566303491592,
0.1798858940601349,
0.027921929955482483,
-0.06157265976071358,
-0.10305018723011017,
0.10064617544412613,
-0.04469102993607521,
0.06838524341583252,
-0.12369509041309357,
0.021130846813321114,
0.09049873054027557,
0.08164099603891373,
0.011007205583155155,
0.06817352026700974,
-0.016652852296829224,
0.012067300267517567,
-0.08803333342075348,
0.018720898777246475,
0.09845291078090668,
0.023675929754972458,
-0.12399162352085114,
0.24828600883483887,
-0.22121906280517578,
0.23131431639194489,
0.20286236703395844,
-0.2072221040725708,
0.034866541624069214,
-0.11290061473846436,
0.008944046683609486,
0.009122604504227638,
0.00658374885097146,
-0.04158162698149681,
0.02772163227200508,
-0.017428645864129066,
0.17939260601997375,
-0.07259617745876312,
-0.01289988774806261,
-0.015334650874137878,
-0.07770753651857376,
-0.0461907684803009,
0.04808717593550682,
0.10923753678798676,
-0.17211377620697021,
0.18076997995376587,
0.2573873698711395,
0.016237052157521248,
0.145608589053154,
-0.02829170413315296,
0.011588740162551403,
0.031093206256628036,
0.051330193877220154,
0.017952319234609604,
-0.03206970542669296,
-0.09656007587909698,
-0.01615116000175476,
0.06039080768823624,
0.00738911610096693,
0.05942045524716377,
-0.14498206973075867,
-0.06179048866033554,
-0.010635633021593094,
-0.054405391216278076,
0.006760948803275824,
0.05542760342359543,
-0.006918448954820633,
0.12766680121421814,
-0.04561823606491089,
-0.06787645071744919,
0.1212611123919487,
-0.018814176321029663,
-0.11172524094581604,
0.16348372399806976,
-0.1353582739830017,
-0.2530854046344757,
-0.1794736236333847,
-0.16009610891342163,
-0.05283679813146591,
0.06249964237213135,
0.12733663618564606,
-0.021347085013985634,
-0.06735935807228088,
-0.08513124287128448,
-0.025516794994473457,
-0.004286123439669609,
-0.0013990295119583607,
-0.030478790402412415,
0.0740688145160675,
-0.04530498385429382,
-0.11495137214660645,
-0.04566733539104462,
0.02995317615568638,
-0.08298885077238083,
0.11527245491743088,
-0.08257944881916046,
0.09218067675828934,
0.14615941047668457,
0.013489138334989548,
-0.0007384112104773521,
-0.06414268910884857,
0.11027345061302185,
-0.05689837411046028,
-0.021911391988396645,
0.19550743699073792,
-0.06030816212296486,
0.054297950118780136,
0.17741087079048157,
0.019415782764554024,
-0.11608056724071503,
0.05627991259098053,
-0.04948503151535988,
-0.08943437039852142,
-0.2322319746017456,
-0.11231391131877899,
-0.09151040762662888,
0.09372707456350327,
0.02010270394384861,
0.06644773483276367,
0.1561785638332367,
0.07071852684020996,
-0.03820497542619705,
-0.003707042895257473,
0.09500505030155182,
0.10332844406366348,
0.26925957202911377,
-0.03648235276341438,
0.1326010376214981,
-0.09864376485347748,
-0.09166394174098969,
0.08177606761455536,
0.09666159003973007,
0.08463814854621887,
0.1308884173631668,
0.10701965540647507,
0.06300177425146103,
0.06616055965423584,
0.13583797216415405,
0.0930822342634201,
0.04747668653726578,
-0.027060847729444504,
-0.021886298432946205,
-0.06314191222190857,
-0.03154398128390312,
0.06433441489934921,
-0.0616009458899498,
-0.1419346034526825,
-0.01669052243232727,
-0.04656673222780228,
0.08677151799201965,
0.12965995073318481,
0.049826931208372116,
-0.20373210310935974,
0.03404708579182625,
0.12731851637363434,
-0.014891417697072029,
-0.0784522220492363,
0.11266084015369415,
0.041655853390693665,
-0.0492335706949234,
0.0968889519572258,
-0.022453399375081062,
0.10560388118028641,
-0.04346970468759537,
0.06621389091014862,
-0.09852857887744904,
-0.09218709915876389,
0.008176117204129696,
0.10175062716007233,
-0.32708150148391724,
0.19553345441818237,
0.025147762149572372,
0.003030276857316494,
-0.07397007197141647,
-0.00860549882054329,
0.006451844237744808,
0.1687123030424118,
0.14335986971855164,
-0.03559787943959236,
-0.11925935745239258,
-0.036741066724061966,
-0.04582417383790016,
0.021652963012456894,
0.114947110414505,
0.004123690538108349,
0.006702055223286152,
-0.060284681618213654,
-0.005426459014415741,
0.013130738399922848,
-0.03866162151098251,
-0.05004815012216568,
-0.18432049453258514,
0.036161795258522034,
0.15424391627311707,
0.105919748544693,
-0.038923926651477814,
0.025927383452653885,
-0.1276562660932541,
0.19439037144184113,
-0.1566786915063858,
-0.05010409653186798,
-0.10357701778411865,
-0.13866664469242096,
-0.010155638679862022,
-0.015465851873159409,
0.06662318855524063,
-0.05540037900209427,
0.05461542308330536,
-0.0985378548502922,
-0.16780197620391846,
0.11566402018070221,
-0.10529130697250366,
-0.03816051036119461,
-0.04050680622458458,
0.15799902379512787,
-0.10722509026527405,
-0.015996212139725685,
0.0695413127541542,
0.027312718331813812,
-0.04470572620630264,
-0.10069151222705841,
-0.01407475396990776,
0.022582462057471275,
0.03730671480298042,
-0.004337325692176819,
-0.1392161101102829,
-0.12033824622631073,
-0.007708914577960968,
-0.08476556092500687,
0.26667433977127075,
0.2823437750339508,
-0.052704282104969025,
0.1610942929983139,
0.18729859590530396,
-0.10833743214607239,
-0.3613292872905731,
-0.1156543493270874,
-0.19261738657951355,
-0.05812794342637062,
-0.0019929534755647182,
-0.09389892220497131,
0.0683603510260582,
0.03696022555232048,
-0.06046358868479729,
0.10693113505840302,
-0.20837023854255676,
-0.11674803495407104,
0.13628779351711273,
0.022366341203451157,
0.3236163854598999,
-0.1799478828907013,
-0.10869350284337997,
-0.1288607120513916,
-0.09091421216726303,
0.17097102105617523,
-0.11236412078142166,
0.09014905989170074,
0.038183145225048065,
0.05064551532268524,
0.019322404637932777,
-0.028820998966693878,
0.11060543358325958,
-0.05023352429270744,
0.05936712771654129,
-0.125066876411438,
-0.007887703366577625,
0.055251557379961014,
-0.021825825795531273,
0.047600388526916504,
-0.17949266731739044,
0.010565537959337234,
-0.02346070110797882,
-0.03863802179694176,
-0.00010827439837157726,
0.07584916800260544,
-0.0018114936538040638,
-0.05403559282422066,
-0.0338921919465065,
-0.07518398016691208,
0.021392593160271645,
-0.005685644224286079,
0.2621508240699768,
-0.09497393667697906,
0.143461674451828,
0.2097931206226349,
0.18348653614521027,
-0.11326921731233597,
0.10850489139556885,
-0.03179696202278137,
-0.10264644771814346,
0.06764788925647736,
-0.13615255057811737,
0.06894791126251221,
0.07885490357875824,
-0.05221731215715408,
0.09386573731899261,
0.07263122498989105,
0.024896983057260513,
0.011904848739504814,
0.15288308262825012,
-0.19621852040290833,
-0.06335753947496414,
-0.02627270668745041,
0.128641277551651,
0.07603052258491516,
0.07438892126083374,
0.19964289665222168,
-0.02363588474690914,
0.024383772164583206,
0.004766607191413641,
0.0470099002122879,
-0.051470812410116196,
0.03084268979728222,
-0.0036850213073194027,
0.009037865325808525,
-0.11732205748558044,
0.11256258189678192,
0.014299282804131508,
-0.13298094272613525,
0.018583929166197777,
0.10019229352474213,
-0.12479248642921448,
-0.1322273164987564,
-0.06397988647222519,
0.09235106408596039,
-0.1650686413049698,
-0.09092073142528534,
-0.035492803901433945,
-0.18309658765792847,
0.023424223065376282,
0.22705577313899994,
0.03624091297388077,
0.09397074580192566,
0.025309177115559578,
-0.050275810062885284,
-0.03333280235528946,
0.044489964842796326,
-0.0776023343205452,
0.026683444157242775,
-0.09567009657621384,
0.016401100903749466,
-0.0327458456158638,
0.04273902624845505,
-0.08503364771604538,
-0.01675247587263584,
-0.13592562079429626,
0.026447130367159843,
-0.1429377794265747,
0.016726532950997353,
-0.09382633119821548,
-0.022313788533210754,
0.025396376848220825,
-0.011300059035420418,
-0.044356830418109894,
-0.0456419438123703,
-0.08871109038591385,
0.022159568965435028,
-0.04314287006855011,
0.0626419335603714,
-0.09819427877664566,
-0.043954771012067795,
0.04525887221097946,
-0.036994993686676025,
0.11698898673057556,
0.04370922967791557,
-0.0972345620393753,
0.07480837404727936,
-0.24783827364444733,
-0.02655193954706192,
0.13125064969062805,
0.015579255297780037,
0.014812005683779716,
0.06315375864505768,
-0.011836273595690727,
0.1417415887117386,
0.009809926152229309,
0.052264366298913956,
0.013304419815540314,
-0.1011492908000946,
-0.001961590489372611,
-0.04560624808073044,
-0.10192413628101349,
-0.018009837716817856,
-0.0727795958518982,
0.1012018471956253,
-0.024938881397247314,
0.17435556650161743,
-0.092584989964962,
0.03382148593664169,
-0.01980471797287464,
0.03883249685168266,
0.005250109359622002,
-0.15618106722831726,
-0.14053024351596832,
-0.07882187515497208,
-0.001336804125458002,
-0.015131816267967224,
0.27824896574020386,
0.008481241762638092,
-0.06852035224437714,
0.09475342184305191,
0.022017354145646095,
0.034207794815301895,
0.04192714765667915,
0.2883480489253998,
0.08536436408758163,
-0.01155257411301136,
-0.15340177714824677,
0.016927950084209442,
0.0375761017203331,
-0.09943119436502457,
0.06783804297447205,
0.09215667098760605,
-0.07752043753862381,
0.10896968841552734,
0.07114557921886444,
0.0072129033505916595,
-0.018362488597631454,
-0.06760571897029877,
-0.055142998695373535,
0.046072885394096375,
-0.049081407487392426,
0.07889051735401154,
0.20583051443099976,
-0.03462229669094086,
-0.015487121418118477,
-0.04577798396348953,
-0.03612710162997246,
-0.18539372086524963,
-0.13403363525867462,
-0.1109817773103714,
-0.11150851845741272,
0.01701277866959572,
-0.08830517530441284,
0.055540017783641815,
0.05017261207103729,
0.054319605231285095,
-0.04291510581970215,
0.06990589201450348,
-0.05907499045133591,
-0.04540880024433136,
0.0442875400185585,
-0.02594553679227829,
0.06306269764900208,
-0.035188764333724976,
-0.06933852285146713,
-0.053066983819007874,
-0.04323925822973251,
-0.02944331243634224,
0.08734598755836487,
0.050012290477752686,
0.06000626087188721,
-0.13524575531482697,
-0.06855671852827072,
-0.032354552298784256,
0.0697886273264885,
-0.008020002394914627,
0.1599365472793579,
0.022465955466032028,
-0.0408162996172905,
0.08315412700176239,
0.1788318008184433,
-0.06562280654907227,
-0.12326370179653168,
-0.031882673501968384,
0.1764252483844757,
0.0044775111600756645,
0.13507580757141113,
-0.040701255202293396,
-0.013290504924952984,
-0.012223204597830772,
0.3287609815597534,
0.25894030928611755,
-0.09469550848007202,
0.03137436881661415,
-0.07478570938110352,
0.04246385395526886,
0.05169496685266495,
0.11560044437646866,
0.0689227506518364,
0.25144103169441223,
-0.03857225552201271,
-0.03432202711701393,
-0.00518826674669981,
0.017930803820490837,
-0.13905346393585205,
0.0863647311925888,
-0.04497753456234932,
-0.05007867515087128,
-0.02601996622979641,
0.09876758605241776,
-0.15639275312423706,
0.06847500801086426,
-0.04330084100365639,
-0.09984272718429565,
0.007430274039506912,
0.007269448135048151,
0.14150775969028473,
-0.019860094413161278,
0.029579713940620422,
-0.03768281638622284,
-0.07121695578098297,
0.01726885512471199,
-0.005519684869796038,
-0.18614327907562256,
0.03858964890241623,
0.023371171206235886,
-0.005911254324018955,
0.07187901437282562,
0.000053612515330314636,
0.06824749708175659,
0.07142221182584763,
0.030734332278370857,
-0.051403749734163284,
0.15974754095077515,
0.026775244623422623,
-0.0868566483259201,
0.050569936633110046,
-0.040603864938020706,
-0.026959538459777832,
0.04548517242074013,
0.06765088438987732,
-0.09669996052980423,
0.06760630011558533,
0.009751532226800919,
-0.11920495331287384,
-0.03518637642264366,
0.010957258753478527,
-0.07562252134084702,
0.07157810032367706,
0.007729232311248779,
-0.028513247147202492,
0.004583296831697226,
-0.016401249915361404,
0.010413174517452717,
-0.018589364364743233,
-0.15951070189476013,
-0.017387747764587402,
-0.12323169410228729,
-0.05553334206342697,
0.11709286272525787,
0.03462449461221695,
-0.23784784972667694,
0.011705098673701286,
-0.0969860851764679,
0.06600788980722427,
-0.19047285616397858,
0.05791420489549637,
0.2114742547273636,
-0.005059413146227598,
-0.031710248440504074,
-0.17269471287727356,
0.06433253735303879,
0.061398617923259735,
-0.036462731659412384,
-0.10163070261478424
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# hubert_RTSPsplit_0208_1
This model is a fine-tuned version of [rinna/japanese-hubert-base](https://huggingface.co/rinna/japanese-hubert-base) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2473
- Wer: 0.5480
- Cer: 0.3836
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 32
- eval_batch_size: 32
- seed: 4
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 1000
- num_epochs: 30
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|
| 13.3667 | 1.0 | 60 | 10.9759 | 0.9650 | 0.9893 |
| 6.7554 | 2.0 | 120 | 5.7617 | 0.9650 | 0.9893 |
| 4.9638 | 3.0 | 180 | 4.5906 | 0.9650 | 0.9893 |
| 3.8281 | 4.0 | 240 | 3.6487 | 0.9650 | 0.9893 |
| 3.2137 | 5.0 | 300 | 3.0965 | 0.9650 | 0.9893 |
| 2.6918 | 6.0 | 360 | 2.5698 | 0.9650 | 0.9893 |
| 1.9548 | 7.0 | 420 | 1.7758 | 1.0 | 0.7781 |
| 1.4576 | 8.0 | 480 | 1.2590 | 1.0 | 0.5555 |
| 1.1469 | 9.0 | 540 | 1.0586 | 1.0 | 0.5343 |
| 0.9577 | 10.0 | 600 | 0.8502 | 0.8095 | 0.4601 |
| 0.9391 | 11.0 | 660 | 0.7605 | 0.8069 | 0.4735 |
| 0.7744 | 12.0 | 720 | 0.7378 | 0.8025 | 0.5018 |
| 0.7492 | 13.0 | 780 | 0.7191 | 0.7920 | 0.5526 |
| 0.683 | 14.0 | 840 | 0.6538 | 0.7827 | 0.5061 |
| 0.6832 | 15.0 | 900 | 0.6730 | 0.7857 | 0.4839 |
| 0.6235 | 16.0 | 960 | 0.5698 | 0.7772 | 0.4543 |
| 0.5675 | 17.0 | 1020 | 0.5220 | 0.6957 | 0.3223 |
| 1.1877 | 18.0 | 1080 | 0.4777 | 0.7861 | 0.4389 |
| 0.498 | 19.0 | 1140 | 0.4616 | 0.7150 | 0.4252 |
| 0.4866 | 20.0 | 1200 | 0.4383 | 0.7180 | 0.3752 |
| 0.4689 | 21.0 | 1260 | 0.4194 | 0.7232 | 0.3532 |
| 0.5681 | 22.0 | 1320 | 0.4201 | 0.6734 | 0.3000 |
| 0.4051 | 23.0 | 1380 | 0.3938 | 0.6845 | 0.3390 |
| 0.3985 | 24.0 | 1440 | 0.3572 | 0.6607 | 0.3760 |
| 0.375 | 25.0 | 1500 | 0.3432 | 0.6332 | 0.3928 |
| 0.3577 | 26.0 | 1560 | 0.3152 | 0.5755 | 0.2824 |
| 0.3265 | 27.0 | 1620 | 0.2835 | 0.5856 | 0.3739 |
| 0.3272 | 28.0 | 1680 | 0.2650 | 0.5539 | 0.3515 |
| 0.2951 | 29.0 | 1740 | 0.2527 | 0.5521 | 0.3830 |
| 0.2845 | 30.0 | 1800 | 0.2473 | 0.5480 | 0.3836 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.14.6
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["wer"], "base_model": "rinna/japanese-hubert-base", "model-index": [{"name": "hubert_RTSPsplit_0208_1", "results": []}]} | automatic-speech-recognition | tndklab/hubert_RTSPsplit_0208_1 | [
"transformers",
"safetensors",
"wav2vec2",
"automatic-speech-recognition",
"generated_from_trainer",
"base_model:rinna/japanese-hubert-base",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-07T15:42:58+00:00 | [] | [] | TAGS
#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-rinna/japanese-hubert-base #license-apache-2.0 #endpoints_compatible #region-us
| hubert\_RTSPsplit\_0208\_1
==========================
This model is a fine-tuned version of rinna/japanese-hubert-base on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.2473
* Wer: 0.5480
* Cer: 0.3836
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0002
* train\_batch\_size: 32
* eval\_batch\_size: 32
* seed: 4
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 1000
* num\_epochs: 30
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.14.6
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 30",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-rinna/japanese-hubert-base #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 30",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
69,
115,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-rinna/japanese-hubert-base #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 30### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
-0.08912018686532974,
0.09964346140623093,
-0.0033935068640857935,
0.07835307717323303,
0.11100562661886215,
-0.015279743820428848,
0.15468066930770874,
0.14713366329669952,
-0.08034640550613403,
0.08372565358877182,
0.10597914457321167,
0.112840436398983,
0.032521359622478485,
0.16359074413776398,
-0.06813818961381912,
-0.2657838761806488,
0.06081047281622887,
0.02039620652794838,
0.002250277902930975,
0.11665598303079605,
0.08827716112136841,
-0.12576954066753387,
0.07531429827213287,
0.015117484144866467,
-0.13691718876361847,
0.01263792347162962,
0.017992425709962845,
-0.1121138334274292,
0.11649101972579956,
0.009819918312132359,
0.08572162687778473,
0.042586952447891235,
0.08152226358652115,
-0.22271576523780823,
0.00626436248421669,
0.039720382541418076,
0.01990896463394165,
0.062482286244630814,
0.035598497837781906,
-0.0148575184866786,
0.09793217480182648,
-0.0820889100432396,
0.05962950736284256,
0.039543960243463516,
-0.11291833221912384,
-0.2624257802963257,
-0.0785074383020401,
0.04150765389204025,
0.10016511380672455,
0.08330240100622177,
-0.02320241928100586,
0.13348306715488434,
-0.04275944083929062,
0.09376586973667145,
0.2787788212299347,
-0.3233463764190674,
-0.044031281024217606,
-0.020638158544898033,
0.04616156220436096,
0.06592807918787003,
-0.09856332838535309,
-0.010138353332877159,
0.0584123432636261,
0.015959860756993294,
0.11718123406171799,
-0.027357488870620728,
-0.06184973940253258,
-0.006013860926032066,
-0.1319805234670639,
-0.021108247339725494,
0.14503560960292816,
0.04638646915555,
-0.05910223722457886,
-0.08908167481422424,
-0.06102704629302025,
-0.15421642363071442,
-0.06252814084291458,
-0.03830194100737572,
0.05099807307124138,
-0.047159913927316666,
-0.06848492473363876,
-0.016352428123354912,
-0.06967291980981827,
-0.09103411436080933,
-0.033557165414094925,
0.22942309081554413,
0.0552532896399498,
-0.004826305899769068,
-0.026827186346054077,
0.06166818365454674,
-0.04789818450808525,
-0.15424244105815887,
-0.02791021205484867,
0.028996849432587624,
0.011138389818370342,
-0.00966371688991785,
-0.02929636649787426,
-0.04863204434514046,
0.04331513121724129,
0.15374656021595,
-0.11410336941480637,
0.08127668499946594,
-0.02524149976670742,
0.012865466997027397,
-0.10317885130643845,
0.1811438351869583,
-0.031404029577970505,
-0.023025749251246452,
0.00798695906996727,
0.08078626543283463,
0.059154145419597626,
-0.015428107231855392,
-0.09492530673742294,
0.015190327540040016,
0.10350283235311508,
0.048551611602306366,
-0.08802325278520584,
0.06695342808961868,
-0.035745538771152496,
0.006254211533814669,
0.011763399466872215,
-0.1253339946269989,
0.024158865213394165,
0.01802748441696167,
-0.06814399361610413,
-0.02335025742650032,
0.000576679187361151,
0.0065943337976932526,
-0.012281801551580429,
0.07134062051773071,
-0.06405007839202881,
0.01419908832758665,
-0.05309118330478668,
-0.1167529746890068,
0.01222726609557867,
-0.11328884214162827,
0.01277919765561819,
-0.11236412823200226,
-0.12796224653720856,
-0.005113035906106234,
0.02999112568795681,
-0.043243519961833954,
0.00454286253079772,
-0.09269628673791885,
-0.09507352858781815,
0.034083399921655655,
-0.02514142170548439,
0.019289467483758926,
-0.08216122537851334,
0.09332530200481415,
0.07954461127519608,
0.09427676349878311,
-0.021097058430314064,
0.03130123391747475,
-0.09180594980716705,
0.0278867669403553,
-0.1989380568265915,
0.04526892676949501,
-0.07207654416561127,
0.038337402045726776,
-0.10891652852296829,
-0.08423776924610138,
0.014870215207338333,
0.010740741156041622,
0.07747002691030502,
0.12251897901296616,
-0.15698596835136414,
-0.08071579039096832,
0.20823590457439423,
-0.12289770692586899,
-0.11463015526533127,
0.11493848264217377,
-0.03606532886624336,
0.048809219151735306,
0.06634943932294846,
0.25124189257621765,
0.01847176067531109,
-0.15381449460983276,
-0.008095423690974712,
-0.04920933023095131,
0.03968138247728348,
0.001616110559552908,
0.0607287771999836,
-0.012739441357553005,
0.02811850607395172,
0.026865608990192413,
-0.03980712592601776,
0.02045668661594391,
-0.07883921265602112,
-0.08341147750616074,
-0.04053812474012375,
-0.09315427392721176,
0.030368687584996223,
0.02375156804919243,
0.059031758457422256,
-0.12982885539531708,
-0.09476731717586517,
0.03272039443254471,
0.09377475082874298,
-0.10132573544979095,
0.04500724747776985,
-0.12306368350982666,
0.0839347168803215,
-0.025993771851062775,
-0.011281754821538925,
-0.14418257772922516,
0.020404363051056862,
0.04522905498743057,
-0.021721262484788895,
0.0328969843685627,
-0.07122265547513962,
0.07724108546972275,
0.06367770582437515,
-0.05901726335287094,
-0.056566204875707626,
-0.019826684147119522,
0.018282949924468994,
-0.059989430010318756,
-0.20031000673770905,
-0.009335056878626347,
-0.04389464482665062,
0.07703087478876114,
-0.15596599876880646,
0.029746633023023605,
0.04023467376828194,
0.10096535086631775,
0.048908453434705734,
-0.01638132520020008,
-0.0018075513653457165,
0.06670243293046951,
-0.030997952446341515,
-0.0540691614151001,
0.039799146354198456,
0.001715353224426508,
-0.08455681800842285,
0.02307084947824478,
-0.16703376173973083,
0.15225955843925476,
0.1418287456035614,
0.0017896622885018587,
-0.07263664901256561,
0.0028941056225448847,
-0.029150158166885376,
-0.02423904836177826,
-0.03525667265057564,
0.019417181611061096,
0.15081816911697388,
-0.016615502536296844,
0.1420743763446808,
-0.10585950314998627,
-0.011506387032568455,
0.051873765885829926,
-0.04822991415858269,
-0.010805563069880009,
0.10107675939798355,
0.03966870903968811,
-0.07689835131168365,
0.1264999806880951,
0.11066096276044846,
-0.0918438509106636,
0.14632411301136017,
-0.0564676895737648,
-0.06343868374824524,
-0.03146776184439659,
0.027518274262547493,
0.01861860230565071,
0.11867987364530563,
-0.12944872677326202,
-0.02651400677859783,
0.009620890952646732,
0.018760548904538155,
0.004128232132643461,
-0.19668681919574738,
-0.0061518908478319645,
0.03190898150205612,
-0.09746778756380081,
-0.014893271960318089,
0.011471518315374851,
-0.006778149865567684,
0.10255691409111023,
-0.00871029868721962,
-0.0976935476064682,
0.01161695271730423,
-0.004645580891519785,
-0.06973129510879517,
0.1748303472995758,
-0.10449196398258209,
-0.1663205772638321,
-0.10845885425806046,
-0.08649501949548721,
-0.05013582110404968,
0.02632283605635166,
0.08427523821592331,
-0.12000507116317749,
-0.04905467852950096,
-0.11196696013212204,
0.0057428618893027306,
0.025925999507308006,
0.038988009095191956,
0.027794448658823967,
0.007775092497467995,
0.06934013962745667,
-0.11404287815093994,
-0.018636345863342285,
-0.04373935982584953,
-0.010490445420145988,
0.03180612996220589,
0.02848033793270588,
0.11796153336763382,
0.13468864560127258,
-0.016655970364809036,
0.034325115382671356,
-0.04449719190597534,
0.19675038754940033,
-0.08047930896282196,
-0.02462032251060009,
0.12970784306526184,
-0.008184609934687614,
0.03704259544610977,
0.15379725396633148,
0.040015216916799545,
-0.11487757414579391,
0.006666544359177351,
0.011877363547682762,
-0.0357089601457119,
-0.21439625322818756,
-0.02932380512356758,
-0.03359195962548256,
0.005390993319451809,
0.08463198691606522,
0.03736631199717522,
0.04084223881363869,
0.03183227404952049,
0.027313100174069405,
0.0051808226853609085,
0.02191844955086708,
0.09415916353464127,
0.12405520677566528,
0.0451899990439415,
0.11587433516979218,
-0.04511338472366333,
-0.051133204251527786,
0.0276540145277977,
0.0072463322430849075,
0.2053804248571396,
0.025843985378742218,
0.15668009221553802,
0.05003946274518967,
0.1563895344734192,
0.03252720460295677,
0.051337551325559616,
0.002665830310434103,
-0.02288062684237957,
-0.006300002336502075,
-0.06620001047849655,
-0.02409963496029377,
0.053521621972322464,
-0.04379360005259514,
0.04318540543317795,
-0.11268191784620285,
0.02577602118253708,
0.05609183385968208,
0.29066526889801025,
0.043272946029901505,
-0.32295989990234375,
-0.07795935869216919,
0.010583191178739071,
-0.06648868322372437,
-0.015318154357373714,
0.06337698549032211,
0.12585876882076263,
-0.05690919607877731,
0.07865207642316818,
-0.046772122383117676,
0.07683679461479187,
-0.041404757648706436,
0.03420745208859444,
0.022992217913269997,
0.07524116337299347,
-0.00427002040669322,
0.02780703268945217,
-0.2634204626083374,
0.29457390308380127,
0.019922442734241486,
0.09628362953662872,
-0.031344447284936905,
0.004283673129975796,
0.02836587280035019,
0.029910098761320114,
0.11662521958351135,
-0.03828331455588341,
-0.12355112284421921,
-0.17329177260398865,
-0.06292030960321426,
0.025872644037008286,
0.1389864832162857,
0.022174960002303123,
0.11336249113082886,
-0.0195149052888155,
-0.024237891659140587,
0.053824808448553085,
-0.07247097045183182,
-0.10646068304777145,
-0.07231006026268005,
-0.025004234164953232,
0.0958649143576622,
0.025717327371239662,
-0.06496690958738327,
-0.08355839550495148,
-0.0946296900510788,
0.10929425805807114,
-0.03907964006066322,
-0.027309024706482887,
-0.10291173309087753,
-0.004278006963431835,
0.10317673534154892,
-0.07827393710613251,
0.060537464916706085,
0.007310779765248299,
0.09741833060979843,
0.01342062558978796,
-0.04124406352639198,
0.11099854111671448,
-0.06895864754915237,
-0.17847003042697906,
-0.0466155931353569,
0.14732380211353302,
0.01635643094778061,
0.044900551438331604,
0.00810579676181078,
0.025302376598119736,
0.01816675253212452,
-0.06726399064064026,
0.03476160764694214,
0.02011297084391117,
0.028187142685055733,
0.0041954051703214645,
-0.052377235144376755,
-0.04076066613197327,
-0.10276062786579132,
-0.03240993618965149,
0.15117909014225006,
0.3037933111190796,
-0.07748232781887054,
0.070234015583992,
0.09224281460046768,
-0.047612518072128296,
-0.1656421422958374,
-0.005619312170892954,
0.036002159118652344,
0.014383256435394287,
-0.010436286218464375,
-0.15419621765613556,
0.046181730926036835,
0.07381287217140198,
-0.02919192425906658,
0.0708683505654335,
-0.27358463406562805,
-0.14290927350521088,
0.12463534623384476,
0.13529707491397858,
0.10407286882400513,
-0.1454513669013977,
-0.05143134668469429,
-0.023296743631362915,
-0.10138633847236633,
0.06918904185295105,
-0.09269919991493225,
0.12883687019348145,
-0.013757997192442417,
0.04540868476033211,
0.014088551513850689,
-0.04440739005804062,
0.12173093855381012,
0.002411806955933571,
0.08555954694747925,
-0.047218143939971924,
-0.0022191060706973076,
0.03175235912203789,
-0.06120797619223595,
0.07283610850572586,
-0.10025770217180252,
0.034920692443847656,
-0.034273356199264526,
-0.03991882503032684,
-0.06648141145706177,
0.018959766253829002,
-0.006404121406376362,
-0.05134222283959389,
-0.05114416033029556,
0.01273806020617485,
0.05192266032099724,
-0.007533463649451733,
0.16948957741260529,
-0.022833697497844696,
0.13146981596946716,
0.1635444164276123,
0.09421101212501526,
-0.08479874581098557,
-0.011607126332819462,
0.01780427247285843,
-0.042419660836458206,
0.07431860268115997,
-0.1379013955593109,
0.05137589946389198,
0.10129624605178833,
0.020455792546272278,
0.15008533000946045,
0.054042112082242966,
-0.05914497375488281,
0.040728650987148285,
0.06646960973739624,
-0.14310042560100555,
-0.11547711491584778,
0.008337637409567833,
-0.01968252658843994,
-0.05697616562247276,
0.08401014655828476,
0.13196998834609985,
-0.07225625962018967,
-0.0014596482506021857,
-0.01682218536734581,
0.017138591036200523,
-0.03902307152748108,
0.18542931973934174,
0.045848164707422256,
0.04783320054411888,
-0.10451627522706985,
0.09496007859706879,
0.028906673192977905,
-0.10621827840805054,
0.06291942298412323,
0.07672789692878723,
-0.09825684130191803,
-0.023666637018322945,
0.024144211784005165,
0.13703279197216034,
-0.010170121677219868,
-0.08676575124263763,
-0.1679791808128357,
-0.12817218899726868,
0.06129283830523491,
0.21752998232841492,
0.07130715996026993,
0.011876986362040043,
-0.02574988454580307,
0.02547907643020153,
-0.11562757939100266,
0.08776987344026566,
0.05001629889011383,
0.06047176942229271,
-0.14679698646068573,
0.1045747920870781,
0.022807450965046883,
0.01375540904700756,
-0.025732168927788734,
0.009379394352436066,
-0.12458935379981995,
0.02587403543293476,
-0.12647724151611328,
0.018656214699149132,
-0.05436423793435097,
-0.0032726628705859184,
0.004011825658380985,
-0.06100153923034668,
-0.07599790394306183,
0.03080832026898861,
-0.10227479040622711,
-0.00918617658317089,
0.01960347220301628,
0.0496690571308136,
-0.1308867484331131,
-0.018312053754925728,
0.024379126727581024,
-0.08767925947904587,
0.0752088874578476,
0.0700036957859993,
-0.032784465700387955,
0.0681251510977745,
-0.11687609553337097,
-0.007625266443938017,
0.09609970450401306,
-0.006737787742167711,
0.03554397076368332,
-0.13156795501708984,
-0.012449695728719234,
0.017279312014579773,
0.048318203538656235,
0.014786184765398502,
0.102851502597332,
-0.10904105752706528,
0.013956419192254543,
-0.04981379956007004,
-0.06005890667438507,
-0.05802452936768532,
0.02788948267698288,
0.1295296549797058,
-0.002361494116485119,
0.17323444783687592,
-0.1103353425860405,
0.014852580614387989,
-0.16532669961452484,
0.0058330753818154335,
-0.015257111750543118,
-0.12049441784620285,
-0.11701415479183197,
-0.02568538300693035,
0.06376186013221741,
-0.06268526613712311,
0.1265282928943634,
-0.023232687264680862,
0.02223757840692997,
0.038805317133665085,
-0.10014605522155762,
-0.021400626748800278,
0.03573395684361458,
0.2264985293149948,
0.041348110884428024,
-0.0440019853413105,
0.046417661011219025,
0.007157388608902693,
0.09804989397525787,
0.09378113597631454,
0.15181879699230194,
0.18867212533950806,
0.021977948024868965,
0.13953571021556854,
0.06767616420984268,
-0.041300032287836075,
-0.14389070868492126,
0.08107193559408188,
-0.05547480285167694,
0.09460028260946274,
0.0044390675611793995,
0.2137097269296646,
0.15666906535625458,
-0.14700697362422943,
0.046249937266111374,
-0.01765417493879795,
-0.07850541919469833,
-0.1335713118314743,
-0.053655095398426056,
-0.11474397778511047,
-0.18957309424877167,
0.0324515663087368,
-0.10792908072471619,
0.05648438632488251,
0.05016424506902695,
0.026370149105787277,
0.009179352782666683,
0.16724050045013428,
0.005614992696791887,
0.0114522073417902,
0.08308503776788712,
0.0042835115455091,
-0.06403017789125443,
-0.04795369878411293,
-0.09161609411239624,
0.03057740069925785,
-0.02804827131330967,
0.024068288505077362,
-0.0075623043812811375,
-0.05540362000465393,
0.04243578389286995,
-0.03002159111201763,
-0.09596681594848633,
0.023412218317389488,
0.02860504388809204,
0.07253047078847885,
0.04878924787044525,
0.03745583817362785,
-0.03804732486605644,
0.009320802055299282,
0.21639949083328247,
-0.08463491499423981,
-0.08662907034158707,
-0.11606244742870331,
0.23331701755523682,
0.05115870013833046,
0.021331438794732094,
0.0018914865795522928,
-0.07684393227100372,
-0.02420777827501297,
0.20735618472099304,
0.1724487841129303,
-0.0226358063519001,
0.0044680992141366005,
-0.041857391595840454,
-0.0036224094219505787,
-0.043105605989694595,
0.07081764191389084,
0.11944852024316788,
0.056730855256319046,
-0.03738010302186012,
-0.045814476907253265,
-0.05332313850522041,
-0.028233719989657402,
-0.04408707097172737,
0.0774768590927124,
0.0032890462316572666,
-0.02799096703529358,
-0.05501900985836983,
0.06032363325357437,
-0.07635039836168289,
-0.10230889916419983,
0.033559802919626236,
-0.19802463054656982,
-0.13854782283306122,
0.006431593559682369,
0.07922535389661789,
0.026432812213897705,
0.036626797169446945,
-0.014948315918445587,
-0.006577673833817244,
0.07152695208787918,
-0.00012433488154783845,
-0.0750819519162178,
-0.0926833301782608,
0.06251803785562515,
-0.12030389904975891,
0.2101035714149475,
-0.019557539373636246,
0.05155015364289284,
0.10681559145450592,
0.0570773184299469,
-0.07263312488794327,
0.1232929453253746,
0.04924149438738823,
-0.10510767996311188,
0.019101129844784737,
0.1394195705652237,
-0.041745152324438095,
0.1420210897922516,
0.04857036471366882,
-0.12613841891288757,
0.012032254599034786,
-0.025163033977150917,
-0.07660960406064987,
-0.0679321363568306,
-0.03400052338838577,
-0.06240237504243851,
0.11890311539173126,
0.16516010463237762,
-0.044023171067237854,
0.010357272811233997,
-0.03897186368703842,
0.06039748340845108,
0.06838653236627579,
0.0023293416015803814,
-0.02905210852622986,
-0.27447354793548584,
0.008695272728800774,
0.0491984486579895,
-0.010766981169581413,
-0.26495885848999023,
-0.0972658321261406,
-0.004903189372271299,
-0.03216979652643204,
-0.09404901415109634,
0.06962807476520538,
0.10307527333498001,
0.05003926157951355,
-0.04834010824561119,
-0.10341013967990875,
-0.022035008296370506,
0.18145865201950073,
-0.1551308035850525,
-0.09028976410627365
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | m-newhauser/mistral_7b_guanaco_test | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-07T15:44:10+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
31,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06646376848220825,
0.2168014943599701,
-0.00225935154594481,
0.023818302899599075,
0.1271018385887146,
-0.001635765191167593,
0.04218708351254463,
0.13324736058712006,
-0.020175931975245476,
0.11144465953111649,
0.046588581055402756,
0.09377603232860565,
0.09928803145885468,
0.18404334783554077,
0.04859916493296623,
-0.2059975117444992,
0.007056170143187046,
-0.09090408682823181,
0.014076028019189835,
0.1116579994559288,
0.13719257712364197,
-0.10291384905576706,
0.08272874355316162,
-0.04045208916068077,
-0.02019004337489605,
0.00012576708104461432,
-0.09259183704853058,
-0.07032395154237747,
0.06885425746440887,
0.06264153122901917,
0.051234472543001175,
0.001456156256608665,
0.09140396863222122,
-0.2864592671394348,
0.017265573143959045,
0.08406311273574829,
0.0027674848679453135,
0.06290827691555023,
0.07236549258232117,
-0.07389893382787704,
0.11328595131635666,
-0.08021481335163116,
0.13019037246704102,
0.08625296503305435,
-0.062064990401268005,
-0.23071379959583282,
-0.07525765895843506,
0.0963398814201355,
0.12251301854848862,
0.06215599179267883,
-0.022921854630112648,
0.15455181896686554,
-0.06248689442873001,
0.012971068732440472,
0.1294165402650833,
-0.11526761949062347,
-0.05572471022605896,
0.061741601675748825,
0.11775490641593933,
0.10740239918231964,
-0.14110268652439117,
-0.0017287094378843904,
0.04900608956813812,
0.029121357947587967,
0.08589313924312592,
0.022661056369543076,
0.12003941088914871,
0.04652795568108559,
-0.13695219159126282,
-0.04037507623434067,
0.12011898308992386,
0.038862764835357666,
-0.06446044892072678,
-0.2168138176202774,
-0.006778308190405369,
-0.0601806715130806,
-0.014732478186488152,
-0.07019448280334473,
0.039128515869379044,
-0.02470310963690281,
0.07317749410867691,
-0.04465159401297569,
-0.1063927412033081,
-0.0421026237308979,
0.0892222449183464,
0.07748593389987946,
0.011527054943144321,
-0.02519804798066616,
0.04627908393740654,
0.13455867767333984,
0.05402068421244621,
-0.10399353504180908,
-0.07017925381660461,
-0.06942764669656754,
-0.09420394152402878,
-0.04035796597599983,
0.056760527193546295,
0.031942449510097504,
0.02665667235851288,
0.22703726589679718,
0.016653569415211678,
0.04155244305729866,
0.0224777739495039,
0.01032855175435543,
0.043662428855895996,
0.0955500528216362,
-0.05303520709276199,
-0.15660029649734497,
-0.04072032496333122,
0.09077946096658707,
-0.0027527001220732927,
-0.036689214408397675,
-0.03966725245118141,
0.03849169611930847,
0.06843466311693192,
0.13122352957725525,
0.07552056759595871,
-0.017929591238498688,
-0.04813180863857269,
-0.030096933245658875,
0.23523783683776855,
-0.1493375599384308,
0.04426715523004532,
-0.02271856553852558,
-0.01804111897945404,
-0.03908449783921242,
0.03597262129187584,
0.022118929773569107,
-0.000004518366949923802,
0.09706240892410278,
-0.058981191366910934,
-0.05378659814596176,
-0.10168042778968811,
-0.03272576630115509,
0.04088849574327469,
-0.013975566253066063,
-0.010589460842311382,
-0.09025166928768158,
-0.09490354359149933,
-0.04766594246029854,
0.05537205561995506,
-0.05123869329690933,
-0.03770573064684868,
0.009465423412621021,
-0.08151785284280777,
-0.005444355774670839,
-0.005417742300778627,
0.10699385404586792,
-0.03222226724028587,
0.04445803165435791,
-0.027600755915045738,
0.05225523188710213,
0.09919606149196625,
0.031576547771692276,
-0.0773419588804245,
0.0561848059296608,
-0.22559374570846558,
0.07503069192171097,
-0.11481974273920059,
0.04335082694888115,
-0.1704932004213333,
-0.042439818382263184,
0.005444696638733149,
0.0139949731528759,
0.013206101022660732,
0.12720820307731628,
-0.19255615770816803,
-0.01654396951198578,
0.13260798156261444,
-0.09212633967399597,
-0.118110790848732,
0.07884611934423447,
-0.029701577499508858,
0.1624738723039627,
0.04682036489248276,
-0.027025915682315826,
0.09224298596382141,
-0.16434773802757263,
-0.07092688232660294,
-0.00949116237461567,
-0.01727987825870514,
0.12109188735485077,
0.07512219995260239,
-0.05991523340344429,
0.046571120619773865,
0.02832140028476715,
-0.038078423589468,
-0.04424772411584854,
-0.050857074558734894,
-0.10884185880422592,
-0.01070026308298111,
-0.08987759798765182,
0.04065500199794769,
-0.01250192429870367,
-0.07916021347045898,
-0.029885273426771164,
-0.18612512946128845,
-0.0030564051121473312,
0.10038342326879501,
0.0035033065360039473,
-0.005652366206049919,
-0.08666291832923889,
0.026358824223279953,
-0.03112892620265484,
-0.008404186926782131,
-0.16764774918556213,
-0.04399421438574791,
0.046902090311050415,
-0.16094985604286194,
0.020117372274398804,
-0.06413903087377548,
0.06334125250577927,
0.03641495108604431,
-0.05590536445379257,
-0.0248766727745533,
-0.01730942726135254,
0.011945613659918308,
-0.05083848536014557,
-0.18994836509227753,
-0.056277405470609665,
-0.037882111966609955,
0.149809330701828,
-0.25956398248672485,
0.032966937869787216,
0.051140617579221725,
0.14649195969104767,
0.00406361510977149,
-0.05115427449345589,
0.01429014839231968,
-0.05360214412212372,
-0.054652128368616104,
-0.06746816635131836,
-0.006135428790003061,
-0.027576493099331856,
-0.05147203803062439,
0.019243421033024788,
-0.1755700707435608,
-0.021410830318927765,
0.09424154460430145,
0.12876708805561066,
-0.1486445665359497,
-0.018640631809830666,
-0.048725154250860214,
-0.06339836865663528,
-0.0715010017156601,
-0.07038594037294388,
0.10712739825248718,
0.0513901449739933,
0.04796046018600464,
-0.07435787469148636,
-0.07092321664094925,
0.02726263552904129,
0.006906150374561548,
-0.03382374346256256,
0.08727246522903442,
0.05199531093239784,
-0.09209315478801727,
0.0756213590502739,
0.1092359870672226,
0.07177663594484329,
0.09363535046577454,
0.01574566215276718,
-0.11756632477045059,
-0.028492970392107964,
0.036266472190618515,
0.02740776725113392,
0.1465986967086792,
-0.05952361226081848,
0.04016614332795143,
0.04494241625070572,
-0.04170418903231621,
0.022319864481687546,
-0.08787637203931808,
0.024075502529740334,
0.025203049182891846,
-0.0034381982404738665,
0.06284574419260025,
-0.02525499276816845,
-0.0050758360885083675,
0.07016654312610626,
0.047779910266399384,
0.04621000960469246,
0.009655474685132504,
-0.01720241829752922,
-0.1047825813293457,
0.16950392723083496,
-0.0951867327094078,
-0.269941508769989,
-0.17632324993610382,
0.026197833940386772,
0.04035249724984169,
-0.022378476336598396,
0.031619444489479065,
-0.07056326419115067,
-0.10630585998296738,
-0.1060405746102333,
-0.002429972169920802,
0.01714223250746727,
-0.06364088505506516,
-0.0741225928068161,
0.07348573952913284,
0.04382912442088127,
-0.14902326464653015,
0.038552410900592804,
0.055694397538900375,
-0.057955220341682434,
-0.0233661737293005,
0.09118817001581192,
0.12397737801074982,
0.14583967626094818,
-0.021366750821471214,
-0.028626007959246635,
0.029004426673054695,
0.19620531797409058,
-0.13469526171684265,
0.10371150821447372,
0.13814030587673187,
-0.04545360431075096,
0.08360563963651657,
0.1560150384902954,
0.029186224564909935,
-0.08317049592733383,
0.05044832453131676,
0.04082648828625679,
-0.043159641325473785,
-0.2666129767894745,
-0.0534592866897583,
0.012832709588110447,
-0.06255637854337692,
0.09786593168973923,
0.10183793306350708,
0.11542957276105881,
0.034910861402750015,
-0.07166364789009094,
-0.043925940990448,
-0.0058974819257855415,
0.11737963557243347,
-0.05490213260054588,
-0.012639665976166725,
0.07686592638492584,
-0.05086168646812439,
0.005355054512619972,
0.10266812145709991,
0.02973790094256401,
0.17442677915096283,
0.020399179309606552,
0.11231429129838943,
0.06195578724145889,
0.08633565157651901,
0.0007386076031252742,
0.02951662428677082,
0.05147615820169449,
0.017203815281391144,
-0.002300140680745244,
-0.10421168059110641,
-0.006156572140753269,
0.1449710875749588,
0.028103826567530632,
0.029669636860489845,
-0.0018948549404740334,
-0.005003341939300299,
0.05121048167347908,
0.1746254414319992,
-0.011592294089496136,
-0.22072425484657288,
-0.0845772922039032,
0.06936841458082199,
-0.06218599155545235,
-0.12968985736370087,
-0.026130788028240204,
0.045467354357242584,
-0.17519839107990265,
0.026703642681241035,
-0.027433741837739944,
0.0919293761253357,
-0.09345759451389313,
-0.02221956104040146,
0.03687324374914169,
0.084866963326931,
-0.014529162086546421,
0.08703910559415817,
-0.14498743414878845,
0.11886418610811234,
0.02978132851421833,
0.09024628251791,
-0.11081171780824661,
0.07909037172794342,
-0.007550720125436783,
0.009180475026369095,
0.19379350543022156,
-0.011335089802742004,
-0.03514958545565605,
-0.08774717897176743,
-0.11210042238235474,
-0.013537433929741383,
0.12687496840953827,
-0.1243172138929367,
0.08773399889469147,
-0.015198243781924248,
-0.044079482555389404,
0.00937260314822197,
-0.12100647389888763,
-0.17273177206516266,
-0.19628387689590454,
0.05585884302854538,
-0.09575839340686798,
0.025643249973654747,
-0.11914430558681488,
-0.07089093327522278,
-0.02952558360993862,
0.241120383143425,
-0.1745356321334839,
-0.06510113179683685,
-0.1468164622783661,
-0.046294767409563065,
0.1662203073501587,
-0.04437198117375374,
0.0718095526099205,
-0.0208172257989645,
0.20345525443553925,
0.005988610442727804,
-0.004939318168908358,
0.06724198162555695,
-0.08892562240362167,
-0.16873881220817566,
-0.06771010160446167,
0.1510489284992218,
0.11680185794830322,
0.04907919466495514,
-0.002248800592496991,
0.0011772146681323647,
-0.016943959519267082,
-0.1137804463505745,
-0.0033210667315870523,
0.16037839651107788,
0.03878779336810112,
0.025986969470977783,
-0.05243593826889992,
-0.08797456324100494,
-0.06899320334196091,
-0.06853509694337845,
0.06221301481127739,
0.19590823352336884,
-0.10376439243555069,
0.1700313836336136,
0.147536963224411,
-0.07305635511875153,
-0.23175598680973053,
0.035342130810022354,
0.04983805492520332,
0.0014306638622656465,
0.04886869341135025,
-0.18252557516098022,
0.10521943867206573,
0.019543392583727837,
-0.05505957826972008,
0.13485197722911835,
-0.1557481735944748,
-0.1552847921848297,
0.0722852572798729,
0.03904085233807564,
-0.22423844039440155,
-0.1354004591703415,
-0.09622503817081451,
-0.05825018882751465,
-0.14065024256706238,
0.06054598465561867,
-0.002136280992999673,
0.015948504209518433,
0.03500790148973465,
-0.0015643214574083686,
0.027123261243104935,
-0.058935679495334625,
0.18609118461608887,
-0.004065449349582195,
0.020676052197813988,
-0.060264769941568375,
-0.0478842556476593,
0.09839435666799545,
-0.06130504235625267,
0.12208222597837448,
0.004057085141539574,
0.01594383642077446,
-0.10362856835126877,
-0.048314861953258514,
-0.04328322783112526,
0.05154227837920189,
-0.07548051327466965,
-0.10070807486772537,
-0.043625857681035995,
0.08841723203659058,
0.07005169242620468,
-0.03383097052574158,
0.00549331633374095,
-0.07189501076936722,
0.10019614547491074,
0.17795267701148987,
0.17573626339435577,
0.009926567785441875,
-0.07241068035364151,
0.01677953451871872,
-0.04142116755247116,
0.044231921434402466,
-0.2513144314289093,
0.03756171092391014,
0.06098250672221184,
0.029438555240631104,
0.09217222779989243,
-0.020435843616724014,
-0.1820858269929886,
-0.04050002992153168,
0.08094815909862518,
-0.05452597141265869,
-0.22617179155349731,
-0.019085140898823738,
0.0954197570681572,
-0.2020406424999237,
-0.007372708059847355,
0.03995226323604584,
-0.048725228756666183,
-0.023169852793216705,
0.00010950004070764408,
0.06317184865474701,
0.002471912419423461,
0.09773622453212738,
0.0735151618719101,
0.09715340286493301,
-0.08337292820215225,
0.10562895983457565,
0.10150538384914398,
-0.09572599828243256,
0.03605884686112404,
0.06754924356937408,
-0.05300498008728027,
-0.043293699622154236,
0.03665391728281975,
0.033023297786712646,
0.005234600510448217,
-0.060321882367134094,
0.013913018628954887,
-0.036497246474027634,
0.044923391193151474,
0.08326134830713272,
0.03754979372024536,
-0.013354414142668247,
0.06462216377258301,
0.03401726484298706,
-0.10898099094629288,
0.10366570204496384,
0.01731540448963642,
0.04105307161808014,
-0.08384523540735245,
-0.019968897104263306,
0.035425446927547455,
0.030576206743717194,
-0.01765924133360386,
-0.02306121215224266,
-0.02860277332365513,
-0.01614218018949032,
-0.14299540221691132,
-0.023106401786208153,
-0.07243485748767853,
0.006181265693157911,
0.014656842686235905,
-0.031884219497442245,
-0.011233693920075893,
0.02475680410861969,
-0.06979699432849884,
-0.07426341623067856,
-0.006949664559215307,
0.09833318740129471,
-0.15115703642368317,
0.008848577737808228,
0.06907843053340912,
-0.11088496446609497,
0.08190931379795074,
-0.008411259390413761,
0.016245156526565552,
0.022527478635311127,
-0.15448406338691711,
0.05601610988378525,
0.0008648968650959432,
0.01916889287531376,
0.025886621326208115,
-0.16471809148788452,
0.004104440100491047,
-0.04661374166607857,
-0.02149827405810356,
-0.00004464812809601426,
-0.02647159807384014,
-0.12325995415449142,
0.06858719140291214,
-0.015622655861079693,
-0.035931166261434555,
-0.02701525390148163,
0.0539589487016201,
0.07888586074113846,
-0.027474910020828247,
0.10445091128349304,
-0.008690856397151947,
0.04941811040043831,
-0.16801609098911285,
-0.02470702864229679,
-0.04982255399227142,
0.019377702847123146,
0.009884213097393513,
-0.007693959400057793,
0.04183054715394974,
-0.00976533442735672,
0.21883612871170044,
-0.05075952783226967,
0.1607085019350052,
0.05847611650824547,
-0.017352959141135216,
-0.0007513365126214921,
0.06180921941995621,
0.05997028574347496,
0.04658793285489082,
0.009480604901909828,
0.023740366101264954,
-0.022450892254710197,
-0.006695089396089315,
-0.15932634472846985,
0.01890849508345127,
0.14999441802501678,
0.06301083415746689,
0.024745315313339233,
0.05866100639104843,
-0.12775006890296936,
-0.12135478109121323,
0.09311001747846603,
-0.026755332946777344,
0.00928465835750103,
-0.08245618641376495,
0.1358020007610321,
0.14980104565620422,
-0.14000412821769714,
0.05256148427724838,
-0.06134212389588356,
-0.05217423290014267,
-0.10388828068971634,
-0.12032219022512436,
-0.05887215584516525,
-0.053666237741708755,
0.002330566756427288,
-0.03760887682437897,
0.054546963423490524,
0.03344334661960602,
-0.009351172484457493,
-0.00022941511997487396,
0.13597318530082703,
-0.019751882180571556,
-0.0028988157864660025,
0.048313532024621964,
0.03693558648228645,
0.02373051457107067,
-0.05275435373187065,
0.02940409444272518,
0.02539868652820587,
0.032232340425252914,
0.06546790152788162,
0.033412106335163116,
-0.047448933124542236,
0.03804153576493263,
-0.0025254099164158106,
-0.11207924783229828,
0.019641218706965446,
-0.00460948096588254,
-0.0742158442735672,
0.1268945336341858,
0.0407399944961071,
0.010224059224128723,
-0.03741471841931343,
0.24361543357372284,
-0.06653323769569397,
-0.06378097087144852,
-0.13251738250255585,
0.10491154342889786,
-0.0027236645109951496,
0.06476365029811859,
0.023412218317389488,
-0.1284150779247284,
0.005243356805294752,
0.13858191668987274,
0.12181595712900162,
0.0045748427510261536,
0.009228081442415714,
0.0518609918653965,
0.0025186820421367884,
-0.06998204439878464,
0.054019294679164886,
0.06992026418447495,
0.12919506430625916,
-0.07847554981708527,
0.07680778950452805,
0.0006860480643808842,
-0.08370215445756912,
-0.02947772853076458,
0.11312682181596756,
-0.0409729965031147,
0.03491825982928276,
-0.047444481402635574,
0.10916327685117722,
-0.05787910893559456,
-0.29412412643432617,
0.02350960113108158,
-0.09588567912578583,
-0.15202060341835022,
-0.018367812037467957,
0.05944539234042168,
-0.02624768204987049,
0.018029648810625076,
0.06971040368080139,
-0.06011629104614258,
0.20098382234573364,
0.0335683599114418,
-0.07864278554916382,
-0.0664360448718071,
0.04837050288915634,
-0.06564252078533173,
0.2949807047843933,
0.008418165147304535,
0.02863333560526371,
0.10770907253026962,
-0.03253700211644173,
-0.18271861970424652,
0.010723991319537163,
0.1133992001414299,
-0.08056149631738663,
0.08200647681951523,
0.19000613689422607,
-0.012578671798110008,
0.1209007054567337,
0.05294662341475487,
-0.047376248985528946,
0.04217283055186272,
-0.03389401361346245,
-0.051268599927425385,
-0.10752558708190918,
0.058453381061553955,
-0.05909625440835953,
0.15447644889354706,
0.10152646154165268,
-0.05671518296003342,
-0.004550917539745569,
-0.05555408447980881,
0.04875178262591362,
0.01804669201374054,
0.12263146042823792,
0.02951994352042675,
-0.1865430772304535,
0.032826557755470276,
-0.01144319772720337,
0.10186848044395447,
-0.25588861107826233,
-0.08421015739440918,
0.08833149075508118,
-0.011924264021217823,
-0.05105875805020332,
0.10560628771781921,
0.057650718837976456,
0.04243382066488266,
-0.043439045548439026,
-0.10480839014053345,
-0.02186836116015911,
0.14663739502429962,
-0.1469624787569046,
-0.025013303384184837
] |
null | null | transformers | # Lao to English Translation Model
Welcome to the forefront of linguistic innovation with our groundbreaking T5 language model designed specifically for Lao to English translation. In a rapidly globalizing world where effective communication is paramount, our T5 model stands as a beacon of excellence, offering unparalleled accuracy, fluency, and efficiency in bridging the language gap between Lao and English.
Built on state-of-the-art deep learning architecture and trained on vast datasets of Lao and English texts, our language model (LLM) harnesses the power of transformer-based technology to deliver seamless and precise translations. Whether you're a business expanding into Laotian markets, a researcher seeking to access Lao-language resources, or an individual connecting with Lao-speaking communities, our T5 model is your ultimate solution for unlocking linguistic barriers and fostering meaningful cross-cultural exchanges.
With a commitment to quality and innovation, our translation model not only translates words but also preserves context, tone, and cultural nuances, ensuring that the essence of the original message remains intact in every translated sentence. Whether it's documents, websites, or multimedia content, our LLM model offers unmatched versatility and reliability, empowering users to communicate effortlessly across languages and borders.
## How to use
### On GPU
```python
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
tokenizer = AutoTokenizer.from_pretrained("minhtoan/t5-translate-lao-english")
model = AutoModelForSeq2SeqLM.from_pretrained("minhtoan/t5-translate-lao-english")
model.cuda()
src = "ຂ້ອຍຮັກເຈົ້າ"
tokenized_text = tokenizer.encode(src, return_tensors="pt").cuda()
model.eval()
translate_ids = model.generate(tokenized_text, max_length=140)
output = tokenizer.decode(translate_ids[0], skip_special_tokens=True)
output
```
'I love you'
### On CPU
```python
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
tokenizer = AutoTokenizer.from_pretrained("minhtoan/t5-translate-lao-english")
model = AutoModelForSeq2SeqLM.from_pretrained("minhtoan/t5-translate-lao-english")
src = "ຂ້ອຍຮັກເຈົ້າ"
input_ids = tokenizer(src, max_length=200, return_tensors="pt", padding="max_length", truncation=True).input_ids
outputs = model.generate(input_ids=input_ids, max_new_tokens=140)
output = tokenizer.batch_decode(outputs, skip_special_tokens=True)[0]
output
```
'I love you'
## Author
`
Phan Minh Toan
` | {"language": ["en", "lo"], "license": "mit", "library_name": "transformers", "tags": ["translation"], "widget": [{"text": "\u0e82\u0ec9\u0ead\u0e8d\u0ea2\u0eb2\u0e81\u0e8a\u0eb7\u0ec9\u0e9b\u0eb6\u0ec9\u0ea1"}], "inference": {"parameters": {"max_length": 140}}, "pipeline_tag": "translation"} | translation | minhtoan/t5-translate-lao-english | [
"transformers",
"pytorch",
"mt5",
"text2text-generation",
"translation",
"en",
"lo",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T15:45:49+00:00 | [] | [
"en",
"lo"
] | TAGS
#transformers #pytorch #mt5 #text2text-generation #translation #en #lo #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # Lao to English Translation Model
Welcome to the forefront of linguistic innovation with our groundbreaking T5 language model designed specifically for Lao to English translation. In a rapidly globalizing world where effective communication is paramount, our T5 model stands as a beacon of excellence, offering unparalleled accuracy, fluency, and efficiency in bridging the language gap between Lao and English.
Built on state-of-the-art deep learning architecture and trained on vast datasets of Lao and English texts, our language model (LLM) harnesses the power of transformer-based technology to deliver seamless and precise translations. Whether you're a business expanding into Laotian markets, a researcher seeking to access Lao-language resources, or an individual connecting with Lao-speaking communities, our T5 model is your ultimate solution for unlocking linguistic barriers and fostering meaningful cross-cultural exchanges.
With a commitment to quality and innovation, our translation model not only translates words but also preserves context, tone, and cultural nuances, ensuring that the essence of the original message remains intact in every translated sentence. Whether it's documents, websites, or multimedia content, our LLM model offers unmatched versatility and reliability, empowering users to communicate effortlessly across languages and borders.
## How to use
### On GPU
'I love you'
### On CPU
'I love you'
## Author
'
Phan Minh Toan
' | [
"# Lao to English Translation Model\nWelcome to the forefront of linguistic innovation with our groundbreaking T5 language model designed specifically for Lao to English translation. In a rapidly globalizing world where effective communication is paramount, our T5 model stands as a beacon of excellence, offering unparalleled accuracy, fluency, and efficiency in bridging the language gap between Lao and English.\n\nBuilt on state-of-the-art deep learning architecture and trained on vast datasets of Lao and English texts, our language model (LLM) harnesses the power of transformer-based technology to deliver seamless and precise translations. Whether you're a business expanding into Laotian markets, a researcher seeking to access Lao-language resources, or an individual connecting with Lao-speaking communities, our T5 model is your ultimate solution for unlocking linguistic barriers and fostering meaningful cross-cultural exchanges.\n\nWith a commitment to quality and innovation, our translation model not only translates words but also preserves context, tone, and cultural nuances, ensuring that the essence of the original message remains intact in every translated sentence. Whether it's documents, websites, or multimedia content, our LLM model offers unmatched versatility and reliability, empowering users to communicate effortlessly across languages and borders.",
"## How to use",
"### On GPU\n\n'I love you'",
"### On CPU\n\n'I love you'",
"## Author\n'\nPhan Minh Toan \n'"
] | [
"TAGS\n#transformers #pytorch #mt5 #text2text-generation #translation #en #lo #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Lao to English Translation Model\nWelcome to the forefront of linguistic innovation with our groundbreaking T5 language model designed specifically for Lao to English translation. In a rapidly globalizing world where effective communication is paramount, our T5 model stands as a beacon of excellence, offering unparalleled accuracy, fluency, and efficiency in bridging the language gap between Lao and English.\n\nBuilt on state-of-the-art deep learning architecture and trained on vast datasets of Lao and English texts, our language model (LLM) harnesses the power of transformer-based technology to deliver seamless and precise translations. Whether you're a business expanding into Laotian markets, a researcher seeking to access Lao-language resources, or an individual connecting with Lao-speaking communities, our T5 model is your ultimate solution for unlocking linguistic barriers and fostering meaningful cross-cultural exchanges.\n\nWith a commitment to quality and innovation, our translation model not only translates words but also preserves context, tone, and cultural nuances, ensuring that the essence of the original message remains intact in every translated sentence. Whether it's documents, websites, or multimedia content, our LLM model offers unmatched versatility and reliability, empowering users to communicate effortlessly across languages and borders.",
"## How to use",
"### On GPU\n\n'I love you'",
"### On CPU\n\n'I love you'",
"## Author\n'\nPhan Minh Toan \n'"
] | [
61,
308,
4,
9,
9,
8
] | [
"passage: TAGS\n#transformers #pytorch #mt5 #text2text-generation #translation #en #lo #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Lao to English Translation Model\nWelcome to the forefront of linguistic innovation with our groundbreaking T5 language model designed specifically for Lao to English translation. In a rapidly globalizing world where effective communication is paramount, our T5 model stands as a beacon of excellence, offering unparalleled accuracy, fluency, and efficiency in bridging the language gap between Lao and English.\n\nBuilt on state-of-the-art deep learning architecture and trained on vast datasets of Lao and English texts, our language model (LLM) harnesses the power of transformer-based technology to deliver seamless and precise translations. Whether you're a business expanding into Laotian markets, a researcher seeking to access Lao-language resources, or an individual connecting with Lao-speaking communities, our T5 model is your ultimate solution for unlocking linguistic barriers and fostering meaningful cross-cultural exchanges.\n\nWith a commitment to quality and innovation, our translation model not only translates words but also preserves context, tone, and cultural nuances, ensuring that the essence of the original message remains intact in every translated sentence. Whether it's documents, websites, or multimedia content, our LLM model offers unmatched versatility and reliability, empowering users to communicate effortlessly across languages and borders.## How to use### On GPU\n\n'I love you'### On CPU\n\n'I love you'## Author\n'\nPhan Minh Toan \n'"
] | [
0.03372125327587128,
-0.0573212206363678,
-0.005400074180215597,
0.10750769823789597,
0.04517608880996704,
-0.05261755362153053,
-0.006309728138148785,
0.08197543770074844,
0.08130926638841629,
0.009339756332337856,
-0.07962865382432938,
-0.041971102356910706,
0.0311452504247427,
0.05777369812130928,
0.03443858399987221,
-0.30664047598838806,
0.03365660086274147,
-0.06596186757087708,
-0.013466586358845234,
0.07832790166139603,
0.1519637405872345,
-0.005639937706291676,
0.14732718467712402,
0.03433961048722267,
0.016471663489937782,
0.014621843583881855,
-0.07348448038101196,
-0.04455286264419556,
0.09996192902326584,
0.05301608890295029,
0.05909593030810356,
0.018963312730193138,
-0.02445397898554802,
-0.16290025413036346,
0.0084601528942585,
0.054492708295583725,
-0.044980764389038086,
-0.012542990036308765,
-0.0009383921860717237,
-0.09253063797950745,
0.15884929895401,
-0.020173367112874985,
0.05721356347203255,
0.06699621677398682,
-0.14831708371639252,
-0.08259810507297516,
-0.05478421226143837,
-0.08093265444040298,
0.027328284457325935,
0.03862482309341431,
-0.0552237406373024,
0.1096881851553917,
-0.12291973829269409,
0.07014238089323044,
0.0560007318854332,
-0.3471423387527466,
0.004986568354070187,
-0.006071870215237141,
0.03689059242606163,
0.13665615022182465,
-0.07450678199529648,
0.06263723224401474,
-0.03148466721177101,
0.020114662125706673,
0.06102726235985756,
-0.042673010379076004,
0.013797352090477943,
-0.03742224723100662,
-0.1563444882631302,
0.036179229617118835,
0.23619568347930908,
-0.0012362995184957981,
-0.05708320438861847,
-0.07547050714492798,
-0.06695622950792313,
0.07354089617729187,
0.0030467617325484753,
-0.11455851048231125,
0.07426268607378006,
0.028772404417395592,
0.16176044940948486,
-0.12200428545475006,
-0.06796955317258835,
0.02298525907099247,
-0.2057432234287262,
0.06569799780845642,
0.0958968997001648,
0.025019017979502678,
0.01918855682015419,
0.004071883857250214,
-0.14999163150787354,
-0.01600019447505474,
-0.08211430162191391,
-0.08328311145305634,
-0.08121577650308609,
0.002440766431391239,
0.020512862130999565,
-0.00848377589136362,
0.026533974334597588,
0.11129077523946762,
-0.09700014442205429,
-0.013440600596368313,
-0.09521631896495819,
0.04056050628423691,
0.07099521905183792,
0.04895029589533806,
-0.0674004778265953,
-0.05992940068244934,
-0.011532234959304333,
-0.09346127510070801,
-0.048108410090208054,
0.017164302989840508,
-0.049702513962984085,
-0.03797624260187149,
-0.04721042886376381,
0.043632227927446365,
0.01811864785850048,
0.07193194329738617,
-0.033063244074583054,
-0.015066047199070454,
0.04827723652124405,
-0.0708416998386383,
-0.04848537966609001,
-0.0603264756500721,
-0.06268363445997238,
0.11206639558076859,
0.06581206619739532,
0.07065628468990326,
-0.07307128608226776,
0.022248784080147743,
-0.03813649341464043,
-0.014109187759459019,
-0.061812058091163635,
-0.10506660491228104,
0.044232308864593506,
-0.04797375574707985,
-0.016978774219751358,
-0.12404786795377731,
-0.09089614450931549,
0.004058192018419504,
0.031081758439540863,
-0.07262788712978363,
-0.03429882973432541,
-0.015576554462313652,
0.010316838510334492,
-0.03442145138978958,
0.02966753952205181,
-0.09577927738428116,
-0.025719616562128067,
-0.02277054265141487,
-0.027456246316432953,
0.09008482098579407,
-0.07999490946531296,
0.049217067658901215,
-0.1210353821516037,
0.02751144953072071,
-0.1929156631231308,
0.09930546581745148,
-0.08345396816730499,
-0.03873284533619881,
-0.0942615196108818,
-0.01320931687951088,
-0.04519888386130333,
0.027190610766410828,
-0.07884947210550308,
0.07587835937738419,
-0.08317911624908447,
-0.01085989736020565,
0.20406576991081238,
-0.11413415521383286,
-0.058409273624420166,
0.1974702626466751,
-0.007014790549874306,
0.14389613270759583,
0.10781138390302658,
0.2117949277162552,
0.07326891273260117,
-0.13169392943382263,
-0.010143827646970749,
0.035343412309885025,
-0.15180054306983948,
0.11075595766305923,
0.15771104395389557,
-0.01738695800304413,
0.035453177988529205,
0.020455680787563324,
0.009584820829331875,
-0.0185542069375515,
-0.04409238323569298,
-0.031082825735211372,
0.015567242167890072,
0.0056399814784526825,
0.13771960139274597,
-0.059259042143821716,
-0.008266051299870014,
-0.009228144772350788,
-0.09888859838247299,
0.11125127971172333,
0.01182171143591404,
-0.007232203148305416,
0.05988308787345886,
-0.1499321162700653,
0.018664279952645302,
0.11709392070770264,
0.00679669389501214,
-0.08055239915847778,
-0.07188510149717331,
0.06563946604728699,
-0.11685008555650711,
0.14348776638507843,
0.10435232520103455,
0.007322650868445635,
0.05120982974767685,
-0.07777876406908035,
0.08002541214227676,
-0.06497013568878174,
-0.001465548062697053,
0.005077204667031765,
-0.1289331167936325,
0.1050153374671936,
-0.033785395324230194,
0.043317053467035294,
0.017835086211562157,
-0.013709534890949726,
0.06398653239011765,
0.07580514252185822,
0.057449422776699066,
0.05429493263363838,
-0.050474077463150024,
0.0474977008998394,
0.01881030946969986,
0.02250157669186592,
0.047347672283649445,
-0.03409063443541527,
0.01072713267058134,
0.23258453607559204,
-0.04683086648583412,
-0.004659099970012903,
0.0796944797039032,
-0.03475772216916084,
-0.07364963740110397,
-0.09039284288883209,
0.0013279860140755773,
-0.07587572932243347,
0.0167816411703825,
-0.04262382909655571,
0.17580491304397583,
0.011966625228524208,
0.06836855411529541,
-0.15695083141326904,
0.01250478345900774,
-0.0049848053604364395,
-0.09029452502727509,
-0.010778275318443775,
0.052150093019008636,
-0.04851103946566582,
-0.23159684240818024,
0.09435784071683884,
0.07946563512086868,
0.06342726200819016,
0.14063407480716705,
0.02596292831003666,
-0.027372587472200394,
-0.076533243060112,
0.08534127473831177,
-0.02595364861190319,
0.007923861034214497,
-0.08135838061571121,
0.021833812817931175,
0.02660949155688286,
0.09258365631103516,
0.04180566594004631,
0.008349878713488579,
0.03685492277145386,
0.03570981323719025,
0.002518625697121024,
0.0032564180437475443,
0.10898806154727936,
0.0017124833539128304,
0.08034645766019821,
0.00994043704122305,
0.06546350568532944,
-0.03249198570847511,
-0.0006910819793120027,
-0.012550155632197857,
0.10144779086112976,
-0.161415234208107,
-0.329855352640152,
-0.0733826532959938,
-0.04621066153049469,
-0.06680967658758163,
-0.05999785289168358,
0.0534215122461319,
-0.02713705040514469,
-0.05166756734251976,
-0.05287923291325569,
0.09081391245126724,
-0.08194845914840698,
-0.13607923686504364,
-0.1479460895061493,
0.07236050814390182,
-0.09363418072462082,
-0.07823753356933594,
-0.03094865381717682,
-0.002082171617075801,
-0.1596151739358902,
0.04577585682272911,
-0.09736329317092896,
0.03489550203084946,
-0.009329081512987614,
0.01413966529071331,
-0.0396680012345314,
-0.06150870770215988,
0.1636948138475418,
-0.051226045936346054,
0.09875167161226273,
0.14484599232673645,
-0.015203752554953098,
0.102227583527565,
0.17130714654922485,
-0.0073365988209843636,
-0.03361513465642929,
0.04552220180630684,
0.0845627635717392,
-0.0269435103982687,
-0.22365310788154602,
-0.11892019957304001,
-0.07056667655706406,
0.011746239848434925,
-0.09577444195747375,
0.010814636014401913,
0.0488818921148777,
-0.010610182769596577,
0.0035362716298550367,
-0.013505310751497746,
0.0525178462266922,
0.04952896386384964,
0.2843591272830963,
-0.062106337398290634,
0.07871796935796738,
-0.03667205572128296,
0.018805358558893204,
0.10468640923500061,
0.12384135276079178,
0.28009510040283203,
0.03756092116236687,
0.21127746999263763,
0.14418400824069977,
0.19081929326057434,
0.09292903542518616,
0.0034026235807687044,
-0.02268049120903015,
0.046956803649663925,
-0.04877275973558426,
-0.04021555930376053,
0.0997893288731575,
0.055980384349823,
0.12203948199748993,
-0.05999055504798889,
-0.0005653324769809842,
0.12513583898544312,
0.08414170891046524,
0.18653687834739685,
-0.05386548489332199,
0.012572027742862701,
0.040850333869457245,
-0.027198294177651405,
-0.04348808526992798,
-0.10033529996871948,
0.06202564388513565,
0.07899339497089386,
-0.17035646736621857,
0.024320874363183975,
0.0198544729501009,
0.06129202991724014,
-0.0643555298447609,
-0.015992579981684685,
-0.048929087817668915,
0.02409781701862812,
0.004465682897716761,
0.09130293130874634,
-0.23351898789405823,
0.2042117863893509,
0.01744009740650654,
-0.015469136647880077,
-0.01748650148510933,
-0.014149890281260014,
-0.021047396585345268,
0.13020510971546173,
0.17309826612472534,
0.04965772107243538,
0.020176207646727562,
0.037996333092451096,
-0.05187565088272095,
0.0051759881898760796,
0.13686923682689667,
-0.007847373373806477,
0.03547642007470131,
-0.050717469304800034,
-0.0072602699510753155,
-0.05790400132536888,
0.08616791665554047,
-0.14669562876224518,
-0.14181305468082428,
0.058747947216033936,
-0.11869128793478012,
-0.034774281084537506,
-0.0258480291813612,
-0.009341660887002945,
-0.0016597169451415539,
0.02836971916258335,
-0.06444116681814194,
-0.11095812916755676,
-0.0976271703839302,
-0.11561203747987747,
0.09517285972833633,
-0.08409811556339264,
0.06088867411017418,
-0.08589781820774078,
0.017722761258482933,
-0.01949080266058445,
-0.04052084684371948,
0.08010531216859818,
-0.006171135231852531,
-0.07804793119430542,
0.0022563596721738577,
0.10631239414215088,
0.05930278077721596,
0.01908051408827305,
0.029142653569579124,
0.01969919167459011,
-0.0011181463487446308,
-0.0907987505197525,
-0.0304420106112957,
0.005417598877102137,
-0.04253637418150902,
0.11763697117567062,
-0.1494866907596588,
-0.022772379219532013,
-0.09747600555419922,
-0.08467914164066315,
0.12923520803451538,
0.14588184654712677,
-0.05301998555660248,
0.20047104358673096,
0.18644729256629944,
-0.10747810453176498,
-0.2021738886833191,
-0.0423702746629715,
0.03361501917243004,
-0.0453476645052433,
0.10363157838582993,
-0.18405947089195251,
-0.011212339624762535,
0.03401011973619461,
-0.0006322890403680503,
-0.10394326597452164,
-0.14274582266807556,
-0.11245878785848618,
-0.004177838563919067,
-0.06066761910915375,
0.06643502414226532,
-0.022184811532497406,
-0.01277963351458311,
0.023707151412963867,
-0.11832290142774582,
0.10663391649723053,
-0.04937302693724632,
0.06187101826071739,
0.01404032576829195,
0.07619494199752808,
0.024689991027116776,
0.05220077559351921,
0.10149642080068588,
-0.08196457475423813,
0.0031791389919817448,
-0.08862204104661942,
-0.02996675856411457,
0.08610871434211731,
0.00916345976293087,
0.0889028012752533,
-0.014309944584965706,
0.007839562371373177,
0.02802695892751217,
-0.07968343794345856,
-0.06188841909170151,
0.08147549629211426,
-0.05662801116704941,
-0.09011887013912201,
-0.0976218581199646,
0.13773007690906525,
0.02862374670803547,
-0.010154257528483868,
-0.06388957053422928,
-0.054898954927921295,
-0.04060909152030945,
0.15381473302841187,
0.1717463582754135,
0.002240060595795512,
-0.06111467629671097,
-0.032745540142059326,
-0.01063529122620821,
0.05294976010918617,
-0.09888350963592529,
0.009238465689122677,
0.12640143930912018,
-0.06927851587533951,
0.09729862213134766,
-0.015712643042206764,
-0.17796485126018524,
0.025957483798265457,
0.06762906163930893,
-0.0033162569161504507,
-0.31290149688720703,
-0.05597994849085808,
-0.03633342310786247,
0.04676370322704315,
-0.06018926948308945,
0.12224537879228592,
-0.0862923115491867,
-0.009326400235295296,
-0.005702167749404907,
0.040065325796604156,
-0.058004237711429596,
-0.05668599531054497,
-0.0552644282579422,
0.0005440599634312093,
-0.043555181473493576,
0.15925081074237823,
0.11970768868923187,
-0.11773168295621872,
-0.002919314196333289,
0.1026068851351738,
-0.07683974504470825,
-0.08643553406000137,
-0.05054979398846626,
0.06620875746011734,
-0.06369476765394211,
-0.03700918331742287,
0.09374134987592697,
-0.09329653531312943,
-0.04512592777609825,
0.22838518023490906,
0.024991007521748543,
0.01850336417555809,
0.023861773312091827,
-0.004653834737837315,
0.03609829023480415,
0.07176948338747025,
-0.005723919719457626,
-0.0460117943584919,
-0.04058140143752098,
-0.020662562921643257,
0.061653804033994675,
0.09032315015792847,
-0.024661563336849213,
-0.10446835309267044,
-0.09420373290777206,
-0.01934485137462616,
-0.0746048092842102,
-0.005673660431057215,
-0.03751048818230629,
-0.025149879977107048,
-0.010288745164871216,
-0.0687071681022644,
-0.042307306081056595,
0.019033024087548256,
-0.08349860459566116,
0.003810401074588299,
0.025305381044745445,
0.1356283277273178,
-0.09243269264698029,
-0.039646707475185394,
0.15728691220283508,
0.019436011090874672,
0.08281882107257843,
-0.0277030561119318,
-0.060185860842466354,
0.05832584947347641,
0.015931546688079834,
0.041385769844055176,
-0.04596225172281265,
0.06699605286121368,
0.044156815856695175,
-0.06789132207632065,
-0.022203145548701286,
-0.02786257490515709,
0.007184147369116545,
-0.011745973490178585,
0.1619829386472702,
-0.0917842835187912,
0.009105890989303589,
0.0022516532335430384,
-0.0914427861571312,
-0.07008139044046402,
-0.013414785265922546,
0.021500656381249428,
0.06687720119953156,
0.04503345862030983,
-0.010966040194034576,
0.031069055199623108,
-0.03374899923801422,
0.011732653714716434,
0.040094077587127686,
0.005269717890769243,
0.09873583912849426,
-0.08323976397514343,
-0.008793612942099571,
0.004053809214383364,
0.22091266512870789,
0.06719561666250229,
-0.05059655010700226,
0.03880271315574646,
-0.020497992634773254,
-0.17391513288021088,
0.0011542087886482477,
-0.10263612121343613,
0.011158916167914867,
-0.011571048758924007,
-0.04262347146868706,
-0.023937534540891647,
-0.14003604650497437,
-0.10669378936290741,
0.08023672550916672,
0.2159835398197174,
0.0994655042886734,
-0.01537330076098442,
0.08899714052677155,
0.032755762338638306,
-0.09473821520805359,
-0.0836741030216217,
0.021417783573269844,
0.04257301241159439,
-0.10471213608980179,
0.038685716688632965,
0.13179875910282135,
-0.11377736926078796,
0.19062434136867523,
0.025677626952528954,
-0.060853779315948486,
-0.12620890140533447,
-0.14121319353580475,
-0.0447828471660614,
-0.04694468528032303,
-0.01354281697422266,
-0.10857269912958145,
0.08113952726125717,
0.010156030766665936,
0.09638277441263199,
-0.045432642102241516,
0.055639371275901794,
-0.24965593218803406,
-0.15739436447620392,
0.036907244473695755,
-0.0024083598982542753,
0.10019305348396301,
0.02232319675385952,
0.02755316160619259,
-0.07787878811359406,
0.06943535059690475,
-0.07211966067552567,
0.06056294962763786,
-0.06349630653858185,
0.00851631909608841,
-0.082961305975914,
-0.012294448912143707,
-0.019367774948477745,
0.025978095829486847,
-0.07154551893472672,
0.08834613859653473,
0.02655155025422573,
-0.07339135557413101,
0.016072912141680717,
0.25268352031707764,
0.03766729682683945,
-0.1008409857749939,
-0.1506992131471634,
0.08146238327026367,
-0.008399500511586666,
0.035380251705646515,
0.0010514471214264631,
0.0019636384677141905,
-0.08120180666446686,
0.1830788105726242,
0.14904436469078064,
-0.04438730701804161,
0.003039531409740448,
0.03402969613671303,
0.04297526553273201,
0.0604628287255764,
0.11322969943284988,
0.02613852731883526,
0.28675714135169983,
-0.07287601381540298,
0.09688354283571243,
-0.09049589186906815,
0.009447674266994,
-0.10640434175729752,
0.0607428252696991,
-0.0018321791430935264,
0.004401843994855881,
-0.038475193083286285,
0.1145009845495224,
-0.049906179308891296,
-0.07243674248456955,
-0.02018863707780838,
-0.0586317777633667,
-0.06936246156692505,
-0.0690690353512764,
0.009814553894102573,
0.020028840750455856,
0.013363653793931007,
0.03921206295490265,
-0.013121993280947208,
0.06130574271082878,
0.08803007751703262,
-0.07420868426561356,
-0.011307175271213055,
0.15547217428684235,
0.02362966723740101,
0.07181357592344284,
0.008755339309573174,
0.08794105052947998,
0.06719262897968292,
0.05391371622681618,
-0.05074896663427353,
0.04257100075483322,
0.011164583265781403,
-0.06375160068273544,
-0.013775059953331947,
0.049612097442150116,
0.0211881585419178,
0.03055688552558422,
0.05440589040517807,
0.008895890787243843,
0.10204961895942688,
0.020863033831119537,
-0.022900966927409172,
-0.040111519396305084,
0.13714390993118286,
-0.1697983294725418,
0.08500316739082336,
0.16636843979358673,
0.016696464270353317,
0.021706748753786087,
-0.07133778929710388,
-0.021530652418732643,
0.025868389755487442,
-0.06480848789215088,
-0.06565581262111664,
-0.04054161161184311,
-0.0031173282768577337,
0.06797724962234497,
0.04159337654709816,
-0.05573279783129692,
-0.065346859395504,
-0.009949873201549053,
0.06710312515497208,
-0.07049136608839035,
0.09211975336074829,
0.12844188511371613,
0.0019111547153443098,
0.0008285208605229855,
-0.22303202748298645,
0.010229049250483513,
0.07146873325109482,
-0.10395625233650208,
-0.007989559322595596
] |
null | null | transformers | # MetricX-23
*This is not an officially supported Google product.*
**GitHub repository: [https://github.com/google-research/metricx](https://github.com/google-research/metricx)**
This repository contains the MetricX-23 models,
a family of models for automatic evaluation of translations that were proposed
in the WMT'23 Metrics Shared Task submission
[MetricX-23: The Google Submission to the WMT 2023 Metrics Shared Task](https://aclanthology.org/2023.wmt-1.63/).
The models were trained in [T5X](https://github.com/google-research/t5x) and
then converted for use in PyTorch.
## Available Models
There are 6 models available on HuggingFace that vary in the number of
parameters and whether or not the model is reference-based or reference-free
(also known as quality estimation, or QE):
* [MetricX-23-XXL](https://huggingface.co/google/metricx-23-large-v2p0)
* [MetricX-23-XL](https://huggingface.co/google/metricx-23-xl-v2p0)
* [MetricX-23-Large](https://huggingface.co/google/metricx-23-xxl-v2p0)
* [MetricX-23-QE-XXL](https://huggingface.co/google/metricx-23-qe-large-v2p0)
* [MetricX-23-QE-XL](https://huggingface.co/google/metricx-23-qe-xl-v2p0)
* [MetricX-23-QE-Large](https://huggingface.co/google/metricx-23-qe-xxl-v2p0)
We recommend using the XXL model versions for the best agreement with human
judgments of translation quality, the Large versions for best speed, and the
XL for an intermediate use case.
## Changes to the WMT'23 Submission
These models available here are most similar to the primary submission to the WMT'23 Metrics
Shared Task. They are initialized with [mT5](https://aclanthology.org/2021.naacl-main.41/)
then fine-tuned on a combination of direct assessment and MQM data. However,
we made some changes that make these models different from the WMT'23 submissions.
First, the models are trained to regress the actual MQM score rather than a
normalized score between 0 and 1. **That means the output from the MetricX-23
models is a score in the range [0, 25] where lower is better (i.e., it predicts
an error score).**
Second, these models were trained with a larger variety of synthetic data that
makes them more robust to translation edge cases like over- and undertranslation,
described in more detail in the following section.
### Synthetic Data
In order for our MetricX models to learn to identify certain types of bad
translations that are not sufficiently (or at all) represented in the regular
training data, we created synthetic examples and mixed them in during training.
The synthetic training data was generated from the DA datasets ranging from
WMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have
the candidate translation manipulated so as to turn it into a bad translation
with a specific issue commonly unrecognized by learned metrics.
The table below provides an overview of the various failure modes that we
considered, including brief descriptions of how we prepared the synthetic data
to address them.
| Failure mode | Synthetic example description |
| ----------- | ----------- |
| Undertranslation | Candidate translation with an arbitrary sentence removed (if multi-sentence); alternatively, candidate with a certain proportion of words removed from the end. |
| Overtranslation | Candidate translation duplicated (with space in between). |
| Fluent but unrelated translation | Arbitrary reference of a similar length from the dataset. |
| Gibberish | Text of a similar length as the reference, generated by sampling words from the reference translation vocabulary (built from all references in the data). |
| Missing punctuation | Reference translation with the end punctuation removed (11 punctuation symbols considered). |
| Latin instead of Chinese/Japanese or Hindi/Bengali punctuation | Candidate translation with the language-specific punctuation symbol at the end replaced with the Latin equivalent (e.g., "." instead of "。" or "।"); alternatively, the punctuation symbol is replaced with the Latin equivalent in the reference, keeping the correct one in the candidate. |
| Reference-matching translation | Reference translation copied as the candidate translation (unlike the rest of the synthetic data, these examples are meant to train the metric to predict a perfect score for candidates matching the reference). |
Examples from the first 4 categories were assigned a label corresponding to the
worst score on the given rating scale (e.g., 25 when mixed with MQM training
data), whereas the reference-matching translation examples are assigned the best
score (e.g., 0 when used with MQM data). The missing/incorrect punctuation
examples were labeled with a score slightly worse than perfect.
Note that some of the synthetic datasets are only meaningful in the
reference-based scenario, and we thus excluded them when training a QE variant
of MetricX. These are the Latin-vs-special punctuation and the
reference-matching translation examples.
Most of the synthetic training sets were created using stratified sampling
across target languages, taking 500 examples per target language. One exception
is the missing punctuation set, which used a stratified sample across different
punctuation symbols instead.
When training MetricX, a small proportion of the synthetic examples was mixed
with the regular training examples. During the first-stage fine-tuning on DA
data, each synthetic training set constituted between 0.1% and 1% of all
training examples, whereas in the second-stage fine-tuning on MQM data we used
an even smaller proportion, around 0.05%.
As for evaluating the effect of the synthetic training data on the model's
performance, the DEMETR challenge set - which we originally used to evaluate the
models submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We
therefore created a new DEMETR-style test set based on the WMT22 DA data, with
examples constructed analogically to the synthetic training examples, as
described above. This test set helped us determine the right proportions of
synthetic data for fine-tuning in order to make MetricX robust for the failure
modes in consideration, without sacrificing the system- and segment-level
correlations with human ratings.
## Usage
The code for using MetricX models can be found at [https://github.com/google-research/metricx](https://github.com/google-research/metricx).
The repository contains example prediction scripts, described below.
The `metricx23/predict.py` script contains an example for how to run inference
on the models.
### Reference-Based
Example usage for a reference-based model:
```bash
python -m metricx23.predict \
--tokenizer google/mt5-xl \
--model_name_or_path google/metricx-23-xl-v2p0 \
--max_input_length 1024 \
--batch_size 1 \
--input_file input.jsonl \
--output_file output.jsonl
```
`input.jsonl` is expected to have 1 serialized JSON object per line with
`"reference"` and `"hypothesis"` fields. The output jsonl will be parallel
to `input.jsonl` but additionally contain a `"prediction"` field with the predicted score.
Note that the model was trained with a maximum input length of 1024 tokens, so
significantly increasing that value may lead to unpredictable behavior.
### Reference-Free
Example usage for a reference-free model:
```bash
python -m metricx23.predict \
--tokenizer google/mt5-xl \
--model_name_or_path google/metricx-23-qe-xl-v2p0 \
--max_input_length 1024 \
--batch_size 1 \
--input_file input.jsonl \
--output_file output.jsonl \
--qe
```
`input.jsonl` is expected to have 1 serialized JSON object per line with
`"source"` and `"hypothesis"` fields. The output jsonl will be parallel
to `input.jsonl` but additionally contain a `"prediction"` field with the predicted score.
## Meta-Evaluation
The `metricx23/evaluate.py` script contains code to calculate various correlations
between the MetricX-23 scores and MQM ratings of translation quality using the
[MT Metrics Eval](https://github.com/google-research/mt-metrics-eval) library.
Example usage:
```bash
python -m metricx23.evaluate \
--dataset wmt22 \
--lp en-de \
--input_file input.jsonl \
--output_file output.json
```
`input.jsonl` is expected to have one JSON object serialized per line.
Each JSON object is expected to contain 4 fields:
* `"system_id"`: The name of the system that generated the translation.
* `"segment_id"`: The 0-based index of the corresponding segment in the MT
Metrics Eval data.
* `"label"`: The ground-truth translation quality score (with higher is better).
* `"prediction"`: The model predicted translation quality score (with lower is
better; the script negates the scores so higher is better).
The script will calculate the 4 agreement/correlations that were used in the
WMT'23 Shared Task. Below are the results for the MetricX-23 models on the
WMT'22 Metrics Shared Task data:
English-German:
| Model | System-Level Accuracy | System-Level Pearson | Segment-Level Pearson | Segment-Level Pairwise Acc |
| ----------- | ----------- | ----------- | ----------- | ----------- |
| MetricX-23-XXL | 0.795 | 0.835 | 0.546 | 0.619 |
| MetricX-23-XL | 0.756 | 0.813 | 0.540 | 0.605 |
| MetricX-23-Large | 0.769 | 0.759 | 0.507 | 0.595 |
| MetricX-23-QE-XXL | 0.769 | 0.830 | 0.490 | 0.606 |
| MetricX-23-QE-XL | 0.718 | 0.684 | 0.421 | 0.594 |
| MetricX-23-QE-Large | 0.744 | 0.671 | 0.387 | 0.579 |
English-Russian:
| Model | System-Level Accuracy | System-Level Pearson | Segment-Level Pearson | Segment-Level Pairwise Acc |
| ----------- | ----------- | ----------- | ----------- | ----------- |
| MetricX-23-XXL | 0.905 | 0.943 | 0.477 | 0.609 |
| MetricX-23-XL | 0.876 | 0.906 | 0.498 | 0.589 |
| MetricX-23-Large | 0.876 | 0.841 | 0.474 | 0.569 |
| MetricX-23-QE-XXL | 0.895 | 0.940 | 0.470 | 0.602 |
| MetricX-23-QE-XL | 0.848 | 0.861 | 0.415 | 0.570 |
| MetricX-23-QE-Large | 0.819 | 0.778 | 0.411 | 0.551 |
Chinese-English:
| Model | System-Level Accuracy | System-Level Pearson | Segment-Level Pearson | Segment-Level Pairwise Acc |
| ----------- | ----------- | ----------- | ----------- | ----------- |
| MetricX-23-XXL | 0.868 | 0.919 | 0.605 | 0.551 |
| MetricX-23-XL | 0.868 | 0.924 | 0.584 | 0.543 |
| MetricX-23-Large | 0.857 | 0.919 | 0.555 | 0.539 |
| MetricX-23-QE-XXL | 0.857 | 0.928 | 0.573 | 0.544 |
| MetricX-23-QE-XL | 0.802 | 0.879 | 0.546 | 0.529 |
| MetricX-23-QE-Large | 0.758 | 0.904 | 0.522 | 0.529 |
The `metricx23/evaluate_wmt23.py` script re-calculates the average correlation
score that was used to rank submissions from the
[WMT'23 Shared Task](https://www2.statmt.org/wmt23/pdf/2023.wmt-1.51.pdf).
Example usage:
```bash
python -m metricx23.evaluate_wmt23 \
--en_de predictions_ende.jsonl \
--he_en predictions_heen.jsonl \
--zh_en predictions_zhen.jsonl \
--output_file output.json
```
Each of the 3 input files is expected to be in the same format as described
above. Each file should correspond to running inference on each of the language
pairs from the WMT'23 dataset.
The results for each of the models is the following:
| Model | Average Correlation |
| ----------- | ----------- |
| MetricX-23-XXL | 0.812 |
| MetricX-23-XL | 0.813 |
| MetricX-23-Large | 0.794 |
| MetricX-23-QE-XXL | 0.797 |
| MetricX-23-QE-XL | 0.767 |
| MetricX-23-QE-Large | 0.762 |
## Citation
If you use MetricX-23 in your research, please cite the following publication:
```bibtex
@inproceedings{juraska-etal-2023-metricx,
title = {{MetricX-23: The Google Submission to the WMT 2023 Metrics Shared Task}},
author = "Juraska, Juraj and
Finkelstein, Mara and
Deutsch, Daniel and
Siddhant, Aditya and
Mirzazadeh, Mehdi and
Freitag, Markus",
editor = "Koehn, Philipp and
Haddow, Barry and
Kocmi, Tom and
Monz, Christof",
booktitle = "Proceedings of the Eighth Conference on Machine Translation",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.wmt-1.63",
doi = "10.18653/v1/2023.wmt-1.63",
pages = "756--767",
}
``` | {"license": "apache-2.0"} | null | google/metricx-23-large-v2p0 | [
"transformers",
"pytorch",
"mt5",
"license:apache-2.0",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T15:47:41+00:00 | [] | [] | TAGS
#transformers #pytorch #mt5 #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us
| MetricX-23
==========
*This is not an officially supported Google product.*
GitHub repository: URL
This repository contains the MetricX-23 models,
a family of models for automatic evaluation of translations that were proposed
in the WMT'23 Metrics Shared Task submission
MetricX-23: The Google Submission to the WMT 2023 Metrics Shared Task.
The models were trained in T5X and
then converted for use in PyTorch.
Available Models
----------------
There are 6 models available on HuggingFace that vary in the number of
parameters and whether or not the model is reference-based or reference-free
(also known as quality estimation, or QE):
* MetricX-23-XXL
* MetricX-23-XL
* MetricX-23-Large
* MetricX-23-QE-XXL
* MetricX-23-QE-XL
* MetricX-23-QE-Large
We recommend using the XXL model versions for the best agreement with human
judgments of translation quality, the Large versions for best speed, and the
XL for an intermediate use case.
Changes to the WMT'23 Submission
--------------------------------
These models available here are most similar to the primary submission to the WMT'23 Metrics
Shared Task. They are initialized with mT5
then fine-tuned on a combination of direct assessment and MQM data. However,
we made some changes that make these models different from the WMT'23 submissions.
First, the models are trained to regress the actual MQM score rather than a
normalized score between 0 and 1. That means the output from the MetricX-23
models is a score in the range [0, 25] where lower is better (i.e., it predicts
an error score).
Second, these models were trained with a larger variety of synthetic data that
makes them more robust to translation edge cases like over- and undertranslation,
described in more detail in the following section.
### Synthetic Data
In order for our MetricX models to learn to identify certain types of bad
translations that are not sufficiently (or at all) represented in the regular
training data, we created synthetic examples and mixed them in during training.
The synthetic training data was generated from the DA datasets ranging from
WMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have
the candidate translation manipulated so as to turn it into a bad translation
with a specific issue commonly unrecognized by learned metrics.
The table below provides an overview of the various failure modes that we
considered, including brief descriptions of how we prepared the synthetic data
to address them.
Examples from the first 4 categories were assigned a label corresponding to the
worst score on the given rating scale (e.g., 25 when mixed with MQM training
data), whereas the reference-matching translation examples are assigned the best
score (e.g., 0 when used with MQM data). The missing/incorrect punctuation
examples were labeled with a score slightly worse than perfect.
Note that some of the synthetic datasets are only meaningful in the
reference-based scenario, and we thus excluded them when training a QE variant
of MetricX. These are the Latin-vs-special punctuation and the
reference-matching translation examples.
Most of the synthetic training sets were created using stratified sampling
across target languages, taking 500 examples per target language. One exception
is the missing punctuation set, which used a stratified sample across different
punctuation symbols instead.
When training MetricX, a small proportion of the synthetic examples was mixed
with the regular training examples. During the first-stage fine-tuning on DA
data, each synthetic training set constituted between 0.1% and 1% of all
training examples, whereas in the second-stage fine-tuning on MQM data we used
an even smaller proportion, around 0.05%.
As for evaluating the effect of the synthetic training data on the model's
performance, the DEMETR challenge set - which we originally used to evaluate the
models submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We
therefore created a new DEMETR-style test set based on the WMT22 DA data, with
examples constructed analogically to the synthetic training examples, as
described above. This test set helped us determine the right proportions of
synthetic data for fine-tuning in order to make MetricX robust for the failure
modes in consideration, without sacrificing the system- and segment-level
correlations with human ratings.
Usage
-----
The code for using MetricX models can be found at URL
The repository contains example prediction scripts, described below.
The 'metricx23/URL' script contains an example for how to run inference
on the models.
### Reference-Based
Example usage for a reference-based model:
'URL' is expected to have 1 serialized JSON object per line with
'"reference"' and '"hypothesis"' fields. The output jsonl will be parallel
to 'URL' but additionally contain a '"prediction"' field with the predicted score.
Note that the model was trained with a maximum input length of 1024 tokens, so
significantly increasing that value may lead to unpredictable behavior.
### Reference-Free
Example usage for a reference-free model:
'URL' is expected to have 1 serialized JSON object per line with
'"source"' and '"hypothesis"' fields. The output jsonl will be parallel
to 'URL' but additionally contain a '"prediction"' field with the predicted score.
Meta-Evaluation
---------------
The 'metricx23/URL' script contains code to calculate various correlations
between the MetricX-23 scores and MQM ratings of translation quality using the
MT Metrics Eval library.
Example usage:
'URL' is expected to have one JSON object serialized per line.
Each JSON object is expected to contain 4 fields:
* '"system\_id"': The name of the system that generated the translation.
* '"segment\_id"': The 0-based index of the corresponding segment in the MT
Metrics Eval data.
* '"label"': The ground-truth translation quality score (with higher is better).
* '"prediction"': The model predicted translation quality score (with lower is
better; the script negates the scores so higher is better).
The script will calculate the 4 agreement/correlations that were used in the
WMT'23 Shared Task. Below are the results for the MetricX-23 models on the
WMT'22 Metrics Shared Task data:
English-German:
English-Russian:
Chinese-English:
The 'metricx23/evaluate\_wmt23.py' script re-calculates the average correlation
score that was used to rank submissions from the
WMT'23 Shared Task.
Example usage:
Each of the 3 input files is expected to be in the same format as described
above. Each file should correspond to running inference on each of the language
pairs from the WMT'23 dataset.
The results for each of the models is the following:
If you use MetricX-23 in your research, please cite the following publication:
| [
"### Synthetic Data\n\n\nIn order for our MetricX models to learn to identify certain types of bad\ntranslations that are not sufficiently (or at all) represented in the regular\ntraining data, we created synthetic examples and mixed them in during training.\nThe synthetic training data was generated from the DA datasets ranging from\nWMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have\nthe candidate translation manipulated so as to turn it into a bad translation\nwith a specific issue commonly unrecognized by learned metrics.\n\n\nThe table below provides an overview of the various failure modes that we\nconsidered, including brief descriptions of how we prepared the synthetic data\nto address them.\n\n\n\nExamples from the first 4 categories were assigned a label corresponding to the\nworst score on the given rating scale (e.g., 25 when mixed with MQM training\ndata), whereas the reference-matching translation examples are assigned the best\nscore (e.g., 0 when used with MQM data). The missing/incorrect punctuation\nexamples were labeled with a score slightly worse than perfect.\n\n\nNote that some of the synthetic datasets are only meaningful in the\nreference-based scenario, and we thus excluded them when training a QE variant\nof MetricX. These are the Latin-vs-special punctuation and the\nreference-matching translation examples.\n\n\nMost of the synthetic training sets were created using stratified sampling\nacross target languages, taking 500 examples per target language. One exception\nis the missing punctuation set, which used a stratified sample across different\npunctuation symbols instead.\n\n\nWhen training MetricX, a small proportion of the synthetic examples was mixed\nwith the regular training examples. During the first-stage fine-tuning on DA\ndata, each synthetic training set constituted between 0.1% and 1% of all\ntraining examples, whereas in the second-stage fine-tuning on MQM data we used\nan even smaller proportion, around 0.05%.\n\n\nAs for evaluating the effect of the synthetic training data on the model's\nperformance, the DEMETR challenge set - which we originally used to evaluate the\nmodels submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We\ntherefore created a new DEMETR-style test set based on the WMT22 DA data, with\nexamples constructed analogically to the synthetic training examples, as\ndescribed above. This test set helped us determine the right proportions of\nsynthetic data for fine-tuning in order to make MetricX robust for the failure\nmodes in consideration, without sacrificing the system- and segment-level\ncorrelations with human ratings.\n\n\nUsage\n-----\n\n\nThe code for using MetricX models can be found at URL\nThe repository contains example prediction scripts, described below.\n\n\nThe 'metricx23/URL' script contains an example for how to run inference\non the models.",
"### Reference-Based\n\n\nExample usage for a reference-based model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"reference\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nNote that the model was trained with a maximum input length of 1024 tokens, so\nsignificantly increasing that value may lead to unpredictable behavior.",
"### Reference-Free\n\n\nExample usage for a reference-free model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"source\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nMeta-Evaluation\n---------------\n\n\nThe 'metricx23/URL' script contains code to calculate various correlations\nbetween the MetricX-23 scores and MQM ratings of translation quality using the\nMT Metrics Eval library.\n\n\nExample usage:\n\n\n'URL' is expected to have one JSON object serialized per line.\nEach JSON object is expected to contain 4 fields:\n\n\n* '\"system\\_id\"': The name of the system that generated the translation.\n* '\"segment\\_id\"': The 0-based index of the corresponding segment in the MT\nMetrics Eval data.\n* '\"label\"': The ground-truth translation quality score (with higher is better).\n* '\"prediction\"': The model predicted translation quality score (with lower is\nbetter; the script negates the scores so higher is better).\n\n\nThe script will calculate the 4 agreement/correlations that were used in the\nWMT'23 Shared Task. Below are the results for the MetricX-23 models on the\nWMT'22 Metrics Shared Task data:\n\n\nEnglish-German:\n\n\n\nEnglish-Russian:\n\n\n\nChinese-English:\n\n\n\nThe 'metricx23/evaluate\\_wmt23.py' script re-calculates the average correlation\nscore that was used to rank submissions from the\nWMT'23 Shared Task.\n\n\nExample usage:\n\n\nEach of the 3 input files is expected to be in the same format as described\nabove. Each file should correspond to running inference on each of the language\npairs from the WMT'23 dataset.\n\n\nThe results for each of the models is the following:\n\n\n\nIf you use MetricX-23 in your research, please cite the following publication:"
] | [
"TAGS\n#transformers #pytorch #mt5 #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us \n",
"### Synthetic Data\n\n\nIn order for our MetricX models to learn to identify certain types of bad\ntranslations that are not sufficiently (or at all) represented in the regular\ntraining data, we created synthetic examples and mixed them in during training.\nThe synthetic training data was generated from the DA datasets ranging from\nWMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have\nthe candidate translation manipulated so as to turn it into a bad translation\nwith a specific issue commonly unrecognized by learned metrics.\n\n\nThe table below provides an overview of the various failure modes that we\nconsidered, including brief descriptions of how we prepared the synthetic data\nto address them.\n\n\n\nExamples from the first 4 categories were assigned a label corresponding to the\nworst score on the given rating scale (e.g., 25 when mixed with MQM training\ndata), whereas the reference-matching translation examples are assigned the best\nscore (e.g., 0 when used with MQM data). The missing/incorrect punctuation\nexamples were labeled with a score slightly worse than perfect.\n\n\nNote that some of the synthetic datasets are only meaningful in the\nreference-based scenario, and we thus excluded them when training a QE variant\nof MetricX. These are the Latin-vs-special punctuation and the\nreference-matching translation examples.\n\n\nMost of the synthetic training sets were created using stratified sampling\nacross target languages, taking 500 examples per target language. One exception\nis the missing punctuation set, which used a stratified sample across different\npunctuation symbols instead.\n\n\nWhen training MetricX, a small proportion of the synthetic examples was mixed\nwith the regular training examples. During the first-stage fine-tuning on DA\ndata, each synthetic training set constituted between 0.1% and 1% of all\ntraining examples, whereas in the second-stage fine-tuning on MQM data we used\nan even smaller proportion, around 0.05%.\n\n\nAs for evaluating the effect of the synthetic training data on the model's\nperformance, the DEMETR challenge set - which we originally used to evaluate the\nmodels submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We\ntherefore created a new DEMETR-style test set based on the WMT22 DA data, with\nexamples constructed analogically to the synthetic training examples, as\ndescribed above. This test set helped us determine the right proportions of\nsynthetic data for fine-tuning in order to make MetricX robust for the failure\nmodes in consideration, without sacrificing the system- and segment-level\ncorrelations with human ratings.\n\n\nUsage\n-----\n\n\nThe code for using MetricX models can be found at URL\nThe repository contains example prediction scripts, described below.\n\n\nThe 'metricx23/URL' script contains an example for how to run inference\non the models.",
"### Reference-Based\n\n\nExample usage for a reference-based model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"reference\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nNote that the model was trained with a maximum input length of 1024 tokens, so\nsignificantly increasing that value may lead to unpredictable behavior.",
"### Reference-Free\n\n\nExample usage for a reference-free model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"source\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nMeta-Evaluation\n---------------\n\n\nThe 'metricx23/URL' script contains code to calculate various correlations\nbetween the MetricX-23 scores and MQM ratings of translation quality using the\nMT Metrics Eval library.\n\n\nExample usage:\n\n\n'URL' is expected to have one JSON object serialized per line.\nEach JSON object is expected to contain 4 fields:\n\n\n* '\"system\\_id\"': The name of the system that generated the translation.\n* '\"segment\\_id\"': The 0-based index of the corresponding segment in the MT\nMetrics Eval data.\n* '\"label\"': The ground-truth translation quality score (with higher is better).\n* '\"prediction\"': The model predicted translation quality score (with lower is\nbetter; the script negates the scores so higher is better).\n\n\nThe script will calculate the 4 agreement/correlations that were used in the\nWMT'23 Shared Task. Below are the results for the MetricX-23 models on the\nWMT'22 Metrics Shared Task data:\n\n\nEnglish-German:\n\n\n\nEnglish-Russian:\n\n\n\nChinese-English:\n\n\n\nThe 'metricx23/evaluate\\_wmt23.py' script re-calculates the average correlation\nscore that was used to rank submissions from the\nWMT'23 Shared Task.\n\n\nExample usage:\n\n\nEach of the 3 input files is expected to be in the same format as described\nabove. Each file should correspond to running inference on each of the language\npairs from the WMT'23 dataset.\n\n\nThe results for each of the models is the following:\n\n\n\nIf you use MetricX-23 in your research, please cite the following publication:"
] | [
42,
666,
111,
457
] | [
"passage: TAGS\n#transformers #pytorch #mt5 #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us \n",
"passage: ### Synthetic Data\n\n\nIn order for our MetricX models to learn to identify certain types of bad\ntranslations that are not sufficiently (or at all) represented in the regular\ntraining data, we created synthetic examples and mixed them in during training.\nThe synthetic training data was generated from the DA datasets ranging from\nWMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have\nthe candidate translation manipulated so as to turn it into a bad translation\nwith a specific issue commonly unrecognized by learned metrics.\n\n\nThe table below provides an overview of the various failure modes that we\nconsidered, including brief descriptions of how we prepared the synthetic data\nto address them.\n\n\n\nExamples from the first 4 categories were assigned a label corresponding to the\nworst score on the given rating scale (e.g., 25 when mixed with MQM training\ndata), whereas the reference-matching translation examples are assigned the best\nscore (e.g., 0 when used with MQM data). The missing/incorrect punctuation\nexamples were labeled with a score slightly worse than perfect.\n\n\nNote that some of the synthetic datasets are only meaningful in the\nreference-based scenario, and we thus excluded them when training a QE variant\nof MetricX. These are the Latin-vs-special punctuation and the\nreference-matching translation examples.\n\n\nMost of the synthetic training sets were created using stratified sampling\nacross target languages, taking 500 examples per target language. One exception\nis the missing punctuation set, which used a stratified sample across different\npunctuation symbols instead.\n\n\nWhen training MetricX, a small proportion of the synthetic examples was mixed\nwith the regular training examples. During the first-stage fine-tuning on DA\ndata, each synthetic training set constituted between 0.1% and 1% of all\ntraining examples, whereas in the second-stage fine-tuning on MQM data we used\nan even smaller proportion, around 0.05%.\n\n\nAs for evaluating the effect of the synthetic training data on the model's\nperformance, the DEMETR challenge set - which we originally used to evaluate the\nmodels submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We\ntherefore created a new DEMETR-style test set based on the WMT22 DA data, with\nexamples constructed analogically to the synthetic training examples, as\ndescribed above. This test set helped us determine the right proportions of\nsynthetic data for fine-tuning in order to make MetricX robust for the failure\nmodes in consideration, without sacrificing the system- and segment-level\ncorrelations with human ratings.\n\n\nUsage\n-----\n\n\nThe code for using MetricX models can be found at URL\nThe repository contains example prediction scripts, described below.\n\n\nThe 'metricx23/URL' script contains an example for how to run inference\non the models.### Reference-Based\n\n\nExample usage for a reference-based model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"reference\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nNote that the model was trained with a maximum input length of 1024 tokens, so\nsignificantly increasing that value may lead to unpredictable behavior."
] | [
-0.08049466460943222,
-0.010928971692919731,
-0.004997325129806995,
0.02674533985555172,
0.10898007452487946,
0.017911424860358238,
0.030734051018953323,
0.10377532243728638,
-0.02402583509683609,
-0.03567541018128395,
0.03877764940261841,
0.08288230746984482,
0.015161662362515926,
0.022808298468589783,
0.01772439479827881,
-0.1944202184677124,
0.05506649613380432,
0.008813947439193726,
-0.07605475932359695,
0.06908859312534332,
0.11011916399002075,
-0.04383927211165428,
0.047993674874305725,
0.01802058517932892,
-0.046737369149923325,
0.03515304625034332,
0.02192578837275505,
-0.05999843776226044,
0.09158408641815186,
0.08443905413150787,
0.04427182674407959,
0.026185670867562294,
-0.03256708011031151,
-0.2039792835712433,
0.018693745136260986,
0.059222154319286346,
-0.04639671370387077,
0.032644130289554596,
0.06700465828180313,
-0.04408042132854462,
0.10856176912784576,
-0.03563929721713066,
-0.038342323154211044,
0.052534811198711395,
-0.08053379505872726,
-0.11484964191913605,
-0.06861758977174759,
0.005672447383403778,
0.05251080170273781,
0.11462049186229706,
0.016137007623910904,
0.08802412450313568,
-0.08221147954463959,
0.039690323173999786,
0.1186930239200592,
-0.24400970339775085,
0.022698264569044113,
0.031076578423380852,
0.0391852930188179,
0.059262461960315704,
-0.02663654275238514,
-0.011056511662900448,
0.04030945897102356,
0.036150794476270676,
0.04718852788209915,
-0.02858263999223709,
0.022121166810393333,
0.03753712773323059,
-0.11128073930740356,
-0.06610941141843796,
0.2194429337978363,
0.009360874071717262,
-0.039162129163742065,
-0.012831603176891804,
-0.040371835231781006,
-0.021052006632089615,
0.02562984824180603,
0.011642597615718842,
0.0126093290746212,
0.06509723514318466,
0.08087117969989777,
-0.06699725240468979,
-0.1469363272190094,
-0.02156376652419567,
-0.13039535284042358,
0.056314483284950256,
0.015906725078821182,
0.0635187178850174,
-0.13009536266326904,
0.06865113973617554,
-0.11303252726793289,
-0.037289686501026154,
-0.044206660240888596,
-0.09330593794584274,
0.027131132781505585,
0.0172458216547966,
-0.0671110600233078,
-0.061645977199077606,
0.01954670250415802,
0.09166590869426727,
-0.03881584480404854,
0.02113628201186657,
0.06317485123872757,
0.08239984512329102,
0.02081705816090107,
0.08037593215703964,
0.01793535053730011,
-0.011132435873150826,
0.08186338096857071,
-0.06666721403598785,
0.03507205471396446,
-0.00009720586240291595,
-0.05460785701870918,
-0.04377254843711853,
-0.02165292575955391,
0.07720592617988586,
0.024521537125110626,
0.024704361334443092,
-0.022436054423451424,
-0.014114882797002792,
0.15771594643592834,
-0.08077961206436157,
-0.031236806884407997,
-0.011595608666539192,
-0.05022282153367996,
0.047523126006126404,
0.06264422833919525,
0.006788937374949455,
-0.11120638251304626,
0.05265801399946213,
-0.03781677410006523,
-0.03154580295085907,
-0.08177147805690765,
-0.060364820063114166,
0.05910094082355499,
-0.0520191490650177,
0.022135227918624878,
-0.12485349923372269,
-0.14817503094673157,
-0.0194524135440588,
0.00905279628932476,
0.017627842724323273,
-0.04404270648956299,
0.030841557309031487,
-0.021325333043932915,
-0.010873598977923393,
-0.03838267922401428,
0.04409907013177872,
-0.0735226571559906,
0.01654006540775299,
-0.05891093611717224,
0.006638679653406143,
-0.12875565886497498,
0.05859975516796112,
-0.07795718312263489,
-0.008538316003978252,
-0.12053868919610977,
0.06858694553375244,
-0.06013733893632889,
0.06287765502929688,
-0.06418386101722717,
-0.059423450380563736,
-0.05324282497167587,
0.03673676773905754,
-0.011127032339572906,
0.08703723549842834,
-0.11388625204563141,
-0.04371771216392517,
0.06918473541736603,
-0.1147933155298233,
-0.10631323605775833,
0.05931851267814636,
-0.015358645468950272,
0.03748219460248947,
0.07437919080257416,
0.06889817118644714,
0.13736900687217712,
-0.022680290043354034,
-0.030354108661413193,
0.0774422138929367,
-0.09487991780042648,
-0.0831356942653656,
0.043639007955789566,
-0.02724086493253708,
-0.027338385581970215,
0.010795310139656067,
-0.05432993173599243,
0.053616154938936234,
0.004504588432610035,
-0.041023075580596924,
-0.03911237046122551,
-0.03953332453966141,
-0.03608177229762077,
-0.007416178472340107,
0.04547516256570816,
-0.030613455921411514,
-0.04433886334300041,
0.014297567307949066,
0.05688025429844856,
0.0016403227346017957,
0.06881103664636612,
-0.06005219742655754,
0.05407850816845894,
-0.018144575878977776,
0.008445029146969318,
-0.16996923089027405,
-0.050326526165008545,
0.03246239572763443,
-0.04993841052055359,
0.021022705361247063,
0.02538926526904106,
0.02754078060388565,
-0.014172155410051346,
-0.03300439566373825,
-0.01649932749569416,
0.008937809616327286,
0.008251310326159,
-0.05732375383377075,
-0.12281204760074615,
-0.0057578012347221375,
-0.030820120126008987,
-0.01729525998234749,
-0.027541520074009895,
0.012479400262236595,
0.07977741956710815,
0.044011034071445465,
0.009504541754722595,
0.03403853625059128,
0.005829425062984228,
0.01828007400035858,
-0.025431593880057335,
-0.01253400556743145,
0.06733429431915283,
0.029115518555045128,
-0.08058406412601471,
0.13161158561706543,
-0.1250324547290802,
0.017595291137695312,
0.13202296197414398,
-0.10783318430185318,
0.04871136322617531,
-0.04314408451318741,
-0.031379710882902145,
-0.0057036676444113255,
-0.03291696310043335,
-0.022622399032115936,
0.12076297402381897,
-0.004013901576399803,
0.0929318368434906,
-0.07880045473575592,
-0.05966495722532272,
-0.0097917215898633,
0.0038842959329485893,
-0.010753002017736435,
0.0713651031255722,
0.12100948393344879,
-0.12231792509555817,
0.08771493285894394,
0.13989409804344177,
-0.003570202738046646,
0.12119580805301666,
-0.040826037526130676,
-0.07305974513292313,
0.03446398302912712,
0.004951390903443098,
-0.01554139144718647,
0.04345749691128731,
-0.12445223331451416,
0.015536606311798096,
0.05309225618839264,
0.08432695269584656,
0.11454753577709198,
-0.11917590349912643,
0.0067930202931165695,
0.015571664087474346,
-0.03209373727440834,
-0.09838879108428955,
-0.0021094102412462234,
-0.01198030449450016,
0.07585512846708298,
-0.010641621425747871,
0.034573666751384735,
0.06810279190540314,
-0.033153194934129715,
-0.1333116739988327,
0.1549171507358551,
-0.16806340217590332,
-0.2015051692724228,
-0.1637459099292755,
0.0050877295434474945,
-0.04799898713827133,
0.007370667066425085,
0.10294272005558014,
-0.07140845060348511,
-0.04824806749820709,
-0.030684374272823334,
0.044100407510995865,
-0.026180272921919823,
-0.004666412249207497,
-0.05338970944285393,
0.042327139526605606,
0.016776274889707565,
-0.12135301530361176,
-0.006609104573726654,
-0.0036472089122980833,
-0.017405901104211807,
0.041763320565223694,
-0.07078999280929565,
0.07727597653865814,
0.08807438611984253,
0.01986617222428322,
-0.01348377950489521,
-0.04507536441087723,
0.21933329105377197,
-0.05761788785457611,
0.04443945735692978,
0.2112521529197693,
-0.011195655912160873,
0.05879829078912735,
0.052676692605018616,
-0.004360849969089031,
-0.02312907949090004,
0.02751847915351391,
0.013980694115161896,
-0.0917525514960289,
-0.284024178981781,
-0.06044050678610802,
-0.09147080034017563,
-0.06417541950941086,
0.0023470930755138397,
0.06487169861793518,
0.03722584992647171,
0.050735119730234146,
-0.04127505049109459,
-0.014369443990290165,
0.041575632989406586,
0.05546524375677109,
0.09428055584430695,
0.013206657022237778,
0.06720979511737823,
-0.11070599406957626,
0.012433091178536415,
0.11600261181592941,
0.09805083274841309,
0.23822110891342163,
-0.026172611862421036,
0.04857305809855461,
0.11312685161828995,
0.11436714231967926,
0.08747904002666473,
0.10965031385421753,
-0.04313330352306366,
0.029491525143384933,
-0.041477788239717484,
-0.02848794311285019,
-0.03056623786687851,
0.08165007829666138,
-0.021050430834293365,
-0.0669928565621376,
-0.06534460186958313,
-0.021561436355113983,
0.06433352082967758,
0.21831148862838745,
-0.005303980782628059,
-0.07421030104160309,
-0.05741162598133087,
0.06067470833659172,
-0.010603219270706177,
-0.03260648995637894,
0.07761399447917938,
0.04806586354970932,
-0.07070834934711456,
0.08121111243963242,
-0.0008651819080114365,
0.11229313164949417,
0.014927586540579796,
0.018110742792487144,
0.03192968666553497,
-0.006562814116477966,
0.018823236227035522,
0.11194905638694763,
-0.26115158200263977,
0.12093576788902283,
-0.011933675035834312,
0.05421082302927971,
-0.0762992650270462,
0.022125624120235443,
0.04246993362903595,
0.1273622065782547,
0.09262282401323318,
0.043345239013433456,
-0.14447379112243652,
-0.0103498175740242,
0.0019879117608070374,
0.056262798607349396,
0.04705921933054924,
0.010509418323636055,
0.011808192357420921,
-0.011274687014520168,
0.013288512825965881,
0.015993541106581688,
0.08853989839553833,
-0.0517207607626915,
-0.16778728365898132,
0.04126080125570297,
-0.006679844111204147,
0.015364071354269981,
-0.07193623483181,
0.005432065576314926,
-0.039848387241363525,
0.08026738464832306,
-0.14921778440475464,
-0.0995694100856781,
-0.09588959813117981,
-0.08473988622426987,
0.08787038922309875,
-0.0654374435544014,
0.06406596302986145,
0.007063437253236771,
0.07009734213352203,
-0.03068877011537552,
-0.14400194585323334,
0.057342350482940674,
-0.11315649747848511,
-0.08588340878486633,
-0.0034783631563186646,
0.1125456690788269,
0.013006072491407394,
0.0326993428170681,
0.019293401390314102,
-0.010193015448749065,
-0.046051137149333954,
-0.11050929129123688,
0.01819794625043869,
0.07997186481952667,
0.05762843042612076,
0.05681586265563965,
-0.09417805820703506,
-0.029595911502838135,
-0.04396835342049599,
-0.06783974170684814,
0.1667174994945526,
0.10756810754537582,
-0.06620179116725922,
0.1170005351305008,
0.16132885217666626,
-0.1295406073331833,
-0.19067947566509247,
-0.043041959404945374,
-0.03716177120804787,
-0.01986772194504738,
-0.02807173877954483,
-0.1475318968296051,
0.07539469003677368,
0.09812235087156296,
-0.0024198442697525024,
0.04476727545261383,
-0.2080758810043335,
-0.08643566817045212,
0.04781382530927658,
0.0001132916659116745,
0.18663114309310913,
-0.10906235128641129,
-0.044622793793678284,
-0.052799079567193985,
-0.11352433264255524,
0.14947330951690674,
-0.06072360277175903,
0.05728284642100334,
0.00732648279517889,
0.0032600341364741325,
-0.006010896060615778,
-0.03202845901250839,
0.13251455128192902,
0.03674418851733208,
0.043941304087638855,
-0.028950639069080353,
0.055670365691185,
0.025221113115549088,
-0.04210628941655159,
0.044331781566143036,
-0.04143795743584633,
0.021573659032583237,
-0.14452756941318512,
-0.045783452689647675,
-0.08391499519348145,
0.04536336287856102,
-0.01962660811841488,
-0.039184920489788055,
-0.042702145874500275,
0.0021056830883026123,
0.08968086540699005,
-0.014609623700380325,
0.054970160126686096,
-0.05837482213973999,
0.059537388384342194,
0.07670764625072479,
0.15645602345466614,
-0.10332942008972168,
-0.1217908263206482,
-0.02813919633626938,
-0.021626751869916916,
0.08217503130435944,
-0.1376548409461975,
0.07488036155700684,
0.11647255718708038,
-0.04862956330180168,
0.08627215772867203,
0.055094070732593536,
-0.07522906363010406,
0.005458813160657883,
0.10446281731128693,
-0.07246147096157074,
-0.06008515506982803,
-0.0390506386756897,
0.08281482756137848,
0.029428621754050255,
0.03669393062591553,
0.12870027124881744,
-0.0009662613738328218,
-0.032006535679101944,
-0.02065277099609375,
0.06402802467346191,
-0.0347125381231308,
0.0842379555106163,
0.039138346910476685,
0.017695903778076172,
-0.08644309639930725,
0.1338755190372467,
0.01591717265546322,
-0.1885780692100525,
-0.009080829098820686,
0.11049890518188477,
-0.11632084846496582,
-0.0807068943977356,
-0.025578761473298073,
0.009648185223340988,
-0.14496804773807526,
-0.10269727557897568,
-0.032402172684669495,
-0.14044955372810364,
0.06719717383384705,
0.16770902276039124,
0.02116461656987667,
0.03302031010389328,
-0.006035410333424807,
-0.028474289923906326,
-0.03122306987643242,
-0.013931620866060257,
-0.0277186818420887,
0.017339259386062622,
-0.05742930248379707,
0.0389898419380188,
0.019052214920520782,
0.06222548335790634,
-0.016825642436742783,
0.015219002030789852,
-0.0635049119591713,
0.010224338620901108,
-0.15755128860473633,
0.024624407291412354,
-0.05316181108355522,
-0.017552291974425316,
-0.01309568714350462,
-0.014190688729286194,
-0.05401631444692612,
0.05547712743282318,
-0.09220357239246368,
-0.020396772772073746,
-0.03640173375606537,
0.027906954288482666,
-0.05868091061711311,
-0.02378096804022789,
0.050178419798612595,
-0.047896455973386765,
0.1031966358423233,
0.06362350285053253,
-0.05172618851065636,
0.0687813088297844,
-0.054818395525217056,
-0.03117077425122261,
0.04752342030405998,
0.04614630341529846,
0.0024338315706700087,
-0.06958136707544327,
0.053022030740976334,
0.0532219260931015,
-0.05545207858085632,
-0.00179347675293684,
-0.00773459579795599,
-0.106658935546875,
-0.026197271421551704,
0.026790166273713112,
-0.007194001227617264,
-0.06978121399879456,
-0.046734243631362915,
0.11590144783258438,
0.04049905017018318,
0.1221245527267456,
-0.006192055530846119,
0.039656929671764374,
-0.0669143944978714,
0.0076656281016767025,
-0.0022521058563143015,
-0.0823831707239151,
-0.09226924180984497,
-0.03779274597764015,
0.01616818644106388,
-0.032287146896123886,
0.20258086919784546,
-0.027631420642137527,
-0.001419171690940857,
0.05651244521141052,
0.0006612688302993774,
-0.019393347203731537,
0.012330395169556141,
0.1308664083480835,
0.009836757555603981,
0.00910251121968031,
-0.02165895141661167,
0.024475062265992165,
-0.016778085380792618,
-0.0041928887367248535,
0.13951833546161652,
0.09686414152383804,
0.16705356538295746,
0.1026294082403183,
0.002041283994913101,
-0.03730851784348488,
-0.01945573464035988,
-0.06291651725769043,
0.028470784425735474,
0.058277349919080734,
0.00701283710077405,
0.01671595126390457,
0.11799974739551544,
-0.09208066761493683,
0.09887351095676422,
-0.06913766264915466,
-0.05150070786476135,
-0.1154223382472992,
-0.09975318610668182,
-0.010319891385734081,
-0.08603835850954056,
-0.029209984466433525,
-0.13546273112297058,
-0.020355189219117165,
0.06884671002626419,
0.038902945816516876,
-0.05183970928192139,
0.007247757166624069,
-0.11964748054742813,
-0.10397026687860489,
0.04866647720336914,
-0.014942724257707596,
0.04801199212670326,
0.04992463439702988,
0.012288471683859825,
0.04020651429891586,
-0.03871965408325195,
-0.02611941285431385,
0.039849359542131424,
0.04274921864271164,
0.045678697526454926,
-0.03452969342470169,
-0.010724259540438652,
-0.028795840218663216,
0.012287003919482231,
0.02913379855453968,
0.22296985983848572,
0.010857714340090752,
-0.031127674505114555,
0.06575135886669159,
0.19841772317886353,
-0.04081185162067413,
-0.09934508055448532,
-0.09630105644464493,
0.1641465723514557,
0.045690394937992096,
0.029041193425655365,
0.016438156366348267,
-0.057314880192279816,
-0.00565880723297596,
0.22985684871673584,
0.17313337326049805,
-0.025465190410614014,
-0.009001774713397026,
-0.006234009750187397,
0.014965811744332314,
0.06765225529670715,
0.08408764004707336,
0.1142108142375946,
0.21731261909008026,
-0.06078370288014412,
0.0639389306306839,
-0.0446958988904953,
0.0698341652750969,
-0.034699127078056335,
0.17769397795200348,
0.001534847542643547,
-0.058601561933755875,
-0.0005737040191888809,
0.06745973229408264,
-0.07404687255620956,
-0.10022060573101044,
-0.03632500767707825,
-0.07546216249465942,
-0.05532532557845116,
-0.0036639878526329994,
0.05928308516740799,
0.022997058928012848,
0.041565168648958206,
-0.0059436894953250885,
-0.044010479003190994,
0.10706667602062225,
0.023562224581837654,
-0.17537671327590942,
-0.05920793116092682,
0.08894800394773483,
0.011057954281568527,
0.08248458802700043,
0.007951926440000534,
0.048754554241895676,
0.07727794349193573,
0.03654977306723595,
-0.11864081025123596,
0.07930269092321396,
-0.00997848343104124,
-0.05712908133864403,
0.007062149234116077,
0.014758620411157608,
-0.00663449801504612,
0.020782694220542908,
0.029319513589143753,
-0.03974604234099388,
0.05011465772986412,
0.03723395615816116,
-0.04944480583071709,
-0.00523018091917038,
0.03254415839910507,
-0.10453616082668304,
0.08378694206476212,
0.07704398781061172,
-0.022326018661260605,
-0.027006199583411217,
-0.05415914207696915,
0.02430926077067852,
0.012829716317355633,
-0.023813046514987946,
-0.024040259420871735,
-0.06499078124761581,
0.017759330570697784,
0.01358121633529663,
0.03267041966319084,
-0.13761897385120392,
-0.03263425827026367,
0.009745022282004356,
-0.028417814522981644,
-0.08249472081661224,
0.06735780090093613,
0.043854910880327225,
0.026215003803372383,
-0.03867888078093529,
0.0033451877534389496,
-0.004438048228621483,
0.04008375108242035,
-0.0645875409245491,
-0.08099328726530075
] |
null | null | generic |
# Fork of [Salesforce/blip-image-captioning-large](https://huggingface.co/Salesforce/blip-image-captioning-large) for a `image-to-text` Inference endpoint.
> Inspired by https://huggingface.co/sergeipetrov/blip_captioning
This repository implements a `custom` task for `image-to-text` for 🤗 Inference Endpoints to allow image capturing.
The code for the customized pipeline is in the handler.py.
To use deploy this model an Inference Endpoint you have to select `Custom` as task to use the `handler.py` file.
### expected Request payload
Image to be labeled as binary.
#### CURL
```
curl URL \
-X POST \
--data-binary @car.png \
-H "Content-Type: image/png"
```
#### Python
```python
requests.post(ENDPOINT_URL, headers={"Content-Type": "image/png"}, data=open("car.png", 'rb').read()).json()
``` | {"library_name": "generic", "tags": ["vision", "image-to-text", "endpoints-template"], "inference": false, "pipeline_tag": "image-to-text", "base_model": "Salesforce/blip-image-captioning-large"} | image-to-text | pimcore/IEP__image-capturing-large | [
"generic",
"vision",
"image-to-text",
"endpoints-template",
"base_model:Salesforce/blip-image-captioning-large",
"endpoints_compatible",
"region:us"
] | 2024-02-07T15:52:17+00:00 | [] | [] | TAGS
#generic #vision #image-to-text #endpoints-template #base_model-Salesforce/blip-image-captioning-large #endpoints_compatible #region-us
|
# Fork of Salesforce/blip-image-captioning-large for a 'image-to-text' Inference endpoint.
> Inspired by URL
This repository implements a 'custom' task for 'image-to-text' for Inference Endpoints to allow image capturing.
The code for the customized pipeline is in the URL.
To use deploy this model an Inference Endpoint you have to select 'Custom' as task to use the 'URL' file.
### expected Request payload
Image to be labeled as binary.
#### CURL
#### Python
| [
"# Fork of Salesforce/blip-image-captioning-large for a 'image-to-text' Inference endpoint.\n\n> Inspired by URL\n\nThis repository implements a 'custom' task for 'image-to-text' for Inference Endpoints to allow image capturing. \nThe code for the customized pipeline is in the URL.\n\nTo use deploy this model an Inference Endpoint you have to select 'Custom' as task to use the 'URL' file.",
"### expected Request payload\n\nImage to be labeled as binary.",
"#### CURL",
"#### Python"
] | [
"TAGS\n#generic #vision #image-to-text #endpoints-template #base_model-Salesforce/blip-image-captioning-large #endpoints_compatible #region-us \n",
"# Fork of Salesforce/blip-image-captioning-large for a 'image-to-text' Inference endpoint.\n\n> Inspired by URL\n\nThis repository implements a 'custom' task for 'image-to-text' for Inference Endpoints to allow image capturing. \nThe code for the customized pipeline is in the URL.\n\nTo use deploy this model an Inference Endpoint you have to select 'Custom' as task to use the 'URL' file.",
"### expected Request payload\n\nImage to be labeled as binary.",
"#### CURL",
"#### Python"
] | [
52,
115,
16,
4,
3
] | [
"passage: TAGS\n#generic #vision #image-to-text #endpoints-template #base_model-Salesforce/blip-image-captioning-large #endpoints_compatible #region-us \n# Fork of Salesforce/blip-image-captioning-large for a 'image-to-text' Inference endpoint.\n\n> Inspired by URL\n\nThis repository implements a 'custom' task for 'image-to-text' for Inference Endpoints to allow image capturing. \nThe code for the customized pipeline is in the URL.\n\nTo use deploy this model an Inference Endpoint you have to select 'Custom' as task to use the 'URL' file.### expected Request payload\n\nImage to be labeled as binary.#### CURL#### Python"
] | [
-0.023671861737966537,
-0.06029039993882179,
-0.0019632112234830856,
0.03561920300126076,
0.12908577919006348,
0.039704855531454086,
-0.01651674322783947,
0.03249632939696312,
0.07754411548376083,
-0.012935532256960869,
0.13124485313892365,
0.13757233321666718,
0.014533348381519318,
0.10433369129896164,
-0.007329554297029972,
-0.2810198664665222,
-0.020349537953734398,
0.08846628665924072,
0.08932026475667953,
0.06524450331926346,
0.04028117656707764,
-0.07848405838012695,
0.16522517800331116,
-0.04320706054568291,
-0.16085195541381836,
0.060058046132326126,
0.006556819658726454,
0.0317915678024292,
0.023954473435878754,
0.07003290206193924,
0.05199255049228668,
-0.030745716765522957,
0.06218409165740013,
-0.11437931656837463,
0.024211393669247627,
0.024132436141371727,
-0.06112481653690338,
0.0032257239799946547,
0.015856362879276276,
-0.025267958641052246,
0.1547260284423828,
-0.09181836992502213,
0.042495716363191605,
0.02799968793988228,
-0.0444226898252964,
0.04039497300982475,
0.11216594278812408,
0.19035504758358002,
0.0921134352684021,
0.00447064358741045,
0.06539551913738251,
-0.06537473946809769,
-0.011130942031741142,
0.13880401849746704,
0.04979242756962776,
-0.1332143396139145,
-0.008618383668363094,
0.18814975023269653,
0.013958434574306011,
0.05388464033603668,
-0.034120842814445496,
-0.0027533918619155884,
-0.0014868996804580092,
0.0675564631819725,
0.053055550903081894,
-0.07659061998128891,
-0.11691651493310928,
-0.015126095153391361,
-0.048140332102775574,
-0.1357792764902115,
0.11132385581731796,
0.024349411949515343,
-0.04257715493440628,
-0.03947238624095917,
-0.08575883507728577,
-0.038559749722480774,
-0.121535524725914,
0.12187273055315018,
0.05414547771215439,
0.03806116804480553,
-0.033498864620923996,
-0.08436595648527145,
-0.08085078001022339,
-0.06386531889438629,
-0.055584345012903214,
-0.011775966733694077,
-0.031740911304950714,
0.12123847007751465,
-0.13051527738571167,
0.13559706509113312,
-0.12269004434347153,
-0.07777904719114304,
-0.038690753281116486,
-0.10311248898506165,
0.014972448348999023,
0.03964782506227493,
0.005807838402688503,
-0.10334571450948715,
-0.00082219741307199,
-0.003377735847607255,
0.13397277891635895,
0.017008468508720398,
-0.04072821885347366,
0.12140771001577377,
0.058046504855155945,
0.07946572452783585,
-0.05068422853946686,
-0.021211769431829453,
0.10732138901948929,
-0.05942130088806152,
0.031061135232448578,
0.02365998551249504,
-0.1511542797088623,
-0.07853150367736816,
-0.053957078605890274,
-0.024433523416519165,
-0.007495776750147343,
0.10194233059883118,
-0.037110816687345505,
-0.08594857901334763,
0.14727503061294556,
-0.022739388048648834,
-0.021112682297825813,
-0.030722079798579216,
-0.03750099241733551,
0.11941317468881607,
0.15332752466201782,
-0.090850830078125,
-0.047930795699357986,
0.11662374436855316,
-0.06015338748693466,
-0.0013449281686916947,
-0.11781532317399979,
-0.03909393772482872,
0.010104197077453136,
-0.06619011610746384,
0.05921836197376251,
-0.12292789667844772,
-0.2851099669933319,
0.026099620386958122,
0.11049546301364899,
0.00805251020938158,
-0.03890664502978325,
0.02123063988983631,
0.03867478668689728,
-0.031915780156850815,
-0.06199822947382927,
0.09741745889186859,
-0.08042184263467789,
0.077731192111969,
-0.09878703951835632,
0.11319907754659653,
-0.17122778296470642,
-0.013530818745493889,
-0.07613649219274521,
0.06663766503334045,
-0.18677064776420593,
0.10558431595563889,
0.05424988269805908,
0.1467798501253128,
-0.02134845219552517,
-0.06537562608718872,
-0.018666386604309082,
0.021602751687169075,
-0.018926482647657394,
0.09022531658411026,
-0.12275189161300659,
0.05258215218782425,
0.1457461565732956,
-0.0562029629945755,
-0.10868656635284424,
0.0018388854805380106,
-0.0555935800075531,
0.06648688018321991,
0.08463338017463684,
0.07196896523237228,
0.005101646296679974,
-0.16296494007110596,
0.03250095620751381,
0.11306656152009964,
-0.19066837430000305,
-0.1327017992734909,
-0.017457667738199234,
0.08317030221223831,
0.008757548406720161,
0.07098551094532013,
-0.027977079153060913,
0.06795158982276917,
0.014120155945420265,
0.00794881209731102,
-0.07771866768598557,
-0.04196691885590553,
-0.12045479565858841,
0.013929402455687523,
-0.0009016047115437686,
0.03571172431111336,
-0.028122780844569206,
0.06677261739969254,
0.03170448914170265,
-0.07116139680147171,
0.02334223873913288,
0.0037619932554662228,
0.09288403391838074,
-0.19122949242591858,
-0.006875311024487019,
-0.12364834547042847,
-0.04520697891712189,
-0.0380469374358654,
0.15213727951049805,
0.07810886949300766,
-0.024756288155913353,
0.022829553112387657,
-0.04779178649187088,
0.017757657915353775,
0.011250494979321957,
0.07624847441911697,
-0.0748424306511879,
-0.048309024423360825,
-0.05581510066986084,
-0.1260078102350235,
-0.0062654572539031506,
-0.11368267238140106,
-0.1333688199520111,
-0.002022158820182085,
0.12331754714250565,
0.06145457178354263,
0.02849404141306877,
0.02098393253982067,
0.026066385209560394,
-0.09759015589952469,
-0.02547597698867321,
-0.046136729419231415,
0.09333175420761108,
0.018239423632621765,
0.03436119854450226,
0.02003890834748745,
0.03241177648305893,
0.08797875791788101,
0.1302814781665802,
-0.19794689118862152,
0.0009487486095167696,
-0.11914163082838058,
-0.04482591897249222,
0.006799470167607069,
0.019432583823800087,
-0.045726530253887177,
0.11547702550888062,
0.06768695265054703,
0.13190403580665588,
-0.0998966172337532,
0.0043375324457883835,
0.03876379877328873,
-0.027319451794028282,
0.020643560215830803,
-0.08834248036146164,
0.0955735445022583,
-0.22937671840190887,
0.059941280633211136,
-0.050365373492240906,
-0.010019732639193535,
0.1754656434059143,
0.05397558957338333,
-0.05475572869181633,
0.0873824879527092,
0.06451716274023056,
-0.034168776124715805,
0.1795569658279419,
-0.17186367511749268,
-0.06352730840444565,
0.07364533841609955,
-0.08904130756855011,
0.07110384851694107,
-0.07395865023136139,
0.02850758656859398,
0.01495310291647911,
-0.007059566676616669,
-0.0011151493526995182,
-0.02056555077433586,
0.020226409658789635,
0.06030028313398361,
0.05416833609342575,
0.029289089143276215,
0.053175654262304306,
-0.05192849412560463,
-0.07282014191150665,
0.1335180550813675,
0.04331649839878082,
-0.20813341438770294,
-0.16401229798793793,
-0.20549342036247253,
-0.11628130823373795,
0.03217528760433197,
-0.0042461007833480835,
0.021464234218001366,
-0.06206526607275009,
-0.03801347687840462,
-0.01235957071185112,
0.05144314467906952,
-0.036746151745319366,
-0.23254455626010895,
0.06394670158624649,
-0.0204977598041296,
-0.08670353144407272,
-0.023116976022720337,
-0.020986763760447502,
0.052280306816101074,
0.038875069469213486,
-0.08741765469312668,
0.10099325329065323,
0.09264669567346573,
-0.03957439213991165,
0.03233887627720833,
0.03210333362221718,
0.186710923910141,
-0.04056460037827492,
0.043435413390398026,
0.1390611082315445,
-0.031173137947916985,
0.0401238389313221,
0.19495739042758942,
0.005983133800327778,
-0.11335336416959763,
0.026667717844247818,
-0.04907280579209328,
-0.10389016568660736,
-0.01448331493884325,
-0.11112259328365326,
-0.13324704766273499,
-0.054307952523231506,
0.21543049812316895,
0.038960281759500504,
-0.06114872917532921,
0.15881559252738953,
-0.023767758160829544,
0.17146223783493042,
-0.05914308503270149,
0.19255641102790833,
0.06894710659980774,
-0.027242589741945267,
0.0247914157807827,
-0.055427562445402145,
-0.0053002373315393925,
0.0705646276473999,
0.17930883169174194,
0.1332392692565918,
-0.1189410462975502,
0.0023660713341087103,
0.055341705679893494,
0.017110709100961685,
0.11563663929700851,
0.17048116028308868,
-0.11758386343717575,
0.009361865930259228,
0.002632454037666321,
-0.0310673788189888,
-0.11257080733776093,
0.07416048645973206,
-0.028060123324394226,
0.029989948496222496,
0.0019798672292381525,
-0.12295717000961304,
-0.005715722218155861,
0.2370331883430481,
0.0029535177163779736,
-0.37141722440719604,
-0.045116301625967026,
0.06508447229862213,
0.012454940937459469,
-0.17041292786598206,
0.013857332989573479,
0.08753946423530579,
-0.09501562267541885,
0.05188047140836716,
-0.01358446292579174,
0.142072394490242,
-0.04164966940879822,
0.013459754176437855,
0.04078945890069008,
-0.01614469103515148,
0.016154112294316292,
0.08677379786968231,
-0.07341936975717545,
0.04434848576784134,
0.03760204836726189,
-0.0062306285835802555,
-0.12576131522655487,
-0.0020591963548213243,
-0.030696172267198563,
0.1698460727930069,
0.12331084907054901,
-0.008987149223685265,
0.14541171491146088,
-0.0226519126445055,
0.02730766125023365,
0.029345104470849037,
-0.019216546788811684,
0.04429344832897186,
0.06223625689744949,
0.0447864793241024,
-0.04276903718709946,
-0.0715332105755806,
0.013899780809879303,
0.07409262657165527,
-0.20375223457813263,
-0.08316027373075485,
-0.04590979218482971,
0.06673616915941238,
0.04094555601477623,
0.1166095957159996,
0.21376369893550873,
0.19861352443695068,
-0.05984988808631897,
-0.14715050160884857,
-0.09083003550767899,
0.09973420947790146,
0.008415725082159042,
0.0004669788759201765,
0.09105373173952103,
-0.0966167077422142,
0.11344186961650848,
-0.03420916199684143,
-0.1562049388885498,
0.07086192816495895,
-0.07908741384744644,
0.07194468379020691,
-0.050950441509485245,
0.06838490813970566,
-0.11146868765354156,
-0.05731847137212753,
0.03818543627858162,
-0.04280506446957588,
-0.07636339217424393,
-0.10015348345041275,
-0.08605839312076569,
0.07650130242109299,
0.17854779958724976,
0.0768832340836525,
-0.11055461317300797,
-0.04046246036887169,
0.06134122610092163,
0.14064516127109528,
0.11463301628828049,
0.013632835820317268,
-0.04718469828367233,
0.016767624765634537,
-0.006818138994276524,
-0.010109703056514263,
-0.19456453621387482,
0.008966520428657532,
0.0005683076451532543,
0.009034642018377781,
-0.10201337188482285,
-0.02759343385696411,
0.1437687873840332,
-0.04921490326523781,
0.00736209936439991,
0.2078542560338974,
-0.15764456987380981,
-0.09041313081979752,
0.013188495300710201,
0.14806890487670898,
0.1577339619398117,
-0.09234273433685303,
-0.004185273312032223,
-0.14441001415252686,
-0.23351912200450897,
0.13100939989089966,
0.015112990513443947,
0.054108526557683945,
-0.042783915996551514,
0.15402428805828094,
0.06308263540267944,
-0.049184251576662064,
0.013826696202158928,
-0.026208901777863503,
0.10670056939125061,
-0.07518362253904343,
0.021343259140849113,
0.08101581037044525,
-0.04563338682055473,
0.14484767615795135,
-0.060992710292339325,
0.039307281374931335,
-0.15609870851039886,
-0.055538076907396317,
-0.06992854923009872,
0.020507700741291046,
0.02663162723183632,
-0.09125884622335434,
-0.07375836372375488,
-0.04911869391798973,
0.1004573330283165,
0.0922984778881073,
0.09259167313575745,
-0.11032414436340332,
-0.12635351717472076,
0.13637438416481018,
0.031592417508363724,
-0.13694800436496735,
-0.22198770940303802,
-0.0480429008603096,
-0.028784889727830887,
0.11099593341350555,
-0.256037175655365,
0.017994727939367294,
0.053863923996686935,
0.042974431067705154,
0.019420843571424484,
0.07839815318584442,
0.024949967861175537,
0.0008318856707774103,
0.07590154558420181,
-0.07524297386407852,
0.06646570563316345,
0.0007319452124647796,
0.11019973456859589,
-0.038304537534713745,
-0.06346607953310013,
0.07513190060853958,
-0.02903033047914505,
-0.003196699544787407,
-0.0045121763832867146,
-0.002514652907848358,
-0.11602677404880524,
-0.04635163024067879,
0.05406928062438965,
0.020359354093670845,
-0.09169749170541763,
-0.012139474973082542,
0.016498107463121414,
-0.12698911130428314,
-0.07631535083055496,
0.07800709456205368,
-0.1206914484500885,
-0.06519512832164764,
-0.029632654041051865,
0.08672525733709335,
-0.13165079057216644,
0.02469293773174286,
0.05184977129101753,
-0.02495027706027031,
0.045844729989767075,
0.11074717342853546,
0.054448023438453674,
-0.025905709713697433,
-0.02102312259376049,
-0.04728824645280838,
-0.0987657755613327,
0.004698189441114664,
-0.057852067053318024,
0.08373120427131653,
-0.138454869389534,
-0.019569920375943184,
0.012197400443255901,
0.1000266969203949,
-0.06284010410308838,
-0.007830563932657242,
-0.130409836769104,
0.03441997990012169,
-0.1534237563610077,
0.12221316248178482,
-0.08358725905418396,
0.013617333024740219,
-0.02523624338209629,
0.09024855494499207,
-0.06266521662473679,
0.012168639339506626,
-0.07071182131767273,
0.007437668275088072,
0.041225600987672806,
0.012320849113166332,
-0.012867847457528114,
-0.013855029828846455,
0.00014547667524311692,
-0.06711805611848831,
0.0357234850525856,
0.005441322922706604,
-0.08013439178466797,
-0.03740400820970535,
-0.08884250372648239,
-0.08852437138557434,
0.08909011632204056,
0.046518344432115555,
0.020736711099743843,
0.21538591384887695,
0.08664534240961075,
0.05180668085813522,
-0.001708638621494174,
-0.060112882405519485,
0.048471059650182724,
-0.09883220493793488,
0.08430328965187073,
0.0003973851562477648,
0.028604568913578987,
-0.07447820156812668,
0.04511638358235359,
-0.002881370484828949,
0.09389191120862961,
0.056574225425720215,
-0.042568549513816833,
0.028890335932374,
-0.10795792937278748,
-0.028411338105797768,
0.003096151864156127,
-0.002988995984196663,
0.0791504830121994,
-0.08148375898599625,
-0.010357601568102837,
-0.023627834394574165,
0.26619815826416016,
0.08971035480499268,
0.027949346229434013,
-0.014326031319797039,
0.14660340547561646,
0.14857284724712372,
0.013475929386913776,
0.22077317535877228,
-0.053389906883239746,
0.02573465369641781,
0.024501658976078033,
0.09485913813114166,
0.07843868434429169,
0.16678614914417267,
0.01997651718556881,
0.11996905505657196,
-0.017016103491187096,
0.06628783792257309,
-0.005338815972208977,
0.06833982467651367,
-0.0037183354143053293,
-0.10392691940069199,
0.07416022568941116,
0.09325261414051056,
-0.11089760065078735,
0.048458706587553024,
0.1416161209344864,
-0.09129258245229721,
0.03883440047502518,
0.11619553714990616,
-0.05254239961504936,
0.019013309851288795,
-0.26313626766204834,
-0.024156464263796806,
-0.12967565655708313,
0.04288984835147858,
-0.05971024930477142,
-0.008368927985429764,
0.11249043047428131,
-0.022449567914009094,
-0.029246043413877487,
0.08570577204227448,
-0.09597659856081009,
-0.07619034498929977,
0.05841304734349251,
0.043067555874586105,
-0.007550784852355719,
0.06764591485261917,
0.1019468829035759,
0.02145514264702797,
-0.11647255718708038,
-0.021902963519096375,
0.029799602925777435,
0.13590523600578308,
-0.017151515930891037,
0.01888764277100563,
0.00027846317971125245,
-0.04497317597270012,
0.05584420636296272,
-0.03996753320097923,
0.200742706656456,
-0.008061278611421585,
-0.011519137769937515,
-0.03984440863132477,
0.12467911839485168,
-0.05071750283241272,
-0.1114867702126503,
-0.042663343250751495,
0.10243181884288788,
-0.03162484988570213,
-0.019597411155700684,
-0.04500656574964523,
-0.07189540565013885,
-0.006760040298104286,
0.3002967834472656,
0.07472484558820724,
-0.13484662771224976,
0.0011234956327825785,
-0.002457322785630822,
0.02025790326297283,
0.019316943362355232,
0.14146830141544342,
0.006111301481723785,
0.38462018966674805,
-0.044279709458351135,
-0.042210426181554794,
0.0010406564688310027,
-0.02893994189798832,
-0.05773166939616203,
-0.058182764798402786,
-0.01432996615767479,
-0.09898021817207336,
-0.05293852463364601,
0.03874496743083,
-0.23558272421360016,
0.08959103375673294,
0.12457654625177383,
-0.07807993143796921,
0.014512649737298489,
-0.06013886630535126,
-0.1190807968378067,
-0.004326879512518644,
0.0448618046939373,
-0.09054749459028244,
-0.052856992930173874,
0.10149355977773666,
-0.03972109779715538,
-0.25745710730552673,
-0.05511453375220299,
0.0197649784386158,
-0.09010159969329834,
0.17192983627319336,
-0.0517912320792675,
0.06894662231206894,
0.03942335769534111,
-0.004182362928986549,
-0.0744398906826973,
-0.05267498642206192,
-0.04241062328219414,
-0.08050375431776047,
0.022594714537262917,
0.08942324668169022,
-0.051174312829971313,
-0.0818038359284401,
-0.06350160390138626,
-0.15441904962062836,
-0.014785490930080414,
0.07743293046951294,
0.06424938142299652,
-0.11731698364019394,
-0.07512339949607849,
-0.10182574391365051,
0.08361619710922241,
0.061896976083517075,
0.0034082443453371525,
-0.054693255573511124,
-0.08696442097425461,
0.11402949690818787,
0.08448220044374466,
0.007548682391643524,
-0.017902826890349388,
0.013771564699709415,
-0.07521448284387589,
-0.039149440824985504,
-0.07741565257310867,
-0.1713232398033142,
0.005593865178525448,
-0.02929646149277687,
-0.0009294296032749116,
-0.05360160022974014,
0.07483431696891785,
0.11703212559223175,
0.06205526739358902,
-0.04970531165599823,
-0.07489441335201263,
0.03401307016611099,
0.060557182878255844,
-0.06672614812850952,
-0.14516586065292358
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->
# Wintersmith/LLM_generated_text_detector
This model is a fine-tuned version of [distilbert/distilbert-base-uncased-finetuned-sst-2-english](https://huggingface.co/distilbert/distilbert-base-uncased-finetuned-sst-2-english) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 0.0082
- Train Accuracy: 0.9974
- Validation Loss: 0.0191
- Validation Accuracy: 0.9941
- Epoch: 1
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- optimizer: {'name': 'Adam', 'weight_decay': None, 'clipnorm': None, 'global_clipnorm': None, 'clipvalue': None, 'use_ema': False, 'ema_momentum': 0.99, 'ema_overwrite_frequency': None, 'jit_compile': True, 'is_legacy_optimizer': False, 'learning_rate': {'module': 'keras.optimizers.schedules', 'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 5e-05, 'decay_steps': 3630, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered_name': None}, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False}
- training_precision: float32
### Training results
| Train Loss | Train Accuracy | Validation Loss | Validation Accuracy | Epoch |
|:----------:|:--------------:|:---------------:|:-------------------:|:-----:|
| 0.0579 | 0.9809 | 0.0272 | 0.9920 | 0 |
| 0.0082 | 0.9974 | 0.0191 | 0.9941 | 1 |
### Framework versions
- Transformers 4.37.0
- TensorFlow 2.15.0
- Datasets 2.15.0
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_keras_callback"], "base_model": "distilbert/distilbert-base-uncased-finetuned-sst-2-english", "model-index": [{"name": "Wintersmith/LLM_generated_text_detector", "results": []}]} | text-classification | Wintersmith/LLM_generated_text_detector | [
"transformers",
"tf",
"distilbert",
"text-classification",
"generated_from_keras_callback",
"base_model:distilbert/distilbert-base-uncased-finetuned-sst-2-english",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] | 2024-02-07T15:55:26+00:00 | [] | [] | TAGS
#transformers #tf #distilbert #text-classification #generated_from_keras_callback #base_model-distilbert/distilbert-base-uncased-finetuned-sst-2-english #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
| Wintersmith/LLM\_generated\_text\_detector
==========================================
This model is a fine-tuned version of distilbert/distilbert-base-uncased-finetuned-sst-2-english on an unknown dataset.
It achieves the following results on the evaluation set:
* Train Loss: 0.0082
* Train Accuracy: 0.9974
* Validation Loss: 0.0191
* Validation Accuracy: 0.9941
* Epoch: 1
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* optimizer: {'name': 'Adam', 'weight\_decay': None, 'clipnorm': None, 'global\_clipnorm': None, 'clipvalue': None, 'use\_ema': False, 'ema\_momentum': 0.99, 'ema\_overwrite\_frequency': None, 'jit\_compile': True, 'is\_legacy\_optimizer': False, 'learning\_rate': {'module': 'keras.optimizers.schedules', 'class\_name': 'PolynomialDecay', 'config': {'initial\_learning\_rate': 5e-05, 'decay\_steps': 3630, 'end\_learning\_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered\_name': None}, 'beta\_1': 0.9, 'beta\_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False}
* training\_precision: float32
### Training results
### Framework versions
* Transformers 4.37.0
* TensorFlow 2.15.0
* Datasets 2.15.0
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* optimizer: {'name': 'Adam', 'weight\\_decay': None, 'clipnorm': None, 'global\\_clipnorm': None, 'clipvalue': None, 'use\\_ema': False, 'ema\\_momentum': 0.99, 'ema\\_overwrite\\_frequency': None, 'jit\\_compile': True, 'is\\_legacy\\_optimizer': False, 'learning\\_rate': {'module': 'keras.optimizers.schedules', 'class\\_name': 'PolynomialDecay', 'config': {'initial\\_learning\\_rate': 5e-05, 'decay\\_steps': 3630, 'end\\_learning\\_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered\\_name': None}, 'beta\\_1': 0.9, 'beta\\_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False}\n* training\\_precision: float32",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.0\n* TensorFlow 2.15.0\n* Datasets 2.15.0\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tf #distilbert #text-classification #generated_from_keras_callback #base_model-distilbert/distilbert-base-uncased-finetuned-sst-2-english #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* optimizer: {'name': 'Adam', 'weight\\_decay': None, 'clipnorm': None, 'global\\_clipnorm': None, 'clipvalue': None, 'use\\_ema': False, 'ema\\_momentum': 0.99, 'ema\\_overwrite\\_frequency': None, 'jit\\_compile': True, 'is\\_legacy\\_optimizer': False, 'learning\\_rate': {'module': 'keras.optimizers.schedules', 'class\\_name': 'PolynomialDecay', 'config': {'initial\\_learning\\_rate': 5e-05, 'decay\\_steps': 3630, 'end\\_learning\\_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered\\_name': None}, 'beta\\_1': 0.9, 'beta\\_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False}\n* training\\_precision: float32",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.0\n* TensorFlow 2.15.0\n* Datasets 2.15.0\n* Tokenizers 0.15.1"
] | [
89,
305,
4,
31
] | [
"passage: TAGS\n#transformers #tf #distilbert #text-classification #generated_from_keras_callback #base_model-distilbert/distilbert-base-uncased-finetuned-sst-2-english #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* optimizer: {'name': 'Adam', 'weight\\_decay': None, 'clipnorm': None, 'global\\_clipnorm': None, 'clipvalue': None, 'use\\_ema': False, 'ema\\_momentum': 0.99, 'ema\\_overwrite\\_frequency': None, 'jit\\_compile': True, 'is\\_legacy\\_optimizer': False, 'learning\\_rate': {'module': 'keras.optimizers.schedules', 'class\\_name': 'PolynomialDecay', 'config': {'initial\\_learning\\_rate': 5e-05, 'decay\\_steps': 3630, 'end\\_learning\\_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered\\_name': None}, 'beta\\_1': 0.9, 'beta\\_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False}\n* training\\_precision: float32### Training results### Framework versions\n\n\n* Transformers 4.37.0\n* TensorFlow 2.15.0\n* Datasets 2.15.0\n* Tokenizers 0.15.1"
] | [
-0.059331197291612625,
0.09972484409809113,
-0.008403055369853973,
0.08544661849737167,
0.10896792262792587,
0.048387814313173294,
0.07808486372232437,
0.13430915772914886,
-0.06341200321912766,
0.1604422926902771,
0.1017378494143486,
0.13323581218719482,
0.052498020231723785,
0.10141566395759583,
-0.07050791382789612,
-0.179926335811615,
0.046800028532743454,
-0.05009843409061432,
-0.037120554596185684,
0.0573892779648304,
0.0717267319560051,
-0.04896648973226547,
0.06960134953260422,
-0.03260934352874756,
-0.04148126766085625,
-0.01177836675196886,
0.010727419517934322,
-0.035843197256326675,
0.05635160207748413,
0.06858517974615097,
0.047783996909856796,
0.0050399270839989185,
-0.006878830026835203,
-0.21385717391967773,
0.0018894332461059093,
0.10980431735515594,
0.007140733301639557,
0.06777910143136978,
0.06121530383825302,
-0.023619547486305237,
0.13063406944274902,
-0.1236693263053894,
0.04433567449450493,
0.03109978884458542,
-0.14901213347911835,
-0.19149664044380188,
-0.10353066772222519,
0.02484552562236786,
0.1042754128575325,
0.0633559376001358,
-0.016945237293839455,
0.129305899143219,
-0.04693286493420601,
0.08939564973115921,
0.11590064316987991,
-0.2716745138168335,
-0.059529248625040054,
0.007309183478355408,
0.043730635195970535,
-0.00961899571120739,
-0.08319591730833054,
-0.007603888865560293,
-0.007721161004155874,
0.022199861705303192,
0.013029600493609905,
-0.030086150392889977,
0.012764044106006622,
-0.06202853098511696,
-0.06217070668935776,
-0.061994943767786026,
0.17400294542312622,
0.0659950003027916,
-0.046809494495391846,
-0.09852346777915955,
-0.03532731533050537,
-0.14078961312770844,
-0.0019879029132425785,
-0.036687616258859634,
0.0027395212091505527,
0.008996655233204365,
0.010764424689114094,
-0.021952860057353973,
-0.040025923401117325,
-0.012421146966516972,
0.01797383278608322,
0.08027325570583344,
0.04330176115036011,
-0.008562497794628143,
0.03513527289032936,
0.06821052730083466,
-0.0272015742957592,
-0.14391273260116577,
-0.026606032624840736,
-0.011740514077246189,
-0.05524059385061264,
-0.023481957614421844,
-0.04583657905459404,
0.018914246931672096,
0.10007000714540482,
0.20439055562019348,
-0.0537957027554512,
0.11207104474306107,
-0.009355706162750721,
0.0233721062541008,
-0.07555293291807175,
0.08760170638561249,
-0.027580251917243004,
-0.08372015506029129,
-0.009520900435745716,
0.0681721419095993,
0.030981777235865593,
-0.04210763797163963,
-0.028068481013178825,
0.04016930237412453,
0.06677975505590439,
0.03925049677491188,
0.02361639216542244,
0.06740102916955948,
-0.09975996613502502,
-0.017460841685533524,
0.04962994530797005,
-0.1225738674402237,
0.06249189004302025,
0.05321674793958664,
-0.07193801552057266,
0.04404508322477341,
0.05057169124484062,
-0.0171787329018116,
-0.10739628970623016,
0.027841513976454735,
-0.0557149238884449,
-0.021388929337263107,
-0.069176584482193,
-0.09729031473398209,
0.01926874928176403,
-0.06472153216600418,
-0.01977953314781189,
-0.05835981294512749,
-0.13727183640003204,
-0.07185547798871994,
0.09054790437221527,
-0.06746702641248703,
-0.04294901341199875,
-0.07330594956874847,
-0.16208995878696442,
0.07048963010311127,
0.004703314509242773,
0.033259905874729156,
-0.05893907696008682,
0.051471151411533356,
-0.010607010684907436,
0.013163326308131218,
0.01728818751871586,
0.020934024825692177,
-0.06589169800281525,
0.06594093888998032,
-0.130449116230011,
0.12723711133003235,
-0.06721072643995285,
0.027202166616916656,
-0.14040471613407135,
-0.0636342391371727,
0.028244372457265854,
0.005749659147113562,
0.10728567093610764,
0.1294248253107071,
-0.16314177215099335,
-0.04087778553366661,
0.112252376973629,
-0.06619834154844284,
-0.07349614053964615,
0.09365807473659515,
-0.02279859595000744,
-0.023004204034805298,
0.06678333878517151,
0.09803302586078644,
0.0877249538898468,
-0.07558846473693848,
-0.01639392226934433,
-0.0697452574968338,
0.032824259251356125,
0.12242413312196732,
0.029935115948319435,
-0.08249922841787338,
-0.02845076657831669,
0.029690152034163475,
-0.025177815929055214,
-0.0048427945002913475,
-0.04624859243631363,
-0.052910175174474716,
-0.034843552857637405,
-0.04028581455349922,
0.053969040513038635,
0.028969697654247284,
-0.005673197563737631,
-0.0852472335100174,
-0.18388395011425018,
0.02378011867403984,
0.04084273800253868,
-0.07106832414865494,
0.009971082210540771,
-0.0511704683303833,
0.07196403294801712,
0.0611538402736187,
0.020624598488211632,
-0.14745774865150452,
-0.10170279443264008,
0.03189873695373535,
-0.017416929826140404,
0.02126447483897209,
-0.044571563601493835,
0.04056408256292343,
0.04736443981528282,
-0.04034307971596718,
-0.03504583239555359,
-0.008354488760232925,
0.010796228423714638,
-0.029880603775382042,
-0.21212896704673767,
-0.013645094819366932,
-0.006931286305189133,
0.09178327023983002,
-0.2675008177757263,
0.0097174858674407,
0.0462375283241272,
0.14965179562568665,
0.022538000717759132,
-0.043188974261283875,
-0.0109042227268219,
0.06387192010879517,
-0.04061739891767502,
-0.07508803904056549,
0.023930834606289864,
0.015118641778826714,
-0.1110989972949028,
-0.05437109246850014,
-0.18694867193698883,
0.07632766664028168,
0.09926939755678177,
-0.04021749645471573,
-0.14396341145038605,
0.024091389030218124,
-0.014119788073003292,
-0.041551947593688965,
0.00916810892522335,
0.02056259661912918,
0.1610104888677597,
0.032743822783231735,
0.11794378608465195,
-0.02848452888429165,
-0.024613497778773308,
0.006943540181964636,
-0.02828812226653099,
-0.030435267835855484,
0.1372208148241043,
-0.04016417637467384,
-0.11429576575756073,
0.09051217138767242,
0.07449200004339218,
-0.08624061197042465,
0.11324749141931534,
-0.03843517228960991,
-0.05297525227069855,
-0.1036921814084053,
0.09753019362688065,
0.06795799732208252,
0.04327204450964928,
-0.09101803600788116,
0.02720368094742298,
0.011006765998899937,
0.004857620224356651,
-0.023754306137561798,
-0.11034534871578217,
0.03226107358932495,
-0.0009348732419312,
-0.046936459839344025,
0.09740587323904037,
0.0050176894292235374,
0.016415957361459732,
0.09829725325107574,
0.01235419325530529,
-0.025766463950276375,
0.023770837113261223,
-0.020224833860993385,
-0.07033586502075195,
0.20934146642684937,
-0.1390170007944107,
-0.1517575979232788,
-0.0870523676276207,
0.009750678203999996,
-0.07851937413215637,
-0.009857802651822567,
0.014451926574110985,
-0.04084976017475128,
-0.06532184034585953,
-0.07935739308595657,
-0.025359589606523514,
0.011837964877486229,
-0.007594863884150982,
-0.013891362585127354,
0.002599261235445738,
0.13741092383861542,
-0.09656572341918945,
-0.033825721591711044,
0.017788032069802284,
-0.09048262238502502,
-0.012988694943487644,
0.05550935119390488,
0.0296221561729908,
0.09466392546892166,
0.02976139262318611,
-0.0011363893281668425,
-0.017989885061979294,
0.19610755145549774,
-0.07609139382839203,
0.03703024983406067,
0.08834118396043777,
-0.03819987550377846,
0.0772084966301918,
0.15952783823013306,
0.0491882748901844,
-0.0832449421286583,
0.0019616857171058655,
0.07284834235906601,
0.0134145338088274,
-0.22515247762203217,
-0.05278502404689789,
-0.05721341818571091,
-0.045605193823575974,
0.08898589760065079,
0.05913923308253288,
0.10657752305269241,
0.011624585837125778,
-0.03529581427574158,
0.06244223192334175,
0.07895781099796295,
0.0925617441534996,
0.1605737954378128,
0.07351064682006836,
0.09544908255338669,
-0.013178732246160507,
-0.009764260612428188,
0.029480984434485435,
0.00588640570640564,
0.16925127804279327,
-0.007604407612234354,
0.11479399353265762,
0.08693726360797882,
0.07813820242881775,
-0.04069234058260918,
-0.03234138339757919,
0.021089233458042145,
0.029726417735219002,
0.024689815938472748,
-0.06559983640909195,
-0.03365950286388397,
0.031114205718040466,
0.050355542451143265,
0.06231456995010376,
-0.07927364110946655,
0.01856185309588909,
0.10164996236562729,
0.16786572337150574,
0.12108392268419266,
-0.29005077481269836,
-0.08869817852973938,
0.02339833974838257,
-0.01972295343875885,
-0.07115083187818527,
-0.0022407113574445248,
0.07767858356237411,
-0.0751434713602066,
0.10119087249040604,
-0.04296369478106499,
0.05725742131471634,
-0.1009116843342781,
0.042917411774396896,
0.08546719700098038,
0.09880125522613525,
0.03149879723787308,
0.020614130422472954,
-0.29067933559417725,
0.21140410006046295,
0.01550650130957365,
0.11592797935009003,
-0.03332967311143875,
0.08211616426706314,
0.04530894011259079,
-0.0490134060382843,
0.09921623766422272,
-0.013869118876755238,
-0.07085967808961868,
-0.16069239377975464,
-0.05542219802737236,
-0.014548171311616898,
0.11153848469257355,
-0.040767520666122437,
0.09666028618812561,
-0.03191486373543739,
-0.02291337214410305,
0.013721109367907047,
-0.023024702444672585,
-0.1951647400856018,
-0.11684194952249527,
0.07552590221166611,
-0.009771429933607578,
-0.0015101638855412602,
-0.04809292405843735,
-0.031690485775470734,
-0.0702933743596077,
0.25347253680229187,
-0.1896386444568634,
-0.05563444644212723,
-0.12065843492746353,
0.06751278787851334,
0.1321922391653061,
-0.07684499770402908,
0.04122922942042351,
-0.01247495412826538,
0.05748749151825905,
0.05713251605629921,
-0.053411904722452164,
0.11049186438322067,
-0.020670481026172638,
-0.19278517365455627,
-0.0679367408156395,
0.11600150913000107,
0.041309554129838943,
0.02121346816420555,
0.00040079077007249,
0.0733334869146347,
0.028547437861561775,
-0.11365284025669098,
0.05834643915295601,
0.04240589961409569,
0.0503234826028347,
0.054571811109781265,
-0.0355311818420887,
-0.05843080207705498,
-0.05623318627476692,
0.019463125616312027,
0.061970800161361694,
0.2944394052028656,
-0.08112646639347076,
0.027571959421038628,
0.004150469787418842,
-0.11245059967041016,
-0.1615842580795288,
-0.006976371631026268,
0.10412122309207916,
-0.0010609697783365846,
-0.037239667028188705,
-0.17225930094718933,
0.05937286838889122,
0.12733857333660126,
0.008397534489631653,
0.0860888734459877,
-0.2915959656238556,
-0.1365148425102234,
0.08560550957918167,
0.07020080834627151,
-0.047880869358778,
-0.2021241933107376,
-0.08697306364774704,
-0.03603417053818703,
-0.02777162380516529,
0.13012245297431946,
-0.046139877289533615,
0.0880035012960434,
0.030491534620523453,
-0.035878315567970276,
0.0339021272957325,
-0.020452383905649185,
0.1791926473379135,
-0.010998898185789585,
0.057238053530454636,
-0.06610451638698578,
-0.012489612214267254,
0.09244024753570557,
-0.10292421281337738,
0.04940709099173546,
-0.07123640924692154,
0.024804214015603065,
-0.12322375178337097,
0.0024370644241571426,
-0.057541415095329285,
0.07556115090847015,
-0.07507851719856262,
0.0018210643902420998,
-0.02140541933476925,
0.0524175688624382,
0.10146740823984146,
0.017646435648202896,
0.1081261858344078,
-0.008822460658848286,
0.17288604378700256,
0.12741057574748993,
0.07253777980804443,
0.06801445037126541,
-0.09486300498247147,
0.0511782169342041,
-0.009705224074423313,
0.03718610107898712,
-0.11398817598819733,
0.045118991285562515,
0.14130799472332,
-0.00679396465420723,
0.14476947486400604,
0.041793134063482285,
-0.046837978065013885,
0.021027784794569016,
0.05762801319360733,
-0.12183681130409241,
-0.06535699218511581,
-0.008392038755118847,
-0.0077551123686134815,
-0.06082577258348465,
-0.017377672716975212,
0.16257740557193756,
-0.02936888486146927,
0.014799563214182854,
0.02981819584965706,
0.06023740768432617,
-0.04148911312222481,
0.13007709383964539,
-0.013459654524922371,
0.07391943037509918,
-0.06829278916120529,
0.12164702266454697,
0.09404310584068298,
-0.12429986894130707,
0.11407686769962311,
0.08498098701238632,
-0.05110267549753189,
-0.04940173402428627,
0.006714733317494392,
0.0909324511885643,
0.06577000766992569,
-0.03028290346264839,
-0.06745284795761108,
-0.12247896194458008,
0.08237714320421219,
0.0925837978720665,
0.018737168982625008,
0.058052223175764084,
-0.028850756585597992,
-0.016100499778985977,
-0.07883039116859436,
0.10168717801570892,
0.09113584458827972,
0.04396554455161095,
-0.11532098799943924,
0.1139768585562706,
0.006760499440133572,
-0.05249383673071861,
0.014586696401238441,
-0.01340479776263237,
-0.19264695048332214,
-0.03157375752925873,
-0.0933612585067749,
0.0487709604203701,
-0.023896601051092148,
-0.0010840181494131684,
0.03301917389035225,
-0.023878900334239006,
-0.0440320186316967,
0.02333167940378189,
-0.0685499757528305,
-0.09290791302919388,
0.031603582203388214,
0.10229568183422089,
-0.134816974401474,
-0.05412079021334648,
0.01784270629286766,
-0.1285742223262787,
0.05328293889760971,
0.015694700181484222,
0.009080138988792896,
-0.0019366088090464473,
-0.11529884487390518,
0.015515219420194626,
0.022290952503681183,
0.004788392223417759,
0.009640799835324287,
-0.16310913860797882,
0.028664149343967438,
-0.0330800861120224,
0.006741994991898537,
0.004858499392867088,
0.06107919663190842,
-0.10557666420936584,
-0.0637081190943718,
-0.02448158711194992,
-0.02563149482011795,
-0.04716281592845917,
0.051339536905288696,
0.13434715569019318,
-0.03290114179253578,
0.15360496938228607,
-0.09720861166715622,
0.053087204694747925,
-0.1848476231098175,
-0.025357360020279884,
0.028647571802139282,
-0.07241090387105942,
-0.06135889142751694,
-0.0014654865954071283,
0.11362513899803162,
-0.07532667368650436,
0.06316238641738892,
-0.050542768090963364,
0.07763100415468216,
0.020174305886030197,
-0.06470053642988205,
-0.09146434813737869,
0.09112917631864548,
0.14071720838546753,
0.05944325029850006,
-0.023615023121237755,
0.008724626153707504,
-0.036624420434236526,
0.049390751868486404,
0.051459770649671555,
0.13995668292045593,
0.09739536046981812,
0.01668451353907585,
0.07001718133687973,
0.04199216142296791,
-0.11691521108150482,
-0.10529305785894394,
0.1464063972234726,
-0.05046498402953148,
0.18168342113494873,
-0.0428844690322876,
0.0475430004298687,
0.04587458446621895,
-0.18846693634986877,
0.046483077108860016,
-0.054064761847257614,
-0.11092972010374069,
-0.07843488454818726,
-0.16998033225536346,
-0.08180276304483414,
-0.07064850628376007,
-0.0033252632711082697,
-0.12616491317749023,
0.039022330194711685,
0.058360692113637924,
0.017318254336714745,
0.0034178136847913265,
0.05888313427567482,
-0.02853725664317608,
-0.007165760733187199,
0.11202149838209152,
0.013380555436015129,
-0.0029230392538011074,
-0.023712534457445145,
-0.06827971339225769,
0.0017847557319328189,
0.01673009619116783,
0.04300425201654434,
0.034719742834568024,
0.01972687430679798,
0.05682208016514778,
-0.027415405958890915,
-0.09154580533504486,
0.0715351477265358,
0.021963974460959435,
-0.02298019826412201,
0.08941774815320969,
0.04304574802517891,
-0.05241674929857254,
-0.012348459102213383,
0.1456887125968933,
-0.0655389279127121,
-0.043910861015319824,
-0.15177574753761292,
0.1904286891222,
0.013883902691304684,
0.019856268540024757,
0.015821659937500954,
-0.07732778042554855,
-0.010346581228077412,
0.11137577891349792,
0.14640700817108154,
-0.02415033057332039,
-0.00958409532904625,
0.07220607995986938,
-0.01185655128210783,
-0.01436493918299675,
0.09478209167718887,
0.06882684677839279,
0.04281517118215561,
-0.031775571405887604,
-0.01056629791855812,
0.0066315485164523125,
-0.0403473787009716,
-0.08990294486284256,
0.06768282502889633,
0.003594751236960292,
0.00047049293061718345,
-0.010757220908999443,
0.07849480211734772,
-0.10069911926984787,
-0.102706678211689,
0.1078123226761818,
-0.1963828206062317,
-0.1776045709848404,
-0.024863289669156075,
0.009043974801898003,
0.0078984210267663,
0.04577596113085747,
0.011566969566047192,
-0.03572343289852142,
0.1467714160680771,
-0.037522606551647186,
-0.02409072034060955,
-0.0690406858921051,
-0.010345366783440113,
-0.009365453384816647,
0.20310352742671967,
0.0006558506283909082,
0.039401520043611526,
0.14295637607574463,
0.023864684626460075,
-0.09667855501174927,
0.03497908264398575,
0.0904531478881836,
-0.09320153295993805,
0.0657622367143631,
0.09084860235452652,
-0.01504660677164793,
0.15278635919094086,
0.10472844541072845,
-0.06728152185678482,
0.013766019605100155,
0.004007885232567787,
-0.040698058903217316,
-0.029475517570972443,
-0.009251938201487064,
-0.07517915219068527,
0.12688104808330536,
0.21968266367912292,
-0.03257209435105324,
-0.0017397545743733644,
-0.022671833634376526,
0.023362692445516586,
0.03268036991357803,
0.07156103849411011,
-0.04292259365320206,
-0.1977330595254898,
0.08548318594694138,
0.02167273871600628,
0.07858043909072876,
-0.11386184394359589,
-0.1020682156085968,
0.022348454222083092,
-0.011652130633592606,
-0.09654995054006577,
0.12162687629461288,
0.03770916908979416,
0.011643048375844955,
-0.06487405300140381,
-0.12705667316913605,
-0.030471013858914375,
0.16960448026657104,
-0.08762355893850327,
-0.0646103173494339
] |
null | null | sentence-transformers |
# OmarAlsaabi/e5-base-mlqa-finetuned-arabic-for-rag-attempt-2
This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
<!--- Describe your model here -->
## Usage (Sentence-Transformers)
Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
```
pip install -U sentence-transformers
```
Then you can use the model like this:
```python
from sentence_transformers import SentenceTransformer
sentences = ["This is an example sentence", "Each sentence is converted"]
model = SentenceTransformer('OmarAlsaabi/e5-base-mlqa-finetuned-arabic-for-rag-attempt-2')
embeddings = model.encode(sentences)
print(embeddings)
```
## Evaluation Results
<!--- Describe how your model was evaluated -->
For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name=OmarAlsaabi/e5-base-mlqa-finetuned-arabic-for-rag-attempt-2)
## Training
The model was trained with the parameters:
**DataLoader**:
`torch.utils.data.dataloader.DataLoader` of length 1334 with parameters:
```
{'batch_size': 4, 'sampler': 'torch.utils.data.sampler.SequentialSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
```
**Loss**:
`sentence_transformers.losses.MultipleNegativesRankingLoss.MultipleNegativesRankingLoss` with parameters:
```
{'scale': 20.0, 'similarity_fct': 'cos_sim'}
```
Parameters of the fit()-Method:
```
{
"epochs": 2,
"evaluation_steps": 50,
"evaluator": "sentence_transformers.evaluation.InformationRetrievalEvaluator.InformationRetrievalEvaluator",
"max_grad_norm": 1,
"optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
"optimizer_params": {
"lr": 1e-05
},
"scheduler": "WarmupLinear",
"steps_per_epoch": null,
"warmup_steps": 266,
"weight_decay": 0.01
}
```
## Full Model Architecture
```
SentenceTransformer(
(0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: XLMRobertaModel
(1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False})
(2): Normalize()
)
```
## Citing & Authors
<!--- Describe where people can find more information --> | {"library_name": "sentence-transformers", "tags": ["sentence-transformers", "feature-extraction", "sentence-similarity"], "pipeline_tag": "sentence-similarity"} | sentence-similarity | OmarAlsaabi/e5-base-mlqa-finetuned-arabic-for-rag-attempt-2 | [
"sentence-transformers",
"safetensors",
"xlm-roberta",
"feature-extraction",
"sentence-similarity",
"endpoints_compatible",
"region:us"
] | 2024-02-07T15:56:35+00:00 | [] | [] | TAGS
#sentence-transformers #safetensors #xlm-roberta #feature-extraction #sentence-similarity #endpoints_compatible #region-us
|
# OmarAlsaabi/e5-base-mlqa-finetuned-arabic-for-rag-attempt-2
This is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
## Usage (Sentence-Transformers)
Using this model becomes easy when you have sentence-transformers installed:
Then you can use the model like this:
## Evaluation Results
For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL
## Training
The model was trained with the parameters:
DataLoader:
'URL.dataloader.DataLoader' of length 1334 with parameters:
Loss:
'sentence_transformers.losses.MultipleNegativesRankingLoss.MultipleNegativesRankingLoss' with parameters:
Parameters of the fit()-Method:
## Full Model Architecture
## Citing & Authors
| [
"# OmarAlsaabi/e5-base-mlqa-finetuned-arabic-for-rag-attempt-2\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.",
"## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:",
"## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL",
"## Training\nThe model was trained with the parameters:\n\nDataLoader:\n\n'URL.dataloader.DataLoader' of length 1334 with parameters:\n\n\nLoss:\n\n'sentence_transformers.losses.MultipleNegativesRankingLoss.MultipleNegativesRankingLoss' with parameters:\n \n\nParameters of the fit()-Method:",
"## Full Model Architecture",
"## Citing & Authors"
] | [
"TAGS\n#sentence-transformers #safetensors #xlm-roberta #feature-extraction #sentence-similarity #endpoints_compatible #region-us \n",
"# OmarAlsaabi/e5-base-mlqa-finetuned-arabic-for-rag-attempt-2\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.",
"## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:",
"## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL",
"## Training\nThe model was trained with the parameters:\n\nDataLoader:\n\n'URL.dataloader.DataLoader' of length 1334 with parameters:\n\n\nLoss:\n\n'sentence_transformers.losses.MultipleNegativesRankingLoss.MultipleNegativesRankingLoss' with parameters:\n \n\nParameters of the fit()-Method:",
"## Full Model Architecture",
"## Citing & Authors"
] | [
44,
71,
38,
29,
86,
5,
6
] | [
"passage: TAGS\n#sentence-transformers #safetensors #xlm-roberta #feature-extraction #sentence-similarity #endpoints_compatible #region-us \n# OmarAlsaabi/e5-base-mlqa-finetuned-arabic-for-rag-attempt-2\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL## Training\nThe model was trained with the parameters:\n\nDataLoader:\n\n'URL.dataloader.DataLoader' of length 1334 with parameters:\n\n\nLoss:\n\n'sentence_transformers.losses.MultipleNegativesRankingLoss.MultipleNegativesRankingLoss' with parameters:\n \n\nParameters of the fit()-Method:## Full Model Architecture## Citing & Authors"
] | [
-0.04529089108109474,
0.09118878841400146,
-0.005426240619271994,
0.06709544360637665,
0.10521475970745087,
0.04291733726859093,
0.16383115947246552,
0.08315861225128174,
-0.07977593690156937,
0.07150132209062576,
0.06664532423019409,
0.1041206568479538,
0.005714823491871357,
0.010549271479249,
-0.015413535758852959,
-0.24740473926067352,
0.03564666584134102,
-0.06881006062030792,
-0.021772058680653572,
0.06142279878258705,
0.1570766568183899,
-0.05176593363285065,
0.067656010389328,
-0.009463428519666195,
-0.059222009032964706,
0.06500285863876343,
-0.01766260527074337,
-0.06463876366615295,
0.09630551189184189,
0.10458171367645264,
0.04984922334551811,
0.017158938571810722,
0.0006875419057905674,
-0.2516482174396515,
0.03302847221493721,
0.04686535894870758,
-0.025769492611289024,
0.03196772187948227,
-0.031598057597875595,
0.0061759608797729015,
0.10760307312011719,
-0.1155075654387474,
0.0356624573469162,
0.015387364663183689,
-0.07949110120534897,
-0.01810324192047119,
-0.006369817536324263,
0.014231264591217041,
0.1297013908624649,
0.08220808953046799,
-0.023187706246972084,
0.12153001874685287,
-0.05031779035925865,
0.10722308605909348,
0.12880952656269073,
-0.2866532802581787,
-0.02710356004536152,
0.0519624687731266,
0.06150070205330849,
0.07907982170581818,
-0.09423451125621796,
0.0282579455524683,
0.019290776923298836,
0.01697942242026329,
0.05990395322442055,
-0.07573086023330688,
-0.07903552055358887,
0.006614535115659237,
-0.06759949773550034,
0.0209050290286541,
0.18829411268234253,
0.04409816116094589,
-0.0301576629281044,
-0.15417683124542236,
-0.05545385181903839,
0.11117196083068848,
-0.05198107287287712,
-0.037908487021923065,
0.03627423569560051,
0.03469407558441162,
0.014231296256184578,
-0.10398431122303009,
-0.10182224959135056,
-0.041520025581121445,
-0.07342556118965149,
0.09766256809234619,
0.02259621024131775,
-0.010973554104566574,
-0.020866699516773224,
0.0157762560993433,
-0.11965395510196686,
-0.10204761475324631,
-0.032316312193870544,
-0.03752459958195686,
-0.09581542760133743,
-0.005564776714891195,
-0.061923544853925705,
-0.04195300117135048,
0.05551768094301224,
0.01995212584733963,
-0.012783531099557877,
0.007853844203054905,
0.03274662792682648,
0.08479398488998413,
-0.0029235875699669123,
0.03360295295715332,
-0.06589773297309875,
-0.06615425646305084,
0.01647304743528366,
0.04825768247246742,
0.058111779391765594,
-0.009566525928676128,
-0.07963726669549942,
-0.04382404312491417,
-0.01017786841839552,
0.07743168622255325,
0.0063562639988958836,
0.06604289263486862,
-0.02638128772377968,
-0.02617928944528103,
0.05472102388739586,
-0.12584351003170013,
-0.0005511401686817408,
0.017837829887866974,
-0.06668365001678467,
-0.03025517240166664,
0.08699999749660492,
-0.014474349096417427,
-0.09471900761127472,
-0.01668955571949482,
-0.09493294358253479,
-0.009245971217751503,
-0.023153113201260567,
-0.16330698132514954,
0.0005339904455468059,
0.02265029214322567,
0.006691708229482174,
-0.15671473741531372,
-0.18981362879276276,
-0.03469143062829971,
0.01655733399093151,
-0.024255774915218353,
-0.05433320254087448,
-0.13464081287384033,
-0.014670629054307938,
-0.00487868906930089,
-0.02958144061267376,
-0.07968052476644516,
-0.03427255153656006,
0.025398967787623405,
-0.037649836391210556,
0.07437752187252045,
-0.014219353906810284,
0.0396946519613266,
-0.058452505618333817,
0.020304393023252487,
-0.043230921030044556,
0.19128520786762238,
-0.0208800807595253,
0.09185431152582169,
-0.12306562811136246,
-0.0032552434131503105,
-0.0033842758275568485,
0.06940095871686935,
0.018900256603956223,
0.18683399260044098,
-0.1621299386024475,
-0.05886053293943405,
0.12033726274967194,
-0.052212782204151154,
-0.18937334418296814,
0.11333363503217697,
-0.05570440739393234,
0.14433163404464722,
0.13580399751663208,
0.1267121583223343,
0.11046739667654037,
-0.03901086747646332,
0.005115773994475603,
0.09242293238639832,
-0.01601269654929638,
0.05409945920109749,
0.0683453381061554,
-0.013178921304643154,
0.07253140956163406,
-0.02186167985200882,
-0.011791852302849293,
0.049205612391233444,
-0.0025539591442793608,
-0.07171407341957092,
0.017258085310459137,
-0.07191593199968338,
0.018688343465328217,
-0.04400019347667694,
0.06037735193967819,
-0.004826208110898733,
-0.06928117573261261,
0.12397114187479019,
0.10822641104459763,
-0.07866884768009186,
0.04999454319477081,
-0.05606994405388832,
0.0013639186508953571,
-0.07418522983789444,
0.010137425735592842,
-0.16639631986618042,
-0.1297372728586197,
-0.005881429649889469,
0.06490922719240189,
0.036399081349372864,
0.024092858657240868,
0.04026643559336662,
0.033893898129463196,
-0.02715124376118183,
0.02220752462744713,
0.07535355538129807,
0.005847143474966288,
-0.1255677342414856,
-0.14016848802566528,
0.006038272753357887,
-0.04655002802610397,
0.029511770233511925,
-0.16531531512737274,
0.026899924501776695,
-0.06838905811309814,
0.021138019859790802,
0.02046659216284752,
0.005648517049849033,
0.02140159346163273,
-0.02923686057329178,
-0.010752207599580288,
-0.044050201773643494,
0.037764228880405426,
0.04885447397828102,
-0.19661198556423187,
0.12035247683525085,
-0.22761034965515137,
-0.023622116073966026,
0.05322514474391937,
0.02050717920064926,
-0.0701138973236084,
-0.0702095702290535,
-0.010348614305257797,
0.0009153799619525671,
-0.04201045259833336,
-0.05152292549610138,
0.1495082676410675,
0.03186678886413574,
0.15524451434612274,
-0.09727029502391815,
-0.010252077132463455,
-0.04999164491891861,
-0.0498465858399868,
-0.03740870580077171,
0.10975142568349838,
-0.07205715030431747,
-0.17507533729076385,
0.07561317831277847,
0.07375264167785645,
-0.09641891717910767,
0.11176919937133789,
0.019398240372538567,
-0.05809434875845909,
-0.02235182374715805,
0.05720909312367439,
0.017595453187823296,
0.03261102735996246,
-0.049616776406764984,
0.017191758379340172,
0.029085686430335045,
0.01769973151385784,
0.025972440838813782,
-0.05747580528259277,
0.021972673013806343,
0.07119246572256088,
-0.04286058992147446,
0.022743437439203262,
0.004654876422137022,
-0.003077416680753231,
0.07716170698404312,
0.016639797016978264,
-0.0010754402028396726,
-0.010096470825374126,
-0.042387768626213074,
-0.126202791929245,
0.20597243309020996,
-0.11683788895606995,
-0.16104447841644287,
-0.12195777893066406,
0.0068493373692035675,
-0.057177767157554626,
0.008231204003095627,
0.09089473634958267,
-0.0564110092818737,
-0.045681849122047424,
-0.10715397447347641,
0.03978154435753822,
0.0317499116063118,
-0.03548946976661682,
0.026814492419362068,
0.012131939642131329,
0.00550021743401885,
-0.12239633500576019,
0.006212232168763876,
-0.008845223113894463,
-0.03133222088217735,
-0.04079502075910568,
-0.0803135484457016,
0.02161019667983055,
0.06620130687952042,
0.03211598098278046,
-0.0011326444800943136,
-0.033737536519765854,
0.23943942785263062,
-0.05143297091126442,
0.046113621443510056,
0.16924765706062317,
0.000399350916268304,
0.06122663989663124,
0.1375432163476944,
0.009400129318237305,
-0.058853235095739365,
0.04722745716571808,
0.06002776324748993,
0.003156871534883976,
-0.13770639896392822,
-0.07350191473960876,
-0.10172960162162781,
-0.05524732172489166,
0.09045298397541046,
0.05489380657672882,
-0.045258428901433945,
0.09323671460151672,
-0.04070817679166794,
0.028163261711597443,
0.07627132534980774,
0.10624733567237854,
0.09333523362874985,
0.00388745847158134,
0.08302570879459381,
-0.05630655586719513,
-0.056999947875738144,
0.06909400969743729,
0.030077271163463593,
0.14524197578430176,
-0.028237104415893555,
0.12913838028907776,
0.0637679472565651,
-0.0037953490391373634,
-0.01807357184588909,
0.0650947317481041,
-0.07427164912223816,
0.010647881776094437,
-0.03308501094579697,
-0.07958134263753891,
-0.04001336917281151,
0.08144406974315643,
0.0417899489402771,
0.00005990039062453434,
-0.07445827126502991,
0.08705201745033264,
0.1318809986114502,
0.14882203936576843,
0.08786845207214355,
-0.26574888825416565,
-0.10332314670085907,
0.057363297790288925,
-0.08350341767072678,
-0.0681571438908577,
0.004679826553910971,
0.11193018406629562,
-0.08287668228149414,
0.03746781125664711,
0.0053994422778487206,
0.11232438683509827,
-0.02317064441740513,
0.01427708100527525,
-0.0870620384812355,
0.039914730936288834,
-0.026018083095550537,
0.07749043405056,
-0.20761291682720184,
0.13806955516338348,
0.0470135398209095,
0.07008707523345947,
-0.037168145179748535,
0.03196724131703377,
0.11945268511772156,
0.0933060497045517,
0.19670264422893524,
-0.020004134625196457,
-0.008394932374358177,
-0.012409022077918053,
-0.06416413933038712,
0.045163121074438095,
0.029313206672668457,
-0.04634098336100578,
0.07669719308614731,
-0.0474480576813221,
0.020923055708408356,
0.02523592859506607,
0.06636190414428711,
-0.059085652232170105,
-0.20565329492092133,
-0.0245138481259346,
0.10033620893955231,
-0.025250930339097977,
-0.012224695645272732,
-0.024385537952184677,
0.007021080236881971,
0.21422018110752106,
-0.0535479374229908,
-0.11879360675811768,
-0.11141247302293777,
0.0350775271654129,
0.07322397828102112,
-0.10866239666938782,
-0.012327787466347218,
-0.022444074973464012,
0.11964421719312668,
-0.05762459710240364,
-0.06686766445636749,
0.047489579766988754,
-0.08357426524162292,
0.022746333852410316,
0.0006667966372333467,
0.046805284917354584,
0.00938023254275322,
0.02430840954184532,
0.06375723332166672,
-0.01228557713329792,
-0.04064187780022621,
-0.09959472715854645,
-0.12167283892631531,
0.0638076588511467,
0.014184092171490192,
0.09095336496829987,
-0.17475281655788422,
-0.015955526381731033,
-0.0779670923948288,
0.02497326023876667,
0.2113547921180725,
0.24671095609664917,
-0.0561644583940506,
0.07527519017457962,
0.19078022241592407,
-0.10395088791847229,
-0.22638720273971558,
-0.11315683275461197,
0.008478901349008083,
0.06513158231973648,
0.07125958055257797,
-0.06611062586307526,
0.04013596475124359,
0.03732818365097046,
0.010554141364991665,
-0.08105781674385071,
-0.2633799910545349,
-0.11756733059883118,
0.15788567066192627,
0.05714668706059456,
0.10657621175050735,
-0.15284337103366852,
-0.0374307706952095,
-0.11042586714029312,
0.01160395797342062,
0.07762804627418518,
-0.06480415165424347,
0.13109736144542694,
0.0375506691634655,
0.005660345312207937,
0.03780905529856682,
0.007910449057817459,
0.1743420511484146,
0.03227368742227554,
0.06891161948442459,
-0.0419318750500679,
-0.015980588272213936,
0.055219538509845734,
-0.09985746443271637,
0.13356003165245056,
-0.11681437492370605,
0.06865439563989639,
-0.15511301159858704,
-0.0412968210875988,
-0.03585217520594597,
0.005535813048481941,
-0.023231294006109238,
-0.03846372291445732,
-0.04117165133357048,
0.047040898352861404,
0.1620558649301529,
0.004182173404842615,
0.05846180394291878,
-0.08532880991697311,
0.07520522177219391,
0.09742173552513123,
0.12421667575836182,
0.027061909437179565,
-0.11270944029092789,
0.06208633631467819,
0.003625876270234585,
0.11026716232299805,
-0.1844533383846283,
0.08015753328800201,
0.06198425963521004,
-0.02976509928703308,
0.13578848540782928,
0.03396047279238701,
-0.01472314540296793,
0.013850603252649307,
0.061543211340904236,
-0.06859269738197327,
-0.14941155910491943,
-0.029880305752158165,
-0.03611136972904205,
-0.12269976735115051,
-0.07836893200874329,
0.1566050797700882,
-0.044047072529792786,
0.020876863971352577,
0.008024671114981174,
0.047767817974090576,
-0.05126107484102249,
0.11822891235351562,
0.02888517826795578,
0.040584497153759,
-0.03979392349720001,
0.11299410462379456,
0.05199385806918144,
-0.07810273766517639,
0.05514223501086235,
0.10830776393413544,
-0.1021203026175499,
-0.06154404208064079,
-0.040876761078834534,
0.07783166319131851,
-0.11770526319742203,
-0.03863515332341194,
-0.0888117328286171,
-0.06114191561937332,
-0.027658313512802124,
0.040536247193813324,
0.03568519279360771,
0.04938892275094986,
-0.06156456470489502,
-0.021192001178860664,
-0.09149666875600815,
0.07743945717811584,
0.07821986079216003,
-0.006020017433911562,
-0.04908966273069382,
0.07957100868225098,
-0.03436259925365448,
0.09775439649820328,
-0.026934700086712837,
0.01622694917023182,
-0.09112810343503952,
0.002425167243927717,
-0.07946672290563583,
0.054280586540699005,
-0.1460900902748108,
0.004143482074141502,
0.028065767139196396,
0.059766028076410294,
-0.05891632288694382,
-0.00213633943349123,
-0.038717009127140045,
-0.03617765009403229,
-0.025437884032726288,
0.08499198406934738,
-0.11768448352813721,
-0.04552681744098663,
0.007112511433660984,
-0.07469736784696579,
0.09751097857952118,
0.017589980736374855,
-0.048516396433115005,
0.017057184129953384,
-0.08567554503679276,
-0.025085318833589554,
0.039252638816833496,
0.0397222638130188,
0.034432366490364075,
-0.10451937466859818,
0.040018822997808456,
-0.00981513038277626,
0.0289300549775362,
-0.01784113235771656,
0.060996055603027344,
-0.0733877643942833,
0.002088902285322547,
-0.07208773493766785,
0.01603752002120018,
-0.08140584826469421,
0.011786092072725296,
0.008208679035305977,
0.028856923803687096,
0.13072921335697174,
-0.07446545362472534,
0.04074866697192192,
-0.09234638512134552,
0.011254019103944302,
0.015985187143087387,
-0.05864373967051506,
0.053446702659130096,
-0.09631726145744324,
0.07088583707809448,
-0.05914117768406868,
0.12355135381221771,
-0.05665584281086922,
0.004381737671792507,
0.06374664604663849,
-0.052999064326286316,
0.101753830909729,
-0.029151853173971176,
0.11157745867967606,
0.033554673194885254,
-0.020780518651008606,
-0.02835126407444477,
0.012003212235867977,
0.07006344944238663,
0.03739508241415024,
0.08108983188867569,
0.13443253934383392,
0.0375523716211319,
0.14477714896202087,
0.042747292667627335,
-0.028054198250174522,
0.057015225291252136,
-0.010739337652921677,
-0.006328025832772255,
0.014745712280273438,
0.003768371185287833,
-0.07226821780204773,
0.2491101622581482,
-0.13395482301712036,
0.11798841506242752,
0.001829849323257804,
-0.08704514056444168,
-0.14278577268123627,
-0.07117234170436859,
-0.06973136961460114,
-0.051301512867212296,
-0.016275988891720772,
-0.15579833090305328,
-0.03113957867026329,
0.07448036968708038,
0.02205541729927063,
0.022297201678156853,
0.14697295427322388,
-0.0889429897069931,
-0.10561913251876831,
0.06817086040973663,
-0.05080126225948334,
0.08753214031457901,
0.045167211443185806,
-0.003103698370978236,
0.056500937789678574,
0.020248230546712875,
0.036659639328718185,
0.03710631653666496,
0.09581172466278076,
0.07862724363803864,
-0.08591166138648987,
-0.05285941809415817,
-0.031052924692630768,
0.01762058585882187,
-0.001996745588257909,
0.11612873524427414,
0.07596610486507416,
-0.07450401782989502,
-0.008599301800131798,
0.20929989218711853,
-0.07351347804069519,
-0.12991875410079956,
-0.21921402215957642,
0.09849566221237183,
0.041634753346443176,
0.023540548980236053,
-0.03424450755119324,
-0.08853548765182495,
-0.04133236035704613,
0.17667627334594727,
0.21071098744869232,
-0.16710463166236877,
0.027506502345204353,
-0.006118424236774445,
0.019862189888954163,
0.026459120213985443,
0.02911669947206974,
0.049367185682058334,
0.17219789326190948,
-0.05317264422774315,
0.036375585943460464,
-0.015628846362233162,
0.012455311603844166,
-0.08987195789813995,
0.16384081542491913,
0.04721735045313835,
0.01903955079615116,
-0.00943874940276146,
0.09346143156290054,
-0.0373983271420002,
-0.06925556808710098,
-0.016174329444766045,
-0.06156914681196213,
-0.10783498734235764,
-0.04400493949651718,
-0.02138960361480713,
0.03428536280989647,
0.06647747755050659,
0.008989970199763775,
-0.03050859086215496,
0.07279065996408463,
-0.0011258699232712388,
-0.09702882170677185,
-0.07681328803300858,
0.06706254929304123,
0.044718265533447266,
0.14761637151241302,
0.007814202457666397,
-0.008110706694424152,
0.10494940727949142,
-0.035284653306007385,
-0.029904667288064957,
0.11892463266849518,
0.05827192962169647,
-0.027487434446811676,
0.123702272772789,
0.05267445370554924,
-0.034172143787145615,
0.09732165187597275,
0.06768006831407547,
-0.1070907935500145,
0.040561363101005554,
0.009779978543519974,
-0.06476156413555145,
-0.06310280412435532,
0.005164246540516615,
-0.07864923030138016,
0.11881201714277267,
0.16110356152057648,
-0.025766916573047638,
0.006837749853730202,
0.009690909646451473,
0.01851848140358925,
0.04548819363117218,
0.011736933141946793,
-0.029904697090387344,
-0.10962698608636856,
0.015643132850527763,
-0.024648461490869522,
0.030125005170702934,
-0.32174333930015564,
-0.08512712270021439,
-0.008064293302595615,
-0.025530239567160606,
-0.03364702686667442,
0.1154986321926117,
0.10544726997613907,
0.006881712470203638,
-0.04951922222971916,
-0.2665429413318634,
0.01614590920507908,
0.0909474566578865,
-0.07430253177881241,
-0.1393526941537857
] |
null | null | transformers | Finetune of miqu-70b-sf dequant of miqudev's leak of Mistral-70B (allegedly an early mistral medium). My diffs are available under CC-0, this is a merge with the leaked model, you can use the other repository to save bandwidth.
EQ-Bench: 84.89
Will run more benches later. | {"license": "cc-by-2.0"} | text-generation | LoneStriker/Senku-70B-Full-6.0bpw-h6-exl2 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"license:cc-by-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T15:57:05+00:00 | [] | [] | TAGS
#transformers #safetensors #llama #text-generation #conversational #license-cc-by-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| Finetune of miqu-70b-sf dequant of miqudev's leak of Mistral-70B (allegedly an early mistral medium). My diffs are available under CC-0, this is a merge with the leaked model, you can use the other repository to save bandwidth.
EQ-Bench: 84.89
Will run more benches later. | [] | [
"TAGS\n#transformers #safetensors #llama #text-generation #conversational #license-cc-by-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
60
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #conversational #license-cc-by-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
-0.006207775790244341,
0.05186443030834198,
-0.005766681861132383,
0.0034429291263222694,
0.09202545881271362,
-0.026229875162243843,
0.21461424231529236,
0.09413789957761765,
-0.012125303968787193,
-0.027240416035056114,
0.15366454422473907,
0.20107468962669373,
-0.030660197138786316,
0.0625392273068428,
-0.12453678250312805,
-0.15658724308013916,
0.07951316237449646,
-0.0074092973954975605,
0.0468166321516037,
0.08121465146541595,
0.11726208031177521,
-0.05624562129378319,
0.07327274978160858,
-0.05116863548755646,
-0.10777227580547333,
0.01719590090215206,
0.07997997105121613,
-0.1362335979938507,
0.10389529168605804,
0.070436492562294,
0.09596491605043411,
0.09478230774402618,
-0.026663783937692642,
-0.2320748120546341,
0.024640627205371857,
-0.013039902783930302,
-0.08545025438070297,
0.02922174707055092,
0.03788268566131592,
-0.042137615382671356,
0.05930780619382858,
0.04187946766614914,
-0.012704534456133842,
0.0843532532453537,
-0.12006022036075592,
0.03492521494626999,
-0.054476991295814514,
0.008457338437438011,
0.11366608738899231,
0.07924187183380127,
0.002242166781798005,
0.1153382733464241,
-0.05578871816396713,
0.08845382928848267,
0.06245912238955498,
-0.3711334466934204,
0.014689184725284576,
0.13522937893867493,
0.07748636603355408,
0.06413708627223969,
-0.05252519249916077,
0.11249673366546631,
0.06629528105258942,
-0.03810290992259979,
0.04312480613589287,
-0.07572442293167114,
-0.08102772384881973,
0.034424908459186554,
-0.04833368957042694,
-0.030284471809864044,
0.23118719458580017,
-0.03052249550819397,
0.012902313843369484,
-0.08178368955850601,
-0.06274984031915665,
0.003528473898768425,
-0.027013633400201797,
0.033650096505880356,
-0.007114062085747719,
0.09828754514455795,
0.0036289477720856667,
-0.04431737959384918,
-0.15287499129772186,
-0.017103472724556923,
-0.17628958821296692,
0.10724018514156342,
-0.0038528726436197758,
0.03643316775560379,
-0.1433914750814438,
0.03731580823659897,
0.01849939487874508,
-0.09380249679088593,
-0.01922903209924698,
-0.06579022109508514,
0.07399444282054901,
-0.015354927629232407,
-0.04418032616376877,
-0.04420563578605652,
0.146459698677063,
0.11814041435718536,
-0.01364491693675518,
0.003440734464675188,
-0.11379846185445786,
0.0978090688586235,
-0.019500313326716423,
-0.0008985437452793121,
0.047480762004852295,
-0.007553805597126484,
0.1015886589884758,
-0.08041958510875702,
0.08773300796747208,
-0.04154345020651817,
-0.16899511218070984,
0.014984861016273499,
-0.009690387174487114,
0.14272445440292358,
0.0067246644757688046,
0.08934520930051804,
-0.03848949074745178,
0.06400565803050995,
0.07495175302028656,
-0.06882821768522263,
0.002561564091593027,
0.008464166894555092,
0.05782800912857056,
0.020744536072015762,
0.0357230082154274,
0.05749142915010452,
-0.053222328424453735,
0.034963082522153854,
-0.05974286422133446,
-0.030262604355812073,
-0.04071192815899849,
-0.05133391171693802,
0.07166485488414764,
-0.026888888329267502,
0.023412950336933136,
-0.1889147162437439,
-0.17520999908447266,
0.017266716808080673,
0.006569101009517908,
-0.016733024269342422,
-0.0409536249935627,
-0.05526142567396164,
-0.029290098696947098,
0.025005020201206207,
-0.08925633877515793,
-0.08324684202671051,
-0.08618094772100449,
0.08566243946552277,
-0.034463070333004,
0.052008256316185,
-0.17173472046852112,
0.030293233692646027,
-0.09709536284208298,
0.015842683613300323,
-0.024641036987304688,
0.05080806463956833,
-0.03994987532496452,
0.1493525505065918,
-0.03130904585123062,
0.028674336150288582,
-0.049635149538517,
0.06748075783252716,
-0.04132431373000145,
0.20219798386096954,
-0.13369248807430267,
-0.02732842043042183,
0.23248150944709778,
-0.1203412264585495,
-0.2215833067893982,
0.09494956582784653,
-0.007244576700031757,
0.06883425265550613,
0.1192723736166954,
0.16428819298744202,
-0.018312968313694,
-0.084824338555336,
0.043344683945178986,
0.10571722686290741,
-0.05737517401576042,
-0.10720136761665344,
0.012378843501210213,
-0.02602921798825264,
-0.10123564302921295,
0.021436169743537903,
0.054138898849487305,
0.03870760649442673,
-0.010303346440196037,
-0.06431794166564941,
-0.04900350421667099,
-0.04556714743375778,
-0.019114533439278603,
-0.05745845288038254,
0.04891825467348099,
-0.10103435069322586,
-0.004564722999930382,
0.02500876598060131,
-0.0006331629119813442,
-0.017695071175694466,
0.03802654147148132,
-0.1019439622759819,
0.05491343140602112,
-0.01049356535077095,
0.05139922350645065,
-0.08506914973258972,
-0.09540282934904099,
-0.01487639732658863,
0.11419088393449783,
0.04237206280231476,
0.03597130998969078,
0.03009572997689247,
-0.0029274821281433105,
-0.016321944072842598,
0.019669566303491592,
0.1798858940601349,
0.027921929955482483,
-0.06157265976071358,
-0.10305018723011017,
0.10064617544412613,
-0.04469102993607521,
0.06838524341583252,
-0.12369509041309357,
0.021130846813321114,
0.09049873054027557,
0.08164099603891373,
0.011007205583155155,
0.06817352026700974,
-0.016652852296829224,
0.012067300267517567,
-0.08803333342075348,
0.018720898777246475,
0.09845291078090668,
0.023675929754972458,
-0.12399162352085114,
0.24828600883483887,
-0.22121906280517578,
0.23131431639194489,
0.20286236703395844,
-0.2072221040725708,
0.034866541624069214,
-0.11290061473846436,
0.008944046683609486,
0.009122604504227638,
0.00658374885097146,
-0.04158162698149681,
0.02772163227200508,
-0.017428645864129066,
0.17939260601997375,
-0.07259617745876312,
-0.01289988774806261,
-0.015334650874137878,
-0.07770753651857376,
-0.0461907684803009,
0.04808717593550682,
0.10923753678798676,
-0.17211377620697021,
0.18076997995376587,
0.2573873698711395,
0.016237052157521248,
0.145608589053154,
-0.02829170413315296,
0.011588740162551403,
0.031093206256628036,
0.051330193877220154,
0.017952319234609604,
-0.03206970542669296,
-0.09656007587909698,
-0.01615116000175476,
0.06039080768823624,
0.00738911610096693,
0.05942045524716377,
-0.14498206973075867,
-0.06179048866033554,
-0.010635633021593094,
-0.054405391216278076,
0.006760948803275824,
0.05542760342359543,
-0.006918448954820633,
0.12766680121421814,
-0.04561823606491089,
-0.06787645071744919,
0.1212611123919487,
-0.018814176321029663,
-0.11172524094581604,
0.16348372399806976,
-0.1353582739830017,
-0.2530854046344757,
-0.1794736236333847,
-0.16009610891342163,
-0.05283679813146591,
0.06249964237213135,
0.12733663618564606,
-0.021347085013985634,
-0.06735935807228088,
-0.08513124287128448,
-0.025516794994473457,
-0.004286123439669609,
-0.0013990295119583607,
-0.030478790402412415,
0.0740688145160675,
-0.04530498385429382,
-0.11495137214660645,
-0.04566733539104462,
0.02995317615568638,
-0.08298885077238083,
0.11527245491743088,
-0.08257944881916046,
0.09218067675828934,
0.14615941047668457,
0.013489138334989548,
-0.0007384112104773521,
-0.06414268910884857,
0.11027345061302185,
-0.05689837411046028,
-0.021911391988396645,
0.19550743699073792,
-0.06030816212296486,
0.054297950118780136,
0.17741087079048157,
0.019415782764554024,
-0.11608056724071503,
0.05627991259098053,
-0.04948503151535988,
-0.08943437039852142,
-0.2322319746017456,
-0.11231391131877899,
-0.09151040762662888,
0.09372707456350327,
0.02010270394384861,
0.06644773483276367,
0.1561785638332367,
0.07071852684020996,
-0.03820497542619705,
-0.003707042895257473,
0.09500505030155182,
0.10332844406366348,
0.26925957202911377,
-0.03648235276341438,
0.1326010376214981,
-0.09864376485347748,
-0.09166394174098969,
0.08177606761455536,
0.09666159003973007,
0.08463814854621887,
0.1308884173631668,
0.10701965540647507,
0.06300177425146103,
0.06616055965423584,
0.13583797216415405,
0.0930822342634201,
0.04747668653726578,
-0.027060847729444504,
-0.021886298432946205,
-0.06314191222190857,
-0.03154398128390312,
0.06433441489934921,
-0.0616009458899498,
-0.1419346034526825,
-0.01669052243232727,
-0.04656673222780228,
0.08677151799201965,
0.12965995073318481,
0.049826931208372116,
-0.20373210310935974,
0.03404708579182625,
0.12731851637363434,
-0.014891417697072029,
-0.0784522220492363,
0.11266084015369415,
0.041655853390693665,
-0.0492335706949234,
0.0968889519572258,
-0.022453399375081062,
0.10560388118028641,
-0.04346970468759537,
0.06621389091014862,
-0.09852857887744904,
-0.09218709915876389,
0.008176117204129696,
0.10175062716007233,
-0.32708150148391724,
0.19553345441818237,
0.025147762149572372,
0.003030276857316494,
-0.07397007197141647,
-0.00860549882054329,
0.006451844237744808,
0.1687123030424118,
0.14335986971855164,
-0.03559787943959236,
-0.11925935745239258,
-0.036741066724061966,
-0.04582417383790016,
0.021652963012456894,
0.114947110414505,
0.004123690538108349,
0.006702055223286152,
-0.060284681618213654,
-0.005426459014415741,
0.013130738399922848,
-0.03866162151098251,
-0.05004815012216568,
-0.18432049453258514,
0.036161795258522034,
0.15424391627311707,
0.105919748544693,
-0.038923926651477814,
0.025927383452653885,
-0.1276562660932541,
0.19439037144184113,
-0.1566786915063858,
-0.05010409653186798,
-0.10357701778411865,
-0.13866664469242096,
-0.010155638679862022,
-0.015465851873159409,
0.06662318855524063,
-0.05540037900209427,
0.05461542308330536,
-0.0985378548502922,
-0.16780197620391846,
0.11566402018070221,
-0.10529130697250366,
-0.03816051036119461,
-0.04050680622458458,
0.15799902379512787,
-0.10722509026527405,
-0.015996212139725685,
0.0695413127541542,
0.027312718331813812,
-0.04470572620630264,
-0.10069151222705841,
-0.01407475396990776,
0.022582462057471275,
0.03730671480298042,
-0.004337325692176819,
-0.1392161101102829,
-0.12033824622631073,
-0.007708914577960968,
-0.08476556092500687,
0.26667433977127075,
0.2823437750339508,
-0.052704282104969025,
0.1610942929983139,
0.18729859590530396,
-0.10833743214607239,
-0.3613292872905731,
-0.1156543493270874,
-0.19261738657951355,
-0.05812794342637062,
-0.0019929534755647182,
-0.09389892220497131,
0.0683603510260582,
0.03696022555232048,
-0.06046358868479729,
0.10693113505840302,
-0.20837023854255676,
-0.11674803495407104,
0.13628779351711273,
0.022366341203451157,
0.3236163854598999,
-0.1799478828907013,
-0.10869350284337997,
-0.1288607120513916,
-0.09091421216726303,
0.17097102105617523,
-0.11236412078142166,
0.09014905989170074,
0.038183145225048065,
0.05064551532268524,
0.019322404637932777,
-0.028820998966693878,
0.11060543358325958,
-0.05023352429270744,
0.05936712771654129,
-0.125066876411438,
-0.007887703366577625,
0.055251557379961014,
-0.021825825795531273,
0.047600388526916504,
-0.17949266731739044,
0.010565537959337234,
-0.02346070110797882,
-0.03863802179694176,
-0.00010827439837157726,
0.07584916800260544,
-0.0018114936538040638,
-0.05403559282422066,
-0.0338921919465065,
-0.07518398016691208,
0.021392593160271645,
-0.005685644224286079,
0.2621508240699768,
-0.09497393667697906,
0.143461674451828,
0.2097931206226349,
0.18348653614521027,
-0.11326921731233597,
0.10850489139556885,
-0.03179696202278137,
-0.10264644771814346,
0.06764788925647736,
-0.13615255057811737,
0.06894791126251221,
0.07885490357875824,
-0.05221731215715408,
0.09386573731899261,
0.07263122498989105,
0.024896983057260513,
0.011904848739504814,
0.15288308262825012,
-0.19621852040290833,
-0.06335753947496414,
-0.02627270668745041,
0.128641277551651,
0.07603052258491516,
0.07438892126083374,
0.19964289665222168,
-0.02363588474690914,
0.024383772164583206,
0.004766607191413641,
0.0470099002122879,
-0.051470812410116196,
0.03084268979728222,
-0.0036850213073194027,
0.009037865325808525,
-0.11732205748558044,
0.11256258189678192,
0.014299282804131508,
-0.13298094272613525,
0.018583929166197777,
0.10019229352474213,
-0.12479248642921448,
-0.1322273164987564,
-0.06397988647222519,
0.09235106408596039,
-0.1650686413049698,
-0.09092073142528534,
-0.035492803901433945,
-0.18309658765792847,
0.023424223065376282,
0.22705577313899994,
0.03624091297388077,
0.09397074580192566,
0.025309177115559578,
-0.050275810062885284,
-0.03333280235528946,
0.044489964842796326,
-0.0776023343205452,
0.026683444157242775,
-0.09567009657621384,
0.016401100903749466,
-0.0327458456158638,
0.04273902624845505,
-0.08503364771604538,
-0.01675247587263584,
-0.13592562079429626,
0.026447130367159843,
-0.1429377794265747,
0.016726532950997353,
-0.09382633119821548,
-0.022313788533210754,
0.025396376848220825,
-0.011300059035420418,
-0.044356830418109894,
-0.0456419438123703,
-0.08871109038591385,
0.022159568965435028,
-0.04314287006855011,
0.0626419335603714,
-0.09819427877664566,
-0.043954771012067795,
0.04525887221097946,
-0.036994993686676025,
0.11698898673057556,
0.04370922967791557,
-0.0972345620393753,
0.07480837404727936,
-0.24783827364444733,
-0.02655193954706192,
0.13125064969062805,
0.015579255297780037,
0.014812005683779716,
0.06315375864505768,
-0.011836273595690727,
0.1417415887117386,
0.009809926152229309,
0.052264366298913956,
0.013304419815540314,
-0.1011492908000946,
-0.001961590489372611,
-0.04560624808073044,
-0.10192413628101349,
-0.018009837716817856,
-0.0727795958518982,
0.1012018471956253,
-0.024938881397247314,
0.17435556650161743,
-0.092584989964962,
0.03382148593664169,
-0.01980471797287464,
0.03883249685168266,
0.005250109359622002,
-0.15618106722831726,
-0.14053024351596832,
-0.07882187515497208,
-0.001336804125458002,
-0.015131816267967224,
0.27824896574020386,
0.008481241762638092,
-0.06852035224437714,
0.09475342184305191,
0.022017354145646095,
0.034207794815301895,
0.04192714765667915,
0.2883480489253998,
0.08536436408758163,
-0.01155257411301136,
-0.15340177714824677,
0.016927950084209442,
0.0375761017203331,
-0.09943119436502457,
0.06783804297447205,
0.09215667098760605,
-0.07752043753862381,
0.10896968841552734,
0.07114557921886444,
0.0072129033505916595,
-0.018362488597631454,
-0.06760571897029877,
-0.055142998695373535,
0.046072885394096375,
-0.049081407487392426,
0.07889051735401154,
0.20583051443099976,
-0.03462229669094086,
-0.015487121418118477,
-0.04577798396348953,
-0.03612710162997246,
-0.18539372086524963,
-0.13403363525867462,
-0.1109817773103714,
-0.11150851845741272,
0.01701277866959572,
-0.08830517530441284,
0.055540017783641815,
0.05017261207103729,
0.054319605231285095,
-0.04291510581970215,
0.06990589201450348,
-0.05907499045133591,
-0.04540880024433136,
0.0442875400185585,
-0.02594553679227829,
0.06306269764900208,
-0.035188764333724976,
-0.06933852285146713,
-0.053066983819007874,
-0.04323925822973251,
-0.02944331243634224,
0.08734598755836487,
0.050012290477752686,
0.06000626087188721,
-0.13524575531482697,
-0.06855671852827072,
-0.032354552298784256,
0.0697886273264885,
-0.008020002394914627,
0.1599365472793579,
0.022465955466032028,
-0.0408162996172905,
0.08315412700176239,
0.1788318008184433,
-0.06562280654907227,
-0.12326370179653168,
-0.031882673501968384,
0.1764252483844757,
0.0044775111600756645,
0.13507580757141113,
-0.040701255202293396,
-0.013290504924952984,
-0.012223204597830772,
0.3287609815597534,
0.25894030928611755,
-0.09469550848007202,
0.03137436881661415,
-0.07478570938110352,
0.04246385395526886,
0.05169496685266495,
0.11560044437646866,
0.0689227506518364,
0.25144103169441223,
-0.03857225552201271,
-0.03432202711701393,
-0.00518826674669981,
0.017930803820490837,
-0.13905346393585205,
0.0863647311925888,
-0.04497753456234932,
-0.05007867515087128,
-0.02601996622979641,
0.09876758605241776,
-0.15639275312423706,
0.06847500801086426,
-0.04330084100365639,
-0.09984272718429565,
0.007430274039506912,
0.007269448135048151,
0.14150775969028473,
-0.019860094413161278,
0.029579713940620422,
-0.03768281638622284,
-0.07121695578098297,
0.01726885512471199,
-0.005519684869796038,
-0.18614327907562256,
0.03858964890241623,
0.023371171206235886,
-0.005911254324018955,
0.07187901437282562,
0.000053612515330314636,
0.06824749708175659,
0.07142221182584763,
0.030734332278370857,
-0.051403749734163284,
0.15974754095077515,
0.026775244623422623,
-0.0868566483259201,
0.050569936633110046,
-0.040603864938020706,
-0.026959538459777832,
0.04548517242074013,
0.06765088438987732,
-0.09669996052980423,
0.06760630011558533,
0.009751532226800919,
-0.11920495331287384,
-0.03518637642264366,
0.010957258753478527,
-0.07562252134084702,
0.07157810032367706,
0.007729232311248779,
-0.028513247147202492,
0.004583296831697226,
-0.016401249915361404,
0.010413174517452717,
-0.018589364364743233,
-0.15951070189476013,
-0.017387747764587402,
-0.12323169410228729,
-0.05553334206342697,
0.11709286272525787,
0.03462449461221695,
-0.23784784972667694,
0.011705098673701286,
-0.0969860851764679,
0.06600788980722427,
-0.19047285616397858,
0.05791420489549637,
0.2114742547273636,
-0.005059413146227598,
-0.031710248440504074,
-0.17269471287727356,
0.06433253735303879,
0.061398617923259735,
-0.036462731659412384,
-0.10163070261478424
] |
null | null | generic |
# Fork of [caidas/swin2SR-classical-sr-x2-64](https://huggingface.co/caidas/swin2SR-classical-sr-x2-64) for a `image-to-image` Inference endpoint.
> Inspired by https://huggingface.co/sergeipetrov/swin2SR-classical-sr-x2-64-IE
This repository implements a `custom` task for `image-to-image` for 🤗 Inference Endpoints to allow image up scaling by doubling image resolution.
The code for the customized pipeline is in the handler.py.
To use deploy this model an Inference Endpoint you have to select `Custom` as task to use the `handler.py` file.
### expected Request payload
Image to be labeled as binary.
#### CURL
```
curl URL \
-X POST \
--data-binary @car.png \
-H "Content-Type: image/png"
```
#### Python
```python
requests.post(ENDPOINT_URL, headers={"Content-Type": "image/png"}, data=open("car.png", 'rb').read()).json()
``` | {"library_name": "generic", "tags": ["vision", "image-to-image", "endpoints-template"], "inference": false, "pipeline_tag": "image-to-image", "base_model": "caidas/swin2SR-classical-sr-x2-64"} | image-to-image | pimcore/IEP__image-upscaling-2x | [
"generic",
"vision",
"image-to-image",
"endpoints-template",
"base_model:caidas/swin2SR-classical-sr-x2-64",
"endpoints_compatible",
"region:us"
] | 2024-02-07T15:57:08+00:00 | [] | [] | TAGS
#generic #vision #image-to-image #endpoints-template #base_model-caidas/swin2SR-classical-sr-x2-64 #endpoints_compatible #region-us
|
# Fork of caidas/swin2SR-classical-sr-x2-64 for a 'image-to-image' Inference endpoint.
> Inspired by URL
This repository implements a 'custom' task for 'image-to-image' for Inference Endpoints to allow image up scaling by doubling image resolution.
The code for the customized pipeline is in the URL.
To use deploy this model an Inference Endpoint you have to select 'Custom' as task to use the 'URL' file.
### expected Request payload
Image to be labeled as binary.
#### CURL
#### Python
| [
"# Fork of caidas/swin2SR-classical-sr-x2-64 for a 'image-to-image' Inference endpoint.\n\n> Inspired by URL\n\nThis repository implements a 'custom' task for 'image-to-image' for Inference Endpoints to allow image up scaling by doubling image resolution. \nThe code for the customized pipeline is in the URL.\n\nTo use deploy this model an Inference Endpoint you have to select 'Custom' as task to use the 'URL' file.",
"### expected Request payload\n\nImage to be labeled as binary.",
"#### CURL",
"#### Python"
] | [
"TAGS\n#generic #vision #image-to-image #endpoints-template #base_model-caidas/swin2SR-classical-sr-x2-64 #endpoints_compatible #region-us \n",
"# Fork of caidas/swin2SR-classical-sr-x2-64 for a 'image-to-image' Inference endpoint.\n\n> Inspired by URL\n\nThis repository implements a 'custom' task for 'image-to-image' for Inference Endpoints to allow image up scaling by doubling image resolution. \nThe code for the customized pipeline is in the URL.\n\nTo use deploy this model an Inference Endpoint you have to select 'Custom' as task to use the 'URL' file.",
"### expected Request payload\n\nImage to be labeled as binary.",
"#### CURL",
"#### Python"
] | [
54,
125,
16,
4,
3
] | [
"passage: TAGS\n#generic #vision #image-to-image #endpoints-template #base_model-caidas/swin2SR-classical-sr-x2-64 #endpoints_compatible #region-us \n# Fork of caidas/swin2SR-classical-sr-x2-64 for a 'image-to-image' Inference endpoint.\n\n> Inspired by URL\n\nThis repository implements a 'custom' task for 'image-to-image' for Inference Endpoints to allow image up scaling by doubling image resolution. \nThe code for the customized pipeline is in the URL.\n\nTo use deploy this model an Inference Endpoint you have to select 'Custom' as task to use the 'URL' file.### expected Request payload\n\nImage to be labeled as binary.#### CURL#### Python"
] | [
-0.06693913787603378,
-0.0041433884762227535,
-0.002260977402329445,
0.05963021144270897,
0.14556115865707397,
0.057247959077358246,
-0.008639280684292316,
0.052945733070373535,
0.10846153646707535,
0.03680424019694328,
0.14731264114379883,
0.1261463314294815,
0.03762045502662659,
0.01681155525147915,
-0.005207451526075602,
-0.2985731065273285,
0.02685234695672989,
0.03625442087650299,
0.04793144389986992,
0.03426756337285042,
0.006279828492552042,
-0.06672459840774536,
0.1807163506746292,
-0.029277892783284187,
-0.14545807242393494,
0.05145743861794472,
0.015864098444581032,
0.03899190574884415,
0.03292962536215782,
0.04556737840175629,
0.1029028668999672,
-0.003298266092315316,
0.10438528656959534,
-0.09349647909402847,
0.03121168166399002,
0.03445516154170036,
-0.04110713303089142,
0.024379456415772438,
0.055163294076919556,
0.004378410521894693,
0.07711748778820038,
-0.0305133406072855,
0.024241752922534943,
0.011015172116458416,
-0.05739142373204231,
0.009528876282274723,
0.0594014897942543,
0.09361255913972855,
0.18472397327423096,
0.007647174876183271,
0.08447608351707458,
-0.10620005428791046,
0.006011030171066523,
0.09345840662717819,
0.07960338145494461,
-0.13513974845409393,
-0.04424747824668884,
0.1603814661502838,
-0.04754423350095749,
0.031793877482414246,
-0.03307407721877098,
0.016114918515086174,
-0.01175383385270834,
0.029856029897928238,
0.02740098349750042,
-0.10834236443042755,
-0.1492241621017456,
-0.021561404690146446,
-0.06342864036560059,
-0.11604137718677521,
0.1071401983499527,
0.008662175387144089,
-0.06037581339478493,
0.00936113204807043,
-0.13824957609176636,
-0.10032826662063599,
-0.13085664808750153,
0.048747468739748,
0.022918758913874626,
0.05612330883741379,
-0.008174669928848743,
-0.04250672459602356,
-0.11136432737112045,
-0.0671117752790451,
-0.0006915290141478181,
-0.02115369588136673,
0.0028223772533237934,
0.1444436013698578,
-0.07773404568433762,
0.13581311702728271,
-0.09725257754325867,
-0.06264530122280121,
-0.024442121386528015,
-0.07669703662395477,
-0.03152015060186386,
0.06604234874248505,
0.01266767643392086,
-0.11725953966379166,
-0.0022793421521782875,
0.13250188529491425,
0.16156965494155884,
0.08340267837047577,
0.05836851894855499,
0.0888928547501564,
0.005506973247975111,
0.0996212586760521,
-0.10367397964000702,
-0.004642967600375414,
0.0664835274219513,
-0.09529320895671844,
0.014863304793834686,
-0.04183289036154747,
-0.1374090313911438,
-0.07435712963342667,
-0.06332894414663315,
-0.02747287042438984,
0.01828615926206112,
0.09573677182197571,
-0.0031261856202036142,
-0.05203595012426376,
0.24794968962669373,
-0.0047577316872775555,
0.035786595195531845,
-0.01683235540986061,
0.023541653528809547,
0.13327623903751373,
0.10827029496431351,
-0.03187798708677292,
-0.062174491584300995,
0.15324866771697998,
-0.05768480524420738,
0.047337256371974945,
-0.10751466453075409,
0.009603613056242466,
-0.01692919060587883,
-0.03110639937222004,
0.02067909948527813,
-0.14435206353664398,
-0.17745891213417053,
-0.016055651009082794,
0.12967747449874878,
-0.004176954738795757,
0.05218569189310074,
0.026978157460689545,
-0.004869039170444012,
0.04121888801455498,
-0.015610828995704651,
-0.008760398253798485,
-0.05915975943207741,
0.022462336346507072,
-0.08020228147506714,
0.07458052039146423,
-0.18620727956295013,
-0.02873825468122959,
-0.056702882051467896,
0.060684140771627426,
-0.15471151471138,
0.07680435478687286,
0.028860753402113914,
0.09422949701547623,
-0.02241458185017109,
-0.04867991432547569,
-0.10082581639289856,
-0.0074436054565012455,
0.01371800061315298,
0.1328507363796234,
-0.13308560848236084,
0.07653623074293137,
0.146951362490654,
-0.08338724821805954,
-0.07237622886896133,
0.06475707143545151,
-0.015652133151888847,
-0.04588814079761505,
0.031910840421915054,
0.09196074306964874,
0.12210071086883545,
-0.185940682888031,
0.029567617923021317,
0.07695165276527405,
-0.13476483523845673,
-0.11110762506723404,
0.008790817111730576,
0.10412785410881042,
0.0389024056494236,
0.07428920269012451,
-0.11333362013101578,
0.020131587982177734,
-0.0016617674846202135,
-0.01868729665875435,
-0.05818462744355202,
0.0261545367538929,
-0.1635758876800537,
0.008153147995471954,
0.004808178171515465,
0.06311509013175964,
-0.010591190308332443,
-0.09451495110988617,
0.045937664806842804,
-0.09191064536571503,
0.06014040485024452,
-0.03631855919957161,
0.12009598314762115,
-0.12991341948509216,
-0.0011674187844619155,
-0.0885075181722641,
-0.017878476530313492,
-0.01878327690064907,
0.22000053524971008,
0.033642735332250595,
-0.07468436658382416,
0.012682858854532242,
-0.003513194853439927,
0.036091335117816925,
-0.040336258709430695,
0.09375958889722824,
-0.08042758703231812,
-0.05865868926048279,
0.009403515607118607,
-0.12358881533145905,
-0.01049888040870428,
-0.11110914498567581,
-0.14398513734340668,
0.0006742064724676311,
0.11264458298683167,
0.038896068930625916,
0.05601884424686432,
-0.02973034977912903,
-0.00849858671426773,
-0.04919428378343582,
-0.04477572441101074,
-0.04097268730401993,
0.06708076596260071,
0.030796820297837257,
0.09883569180965424,
0.013606867752969265,
-0.0157859455794096,
0.11412647366523743,
0.0648828074336052,
-0.2536700963973999,
-0.02291499637067318,
-0.031465012580156326,
-0.0017150196945294738,
0.026090048253536224,
0.009341070428490639,
-0.03196193277835846,
0.11608889698982239,
0.029611920937895775,
0.0963590145111084,
-0.03463258966803551,
0.0821504145860672,
0.04128376394510269,
-0.0774071142077446,
0.0166129469871521,
-0.08461975306272507,
0.13881820440292358,
-0.09880848228931427,
0.02753310278058052,
-0.05855260789394379,
-0.05651593953371048,
0.05225067585706711,
0.07713096588850021,
-0.07300319522619247,
0.06218515336513519,
0.0643322616815567,
0.012706167995929718,
0.12605810165405273,
-0.08088912069797516,
-0.07233750820159912,
0.04689270630478859,
-0.037650227546691895,
0.03271371126174927,
-0.13021115958690643,
0.039065103977918625,
0.012085186317563057,
0.0033481521531939507,
0.0375317819416523,
-0.038673292845487595,
-0.04518752545118332,
0.0591573528945446,
0.0008197541465051472,
0.001643747789785266,
-0.011099739000201225,
-0.05552441254258156,
-0.07947666198015213,
0.147177591919899,
0.026857856661081314,
-0.2156434804201126,
-0.18930280208587646,
-0.13820022344589233,
-0.04936189576983452,
0.0666986033320427,
0.03236426040530205,
-0.0013452486600726843,
-0.07264415174722672,
-0.030892308801412582,
-0.05530385673046112,
0.027874350547790527,
-0.006284653674811125,
-0.15364323556423187,
0.006970665883272886,
-0.019388608634471893,
-0.014200902543962002,
-0.023284809663891792,
-0.01795484870672226,
-0.012857697904109955,
0.11816821992397308,
-0.07415545731782913,
0.15361058712005615,
0.08010262250900269,
-0.00938017200678587,
0.03166640177369118,
0.04110465198755264,
0.14919069409370422,
-0.0402364581823349,
0.0764412134885788,
0.17002016305923462,
-0.03941834717988968,
0.08031053096055984,
0.14561709761619568,
0.005684259347617626,
-0.11538714170455933,
-0.01930968277156353,
-0.06174369156360626,
-0.126406729221344,
0.02134748362004757,
-0.08060798794031143,
-0.14176379144191742,
-0.0019844910129904747,
0.20333990454673767,
0.051972296088933945,
-0.0763695165514946,
0.19548705220222473,
-0.050350621342659,
0.13763783872127533,
-0.05008550360798836,
0.17879155278205872,
0.021287761628627777,
0.012257558293640614,
0.03295713663101196,
-0.0752897709608078,
-0.012006550095975399,
0.0751909390091896,
0.24256424605846405,
0.10788986086845398,
-0.04263107106089592,
0.04270975664258003,
0.009720551781356335,
0.1312185525894165,
0.08917851001024246,
0.1126234382390976,
-0.0610048808157444,
-0.011323051527142525,
-0.0015104861231520772,
-0.01811140403151512,
-0.10350779443979263,
0.09458165615797043,
-0.08089238405227661,
-0.09779275953769684,
0.04171326383948326,
0.01225131656974554,
0.0036657899618148804,
0.24242444336414337,
0.016924051567912102,
-0.3782619535923004,
-0.019347593188285828,
0.043516699224710464,
0.04461121931672096,
-0.15913861989974976,
0.03509293869137764,
0.06060692295432091,
-0.06961794942617416,
0.033946651965379715,
-0.07615305483341217,
0.13193368911743164,
-0.08428370952606201,
-0.0013134469045326114,
0.06270931661128998,
-0.04082690551877022,
0.06122850999236107,
0.039021704345941544,
-0.0355912446975708,
0.032355133444070816,
0.0003561267803888768,
-0.02248118817806244,
-0.0819798931479454,
0.02267719805240631,
-0.019733065739274025,
0.2085995227098465,
0.16030913591384888,
0.002423460828140378,
-0.04658833518624306,
-0.0698075070977211,
0.03528856858611107,
0.04958919435739517,
0.04837796464562416,
-0.0015457698609679937,
0.054512832313776016,
0.005304287187755108,
-0.04791666194796562,
-0.06070824712514877,
0.05831882730126381,
0.07712744921445847,
-0.21346387267112732,
-0.029310286045074463,
0.010252888314425945,
0.034404076635837555,
0.03720584884285927,
0.13401877880096436,
0.18558982014656067,
0.17589649558067322,
-0.09586449712514877,
-0.08395881950855255,
-0.13872382044792175,
0.03416645526885986,
-0.009914983995258808,
-0.05197136104106903,
0.060299161821603775,
-0.07792779058218002,
0.10304506123065948,
0.02002228982746601,
-0.16139452159404755,
0.09836923331022263,
-0.1238815039396286,
0.060958120971918106,
-0.08506745845079422,
0.02618529088795185,
-0.11842751502990723,
-0.07761989533901215,
0.014020400121808052,
0.02773369289934635,
-0.05808291956782341,
-0.1201038733124733,
-0.0314796045422554,
0.07088839262723923,
0.1403512805700302,
0.03841858357191086,
-0.053538739681243896,
-0.034228138625621796,
0.04889403656125069,
0.13734272122383118,
0.05045393481850624,
-0.013622811995446682,
-0.0859973281621933,
-0.015849251300096512,
0.012743399478495121,
-0.029582304880023003,
-0.3022732734680176,
-0.039335861802101135,
0.02086002007126808,
-0.01997295953333378,
-0.09545837342739105,
-0.09863337874412537,
0.16529981791973114,
0.004225769080221653,
-0.020692814141511917,
0.19486042857170105,
-0.17630095779895782,
-0.08051305264234543,
0.012031438760459423,
0.1334410160779953,
0.16655242443084717,
-0.1461065709590912,
-0.0061837113462388515,
-0.10554847866296768,
-0.23138979077339172,
0.07189277559518814,
0.10685449093580246,
0.03786536306142807,
-0.03760194405913353,
0.11567356437444687,
0.018216686323285103,
-0.05388805642724037,
0.044572312384843826,
-0.0033231363631784916,
0.1122170239686966,
-0.05326206237077713,
0.025702212005853653,
0.03932485729455948,
-0.06116383895277977,
0.10869092494249344,
0.0105521772056818,
0.06340159475803375,
-0.1468990594148636,
-0.0221844632178545,
-0.0029654938261955976,
-0.020929381251335144,
0.0572519488632679,
-0.05091764032840729,
-0.10663443058729172,
-0.08446939289569855,
0.08146309107542038,
0.06491707265377045,
0.14796267449855804,
-0.06794693320989609,
-0.1288355141878128,
0.14606134593486786,
-0.04938235506415367,
-0.12378761172294617,
-0.19789841771125793,
-0.05776523798704147,
-0.026875479146838188,
0.12432372570037842,
-0.1760370433330536,
0.0561564639210701,
0.04754989966750145,
0.014900045469403267,
0.0017520220717415214,
0.06936104595661163,
0.07410164922475815,
0.037382822483778,
0.07907035946846008,
-0.09007231891155243,
0.01681864634156227,
0.002194778760895133,
0.0823659747838974,
-0.008890632539987564,
0.013013429008424282,
0.08436139672994614,
-0.02532750554382801,
0.004836909472942352,
-0.012077322229743004,
0.034392233937978745,
-0.060790859162807465,
-0.018002847209572792,
0.03677808493375778,
-0.011751016601920128,
-0.07278196513652802,
0.007674895226955414,
-0.008142583072185516,
-0.11827313154935837,
-0.08027950674295425,
0.07949895411729813,
-0.13501349091529846,
-0.07757915556430817,
-0.07396538555622101,
0.01836106739938259,
-0.1938866376876831,
-0.01689557172358036,
0.02050480991601944,
0.018963279202580452,
0.012800164520740509,
-0.030743269249796867,
0.044037625193595886,
-0.02174190618097782,
-0.0234469436109066,
-0.007591721601784229,
-0.0757172629237175,
-0.030128730461001396,
-0.07658638060092926,
0.12842650711536407,
-0.18383902311325073,
-0.059120532125234604,
-0.0278671495616436,
0.0715116485953331,
-0.07274185866117477,
-0.01728520728647709,
-0.09634985774755478,
0.03127618879079819,
-0.10520416498184204,
0.17443710565567017,
-0.09446162730455399,
-0.005381101742386818,
0.006697372067719698,
0.018035240471363068,
-0.07712739706039429,
0.029582208022475243,
-0.06320430338382721,
-0.016603000462055206,
-0.03789582476019859,
0.0003835779498331249,
-0.008055565878748894,
0.01148257590830326,
0.025427725166082382,
-0.06064106523990631,
-0.00603861641138792,
0.041521117091178894,
-0.10271979868412018,
-0.013835802674293518,
-0.12657099962234497,
-0.07125051319599152,
0.1155666932463646,
0.06024521216750145,
-0.022431155666708946,
0.22552454471588135,
0.046649277210235596,
0.05034082382917404,
0.004749345127493143,
-0.08218506723642349,
0.11789004504680634,
-0.07168374210596085,
0.0554322674870491,
-0.11913074553012848,
-0.02606702409684658,
-0.05567028746008873,
0.027604861184954643,
0.05552225932478905,
0.09850256145000458,
0.055214621126651764,
-0.044402070343494415,
0.03513985127210617,
-0.11474829167127609,
-0.01077771745622158,
0.05650679022073746,
0.012168405577540398,
0.005093224812299013,
-0.12142189592123032,
0.008427305147051811,
-0.028616653755307198,
0.1902875155210495,
0.008318495936691761,
0.004581509158015251,
-0.03823390230536461,
0.1471363604068756,
0.14451807737350464,
0.03666907176375389,
0.12476859986782074,
0.013967765495181084,
-0.03739316761493683,
-0.033343009650707245,
0.09010767191648483,
0.061881061643362045,
0.1917669028043747,
-0.013514860533177853,
0.06294067949056625,
-0.011286824010312557,
0.09491283446550369,
-0.1014958992600441,
0.06874537467956543,
-0.07888410985469818,
-0.13679267466068268,
-0.07260652631521225,
0.05734629184007645,
-0.08852363377809525,
0.02496856078505516,
0.23204131424427032,
-0.06507746875286102,
0.020409567281603813,
0.06933721154928207,
-0.051636360585689545,
-0.037432555109262466,
-0.3618152439594269,
-0.0821942389011383,
-0.1668490767478943,
0.02981492690742016,
-0.03379368036985397,
0.024809667840600014,
0.1532903015613556,
-0.032377537339925766,
0.03741628676652908,
0.1380728930234909,
-0.05360892042517662,
-0.06147210672497749,
0.02707790769636631,
0.03196599707007408,
-0.003144542220979929,
0.13240686058998108,
0.058212362229824066,
0.05442290008068085,
-0.1018301248550415,
0.04036681354045868,
0.027925940230488777,
0.14376750588417053,
0.04611864686012268,
-0.007502784952521324,
-0.03777911141514778,
-0.050484634935855865,
0.03786375746130943,
-0.06631900370121002,
0.1526602804660797,
-0.006968729663640261,
0.0057181729935109615,
0.00041019919444806874,
0.14633673429489136,
-0.10020656883716583,
0.030390823259949684,
-0.1012955829501152,
0.07531630992889404,
-0.036413826048374176,
0.020853666588664055,
-0.033795442432165146,
-0.05074675381183624,
-0.0037557899486273527,
0.26617345213890076,
0.12049680203199387,
-0.09705595672130585,
-0.014737913385033607,
-0.00332077220082283,
0.009432926774024963,
-0.008835175074636936,
0.1881399005651474,
-0.008531241677701473,
0.3794819116592407,
-0.060777697712183,
-0.11971525102853775,
-0.0007811291725374758,
-0.02628098428249359,
-0.10383965075016022,
-0.058653753250837326,
-0.046226851642131805,
-0.0694609209895134,
-0.0767151415348053,
0.06935618072748184,
-0.14593714475631714,
0.030634110793471336,
0.18783877789974213,
-0.02156393975019455,
0.017189139500260353,
-0.0040845186449587345,
-0.08187869936227798,
0.0114095788449049,
0.05832456424832344,
-0.09095168113708496,
-0.05675622075796127,
0.09542521834373474,
-0.01077845599502325,
-0.19520831108093262,
-0.004989562556147575,
-0.047560617327690125,
-0.10117153078317642,
0.16839443147182465,
-0.007639065384864807,
0.027136119082570076,
0.02525664120912552,
0.03756169229745865,
-0.07154884934425354,
-0.028974352404475212,
-0.06134429946541786,
-0.06579495966434479,
-0.009909897111356258,
0.08913205564022064,
-0.05613451823592186,
-0.06636522710323334,
-0.021208135411143303,
-0.10834996402263641,
-0.014660806395113468,
0.059458959847688675,
0.0872168019413948,
-0.07755196839570999,
-0.07164224982261658,
-0.11525832116603851,
0.09434299170970917,
0.03291524201631546,
0.02115311659872532,
-0.02133815363049507,
-0.039976924657821655,
0.05911771580576897,
0.04716811329126358,
-0.019631318747997284,
0.0018823215505108237,
-0.020415551960468292,
-0.07133488357067108,
-0.030095074325799942,
-0.005791961215436459,
-0.03650107607245445,
0.0043386900797486305,
-0.03394775092601776,
0.012246310710906982,
-0.03559291362762451,
0.09088056534528732,
0.0637965202331543,
0.015351086854934692,
-0.04131801053881645,
-0.122963547706604,
0.07831079512834549,
0.09541017562150955,
-0.03223400190472603,
-0.16223010420799255
] |
null | null | null |
# NeverSleep/MiquMaid-v2-70B-DPO Imatrix GGUF
## Description
Imatrix GGUF quants of [NeverSleep/MiquMaid-v2-70B](https://huggingface.co/NeverSleep/MiquMaid-v2-70B)
[IQ2-XS](https://huggingface.co/Kooten/MiquMaid-v2-70B-Imatrix-GGUF/blob/main/MiquMaid-v2-70B-IQ2_XS.gguf), [IQ2-XXS](https://huggingface.co/Kooten/MiquMaid-v2-70B-Imatrix-GGUF/blob/main/MiquMaid-v2-70B-IQ2_XXS.gguf), [IQ3-XXS](https://huggingface.co/Kooten/MiquMaid-v2-70B-Imatrix-GGUF/blob/main/MiquMaid-v2-70B-IQ3_XXS.gguf)
### Custom format:
```
### Instruction:
{system prompt}
### Input:
{input}
### Response:
{reply}
```
## Contact
Kooten on discord
[ko-fi.com/kooten](https://ko-fi.com/kooten) | {"license": "cc-by-nc-4.0", "tags": ["not-for-all-audiences", "nsfw"]} | null | Kooten/MiquMaid-v2-70B-Imatrix-GGUF | [
"gguf",
"not-for-all-audiences",
"nsfw",
"license:cc-by-nc-4.0",
"region:us"
] | 2024-02-07T15:57:37+00:00 | [] | [] | TAGS
#gguf #not-for-all-audiences #nsfw #license-cc-by-nc-4.0 #region-us
|
# NeverSleep/MiquMaid-v2-70B-DPO Imatrix GGUF
## Description
Imatrix GGUF quants of NeverSleep/MiquMaid-v2-70B
IQ2-XS, IQ2-XXS, IQ3-XXS
### Custom format:
## Contact
Kooten on discord
URL | [
"# NeverSleep/MiquMaid-v2-70B-DPO Imatrix GGUF",
"## Description\nImatrix GGUF quants of NeverSleep/MiquMaid-v2-70B\n\nIQ2-XS, IQ2-XXS, IQ3-XXS",
"### Custom format:",
"## Contact\nKooten on discord\n\nURL"
] | [
"TAGS\n#gguf #not-for-all-audiences #nsfw #license-cc-by-nc-4.0 #region-us \n",
"# NeverSleep/MiquMaid-v2-70B-DPO Imatrix GGUF",
"## Description\nImatrix GGUF quants of NeverSleep/MiquMaid-v2-70B\n\nIQ2-XS, IQ2-XXS, IQ3-XXS",
"### Custom format:",
"## Contact\nKooten on discord\n\nURL"
] | [
33,
23,
38,
5,
7
] | [
"passage: TAGS\n#gguf #not-for-all-audiences #nsfw #license-cc-by-nc-4.0 #region-us \n# NeverSleep/MiquMaid-v2-70B-DPO Imatrix GGUF## Description\nImatrix GGUF quants of NeverSleep/MiquMaid-v2-70B\n\nIQ2-XS, IQ2-XXS, IQ3-XXS### Custom format:## Contact\nKooten on discord\n\nURL"
] | [
-0.06136941537261009,
0.11133108288049698,
-0.006444139406085014,
0.09798911958932877,
-0.02592054195702076,
0.07480200380086899,
0.2010866105556488,
0.08492279052734375,
0.12762531638145447,
-0.014663360081613064,
0.0747612714767456,
-0.036147940903902054,
0.06280605494976044,
0.02867213264107704,
-0.01995835080742836,
0.0007472994038835168,
0.0286093782633543,
0.027168769389390945,
0.06942065060138702,
0.04823126271367073,
0.029519127681851387,
0.005569336004555225,
0.0001170270043076016,
-0.012080526910722256,
-0.06746985763311386,
-0.11461964249610901,
-0.0265667662024498,
-0.0067425197921693325,
0.030714277178049088,
0.023854965344071388,
-0.04005401208996773,
0.15124531090259552,
-0.04231675714254379,
-0.11919476836919785,
0.019010303542017937,
-0.0110403373837471,
-0.04148271307349205,
0.011710683815181255,
-0.007002673577517271,
0.01732245273888111,
0.105200856924057,
0.09004223346710205,
-0.09940118342638016,
0.0541895255446434,
-0.17872413992881775,
-0.07483775913715363,
-0.028451943770051003,
0.1152937188744545,
-0.020176542922854424,
0.10161774605512619,
-0.0027587958611547947,
0.14638464152812958,
-0.10285426676273346,
0.02660905010998249,
0.19199424982070923,
-0.23142419755458832,
0.017223909497261047,
0.18554960191249847,
-0.09956266731023788,
0.027896763756871223,
-0.11917639523744583,
0.03871836140751839,
-0.003879185765981674,
-0.051135607063770294,
-0.08212362974882126,
-0.0327770933508873,
0.06693283468484879,
-0.054907456040382385,
-0.023647025227546692,
0.007721752859652042,
0.16595755517482758,
0.05135667696595192,
-0.04907706007361412,
0.09946531057357788,
0.018001459538936615,
-0.12642242014408112,
-0.05902090668678284,
-0.008579040877521038,
0.06235537678003311,
0.018307054415345192,
0.02529558353126049,
-0.030866945162415504,
-0.101947121322155,
-0.055563785135746,
-0.10761075466871262,
0.15846523642539978,
-0.07526582479476929,
0.0364486463367939,
-0.019239770248532295,
-0.02754756063222885,
-0.23469974100589752,
-0.06765130907297134,
-0.12320663034915924,
-0.07257089018821716,
0.023245375603437424,
-0.043857499957084656,
0.013590001501142979,
0.03361644968390465,
0.14275400340557098,
0.13324303925037384,
-0.020801573991775513,
0.03848160058259964,
-0.000155042958795093,
0.07791562378406525,
-0.012400696985423565,
-0.02209591306746006,
0.04792540892958641,
0.020313194021582603,
0.06536277383565903,
-0.027590736746788025,
0.025550909340381622,
-0.01988261193037033,
-0.03672543540596962,
-0.03407769650220871,
-0.07035902887582779,
0.05061522126197815,
0.0384410135447979,
0.03215595707297325,
-0.09057943522930145,
-0.015380644239485264,
0.0881158709526062,
0.05576673895120621,
-0.026757556945085526,
0.055610936135053635,
0.026575252413749695,
0.02016337774693966,
-0.010059951804578304,
0.05946793407201767,
0.04165071249008179,
0.023782962933182716,
-0.08763033151626587,
0.052009280771017075,
0.03870197758078575,
0.09638656675815582,
0.08707785606384277,
0.061280086636543274,
-0.0005390362348407507,
-0.10313710570335388,
-0.09561987966299057,
0.0179695226252079,
0.01908082701265812,
-0.03152807801961899,
0.016219349578022957,
0.01894591934978962,
-0.0374375581741333,
-0.038761816918849945,
-0.0075719887390732765,
-0.13148018717765808,
-0.06351203471422195,
0.08924635499715805,
0.08196493238210678,
0.09404625743627548,
-0.10458914935588837,
-0.014829362742602825,
-0.09015262871980667,
0.02303837053477764,
0.025757521390914917,
-0.011473345570266247,
-0.1542350798845291,
-0.024138571694493294,
0.01071260217577219,
0.004982727579772472,
-0.04731391370296478,
-0.062097083777189255,
-0.00015251646982505918,
0.2266242504119873,
-0.06456100195646286,
-0.021415283903479576,
0.1383516788482666,
-0.12946638464927673,
-0.15179872512817383,
0.11566281318664551,
0.13984930515289307,
-0.08284933120012283,
0.025759709998965263,
0.23813292384147644,
-0.03887421265244484,
-0.10061229765415192,
-0.08646579831838608,
0.062391359359025955,
-0.004066126886755228,
-0.08105790615081787,
0.13037678599357605,
-0.026051297783851624,
-0.0221256110817194,
0.055381134152412415,
0.06157481670379639,
0.0035504403058439493,
-0.002358662663027644,
-0.11670275777578354,
-0.02608482725918293,
-0.0514158196747303,
0.09580500423908234,
0.038196057081222534,
0.0024136591237038374,
-0.058154817670583725,
0.009558508172631264,
-0.19589635729789734,
0.04672807455062866,
0.08846769481897354,
-0.014243298210203648,
-0.09364300966262817,
0.060741618275642395,
-0.05523613840341568,
-0.013888485729694366,
0.012718265876173973,
-0.052193429321050644,
-0.00506248464807868,
-0.0226367749273777,
0.14909109473228455,
0.050533395260572433,
0.06657521426677704,
-0.0004732785455416888,
-0.10095172375440598,
0.03928590193390846,
0.015243690460920334,
0.029924169182777405,
0.00473427027463913,
-0.1116166040301323,
0.151590496301651,
0.01530588511377573,
0.13621489703655243,
-0.214121013879776,
-0.04391840845346451,
0.08246861398220062,
0.034110020846128464,
0.010084609501063824,
-0.12122289836406708,
0.11406972259283066,
0.014840980060398579,
0.031671781092882156,
0.003222990781068802,
0.08165621012449265,
0.01136988215148449,
-0.12040521949529648,
0.05408915877342224,
-0.09415101259946823,
0.12625427544116974,
0.09240605682134628,
0.07527066767215729,
-0.028806546702980995,
0.023142799735069275,
-0.031198587268590927,
0.006977295037358999,
0.08517684787511826,
0.06331785023212433,
0.06694085150957108,
-0.0934256762266159,
0.013874039053916931,
-0.039230216294527054,
-0.02281942404806614,
0.049931127578020096,
-0.08493214100599289,
-0.07340934127569199,
0.06696228682994843,
0.09088139235973358,
-0.11659235507249832,
0.11498679220676422,
0.14686764776706696,
-0.04472476989030838,
0.10076157748699188,
-0.012307546101510525,
0.024485256522893906,
-0.12711256742477417,
0.021500810980796814,
0.046676963567733765,
0.19601941108703613,
-0.24010294675827026,
0.10377997159957886,
0.04278215020895004,
0.031561557203531265,
0.026211131364107132,
-0.11223777383565903,
-0.09094741940498352,
-0.030050627887248993,
-0.09167512506246567,
-0.17993897199630737,
0.041634999215602875,
-0.0720340758562088,
0.0791526734828949,
0.002149117412045598,
0.030947549268603325,
0.09085829555988312,
0.03537265956401825,
-0.08743555098772049,
0.0956353172659874,
-0.09324093908071518,
-0.17959906160831451,
0.026458194479346275,
-0.02622576430439949,
-0.17484061419963837,
0.015947695821523666,
0.058710575103759766,
-0.07178009301424026,
-0.01805531419813633,
-0.03684346005320549,
-0.03996521234512329,
-0.056145358830690384,
0.045730218291282654,
0.06374558061361313,
0.014894900843501091,
0.05927208438515663,
-0.01511429063975811,
-0.056713275611400604,
-0.012700475752353668,
-0.043319299817085266,
0.15958188474178314,
-0.01188455056399107,
0.0732010006904602,
0.08141420036554337,
0.055351726710796356,
-0.025495685636997223,
0.028201721608638763,
0.27709710597991943,
-0.1296452283859253,
-0.0017742557683959603,
0.09385092556476593,
0.08524345606565475,
0.016574181616306305,
0.15111978352069855,
0.07551433145999908,
-0.07768746465444565,
-0.06853824108839035,
0.02494824305176735,
-0.10652092099189758,
-0.1380564272403717,
-0.026229253038764,
-0.08717323839664459,
0.1188778206706047,
-0.08908595144748688,
0.05295970290899277,
0.13455422222614288,
0.030165767297148705,
0.008344498462975025,
-0.041799742728471756,
-0.022536804899573326,
0.044960375875234604,
0.08741138875484467,
0.05234315246343613,
-0.0063356622122228146,
-0.027863066643476486,
0.026191486045718193,
0.13443513214588165,
0.0733209028840065,
0.007477644365280867,
0.028715332970023155,
0.1447247862815857,
0.08261129260063171,
0.038735173642635345,
0.06718779355287552,
-0.0316980816423893,
0.015059364028275013,
-0.10372171550989151,
-0.0022508709225803614,
-0.06704461574554443,
-0.03540851175785065,
-0.015839053317904472,
0.10461881011724472,
-0.16809788346290588,
0.06538590788841248,
-0.17093291878700256,
0.07894236594438553,
-0.07925427705049515,
0.10948961973190308,
-0.006697706412523985,
-0.0492614209651947,
0.002209477126598358,
0.06579918414354324,
0.006783086806535721,
0.012915381230413914,
-0.05966593325138092,
-0.056114863604307175,
0.08801327645778656,
0.06473079323768616,
0.030787033960223198,
-0.030937831848859787,
0.012200151570141315,
-0.17843273282051086,
-0.06240028142929077,
-0.03982189670205116,
0.07490493357181549,
-0.3036797046661377,
0.13606496155261993,
0.053549956530332565,
0.08939732611179352,
-0.056977834552526474,
-0.07947979122400284,
0.04713628068566322,
0.09173867851495743,
0.13756594061851501,
0.07645595818758011,
0.12801460921764374,
-0.1209874376654625,
-0.04039662703871727,
0.02601849101483822,
0.05573543906211853,
-0.005975381005555391,
-0.021057449281215668,
0.07099264860153198,
0.05080174282193184,
-0.0036198985762894154,
0.05682520940899849,
-0.15094293653964996,
-0.02308906801044941,
0.19459441304206848,
0.1713649481534958,
0.019842317327857018,
-0.0173236932605505,
0.014519482851028442,
-0.17953252792358398,
0.21316221356391907,
-0.13855019211769104,
-0.05026520788669586,
-0.10356033593416214,
-0.060828760266304016,
-0.0026007499545812607,
-0.029995402321219444,
-0.07111256569623947,
-0.0881228819489479,
0.030661696568131447,
-0.012593363411724567,
-0.02413913980126381,
0.05497469753026962,
-0.10356134176254272,
-0.17706778645515442,
-0.07866661250591278,
0.13787385821342468,
-0.04875003919005394,
0.0602494440972805,
0.020786339417099953,
-0.0006236707558855414,
0.02399154007434845,
-0.13892550766468048,
0.1475113183259964,
-0.14113003015518188,
-0.1586979776620865,
0.0004864647926297039,
0.04932813346385956,
0.016059769317507744,
-0.04099304974079132,
-0.11147284507751465,
0.08743306994438171,
0.5538868308067322,
0.03938773646950722,
0.1569063663482666,
0.20365941524505615,
0.015872454270720482,
-0.20781053602695465,
-0.10624918341636658,
-0.056101128458976746,
-0.13230223953723907,
-0.046490393579006195,
-0.14799439907073975,
0.033900436013936996,
0.17447912693023682,
-0.06430035084486008,
0.2624679207801819,
-0.08193640410900116,
-0.08743710815906525,
-0.013366875238716602,
0.05343803018331528,
0.3357769250869751,
-0.10280048102140427,
-0.07101985812187195,
0.036907024681568146,
-0.09264104068279266,
0.20172454416751862,
0.062447261065244675,
0.1667202115058899,
-0.020850686356425285,
-0.006943884771317244,
0.024641599506139755,
-0.026889117434620857,
0.16363458335399628,
0.03102937527000904,
-0.010678024962544441,
-0.06207979843020439,
-0.028328027576208115,
0.12648944556713104,
0.019456591457128525,
-0.01143435575067997,
-0.08007180690765381,
0.020676231011748314,
-0.0623646043241024,
-0.039432402700185776,
-0.08363662660121918,
0.03980334848165512,
-0.06060618907213211,
-0.07365337759256363,
-0.07609324157238007,
0.09168534725904465,
0.0012835769448429346,
-0.00857966672629118,
0.06740006804466248,
-0.08457598835229874,
0.01673279143869877,
0.21315909922122955,
0.021540245041251183,
-0.10495131462812424,
-0.032008904963731766,
-0.05915224924683571,
-0.082825668156147,
0.002618438797071576,
-0.054380785673856735,
0.019696755334734917,
0.0675007700920105,
0.07770814001560211,
0.06098489835858345,
0.04168868437409401,
-0.07563949376344681,
0.04553954303264618,
0.1437787115573883,
-0.08398330956697464,
-0.11874496936798096,
0.006481909193098545,
-0.056614950299263,
0.025024328380823135,
0.03941596671938896,
0.11241941154003143,
-0.006202193908393383,
0.026160968467593193,
0.033983927220106125,
0.027315331622958183,
-0.03819100558757782,
0.014522482641041279,
-0.018991226330399513,
0.04632114991545677,
-0.11810903251171112,
0.0831277072429657,
0.037460945546627045,
0.025798233225941658,
-0.057548440992832184,
0.10596493631601334,
-0.07359784841537476,
-0.11286573112010956,
0.002988805528730154,
-0.031099295243620872,
-0.0916275605559349,
0.014341100119054317,
-0.003965430427342653,
-0.0017353302100673318,
-0.0262734517455101,
0.10767745226621628,
-0.02649151347577572,
0.03154163062572479,
0.08838517963886261,
0.002178235212340951,
-0.016484567895531654,
0.03573266416788101,
-0.15086907148361206,
0.057429779320955276,
-0.029458465054631233,
-0.02560274489223957,
-0.03705795854330063,
0.059503115713596344,
-0.03740110993385315,
-0.023088760673999786,
-0.18709924817085266,
-0.051496610045433044,
0.05502872169017792,
0.0067373174242675304,
-0.07473207265138626,
0.003205937799066305,
0.021113134920597076,
-0.022507036104798317,
-0.040056534111499786,
0.00663637463003397,
-0.07146163284778595,
-0.02673460729420185,
0.07106157392263412,
0.04251774027943611,
-0.04856690391898155,
-0.04343724995851517,
0.04583821818232536,
0.01934688165783882,
0.11534145474433899,
0.1183481439948082,
0.022745905444025993,
0.0021655240561813116,
-0.26637083292007446,
0.05469798296689987,
-0.013005106709897518,
0.06716962158679962,
0.0019679777324199677,
0.02387879602611065,
-0.030750541016459465,
0.00006218321505002677,
0.02487774007022381,
0.07429797947406769,
0.09864679723978043,
-0.08168424665927887,
-0.1303018480539322,
-0.06908093392848969,
-0.08809448778629303,
-0.002545897848904133,
-0.06406357139348984,
0.07152220606803894,
-0.04560951888561249,
0.09223967790603638,
-0.056975774466991425,
0.03019026853144169,
-0.03069373033940792,
0.007108905352652073,
0.033682212233543396,
-0.0614018514752388,
-0.021640390157699585,
-0.033580221235752106,
0.015771161764860153,
0.0009335819049738348,
0.09929785877466202,
-0.0878136157989502,
-0.13831017911434174,
0.02238752320408821,
-0.0886693075299263,
-0.0168844535946846,
-0.0969071090221405,
0.184371680021286,
0.12223026156425476,
-0.0706624686717987,
-0.03958067670464516,
0.05892564728856087,
-0.024205690249800682,
-0.09186212718486786,
0.1267533004283905,
-0.0012600341578945518,
-0.03483147174119949,
-0.00966996606439352,
-0.03510632365942001,
-0.10618836432695389,
-0.08572696149349213,
-0.010662121698260307,
-0.08286168426275253,
0.03230459615588188,
-0.059760306030511856,
-0.0076285856775939465,
0.0941876471042633,
-0.004986080806702375,
0.03553386032581329,
-0.08724702149629593,
-0.03753215819597244,
-0.006352781318128109,
-0.10488346964120865,
-0.036176614463329315,
-0.07812035083770752,
0.07673940062522888,
-0.06220331788063049,
0.021970830857753754,
0.007494697347283363,
0.05222199112176895,
-0.0333687849342823,
0.11490070819854736,
-0.07915503531694412,
-0.011355921626091003,
0.03482864052057266,
-0.03251298516988754,
-0.06638863682746887,
-0.016907675191760063,
0.01042974554002285,
0.04407918080687523,
-0.05501178652048111,
0.055242493748664856,
0.015005955472588539,
0.004185464233160019,
-0.06999317556619644,
-0.002816815860569477,
-0.13630835711956024,
0.005261213984340429,
0.014738336205482483,
0.11217841506004333,
0.21500490605831146,
-0.030547888949513435,
0.0451803021132946,
0.0261512603610754,
0.09196694195270538,
0.00705724535509944,
-0.04425252601504326,
-0.07855186611413956,
0.16019092500209808,
-0.1206061840057373,
0.018399380147457123,
-0.08634801208972931,
-0.06083134561777115,
0.012054099701344967,
0.19689153134822845,
0.026902232319116592,
-0.09622456133365631,
0.01630576327443123,
-0.01820019632577896,
0.008160698227584362,
0.0177694633603096,
0.04120367765426636,
-0.005086146295070648,
0.1829162836074829,
-0.0641474574804306,
0.0032346732914447784,
-0.002354204189032316,
0.02484956756234169,
-0.06250714510679245,
0.18808166682720184,
0.0587601363658905,
-0.0303767379373312,
-0.12846055626869202,
0.04795662313699722,
-0.19028973579406738,
-0.027738898992538452,
0.08422661572694778,
-0.09906638413667679,
-0.07191913574934006,
-0.02576468139886856,
-0.10980243235826492,
0.0218777135014534,
0.028630858287215233,
-0.010616161860525608,
-0.021364031359553337,
-0.19630365073680878,
-0.015836969017982483,
-0.0944652184844017,
0.05283988639712334,
-0.02218419872224331,
0.18961510062217712,
0.24159394204616547,
0.008371911011636257,
0.13129159808158875,
0.0750129297375679,
-0.03682859241962433,
-0.11622646450996399,
0.16757860779762268,
0.010563344694674015,
-0.16288375854492188,
-0.04034043103456497,
0.10940545052289963,
0.009539967402815819,
-0.042498499155044556,
0.12502063810825348,
0.02547432668507099,
-0.0033530613873153925,
0.32186681032180786,
-0.034190692007541656,
-0.006212572567164898,
0.07450693845748901,
-0.05870545282959938,
0.05930255353450775,
0.014592810533940792,
-0.03629843518137932,
-0.09658385813236237,
0.014061534777283669,
0.07231281697750092,
0.036429453641176224,
-0.05117250978946686,
-0.08318722248077393,
-0.07973888516426086,
0.04195398837327957,
0.06528086960315704,
0.07059099525213242,
-0.1723967343568802,
-0.010470394976437092,
-0.10875795036554337,
0.04149572551250458,
-0.039310332387685776,
0.023438863456249237,
0.08727719634771347,
-0.07112095504999161,
0.005378099624067545,
-0.16542436182498932,
0.004918581340461969,
-0.00442963233217597,
-0.042359571903944016,
-0.09819452464580536
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# robust_llm_pythia-tt-410m-mz-v0
This model is a fine-tuned version of [EleutherAI/pythia-410m-deduped](https://huggingface.co/EleutherAI/pythia-410m-deduped) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 8
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 1
### Training results
### Framework versions
- Transformers 4.37.1
- Pytorch 2.1.2
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "EleutherAI/pythia-410m-deduped", "model-index": [{"name": "robust_llm_pythia-tt-410m-mz-v0", "results": []}]} | text-classification | AlignmentResearch/robust_llm_pythia-tt-410m-mz-v0 | [
"transformers",
"tensorboard",
"safetensors",
"gpt_neox",
"text-classification",
"generated_from_trainer",
"base_model:EleutherAI/pythia-410m-deduped",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T15:57:48+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #gpt_neox #text-classification #generated_from_trainer #base_model-EleutherAI/pythia-410m-deduped #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# robust_llm_pythia-tt-410m-mz-v0
This model is a fine-tuned version of EleutherAI/pythia-410m-deduped on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 8
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 1
### Training results
### Framework versions
- Transformers 4.37.1
- Pytorch 2.1.2
- Datasets 2.16.1
- Tokenizers 0.15.1
| [
"# robust_llm_pythia-tt-410m-mz-v0\n\nThis model is a fine-tuned version of EleutherAI/pythia-410m-deduped on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 8\n- eval_batch_size: 64\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.1\n- Pytorch 2.1.2\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #gpt_neox #text-classification #generated_from_trainer #base_model-EleutherAI/pythia-410m-deduped #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# robust_llm_pythia-tt-410m-mz-v0\n\nThis model is a fine-tuned version of EleutherAI/pythia-410m-deduped on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 8\n- eval_batch_size: 64\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.1\n- Pytorch 2.1.2\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
88,
51,
6,
12,
8,
3,
90,
4,
30
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #gpt_neox #text-classification #generated_from_trainer #base_model-EleutherAI/pythia-410m-deduped #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# robust_llm_pythia-tt-410m-mz-v0\n\nThis model is a fine-tuned version of EleutherAI/pythia-410m-deduped on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 8\n- eval_batch_size: 64\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1### Training results### Framework versions\n\n- Transformers 4.37.1\n- Pytorch 2.1.2\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.10561726242303848,
0.171852245926857,
-0.00273716077208519,
0.08492659032344818,
0.13674034178256989,
0.005278249271214008,
0.11791041493415833,
0.1464013308286667,
-0.10233872383832932,
0.0905134379863739,
0.09618810564279556,
0.08088137954473495,
0.06161019951105118,
0.14429478347301483,
-0.027833707630634308,
-0.23780399560928345,
0.02081311121582985,
0.009791475720703602,
-0.03788036108016968,
0.1026669591665268,
0.1134067252278328,
-0.08096183091402054,
0.09728512912988663,
0.041577860713005066,
-0.1328280121088028,
0.021859215572476387,
-0.027062509208917618,
-0.07858627289533615,
0.07991960644721985,
0.018646959215402603,
0.043007854372262955,
0.005170444492250681,
0.06537466496229172,
-0.181588813662529,
-0.004032128490507603,
0.06176597252488136,
0.005565493833273649,
0.09538514167070389,
0.04962676391005516,
0.021099643781781197,
0.061165425926446915,
-0.15357783436775208,
0.08280495554208755,
0.0365191251039505,
-0.0817311555147171,
-0.1950533390045166,
-0.09157554805278778,
0.08138673007488251,
0.062144290655851364,
0.09453961253166199,
0.01655523292720318,
0.20810867846012115,
0.022285573184490204,
0.0822710320353508,
0.1954139620065689,
-0.2984469532966614,
-0.04407013952732086,
-0.006515359040349722,
0.07432913035154343,
0.08562015742063522,
-0.09003469347953796,
0.006253803614526987,
0.04966868832707405,
0.028093017637729645,
0.10819026082754135,
-0.01061047613620758,
0.012563387863337994,
-0.04397730529308319,
-0.12210652977228165,
-0.06022383272647858,
0.1998109668493271,
0.0543220154941082,
-0.07269081473350525,
-0.13153840601444244,
-0.0702613890171051,
-0.11353658884763718,
-0.01706988923251629,
-0.040858928114175797,
0.026782706379890442,
-0.0382627509534359,
-0.05226621404290199,
-0.06089397519826889,
-0.06378825753927231,
-0.02940729819238186,
0.018988830968737602,
0.14430569112300873,
0.050460558384656906,
0.031097451224923134,
-0.021031677722930908,
0.07489471137523651,
-0.03180919215083122,
-0.1510993242263794,
-0.03848277032375336,
0.0004753661050926894,
-0.02538912370800972,
-0.04029351845383644,
-0.02380356937646866,
-0.03955686464905739,
0.009159831330180168,
0.15878133475780487,
-0.061947062611579895,
0.0658549815416336,
0.01707812026143074,
-0.0036197430454194546,
-0.020092427730560303,
0.15089534223079681,
-0.024378718808293343,
-0.0343584306538105,
0.03339407220482826,
0.09920869022607803,
0.05700018256902695,
-0.0242936909198761,
-0.08445925265550613,
-0.02083476446568966,
0.12482673674821854,
0.0673050582408905,
-0.004174943547695875,
0.030187997967004776,
-0.055932529270648956,
-0.050604742020368576,
0.12383820116519928,
-0.12650805711746216,
0.02441304735839367,
-0.04004133492708206,
-0.07593190670013428,
-0.06660901010036469,
0.020288482308387756,
0.00476053124293685,
-0.051294248551130295,
-0.0010380977764725685,
-0.08801109343767166,
-0.02447669580578804,
-0.05471597611904144,
-0.04996403306722641,
0.006099442485719919,
-0.0675334706902504,
0.013709411956369877,
-0.09432511776685715,
-0.19519367814064026,
-0.025649720802903175,
0.012076782993972301,
-0.05532321706414223,
-0.07523468881845474,
0.009823127649724483,
-0.05046799033880234,
0.034687984734773636,
-0.008314640261232853,
0.053864192217588425,
-0.035886213183403015,
0.057389989495277405,
0.056479956954717636,
0.021983571350574493,
0.032278478145599365,
0.04464644938707352,
-0.10269153863191605,
0.04933593422174454,
-0.09461299329996109,
0.08659478276968002,
-0.06338634341955185,
0.03950100392103195,
-0.13602890074253082,
-0.09638211876153946,
0.02092244103550911,
-0.03534207120537758,
0.0701901838183403,
0.10895276069641113,
-0.09856961667537689,
-0.025930503383278847,
0.11599332094192505,
-0.09235028922557831,
-0.12754587829113007,
0.09045040607452393,
-0.003459494560956955,
0.019020425155758858,
0.06268047541379929,
0.15182079374790192,
0.12309393286705017,
-0.10842721909284592,
-0.050895076245069504,
0.013984429650008678,
0.04856077954173088,
-0.005826503969728947,
0.08604533970355988,
0.0019849881064146757,
0.034489959478378296,
0.01389169879257679,
-0.04538732022047043,
-0.024534322321414948,
-0.05967036634683609,
-0.08286120742559433,
-0.0757111981511116,
-0.07689732313156128,
0.03653485327959061,
0.018559275195002556,
0.034725435078144073,
-0.07456431537866592,
-0.125461146235466,
0.06364457309246063,
0.1355142891407013,
-0.054771434515714645,
0.022216718643903732,
-0.09243461489677429,
0.045693378895521164,
-0.06581196933984756,
-0.016153834760189056,
-0.19958069920539856,
-0.09121906757354736,
0.07093065232038498,
-0.11795338988304138,
0.02277310937643051,
-0.03125784173607826,
0.047121644020080566,
0.08534682542085648,
-0.025054261088371277,
-0.05845755711197853,
-0.07946839928627014,
-0.013794535771012306,
-0.1091054305434227,
-0.14033161103725433,
-0.051427535712718964,
-0.030910683795809746,
0.09438572078943253,
-0.17406021058559418,
0.03199819102883339,
0.05345383659005165,
0.13235384225845337,
0.033746615052223206,
-0.05734450742602348,
0.022537024691700935,
0.014960168860852718,
-0.011860269121825695,
-0.11037762463092804,
0.029715364798903465,
0.0217103473842144,
-0.10890112817287445,
-0.012235860340297222,
-0.17543910443782806,
0.1366472989320755,
0.07502447813749313,
0.10610296577215195,
-0.0549837090075016,
-0.010352497920393944,
-0.05557996407151222,
-0.05551328510046005,
-0.04352600872516632,
-0.020440073683857918,
0.18378806114196777,
-0.001982462126761675,
0.1420019268989563,
-0.07275193929672241,
-0.06753586232662201,
0.01962852291762829,
-0.001974568236619234,
-0.04854584485292435,
0.0665561780333519,
-0.04978277534246445,
-0.176454558968544,
0.11461227387189865,
0.11014319211244583,
-0.014823252335190773,
0.1332104653120041,
-0.055381692945957184,
-0.07701659947633743,
-0.04548100382089615,
0.0070168329402804375,
0.02250489592552185,
0.0822962075471878,
-0.07050507515668869,
0.0013735832180827856,
0.052357789129018784,
0.008275855332612991,
0.01522634457796812,
-0.13493677973747253,
0.015083769336342812,
0.03829900175333023,
-0.039998430758714676,
0.02850000187754631,
0.007141510024666786,
-0.017440982162952423,
0.06442281603813171,
0.021866222843527794,
0.0052825515158474445,
0.046616457402706146,
0.0019375222036615014,
-0.06857980042695999,
0.16781368851661682,
-0.1141677275300026,
-0.19579386711120605,
-0.1540682464838028,
0.04591953381896019,
-0.08341476321220398,
-0.00936385802924633,
0.04226602986454964,
-0.05533507466316223,
-0.07346420735120773,
-0.08624986559152603,
-0.04895868897438049,
-0.058250561356544495,
0.008907714858651161,
0.07624241709709167,
0.0016830054810270667,
0.1199786514043808,
-0.1322876214981079,
-0.002735236892476678,
0.02147541753947735,
-0.059688374400138855,
-0.015538745559751987,
0.048332780599594116,
0.10917292535305023,
0.06046278402209282,
-0.01871388964354992,
0.010058260522782803,
-0.020758410915732384,
0.2608772814273834,
-0.07910621166229248,
-0.01078766118735075,
0.18200817704200745,
0.013986070640385151,
0.09709488600492477,
0.09319042414426804,
0.02724536508321762,
-0.0811530351638794,
0.024603869765996933,
0.01966765709221363,
-0.025868235155940056,
-0.24491345882415771,
-0.028222251683473587,
-0.01595315709710121,
-0.055584866553545,
0.11963540315628052,
0.073969766497612,
0.07772541791200638,
0.07363541424274445,
-0.04880288988351822,
0.06940218061208725,
-0.027060765773057938,
0.12167328596115112,
0.13893358409404755,
0.0607127882540226,
0.10101298987865448,
-0.035369325429201126,
-0.02210097573697567,
0.05790127441287041,
0.035467565059661865,
0.21387846767902374,
-0.013013073243200779,
0.23070231080055237,
0.007246182765811682,
0.1354256123304367,
0.0037906323559582233,
0.025995414704084396,
0.016698062419891357,
0.019115598872303963,
0.014054564759135246,
-0.07204366475343704,
-0.033137768507003784,
0.03849177062511444,
-0.04133007675409317,
0.053675152361392975,
-0.07098757475614548,
0.06448445469141006,
0.016657181084156036,
0.25281068682670593,
0.019625673070549965,
-0.3335632383823395,
-0.09250140935182571,
0.017521072179079056,
-0.026475677266716957,
-0.08485762029886246,
-0.015664363279938698,
0.07180937379598618,
-0.13628274202346802,
0.08653796464204788,
-0.06771975755691528,
0.093207448720932,
-0.0886773094534874,
0.001638231799006462,
0.05380702763795853,
0.11608710885047913,
-0.0026401160284876823,
0.09010764211416245,
-0.19783683121204376,
0.17194975912570953,
0.042593616992235184,
0.09422273933887482,
-0.0675434023141861,
0.05238184332847595,
0.0044057415798306465,
0.09004747122526169,
0.1063770204782486,
-0.00808937381953001,
-0.07276100665330887,
-0.15369565784931183,
-0.12945421040058136,
0.0048260875046253204,
0.09863140434026718,
-0.050797201693058014,
0.06581763178110123,
-0.04595096781849861,
0.007526066154241562,
0.015097824856638908,
-0.08709919452667236,
-0.15382130444049835,
-0.1366850733757019,
0.05039529129862785,
0.02879462018609047,
0.002681741490960121,
-0.08679834753274918,
-0.10196366906166077,
-0.006646587513387203,
0.1776142567396164,
-0.012378503568470478,
-0.09400147944688797,
-0.15164533257484436,
0.05247272923588753,
0.1146719679236412,
-0.09141425043344498,
0.029892371967434883,
-0.007794850505888462,
0.15666405856609344,
0.027799539268016815,
-0.0818222165107727,
0.06062699109315872,
-0.06406120955944061,
-0.1802753508090973,
-0.028217758983373642,
0.12623047828674316,
0.01458729337900877,
0.03548213839530945,
0.007599939592182636,
0.03118874877691269,
-0.008276291191577911,
-0.075143963098526,
0.012772939167916775,
0.03136937692761421,
0.09817986935377121,
0.01765836775302887,
-0.010546901263296604,
0.05588370934128761,
-0.06065935641527176,
-0.017274171113967896,
0.11768961697816849,
0.21294426918029785,
-0.06736836582422256,
0.06369448453187943,
0.07021353393793106,
-0.06899640709161758,
-0.15484921634197235,
0.0033233684953302145,
0.1088787391781807,
0.00990114826709032,
0.057593412697315216,
-0.1642095446586609,
0.0700412169098854,
0.09566235542297363,
-0.039502520114183426,
0.07235908508300781,
-0.28746989369392395,
-0.1347283571958542,
0.06185326725244522,
0.10839872062206268,
0.03415350988507271,
-0.1570444405078888,
-0.07399630546569824,
-0.0410890057682991,
-0.09055258333683014,
0.06838428229093552,
-0.04604938626289368,
0.10494453459978104,
-0.0404692143201828,
0.04958501085639,
0.02201208844780922,
-0.029720785096287727,
0.16228662431240082,
0.017835404723882675,
0.06845932453870773,
-0.07075048238039017,
0.07133977860212326,
0.12569667398929596,
-0.08330589532852173,
0.09384418278932571,
-0.08113088458776474,
0.07988444715738297,
-0.1597341001033783,
-0.013137947767972946,
-0.033501334488391876,
0.06497646123170853,
-0.04580894112586975,
-0.04500969871878624,
-0.03578805923461914,
0.01857098750770092,
0.04401330277323723,
-0.03604796156287193,
0.15135496854782104,
0.06584823131561279,
0.060963138937950134,
0.1770731806755066,
0.07531896978616714,
0.0070027075707912445,
-0.16812016069889069,
-0.022063609212636948,
-0.021412020549178123,
0.07408542931079865,
-0.12781576812267303,
0.025752799585461617,
0.09816677868366241,
0.025871766731142998,
0.11413931101560593,
0.008113701827824116,
-0.07353976368904114,
0.007155396044254303,
0.029332010075449944,
-0.09943033009767532,
-0.15914329886436462,
-0.052092913538217545,
0.032309211790561676,
-0.1404508799314499,
0.042972564697265625,
0.1328319013118744,
-0.054255031049251556,
-0.027524935081601143,
-0.012652020901441574,
0.014511575922369957,
0.010174164548516273,
0.17314328253269196,
0.05418986827135086,
0.0739729031920433,
-0.08190012723207474,
0.11968778073787689,
0.08430419862270355,
-0.05868859589099884,
0.06738881021738052,
0.04947996884584427,
-0.08798981457948685,
-0.025760289281606674,
0.051118891686201096,
0.11021410673856735,
-0.03971346467733383,
-0.06105243042111397,
-0.082502581179142,
-0.0491718053817749,
0.04295147582888603,
0.051419053226709366,
0.05781413987278938,
0.0026841191574931145,
-0.0283208005130291,
-0.004363712854683399,
-0.11762119084596634,
0.10689020901918411,
0.02974054217338562,
0.09024776518344879,
-0.19681525230407715,
0.01584392972290516,
0.021975088864564896,
0.05477837845683098,
-0.014590799808502197,
-0.0005405503325164318,
-0.06665213406085968,
-0.041394613683223724,
-0.10089718550443649,
0.020907118916511536,
-0.03413388505578041,
0.007426350377500057,
-0.03276745229959488,
-0.06077922508120537,
-0.042267706245183945,
0.07963752001523972,
-0.043058112263679504,
-0.08832057565450668,
0.018439382314682007,
0.04462283477187157,
-0.12804335355758667,
-0.014092822559177876,
0.02057083696126938,
-0.10175389051437378,
0.09057392925024033,
0.041731689125299454,
0.020661359652876854,
-0.012430793605744839,
0.024288814514875412,
0.04343230277299881,
0.03697090968489647,
0.017888229340314865,
0.038226719945669174,
-0.09971849620342255,
-0.011616895906627178,
-0.018833642825484276,
0.0045390259474515915,
0.007098729256540537,
0.09611557424068451,
-0.14055129885673523,
-0.036709681153297424,
-0.04075096920132637,
-0.01798415184020996,
-0.06572268903255463,
0.05037117749452591,
0.10259649157524109,
0.010493393987417221,
0.16913247108459473,
-0.057074349373579025,
0.02516871877014637,
-0.1930965781211853,
-0.01055690087378025,
0.003646446391940117,
-0.07081033289432526,
-0.04144757613539696,
-0.02275857701897621,
0.06181395798921585,
-0.056015368551015854,
0.11074884235858917,
-0.012964884750545025,
0.10877122730016708,
0.04742545261979103,
0.01861216500401497,
-0.011262191459536552,
-0.004376330878585577,
0.1639181524515152,
0.051082100719213486,
0.002027111826464534,
0.11219839006662369,
-0.021386103704571724,
0.07113958895206451,
0.011384670622646809,
0.06882230192422867,
0.11112651228904724,
-0.030884578824043274,
0.07386604696512222,
0.05019963160157204,
-0.05399257689714432,
-0.21459750831127167,
0.033687006682157516,
-0.01660013757646084,
0.1197255477309227,
-0.01150808110833168,
0.09063667804002762,
0.13287140429019928,
-0.1422809511423111,
0.03413408622145653,
-0.02209225669503212,
-0.1013811007142067,
-0.09394799172878265,
-0.1130417063832283,
-0.08831276744604111,
-0.11595712602138519,
0.008506055921316147,
-0.11661233007907867,
0.020005883648991585,
0.06744245439767838,
-0.01729024201631546,
-0.009171965532004833,
0.1582811325788498,
-0.003772997995838523,
-0.0059605794958770275,
0.04644881561398506,
0.0008732642163522542,
-0.017525361850857735,
-0.0022089420817792416,
-0.05351274088025093,
0.045319296419620514,
0.02918233722448349,
0.0775495171546936,
-0.02417704463005066,
0.021215716376900673,
0.042911022901535034,
-0.021120287477970123,
-0.09410669654607773,
0.01098352950066328,
0.03622112050652504,
0.003154211677610874,
0.038516271859407425,
0.035157982259988785,
-0.00424905214458704,
-0.039003800600767136,
0.24337634444236755,
-0.052598532289266586,
-0.03451705351471901,
-0.11017509549856186,
0.14647041261196136,
0.04697035625576973,
-0.027360640466213226,
0.07429580390453339,
-0.10120001435279846,
0.0010054762242361903,
0.11590419709682465,
0.12498392164707184,
-0.015355641953647137,
-0.016114484518766403,
-0.02349325269460678,
-0.02184039168059826,
-0.0304032601416111,
0.09067756682634354,
0.09671121835708618,
-0.01480698212981224,
-0.051716551184654236,
0.017986714839935303,
-0.021421659737825394,
-0.03183663636445999,
-0.10746892541646957,
0.07489079236984253,
-0.009950831532478333,
0.01290216576308012,
-0.022614778950810432,
0.05977059155702591,
0.029490474611520767,
-0.1343683898448944,
0.025229157879948616,
-0.14385594427585602,
-0.16676746308803558,
0.001815533614717424,
0.07475912570953369,
-0.006019333843141794,
0.043641235679388046,
0.026727618649601936,
0.004402189515531063,
0.0936569944024086,
-0.013680989854037762,
-0.08789470791816711,
-0.07886925339698792,
0.07925428450107574,
-0.11596812307834625,
0.2092898041009903,
-0.004428068175911903,
0.07906520366668701,
0.10981132835149765,
-0.005871836561709642,
-0.16449831426143646,
0.0372631810605526,
0.058991044759750366,
-0.021994473412632942,
0.047698311507701874,
0.1568998247385025,
-0.01613222435116768,
0.055193036794662476,
0.035965628921985626,
-0.10684987157583237,
-0.06953977048397064,
-0.027384048327803612,
0.041094448417425156,
-0.06525472551584244,
-0.0074331751093268394,
-0.06173979490995407,
0.1589069962501526,
0.15776368975639343,
-0.05882057175040245,
-0.02189815416932106,
-0.048769209533929825,
0.025286458432674408,
0.06719450652599335,
0.041848309338092804,
-0.002629660302773118,
-0.20940497517585754,
0.0058960141614079475,
0.03853709623217583,
0.03727302700281143,
-0.2207089364528656,
-0.10942442715167999,
0.02263503335416317,
-0.0558357872068882,
-0.06789369881153107,
0.08773741126060486,
0.03995434194803238,
0.002638207981362939,
-0.04001977667212486,
-0.06430051475763321,
-0.0662650391459465,
0.12197888642549515,
-0.16149650514125824,
-0.055369842797517776
] |
null | null | null |
# NeverSleep/MiquMaid-v2-70B-DPO Imatrix GGUF
## Description
Imatrix GGUF quants of [NeverSleep/MiquMaid-v2-70B-DPO](https://huggingface.co/NeverSleep/MiquMaid-v2-70B-DPO)
[IQ2-XS](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-Imatrix-GGUF/blob/main/MiquMaid-v2-70B-DPO-IQ2_XS.gguf), [IQ2-XXS](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-Imatrix-GGUF/blob/main/MiquMaid-v2-70B-DPO-IQ2_XXS.gguf), [IQ3-XXS](https://huggingface.co/Kooten/MiquMaid-v2-70B-DPO-Imatrix-GGUF/blob/main/MiquMaid-v2-70B-DPO-IQ3_XXS.gguf)
### Custom format:
```
### Instruction:
{system prompt}
### Input:
{input}
### Response:
{reply}
```
## Contact
Kooten on discord
[ko-fi.com/kooten](https://ko-fi.com/kooten) | {"license": "cc-by-nc-4.0", "tags": ["not-for-all-audiences", "nsfw"]} | null | Kooten/MiquMaid-v2-70B-DPO-Imatrix-GGUF | [
"gguf",
"not-for-all-audiences",
"nsfw",
"license:cc-by-nc-4.0",
"region:us"
] | 2024-02-07T15:57:53+00:00 | [] | [] | TAGS
#gguf #not-for-all-audiences #nsfw #license-cc-by-nc-4.0 #region-us
|
# NeverSleep/MiquMaid-v2-70B-DPO Imatrix GGUF
## Description
Imatrix GGUF quants of NeverSleep/MiquMaid-v2-70B-DPO
IQ2-XS, IQ2-XXS, IQ3-XXS
### Custom format:
## Contact
Kooten on discord
URL | [
"# NeverSleep/MiquMaid-v2-70B-DPO Imatrix GGUF",
"## Description\nImatrix GGUF quants of NeverSleep/MiquMaid-v2-70B-DPO\n\nIQ2-XS, IQ2-XXS, IQ3-XXS",
"### Custom format:",
"## Contact\nKooten on discord\n\nURL"
] | [
"TAGS\n#gguf #not-for-all-audiences #nsfw #license-cc-by-nc-4.0 #region-us \n",
"# NeverSleep/MiquMaid-v2-70B-DPO Imatrix GGUF",
"## Description\nImatrix GGUF quants of NeverSleep/MiquMaid-v2-70B-DPO\n\nIQ2-XS, IQ2-XXS, IQ3-XXS",
"### Custom format:",
"## Contact\nKooten on discord\n\nURL"
] | [
33,
23,
41,
5,
7
] | [
"passage: TAGS\n#gguf #not-for-all-audiences #nsfw #license-cc-by-nc-4.0 #region-us \n# NeverSleep/MiquMaid-v2-70B-DPO Imatrix GGUF## Description\nImatrix GGUF quants of NeverSleep/MiquMaid-v2-70B-DPO\n\nIQ2-XS, IQ2-XXS, IQ3-XXS### Custom format:## Contact\nKooten on discord\n\nURL"
] | [
-0.06958597898483276,
0.11813756078481674,
-0.006708186585456133,
0.10240675508975983,
-0.016641097143292427,
0.07521188259124756,
0.19953849911689758,
0.0846794918179512,
0.13095231354236603,
-0.007534670177847147,
0.06944800168275833,
-0.016475141048431396,
0.06974062323570251,
0.026760267093777657,
-0.015653377398848534,
-0.0031161634251475334,
0.02530612237751484,
0.032479893416166306,
0.07100807875394821,
0.044242143630981445,
0.031263694167137146,
0.004130910616368055,
-0.0023783231154084206,
-0.01543028187006712,
-0.06765978038311005,
-0.11257842183113098,
-0.026383332908153534,
-0.007048073690384626,
0.03801547735929489,
0.017782563343644142,
-0.046400874853134155,
0.1494089812040329,
-0.038184549659490585,
-0.11209950596094131,
0.016631310805678368,
-0.007444179151207209,
-0.03690542280673981,
0.011049306020140648,
0.0016098138876259327,
0.022358065471053123,
0.09498269855976105,
0.09597228467464447,
-0.09199508279561996,
0.05568252131342888,
-0.17545755207538605,
-0.08297435194253922,
-0.02978583611547947,
0.1209636852145195,
-0.020938878878951073,
0.10737888514995575,
-0.00015959639858920127,
0.1544869840145111,
-0.08866605162620544,
0.021610483527183533,
0.19204501807689667,
-0.2301158308982849,
0.022287718951702118,
0.1873643547296524,
-0.10230527073144913,
0.021611418575048447,
-0.11224717646837234,
0.02456893026828766,
-0.008020065724849701,
-0.048192474991083145,
-0.09368938207626343,
-0.03405431658029556,
0.07487794011831284,
-0.0599408820271492,
-0.019157087430357933,
0.010263659060001373,
0.1689848005771637,
0.05625251308083534,
-0.05893359333276749,
0.10719992965459824,
0.023116668686270714,
-0.13025428354740143,
-0.05813465267419815,
-0.01946137100458145,
0.07200653851032257,
0.01980843022465706,
0.027863331139087677,
-0.016729045659303665,
-0.09910350292921066,
-0.05773720145225525,
-0.10628513991832733,
0.15919239819049835,
-0.07604163885116577,
0.030811937525868416,
-0.022340280935168266,
-0.02321595512330532,
-0.226840540766716,
-0.06592099368572235,
-0.12327763438224792,
-0.06469738483428955,
0.02736750803887844,
-0.044966921210289,
0.01860283501446247,
0.032505836337804794,
0.13935446739196777,
0.12316292524337769,
-0.011166155338287354,
0.027666239067912102,
-0.009008875116705894,
0.07709483057260513,
-0.01562906615436077,
-0.012039696797728539,
0.04578515514731407,
0.035912204533815384,
0.06773509830236435,
-0.02690245769917965,
0.03317601606249809,
-0.016591602936387062,
-0.036714158952236176,
-0.027284229174256325,
-0.06261839717626572,
0.05227307230234146,
0.04613381251692772,
0.030103599652647972,
-0.10189370810985565,
-0.019384659826755524,
0.0816153734922409,
0.05380946770310402,
-0.029268009588122368,
0.057904887944459915,
0.021455863490700722,
0.03456961736083031,
-0.009329176507890224,
0.054953720420598984,
0.03835340589284897,
0.024834884330630302,
-0.09207714349031448,
0.0517650730907917,
0.042351171374320984,
0.09903256595134735,
0.09212676435709,
0.0653764009475708,
0.0031798211857676506,
-0.1105627790093422,
-0.10143677145242691,
0.01343531534075737,
0.009829297661781311,
-0.03366287425160408,
0.00895904004573822,
0.021313142031431198,
-0.033635709434747696,
-0.04944543540477753,
-0.010353176854550838,
-0.13480907678604126,
-0.05886983498930931,
0.0974670797586441,
0.08239800482988358,
0.08606381714344025,
-0.10670758038759232,
-0.009479096159338951,
-0.09286972880363464,
0.022064656019210815,
0.01789749786257744,
-0.016618359833955765,
-0.15204745531082153,
-0.023432167246937752,
0.012950198724865913,
0.005285379011183977,
-0.042524877935647964,
-0.0620105005800724,
0.006487763021141291,
0.23425829410552979,
-0.06638403236865997,
-0.021153878420591354,
0.12734687328338623,
-0.12580175697803497,
-0.15379269421100616,
0.10712278634309769,
0.13881944119930267,
-0.07119099050760269,
0.034400176256895065,
0.24148616194725037,
-0.049304891377687454,
-0.11663176864385605,
-0.09238090366125107,
0.058658938854932785,
-0.007259405683726072,
-0.08913443982601166,
0.1308172345161438,
-0.03619234263896942,
-0.018405912443995476,
0.057488538324832916,
0.05182415619492531,
0.00041634758235886693,
0.001387989497743547,
-0.11063368618488312,
-0.02299039252102375,
-0.05155320093035698,
0.08274596184492111,
0.044482652097940445,
0.0025835952255874872,
-0.05505656823515892,
0.007313697133213282,
-0.19700241088867188,
0.053226880729198456,
0.0925600603222847,
-0.018899627029895782,
-0.08799299597740173,
0.05471368879079819,
-0.053775932639837265,
-0.014339548535645008,
0.013352809473872185,
-0.05176255479454994,
-0.005005914252251387,
-0.02503916621208191,
0.14743025600910187,
0.04731094464659691,
0.06113729625940323,
-0.005408823490142822,
-0.08869422972202301,
0.03942732512950897,
0.0056998711079359055,
0.0320800319314003,
0.006680487189441919,
-0.1048283651471138,
0.1529587358236313,
0.013365251943469048,
0.13945114612579346,
-0.20888710021972656,
-0.04535660147666931,
0.07391033321619034,
0.031937405467033386,
0.00840369239449501,
-0.12727589905261993,
0.11252642422914505,
0.013309312053024769,
0.037548236548900604,
-0.0008245877688750625,
0.08539844304323196,
0.021075598895549774,
-0.11641304939985275,
0.05073419213294983,
-0.08883889764547348,
0.13205935060977936,
0.09366051852703094,
0.07581279426813126,
-0.03718048334121704,
0.022589128464460373,
-0.02992749772965908,
0.007448245771229267,
0.08305707573890686,
0.05834341049194336,
0.06724437326192856,
-0.08881893008947372,
0.015268553979694843,
-0.034770749509334564,
-0.016131283715367317,
0.05016956478357315,
-0.07912853360176086,
-0.06938455253839493,
0.0645163282752037,
0.09699681401252747,
-0.10580737888813019,
0.12482395023107529,
0.13418959081172943,
-0.03919493779540062,
0.0890125259757042,
-0.018510065972805023,
0.020179508253932,
-0.13135211169719696,
0.02356553077697754,
0.043963462114334106,
0.204131081700325,
-0.23918522894382477,
0.10704392194747925,
0.04430485516786575,
0.028414491564035416,
0.020320704206824303,
-0.11256587505340576,
-0.09339595586061478,
-0.02863829769194126,
-0.08914284408092499,
-0.18663039803504944,
0.03899753466248512,
-0.07424749433994293,
0.08668401092290878,
0.006491639651358128,
0.020442001521587372,
0.09962709248065948,
0.03930266574025154,
-0.08938082307577133,
0.09104825556278229,
-0.09019682556390762,
-0.17753489315509796,
0.037408944219350815,
-0.015321152284741402,
-0.17542453110218048,
0.022196084260940552,
0.056775547564029694,
-0.06901320815086365,
-0.013054500333964825,
-0.024153783917427063,
-0.04466938599944115,
-0.06398507952690125,
0.047080691903829575,
0.06431183218955994,
0.014956665225327015,
0.06068427115678787,
-0.012696726247668266,
-0.05925711989402771,
-0.012314355932176113,
-0.03581321984529495,
0.14830294251441956,
-0.007391644641757011,
0.06646008789539337,
0.08697699755430222,
0.0658838301897049,
-0.025072624906897545,
0.028814617544412613,
0.26050394773483276,
-0.1299629509449005,
-0.008340800181031227,
0.09325709193944931,
0.08837428689002991,
0.010485499165952206,
0.1538276970386505,
0.07343991845846176,
-0.08337726444005966,
-0.06560104340314865,
0.021340062841773033,
-0.09702056646347046,
-0.14225256443023682,
-0.022474665194749832,
-0.0843707025051117,
0.1326255053281784,
-0.08837694674730301,
0.05679165944457054,
0.12315074354410172,
0.02975943684577942,
0.013651078566908836,
-0.04348503425717354,
-0.027150889858603477,
0.04930756241083145,
0.08782899379730225,
0.04431922361254692,
-0.010462545789778233,
-0.02721339650452137,
0.02050705999135971,
0.1283283680677414,
0.08441898226737976,
-0.001197201432660222,
0.0329945906996727,
0.13832509517669678,
0.07589580863714218,
0.031368549913167953,
0.055586401373147964,
-0.03597624972462654,
0.032325103878974915,
-0.10518389940261841,
-0.004691782873123884,
-0.06579972058534622,
-0.03845393285155296,
-0.017763953655958176,
0.10670024156570435,
-0.16418924927711487,
0.053219784051179886,
-0.18115079402923584,
0.06662186980247498,
-0.07437989115715027,
0.10170301049947739,
-0.006531459745019674,
-0.0509890615940094,
-0.0036964353639632463,
0.06514254212379456,
0.012981363572180271,
0.009658382274210453,
-0.08531422913074493,
-0.057947199791669846,
0.08249378204345703,
0.06933286041021347,
0.02343781851232052,
-0.03954406827688217,
0.013391917571425438,
-0.1829402595758438,
-0.06885886192321777,
-0.03790002688765526,
0.07289756089448929,
-0.3169785737991333,
0.12884311378002167,
0.054107293486595154,
0.09445417672395706,
-0.058828286826610565,
-0.07537312060594559,
0.04298360273241997,
0.08008411526679993,
0.14572733640670776,
0.07118359208106995,
0.14496341347694397,
-0.12487588077783585,
-0.03489181399345398,
0.02380094863474369,
0.04788793995976448,
-0.0044497051276266575,
-0.025021588429808617,
0.0727430135011673,
0.057677969336509705,
-0.005241530016064644,
0.06369869410991669,
-0.14255686104297638,
-0.009525504894554615,
0.20199859142303467,
0.15942728519439697,
0.03310457989573479,
-0.017169149592518806,
0.01536891795694828,
-0.18341319262981415,
0.20692317187786102,
-0.13471068441867828,
-0.04908636957406998,
-0.0971679538488388,
-0.04997089132666588,
-0.02049281634390354,
-0.025935426354408264,
-0.07477767020463943,
-0.08294866234064102,
0.02643086016178131,
-0.010983799584209919,
-0.022427255287766457,
0.05735046789050102,
-0.10327262431383133,
-0.1717165857553482,
-0.08259379863739014,
0.14061060547828674,
-0.04973713308572769,
0.058586400002241135,
0.020876161754131317,
-0.006716446951031685,
0.025796914473176003,
-0.12924961745738983,
0.15648335218429565,
-0.14255665242671967,
-0.15297885239124298,
-0.00006922556349309161,
0.05274162441492081,
0.025707364082336426,
-0.047089193016290665,
-0.10947242379188538,
0.08353504538536072,
0.5491383671760559,
0.04256271570920944,
0.15176963806152344,
0.19224657118320465,
0.01599702052772045,
-0.19779250025749207,
-0.0995451956987381,
-0.0509810596704483,
-0.1340470016002655,
-0.042941153049468994,
-0.14534808695316315,
0.03342019021511078,
0.18298490345478058,
-0.07187815755605698,
0.2723676860332489,
-0.0885716900229454,
-0.07973220944404602,
0.0022031613625586033,
0.0562569722533226,
0.33682307600975037,
-0.11074879765510559,
-0.06683191657066345,
0.03705092519521713,
-0.09182582795619965,
0.20704646408557892,
0.07004863023757935,
0.16490842401981354,
-0.023349201306700706,
0.00030362160759977996,
0.03255615383386612,
-0.024114925414323807,
0.15965069830417633,
0.038498178124427795,
-0.014062250033020973,
-0.06037494167685509,
-0.02945791184902191,
0.11186660826206207,
0.02490013651549816,
-0.0027117906138300896,
-0.06163894757628441,
0.021015511825680733,
-0.06630419194698334,
-0.04091214761137962,
-0.08296246081590652,
0.04134111478924751,
-0.05970713123679161,
-0.07574968785047531,
-0.07548683881759644,
0.09323174506425858,
-0.006052054464817047,
-0.008572018705308437,
0.061141788959503174,
-0.08396442979574203,
0.01497417502105236,
0.20974579453468323,
0.011077583767473698,
-0.0943315178155899,
-0.0635024830698967,
-0.060000836849212646,
-0.08490484952926636,
0.007335032802075148,
-0.04651283845305443,
0.013910796493291855,
0.07085282355546951,
0.0834137499332428,
0.05804062634706497,
0.04410867393016815,
-0.08143637329339981,
0.04245531186461449,
0.13592776656150818,
-0.08124414831399918,
-0.12195437401533127,
0.00808438565582037,
-0.04962233453989029,
0.023163316771388054,
0.04774436727166176,
0.11254063993692398,
-0.011070625856518745,
0.02401946298778057,
0.041806548833847046,
0.028922678902745247,
-0.04197254776954651,
0.020022369921207428,
-0.01319898758083582,
0.04238300025463104,
-0.11658086627721786,
0.0967806801199913,
0.03765223175287247,
0.017555424943566322,
-0.05786796286702156,
0.09909791499376297,
-0.06549282371997833,
-0.1113213524222374,
-0.01305398903787136,
-0.032488252967596054,
-0.09387429058551788,
0.016348708420991898,
-0.015161080285906792,
-0.003209958551451564,
-0.02450127899646759,
0.07889245450496674,
-0.027259059250354767,
0.03444192185997963,
0.08134415745735168,
0.004210307262837887,
-0.02416612207889557,
0.029343536123633385,
-0.1502898633480072,
0.06079152971506119,
-0.035946574062108994,
-0.022015446797013283,
-0.03647680953145027,
0.06230900436639786,
-0.033106524497270584,
-0.025802969932556152,
-0.19184035062789917,
-0.0487334318459034,
0.07077541202306747,
-0.0016334963729605079,
-0.07315029203891754,
0.001310011837631464,
0.024726439267396927,
-0.01817411743104458,
-0.033658869564533234,
0.005537326447665691,
-0.06910553574562073,
-0.028738532215356827,
0.07611871510744095,
0.04325120151042938,
-0.04453157261013985,
-0.0479283444583416,
0.03904154524207115,
0.01748739928007126,
0.11532656848430634,
0.12575788795948029,
0.03019719012081623,
-0.007366720121353865,
-0.259486585855484,
0.06042006239295006,
-0.01048726961016655,
0.06429697573184967,
0.003133725607767701,
0.014882143586874008,
-0.03487616777420044,
0.003354353131726384,
0.01465523336082697,
0.07139559090137482,
0.10117606818675995,
-0.08494284749031067,
-0.14037996530532837,
-0.061244770884513855,
-0.0765756294131279,
-0.005053704138845205,
-0.06881289184093475,
0.0679543986916542,
-0.03937363624572754,
0.08563478291034698,
-0.05803963169455528,
0.03669266775250435,
-0.03867091238498688,
0.006090876180678606,
0.026323063299059868,
-0.05630730837583542,
-0.008157022297382355,
-0.0281312745064497,
0.01875070296227932,
0.0022149954456835985,
0.083016037940979,
-0.08704541623592377,
-0.13408073782920837,
0.025090305134654045,
-0.08527383208274841,
-0.006740836892277002,
-0.0981672927737236,
0.18910656869411469,
0.11831270158290863,
-0.06948768347501755,
-0.03705466166138649,
0.06337236613035202,
-0.018788045272231102,
-0.07854723185300827,
0.12324610352516174,
0.0034448192454874516,
-0.04146568104624748,
-0.007537037134170532,
-0.027191083878278732,
-0.11375672370195389,
-0.09079573303461075,
-0.011280655860900879,
-0.07512028515338898,
0.031404685229063034,
-0.05715370923280716,
0.00012148995301686227,
0.08872322738170624,
-0.0004990851739421487,
0.031185347586870193,
-0.08709913492202759,
-0.03321842476725578,
-0.00281333620660007,
-0.09657162427902222,
-0.02797984704375267,
-0.08457280695438385,
0.07804512232542038,
-0.055157922208309174,
0.023403966799378395,
0.01672516018152237,
0.04612213000655174,
-0.04236598685383797,
0.10902613401412964,
-0.07943501323461533,
-0.019031794741749763,
0.03656202554702759,
-0.0330231711268425,
-0.07525346428155899,
-0.02824619598686695,
0.005770452320575714,
0.040216438472270966,
-0.048597030341625214,
0.054622575640678406,
0.008492148481309414,
0.00905416440218687,
-0.07327476143836975,
0.0013987537240609527,
-0.12741239368915558,
0.011307969689369202,
0.01734832115471363,
0.10615243762731552,
0.23527804017066956,
-0.03460415080189705,
0.04712940752506256,
0.029701780527830124,
0.08320019394159317,
0.007671938743442297,
-0.030817829072475433,
-0.07545819878578186,
0.17046542465686798,
-0.11815965175628662,
0.012096421793103218,
-0.0825771763920784,
-0.06288855522871017,
0.01198521163314581,
0.18756353855133057,
0.016860343515872955,
-0.09780240803956985,
0.013623732142150402,
-0.02358251065015793,
0.004246414173394442,
0.011925132013857365,
0.03912976011633873,
-0.0009647512924857438,
0.18832992017269135,
-0.06422639638185501,
0.015279581770300865,
-0.003263740101829171,
0.03182072192430496,
-0.0571514293551445,
0.18879637122154236,
0.06476069241762161,
-0.027777735143899918,
-0.12459591031074524,
0.04851110279560089,
-0.19163201749324799,
-0.03625401109457016,
0.08143792301416397,
-0.09731447696685791,
-0.07034134119749069,
-0.028791552409529686,
-0.1331147998571396,
0.017952997237443924,
0.03168674558401108,
-0.015404485166072845,
-0.012879686430096626,
-0.1888996958732605,
-0.018380895256996155,
-0.09665103256702423,
0.04364268481731415,
-0.022310543805360794,
0.1996811479330063,
0.24381089210510254,
0.000021653610019711778,
0.14423495531082153,
0.07273070514202118,
-0.02276396006345749,
-0.12140430510044098,
0.16600878536701202,
0.008939906023442745,
-0.16769994795322418,
-0.036239419132471085,
0.09247350692749023,
0.00871390663087368,
-0.04694471135735512,
0.12510672211647034,
0.02215687185525894,
-0.012139556929469109,
0.33245158195495605,
-0.03444315493106842,
-0.017186680808663368,
0.07634903490543365,
-0.06349294632673264,
0.05587043985724449,
0.010524539276957512,
-0.03814045339822769,
-0.09909024089574814,
0.010428141802549362,
0.07050155103206635,
0.048877160996198654,
-0.055133070796728134,
-0.07639805227518082,
-0.08090100437402725,
0.04865029826760292,
0.06454959511756897,
0.061779435724020004,
-0.1711442619562149,
-0.011827843263745308,
-0.10334401577711105,
0.03218163177371025,
-0.0343395434319973,
0.011541460640728474,
0.06761627644300461,
-0.07305824756622314,
-0.0007814468699507415,
-0.15651236474514008,
-0.0067059933207929134,
-0.00014028527948539704,
-0.04120233282446861,
-0.10108377039432526
] |
null | null | null |
OpenHermes-2.5-Mistral-7B by teknium converted to f16 gguf for easier tinkering;
original model at https://huggingface.co/teknium/OpenHermes-2.5-Mistral-7B | {"license": "apache-2.0"} | null | interrobang/OpenHermes-2.5-Mistral-7B-GGUF-f16 | [
"gguf",
"license:apache-2.0",
"region:us"
] | 2024-02-07T16:03:14+00:00 | [] | [] | TAGS
#gguf #license-apache-2.0 #region-us
|
OpenHermes-2.5-Mistral-7B by teknium converted to f16 gguf for easier tinkering;
original model at URL | [] | [
"TAGS\n#gguf #license-apache-2.0 #region-us \n"
] | [
17
] | [
"passage: TAGS\n#gguf #license-apache-2.0 #region-us \n"
] | [
-0.01330315787345171,
0.15578481554985046,
-0.008150381036102772,
0.0016573866596445441,
-0.04673202335834503,
0.030570833012461662,
0.13376784324645996,
0.09194369614124298,
0.128515362739563,
-0.07453092932701111,
0.17055568099021912,
0.028890712186694145,
0.03800550848245621,
0.03976481780409813,
0.025271227583289146,
-0.11722369492053986,
0.10813431441783905,
-0.03824392333626747,
-0.09146501868963242,
-0.0007550549926236272,
0.0461740680038929,
0.03938475623726845,
-0.020297255367040634,
-0.012304259464144707,
-0.04646717384457588,
-0.01807590387761593,
0.061380259692668915,
-0.021618852391839027,
0.0584920234978199,
0.044396135956048965,
-0.013440944254398346,
0.02849559485912323,
-0.044239118695259094,
-0.20778964459896088,
0.01814555749297142,
-0.030956419184803963,
-0.0963222086429596,
0.034455686807632446,
-0.002720483811572194,
0.03162013739347458,
0.061540063470602036,
0.11115614324808121,
-0.11918631196022034,
0.037214748561382294,
-0.13505050539970398,
-0.30473440885543823,
-0.18398606777191162,
0.06021247059106827,
0.0028733352664858103,
0.041688770055770874,
0.07521043717861176,
0.05276030674576759,
-0.15611688792705536,
-0.04681211709976196,
0.06740783900022507,
-0.35676148533821106,
0.026624629274010658,
0.1421503871679306,
-0.04767010360956192,
0.03409823030233383,
-0.016959721222519875,
0.10372579842805862,
0.08160287141799927,
-0.0309890303760767,
-0.11352469027042389,
-0.04248137027025223,
-0.07062457501888275,
0.11909499764442444,
-0.022388093173503876,
-0.10835286229848862,
0.299734890460968,
0.07829401642084122,
-0.03949820250272751,
0.13934776186943054,
-0.018232177942991257,
0.12180425226688385,
0.0037959686014801264,
0.07966145128011703,
0.07500946521759033,
0.2106771469116211,
0.1879701316356659,
-0.1285460740327835,
-0.15332761406898499,
-0.07658189535140991,
-0.18097953498363495,
0.09647365659475327,
0.00481637567281723,
0.15394099056720734,
-0.12812361121177673,
0.004832482896745205,
-0.1665661633014679,
-0.05767213553190231,
-0.09051498025655746,
-0.052037935703992844,
0.14505687355995178,
0.09770436584949493,
-0.07494958490133286,
0.17277710139751434,
0.21161453425884247,
0.26735973358154297,
-0.03192757070064545,
0.006477556191384792,
-0.06420958042144775,
0.15842780470848083,
-0.0583864264190197,
0.019410474225878716,
0.05191675201058388,
0.11087486892938614,
0.13493096828460693,
-0.1719009280204773,
0.09378605335950851,
-0.016259867697954178,
-0.11252851784229279,
-0.011826573871076107,
-0.1737414002418518,
0.1333753764629364,
0.046528834849596024,
-0.08763917535543442,
-0.02980431728065014,
0.06516948342323303,
0.13636615872383118,
-0.016305889934301376,
-0.012872343882918358,
-0.008937547914683819,
0.009359321556985378,
-0.06836733967065811,
0.0008180695003829896,
0.060471873730421066,
0.08052167296409607,
-0.010497825220227242,
-0.09346228837966919,
-0.02963828109204769,
0.03977816551923752,
0.13056598603725433,
0.12507236003875732,
-0.03587067127227783,
0.04867643862962723,
-0.07281927764415741,
-0.1537899523973465,
0.043609101325273514,
0.0720045417547226,
0.01405393797904253,
-0.02105209231376648,
0.12900052964687347,
0.02947195991873741,
-0.002579237800091505,
-0.07217522710561752,
-0.019955281168222427,
-0.08173951506614685,
0.05495830997824669,
-0.0830233246088028,
-0.021544981747865677,
-0.25846993923187256,
-0.004414012189954519,
-0.08588613569736481,
0.05437181144952774,
0.026437964290380478,
-0.07966095209121704,
-0.15488311648368835,
0.15681788325309753,
-0.04755254089832306,
0.0606415681540966,
-0.055992551147937775,
-0.019309088587760925,
-0.04433238506317139,
0.07082654535770416,
-0.06881338357925415,
-0.034905560314655304,
0.21830801665782928,
-0.14372503757476807,
-0.12107643485069275,
0.045915182679891586,
0.058830294758081436,
-0.05773618072271347,
0.03722799941897392,
0.3155798017978668,
-0.033125780522823334,
-0.02127721533179283,
0.10441244393587112,
0.17700493335723877,
-0.10252934694290161,
-0.19990848004817963,
0.1626713126897812,
-0.18464820086956024,
-0.21314117312431335,
0.03374022617936134,
-0.1354188770055771,
0.14721231162548065,
0.029108481481671333,
-0.10722507536411285,
-0.042635608464479446,
-0.08002658933401108,
-0.048498500138521194,
-0.0413840115070343,
0.047936778515577316,
-0.04613621532917023,
0.04448116570711136,
-0.18495601415634155,
0.06669430434703827,
0.12992608547210693,
0.04142272472381592,
-0.04430747032165527,
0.06424026191234589,
0.018889864906668663,
0.04718081280589104,
0.014743788167834282,
-0.021515699103474617,
0.02174583449959755,
-0.09805754572153091,
0.07063749432563782,
0.08627311885356903,
0.059764452278614044,
-0.06353570520877838,
0.012208250351250172,
0.06167324259877205,
0.0011235947022214532,
0.03216798976063728,
0.06799250841140747,
-0.09858707338571548,
0.1041717529296875,
0.003250909736379981,
0.08585058897733688,
0.03630579635500908,
-0.01942511461675167,
0.1006988137960434,
-0.05730016902089119,
-0.06386759132146835,
0.0249167513102293,
0.008679588325321674,
-0.1123015284538269,
0.05832165852189064,
-0.01720862090587616,
0.08887401223182678,
0.0644783154129982,
-0.10543974488973618,
0.169432133436203,
0.06818334013223648,
0.20611946284770966,
0.1413285881280899,
0.03921271115541458,
0.09898378700017929,
-0.015305405482649803,
-0.01739814504981041,
-0.0033933527301996946,
0.0786864310503006,
0.03553178906440735,
-0.018575791269540787,
-0.03419099375605583,
-0.01237060222774744,
-0.028527243062853813,
0.009019540622830391,
-0.028413871303200722,
-0.08427856862545013,
-0.038365963846445084,
-0.00293926103040576,
0.18400533497333527,
-0.1352609246969223,
0.15670859813690186,
0.39662665128707886,
0.05809931084513664,
0.10376974195241928,
-0.1573796272277832,
-0.04060497134923935,
-0.04825512692332268,
0.03319299966096878,
-0.02701469324529171,
0.184132382273674,
-0.10087321698665619,
0.03391242399811745,
0.07232333719730377,
0.057861048728227615,
0.06244037300348282,
-0.17538565397262573,
-0.14002028107643127,
-0.02046295255422592,
-0.09578375518321991,
-0.104468435049057,
0.024245386943221092,
-0.16045768558979034,
0.026515116915106773,
0.026555776596069336,
-0.06167523190379143,
0.1598798930644989,
-0.008611330762505531,
-0.04725697636604309,
0.07082630693912506,
-0.20357191562652588,
-0.12087341398000717,
-0.07626573741436005,
-0.0042744106613099575,
-0.05095795914530754,
0.023625193163752556,
0.08465500921010971,
-0.09592214971780777,
-0.05633152276277542,
0.07234445214271545,
-0.09036894887685776,
-0.06662289053201675,
0.015421921387314796,
0.07410863041877747,
0.01950029656291008,
0.031299129128456116,
-0.10658411681652069,
-0.05114024132490158,
-0.0207142923027277,
-0.06643637269735336,
0.043042417615652084,
-0.029125841334462166,
0.0803024172782898,
0.09105047583580017,
0.08921805769205093,
0.06849837303161621,
-0.0003459077561274171,
0.14692601561546326,
-0.04780511185526848,
-0.07088837027549744,
0.1597827970981598,
0.02298906072974205,
0.03533780947327614,
0.10251420736312866,
0.07653065770864487,
-0.10019218176603317,
-0.033925220370292664,
-0.03363695368170738,
-0.11724339425563812,
-0.2371799200773239,
-0.02427528239786625,
-0.08114678412675858,
0.09731483459472656,
-0.03697621077299118,
0.1421261876821518,
0.1297444850206375,
0.07038690894842148,
0.01488346979022026,
-0.02049473114311695,
-0.0003764019056688994,
-0.03656056523323059,
0.1047646775841713,
-0.040648601949214935,
-0.04150701314210892,
-0.13765324652194977,
0.05280669033527374,
0.2123951017856598,
0.139052614569664,
0.13875830173492432,
0.2556404173374176,
0.0865810289978981,
0.14902375638484955,
0.1700367033481598,
0.043575435876846313,
0.01834452524781227,
0.02130916342139244,
-0.025624319911003113,
-0.0733172670006752,
-0.021722665056586266,
-0.024601932615041733,
0.060184698551893234,
-0.011125342920422554,
-0.20776666700839996,
0.06618745625019073,
-0.23978964984416962,
0.06574483960866928,
0.0798097774386406,
0.06448644399642944,
0.07763633877038956,
0.05836908146739006,
0.08228960633277893,
0.07039085030555725,
0.030951116234064102,
0.1121925488114357,
-0.05280335620045662,
-0.07145904004573822,
0.08945973217487335,
0.033000752329826355,
0.055582351982593536,
0.05654873698949814,
0.018150946125388145,
-0.09060613065958023,
-0.1282111406326294,
0.06521598249673843,
0.14194566011428833,
-0.2339191883802414,
0.21457308530807495,
0.023072047159075737,
-0.07668514549732208,
-0.039042409509420395,
-0.00877209473401308,
0.10390420258045197,
0.15340037643909454,
0.12154663354158401,
0.0913349986076355,
-0.1597919762134552,
0.06428830325603485,
-0.07498005777597427,
0.06961594521999359,
-0.0024978057481348515,
-0.04789665341377258,
-0.1743926703929901,
-0.04370199516415596,
0.05610458180308342,
0.02999473176896572,
0.14242549240589142,
-0.15068703889846802,
-0.08502397686243057,
0.047799695283174515,
0.13189667463302612,
-0.012925185263156891,
-0.1351172924041748,
0.07789962738752365,
-0.0223079863935709,
0.13693971931934357,
-0.08600178360939026,
-0.009268487803637981,
-0.06393920630216599,
-0.11813131719827652,
0.03256971016526222,
-0.03487817198038101,
0.030751710757613182,
-0.08259879052639008,
-0.11836646497249603,
-0.0880444347858429,
-0.19962769746780396,
0.0970500111579895,
-0.0914505124092102,
0.011035526171326637,
-0.007812321186065674,
0.10870376229286194,
-0.04637550935149193,
0.011202405206859112,
-0.014477218501269817,
-0.005613986402750015,
-0.037054069340229034,
-0.18300330638885498,
0.12522146105766296,
-0.0354662649333477,
-0.038446735590696335,
0.010260424576699734,
0.029084086418151855,
0.08555833250284195,
0.07213917374610901,
-0.11324945837259293,
0.17003758251667023,
0.28164100646972656,
-0.06761249154806137,
0.21416988968849182,
0.31458553671836853,
-0.0837627723813057,
-0.2168940305709839,
-0.20412485301494598,
-0.243454247713089,
-0.1341201663017273,
0.013469581492245197,
-0.22552542388439178,
0.0766766369342804,
0.20877110958099365,
-0.1714908480644226,
0.2946220934391022,
-0.1911279857158661,
-0.002857143757864833,
0.18236051499843597,
-0.052634578198194504,
0.44611573219299316,
-0.171253964304924,
-0.12041395157575607,
-0.033438790589571,
-0.19569972157478333,
0.12554602324962616,
-0.16359169781208038,
0.05501075088977814,
0.024149509146809578,
-0.08415087312459946,
-0.054437655955553055,
-0.02173338644206524,
0.2504390478134155,
-0.003609958803281188,
0.04743576794862747,
-0.08139622211456299,
0.020269259810447693,
0.1750870943069458,
0.0035608301404863596,
0.017100313678383827,
-0.1639900505542755,
-0.01987898163497448,
-0.011092170141637325,
0.017119288444519043,
-0.04806634411215782,
0.1080654114484787,
0.021870044991374016,
-0.07784450799226761,
-0.11714762449264526,
0.001524197170510888,
-0.08141826093196869,
0.01126081496477127,
0.1802481710910797,
0.04831566661596298,
-0.022556783631443977,
0.03305482119321823,
-0.11502696573734283,
-0.21776944398880005,
-0.01617501676082611,
-0.11659006774425507,
-0.05422298610210419,
0.07242408394813538,
-0.22702434659004211,
0.00973254069685936,
0.03509848192334175,
-0.021992651745676994,
0.05734172463417053,
0.0652010440826416,
-0.10430524498224258,
0.01023867167532444,
0.1239415779709816,
-0.10563014447689056,
-0.1211216077208519,
0.024934910237789154,
0.025935843586921692,
0.17365825176239014,
0.050733763724565506,
0.07385742664337158,
0.06699050962924957,
0.004279726184904575,
0.016696322709321976,
0.07876995950937271,
-0.16484880447387695,
-0.029713543131947517,
0.07098285108804703,
-0.040593817830085754,
-0.1394682228565216,
0.19541186094284058,
0.028642958030104637,
0.029897678643465042,
-0.024543415755033493,
0.035570524632930756,
-0.05335932597517967,
-0.09447703510522842,
-0.13994671404361725,
-0.020488658919930458,
-0.11634241044521332,
-0.14416252076625824,
0.04803739860653877,
-0.08919020742177963,
-0.02546677179634571,
-0.03999754413962364,
0.07524143159389496,
0.13654081523418427,
0.058341559022665024,
0.013631465844810009,
0.16242444515228271,
-0.09581738710403442,
-0.18957501649856567,
-0.020761094987392426,
-0.06745804101228714,
-0.20285774767398834,
0.02592097781598568,
0.08869154006242752,
-0.020691489800810814,
-0.03635775297880173,
-0.09561870247125626,
0.04295523837208748,
-0.1443842053413391,
-0.008468776941299438,
-0.06995626538991928,
0.011855300515890121,
0.06764024496078491,
-0.09456440061330795,
0.010382445529103279,
0.04733350872993469,
-0.14076519012451172,
-0.057332105934619904,
0.0022140543442219496,
0.03670362010598183,
-0.10423135757446289,
-0.05571652576327324,
0.12314850091934204,
0.0703822523355484,
0.14754506945610046,
0.1352262943983078,
0.05553649738430977,
0.14190629124641418,
-0.22292354702949524,
-0.05461675673723221,
0.05658799037337303,
-0.004033025354146957,
-0.0409574992954731,
0.0034502120688557625,
-0.03508859872817993,
0.03941653296351433,
-0.07560205459594727,
0.042029887437820435,
-0.01114447321742773,
-0.131355419754982,
-0.16204427182674408,
-0.020497437566518784,
-0.1109241396188736,
0.033404845744371414,
-0.17885321378707886,
0.17627950012683868,
0.07166852056980133,
0.051701869815588,
0.07173995673656464,
-0.017255190759897232,
0.0028538221959024668,
0.030197259038686752,
-0.0401480607688427,
-0.08765622228384018,
-0.13222530484199524,
-0.008726547472178936,
-0.12079193443059921,
-0.03162060305476189,
0.33216890692710876,
-0.030980104580521584,
-0.1621956080198288,
0.027868445962667465,
0.07404996454715729,
0.12636078894138336,
0.0026254802942276,
0.2259112149477005,
0.054794542491436005,
0.014137940481305122,
-0.13172785937786102,
0.009069803170859814,
0.005791394039988518,
-0.27599358558654785,
0.0624869205057621,
0.0013532752636820078,
0.08246773481369019,
0.020798001438379288,
0.040949176996946335,
-0.08768724650144577,
-0.01853700540959835,
-0.03068496286869049,
0.07054902613162994,
0.011548550799489021,
0.041395153850317,
0.05476638674736023,
0.1752733737230301,
-0.024234607815742493,
0.014738751575350761,
-0.019994881004095078,
0.01285971887409687,
-0.1276385486125946,
-0.13978463411331177,
0.014189327135682106,
-0.15848460793495178,
0.045844435691833496,
-0.010509041137993336,
0.039326947182416916,
0.20746053755283356,
0.04208739846944809,
-0.06142117455601692,
-0.0455465130507946,
-0.08228625357151031,
-0.043509144335985184,
-0.009183176793158054,
-0.02066180668771267,
-0.06648684293031693,
-0.04331827908754349,
-0.09308581799268723,
-0.015449702739715576,
-0.06880850344896317,
-0.03242972493171692,
0.01768173836171627,
0.029128799214959145,
0.01990487426519394,
-0.08321307599544525,
0.002178385853767395,
-0.09582957625389099,
0.05271981284022331,
0.004241388291120529,
0.16314728558063507,
0.026698041707277298,
0.024437984451651573,
0.13244377076625824,
0.0729329064488411,
-0.013862117193639278,
-0.10025795549154282,
-0.05922839790582657,
0.03516887500882149,
-0.07846292853355408,
0.0765465646982193,
-0.0554979145526886,
0.0003111901751253754,
-0.007978007197380066,
0.23187364637851715,
0.18745596706867218,
-0.1058710440993309,
0.010778109543025494,
-0.024767739698290825,
0.011735105887055397,
0.06795872747898102,
0.13705003261566162,
0.05810157582163811,
0.1847129464149475,
-0.07218676060438156,
-0.043915461748838425,
0.00902556348592043,
0.017504174262285233,
-0.18057847023010254,
0.0833623856306076,
-0.00872169528156519,
-0.07224971055984497,
-0.03864730894565582,
0.10137778520584106,
-0.0695849135518074,
0.07668505609035492,
0.06028321757912636,
-0.028682032600045204,
0.027593085542321205,
-0.005067290738224983,
0.08621590584516525,
0.03740588203072548,
0.053074054419994354,
-0.1063728779554367,
-0.07684344053268433,
0.008703646250069141,
0.021847110241651535,
-0.3201918601989746,
-0.17543719708919525,
0.0938958078622818,
0.07460742443799973,
0.2910914719104767,
0.00119002943392843,
0.04769929125905037,
0.03156432509422302,
0.0401572659611702,
-0.12241633981466293,
0.12800614535808563,
0.012672079727053642,
-0.04014110192656517,
-0.1387740820646286,
-0.1924290657043457,
-0.032277900725603104,
-0.11549804359674454,
0.052928097546100616,
0.100664421916008,
0.012590800411999226,
0.21247832477092743,
-0.049709733575582504,
-0.022715643048286438,
-0.0057846843264997005,
-0.1684456169605255,
0.04534858837723732,
-0.063796766102314,
-0.03746533393859863,
-0.07755741477012634,
-0.06920577585697174,
0.006900185719132423,
0.07946354150772095,
-0.17020881175994873,
-0.052895352244377136,
0.2066379338502884,
0.04911750182509422,
0.1324104219675064,
0.011248554103076458,
0.05078674107789993,
-0.029439883306622505,
-0.07551411539316177,
0.09735040366649628,
-0.045069485902786255,
0.01129083801060915,
0.13425007462501526,
-0.018308015540242195,
0.02112884260714054,
-0.166777104139328,
0.02545231580734253,
-0.06915885210037231,
-0.0002544210001360625,
-0.06249799206852913
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# robbert-2023-dutch-base-gender
This model is a fine-tuned version of [DTAI-KULeuven/robbert-2023-dutch-base](https://huggingface.co/DTAI-KULeuven/robbert-2023-dutch-base) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.6591
- Precision: 0.6282
- Recall: 0.6290
- Fscore: 0.6278
- Accuracy: 0.6285
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 128
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 1.0
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | Fscore | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
| 0.616 | 0.29 | 2000 | 0.6498 | 0.6295 | 0.6299 | 0.6266 | 0.6267 |
| 0.6033 | 0.59 | 4000 | 0.6584 | 0.6278 | 0.6274 | 0.6228 | 0.6228 |
| 0.5896 | 0.88 | 6000 | 0.6600 | 0.6285 | 0.6293 | 0.6282 | 0.6290 |
### Framework versions
- Transformers 4.36.2
- Pytorch 2.1.2+cu121
- Datasets 2.14.5
- Tokenizers 0.15.0
| {"license": "mit", "tags": ["generated_from_trainer"], "metrics": ["precision", "recall", "accuracy"], "base_model": "DTAI-KULeuven/robbert-2023-dutch-base", "model-index": [{"name": "robbert-2023-dutch-base-gender", "results": []}]} | text-classification | clips/robbert-2023-dutch-base-gender | [
"transformers",
"safetensors",
"roberta",
"text-classification",
"generated_from_trainer",
"base_model:DTAI-KULeuven/robbert-2023-dutch-base",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T16:04:08+00:00 | [] | [] | TAGS
#transformers #safetensors #roberta #text-classification #generated_from_trainer #base_model-DTAI-KULeuven/robbert-2023-dutch-base #license-mit #autotrain_compatible #endpoints_compatible #region-us
| robbert-2023-dutch-base-gender
==============================
This model is a fine-tuned version of DTAI-KULeuven/robbert-2023-dutch-base on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 0.6591
* Precision: 0.6282
* Recall: 0.6290
* Fscore: 0.6278
* Accuracy: 0.6285
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 1e-05
* train\_batch\_size: 64
* eval\_batch\_size: 64
* seed: 42
* gradient\_accumulation\_steps: 2
* total\_train\_batch\_size: 128
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 1.0
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.36.2
* Pytorch 2.1.2+cu121
* Datasets 2.14.5
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1.0\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.36.2\n* Pytorch 2.1.2+cu121\n* Datasets 2.14.5\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #safetensors #roberta #text-classification #generated_from_trainer #base_model-DTAI-KULeuven/robbert-2023-dutch-base #license-mit #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1.0\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.36.2\n* Pytorch 2.1.2+cu121\n* Datasets 2.14.5\n* Tokenizers 0.15.0"
] | [
71,
141,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #roberta #text-classification #generated_from_trainer #base_model-DTAI-KULeuven/robbert-2023-dutch-base #license-mit #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1.0\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.36.2\n* Pytorch 2.1.2+cu121\n* Datasets 2.14.5\n* Tokenizers 0.15.0"
] | [
-0.13086581230163574,
0.07821930199861526,
-0.0012547927908599377,
0.0759502649307251,
0.1567527949810028,
0.015515098348259926,
0.1293575018644333,
0.12184786796569824,
-0.13017216324806213,
0.05912942439317703,
0.10590435564517975,
0.08668278157711029,
0.02909673936665058,
0.13178417086601257,
-0.0362340584397316,
-0.30875182151794434,
0.01106578391045332,
0.025230715051293373,
-0.16889360547065735,
0.13198162615299225,
0.1203911229968071,
-0.12577584385871887,
0.05803094431757927,
0.029527045786380768,
-0.18535134196281433,
0.010503770783543587,
0.003433637786656618,
-0.07367699593305588,
0.12118839472532272,
0.03545995056629181,
0.13142631947994232,
0.036007389426231384,
0.10210493206977844,
-0.17426499724388123,
0.0101740388199687,
0.06782687455415726,
0.025353150442242622,
0.09471776336431503,
0.07354054600000381,
-0.021703040227293968,
0.11835868656635284,
-0.07088198512792587,
0.07726021111011505,
0.03970130905508995,
-0.11748363077640533,
-0.28191888332366943,
-0.10329033434391022,
0.08780095726251602,
0.13143257796764374,
0.06393289566040039,
-0.014062770642340183,
0.1043994203209877,
-0.08871722221374512,
0.0770627036690712,
0.2291768342256546,
-0.274977445602417,
-0.09560250490903854,
-0.0010786051861941814,
0.03829753026366234,
0.022480430081486702,
-0.11932732909917831,
-0.033964600414037704,
0.06023487076163292,
0.023132136091589928,
0.09920848906040192,
0.011869196780025959,
-0.008204825222492218,
0.005549279041588306,
-0.15218864381313324,
-0.03481662645936012,
0.13795912265777588,
0.06569228321313858,
-0.05983097478747368,
-0.0427062101662159,
-0.0355544313788414,
-0.17267678678035736,
-0.039763446897268295,
-0.004347787704318762,
0.024044346064329147,
-0.06792860478162766,
-0.14813551306724548,
0.01736997254192829,
-0.09331003576517105,
-0.09800378233194351,
-0.005741526838392019,
0.21981026232242584,
0.05687350407242775,
-0.0052800788544118404,
-0.012737167067825794,
0.1419021338224411,
0.0444413460791111,
-0.1561744660139084,
0.00016662455163896084,
0.013380507007241249,
-0.04483582079410553,
-0.040484461933374405,
-0.04360441118478775,
-0.019366426393389702,
-0.0026268966030329466,
0.1759459525346756,
-0.07372535765171051,
0.029306208714842796,
0.02907419763505459,
0.022299161180853844,
-0.10081178694963455,
0.1824224293231964,
-0.07517781853675842,
-0.020396489650011063,
-0.0015290130395442247,
0.10953264683485031,
0.02093260921537876,
0.0008486994775012136,
-0.08710014820098877,
0.01332441158592701,
0.11653968691825867,
0.043156079947948456,
-0.03223428130149841,
0.034215930849313736,
-0.046150047332048416,
-0.027803989127278328,
0.0642428770661354,
-0.08183722198009491,
0.02670194022357464,
0.0052803074941039085,
-0.10467435419559479,
-0.027914628386497498,
-0.0005249577225185931,
0.019384322687983513,
0.0207339096814394,
0.16028232872486115,
-0.10392503440380096,
-0.005134605336934328,
-0.11058183014392853,
-0.11155825108289719,
0.022426530718803406,
-0.020460842177271843,
0.0023275893181562424,
-0.09562398493289948,
-0.15167099237442017,
-0.023759525269269943,
0.03890297934412956,
-0.03103785775601864,
-0.07293651252985,
-0.03367259353399277,
-0.09146220237016678,
0.036434147506952286,
-0.014822016470134258,
0.14338771998882294,
-0.04552654176950455,
0.12605924904346466,
0.07052191346883774,
0.049735479056835175,
0.00963598769158125,
0.03893806040287018,
-0.09348256886005402,
0.0443117655813694,
-0.19265595078468323,
0.03561262786388397,
-0.07035980373620987,
0.06853053718805313,
-0.10365822911262512,
-0.13081181049346924,
0.014220809563994408,
-0.009783544577658176,
0.09697821736335754,
0.11862847954034805,
-0.11890465021133423,
-0.1034439280629158,
0.15443184971809387,
-0.10101145505905151,
-0.13227565586566925,
0.1126980260014534,
-0.0324343703687191,
0.0039150649681687355,
0.04613446071743965,
0.10456015914678574,
0.10038045048713684,
-0.07287974655628204,
-0.025478187948465347,
-0.054992690682411194,
0.11410200595855713,
-0.014181329868733883,
0.1103687658905983,
-0.003291028318926692,
-0.019464395940303802,
0.017842218279838562,
-0.06598074734210968,
0.04508567601442337,
-0.12341071665287018,
-0.08616352826356888,
-0.026940852403640747,
-0.09693437069654465,
0.08778742700815201,
0.06404495984315872,
0.07127592712640762,
-0.10983970016241074,
-0.12904585897922516,
0.09678813070058823,
0.1198919415473938,
-0.059211134910583496,
0.01988127827644348,
-0.07155536860227585,
0.07935231178998947,
-0.05869276821613312,
-0.021925652399659157,
-0.17493605613708496,
-0.06806153804063797,
0.017064379528164864,
-0.021477824077010155,
0.0023569173645228148,
-0.04785192757844925,
0.07122153043746948,
0.08604075014591217,
-0.08972954750061035,
-0.06716568768024445,
-0.09733669459819794,
-0.018047267571091652,
-0.08831942081451416,
-0.2278314083814621,
-0.08713375777006149,
-0.015869714319705963,
0.16232769191265106,
-0.20964722335338593,
0.041654884815216064,
0.002053436590358615,
0.14924103021621704,
0.04262740537524223,
-0.02135196328163147,
-0.03681125119328499,
0.0904250219464302,
-0.03850700706243515,
-0.05710078775882721,
0.03252588212490082,
0.0028490147087723017,
-0.10304009914398193,
-0.02490640990436077,
-0.10125579684972763,
0.15885809063911438,
0.11395389586687088,
-0.015077830292284489,
-0.10130622237920761,
-0.008247082121670246,
-0.09361138939857483,
-0.039467763155698776,
-0.04584067687392235,
-0.0033732519950717688,
0.07916174083948135,
0.016815640032291412,
0.13506196439266205,
-0.07901337742805481,
-0.06127583980560303,
0.03284044936299324,
-0.0194097887724638,
0.003141186200082302,
0.1153317391872406,
0.0688704326748848,
-0.06340078264474869,
0.13028421998023987,
0.11387784034013748,
-0.10306253284215927,
0.14007407426834106,
-0.06713917851448059,
-0.09018012881278992,
-0.028927434235811234,
0.0014280433533713222,
0.038937367498874664,
0.1463891565799713,
-0.08849594742059708,
-0.008743883110582829,
0.011541953310370445,
0.007343711797147989,
0.01733292080461979,
-0.21482492983341217,
-0.027608955278992653,
0.0247696191072464,
-0.03577110543847084,
-0.022907579317688942,
-0.012637068517506123,
0.0052387830801308155,
0.10725394636392593,
0.017560238018631935,
-0.038823824375867844,
0.01631348580121994,
0.01150698121637106,
-0.07149933278560638,
0.2221580147743225,
-0.07685927301645279,
-0.12262549251317978,
-0.1597745567560196,
-0.008677094243466854,
-0.060363829135894775,
-0.003367532277479768,
0.040697187185287476,
-0.10963331907987595,
-0.01602255553007126,
-0.048661358654499054,
0.046760834753513336,
-0.024737318977713585,
0.04642050713300705,
-0.013407343067228794,
0.027809517458081245,
0.08696456253528595,
-0.11463482677936554,
0.029187064617872238,
-0.027871014550328255,
-0.0697370395064354,
0.03221089392900467,
0.04370163753628731,
0.1120673343539238,
0.16230519115924835,
-0.01754286140203476,
0.006844505202025175,
-0.04377729818224907,
0.1655423641204834,
-0.08531089872121811,
-0.042274925857782364,
0.13539260625839233,
0.006438383366912603,
0.04768076911568642,
0.1294020414352417,
0.06186244264245033,
-0.07639440149068832,
0.03462115302681923,
0.050188254565000534,
-0.021686019375920296,
-0.23892655968666077,
-0.022991450503468513,
-0.03815894201397896,
0.0011720050824806094,
0.10607919842004776,
0.030741862952709198,
0.019976798444986343,
0.06599155813455582,
-0.038617879152297974,
0.04750266298651695,
-0.03524177893996239,
0.09751827269792557,
0.07583782076835632,
0.06006091833114624,
0.14319710433483124,
-0.03304452449083328,
-0.05927104130387306,
0.024246886372566223,
-0.03402385860681534,
0.21458478271961212,
-0.0323030948638916,
0.08570287376642227,
0.05407450720667839,
0.1654246300458908,
0.014260286465287209,
0.07789234817028046,
0.021676402539014816,
-0.04048211872577667,
0.020250100642442703,
-0.05889284238219261,
-0.015363656915724277,
0.03150413557887077,
-0.014475654810667038,
0.09257005155086517,
-0.15507341921329498,
-0.009834565222263336,
0.04032721370458603,
0.29295143485069275,
0.08568771183490753,
-0.32690733671188354,
-0.13902351260185242,
0.008382688276469707,
-0.0511028952896595,
-0.03261216729879379,
0.0078102098777890205,
0.08400380611419678,
-0.10018840432167053,
0.05557524040341377,
-0.0853099599480629,
0.09251634776592255,
-0.018368937075138092,
-0.004588152747601271,
0.07506521791219711,
0.10308412462472916,
-0.021628297865390778,
0.05922611430287361,
-0.2553474009037018,
0.2897438406944275,
0.0006716098287142813,
0.09211508929729462,
-0.024821316823363304,
0.020182760432362556,
0.04165443778038025,
0.04245422035455704,
0.03942260146141052,
-0.019005466252565384,
-0.0633031502366066,
-0.20293016731739044,
-0.0716431587934494,
0.02496321313083172,
0.12353941053152084,
-0.0919005423784256,
0.13964560627937317,
-0.02526397630572319,
-0.014793203212320805,
0.05606139823794365,
-0.046261344105005264,
-0.11589914560317993,
-0.07048032432794571,
0.00983769167214632,
-0.02930830791592598,
0.06261073052883148,
-0.12411081790924072,
-0.13426505029201508,
-0.03511491417884827,
0.16272252798080444,
-0.07080979645252228,
-0.038231831043958664,
-0.12658265233039856,
0.12287084013223648,
0.1441088616847992,
-0.07628265768289566,
0.04826396703720093,
-0.004254498519003391,
0.13311107456684113,
0.01349963340908289,
-0.02869594469666481,
0.08969056606292725,
-0.07717625051736832,
-0.2508983612060547,
-0.06274187564849854,
0.14792142808437347,
0.02902764081954956,
0.05863138288259506,
-0.025605246424674988,
0.033397600054740906,
-0.0043769050389528275,
-0.0857555940747261,
0.029873838648200035,
-0.027501456439495087,
0.044574104249477386,
0.04319927841424942,
-0.03546936810016632,
0.013770028948783875,
-0.057081177830696106,
-0.04560122638940811,
0.1061096265912056,
0.3061007857322693,
-0.08952012658119202,
-0.025264378637075424,
0.05834786966443062,
-0.03464591130614281,
-0.16429299116134644,
0.058694563806056976,
0.1171749159693718,
0.028771517798304558,
0.0015389893669635057,
-0.18099863827228546,
0.08731470257043839,
0.10788007080554962,
-0.045567892491817474,
0.10845699906349182,
-0.27664318680763245,
-0.14471279084682465,
0.08741334080696106,
0.13432206213474274,
0.04337410628795624,
-0.1719551831483841,
-0.04999537020921707,
-0.025681255385279655,
-0.10797721892595291,
0.11169348657131195,
-0.059029899537563324,
0.10050521045923233,
-0.024285754188895226,
0.05048025771975517,
0.01234322227537632,
-0.053977012634277344,
0.14329300820827484,
-0.030955592170357704,
0.08177076280117035,
-0.012758509255945683,
-0.009160056710243225,
0.10059704631567001,
-0.062248315662145615,
-0.002475518500432372,
-0.0487741194665432,
0.04182303696870804,
-0.0941992700099945,
-0.022114552557468414,
-0.10130058228969574,
0.05420704931020737,
-0.06246989965438843,
-0.06360097974538803,
-0.035799793899059296,
0.061080820858478546,
0.0017322872299700975,
-0.02282794564962387,
0.16452370584011078,
0.013166195712983608,
0.21665994822978973,
0.08341404795646667,
0.07195836305618286,
0.0033315818291157484,
-0.025268401950597763,
0.007640501484274864,
-0.0368461087346077,
0.07249587774276733,
-0.1652161329984665,
0.014687794260680676,
0.13013479113578796,
0.06482760608196259,
0.12421763688325882,
0.07926011085510254,
-0.06220218166708946,
0.03230256214737892,
0.10532674193382263,
-0.11891993135213852,
-0.0833732932806015,
-0.0303184874355793,
0.012888580560684204,
-0.1691206991672516,
0.07400080561637878,
0.10249589383602142,
-0.0855310782790184,
-0.029143473133444786,
0.009428909979760647,
-0.007777581922709942,
-0.037480298429727554,
0.21747899055480957,
0.08657132834196091,
0.09117259085178375,
-0.08632206916809082,
0.07541195303201675,
0.053804732859134674,
-0.12030886858701706,
-0.015308630652725697,
0.06566809862852097,
-0.059472016990184784,
-0.01122742984443903,
0.013550451956689358,
0.08908527344465256,
-0.06750455498695374,
-0.049585916101932526,
-0.16820620000362396,
-0.12326779961585999,
0.06025807559490204,
0.14578568935394287,
0.07246541976928711,
0.03912634775042534,
-0.010155299678444862,
0.06267386674880981,
-0.13925336301326752,
0.12219260632991791,
0.08426369726657867,
0.10817307978868484,
-0.15558390319347382,
0.17369680106639862,
0.0012418575352057815,
0.02649800106883049,
-0.005150660406798124,
0.01825544983148575,
-0.11256659030914307,
-0.0024051712825894356,
-0.14451710879802704,
-0.06210722774267197,
-0.051303233951330185,
-0.009759747423231602,
-0.003158682258799672,
-0.05724729970097542,
-0.07423818111419678,
0.024831626564264297,
-0.11232290416955948,
-0.03957640752196312,
0.011076501570641994,
0.03742809221148491,
-0.12362999469041824,
-0.009594609960913658,
0.04610943794250488,
-0.10980743169784546,
0.07795237749814987,
0.0422351211309433,
0.06630025804042816,
0.055450428277254105,
-0.07688035070896149,
0.01828734576702118,
0.02775360457599163,
-0.024383926764130592,
0.0455051064491272,
-0.12652771174907684,
0.0043456051498651505,
-0.033519092947244644,
0.057498201727867126,
0.010341516695916653,
0.03032054379582405,
-0.14539116621017456,
-0.00785613153129816,
-0.013203930109739304,
-0.05682849511504173,
-0.05547642707824707,
0.04458572715520859,
0.06418178975582123,
0.027025289833545685,
0.1748506873846054,
-0.08392956107854843,
0.024264223873615265,
-0.2320726215839386,
-0.0033628346864134073,
-0.03757196664810181,
-0.09031985700130463,
-0.11436391621828079,
-0.02250746451318264,
0.0662536770105362,
-0.0517071932554245,
0.07885710895061493,
-0.030473798513412476,
0.10114232450723648,
0.04937031492590904,
-0.06441426277160645,
0.044389598071575165,
0.051880378276109695,
0.21284569799900055,
0.04466458782553673,
-0.022424980998039246,
0.045437972992658615,
0.030506538227200508,
0.06739594787359238,
0.07860774546861649,
0.18667131662368774,
0.11929991096258163,
-0.018867220729589462,
0.08997625857591629,
0.053724680095911026,
-0.06567691266536713,
-0.1545581966638565,
0.0294464323669672,
-0.01616017334163189,
0.09054303914308548,
-0.0041166567243635654,
0.1772710680961609,
0.14399561285972595,
-0.17938122153282166,
0.022549211978912354,
-0.03447059541940689,
-0.07320693880319595,
-0.10205313563346863,
-0.004481382202357054,
-0.08020894229412079,
-0.18882347643375397,
0.024367455393075943,
-0.12741276621818542,
0.010193931870162487,
0.07022421807050705,
0.014413589611649513,
0.006895388942211866,
0.14815068244934082,
0.07018423825502396,
0.027640443295240402,
0.09454190731048584,
0.015518196858465672,
-0.01167729590088129,
-0.015985755249857903,
-0.0886792466044426,
0.010876605287194252,
-0.042610377073287964,
0.034238144755363464,
-0.06883325427770615,
-0.11579473316669464,
0.054662447422742844,
0.031040754169225693,
-0.11159507185220718,
0.03182373568415642,
0.016795488074421883,
0.09592247754335403,
0.05262168124318123,
0.0058951182290911674,
0.0214091707020998,
-0.022297825664281845,
0.2733765244483948,
-0.09956756234169006,
-0.06127602979540825,
-0.13379283249378204,
0.29841887950897217,
0.0305323526263237,
-0.019816432148218155,
0.04288030415773392,
-0.10403207689523697,
-0.009112406522035599,
0.13439713418483734,
0.14457708597183228,
-0.011129789054393768,
-0.019910713657736778,
0.010614019818603992,
-0.02759765088558197,
-0.05608629435300827,
0.08840280026197433,
0.09443257749080658,
0.07879709452390671,
-0.0849958136677742,
-0.04286005720496178,
-0.04574080556631088,
-0.0382375530898571,
-0.003747136564925313,
0.07746180146932602,
0.021529749035835266,
-0.018098779022693634,
-0.047539349645376205,
0.09120194613933563,
-0.041579797863960266,
-0.12200284749269485,
0.07928809523582458,
-0.18770158290863037,
-0.18825459480285645,
-0.027300477027893066,
0.06062209606170654,
0.0064565581269562244,
0.07859613746404648,
-0.010601844638586044,
-0.024686157703399658,
0.07143161445856094,
-0.00843722652643919,
-0.03779034689068794,
-0.1421317458152771,
0.11073374003171921,
-0.030543535947799683,
0.2043776512145996,
-0.04897521808743477,
0.045519474893808365,
0.1326632797718048,
0.04045087471604347,
-0.1017102301120758,
0.021226519718766212,
0.07579539716243744,
-0.12293883413076401,
0.039406903088092804,
0.16526491940021515,
-0.028072450309991837,
0.10479696840047836,
0.03639228641986847,
-0.14961330592632294,
0.015570798888802528,
-0.07163846492767334,
-0.05951276421546936,
-0.08633505553007126,
0.0004316022095736116,
-0.04243799299001694,
0.12098683416843414,
0.2358609139919281,
-0.06569207459688187,
0.000737420457880944,
-0.06067411229014397,
0.02896081656217575,
0.0775570198893547,
0.10241622477769852,
-0.023016363382339478,
-0.2830873131752014,
0.019926870241761208,
0.07302338629961014,
-0.014291021972894669,
-0.2912633419036865,
-0.07474648207426071,
0.04117829352617264,
-0.06605277955532074,
-0.06907451152801514,
0.11196913570165634,
0.07904025167226791,
0.04674350097775459,
-0.05011817812919617,
-0.09744592756032944,
-0.08692199736833572,
0.17422188818454742,
-0.17271502315998077,
-0.08131639659404755
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->
# TesterGG/sequence_classification_model
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the 'act' classification labels in 'daily_dialog' dataset.
It achieves the following results on the evaluation set:
- Train Loss: 0.3870
- Validation Loss: 0.5128
- Train Accuracy: 0.8059
- Epoch: 2
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- optimizer: {'name': 'Adam', 'weight_decay': None, 'clipnorm': None, 'global_clipnorm': None, 'clipvalue': None, 'use_ema': False, 'ema_momentum': 0.99, 'ema_overwrite_frequency': None, 'jit_compile': True, 'is_legacy_optimizer': False, 'learning_rate': {'module': 'keras.optimizers.schedules', 'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 9080, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered_name': None}, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False}
- training_precision: float32
### Training results
| Train Loss | Validation Loss | Train Accuracy | Epoch |
|:----------:|:---------------:|:--------------:|:-----:|
| 0.5076 | 0.5274 | 0.7987 | 0 |
| 0.4112 | 0.5128 | 0.8059 | 1 |
| 0.3870 | 0.5128 | 0.8059 | 2 |
### Framework versions
- Transformers 4.38.0.dev0
- TensorFlow 2.15.0
- Datasets 2.16.1
- Tokenizers 0.15.1 | {"license": "apache-2.0", "tags": ["generated_from_keras_callback"], "datasets": ["daily_dialog"], "base_model": "distilbert-base-uncased", "model-index": [{"name": "TesterGG/sequence_classification_model", "results": []}]} | text-classification | TesterGG/sequence_classification_model | [
"transformers",
"tf",
"distilbert",
"text-classification",
"generated_from_keras_callback",
"dataset:daily_dialog",
"base_model:distilbert-base-uncased",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T16:04:25+00:00 | [] | [] | TAGS
#transformers #tf #distilbert #text-classification #generated_from_keras_callback #dataset-daily_dialog #base_model-distilbert-base-uncased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| TesterGG/sequence\_classification\_model
========================================
This model is a fine-tuned version of distilbert-base-uncased on the 'act' classification labels in 'daily\_dialog' dataset.
It achieves the following results on the evaluation set:
* Train Loss: 0.3870
* Validation Loss: 0.5128
* Train Accuracy: 0.8059
* Epoch: 2
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* optimizer: {'name': 'Adam', 'weight\_decay': None, 'clipnorm': None, 'global\_clipnorm': None, 'clipvalue': None, 'use\_ema': False, 'ema\_momentum': 0.99, 'ema\_overwrite\_frequency': None, 'jit\_compile': True, 'is\_legacy\_optimizer': False, 'learning\_rate': {'module': 'keras.optimizers.schedules', 'class\_name': 'PolynomialDecay', 'config': {'initial\_learning\_rate': 2e-05, 'decay\_steps': 9080, 'end\_learning\_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered\_name': None}, 'beta\_1': 0.9, 'beta\_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False}
* training\_precision: float32
### Training results
### Framework versions
* Transformers 4.38.0.dev0
* TensorFlow 2.15.0
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* optimizer: {'name': 'Adam', 'weight\\_decay': None, 'clipnorm': None, 'global\\_clipnorm': None, 'clipvalue': None, 'use\\_ema': False, 'ema\\_momentum': 0.99, 'ema\\_overwrite\\_frequency': None, 'jit\\_compile': True, 'is\\_legacy\\_optimizer': False, 'learning\\_rate': {'module': 'keras.optimizers.schedules', 'class\\_name': 'PolynomialDecay', 'config': {'initial\\_learning\\_rate': 2e-05, 'decay\\_steps': 9080, 'end\\_learning\\_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered\\_name': None}, 'beta\\_1': 0.9, 'beta\\_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False}\n* training\\_precision: float32",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* TensorFlow 2.15.0\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tf #distilbert #text-classification #generated_from_keras_callback #dataset-daily_dialog #base_model-distilbert-base-uncased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* optimizer: {'name': 'Adam', 'weight\\_decay': None, 'clipnorm': None, 'global\\_clipnorm': None, 'clipvalue': None, 'use\\_ema': False, 'ema\\_momentum': 0.99, 'ema\\_overwrite\\_frequency': None, 'jit\\_compile': True, 'is\\_legacy\\_optimizer': False, 'learning\\_rate': {'module': 'keras.optimizers.schedules', 'class\\_name': 'PolynomialDecay', 'config': {'initial\\_learning\\_rate': 2e-05, 'decay\\_steps': 9080, 'end\\_learning\\_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered\\_name': None}, 'beta\\_1': 0.9, 'beta\\_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False}\n* training\\_precision: float32",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* TensorFlow 2.15.0\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
78,
304,
4,
36
] | [
"passage: TAGS\n#transformers #tf #distilbert #text-classification #generated_from_keras_callback #dataset-daily_dialog #base_model-distilbert-base-uncased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* optimizer: {'name': 'Adam', 'weight\\_decay': None, 'clipnorm': None, 'global\\_clipnorm': None, 'clipvalue': None, 'use\\_ema': False, 'ema\\_momentum': 0.99, 'ema\\_overwrite\\_frequency': None, 'jit\\_compile': True, 'is\\_legacy\\_optimizer': False, 'learning\\_rate': {'module': 'keras.optimizers.schedules', 'class\\_name': 'PolynomialDecay', 'config': {'initial\\_learning\\_rate': 2e-05, 'decay\\_steps': 9080, 'end\\_learning\\_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered\\_name': None}, 'beta\\_1': 0.9, 'beta\\_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False}\n* training\\_precision: float32### Training results### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* TensorFlow 2.15.0\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.07774730026721954,
0.10574910044670105,
-0.008158856071531773,
0.07070817053318024,
0.11575765162706375,
0.05190892145037651,
0.08363787829875946,
0.14179934561252594,
-0.05340781435370445,
0.15820519626140594,
0.08424398303031921,
0.13995261490345,
0.04692290723323822,
0.1540057361125946,
-0.07283396273851395,
-0.1579960137605667,
0.04228971526026726,
-0.043870728462934494,
-0.04223214462399483,
0.06516209244728088,
0.06946312636137009,
-0.05634510889649391,
0.07209241390228271,
-0.03453405946493149,
-0.056136246770620346,
-0.015956133604049683,
0.012714691460132599,
-0.05022758990526199,
0.06040280684828758,
0.06957722455263138,
0.044041574001312256,
0.006845845375210047,
-0.005410291254520416,
-0.22562919557094574,
0.0027623986825346947,
0.10994421690702438,
0.012730865739285946,
0.06862630695104599,
0.04890686646103859,
-0.031099608168005943,
0.12570792436599731,
-0.1110650897026062,
0.05936483293771744,
0.036391545087099075,
-0.15162625908851624,
-0.21309004724025726,
-0.08487389981746674,
0.018886391073465347,
0.11387686431407928,
0.06588197499513626,
-0.013908346183598042,
0.11949847638607025,
-0.05551391839981079,
0.08551519364118576,
0.1293277144432068,
-0.2567313313484192,
-0.0478442944586277,
0.03043389320373535,
0.018649473786354065,
-0.01113654114305973,
-0.08850304037332535,
-0.01895212195813656,
-0.0021005570888519287,
0.01532662846148014,
0.024762049317359924,
-0.015565074048936367,
0.002459743293002248,
-0.058104630559682846,
-0.06576915830373764,
-0.06467152386903763,
0.17384783923625946,
0.06746620684862137,
-0.03301641345024109,
-0.08932512253522873,
-0.04484854266047478,
-0.14877070486545563,
-0.026904001832008362,
-0.01800176501274109,
0.013454289175570011,
0.008653407916426659,
-0.01205387618392706,
0.0036522741429507732,
-0.040127404034137726,
-0.022393105551600456,
0.03595532849431038,
0.0512087307870388,
0.042346782982349396,
-0.004631820600479841,
0.010901099070906639,
0.06798297166824341,
-0.019567320123314857,
-0.14385555684566498,
-0.039605479687452316,
-0.0007797391153872013,
-0.07068117707967758,
-0.03307866305112839,
-0.039100054651498795,
0.04230089858174324,
0.1083671972155571,
0.2138673961162567,
-0.06321843713521957,
0.1360434591770172,
-0.002332122763618827,
0.015350489877164364,
-0.06630542874336243,
0.0785406231880188,
0.012571241706609726,
-0.090126633644104,
-0.015856796875596046,
0.058878690004348755,
0.014994216151535511,
-0.03495955839753151,
-0.013867349363863468,
0.04998065531253815,
0.07326504588127136,
0.04376902058720589,
0.01515139453113079,
0.08162081241607666,
-0.07587049901485443,
-0.012544896453619003,
0.0260930173099041,
-0.13026444613933563,
0.047075092792510986,
0.061556074768304825,
-0.05250602215528488,
0.041276950389146805,
0.059300534427165985,
-0.029210271313786507,
-0.10367657989263535,
0.03938217833638191,
-0.058092568069696426,
-0.036819085478782654,
-0.06900452822446823,
-0.06870625913143158,
0.0322229377925396,
-0.071871317923069,
-0.02154369093477726,
-0.05564020201563835,
-0.1297616809606552,
-0.07388157397508621,
0.09607928991317749,
-0.05953991040587425,
-0.05829433351755142,
-0.08623868227005005,
-0.15959972143173218,
0.07614070922136307,
-0.005745348520576954,
0.09145969152450562,
-0.06696054339408875,
0.05293913558125496,
-0.007046624086797237,
0.013286626897752285,
0.05295317620038986,
0.027506239712238312,
-0.07199845463037491,
0.06257922947406769,
-0.1633993536233902,
0.12381929904222488,
-0.08322086185216904,
0.040129404515028,
-0.1573028266429901,
-0.040520891547203064,
0.035477425903081894,
0.009051051922142506,
0.09829186648130417,
0.13026905059814453,
-0.15144692361354828,
-0.05516982823610306,
0.14714586734771729,
-0.08352099359035492,
-0.09711962193250656,
0.08001886308193207,
-0.02665642276406288,
-0.020774004980921745,
0.07778452336788177,
0.10028531402349472,
0.0828220471739769,
-0.07780411839485168,
-0.019776439294219017,
-0.0545264296233654,
0.0468716099858284,
0.06242470070719719,
0.02908620610833168,
-0.07888520509004593,
-0.03512758016586304,
0.026300137862563133,
-0.020785437896847725,
-0.004496135748922825,
-0.04839181900024414,
-0.04981463402509689,
-0.045724544674158096,
-0.051224011927843094,
0.025617925450205803,
0.04071805998682976,
-0.01207748893648386,
-0.08563797175884247,
-0.17921775579452515,
0.029852116480469704,
0.04601450636982918,
-0.07950174808502197,
0.0186599250882864,
-0.06369806826114655,
0.08518249541521072,
0.054040901362895966,
0.020167961716651917,
-0.16337096691131592,
-0.10014183074235916,
0.020542189478874207,
-0.021478304639458656,
0.006309959571808577,
-0.054659150540828705,
0.044215697795152664,
0.02871682494878769,
-0.04590366408228874,
-0.033437687903642654,
-0.00027236680034548044,
0.009525690227746964,
-0.040196649730205536,
-0.23158986866474152,
-0.012691738083958626,
-0.014623722992837429,
0.10316125303506851,
-0.27434220910072327,
0.002929740585386753,
0.03898952528834343,
0.14677326381206512,
0.03050752356648445,
-0.04749526455998421,
-0.010429074056446552,
0.04282578080892563,
-0.029324911534786224,
-0.07169435173273087,
0.03351202607154846,
0.02319478802382946,
-0.10066024959087372,
-0.04306202009320259,
-0.18013857305049896,
0.04139450937509537,
0.10107895731925964,
-0.04476324841380119,
-0.13567303121089935,
0.020758219063282013,
-0.018782880157232285,
-0.034971315413713455,
-0.009589316323399544,
0.026926314458251,
0.14847970008850098,
0.03994465619325638,
0.11552432924509048,
-0.037545520812273026,
-0.04879908263683319,
0.011205056682229042,
-0.028247525915503502,
-0.011349903419613838,
0.14346973598003387,
0.011062300764024258,
-0.1060786172747612,
0.09918452799320221,
0.09688889980316162,
-0.07952714711427689,
0.11919324100017548,
-0.06402672827243805,
-0.06418436765670776,
-0.08144456148147583,
0.08126837015151978,
0.0488063208758831,
0.03200596570968628,
-0.11604169756174088,
0.026255950331687927,
0.01742420718073845,
0.01885264366865158,
-0.014719326980412006,
-0.09984295070171356,
0.04335372895002365,
-0.008168218657374382,
-0.05202748626470566,
0.08969085663557053,
0.0033803631085902452,
0.010502864606678486,
0.09425188601016998,
0.019448377192020416,
-0.02336723357439041,
0.025873398408293724,
-0.023847948759794235,
-0.09050871431827545,
0.21972328424453735,
-0.11854285001754761,
-0.11627321690320969,
-0.07218927145004272,
-0.027697192505002022,
-0.05304312705993652,
-0.025656849145889282,
0.02315293252468109,
-0.07154736667871475,
-0.06711115688085556,
-0.07540933787822723,
-0.011978589929640293,
0.015205079689621925,
-0.002695584436878562,
-0.0012909133220091462,
0.029391437768936157,
0.13169832527637482,
-0.09689055383205414,
-0.033268630504608154,
0.00032280437881127,
-0.0853104293346405,
-0.01710418425500393,
0.043273746967315674,
0.00936448760330677,
0.1180504709482193,
0.032368507236242294,
0.015499059110879898,
-0.018422819674015045,
0.21976599097251892,
-0.07317116856575012,
0.030370116233825684,
0.10031144320964813,
-0.03286753594875336,
0.06950006633996964,
0.1577829122543335,
0.04512891173362732,
-0.0991876870393753,
0.009688668884336948,
0.093399778008461,
0.0036876739468425512,
-0.24730394780635834,
-0.048101361840963364,
-0.05470673367381096,
-0.04066646844148636,
0.08526456356048584,
0.05795472487807274,
0.0794835090637207,
0.015197096392512321,
-0.02446240931749344,
0.03875012695789337,
0.07072137296199799,
0.08396655321121216,
0.14037275314331055,
0.08114694803953171,
0.09018979221582413,
-0.0341620072722435,
0.01292745117098093,
0.028431406244635582,
0.0070855156518518925,
0.2015482634305954,
0.005227058660238981,
0.10571970790624619,
0.10054738819599152,
0.05946933478116989,
-0.027968116104602814,
-0.023330580443143845,
0.005535797215998173,
0.028657589107751846,
0.02923249825835228,
-0.06527862697839737,
-0.046434417366981506,
0.020658105611801147,
0.030992677435278893,
0.03842128813266754,
-0.09256678074598312,
0.03555486351251602,
0.08978327363729477,
0.20203237235546112,
0.13751885294914246,
-0.2890785038471222,
-0.0866565853357315,
0.02068001590669155,
-0.03887647017836571,
-0.05784115567803383,
-0.0008140064892359078,
0.05405097082257271,
-0.07771382480859756,
0.11094246059656143,
-0.035263799130916595,
0.06929757446050644,
-0.1032339483499527,
0.03274131193757057,
0.07069571316242218,
0.09534201771020889,
0.026085911318659782,
0.021584372967481613,
-0.28893136978149414,
0.24635347723960876,
0.004962549079209566,
0.11974135786294937,
-0.03919919207692146,
0.07814405858516693,
0.03663793206214905,
-0.04894234985113144,
0.10031946748495102,
-0.018766915425658226,
-0.12738655507564545,
-0.13499009609222412,
-0.07052647322416306,
0.003131621051579714,
0.08923080563545227,
-0.0521479956805706,
0.09552870690822601,
-0.03683444857597351,
-0.027481796219944954,
0.020078832283616066,
-0.03261717036366463,
-0.1759224236011505,
-0.10694216191768646,
0.056086938828229904,
-0.00690166000276804,
-0.020566698163747787,
-0.044338829815387726,
-0.03244609758257866,
-0.05043283477425575,
0.23246490955352783,
-0.16329094767570496,
-0.06486227363348007,
-0.1300567388534546,
0.06876814365386963,
0.11325623840093613,
-0.0832439661026001,
0.049104977399110794,
-0.010786480270326138,
0.07982037961483002,
0.06281442195177078,
-0.0650440901517868,
0.1172749325633049,
-0.03189525380730629,
-0.19491983950138092,
-0.06986667215824127,
0.13358774781227112,
0.04330341890454292,
0.020376399159431458,
-0.0006148128886707127,
0.073126420378685,
0.038078129291534424,
-0.09839823842048645,
0.06577574461698532,
0.04149354249238968,
0.03909693658351898,
0.03971763327717781,
-0.03668244555592537,
-0.07664118707180023,
-0.0464867502450943,
0.007459315937012434,
0.0769202783703804,
0.2728872299194336,
-0.07466403394937515,
0.04267203435301781,
0.040099553763866425,
-0.11493334919214249,
-0.16987989842891693,
-0.004121795296669006,
0.10813484340906143,
-0.020320141687989235,
-0.05453525483608246,
-0.18128788471221924,
0.09678055346012115,
0.1183432936668396,
0.0003874623216688633,
0.05462620034813881,
-0.2852826416492462,
-0.1391339898109436,
0.0797504186630249,
0.07488420605659485,
-0.04425739869475365,
-0.1858251541852951,
-0.08200868219137192,
-0.036868032068014145,
-0.06318087875843048,
0.14203697443008423,
-0.056579697877168655,
0.08349967747926712,
0.0338272862136364,
-0.02678772620856762,
0.027716564014554024,
-0.02826942503452301,
0.142872616648674,
0.014661709778010845,
0.05028991773724556,
-0.06679965555667877,
0.005560463294386864,
0.07379142940044403,
-0.11729700863361359,
0.04313766211271286,
-0.080410435795784,
0.03266455978155136,
-0.13020244240760803,
-0.006608844734728336,
-0.06415785849094391,
0.06616725772619247,
-0.07912872731685638,
0.006487547419965267,
-0.010030671954154968,
0.05231539532542229,
0.1143684983253479,
0.009773015044629574,
0.11852502822875977,
-0.02032836340367794,
0.18056131899356842,
0.11722438782453537,
0.09351534396409988,
0.03405963256955147,
-0.12972185015678406,
0.04329216852784157,
-0.013614975847303867,
0.0336824506521225,
-0.17023637890815735,
0.05185184255242348,
0.12854228913784027,
0.0009154501021839678,
0.1254473477602005,
0.04740287736058235,
-0.053609129041433334,
0.013077101670205593,
0.06723110377788544,
-0.11405894160270691,
-0.07746586948633194,
0.012705037370324135,
-0.005289880558848381,
-0.09047608822584152,
-0.005854452028870583,
0.15398161113262177,
-0.013797057792544365,
0.022682448849081993,
0.020198417827486992,
0.06536699831485748,
-0.048144709318876266,
0.14230000972747803,
-0.016358960419893265,
0.08674614131450653,
-0.08674901723861694,
0.1382737010717392,
0.09850294888019562,
-0.12548089027404785,
0.12262873351573944,
0.09717957675457001,
-0.06139861419796944,
-0.05004679784178734,
0.0379842109978199,
0.1297493726015091,
0.057716768234968185,
-0.03277844190597534,
-0.06270662695169449,
-0.12281104922294617,
0.10617484897375107,
0.12139375507831573,
0.016024142503738403,
0.05369870737195015,
-0.019209319725632668,
-0.016171323135495186,
-0.08217369019985199,
0.10431857407093048,
0.08896709233522415,
0.04613083600997925,
-0.09917156398296356,
0.10854923725128174,
0.016859600320458412,
-0.0448724739253521,
0.016779251396656036,
-0.0044388240203261375,
-0.18704979121685028,
-0.009175078943371773,
-0.07630226761102676,
0.048298295587301254,
-0.01717519946396351,
0.005708232056349516,
0.023297324776649475,
-0.02530379593372345,
-0.05239950865507126,
0.026435308158397675,
-0.08370426297187805,
-0.07772787660360336,
0.02275405079126358,
0.10649572312831879,
-0.13770386576652527,
-0.037507086992263794,
0.017745764926075935,
-0.13629859685897827,
0.06358827650547028,
0.014984090812504292,
0.0027232978027313948,
0.019816366955637932,
-0.11838747560977936,
-0.001095362938940525,
0.01746233180165291,
0.006112504750490189,
0.017302609980106354,
-0.1649717539548874,
0.031738150864839554,
-0.03292548283934593,
0.01648794300854206,
0.006148680113255978,
0.05733721703290939,
-0.11062226444482803,
-0.03579737991094589,
-0.016438227146863937,
-0.033939428627491,
-0.042370617389678955,
0.0352480374276638,
0.1373472362756729,
-0.037345513701438904,
0.15728172659873962,
-0.08956263214349747,
0.04723939672112465,
-0.1940249353647232,
-0.02448590099811554,
0.03606433793902397,
-0.06534770131111145,
-0.077923983335495,
-0.006822157651185989,
0.12148495763540268,
-0.08728362619876862,
0.05329954996705055,
-0.051538169384002686,
0.08766894787549973,
0.03239377215504646,
-0.0763297975063324,
-0.08697157353162766,
0.09401517361402512,
0.152584969997406,
0.06299066543579102,
-0.01255835872143507,
0.04781258851289749,
-0.04341723769903183,
0.04980199784040451,
0.05399581044912338,
0.16998319327831268,
0.11354082822799683,
0.021004365757107735,
0.07617626339197159,
0.04668513685464859,
-0.11961714178323746,
-0.11405239999294281,
0.16219006478786469,
-0.07996246218681335,
0.18757817149162292,
-0.04469531401991844,
0.07492951303720474,
0.04162956774234772,
-0.17328421771526337,
0.044555291533470154,
-0.0637274757027626,
-0.10226859897375107,
-0.09937252104282379,
-0.1577668935060501,
-0.09524936228990555,
-0.08318337053060532,
0.0100284768268466,
-0.12653179466724396,
0.033230215311050415,
0.07994416356086731,
0.02684730477631092,
0.011508842930197716,
0.055379558354616165,
-0.016864432021975517,
0.008110295981168747,
0.10479887574911118,
0.012879331596195698,
-0.00996268168091774,
-0.04234303534030914,
-0.05482610687613487,
0.014239970594644547,
0.029740076512098312,
0.03615067899227142,
0.03620925545692444,
0.0012871702201664448,
0.07132912427186966,
-0.008692658506333828,
-0.08867858350276947,
0.0647398829460144,
0.015261881053447723,
-0.015384726226329803,
0.08580727875232697,
0.03769887983798981,
-0.042476460337638855,
-0.007515812758356333,
0.15037016570568085,
-0.06919574737548828,
-0.06396187841892242,
-0.17344078421592712,
0.20247825980186462,
0.010695491917431355,
0.010110432282090187,
0.02182510681450367,
-0.07535265386104584,
-0.017515011131763458,
0.1354520320892334,
0.12645772099494934,
0.002864477690309286,
-0.010618750937283039,
0.07611201703548431,
-0.00928711798042059,
-0.017264623194932938,
0.11458922177553177,
0.07659166306257248,
0.038846999406814575,
-0.03590790554881096,
-0.01938514970242977,
0.013320986181497574,
-0.046116143465042114,
-0.07634935528039932,
0.08627268671989441,
-0.004223949275910854,
-0.007549954112619162,
0.001459969556890428,
0.08289570361375809,
-0.10777878016233444,
-0.1294747292995453,
0.08414226770401001,
-0.20360635221004486,
-0.18372485041618347,
-0.044721417129039764,
0.019056377932429314,
0.016930928453803062,
0.03519798070192337,
0.02153228409588337,
-0.04542827233672142,
0.14694492518901825,
-0.0374617725610733,
-0.02826634980738163,
-0.10188255459070206,
-0.003834058763459325,
-0.023029014468193054,
0.22251223027706146,
0.0034173058811575174,
0.01290206890553236,
0.15062187612056732,
0.02619619108736515,
-0.1035720705986023,
0.03733059763908386,
0.09458325058221817,
-0.09402383118867874,
0.06961144506931305,
0.09008972346782684,
-0.03379746526479721,
0.16478106379508972,
0.10525649785995483,
-0.0813903957605362,
0.0027109053917229176,
0.011385582387447357,
-0.03450242429971695,
-0.036573171615600586,
-0.0045868465676903725,
-0.07512184232473373,
0.12173906713724136,
0.2263403981924057,
-0.04289817810058594,
-0.011139795184135437,
-0.03526294603943825,
0.039108797907829285,
0.016977490857243538,
0.035369593650102615,
-0.04800283536314964,
-0.21897096931934357,
0.09267503768205643,
0.030723396688699722,
0.07415258884429932,
-0.11965161561965942,
-0.08723890781402588,
0.03015359863638878,
-0.002416804200038314,
-0.11028171330690384,
0.13323338329792023,
0.04768044129014015,
0.018203740939497948,
-0.06721846014261246,
-0.12574872374534607,
-0.02961435355246067,
0.18545964360237122,
-0.10528284311294556,
-0.07222872227430344
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# hubert_RTSPsplit_0208_2
This model is a fine-tuned version of [rinna/japanese-hubert-base](https://huggingface.co/rinna/japanese-hubert-base) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.1511
- Wer: 0.1981
- Cer: 0.0674
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.00025
- train_batch_size: 32
- eval_batch_size: 32
- seed: 4
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 1000
- num_epochs: 40
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|
| 12.328 | 1.0 | 60 | 10.3374 | 1.0 | 1.0 |
| 6.4126 | 2.0 | 120 | 5.9206 | 1.0 | 1.0 |
| 4.7432 | 3.0 | 180 | 4.5414 | 1.0 | 1.0 |
| 3.6044 | 4.0 | 240 | 3.4580 | 1.0 | 1.0 |
| 3.0143 | 5.0 | 300 | 2.9594 | 1.0 | 1.0 |
| 2.2541 | 6.0 | 360 | 2.1267 | 1.0 | 0.9106 |
| 1.4951 | 7.0 | 420 | 1.3833 | 1.0 | 0.5168 |
| 1.1808 | 8.0 | 480 | 1.0881 | 1.0 | 0.5387 |
| 0.9955 | 9.0 | 540 | 0.8620 | 0.6715 | 0.4108 |
| 0.8934 | 10.0 | 600 | 0.8027 | 0.6741 | 0.4132 |
| 0.8321 | 11.0 | 660 | 0.7528 | 0.7162 | 0.4529 |
| 0.866 | 12.0 | 720 | 0.7082 | 0.6592 | 0.4064 |
| 0.7158 | 13.0 | 780 | 0.6850 | 0.6469 | 0.4024 |
| 0.6976 | 14.0 | 840 | 0.6642 | 0.6734 | 0.4207 |
| 0.6829 | 15.0 | 900 | 0.6459 | 0.6689 | 0.4104 |
| 0.6748 | 16.0 | 960 | 0.6891 | 0.7162 | 0.4496 |
| 0.6816 | 17.0 | 1020 | 0.7691 | 0.6678 | 0.4168 |
| 1.2144 | 18.0 | 1080 | 0.5982 | 0.6149 | 0.3510 |
| 0.6285 | 19.0 | 1140 | 0.5844 | 0.6879 | 0.4058 |
| 0.5805 | 20.0 | 1200 | 0.5947 | 0.5981 | 0.3338 |
| 0.5707 | 21.0 | 1260 | 0.5202 | 0.5568 | 0.2629 |
| 0.6479 | 22.0 | 1320 | 0.6423 | 0.6495 | 0.2831 |
| 0.5267 | 23.0 | 1380 | 0.4946 | 0.5609 | 0.2582 |
| 0.5049 | 24.0 | 1440 | 0.4816 | 0.5255 | 0.2463 |
| 0.4922 | 25.0 | 1500 | 0.4518 | 0.5285 | 0.2442 |
| 0.4669 | 26.0 | 1560 | 0.4594 | 0.5151 | 0.2329 |
| 0.4503 | 27.0 | 1620 | 0.4018 | 0.5058 | 0.2227 |
| 0.447 | 28.0 | 1680 | 0.3785 | 0.4682 | 0.1980 |
| 0.3744 | 29.0 | 1740 | 0.3351 | 0.3847 | 0.1432 |
| 0.3516 | 30.0 | 1800 | 0.2866 | 0.3456 | 0.1238 |
| 0.335 | 31.0 | 1860 | 0.2582 | 0.3128 | 0.1152 |
| 0.3282 | 32.0 | 1920 | 0.2578 | 0.2987 | 0.1063 |
| 0.3113 | 33.0 | 1980 | 0.2272 | 0.2436 | 0.0850 |
| 0.2812 | 34.0 | 2040 | 0.2112 | 0.2410 | 0.0846 |
| 0.3105 | 35.0 | 2100 | 0.1911 | 0.2253 | 0.0774 |
| 0.2225 | 36.0 | 2160 | 0.1751 | 0.2089 | 0.0719 |
| 0.2351 | 37.0 | 2220 | 0.1838 | 0.2291 | 0.0781 |
| 0.2028 | 38.0 | 2280 | 0.1583 | 0.2037 | 0.0686 |
| 0.217 | 39.0 | 2340 | 0.1509 | 0.1918 | 0.0651 |
| 0.2698 | 40.0 | 2400 | 0.1511 | 0.1981 | 0.0674 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.14.6
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["wer"], "base_model": "rinna/japanese-hubert-base", "model-index": [{"name": "hubert_RTSPsplit_0208_2", "results": []}]} | automatic-speech-recognition | tndklab/hubert_RTSPsplit_0208_2 | [
"transformers",
"safetensors",
"wav2vec2",
"automatic-speech-recognition",
"generated_from_trainer",
"base_model:rinna/japanese-hubert-base",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-07T16:04:27+00:00 | [] | [] | TAGS
#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-rinna/japanese-hubert-base #license-apache-2.0 #endpoints_compatible #region-us
| hubert\_RTSPsplit\_0208\_2
==========================
This model is a fine-tuned version of rinna/japanese-hubert-base on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.1511
* Wer: 0.1981
* Cer: 0.0674
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.00025
* train\_batch\_size: 32
* eval\_batch\_size: 32
* seed: 4
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 1000
* num\_epochs: 40
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.14.6
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.00025\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 40",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-rinna/japanese-hubert-base #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.00025\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 40",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
69,
116,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-rinna/japanese-hubert-base #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.00025\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 40### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
-0.09076754748821259,
0.09650047868490219,
-0.0033428561873733997,
0.08424907922744751,
0.09868137538433075,
-0.02092755027115345,
0.159502774477005,
0.14625120162963867,
-0.08279230445623398,
0.08301323652267456,
0.11080119758844376,
0.11794930696487427,
0.036990974098443985,
0.17098651826381683,
-0.07055514305830002,
-0.2638213634490967,
0.06775746494531631,
0.02182844467461109,
0.02323601022362709,
0.1202562004327774,
0.09016687422990799,
-0.12968774139881134,
0.07683330774307251,
0.018725385889410973,
-0.1415586918592453,
-0.0005481559783220291,
0.017017921432852745,
-0.11343399435281754,
0.1161469891667366,
0.0006469659856520593,
0.07444359362125397,
0.046730875968933105,
0.07879847288131714,
-0.20834772288799286,
0.0047084069810807705,
0.0334184467792511,
0.02232973463833332,
0.06521433591842651,
0.03220049664378166,
-0.010932867415249348,
0.08635362237691879,
-0.08046294003725052,
0.05938517302274704,
0.03734159097075462,
-0.1184881180524826,
-0.2750934958457947,
-0.08543279021978378,
0.04364733770489693,
0.1020745038986206,
0.07460778206586838,
-0.02363527938723564,
0.15434154868125916,
-0.03545349836349487,
0.09986387938261032,
0.2773304879665375,
-0.3426224887371063,
-0.04521939903497696,
-0.016537832096219063,
0.05831783264875412,
0.06824740022420883,
-0.09751679003238678,
-0.010444448329508305,
0.051367372274398804,
0.025206033140420914,
0.12729260325431824,
-0.029278194531798363,
-0.04978132247924805,
-0.01198301650583744,
-0.13151754438877106,
-0.030727224424481392,
0.14924085140228271,
0.04110560193657875,
-0.06572528183460236,
-0.09770222008228302,
-0.06409089267253876,
-0.16038671135902405,
-0.06597151607275009,
-0.042031221091747284,
0.05187009647488594,
-0.044124819338321686,
-0.07119421660900116,
-0.010516375303268433,
-0.06668243557214737,
-0.08210305869579315,
-0.03156432881951332,
0.2252018302679062,
0.05418338254094124,
0.00023510772734880447,
-0.03257542476058006,
0.0601612813770771,
-0.05911577120423317,
-0.16069377958774567,
-0.031091604381799698,
0.022678086534142494,
0.021006325259804726,
-0.013912255875766277,
-0.027761202305555344,
-0.05756629258394241,
0.040323615074157715,
0.16373705863952637,
-0.12206094712018967,
0.08184075355529785,
-0.035283736884593964,
0.01649189367890358,
-0.09859190136194229,
0.18269869685173035,
-0.02153027430176735,
-0.016051873564720154,
0.009559266269207,
0.08171551674604416,
0.061938073486089706,
-0.015738358721137047,
-0.0938335433602333,
0.020426105707883835,
0.11515917629003525,
0.049287088215351105,
-0.08233338594436646,
0.07076936960220337,
-0.034688062965869904,
0.007837841287255287,
0.0214079562574625,
-0.12311156839132309,
0.02697054110467434,
0.010212010703980923,
-0.06819657981395721,
-0.038666266947984695,
-0.009780277498066425,
0.010002934373915195,
-0.0172134842723608,
0.06571757048368454,
-0.06107781454920769,
0.020252089947462082,
-0.04711335524916649,
-0.11604462563991547,
0.010129078291356564,
-0.11582490056753159,
0.008928044699132442,
-0.10997232049703598,
-0.13179393112659454,
0.0005179750733077526,
0.03331257030367851,
-0.0457293801009655,
0.006465911865234375,
-0.0847645178437233,
-0.0958871990442276,
0.03450998291373253,
-0.026085926219820976,
0.005918074399232864,
-0.08305571973323822,
0.09000366181135178,
0.08087744563817978,
0.09626598656177521,
-0.020353903993964195,
0.022401956841349602,
-0.10113294422626495,
0.03260631486773491,
-0.22008277475833893,
0.04611906036734581,
-0.06748747080564499,
0.05170056223869324,
-0.10947106778621674,
-0.08673832565546036,
0.02019610069692135,
-0.0013639226090162992,
0.07877123355865479,
0.12721586227416992,
-0.1661471426486969,
-0.07161231338977814,
0.22067014873027802,
-0.12198273837566376,
-0.1217501237988472,
0.12459380179643631,
-0.03701808676123619,
0.04827626422047615,
0.06962155550718307,
0.26579299569129944,
0.0033886400051414967,
-0.1477050483226776,
-0.009249582886695862,
-0.057774245738983154,
0.03343565762042999,
0.0027675293385982513,
0.05907359719276428,
-0.009561100043356419,
0.04167183116078377,
0.02618139423429966,
-0.024510744959115982,
0.011248024180531502,
-0.07831777632236481,
-0.08467549085617065,
-0.04913467913866043,
-0.08772718906402588,
0.03219754621386528,
0.02104218304157257,
0.0596381314098835,
-0.1414564996957779,
-0.09510131925344467,
0.009114265441894531,
0.07967653125524521,
-0.09519398957490921,
0.047460075467824936,
-0.13422074913978577,
0.09082935005426407,
-0.027988851070404053,
-0.0071525974199175835,
-0.14752285182476044,
0.03215097635984421,
0.05309880152344704,
-0.022841084748506546,
0.026859527453780174,
-0.08658614009618759,
0.08021628111600876,
0.06701115518808365,
-0.05696418881416321,
-0.05557432770729065,
-0.007665877230465412,
0.019883280619978905,
-0.06917889416217804,
-0.19578348100185394,
-0.012857132591307163,
-0.045780811458826065,
0.0583973191678524,
-0.1511320173740387,
0.03644385188817978,
0.05483260750770569,
0.10416541993618011,
0.04631131514906883,
-0.016832875087857246,
-0.002427011262625456,
0.060403041541576385,
-0.030751878395676613,
-0.056479763239622116,
0.0392739363014698,
0.008541484363377094,
-0.08372173458337784,
0.024052277207374573,
-0.17099301517009735,
0.15429162979125977,
0.14338897168636322,
0.008190275169909,
-0.07147093117237091,
0.004997055511921644,
-0.03053557686507702,
-0.02288966067135334,
-0.025797948241233826,
0.022867990657687187,
0.14865577220916748,
-0.01391206867992878,
0.14144639670848846,
-0.10951705276966095,
-0.014976494945585728,
0.05980069562792778,
-0.05280771106481552,
-0.005313358269631863,
0.09506470710039139,
0.012865600176155567,
-0.09282302856445312,
0.12431523948907852,
0.11715414375066757,
-0.07742686569690704,
0.13277654349803925,
-0.061990056186914444,
-0.053085893392562866,
-0.038096923381090164,
0.026637643575668335,
0.021416286006569862,
0.12279482185840607,
-0.1218685582280159,
-0.030751559883356094,
0.013119139708578587,
0.021711573004722595,
-0.006022168789058924,
-0.1947011649608612,
-0.0008480838732793927,
0.027256306260824203,
-0.09174544364213943,
-0.020446449518203735,
0.011891341768205166,
-0.00633715745061636,
0.10245800763368607,
-0.004402608145028353,
-0.09503260999917984,
0.015922600403428078,
-0.000706344551872462,
-0.06695152074098587,
0.17280516028404236,
-0.1032729521393776,
-0.164493128657341,
-0.0972789078950882,
-0.09136322885751724,
-0.05287499353289604,
0.026287397369742393,
0.09286163002252579,
-0.11361353099346161,
-0.04998425394296646,
-0.11768829822540283,
-0.011328049004077911,
0.031144099310040474,
0.03829444572329521,
0.038282137364149094,
0.0008093942306004465,
0.07277143746614456,
-0.11199160665273666,
-0.0268242247402668,
-0.03612270578742027,
0.0009095185669139028,
0.04101545736193657,
0.028809502720832825,
0.11598572880029678,
0.12370479851961136,
-0.01784447953104973,
0.03705060854554176,
-0.041898149996995926,
0.2132953405380249,
-0.08003704994916916,
-0.01811544969677925,
0.12412244081497192,
-0.009297134354710579,
0.04665033891797066,
0.15776418149471283,
0.04235414043068886,
-0.11828933656215668,
-0.000004179509687673999,
-0.004547055345028639,
-0.03976408764719963,
-0.208200603723526,
-0.02529434859752655,
-0.030825819820165634,
0.01285544503480196,
0.08601894229650497,
0.03875606507062912,
0.04723159596323967,
0.031910039484500885,
0.02762633189558983,
0.008648223243653774,
0.030122920870780945,
0.10462893545627594,
0.11685993522405624,
0.04964425787329674,
0.12043797969818115,
-0.04485546424984932,
-0.05032942444086075,
0.02544115111231804,
0.008554297499358654,
0.20379450917243958,
0.031524624675512314,
0.16480909287929535,
0.04754643514752388,
0.16421902179718018,
0.035905685275793076,
0.05339624732732773,
-0.00014810424181632698,
-0.021624311804771423,
-0.004989124368876219,
-0.065479576587677,
-0.026477476581931114,
0.04993575066328049,
-0.050491053611040115,
0.040491703897714615,
-0.1051715537905693,
0.025360124185681343,
0.06700073182582855,
0.3053057789802551,
0.03769852593541145,
-0.3353784382343292,
-0.08466922491788864,
0.012020593509078026,
-0.06682687252759933,
-0.01894277147948742,
0.06137808412313461,
0.1157044768333435,
-0.057207949459552765,
0.08062201738357544,
-0.053610265254974365,
0.078513503074646,
-0.039375800639390945,
0.04137694835662842,
0.021719451993703842,
0.06565678864717484,
-0.0041731479577720165,
0.021358724683523178,
-0.2672540247440338,
0.2925756573677063,
0.026840442791581154,
0.09990954399108887,
-0.03799224644899368,
0.00482309702783823,
0.024224141612648964,
0.04169187322258949,
0.12100731581449509,
-0.03424684330821037,
-0.1382378786802292,
-0.16999147832393646,
-0.06684046983718872,
0.021379034966230392,
0.1398986279964447,
0.02685120515525341,
0.11325069516897202,
-0.013440483249723911,
-0.02028021775186062,
0.047678928822278976,
-0.08361915498971939,
-0.10384678840637207,
-0.06517406553030014,
-0.018302500247955322,
0.0970185324549675,
0.03229420632123947,
-0.06617948412895203,
-0.07938079535961151,
-0.09486231952905655,
0.11777379363775253,
-0.03138347342610359,
-0.031805213540792465,
-0.10238923132419586,
-0.017268817871809006,
0.0995723232626915,
-0.07308880984783173,
0.05720553174614906,
0.0038140772376209497,
0.10109230130910873,
0.009897560812532902,
-0.04123402386903763,
0.11968803405761719,
-0.06922341138124466,
-0.18261907994747162,
-0.045109279453754425,
0.1501799076795578,
0.018063733354210854,
0.042182084172964096,
0.015036439523100853,
0.035241078585386276,
0.012360399588942528,
-0.06739451736211777,
0.040463436394929886,
0.0016652561025694013,
0.04370736703276634,
-0.015814252197742462,
-0.037909768521785736,
-0.03911634162068367,
-0.1065296158194542,
-0.02880173735320568,
0.15052948892116547,
0.31075647473335266,
-0.08249593526124954,
0.0740731805562973,
0.09619767963886261,
-0.04227391257882118,
-0.1565513163805008,
-0.010144059546291828,
0.027306824922561646,
0.006517363246530294,
-0.011724102310836315,
-0.1439572423696518,
0.03592405095696449,
0.0722443237900734,
-0.0343514010310173,
0.06893078982830048,
-0.2755347788333893,
-0.14260803163051605,
0.11895620822906494,
0.12946541607379913,
0.10634376108646393,
-0.14984147250652313,
-0.05333651229739189,
-0.020232046023011208,
-0.114122174680233,
0.06753174960613251,
-0.09812813252210617,
0.12219687551259995,
-0.01498604565858841,
0.05178532749414444,
0.012704776600003242,
-0.050019390881061554,
0.12111520022153854,
0.004216216970235109,
0.08872837573289871,
-0.05490025877952576,
0.0037074561696499586,
0.052155859768390656,
-0.06886352598667145,
0.07776550948619843,
-0.098098523914814,
0.03724490851163864,
-0.03720570728182793,
-0.03295565024018288,
-0.06607866287231445,
0.013052066788077354,
-0.008056010119616985,
-0.05031762644648552,
-0.05136627331376076,
0.013505234383046627,
0.05642545968294144,
-0.01077979989349842,
0.17928393185138702,
-0.01344696432352066,
0.1387026309967041,
0.17939114570617676,
0.09950783103704453,
-0.10447637736797333,
-0.008546764962375164,
0.02116118185222149,
-0.04371077939867973,
0.06841384619474411,
-0.13557465374469757,
0.055808909237384796,
0.10333482921123505,
0.013483012095093727,
0.1441841423511505,
0.05373626574873924,
-0.054046470671892166,
0.03971727937459946,
0.06024925038218498,
-0.15721353888511658,
-0.12320055067539215,
0.01612766832113266,
-0.003427485702559352,
-0.05756830796599388,
0.09543567895889282,
0.12863288819789886,
-0.07225392758846283,
-0.005671200342476368,
-0.01507823821157217,
0.024134501814842224,
-0.03480587527155876,
0.18110120296478271,
0.0357988178730011,
0.05396825075149536,
-0.1066848635673523,
0.09100915491580963,
0.03014669194817543,
-0.10466565936803818,
0.06372266262769699,
0.08244134485721588,
-0.10948849469423294,
-0.024639222770929337,
0.020477788522839546,
0.13083231449127197,
-0.0036569905932992697,
-0.08208739757537842,
-0.16799327731132507,
-0.13056683540344238,
0.0601874515414238,
0.21397186815738678,
0.07090865820646286,
0.010078741237521172,
-0.024619996547698975,
0.02223975770175457,
-0.11493024975061417,
0.0896109789609909,
0.047990504652261734,
0.0652262344956398,
-0.14332805573940277,
0.10191895067691803,
0.017509564757347107,
0.015625476837158203,
-0.02564758062362671,
0.008697699755430222,
-0.12364467233419418,
0.02818440832197666,
-0.1287078708410263,
0.019236918538808823,
-0.0631856694817543,
-0.0034915076103061438,
-0.0041174194775521755,
-0.06012892350554466,
-0.07174940407276154,
0.02974560484290123,
-0.10296472162008286,
-0.007540683727711439,
0.023884933441877365,
0.04289931803941727,
-0.12886862456798553,
-0.017154347151517868,
0.020528584718704224,
-0.0876479521393776,
0.0786827877163887,
0.07217193394899368,
-0.04153727740049362,
0.06056392192840576,
-0.09446529299020767,
-0.0050223651342093945,
0.09406949579715729,
-0.010635033249855042,
0.038256727159023285,
-0.12999311089515686,
-0.020699549466371536,
0.02389759011566639,
0.04417278617620468,
0.02062789537012577,
0.11453025043010712,
-0.10760525614023209,
0.01769082620739937,
-0.039378631860017776,
-0.056341756135225296,
-0.06040617823600769,
0.028662171214818954,
0.13391759991645813,
-0.004385863430798054,
0.17714865505695343,
-0.10838685929775238,
0.011343085207045078,
-0.16391178965568542,
0.0014243748737499118,
-0.00818479061126709,
-0.13408461213111877,
-0.10864001512527466,
-0.026173124089837074,
0.061637211591005325,
-0.06841031461954117,
0.13555201888084412,
-0.022439811378717422,
0.019817326217889786,
0.04452769085764885,
-0.09903115034103394,
-0.0400998517870903,
0.0350247398018837,
0.21636556088924408,
0.03611231967806816,
-0.051439713686704636,
0.05002468824386597,
0.006056714337319136,
0.09136321395635605,
0.08385773748159409,
0.1491410881280899,
0.18283973634243011,
0.04182782396674156,
0.13473746180534363,
0.06388287991285324,
-0.0317324660718441,
-0.16068775951862335,
0.07656707614660263,
-0.058414846658706665,
0.10364843159914017,
0.0037498031742870808,
0.21509389579296112,
0.15871641039848328,
-0.14392490684986115,
0.04542945697903633,
-0.024350127205252647,
-0.08044745028018951,
-0.12533226609230042,
-0.06336074322462082,
-0.11165053397417068,
-0.1834794133901596,
0.02809632197022438,
-0.10398516803979874,
0.06006649509072304,
0.0501505509018898,
0.026290519163012505,
0.005053517874330282,
0.1636599600315094,
0.0047029959969222546,
0.021281149238348007,
0.08211172372102737,
0.00016928836703300476,
-0.06842047721147537,
-0.037426698952913284,
-0.09321928769350052,
0.024791670963168144,
-0.023142948746681213,
0.030486466363072395,
-0.002637891098856926,
-0.051231302320957184,
0.03415912389755249,
-0.032717134803533554,
-0.09899388998746872,
0.024341104552149773,
0.03300834447145462,
0.06816067546606064,
0.03220058232545853,
0.03673543781042099,
-0.039014607667922974,
0.010789117775857449,
0.21017853915691376,
-0.08570651710033417,
-0.08999954909086227,
-0.114335797727108,
0.22753579914569855,
0.04542962461709976,
0.010367068462073803,
0.002343267435207963,
-0.07579712569713593,
-0.019343581050634384,
0.19747412204742432,
0.17594324052333832,
-0.015881121158599854,
0.0038979812525212765,
-0.048903342336416245,
-0.00483758794143796,
-0.040050338953733444,
0.06908921152353287,
0.12112180143594742,
0.041204918175935745,
-0.04203001409769058,
-0.03244589641690254,
-0.06198719143867493,
-0.03215482085943222,
-0.05423877760767937,
0.08034149557352066,
-0.004049410577863455,
-0.023242289200425148,
-0.05917009711265564,
0.06215085834264755,
-0.07349871098995209,
-0.09329373389482498,
0.028861206024885178,
-0.1963609904050827,
-0.13449671864509583,
0.01648113504052162,
0.08295255899429321,
0.020389661192893982,
0.034627366811037064,
-0.009031512774527073,
-0.0032623554579913616,
0.06270983070135117,
-0.00290282117202878,
-0.06032957509160042,
-0.08652309328317642,
0.0661136582493782,
-0.13764338195323944,
0.2122335582971573,
-0.017374807968735695,
0.05229915305972099,
0.11168141663074493,
0.05160718411207199,
-0.08112607151269913,
0.1188230961561203,
0.050016701221466064,
-0.09429468214511871,
0.01274859718978405,
0.14208745956420898,
-0.033798519521951675,
0.13448533415794373,
0.049373526126146317,
-0.1241089329123497,
0.005749570671468973,
-0.0295856986194849,
-0.0636793002486229,
-0.06544361263513565,
-0.03670845180749893,
-0.05394772067666054,
0.11354691535234451,
0.15807540714740753,
-0.0463094636797905,
0.013625058345496655,
-0.04002954438328743,
0.05588362738490105,
0.0704536959528923,
-0.015496640466153622,
-0.036287933588027954,
-0.28634515404701233,
0.01120763085782528,
0.07329150289297104,
-0.007524858694523573,
-0.2562696933746338,
-0.10188130289316177,
-0.005377068184316158,
-0.02993662841618061,
-0.09273763000965118,
0.06989444047212601,
0.10787782818078995,
0.05087282881140709,
-0.050819527357816696,
-0.09196239709854126,
-0.027425555512309074,
0.18270957469940186,
-0.14832474291324615,
-0.08652951568365097
] |
null | null | transformers | Repository for my nlp4web model created for homework 6.
| {"license": "apache-2.0"} | question-answering | Bugtus/bugtus_nlp4web | [
"transformers",
"safetensors",
"bert",
"question-answering",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-07T16:04:52+00:00 | [] | [] | TAGS
#transformers #safetensors #bert #question-answering #license-apache-2.0 #endpoints_compatible #region-us
| Repository for my nlp4web model created for homework 6.
| [] | [
"TAGS\n#transformers #safetensors #bert #question-answering #license-apache-2.0 #endpoints_compatible #region-us \n"
] | [
38
] | [
"passage: TAGS\n#transformers #safetensors #bert #question-answering #license-apache-2.0 #endpoints_compatible #region-us \n"
] | [
-0.04257528483867645,
0.09744332730770111,
-0.007764603942632675,
-0.016286877915263176,
-0.009094440378248692,
-0.020812271162867546,
0.13347093760967255,
0.10591067373752594,
0.055131059139966965,
-0.032628897577524185,
0.19441652297973633,
0.16227160394191742,
-0.028531920164823532,
-0.0005682770861312747,
-0.08701175451278687,
-0.08275408297777176,
0.1424921452999115,
0.023019658401608467,
-0.06485406309366226,
0.08241017907857895,
0.10141376405954361,
-0.0555906742811203,
0.0353364534676075,
-0.028974834829568863,
-0.002630125265568495,
0.029279544949531555,
0.058496907353401184,
-0.08221237361431122,
0.09896499663591385,
0.05031503364443779,
0.08660851418972015,
0.051322732120752335,
-0.04620731994509697,
-0.22003868222236633,
0.02162025310099125,
0.006809709127992392,
-0.06359139829874039,
0.032604169100522995,
0.0059730601496994495,
-0.008515969850122929,
-0.0819128155708313,
0.02574734203517437,
-0.026947226375341415,
0.06279579550027847,
-0.07636083662509918,
-0.26170429587364197,
-0.1271362602710724,
0.035087283700704575,
0.10841421782970428,
0.10228534787893295,
0.027191318571567535,
0.16943702101707458,
-0.11529746651649475,
0.040375422686338425,
0.14055263996124268,
-0.378673791885376,
0.008744551800191402,
0.0671764463186264,
0.0812278464436531,
-0.006373337935656309,
-0.0012481545563787222,
0.056456249207258224,
0.08772512525320053,
-0.006043397821485996,
-0.011010955087840557,
-0.06157419830560684,
-0.03844316303730011,
0.06924300640821457,
-0.04088379815220833,
-0.0776069313287735,
0.2724177837371826,
0.0748060792684555,
-0.013553427532315254,
-0.0050269728526473045,
-0.08200839161872864,
0.11284158378839493,
0.0024881206918507814,
0.015266841277480125,
0.0489053837954998,
0.12754309177398682,
0.037996336817741394,
-0.02323281764984131,
-0.1462320238351822,
-0.02235220931470394,
-0.20547564327716827,
0.12328138947486877,
0.012697963044047356,
0.10442978888750076,
-0.1654789000749588,
0.01019567996263504,
0.005912934895604849,
-0.10506952553987503,
-0.06312784552574158,
-0.07438996434211731,
0.1192413866519928,
0.04025746136903763,
-0.060841698199510574,
0.10710135102272034,
0.1951654553413391,
0.29413795471191406,
0.017302000895142555,
-0.006239267997443676,
-0.07154183834791183,
0.10474302619695663,
-0.05205058306455612,
0.06497589498758316,
0.024410054087638855,
-0.01469916570931673,
0.09001368284225464,
-0.12309235334396362,
0.08254684507846832,
-0.005150670185685158,
-0.05884052440524101,
-0.02283143624663353,
-0.02611115202307701,
0.16455724835395813,
0.09933101385831833,
0.008815774694085121,
-0.04095443710684776,
0.07028356939554214,
0.12219031900167465,
-0.03866266459226608,
-0.013725689612329006,
-0.007916290313005447,
0.06146640330553055,
0.03529832512140274,
0.01705501228570938,
0.05963645130395889,
0.010030368342995644,
0.04039086028933525,
-0.05014842003583908,
-0.06663694232702255,
-0.004854843020439148,
0.03938334807753563,
0.12471140176057816,
-0.04405916482210159,
0.09482230246067047,
-0.1370827704668045,
-0.15182530879974365,
0.051302410662174225,
0.07232639938592911,
0.022397032007575035,
-0.02552621066570282,
0.05928892269730568,
-0.040057938545942307,
-0.012020999565720558,
-0.10096872597932816,
-0.049772150814533234,
-0.08846493810415268,
0.04171490669250488,
-0.04599250108003616,
-0.014823504723608494,
-0.1402961164712906,
0.00812382809817791,
-0.13303716480731964,
0.04052425175905228,
0.007569936104118824,
-0.08998667448759079,
-0.0928901731967926,
0.20995356142520905,
-0.09159921854734421,
0.00014760735211893916,
-0.026933209970593452,
0.018796361982822418,
-0.03550633788108826,
0.14161014556884766,
-0.061781689524650574,
0.015713118016719818,
0.23065873980522156,
-0.12282855063676834,
-0.28133854269981384,
0.0657014548778534,
0.025812629610300064,
0.0071942792274057865,
0.08176320046186447,
0.141417995095253,
-0.00012512433750089258,
-0.09574642777442932,
0.042725659906864166,
0.1407223492860794,
-0.1038617417216301,
-0.16199199855327606,
0.0827714055776596,
-0.0876694843173027,
-0.12695607542991638,
0.030568912625312805,
-0.06103099137544632,
0.06568904966115952,
-0.004821790847927332,
-0.09238865226507187,
-0.08593674004077911,
-0.06409331411123276,
-0.030783340334892273,
-0.008418955840170383,
-0.009432157501578331,
-0.1151609867811203,
-0.013827993534505367,
-0.09654507040977478,
0.015795424580574036,
0.047234002500772476,
0.049886953085660934,
-0.1316417008638382,
0.027745846658945084,
-0.0495578795671463,
0.028752528131008148,
-0.08339344710111618,
-0.07400940358638763,
0.0044556488282978535,
-0.010289590805768967,
-0.026013806462287903,
0.1118559017777443,
0.08440300077199936,
-0.11986441910266876,
-0.010429858230054379,
-0.033031389117240906,
0.12504495680332184,
0.0867590606212616,
0.011795984581112862,
-0.1261163055896759,
0.05346161872148514,
-0.04628949612379074,
-0.030454538762569427,
0.017299463972449303,
0.0009285899577662349,
0.10060951858758926,
0.08954713493585587,
-0.03199011832475662,
0.09894795715808868,
-0.04033000022172928,
-0.06863559782505035,
-0.009727945551276207,
-0.012963189743459225,
0.10319146513938904,
0.04578466713428497,
-0.11492794007062912,
0.1593065857887268,
-0.025342725217342377,
0.2992599904537201,
0.1820358783006668,
-0.10480770468711853,
0.09942416101694107,
0.014168202877044678,
-0.03782559558749199,
-0.0006045532645657659,
0.03672179579734802,
-0.006802912335842848,
-0.047446757555007935,
0.0221259668469429,
0.06712666153907776,
-0.05935444310307503,
-0.05313771590590477,
-0.02455860935151577,
-0.07769802957773209,
-0.020979298278689384,
0.003872235305607319,
0.07324671000242233,
-0.1737745851278305,
0.1771286129951477,
0.39303725957870483,
0.04264039546251297,
-0.009383480064570904,
-0.14140768349170685,
-0.03147420287132263,
-0.006490751169621944,
0.052516646683216095,
-0.012633012607693672,
0.09073569625616074,
-0.14106638729572296,
0.04585237056016922,
0.09059111028909683,
0.04386374354362488,
0.021774986758828163,
-0.134479358792305,
-0.08126591891050339,
0.02071574702858925,
-0.034667354077100754,
-0.056279413402080536,
0.05729258805513382,
-0.05000143498182297,
0.08264435827732086,
-0.009192116558551788,
-0.08472298830747604,
0.14404349029064178,
-0.022974936291575432,
-0.07054123282432556,
0.138668030500412,
-0.1388266533613205,
-0.16135498881340027,
-0.08292821794748306,
-0.0703963115811348,
0.0010421162005513906,
0.009063811972737312,
0.1365727037191391,
-0.05112003535032272,
-0.08023528754711151,
0.0063685206696391106,
-0.07916377484798431,
-0.01846579648554325,
0.031907644122838974,
0.06317482888698578,
0.06692055612802505,
0.0391022153198719,
-0.11974107474088669,
-0.051322679966688156,
0.03995122015476227,
-0.06641329824924469,
0.04629448056221008,
-0.06119416281580925,
0.10507897287607193,
0.049231283366680145,
0.05726221203804016,
0.004787418060004711,
-0.030759748071432114,
0.16386063396930695,
-0.04699980095028877,
0.01540085393935442,
0.21330972015857697,
-0.030449505895376205,
0.07023324072360992,
0.182051882147789,
0.036907486617565155,
-0.09695345908403397,
0.035135574638843536,
-0.03883608058094978,
-0.07916166633367538,
-0.2813175916671753,
-0.036595217883586884,
-0.07151894271373749,
0.0820142850279808,
0.007187908515334129,
0.07737462967634201,
0.11538971215486526,
0.10731877386569977,
-0.02431335859000683,
-0.0916600450873375,
0.029921814799308777,
0.05967087298631668,
0.21335233747959137,
-0.04463474079966545,
0.07716748118400574,
-0.1388511210680008,
-0.044021766632795334,
0.11272060126066208,
0.11140944063663483,
0.11974598467350006,
0.18641206622123718,
0.030753010883927345,
0.12927518784999847,
0.1911485344171524,
0.05023306980729103,
0.11409398168325424,
0.028065279126167297,
-0.05212026089429855,
-0.04774462431669235,
-0.00922818947583437,
-0.04545150324702263,
0.08015001565217972,
-0.062310464680194855,
-0.09326594322919846,
0.01659342460334301,
-0.1524081528186798,
0.0901624783873558,
0.19744738936424255,
0.05877711623907089,
-0.08518442511558533,
0.01160811074078083,
0.12655363976955414,
-0.005665971897542477,
0.008875403553247452,
0.13016094267368317,
-0.030699383467435837,
-0.05257776752114296,
0.09844331443309784,
-0.03467730060219765,
0.11346503347158432,
0.09538062661886215,
0.04914514720439911,
-0.04322628304362297,
-0.16728752851486206,
0.06711546331644058,
0.13458038866519928,
-0.28636452555656433,
0.22808802127838135,
0.004451594781130552,
-0.007343452423810959,
-0.0697663202881813,
0.01068166084587574,
0.011568901129066944,
0.1933901607990265,
0.19277654588222504,
0.015133768320083618,
-0.1606469750404358,
0.03570063039660454,
-0.03637326508760452,
0.0802968218922615,
0.01542493887245655,
0.013686014339327812,
-0.03718005120754242,
-0.06096890941262245,
0.01184647623449564,
0.03331141546368599,
0.11017356812953949,
-0.06787541508674622,
-0.10712948441505432,
-0.0035330455284565687,
0.11641179770231247,
0.09135604649782181,
-0.12772893905639648,
0.05135580524802208,
-0.07390788197517395,
0.13041271269321442,
-0.11899840086698532,
-0.04291342943906784,
-0.07400790601968765,
-0.18751657009124756,
0.06911249458789825,
-0.04007180407643318,
0.0484190508723259,
-0.06268440932035446,
-0.01642044633626938,
-0.05370626598596573,
-0.1689174473285675,
0.11281798779964447,
-0.16903150081634521,
-0.04328068718314171,
-0.047091204673051834,
0.1420503556728363,
-0.08661983907222748,
-0.0013672312488779426,
0.050707969814538956,
0.019904954358935356,
-0.09327593445777893,
-0.12971024215221405,
-0.005414990242570639,
0.002393507631495595,
0.07064947485923767,
-0.04269363731145859,
-0.024543937295675278,
0.006756705231964588,
0.04937254637479782,
-0.048281412571668625,
0.14001767337322235,
0.2447921633720398,
-0.06561994552612305,
0.09972704946994781,
0.24285350739955902,
-0.035112932324409485,
-0.28262606263160706,
-0.1464972347021103,
-0.21898017823696136,
-0.12775881588459015,
0.015031611546874046,
-0.04468406364321709,
0.18120071291923523,
0.08146359026432037,
-0.11408785730600357,
0.05150088295340538,
-0.1440686583518982,
-0.04447482153773308,
0.1879817098379135,
0.004087146371603012,
0.3903743326663971,
-0.17196790874004364,
-0.08341596275568008,
-0.08288178592920303,
-0.2672707438468933,
0.09072936326265335,
-0.15829260647296906,
-0.005909902509301901,
0.032612498849630356,
-0.030140899121761322,
-0.04128895699977875,
-0.07657807320356369,
0.18305306136608124,
-0.03205694630742073,
0.055826231837272644,
-0.1029588058590889,
0.044497281312942505,
0.11663030087947845,
-0.02440498024225235,
0.06310397386550903,
-0.1287553608417511,
0.051334090530872345,
-0.03580524027347565,
-0.022590044885873795,
-0.06313583254814148,
0.08722580969333649,
-0.009885485284030437,
-0.07020280510187149,
-0.057084184139966965,
-0.032403022050857544,
0.03848240524530411,
-0.012391003780066967,
0.2978104054927826,
0.014426864683628082,
0.06653814017772675,
0.08110000938177109,
0.0917278379201889,
-0.23842009902000427,
-0.005315285176038742,
-0.059958070516586304,
-0.10100015252828598,
0.0794772058725357,
-0.14935922622680664,
0.08340612053871155,
0.0958457663655281,
-0.05672338977456093,
0.021417435258626938,
0.0557316355407238,
0.009354016743600368,
-0.049156468361616135,
0.10233937203884125,
-0.1316542774438858,
-0.1417246311903,
0.041288260370492935,
0.13993319869041443,
0.036222923547029495,
0.1309407353401184,
0.0989498421549797,
0.04143310338258743,
-0.008270792663097382,
0.0013773143291473389,
0.0474383570253849,
-0.09991513937711716,
-0.0017854869365692139,
0.050288643687963486,
0.0244634747505188,
-0.10220003128051758,
0.1348164677619934,
-0.05293360352516174,
-0.1823752522468567,
-0.01995881460607052,
-0.03446609899401665,
-0.12716466188430786,
-0.09704139083623886,
0.00416854303330183,
0.036223191767930984,
-0.10744093358516693,
-0.14810140430927277,
-0.03481503576040268,
-0.1530139297246933,
0.04267725348472595,
0.1601562798023224,
0.10946859419345856,
0.10841166228055954,
0.07612371444702148,
-0.035971689969301224,
0.062281426042318344,
-0.014179655350744724,
-0.10911480337381363,
0.006112702656537294,
-0.1025293692946434,
-0.17724281549453735,
-0.011860080994665623,
0.05521996319293976,
-0.043859973549842834,
0.010973007418215275,
-0.1023971363902092,
0.05696243792772293,
-0.22269921004772186,
-0.0076432377099990845,
-0.12649676203727722,
-0.013976792804896832,
0.03806402534246445,
-0.1334836632013321,
-0.03614838793873787,
0.025377236306667328,
-0.09684576094150543,
-0.00546534638851881,
0.02896234579384327,
0.04323272779583931,
-0.1416667103767395,
-0.05754978209733963,
0.10677729547023773,
-0.023856716230511665,
0.08638402819633484,
0.07124603539705276,
-0.08423088490962982,
0.08105576783418655,
-0.1840389370918274,
-0.08980349451303482,
0.06342124193906784,
0.018933966755867004,
0.012605862691998482,
-0.016173217445611954,
-0.03570990636944771,
0.1476832926273346,
-0.017566533759236336,
0.038707245141267776,
-0.02907176874577999,
-0.11187835037708282,
-0.06944091618061066,
0.007569264154881239,
-0.10242627561092377,
0.006935481447726488,
-0.1663556545972824,
0.17976509034633636,
-0.004800746217370033,
0.18632137775421143,
0.014123402535915375,
0.016644947230815887,
-0.044751010835170746,
0.030241433531045914,
-0.024874627590179443,
-0.134684219956398,
-0.09783414751291275,
-0.022047001868486404,
-0.07435352355241776,
-0.05969267711043358,
0.2874581515789032,
-0.10122119635343552,
-0.019822075963020325,
0.10761912912130356,
-0.0034070159308612347,
0.04988741874694824,
0.023176712915301323,
0.3140697479248047,
0.035458970814943314,
-0.010180986486375332,
-0.10740689188241959,
-0.015537165105342865,
0.006040047854185104,
-0.2029706835746765,
0.04288270324468613,
0.1739046275615692,
0.058566417545080185,
0.08491407334804535,
0.040666431188583374,
0.022282075136899948,
-0.04964112117886543,
-0.15732857584953308,
0.025158831849694252,
0.028676703572273254,
0.0362095981836319,
0.11877521127462387,
0.173024520277977,
-0.029261140152812004,
-0.013262822292745113,
-0.0796748548746109,
0.01901201717555523,
-0.16418924927711487,
-0.08834441751241684,
-0.07309039682149887,
-0.138799250125885,
0.027474066242575645,
-0.06293032318353653,
-0.010509610176086426,
0.12102080881595612,
0.059301216155290604,
-0.054558902978897095,
0.06088368594646454,
-0.0589214563369751,
0.006682641338557005,
-0.00389956496655941,
0.017612019553780556,
-0.00676383450627327,
0.05328623577952385,
-0.04322585463523865,
-0.06479064375162125,
-0.051306769251823425,
-0.06380646675825119,
0.04161909222602844,
-0.022518789395689964,
0.04684540629386902,
-0.09517160058021545,
-0.0516262948513031,
-0.08093974739313126,
0.06975723803043365,
-0.026912478730082512,
0.1572774052619934,
0.0032235898543149233,
0.043486643582582474,
0.1262296438217163,
0.18538209795951843,
-0.09463679045438766,
-0.17440453171730042,
-0.09139855951070786,
0.08722901344299316,
-0.012067507952451706,
0.09016337990760803,
-0.008652878925204277,
0.0213534627109766,
-0.011081459000706673,
0.28877511620521545,
0.15066412091255188,
-0.03270269185304642,
0.04776548221707344,
-0.022127574309706688,
0.019428303465247154,
0.00932922400534153,
0.055913764983415604,
0.10395582765340805,
0.1643616408109665,
-0.09416872262954712,
-0.04096146300435066,
-0.0008555143140256405,
0.008221208117902279,
-0.14481529593467712,
0.04450929909944534,
-0.05900419503450394,
-0.07093822956085205,
-0.01863139122724533,
0.11166054755449295,
-0.022354191169142723,
0.04642272740602493,
0.04517446458339691,
-0.06398096680641174,
0.004892091732472181,
0.005991757847368717,
0.22010456025600433,
0.006918218918144703,
0.04005206748843193,
-0.04808815196156502,
-0.0587431862950325,
0.0930517166852951,
0.006003737449645996,
-0.17306779325008392,
-0.06521044671535492,
0.08888675272464752,
0.03338847681879997,
0.19412977993488312,
0.03372453898191452,
0.04617702215909958,
0.09806223958730698,
0.02453777939081192,
-0.14061148464679718,
0.12112400680780411,
0.08484036475419998,
-0.09934845566749573,
-0.04926932603120804,
-0.18325603008270264,
-0.026295624673366547,
-0.005351042374968529,
0.05059859901666641,
-0.10902675241231918,
0.0431249663233757,
0.07689543068408966,
-0.06471244245767593,
-0.087420754134655,
0.049759216606616974,
-0.09211232513189316,
0.05368591472506523,
-0.04437362402677536,
-0.059448473155498505,
-0.05256794020533562,
-0.022707249969244003,
0.04332459345459938,
0.04438557103276253,
-0.16133227944374084,
-0.04639968276023865,
0.03358475863933563,
0.02418488822877407,
0.10541301965713501,
0.01425748597830534,
-0.0337185338139534,
-0.052675116807222366,
-0.025110891088843346,
0.02162259817123413,
-0.10478927940130234,
-0.013846766203641891,
0.11996766179800034,
0.010619251057505608,
-0.0065196179784834385,
-0.15163956582546234,
0.015622405335307121,
0.03065803460776806,
-0.036138880997896194,
-0.09155476838350296
] |
null | null | null |
# BagelMIsteryTour-v2-8x7B 3.5bpw
Imatrix GGUF quant of [ycros/BagelMIsteryTour-v2-8x7B](https://huggingface.co/ycros/BagelMIsteryTour-v2-8x7B)
## Other quants:
EXL2: [5bpw](https://huggingface.co/Kooten/BagelMIsteryTour-v2-8x7B-5bpw-exl2), [3.5bpw](https://huggingface.co/Kooten/BagelMIsteryTour-v2-8x7B-3.5bpw-exl2)
[GGUF](https://huggingface.co/Kooten/BagelMIsteryTour-v2-8x7B-Imatrix-GGUF): [IQ3_XXS](https://huggingface.co/Kooten/BagelMIsteryTour-v2-8x7B-Imatrix-GGUF/blob/main/BagelMIsteryTour-v2-8x7B-IQ3_XXS.gguf), [IQ2_XS](https://huggingface.co/Kooten/BagelMIsteryTour-v2-8x7B-Imatrix-GGUF/blob/main/BagelMIsteryTour-v2-8x7B-IQ2_XS.gguf), [IQ2_XXS](https://huggingface.co/Kooten/BagelMIsteryTour-v2-8x7B-Imatrix-GGUF/blob/main/BagelMIsteryTour-v2-8x7B-IQ2_XXS.gguf)
## Prompt format: Alpaca
It is noted to also work with mistral
```
Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
{prompt}
### Input:
{input}
### Response:
```
## Contact
Kooten on discord
[ko-fi.com/kooten](https://ko-fi.com/kooten) if you would like to support me
| {"license": "cc-by-nc-4.0", "tags": ["mergekit", "merge"], "base_model": ["mistralai/Mixtral-8x7B-v0.1", "jondurbin/bagel-dpo-8x7b-v0.2", "Sao10K/Sensualize-Mixtral-bf16", "mistralai/Mixtral-8x7B-v0.1", "Doctor-Shotgun/limarp-zloss-mixtral-8x7b-qlora", "mistralai/Mixtral-8x7B-Instruct-v0.1"]} | null | Kooten/BagelMIsteryTour-v2-8x7B-Imatrix-GGUF | [
"gguf",
"mergekit",
"merge",
"base_model:mistralai/Mixtral-8x7B-v0.1",
"base_model:jondurbin/bagel-dpo-8x7b-v0.2",
"base_model:Sao10K/Sensualize-Mixtral-bf16",
"base_model:Doctor-Shotgun/limarp-zloss-mixtral-8x7b-qlora",
"base_model:mistralai/Mixtral-8x7B-Instruct-v0.1",
"license:cc-by-nc-4.0",
"region:us"
] | 2024-02-07T16:06:33+00:00 | [] | [] | TAGS
#gguf #mergekit #merge #base_model-mistralai/Mixtral-8x7B-v0.1 #base_model-jondurbin/bagel-dpo-8x7b-v0.2 #base_model-Sao10K/Sensualize-Mixtral-bf16 #base_model-Doctor-Shotgun/limarp-zloss-mixtral-8x7b-qlora #base_model-mistralai/Mixtral-8x7B-Instruct-v0.1 #license-cc-by-nc-4.0 #region-us
|
# BagelMIsteryTour-v2-8x7B 3.5bpw
Imatrix GGUF quant of ycros/BagelMIsteryTour-v2-8x7B
## Other quants:
EXL2: 5bpw, 3.5bpw
GGUF: IQ3_XXS, IQ2_XS, IQ2_XXS
## Prompt format: Alpaca
It is noted to also work with mistral
## Contact
Kooten on discord
URL if you would like to support me
| [
"# BagelMIsteryTour-v2-8x7B 3.5bpw\nImatrix GGUF quant of ycros/BagelMIsteryTour-v2-8x7B",
"## Other quants:\n\nEXL2: 5bpw, 3.5bpw\n\nGGUF: IQ3_XXS, IQ2_XS, IQ2_XXS",
"## Prompt format: Alpaca\nIt is noted to also work with mistral",
"## Contact\nKooten on discord\n\nURL if you would like to support me"
] | [
"TAGS\n#gguf #mergekit #merge #base_model-mistralai/Mixtral-8x7B-v0.1 #base_model-jondurbin/bagel-dpo-8x7b-v0.2 #base_model-Sao10K/Sensualize-Mixtral-bf16 #base_model-Doctor-Shotgun/limarp-zloss-mixtral-8x7b-qlora #base_model-mistralai/Mixtral-8x7B-Instruct-v0.1 #license-cc-by-nc-4.0 #region-us \n",
"# BagelMIsteryTour-v2-8x7B 3.5bpw\nImatrix GGUF quant of ycros/BagelMIsteryTour-v2-8x7B",
"## Other quants:\n\nEXL2: 5bpw, 3.5bpw\n\nGGUF: IQ3_XXS, IQ2_XS, IQ2_XXS",
"## Prompt format: Alpaca\nIt is noted to also work with mistral",
"## Contact\nKooten on discord\n\nURL if you would like to support me"
] | [
137,
42,
39,
18,
14
] | [
"passage: TAGS\n#gguf #mergekit #merge #base_model-mistralai/Mixtral-8x7B-v0.1 #base_model-jondurbin/bagel-dpo-8x7b-v0.2 #base_model-Sao10K/Sensualize-Mixtral-bf16 #base_model-Doctor-Shotgun/limarp-zloss-mixtral-8x7b-qlora #base_model-mistralai/Mixtral-8x7B-Instruct-v0.1 #license-cc-by-nc-4.0 #region-us \n# BagelMIsteryTour-v2-8x7B 3.5bpw\nImatrix GGUF quant of ycros/BagelMIsteryTour-v2-8x7B## Other quants:\n\nEXL2: 5bpw, 3.5bpw\n\nGGUF: IQ3_XXS, IQ2_XS, IQ2_XXS## Prompt format: Alpaca\nIt is noted to also work with mistral## Contact\nKooten on discord\n\nURL if you would like to support me"
] | [
-0.1165803074836731,
0.11884033679962158,
-0.005602696444839239,
0.0798688530921936,
0.031155193224549294,
0.07858549058437347,
0.1823517084121704,
0.14818935096263885,
0.1791534423828125,
0.07615579664707184,
0.055597178637981415,
0.1019953042268753,
0.12104824185371399,
0.23308677971363068,
-0.024549169465899467,
-0.10977035015821457,
0.02157406322658062,
0.027980320155620575,
0.12701857089996338,
0.0761202722787857,
0.09107175469398499,
-0.0002627621579449624,
0.06390511244535446,
0.003025379031896591,
-0.05197371914982796,
-0.0012811953201889992,
-0.006580776534974575,
0.023218804970383644,
0.04432832449674606,
0.08728566765785217,
0.005432053003460169,
0.04386131837964058,
-0.008394408971071243,
-0.21475237607955933,
0.031711552292108536,
-0.019208816811442375,
-0.05556857958436012,
0.04784456267952919,
0.06661388278007507,
-0.10029628127813339,
0.07055430114269257,
-0.08147652447223663,
-0.021173443645238876,
0.05336809158325195,
-0.13463157415390015,
-0.09311920404434204,
-0.15316158533096313,
0.23224519193172455,
0.06908133625984192,
0.01916227489709854,
-0.011906187981367111,
0.11640950292348862,
0.10456974804401398,
0.08892492204904556,
0.28816673159599304,
-0.22444814443588257,
-0.05258231237530708,
0.1681092083454132,
-0.0014567484613507986,
0.038050927221775055,
-0.05180799961090088,
-0.0004473428416531533,
0.009768047370016575,
-0.03001796081662178,
-0.05681557580828667,
-0.0749414786696434,
0.09862307459115982,
-0.023784060031175613,
-0.07525158673524857,
-0.05243098363280296,
0.16216255724430084,
0.04597064107656479,
-0.050049107521772385,
0.010008452460169792,
-0.02121049165725708,
-0.023617910221219063,
-0.06145532801747322,
-0.020113877952098846,
0.036436475813388824,
-0.022528694942593575,
0.08903488516807556,
-0.05303960666060448,
-0.04278319329023361,
-0.04835499823093414,
-0.055158041417598724,
0.17187891900539398,
-0.018709229305386543,
0.008770646527409554,
0.018129266798496246,
0.03948655351996422,
-0.2921903133392334,
-0.12872837483882904,
-0.09648498892784119,
-0.05101025849580765,
-0.05646925047039986,
-0.040244556963443756,
0.021600855514407158,
0.04427219554781914,
0.0765998587012291,
0.2512817680835724,
-0.06867996603250504,
0.11564390361309052,
0.04508660361170769,
0.0019487565150484443,
-0.005216670222580433,
0.03616343066096306,
-0.03403959795832634,
-0.08422447741031647,
0.023290231823921204,
0.0773882046341896,
0.03149774670600891,
0.002615987788885832,
-0.09492280334234238,
-0.06916794925928116,
-0.039363496005535126,
0.03232378885149956,
0.01923971436917782,
0.009856928139925003,
-0.1456775665283203,
-0.07913241535425186,
0.2067272961139679,
-0.06952337920665741,
0.053248096257448196,
0.024251801893115044,
0.01831839792430401,
0.02559080719947815,
0.032511211931705475,
0.032594479620456696,
0.010010374709963799,
0.007279818877577782,
-0.07236307114362717,
0.07479239255189896,
-0.03521006181836128,
-0.04885760322213173,
0.03635367006063461,
-0.02207791805267334,
-0.02691466733813286,
-0.13667848706245422,
-0.12693670392036438,
-0.042527563869953156,
0.03596194460988045,
-0.08510461449623108,
0.03415664657950401,
-0.006782068405300379,
-0.011607427150011063,
0.035558369010686874,
0.0018802884733304381,
0.02115570940077305,
-0.054079484194517136,
0.009202186018228531,
0.04045940935611725,
0.11422338336706161,
-0.0820651575922966,
0.000243389600655064,
0.004855665378272533,
0.0675172507762909,
-0.19555403292179108,
0.04884298890829086,
-0.16094225645065308,
0.012921078130602837,
-0.14053498208522797,
0.008572818711400032,
0.015086687169969082,
-0.050345368683338165,
0.1089949831366539,
0.1441403031349182,
-0.13478800654411316,
-0.03354720398783684,
0.12760743498802185,
-0.09522122889757156,
-0.11519201844930649,
0.07904602587223053,
0.018423201516270638,
-0.03120037540793419,
0.048517972230911255,
0.22282344102859497,
0.07884013652801514,
-0.1468920260667801,
-0.10475785285234451,
-0.01179350633174181,
0.0755598247051239,
0.004634392447769642,
0.08577938377857208,
-0.02281099185347557,
-0.07479002326726913,
-0.0015399233670905232,
-0.07145718485116959,
0.03612261638045311,
-0.003884541802108288,
-0.04946034401655197,
-0.028458233922719955,
-0.03691858798265457,
0.0938141867518425,
0.01283323485404253,
-0.05890346318483353,
-0.04566064849495888,
-0.036648619920015335,
-0.028151843696832657,
0.11562543362379074,
0.026109056547284126,
-0.01362796314060688,
-0.08167221397161484,
0.09906134009361267,
-0.09070068597793579,
-0.005799047648906708,
-0.10573716461658478,
-0.0423613116145134,
0.007854128256440163,
-0.05546623840928078,
0.04604802653193474,
-0.012843737378716469,
0.09368724375963211,
0.004508144222199917,
-0.07870858162641525,
-0.033537693321704865,
0.014756538905203342,
-0.015084592625498772,
-0.06582128256559372,
-0.12785010039806366,
-0.028508061543107033,
-0.06720428913831711,
0.1369793862104416,
-0.06831394881010056,
0.029846103861927986,
0.06577911972999573,
0.1255711168050766,
-0.0018143013585358858,
-0.08893841505050659,
0.10748368501663208,
0.01078372448682785,
0.010786776430904865,
-0.04056293144822121,
0.050037309527397156,
-0.020385943353176117,
-0.13950058817863464,
0.016303787007927895,
-0.12681415677070618,
0.006673394236713648,
0.06578317284584045,
0.16248412430286407,
-0.0754515752196312,
-0.09976495057344437,
0.008311107754707336,
-0.009780733846127987,
0.05882987380027771,
-0.042955610901117325,
0.13825148344039917,
-0.011958640068769455,
0.033764734864234924,
-0.06964404135942459,
-0.06595464795827866,
-0.005365574732422829,
-0.07345558702945709,
-0.041535425931215286,
0.12107101082801819,
-0.041883811354637146,
-0.13622577488422394,
0.0798071101307869,
0.11727002263069153,
-0.012361116707324982,
0.1375153362751007,
0.064854197204113,
-0.018268493935465813,
-0.10600601136684418,
0.05228583142161369,
0.06200866773724556,
0.1010906845331192,
-0.18711808323860168,
0.04957710579037666,
0.06174331530928612,
-0.00837097316980362,
-0.01892567239701748,
-0.08993639796972275,
-0.010393922217190266,
-0.002192644402384758,
-0.07349859923124313,
-0.028434177860617638,
0.05326405167579651,
-0.027535483241081238,
0.07925325632095337,
0.025487670674920082,
0.06582964211702347,
-0.0262879878282547,
0.0023279923480004072,
-0.07788249105215073,
0.16827264428138733,
-0.08228227496147156,
-0.12018289417028427,
-0.06348425149917603,
-0.14433111250400543,
-0.11981762200593948,
0.0034626389387995005,
0.04769999161362648,
-0.07596947997808456,
-0.03995073214173317,
-0.035508058965206146,
0.03384024649858475,
0.043814145028591156,
-0.0041446625255048275,
0.034678589552640915,
-0.05667269602417946,
0.0678839236497879,
-0.09324400871992111,
-0.05799451470375061,
0.008861595764756203,
0.011826533824205399,
0.07407248020172119,
-0.0491187646985054,
0.07584014534950256,
0.07968933880329132,
0.03244245797395706,
0.03822724521160126,
0.00820339284837246,
0.2751885652542114,
-0.06568396091461182,
0.0788097232580185,
0.12073825299739838,
0.042339324951171875,
0.09847554564476013,
0.15194953978061676,
0.02490082196891308,
-0.06736604124307632,
-0.03696727007627487,
0.012675208039581776,
-0.03453139215707779,
-0.22618471086025238,
-0.049894578754901886,
-0.05948510393500328,
0.007736529223620892,
0.04247048869729042,
0.03344837948679924,
0.016872167587280273,
0.12050795555114746,
-0.06757842749357224,
0.03965456038713455,
-0.03398604318499565,
0.0882570743560791,
0.04098854586482048,
0.04864379018545151,
0.06901244819164276,
-0.04970182105898857,
0.039546333253383636,
0.10645613074302673,
0.14617863297462463,
0.11939670890569687,
0.03498115763068199,
0.12629611790180206,
0.05596471577882767,
0.051399968564510345,
0.005745965987443924,
0.029910800978541374,
-0.020990518853068352,
-0.05160486698150635,
-0.021027516573667526,
-0.08959080278873444,
-0.03778872638940811,
0.06948030740022659,
0.012051018886268139,
0.006800311151891947,
-0.015645502135157585,
-0.08577033132314682,
0.06849173456430435,
-0.018468322232365608,
0.08303165435791016,
-0.225714772939682,
-0.06137498840689659,
0.08104152977466583,
0.0309898741543293,
-0.05119834840297699,
-0.08010785281658173,
0.017638428136706352,
-0.06355611234903336,
0.1758331060409546,
0.027007250115275383,
0.08674764633178711,
-0.039765629917383194,
-0.038767650723457336,
-0.055350273847579956,
0.06523626297712326,
-0.019967664033174515,
0.036930397152900696,
-0.13357992470264435,
0.21213378012180328,
0.053434018045663834,
0.028877925127744675,
-0.01739276945590973,
0.014165874570608139,
0.11272069066762924,
0.07270344346761703,
0.12999939918518066,
0.06649857759475708,
-0.040221843868494034,
-0.13444271683692932,
-0.06933474540710449,
-0.003948896657675505,
0.026366950944066048,
-0.05670600011944771,
0.06552109867334366,
0.05219539254903793,
-0.05162203684449196,
-0.07553151994943619,
0.06300123780965805,
-0.19148044288158417,
-0.09958329051733017,
0.11610746383666992,
0.052743151783943176,
0.03452513366937637,
-0.030330875888466835,
-0.03652476146817207,
-0.03923114389181137,
0.1658017337322235,
-0.030705450102686882,
-0.10759366303682327,
-0.10708415508270264,
-0.002522590570151806,
0.09013240039348602,
-0.1014038622379303,
-0.011779180727899075,
-0.0650152713060379,
0.057762354612350464,
-0.06934870034456253,
-0.04896225780248642,
0.06751567125320435,
-0.09090453386306763,
-0.09574732929468155,
-0.037200409919023514,
0.15175727009773254,
-0.020064759999513626,
0.06049000471830368,
0.03571192920207977,
0.021870505064725876,
-0.003980302717536688,
-0.12929940223693848,
0.09194474667310715,
0.056881967931985855,
0.009469174779951572,
0.08244170248508453,
-0.0004255033563822508,
-0.10442136973142624,
-0.07974117994308472,
0.0037647353019565344,
0.06763916462659836,
0.3768690228462219,
-0.027770590037107468,
0.06414937227964401,
0.07509277760982513,
-0.08939444273710251,
-0.13783049583435059,
-0.05008196458220482,
0.058523084968328476,
-0.015596742741763592,
0.01581098511815071,
-0.10765153169631958,
-0.01562754064798355,
0.12490451335906982,
-0.007756972219794989,
0.11810017377138138,
-0.1877918690443039,
-0.09563376009464264,
-0.019606349989771843,
0.08041784167289734,
0.09944824874401093,
-0.1798149049282074,
-0.11309517174959183,
-0.0663212388753891,
-0.23948636651039124,
0.0007936347974464297,
0.028159791603684425,
0.12699846923351288,
-0.08194080740213394,
0.05884108319878578,
0.01699678972363472,
-0.037577640265226364,
0.2258811891078949,
0.1259322166442871,
0.014052270911633968,
-0.06225309520959854,
-0.046821143478155136,
0.08659927546977997,
-0.07151835411787033,
0.10690745711326599,
-0.0553089901804924,
0.012448788620531559,
-0.22175736725330353,
-0.02417410910129547,
-0.05571070685982704,
0.013624373823404312,
-0.02965966984629631,
-0.027831539511680603,
-0.04135103151202202,
0.07061430811882019,
0.003643179079517722,
0.004033979028463364,
0.02849944680929184,
-0.0846557691693306,
0.09161283820867538,
0.14869050681591034,
0.0732971653342247,
-0.11602126806974411,
-0.08885932713747025,
-0.003531087888404727,
-0.05353984236717224,
0.012644978240132332,
-0.1933712661266327,
0.017820067703723907,
0.08273966610431671,
0.026197465136647224,
0.12614323198795319,
-0.00825051125138998,
-0.08192796260118484,
0.057813145220279694,
0.08815336227416992,
-0.11899416893720627,
-0.11397094279527664,
-0.0011038288939744234,
0.0858611911535263,
-0.10474865138530731,
0.01125967688858509,
0.16276739537715912,
0.03272630274295807,
-0.05170499533414841,
0.007938053458929062,
0.028396746143698692,
-0.05308340862393379,
0.11990255117416382,
0.05847933515906334,
0.08078522235155106,
-0.082878977060318,
0.01715988852083683,
-0.01268567144870758,
-0.03158688545227051,
-0.006693575996905565,
0.10376071184873581,
-0.06891518831253052,
-0.09641774743795395,
-0.06236611306667328,
0.1285003423690796,
-0.06263599544763565,
0.018808988854289055,
-0.04068247973918915,
-0.037872735410928726,
0.008353921584784985,
0.09227624535560608,
0.04889284819364548,
-0.04421950876712799,
0.07892858982086182,
-0.05170538276433945,
0.0328533798456192,
0.055920958518981934,
0.0574713796377182,
0.09272544831037521,
-0.06187326833605766,
-0.06180214509367943,
-0.05971689522266388,
0.04235082492232323,
0.00775967026129365,
0.01141548901796341,
-0.19127662479877472,
-0.0486447811126709,
-0.10541307926177979,
0.041592348366975784,
-0.14925125241279602,
0.008711566217243671,
-0.026240911334753036,
-0.057749416679143906,
-0.01211455836892128,
0.020516881719231606,
-0.08581246435642242,
-0.04819982126355171,
-0.01665080524981022,
0.0892716497182846,
-0.061670154333114624,
-0.040832873433828354,
0.057610344141721725,
-0.06143781915307045,
0.09162692725658417,
0.06188107281923294,
0.01949317939579487,
-0.04736990109086037,
-0.22151495516300201,
-0.0035735636483877897,
0.0008567813201807439,
0.03766701743006706,
0.04120069742202759,
-0.17137083411216736,
0.020872073248028755,
-0.044450026005506516,
0.029085155576467514,
0.0399484783411026,
0.12315607070922852,
-0.0805591270327568,
-0.005704253446310759,
-0.08689969033002853,
-0.0806269496679306,
-0.046914778649806976,
0.03829873725771904,
0.14311116933822632,
0.018396321684122086,
0.054816581308841705,
-0.03983290493488312,
0.023197464644908905,
-0.158208966255188,
0.011475133709609509,
0.009669404476881027,
-0.08019918948411942,
-0.02022671326994896,
-0.09101837873458862,
0.054845090955495834,
0.023324089124798775,
0.046111490577459335,
-0.08221463859081268,
-0.009563670493662357,
0.004033737815916538,
-0.018501386046409607,
0.12988793849945068,
-0.06890272349119186,
0.054300110787153244,
0.0016304274322465062,
-0.022952144965529442,
0.02476196549832821,
0.0541321262717247,
0.01779782772064209,
0.01699192076921463,
0.10551968216896057,
0.13059572875499725,
0.027467120438814163,
0.065297432243824,
0.02287529595196247,
-0.019429635256528854,
-0.08696328848600388,
-0.005907702259719372,
-0.038429003208875656,
0.02712269127368927,
-0.0631893202662468,
0.06853368878364563,
0.12095429748296738,
-0.18783780932426453,
0.0726536363363266,
-0.00494239991530776,
-0.04534206911921501,
-0.04064735025167465,
-0.1595098227262497,
-0.08864066749811172,
-0.12193790078163147,
0.015340843237936497,
-0.10545007139444351,
0.0665072575211525,
0.11212258785963058,
0.01769174262881279,
0.059299565851688385,
0.11748882383108139,
-0.0769413486123085,
0.0006913888501003385,
0.050459835678339005,
-0.007054103538393974,
-0.02828427217900753,
0.013565484434366226,
0.002370955888181925,
0.08256891369819641,
-0.06285873800516129,
0.02590852417051792,
0.04212556034326553,
0.0364835187792778,
-0.02315400540828705,
-0.02799167111515999,
-0.13019125163555145,
-0.030707720667123795,
0.058498650789260864,
0.05974850431084633,
0.15126579999923706,
0.04387611895799637,
0.009670430794358253,
-0.0020152453798800707,
0.04852631315588951,
0.008280515670776367,
-0.043476320803165436,
-0.10997369140386581,
0.04951239377260208,
-0.0588235966861248,
0.010881329886615276,
-0.029960015788674355,
-0.06009511649608612,
0.060213103890419006,
0.10561421513557434,
0.0325893796980381,
-0.026124052703380585,
0.00613069161772728,
0.012774604372680187,
0.01794673502445221,
0.03752630576491356,
0.028141485527157784,
0.059267930686473846,
0.08627058565616608,
-0.05206875875592232,
-0.0019552719313651323,
-0.012559688650071621,
-0.04036274924874306,
-0.03819507360458374,
-0.011538413353264332,
-0.0021205972880125046,
-0.026004403829574585,
-0.03423603996634483,
0.10639259964227676,
-0.03337385132908821,
-0.0239996500313282,
0.10092882812023163,
-0.15857243537902832,
-0.12876510620117188,
-0.01595480553805828,
0.04613907262682915,
0.06254546344280243,
0.07029666006565094,
-0.02733978070318699,
-0.02960285171866417,
0.10611499845981598,
-0.009508213959634304,
-0.13722555339336395,
-0.10260715335607529,
0.02959120087325573,
0.010132133029401302,
0.15079723298549652,
-0.017897026613354683,
0.08205781131982803,
0.14044161140918732,
-0.03626738488674164,
-0.14954456686973572,
0.07652508467435837,
0.056327253580093384,
-0.08398249000310898,
0.03135758638381958,
0.13136880099773407,
-0.007989316247403622,
0.030375974252820015,
0.0578334666788578,
-0.06311666965484619,
0.010127642191946507,
0.11694394797086716,
-0.0014173660892993212,
-0.05868315324187279,
0.09943219274282455,
-0.10168743133544922,
0.10563118010759354,
0.14157913625240326,
-0.031092550605535507,
-0.05871187523007393,
-0.024766134098172188,
0.05611953139305115,
0.0570615790784359,
0.048267364501953125,
-0.0812348872423172,
-0.1308322250843048,
0.03231794387102127,
-0.0002589701907709241,
0.0787927433848381,
-0.17796193063259125,
-0.08306116610765457,
-0.1061439961194992,
0.030828440561890602,
-0.0450717993080616,
0.11459917575120926,
0.06655435264110565,
0.008609464392066002,
-0.0021567167714238167,
-0.2692579925060272,
0.011895367875695229,
0.10874278843402863,
-0.06626970320940018,
-0.07257457822561264
] |
null | null | transformers | # Steelskull/Etheria-55b-v0.1

## Merge Details
An attempt to make a functional goliath style merge to create a [Etheria] 55b-200k with two yi-34b-200k models.
due to the merge it 'theoretically' should have a context of 200k but I recommend starting at 32k and moveing up,
as it is unknown (at this time) what the merge has done to the context length.
This is a merge of both VerA and VerB of Etheria-55b (There numbers were surprisingly good), I then created a sacrificial 55B out of the most performant yi-34b-200k Model
and performed a Dare_ties merge and equalize the model into its current state.
### recommended settings and Prompt Format:
Ive tested it up to 32k context using exl2 using these settings:
```
"temp": 0.7,
"temperature_last": true,
"top_p": 1,
"top_k": 0,
"top_a": 0,
"tfs": 1,
"epsilon_cutoff": 0,
"eta_cutoff": 0,
"typical_p": 1,
"min_p": 0.1,
"rep_pen": 1.1,
"rep_pen_range": 8192,
"no_repeat_ngram_size": 0,
"penalty_alpha": 0,
"num_beams": 1,
"length_penalty": 1,
"min_length": 0,
"encoder_rep_pen": 1,
"freq_pen": 0,
"presence_pen": 0,
"do_sample": true,
"early_stopping": false,
"add_bos_token": false,
"truncation_length": 2048,
"ban_eos_token": true,
"skip_special_tokens": true,
"streaming": true,
"mirostat_mode": 0,
"mirostat_tau": 5,
"mirostat_eta": 0.1,
```
Prompt format that work well
```
ChatML & Alpaca
```
### Merge Method
This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using Merged-Etheria-55b as a base.
### Configuration
The following YAML configuration was used to produce this model:
```yaml
base_model: Merged-Etheria-55b
models:
- model: Sacr-Etheria-55b
parameters:
weight: [0.22, 0.113, 0.113, 0.113, 0.113, 0.113]
density: 0.61
- model: Merged-Etheria-55b
parameters:
weight: [0.22, 0.113, 0.113, 0.113, 0.113, 0.113]
density: 0.61
merge_method: dare_ties
tokenizer_source: union
parameters:
int8_mask: true
dtype: bfloat16
``` | {"license": "apache-2.0", "tags": ["mergekit", "Etheria"], "base_model": []} | text-generation | waldie/Etheria-55b-v0.1-2.5bpw-h6-exl2 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"mergekit",
"Etheria",
"arxiv:2311.03099",
"arxiv:2306.01708",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T16:09:57+00:00 | [
"2311.03099",
"2306.01708"
] | [] | TAGS
#transformers #safetensors #llama #text-generation #mergekit #Etheria #arxiv-2311.03099 #arxiv-2306.01708 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # Steelskull/Etheria-55b-v0.1
!image/png
## Merge Details
An attempt to make a functional goliath style merge to create a [Etheria] 55b-200k with two yi-34b-200k models.
due to the merge it 'theoretically' should have a context of 200k but I recommend starting at 32k and moveing up,
as it is unknown (at this time) what the merge has done to the context length.
This is a merge of both VerA and VerB of Etheria-55b (There numbers were surprisingly good), I then created a sacrificial 55B out of the most performant yi-34b-200k Model
and performed a Dare_ties merge and equalize the model into its current state.
### recommended settings and Prompt Format:
Ive tested it up to 32k context using exl2 using these settings:
Prompt format that work well
### Merge Method
This model was merged using the DARE TIES merge method using Merged-Etheria-55b as a base.
### Configuration
The following YAML configuration was used to produce this model:
| [
"# Steelskull/Etheria-55b-v0.1\n\n!image/png",
"## Merge Details\n\nAn attempt to make a functional goliath style merge to create a [Etheria] 55b-200k with two yi-34b-200k models.\n\ndue to the merge it 'theoretically' should have a context of 200k but I recommend starting at 32k and moveing up,\nas it is unknown (at this time) what the merge has done to the context length.\n\nThis is a merge of both VerA and VerB of Etheria-55b (There numbers were surprisingly good), I then created a sacrificial 55B out of the most performant yi-34b-200k Model\nand performed a Dare_ties merge and equalize the model into its current state.",
"### recommended settings and Prompt Format:\n\nIve tested it up to 32k context using exl2 using these settings:\n\n\n\nPrompt format that work well",
"### Merge Method\n\nThis model was merged using the DARE TIES merge method using Merged-Etheria-55b as a base.",
"### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #mergekit #Etheria #arxiv-2311.03099 #arxiv-2306.01708 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Steelskull/Etheria-55b-v0.1\n\n!image/png",
"## Merge Details\n\nAn attempt to make a functional goliath style merge to create a [Etheria] 55b-200k with two yi-34b-200k models.\n\ndue to the merge it 'theoretically' should have a context of 200k but I recommend starting at 32k and moveing up,\nas it is unknown (at this time) what the merge has done to the context length.\n\nThis is a merge of both VerA and VerB of Etheria-55b (There numbers were surprisingly good), I then created a sacrificial 55B out of the most performant yi-34b-200k Model\nand performed a Dare_ties merge and equalize the model into its current state.",
"### recommended settings and Prompt Format:\n\nIve tested it up to 32k context using exl2 using these settings:\n\n\n\nPrompt format that work well",
"### Merge Method\n\nThis model was merged using the DARE TIES merge method using Merged-Etheria-55b as a base.",
"### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
80,
17,
150,
37,
31,
17
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #mergekit #Etheria #arxiv-2311.03099 #arxiv-2306.01708 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Steelskull/Etheria-55b-v0.1\n\n!image/png## Merge Details\n\nAn attempt to make a functional goliath style merge to create a [Etheria] 55b-200k with two yi-34b-200k models.\n\ndue to the merge it 'theoretically' should have a context of 200k but I recommend starting at 32k and moveing up,\nas it is unknown (at this time) what the merge has done to the context length.\n\nThis is a merge of both VerA and VerB of Etheria-55b (There numbers were surprisingly good), I then created a sacrificial 55B out of the most performant yi-34b-200k Model\nand performed a Dare_ties merge and equalize the model into its current state.### recommended settings and Prompt Format:\n\nIve tested it up to 32k context using exl2 using these settings:\n\n\n\nPrompt format that work well### Merge Method\n\nThis model was merged using the DARE TIES merge method using Merged-Etheria-55b as a base.### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
-0.10943952947854996,
-0.10446395725011826,
-0.005232003517448902,
0.018328674137592316,
0.061857376247644424,
0.0309735257178545,
0.07275042682886124,
0.06529273837804794,
-0.004784960299730301,
0.09383053332567215,
-0.03628240525722504,
0.05124363303184509,
0.08081107586622238,
0.2208957076072693,
-0.019696379080414772,
-0.1235513985157013,
0.0428592674434185,
-0.042016781866550446,
-0.05899042636156082,
0.0929655134677887,
0.10397744178771973,
-0.11195550858974457,
0.06682460010051727,
0.01345411129295826,
-0.08024806529283524,
0.013378432020545006,
-0.03744255006313324,
0.04409795254468918,
0.11004608124494553,
0.06468333303928375,
0.07232252508401871,
0.006391345988959074,
0.011162620969116688,
-0.1397426575422287,
0.029614903032779694,
0.02156199887394905,
-0.0021985822822898626,
0.05594278499484062,
0.05943549424409866,
0.03990740701556206,
0.08210603892803192,
-0.11515222489833832,
-0.02958434447646141,
0.06029503792524338,
-0.09756199270486832,
-0.08999074250459671,
-0.13822636008262634,
0.07144752889871597,
0.09960167855024338,
0.04227045550942421,
-0.036004941910505295,
0.10689269751310349,
0.043698523193597794,
0.038257576525211334,
0.023763267323374748,
-0.23035019636154175,
-0.006014341954141855,
0.05385405197739601,
0.06749485433101654,
-0.0016338168643414974,
-0.02997344732284546,
-0.0070805419236421585,
0.05662671476602554,
0.02116132713854313,
0.056398361921310425,
-0.02656618505716324,
0.12347929179668427,
-0.07125341147184372,
-0.12768933176994324,
-0.020250670611858368,
0.12890073657035828,
0.03133610635995865,
-0.0464356429874897,
-0.10400573164224625,
-0.14717699587345123,
0.009162669070065022,
0.024696001783013344,
-0.08530580997467041,
0.024231137707829475,
0.018559105694293976,
0.1307634562253952,
-0.10644713044166565,
-0.1126585602760315,
0.01066503394395113,
-0.1331484466791153,
0.14835798740386963,
0.04433973506093025,
0.02639656513929367,
-0.01576913706958294,
0.057041071355342865,
-0.19743794202804565,
-0.08305581659078598,
-0.050163526087999344,
-0.05851835757493973,
-0.12632329761981964,
-0.049696095287799835,
-0.02140875905752182,
-0.11188270896673203,
0.028971930965781212,
0.18873755633831024,
-0.1569812297821045,
0.020325371995568275,
0.03784731402993202,
0.06138412281870842,
0.05750202015042305,
0.020343760028481483,
-0.056660499423742294,
-0.08361253142356873,
0.03849118575453758,
0.04767495393753052,
0.10712165385484695,
0.04209711030125618,
-0.003629756160080433,
-0.0572672076523304,
-0.03310541436076164,
-0.0011296679731458426,
0.03144507855176926,
0.02191038243472576,
-0.061200592666864395,
-0.030871395021677017,
0.12145324796438217,
-0.12550033628940582,
-0.0007464709342457354,
-0.04235061630606651,
-0.017304591834545135,
-0.0033838360104709864,
0.0659521296620369,
-0.0006404850864782929,
-0.045492589473724365,
0.10302181541919708,
-0.05117880553007126,
-0.014570019207894802,
-0.06828414648771286,
-0.08686903864145279,
0.03259936347603798,
-0.04211961477994919,
-0.04995495080947876,
-0.10581278055906296,
-0.17487038671970367,
-0.043931663036346436,
0.03605557605624199,
-0.023251157253980637,
-0.005304520484060049,
0.01187900546938181,
0.002705501625314355,
-0.01417906116694212,
-0.00769998412579298,
0.0028228533919900656,
-0.018960488960146904,
-0.0023094331845641136,
0.030766155570745468,
0.0429832823574543,
-0.06257209926843643,
0.007821145467460155,
-0.07800941914319992,
0.07915262877941132,
-0.17461510002613068,
0.08633539825677872,
-0.03827532008290291,
-0.030622348189353943,
-0.1349686086177826,
-0.013248167932033539,
-0.10293876379728317,
0.02233383059501648,
0.0605028010904789,
0.12772876024246216,
-0.1602603942155838,
-0.02732502482831478,
0.1890861988067627,
-0.15286654233932495,
-0.10599873960018158,
0.10490400344133377,
0.0036951242946088314,
0.05537373572587967,
0.06145002320408821,
0.05630245804786682,
0.14451774954795837,
-0.12823502719402313,
-0.08591429144144058,
-0.014173904433846474,
-0.05243324115872383,
0.029766641557216644,
0.04798873886466026,
-0.03619996830821037,
0.045599061995744705,
-0.0014145770110189915,
-0.015943868085741997,
0.06110193580389023,
-0.005981776397675276,
-0.040823835879564285,
-0.05430952087044716,
-0.04028356820344925,
0.05191149562597275,
-0.032790619879961014,
-0.003258726093918085,
-0.010376677848398685,
-0.07458120584487915,
-0.02369185909628868,
0.11761699616909027,
0.006007341668009758,
0.03512540087103844,
-0.058320965617895126,
0.11811016499996185,
-0.09691473841667175,
0.028743892908096313,
-0.13570477068424225,
-0.0492108054459095,
0.01625555008649826,
-0.038357265293598175,
0.03522238880395889,
0.08505421131849289,
0.07327118515968323,
0.050320692360401154,
-0.06564534455537796,
-0.05887473002076149,
0.0025470778346061707,
-0.01977710984647274,
-0.07093407213687897,
-0.11775345355272293,
-0.052384622395038605,
-0.07004649192094803,
0.09775111824274063,
-0.042414695024490356,
0.04114162549376488,
0.010192375630140305,
0.14680826663970947,
-0.018884435296058655,
-0.025101639330387115,
0.02020084112882614,
-0.0011204986367374659,
-0.04900673031806946,
-0.026770571246743202,
0.019641438499093056,
0.0056570847518742085,
-0.08374296873807907,
0.10508554428815842,
-0.17906935513019562,
-0.0711081400513649,
0.07735289633274078,
0.04172023758292198,
-0.049782827496528625,
-0.081593818962574,
-0.042073123157024384,
-0.0367065891623497,
-0.06851647794246674,
-0.09333997964859009,
0.12004616856575012,
0.04410794749855995,
0.09481856226921082,
-0.04353799670934677,
-0.06190011650323868,
0.004387145861983299,
0.007205356378108263,
-0.010883238166570663,
0.04327830672264099,
0.015472795814275742,
-0.1709488332271576,
0.06793433427810669,
0.1339443475008011,
0.02820632793009281,
0.04926494136452675,
-0.00037272757617756724,
-0.06710237264633179,
-0.07205095142126083,
0.11168516427278519,
-0.015369158238172531,
0.03600122034549713,
-0.016130799427628517,
0.044063352048397064,
0.02909441851079464,
-0.0024773008190095425,
0.04419197887182236,
-0.0863647609949112,
0.07600856572389603,
0.03272629156708717,
-0.01084850262850523,
0.10669423639774323,
0.053150616586208344,
0.01823967508971691,
0.05238606035709381,
0.024264710023999214,
0.05880584195256233,
0.006869439035654068,
-0.006282936315983534,
-0.09690307825803757,
0.15447427332401276,
-0.08835307508707047,
-0.121907077729702,
-0.17632274329662323,
0.01999291405081749,
-0.03571661561727524,
0.003110327757894993,
0.03562290966510773,
-0.03928403928875923,
-0.08989109843969345,
-0.05200701206922531,
0.12358148396015167,
0.015124117955565453,
-0.017785264179110527,
0.011853360570967197,
-0.010753941722214222,
0.08811288326978683,
-0.08653265237808228,
-0.025031544268131256,
0.012974771670997143,
-0.008839976973831654,
0.04533082991838455,
0.07703752815723419,
0.09134586900472641,
0.05054210498929024,
-0.025393137708306313,
-0.02338995411992073,
0.01539660058915615,
0.22042019665241241,
-0.062169838696718216,
0.11229768395423889,
0.17903335392475128,
-0.030679745599627495,
0.11189977824687958,
0.17240828275680542,
0.022464970126748085,
-0.038127392530441284,
0.011109970510005951,
0.030805833637714386,
-0.023231830447912216,
-0.22295406460762024,
-0.09294198453426361,
-0.09522822499275208,
-0.08801252394914627,
0.01754019223153591,
0.05335405096411705,
-0.04569769278168678,
0.06716098636388779,
-0.12401210516691208,
0.019167670980095863,
0.018258048221468925,
0.058812472969293594,
0.17871776223182678,
-0.013314965181052685,
0.05517899617552757,
-0.020888157188892365,
-0.034729741513729095,
0.10491572320461273,
0.02704397588968277,
0.21006418764591217,
-0.0009596895542927086,
0.08474936336278915,
0.08947689086198807,
0.041876956820487976,
-0.012397797778248787,
0.08698198199272156,
-0.023151731118559837,
0.04385404288768768,
-0.023138446733355522,
-0.07412482053041458,
0.02194851264357567,
0.11354448646306992,
-0.020995093509554863,
0.037039946764707565,
0.002639985177665949,
0.07504837960004807,
0.09565421938896179,
0.1058611273765564,
0.05912692844867706,
-0.201351597905159,
-0.10644569993019104,
0.041142866015434265,
-0.015804750844836235,
-0.050857942551374435,
-0.04924383386969566,
0.09605693072080612,
-0.0276027899235487,
0.1396915465593338,
-0.05209039896726608,
0.0765918493270874,
-0.03845449164509773,
0.004542296286672354,
-0.019064296036958694,
0.1283712536096573,
0.019018065184354782,
0.05205632001161575,
-0.09524990618228912,
0.15999630093574524,
0.028904184699058533,
0.05501687899231911,
-0.009830045513808727,
0.06387419253587723,
0.04290340468287468,
0.05838203430175781,
0.08944016695022583,
0.04829339310526848,
0.01591643877327442,
-0.06741682440042496,
-0.09917420148849487,
-0.013835197314620018,
0.058088816702365875,
-0.014466996304690838,
0.09875226765871048,
-0.03281408175826073,
-0.02245858684182167,
-0.011051773093640804,
0.11747866123914719,
-0.09406507015228271,
-0.11116848886013031,
0.08485017716884613,
0.017779096961021423,
-0.03359028697013855,
-0.07368392497301102,
-0.024709919467568398,
-0.0724799707531929,
0.23719869554042816,
-0.06952031701803207,
-0.09528033435344696,
-0.11833707988262177,
0.04836568981409073,
0.14628329873085022,
-0.09464559704065323,
0.04847116395831108,
-0.006792669650167227,
0.08983587473630905,
-0.10221723467111588,
-0.1008136123418808,
0.03247491270303726,
-0.0706489160656929,
-0.11615795642137527,
0.0008078424725681543,
0.13381585478782654,
0.002635522512719035,
0.018508825451135635,
0.0073898169212043285,
0.039598166942596436,
-0.022691717371344566,
-0.07262139767408371,
0.05046295002102852,
0.16029618680477142,
0.0505424328148365,
0.04811425134539604,
-0.03659951314330101,
-0.15673859417438507,
-0.026911193504929543,
-0.03115687146782875,
0.0925721824169159,
0.24508625268936157,
-0.03751545026898384,
0.0637691393494606,
0.13594430685043335,
-0.07402586936950684,
-0.1694311499595642,
-0.05999257415533066,
0.02900422178208828,
0.06619575619697571,
0.022504322230815887,
-0.07307258248329163,
0.08633725345134735,
0.08359691500663757,
-0.014108659699559212,
0.04309108480811119,
-0.20058414340019226,
-0.12942445278167725,
-0.020861325785517693,
-0.013379338197410107,
0.17127227783203125,
-0.06713000684976578,
-0.09872110188007355,
-0.019538618624210358,
-0.06687154620885849,
-0.007139192894101143,
-0.07655689865350723,
0.0597379095852375,
-0.019155355170369148,
0.05177086591720581,
0.04078345373272896,
-0.05566057190299034,
0.13806800544261932,
-0.041897669434547424,
0.018457898870110512,
-0.06341537088155746,
0.011214425787329674,
0.07800929248332977,
-0.07831251621246338,
0.10354043543338776,
-0.04776172712445259,
0.03017532080411911,
-0.09958693385124207,
-0.03354588896036148,
-0.06237098574638367,
0.07512446492910385,
-0.03139052167534828,
-0.039742108434438705,
-0.05994221195578575,
0.04833868891000748,
0.09459476917982101,
-0.01931736245751381,
0.04818810522556305,
0.019879251718521118,
0.046284399926662445,
0.16048157215118408,
0.10322637856006622,
0.022586697712540627,
-0.10162033885717392,
0.008304250426590443,
-0.020696785300970078,
0.03956984356045723,
-0.13094034790992737,
0.027177954092621803,
0.12577833235263824,
-0.014104821719229221,
0.11395157128572464,
0.006588481366634369,
-0.13671652972698212,
0.02129327319562435,
0.1076834574341774,
-0.10904709249734879,
-0.2416590005159378,
-0.013548976741731167,
0.019171353429555893,
-0.09254088997840881,
0.0161429475992918,
0.16323114931583405,
-0.030513957142829895,
-0.010014905594289303,
0.007125226315110922,
0.06654992699623108,
-0.03551382198929787,
0.12911564111709595,
-0.008959786035120487,
0.06030282750725746,
-0.05704895034432411,
0.06408479064702988,
0.05001753941178322,
-0.11317332834005356,
0.004510906524956226,
0.11660301685333252,
-0.08863106369972229,
-0.07743901759386063,
-0.04510286822915077,
0.018556982278823853,
-0.023153379559516907,
-0.03041745349764824,
-0.018580235540866852,
-0.07938505709171295,
0.006095242220908403,
0.08030784130096436,
0.024216029793024063,
-0.008510199375450611,
0.02436128258705139,
0.006168700754642487,
-0.02513054572045803,
0.09664779156446457,
0.09204736351966858,
0.07062084227800369,
-0.11632400751113892,
0.10487501323223114,
-0.04324759915471077,
0.023008355870842934,
-0.01796892285346985,
0.03780508413910866,
-0.024001432582736015,
-0.034401603043079376,
-0.15718227624893188,
0.022654835134744644,
-0.09417688101530075,
-0.05580664053559303,
-0.03259299322962761,
-0.024696655571460724,
-0.02933518961071968,
0.038989488035440445,
-0.0574825145304203,
-0.05095026269555092,
-0.04321815446019173,
-0.0003308576997369528,
-0.1038406640291214,
-0.019357386976480484,
0.020932210609316826,
-0.07826444506645203,
0.073189377784729,
0.0041158911772072315,
-0.01616940274834633,
-0.028351860120892525,
-0.051958758383989334,
-0.04560166224837303,
0.08830931037664413,
0.04237137734889984,
0.0346984900534153,
-0.097996786236763,
0.01593981496989727,
-0.00964298378676176,
-0.027857769280672073,
-0.041098594665527344,
0.07088202238082886,
-0.07738762348890305,
0.015946241095662117,
-0.013717123307287693,
-0.02868007682263851,
-0.07109881937503815,
-0.010687325149774551,
0.06817642599344254,
0.06875927001237869,
0.11901332437992096,
-0.04505154490470886,
0.028652438893914223,
-0.1726224720478058,
0.010070445016026497,
0.0042783948592841625,
-0.05271821841597557,
-0.010835321620106697,
-0.08382751047611237,
0.036409053951501846,
-0.04532187432050705,
0.08589673787355423,
-0.03201473504304886,
-0.04314291477203369,
0.05484916642308235,
-0.0652681365609169,
-0.1186334565281868,
-0.011840811930596828,
0.05608929321169853,
-0.019104989245533943,
0.002434992929920554,
-0.05894865468144417,
-0.005081210285425186,
0.014833562076091766,
-0.11153209954500198,
0.11392863094806671,
0.11728785932064056,
0.09024875611066818,
0.051291488111019135,
0.05854247510433197,
-0.11190133541822433,
-0.0790833979845047,
-0.08576387166976929,
-0.09098249673843384,
0.04790103808045387,
0.0036997413262724876,
0.14253100752830505,
0.0846179872751236,
-0.13174018263816833,
0.08109606802463531,
-0.03637560084462166,
-0.03083486109972,
-0.06248346343636513,
-0.06026808172464371,
-0.061740562319755554,
-0.03490849956870079,
-0.008822082541882992,
-0.08993697166442871,
0.004471474792808294,
0.012414501048624516,
-0.0003331670304760337,
-0.013066901825368404,
0.14888089895248413,
-0.0039777145721018314,
-0.07642006874084473,
0.03592916205525398,
0.028719371184706688,
0.012015021406114101,
0.07446130365133286,
-0.045931290835142136,
0.062137238681316376,
0.026723111048340797,
0.03696955740451813,
0.035233817994594574,
0.06793112307786942,
0.027518341317772865,
-0.029728123918175697,
-0.07325179129838943,
-0.023526139557361603,
0.02992156520485878,
0.024214569479227066,
0.007816719822585583,
0.018206244334578514,
-0.010703833773732185,
-0.043162792921066284,
0.06574855744838715,
-0.0441068559885025,
-0.081887386739254,
-0.08937965333461761,
0.17699319124221802,
0.0174170583486557,
0.008383885957300663,
0.025799142196774483,
-0.14044618606567383,
0.026106080040335655,
0.13566288352012634,
0.17131853103637695,
0.0369759202003479,
-0.005929701961576939,
0.027940889820456505,
0.0027207843959331512,
0.028013551607728004,
0.021570732817053795,
0.04326905310153961,
0.16356715559959412,
-0.03867912292480469,
0.13342249393463135,
-0.027226878330111504,
-0.034634802490472794,
-0.011825336143374443,
0.07372632622718811,
-0.06472178548574448,
0.04661593586206436,
-0.012737949378788471,
0.06398273259401321,
0.031252484768629074,
-0.21528935432434082,
0.07399392127990723,
-0.10734318196773529,
-0.09675508737564087,
-0.039480190724134445,
0.05550624430179596,
0.013585202395915985,
0.05593801289796829,
-0.01490672305226326,
0.005558918695896864,
0.14389647543430328,
0.007989275269210339,
-0.11275006830692291,
-0.032978836447000504,
0.037572525441646576,
-0.12086831778287888,
0.04667912423610687,
0.010508783161640167,
0.05314922332763672,
0.10324297845363617,
-0.04140359163284302,
-0.1340358704328537,
0.05615480989217758,
0.020454345270991325,
-0.06250754743814468,
0.02777022495865822,
0.12130796909332275,
-0.024158578366041183,
0.0022998186759650707,
0.04669337347149849,
-0.10290127247571945,
-0.006495468784123659,
0.07108736038208008,
-0.06093502789735794,
-0.07835794985294342,
0.0914028212428093,
-0.06830094754695892,
0.13738517463207245,
0.17387232184410095,
-0.01856829784810543,
0.014144426211714745,
-0.059187255799770355,
0.05226492881774902,
0.06078827381134033,
0.14278438687324524,
-0.019479434937238693,
-0.19742940366268158,
0.04307175427675247,
0.051406946033239365,
0.09401006251573563,
-0.22785963118076324,
-0.07901472598314285,
-0.035550858825445175,
0.0030907941982150078,
-0.04726891219615936,
0.13595646619796753,
0.09779863804578781,
0.012836464680731297,
0.009164680726826191,
-0.1028255969285965,
-0.021083468571305275,
0.04950213432312012,
-0.08770444244146347,
-0.008882980793714523
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# roberta-base-detect-cheapfake-combined-train-test-contradict
This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.4261
- Accuracy: 0.89
- F1: 0.8817
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-06
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 5
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
| No log | 1.0 | 166 | 0.4435 | 0.84 | 0.8333 |
| No log | 2.0 | 332 | 0.6567 | 0.835 | 0.8374 |
| No log | 3.0 | 498 | 0.3563 | 0.895 | 0.88 |
| 0.2851 | 4.0 | 664 | 0.3671 | 0.895 | 0.8814 |
| 0.2851 | 5.0 | 830 | 0.4261 | 0.89 | 0.8817 |
### Framework versions
- Transformers 4.37.0
- Pytorch 2.1.2
- Datasets 2.1.0
- Tokenizers 0.15.1
| {"license": "mit", "tags": ["generated_from_trainer"], "metrics": ["accuracy", "f1"], "base_model": "roberta-base", "model-index": [{"name": "roberta-base-detect-cheapfake-combined-train-test-contradict", "results": []}]} | text-classification | hoanghoavienvo/roberta-base-detect-cheapfake-combined-train-test-contradict | [
"transformers",
"tensorboard",
"safetensors",
"roberta",
"text-classification",
"generated_from_trainer",
"base_model:roberta-base",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T16:10:05+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #roberta #text-classification #generated_from_trainer #base_model-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us
| roberta-base-detect-cheapfake-combined-train-test-contradict
============================================================
This model is a fine-tuned version of roberta-base on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.4261
* Accuracy: 0.89
* F1: 0.8817
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-06
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 5
### Training results
### Framework versions
* Transformers 4.37.0
* Pytorch 2.1.2
* Datasets 2.1.0
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.1.0\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #roberta #text-classification #generated_from_trainer #base_model-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.1.0\n* Tokenizers 0.15.1"
] | [
63,
98,
4,
30
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #roberta #text-classification #generated_from_trainer #base_model-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5### Training results### Framework versions\n\n\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.1.0\n* Tokenizers 0.15.1"
] | [
-0.08600206673145294,
0.06570498645305634,
-0.0019204990239813924,
0.10113903135061264,
0.16684317588806152,
0.014031677506864071,
0.1624070405960083,
0.10871861129999161,
-0.10294260829687119,
0.037401098757982254,
0.12606362998485565,
0.15590128302574158,
-0.0037771807983517647,
0.13376674056053162,
-0.07749587297439575,
-0.24532493948936462,
0.0018245121464133263,
0.032166458666324615,
-0.08469920605421066,
0.11326869577169418,
0.1062554121017456,
-0.13537107408046722,
0.08651774376630783,
-0.008616765961050987,
-0.20578622817993164,
0.036329545080661774,
0.043829046189785004,
-0.06629844754934311,
0.13888169825077057,
0.04477453976869583,
0.13803698122501373,
0.0323466919362545,
0.0825287401676178,
-0.18861161172389984,
0.01921221986413002,
0.060024429112672806,
-0.016164105385541916,
0.08551991730928421,
0.046420611441135406,
-0.02915424481034279,
0.1088266521692276,
-0.09399721026420593,
0.06470876932144165,
0.022912081331014633,
-0.13064663112163544,
-0.2047857642173767,
-0.07178008556365967,
0.03512466326355934,
0.09130682051181793,
0.07980083674192429,
-0.01903410442173481,
0.15087150037288666,
-0.06302589178085327,
0.10190152376890182,
0.20320191979408264,
-0.30903124809265137,
-0.06929448246955872,
0.05669908970594406,
0.026419779285788536,
0.08606826514005661,
-0.11623882502317429,
0.0033263445366173983,
0.07934031635522842,
0.020229138433933258,
0.1229044497013092,
-0.03699538856744766,
-0.06325984001159668,
0.01344413310289383,
-0.14060814678668976,
0.0037721828557550907,
0.14584200084209442,
0.04409114271402359,
-0.04269495978951454,
-0.041655585169792175,
-0.05720981955528259,
-0.12549518048763275,
-0.04125942289829254,
-0.03147982433438301,
0.04226845130324364,
-0.028866339474916458,
-0.10943150520324707,
-0.019738169386982918,
-0.1176222413778305,
-0.07366719096899033,
-0.062438610941171646,
0.17537833750247955,
0.03380125015974045,
0.0019474402070045471,
-0.0319366417825222,
0.09073222428560257,
-0.04038610681891441,
-0.1266063004732132,
0.018815528601408005,
0.02527826838195324,
0.004162331577390432,
-0.07341950386762619,
-0.06586432456970215,
-0.09630368649959564,
0.03010738454759121,
0.1509854644536972,
-0.06678813695907593,
0.04953448474407196,
0.008400734513998032,
0.04389837756752968,
-0.09280949085950851,
0.15940865874290466,
-0.044783689081668854,
-0.0169754009693861,
0.029429223388433456,
0.05615467578172684,
0.03744379058480263,
-0.0024727757554501295,
-0.1246785968542099,
0.01947326585650444,
0.12125901877880096,
0.014644557610154152,
-0.07602428644895554,
0.08613777905702591,
-0.037704285234212875,
0.004385835491120815,
0.013026240281760693,
-0.09181790053844452,
0.03402935713529587,
0.0032550308387726545,
-0.060247208923101425,
-0.06886418163776398,
0.02590356580913067,
0.01778290420770645,
0.011470084078609943,
0.10555706173181534,
-0.093335822224617,
0.020108036696910858,
-0.09201612323522568,
-0.13197040557861328,
0.008743475191295147,
-0.06267327070236206,
0.03801275044679642,
-0.11749759316444397,
-0.15898963809013367,
-0.015586327761411667,
0.04713701456785202,
-0.027066528797149658,
-0.02102554216980934,
-0.055265914648771286,
-0.08224213123321533,
0.011702914722263813,
-0.015468253754079342,
0.09630343317985535,
-0.05475189536809921,
0.09865522384643555,
0.06298230588436127,
0.06788348406553268,
-0.05591602250933647,
0.03211529180407524,
-0.1029910072684288,
0.015147479251027107,
-0.21370276808738708,
0.023109402507543564,
-0.06085566058754921,
0.07834313809871674,
-0.07653170824050903,
-0.0808948501944542,
-0.006574357859790325,
0.024290138855576515,
0.07191057503223419,
0.0894995853304863,
-0.14577224850654602,
-0.07529686391353607,
0.16337785124778748,
-0.10338330268859863,
-0.13688550889492035,
0.12065047770738602,
-0.07153736054897308,
0.07209628075361252,
0.0748576670885086,
0.19548174738883972,
0.06132514029741287,
-0.08926330506801605,
0.0066939485259354115,
-0.01213493850082159,
0.03845352679491043,
-0.04769308492541313,
0.057879090309143066,
0.008526604622602463,
-0.0030283082742244005,
0.012297124601900578,
-0.022464055567979813,
0.04851769283413887,
-0.093661829829216,
-0.07932521402835846,
-0.0387558713555336,
-0.10662782937288284,
0.051617611199617386,
0.06220657005906105,
0.0837467759847641,
-0.13305789232254028,
-0.08764820545911789,
0.10104424506425858,
0.07147590816020966,
-0.0685998797416687,
0.01935453526675701,
-0.07937340438365936,
0.07020233571529388,
-0.06111475080251694,
-0.029131611809134483,
-0.16314132511615753,
-0.04840550571680069,
-0.003482039552181959,
0.03194485232234001,
0.033155616372823715,
0.014387257397174835,
0.08066859096288681,
0.07226605713367462,
-0.0763646587729454,
-0.02588605135679245,
0.0032573642674833536,
0.015534065663814545,
-0.12729492783546448,
-0.20918072760105133,
-0.0013232182245701551,
-0.04167552292346954,
0.13347198069095612,
-0.25410863757133484,
0.05544235184788704,
0.006287914700806141,
0.08696035295724869,
0.04103462025523186,
0.003737021004781127,
-0.042151667177677155,
0.07147654891014099,
-0.051059044897556305,
-0.05177261680364609,
0.05058343708515167,
0.005600373726338148,
-0.08958333730697632,
-0.040142275393009186,
-0.15791749954223633,
0.20134836435317993,
0.14240700006484985,
-0.11618336290121078,
-0.10778796672821045,
-0.0007763280882500112,
-0.039731454104185104,
-0.02182142436504364,
-0.04892623797059059,
0.014210013672709465,
0.12411560118198395,
-0.0278097465634346,
0.1544010490179062,
-0.07313669472932816,
-0.036437321454286575,
0.02126617170870304,
-0.06443209946155548,
0.009150391444563866,
0.10837212949991226,
0.0852319747209549,
-0.13530895113945007,
0.1517031043767929,
0.14953050017356873,
-0.1082901656627655,
0.15182219445705414,
-0.03230705484747887,
-0.05271834507584572,
-0.021004976704716682,
0.0012564844219014049,
0.015378988347947598,
0.10907047241926193,
-0.11395707726478577,
-0.010751946829259396,
-0.0005980590940453112,
0.0077028642408549786,
0.022165216505527496,
-0.22158801555633545,
-0.035577163100242615,
0.029754944145679474,
-0.03274324908852577,
0.014843755401670933,
-0.020333606749773026,
-0.008372385986149311,
0.10386443883180618,
-0.003663548966869712,
-0.08077026158571243,
0.043309714645147324,
0.0015276604099199176,
-0.09149713814258575,
0.22322238981723785,
-0.07153813540935516,
-0.12074525654315948,
-0.13137632608413696,
-0.06308228522539139,
-0.043772973120212555,
0.03408372402191162,
0.06822584569454193,
-0.08607363700866699,
-0.039108362048864365,
-0.09637460857629776,
0.009059542790055275,
0.020781824365258217,
0.033156901597976685,
-0.00290565169416368,
0.01552219782024622,
0.07988345623016357,
-0.10835301876068115,
-0.007525399327278137,
-0.055700477212667465,
-0.07228939235210419,
0.04614802822470665,
0.0309445820748806,
0.11669643223285675,
0.14390969276428223,
-0.04455425962805748,
-0.0016189898597076535,
-0.044340264052152634,
0.2224777489900589,
-0.07009442150592804,
-0.017771869897842407,
0.11954013258218765,
-0.018320824950933456,
0.03973684832453728,
0.13595901429653168,
0.06311015039682388,
-0.09661594033241272,
0.03393080458045006,
0.04356337711215019,
-0.0318731926381588,
-0.21769411861896515,
-0.029880298301577568,
-0.029461553320288658,
-0.01112416572868824,
0.08454971760511398,
0.041074689477682114,
0.0506003312766552,
0.07802299410104752,
0.03330901265144348,
0.067042775452137,
-0.005600837990641594,
0.07555030286312103,
0.10777422785758972,
0.0470406748354435,
0.13668203353881836,
-0.058996137231588364,
-0.07814720273017883,
0.027194490656256676,
-0.016419416293501854,
0.19305957853794098,
0.022849194705486298,
0.11025597900152206,
0.05608442425727844,
0.14252160489559174,
0.01256171241402626,
0.06686040014028549,
-0.0024771210737526417,
-0.06585650146007538,
-0.002100203651934862,
-0.0458076149225235,
-0.010579045861959457,
0.04101809486746788,
-0.09778973460197449,
0.05213257297873497,
-0.11695561558008194,
0.013808807358145714,
0.06934473663568497,
0.20384742319583893,
0.05549527332186699,
-0.32642295956611633,
-0.09619461745023727,
0.02715008333325386,
-0.025049269199371338,
-0.016552869230508804,
0.023795509710907936,
0.11624331772327423,
-0.04708203300833702,
0.04273158684372902,
-0.06883805245161057,
0.0767572820186615,
-0.023941799998283386,
0.04416705295443535,
0.03955771028995514,
0.09773968905210495,
-0.028529426082968712,
0.06964021921157837,
-0.2960733473300934,
0.2824225425720215,
0.02115178294479847,
0.08935544639825821,
-0.04636169970035553,
-0.010336963459849358,
0.02977616898715496,
0.10161649435758591,
0.06359560042619705,
-0.030371813103556633,
-0.10210783779621124,
-0.19828763604164124,
-0.029646601527929306,
0.03092193603515625,
0.11082546412944794,
-0.017928892746567726,
0.10420101881027222,
-0.03447194769978523,
0.004432912915945053,
0.09625886380672455,
-0.03452983498573303,
-0.09112965315580368,
-0.08752024173736572,
-0.033395931124687195,
0.024043293669819832,
-0.029478803277015686,
-0.0796312764286995,
-0.1015576496720314,
-0.12886832654476166,
0.17404265701770782,
-0.028141295537352562,
-0.016257336363196373,
-0.10575422644615173,
0.08967835456132889,
0.03877834230661392,
-0.08467373251914978,
0.043005988001823425,
0.015471834689378738,
0.08408457040786743,
0.02050905115902424,
-0.06294377148151398,
0.1347808837890625,
-0.06644392013549805,
-0.16471487283706665,
-0.06261825561523438,
0.0933605208992958,
0.01980728656053543,
0.042070500552654266,
0.007190473843365908,
0.014069028198719025,
-0.011201734654605389,
-0.07608999311923981,
0.03024030104279518,
-0.036660850048065186,
0.05585634335875511,
0.014291519299149513,
-0.06401387602090836,
-0.0249161459505558,
-0.06267979741096497,
-0.03186555951833725,
0.16774354875087738,
0.2925034761428833,
-0.08705956488847733,
-0.012890418991446495,
0.06667302548885345,
-0.06671396642923355,
-0.21791715919971466,
0.07012869417667389,
0.015688590705394745,
-0.0005508727626875043,
0.05315018072724342,
-0.13830730319023132,
0.12217450886964798,
0.0976458191871643,
-0.024142201989889145,
0.09227113425731659,
-0.2631303668022156,
-0.1461210548877716,
0.1385219544172287,
0.1681899130344391,
0.13642407953739166,
-0.167319193482399,
-0.02761927805840969,
-0.04189432039856911,
-0.07488276809453964,
0.10092807561159134,
-0.1506287008523941,
0.11026853322982788,
0.00132397108245641,
0.05847523733973503,
0.006301793269813061,
-0.05753282830119133,
0.11645213514566422,
-0.017304955050349236,
0.12969911098480225,
-0.069461889564991,
-0.03683657944202423,
0.06825874745845795,
-0.047424640506505966,
0.018676109611988068,
-0.09592319279909134,
0.030861686915159225,
-0.04155285283923149,
-0.03374765068292618,
-0.0543719157576561,
0.05129338800907135,
-0.036910977214574814,
-0.0713779628276825,
-0.05155064910650253,
0.030998878180980682,
0.019162297248840332,
-0.018991630524396896,
0.17431388795375824,
-0.0021916257683187723,
0.18378004431724548,
0.14433836936950684,
0.0918685793876648,
-0.061730384826660156,
0.011522401124238968,
0.011672493070363998,
-0.04255223646759987,
0.05861412733793259,
-0.15848009288311005,
0.04012709856033325,
0.11142519861459732,
0.011596500873565674,
0.15295696258544922,
0.08273988962173462,
-0.035484690219163895,
0.028687266632914543,
0.08806952834129333,
-0.1643003672361374,
-0.11249551177024841,
-0.008491411805152893,
-0.06099560856819153,
-0.10167305171489716,
0.08508658409118652,
0.12067213654518127,
-0.07675067335367203,
-0.000010228853170701768,
-0.013357295654714108,
0.0015474603278562427,
-0.05025937408208847,
0.18463782966136932,
0.09248585999011993,
0.04676408693194389,
-0.07371046394109726,
0.07345208525657654,
0.044360287487506866,
-0.06863001734018326,
0.016674334183335304,
0.056955575942993164,
-0.07227350771427155,
-0.053280171006917953,
0.07179713994264603,
0.21511980891227722,
-0.059403084218502045,
-0.05278991535305977,
-0.16287967562675476,
-0.11201624572277069,
0.04304194450378418,
0.21580280363559723,
0.10082872211933136,
0.001002285280264914,
-0.022358136251568794,
0.027397949248552322,
-0.13861262798309326,
0.10798978805541992,
0.03484213352203369,
0.0879274234175682,
-0.15461449325084686,
0.16512612998485565,
-0.005360427778214216,
0.010019737295806408,
-0.034668974578380585,
0.039581622928380966,
-0.13488544523715973,
-0.0011739585315808654,
-0.12381092458963394,
-0.016641924157738686,
-0.03568603843450546,
0.0069932653568685055,
0.012416807003319263,
-0.06153338402509689,
-0.07457449287176132,
0.010409997776150703,
-0.1073962152004242,
-0.005558272358030081,
0.04493274912238121,
0.05002908036112785,
-0.1295691877603531,
-0.036683935672044754,
0.014733103103935719,
-0.0583021305501461,
0.06226504221558571,
0.01783144660294056,
0.020905349403619766,
0.06738348305225372,
-0.20346784591674805,
0.033343348652124405,
0.07544629275798798,
-0.012910076417028904,
0.051609188318252563,
-0.07588769495487213,
-0.005592795088887215,
-0.0005349900457076728,
0.07422737032175064,
0.024595137685537338,
0.07319614291191101,
-0.12298613041639328,
0.02098679356276989,
-0.04210827872157097,
-0.07647592574357986,
-0.061454273760318756,
0.03553742542862892,
0.0794142559170723,
-0.010964000597596169,
0.2014540135860443,
-0.11333870142698288,
0.006374384742230177,
-0.20699836313724518,
0.01093626581132412,
-0.007388307247310877,
-0.12020152062177658,
-0.11779691278934479,
-0.05315009504556656,
0.055490847676992416,
-0.06340280175209045,
0.1464969515800476,
0.03029843419790268,
0.02339332364499569,
0.04629987105727196,
-0.04580806940793991,
0.04256230965256691,
0.039574261754751205,
0.23725154995918274,
0.022547436878085136,
-0.039084210991859436,
0.010037369094789028,
0.05549633502960205,
0.11830946803092957,
0.07737688720226288,
0.18068163096904755,
0.1652003973722458,
-0.08182809501886368,
0.10699618607759476,
0.045044686645269394,
-0.04366864636540413,
-0.11100716143846512,
0.05670735239982605,
-0.040840450674295425,
0.07038306444883347,
-0.02775939740240574,
0.1650652289390564,
0.13920854032039642,
-0.15198780596256256,
0.010288674384355545,
-0.06569299101829529,
-0.08124461770057678,
-0.12476152181625366,
-0.046910252422094345,
-0.11222826689481735,
-0.1536119431257248,
0.008415354415774345,
-0.11611650884151459,
0.0036440815310925245,
0.086429163813591,
0.009501435793936253,
-0.022987065836787224,
0.17768220603466034,
0.017731500789523125,
0.03962316736578941,
0.043358802795410156,
-0.005782541818916798,
-0.03358069434762001,
-0.08191254734992981,
-0.0794321671128273,
0.0008400292717851698,
-0.03245889022946358,
0.011259132996201515,
-0.052440106868743896,
-0.03617175295948982,
0.04467073455452919,
-0.023869497701525688,
-0.10513920336961746,
0.016769764944911003,
0.04167875647544861,
0.05690552294254303,
0.037040453404188156,
0.010846187360584736,
-0.0034108508843928576,
-0.00277092307806015,
0.23805096745491028,
-0.07271870970726013,
-0.07728935778141022,
-0.09867823123931885,
0.25895071029663086,
0.04562653973698616,
0.030735952779650688,
0.0007592852925881743,
-0.10428781062364578,
0.038160596042871475,
0.2237347662448883,
0.21132224798202515,
-0.06923241168260574,
0.014701937325298786,
-0.020957620814442635,
-0.010034875944256783,
-0.03016051836311817,
0.09847007691860199,
0.10234076529741287,
0.017195994034409523,
-0.08125729113817215,
-0.05019671469926834,
-0.03245088830590248,
-0.0028912608977407217,
-0.04793892800807953,
0.06416001170873642,
0.047758858650922775,
0.020032009109854698,
-0.047392308712005615,
0.06595110893249512,
-0.0373816043138504,
-0.11392609775066376,
0.054299354553222656,
-0.19524027407169342,
-0.13969743251800537,
-0.012957870960235596,
0.11331021040678024,
-0.014619627967476845,
0.06478769332170486,
-0.033419109880924225,
-0.005863076075911522,
0.02170657366514206,
-0.017722755670547485,
-0.07714571803808212,
-0.06826706975698471,
0.0641050711274147,
-0.1041223555803299,
0.21739283204078674,
-0.04525701329112053,
0.04107126593589783,
0.14443112909793854,
0.033774234354496,
-0.06824027746915817,
0.10233456641435623,
0.037964221090078354,
-0.07603737711906433,
0.044372472912073135,
0.09218636155128479,
-0.04564828798174858,
0.10568023473024368,
0.054812751710414886,
-0.15010862052440643,
0.031249772757291794,
-0.09510667622089386,
-0.08479778468608856,
-0.05987589433789253,
-0.0357198528945446,
-0.06428837776184082,
0.12451237440109253,
0.18729613721370697,
-0.029660271480679512,
0.02387276105582714,
-0.05230302736163139,
0.02383478917181492,
0.07828881591558456,
0.028414230793714523,
-0.04396243020892143,
-0.25295016169548035,
0.029862873256206512,
0.09130168706178665,
-0.01637539081275463,
-0.3059150278568268,
-0.07215019315481186,
-0.011401300318539143,
-0.046782925724983215,
-0.09848374873399734,
0.08580055832862854,
0.14107194542884827,
0.05782221257686615,
-0.06355797499418259,
-0.11842998117208481,
-0.07799544930458069,
0.16384336352348328,
-0.13390304148197174,
-0.10499555617570877
] |
null | null | transformers |

---
Presenting the FP16 files for Psyonic-Cetacean-20B! This is an experimental Llama2-based stack merge based on the models and recipe below:
- [KoboldAI/PsyFighter-2-13b](https://huggingface.co/KoboldAI/LLaMA2-13B-Psyfighter2-GGUF)
- [microsoft/Orca-2-13b](https://huggingface.co/microsoft/Orca-2-13b)
```yaml
slices:
- sources:
- model: Orca2flat
layer_range: [0, 16]
- sources:
- model: LLaMA2-13B-Psyfighter2 (FP16 not yet available)
layer_range: [8, 24]
- sources:
- model: Orca2flat
layer_range: [17, 32]
- sources:
- model: LLaMA2-13B-Psyfighter2 (FP16 not yet available)
layer_range: [25, 40]
merge_method: passthrough
dtype: float16
```
Note: while we did run an inverted merge the output was not satisfactory and will not be released.
We first flatted the additional ChatML vocabulary tokens out of Orca-2-13B, then performed a stack merge with Psyfighter-2-13B. The results surprised us with their vividness, freshness of prose, obedience to instruction prompting, and formatting cohesion.
This model is focused on storywriting and text adventure, with a side order of Assistant and Chat functionality. Like its ancestor Psyfighter-2 this model will function better if you let it improvise and riff on your concepts rather than feeding it an excess of detail.
Additionally, either the removal of the ChatML vocab or the stack merging process itself has resulted in not only an uncensored model but an actively anti-censored model, so please be aware that this model can and will kill you during adventures or output NSFW material if prompted accordingly.
During testing, the model exhibited an especially strong affinity for science fiction and space opera writing, while handling fantasy elements quite well and horror elements slightly less so. Refer to the Psyfighter-2 model card for best prompting practices.
Despite that, we have tested the model out to 16000 context via Rope scaling and the model does not drive towards NSFW on its own. It will follow your tone and style very well.
Please enjoy, and if you encounter anything exciting or weird, please reach out to me at [[email protected]].
Special thanks as always to the KoboldAI crew who provided the mergebox, testing, and feedback on this model, and to gelukuMLG for the model mascot! | {"license": "other", "tags": ["storywriting", "text adventure", "not-for-all-audiences"], "license_name": "microsoft-research-license"} | text-generation | zaq-hack/psyonic-cetacean-20B-bpw300-h6-exl2-rpcal | [
"transformers",
"safetensors",
"llama",
"text-generation",
"storywriting",
"text adventure",
"not-for-all-audiences",
"license:other",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T16:12:23+00:00 | [] | [] | TAGS
#transformers #safetensors #llama #text-generation #storywriting #text adventure #not-for-all-audiences #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
!image/png
---
Presenting the FP16 files for Psyonic-Cetacean-20B! This is an experimental Llama2-based stack merge based on the models and recipe below:
- KoboldAI/PsyFighter-2-13b
- microsoft/Orca-2-13b
Note: while we did run an inverted merge the output was not satisfactory and will not be released.
We first flatted the additional ChatML vocabulary tokens out of Orca-2-13B, then performed a stack merge with Psyfighter-2-13B. The results surprised us with their vividness, freshness of prose, obedience to instruction prompting, and formatting cohesion.
This model is focused on storywriting and text adventure, with a side order of Assistant and Chat functionality. Like its ancestor Psyfighter-2 this model will function better if you let it improvise and riff on your concepts rather than feeding it an excess of detail.
Additionally, either the removal of the ChatML vocab or the stack merging process itself has resulted in not only an uncensored model but an actively anti-censored model, so please be aware that this model can and will kill you during adventures or output NSFW material if prompted accordingly.
During testing, the model exhibited an especially strong affinity for science fiction and space opera writing, while handling fantasy elements quite well and horror elements slightly less so. Refer to the Psyfighter-2 model card for best prompting practices.
Despite that, we have tested the model out to 16000 context via Rope scaling and the model does not drive towards NSFW on its own. It will follow your tone and style very well.
Please enjoy, and if you encounter anything exciting or weird, please reach out to me at [jebcarter@URL].
Special thanks as always to the KoboldAI crew who provided the mergebox, testing, and feedback on this model, and to gelukuMLG for the model mascot! | [] | [
"TAGS\n#transformers #safetensors #llama #text-generation #storywriting #text adventure #not-for-all-audiences #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
67
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #storywriting #text adventure #not-for-all-audiences #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
0.024552643299102783,
0.06800024956464767,
-0.006275040563195944,
0.025086428970098495,
0.07575658708810806,
0.03634592890739441,
0.25202906131744385,
0.08515089005231857,
-0.011156481690704823,
-0.05512445047497749,
0.17021185159683228,
0.13841918110847473,
-0.03512859344482422,
0.1285029947757721,
-0.11925232410430908,
-0.20949257910251617,
0.03797439858317375,
-0.015636201947927475,
0.13904894888401031,
0.0971633568406105,
0.11016163229942322,
-0.045477233827114105,
0.031189212575554848,
-0.029684241861104965,
-0.049808647483587265,
0.012434341013431549,
0.0614062175154686,
-0.1400814801454544,
0.12151455879211426,
0.03631461784243584,
0.09849084913730621,
0.057987798005342484,
0.01387382484972477,
-0.19102664291858673,
0.04213429614901543,
-0.026659240946173668,
-0.07191905379295349,
-0.011026190593838692,
0.049958232790231705,
-0.06349973380565643,
0.13821005821228027,
0.1146688461303711,
-0.003057145746424794,
0.0849781259894371,
-0.13439074158668518,
0.019003089517354965,
-0.005182833876460791,
0.01815907470881939,
0.0509924553334713,
0.12997621297836304,
-0.036264266818761826,
0.0909242257475853,
-0.043522339314222336,
0.07332784682512283,
0.12362038344144821,
-0.29529646039009094,
-0.04517381265759468,
0.07538816332817078,
0.14856989681720734,
-0.0027046469040215015,
-0.11599653214216232,
0.12709647417068481,
0.07879739999771118,
-0.000878071878105402,
0.012254495173692703,
-0.05533166229724884,
-0.02102483995258808,
-0.005984366871416569,
-0.048018649220466614,
-0.023642435669898987,
0.22733260691165924,
-0.04546692222356796,
0.014757025055587292,
-0.12295781075954437,
-0.04587619751691818,
0.014109168201684952,
-0.026046155020594597,
0.07250696420669556,
-0.02588612586259842,
0.07514931261539459,
-0.004969054367393255,
-0.02780364267528057,
-0.15780109167099,
-0.01355698425322771,
-0.0973406583070755,
0.06659654527902603,
-0.006334282457828522,
0.023584574460983276,
-0.18127739429473877,
0.04526813328266144,
0.07091961055994034,
-0.11227606236934662,
0.007962772622704506,
-0.09481605887413025,
0.08857917040586472,
-0.03652798384428024,
-0.013635757379233837,
-0.05076047033071518,
0.1406492292881012,
0.08988172560930252,
0.04714246094226837,
0.0011104901786893606,
-0.059307802468538284,
0.10416729003190994,
0.03554194048047066,
-0.03440386801958084,
0.11109348386526108,
-0.012477336451411247,
0.06036852300167084,
-0.04641164094209671,
0.08674035966396332,
-0.029425574466586113,
-0.12455346435308456,
0.040593862533569336,
-0.055759038776159286,
0.09845635294914246,
0.04699227958917618,
0.11172306537628174,
-0.03765106573700905,
0.0836324542760849,
-0.024948077276349068,
-0.019617557525634766,
-0.04803768917918205,
-0.005973572842776775,
0.047400686889886856,
0.045478954911231995,
-0.005817574914544821,
0.03922653570771217,
-0.023397700861096382,
0.12197400629520416,
-0.048147112131118774,
-0.058644529432058334,
-0.035749319940805435,
-0.012386869639158249,
0.10729731619358063,
-0.014203619211912155,
0.04573672637343407,
-0.14261916279792786,
-0.22342057526111603,
-0.020009182393550873,
0.021793736144900322,
-0.004237415734678507,
-0.05917438492178917,
-0.08977658301591873,
0.006093802861869335,
0.03232559561729431,
-0.05551949515938759,
-0.07079597562551498,
-0.10328826308250427,
0.09764797240495682,
-0.06331544369459152,
0.0504193939268589,
-0.08403073251247406,
0.0056554777547717094,
-0.10921580344438553,
0.013635451905429363,
0.023804308846592903,
0.011561618186533451,
-0.0031674180645495653,
0.21000196039676666,
0.08261462301015854,
0.06357459723949432,
-0.006601409986615181,
0.09485006332397461,
-0.11328576505184174,
0.2697242498397827,
-0.11523713171482086,
-0.023674432188272476,
0.23737502098083496,
-0.103493832051754,
-0.24764861166477203,
0.10483752191066742,
-0.027484051883220673,
0.17128965258598328,
0.1291683465242386,
0.21719159185886383,
0.018325531855225563,
-0.09560929238796234,
0.06064813956618309,
0.041477859020233154,
-0.09068485349416733,
-0.007705449126660824,
0.03005879372358322,
0.020218372344970703,
-0.04744383320212364,
0.0279146246612072,
0.13358613848686218,
0.0001401962072122842,
-0.015874184668064117,
-0.047281987965106964,
-0.01514775026589632,
-0.030335836112499237,
0.062154728919267654,
-0.02388199418783188,
0.03735813871026039,
-0.1371573954820633,
-0.031697358936071396,
-0.056911736726760864,
0.02699356898665428,
0.009886167012155056,
0.03861241042613983,
-0.06986287981271744,
0.03168988600373268,
0.10698442906141281,
0.061792951077222824,
-0.04639063775539398,
-0.02766101434826851,
-0.07580327242612839,
0.13158337771892548,
0.028735458850860596,
0.11764121800661087,
0.03723444417119026,
-0.020244674757122993,
-0.076547771692276,
0.004391200840473175,
0.14906659722328186,
-0.00561947375535965,
-0.03332475945353508,
-0.11564343422651291,
0.11601313948631287,
-0.06088661774992943,
0.026619553565979004,
-0.19724568724632263,
0.039403051137924194,
0.007507374510169029,
0.040395718067884445,
-0.00899811927229166,
0.04089222475886345,
0.008875788189470768,
-0.03261496126651764,
-0.09673244506120682,
-0.02243170700967312,
0.09705884754657745,
0.004469010978937149,
-0.1396852284669876,
0.17839373648166656,
-0.24855850636959076,
0.2090110033750534,
0.16184480488300323,
-0.27066341042518616,
-0.002455675508826971,
-0.08350001275539398,
0.0007730357465334237,
0.03313853219151497,
-0.04027766361832619,
0.02308301441371441,
-0.005815391428768635,
-0.009480372071266174,
0.16121219098567963,
-0.04701114445924759,
-0.07452880591154099,
-0.027887996286153793,
-0.0807064026594162,
-0.03967132046818733,
0.0888039767742157,
0.00938088446855545,
-0.1790393441915512,
0.17462293803691864,
0.27154117822647095,
0.04770695045590401,
0.26476940512657166,
0.0037540269549936056,
0.0604354627430439,
0.050089988857507706,
-0.020338168367743492,
0.015373273752629757,
-0.020795566961169243,
-0.14726382493972778,
0.013187559321522713,
0.04597891867160797,
-0.013240544125437737,
0.02476980723440647,
-0.09892520308494568,
-0.141377255320549,
-0.020208869129419327,
-0.02446736767888069,
-0.07478200644254684,
0.09729230403900146,
-0.04043428972363472,
0.09148257225751877,
-0.04148007929325104,
0.024069618433713913,
0.10977192223072052,
-0.03983813151717186,
-0.08527882397174835,
0.14562007784843445,
-0.06796591728925705,
-0.25092753767967224,
-0.20157302916049957,
-0.11487630754709244,
-0.009703908115625381,
0.052090249955654144,
0.13602185249328613,
-0.04231968894600868,
-0.038692548871040344,
-0.06209754943847656,
-0.007158034481108189,
-0.09475696831941605,
-0.034049250185489655,
-0.016210326924920082,
0.12361236661672592,
-0.06147293373942375,
-0.06062747910618782,
-0.022181125357747078,
0.00789758749306202,
-0.07334814220666885,
0.13340462744235992,
-0.10077392309904099,
0.10421361029148102,
0.08506505936384201,
0.041940346360206604,
-0.0021395133808255196,
-0.09493318945169449,
0.0940205529332161,
-0.14070361852645874,
0.001426322152838111,
0.20865778625011444,
-0.09155501425266266,
0.05357542261481285,
0.17959506809711456,
0.02837812341749668,
-0.10267776250839233,
0.09350911527872086,
-0.023243924602866173,
-0.08506158739328384,
-0.24538429081439972,
-0.040115781128406525,
-0.07560070604085922,
0.16982731223106384,
0.01556074433028698,
0.08657411485910416,
0.1438826024532318,
0.07791079580783844,
-0.06400014460086823,
-0.06258761882781982,
0.09431915730237961,
0.10531551390886307,
0.25717130303382874,
-0.011607469990849495,
0.08263862133026123,
-0.10608244687318802,
-0.08253690600395203,
0.0807003527879715,
0.002981617348268628,
0.06459205597639084,
0.12218087166547775,
0.09655627608299255,
0.0761043056845665,
0.004725807346403599,
0.09874024242162704,
0.11056267470121384,
0.09913510829210281,
-0.04518439620733261,
-0.03766189143061638,
-0.037798311561346054,
-0.03452032804489136,
0.06429759413003922,
-0.09028049558401108,
-0.11083921045064926,
-0.042543478310108185,
-0.04930552467703819,
0.06531030684709549,
0.10639654099941254,
0.050778668373823166,
-0.05764404311776161,
0.05022019147872925,
0.14868196845054626,
-0.04829220101237297,
-0.12058867514133453,
0.1110382154583931,
0.03420580178499222,
-0.11443531513214111,
0.14529183506965637,
-0.011771789751946926,
0.1382288634777069,
-0.05290690436959267,
0.0881744995713234,
-0.11976394057273865,
-0.15380679070949554,
-0.03562243655323982,
0.12125004827976227,
-0.3409517705440521,
0.2139718234539032,
0.022577650845050812,
0.04588121548295021,
-0.08311774581670761,
-0.06715845316648483,
0.024918219074606895,
0.16455893218517303,
0.1975151002407074,
-0.0301362257450819,
-0.05134611576795578,
-0.0412110760807991,
0.00003511626709951088,
0.019543228670954704,
0.08700890094041824,
-0.029512539505958557,
0.022741146385669708,
-0.08478443324565887,
0.018364109098911285,
-0.022697892040014267,
-0.008467350155115128,
-0.05338098108768463,
-0.15789566934108734,
0.02189684472978115,
0.1127079576253891,
0.13133300840854645,
-0.029481645673513412,
-0.0032808587420731783,
-0.08723343908786774,
0.07290764153003693,
-0.07489017397165298,
-0.027128152549266815,
-0.08750085532665253,
-0.10792524367570877,
-0.058988261967897415,
0.0026672598905861378,
0.010724034160375595,
-0.040624164044857025,
0.0680738240480423,
-0.14056387543678284,
-0.12983058393001556,
0.09150215238332748,
-0.07923561334609985,
-0.11579445004463196,
-0.05087887495756149,
0.12757492065429688,
-0.06683522462844849,
0.03861720487475395,
0.048804886639118195,
0.010099723003804684,
-0.08903469890356064,
-0.1087338775396347,
0.004107189830392599,
-0.0766809955239296,
-0.007311676628887653,
-0.05894060432910919,
-0.07814282178878784,
-0.06319831311702728,
-0.025133220478892326,
-0.1661362498998642,
0.22321660816669464,
0.34867510199546814,
-0.010433822870254517,
0.18686027824878693,
0.14680537581443787,
-0.12579557299613953,
-0.3470672369003296,
-0.13655027747154236,
-0.21006782352924347,
-0.07806874066591263,
0.04293923079967499,
-0.055371955037117004,
0.023506056517362595,
-0.01795108988881111,
-0.02330568991601467,
0.1137828528881073,
-0.17136253416538239,
-0.11587657034397125,
0.12197116017341614,
-0.010856055654585361,
0.32602739334106445,
-0.1982441544532776,
-0.0487438440322876,
-0.13322415947914124,
-0.07285565137863159,
0.055188070982694626,
-0.1204465925693512,
0.12010639160871506,
-0.007437644060701132,
0.09034236520528793,
0.014261064119637012,
-0.024851521477103233,
0.06800531595945358,
-0.13509199023246765,
0.05364105477929115,
-0.14163747429847717,
-0.0014003977412357926,
0.10801897943019867,
-0.03557075187563896,
-0.003281416604295373,
-0.22687503695487976,
-0.05253811553120613,
-0.060481809079647064,
-0.07031818479299545,
-0.03123307228088379,
0.08795113116502762,
-0.03486017882823944,
-0.05428633093833923,
-0.03695521131157875,
-0.04518020898103714,
-0.013758636079728603,
-0.013743762858211994,
0.26902616024017334,
-0.1369444578886032,
0.19221246242523193,
0.023955337703227997,
0.22811441123485565,
-0.07853924483060837,
0.08864311873912811,
-0.05528164282441139,
-0.08313252776861191,
0.00615558261051774,
-0.13969995081424713,
0.023851079866290092,
0.07635126262903214,
-0.05121266841888428,
0.062188442796468735,
0.0992860421538353,
0.01268355268985033,
0.0430474691092968,
0.15104900300502777,
-0.18661701679229736,
-0.08694476634263992,
-0.05470138043165207,
0.12216684967279434,
0.021877799183130264,
0.023732110857963562,
0.13673578202724457,
0.0038171184714883566,
0.028363779187202454,
0.0015041254227980971,
0.02920924685895443,
0.014596535824239254,
-0.003732886165380478,
-0.055541325360536575,
0.016229065135121346,
-0.09915612637996674,
0.0809159055352211,
0.0294171292334795,
-0.11212829500436783,
0.025413040071725845,
0.15533261001110077,
-0.0650755912065506,
-0.11768341064453125,
0.0057175676338374615,
0.148555189371109,
-0.10374071449041367,
-0.04402026906609535,
-0.062073465436697006,
-0.17184202373027802,
0.038068026304244995,
0.2248273491859436,
0.027189888060092926,
0.0705033466219902,
0.006424351595342159,
-0.008606072515249252,
-0.008470558561384678,
0.05588870123028755,
-0.016039615496993065,
-0.0008846343262121081,
-0.1332293599843979,
0.03644512966275215,
-0.022411677986383438,
0.053814224898815155,
-0.09663651138544083,
-0.03823291137814522,
-0.12467819452285767,
0.06714051961898804,
-0.1192275732755661,
0.00023566542949993163,
-0.12214960157871246,
-0.006123457103967667,
-0.0014943060232326388,
0.021250415593385696,
-0.062156952917575836,
-0.07691515982151031,
-0.08534861356019974,
-0.006057679653167725,
-0.002577902050688863,
0.04671122878789902,
-0.08457479625940323,
-0.020621569827198982,
0.06936324387788773,
-0.0422893688082695,
0.09338672459125519,
0.02773372270166874,
-0.0844649076461792,
0.05917053669691086,
-0.272941917181015,
0.0053605628199875355,
0.0876004695892334,
-0.012540568597614765,
-0.04511613771319389,
0.18080012500286102,
-0.04972167685627937,
0.0834890827536583,
0.013573896139860153,
0.08494513481855392,
-0.0717923641204834,
-0.10856453329324722,
0.00018396993982605636,
0.06715316325426102,
-0.1337595134973526,
0.01886873133480549,
-0.06925108283758163,
0.1079791858792305,
-0.10989644378423691,
0.1877453625202179,
-0.10827895998954773,
0.07824420928955078,
-0.02673555351793766,
0.05858476459980011,
0.04249152913689613,
-0.1291210651397705,
-0.08001314848661423,
-0.11095483601093292,
-0.021375169977545738,
-0.004540305119007826,
0.21591472625732422,
0.012018105946481228,
-0.05183045566082001,
0.09976179152727127,
0.0533432699739933,
0.01753421686589718,
0.010756858624517918,
0.25596606731414795,
0.05716362223029137,
-0.0705529972910881,
-0.14858059585094452,
0.0044577959924936295,
0.05593341588973999,
-0.12991869449615479,
0.08481264859437943,
0.03308482840657234,
-0.06647837162017822,
0.04766938090324402,
0.00456581823527813,
0.0951186940073967,
-0.022376136854290962,
-0.1433664709329605,
0.007966206409037113,
0.043242935091257095,
0.00026994472136721015,
0.025908365845680237,
0.17714416980743408,
0.03602594509720802,
-0.02349943108856678,
-0.11821579933166504,
-0.039522428065538406,
-0.1723768413066864,
-0.10499744117259979,
-0.07720645517110825,
-0.07751176506280899,
0.04657287895679474,
-0.06475445628166199,
0.08908320218324661,
0.02177288942039013,
0.07714133709669113,
-0.10031405091285706,
0.09067082405090332,
0.028232000768184662,
-0.04782482236623764,
0.0918969139456749,
-0.03279038146138191,
0.036186881363391876,
-0.07358130812644958,
0.02436944469809532,
-0.06267331540584564,
-0.04309646040201187,
-0.03310341015458107,
0.07883191108703613,
-0.02015214040875435,
0.01041588094085455,
-0.15917716920375824,
-0.07245973497629166,
-0.007560403551906347,
0.0715281292796135,
0.017882371321320534,
0.14909875392913818,
-0.0021014593075960875,
-0.030547933652997017,
0.07310996949672699,
0.2093002051115036,
-0.028740087524056435,
-0.12363199144601822,
-0.00708317244425416,
0.0667002946138382,
-0.040854401886463165,
0.08084486424922943,
-0.09793492406606674,
0.0012357545783743262,
0.016111919656395912,
0.33123570680618286,
0.25315988063812256,
-0.06639370322227478,
0.012642205692827702,
-0.09266163408756256,
0.042206183075904846,
0.03813379630446434,
0.07809721678495407,
0.049722738564014435,
0.18034042418003082,
-0.08328406512737274,
0.04224809631705284,
0.014901530928909779,
0.013792493380606174,
-0.1341090351343155,
0.11629787087440491,
0.015386324375867844,
0.03003045916557312,
-0.052709948271512985,
0.13859224319458008,
-0.20088759064674377,
0.1125604584813118,
-0.10647700726985931,
-0.06685298681259155,
0.01139877736568451,
-0.024373527616262436,
0.17648883163928986,
-0.0187532976269722,
0.04782719537615776,
0.02296639233827591,
-0.024096770212054253,
-0.06161698326468468,
0.004831473808735609,
-0.17059317231178284,
0.015253123827278614,
0.07425881922245026,
-0.09189016371965408,
0.1422913670539856,
-0.0476461723446846,
-0.054282136261463165,
0.06194471940398216,
-0.02206886000931263,
-0.03487440198659897,
0.13465729355812073,
0.02734525501728058,
-0.07007022947072983,
0.050422780215740204,
0.03805816173553467,
-0.017948413267731667,
-0.03648299723863602,
0.09214751422405243,
-0.0532359816133976,
0.06684883683919907,
0.027004847303032875,
-0.06449272483587265,
-0.017657961696386337,
0.11114069074392319,
-0.06183087080717087,
0.072293721139431,
0.02306167222559452,
-0.03551258519291878,
0.0022191223688423634,
-0.02228567749261856,
0.009448587894439697,
-0.02391415275633335,
-0.14522698521614075,
-0.013912297785282135,
-0.0681820958852768,
-0.07442320883274078,
0.12633441388607025,
0.00903337448835373,
-0.3678169548511505,
0.01395304687321186,
-0.11143926531076431,
0.03296716511249542,
-0.16528530418872833,
0.015266507863998413,
0.1719563752412796,
0.00485230190679431,
0.0060460832901299,
-0.15911318361759186,
0.059644103050231934,
0.06673851609230042,
-0.025990379974246025,
-0.0944560095667839
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | manche/gpt2-safeguard-zs | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T16:13:18+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #gpt2 #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #gpt2 #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
57,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #gpt2 #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.05622259899973869,
0.16002345085144043,
-0.004987028427422047,
0.023115945979952812,
0.0962471067905426,
0.011845538392663002,
0.06785304099321365,
0.11496778577566147,
-0.020396295934915543,
0.11142492294311523,
0.03292480856180191,
0.0972127765417099,
0.11474913358688354,
0.16215258836746216,
0.004439093638211489,
-0.23455148935317993,
0.04782992601394653,
-0.12695099413394928,
-0.033447545021772385,
0.11785799264907837,
0.14491069316864014,
-0.10402194410562515,
0.07766910642385483,
-0.030544815585017204,
-0.009361269883811474,
-0.03290390968322754,
-0.06365230679512024,
-0.05152205005288124,
0.05037128925323486,
0.06932847946882248,
0.06591591984033585,
0.007509593386203051,
0.09122733771800995,
-0.2655104100704193,
0.02280162274837494,
0.07630051672458649,
-0.0015554219717159867,
0.07497020810842514,
0.048351652920246124,
-0.08209776133298874,
0.0788840726017952,
-0.05696587264537811,
0.14718368649482727,
0.08216129243373871,
-0.08924587815999985,
-0.1965435892343521,
-0.08464295417070389,
0.10284840315580368,
0.18357418477535248,
0.05158785358071327,
-0.024141347035765648,
0.10476154088973999,
-0.08419200032949448,
0.008797040209174156,
0.06024181470274925,
-0.06443428993225098,
-0.05412506312131882,
0.06934051215648651,
0.07975570857524872,
0.07967228442430496,
-0.13025140762329102,
-0.014651902951300144,
0.011243549175560474,
0.007594773545861244,
0.08504551649093628,
0.022028017789125443,
0.14595499634742737,
0.04393624886870384,
-0.13030564785003662,
-0.044304780662059784,
0.09771761298179626,
0.04345165938138962,
-0.053857799619436264,
-0.2537047266960144,
-0.024983759969472885,
-0.03927002474665642,
-0.03094942681491375,
-0.038562554866075516,
0.04431856796145439,
-0.011080716736614704,
0.08032315224409103,
-0.01118796318769455,
-0.08149448037147522,
-0.041395120322704315,
0.06544242054224014,
0.062143467366695404,
0.026896316558122635,
-0.01158317644149065,
0.00973866879940033,
0.1224486380815506,
0.10907839238643646,
-0.12763150036334991,
-0.05768941715359688,
-0.06755511462688446,
-0.08307720720767975,
-0.04300352931022644,
0.03337155282497406,
0.044020529836416245,
0.04436098039150238,
0.2466370165348053,
0.01108562108129263,
0.05453123152256012,
0.045806169509887695,
0.010608446784317493,
0.06787561625242233,
0.11606968939304352,
-0.062306761741638184,
-0.09178462624549866,
-0.029058339074254036,
0.09215214103460312,
0.006741520017385483,
-0.042814407497644424,
-0.060904473066329956,
0.06479041278362274,
0.012608112767338753,
0.12110785394906998,
0.08444269746541977,
0.0026690615341067314,
-0.07305197417736053,
-0.06963318586349487,
0.18848419189453125,
-0.1598394364118576,
0.047875016927719116,
0.031182926148176193,
-0.038971830159425735,
-0.0014042917173355818,
0.008752269670367241,
0.02394084818661213,
-0.020246321335434914,
0.08923295140266418,
-0.05574449151754379,
-0.03784004598855972,
-0.11079790443181992,
-0.03252100944519043,
0.030985163524746895,
0.0051483530551195145,
-0.027043871581554413,
-0.033837489783763885,
-0.09040277451276779,
-0.059588029980659485,
0.0922931432723999,
-0.07471107691526413,
-0.04984431713819504,
-0.013726521283388138,
-0.07691634446382523,
0.023329194635152817,
0.016799474135041237,
0.08357251435518265,
-0.02157396264374256,
0.0384126678109169,
-0.0560205839574337,
0.0631464347243309,
0.11269522458314896,
0.029363946989178658,
-0.053069718182086945,
0.05750001594424248,
-0.24315528571605682,
0.10326608270406723,
-0.07320205867290497,
0.050549428910017014,
-0.15059062838554382,
-0.026000602170825005,
0.044471126049757004,
0.00805877335369587,
-0.013138634152710438,
0.14088952541351318,
-0.21621745824813843,
-0.0323486253619194,
0.16741067171096802,
-0.0939871072769165,
-0.07602590322494507,
0.059108685702085495,
-0.05233629792928696,
0.10869261622428894,
0.04351044446229935,
-0.02232111617922783,
0.060673557221889496,
-0.14475463330745697,
-0.01067100279033184,
-0.04139741137623787,
-0.02402937039732933,
0.16397778689861298,
0.07567544281482697,
-0.06286642700433731,
0.08052356541156769,
0.024165838956832886,
-0.017831770703196526,
-0.04484899342060089,
-0.023361295461654663,
-0.10819391161203384,
0.009856974706053734,
-0.06032416597008705,
0.02424289658665657,
-0.025761527940630913,
-0.09367526322603226,
-0.02868773601949215,
-0.1802000105381012,
-0.009223134256899357,
0.0881323292851448,
-0.011722641065716743,
-0.021903391927480698,
-0.12039245665073395,
0.011948852799832821,
0.031212422996759415,
0.002984174294397235,
-0.13029038906097412,
-0.05838731303811073,
0.027675874531269073,
-0.16422230005264282,
0.03272955119609833,
-0.05597274377942085,
0.05056252330541611,
0.03445037454366684,
-0.03187771514058113,
-0.033117350190877914,
0.009550533257424831,
0.006354342680424452,
-0.010578392073512077,
-0.2502359449863434,
-0.02440580166876316,
-0.0219739843159914,
0.17386503517627716,
-0.21793730556964874,
0.04213962331414223,
0.07686693966388702,
0.14929872751235962,
0.006240781396627426,
-0.038500864058732986,
0.010139784775674343,
-0.08222103863954544,
-0.030560437589883804,
-0.0643099993467331,
-0.012082485482096672,
-0.03717579320073128,
-0.05608142167329788,
0.05165567249059677,
-0.16133594512939453,
-0.028727244585752487,
0.1057019829750061,
0.06860516220331192,
-0.14001330733299255,
-0.019125886261463165,
-0.04171464592218399,
-0.043496038764715195,
-0.05877087265253067,
-0.0552728995680809,
0.1185101792216301,
0.05596614256501198,
0.04696191847324371,
-0.06956122815608978,
-0.07775315642356873,
0.007865429855883121,
-0.017090093344449997,
-0.017978519201278687,
0.08920905739068985,
0.07311701774597168,
-0.12023317068815231,
0.09247473627328873,
0.10194233059883118,
0.09365488588809967,
0.108615942299366,
-0.017981963232159615,
-0.08929306268692017,
-0.04584396257996559,
0.02045595459640026,
0.013332244008779526,
0.14797501266002655,
-0.01403066236525774,
0.056954506784677505,
0.03922648727893829,
-0.01123172789812088,
0.012020308524370193,
-0.09384570270776749,
0.027314940467476845,
0.034342724829912186,
-0.020308034494519234,
0.03796098753809929,
-0.04001156985759735,
0.019826533272862434,
0.08712323755025864,
0.04676510766148567,
0.04415108636021614,
0.011758276261389256,
-0.04233846068382263,
-0.10904491692781448,
0.173858180642128,
-0.12615609169006348,
-0.24583272635936737,
-0.14115718007087708,
0.0015609683468937874,
0.04152948409318924,
-0.009671499952673912,
0.003867273684591055,
-0.07054664939641953,
-0.11710625886917114,
-0.0934595838189125,
0.018713686615228653,
0.04491026699542999,
-0.07426843047142029,
-0.0596279613673687,
0.059872306883335114,
0.03894329443573952,
-0.14430272579193115,
0.022237464785575867,
0.047419775277376175,
-0.09032250195741653,
-0.006925572175532579,
0.08398029953241348,
0.06729988008737564,
0.17764869332313538,
0.009659109637141228,
-0.021044570952653885,
0.03080335259437561,
0.21258224546909332,
-0.14283664524555206,
0.11252175271511078,
0.14021345973014832,
-0.09024007618427277,
0.08099348843097687,
0.1948828399181366,
0.039186809211969376,
-0.10478170961141586,
0.03259138762950897,
0.02489176020026207,
-0.028939135372638702,
-0.25018003582954407,
-0.0680207833647728,
0.002590036718174815,
-0.04892077296972275,
0.07092583924531937,
0.0918794497847557,
0.09946957975625992,
0.015428726561367512,
-0.09732488542795181,
-0.08017807453870773,
0.0468163788318634,
0.10640767961740494,
0.0070237633772194386,
-0.01532268337905407,
0.08905128389596939,
-0.03260866180062294,
0.018378758803009987,
0.0954233929514885,
0.00412675691768527,
0.17459604144096375,
0.05586163327097893,
0.17767499387264252,
0.07751350849866867,
0.06634163856506348,
0.019167855381965637,
0.0069374511949718,
0.02067388966679573,
0.017508454620838165,
-0.004214957356452942,
-0.08522020280361176,
-0.00457410141825676,
0.12029227614402771,
0.06321834027767181,
0.024303704500198364,
0.0137604009360075,
-0.03941800817847252,
0.08438141644001007,
0.17332784831523895,
0.0020201504230499268,
-0.18486954271793365,
-0.07240456342697144,
0.07921045273542404,
-0.0910051167011261,
-0.10552998632192612,
-0.03353073075413704,
0.03346012532711029,
-0.1747758537530899,
0.02097497321665287,
-0.017018353566527367,
0.10809773951768875,
-0.13855572044849396,
-0.018670624122023582,
0.06328251957893372,
0.07232730835676193,
-0.0028869258239865303,
0.06308864802122116,
-0.153975248336792,
0.1050168052315712,
0.016289174556732178,
0.06754438579082489,
-0.09747608006000519,
0.10138221830129623,
-0.006303760688751936,
-0.007241528946906328,
0.13875643908977509,
0.010596190579235554,
-0.05694379657506943,
-0.08987913280725479,
-0.10555228590965271,
-0.008462639525532722,
0.12933635711669922,
-0.15157614648342133,
0.0847775787115097,
-0.028662750497460365,
-0.043171048164367676,
0.0024383023846894503,
-0.1199452206492424,
-0.1302652359008789,
-0.1875755488872528,
0.058235347270965576,
-0.1366453617811203,
0.039557021111249924,
-0.10582595318555832,
-0.04340389743447304,
-0.028466427698731422,
0.2041483372449875,
-0.2317875325679779,
-0.0682469978928566,
-0.1541893482208252,
-0.08429346233606339,
0.14446710050106049,
-0.04730919376015663,
0.08914490789175034,
-0.0013825427740812302,
0.19013537466526031,
0.024473950266838074,
-0.02387205697596073,
0.10308998823165894,
-0.09543927758932114,
-0.19450686872005463,
-0.08603953570127487,
0.15582145750522614,
0.13931062817573547,
0.03702725097537041,
-0.004593946039676666,
0.029260434210300446,
-0.020000332966446877,
-0.12535293400287628,
0.025526588782668114,
0.1793687790632248,
0.07859015464782715,
0.023437971249222755,
-0.025896867737174034,
-0.10993997752666473,
-0.06524094194173813,
-0.0335373692214489,
0.02718053013086319,
0.18264614045619965,
-0.07421271502971649,
0.1900695115327835,
0.13626199960708618,
-0.05445687845349312,
-0.1955246478319168,
0.018216576427221298,
0.040417760610580444,
0.010847307741641998,
0.03138056397438049,
-0.2078717201948166,
0.09027513861656189,
0.0014845491386950016,
-0.05172133818268776,
0.141556978225708,
-0.174949511885643,
-0.1512570083141327,
0.06491631269454956,
0.0364508256316185,
-0.19348180294036865,
-0.117862768471241,
-0.08817066252231598,
-0.046907443553209305,
-0.17498233914375305,
0.10519181191921234,
0.016932250931859016,
0.009516867808997631,
0.03492651879787445,
0.02640140987932682,
0.011080757714807987,
-0.03873949125409126,
0.19461296498775482,
-0.02505207620561123,
0.029532426968216896,
-0.08079101145267487,
-0.06136554479598999,
0.0607450045645237,
-0.05577658861875534,
0.07896649837493896,
-0.020188091322779655,
0.012835816480219364,
-0.1100873053073883,
-0.0468425452709198,
-0.027396185323596,
0.017321845516562462,
-0.09195652604103088,
-0.09473495930433273,
-0.05146971344947815,
0.09373841434717178,
0.08845265954732895,
-0.036603908985853195,
-0.04043547809123993,
-0.07348548620939255,
0.0325477197766304,
0.17183002829551697,
0.17659065127372742,
0.038550034165382385,
-0.08084331452846527,
-0.005880105309188366,
-0.01188716571778059,
0.04436201974749565,
-0.22519725561141968,
0.06208868324756622,
0.04557957127690315,
0.015879612416028976,
0.11362850666046143,
-0.018783990293741226,
-0.16298477351665497,
-0.06594224274158478,
0.06143777072429657,
-0.06664001196622849,
-0.18599680066108704,
0.0032026967965066433,
0.058006007224321365,
-0.1646854728460312,
-0.037671029567718506,
0.042260222136974335,
-0.0045668939128518105,
-0.04300284758210182,
0.01627597212791443,
0.08071378618478775,
0.005054219625890255,
0.07112491130828857,
0.05733523517847061,
0.0842885971069336,
-0.10417009145021439,
0.07519911974668503,
0.08007751405239105,
-0.08229218423366547,
0.031453702598810196,
0.08910130709409714,
-0.061817802488803864,
-0.03069761022925377,
0.032593827694654465,
0.07753410935401917,
0.019773589447140694,
-0.041717879474163055,
0.008655321784317493,
-0.09745000302791595,
0.06339588761329651,
0.09504765272140503,
0.03549657016992569,
0.014742289669811726,
0.034356739372015,
0.04988397657871246,
-0.07460241764783859,
0.11766603589057922,
0.022336218506097794,
0.01780087500810623,
-0.044981084764003754,
-0.05459042266011238,
0.032110098749399185,
-0.022974027320742607,
-0.010163158178329468,
-0.03885438293218613,
-0.07015778869390488,
-0.018130742013454437,
-0.15929651260375977,
-0.014899281784892082,
-0.04085385054349899,
0.007158880587667227,
0.02551902085542679,
-0.03834335505962372,
0.007963370531797409,
0.012195355258882046,
-0.07085035741329193,
-0.061454467475414276,
-0.022903166711330414,
0.09224231541156769,
-0.16436699032783508,
0.025155464187264442,
0.08285263180732727,
-0.12099926173686981,
0.09775067120790482,
0.021939631551504135,
0.0031351554207503796,
0.028338242322206497,
-0.1542527824640274,
0.04096807911992073,
-0.024365095421671867,
0.01272035762667656,
0.04409142583608627,
-0.22033950686454773,
0.001463581225834787,
-0.03818526118993759,
-0.05954346805810928,
-0.010227864608168602,
-0.033079732209444046,
-0.11291328817605972,
0.09883669763803482,
0.008058897219598293,
-0.08219768106937408,
-0.030809206888079643,
0.03451729565858841,
0.08243680745363235,
-0.02608415111899376,
0.15152283012866974,
0.0016822130419313908,
0.07172226905822754,
-0.17519205808639526,
-0.021702464669942856,
-0.011611736379563808,
0.02207101881504059,
-0.014536668546497822,
-0.015496513806283474,
0.042471300810575485,
-0.02421419881284237,
0.19108575582504272,
-0.026401294395327568,
0.038726791739463806,
0.06405707448720932,
0.01593620702624321,
-0.014801506884396076,
0.10957890748977661,
0.05975057929754257,
0.02399693801999092,
0.022115202620625496,
0.007329683285206556,
-0.039842452853918076,
-0.014149460941553116,
-0.19538825750350952,
0.06474217027425766,
0.1377464383840561,
0.08781574666500092,
-0.01322576031088829,
0.07683692127466202,
-0.10024392604827881,
-0.12397097796201706,
0.11215250939130783,
-0.06283260136842728,
-0.007701667957007885,
-0.06531554460525513,
0.13346771895885468,
0.14944057166576385,
-0.18992236256599426,
0.06835456937551498,
-0.06228158622980118,
-0.05332518368959427,
-0.11744599789381027,
-0.1957325041294098,
-0.055616896599531174,
-0.056456826627254486,
-0.014700124971568584,
-0.048795297741889954,
0.07307228446006775,
0.05693497136235237,
0.012962869368493557,
0.003600025549530983,
0.0766802653670311,
-0.015357231721282005,
0.0008028073934838176,
0.03077360987663269,
0.06600049883127213,
0.013312965631484985,
-0.02929985709488392,
0.020537450909614563,
-0.007275243755429983,
0.04005419462919235,
0.06378308683633804,
0.038119763135910034,
-0.02801438421010971,
0.01591232419013977,
-0.03770609200000763,
-0.10940317064523697,
0.0409080907702446,
-0.028551526367664337,
-0.08112191408872604,
0.13721226155757904,
0.02428387477993965,
0.005870606284588575,
-0.02180131897330284,
0.24582624435424805,
-0.07231455296278,
-0.09001907706260681,
-0.1473579704761505,
0.10211005061864853,
-0.04095151647925377,
0.06560079753398895,
0.04110138490796089,
-0.10732010751962662,
0.013498948886990547,
0.12688814103603363,
0.15896959602832794,
-0.044884394854307175,
0.020156091079115868,
0.03252736106514931,
0.003683826420456171,
-0.04006262496113777,
0.05253688618540764,
0.0694650411605835,
0.14883354306221008,
-0.04907030612230301,
0.08928520232439041,
0.005485867150127888,
-0.10256236046552658,
-0.03822692111134529,
0.11808354407548904,
-0.017866896465420723,
0.018703164532780647,
-0.057248231023550034,
0.11889533698558807,
-0.059861693531274796,
-0.23005777597427368,
0.06317704170942307,
-0.0720362737774849,
-0.14286935329437256,
-0.021647587418556213,
0.07456772774457932,
-0.017636949196457863,
0.02658887766301632,
0.07326807081699371,
-0.07681973278522491,
0.19899281859397888,
0.038975972682237625,
-0.05729197710752487,
-0.05658522993326187,
0.0789351835846901,
-0.114089734852314,
0.2792985737323761,
0.01164181251078844,
0.04984506592154503,
0.10365619510412216,
-0.016686614602804184,
-0.13768579065799713,
0.015234606340527534,
0.09244892746210098,
-0.09004336595535278,
0.03869183734059334,
0.2132277488708496,
-0.002569539239630103,
0.1152428612112999,
0.07714667171239853,
-0.07265080511569977,
0.04592108353972435,
-0.1130065843462944,
-0.0718315914273262,
-0.086885966360569,
0.09441597014665604,
-0.07240451127290726,
0.14123490452766418,
0.12318195402622223,
-0.053516924381256104,
0.010368985123932362,
-0.031209774315357208,
0.04651070013642311,
0.007842876948416233,
0.10365527868270874,
0.010769560933113098,
-0.18099099397659302,
0.022656621411442757,
0.018202748149633408,
0.10856854915618896,
-0.17241089046001434,
-0.09672945737838745,
0.04725200682878494,
0.001958663808181882,
-0.059874359518289566,
0.1282012164592743,
0.057909298688173294,
0.04923510178923607,
-0.043742597103118896,
-0.017267800867557526,
-0.009560109116137028,
0.13584671914577484,
-0.10737434774637222,
-0.0021453071385622025
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-finetuned-emotion
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2205
- Accuracy: 0.925
- F1: 0.9252
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
| 0.8142 | 1.0 | 250 | 0.3149 | 0.9095 | 0.9079 |
| 0.2443 | 2.0 | 500 | 0.2205 | 0.925 | 0.9252 |
### Framework versions
- Transformers 4.13.0
- Pytorch 1.12.1
- Datasets 2.8.0
- Tokenizers 0.10.3
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["emotion"], "metrics": ["accuracy", "f1"], "model-index": [{"name": "distilbert-base-uncased-finetuned-emotion", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "emotion", "type": "emotion", "args": "split"}, "metrics": [{"type": "accuracy", "value": 0.925, "name": "Accuracy"}, {"type": "f1", "value": 0.925235539384546, "name": "F1"}]}]}]} | text-classification | Schnatz65/distilbert-base-uncased-finetuned-emotion | [
"transformers",
"pytorch",
"tensorboard",
"distilbert",
"text-classification",
"generated_from_trainer",
"dataset:emotion",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T16:18:49+00:00 | [] | [] | TAGS
#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-emotion #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| distilbert-base-uncased-finetuned-emotion
=========================================
This model is a fine-tuned version of distilbert-base-uncased on the emotion dataset.
It achieves the following results on the evaluation set:
* Loss: 0.2205
* Accuracy: 0.925
* F1: 0.9252
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 64
* eval\_batch\_size: 64
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 2
### Training results
### Framework versions
* Transformers 4.13.0
* Pytorch 1.12.1
* Datasets 2.8.0
* Tokenizers 0.10.3
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.13.0\n* Pytorch 1.12.1\n* Datasets 2.8.0\n* Tokenizers 0.10.3"
] | [
"TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-emotion #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.13.0\n* Pytorch 1.12.1\n* Datasets 2.8.0\n* Tokenizers 0.10.3"
] | [
67,
98,
4,
31
] | [
"passage: TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-emotion #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2### Training results### Framework versions\n\n\n* Transformers 4.13.0\n* Pytorch 1.12.1\n* Datasets 2.8.0\n* Tokenizers 0.10.3"
] | [
-0.100026436150074,
0.11341743916273117,
-0.0028740314301103354,
0.131758913397789,
0.16103368997573853,
0.043246444314718246,
0.11114879697561264,
0.12988325953483582,
-0.08673057705163956,
0.025439342483878136,
0.10774272680282593,
0.16366703808307648,
0.02127249725162983,
0.09792694449424744,
-0.05628207325935364,
-0.2832159399986267,
-0.0137634864076972,
0.049597159028053284,
-0.02019539289176464,
0.13457055389881134,
0.0966721773147583,
-0.12457016855478287,
0.09381760656833649,
0.0050198109820485115,
-0.17160683870315552,
0.008332164958119392,
0.0007676285458728671,
-0.044595785439014435,
0.14576171338558197,
0.016878683120012283,
0.10081198811531067,
0.007842480204999447,
0.08257234841585159,
-0.22903293371200562,
0.018515946343541145,
0.036652207374572754,
0.0014993117656558752,
0.08650070428848267,
0.03964383900165558,
-0.015507402829825878,
0.16690926253795624,
-0.06421086937189102,
0.05300555378198624,
0.02129579521715641,
-0.11147003620862961,
-0.24078840017318726,
-0.07728250324726105,
0.04542425274848938,
0.06929304450750351,
0.1180371567606926,
-0.02113163098692894,
0.13467101752758026,
-0.09524886310100555,
0.09733443707227707,
0.2401663213968277,
-0.24164080619812012,
-0.06713058799505234,
0.018191754817962646,
0.015891341492533684,
0.04541725292801857,
-0.11857499182224274,
-0.03709149733185768,
0.04601196199655533,
0.05302303656935692,
0.12156891822814941,
-0.036937516182661057,
-0.08929795771837234,
0.012711801566183567,
-0.1301155835390091,
-0.047058526426553726,
0.17659679055213928,
0.061675798147916794,
-0.026434261351823807,
-0.05930037423968315,
-0.05689476802945137,
-0.15068809688091278,
-0.030614491552114487,
-0.019334450364112854,
0.05535327270627022,
-0.014435459859669209,
-0.0675533339381218,
0.0027175152208656073,
-0.12268251180648804,
-0.04259030520915985,
-0.056146811693906784,
0.11194854229688644,
0.02373482659459114,
0.009815054014325142,
-0.019843967631459236,
0.10252010077238083,
0.0002401281235506758,
-0.12112302333116531,
0.01832869090139866,
0.016359521076083183,
0.03215824440121651,
-0.02485841140151024,
-0.07290627807378769,
-0.0558636337518692,
-0.0019483197247609496,
0.10257221758365631,
-0.07286717742681503,
0.048928845673799515,
0.04536154866218567,
0.03794239088892937,
-0.06703442335128784,
0.19235554337501526,
-0.028765590861439705,
-0.02911909483373165,
-0.015193224884569645,
0.0597110390663147,
0.024294359609484673,
-0.005251770373433828,
-0.12107451260089874,
0.026257270947098732,
0.0904218852519989,
0.00028846412897109985,
-0.09334573894739151,
0.08042477071285248,
-0.07781172543764114,
-0.025232046842575073,
-0.02331613376736641,
-0.07680881023406982,
0.028548309579491615,
0.021395336836576462,
-0.07604943960905075,
0.0016866173828020692,
0.03168899565935135,
0.010868197306990623,
-0.01735955849289894,
0.08545901626348495,
-0.07252000272274017,
0.02812882512807846,
-0.09319796413183212,
-0.101068876683712,
0.027278872206807137,
-0.0842718854546547,
0.03825293108820915,
-0.0906146764755249,
-0.19988571107387543,
-0.017813045531511307,
0.07487019896507263,
-0.02157401107251644,
-0.0491173192858696,
-0.06879479438066483,
-0.061445336788892746,
0.015238006599247456,
-0.005260469391942024,
0.09810610860586166,
-0.06536948680877686,
0.08833594620227814,
0.031214280053973198,
0.0834539532661438,
-0.03015952929854393,
0.05376565456390381,
-0.11459598690271378,
0.0053842817433178425,
-0.13874001801013947,
0.052186403423547745,
-0.0464051179587841,
0.07081898301839828,
-0.06729019433259964,
-0.11467760056257248,
0.017973121255636215,
-0.006344716530293226,
0.06170157715678215,
0.11172439903020859,
-0.1894358992576599,
-0.09593353420495987,
0.1648629754781723,
-0.07081681489944458,
-0.1083587110042572,
0.1308870017528534,
-0.06907671689987183,
0.07201527804136276,
0.07130590826272964,
0.18856000900268555,
0.05285041034221649,
-0.07438112795352936,
-0.005385520402342081,
0.01793282851576805,
0.052360545843839645,
-0.027138376608490944,
0.05887956917285919,
0.024229886010289192,
0.038481682538986206,
0.03922762721776962,
-0.016652459278702736,
0.07153616100549698,
-0.09220230579376221,
-0.10333997011184692,
-0.03472093865275383,
-0.09258085489273071,
0.05126404017210007,
0.09311733394861221,
0.06332805007696152,
-0.11061637848615646,
-0.07480276376008987,
0.02934395894408226,
0.09542390704154968,
-0.06665933132171631,
0.03236960992217064,
-0.062239013612270355,
0.05876706540584564,
0.008145982399582863,
-0.014063629321753979,
-0.17498643696308136,
0.017595024779438972,
0.007257543038576841,
0.022722741588950157,
0.005284500773996115,
0.03453248739242554,
0.06667785346508026,
0.042788680642843246,
-0.0602128766477108,
-0.027677714824676514,
-0.04839847981929779,
-0.0016005228972062469,
-0.10905798524618149,
-0.22213751077651978,
-0.018838463351130486,
-0.025550419464707375,
0.17125290632247925,
-0.21083836257457733,
0.04711340740323067,
-0.004774345550686121,
0.05487998574972153,
0.013949050568044186,
-0.019272655248641968,
-0.032000016421079636,
0.05793578550219536,
-0.05398781597614288,
-0.041212283074855804,
0.08115138858556747,
0.012199059128761292,
-0.09546120464801788,
-0.037633080035448074,
-0.10417120158672333,
0.1421249806880951,
0.12738904356956482,
-0.11672405898571014,
-0.0657096654176712,
-0.018910061568021774,
-0.06862234324216843,
-0.018341777846217155,
-0.035702895373106,
0.03848552331328392,
0.19515757262706757,
-0.008584443479776382,
0.13840419054031372,
-0.06507367640733719,
-0.02451181225478649,
0.019798051565885544,
-0.04959966242313385,
0.003685327945277095,
0.135342538356781,
0.10803196579217911,
-0.06439536809921265,
0.1471097469329834,
0.1478729397058487,
-0.08989027142524719,
0.164076566696167,
-0.037957530468702316,
-0.05700278654694557,
-0.027241302654147148,
-0.0455855056643486,
-0.021566646173596382,
0.10287421941757202,
-0.18515707552433014,
-0.007976274006068707,
0.02519763819873333,
0.0035888594575226307,
0.005219319835305214,
-0.22645951807498932,
-0.053650327026844025,
0.05109867826104164,
-0.04457620903849602,
-0.003543492406606674,
-0.007423615548759699,
0.002233544597402215,
0.10206222534179688,
-0.00455425726249814,
-0.08455555886030197,
0.033141255378723145,
-0.003937066998332739,
-0.08631489425897598,
0.2052212953567505,
-0.09385253489017487,
-0.17771078646183014,
-0.10565033555030823,
-0.07036307454109192,
-0.05215739086270332,
0.007056060247123241,
0.07503115385770798,
-0.11951767653226852,
-0.01945514790713787,
-0.0816049799323082,
0.0286667812615633,
0.0050012776628136635,
0.0176655575633049,
0.0312381312251091,
-0.002112192567437887,
0.0522066168487072,
-0.10822886228561401,
-0.020700031891465187,
-0.06309893727302551,
-0.04650511220097542,
0.05465009808540344,
0.018853912129998207,
0.1166713535785675,
0.16413907706737518,
-0.002338346792384982,
0.014120565727353096,
-0.03649786487221718,
0.23147641122341156,
-0.07207375019788742,
-0.019283760339021683,
0.14724275469779968,
-0.012881746515631676,
0.05390378087759018,
0.11196215450763702,
0.06766097992658615,
-0.0916331559419632,
0.014643266797065735,
0.04638875275850296,
-0.03918801620602608,
-0.22066658735275269,
-0.04066229984164238,
-0.04911269620060921,
0.02437387779355049,
0.06839477270841599,
0.022736484184861183,
0.04428285360336304,
0.07435554265975952,
0.04079615697264671,
0.03840954601764679,
-0.04271429032087326,
0.054471250623464584,
0.13246268033981323,
0.0186954103410244,
0.10313186794519424,
-0.03513602539896965,
-0.05326450616121292,
0.05883491784334183,
-0.01659216918051243,
0.2144286036491394,
0.0024849872570484877,
0.14159351587295532,
0.06154998391866684,
0.16980677843093872,
-0.035185158252716064,
0.06866680085659027,
-0.01617305912077427,
-0.040746890008449554,
-0.03582670912146568,
-0.029570119455456734,
-0.059687525033950806,
0.03902768716216087,
-0.056811969727277756,
0.08399348706007004,
-0.13915759325027466,
0.012684871442615986,
0.06727787852287292,
0.2794789671897888,
0.025495458394289017,
-0.3152979612350464,
-0.1123858317732811,
0.011033684015274048,
-0.03919573500752449,
-0.00009014346869662404,
0.024136049672961235,
0.08933492004871368,
-0.09758521616458893,
0.04235730320215225,
-0.06414762139320374,
0.08287368714809418,
-0.06449362635612488,
0.06592591106891632,
0.0427221842110157,
0.06850024312734604,
0.014322699047625065,
0.09088566899299622,
-0.28991037607192993,
0.2705853283405304,
-0.010038765147328377,
0.059155408293008804,
-0.08646770566701889,
0.0014503052225336432,
0.0632866844534874,
0.07092951238155365,
0.07169803977012634,
-0.006860123481601477,
-0.003236048622056842,
-0.17326349020004272,
-0.03280024230480194,
0.0328071229159832,
0.06368552148342133,
-0.034936632961034775,
0.08421660959720612,
-0.025439860299229622,
0.011094950139522552,
0.07892243564128876,
0.040938518941402435,
-0.05315690487623215,
-0.10047648102045059,
-0.01225526724010706,
0.031458813697099686,
-0.05944078415632248,
-0.054294634610414505,
-0.12591932713985443,
-0.10484599322080612,
0.1490866094827652,
0.0036942327860742807,
-0.02877337858080864,
-0.10381688177585602,
0.08178707957267761,
0.03828531876206398,
-0.08651566505432129,
0.02264222502708435,
0.009436669759452343,
0.08391891419887543,
0.02367416024208069,
-0.07123027741909027,
0.10834020376205444,
-0.08028549700975418,
-0.1735558807849884,
-0.06549074500799179,
0.0975092351436615,
0.05657711625099182,
0.07783655822277069,
-0.004133510868996382,
-0.0109915966168046,
-0.05143224447965622,
-0.08574846386909485,
0.042248886078596115,
0.028024159371852875,
0.05896523967385292,
0.014072662219405174,
-0.04918627813458443,
0.013356192037463188,
-0.07089504599571228,
-0.03843580186367035,
0.19799664616584778,
0.24020572006702423,
-0.08604232966899872,
0.036707401275634766,
0.034249063581228256,
-0.07560860365629196,
-0.19228993356227875,
0.044577013701200485,
0.05789736658334732,
0.0027150819078087807,
0.05362004041671753,
-0.19754868745803833,
0.12147253751754761,
0.08115622401237488,
-0.013159049674868584,
0.08600065857172012,
-0.30700427293777466,
-0.11319558322429657,
0.1399933248758316,
0.1462010145187378,
0.12088215351104736,
-0.1438206136226654,
-0.0007132498431019485,
-0.031461991369724274,
-0.11460349708795547,
0.1143212839961052,
-0.09343302249908447,
0.12072616815567017,
-0.023526210337877274,
0.12123940140008926,
0.009368562139570713,
-0.04576881229877472,
0.12017424404621124,
0.02092742919921875,
0.09840598702430725,
-0.07180329412221909,
-0.021935632452368736,
0.027178745716810226,
-0.041179485619068146,
0.03022254817187786,
-0.10450135916471481,
0.019567683339118958,
-0.12448035180568695,
-0.03249359503388405,
-0.09228221327066422,
0.03547781705856323,
-0.03915230929851532,
-0.07729648798704147,
-0.05163641273975372,
0.029966028407216072,
0.08319655805826187,
-0.002040472347289324,
0.09379718452692032,
0.021842876449227333,
0.11600504070520401,
0.1119164302945137,
0.09727758169174194,
-0.054947178810834885,
-0.07551794499158859,
-0.027251025661826134,
-0.0073924134485423565,
0.04785503074526787,
-0.15233701467514038,
0.01881331391632557,
0.13489313423633575,
0.018005821853876114,
0.168056920170784,
0.0862375870347023,
-0.040682923048734665,
0.01926465518772602,
0.057832982391119,
-0.15444903075695038,
-0.09576456248760223,
-0.022767947986721992,
-0.06897591799497604,
-0.1220385879278183,
0.035213541239500046,
0.08296231925487518,
-0.07164204120635986,
0.00046931105316616595,
-0.01585174724459648,
0.021240264177322388,
-0.04392145201563835,
0.1647665798664093,
0.04741707071661949,
0.030088059604167938,
-0.10156957805156708,
0.08041085302829742,
0.015862122178077698,
-0.10718400776386261,
0.031965747475624084,
0.07418500632047653,
-0.07486206293106079,
-0.057788241654634476,
0.06869816780090332,
0.21087154746055603,
-0.05970010161399841,
-0.053281500935554504,
-0.15095461905002594,
-0.12610001862049103,
0.08376026898622513,
0.1294068992137909,
0.11731436103582382,
0.0029358728788793087,
-0.08841109275817871,
0.025240829214453697,
-0.11098866164684296,
0.09220398217439651,
0.06049072742462158,
0.038271620869636536,
-0.1360117346048355,
0.1157979890704155,
0.010371189564466476,
0.04418357089161873,
-0.02166377194225788,
0.00827308464795351,
-0.08928628265857697,
0.012144476175308228,
-0.12207236886024475,
-0.029520483687520027,
-0.03976331278681755,
0.012406648136675358,
0.0017081927508115768,
-0.04533249884843826,
-0.04536421597003937,
0.008422994054853916,
-0.11959463357925415,
-0.015581727027893066,
0.037141527980566025,
0.07586314529180527,
-0.11332742124795914,
-0.04161620885133743,
0.026485569775104523,
-0.065317802131176,
0.09352362900972366,
0.061750903725624084,
0.008706808090209961,
0.05923302471637726,
-0.1633487194776535,
0.024299323558807373,
0.0932694524526596,
0.014035869389772415,
0.0499061644077301,
-0.08308964222669601,
-0.014670866541564465,
-0.007683380972594023,
0.039766933768987656,
0.014028651639819145,
0.08129691332578659,
-0.1271032691001892,
0.01744823530316353,
0.003536345437169075,
-0.09294671565294266,
-0.06998240947723389,
0.03262832760810852,
0.0747622698545456,
0.014311526902019978,
0.20272651314735413,
-0.07816266268491745,
0.04452076181769371,
-0.21731480956077576,
0.007282927632331848,
0.0026926747523248196,
-0.10149795562028885,
-0.13229356706142426,
-0.0755867213010788,
0.05521456152200699,
-0.059817880392074585,
0.1352355182170868,
0.04646427929401398,
0.012890752404928207,
0.013127532787621021,
-0.01413293182849884,
0.03143651783466339,
0.00099749187938869,
0.1896950602531433,
0.03029065579175949,
-0.05000760406255722,
0.06581796705722809,
0.05199793726205826,
0.1205286830663681,
0.129627525806427,
0.1982567310333252,
0.14593224227428436,
0.020586853846907616,
0.10959229618310928,
0.030338430777192116,
-0.030545806512236595,
-0.1581438183784485,
0.025118879973888397,
-0.04212102293968201,
0.11920706927776337,
-0.016752582043409348,
0.24379697442054749,
0.06891170889139175,
-0.16222605109214783,
0.0646301656961441,
-0.06454839557409286,
-0.07948721200227737,
-0.10633296519517899,
-0.06990300118923187,
-0.07938313484191895,
-0.14934886991977692,
0.001483576255850494,
-0.1355486959218979,
0.007500220090150833,
0.09190927445888519,
0.01171383261680603,
-0.04385629668831825,
0.1370875984430313,
0.01583770290017128,
0.017653688788414,
0.08805688470602036,
0.003928620833903551,
-0.06791482865810394,
-0.12212827801704407,
-0.0566425696015358,
-0.012766827829182148,
-0.016638727858662605,
0.03842410445213318,
-0.05123954638838768,
-0.06654199957847595,
0.020417416468262672,
-0.024185799062252045,
-0.10174098610877991,
0.008178479969501495,
0.006598432082682848,
0.06464841216802597,
0.04309394583106041,
0.001111063756980002,
0.0173525158315897,
0.002039686543866992,
0.19591140747070312,
-0.07755152881145477,
-0.025873661041259766,
-0.1032378226518631,
0.2315923273563385,
0.020904771983623505,
-0.015406524762511253,
0.03375616669654846,
-0.07141432166099548,
-0.011106031015515327,
0.24757970869541168,
0.21281060576438904,
-0.08773618936538696,
-0.005825194995850325,
-0.0001868423423729837,
0.004327478352934122,
-0.045709189027547836,
0.09868420660495758,
0.15198321640491486,
0.014892181381583214,
-0.09792841225862503,
-0.019693568348884583,
-0.059453465044498444,
-0.024055341258645058,
-0.0268892589956522,
0.06228121370077133,
0.06803416460752487,
0.012797398492693901,
-0.04213897883892059,
0.05411719158291817,
-0.08928489685058594,
-0.10218896716833115,
0.06939391046762466,
-0.2157963663339615,
-0.15352196991443634,
-0.01708640158176422,
0.10154250264167786,
0.03206929937005043,
0.0765378400683403,
-0.019011659547686577,
-0.0011399310315027833,
0.11045973002910614,
-0.018160734325647354,
-0.11844394356012344,
-0.07104877382516861,
0.10842294245958328,
-0.1382681131362915,
0.20610204339027405,
-0.062369704246520996,
0.038563597947359085,
0.12494189292192459,
0.06413519382476807,
-0.052580077201128006,
0.0743044763803482,
0.051140692085027695,
-0.05081452801823616,
0.004815354943275452,
0.10779500752687454,
-0.029796680435538292,
0.07655415683984756,
0.050214383751153946,
-0.1531405746936798,
0.01887759380042553,
-0.0399131253361702,
-0.06630837917327881,
-0.044367291033267975,
-0.007926621474325657,
-0.06923356652259827,
0.11583676934242249,
0.21901483833789825,
-0.027424341067671776,
-0.004762118216603994,
-0.07405556738376617,
0.009370840154588223,
0.051283206790685654,
0.006324347108602524,
-0.05417166277766228,
-0.20606543123722076,
0.011108332313597202,
0.0710698664188385,
-0.013021275401115417,
-0.2565736174583435,
-0.1050436720252037,
0.0037766776513308287,
-0.06672561913728714,
-0.08896733075380325,
0.061359304934740067,
0.0663367435336113,
0.05603806674480438,
-0.04747587442398071,
-0.05951978638768196,
-0.06421943008899689,
0.1677783578634262,
-0.13869357109069824,
-0.08040347695350647
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# wav2vec2-300m-england-0208-ladderside_gate_adapter_attempt-avatar
This model is a fine-tuned version of [vitouphy/wav2vec2-xls-r-300m-english](https://huggingface.co/vitouphy/wav2vec2-xls-r-300m-english) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2758
- Wer: 0.2621
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.001
- train_batch_size: 16
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 1227
- num_epochs: 15
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer |
|:-------------:|:-----:|:-----:|:---------------:|:------:|
| 1.5967 | 1.0 | 1227 | 0.3236 | 0.3103 |
| 0.3438 | 2.0 | 2454 | 0.2916 | 0.2903 |
| 0.3137 | 3.0 | 3681 | 0.2763 | 0.2838 |
| 0.2921 | 4.0 | 4908 | 0.2664 | 0.2740 |
| 0.275 | 5.0 | 6135 | 0.2611 | 0.2677 |
| 0.2601 | 6.0 | 7362 | 0.2560 | 0.2630 |
| 0.2466 | 7.0 | 8589 | 0.2541 | 0.2640 |
| 0.2334 | 8.0 | 9816 | 0.2565 | 0.2635 |
| 0.2212 | 9.0 | 11043 | 0.2568 | 0.2655 |
| 0.21 | 10.0 | 12270 | 0.2582 | 0.2617 |
| 0.1991 | 11.0 | 13497 | 0.2596 | 0.2611 |
| 0.1894 | 12.0 | 14724 | 0.2648 | 0.2598 |
| 0.1805 | 13.0 | 15951 | 0.2726 | 0.2612 |
| 0.1732 | 14.0 | 17178 | 0.2728 | 0.2631 |
| 0.1669 | 15.0 | 18405 | 0.2758 | 0.2621 |
### Framework versions
- Transformers 4.36.0.dev0
- Pytorch 1.12.1+cu113
- Datasets 2.14.7
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["wer"], "base_model": "vitouphy/wav2vec2-xls-r-300m-english", "model-index": [{"name": "wav2vec2-300m-england-0208-ladderside_gate_adapter_attempt-avatar", "results": []}]} | automatic-speech-recognition | Lin25/wav2vec2-300m-england-0208-ladderside_gate_adapter_attempt-avatar | [
"transformers",
"tensorboard",
"safetensors",
"wav2vec2",
"automatic-speech-recognition",
"generated_from_trainer",
"base_model:vitouphy/wav2vec2-xls-r-300m-english",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-07T16:19:51+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-vitouphy/wav2vec2-xls-r-300m-english #license-apache-2.0 #endpoints_compatible #region-us
| wav2vec2-300m-england-0208-ladderside\_gate\_adapter\_attempt-avatar
====================================================================
This model is a fine-tuned version of vitouphy/wav2vec2-xls-r-300m-english on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.2758
* Wer: 0.2621
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.001
* train\_batch\_size: 16
* eval\_batch\_size: 8
* seed: 42
* gradient\_accumulation\_steps: 2
* total\_train\_batch\_size: 32
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 1227
* num\_epochs: 15
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.36.0.dev0
* Pytorch 1.12.1+cu113
* Datasets 2.14.7
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1227\n* num\\_epochs: 15\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.36.0.dev0\n* Pytorch 1.12.1+cu113\n* Datasets 2.14.7\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-vitouphy/wav2vec2-xls-r-300m-english #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1227\n* num\\_epochs: 15\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.36.0.dev0\n* Pytorch 1.12.1+cu113\n* Datasets 2.14.7\n* Tokenizers 0.15.0"
] | [
80,
159,
4,
40
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-vitouphy/wav2vec2-xls-r-300m-english #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1227\n* num\\_epochs: 15\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.36.0.dev0\n* Pytorch 1.12.1+cu113\n* Datasets 2.14.7\n* Tokenizers 0.15.0"
] | [
-0.12465691566467285,
0.13422846794128418,
-0.0033921669237315655,
0.04942885786294937,
0.08694503456354141,
0.02121187187731266,
0.10567963868379593,
0.14321152865886688,
-0.05826539918780327,
0.12673917412757874,
0.11136090010404587,
0.0845835879445076,
0.07596516609191895,
0.1459297239780426,
-0.02841995656490326,
-0.29495757818222046,
0.0224810428917408,
-0.016161799430847168,
-0.1142239198088646,
0.10135062783956528,
0.08888418972492218,
-0.10815936326980591,
0.030373359099030495,
0.006174994166940451,
-0.08791209757328033,
-0.013285565190017223,
-0.03192336484789848,
-0.0618252195417881,
0.10925420373678207,
0.05721037834882736,
0.0789627730846405,
0.038640573620796204,
0.0832996517419815,
-0.27146294713020325,
0.013236995786428452,
0.05100950598716736,
0.02037181332707405,
0.07485561817884445,
0.09717301279306412,
-0.01830236427485943,
0.10736017674207687,
-0.10179122537374496,
0.07814016193151474,
0.03811271861195564,
-0.09145806729793549,
-0.3031558394432068,
-0.07879561185836792,
0.05247556045651436,
0.14443330466747284,
0.08126156777143478,
-0.03472461551427841,
0.07135944068431854,
-0.056889958679676056,
0.07816186547279358,
0.22543039917945862,
-0.2625831067562103,
-0.06478251516819,
-0.01270261686295271,
0.046449512243270874,
0.05287967249751091,
-0.1144741028547287,
-0.019110586494207382,
0.020056825131177902,
0.01791469193994999,
0.08630358427762985,
0.01640322245657444,
0.05626271665096283,
0.019265267997980118,
-0.1485597789287567,
-0.031407251954078674,
0.11930002272129059,
0.09399271011352539,
-0.01598619483411312,
-0.1202399805188179,
-0.033291932195425034,
-0.1576727032661438,
-0.05974581465125084,
-0.011626476421952248,
0.0199188981205225,
-0.035975679755210876,
-0.08254094421863556,
0.020756477490067482,
-0.06593198329210281,
-0.07009761035442352,
0.013556314632296562,
0.13418807089328766,
0.04965931922197342,
-0.03940175846219063,
0.029303492978215218,
0.08000955730676651,
0.039811696857213974,
-0.1509731411933899,
0.0010899495100602508,
0.030081072822213173,
-0.10689733177423477,
-0.012487096711993217,
-0.01752214878797531,
-0.0037971828132867813,
0.031802933663129807,
0.14865058660507202,
-0.02718919701874256,
0.09470890462398529,
0.02501765266060829,
0.010753295384347439,
-0.08100786060094833,
0.14188045263290405,
-0.06458915770053864,
-0.08041055500507355,
-0.04937044158577919,
0.11280500143766403,
0.023620815947651863,
-0.016562215983867645,
-0.07768469303846359,
0.02609618753194809,
0.0910460501909256,
0.04789042845368385,
-0.002183925360441208,
0.00941375084221363,
-0.07486971467733383,
-0.023043794557452202,
0.04457709193229675,
-0.10666938871145248,
0.05648881942033768,
0.040808551013469696,
-0.0424368754029274,
-0.005682796239852905,
-0.0050771646201610565,
0.03140757232904434,
-0.005957894492894411,
0.11618456989526749,
-0.06692945957183838,
-0.01908540353178978,
-0.052657317370176315,
-0.09988176077604294,
0.03331562876701355,
-0.03511057794094086,
-0.0008417097269557416,
-0.07353773713111877,
-0.08586865663528442,
-0.05412377789616585,
0.05573324114084244,
-0.057308558374643326,
-0.062025632709264755,
-0.07954680919647217,
-0.05664779990911484,
0.0692707896232605,
-0.009371276944875717,
0.1232767179608345,
-0.0534224733710289,
0.09213440120220184,
0.004930954892188311,
0.0687665119767189,
0.05363429710268974,
0.05418974533677101,
-0.03217229247093201,
0.04693179205060005,
-0.1655811071395874,
0.07958008348941803,
-0.10107725858688354,
0.0462600402534008,
-0.16477049887180328,
-0.08749037981033325,
-0.010742590762674809,
0.0037504418287426233,
0.0900646448135376,
0.11584417521953583,
-0.1839005947113037,
-0.09751975536346436,
0.1800609976053238,
-0.08436178416013718,
-0.10292261838912964,
0.14855927228927612,
-0.018097100779414177,
-0.04493969306349754,
0.03095185197889805,
0.18467818200588226,
0.09480796754360199,
-0.1019318625330925,
-0.014242682605981827,
-0.048202622681856155,
0.125900000333786,
0.030600544065237045,
0.11503001302480698,
-0.055704984813928604,
0.015985598787665367,
-0.006210555788129568,
-0.022121546790003777,
0.05842384323477745,
-0.07523756474256516,
-0.08422631770372391,
-0.013014234602451324,
-0.07502517849206924,
0.026611171662807465,
0.05130164325237274,
0.025490496307611465,
-0.08779723197221756,
-0.13890734314918518,
0.009012717753648758,
0.112187460064888,
-0.09905129671096802,
0.02573724091053009,
-0.07166428864002228,
0.06583460420370102,
-0.02471533976495266,
-0.005061803851276636,
-0.13695982098579407,
-0.011253075674176216,
0.02891036868095398,
-0.04860220104455948,
0.006382744759321213,
-0.023464569821953773,
0.07495082914829254,
0.05582648143172264,
-0.0624234601855278,
-0.06751437485218048,
-0.03387540578842163,
0.010957157239317894,
-0.07078094780445099,
-0.2536008656024933,
-0.04739753529429436,
-0.041560348123311996,
0.17449812591075897,
-0.23287859559059143,
0.007978971116244793,
0.009731430560350418,
0.14297343790531158,
0.0404483899474144,
-0.049616739153862,
-0.004274751991033554,
0.05847722664475441,
-0.029938064515590668,
-0.06424052268266678,
0.03185473382472992,
-0.012096774764358997,
-0.13092069327831268,
0.010333947837352753,
-0.1443648487329483,
0.09557273238897324,
0.10555122047662735,
0.043106622993946075,
-0.08215141296386719,
-0.08893183618783951,
-0.056019674986600876,
-0.04625513032078743,
-0.032676562666893005,
-0.005051587242633104,
0.1368291676044464,
0.022296762093901634,
0.09638182073831558,
-0.07204465568065643,
-0.03881654888391495,
0.03599295765161514,
0.014294483698904514,
-0.04407728835940361,
0.16101762652397156,
0.07079316675662994,
-0.07004653662443161,
0.10077449679374695,
0.13098689913749695,
-0.04669530689716339,
0.12386652827262878,
-0.06117767095565796,
-0.09604673832654953,
-0.03909028694033623,
0.028104135766625404,
0.038013212382793427,
0.10401234775781631,
-0.12499450147151947,
0.00011558888218132779,
0.020688150078058243,
0.02525905705988407,
0.0072943586856126785,
-0.17663924396038055,
-0.01109160203486681,
0.051894038915634155,
-0.05910249054431915,
-0.0072919102385640144,
-0.014158312231302261,
-0.018293175846338272,
0.08396372199058533,
0.013992696069180965,
-0.06050700694322586,
-0.02008850686252117,
-0.015141600742936134,
-0.10052376240491867,
0.18735739588737488,
-0.12100609391927719,
-0.13682174682617188,
-0.1107015609741211,
-0.024012308567762375,
-0.004433472640812397,
-0.013666593469679356,
0.054017987102270126,
-0.11229386925697327,
-0.042320843786001205,
-0.08469092100858688,
0.02971469797194004,
-0.059263020753860474,
0.0501706637442112,
0.024538688361644745,
0.006641092710196972,
0.04327184706926346,
-0.0882573127746582,
0.021516606211662292,
-0.019353307783603668,
0.006218044552952051,
0.01273646391928196,
0.013273934833705425,
0.09901361912488937,
0.16751913726329803,
0.051268115639686584,
0.025012869387865067,
-0.047743625938892365,
0.17528291046619415,
-0.1031772792339325,
0.005883621983230114,
0.09682352095842361,
0.0012472504749894142,
0.04982517287135124,
0.16714616119861603,
0.04836713522672653,
-0.08160148561000824,
0.02094469591975212,
0.02905653603374958,
-0.010264560580253601,
-0.23609279096126556,
-0.04486323148012161,
-0.05981893837451935,
-0.008341739885509014,
0.11863920092582703,
0.040315914899110794,
-0.02237948402762413,
0.03303240239620209,
-0.014322753064334393,
-0.004443172365427017,
0.014414799399673939,
0.06731575727462769,
0.08740384131669998,
0.04183557257056236,
0.12013236433267593,
-0.025197764858603477,
-0.028503015637397766,
0.039633914828300476,
-0.006057452410459518,
0.22490598261356354,
0.013164778240025043,
0.15907412767410278,
0.03730512037873268,
0.14815069735050201,
0.013581855222582817,
0.04572770744562149,
0.013271297328174114,
-0.025757092982530594,
0.0042893411591649055,
-0.06349009275436401,
-0.0153342979028821,
0.06815525889396667,
0.10480351746082306,
0.015667922794818878,
-0.11365535855293274,
0.016864264383912086,
0.028722455725073814,
0.2802526652812958,
0.10110834985971451,
-0.2881394326686859,
-0.08494052290916443,
0.024226831272244453,
-0.06429651379585266,
-0.023689718917012215,
0.030396588146686554,
0.10517676174640656,
-0.055776551365852356,
0.08272742480039597,
-0.05832483991980553,
0.07868875563144684,
-0.05875645577907562,
-0.0075154732912778854,
0.040583688765764236,
0.0838610976934433,
-0.011864845640957355,
0.05436096340417862,
-0.23350310325622559,
0.300857812166214,
0.002144297119230032,
0.06185218319296837,
-0.041023120284080505,
0.02903771586716175,
0.023582953959703445,
-0.02367059886455536,
0.09743359684944153,
-0.012309964746236801,
-0.14949147403240204,
-0.1586543619632721,
-0.10755860805511475,
0.023196902126073837,
0.11868114769458771,
-0.06869807839393616,
0.10237953066825867,
-0.022582538425922394,
-0.035772960633039474,
0.06107112765312195,
-0.0437554307281971,
-0.11314672976732254,
-0.13791383802890778,
0.01837385818362236,
0.02390027418732643,
0.04385851323604584,
-0.08861307799816132,
-0.11458326876163483,
-0.09115555882453918,
0.15207993984222412,
-0.09642759710550308,
-0.008278883993625641,
-0.13828150928020477,
0.07758791744709015,
0.1609092801809311,
-0.08590144664049149,
0.04970823600888252,
0.006224216427654028,
0.12112695723772049,
-0.004389591049402952,
-0.021128958091139793,
0.12441623210906982,
-0.08886036276817322,
-0.19982311129570007,
-0.07579217851161957,
0.16541942954063416,
0.039482783526182175,
0.06833011656999588,
-0.020997148007154465,
0.041578106582164764,
-0.009769851341843605,
-0.0781874731183052,
0.08743233233690262,
0.05540665239095688,
0.022205648943781853,
0.038115572184324265,
-0.023264657706022263,
-0.03284603729844093,
-0.06224752217531204,
-0.07518796622753143,
0.13796380162239075,
0.3099845349788666,
-0.09964250773191452,
0.054175637662410736,
0.07267068326473236,
-0.04165518283843994,
-0.14749953150749207,
-0.011213596910238266,
0.11096030473709106,
0.032545384019613266,
0.019626058638095856,
-0.1909829080104828,
0.04654333367943764,
0.08012045174837112,
-0.022109389305114746,
0.054848238825798035,
-0.299514502286911,
-0.13938665390014648,
0.11130546778440475,
0.0951491966843605,
-0.023050449788570404,
-0.1629284918308258,
-0.07291495054960251,
-0.01841890625655651,
-0.08837102353572845,
0.058659877628088,
-0.021678363904356956,
0.10548903048038483,
0.0014504729770123959,
0.0050860196352005005,
0.014934753999114037,
-0.056419748812913895,
0.15836037695407867,
-0.0077098277397453785,
0.03126494586467743,
-0.0105955321341753,
0.022249100729823112,
-0.03787294030189514,
-0.0649576187133789,
0.0032912935130298138,
-0.08965753763914108,
0.03260745853185654,
-0.11746523529291153,
-0.03461860492825508,
-0.06254439055919647,
0.013990161940455437,
-0.04498837888240814,
-0.03865162655711174,
-0.041557587683200836,
0.049785126000642776,
0.07602188736200333,
-0.00866411067545414,
0.13438765704631805,
-0.03367864340543747,
0.1528163105249405,
0.09713498502969742,
0.08908714354038239,
0.0037102289497852325,
-0.06963146477937698,
-0.010377682745456696,
-0.034309402108192444,
0.03859543427824974,
-0.1360805630683899,
0.026729537174105644,
0.14457173645496368,
0.035194989293813705,
0.1566532403230667,
0.04956107586622238,
-0.08772237598896027,
0.010831729508936405,
0.07128259539604187,
-0.08107803761959076,
-0.17113368213176727,
-0.016871606931090355,
0.044225748628377914,
-0.14615696668624878,
0.002111061243340373,
0.10808148235082626,
-0.034818731248378754,
-0.008472893387079239,
0.009534978307783604,
0.041008636355400085,
-0.017468124628067017,
0.21549323201179504,
0.035228949040174484,
0.07741193473339081,
-0.08696434646844864,
0.0661788135766983,
0.0627426728606224,
-0.18075448274612427,
0.048869747668504715,
0.08888214826583862,
-0.05956956371665001,
-0.0223891269415617,
0.034298304468393326,
0.08823622018098831,
0.013687703758478165,
-0.05001280456781387,
-0.10624095052480698,
-0.1416071206331253,
0.09546911716461182,
0.08980558812618256,
0.0286586731672287,
0.010445079766213894,
-0.018578147515654564,
0.028939250856637955,
-0.08818807452917099,
0.11803793907165527,
0.08050849288702011,
0.06904196739196777,
-0.13148103654384613,
0.0962539091706276,
0.005609280429780483,
-0.014090328477323055,
0.0027510446961969137,
0.016220945864915848,
-0.12626852095127106,
0.00307014980353415,
-0.10474570840597153,
-0.010948458686470985,
-0.08163022249937057,
-0.004538469947874546,
0.007960007525980473,
-0.06569186598062515,
-0.044137779623270035,
0.0034379728604108095,
-0.10066719353199005,
-0.04576309770345688,
-0.022153012454509735,
0.06914004683494568,
-0.11724007874727249,
-0.020933514460921288,
0.033753279596567154,
-0.1112338975071907,
0.09764771163463593,
0.03073752298951149,
0.036367110908031464,
0.019938191398978233,
-0.09965749830007553,
0.021258065477013588,
0.032896630465984344,
-0.006201489828526974,
0.02797832153737545,
-0.18904319405555725,
-0.016797177493572235,
-0.026024438440799713,
0.01231157872825861,
0.0006191044813022017,
0.042001478374004364,
-0.1148499995470047,
-0.0026931690517812967,
-0.06373682618141174,
-0.07059185951948166,
-0.054164156317710876,
0.05040789768099785,
0.0708576962351799,
0.01201667357236147,
0.14907217025756836,
-0.0870274007320404,
0.053721170872449875,
-0.2193002998828888,
0.0027807094156742096,
-0.031426187604665756,
-0.05132703110575676,
-0.05167054384946823,
-0.019764194265007973,
0.08048106729984283,
-0.054605633020401,
0.0773584246635437,
-0.06151328608393669,
0.0437510684132576,
0.04007377475500107,
-0.10574456304311752,
0.025782490149140358,
0.04471040144562721,
0.19997578859329224,
0.05083022639155388,
-0.02550613135099411,
0.04279240965843201,
0.0024539080914109945,
0.07178132981061935,
0.137655571103096,
0.13864757120609283,
0.16311699151992798,
0.04619503393769264,
0.08988433331251144,
0.056616492569446564,
-0.12446460127830505,
-0.14952975511550903,
0.13274778425693512,
-0.0505509190261364,
0.12263701856136322,
-0.0036637966986745596,
0.19783629477024078,
0.1195773184299469,
-0.19866126775741577,
0.0333433635532856,
-0.03165009990334511,
-0.0888478234410286,
-0.11334225535392761,
-0.0694054514169693,
-0.09658876061439514,
-0.18413053452968597,
0.0030663402285426855,
-0.10130501538515091,
0.043609704822301865,
0.02714058943092823,
0.04914320260286331,
0.05326608940958977,
0.09732173383235931,
0.05968717113137245,
0.016358498483896255,
0.09090957045555115,
0.02789892442524433,
-0.020201245322823524,
-0.025575658306479454,
-0.08776678144931793,
0.03609349951148033,
-0.04193320870399475,
0.04367658123373985,
-0.04059930145740509,
-0.0959504172205925,
0.07702730596065521,
0.020310360938310623,
-0.10311459749937057,
0.01788952387869358,
-0.006651570554822683,
0.05427092686295509,
0.10480048507452011,
0.040353260934352875,
-0.016790350899100304,
-0.015592037700116634,
0.21644027531147003,
-0.09369787573814392,
-0.04794204607605934,
-0.13067875802516937,
0.21886341273784637,
-0.002035768935456872,
0.00789148174226284,
0.018390290439128876,
-0.08484547585248947,
-0.001614079112187028,
0.1462174952030182,
0.14836829900741577,
-0.008537434972822666,
-0.0128974923864007,
0.03556783124804497,
-0.00969971064478159,
-0.03504331409931183,
0.060224808752536774,
0.12247539311647415,
0.08845093101263046,
-0.05734921619296074,
-0.047053441405296326,
-0.045989684760570526,
-0.054885007441043854,
-0.028954679146409035,
0.07112771272659302,
0.02638692781329155,
-0.013109331950545311,
-0.007614858448505402,
0.11813981086015701,
-0.036612775176763535,
-0.13547247648239136,
0.03129667788743973,
-0.19458259642124176,
-0.18193116784095764,
-0.030999675393104553,
0.0831923708319664,
0.03019033372402191,
0.04178052768111229,
0.004377693869173527,
-0.032944366335868835,
0.1087295338511467,
0.005747510585933924,
-0.061776306480169296,
-0.09095799177885056,
0.07698226720094681,
-0.06195938214659691,
0.17110060155391693,
-0.03269008919596672,
0.02339637652039528,
0.13258329033851624,
0.07403609901666641,
-0.08537306636571884,
0.04070720076560974,
0.0881977528333664,
-0.09510976076126099,
0.06531482934951782,
0.16427946090698242,
-0.043467238545417786,
0.1507003903388977,
0.06624232977628708,
-0.10473897308111191,
0.03149082884192467,
-0.09452635049819946,
-0.06903994083404541,
-0.05181585252285004,
0.028023090213537216,
-0.04808000475168228,
0.15068504214286804,
0.18012842535972595,
-0.06588926166296005,
-0.02063954994082451,
-0.028162136673927307,
0.01489912997931242,
0.0314309261739254,
0.14450062811374664,
-0.02213136851787567,
-0.2586686909198761,
0.026346303522586823,
0.005942780990153551,
0.03500347584486008,
-0.24705246090888977,
-0.09849528968334198,
0.018885361030697823,
-0.04813682660460472,
-0.07339941710233688,
0.11952648311853409,
0.08413115888834,
0.0388621985912323,
-0.06799168139696121,
-0.1152370497584343,
-0.0226230900734663,
0.16855163872241974,
-0.16250088810920715,
-0.05446529760956764
] |
null | null | diffusers |
# LoRA DreamBooth - danaleee/CL_rank4_iter800_valprompt
These are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using [DreamBooth](https://dreambooth.github.io/). You can find some example images in the following.




LoRA for the text encoder was enabled: False.
| {"license": "creativeml-openrail-m", "tags": ["stable-diffusion", "stable-diffusion-diffusers", "text-to-image", "diffusers", "lora"], "base_model": "CompVis/stable-diffusion-v1-4", "instance_prompt": "a photo of sks teddybear", "inference": true} | text-to-image | danaleee/CL_rank4_iter800_valprompt | [
"diffusers",
"tensorboard",
"stable-diffusion",
"stable-diffusion-diffusers",
"text-to-image",
"lora",
"base_model:CompVis/stable-diffusion-v1-4",
"license:creativeml-openrail-m",
"region:us"
] | 2024-02-07T16:20:41+00:00 | [] | [] | TAGS
#diffusers #tensorboard #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us
|
# LoRA DreamBooth - danaleee/CL_rank4_iter800_valprompt
These are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using DreamBooth. You can find some example images in the following.
!img_0
!img_1
!img_2
!img_3
LoRA for the text encoder was enabled: False.
| [
"# LoRA DreamBooth - danaleee/CL_rank4_iter800_valprompt\n\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
"TAGS\n#diffusers #tensorboard #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us \n",
"# LoRA DreamBooth - danaleee/CL_rank4_iter800_valprompt\n\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
70,
109
] | [
"passage: TAGS\n#diffusers #tensorboard #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #region-us \n# LoRA DreamBooth - danaleee/CL_rank4_iter800_valprompt\n\nThese are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks teddybear using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
-0.06403809040784836,
-0.01912468485534191,
-0.0027079954743385315,
0.0819823294878006,
0.10148093104362488,
0.02776934951543808,
0.21502014994621277,
0.08201627433300018,
0.03845279663801193,
0.0702517181634903,
0.11390383541584015,
0.07433994114398956,
-0.005233790259808302,
0.09161096811294556,
-0.015448933467268944,
-0.19556494057178497,
-0.01394282653927803,
-0.01825922727584839,
-0.06194775551557541,
0.045116912573575974,
0.0458020381629467,
-0.04542998969554901,
0.12057823687791824,
-0.019657842814922333,
-0.1661662757396698,
0.03075687400996685,
-0.0040834746323525906,
-0.057399895042181015,
0.08059728145599365,
0.06897717714309692,
0.033515460789203644,
0.06173942610621452,
0.04516362398862839,
-0.13466651737689972,
0.037042032927274704,
0.020500198006629944,
-0.05727791786193848,
0.08277659863233566,
-0.04625704139471054,
-0.024833759292960167,
0.11181218177080154,
-0.0191158726811409,
0.003722669091075659,
0.017441706731915474,
-0.0811433270573616,
-0.05281444638967514,
-0.021313855424523354,
0.007304344791918993,
0.035774972289800644,
0.037050750106573105,
0.01162349246442318,
0.08957083523273468,
-0.022555101662874222,
0.08422932028770447,
0.26362425088882446,
-0.2161150723695755,
-0.040414001792669296,
0.21261288225650787,
0.012910652905702591,
0.08581091463565826,
-0.04512929543852806,
0.09801483154296875,
0.09915975481271744,
-0.031582072377204895,
0.04734696447849274,
-0.04597477242350578,
-0.005726148374378681,
-0.049544841051101685,
-0.10984606295824051,
0.06504223495721817,
0.1409047394990921,
0.006971485447138548,
-0.048718683421611786,
-0.15149550139904022,
-0.03959684073925018,
0.00257797259837389,
-0.02574186958372593,
0.0466853603720665,
0.006212103646248579,
-0.008213100023567677,
-0.03954506292939186,
-0.0405924990773201,
-0.09285756200551987,
-0.05998034402728081,
-0.017353584989905357,
0.0704701840877533,
-0.006895981263369322,
0.05989989638328552,
-0.01428080815821886,
0.11205750703811646,
-0.098143070936203,
-0.1281043142080307,
0.0507097989320755,
-0.0671081468462944,
0.015217331238090992,
0.06588982045650482,
-0.012961341999471188,
-0.15642733871936798,
0.04826904088258743,
0.00889264140278101,
0.14836087822914124,
0.01905730366706848,
-0.010243989527225494,
0.1286815106868744,
-0.00557751813903451,
0.04606293514370918,
-0.05553492158651352,
-0.004334692843258381,
0.014335514977574348,
0.027929754927754402,
0.08992905169725418,
-0.07960252463817596,
-0.13882891833782196,
-0.013746518641710281,
-0.04508064314723015,
0.02152709849178791,
-0.08297059684991837,
0.028232822194695473,
-0.09434807300567627,
-0.013114052824676037,
0.08286312967538834,
0.0003270331071689725,
0.04328123852610588,
-0.02405433915555477,
-0.023418210446834564,
0.1230260506272316,
0.16339431703090668,
0.019526487216353416,
0.018288109451532364,
0.08981496095657349,
-0.0655788853764534,
0.04973616823554039,
-0.024370761588215828,
-0.11493290215730667,
0.017808018252253532,
-0.16341452300548553,
-0.008707420900464058,
-0.13552238047122955,
-0.02857116237282753,
-0.003302348079159856,
0.041668541729450226,
-0.04682629555463791,
0.06026662886142731,
-0.058109305799007416,
-0.08863845467567444,
-0.007945443503558636,
0.061928994953632355,
0.00666175689548254,
-0.023748598992824554,
0.05410725250840187,
-0.02702663466334343,
0.1431545466184616,
-0.10466044396162033,
-0.03136221319437027,
-0.0645868182182312,
0.008374386467039585,
-0.14520969986915588,
0.06668452173471451,
-0.05664831027388573,
0.059087272733449936,
-0.0663350448012352,
-0.04498099163174629,
-0.05097215622663498,
0.05490422248840332,
0.035455211997032166,
0.13559073209762573,
-0.25775545835494995,
-0.0659942701458931,
0.0988934114575386,
-0.15342940390110016,
-0.0840730220079422,
0.05597928538918495,
-0.019737768918275833,
0.10469520092010498,
0.05975286662578583,
0.10876773297786713,
0.07410487532615662,
-0.29768913984298706,
-0.0006618750630877912,
-0.04755578190088272,
-0.03150198981165886,
-0.05140906572341919,
-0.00028375457623042166,
0.06416554003953934,
0.006773205939680338,
0.03413834422826767,
-0.04580661281943321,
0.07737613469362259,
-0.041915033012628555,
-0.011432375758886337,
-0.021972807124257088,
-0.0342831052839756,
-0.009799250401556492,
0.0037940118927508593,
0.06345032155513763,
-0.01870134100317955,
-0.04630181938409805,
0.0034789645578712225,
0.0353398434817791,
-0.06441408395767212,
0.024669727310538292,
-0.0268857441842556,
0.09516798704862595,
-0.07860253751277924,
-0.016681304201483727,
-0.10060384124517441,
0.01891675405204296,
0.040081292390823364,
0.124611496925354,
0.09219381213188171,
0.05741831660270691,
0.10061176121234894,
0.06420552730560303,
0.0005364702083170414,
0.015330390073359013,
0.07087218016386032,
-0.024467214941978455,
-0.0554451122879982,
-0.20251840353012085,
0.027448482811450958,
-0.0866859033703804,
0.046237409114837646,
-0.20928120613098145,
0.02281379885971546,
0.07146544754505157,
0.1806805282831192,
0.11007224768400192,
-0.029057450592517853,
0.06384287029504776,
0.023045659065246582,
-0.06231020390987396,
-0.0483199879527092,
0.03972896188497543,
-0.007400625851005316,
-0.13515974581241608,
0.1649106740951538,
-0.15838713943958282,
0.06789637356996536,
0.1294366419315338,
-0.03711557388305664,
-0.07396496832370758,
-0.07614585012197495,
-0.005144384689629078,
0.03765964135527611,
-0.060470789670944214,
-0.019587332382798195,
0.09076426923274994,
-0.0034756192471832037,
0.1578441858291626,
-0.022008264437317848,
0.0029480031225830317,
0.07042249292135239,
-0.03163981810212135,
-0.08039087057113647,
0.10068253427743912,
0.1123909056186676,
0.01021781750023365,
0.013487220741808414,
0.06899360567331314,
-0.018049826845526695,
0.1459129899740219,
-0.002208979334682226,
-0.0474790558218956,
-0.04496482014656067,
0.016903219744563103,
0.056954074651002884,
0.14856073260307312,
-0.015561042353510857,
-0.03191840648651123,
-0.007485839072614908,
-0.07780373096466064,
0.009159459732472897,
-0.15655076503753662,
-0.026459986343979836,
0.029839573428034782,
-0.032735425978899,
0.12642532587051392,
0.10329664498567581,
-0.08949295431375504,
0.09635338187217712,
-0.1102837324142456,
-0.07886908203363419,
-0.030347175896167755,
-0.021086134016513824,
-0.06439391523599625,
0.11569448560476303,
-0.05298544093966484,
-0.1547464281320572,
-0.1781131774187088,
-0.0080312704667449,
0.008416324853897095,
-0.010307404212653637,
0.04804869741201401,
-0.08926640450954437,
-0.06611368805170059,
-0.11586663872003555,
0.04077331721782684,
0.037470053881406784,
0.048737313598394394,
0.04871385172009468,
-0.021726233884692192,
-0.006538061425089836,
-0.11117211729288101,
0.006342211272567511,
-0.08251069486141205,
0.09787896275520325,
0.07747673243284225,
-0.004547393415123224,
0.11337340623140335,
0.12908640503883362,
0.02779196947813034,
0.04989927262067795,
0.01005121972411871,
0.22071242332458496,
-0.032919853925704956,
0.08564963936805725,
0.11564260721206665,
-0.012399379163980484,
0.061680976301431656,
0.10612256079912186,
0.056032899767160416,
-0.07105161994695663,
0.048993002623319626,
-0.003030498279258609,
-0.15001854300498962,
-0.07217659801244736,
-0.0467098169028759,
-0.03957079350948334,
-0.03749268501996994,
0.0849454253911972,
0.04893402382731438,
0.10388457030057907,
0.09535162150859833,
0.06329978257417679,
0.11502949893474579,
0.059253957122564316,
0.07928356528282166,
0.05794480815529823,
-0.04317384585738182,
0.04551837965846062,
-0.08380094170570374,
-0.11263305693864822,
0.10672270506620407,
-0.07184915989637375,
0.20058824121952057,
-0.06643599271774292,
0.04637473449110985,
0.05345495790243149,
-0.051749151200056076,
0.0974617600440979,
0.041994646191596985,
-0.06905864924192429,
-0.01166562456637621,
-0.042756546288728714,
-0.1289643943309784,
0.09512977302074432,
0.08718732744455338,
-0.026657065376639366,
-0.04318408668041229,
-0.00819322932511568,
0.03656448423862457,
0.013363540172576904,
0.013277050107717514,
0.1431867480278015,
-0.26281505823135376,
0.038054801523685455,
0.012084228917956352,
0.06484752148389816,
-0.0004629245086107403,
0.007180110551416874,
0.22001037001609802,
0.0047096251510083675,
0.08374249935150146,
-0.06997823715209961,
0.05791766196489334,
-0.027488617226481438,
-0.024806804955005646,
-0.012226812541484833,
0.12402921915054321,
-0.029434500262141228,
-0.05148465558886528,
-0.2036358118057251,
0.08324871957302094,
0.00878535769879818,
-0.010234410874545574,
-0.07041724026203156,
-0.026224952191114426,
0.03248727694153786,
-0.00752894626930356,
0.08426511287689209,
0.005954920779913664,
0.010060069151222706,
-0.10802402347326279,
-0.1486591100692749,
-0.04082779958844185,
0.08773882687091827,
-0.03321250155568123,
0.046611931174993515,
0.056865669786930084,
-0.03996516391634941,
0.015953104943037033,
0.012187889777123928,
-0.12881436944007874,
-0.10952196270227432,
0.010912762023508549,
0.1555682122707367,
-0.0813402533531189,
-0.026349544525146484,
-0.10340709239244461,
-0.04820692166686058,
0.05996250733733177,
-0.04283863678574562,
-0.048442158848047256,
-0.06449680775403976,
0.013150354847311974,
0.08591903001070023,
-0.0281917043030262,
0.005537814926356077,
-0.038593485951423645,
0.07755585014820099,
-0.08191221952438354,
-0.14968310296535492,
0.0775071308016777,
-0.019168537110090256,
-0.1198141798377037,
-0.09681004285812378,
0.10655517876148224,
-0.00758502259850502,
-0.0035073726903647184,
-0.01501635741442442,
0.03575945273041725,
0.02354598231613636,
-0.09390327334403992,
0.08201109617948532,
0.1418909728527069,
-0.1187058612704277,
0.08629121631383896,
-0.035414498299360275,
-0.006325878202915192,
-0.05165180191397667,
0.007016763556748629,
0.11422546207904816,
0.23664410412311554,
-0.10256873816251755,
0.09378791600465775,
0.02340368926525116,
-0.08606907725334167,
-0.19479864835739136,
-0.015544246882200241,
-0.0023345842491835356,
0.03728392720222473,
0.017479248344898224,
-0.09042929857969284,
0.10055974870920181,
0.05373078212141991,
0.0032882175873965025,
0.2544887959957123,
-0.37809187173843384,
-0.14377321302890778,
-0.018459968268871307,
0.1592307984828949,
0.23886846005916595,
-0.17076843976974487,
-0.0791986957192421,
-0.02622740902006626,
-0.07412165403366089,
0.15293575823307037,
-0.046558499336242676,
0.09798591583967209,
-0.032230835407972336,
-0.03402456268668175,
0.027583377435803413,
-0.0384509302675724,
0.11660727858543396,
-0.0070481677539646626,
0.04699692502617836,
-0.08609452843666077,
-0.08124737441539764,
0.08245531469583511,
-0.04232442006468773,
0.019716650247573853,
-0.14503423869609833,
0.01048110518604517,
-0.0691995695233345,
-0.012037270702421665,
0.012544753029942513,
-0.0019452450796961784,
-0.02225651778280735,
-0.03897279128432274,
-0.10981116443872452,
0.018080491572618484,
-0.010515352711081505,
-0.012333390302956104,
0.0770924836397171,
-0.021311555057764053,
0.042737022042274475,
0.14731469750404358,
-0.03175806999206543,
0.0591786652803421,
-0.06786045432090759,
-0.015309245325624943,
-0.03989124670624733,
0.10987086594104767,
-0.16638875007629395,
-0.0018515256233513355,
0.15063625574111938,
0.0842001736164093,
0.09925667196512222,
0.021799137815833092,
-0.09589464962482452,
0.09262481331825256,
0.13114354014396667,
-0.08792462199926376,
-0.009335857816040516,
-0.02910660393536091,
-0.036676760762929916,
0.11080039292573929,
0.018999915570020676,
0.18706001341342926,
-0.07697457075119019,
0.041624557226896286,
0.01709686778485775,
0.0066497912630438805,
-0.017004067078232765,
0.08545937389135361,
0.04088068753480911,
0.00535707687959075,
-0.06761232018470764,
0.05493567883968353,
-0.01299224141985178,
0.015374132432043552,
0.010864801704883575,
0.07163301110267639,
-0.06333953887224197,
-0.011716494336724281,
0.006595042999833822,
0.20309200882911682,
-0.10569093376398087,
0.008596858941018581,
-0.10688471049070358,
-0.10488927364349365,
0.022659244015812874,
0.15920978784561157,
0.05492470785975456,
0.007096205372363329,
-0.04050235077738762,
-0.024366598576307297,
-0.05520016327500343,
0.03864378482103348,
0.05910024046897888,
0.07781106978654861,
-0.23290492594242096,
-0.027771761640906334,
0.015064816921949387,
-0.03199734166264534,
-0.07129121571779251,
-0.03907070681452751,
-0.111953005194664,
-0.021807221695780754,
-0.039924751967191696,
0.10979276150465012,
-0.056607093662023544,
-0.02088336832821369,
-0.013066990301012993,
-0.0397760309278965,
-0.019163357093930244,
0.013350007124245167,
-0.038817211985588074,
-0.01609853096306324,
0.0016378369182348251,
-0.012154461815953255,
-0.052674002945423126,
-0.08565595000982285,
-0.015558294020593166,
-0.09302201867103577,
0.04213583841919899,
-0.016844207420945168,
-0.07665203511714935,
0.01447745505720377,
-0.22398485243320465,
0.0351681150496006,
0.18199679255485535,
-0.0039426484145224094,
-0.011056573130190372,
0.01795199327170849,
0.01671888493001461,
-0.034877121448516846,
0.07718933373689651,
0.0012086834758520126,
0.058319393545389175,
-0.07721461355686188,
-0.010127730667591095,
-0.07276295125484467,
-0.007137161213904619,
-0.05593208596110344,
0.0713859274983406,
0.12814286351203918,
0.14264225959777832,
0.12549665570259094,
-0.1029796153306961,
0.08036024868488312,
-0.11136461794376373,
0.014355169609189034,
0.020769376307725906,
-0.06955956667661667,
0.030121371150016785,
-0.025551963597536087,
-0.013226136565208435,
-0.04952284321188927,
0.14608395099639893,
0.010592637583613396,
-0.1440640091896057,
-0.021400373429059982,
0.050213418900966644,
-0.0685553178191185,
0.02321581169962883,
0.1782781481742859,
0.03464437648653984,
0.05561001971364021,
-0.11280795931816101,
0.057346172630786896,
0.166572704911232,
0.08082842081785202,
0.06835076957941055,
0.03473252058029175,
-0.0036776827182620764,
0.09494494646787643,
0.0952463373541832,
0.01319580152630806,
0.025083716958761215,
0.13967564702033997,
-0.04643873870372772,
0.12644967436790466,
-0.06309179216623306,
-0.008959080092608929,
0.11123735457658768,
-0.028462553396821022,
-0.030151015147566795,
0.0784590020775795,
-0.04807315021753311,
-0.06953224539756775,
-0.13671879470348358,
-0.06719990819692612,
-0.14941096305847168,
0.03391377627849579,
-0.04388823360204697,
-0.003321397816762328,
-0.006885242648422718,
0.05409400165081024,
0.061085741966962814,
0.03724924847483635,
-0.017369167879223824,
-0.04729431867599487,
0.14202633500099182,
-0.017148736864328384,
-0.10454820841550827,
0.014363348484039307,
0.025181658565998077,
0.0470464825630188,
0.02305086888372898,
-0.04195687174797058,
0.07104811072349548,
0.06936251372098923,
-0.00400356063619256,
-0.02012818679213524,
-0.08294057846069336,
-0.020520759746432304,
-0.0007324865437112749,
0.005146214738488197,
0.1315651684999466,
0.08249706029891968,
-0.04633573070168495,
-0.04328155890107155,
0.12504301965236664,
-0.04630393162369728,
-0.01748707890510559,
-0.1259365975856781,
0.014430641196668148,
-0.13596291840076447,
0.029018061235547066,
-0.07382317632436752,
-0.0983690544962883,
-0.044682830572128296,
0.20010647177696228,
0.21507993340492249,
-0.07541535794734955,
0.03557349368929863,
-0.02410084567964077,
-0.0061414302326738834,
0.009276269935071468,
0.08160758018493652,
0.026782028377056122,
0.23757103085517883,
-0.04206646978855133,
-0.01700858399271965,
-0.08236169815063477,
-0.05179218202829361,
-0.042406901717185974,
-0.06910549104213715,
0.024620655924081802,
-0.04555794969201088,
-0.09066620469093323,
0.06677846610546112,
-0.1422184556722641,
-0.13631173968315125,
0.14665105938911438,
-0.1942230761051178,
-0.044230226427316666,
-0.05821848288178444,
0.052333056926727295,
0.039424777030944824,
0.025306180119514465,
-0.07441545277833939,
-0.021235553547739983,
0.0016368570504710078,
-0.008903192356228828,
-0.15969161689281464,
-0.015418264083564281,
-0.08494534343481064,
-0.16938209533691406,
0.07291220873594284,
-0.022724861279129982,
0.06950244307518005,
0.049689505249261856,
0.01920454204082489,
-0.051952362060546875,
0.03333878517150879,
-0.040540024638175964,
-0.07047959417104721,
-0.09333635866641998,
0.09475483745336533,
-0.024311954155564308,
0.08553284406661987,
0.030453944578766823,
-0.09146782010793686,
0.024030776694417,
0.07015589624643326,
-0.09184639155864716,
-0.09294628351926804,
-0.024570293724536896,
-0.07836069911718369,
0.0950201079249382,
0.08755247294902802,
-0.01698891818523407,
-0.005184739362448454,
0.007419266272336245,
0.013515433296561241,
0.012721982784569263,
-0.09620300680398941,
0.03416696563363075,
-0.07623768597841263,
-0.021107295528054237,
-0.0060608647763729095,
0.02621421404182911,
-0.2390073835849762,
-0.07785071432590485,
-0.15479840338230133,
-0.013186748139560223,
-0.01822369359433651,
0.07680783420801163,
0.23109517991542816,
0.0299867931753397,
0.002846379764378071,
-0.16348156332969666,
0.054884523153305054,
0.08326030522584915,
-0.05398024991154671,
-0.05731700360774994
] |
null | null | transformers |
# Model Trained Using AutoTrain
This model was trained using AutoTrain. For more information, please visit [AutoTrain](https://hf.co/docs/autotrain).
# Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_path = "PATH_TO_THIS_REPO"
tokenizer = AutoTokenizer.from_pretrained(model_path)
model = AutoModelForCausalLM.from_pretrained(
model_path,
device_map="auto",
torch_dtype='auto'
).eval()
# Prompt content: "hi"
messages = [
{"role": "user", "content": "hi"}
]
input_ids = tokenizer.apply_chat_template(conversation=messages, tokenize=True, add_generation_prompt=True, return_tensors='pt')
output_ids = model.generate(input_ids.to('cuda'))
response = tokenizer.decode(output_ids[0][input_ids.shape[1]:], skip_special_tokens=True)
# Model response: "Hello! How can I assist you today?"
print(response)
``` | {"license": "other", "tags": ["autotrain", "text-generation"], "widget": [{"text": "I love AutoTrain because "}]} | text-generation | paulux84/autotrain-z58fs-z9tot | [
"transformers",
"tensorboard",
"safetensors",
"mistral",
"text-generation",
"autotrain",
"conversational",
"license:other",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T16:21:47+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #mistral #text-generation #autotrain #conversational #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Trained Using AutoTrain
This model was trained using AutoTrain. For more information, please visit AutoTrain.
# Usage
| [
"# Model Trained Using AutoTrain\n\nThis model was trained using AutoTrain. For more information, please visit AutoTrain.",
"# Usage"
] | [
"TAGS\n#transformers #tensorboard #safetensors #mistral #text-generation #autotrain #conversational #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Trained Using AutoTrain\n\nThis model was trained using AutoTrain. For more information, please visit AutoTrain.",
"# Usage"
] | [
64,
29,
3
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #mistral #text-generation #autotrain #conversational #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Trained Using AutoTrain\n\nThis model was trained using AutoTrain. For more information, please visit AutoTrain.# Usage"
] | [
-0.02699541114270687,
0.052525196224451065,
-0.0015375087969005108,
0.05149228125810623,
0.12088289856910706,
-0.04163394495844841,
0.2625403106212616,
0.05558646097779274,
-0.04610184207558632,
-0.06594274193048477,
0.17726024985313416,
0.18847188353538513,
-0.03057561069726944,
0.1602223962545395,
-0.04588054493069649,
-0.2628976106643677,
0.01935308612883091,
0.002317485399544239,
0.0524492971599102,
0.11028582602739334,
0.1389206349849701,
-0.07855317741632462,
0.0668196976184845,
0.041918426752090454,
-0.21349984407424927,
0.03016752563416958,
0.0596829317510128,
-0.12534905970096588,
0.17129358649253845,
0.07529635727405548,
0.117461197078228,
0.060434285551309586,
0.10621046274900436,
-0.09247836470603943,
0.031026212498545647,
0.01950329728424549,
-0.020604576915502548,
0.09757140278816223,
0.07917279005050659,
-0.0350341871380806,
0.07296935468912125,
0.14127607643604279,
0.10844720155000687,
0.039640773087739944,
-0.09329894930124283,
0.02944805659353733,
0.0020703261252492666,
0.022427711635828018,
0.11719834059476852,
0.09943539649248123,
-0.01877782493829727,
0.18181537091732025,
-0.08212538808584213,
0.10259909927845001,
-0.045457057654857635,
-0.28803253173828125,
-0.03986995294690132,
0.1873147338628769,
0.07420988380908966,
0.013626144267618656,
-0.12427637726068497,
0.078909732401371,
0.1008378192782402,
-0.004938090685755014,
0.0701189786195755,
-0.033190663903951645,
-0.0393044613301754,
-0.01346403919160366,
-0.09974315762519836,
0.05136822164058685,
0.16537895798683167,
-0.07599294185638428,
-0.009413180872797966,
-0.12710975110530853,
-0.05052062124013901,
0.021240994334220886,
-0.011423724703490734,
-0.12030315399169922,
-0.03188876062631607,
0.08303900808095932,
-0.05930151045322418,
-0.0464908592402935,
-0.14200998842716217,
-0.04042978212237358,
-0.11552562564611435,
0.057527095079422,
-0.006070165429264307,
-0.016220789402723312,
-0.12115611881017685,
0.09469593316316605,
0.03101532533764839,
-0.11626887321472168,
0.060420598834753036,
-0.0741722360253334,
0.001140298554673791,
-0.10681776702404022,
-0.023332959040999413,
-0.1952218860387802,
0.030099496245384216,
0.17851774394512177,
0.15494675934314728,
0.020939888432621956,
-0.07886002957820892,
0.048338331282138824,
0.013530293479561806,
0.12733060121536255,
0.02140798792243004,
-0.046359043568372726,
0.05125835910439491,
-0.04296070709824562,
-0.009869301691651344,
-0.0485469251871109,
-0.20714326202869415,
0.05008949339389801,
0.03269507363438606,
0.03291710838675499,
-0.04729897156357765,
0.1083824411034584,
-0.028031980618834496,
0.016060568392276764,
0.033669810742139816,
-0.03928220644593239,
0.039990223944187164,
-0.050537772476673126,
0.008731657639145851,
-0.03720587119460106,
0.04612233489751816,
0.10477311909198761,
0.021429337561130524,
0.10337132215499878,
-0.07821279764175415,
-0.02858932502567768,
-0.09844601899385452,
-0.08640087395906448,
0.01666846126317978,
0.011332022026181221,
0.03232283517718315,
-0.18575049936771393,
-0.23881326615810394,
-0.024755913764238358,
0.06387589126825333,
-0.03176957741379738,
-0.06939536333084106,
-0.09116985648870468,
0.008387977257370949,
0.06912240386009216,
-0.02539674937725067,
0.06623683124780655,
-0.004402702208608389,
0.020224828273057938,
-0.07790552824735641,
0.038042839616537094,
-0.06866512447595596,
0.015649111941456795,
-0.11430995166301727,
-0.022092711180448532,
-0.03882748633623123,
0.04329218342900276,
-0.03006068430840969,
0.14419233798980713,
-0.03842197358608246,
0.04526497796177864,
-0.044787291437387466,
0.056167472153902054,
0.02476482465863228,
0.1392175555229187,
-0.16037936508655548,
-0.043978430330753326,
0.14509063959121704,
-0.10600236058235168,
-0.11773402988910675,
0.12836679816246033,
-0.09591180086135864,
0.24804648756980896,
0.1129641905426979,
0.11147007346153259,
0.05363072082400322,
-0.08599360287189484,
0.10796362906694412,
0.030262673273682594,
-0.08265260607004166,
-0.029414139688014984,
-0.007092099171131849,
0.008056282065808773,
-0.1482926309108734,
0.02850986458361149,
0.14341646432876587,
0.06641319394111633,
-0.039340823888778687,
-0.07952730357646942,
-0.011220400221645832,
-0.04146914929151535,
0.08381195366382599,
0.012549158185720444,
0.14203421771526337,
-0.04933689907193184,
-0.04913952201604843,
0.08753503113985062,
0.02959383651614189,
0.01807376556098461,
-0.033537544310092926,
-0.09085430204868317,
-0.024506034329533577,
-0.05371454358100891,
0.021011503413319588,
-0.11492293328046799,
-0.05873663350939751,
-0.027979671955108643,
0.09379857778549194,
0.05248923599720001,
0.12190566211938858,
0.0553957037627697,
0.03518616035580635,
-0.016852382570505142,
0.03468210995197296,
0.19218511879444122,
0.03407968580722809,
-0.1401698738336563,
-0.14350873231887817,
0.11236314475536346,
-0.09722626954317093,
0.1408296823501587,
-0.24620012938976288,
0.040916357189416885,
-0.06338175386190414,
0.09650067985057831,
0.01381456758826971,
0.06793159246444702,
-0.07225442677736282,
0.016335727646946907,
-0.11225221306085587,
-0.0067980182357132435,
0.0727112665772438,
0.047955222427845,
-0.07784378528594971,
0.1471128612756729,
-0.18932963907718658,
0.22249622642993927,
0.1227576732635498,
-0.13205888867378235,
-0.09022273123264313,
-0.09933790564537048,
0.007844572886824608,
-0.024010734632611275,
-0.07395610213279724,
-0.011356044560670853,
0.13247385621070862,
-0.02462097257375717,
0.1890343278646469,
-0.02045755833387375,
-0.04548988863825798,
-0.001989134354516864,
-0.09136506915092468,
-0.017559170722961426,
0.031994812190532684,
0.06634363532066345,
-0.22326712310314178,
0.13752993941307068,
0.1406964212656021,
-0.05221952497959137,
0.22264766693115234,
0.032917723059654236,
0.026482457295060158,
0.003446560585871339,
-0.049891844391822815,
0.020148031413555145,
-0.009107460267841816,
-0.06893884390592575,
-0.049932919442653656,
0.014478646218776703,
-0.0142153175547719,
0.03317056596279144,
-0.12587253749370575,
-0.01182317454367876,
0.026522744446992874,
0.0580943338572979,
0.06508933752775192,
0.06177165359258652,
-0.08519607782363892,
0.08249333500862122,
-0.04087919741868973,
-0.14336703717708588,
0.1108337789773941,
0.009114470332860947,
-0.12669003009796143,
0.18299083411693573,
-0.0758049413561821,
-0.23183932900428772,
-0.18516750633716583,
-0.17221717536449432,
-0.00723571190610528,
0.062238529324531555,
0.059693045914173126,
-0.07341434806585312,
-0.06569252163171768,
-0.020864974707365036,
-0.06828463077545166,
0.01007883995771408,
-0.002030706498771906,
-0.06883236765861511,
0.03270689398050308,
-0.014208085834980011,
-0.10116046667098999,
-0.029029831290245056,
-0.009190062060952187,
-0.04901104420423508,
0.0692099779844284,
-0.052616529166698456,
0.060191500931978226,
0.18694517016410828,
-0.01489932555705309,
0.023964669555425644,
-0.04143443703651428,
0.12997864186763763,
-0.07899091392755508,
0.027894146740436554,
0.08909832686185837,
-0.07445263862609863,
0.03881121054291725,
0.20737601816654205,
0.029496803879737854,
-0.05632537975907326,
0.07801061868667603,
-0.028718801215291023,
-0.0774407610297203,
-0.20642608404159546,
-0.08684719353914261,
-0.020633254200220108,
0.019245212897658348,
0.08364695310592651,
0.057998914271593094,
0.26918041706085205,
0.12964649498462677,
0.07459327578544617,
0.06608638167381287,
0.04691487178206444,
0.08669880777597427,
0.13618479669094086,
-0.02930464781820774,
0.1926214098930359,
-0.07248775660991669,
-0.17590169608592987,
0.04103903844952583,
-0.019386671483516693,
0.0760020986199379,
0.15984705090522766,
0.053699299693107605,
0.029926501214504242,
0.05818108469247818,
0.14610257744789124,
0.11831671744585037,
0.05941009148955345,
-0.06771506369113922,
-0.008292550221085548,
-0.025469649583101273,
-0.02513192594051361,
0.11616689711809158,
-0.031084444373846054,
-0.0541393868625164,
-0.003292324487119913,
0.06276321411132812,
0.03913247585296631,
0.08445098996162415,
0.011915687471628189,
-0.2954202890396118,
0.010847505182027817,
0.035310279577970505,
-0.05624374374747276,
-0.08180417120456696,
0.0618964284658432,
0.019216833636164665,
-0.14484640955924988,
0.0181722529232502,
-0.06784247606992722,
0.08817319571971893,
-0.008300479501485825,
0.06697096675634384,
-0.06832937151193619,
-0.007470566779375076,
-0.04647418111562729,
0.12014855444431305,
-0.38092684745788574,
0.2078399658203125,
-0.006222316529601812,
0.0456739105284214,
-0.11163492500782013,
0.020704366266727448,
0.08321543782949448,
0.15837235748767853,
0.09556505084037781,
-0.0660763531923294,
-0.14684933423995972,
-0.09546715766191483,
-0.10107629001140594,
-0.014514556154608727,
0.02460043877363205,
-0.01051158644258976,
0.012242973782122135,
-0.10082778334617615,
-0.013952648267149925,
0.059524063020944595,
-0.008193613961338997,
-0.13010439276695251,
-0.17529445886611938,
0.021371234208345413,
0.04099329933524132,
0.03484678268432617,
-0.020617537200450897,
-0.10272130370140076,
-0.11836910247802734,
0.1934756636619568,
0.006529111415147781,
-0.002373364521190524,
-0.13872013986110687,
-0.012368431314826012,
-0.055968910455703735,
-0.04413953423500061,
0.06175634264945984,
0.00045833707554265857,
0.1418440192937851,
-0.08526213467121124,
-0.09989900887012482,
0.12388047575950623,
-0.11607690155506134,
-0.05472702533006668,
-0.11755654960870743,
0.020144371315836906,
-0.012329724617302418,
0.0054936837404966354,
0.0963820368051529,
0.03955746069550514,
-0.07033362984657288,
-0.05635528638958931,
-0.016631219536066055,
-0.011973624117672443,
-0.030847571790218353,
-0.13556155562400818,
-0.09645826369524002,
-0.11676975339651108,
-0.021463563665747643,
-0.09085120260715485,
0.2225944697856903,
0.1282944530248642,
-0.09359250962734222,
0.13654381036758423,
0.2120448350906372,
-0.1091618612408638,
-0.3224632143974304,
-0.04555538296699524,
-0.031316548585891724,
0.015949100255966187,
0.044650688767433167,
-0.12177825719118118,
0.10702518373727798,
-0.004046852700412273,
-0.07988344132900238,
-0.02333424612879753,
-0.15922191739082336,
-0.17143768072128296,
0.2526555061340332,
0.02667848765850067,
0.21582216024398804,
-0.1036934107542038,
-0.06667744368314743,
-0.12322782725095749,
0.033653486520051956,
0.07333366572856903,
-0.12631624937057495,
0.06262853741645813,
0.0570843331515789,
0.0673443153500557,
0.029383040964603424,
-0.0368652381002903,
0.0476229153573513,
-0.051926832646131516,
0.05953340232372284,
-0.1630999594926834,
-0.03829682990908623,
0.06113021448254585,
-0.02620183303952217,
0.08101767301559448,
-0.06936794519424438,
0.011851944960653782,
-0.012223218567669392,
-0.06766347587108612,
0.029848668724298477,
0.07275693863630295,
-0.001643786090426147,
-0.11137469112873077,
0.005139535292983055,
-0.010434017516672611,
-0.0038823471404612064,
-0.05620753765106201,
0.09866064786911011,
-0.03227032721042633,
0.1555190235376358,
0.2008318305015564,
0.24337852001190186,
-0.06256049871444702,
0.06945628672838211,
-0.02731485851109028,
-0.11183883994817734,
0.08404701948165894,
-0.07549898326396942,
0.02428785152733326,
0.0827280580997467,
-0.040800586342811584,
0.15702922642230988,
0.0655057430267334,
-0.013254620134830475,
-0.005723502021282911,
0.14080925285816193,
-0.1530471295118332,
-0.0159621424973011,
-0.06955518573522568,
0.10592132806777954,
0.02018340304493904,
0.026091108098626137,
0.14143383502960205,
-0.07682769000530243,
-0.009049240499734879,
0.04147037863731384,
-0.003871590131893754,
-0.0168798565864563,
0.11431785672903061,
0.035309236496686935,
0.019573885947465897,
-0.07142370939254761,
0.02572551742196083,
0.07623033225536346,
-0.015167484059929848,
0.0498647503554821,
0.03143294155597687,
-0.09164922684431076,
-0.11104477196931839,
0.022583844140172005,
0.2792930006980896,
-0.1597553789615631,
-0.0767328143119812,
-0.03942285478115082,
-0.12160889804363251,
0.014694852754473686,
0.11444544792175293,
0.0762668251991272,
0.03335021063685417,
-0.08444573730230331,
-0.01821875013411045,
-0.11082186549901962,
0.11144796013832092,
0.020362544804811478,
0.06067124754190445,
-0.14576083421707153,
0.08216147124767303,
-0.026234211400151253,
-0.004679789766669273,
-0.09952213615179062,
-0.031425606459379196,
-0.13179950416088104,
0.013805066235363483,
-0.1193600445985794,
-0.034559521824121475,
-0.044988371431827545,
-0.02029399573802948,
0.05154072120785713,
-0.012474157847464085,
-0.02804945409297943,
-0.03174898400902748,
-0.08728643506765366,
0.03447297215461731,
0.00530976289883256,
0.0416826456785202,
-0.04762420803308487,
-0.022761870175600052,
0.01943979226052761,
-0.005551697686314583,
0.0724945217370987,
-0.00610021548345685,
-0.023790907114744186,
0.042491160333156586,
-0.19401361048221588,
0.0373309850692749,
0.07397368550300598,
0.0035578045062720776,
0.03805038332939148,
-0.022769832983613014,
-0.012226573191583157,
0.08592986315488815,
0.05349747836589813,
0.044497326016426086,
0.00602300651371479,
-0.08010756224393845,
0.06233002617955208,
0.028811337426304817,
-0.11983266472816467,
-0.04654451459646225,
-0.0010786387138068676,
-0.008851143531501293,
-0.01833510212600231,
0.21574845910072327,
-0.12357861548662186,
0.04714543744921684,
-0.06649184226989746,
0.03950250893831253,
-0.01034714374691248,
-0.1332462728023529,
-0.07150128483772278,
-0.11829763650894165,
-0.018726184964179993,
-0.0180298313498497,
0.22265899181365967,
0.13177891075611115,
-0.026160845533013344,
0.03662189468741417,
0.065272256731987,
0.05387585982680321,
0.01899443380534649,
0.18585743010044098,
0.09506116062402725,
0.006766694597899914,
-0.13416342437267303,
0.07946646958589554,
0.04529587924480438,
-0.0195454154163599,
0.02385331131517887,
-0.009406870231032372,
-0.10800548642873764,
0.08788439631462097,
0.06324751675128937,
0.0019808029755949974,
-0.08126944303512573,
-0.07060925662517548,
-0.13577310740947723,
0.05875951424241066,
-0.0824219211935997,
0.0296943336725235,
0.18155713379383087,
-0.0420151986181736,
-0.022988205775618553,
-0.048243507742881775,
-0.061702877283096313,
-0.21090053021907806,
-0.19175760447978973,
-0.11900339275598526,
-0.06937247514724731,
0.009304484352469444,
-0.04402853175997734,
0.040969494730234146,
0.012307711876928806,
0.07516354322433472,
-0.05410514771938324,
0.0900542289018631,
-0.06605858355760574,
0.0025424438063055277,
0.029864680022001266,
-0.048153914511203766,
0.0048726750537753105,
-0.19374705851078033,
-0.004350534174591303,
-0.12837374210357666,
0.04954437166452408,
-0.03329106420278549,
-0.026907531544566154,
0.023711740970611572,
0.007632098160684109,
-0.029052425175905228,
-0.049126382917165756,
-0.0118684908375144,
0.037363916635513306,
0.009906113147735596,
0.03080584853887558,
0.013063051737844944,
-0.009350288659334183,
0.0260739978402853,
0.18832696974277496,
-0.04662957787513733,
-0.15130199491977692,
-0.14174605906009674,
0.20150485634803772,
-0.003776230849325657,
0.12395143508911133,
-0.0641348734498024,
-0.020259689539670944,
0.03532780706882477,
0.30355772376060486,
0.28879380226135254,
-0.05663159862160683,
0.014149083755910397,
-0.002601712942123413,
-0.004551102872937918,
-0.01731131225824356,
0.15413738787174225,
0.0075277178548276424,
0.17758780717849731,
-0.05308525264263153,
0.029151551425457,
-0.009426014497876167,
-0.09109172224998474,
-0.0567401684820652,
0.1175808534026146,
0.004690850619226694,
-0.0003956313303206116,
-0.018972044810652733,
0.07927758991718292,
-0.09866127371788025,
0.09541647881269455,
-0.1547948718070984,
-0.06571374088525772,
-0.07504004240036011,
0.03635769337415695,
0.1304396241903305,
-0.017138052731752396,
0.046681176871061325,
-0.020876076072454453,
-0.009447058662772179,
0.027952050790190697,
-0.04140012711286545,
-0.08449426293373108,
0.02128974162042141,
0.06701259315013885,
-0.06004859134554863,
0.18278516829013824,
-0.027691975235939026,
0.048137884587049484,
0.0995766744017601,
0.003908407874405384,
-0.08403819054365158,
0.06885812431573868,
-0.008872791193425655,
-0.05479026958346367,
0.12772956490516663,
0.008756602182984352,
0.012023258022964,
-0.009214076213538647,
0.019285310059785843,
-0.1543876826763153,
0.1239546611905098,
-0.1478824019432068,
-0.08811484277248383,
-0.04636934772133827,
0.10732483863830566,
-0.02003641426563263,
0.15155507624149323,
0.0726822093129158,
-0.010221063159406185,
0.01297276746481657,
-0.019649337977170944,
0.03318748623132706,
-0.022814597934484482,
-0.08245015144348145,
-0.043602608144283295,
-0.2061813473701477,
-0.020260902121663094,
0.035927873104810715,
-0.012469958513975143,
-0.2885707914829254,
-0.07848802208900452,
-0.09369978308677673,
-0.03648434951901436,
-0.13286547362804413,
0.06016679108142853,
0.23642539978027344,
0.03423815220594406,
-0.01795719936490059,
-0.08926072716712952,
-0.010259761475026608,
0.0424993596971035,
-0.06064554303884506,
-0.09056360274553299
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# gpt-finetuned-wikitext2
This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 6.3613
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| 6.7968 | 1.0 | 1123 | 6.6783 |
| 6.4613 | 2.0 | 2246 | 6.4416 |
| 6.3067 | 3.0 | 3369 | 6.3613 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "mit", "tags": ["generated_from_trainer"], "base_model": "gpt2", "model-index": [{"name": "gpt-finetuned-wikitext2", "results": []}]} | text-generation | GavinQiangLi/gpt-finetuned-wikitext2 | [
"transformers",
"tensorboard",
"safetensors",
"gpt2",
"text-generation",
"generated_from_trainer",
"base_model:gpt2",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T16:25:16+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #gpt2 #text-generation #generated_from_trainer #base_model-gpt2 #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| gpt-finetuned-wikitext2
=======================
This model is a fine-tuned version of gpt2 on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 6.3613
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3.0
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #gpt2 #text-generation #generated_from_trainer #base_model-gpt2 #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
72,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #gpt2 #text-generation #generated_from_trainer #base_model-gpt2 #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.09056128561496735,
0.06809527426958084,
-0.0025528871919959784,
0.10096219927072525,
0.1381831020116806,
0.017848413437604904,
0.16674168407917023,
0.12242098152637482,
-0.07412248104810715,
0.05490725859999657,
0.1460796445608139,
0.12613141536712646,
0.02924579195678234,
0.13790875673294067,
-0.06820142269134521,
-0.23379817605018616,
0.006077175494283438,
0.02917812392115593,
-0.05179181322455406,
0.12489522248506546,
0.09390316158533096,
-0.12173109501600266,
0.09821289777755737,
-0.011401938274502754,
-0.1863936483860016,
0.0021493518725037575,
0.02595284767448902,
-0.05243686959147453,
0.13893486559391022,
0.04182766377925873,
0.11320293694734573,
0.03709891438484192,
0.07697579264640808,
-0.18702448904514313,
0.0141896428540349,
0.06087725609540939,
-0.00549273518845439,
0.08726999908685684,
0.04951661452651024,
-0.000012706768757198006,
0.11250170320272446,
-0.08061284571886063,
0.05145649611949921,
0.017947662621736526,
-0.1429160088300705,
-0.19648215174674988,
-0.08166409283876419,
0.028207262977957726,
0.09511610120534897,
0.08330243080854416,
-0.02246807888150215,
0.12686599791049957,
-0.03957941755652428,
0.09657186269760132,
0.23117992281913757,
-0.32891371846199036,
-0.06398431956768036,
0.06295526772737503,
0.05190695822238922,
0.09522278606891632,
-0.09870574623346329,
0.009848032146692276,
0.061045870184898376,
0.025770969688892365,
0.1262919008731842,
-0.025607628747820854,
-0.013934186659753323,
0.01095869392156601,
-0.1446661651134491,
-0.032961372286081314,
0.15285944938659668,
0.040089432150125504,
-0.043372850865125656,
-0.06932097673416138,
-0.07136590033769608,
-0.1448194831609726,
-0.03480764105916023,
-0.019542895257472992,
0.03927380219101906,
-0.016807537525892258,
-0.08658316731452942,
-0.04665970057249069,
-0.11186057329177856,
-0.08426578342914581,
-0.05129887908697128,
0.13720689713954926,
0.032003648579120636,
0.004749617073684931,
-0.02339453622698784,
0.11595379561185837,
-0.052253127098083496,
-0.13509170711040497,
0.004059792496263981,
0.025129398331046104,
0.013984785415232182,
-0.047912776470184326,
-0.048950932919979095,
-0.10673307627439499,
0.021154658868908882,
0.13801515102386475,
-0.060942381620407104,
0.05338501185178757,
0.006936258636415005,
0.049335777759552,
-0.0995190367102623,
0.1683686375617981,
-0.037785619497299194,
-0.024068692699074745,
0.019097251817584038,
0.07289014756679535,
0.051110316067934036,
-0.02285294234752655,
-0.1341378390789032,
0.021184759214520454,
0.10754746943712234,
0.020990097895264626,
-0.0561378076672554,
0.08088257908821106,
-0.04687834531068802,
-0.004793973173946142,
0.024190567433834076,
-0.08290587365627289,
0.023597650229930878,
-0.0057591903023421764,
-0.054636456072330475,
-0.06249101087450981,
0.021960048004984856,
0.021696727722883224,
0.00041165912989526987,
0.10452143102884293,
-0.0860556811094284,
0.009091547690331936,
-0.08142657577991486,
-0.12781862914562225,
0.01262646447867155,
-0.07894767075777054,
0.012721879407763481,
-0.10412304848432541,
-0.19083257019519806,
-0.012429673224687576,
0.049482427537441254,
-0.036995548754930496,
-0.029123913496732712,
-0.06562601029872894,
-0.07668423652648926,
0.01880665309727192,
-0.02080162614583969,
0.08955041319131851,
-0.06315923482179642,
0.10288941115140915,
0.044523194432258606,
0.07134629040956497,
-0.055453114211559296,
0.031224871054291725,
-0.09264518320560455,
0.0354946069419384,
-0.17711801826953888,
0.03808043897151947,
-0.04268611595034599,
0.05735969543457031,
-0.080870121717453,
-0.07864441722631454,
-0.010458719916641712,
0.006909788120537996,
0.07786839455366135,
0.10634377598762512,
-0.15960800647735596,
-0.07473272830247879,
0.19582629203796387,
-0.08860130608081818,
-0.14866942167282104,
0.14246708154678345,
-0.05571338161826134,
0.05093294754624367,
0.07529725879430771,
0.2063729166984558,
0.041722193360328674,
-0.10098408162593842,
0.008962782099843025,
-0.004913205746561289,
0.04729123413562775,
-0.03098015859723091,
0.07477094233036041,
-0.0049115996807813644,
0.01690625213086605,
0.012815915048122406,
-0.04358939453959465,
0.04775530844926834,
-0.08757823705673218,
-0.07703150808811188,
-0.036710213869810104,
-0.09226883202791214,
0.047793589532375336,
0.04382869973778725,
0.07111519575119019,
-0.12602275609970093,
-0.09935164451599121,
0.04123856872320175,
0.0594940111041069,
-0.08133374154567719,
0.025141526013612747,
-0.07108814269304276,
0.09455830603837967,
-0.07761933654546738,
-0.011343000456690788,
-0.13599015772342682,
-0.06046932190656662,
0.012660862877964973,
0.02043353021144867,
0.022248849272727966,
0.00944613665342331,
0.08080432564020157,
0.08685511350631714,
-0.06672825664281845,
-0.021333543583750725,
-0.002270165365189314,
0.000021162144548725337,
-0.13053777813911438,
-0.19118475914001465,
-0.00891368743032217,
-0.031045112758874893,
0.13987098634243011,
-0.24124111235141754,
0.05555005371570587,
0.017107510939240456,
0.07744037359952927,
0.03286753594875336,
-0.026775531470775604,
-0.03544329106807709,
0.05084521323442459,
-0.05110393837094307,
-0.06824219226837158,
0.06252512335777283,
0.010483325459063053,
-0.10839737951755524,
-0.01911192014813423,
-0.18453916907310486,
0.19159196317195892,
0.1364714652299881,
-0.08574911206960678,
-0.08612702041864395,
-0.0077362083829939365,
-0.04206875339150429,
-0.021902862936258316,
-0.04470087215304375,
-0.011150618083775043,
0.14737637341022491,
-0.00892745889723301,
0.15799827873706818,
-0.09010282903909683,
-0.04528059810400009,
0.028837129473686218,
-0.05360740050673485,
0.014661007560789585,
0.11146004498004913,
0.08118748664855957,
-0.10671050101518631,
0.15006029605865479,
0.1602982133626938,
-0.06470346450805664,
0.15555226802825928,
-0.027200933545827866,
-0.0553949698805809,
-0.02943415939807892,
0.025611108168959618,
0.017484324052929878,
0.10411196202039719,
-0.1318131685256958,
-0.009252929128706455,
0.007513432763516903,
0.021311478689312935,
0.02193121425807476,
-0.22756941616535187,
-0.03885500878095627,
0.045263830572366714,
-0.0621686689555645,
0.0011787699768319726,
-0.01090402901172638,
-0.01659265346825123,
0.10500795394182205,
0.008806956000626087,
-0.06349760293960571,
0.042197126895189285,
-0.0018732607131823897,
-0.08807819336652756,
0.21039322018623352,
-0.06608946621417999,
-0.15328145027160645,
-0.13332532346248627,
-0.07209832221269608,
-0.05359160527586937,
0.03819100931286812,
0.06818952411413193,
-0.0816270112991333,
-0.035627465695142746,
-0.11244255304336548,
0.02511731907725334,
0.008616508916020393,
0.023714367300271988,
0.008451751433312893,
-0.01478289533406496,
0.0575849786400795,
-0.1049032136797905,
-0.01632842794060707,
-0.0474519319832325,
-0.06806967407464981,
0.04590265825390816,
0.01793626695871353,
0.11110923439264297,
0.1528436243534088,
-0.023843152448534966,
0.01168255228549242,
-0.042298417538404465,
0.23634184896945953,
-0.07910632342100143,
-0.016219785436987877,
0.1260772943496704,
-0.016036992892622948,
0.055155858397483826,
0.1237182542681694,
0.05677172914147377,
-0.10750941187143326,
0.019709844142198563,
0.026917580515146255,
-0.050928495824337006,
-0.20891758799552917,
-0.022129304707050323,
-0.042331814765930176,
0.019493846222758293,
0.08914721757173538,
0.03668850660324097,
0.03725428506731987,
0.07465830445289612,
0.015163272619247437,
0.08140446245670319,
0.00318527570925653,
0.08147349208593369,
0.10528820008039474,
0.03734028339385986,
0.13583609461784363,
-0.046897053718566895,
-0.06294741481542587,
0.04246792197227478,
-0.0036438447423279285,
0.20714810490608215,
0.01751178503036499,
0.15144850313663483,
0.05160778388381004,
0.13809625804424286,
0.006622801534831524,
0.060332540422677994,
-0.017757073044776917,
-0.048486772924661636,
-0.016539573669433594,
-0.05041869357228279,
-0.022022662684321404,
0.03814598172903061,
-0.08572392165660858,
0.03506161645054817,
-0.10535914450883865,
0.003437948180362582,
0.05607115477323532,
0.20601679384708405,
0.062258031219244,
-0.33750119805336,
-0.0905357152223587,
0.03155820444226265,
-0.01795387454330921,
-0.02763739414513111,
0.026251671835780144,
0.13911432027816772,
-0.06257518380880356,
0.04603578895330429,
-0.08266576379537582,
0.08290748298168182,
-0.04460844770073891,
0.05276663973927498,
0.04142896831035614,
0.08337219059467316,
-0.026498882099986076,
0.07271967828273773,
-0.2894154489040375,
0.27676719427108765,
0.018017452210187912,
0.07435306161642075,
-0.055272504687309265,
0.0007471432909369469,
0.018972313031554222,
0.07249210774898529,
0.08417865633964539,
-0.02012914977967739,
-0.06388355791568756,
-0.1924000382423401,
-0.0549270398914814,
0.030810823664069176,
0.11773815751075745,
-0.04370654746890068,
0.11185532063245773,
-0.02936672605574131,
0.007648822385817766,
0.07989225536584854,
-0.010275938548147678,
-0.06808256357908249,
-0.09898540377616882,
0.002041063504293561,
0.02978076972067356,
-0.022048218175768852,
-0.07790572941303253,
-0.09076252579689026,
-0.13601510226726532,
0.17920134961605072,
-0.047778528183698654,
-0.04375506937503815,
-0.1043119728565216,
0.07061570137739182,
0.045220889151096344,
-0.08187993615865707,
0.035630207508802414,
0.011203563772141933,
0.08495581150054932,
0.015212233178317547,
-0.056484006345272064,
0.13421152532100677,
-0.062260087579488754,
-0.18268002569675446,
-0.06536144763231277,
0.12325260043144226,
0.015568505972623825,
0.04536212980747223,
0.0017954764189198613,
0.018502365797758102,
-0.027797820046544075,
-0.08970554172992706,
0.047019112855196,
-0.017695406451821327,
0.04656846076250076,
0.008819486945867538,
-0.027270570397377014,
0.022180253639817238,
-0.05409298092126846,
-0.04043448716402054,
0.17282778024673462,
0.2988189160823822,
-0.078043632209301,
0.01904793083667755,
0.04385428875684738,
-0.07089382410049438,
-0.2010335475206375,
0.03851896896958351,
0.018006207421422005,
0.004344097338616848,
0.04399720951914787,
-0.1487164944410324,
0.07331637293100357,
0.10179731994867325,
-0.027309533208608627,
0.12998180091381073,
-0.31497567892074585,
-0.14281047880649567,
0.10411899536848068,
0.1542544811964035,
0.13861362636089325,
-0.17136944830417633,
-0.04761020466685295,
-0.03401949629187584,
-0.11041120439767838,
0.1071501299738884,
-0.12765151262283325,
0.12554749846458435,
-0.01118628028780222,
0.07108836621046066,
0.011067335493862629,
-0.06272457540035248,
0.12952494621276855,
-0.0145385367795825,
0.1060604378581047,
-0.07478170841932297,
-0.010769554413855076,
0.0672825500369072,
-0.053607914596796036,
0.025783395394682884,
-0.12371992319822311,
0.029032545164227486,
-0.05784471333026886,
-0.038851335644721985,
-0.05079234763979912,
0.03554591163992882,
-0.032128386199474335,
-0.07073725014925003,
-0.0500924326479435,
0.011431172490119934,
0.03173641487956047,
-0.008821782656013966,
0.15905970335006714,
-0.0052466788329184055,
0.16920673847198486,
0.1288958340883255,
0.0797029584646225,
-0.07538238167762756,
-0.013774005696177483,
-0.0014004242839291692,
-0.029799221083521843,
0.05267246067523956,
-0.1646360158920288,
0.03039737232029438,
0.12231360375881195,
0.012851564213633537,
0.14915534853935242,
0.08203063160181046,
-0.04793400317430496,
0.036409731954336166,
0.06678934395313263,
-0.17571373283863068,
-0.13189390301704407,
-0.017248131334781647,
-0.049651842564344406,
-0.10032529383897781,
0.07216271758079529,
0.13144557178020477,
-0.0668412446975708,
0.008448099717497826,
-0.012336189858615398,
0.016800418496131897,
-0.03731068596243858,
0.18644404411315918,
0.05279620736837387,
0.04741140455007553,
-0.084463931620121,
0.06486725062131882,
0.03883186727762222,
-0.06267134100198746,
0.02677386812865734,
0.059039585292339325,
-0.07989566028118134,
-0.04208127036690712,
0.03258395195007324,
0.19104057550430298,
-0.046127211302518845,
-0.040237218141555786,
-0.15060527622699738,
-0.120575912296772,
0.05237482488155365,
0.16726326942443848,
0.08303184062242508,
0.016621584072709084,
-0.031216012313961983,
0.02204933390021324,
-0.12268561869859695,
0.1078961193561554,
0.040421389043331146,
0.08812212944030762,
-0.14250284433364868,
0.14692163467407227,
-0.004924541339278221,
0.014427771791815758,
-0.03082052245736122,
0.03878059983253479,
-0.12027952820062637,
-0.0015994077548384666,
-0.1282837837934494,
-0.017664724960923195,
-0.027912219986319542,
-0.007076366804540157,
-0.00546577712520957,
-0.05050886049866676,
-0.06119975075125694,
0.007058986462652683,
-0.10402026027441025,
-0.02127249725162983,
0.02607150375843048,
0.036142077296972275,
-0.12331508100032806,
-0.03630498796701431,
0.015352361835539341,
-0.06681694090366364,
0.08183576911687851,
0.021490009501576424,
0.020336482673883438,
0.054485443979501724,
-0.19135692715644836,
0.039559513330459595,
0.059929076582193375,
0.0028611598536372185,
0.03701378405094147,
-0.06787173449993134,
-0.01427218783646822,
-0.0071058799512684345,
0.0509757399559021,
0.031155381351709366,
0.0721600204706192,
-0.12318484485149384,
0.018655413761734962,
-0.028755316510796547,
-0.060507699847221375,
-0.054549191147089005,
0.042401108890771866,
0.05466369912028313,
0.010212392546236515,
0.1842985451221466,
-0.10302137583494186,
0.02336595207452774,
-0.21337299048900604,
0.005976200569421053,
0.013741832226514816,
-0.12292787432670593,
-0.10181016474962234,
-0.05570589005947113,
0.06211017072200775,
-0.057244278490543365,
0.1531209796667099,
0.021311698481440544,
0.019158506765961647,
0.034917816519737244,
-0.018132267519831657,
0.06269442290067673,
0.013808820396661758,
0.22580699622631073,
0.03301788121461868,
-0.048795778304338455,
0.009688337333500385,
0.044825758785009384,
0.11917158961296082,
0.058472469449043274,
0.18908384442329407,
0.13387775421142578,
-0.05053223296999931,
0.10243424028158188,
0.03938594087958336,
-0.04970654845237732,
-0.151210755109787,
0.048389703035354614,
-0.04466920718550682,
0.09982673823833466,
-0.021789032965898514,
0.19295869767665863,
0.12047889083623886,
-0.15374812483787537,
0.008750703185796738,
-0.04772496223449707,
-0.08284588158130646,
-0.12173067778348923,
-0.08467255532741547,
-0.0985812395811081,
-0.1507975161075592,
0.007698882836848497,
-0.11666832119226456,
0.030272284522652626,
0.10649031400680542,
0.01646130532026291,
-0.01728101633489132,
0.18084026873111725,
0.02077774703502655,
0.02425394020974636,
0.05277477949857712,
-0.00184342497959733,
-0.025302186608314514,
-0.07665886729955673,
-0.08354206383228302,
0.00913175754249096,
-0.019609013572335243,
0.03542591258883476,
-0.03976646065711975,
-0.024681102484464645,
0.03596204146742821,
-0.016014715656638145,
-0.10675211995840073,
-0.0006488219951279461,
0.045256998389959335,
0.05810665339231491,
0.03528144955635071,
0.006442788057029247,
-0.0003661039227154106,
-0.00474885618314147,
0.22563748061656952,
-0.07323932647705078,
-0.060973621904850006,
-0.08023044466972351,
0.22699740529060364,
0.02163662202656269,
0.015002545900642872,
0.0035311190877109766,
-0.09128163754940033,
0.03555387631058693,
0.23281505703926086,
0.2059980183839798,
-0.0837680846452713,
0.0012300063390284777,
-0.014407549053430557,
-0.005087652709335089,
-0.01591895893216133,
0.10233580321073532,
0.09856351464986801,
0.049174368381500244,
-0.07895512878894806,
-0.03856958821415901,
-0.02771371603012085,
-0.002243949566036463,
-0.041175276041030884,
0.06611569225788116,
0.04436876252293587,
0.019829751923680305,
-0.04603514075279236,
0.06648112833499908,
-0.04525080323219299,
-0.09842053055763245,
0.02587287127971649,
-0.2030579000711441,
-0.13890999555587769,
-0.002745509147644043,
0.10673332214355469,
-0.01768561452627182,
0.06437819451093674,
-0.0334940105676651,
0.0010927028488367796,
0.04496816545724869,
-0.018601780757308006,
-0.08722289651632309,
-0.052608974277973175,
0.060956500470638275,
-0.11497385054826736,
0.22473330795764923,
-0.04154995083808899,
0.045806147158145905,
0.13082513213157654,
0.02837492898106575,
-0.07280921190977097,
0.09297476708889008,
0.04155753552913666,
-0.06319914013147354,
0.031915824860334396,
0.0958428606390953,
-0.035735636949539185,
0.09097802639007568,
0.061976008117198944,
-0.13848306238651276,
0.01646588370203972,
-0.047426506876945496,
-0.07511083781719208,
-0.053246431052684784,
-0.04821081459522247,
-0.061395712196826935,
0.13512717187404633,
0.18378499150276184,
-0.03449869528412819,
0.014288014732301235,
-0.048711732029914856,
0.019743043929338455,
0.06974764913320541,
0.05668903887271881,
-0.02842186577618122,
-0.24795599281787872,
0.011637489311397076,
0.09647117555141449,
-0.01345064677298069,
-0.3207091689109802,
-0.07913479954004288,
-0.015894245356321335,
-0.03357844054698944,
-0.09838392585515976,
0.08659610152244568,
0.13363921642303467,
0.05246322229504585,
-0.061531953513622284,
-0.0848599448800087,
-0.06982412934303284,
0.16599711775779724,
-0.12612435221672058,
-0.09700369089841843
] |
null | null | stable-baselines3 |
# **DQN** Agent playing **SpaceInvadersNoFrameskip-v4**
This is a trained model of a **DQN** agent playing **SpaceInvadersNoFrameskip-v4**
using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3)
and the [RL Zoo](https://github.com/DLR-RM/rl-baselines3-zoo).
The RL Zoo is a training framework for Stable Baselines3
reinforcement learning agents,
with hyperparameter optimization and pre-trained agents included.
## Usage (with SB3 RL Zoo)
RL Zoo: https://github.com/DLR-RM/rl-baselines3-zoo<br/>
SB3: https://github.com/DLR-RM/stable-baselines3<br/>
SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib
Install the RL Zoo (with SB3 and SB3-Contrib):
```bash
pip install rl_zoo3
```
```
# Download model and save it into the logs/ folder
python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga flipchip167 -f logs/
python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/
```
If you installed the RL Zoo3 via pip (`pip install rl_zoo3`), from anywhere you can do:
```
python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga flipchip167 -f logs/
python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/
```
## Training (with the RL Zoo)
```
python -m rl_zoo3.train --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/
# Upload the model and generate video (when possible)
python -m rl_zoo3.push_to_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ -orga flipchip167
```
## Hyperparameters
```python
OrderedDict([('batch_size', 32),
('buffer_size', 100000),
('env_wrapper',
['stable_baselines3.common.atari_wrappers.AtariWrapper']),
('exploration_final_eps', 0.01),
('exploration_fraction', 0.1),
('frame_stack', 4),
('gradient_steps', 1),
('learning_rate', 0.0001),
('learning_starts', 100000),
('n_timesteps', 1000000.0),
('optimize_memory_usage', False),
('policy', 'CnnPolicy'),
('target_update_interval', 1000),
('train_freq', 4),
('normalize', False)])
```
# Environment Arguments
```python
{'render_mode': 'rgb_array'}
```
| {"library_name": "stable-baselines3", "tags": ["SpaceInvadersNoFrameskip-v4", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "DQN", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "SpaceInvadersNoFrameskip-v4", "type": "SpaceInvadersNoFrameskip-v4"}, "metrics": [{"type": "mean_reward", "value": "758.50 +/- 201.67", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | flipchip167/dqn-SpaceInvadersNoFrameskip-v4 | [
"stable-baselines3",
"SpaceInvadersNoFrameskip-v4",
"deep-reinforcement-learning",
"reinforcement-learning",
"model-index",
"region:us"
] | 2024-02-07T16:27:11+00:00 | [] | [] | TAGS
#stable-baselines3 #SpaceInvadersNoFrameskip-v4 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
|
# DQN Agent playing SpaceInvadersNoFrameskip-v4
This is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4
using the stable-baselines3 library
and the RL Zoo.
The RL Zoo is a training framework for Stable Baselines3
reinforcement learning agents,
with hyperparameter optimization and pre-trained agents included.
## Usage (with SB3 RL Zoo)
RL Zoo: URL
SB3: URL
SB3 Contrib: URL
Install the RL Zoo (with SB3 and SB3-Contrib):
If you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do:
## Training (with the RL Zoo)
## Hyperparameters
# Environment Arguments
| [
"# DQN Agent playing SpaceInvadersNoFrameskip-v4\nThis is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4\nusing the stable-baselines3 library\nand the RL Zoo.\n\nThe RL Zoo is a training framework for Stable Baselines3\nreinforcement learning agents,\nwith hyperparameter optimization and pre-trained agents included.",
"## Usage (with SB3 RL Zoo)\n\nRL Zoo: URL\nSB3: URL\nSB3 Contrib: URL\n\nInstall the RL Zoo (with SB3 and SB3-Contrib):\n\n\n\n\nIf you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do:",
"## Training (with the RL Zoo)",
"## Hyperparameters",
"# Environment Arguments"
] | [
"TAGS\n#stable-baselines3 #SpaceInvadersNoFrameskip-v4 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n",
"# DQN Agent playing SpaceInvadersNoFrameskip-v4\nThis is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4\nusing the stable-baselines3 library\nand the RL Zoo.\n\nThe RL Zoo is a training framework for Stable Baselines3\nreinforcement learning agents,\nwith hyperparameter optimization and pre-trained agents included.",
"## Usage (with SB3 RL Zoo)\n\nRL Zoo: URL\nSB3: URL\nSB3 Contrib: URL\n\nInstall the RL Zoo (with SB3 and SB3-Contrib):\n\n\n\n\nIf you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do:",
"## Training (with the RL Zoo)",
"## Hyperparameters",
"# Environment Arguments"
] | [
43,
90,
73,
9,
5,
7
] | [
"passage: TAGS\n#stable-baselines3 #SpaceInvadersNoFrameskip-v4 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# DQN Agent playing SpaceInvadersNoFrameskip-v4\nThis is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4\nusing the stable-baselines3 library\nand the RL Zoo.\n\nThe RL Zoo is a training framework for Stable Baselines3\nreinforcement learning agents,\nwith hyperparameter optimization and pre-trained agents included.## Usage (with SB3 RL Zoo)\n\nRL Zoo: URL\nSB3: URL\nSB3 Contrib: URL\n\nInstall the RL Zoo (with SB3 and SB3-Contrib):\n\n\n\n\nIf you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do:## Training (with the RL Zoo)## Hyperparameters# Environment Arguments"
] | [
0.043572068214416504,
0.2414778620004654,
-0.0026879787910729647,
0.012635791674256325,
0.05784223601222038,
0.0030472534708678722,
0.08585051447153091,
0.10650663822889328,
0.024212315678596497,
-0.001382096204906702,
0.003954293206334114,
0.17533031105995178,
0.03632635250687599,
0.13125447928905487,
-0.018073517829179764,
-0.2066594809293747,
-0.013479253277182579,
-0.06247470900416374,
-0.07153085619211197,
0.036099132150411606,
0.07206681370735168,
-0.030116932466626167,
0.036061208695173264,
-0.051406677812337875,
-0.057161085307598114,
0.036824777722358704,
-0.03157254680991173,
0.007067287806421518,
0.15158706903457642,
-0.1222257912158966,
0.12329676002264023,
0.020955175161361694,
0.1896144151687622,
-0.12332789599895477,
0.0339222252368927,
0.08982209116220474,
-0.036988191306591034,
0.013221588917076588,
0.00975361280143261,
-0.052562564611434937,
0.1590864509344101,
-0.09371145814657211,
0.07146181166172028,
0.010926910676062107,
-0.07592244446277618,
-0.1774153709411621,
-0.09356249868869781,
0.07947742193937302,
0.0617753230035305,
0.005319166928529739,
0.03726791962981224,
0.11306490749120712,
-0.020991774275898933,
0.06488905102014542,
0.11562903225421906,
-0.17549200356006622,
0.013578375801444054,
0.17859570682048798,
0.003242473118007183,
0.15767055749893188,
-0.05546637624502182,
0.019877681508660316,
0.02752300351858139,
0.04758313298225403,
0.06873945891857147,
-0.08186400681734085,
-0.1364826112985611,
-0.056155186146497726,
-0.15456219017505646,
-0.03352400287985802,
0.05195203423500061,
-0.011860138736665249,
-0.05783402919769287,
-0.010724928230047226,
-0.04010869935154915,
0.0008851495804265141,
-0.028637725859880447,
0.01805497519671917,
0.07031578570604324,
-0.01226285845041275,
0.02092539705336094,
-0.08391954004764557,
-0.0390290804207325,
-0.038563769310712814,
-0.018022390082478523,
0.12054917961359024,
0.08285853266716003,
0.0266572255641222,
-0.04135355353355408,
0.10274127870798111,
-0.07091585546731949,
-0.05454207584261894,
0.04555258899927139,
-0.03786851093173027,
-0.10615779459476471,
0.02120024710893631,
-0.05905991420149803,
0.026879185810685158,
0.09943640232086182,
0.18048083782196045,
-0.09862488508224487,
0.012620617635548115,
-0.03430783003568649,
0.08121664822101593,
-0.03196052461862564,
0.03197542577981949,
-0.0840383991599083,
-0.016251085326075554,
0.17835216224193573,
0.0030782297253608704,
0.022272996604442596,
0.002074616262689233,
-0.049819961190223694,
-0.02881433069705963,
-0.017756454646587372,
0.06631895154714584,
0.07032092660665512,
0.010587303899228573,
-0.0037596761249005795,
-0.027667716145515442,
-0.036921944469213486,
-0.05629328638315201,
-0.04952820762991905,
0.018803736194968224,
-0.04712437093257904,
-0.047942135483026505,
0.06027210131287575,
-0.005624116864055395,
0.11337806284427643,
-0.025607796385884285,
0.026316547766327858,
-0.019410157576203346,
-0.07494441419839859,
-0.13221681118011475,
-0.0304415225982666,
0.0691632330417633,
0.04371757060289383,
-0.22497159242630005,
-0.16994807124137878,
-0.008539012633264065,
0.017946386709809303,
-0.018741264939308167,
-0.11334165185689926,
0.02453240379691124,
-0.007166135590523481,
-0.049758363515138626,
-0.01601579785346985,
0.10474669933319092,
-0.020438622683286667,
0.018010856583714485,
-0.05593825876712799,
0.16603368520736694,
-0.14290283620357513,
0.031004127115011215,
-0.08706212788820267,
0.023509707301855087,
-0.21286657452583313,
0.041208744049072266,
-0.177636057138443,
0.04863585904240608,
-0.08500861376523972,
0.02327173389494419,
0.021320728585124016,
0.01968831568956375,
0.08580207824707031,
0.10143322497606277,
-0.23631145060062408,
0.05405791476368904,
0.07900930196046829,
-0.022739801555871964,
-0.04218491166830063,
0.06798892468214035,
-0.06558530032634735,
0.1382148116827011,
0.046505436301231384,
0.24831900000572205,
0.10361487418413162,
-0.2036508023738861,
0.061786454170942307,
0.0578593946993351,
-0.08880111575126648,
-0.004730981774628162,
-0.020022382959723473,
0.11598580330610275,
-0.01114928349852562,
0.03338807821273804,
-0.12186288088560104,
0.1456439197063446,
0.02738998830318451,
-0.0165485180914402,
-0.04454165697097778,
-0.1614885926246643,
0.10309953987598419,
-0.015504824928939342,
0.09532155096530914,
-0.042415786534547806,
0.0001161050095106475,
-0.011168917641043663,
0.18012429773807526,
-0.043841805309057236,
0.0007168867159634829,
0.07871408760547638,
0.10895700752735138,
0.028009075671434402,
-0.020230965688824654,
-0.20380273461341858,
-0.0423048660159111,
0.02367858961224556,
0.044489551335573196,
0.2190362960100174,
0.19936694204807281,
0.07770156860351562,
-0.022313760593533516,
-0.025487221777439117,
-0.003248062450438738,
-0.05106664076447487,
0.03467361256480217,
-0.027858436107635498,
-0.024532482028007507,
0.06065356358885765,
-0.09305168688297272,
0.02817818708717823,
-0.13112716376781464,
0.06307920068502426,
-0.17345242202281952,
0.06863926351070404,
0.021998396143317223,
-0.005436043255031109,
0.024577690288424492,
-0.011292695067822933,
-0.034188106656074524,
-0.06233125180006027,
0.07110602408647537,
0.06098933145403862,
0.014702376909554005,
0.0021991983521729708,
-0.0683600977063179,
-0.13828523457050323,
0.08231553435325623,
-0.04042381793260574,
-0.14305958151817322,
0.06392676383256912,
0.011172642931342125,
0.04875864461064339,
-0.05975872278213501,
0.016254881396889687,
0.22900153696537018,
0.05321883037686348,
0.09785865992307663,
-0.04092191904783249,
-0.022525805979967117,
-0.06617844104766846,
-0.06677833944559097,
0.09694591909646988,
0.10812206566333771,
0.060318704694509506,
-0.0030071530491113663,
0.07626225054264069,
0.10942911356687546,
-0.1035122498869896,
-0.0651884600520134,
0.03220061957836151,
-0.05973697826266289,
0.019652515649795532,
0.049140311777591705,
0.02971293032169342,
0.08619047701358795,
0.1833551675081253,
0.008245792239904404,
0.0386311337351799,
-0.025997694581747055,
0.026109617203474045,
-0.15547916293144226,
-0.03145433962345123,
0.04308181628584862,
0.00886955764144659,
-0.07408110797405243,
0.04994636029005051,
0.051439400762319565,
0.13607151806354523,
-0.08217083662748337,
-0.13170577585697174,
-0.059745315462350845,
-0.03804200142621994,
-0.04239124804735184,
0.14975430071353912,
-0.08507520705461502,
-0.19221234321594238,
-0.017164425924420357,
-0.15751953423023224,
-0.02518727444112301,
-0.005179801490157843,
0.002318724524229765,
-0.08325926214456558,
0.017780914902687073,
0.010001576505601406,
-0.03129372000694275,
-0.0684933215379715,
-0.06596160680055618,
-0.05786636844277382,
0.09124112874269485,
0.06932931393384933,
-0.12240120023488998,
-0.00961651187390089,
-0.03742414712905884,
-0.020465577021241188,
0.04516167193651199,
0.08452648669481277,
-0.007267598994076252,
0.07773483544588089,
-0.13209199905395508,
-0.06962883472442627,
0.02834828943014145,
0.2766247093677521,
0.02882981114089489,
0.004668009467422962,
0.17051753401756287,
-0.03629542142152786,
0.04912714660167694,
0.16181479394435883,
0.030781643465161324,
-0.14196757972240448,
0.07090470939874649,
-0.011341600678861141,
-0.09542687982320786,
-0.1706860214471817,
-0.10215658694505692,
-0.037867411971092224,
-0.05015881359577179,
0.05638284236192703,
0.004951419774442911,
-0.04476970434188843,
0.05910305306315422,
0.08782228082418442,
-0.017004497349262238,
-0.06151578947901726,
0.11129767447710037,
0.032263003289699554,
-0.030136963352560997,
0.08078382909297943,
-0.042354047298431396,
-0.04206389561295509,
0.0032403599470853806,
0.22643887996673584,
0.0937788337469101,
-0.01775507442653179,
-0.042567066848278046,
0.019317636266350746,
0.05095715448260307,
0.03613382205367088,
0.11312435567378998,
-0.06975842267274857,
-0.06826137751340866,
-0.035185977816581726,
0.027829548344016075,
-0.02945687249302864,
0.08205190300941467,
0.0630207508802414,
0.005563626065850258,
-0.04653681069612503,
-0.07972332090139389,
-0.04849022626876831,
0.08408913016319275,
-0.027642227709293365,
-0.10093270242214203,
0.09321888536214828,
0.048575710505247116,
0.0016974330646917224,
0.03055831417441368,
0.027994604781270027,
0.01462269201874733,
-0.07982148975133896,
-0.06775744259357452,
0.011468625627458096,
0.07076629996299744,
-0.06822766363620758,
-0.027886953204870224,
-0.19817815721035004,
0.14578363299369812,
0.010630400851368904,
0.04118429124355316,
-0.13048617541790009,
0.1209396943449974,
-0.023116756230592728,
-0.026430301368236542,
0.013811616227030754,
0.0014643745962530375,
0.08203291147947311,
-0.04806509613990784,
0.15762180089950562,
0.009528410620987415,
-0.28092408180236816,
-0.1418946087360382,
-0.08416824042797089,
-0.051183976233005524,
-0.022873088717460632,
0.014752174727618694,
0.0642135739326477,
0.01516205258667469,
0.003868846921250224,
-0.013076163828372955,
0.03185269236564636,
-0.09826882928609848,
-0.06493937969207764,
-0.04839126765727997,
-0.02250157669186592,
-0.06525848805904388,
-0.05647949501872063,
-0.0006809153710491955,
-0.17226077616214752,
0.12522587180137634,
0.11787347495555878,
-0.06451737880706787,
-0.041814323514699936,
-0.06554657220840454,
0.046191465109586716,
-0.07571537792682648,
0.0469326451420784,
0.003414976177737117,
0.019198855385184288,
-0.06806991249322891,
-0.17922484874725342,
0.016097763553261757,
-0.10899919271469116,
0.03772687539458275,
-0.05070559307932854,
0.020257100462913513,
0.08594245463609695,
0.17520126700401306,
0.05856714025139809,
0.01460097823292017,
-0.07239776104688644,
-0.07543374598026276,
-0.0017121878918260336,
-0.06344114243984222,
0.05762333422899246,
-0.009151889942586422,
-0.20333483815193176,
0.02763226442039013,
-0.11414948850870132,
0.06860900670289993,
0.3310066759586334,
0.3324824273586273,
-0.10698744654655457,
0.1177443116903305,
0.04819539934396744,
-0.042202454060316086,
-0.21051374077796936,
-0.002244179602712393,
0.012272895313799381,
0.024992236867547035,
0.13725964725017548,
-0.12924811244010925,
0.05453680083155632,
0.0794181227684021,
-0.024458877742290497,
0.01456840243190527,
-0.09078162908554077,
-0.10816970467567444,
0.20847418904304504,
0.14226987957954407,
0.04421741142868996,
-0.09421348571777344,
0.08391669392585754,
0.004295284394174814,
0.08375877887010574,
0.2107764035463333,
-0.052112679928541183,
0.10695768147706985,
0.005195184610784054,
0.19852910935878754,
0.0328996516764164,
-0.023768596351146698,
0.10834760218858719,
-0.009801650419831276,
0.07911337912082672,
0.03985166177153587,
-0.007676942739635706,
0.010487722232937813,
-0.04522453248500824,
0.014148596674203873,
-0.028376007452607155,
0.010284217074513435,
-0.2274095118045807,
0.0582297146320343,
-0.06368855386972427,
0.04604509472846985,
0.008256820961833,
-0.0999874547123909,
-0.03583388403058052,
0.06431841105222702,
0.08014573156833649,
0.01975327916443348,
0.0436067171394825,
-0.03867863491177559,
0.11051398515701294,
0.20660489797592163,
-0.009811338968575,
0.17751595377922058,
-0.0615963339805603,
0.01464168168604374,
-0.023011628538370132,
-0.04223164543509483,
-0.1462583988904953,
-0.035259708762168884,
0.03498423472046852,
0.057734888046979904,
0.015203364193439484,
0.049647457897663116,
-0.05656236410140991,
0.08498423546552658,
0.021687336266040802,
-0.041541360318660736,
0.033579520881175995,
0.08835696429014206,
0.12415177375078201,
0.010754258371889591,
-0.030121933668851852,
0.06147436052560806,
-0.08128108084201813,
-0.09446098655462265,
-0.004497923422604799,
-0.029991207644343376,
-0.1083834245800972,
0.11353230476379395,
0.16914646327495575,
0.039594944566488266,
-0.057076629251241684,
0.10688766092061996,
-0.02768099494278431,
0.10047874599695206,
0.009198128245770931,
0.06507332623004913,
-0.014091075398027897,
-0.03691792115569115,
0.10611724853515625,
-0.05442855879664421,
-0.01637818105518818,
0.07645545154809952,
-0.06522727757692337,
-0.023877469822764397,
-0.0801999643445015,
0.06034626066684723,
0.09222240000963211,
-0.16854619979858398,
-0.0639432892203331,
-0.032122284173965454,
-0.08628080040216446,
0.013965039514005184,
0.012447911314666271,
0.0710059329867363,
-0.08589600026607513,
0.06316167116165161,
-0.024337708950042725,
0.015639442950487137,
-0.03689891844987869,
0.019222697243094444,
-0.19525384902954102,
-0.002140450058504939,
-0.11280795186758041,
-0.00348020251840353,
-0.002931603929027915,
0.04463808611035347,
-0.04961875081062317,
-0.029358822852373123,
-0.0030675032176077366,
0.044366419315338135,
-0.16609135270118713,
0.002798673929646611,
-0.011639905162155628,
0.03210212290287018,
-0.0002893915225286037,
-0.0983390137553215,
0.014195028692483902,
-0.04294256120920181,
-0.04198618605732918,
0.04925514757633209,
0.009436776861548424,
0.06470516324043274,
-0.2795179784297943,
-0.14905457198619843,
0.030816160142421722,
0.0683867484331131,
0.05483196675777435,
-0.1830425262451172,
0.03568267077207565,
-0.08042316138744354,
-0.02253127470612526,
-0.037770628929138184,
0.018491698428988457,
-0.0539514496922493,
0.0018174031283706427,
-0.04225044324994087,
-0.023033907637000084,
-0.028055014088749886,
-0.07556360960006714,
0.0826747715473175,
0.12462522834539413,
0.07555580884218216,
-0.03807181864976883,
0.09595896303653717,
-0.10009756684303284,
-0.04657831788063049,
-0.04052736237645149,
-0.036951083689928055,
0.017965637147426605,
-0.0870552659034729,
0.048530060797929764,
0.05188591405749321,
0.18719671666622162,
-0.08520494401454926,
-0.058800119906663895,
-0.014255574904382229,
0.0746525228023529,
0.07849094271659851,
0.005095830652862787,
0.17779210209846497,
-0.045693784952163696,
0.05693846940994263,
0.021304311230778694,
0.046699028462171555,
0.10497613251209259,
-0.023569339886307716,
0.14490213990211487,
0.21171095967292786,
-0.037196725606918335,
-0.11048602312803268,
0.043668005615472794,
0.01745123788714409,
-0.002401199424639344,
0.05968761444091797,
0.11983796209096909,
-0.050589341670274734,
-0.10903856158256531,
0.23442286252975464,
0.054169271141290665,
-0.11218088120222092,
0.09546315670013428,
0.039532262831926346,
-0.015890996903181076,
-0.1301896870136261,
0.010444961488246918,
-0.0013640925753861666,
-0.11233190447092056,
0.03386834263801575,
-0.06087532266974449,
-0.025547027587890625,
0.11809267848730087,
0.008789865300059319,
0.03317064419388771,
-0.04139537364244461,
-0.03756232187151909,
-0.04352104663848877,
-0.04273213446140289,
-0.012549578212201595,
-0.02991986647248268,
-0.030186517164111137,
-0.07621737569570541,
-0.007770835887640715,
-0.012012424878776073,
0.030795488506555557,
-0.015285328030586243,
-0.02503054589033127,
-0.021192016080021858,
-0.06697061657905579,
-0.0026312144473195076,
-0.008178025484085083,
0.015549594536423683,
0.010121971368789673,
0.2358063906431198,
0.07042546570301056,
-0.10260069370269775,
-0.01036880537867546,
0.22197756171226501,
-0.03853277862071991,
-0.06528383493423462,
-0.07849395275115967,
0.25128230452537537,
-0.10482002794742584,
0.051095426082611084,
-0.005819917656481266,
-0.06550488620996475,
-0.07153836637735367,
0.2309868484735489,
0.13502730429172516,
-0.1677926480770111,
0.06329060345888138,
-0.0368385910987854,
-0.009490780532360077,
-0.14286863803863525,
0.16013580560684204,
0.1865294873714447,
0.09480160474777222,
-0.12259847670793533,
0.0023130534682422876,
-0.03518044203519821,
-0.018328361213207245,
-0.1660851687192917,
-0.004593863617628813,
-0.029364850372076035,
-0.0427238829433918,
-0.050771355628967285,
0.029773715883493423,
-0.15205919742584229,
-0.0927426889538765,
-0.1916799396276474,
-0.11482496559619904,
-0.12386849522590637,
-0.04549141973257065,
-0.11142764985561371,
-0.0019938007462769747,
0.02257080189883709,
-0.0641874223947525,
0.021061956882476807,
-0.0212461706250906,
-0.05887424945831299,
0.015386379323899746,
-0.08395619690418243,
0.0674985870718956,
0.06488548219203949,
0.15327942371368408,
-0.0790991559624672,
0.025424562394618988,
0.07090727984905243,
-0.057595450431108475,
-0.10164349526166916,
0.06067253649234772,
0.015708057209849358,
-0.1972588747739792,
0.007548294495791197,
0.17712996900081635,
-0.10420889407396317,
0.09745754301548004,
0.048501528799533844,
-0.012951982207596302,
0.0867827981710434,
-0.024721821770071983,
-0.016682926565408707,
-0.04852180927991867,
-0.011212974786758423,
-0.10143939405679703,
0.09892100840806961,
0.0876845121383667,
-0.0517118014395237,
0.07436849176883698,
-0.09508965909481049,
-0.04068392515182495,
0.13103286921977997,
-0.010057874955236912,
-0.08450483530759811,
-0.11667824536561966,
-0.04081142693758011,
0.09684515744447708,
-0.018041390925645828,
-0.20185889303684235,
-0.11639472097158432,
-0.11752668023109436,
-0.00014377340266946703,
-0.03563340753316879,
0.061800602823495865,
0.02430674433708191,
-0.02556120604276657,
-0.008150683715939522,
-0.17615078389644623,
-0.06614746153354645,
0.13479791581630707,
-0.10176112502813339,
-0.07456064969301224
] |
null | null | stable-baselines3 |
# **PPO** Agent playing **LunarLander-v2**
This is a trained model of a **PPO** agent playing **LunarLander-v2**
using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3).
## Usage (with Stable-baselines3)
TODO: Add your code
```python
from stable_baselines3 import ...
from huggingface_sb3 import load_from_hub
...
```
| {"library_name": "stable-baselines3", "tags": ["LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "PPO", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "LunarLander-v2", "type": "LunarLander-v2"}, "metrics": [{"type": "mean_reward", "value": "270.54 +/- 13.03", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | Scott617/ppo-LunarLander-v2 | [
"stable-baselines3",
"LunarLander-v2",
"deep-reinforcement-learning",
"reinforcement-learning",
"model-index",
"region:us"
] | 2024-02-07T16:28:45+00:00 | [] | [] | TAGS
#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
|
# PPO Agent playing LunarLander-v2
This is a trained model of a PPO agent playing LunarLander-v2
using the stable-baselines3 library.
## Usage (with Stable-baselines3)
TODO: Add your code
| [
"# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
"TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n",
"# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
39,
41,
17
] | [
"passage: TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
0.03942384943366051,
0.04900386184453964,
-0.005304091144353151,
0.026427261531352997,
0.107408307492733,
-0.026511888951063156,
0.11188238859176636,
0.0814051404595375,
0.10722193866968155,
0.04762078449130058,
0.08338645845651627,
0.06030960753560066,
0.05080918222665787,
0.2571701407432556,
0.04754156619310379,
-0.22987541556358337,
0.036159250885248184,
-0.04869936779141426,
0.12395193427801132,
0.07178173214197159,
-0.0038484656251966953,
-0.06485428661108017,
0.020415637642145157,
-0.013290755450725555,
0.05367108806967735,
0.04282612353563309,
-0.01716216839849949,
-0.08207534998655319,
0.07169748842716217,
-0.06345846503973007,
0.06986866891384125,
0.07677983492612839,
0.13218913972377777,
-0.17832116782665253,
0.029566360637545586,
0.02571309357881546,
-0.07189024239778519,
0.01342033501714468,
0.008019951172173023,
0.05120139941573143,
0.17303818464279175,
0.019879888743162155,
0.07844575494527817,
-0.0025605305563658476,
-0.15412317216396332,
-0.018950799480080605,
0.0436202734708786,
0.12546207010746002,
0.08808347582817078,
0.04605821147561073,
0.01970590092241764,
0.17503218352794647,
-0.054352790117263794,
-0.028833400458097458,
0.21759237349033356,
-0.2881564497947693,
-0.031460098922252655,
0.321048766374588,
0.06997483223676682,
0.09725230932235718,
-0.07540661096572876,
-0.03619609400629997,
0.007783263456076384,
-0.013137873262166977,
-0.028666524216532707,
-0.07447073608636856,
0.17313385009765625,
0.05152064561843872,
-0.05057951435446739,
-0.09541505575180054,
0.16948209702968597,
0.006921638268977404,
0.0018855923553928733,
-0.019282981753349304,
0.009060598909854889,
0.07402525842189789,
-0.016097044572234154,
-0.07255112379789352,
0.057438433170318604,
0.05330665782094002,
0.019649166613817215,
-0.1435653269290924,
-0.10762494057416916,
-0.022740179672837257,
-0.008012006990611553,
0.17786912620067596,
-0.009255532175302505,
0.042902372777462006,
0.003065188182517886,
0.10384012013673782,
-0.12480384111404419,
-0.03354184702038765,
-0.0454259067773819,
-0.07565800100564957,
-0.0223417766392231,
-0.02058211714029312,
-0.03580251708626747,
0.07184842973947525,
0.11971849203109741,
0.027368178591132164,
0.09350208193063736,
0.047715865075588226,
-0.03206788748502731,
0.06343851238489151,
0.05555703118443489,
0.14222665131092072,
0.05807621404528618,
0.012854371219873428,
0.13179877400398254,
0.055213116109371185,
0.033023182302713394,
-0.0613492950797081,
-0.18252409994602203,
0.07489913702011108,
-0.07031869143247604,
0.007941240444779396,
0.12051256000995636,
-0.04480670019984245,
-0.1183447614312172,
-0.037500523030757904,
-0.017392054200172424,
-0.06224250793457031,
-0.025395862758159637,
0.0547584593296051,
-0.02883218228816986,
-0.03973718360066414,
0.0011496668448671699,
0.09384800493717194,
0.00953749567270279,
-0.1752052903175354,
0.03303423151373863,
-0.025042934343218803,
-0.10782608389854431,
0.009975161403417587,
0.0022444494534283876,
0.03394931182265282,
0.04408763721585274,
-0.11822668462991714,
-0.30899152159690857,
-0.07652641832828522,
0.05490870401263237,
-0.06516939401626587,
-0.18425025045871735,
-0.13193942606449127,
0.02454492449760437,
-0.09037084132432938,
-0.044885024428367615,
-0.12759265303611755,
-0.028549788519740105,
0.01743689924478531,
0.011519349180161953,
0.10758619755506516,
-0.0106219332665205,
-0.012188062071800232,
-0.1571401208639145,
0.008273907005786896,
-0.20951123535633087,
0.0890483483672142,
-0.019150104373693466,
0.037884220480918884,
-0.032381169497966766,
-0.07404014468193054,
0.030707746744155884,
0.052499737590551376,
-0.01474119070917368,
0.13510210812091827,
-0.15592676401138306,
-0.03691192343831062,
-0.007996266707777977,
-0.13611900806427002,
-0.04786273464560509,
-0.10358831286430359,
-0.04357128217816353,
0.13354332745075226,
0.018664736300706863,
0.15356586873531342,
-0.08709818124771118,
-0.0722038671374321,
0.20489206910133362,
-0.010411538183689117,
-0.12820468842983246,
-0.076752208173275,
0.10165707021951675,
0.021510310471057892,
-0.056606587022542953,
-0.02523270808160305,
-0.1839766949415207,
-0.0152357779443264,
-0.04550420492887497,
-0.047039128839969635,
0.01796751655638218,
-0.010888241231441498,
0.13837894797325134,
0.08494598418474197,
0.05018039792776108,
-0.06086122244596481,
-0.006730288732796907,
0.10779471695423126,
0.08823856711387634,
0.008680110797286034,
0.023406028747558594,
-0.05774238705635071,
0.09552932530641556,
-0.04003755748271942,
-0.0142367510125041,
-0.08283266425132751,
-0.036246106028556824,
-0.026256313547492027,
0.17507147789001465,
0.09440762549638748,
0.2257927656173706,
0.09567736834287643,
0.039160262793302536,
0.031270865350961685,
-0.13181598484516144,
-0.1425403207540512,
-0.0017254541162401438,
0.09020978957414627,
-0.14270411431789398,
-0.04119925573468208,
-0.08974775671958923,
-0.17768175899982452,
-0.12202505767345428,
0.0006432619411498308,
-0.17960017919540405,
0.06390921026468277,
0.05408334732055664,
-0.035177867859601974,
0.03272094577550888,
0.13032332062721252,
-0.011533179320394993,
-0.03967514634132385,
0.0831870287656784,
0.0379033200442791,
-0.041234664618968964,
-0.021742934361100197,
0.11885567009449005,
0.15673065185546875,
0.13124459981918335,
-0.03511447086930275,
0.004914294462651014,
0.07076404243707657,
-0.02309088408946991,
0.06539414077997208,
0.0558244064450264,
0.20973342657089233,
0.188301220536232,
0.038996949791908264,
0.008822928182780743,
-0.07048165798187256,
0.0855446457862854,
-0.0742373839020729,
-0.14302679896354675,
-0.05579735338687897,
0.08729292452335358,
0.016605578362941742,
0.023469142615795135,
0.08711627870798111,
0.024545932188630104,
0.09132762253284454,
0.15968108177185059,
0.01990218088030815,
-0.09659269452095032,
-0.050218869000673294,
0.01175848301500082,
0.027713103219866753,
0.04794301092624664,
-0.04514073207974434,
-0.00937939714640379,
0.017020760104060173,
-0.10303554683923721,
0.031789086759090424,
-0.1413339376449585,
-0.1358717679977417,
0.044326696544885635,
0.003906996920704842,
0.010907664895057678,
0.02786896750330925,
-0.0038291432429105043,
0.019039705395698547,
0.04351753741502762,
-0.06975466758012772,
0.047416772693395615,
-0.024745507165789604,
-0.020031947642564774,
0.03340689837932587,
-0.057257164269685745,
-0.205775648355484,
-0.17696654796600342,
0.00013708483311347663,
-0.09910997003316879,
0.10194740444421768,
0.018308809027075768,
-0.12373185902833939,
0.047737859189510345,
-0.05822649225592613,
0.027574289590120316,
-0.01875593699514866,
-0.049130141735076904,
0.10507171601057053,
0.1525275856256485,
-0.016146350651979446,
0.018018173053860664,
-0.04865182936191559,
-0.10157987475395203,
-0.19632206857204437,
0.0691583976149559,
0.04680244252085686,
0.014610917307436466,
0.10669491440057755,
0.018072687089443207,
0.02367905154824257,
-0.007674071006476879,
-0.016521066427230835,
-0.011659215204417706,
-0.08781040459871292,
0.31909599900245667,
0.04510033503174782,
-0.025173069909214973,
0.02041010931134224,
-0.0043001663871109486,
-0.028083480894565582,
0.03263787180185318,
-0.0985708013176918,
-0.07548979669809341,
-0.08774089068174362,
-0.04367410019040108,
-0.09784720093011856,
0.053299110382795334,
0.05916472524404526,
0.003188040340319276,
-0.07727594673633575,
0.04221395403146744,
0.11369874328374863,
-0.0923808291554451,
-0.07137343287467957,
0.07477962225675583,
0.0972946360707283,
-0.07331304252147675,
0.00012658814375754446,
0.00874367356300354,
0.023951783776283264,
0.037102166563272476,
0.06778035312891006,
-0.03966575115919113,
0.08589404821395874,
-0.19917890429496765,
0.0372927263379097,
0.106058269739151,
0.023754918947815895,
0.0638108178973198,
0.07643651217222214,
-0.1058402881026268,
-0.008500572293996811,
-0.032518330961465836,
-0.21341575682163239,
0.1668180525302887,
0.1355515867471695,
0.06788124144077301,
-0.025637222453951836,
-0.00461410591378808,
-0.0649740919470787,
0.05773647129535675,
0.02723747305572033,
-0.14758841693401337,
0.004883295856416225,
0.06064270809292793,
0.026899009943008423,
0.01614922471344471,
0.07971042394638062,
0.014697225764393806,
-0.1801026314496994,
-0.014406266622245312,
0.10730406641960144,
0.002390873385593295,
0.0053148469887673855,
-0.03175045922398567,
-0.1755964607000351,
0.0751047357916832,
0.004285442177206278,
0.07233936339616776,
-0.1676585078239441,
0.14297930896282196,
-0.10089799761772156,
0.07726949453353882,
-0.004285062663257122,
-0.021311495453119278,
0.02507244050502777,
-0.0541163794696331,
0.15163759887218475,
0.01058570109307766,
-0.021810131147503853,
-0.1200498715043068,
-0.1717042326927185,
-0.019227758049964905,
-0.11788936704397202,
-0.11679866164922714,
0.050424277782440186,
0.062185097485780716,
0.04923136904835701,
-0.061147067695856094,
0.1518532931804657,
-0.047422297298908234,
0.060713399201631546,
-0.06893875449895859,
-0.06755045056343079,
0.03764858841896057,
-0.12588608264923096,
-0.08176055550575256,
0.05573027580976486,
0.19166934490203857,
0.15833087265491486,
-0.02816431224346161,
-0.03472423925995827,
-0.047419581562280655,
-0.006212298292666674,
-0.007802055217325687,
0.0275666993111372,
0.023223137483000755,
0.07315318286418915,
-0.07681374251842499,
-0.11649256944656372,
0.033787861466407776,
-0.06713802367448807,
-0.055589709430933,
-0.015439179725944996,
0.1513158082962036,
0.04671623185276985,
0.07720734924077988,
-0.018946662545204163,
0.03887668624520302,
-0.001724981120787561,
-0.056474871933460236,
0.16197094321250916,
0.03885216265916824,
-0.05193585529923439,
0.06837689876556396,
0.053174007683992386,
0.043745119124650955,
0.03011113777756691,
-0.026783017441630363,
0.206032395362854,
0.1980147808790207,
0.014206883497536182,
0.2175983190536499,
0.03177616000175476,
-0.03772832080721855,
-0.1300560086965561,
-0.065880686044693,
-0.006372632458806038,
0.03559038043022156,
0.08070417493581772,
-0.18207235634326935,
-0.015011128038167953,
-0.05689644813537598,
-0.034518610686063766,
-0.15059494972229004,
-0.28553900122642517,
-0.05957856774330139,
0.20075850188732147,
0.14706264436244965,
0.27519428730010986,
-0.10432573407888412,
0.035197313874959946,
0.02663275972008705,
-0.04912831634283066,
-0.006501141935586929,
0.00018665487004909664,
0.10268618166446686,
-0.15421873331069946,
0.1176437959074974,
0.08486983180046082,
-0.019002694636583328,
0.01058861706405878,
-0.1619086116552353,
0.00936629343777895,
-0.12191236019134521,
0.05354422330856323,
0.1400289237499237,
-0.048128653317689896,
-0.054873593151569366,
0.14033560454845428,
-0.024562934413552284,
-0.22685599327087402,
-0.04648222774267197,
-0.043600670993328094,
-0.010640020482242107,
0.026607351377606392,
-0.1013401448726654,
0.04101909324526787,
0.1330099105834961,
0.009380043484270573,
0.1147187277674675,
0.11749245226383209,
-0.052566803991794586,
0.10792597383260727,
0.2257719188928604,
-0.018785694614052773,
0.04689010605216026,
-0.12743118405342102,
-0.0012336712097749114,
-0.028270328417420387,
0.013657891191542149,
-0.09504974633455276,
-0.09938385337591171,
0.02366873063147068,
0.02872389927506447,
0.009118586778640747,
0.0921793207526207,
-0.029922157526016235,
0.0759170651435852,
0.06817561388015747,
-0.13014446198940277,
-0.16288450360298157,
0.015828335657715797,
-0.007344507612287998,
0.08354310691356659,
0.00027861111448146403,
0.08878035843372345,
-0.11932205408811569,
-0.018093237653374672,
-0.03153328225016594,
-0.03319635987281799,
-0.130486860871315,
-0.07138993591070175,
0.06156524643301964,
0.028095467016100883,
-0.06602972000837326,
0.1398407518863678,
0.026440169662237167,
0.15942534804344177,
0.049197953194379807,
0.012499804608523846,
0.07227300107479095,
-0.05345509201288223,
0.1283530443906784,
0.13818155229091644,
-0.00868943240493536,
-0.05460423603653908,
-0.1013643890619278,
-0.10236792266368866,
0.08925779908895493,
-0.05773641914129257,
0.07476430386304855,
-0.14885357022285461,
-0.06675903499126434,
0.015772046521306038,
0.016141414642333984,
-0.09562095999717712,
0.02571965754032135,
-0.01625603251159191,
-0.18119946122169495,
0.056570518761873245,
-0.048285093158483505,
0.0440407395362854,
-0.06347788125276566,
-0.1110161691904068,
-0.17226378619670868,
0.06091433763504028,
0.08593481779098511,
-0.053876690566539764,
-0.12229149043560028,
0.011023230850696564,
-0.00012518465518951416,
-0.06341652572154999,
-0.05023367330431938,
0.09722746908664703,
-0.11020902544260025,
0.031452205032110214,
-0.012567701749503613,
0.08853451162576675,
-0.03510405123233795,
-0.011538895778357983,
0.044220831245183945,
-0.08039166033267975,
-0.009481523185968399,
0.03534642979502678,
-0.026372017338871956,
-0.04127239063382149,
-0.2689029574394226,
0.0036654395516961813,
0.0341104120016098,
0.02497158572077751,
0.07856601476669312,
0.011906822212040424,
0.021174922585487366,
0.03993808850646019,
-0.15396519005298615,
-0.013395369984209538,
0.14574195444583893,
-0.07689505815505981,
-0.022186370566487312,
0.05703273415565491,
-0.09054436534643173,
0.013882770203053951,
-0.030287226662039757,
0.1345842480659485,
0.023923413828015327,
0.06404478847980499,
-0.0851147472858429,
0.10106813907623291,
-0.1451139897108078,
-0.04998219385743141,
-0.01244612317532301,
0.09761348366737366,
0.07019034773111343,
-0.10272270441055298,
0.014697125181555748,
0.04210108891129494,
0.19416837394237518,
0.016384804621338844,
-0.0356343574821949,
-0.03396720811724663,
0.004015897400677204,
0.22076453268527985,
0.03044266067445278,
0.10457023978233337,
0.07281364500522614,
-0.026583973318338394,
0.12624378502368927,
0.09929762035608292,
0.11280370503664017,
-0.055645186454057693,
0.13904185593128204,
0.04667386785149574,
0.038641396909952164,
0.0614289753139019,
0.06836545467376709,
0.09098632633686066,
-0.0008288522367365658,
0.1138714924454689,
0.013811973854899406,
-0.02422109805047512,
-0.021335409954190254,
0.17759373784065247,
0.10501719266176224,
-0.14769648015499115,
0.029047364369034767,
-0.01258957851678133,
0.039933037012815475,
-0.014194529503583908,
-0.15634691715240479,
-0.07240267097949982,
-0.3315149247646332,
0.1226184144616127,
-0.07119352370500565,
0.019930170848965645,
0.007913772016763687,
-0.037425633519887924,
-0.03296699747443199,
-0.04477746784687042,
0.13151589035987854,
-0.013641550205647945,
-0.006079165264964104,
-0.04815853759646416,
-0.015360191464424133,
-0.11607866734266281,
-0.11200575530529022,
-0.013207737356424332,
-0.13671602308750153,
-0.010119039565324783,
0.05595948174595833,
0.003977729007601738,
0.01821410097181797,
-0.03142618387937546,
0.0024383175186812878,
0.06541839241981506,
-0.05751744285225868,
0.056182678788900375,
0.12097269296646118,
0.08766137808561325,
-0.1058853268623352,
0.031048951670527458,
0.2011747509241104,
0.04359564557671547,
-0.12483977526426315,
0.01449228823184967,
0.1819491684436798,
0.004885740112513304,
0.017068125307559967,
-0.006097703706473112,
-0.0540788508951664,
-0.07554277032613754,
0.1251034289598465,
0.08296554535627365,
-0.09985227137804031,
0.015833314508199692,
-0.0726347416639328,
-0.01594804972410202,
-0.06374675035476685,
0.10130585730075836,
0.09538925439119339,
0.04440245032310486,
-0.10621760785579681,
-0.08487539738416672,
-0.10891728103160858,
0.040588874369859695,
-0.08629853278398514,
-0.07311757653951645,
0.09629398584365845,
-0.07057105004787445,
-0.07029950618743896,
0.025521177798509598,
-0.17978744208812714,
-0.009467960335314274,
0.1711762249469757,
-0.24654000997543335,
-0.0916430801153183,
-0.10857923328876495,
0.14477859437465668,
0.016497576609253883,
0.1013975441455841,
-0.006207061931490898,
-0.007889035157859325,
-0.20577777922153473,
0.024890204891562462,
-0.05293011665344238,
-0.02073732763528824,
0.07814782857894897,
-0.09476397186517715,
0.22629831731319427,
-0.08276885002851486,
0.020940175279974937,
0.012659613974392414,
0.0870661810040474,
-0.030675338581204414,
0.09283176809549332,
-0.03660329803824425,
-0.12576518952846527,
-0.03620953485369682,
0.03001813031733036,
0.013904244638979435,
0.10071761906147003,
0.09772487729787827,
-0.03414725139737129,
0.03389119729399681,
0.09747414290904999,
0.04172342270612717,
-0.023843804374337196,
0.0360250361263752,
-0.17077107727527618,
0.02182629331946373,
-0.018498148769140244,
-0.06935930997133255,
0.03687669709324837,
-0.06603235751390457,
0.1639697551727295,
0.04022442549467087,
0.0670473501086235,
-0.036152735352516174,
0.0073931049555540085,
-0.014454689808189869,
-0.013775371946394444,
-0.026180334389209747,
-0.17259705066680908,
-0.10422050207853317,
-0.1347656100988388,
-0.012701659463346004,
-0.034971047192811966,
0.04591470584273338,
0.023234914988279343,
-0.0003200018545612693,
-0.014577031135559082,
-0.12090865522623062,
0.04360328987240791,
0.11146783083677292,
-0.04631396010518074,
-0.026193076744675636
] |
null | null | transformers |
**NOTE:** This is a work-in-progress model that is **not** considered finished. Keep this in mind when using this model, or continue training this model.
# Model Card for mt5-small nl-en translation
The mt5-small nl-en translation model is a finetuned version of [google/mt5-small](https://huggingface.co/google/mt5-small).
It was finetuned on 237k rows of the [iwslt2017](https://huggingface.co/datasets/iwslt2017/viewer/iwslt2017-en-nl) dataset and roughly 38k rows of the [opus_books](https://huggingface.co/datasets/opus_books/viewer/en-nl) dataset. The model was trained for 15 epochs with a batchsize of 8 and is currently **not** considered finished.
## How to use
**Install dependencies**
```bash
pip install transformers
```
You can use the following code for model inference. This model was finetuned to work with an identifier when prompted that needs to be present for the best results.
```Python
from transformers import AutoModelForSeq2SeqLM, AutoTokenizer, GenerationConfig
tokenizer = AutoTokenizer.from_pretrained("Michielo/mt5-small_nl-en_translation")
model = AutoModelForSeq2SeqLM.from_pretrained("Michielo/mt5-small_nl-en_translation")
translation_generation_config = GenerationConfig(
num_beams=4,
early_stopping=True,
decoder_start_token_id=0,
eos_token_id=model.config.eos_token_id,
pad_token=model.config.pad_token_id,
)
translation_generation_config.save_pretrained("/tmp", "translation_generation_config.json")
generation_config = GenerationConfig.from_pretrained("/tmp", "translation_generation_config.json")
inputs = tokenizer(">>en<< Your dutch text here", return_tensors="pt")
outputs = model.generate(**inputs, generation_config=generation_config)
print(tokenizer.batch_decode(outputs, skip_special_tokens=True))
```
## License
This project is licensed under the Apache License 2.0 - see the [LICENSE](https://www.apache.org/licenses/LICENSE-2.0) file for details. | {"language": ["en", "nl"], "license": "apache-2.0", "tags": ["translation"], "datasets": ["opus_books", "iwslt2017"], "metrics": ["sacrebleu"], "pipeline_tag": "text2text-generation", "widget": [{"text": ">>en<< Was het leuk?"}]} | text2text-generation | Michielo/mt5-small_nl-en_translation | [
"transformers",
"safetensors",
"mt5",
"text2text-generation",
"translation",
"en",
"nl",
"dataset:opus_books",
"dataset:iwslt2017",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T16:29:32+00:00 | [] | [
"en",
"nl"
] | TAGS
#transformers #safetensors #mt5 #text2text-generation #translation #en #nl #dataset-opus_books #dataset-iwslt2017 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
NOTE: This is a work-in-progress model that is not considered finished. Keep this in mind when using this model, or continue training this model.
# Model Card for mt5-small nl-en translation
The mt5-small nl-en translation model is a finetuned version of google/mt5-small.
It was finetuned on 237k rows of the iwslt2017 dataset and roughly 38k rows of the opus_books dataset. The model was trained for 15 epochs with a batchsize of 8 and is currently not considered finished.
## How to use
Install dependencies
You can use the following code for model inference. This model was finetuned to work with an identifier when prompted that needs to be present for the best results.
## License
This project is licensed under the Apache License 2.0 - see the LICENSE file for details. | [
"# Model Card for mt5-small nl-en translation\n\nThe mt5-small nl-en translation model is a finetuned version of google/mt5-small.\n\nIt was finetuned on 237k rows of the iwslt2017 dataset and roughly 38k rows of the opus_books dataset. The model was trained for 15 epochs with a batchsize of 8 and is currently not considered finished.",
"## How to use\n\nInstall dependencies\n\n\nYou can use the following code for model inference. This model was finetuned to work with an identifier when prompted that needs to be present for the best results.",
"## License\nThis project is licensed under the Apache License 2.0 - see the LICENSE file for details."
] | [
"TAGS\n#transformers #safetensors #mt5 #text2text-generation #translation #en #nl #dataset-opus_books #dataset-iwslt2017 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for mt5-small nl-en translation\n\nThe mt5-small nl-en translation model is a finetuned version of google/mt5-small.\n\nIt was finetuned on 237k rows of the iwslt2017 dataset and roughly 38k rows of the opus_books dataset. The model was trained for 15 epochs with a batchsize of 8 and is currently not considered finished.",
"## How to use\n\nInstall dependencies\n\n\nYou can use the following code for model inference. This model was finetuned to work with an identifier when prompted that needs to be present for the best results.",
"## License\nThis project is licensed under the Apache License 2.0 - see the LICENSE file for details."
] | [
81,
102,
44,
23
] | [
"passage: TAGS\n#transformers #safetensors #mt5 #text2text-generation #translation #en #nl #dataset-opus_books #dataset-iwslt2017 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for mt5-small nl-en translation\n\nThe mt5-small nl-en translation model is a finetuned version of google/mt5-small.\n\nIt was finetuned on 237k rows of the iwslt2017 dataset and roughly 38k rows of the opus_books dataset. The model was trained for 15 epochs with a batchsize of 8 and is currently not considered finished.## How to use\n\nInstall dependencies\n\n\nYou can use the following code for model inference. This model was finetuned to work with an identifier when prompted that needs to be present for the best results.## License\nThis project is licensed under the Apache License 2.0 - see the LICENSE file for details."
] | [
-0.12667806446552277,
0.13236188888549805,
-0.0012636033352464437,
0.05256720259785652,
0.060171838849782944,
-0.014785634353756905,
0.14219346642494202,
0.08673228323459625,
0.04657646641135216,
0.020011235028505325,
0.08810403198003769,
0.1001017838716507,
0.046548742800951004,
0.1310780793428421,
-0.013927494175732136,
-0.12012605369091034,
0.055878765881061554,
-0.02689828723669052,
0.017081404104828835,
0.05641801282763481,
0.07780534774065018,
-0.029526283964514732,
0.04965876042842865,
-0.018057359382510185,
-0.036607831716537476,
0.013410786166787148,
0.01183987408876419,
-0.05670010298490524,
0.04813545569777489,
0.08209015429019928,
0.010106283240020275,
0.07020200043916702,
0.0436658039689064,
-0.17670106887817383,
0.002241133712232113,
0.04773822799324989,
-0.02756079100072384,
0.03491117060184479,
0.06999780237674713,
0.006108568049967289,
0.13267695903778076,
-0.1053943783044815,
-0.020342497155070305,
0.05674254894256592,
-0.08216043561697006,
-0.11279468238353729,
-0.11660358309745789,
0.08278270810842514,
0.0718793198466301,
0.07218249142169952,
0.00002140202559530735,
0.20316995680332184,
-0.01933985762298107,
0.07308041304349899,
0.0634491816163063,
-0.2960214912891388,
0.0055651068687438965,
0.12099352478981018,
0.04960491508245468,
0.06294824928045273,
-0.026742445304989815,
0.05855882540345192,
0.09502209722995758,
0.002093245042487979,
0.07868240028619766,
-0.050095997750759125,
0.008995517157018185,
-0.003514242824167013,
-0.16203510761260986,
-0.06322525441646576,
0.2769603431224823,
0.03174334764480591,
-0.0930127277970314,
-0.08994749188423157,
-0.07697403430938721,
0.04733423516154289,
0.023907149210572243,
-0.026199307292699814,
0.06655942648649216,
0.05967603623867035,
0.06574374437332153,
-0.05639088898897171,
-0.11716899275779724,
-0.013027546927332878,
-0.0713256299495697,
-0.03605830296874046,
-0.011645122431218624,
0.05817679315805435,
-0.035682328045368195,
0.04455142468214035,
-0.20766617357730865,
-0.03295809030532837,
-0.11040260642766953,
-0.07645940035581589,
-0.03412463143467903,
-0.0544562004506588,
-0.002926274435594678,
-0.07618919759988785,
0.005059846211224794,
0.17203810811042786,
-0.028051171451807022,
-0.018642708659172058,
0.021623672917485237,
0.04709101468324661,
0.04073509946465492,
0.06157979369163513,
-0.015388821251690388,
-0.061490342020988464,
0.12413103878498077,
-0.016552641987800598,
0.09414359927177429,
0.010270453058183193,
-0.0361766591668129,
-0.08215627819299698,
0.0013684886507689953,
0.07439398765563965,
0.07597862184047699,
0.030103763565421104,
0.006648764945566654,
0.00288074417039752,
0.15398824214935303,
-0.07706987857818604,
0.013827070593833923,
-0.008739199489355087,
-0.050854966044425964,
0.02177075296640396,
0.15769417583942413,
0.018212009221315384,
-0.0513724684715271,
-0.011420158669352531,
-0.02951178327202797,
-0.0341062918305397,
-0.049518343061208725,
-0.09511101245880127,
0.10103340446949005,
0.0007252430077642202,
-0.019378697499632835,
-0.14510710537433624,
-0.3020976483821869,
-0.04002239927649498,
0.026360880583524704,
0.008073129691183567,
0.00674119358882308,
0.045484356582164764,
0.030177351087331772,
0.026777667924761772,
-0.058594487607479095,
0.0043215700425207615,
-0.05622619017958641,
0.019796378910541534,
0.04760581627488136,
0.06483043730258942,
-0.10389375686645508,
0.021376579999923706,
-0.14187626540660858,
0.029082270339131355,
-0.10434022545814514,
0.070773184299469,
-0.05021115764975548,
0.0687628984451294,
-0.11179624497890472,
-0.031226390972733498,
-0.017363619059324265,
0.029886281117796898,
0.01800810545682907,
0.20103105902671814,
-0.12239785492420197,
0.005922118667513132,
0.13284792006015778,
-0.1494981050491333,
-0.17272982001304626,
0.10539679229259491,
0.026150139048695564,
0.10744128376245499,
0.11839158087968826,
0.08578461408615112,
0.1091194674372673,
-0.0976952537894249,
-0.09829472005367279,
0.05939154326915741,
-0.014996230602264404,
-0.14161860942840576,
0.09651124477386475,
0.0007545487605966628,
-0.0426170788705349,
0.009007329121232033,
-0.11297918111085892,
0.060963988304138184,
0.02116566151380539,
-0.07167940586805344,
-0.07741729170084,
-0.08962886035442352,
-0.06811430305242538,
0.027745315805077553,
0.01897379197180271,
-0.058068301528692245,
-0.03314420208334923,
0.11191950738430023,
0.07013468444347382,
0.0037998498883098364,
0.02942219004034996,
-0.0452997088432312,
0.0709814578294754,
-0.12537790834903717,
-0.0010715869721025229,
-0.1107921153306961,
-0.03659876808524132,
0.015609567053616047,
-0.036815665662288666,
0.02387712523341179,
-0.018026353791356087,
0.06452662497758865,
0.08663415163755417,
-0.04776909947395325,
-0.03664878010749817,
0.09781678766012192,
-0.008256616070866585,
-0.10933470726013184,
-0.1075122207403183,
-0.020500946789979935,
-0.032065343111753464,
0.07956799864768982,
-0.23765531182289124,
0.03553716093301773,
-0.04126828908920288,
0.05484938248991966,
-0.01933472976088524,
-0.01039193570613861,
0.09818318486213684,
-0.04022027179598808,
-0.03319681063294411,
-0.07038625329732895,
0.07826917618513107,
0.04513968154788017,
-0.06434841454029083,
0.1289748251438141,
-0.17343676090240479,
0.011912857182323933,
0.11706167459487915,
0.015490401536226273,
0.01853756420314312,
-0.047582514584064484,
-0.04466214030981064,
-0.010637443512678146,
-0.08150020986795425,
-0.026868833228945732,
-0.05202886089682579,
-0.012977110221982002,
0.08313097059726715,
-0.09070809185504913,
-0.030825216323137283,
-0.02057788334786892,
-0.05787764489650726,
-0.06824801117181778,
0.08440333604812622,
0.10612998902797699,
-0.16736054420471191,
0.08521949499845505,
0.14129199087619781,
-0.019019601866602898,
0.14413368701934814,
-0.007226441986858845,
-0.07051829993724823,
-0.006161030847579241,
0.07653103768825531,
0.050030939280986786,
0.016807403415441513,
-0.07433681935071945,
-0.007400242146104574,
0.06434439867734909,
0.0036005310248583555,
0.021221784874796867,
-0.11480928957462311,
-0.009042855352163315,
0.03584479168057442,
-0.04806063324213028,
-0.07167869806289673,
0.02302698791027069,
-0.08838982880115509,
0.0454610139131546,
-0.06057444214820862,
0.06411277502775192,
0.030967172235250473,
-0.0024673708248883486,
-0.16511355340480804,
0.16610665619373322,
-0.07488195598125458,
-0.1641143411397934,
-0.14946669340133667,
0.02854088880121708,
-0.05103430524468422,
0.046193547546863556,
0.09126345813274384,
-0.018577171489596367,
-0.08601599931716919,
-0.12523896992206573,
-0.0072041042149066925,
-0.005248572211712599,
-0.01113195437937975,
-0.07834244519472122,
-0.0054236482828855515,
0.03650178015232086,
-0.164511039853096,
-0.005774591583758593,
0.04508224502205849,
-0.07824806123971939,
0.04892114922404289,
-0.07423193007707596,
0.11775196343660355,
0.07201910018920898,
-0.0002905675210058689,
-0.027381064370274544,
-0.020712872967123985,
0.1788225769996643,
-0.019702447578310966,
0.05879422649741173,
0.22131049633026123,
0.11262766271829605,
0.022234216332435608,
0.09020145237445831,
-0.014618366956710815,
-0.07693943381309509,
0.04625624045729637,
-0.01124713383615017,
-0.08745957911014557,
-0.27989667654037476,
-0.05183015763759613,
-0.0733678936958313,
0.05379695072770119,
0.060705721378326416,
0.053397443145513535,
0.030082838609814644,
0.10430311411619186,
-0.11658564954996109,
0.0981840193271637,
0.013570606708526611,
0.07970201224088669,
0.07150036096572876,
0.044107913970947266,
0.07951415330171585,
-0.1247798353433609,
0.06063707172870636,
0.13540643453598022,
-0.007010553032159805,
0.12431793659925461,
-0.041230447590351105,
0.10320742428302765,
0.10770545899868011,
0.11163999140262604,
0.036008477210998535,
0.14742304384708405,
0.027385149151086807,
0.03400014340877533,
-0.01210331916809082,
-0.12097682058811188,
-0.05635172873735428,
0.12943993508815765,
-0.056133102625608444,
-0.018430642783641815,
-0.03812691941857338,
0.05159873142838478,
0.027900202199816704,
0.1161862313747406,
0.06095987185835838,
-0.2653510272502899,
-0.07872358709573746,
0.09035108983516693,
0.035855699330568314,
-0.02870110422372818,
0.015685809776186943,
0.0563168004155159,
-0.09826799482107162,
0.13387839496135712,
0.01600331813097,
0.11293358355760574,
0.0005704759387299418,
-0.03738722205162048,
-0.05048390105366707,
0.023333828896284103,
-0.0029692542739212513,
0.10265000909566879,
-0.2580508291721344,
0.09134288877248764,
0.04307052493095398,
0.10210880637168884,
-0.053227588534355164,
0.041553471237421036,
0.0630747452378273,
0.18174763023853302,
0.12308064103126526,
0.0646238774061203,
-0.07038266211748123,
-0.029462512582540512,
-0.053122926503419876,
0.006267949938774109,
-0.06924857944250107,
0.043330222368240356,
0.03279302641749382,
0.018834590911865234,
-0.006911795120686293,
0.01673041842877865,
0.17025436460971832,
-0.1999993771314621,
-0.10901730507612228,
0.022742029279470444,
0.08338017016649246,
0.023600423708558083,
-0.07317411154508591,
-0.021892478689551353,
-0.05555485934019089,
0.12087453156709671,
-0.007980003021657467,
-0.08701952546834946,
-0.1190907210111618,
-0.009543485939502716,
0.08019476383924484,
-0.06377124041318893,
0.001245790277607739,
-0.02266741916537285,
0.08843941986560822,
-0.047269757837057114,
-0.09571850299835205,
0.08352566510438919,
-0.15405316650867462,
-0.08375263214111328,
-0.026018399745225906,
0.06235853582620621,
-0.01647355407476425,
0.03609852492809296,
0.05386598780751228,
0.005901367403566837,
-0.006306334398686886,
-0.10824448615312576,
0.0567815788090229,
0.13575860857963562,
0.06376583874225616,
0.006705634295940399,
-0.03548886254429817,
-0.07678928226232529,
0.018056027591228485,
-0.09668073058128357,
0.07724794745445251,
0.17467913031578064,
-0.02683734893798828,
0.07010143995285034,
0.19713599979877472,
-0.1365530639886856,
-0.18592897057533264,
-0.007658097427338362,
-0.029647447168827057,
-0.007356907706707716,
-0.030482197180390358,
-0.07130676507949829,
0.15280866622924805,
0.07955344021320343,
-0.03322422131896019,
0.05671115964651108,
-0.22575564682483673,
-0.12640368938446045,
0.05344041436910629,
0.07241325825452805,
0.17226947844028473,
-0.18070575594902039,
-0.07691451162099838,
-0.10442972183227539,
-0.15459731221199036,
0.15138885378837585,
-0.2559758424758911,
0.04302368685603142,
0.008492110297083855,
-0.027547379955649376,
-0.02233302965760231,
-0.057907961308956146,
0.11582447588443756,
0.026688702404499054,
0.045822881162166595,
-0.06647312641143799,
0.18558552861213684,
0.10044562816619873,
-0.06667258590459824,
0.18876045942306519,
-0.07752823829650879,
0.03576499596238136,
-0.1652960628271103,
-0.06526767462491989,
-0.03431292250752449,
0.12091987580060959,
-0.06318725645542145,
-0.07767423242330551,
-0.017039747908711433,
0.004848269280046225,
0.10286883264780045,
-0.015103135257959366,
-0.01762903854250908,
-0.004158231429755688,
0.04070555046200752,
0.15446634590625763,
0.1409449577331543,
-0.02884797751903534,
-0.07665041089057922,
-0.03389711678028107,
-0.041564345359802246,
0.06934721022844315,
-0.16820146143436432,
0.025597866624593735,
0.04682541638612747,
0.012246322818100452,
0.09923788160085678,
-0.0034461584873497486,
-0.06952887028455734,
-0.003821407677605748,
0.058685071766376495,
-0.08973260223865509,
-0.06127842515707016,
0.0011404326651245356,
0.08188453316688538,
0.007274127099663019,
0.028495755046606064,
0.12031800299882889,
-0.09629704803228378,
-0.007402781862765551,
-0.018648935481905937,
0.05303221568465233,
0.0044978223741054535,
0.15131410956382751,
0.03246035799384117,
0.04442526400089264,
-0.07988431304693222,
0.13373389840126038,
0.04409707710146904,
-0.09173049032688141,
0.022111138328909874,
0.11637374758720398,
-0.1119455099105835,
-0.10497206449508667,
0.03208277001976967,
0.02808602713048458,
-0.14950169622898102,
-0.11406315863132477,
0.017979292199015617,
-0.03486168757081032,
0.011943262070417404,
0.11074745655059814,
0.010196712799370289,
-0.020253846421837807,
0.01702350378036499,
-0.04180239140987396,
-0.04338240623474121,
0.03600688278675079,
0.015335525386035442,
0.06041274219751358,
-0.1081443801522255,
0.09670546650886536,
-0.022075925022363663,
-0.04748301953077316,
-0.021460220217704773,
0.02786090411245823,
-0.055428966879844666,
-0.013590292073786259,
-0.1580316722393036,
0.07021356374025345,
-0.12042330950498581,
-0.006026202347129583,
-0.04351057484745979,
-0.04700944572687149,
-0.040614478290081024,
0.03709770366549492,
-0.06896698474884033,
-0.01095614768564701,
-0.03400997072458267,
0.07329722493886948,
-0.12601618468761444,
-0.05632153898477554,
0.021903082728385925,
-0.0540698878467083,
0.04742541164159775,
0.006852814462035894,
-0.05738476663827896,
0.0022225105203688145,
-0.16498029232025146,
-0.016133029013872147,
0.03768031671643257,
0.03415520116686821,
-0.024371270090341568,
-0.04730493575334549,
0.01934507116675377,
0.1070832684636116,
0.022649727761745453,
0.028321275487542152,
0.004902216140180826,
-0.0789564773440361,
-0.04726323485374451,
-0.022000037133693695,
-0.0002515676897019148,
-0.05372808501124382,
0.019425619393587112,
0.07964776456356049,
0.028401488438248634,
0.1631712019443512,
-0.09945983439683914,
-0.015611539594829082,
-0.15366345643997192,
0.0061886487528681755,
-0.005644343793392181,
-0.04487114027142525,
-0.1126706451177597,
-0.0017370700370520353,
0.03491324186325073,
0.0016105493996292353,
0.20430198311805725,
-0.0035914196632802486,
0.02605121210217476,
0.028506387025117874,
0.03371834382414818,
0.08021865785121918,
-0.04278986155986786,
0.23263175785541534,
0.045185938477516174,
0.04929738491773605,
0.027799436822533607,
0.06587721407413483,
0.08665909618139267,
-0.013986269012093544,
0.11920662224292755,
0.10431364923715591,
0.04458091780543327,
0.1317189484834671,
-0.0330849289894104,
-0.03633591905236244,
-0.05951883643865585,
-0.139553502202034,
-0.011038719676434994,
0.10032914578914642,
-0.058573734015226364,
0.061237581074237823,
0.15546081960201263,
-0.1094222143292427,
0.00029526720754802227,
-0.05020494759082794,
-0.047300174832344055,
-0.16036483645439148,
-0.2048703283071518,
-0.04618433117866516,
-0.06978347152471542,
0.0021671676076948643,
-0.08850673586130142,
0.0010969856521114707,
0.019703274592757225,
0.04172495752573013,
-0.01795741729438305,
0.022129112854599953,
-0.062182534486055374,
-0.04992608726024628,
-0.0038859383203089237,
-0.044573619961738586,
0.00804160162806511,
0.009728200733661652,
-0.010472031310200691,
0.06481392681598663,
-0.04569113999605179,
0.01212249044328928,
0.05519098788499832,
0.06778331845998764,
0.08523581922054291,
0.005087598226964474,
-0.07198953628540039,
-0.060971569269895554,
0.07240062952041626,
0.033219240605831146,
0.14512647688388824,
0.05249367654323578,
-0.04299930855631828,
0.04098387062549591,
0.13578522205352783,
-0.03786128759384155,
-0.1239066794514656,
-0.07485013455152512,
0.13918134570121765,
-0.04714304953813553,
-0.008133641444146633,
-0.0012509443331509829,
-0.09781021624803543,
0.06198430806398392,
0.20039379596710205,
0.13963142037391663,
0.0544414222240448,
-0.040173232555389404,
-0.06456339359283447,
-0.016225317493081093,
-0.015186852775514126,
0.09778079390525818,
0.07873106747865677,
0.19204363226890564,
-0.06638165563344955,
0.0798153355717659,
-0.023586496710777283,
-0.022792618721723557,
-0.07924088835716248,
0.09859365224838257,
-0.057488784193992615,
-0.029472600668668747,
0.029839154332876205,
0.07973560690879822,
-0.013292107731103897,
-0.11271389573812485,
-0.018224362283945084,
0.02449192851781845,
-0.029809240251779556,
-0.0463034026324749,
0.0037732908967882395,
0.05646659433841705,
0.01754075475037098,
-0.047449998557567596,
0.020098742097616196,
0.00936143659055233,
-0.018275367096066475,
-0.22391483187675476,
-0.046977441757917404,
0.025957489386200905,
0.058329518884420395,
0.21760191023349762,
0.021345090121030807,
0.08637376874685287,
0.09719894826412201,
-0.03183210641145706,
-0.16923293471336365,
0.10072536766529083,
0.012195917777717113,
-0.04196782782673836,
0.050658609718084335,
0.05205957219004631,
-0.044691238552331924,
0.08827175199985504,
0.030940726399421692,
-0.07163619250059128,
0.0061448668129742146,
0.1098189577460289,
-0.03474125266075134,
-0.13109435141086578,
0.0151645103469491,
-0.11254815757274628,
0.13195832073688507,
0.05872110277414322,
-0.060019802302122116,
-0.01173887774348259,
-0.0960163027048111,
0.0823776125907898,
-0.0061384704895317554,
-0.07605397701263428,
0.03436686098575592,
-0.10571382194757462,
0.04027900844812393,
0.02301301434636116,
0.08168789744377136,
-0.22508004307746887,
-0.021684855222702026,
-0.06253199279308319,
-0.05406441166996956,
-0.10390469431877136,
0.08197741210460663,
0.12223770469427109,
0.006449530366808176,
-0.03576577827334404,
-0.14249970018863678,
-0.014823384582996368,
0.08004111051559448,
-0.062110815197229385,
-0.09766340255737305
] |
null | null | transformers |

# Polka-1.1B-Chat
`eryk-mazus/polka-1.1b-chat` **is the first Polish model trained to act as a helpful, conversational assistant that can be run locally.**
The model is based on [TinyLlama-1.1B](https://huggingface.co/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T) with the custom, extended tokenizer for more efficient Polish text generation, that was additionally pretrained on 5.7 billion tokens. **It was then fine-tuned on around 60k synthetically generated and machine-translated multi-turn conversations with the [Direct Preference Optimization (DPO)](https://arxiv.org/abs/2305.18290) performed on top of it.**
Context size: 4,096 tokens
In addition, we're releasing:
* [polka-1.1b](https://huggingface.co/eryk-mazus/polka-1.1b) - our base model with an extended tokenizer and additional pre-training on Polish corpus sampled using [DSIR](https://github.com/p-lambda/dsir)
* [polka-pretrain-en-pl-v1](https://huggingface.co/datasets/eryk-mazus/polka-pretrain-en-pl-v1) - the pre-training dataset
* [polka-dpo-v1](https://huggingface.co/datasets/eryk-mazus/polka-dpo-v1) - dataset of DPO pairs
* [polka-1.1b-chat-gguf](https://huggingface.co/eryk-mazus/polka-1.1b-chat-gguf) - GGUF files for the chat model
## Usage
Sample code:
```python
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM, TextStreamer
model_name = "eryk-mazus/polka-1.1b-chat"
tokenizer = AutoTokenizer.from_pretrained(model_name, use_fast=True)
tokenizer.pad_token = tokenizer.eos_token
model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype=torch.bfloat16 if torch.cuda.is_bf16_supported() else torch.float16,
device_map="auto"
)
streamer = TextStreamer(tokenizer, skip_prompt=True)
# You are a helpful assistant.
system_prompt = "Jesteś pomocnym asystentem."
chat = [{"role": "system", "content": system_prompt}]
# Compose a short song on programming.
user_input = "Napisz krótką piosenkę o programowaniu."
chat.append({"role": "user", "content": user_input})
# Generate - add_generation_prompt to make sure it continues as assistant
inputs = tokenizer.apply_chat_template(chat, add_generation_prompt=True, return_tensors="pt")
# For multi-GPU, find the device of the first parameter of the model
first_param_device = next(model.parameters()).device
inputs = inputs.to(first_param_device)
with torch.no_grad():
outputs = model.generate(
inputs,
pad_token_id=tokenizer.eos_token_id,
max_new_tokens=512,
temperature=0.2,
repetition_penalty=1.15,
top_p=0.95,
do_sample=True,
streamer=streamer,
)
# Add just the new tokens to our chat
new_tokens = outputs[0, inputs.size(1):]
response = tokenizer.decode(new_tokens, skip_special_tokens=True)
chat.append({"role": "assistant", "content": response})
```
The model works seamlessly with [vLLM](https://github.com/vllm-project/vllm) as well.
## Prompt format
This model uses ChatML as the prompt format:
```
<|im_start|>system
Jesteś pomocnym asystentem.
<|im_start|>user
Jakie jest dzienne zapotrzebowanie kaloryczne dorosłej osoby?<|im_end|>
<|im_start|>assistant
Dla dorosłych osób zaleca się spożywanie około 2000-3000 kcal dziennie, aby utrzymać optymalne zdrowie i dobre samopoczucie.<|im_end|>
```
This prompt is available as a [chat template](https://huggingface.co/docs/transformers/chat_templating), which means you can format messages using the `tokenizer.apply_chat_template()` method, as demonstrated in the example above.
***
We've actively looking for additional compute to train better and larger models for this project. If you want to collaborate, please reach out at: eryk.mazus at gmail dot com | {"language": ["pl"], "license": "mit", "tags": ["generated_from_trainer", "conversational", "polish"], "datasets": ["eryk-mazus/polka-dpo-v1"], "pipeline_tag": "text-generation", "inference": false} | text-generation | eryk-mazus/polka-1.1b-chat | [
"transformers",
"safetensors",
"llama",
"text-generation",
"generated_from_trainer",
"conversational",
"polish",
"pl",
"dataset:eryk-mazus/polka-dpo-v1",
"arxiv:2305.18290",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T16:33:52+00:00 | [
"2305.18290"
] | [
"pl"
] | TAGS
#transformers #safetensors #llama #text-generation #generated_from_trainer #conversational #polish #pl #dataset-eryk-mazus/polka-dpo-v1 #arxiv-2305.18290 #license-mit #autotrain_compatible #text-generation-inference #region-us
|
!image/png
# Polka-1.1B-Chat
'eryk-mazus/polka-1.1b-chat' is the first Polish model trained to act as a helpful, conversational assistant that can be run locally.
The model is based on TinyLlama-1.1B with the custom, extended tokenizer for more efficient Polish text generation, that was additionally pretrained on 5.7 billion tokens. It was then fine-tuned on around 60k synthetically generated and machine-translated multi-turn conversations with the Direct Preference Optimization (DPO) performed on top of it.
Context size: 4,096 tokens
In addition, we're releasing:
* polka-1.1b - our base model with an extended tokenizer and additional pre-training on Polish corpus sampled using DSIR
* polka-pretrain-en-pl-v1 - the pre-training dataset
* polka-dpo-v1 - dataset of DPO pairs
* polka-1.1b-chat-gguf - GGUF files for the chat model
## Usage
Sample code:
The model works seamlessly with vLLM as well.
## Prompt format
This model uses ChatML as the prompt format:
This prompt is available as a chat template, which means you can format messages using the 'tokenizer.apply_chat_template()' method, as demonstrated in the example above.
*
We've actively looking for additional compute to train better and larger models for this project. If you want to collaborate, please reach out at: URL at gmail dot com | [
"# Polka-1.1B-Chat\n\n'eryk-mazus/polka-1.1b-chat' is the first Polish model trained to act as a helpful, conversational assistant that can be run locally.\n\nThe model is based on TinyLlama-1.1B with the custom, extended tokenizer for more efficient Polish text generation, that was additionally pretrained on 5.7 billion tokens. It was then fine-tuned on around 60k synthetically generated and machine-translated multi-turn conversations with the Direct Preference Optimization (DPO) performed on top of it.\n\nContext size: 4,096 tokens\n\nIn addition, we're releasing:\n* polka-1.1b - our base model with an extended tokenizer and additional pre-training on Polish corpus sampled using DSIR\n* polka-pretrain-en-pl-v1 - the pre-training dataset\n* polka-dpo-v1 - dataset of DPO pairs\n* polka-1.1b-chat-gguf - GGUF files for the chat model",
"## Usage\n\nSample code:\n\n\n\nThe model works seamlessly with vLLM as well.",
"## Prompt format\n\nThis model uses ChatML as the prompt format:\n\n\nThis prompt is available as a chat template, which means you can format messages using the 'tokenizer.apply_chat_template()' method, as demonstrated in the example above.\n\n*\nWe've actively looking for additional compute to train better and larger models for this project. If you want to collaborate, please reach out at: URL at gmail dot com"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #generated_from_trainer #conversational #polish #pl #dataset-eryk-mazus/polka-dpo-v1 #arxiv-2305.18290 #license-mit #autotrain_compatible #text-generation-inference #region-us \n",
"# Polka-1.1B-Chat\n\n'eryk-mazus/polka-1.1b-chat' is the first Polish model trained to act as a helpful, conversational assistant that can be run locally.\n\nThe model is based on TinyLlama-1.1B with the custom, extended tokenizer for more efficient Polish text generation, that was additionally pretrained on 5.7 billion tokens. It was then fine-tuned on around 60k synthetically generated and machine-translated multi-turn conversations with the Direct Preference Optimization (DPO) performed on top of it.\n\nContext size: 4,096 tokens\n\nIn addition, we're releasing:\n* polka-1.1b - our base model with an extended tokenizer and additional pre-training on Polish corpus sampled using DSIR\n* polka-pretrain-en-pl-v1 - the pre-training dataset\n* polka-dpo-v1 - dataset of DPO pairs\n* polka-1.1b-chat-gguf - GGUF files for the chat model",
"## Usage\n\nSample code:\n\n\n\nThe model works seamlessly with vLLM as well.",
"## Prompt format\n\nThis model uses ChatML as the prompt format:\n\n\nThis prompt is available as a chat template, which means you can format messages using the 'tokenizer.apply_chat_template()' method, as demonstrated in the example above.\n\n*\nWe've actively looking for additional compute to train better and larger models for this project. If you want to collaborate, please reach out at: URL at gmail dot com"
] | [
87,
244,
21,
97
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #generated_from_trainer #conversational #polish #pl #dataset-eryk-mazus/polka-dpo-v1 #arxiv-2305.18290 #license-mit #autotrain_compatible #text-generation-inference #region-us \n# Polka-1.1B-Chat\n\n'eryk-mazus/polka-1.1b-chat' is the first Polish model trained to act as a helpful, conversational assistant that can be run locally.\n\nThe model is based on TinyLlama-1.1B with the custom, extended tokenizer for more efficient Polish text generation, that was additionally pretrained on 5.7 billion tokens. It was then fine-tuned on around 60k synthetically generated and machine-translated multi-turn conversations with the Direct Preference Optimization (DPO) performed on top of it.\n\nContext size: 4,096 tokens\n\nIn addition, we're releasing:\n* polka-1.1b - our base model with an extended tokenizer and additional pre-training on Polish corpus sampled using DSIR\n* polka-pretrain-en-pl-v1 - the pre-training dataset\n* polka-dpo-v1 - dataset of DPO pairs\n* polka-1.1b-chat-gguf - GGUF files for the chat model## Usage\n\nSample code:\n\n\n\nThe model works seamlessly with vLLM as well.## Prompt format\n\nThis model uses ChatML as the prompt format:\n\n\nThis prompt is available as a chat template, which means you can format messages using the 'tokenizer.apply_chat_template()' method, as demonstrated in the example above.\n\n*\nWe've actively looking for additional compute to train better and larger models for this project. If you want to collaborate, please reach out at: URL at gmail dot com"
] | [
-0.08273912966251373,
-0.003965835087001324,
-0.00449805473908782,
0.003659808076918125,
0.07618175446987152,
0.00016324703756254166,
0.1439794898033142,
0.0951400026679039,
0.003750921692699194,
0.06040632724761963,
0.027985593304038048,
0.038818612694740295,
0.05469865724444389,
0.18518874049186707,
0.09097550809383392,
-0.2699148654937744,
0.07215029001235962,
-0.05587564781308174,
-0.03551819175481796,
0.07291437685489655,
0.13630953431129456,
-0.08613952249288559,
0.04477264732122421,
-0.004477477632462978,
-0.04251640662550926,
0.018131239339709282,
-0.06270356476306915,
-0.04315338283777237,
0.04575018957257271,
0.013508168049156666,
0.049899544566869736,
-0.010189641267061234,
0.05218726769089699,
-0.13184954226016998,
0.011093567125499249,
0.08699352294206619,
0.04486449435353279,
0.048463061451911926,
0.09845517575740814,
0.0276996660977602,
0.22637073695659637,
-0.14046666026115417,
0.03957784175872803,
0.019160039722919464,
-0.041897859424352646,
-0.1230645701289177,
-0.07626709342002869,
0.07041642814874649,
0.08232951909303665,
0.07137782871723175,
-0.049799222499132156,
0.17660589516162872,
-0.02223450317978859,
0.05449843779206276,
-0.0038896524347364902,
-0.21654930710792542,
-0.048737820237874985,
0.0920703262090683,
0.06012338027358055,
0.08869586884975433,
-0.056418828666210175,
-0.07050519436597824,
-0.02486901357769966,
-0.009562777355313301,
-0.07446164637804031,
-0.03645233437418938,
0.06553322821855545,
-0.07440199702978134,
-0.15490856766700745,
-0.025406379252672195,
0.2876354157924652,
0.007997294887900352,
-0.08686116337776184,
-0.1596451848745346,
-0.02422369457781315,
0.015216481871902943,
-0.011767330579459667,
-0.025938600301742554,
0.026184245944023132,
0.01416011806577444,
0.02262112870812416,
-0.015350832603871822,
-0.11994842439889908,
-0.007356786169111729,
-0.008068743161857128,
0.12530329823493958,
0.027297643944621086,
0.02734454721212387,
-0.014176275581121445,
0.03226150572299957,
-0.02182244136929512,
-0.0608307346701622,
-0.016122279688715935,
-0.07537779957056046,
-0.09847000986337662,
-0.06746483594179153,
-0.05960353836417198,
-0.20385536551475525,
0.04518963396549225,
0.17092116177082062,
0.026366515085101128,
0.0642910748720169,
0.03286328166723251,
0.028688546270132065,
0.1217912882566452,
0.07404688000679016,
-0.09344888478517532,
0.05149640887975693,
0.0026854409370571375,
-0.058875326067209244,
0.06187157332897186,
0.040463317185640335,
-0.07329142093658447,
-0.01141287013888359,
-0.023354491218924522,
0.0008859719382598996,
0.02597895637154579,
0.07965084165334702,
0.015799779444932938,
-0.1018119528889656,
0.06120923534035683,
-0.12053443491458893,
-0.019387900829315186,
0.004346772097051144,
-0.0646962821483612,
0.1062498688697815,
-0.008751428686082363,
0.010492815636098385,
-0.09041455388069153,
0.04704240709543228,
-0.010271655395627022,
-0.004045618698000908,
-0.06058612838387489,
-0.12122492492198944,
0.051883671432733536,
0.0068145147524774075,
-0.02029755711555481,
-0.1406364142894745,
-0.07134713232517242,
-0.05293552577495575,
0.06981248408555984,
-0.01825064606964588,
-0.02389722317457199,
-0.03226547688245773,
-0.004894920624792576,
0.006733774673193693,
0.013811658136546612,
0.04823380336165428,
-0.02769852988421917,
0.058451097458601,
-0.11981216073036194,
0.04712250083684921,
-0.014635729603469372,
0.010382551699876785,
-0.12445144355297089,
-0.029174242168664932,
-0.2638566792011261,
0.08057851344347,
-0.037776604294776917,
0.05282418429851532,
-0.10609043389558792,
-0.012013738974928856,
-0.04488212987780571,
-0.003924683202058077,
0.043306611478328705,
0.13737629354000092,
-0.18364493548870087,
0.024620184674859047,
0.12132144719362259,
-0.1270851343870163,
0.009452496655285358,
0.08930313587188721,
-0.0013525463873520494,
0.1253354847431183,
0.06637247651815414,
0.2470482736825943,
0.1337313950061798,
-0.1308099925518036,
-0.0464143343269825,
0.019313622266054153,
-0.048830900341272354,
0.023618947714567184,
0.07930722832679749,
-0.07376907020807266,
0.05666846036911011,
0.00952586904168129,
-0.05175460875034332,
0.022212760522961617,
-0.0071721384301781654,
-0.02061847783625126,
-0.007690620608627796,
-0.03457414731383324,
0.004753364250063896,
0.012705937027931213,
0.001426342292688787,
-0.062120456248521805,
-0.09314271807670593,
0.0003924741758964956,
0.1636781394481659,
-0.036529071629047394,
0.021624166518449783,
-0.09197795391082764,
0.007433807943016291,
-0.06436630338430405,
0.01969161443412304,
-0.15276914834976196,
-0.01739874668419361,
0.055583272129297256,
-0.02210049144923687,
0.03424079343676567,
0.022958258166909218,
0.08004452288150787,
0.007573510054498911,
0.025317145511507988,
-0.01166832260787487,
-0.004779485519975424,
-0.031229162588715553,
-0.049096688628196716,
-0.1198543831706047,
-0.07301818579435349,
-0.05095897987484932,
0.11011062562465668,
-0.03462432697415352,
0.012205777689814568,
0.02701430954039097,
0.11437804996967316,
0.008314601145684719,
-0.06405419856309891,
0.028839748352766037,
0.00908930879086256,
0.012462605722248554,
-0.061246681958436966,
0.06519683450460434,
0.08333362638950348,
0.02014809474349022,
0.026181448251008987,
-0.09344809502363205,
-0.18103453516960144,
0.08879762887954712,
0.04998372122645378,
-0.0964835062623024,
-0.07369614392518997,
-0.047548335045576096,
0.013182415626943111,
-0.024837860837578773,
-0.06034252420067787,
0.10372000187635422,
0.011135216802358627,
0.05008602887392044,
-0.04426691308617592,
-0.040569860488176346,
0.018099313601851463,
-0.03289884701371193,
-0.04403186962008476,
0.0818798840045929,
0.053132835775613785,
-0.0873103216290474,
0.05413360148668289,
-0.014975281432271004,
0.08140140026807785,
0.20000126957893372,
-0.06340289115905762,
-0.1184471920132637,
-0.03161878511309624,
0.10286808013916016,
0.01317842211574316,
0.09733746200799942,
-0.1334386020898819,
0.014402183704078197,
0.03550305217504501,
0.03969769924879074,
0.036126211285591125,
-0.022357549518346786,
0.05233754962682724,
0.008862562477588654,
-0.03300746530294418,
0.0036952372174710035,
0.034518029540777206,
-0.001815463649109006,
0.09055407345294952,
0.03982096165418625,
0.07713568210601807,
-0.015432610176503658,
-0.039800964295864105,
-0.0915534719824791,
0.11281697452068329,
-0.11236561834812164,
-0.14286865293979645,
-0.11727980524301529,
0.004735963419079781,
-0.05223643407225609,
0.03120242990553379,
0.012742063961923122,
-0.0969022661447525,
-0.053423792123794556,
-0.05228080227971077,
0.11061934381723404,
-0.02268052101135254,
-0.004598450846970081,
-0.0696701854467392,
0.06215521693229675,
0.05123482272028923,
-0.11994766443967819,
0.016743091866374016,
-0.021098440513014793,
-0.06065861135721207,
0.03919105604290962,
0.015304570086300373,
0.047246359288692474,
0.0812452957034111,
-0.012224443256855011,
0.018277272582054138,
0.016098961234092712,
0.13820965588092804,
-0.09714162349700928,
0.05525987967848778,
0.11497525870800018,
0.019573915749788284,
-0.005652518942952156,
0.11493588984012604,
-0.009240119718015194,
-0.022242644801735878,
0.059393174946308136,
0.0003106806834693998,
-0.03655213490128517,
-0.1517212986946106,
-0.13484570384025574,
-0.04193302243947983,
0.0341176763176918,
0.10409894585609436,
0.04340029135346413,
-0.14936010539531708,
0.06386607140302658,
-0.08819716423749924,
-0.020938711240887642,
0.028937270864844322,
0.07232043147087097,
0.031785935163497925,
-0.003777695819735527,
0.03241037204861641,
-0.0731046199798584,
-0.06057340279221535,
0.08542568236589432,
-0.008861393667757511,
0.17824417352676392,
-0.035738490521907806,
0.08261891454458237,
0.04789984971284866,
0.08650276064872742,
-0.007894724607467651,
0.11461316049098969,
0.026209669187664986,
0.03879605978727341,
0.017481351271271706,
-0.09072237461805344,
-0.03204677999019623,
0.01575423590838909,
0.015107648447155952,
-0.03960580751299858,
-0.08924250304698944,
-0.04364500939846039,
0.06379256397485733,
0.2176091969013214,
0.03144023194909096,
-0.1454017013311386,
-0.07204720377922058,
-0.003463789587840438,
-0.030242525041103363,
-0.06352722644805908,
-0.02597322128713131,
0.10760580003261566,
-0.1179785281419754,
0.048502057790756226,
-0.013397583737969398,
0.018480287864804268,
-0.10355378687381744,
-0.007269140332937241,
0.04575803875923157,
0.07334686815738678,
-0.01213457528501749,
0.09650690853595734,
-0.1566724330186844,
0.08138863742351532,
0.023525750264525414,
0.12279806286096573,
-0.09562399238348007,
0.03306424245238304,
-0.016356592997908592,
-0.057831522077322006,
0.11684495210647583,
0.029377078637480736,
0.011818052269518375,
-0.07886597514152527,
-0.14061787724494934,
0.028582239523530006,
0.08754684776067734,
-0.02836170792579651,
0.09317591786384583,
-0.02271290123462677,
0.023332633078098297,
-0.0492425337433815,
0.04644366353750229,
-0.1417306512594223,
-0.16027489304542542,
0.04918774217367172,
-0.02066909335553646,
0.013706741854548454,
-0.05345046520233154,
0.006308514624834061,
0.027428990229964256,
0.1830407828092575,
-0.012202030047774315,
-0.08899802714586258,
-0.0892716720700264,
-0.006310628727078438,
0.11828059703111649,
-0.03220729902386665,
0.007332052569836378,
0.00433721300214529,
0.13048261404037476,
-0.026734840124845505,
-0.10167686641216278,
0.017987804487347603,
-0.05094297602772713,
-0.03277680277824402,
0.011904079467058182,
0.12022987008094788,
0.10510373115539551,
0.015707524493336678,
0.026472825556993484,
0.02031678333878517,
-0.007381503004580736,
-0.07009556889533997,
-0.04134135693311691,
0.06627863645553589,
0.00720288185402751,
0.05960048735141754,
-0.114244244992733,
-0.017755024135112762,
-0.04093371331691742,
-0.050579994916915894,
0.10786627233028412,
0.16540725529193878,
-0.03904533013701439,
0.12396320700645447,
0.06426212936639786,
-0.06419940292835236,
-0.21073849499225616,
0.02204974740743637,
0.08406835049390793,
0.03819461911916733,
-0.07716318219900131,
-0.20693178474903107,
0.034303631633520126,
0.08798066526651382,
0.01666790060698986,
0.0287761390209198,
-0.3101402521133423,
-0.09633640199899673,
0.10949886590242386,
0.002657705917954445,
0.06672380864620209,
-0.1025390774011612,
-0.030112097039818764,
-0.05376284196972847,
0.048988983035087585,
0.13793514668941498,
-0.1381477564573288,
0.08580786734819412,
0.03533896803855896,
0.035610999912023544,
0.08434563875198364,
-0.021383684128522873,
0.10557443648576736,
0.05544913187623024,
-0.0010224275756627321,
-0.10333486646413803,
0.09593517333269119,
0.07074624300003052,
-0.033078067004680634,
0.1398986130952835,
0.06927859783172607,
-0.013779866509139538,
-0.18592968583106995,
-0.035789381712675095,
-0.044343557208776474,
0.03579125925898552,
-0.0333961546421051,
-0.07866302877664566,
-0.044249046593904495,
0.05443940684199333,
0.0325724221765995,
-0.016550246626138687,
-0.03318292647600174,
-0.04092468321323395,
0.003873493755236268,
0.06332162022590637,
0.13526605069637299,
-0.01982305757701397,
-0.060868941247463226,
0.07051510363817215,
-0.006783135700970888,
0.07024014741182327,
-0.022005844861268997,
0.004589373245835304,
0.06301821768283844,
0.0006152428104542196,
0.02379220724105835,
0.005299641750752926,
-0.12579768896102905,
-0.037844687700271606,
0.06076119467616081,
-0.13063500821590424,
-0.07887830585241318,
0.01927640289068222,
0.11153823882341385,
-0.12125509232282639,
0.018224013969302177,
0.14214232563972473,
-0.050968341529369354,
-0.015306923538446426,
-0.005251398775726557,
0.023040078580379486,
-0.03832532465457916,
0.1537145972251892,
0.0047228531911969185,
0.011817438527941704,
-0.05737627297639847,
0.14535623788833618,
0.0585198774933815,
-0.10918283462524414,
0.02047138288617134,
0.10624897480010986,
-0.10500096529722214,
-0.08572961390018463,
-0.01830679178237915,
0.026304034516215324,
-0.0014778503682464361,
-0.020732907578349113,
0.010710274800658226,
-0.08995316922664642,
0.012607530690729618,
-0.06334225833415985,
0.03691384568810463,
0.009990799240767956,
-0.0125834746286273,
-0.004723512101918459,
-0.1694517880678177,
0.044372186064720154,
0.08628804981708527,
-0.008016555570065975,
-0.041706498712301254,
0.1520763635635376,
-0.04865393415093422,
-0.030411064624786377,
-0.0010586368152871728,
-0.046670280396938324,
-0.09457731246948242,
-0.007236518431454897,
-0.014203961938619614,
-0.049453165382146835,
-0.06693527102470398,
-0.04468638822436333,
0.009824493899941444,
0.06707634031772614,
0.010365715250372887,
-0.010721112601459026,
-0.048108555376529694,
-0.035151589661836624,
-0.08502066135406494,
0.024091023951768875,
-0.04930851608514786,
0.008875293657183647,
-0.014495696872472763,
-0.09644651412963867,
0.10170380026102066,
0.09933283925056458,
0.004250870551913977,
-0.05850907042622566,
-0.11622339487075806,
-0.0030120681039988995,
-0.016708912327885628,
0.009961966425180435,
-0.008477292954921722,
-0.10900791734457016,
0.019471224397420883,
0.01662050560116768,
-0.010894077830016613,
0.00414300337433815,
0.13099415600299835,
-0.08403223752975464,
0.09296632558107376,
-0.010157707147300243,
0.01842442899942398,
-0.09262196719646454,
0.0498511865735054,
0.027434037998318672,
0.10756634175777435,
0.15248748660087585,
-0.09999336302280426,
0.01849859021604061,
-0.0958697497844696,
-0.006504926364868879,
0.003548697102814913,
0.01586013473570347,
0.050711490213871,
-0.019055930897593498,
0.05347010865807533,
0.006662073079496622,
0.0792412981390953,
0.03055768646299839,
0.01591394655406475,
0.054894451051950455,
-0.034444283694028854,
0.01563849113881588,
0.033948060125112534,
0.15535211563110352,
0.006373549345880747,
0.01809001713991165,
0.0018665792886167765,
0.05833031237125397,
0.029395626857876778,
0.032784976065158844,
0.05661468580365181,
0.16925959289073944,
0.03815428912639618,
0.03533713519573212,
0.020588399842381477,
-0.08629065752029419,
0.031148070469498634,
0.06454948335886002,
0.003686654381453991,
0.06131155416369438,
-0.07371518760919571,
0.16741321980953217,
0.04925913363695145,
-0.14967334270477295,
0.01572221890091896,
-0.027818424627184868,
-0.06745366752147675,
-0.05842633917927742,
-0.17055372893810272,
-0.01623351313173771,
-0.10507941246032715,
-0.02083786390721798,
-0.09858498722314835,
0.016996413469314575,
0.0115360077470541,
0.03965269774198532,
-0.017279211431741714,
0.06354758888483047,
-0.16170483827590942,
-0.10564494878053665,
0.07303185015916824,
-0.013888208195567131,
0.003991093952208757,
-0.0024557465221732855,
-0.04426704719662666,
0.0017020516097545624,
0.004224339500069618,
0.07117775827646255,
0.058419279754161835,
0.05060899630188942,
0.020375380292534828,
-0.0575362965464592,
-0.037052083760499954,
-0.003835340728983283,
-0.004159821663051844,
0.05574993044137955,
0.14988449215888977,
0.06469666957855225,
-0.08802225440740585,
-0.03738868236541748,
0.21350789070129395,
0.0008516860543750226,
-0.12777221202850342,
-0.1262105405330658,
-0.008798614144325256,
0.010073450393974781,
0.024803509935736656,
-0.010491599328815937,
-0.1119254007935524,
-0.026567354798316956,
0.15841296315193176,
0.17736904323101044,
0.04499954730272293,
-0.007307432126253843,
-0.01705983281135559,
-0.014567465521395206,
0.009846068918704987,
0.13364578783512115,
0.052139826118946075,
0.15455341339111328,
-0.043130677193403244,
0.13044387102127075,
-0.012994330376386642,
0.020800935104489326,
-0.10313376784324646,
0.11282999068498611,
-0.013067289255559444,
-0.025724755600094795,
-0.007094394415616989,
0.0973067432641983,
-0.0583014041185379,
-0.12537074089050293,
-0.0075793140567839146,
-0.01631559431552887,
-0.08430925011634827,
-0.0002123237936757505,
-0.034411221742630005,
0.014738231897354126,
0.11123405396938324,
-0.04589235410094261,
0.014290310442447662,
0.10326236486434937,
0.021620649844408035,
-0.08179891109466553,
-0.14199376106262207,
0.048773445188999176,
-0.07397980988025665,
0.18161816895008087,
-0.0032887367997318506,
0.07153769582509995,
0.0676192045211792,
0.010994473472237587,
-0.17480866611003876,
0.010236227884888649,
-0.0048577249981462955,
-0.08400978147983551,
0.06278913468122482,
0.06979759782552719,
-0.05395401269197464,
0.12548764050006866,
0.03410151228308678,
-0.12664830684661865,
0.029267624020576477,
0.06253572553396225,
0.08302515745162964,
-0.08244921267032623,
0.10277029871940613,
-0.12562218308448792,
0.13445937633514404,
0.15542945265769958,
-0.04468894004821777,
-0.006570967845618725,
-0.08404723554849625,
0.014574049040675163,
0.048467230051755905,
0.00009592444257577881,
-0.008349238894879818,
-0.20007644593715668,
-0.04326827451586723,
-0.04430367052555084,
0.02940765582025051,
-0.2004467248916626,
-0.051035407930612564,
-0.017452964559197426,
-0.008012580685317516,
-0.03561271354556084,
0.08524847030639648,
-0.02187514677643776,
0.034264761954545975,
-0.049291085451841354,
0.12361360341310501,
0.023323796689510345,
0.12148483097553253,
-0.0921144038438797,
-0.08061150461435318
] |
null | null | transformers | # MetricX-23
*This is not an officially supported Google product.*
**GitHub repository: [https://github.com/google-research/metricx](https://github.com/google-research/metricx)**
This repository contains the MetricX-23 models,
a family of models for automatic evaluation of translations that were proposed
in the WMT'23 Metrics Shared Task submission
[MetricX-23: The Google Submission to the WMT 2023 Metrics Shared Task](https://aclanthology.org/2023.wmt-1.63/).
The models were trained in [T5X](https://github.com/google-research/t5x) and
then converted for use in PyTorch.
## Available Models
There are 6 models available on HuggingFace that vary in the number of
parameters and whether or not the model is reference-based or reference-free
(also known as quality estimation, or QE):
* [MetricX-23-XXL](https://huggingface.co/google/metricx-23-large-v2p0)
* [MetricX-23-XL](https://huggingface.co/google/metricx-23-xl-v2p0)
* [MetricX-23-Large](https://huggingface.co/google/metricx-23-xxl-v2p0)
* [MetricX-23-QE-XXL](https://huggingface.co/google/metricx-23-qe-large-v2p0)
* [MetricX-23-QE-XL](https://huggingface.co/google/metricx-23-qe-xl-v2p0)
* [MetricX-23-QE-Large](https://huggingface.co/google/metricx-23-qe-xxl-v2p0)
We recommend using the XXL model versions for the best agreement with human
judgments of translation quality, the Large versions for best speed, and the
XL for an intermediate use case.
## Changes to the WMT'23 Submission
These models available here are most similar to the primary submission to the WMT'23 Metrics
Shared Task. They are initialized with [mT5](https://aclanthology.org/2021.naacl-main.41/)
then fine-tuned on a combination of direct assessment and MQM data. However,
we made some changes that make these models different from the WMT'23 submissions.
First, the models are trained to regress the actual MQM score rather than a
normalized score between 0 and 1. **That means the output from the MetricX-23
models is a score in the range [0, 25] where lower is better (i.e., it predicts
an error score).**
Second, these models were trained with a larger variety of synthetic data that
makes them more robust to translation edge cases like over- and undertranslation,
described in more detail in the following section.
### Synthetic Data
In order for our MetricX models to learn to identify certain types of bad
translations that are not sufficiently (or at all) represented in the regular
training data, we created synthetic examples and mixed them in during training.
The synthetic training data was generated from the DA datasets ranging from
WMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have
the candidate translation manipulated so as to turn it into a bad translation
with a specific issue commonly unrecognized by learned metrics.
The table below provides an overview of the various failure modes that we
considered, including brief descriptions of how we prepared the synthetic data
to address them.
| Failure mode | Synthetic example description |
| ----------- | ----------- |
| Undertranslation | Candidate translation with an arbitrary sentence removed (if multi-sentence); alternatively, candidate with a certain proportion of words removed from the end. |
| Overtranslation | Candidate translation duplicated (with space in between). |
| Fluent but unrelated translation | Arbitrary reference of a similar length from the dataset. |
| Gibberish | Text of a similar length as the reference, generated by sampling words from the reference translation vocabulary (built from all references in the data). |
| Missing punctuation | Reference translation with the end punctuation removed (11 punctuation symbols considered). |
| Latin instead of Chinese/Japanese or Hindi/Bengali punctuation | Candidate translation with the language-specific punctuation symbol at the end replaced with the Latin equivalent (e.g., "." instead of "。" or "।"); alternatively, the punctuation symbol is replaced with the Latin equivalent in the reference, keeping the correct one in the candidate. |
| Reference-matching translation | Reference translation copied as the candidate translation (unlike the rest of the synthetic data, these examples are meant to train the metric to predict a perfect score for candidates matching the reference). |
Examples from the first 4 categories were assigned a label corresponding to the
worst score on the given rating scale (e.g., 25 when mixed with MQM training
data), whereas the reference-matching translation examples are assigned the best
score (e.g., 0 when used with MQM data). The missing/incorrect punctuation
examples were labeled with a score slightly worse than perfect.
Note that some of the synthetic datasets are only meaningful in the
reference-based scenario, and we thus excluded them when training a QE variant
of MetricX. These are the Latin-vs-special punctuation and the
reference-matching translation examples.
Most of the synthetic training sets were created using stratified sampling
across target languages, taking 500 examples per target language. One exception
is the missing punctuation set, which used a stratified sample across different
punctuation symbols instead.
When training MetricX, a small proportion of the synthetic examples was mixed
with the regular training examples. During the first-stage fine-tuning on DA
data, each synthetic training set constituted between 0.1% and 1% of all
training examples, whereas in the second-stage fine-tuning on MQM data we used
an even smaller proportion, around 0.05%.
As for evaluating the effect of the synthetic training data on the model's
performance, the DEMETR challenge set - which we originally used to evaluate the
models submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We
therefore created a new DEMETR-style test set based on the WMT22 DA data, with
examples constructed analogically to the synthetic training examples, as
described above. This test set helped us determine the right proportions of
synthetic data for fine-tuning in order to make MetricX robust for the failure
modes in consideration, without sacrificing the system- and segment-level
correlations with human ratings.
## Usage
The code for using MetricX models can be found at [https://github.com/google-research/metricx](https://github.com/google-research/metricx).
The repository contains example prediction scripts, described below.
The `metricx23/predict.py` script contains an example for how to run inference
on the models.
### Reference-Based
Example usage for a reference-based model:
```bash
python -m metricx23.predict \
--tokenizer google/mt5-xl \
--model_name_or_path google/metricx-23-xl-v2p0 \
--max_input_length 1024 \
--batch_size 1 \
--input_file input.jsonl \
--output_file output.jsonl
```
`input.jsonl` is expected to have 1 serialized JSON object per line with
`"reference"` and `"hypothesis"` fields. The output jsonl will be parallel
to `input.jsonl` but additionally contain a `"prediction"` field with the predicted score.
Note that the model was trained with a maximum input length of 1024 tokens, so
significantly increasing that value may lead to unpredictable behavior.
### Reference-Free
Example usage for a reference-free model:
```bash
python -m metricx23.predict \
--tokenizer google/mt5-xl \
--model_name_or_path google/metricx-23-qe-xl-v2p0 \
--max_input_length 1024 \
--batch_size 1 \
--input_file input.jsonl \
--output_file output.jsonl \
--qe
```
`input.jsonl` is expected to have 1 serialized JSON object per line with
`"source"` and `"hypothesis"` fields. The output jsonl will be parallel
to `input.jsonl` but additionally contain a `"prediction"` field with the predicted score.
## Meta-Evaluation
The `metricx23/evaluate.py` script contains code to calculate various correlations
between the MetricX-23 scores and MQM ratings of translation quality using the
[MT Metrics Eval](https://github.com/google-research/mt-metrics-eval) library.
Example usage:
```bash
python -m metricx23.evaluate \
--dataset wmt22 \
--lp en-de \
--input_file input.jsonl \
--output_file output.json
```
`input.jsonl` is expected to have one JSON object serialized per line.
Each JSON object is expected to contain 4 fields:
* `"system_id"`: The name of the system that generated the translation.
* `"segment_id"`: The 0-based index of the corresponding segment in the MT
Metrics Eval data.
* `"label"`: The ground-truth translation quality score (with higher is better).
* `"prediction"`: The model predicted translation quality score (with lower is
better; the script negates the scores so higher is better).
The script will calculate the 4 agreement/correlations that were used in the
WMT'23 Shared Task. Below are the results for the MetricX-23 models on the
WMT'22 Metrics Shared Task data:
English-German:
| Model | System-Level Accuracy | System-Level Pearson | Segment-Level Pearson | Segment-Level Pairwise Acc |
| ----------- | ----------- | ----------- | ----------- | ----------- |
| MetricX-23-XXL | 0.795 | 0.835 | 0.546 | 0.619 |
| MetricX-23-XL | 0.756 | 0.813 | 0.540 | 0.605 |
| MetricX-23-Large | 0.769 | 0.759 | 0.507 | 0.595 |
| MetricX-23-QE-XXL | 0.769 | 0.830 | 0.490 | 0.606 |
| MetricX-23-QE-XL | 0.718 | 0.684 | 0.421 | 0.594 |
| MetricX-23-QE-Large | 0.744 | 0.671 | 0.387 | 0.579 |
English-Russian:
| Model | System-Level Accuracy | System-Level Pearson | Segment-Level Pearson | Segment-Level Pairwise Acc |
| ----------- | ----------- | ----------- | ----------- | ----------- |
| MetricX-23-XXL | 0.905 | 0.943 | 0.477 | 0.609 |
| MetricX-23-XL | 0.876 | 0.906 | 0.498 | 0.589 |
| MetricX-23-Large | 0.876 | 0.841 | 0.474 | 0.569 |
| MetricX-23-QE-XXL | 0.895 | 0.940 | 0.470 | 0.602 |
| MetricX-23-QE-XL | 0.848 | 0.861 | 0.415 | 0.570 |
| MetricX-23-QE-Large | 0.819 | 0.778 | 0.411 | 0.551 |
Chinese-English:
| Model | System-Level Accuracy | System-Level Pearson | Segment-Level Pearson | Segment-Level Pairwise Acc |
| ----------- | ----------- | ----------- | ----------- | ----------- |
| MetricX-23-XXL | 0.868 | 0.919 | 0.605 | 0.551 |
| MetricX-23-XL | 0.868 | 0.924 | 0.584 | 0.543 |
| MetricX-23-Large | 0.857 | 0.919 | 0.555 | 0.539 |
| MetricX-23-QE-XXL | 0.857 | 0.928 | 0.573 | 0.544 |
| MetricX-23-QE-XL | 0.802 | 0.879 | 0.546 | 0.529 |
| MetricX-23-QE-Large | 0.758 | 0.904 | 0.522 | 0.529 |
The `metricx23/evaluate_wmt23.py` script re-calculates the average correlation
score that was used to rank submissions from the
[WMT'23 Shared Task](https://www2.statmt.org/wmt23/pdf/2023.wmt-1.51.pdf).
Example usage:
```bash
python -m metricx23.evaluate_wmt23 \
--en_de predictions_ende.jsonl \
--he_en predictions_heen.jsonl \
--zh_en predictions_zhen.jsonl \
--output_file output.json
```
Each of the 3 input files is expected to be in the same format as described
above. Each file should correspond to running inference on each of the language
pairs from the WMT'23 dataset.
The results for each of the models is the following:
| Model | Average Correlation |
| ----------- | ----------- |
| MetricX-23-XXL | 0.812 |
| MetricX-23-XL | 0.813 |
| MetricX-23-Large | 0.794 |
| MetricX-23-QE-XXL | 0.797 |
| MetricX-23-QE-XL | 0.767 |
| MetricX-23-QE-Large | 0.762 |
## Citation
If you use MetricX-23 in your research, please cite the following publication:
```bibtex
@inproceedings{juraska-etal-2023-metricx,
title = {{MetricX-23: The Google Submission to the WMT 2023 Metrics Shared Task}},
author = "Juraska, Juraj and
Finkelstein, Mara and
Deutsch, Daniel and
Siddhant, Aditya and
Mirzazadeh, Mehdi and
Freitag, Markus",
editor = "Koehn, Philipp and
Haddow, Barry and
Kocmi, Tom and
Monz, Christof",
booktitle = "Proceedings of the Eighth Conference on Machine Translation",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.wmt-1.63",
doi = "10.18653/v1/2023.wmt-1.63",
pages = "756--767",
}
``` | {"license": "apache-2.0"} | null | google/metricx-23-xl-v2p0 | [
"transformers",
"pytorch",
"mt5",
"license:apache-2.0",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T16:34:17+00:00 | [] | [] | TAGS
#transformers #pytorch #mt5 #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us
| MetricX-23
==========
*This is not an officially supported Google product.*
GitHub repository: URL
This repository contains the MetricX-23 models,
a family of models for automatic evaluation of translations that were proposed
in the WMT'23 Metrics Shared Task submission
MetricX-23: The Google Submission to the WMT 2023 Metrics Shared Task.
The models were trained in T5X and
then converted for use in PyTorch.
Available Models
----------------
There are 6 models available on HuggingFace that vary in the number of
parameters and whether or not the model is reference-based or reference-free
(also known as quality estimation, or QE):
* MetricX-23-XXL
* MetricX-23-XL
* MetricX-23-Large
* MetricX-23-QE-XXL
* MetricX-23-QE-XL
* MetricX-23-QE-Large
We recommend using the XXL model versions for the best agreement with human
judgments of translation quality, the Large versions for best speed, and the
XL for an intermediate use case.
Changes to the WMT'23 Submission
--------------------------------
These models available here are most similar to the primary submission to the WMT'23 Metrics
Shared Task. They are initialized with mT5
then fine-tuned on a combination of direct assessment and MQM data. However,
we made some changes that make these models different from the WMT'23 submissions.
First, the models are trained to regress the actual MQM score rather than a
normalized score between 0 and 1. That means the output from the MetricX-23
models is a score in the range [0, 25] where lower is better (i.e., it predicts
an error score).
Second, these models were trained with a larger variety of synthetic data that
makes them more robust to translation edge cases like over- and undertranslation,
described in more detail in the following section.
### Synthetic Data
In order for our MetricX models to learn to identify certain types of bad
translations that are not sufficiently (or at all) represented in the regular
training data, we created synthetic examples and mixed them in during training.
The synthetic training data was generated from the DA datasets ranging from
WMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have
the candidate translation manipulated so as to turn it into a bad translation
with a specific issue commonly unrecognized by learned metrics.
The table below provides an overview of the various failure modes that we
considered, including brief descriptions of how we prepared the synthetic data
to address them.
Examples from the first 4 categories were assigned a label corresponding to the
worst score on the given rating scale (e.g., 25 when mixed with MQM training
data), whereas the reference-matching translation examples are assigned the best
score (e.g., 0 when used with MQM data). The missing/incorrect punctuation
examples were labeled with a score slightly worse than perfect.
Note that some of the synthetic datasets are only meaningful in the
reference-based scenario, and we thus excluded them when training a QE variant
of MetricX. These are the Latin-vs-special punctuation and the
reference-matching translation examples.
Most of the synthetic training sets were created using stratified sampling
across target languages, taking 500 examples per target language. One exception
is the missing punctuation set, which used a stratified sample across different
punctuation symbols instead.
When training MetricX, a small proportion of the synthetic examples was mixed
with the regular training examples. During the first-stage fine-tuning on DA
data, each synthetic training set constituted between 0.1% and 1% of all
training examples, whereas in the second-stage fine-tuning on MQM data we used
an even smaller proportion, around 0.05%.
As for evaluating the effect of the synthetic training data on the model's
performance, the DEMETR challenge set - which we originally used to evaluate the
models submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We
therefore created a new DEMETR-style test set based on the WMT22 DA data, with
examples constructed analogically to the synthetic training examples, as
described above. This test set helped us determine the right proportions of
synthetic data for fine-tuning in order to make MetricX robust for the failure
modes in consideration, without sacrificing the system- and segment-level
correlations with human ratings.
Usage
-----
The code for using MetricX models can be found at URL
The repository contains example prediction scripts, described below.
The 'metricx23/URL' script contains an example for how to run inference
on the models.
### Reference-Based
Example usage for a reference-based model:
'URL' is expected to have 1 serialized JSON object per line with
'"reference"' and '"hypothesis"' fields. The output jsonl will be parallel
to 'URL' but additionally contain a '"prediction"' field with the predicted score.
Note that the model was trained with a maximum input length of 1024 tokens, so
significantly increasing that value may lead to unpredictable behavior.
### Reference-Free
Example usage for a reference-free model:
'URL' is expected to have 1 serialized JSON object per line with
'"source"' and '"hypothesis"' fields. The output jsonl will be parallel
to 'URL' but additionally contain a '"prediction"' field with the predicted score.
Meta-Evaluation
---------------
The 'metricx23/URL' script contains code to calculate various correlations
between the MetricX-23 scores and MQM ratings of translation quality using the
MT Metrics Eval library.
Example usage:
'URL' is expected to have one JSON object serialized per line.
Each JSON object is expected to contain 4 fields:
* '"system\_id"': The name of the system that generated the translation.
* '"segment\_id"': The 0-based index of the corresponding segment in the MT
Metrics Eval data.
* '"label"': The ground-truth translation quality score (with higher is better).
* '"prediction"': The model predicted translation quality score (with lower is
better; the script negates the scores so higher is better).
The script will calculate the 4 agreement/correlations that were used in the
WMT'23 Shared Task. Below are the results for the MetricX-23 models on the
WMT'22 Metrics Shared Task data:
English-German:
English-Russian:
Chinese-English:
The 'metricx23/evaluate\_wmt23.py' script re-calculates the average correlation
score that was used to rank submissions from the
WMT'23 Shared Task.
Example usage:
Each of the 3 input files is expected to be in the same format as described
above. Each file should correspond to running inference on each of the language
pairs from the WMT'23 dataset.
The results for each of the models is the following:
If you use MetricX-23 in your research, please cite the following publication:
| [
"### Synthetic Data\n\n\nIn order for our MetricX models to learn to identify certain types of bad\ntranslations that are not sufficiently (or at all) represented in the regular\ntraining data, we created synthetic examples and mixed them in during training.\nThe synthetic training data was generated from the DA datasets ranging from\nWMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have\nthe candidate translation manipulated so as to turn it into a bad translation\nwith a specific issue commonly unrecognized by learned metrics.\n\n\nThe table below provides an overview of the various failure modes that we\nconsidered, including brief descriptions of how we prepared the synthetic data\nto address them.\n\n\n\nExamples from the first 4 categories were assigned a label corresponding to the\nworst score on the given rating scale (e.g., 25 when mixed with MQM training\ndata), whereas the reference-matching translation examples are assigned the best\nscore (e.g., 0 when used with MQM data). The missing/incorrect punctuation\nexamples were labeled with a score slightly worse than perfect.\n\n\nNote that some of the synthetic datasets are only meaningful in the\nreference-based scenario, and we thus excluded them when training a QE variant\nof MetricX. These are the Latin-vs-special punctuation and the\nreference-matching translation examples.\n\n\nMost of the synthetic training sets were created using stratified sampling\nacross target languages, taking 500 examples per target language. One exception\nis the missing punctuation set, which used a stratified sample across different\npunctuation symbols instead.\n\n\nWhen training MetricX, a small proportion of the synthetic examples was mixed\nwith the regular training examples. During the first-stage fine-tuning on DA\ndata, each synthetic training set constituted between 0.1% and 1% of all\ntraining examples, whereas in the second-stage fine-tuning on MQM data we used\nan even smaller proportion, around 0.05%.\n\n\nAs for evaluating the effect of the synthetic training data on the model's\nperformance, the DEMETR challenge set - which we originally used to evaluate the\nmodels submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We\ntherefore created a new DEMETR-style test set based on the WMT22 DA data, with\nexamples constructed analogically to the synthetic training examples, as\ndescribed above. This test set helped us determine the right proportions of\nsynthetic data for fine-tuning in order to make MetricX robust for the failure\nmodes in consideration, without sacrificing the system- and segment-level\ncorrelations with human ratings.\n\n\nUsage\n-----\n\n\nThe code for using MetricX models can be found at URL\nThe repository contains example prediction scripts, described below.\n\n\nThe 'metricx23/URL' script contains an example for how to run inference\non the models.",
"### Reference-Based\n\n\nExample usage for a reference-based model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"reference\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nNote that the model was trained with a maximum input length of 1024 tokens, so\nsignificantly increasing that value may lead to unpredictable behavior.",
"### Reference-Free\n\n\nExample usage for a reference-free model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"source\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nMeta-Evaluation\n---------------\n\n\nThe 'metricx23/URL' script contains code to calculate various correlations\nbetween the MetricX-23 scores and MQM ratings of translation quality using the\nMT Metrics Eval library.\n\n\nExample usage:\n\n\n'URL' is expected to have one JSON object serialized per line.\nEach JSON object is expected to contain 4 fields:\n\n\n* '\"system\\_id\"': The name of the system that generated the translation.\n* '\"segment\\_id\"': The 0-based index of the corresponding segment in the MT\nMetrics Eval data.\n* '\"label\"': The ground-truth translation quality score (with higher is better).\n* '\"prediction\"': The model predicted translation quality score (with lower is\nbetter; the script negates the scores so higher is better).\n\n\nThe script will calculate the 4 agreement/correlations that were used in the\nWMT'23 Shared Task. Below are the results for the MetricX-23 models on the\nWMT'22 Metrics Shared Task data:\n\n\nEnglish-German:\n\n\n\nEnglish-Russian:\n\n\n\nChinese-English:\n\n\n\nThe 'metricx23/evaluate\\_wmt23.py' script re-calculates the average correlation\nscore that was used to rank submissions from the\nWMT'23 Shared Task.\n\n\nExample usage:\n\n\nEach of the 3 input files is expected to be in the same format as described\nabove. Each file should correspond to running inference on each of the language\npairs from the WMT'23 dataset.\n\n\nThe results for each of the models is the following:\n\n\n\nIf you use MetricX-23 in your research, please cite the following publication:"
] | [
"TAGS\n#transformers #pytorch #mt5 #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us \n",
"### Synthetic Data\n\n\nIn order for our MetricX models to learn to identify certain types of bad\ntranslations that are not sufficiently (or at all) represented in the regular\ntraining data, we created synthetic examples and mixed them in during training.\nThe synthetic training data was generated from the DA datasets ranging from\nWMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have\nthe candidate translation manipulated so as to turn it into a bad translation\nwith a specific issue commonly unrecognized by learned metrics.\n\n\nThe table below provides an overview of the various failure modes that we\nconsidered, including brief descriptions of how we prepared the synthetic data\nto address them.\n\n\n\nExamples from the first 4 categories were assigned a label corresponding to the\nworst score on the given rating scale (e.g., 25 when mixed with MQM training\ndata), whereas the reference-matching translation examples are assigned the best\nscore (e.g., 0 when used with MQM data). The missing/incorrect punctuation\nexamples were labeled with a score slightly worse than perfect.\n\n\nNote that some of the synthetic datasets are only meaningful in the\nreference-based scenario, and we thus excluded them when training a QE variant\nof MetricX. These are the Latin-vs-special punctuation and the\nreference-matching translation examples.\n\n\nMost of the synthetic training sets were created using stratified sampling\nacross target languages, taking 500 examples per target language. One exception\nis the missing punctuation set, which used a stratified sample across different\npunctuation symbols instead.\n\n\nWhen training MetricX, a small proportion of the synthetic examples was mixed\nwith the regular training examples. During the first-stage fine-tuning on DA\ndata, each synthetic training set constituted between 0.1% and 1% of all\ntraining examples, whereas in the second-stage fine-tuning on MQM data we used\nan even smaller proportion, around 0.05%.\n\n\nAs for evaluating the effect of the synthetic training data on the model's\nperformance, the DEMETR challenge set - which we originally used to evaluate the\nmodels submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We\ntherefore created a new DEMETR-style test set based on the WMT22 DA data, with\nexamples constructed analogically to the synthetic training examples, as\ndescribed above. This test set helped us determine the right proportions of\nsynthetic data for fine-tuning in order to make MetricX robust for the failure\nmodes in consideration, without sacrificing the system- and segment-level\ncorrelations with human ratings.\n\n\nUsage\n-----\n\n\nThe code for using MetricX models can be found at URL\nThe repository contains example prediction scripts, described below.\n\n\nThe 'metricx23/URL' script contains an example for how to run inference\non the models.",
"### Reference-Based\n\n\nExample usage for a reference-based model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"reference\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nNote that the model was trained with a maximum input length of 1024 tokens, so\nsignificantly increasing that value may lead to unpredictable behavior.",
"### Reference-Free\n\n\nExample usage for a reference-free model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"source\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nMeta-Evaluation\n---------------\n\n\nThe 'metricx23/URL' script contains code to calculate various correlations\nbetween the MetricX-23 scores and MQM ratings of translation quality using the\nMT Metrics Eval library.\n\n\nExample usage:\n\n\n'URL' is expected to have one JSON object serialized per line.\nEach JSON object is expected to contain 4 fields:\n\n\n* '\"system\\_id\"': The name of the system that generated the translation.\n* '\"segment\\_id\"': The 0-based index of the corresponding segment in the MT\nMetrics Eval data.\n* '\"label\"': The ground-truth translation quality score (with higher is better).\n* '\"prediction\"': The model predicted translation quality score (with lower is\nbetter; the script negates the scores so higher is better).\n\n\nThe script will calculate the 4 agreement/correlations that were used in the\nWMT'23 Shared Task. Below are the results for the MetricX-23 models on the\nWMT'22 Metrics Shared Task data:\n\n\nEnglish-German:\n\n\n\nEnglish-Russian:\n\n\n\nChinese-English:\n\n\n\nThe 'metricx23/evaluate\\_wmt23.py' script re-calculates the average correlation\nscore that was used to rank submissions from the\nWMT'23 Shared Task.\n\n\nExample usage:\n\n\nEach of the 3 input files is expected to be in the same format as described\nabove. Each file should correspond to running inference on each of the language\npairs from the WMT'23 dataset.\n\n\nThe results for each of the models is the following:\n\n\n\nIf you use MetricX-23 in your research, please cite the following publication:"
] | [
42,
666,
111,
457
] | [
"passage: TAGS\n#transformers #pytorch #mt5 #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us \n",
"passage: ### Synthetic Data\n\n\nIn order for our MetricX models to learn to identify certain types of bad\ntranslations that are not sufficiently (or at all) represented in the regular\ntraining data, we created synthetic examples and mixed them in during training.\nThe synthetic training data was generated from the DA datasets ranging from\nWMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have\nthe candidate translation manipulated so as to turn it into a bad translation\nwith a specific issue commonly unrecognized by learned metrics.\n\n\nThe table below provides an overview of the various failure modes that we\nconsidered, including brief descriptions of how we prepared the synthetic data\nto address them.\n\n\n\nExamples from the first 4 categories were assigned a label corresponding to the\nworst score on the given rating scale (e.g., 25 when mixed with MQM training\ndata), whereas the reference-matching translation examples are assigned the best\nscore (e.g., 0 when used with MQM data). The missing/incorrect punctuation\nexamples were labeled with a score slightly worse than perfect.\n\n\nNote that some of the synthetic datasets are only meaningful in the\nreference-based scenario, and we thus excluded them when training a QE variant\nof MetricX. These are the Latin-vs-special punctuation and the\nreference-matching translation examples.\n\n\nMost of the synthetic training sets were created using stratified sampling\nacross target languages, taking 500 examples per target language. One exception\nis the missing punctuation set, which used a stratified sample across different\npunctuation symbols instead.\n\n\nWhen training MetricX, a small proportion of the synthetic examples was mixed\nwith the regular training examples. During the first-stage fine-tuning on DA\ndata, each synthetic training set constituted between 0.1% and 1% of all\ntraining examples, whereas in the second-stage fine-tuning on MQM data we used\nan even smaller proportion, around 0.05%.\n\n\nAs for evaluating the effect of the synthetic training data on the model's\nperformance, the DEMETR challenge set - which we originally used to evaluate the\nmodels submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We\ntherefore created a new DEMETR-style test set based on the WMT22 DA data, with\nexamples constructed analogically to the synthetic training examples, as\ndescribed above. This test set helped us determine the right proportions of\nsynthetic data for fine-tuning in order to make MetricX robust for the failure\nmodes in consideration, without sacrificing the system- and segment-level\ncorrelations with human ratings.\n\n\nUsage\n-----\n\n\nThe code for using MetricX models can be found at URL\nThe repository contains example prediction scripts, described below.\n\n\nThe 'metricx23/URL' script contains an example for how to run inference\non the models.### Reference-Based\n\n\nExample usage for a reference-based model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"reference\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nNote that the model was trained with a maximum input length of 1024 tokens, so\nsignificantly increasing that value may lead to unpredictable behavior."
] | [
-0.08049466460943222,
-0.010928971692919731,
-0.004997325129806995,
0.02674533985555172,
0.10898007452487946,
0.017911424860358238,
0.030734051018953323,
0.10377532243728638,
-0.02402583509683609,
-0.03567541018128395,
0.03877764940261841,
0.08288230746984482,
0.015161662362515926,
0.022808298468589783,
0.01772439479827881,
-0.1944202184677124,
0.05506649613380432,
0.008813947439193726,
-0.07605475932359695,
0.06908859312534332,
0.11011916399002075,
-0.04383927211165428,
0.047993674874305725,
0.01802058517932892,
-0.046737369149923325,
0.03515304625034332,
0.02192578837275505,
-0.05999843776226044,
0.09158408641815186,
0.08443905413150787,
0.04427182674407959,
0.026185670867562294,
-0.03256708011031151,
-0.2039792835712433,
0.018693745136260986,
0.059222154319286346,
-0.04639671370387077,
0.032644130289554596,
0.06700465828180313,
-0.04408042132854462,
0.10856176912784576,
-0.03563929721713066,
-0.038342323154211044,
0.052534811198711395,
-0.08053379505872726,
-0.11484964191913605,
-0.06861758977174759,
0.005672447383403778,
0.05251080170273781,
0.11462049186229706,
0.016137007623910904,
0.08802412450313568,
-0.08221147954463959,
0.039690323173999786,
0.1186930239200592,
-0.24400970339775085,
0.022698264569044113,
0.031076578423380852,
0.0391852930188179,
0.059262461960315704,
-0.02663654275238514,
-0.011056511662900448,
0.04030945897102356,
0.036150794476270676,
0.04718852788209915,
-0.02858263999223709,
0.022121166810393333,
0.03753712773323059,
-0.11128073930740356,
-0.06610941141843796,
0.2194429337978363,
0.009360874071717262,
-0.039162129163742065,
-0.012831603176891804,
-0.040371835231781006,
-0.021052006632089615,
0.02562984824180603,
0.011642597615718842,
0.0126093290746212,
0.06509723514318466,
0.08087117969989777,
-0.06699725240468979,
-0.1469363272190094,
-0.02156376652419567,
-0.13039535284042358,
0.056314483284950256,
0.015906725078821182,
0.0635187178850174,
-0.13009536266326904,
0.06865113973617554,
-0.11303252726793289,
-0.037289686501026154,
-0.044206660240888596,
-0.09330593794584274,
0.027131132781505585,
0.0172458216547966,
-0.0671110600233078,
-0.061645977199077606,
0.01954670250415802,
0.09166590869426727,
-0.03881584480404854,
0.02113628201186657,
0.06317485123872757,
0.08239984512329102,
0.02081705816090107,
0.08037593215703964,
0.01793535053730011,
-0.011132435873150826,
0.08186338096857071,
-0.06666721403598785,
0.03507205471396446,
-0.00009720586240291595,
-0.05460785701870918,
-0.04377254843711853,
-0.02165292575955391,
0.07720592617988586,
0.024521537125110626,
0.024704361334443092,
-0.022436054423451424,
-0.014114882797002792,
0.15771594643592834,
-0.08077961206436157,
-0.031236806884407997,
-0.011595608666539192,
-0.05022282153367996,
0.047523126006126404,
0.06264422833919525,
0.006788937374949455,
-0.11120638251304626,
0.05265801399946213,
-0.03781677410006523,
-0.03154580295085907,
-0.08177147805690765,
-0.060364820063114166,
0.05910094082355499,
-0.0520191490650177,
0.022135227918624878,
-0.12485349923372269,
-0.14817503094673157,
-0.0194524135440588,
0.00905279628932476,
0.017627842724323273,
-0.04404270648956299,
0.030841557309031487,
-0.021325333043932915,
-0.010873598977923393,
-0.03838267922401428,
0.04409907013177872,
-0.0735226571559906,
0.01654006540775299,
-0.05891093611717224,
0.006638679653406143,
-0.12875565886497498,
0.05859975516796112,
-0.07795718312263489,
-0.008538316003978252,
-0.12053868919610977,
0.06858694553375244,
-0.06013733893632889,
0.06287765502929688,
-0.06418386101722717,
-0.059423450380563736,
-0.05324282497167587,
0.03673676773905754,
-0.011127032339572906,
0.08703723549842834,
-0.11388625204563141,
-0.04371771216392517,
0.06918473541736603,
-0.1147933155298233,
-0.10631323605775833,
0.05931851267814636,
-0.015358645468950272,
0.03748219460248947,
0.07437919080257416,
0.06889817118644714,
0.13736900687217712,
-0.022680290043354034,
-0.030354108661413193,
0.0774422138929367,
-0.09487991780042648,
-0.0831356942653656,
0.043639007955789566,
-0.02724086493253708,
-0.027338385581970215,
0.010795310139656067,
-0.05432993173599243,
0.053616154938936234,
0.004504588432610035,
-0.041023075580596924,
-0.03911237046122551,
-0.03953332453966141,
-0.03608177229762077,
-0.007416178472340107,
0.04547516256570816,
-0.030613455921411514,
-0.04433886334300041,
0.014297567307949066,
0.05688025429844856,
0.0016403227346017957,
0.06881103664636612,
-0.06005219742655754,
0.05407850816845894,
-0.018144575878977776,
0.008445029146969318,
-0.16996923089027405,
-0.050326526165008545,
0.03246239572763443,
-0.04993841052055359,
0.021022705361247063,
0.02538926526904106,
0.02754078060388565,
-0.014172155410051346,
-0.03300439566373825,
-0.01649932749569416,
0.008937809616327286,
0.008251310326159,
-0.05732375383377075,
-0.12281204760074615,
-0.0057578012347221375,
-0.030820120126008987,
-0.01729525998234749,
-0.027541520074009895,
0.012479400262236595,
0.07977741956710815,
0.044011034071445465,
0.009504541754722595,
0.03403853625059128,
0.005829425062984228,
0.01828007400035858,
-0.025431593880057335,
-0.01253400556743145,
0.06733429431915283,
0.029115518555045128,
-0.08058406412601471,
0.13161158561706543,
-0.1250324547290802,
0.017595291137695312,
0.13202296197414398,
-0.10783318430185318,
0.04871136322617531,
-0.04314408451318741,
-0.031379710882902145,
-0.0057036676444113255,
-0.03291696310043335,
-0.022622399032115936,
0.12076297402381897,
-0.004013901576399803,
0.0929318368434906,
-0.07880045473575592,
-0.05966495722532272,
-0.0097917215898633,
0.0038842959329485893,
-0.010753002017736435,
0.0713651031255722,
0.12100948393344879,
-0.12231792509555817,
0.08771493285894394,
0.13989409804344177,
-0.003570202738046646,
0.12119580805301666,
-0.040826037526130676,
-0.07305974513292313,
0.03446398302912712,
0.004951390903443098,
-0.01554139144718647,
0.04345749691128731,
-0.12445223331451416,
0.015536606311798096,
0.05309225618839264,
0.08432695269584656,
0.11454753577709198,
-0.11917590349912643,
0.0067930202931165695,
0.015571664087474346,
-0.03209373727440834,
-0.09838879108428955,
-0.0021094102412462234,
-0.01198030449450016,
0.07585512846708298,
-0.010641621425747871,
0.034573666751384735,
0.06810279190540314,
-0.033153194934129715,
-0.1333116739988327,
0.1549171507358551,
-0.16806340217590332,
-0.2015051692724228,
-0.1637459099292755,
0.0050877295434474945,
-0.04799898713827133,
0.007370667066425085,
0.10294272005558014,
-0.07140845060348511,
-0.04824806749820709,
-0.030684374272823334,
0.044100407510995865,
-0.026180272921919823,
-0.004666412249207497,
-0.05338970944285393,
0.042327139526605606,
0.016776274889707565,
-0.12135301530361176,
-0.006609104573726654,
-0.0036472089122980833,
-0.017405901104211807,
0.041763320565223694,
-0.07078999280929565,
0.07727597653865814,
0.08807438611984253,
0.01986617222428322,
-0.01348377950489521,
-0.04507536441087723,
0.21933329105377197,
-0.05761788785457611,
0.04443945735692978,
0.2112521529197693,
-0.011195655912160873,
0.05879829078912735,
0.052676692605018616,
-0.004360849969089031,
-0.02312907949090004,
0.02751847915351391,
0.013980694115161896,
-0.0917525514960289,
-0.284024178981781,
-0.06044050678610802,
-0.09147080034017563,
-0.06417541950941086,
0.0023470930755138397,
0.06487169861793518,
0.03722584992647171,
0.050735119730234146,
-0.04127505049109459,
-0.014369443990290165,
0.041575632989406586,
0.05546524375677109,
0.09428055584430695,
0.013206657022237778,
0.06720979511737823,
-0.11070599406957626,
0.012433091178536415,
0.11600261181592941,
0.09805083274841309,
0.23822110891342163,
-0.026172611862421036,
0.04857305809855461,
0.11312685161828995,
0.11436714231967926,
0.08747904002666473,
0.10965031385421753,
-0.04313330352306366,
0.029491525143384933,
-0.041477788239717484,
-0.02848794311285019,
-0.03056623786687851,
0.08165007829666138,
-0.021050430834293365,
-0.0669928565621376,
-0.06534460186958313,
-0.021561436355113983,
0.06433352082967758,
0.21831148862838745,
-0.005303980782628059,
-0.07421030104160309,
-0.05741162598133087,
0.06067470833659172,
-0.010603219270706177,
-0.03260648995637894,
0.07761399447917938,
0.04806586354970932,
-0.07070834934711456,
0.08121111243963242,
-0.0008651819080114365,
0.11229313164949417,
0.014927586540579796,
0.018110742792487144,
0.03192968666553497,
-0.006562814116477966,
0.018823236227035522,
0.11194905638694763,
-0.26115158200263977,
0.12093576788902283,
-0.011933675035834312,
0.05421082302927971,
-0.0762992650270462,
0.022125624120235443,
0.04246993362903595,
0.1273622065782547,
0.09262282401323318,
0.043345239013433456,
-0.14447379112243652,
-0.0103498175740242,
0.0019879117608070374,
0.056262798607349396,
0.04705921933054924,
0.010509418323636055,
0.011808192357420921,
-0.011274687014520168,
0.013288512825965881,
0.015993541106581688,
0.08853989839553833,
-0.0517207607626915,
-0.16778728365898132,
0.04126080125570297,
-0.006679844111204147,
0.015364071354269981,
-0.07193623483181,
0.005432065576314926,
-0.039848387241363525,
0.08026738464832306,
-0.14921778440475464,
-0.0995694100856781,
-0.09588959813117981,
-0.08473988622426987,
0.08787038922309875,
-0.0654374435544014,
0.06406596302986145,
0.007063437253236771,
0.07009734213352203,
-0.03068877011537552,
-0.14400194585323334,
0.057342350482940674,
-0.11315649747848511,
-0.08588340878486633,
-0.0034783631563186646,
0.1125456690788269,
0.013006072491407394,
0.0326993428170681,
0.019293401390314102,
-0.010193015448749065,
-0.046051137149333954,
-0.11050929129123688,
0.01819794625043869,
0.07997186481952667,
0.05762843042612076,
0.05681586265563965,
-0.09417805820703506,
-0.029595911502838135,
-0.04396835342049599,
-0.06783974170684814,
0.1667174994945526,
0.10756810754537582,
-0.06620179116725922,
0.1170005351305008,
0.16132885217666626,
-0.1295406073331833,
-0.19067947566509247,
-0.043041959404945374,
-0.03716177120804787,
-0.01986772194504738,
-0.02807173877954483,
-0.1475318968296051,
0.07539469003677368,
0.09812235087156296,
-0.0024198442697525024,
0.04476727545261383,
-0.2080758810043335,
-0.08643566817045212,
0.04781382530927658,
0.0001132916659116745,
0.18663114309310913,
-0.10906235128641129,
-0.044622793793678284,
-0.052799079567193985,
-0.11352433264255524,
0.14947330951690674,
-0.06072360277175903,
0.05728284642100334,
0.00732648279517889,
0.0032600341364741325,
-0.006010896060615778,
-0.03202845901250839,
0.13251455128192902,
0.03674418851733208,
0.043941304087638855,
-0.028950639069080353,
0.055670365691185,
0.025221113115549088,
-0.04210628941655159,
0.044331781566143036,
-0.04143795743584633,
0.021573659032583237,
-0.14452756941318512,
-0.045783452689647675,
-0.08391499519348145,
0.04536336287856102,
-0.01962660811841488,
-0.039184920489788055,
-0.042702145874500275,
0.0021056830883026123,
0.08968086540699005,
-0.014609623700380325,
0.054970160126686096,
-0.05837482213973999,
0.059537388384342194,
0.07670764625072479,
0.15645602345466614,
-0.10332942008972168,
-0.1217908263206482,
-0.02813919633626938,
-0.021626751869916916,
0.08217503130435944,
-0.1376548409461975,
0.07488036155700684,
0.11647255718708038,
-0.04862956330180168,
0.08627215772867203,
0.055094070732593536,
-0.07522906363010406,
0.005458813160657883,
0.10446281731128693,
-0.07246147096157074,
-0.06008515506982803,
-0.0390506386756897,
0.08281482756137848,
0.029428621754050255,
0.03669393062591553,
0.12870027124881744,
-0.0009662613738328218,
-0.032006535679101944,
-0.02065277099609375,
0.06402802467346191,
-0.0347125381231308,
0.0842379555106163,
0.039138346910476685,
0.017695903778076172,
-0.08644309639930725,
0.1338755190372467,
0.01591717265546322,
-0.1885780692100525,
-0.009080829098820686,
0.11049890518188477,
-0.11632084846496582,
-0.0807068943977356,
-0.025578761473298073,
0.009648185223340988,
-0.14496804773807526,
-0.10269727557897568,
-0.032402172684669495,
-0.14044955372810364,
0.06719717383384705,
0.16770902276039124,
0.02116461656987667,
0.03302031010389328,
-0.006035410333424807,
-0.028474289923906326,
-0.03122306987643242,
-0.013931620866060257,
-0.0277186818420887,
0.017339259386062622,
-0.05742930248379707,
0.0389898419380188,
0.019052214920520782,
0.06222548335790634,
-0.016825642436742783,
0.015219002030789852,
-0.0635049119591713,
0.010224338620901108,
-0.15755128860473633,
0.024624407291412354,
-0.05316181108355522,
-0.017552291974425316,
-0.01309568714350462,
-0.014190688729286194,
-0.05401631444692612,
0.05547712743282318,
-0.09220357239246368,
-0.020396772772073746,
-0.03640173375606537,
0.027906954288482666,
-0.05868091061711311,
-0.02378096804022789,
0.050178419798612595,
-0.047896455973386765,
0.1031966358423233,
0.06362350285053253,
-0.05172618851065636,
0.0687813088297844,
-0.054818395525217056,
-0.03117077425122261,
0.04752342030405998,
0.04614630341529846,
0.0024338315706700087,
-0.06958136707544327,
0.053022030740976334,
0.0532219260931015,
-0.05545207858085632,
-0.00179347675293684,
-0.00773459579795599,
-0.106658935546875,
-0.026197271421551704,
0.026790166273713112,
-0.007194001227617264,
-0.06978121399879456,
-0.046734243631362915,
0.11590144783258438,
0.04049905017018318,
0.1221245527267456,
-0.006192055530846119,
0.039656929671764374,
-0.0669143944978714,
0.0076656281016767025,
-0.0022521058563143015,
-0.0823831707239151,
-0.09226924180984497,
-0.03779274597764015,
0.01616818644106388,
-0.032287146896123886,
0.20258086919784546,
-0.027631420642137527,
-0.001419171690940857,
0.05651244521141052,
0.0006612688302993774,
-0.019393347203731537,
0.012330395169556141,
0.1308664083480835,
0.009836757555603981,
0.00910251121968031,
-0.02165895141661167,
0.024475062265992165,
-0.016778085380792618,
-0.0041928887367248535,
0.13951833546161652,
0.09686414152383804,
0.16705356538295746,
0.1026294082403183,
0.002041283994913101,
-0.03730851784348488,
-0.01945573464035988,
-0.06291651725769043,
0.028470784425735474,
0.058277349919080734,
0.00701283710077405,
0.01671595126390457,
0.11799974739551544,
-0.09208066761493683,
0.09887351095676422,
-0.06913766264915466,
-0.05150070786476135,
-0.1154223382472992,
-0.09975318610668182,
-0.010319891385734081,
-0.08603835850954056,
-0.029209984466433525,
-0.13546273112297058,
-0.020355189219117165,
0.06884671002626419,
0.038902945816516876,
-0.05183970928192139,
0.007247757166624069,
-0.11964748054742813,
-0.10397026687860489,
0.04866647720336914,
-0.014942724257707596,
0.04801199212670326,
0.04992463439702988,
0.012288471683859825,
0.04020651429891586,
-0.03871965408325195,
-0.02611941285431385,
0.039849359542131424,
0.04274921864271164,
0.045678697526454926,
-0.03452969342470169,
-0.010724259540438652,
-0.028795840218663216,
0.012287003919482231,
0.02913379855453968,
0.22296985983848572,
0.010857714340090752,
-0.031127674505114555,
0.06575135886669159,
0.19841772317886353,
-0.04081185162067413,
-0.09934508055448532,
-0.09630105644464493,
0.1641465723514557,
0.045690394937992096,
0.029041193425655365,
0.016438156366348267,
-0.057314880192279816,
-0.00565880723297596,
0.22985684871673584,
0.17313337326049805,
-0.025465190410614014,
-0.009001774713397026,
-0.006234009750187397,
0.014965811744332314,
0.06765225529670715,
0.08408764004707336,
0.1142108142375946,
0.21731261909008026,
-0.06078370288014412,
0.0639389306306839,
-0.0446958988904953,
0.0698341652750969,
-0.034699127078056335,
0.17769397795200348,
0.001534847542643547,
-0.058601561933755875,
-0.0005737040191888809,
0.06745973229408264,
-0.07404687255620956,
-0.10022060573101044,
-0.03632500767707825,
-0.07546216249465942,
-0.05532532557845116,
-0.0036639878526329994,
0.05928308516740799,
0.022997058928012848,
0.041565168648958206,
-0.0059436894953250885,
-0.044010479003190994,
0.10706667602062225,
0.023562224581837654,
-0.17537671327590942,
-0.05920793116092682,
0.08894800394773483,
0.011057954281568527,
0.08248458802700043,
0.007951926440000534,
0.048754554241895676,
0.07727794349193573,
0.03654977306723595,
-0.11864081025123596,
0.07930269092321396,
-0.00997848343104124,
-0.05712908133864403,
0.007062149234116077,
0.014758620411157608,
-0.00663449801504612,
0.020782694220542908,
0.029319513589143753,
-0.03974604234099388,
0.05011465772986412,
0.03723395615816116,
-0.04944480583071709,
-0.00523018091917038,
0.03254415839910507,
-0.10453616082668304,
0.08378694206476212,
0.07704398781061172,
-0.022326018661260605,
-0.027006199583411217,
-0.05415914207696915,
0.02430926077067852,
0.012829716317355633,
-0.023813046514987946,
-0.024040259420871735,
-0.06499078124761581,
0.017759330570697784,
0.01358121633529663,
0.03267041966319084,
-0.13761897385120392,
-0.03263425827026367,
0.009745022282004356,
-0.028417814522981644,
-0.08249472081661224,
0.06735780090093613,
0.043854910880327225,
0.026215003803372383,
-0.03867888078093529,
0.0033451877534389496,
-0.004438048228621483,
0.04008375108242035,
-0.0645875409245491,
-0.08099328726530075
] |
null | null | transformers | # MetricX-23
*This is not an officially supported Google product.*
**GitHub repository: [https://github.com/google-research/metricx](https://github.com/google-research/metricx)**
This repository contains the MetricX-23 models,
a family of models for automatic evaluation of translations that were proposed
in the WMT'23 Metrics Shared Task submission
[MetricX-23: The Google Submission to the WMT 2023 Metrics Shared Task](https://aclanthology.org/2023.wmt-1.63/).
The models were trained in [T5X](https://github.com/google-research/t5x) and
then converted for use in PyTorch.
## Available Models
There are 6 models available on HuggingFace that vary in the number of
parameters and whether or not the model is reference-based or reference-free
(also known as quality estimation, or QE):
* [MetricX-23-XXL](https://huggingface.co/google/metricx-23-large-v2p0)
* [MetricX-23-XL](https://huggingface.co/google/metricx-23-xl-v2p0)
* [MetricX-23-Large](https://huggingface.co/google/metricx-23-xxl-v2p0)
* [MetricX-23-QE-XXL](https://huggingface.co/google/metricx-23-qe-large-v2p0)
* [MetricX-23-QE-XL](https://huggingface.co/google/metricx-23-qe-xl-v2p0)
* [MetricX-23-QE-Large](https://huggingface.co/google/metricx-23-qe-xxl-v2p0)
We recommend using the XXL model versions for the best agreement with human
judgments of translation quality, the Large versions for best speed, and the
XL for an intermediate use case.
## Changes to the WMT'23 Submission
These models available here are most similar to the primary submission to the WMT'23 Metrics
Shared Task. They are initialized with [mT5](https://aclanthology.org/2021.naacl-main.41/)
then fine-tuned on a combination of direct assessment and MQM data. However,
we made some changes that make these models different from the WMT'23 submissions.
First, the models are trained to regress the actual MQM score rather than a
normalized score between 0 and 1. **That means the output from the MetricX-23
models is a score in the range [0, 25] where lower is better (i.e., it predicts
an error score).**
Second, these models were trained with a larger variety of synthetic data that
makes them more robust to translation edge cases like over- and undertranslation,
described in more detail in the following section.
### Synthetic Data
In order for our MetricX models to learn to identify certain types of bad
translations that are not sufficiently (or at all) represented in the regular
training data, we created synthetic examples and mixed them in during training.
The synthetic training data was generated from the DA datasets ranging from
WMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have
the candidate translation manipulated so as to turn it into a bad translation
with a specific issue commonly unrecognized by learned metrics.
The table below provides an overview of the various failure modes that we
considered, including brief descriptions of how we prepared the synthetic data
to address them.
| Failure mode | Synthetic example description |
| ----------- | ----------- |
| Undertranslation | Candidate translation with an arbitrary sentence removed (if multi-sentence); alternatively, candidate with a certain proportion of words removed from the end. |
| Overtranslation | Candidate translation duplicated (with space in between). |
| Fluent but unrelated translation | Arbitrary reference of a similar length from the dataset. |
| Gibberish | Text of a similar length as the reference, generated by sampling words from the reference translation vocabulary (built from all references in the data). |
| Missing punctuation | Reference translation with the end punctuation removed (11 punctuation symbols considered). |
| Latin instead of Chinese/Japanese or Hindi/Bengali punctuation | Candidate translation with the language-specific punctuation symbol at the end replaced with the Latin equivalent (e.g., "." instead of "。" or "।"); alternatively, the punctuation symbol is replaced with the Latin equivalent in the reference, keeping the correct one in the candidate. |
| Reference-matching translation | Reference translation copied as the candidate translation (unlike the rest of the synthetic data, these examples are meant to train the metric to predict a perfect score for candidates matching the reference). |
Examples from the first 4 categories were assigned a label corresponding to the
worst score on the given rating scale (e.g., 25 when mixed with MQM training
data), whereas the reference-matching translation examples are assigned the best
score (e.g., 0 when used with MQM data). The missing/incorrect punctuation
examples were labeled with a score slightly worse than perfect.
Note that some of the synthetic datasets are only meaningful in the
reference-based scenario, and we thus excluded them when training a QE variant
of MetricX. These are the Latin-vs-special punctuation and the
reference-matching translation examples.
Most of the synthetic training sets were created using stratified sampling
across target languages, taking 500 examples per target language. One exception
is the missing punctuation set, which used a stratified sample across different
punctuation symbols instead.
When training MetricX, a small proportion of the synthetic examples was mixed
with the regular training examples. During the first-stage fine-tuning on DA
data, each synthetic training set constituted between 0.1% and 1% of all
training examples, whereas in the second-stage fine-tuning on MQM data we used
an even smaller proportion, around 0.05%.
As for evaluating the effect of the synthetic training data on the model's
performance, the DEMETR challenge set - which we originally used to evaluate the
models submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We
therefore created a new DEMETR-style test set based on the WMT22 DA data, with
examples constructed analogically to the synthetic training examples, as
described above. This test set helped us determine the right proportions of
synthetic data for fine-tuning in order to make MetricX robust for the failure
modes in consideration, without sacrificing the system- and segment-level
correlations with human ratings.
## Usage
The code for using MetricX models can be found at [https://github.com/google-research/metricx](https://github.com/google-research/metricx).
The repository contains example prediction scripts, described below.
The `metricx23/predict.py` script contains an example for how to run inference
on the models.
### Reference-Based
Example usage for a reference-based model:
```bash
python -m metricx23.predict \
--tokenizer google/mt5-xl \
--model_name_or_path google/metricx-23-xl-v2p0 \
--max_input_length 1024 \
--batch_size 1 \
--input_file input.jsonl \
--output_file output.jsonl
```
`input.jsonl` is expected to have 1 serialized JSON object per line with
`"reference"` and `"hypothesis"` fields. The output jsonl will be parallel
to `input.jsonl` but additionally contain a `"prediction"` field with the predicted score.
Note that the model was trained with a maximum input length of 1024 tokens, so
significantly increasing that value may lead to unpredictable behavior.
### Reference-Free
Example usage for a reference-free model:
```bash
python -m metricx23.predict \
--tokenizer google/mt5-xl \
--model_name_or_path google/metricx-23-qe-xl-v2p0 \
--max_input_length 1024 \
--batch_size 1 \
--input_file input.jsonl \
--output_file output.jsonl \
--qe
```
`input.jsonl` is expected to have 1 serialized JSON object per line with
`"source"` and `"hypothesis"` fields. The output jsonl will be parallel
to `input.jsonl` but additionally contain a `"prediction"` field with the predicted score.
## Meta-Evaluation
The `metricx23/evaluate.py` script contains code to calculate various correlations
between the MetricX-23 scores and MQM ratings of translation quality using the
[MT Metrics Eval](https://github.com/google-research/mt-metrics-eval) library.
Example usage:
```bash
python -m metricx23.evaluate \
--dataset wmt22 \
--lp en-de \
--input_file input.jsonl \
--output_file output.json
```
`input.jsonl` is expected to have one JSON object serialized per line.
Each JSON object is expected to contain 4 fields:
* `"system_id"`: The name of the system that generated the translation.
* `"segment_id"`: The 0-based index of the corresponding segment in the MT
Metrics Eval data.
* `"label"`: The ground-truth translation quality score (with higher is better).
* `"prediction"`: The model predicted translation quality score (with lower is
better; the script negates the scores so higher is better).
The script will calculate the 4 agreement/correlations that were used in the
WMT'23 Shared Task. Below are the results for the MetricX-23 models on the
WMT'22 Metrics Shared Task data:
English-German:
| Model | System-Level Accuracy | System-Level Pearson | Segment-Level Pearson | Segment-Level Pairwise Acc |
| ----------- | ----------- | ----------- | ----------- | ----------- |
| MetricX-23-XXL | 0.795 | 0.835 | 0.546 | 0.619 |
| MetricX-23-XL | 0.756 | 0.813 | 0.540 | 0.605 |
| MetricX-23-Large | 0.769 | 0.759 | 0.507 | 0.595 |
| MetricX-23-QE-XXL | 0.769 | 0.830 | 0.490 | 0.606 |
| MetricX-23-QE-XL | 0.718 | 0.684 | 0.421 | 0.594 |
| MetricX-23-QE-Large | 0.744 | 0.671 | 0.387 | 0.579 |
English-Russian:
| Model | System-Level Accuracy | System-Level Pearson | Segment-Level Pearson | Segment-Level Pairwise Acc |
| ----------- | ----------- | ----------- | ----------- | ----------- |
| MetricX-23-XXL | 0.905 | 0.943 | 0.477 | 0.609 |
| MetricX-23-XL | 0.876 | 0.906 | 0.498 | 0.589 |
| MetricX-23-Large | 0.876 | 0.841 | 0.474 | 0.569 |
| MetricX-23-QE-XXL | 0.895 | 0.940 | 0.470 | 0.602 |
| MetricX-23-QE-XL | 0.848 | 0.861 | 0.415 | 0.570 |
| MetricX-23-QE-Large | 0.819 | 0.778 | 0.411 | 0.551 |
Chinese-English:
| Model | System-Level Accuracy | System-Level Pearson | Segment-Level Pearson | Segment-Level Pairwise Acc |
| ----------- | ----------- | ----------- | ----------- | ----------- |
| MetricX-23-XXL | 0.868 | 0.919 | 0.605 | 0.551 |
| MetricX-23-XL | 0.868 | 0.924 | 0.584 | 0.543 |
| MetricX-23-Large | 0.857 | 0.919 | 0.555 | 0.539 |
| MetricX-23-QE-XXL | 0.857 | 0.928 | 0.573 | 0.544 |
| MetricX-23-QE-XL | 0.802 | 0.879 | 0.546 | 0.529 |
| MetricX-23-QE-Large | 0.758 | 0.904 | 0.522 | 0.529 |
The `metricx23/evaluate_wmt23.py` script re-calculates the average correlation
score that was used to rank submissions from the
[WMT'23 Shared Task](https://www2.statmt.org/wmt23/pdf/2023.wmt-1.51.pdf).
Example usage:
```bash
python -m metricx23.evaluate_wmt23 \
--en_de predictions_ende.jsonl \
--he_en predictions_heen.jsonl \
--zh_en predictions_zhen.jsonl \
--output_file output.json
```
Each of the 3 input files is expected to be in the same format as described
above. Each file should correspond to running inference on each of the language
pairs from the WMT'23 dataset.
The results for each of the models is the following:
| Model | Average Correlation |
| ----------- | ----------- |
| MetricX-23-XXL | 0.812 |
| MetricX-23-XL | 0.813 |
| MetricX-23-Large | 0.794 |
| MetricX-23-QE-XXL | 0.797 |
| MetricX-23-QE-XL | 0.767 |
| MetricX-23-QE-Large | 0.762 |
## Citation
If you use MetricX-23 in your research, please cite the following publication:
```bibtex
@inproceedings{juraska-etal-2023-metricx,
title = {{MetricX-23: The Google Submission to the WMT 2023 Metrics Shared Task}},
author = "Juraska, Juraj and
Finkelstein, Mara and
Deutsch, Daniel and
Siddhant, Aditya and
Mirzazadeh, Mehdi and
Freitag, Markus",
editor = "Koehn, Philipp and
Haddow, Barry and
Kocmi, Tom and
Monz, Christof",
booktitle = "Proceedings of the Eighth Conference on Machine Translation",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.wmt-1.63",
doi = "10.18653/v1/2023.wmt-1.63",
pages = "756--767",
}
``` | {"license": "apache-2.0"} | null | google/metricx-23-xxl-v2p0 | [
"transformers",
"pytorch",
"mt5",
"license:apache-2.0",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T16:34:37+00:00 | [] | [] | TAGS
#transformers #pytorch #mt5 #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us
| MetricX-23
==========
*This is not an officially supported Google product.*
GitHub repository: URL
This repository contains the MetricX-23 models,
a family of models for automatic evaluation of translations that were proposed
in the WMT'23 Metrics Shared Task submission
MetricX-23: The Google Submission to the WMT 2023 Metrics Shared Task.
The models were trained in T5X and
then converted for use in PyTorch.
Available Models
----------------
There are 6 models available on HuggingFace that vary in the number of
parameters and whether or not the model is reference-based or reference-free
(also known as quality estimation, or QE):
* MetricX-23-XXL
* MetricX-23-XL
* MetricX-23-Large
* MetricX-23-QE-XXL
* MetricX-23-QE-XL
* MetricX-23-QE-Large
We recommend using the XXL model versions for the best agreement with human
judgments of translation quality, the Large versions for best speed, and the
XL for an intermediate use case.
Changes to the WMT'23 Submission
--------------------------------
These models available here are most similar to the primary submission to the WMT'23 Metrics
Shared Task. They are initialized with mT5
then fine-tuned on a combination of direct assessment and MQM data. However,
we made some changes that make these models different from the WMT'23 submissions.
First, the models are trained to regress the actual MQM score rather than a
normalized score between 0 and 1. That means the output from the MetricX-23
models is a score in the range [0, 25] where lower is better (i.e., it predicts
an error score).
Second, these models were trained with a larger variety of synthetic data that
makes them more robust to translation edge cases like over- and undertranslation,
described in more detail in the following section.
### Synthetic Data
In order for our MetricX models to learn to identify certain types of bad
translations that are not sufficiently (or at all) represented in the regular
training data, we created synthetic examples and mixed them in during training.
The synthetic training data was generated from the DA datasets ranging from
WMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have
the candidate translation manipulated so as to turn it into a bad translation
with a specific issue commonly unrecognized by learned metrics.
The table below provides an overview of the various failure modes that we
considered, including brief descriptions of how we prepared the synthetic data
to address them.
Examples from the first 4 categories were assigned a label corresponding to the
worst score on the given rating scale (e.g., 25 when mixed with MQM training
data), whereas the reference-matching translation examples are assigned the best
score (e.g., 0 when used with MQM data). The missing/incorrect punctuation
examples were labeled with a score slightly worse than perfect.
Note that some of the synthetic datasets are only meaningful in the
reference-based scenario, and we thus excluded them when training a QE variant
of MetricX. These are the Latin-vs-special punctuation and the
reference-matching translation examples.
Most of the synthetic training sets were created using stratified sampling
across target languages, taking 500 examples per target language. One exception
is the missing punctuation set, which used a stratified sample across different
punctuation symbols instead.
When training MetricX, a small proportion of the synthetic examples was mixed
with the regular training examples. During the first-stage fine-tuning on DA
data, each synthetic training set constituted between 0.1% and 1% of all
training examples, whereas in the second-stage fine-tuning on MQM data we used
an even smaller proportion, around 0.05%.
As for evaluating the effect of the synthetic training data on the model's
performance, the DEMETR challenge set - which we originally used to evaluate the
models submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We
therefore created a new DEMETR-style test set based on the WMT22 DA data, with
examples constructed analogically to the synthetic training examples, as
described above. This test set helped us determine the right proportions of
synthetic data for fine-tuning in order to make MetricX robust for the failure
modes in consideration, without sacrificing the system- and segment-level
correlations with human ratings.
Usage
-----
The code for using MetricX models can be found at URL
The repository contains example prediction scripts, described below.
The 'metricx23/URL' script contains an example for how to run inference
on the models.
### Reference-Based
Example usage for a reference-based model:
'URL' is expected to have 1 serialized JSON object per line with
'"reference"' and '"hypothesis"' fields. The output jsonl will be parallel
to 'URL' but additionally contain a '"prediction"' field with the predicted score.
Note that the model was trained with a maximum input length of 1024 tokens, so
significantly increasing that value may lead to unpredictable behavior.
### Reference-Free
Example usage for a reference-free model:
'URL' is expected to have 1 serialized JSON object per line with
'"source"' and '"hypothesis"' fields. The output jsonl will be parallel
to 'URL' but additionally contain a '"prediction"' field with the predicted score.
Meta-Evaluation
---------------
The 'metricx23/URL' script contains code to calculate various correlations
between the MetricX-23 scores and MQM ratings of translation quality using the
MT Metrics Eval library.
Example usage:
'URL' is expected to have one JSON object serialized per line.
Each JSON object is expected to contain 4 fields:
* '"system\_id"': The name of the system that generated the translation.
* '"segment\_id"': The 0-based index of the corresponding segment in the MT
Metrics Eval data.
* '"label"': The ground-truth translation quality score (with higher is better).
* '"prediction"': The model predicted translation quality score (with lower is
better; the script negates the scores so higher is better).
The script will calculate the 4 agreement/correlations that were used in the
WMT'23 Shared Task. Below are the results for the MetricX-23 models on the
WMT'22 Metrics Shared Task data:
English-German:
English-Russian:
Chinese-English:
The 'metricx23/evaluate\_wmt23.py' script re-calculates the average correlation
score that was used to rank submissions from the
WMT'23 Shared Task.
Example usage:
Each of the 3 input files is expected to be in the same format as described
above. Each file should correspond to running inference on each of the language
pairs from the WMT'23 dataset.
The results for each of the models is the following:
If you use MetricX-23 in your research, please cite the following publication:
| [
"### Synthetic Data\n\n\nIn order for our MetricX models to learn to identify certain types of bad\ntranslations that are not sufficiently (or at all) represented in the regular\ntraining data, we created synthetic examples and mixed them in during training.\nThe synthetic training data was generated from the DA datasets ranging from\nWMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have\nthe candidate translation manipulated so as to turn it into a bad translation\nwith a specific issue commonly unrecognized by learned metrics.\n\n\nThe table below provides an overview of the various failure modes that we\nconsidered, including brief descriptions of how we prepared the synthetic data\nto address them.\n\n\n\nExamples from the first 4 categories were assigned a label corresponding to the\nworst score on the given rating scale (e.g., 25 when mixed with MQM training\ndata), whereas the reference-matching translation examples are assigned the best\nscore (e.g., 0 when used with MQM data). The missing/incorrect punctuation\nexamples were labeled with a score slightly worse than perfect.\n\n\nNote that some of the synthetic datasets are only meaningful in the\nreference-based scenario, and we thus excluded them when training a QE variant\nof MetricX. These are the Latin-vs-special punctuation and the\nreference-matching translation examples.\n\n\nMost of the synthetic training sets were created using stratified sampling\nacross target languages, taking 500 examples per target language. One exception\nis the missing punctuation set, which used a stratified sample across different\npunctuation symbols instead.\n\n\nWhen training MetricX, a small proportion of the synthetic examples was mixed\nwith the regular training examples. During the first-stage fine-tuning on DA\ndata, each synthetic training set constituted between 0.1% and 1% of all\ntraining examples, whereas in the second-stage fine-tuning on MQM data we used\nan even smaller proportion, around 0.05%.\n\n\nAs for evaluating the effect of the synthetic training data on the model's\nperformance, the DEMETR challenge set - which we originally used to evaluate the\nmodels submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We\ntherefore created a new DEMETR-style test set based on the WMT22 DA data, with\nexamples constructed analogically to the synthetic training examples, as\ndescribed above. This test set helped us determine the right proportions of\nsynthetic data for fine-tuning in order to make MetricX robust for the failure\nmodes in consideration, without sacrificing the system- and segment-level\ncorrelations with human ratings.\n\n\nUsage\n-----\n\n\nThe code for using MetricX models can be found at URL\nThe repository contains example prediction scripts, described below.\n\n\nThe 'metricx23/URL' script contains an example for how to run inference\non the models.",
"### Reference-Based\n\n\nExample usage for a reference-based model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"reference\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nNote that the model was trained with a maximum input length of 1024 tokens, so\nsignificantly increasing that value may lead to unpredictable behavior.",
"### Reference-Free\n\n\nExample usage for a reference-free model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"source\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nMeta-Evaluation\n---------------\n\n\nThe 'metricx23/URL' script contains code to calculate various correlations\nbetween the MetricX-23 scores and MQM ratings of translation quality using the\nMT Metrics Eval library.\n\n\nExample usage:\n\n\n'URL' is expected to have one JSON object serialized per line.\nEach JSON object is expected to contain 4 fields:\n\n\n* '\"system\\_id\"': The name of the system that generated the translation.\n* '\"segment\\_id\"': The 0-based index of the corresponding segment in the MT\nMetrics Eval data.\n* '\"label\"': The ground-truth translation quality score (with higher is better).\n* '\"prediction\"': The model predicted translation quality score (with lower is\nbetter; the script negates the scores so higher is better).\n\n\nThe script will calculate the 4 agreement/correlations that were used in the\nWMT'23 Shared Task. Below are the results for the MetricX-23 models on the\nWMT'22 Metrics Shared Task data:\n\n\nEnglish-German:\n\n\n\nEnglish-Russian:\n\n\n\nChinese-English:\n\n\n\nThe 'metricx23/evaluate\\_wmt23.py' script re-calculates the average correlation\nscore that was used to rank submissions from the\nWMT'23 Shared Task.\n\n\nExample usage:\n\n\nEach of the 3 input files is expected to be in the same format as described\nabove. Each file should correspond to running inference on each of the language\npairs from the WMT'23 dataset.\n\n\nThe results for each of the models is the following:\n\n\n\nIf you use MetricX-23 in your research, please cite the following publication:"
] | [
"TAGS\n#transformers #pytorch #mt5 #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us \n",
"### Synthetic Data\n\n\nIn order for our MetricX models to learn to identify certain types of bad\ntranslations that are not sufficiently (or at all) represented in the regular\ntraining data, we created synthetic examples and mixed them in during training.\nThe synthetic training data was generated from the DA datasets ranging from\nWMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have\nthe candidate translation manipulated so as to turn it into a bad translation\nwith a specific issue commonly unrecognized by learned metrics.\n\n\nThe table below provides an overview of the various failure modes that we\nconsidered, including brief descriptions of how we prepared the synthetic data\nto address them.\n\n\n\nExamples from the first 4 categories were assigned a label corresponding to the\nworst score on the given rating scale (e.g., 25 when mixed with MQM training\ndata), whereas the reference-matching translation examples are assigned the best\nscore (e.g., 0 when used with MQM data). The missing/incorrect punctuation\nexamples were labeled with a score slightly worse than perfect.\n\n\nNote that some of the synthetic datasets are only meaningful in the\nreference-based scenario, and we thus excluded them when training a QE variant\nof MetricX. These are the Latin-vs-special punctuation and the\nreference-matching translation examples.\n\n\nMost of the synthetic training sets were created using stratified sampling\nacross target languages, taking 500 examples per target language. One exception\nis the missing punctuation set, which used a stratified sample across different\npunctuation symbols instead.\n\n\nWhen training MetricX, a small proportion of the synthetic examples was mixed\nwith the regular training examples. During the first-stage fine-tuning on DA\ndata, each synthetic training set constituted between 0.1% and 1% of all\ntraining examples, whereas in the second-stage fine-tuning on MQM data we used\nan even smaller proportion, around 0.05%.\n\n\nAs for evaluating the effect of the synthetic training data on the model's\nperformance, the DEMETR challenge set - which we originally used to evaluate the\nmodels submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We\ntherefore created a new DEMETR-style test set based on the WMT22 DA data, with\nexamples constructed analogically to the synthetic training examples, as\ndescribed above. This test set helped us determine the right proportions of\nsynthetic data for fine-tuning in order to make MetricX robust for the failure\nmodes in consideration, without sacrificing the system- and segment-level\ncorrelations with human ratings.\n\n\nUsage\n-----\n\n\nThe code for using MetricX models can be found at URL\nThe repository contains example prediction scripts, described below.\n\n\nThe 'metricx23/URL' script contains an example for how to run inference\non the models.",
"### Reference-Based\n\n\nExample usage for a reference-based model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"reference\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nNote that the model was trained with a maximum input length of 1024 tokens, so\nsignificantly increasing that value may lead to unpredictable behavior.",
"### Reference-Free\n\n\nExample usage for a reference-free model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"source\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nMeta-Evaluation\n---------------\n\n\nThe 'metricx23/URL' script contains code to calculate various correlations\nbetween the MetricX-23 scores and MQM ratings of translation quality using the\nMT Metrics Eval library.\n\n\nExample usage:\n\n\n'URL' is expected to have one JSON object serialized per line.\nEach JSON object is expected to contain 4 fields:\n\n\n* '\"system\\_id\"': The name of the system that generated the translation.\n* '\"segment\\_id\"': The 0-based index of the corresponding segment in the MT\nMetrics Eval data.\n* '\"label\"': The ground-truth translation quality score (with higher is better).\n* '\"prediction\"': The model predicted translation quality score (with lower is\nbetter; the script negates the scores so higher is better).\n\n\nThe script will calculate the 4 agreement/correlations that were used in the\nWMT'23 Shared Task. Below are the results for the MetricX-23 models on the\nWMT'22 Metrics Shared Task data:\n\n\nEnglish-German:\n\n\n\nEnglish-Russian:\n\n\n\nChinese-English:\n\n\n\nThe 'metricx23/evaluate\\_wmt23.py' script re-calculates the average correlation\nscore that was used to rank submissions from the\nWMT'23 Shared Task.\n\n\nExample usage:\n\n\nEach of the 3 input files is expected to be in the same format as described\nabove. Each file should correspond to running inference on each of the language\npairs from the WMT'23 dataset.\n\n\nThe results for each of the models is the following:\n\n\n\nIf you use MetricX-23 in your research, please cite the following publication:"
] | [
42,
666,
111,
457
] | [
"passage: TAGS\n#transformers #pytorch #mt5 #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us \n",
"passage: ### Synthetic Data\n\n\nIn order for our MetricX models to learn to identify certain types of bad\ntranslations that are not sufficiently (or at all) represented in the regular\ntraining data, we created synthetic examples and mixed them in during training.\nThe synthetic training data was generated from the DA datasets ranging from\nWMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have\nthe candidate translation manipulated so as to turn it into a bad translation\nwith a specific issue commonly unrecognized by learned metrics.\n\n\nThe table below provides an overview of the various failure modes that we\nconsidered, including brief descriptions of how we prepared the synthetic data\nto address them.\n\n\n\nExamples from the first 4 categories were assigned a label corresponding to the\nworst score on the given rating scale (e.g., 25 when mixed with MQM training\ndata), whereas the reference-matching translation examples are assigned the best\nscore (e.g., 0 when used with MQM data). The missing/incorrect punctuation\nexamples were labeled with a score slightly worse than perfect.\n\n\nNote that some of the synthetic datasets are only meaningful in the\nreference-based scenario, and we thus excluded them when training a QE variant\nof MetricX. These are the Latin-vs-special punctuation and the\nreference-matching translation examples.\n\n\nMost of the synthetic training sets were created using stratified sampling\nacross target languages, taking 500 examples per target language. One exception\nis the missing punctuation set, which used a stratified sample across different\npunctuation symbols instead.\n\n\nWhen training MetricX, a small proportion of the synthetic examples was mixed\nwith the regular training examples. During the first-stage fine-tuning on DA\ndata, each synthetic training set constituted between 0.1% and 1% of all\ntraining examples, whereas in the second-stage fine-tuning on MQM data we used\nan even smaller proportion, around 0.05%.\n\n\nAs for evaluating the effect of the synthetic training data on the model's\nperformance, the DEMETR challenge set - which we originally used to evaluate the\nmodels submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We\ntherefore created a new DEMETR-style test set based on the WMT22 DA data, with\nexamples constructed analogically to the synthetic training examples, as\ndescribed above. This test set helped us determine the right proportions of\nsynthetic data for fine-tuning in order to make MetricX robust for the failure\nmodes in consideration, without sacrificing the system- and segment-level\ncorrelations with human ratings.\n\n\nUsage\n-----\n\n\nThe code for using MetricX models can be found at URL\nThe repository contains example prediction scripts, described below.\n\n\nThe 'metricx23/URL' script contains an example for how to run inference\non the models.### Reference-Based\n\n\nExample usage for a reference-based model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"reference\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nNote that the model was trained with a maximum input length of 1024 tokens, so\nsignificantly increasing that value may lead to unpredictable behavior."
] | [
-0.08049466460943222,
-0.010928971692919731,
-0.004997325129806995,
0.02674533985555172,
0.10898007452487946,
0.017911424860358238,
0.030734051018953323,
0.10377532243728638,
-0.02402583509683609,
-0.03567541018128395,
0.03877764940261841,
0.08288230746984482,
0.015161662362515926,
0.022808298468589783,
0.01772439479827881,
-0.1944202184677124,
0.05506649613380432,
0.008813947439193726,
-0.07605475932359695,
0.06908859312534332,
0.11011916399002075,
-0.04383927211165428,
0.047993674874305725,
0.01802058517932892,
-0.046737369149923325,
0.03515304625034332,
0.02192578837275505,
-0.05999843776226044,
0.09158408641815186,
0.08443905413150787,
0.04427182674407959,
0.026185670867562294,
-0.03256708011031151,
-0.2039792835712433,
0.018693745136260986,
0.059222154319286346,
-0.04639671370387077,
0.032644130289554596,
0.06700465828180313,
-0.04408042132854462,
0.10856176912784576,
-0.03563929721713066,
-0.038342323154211044,
0.052534811198711395,
-0.08053379505872726,
-0.11484964191913605,
-0.06861758977174759,
0.005672447383403778,
0.05251080170273781,
0.11462049186229706,
0.016137007623910904,
0.08802412450313568,
-0.08221147954463959,
0.039690323173999786,
0.1186930239200592,
-0.24400970339775085,
0.022698264569044113,
0.031076578423380852,
0.0391852930188179,
0.059262461960315704,
-0.02663654275238514,
-0.011056511662900448,
0.04030945897102356,
0.036150794476270676,
0.04718852788209915,
-0.02858263999223709,
0.022121166810393333,
0.03753712773323059,
-0.11128073930740356,
-0.06610941141843796,
0.2194429337978363,
0.009360874071717262,
-0.039162129163742065,
-0.012831603176891804,
-0.040371835231781006,
-0.021052006632089615,
0.02562984824180603,
0.011642597615718842,
0.0126093290746212,
0.06509723514318466,
0.08087117969989777,
-0.06699725240468979,
-0.1469363272190094,
-0.02156376652419567,
-0.13039535284042358,
0.056314483284950256,
0.015906725078821182,
0.0635187178850174,
-0.13009536266326904,
0.06865113973617554,
-0.11303252726793289,
-0.037289686501026154,
-0.044206660240888596,
-0.09330593794584274,
0.027131132781505585,
0.0172458216547966,
-0.0671110600233078,
-0.061645977199077606,
0.01954670250415802,
0.09166590869426727,
-0.03881584480404854,
0.02113628201186657,
0.06317485123872757,
0.08239984512329102,
0.02081705816090107,
0.08037593215703964,
0.01793535053730011,
-0.011132435873150826,
0.08186338096857071,
-0.06666721403598785,
0.03507205471396446,
-0.00009720586240291595,
-0.05460785701870918,
-0.04377254843711853,
-0.02165292575955391,
0.07720592617988586,
0.024521537125110626,
0.024704361334443092,
-0.022436054423451424,
-0.014114882797002792,
0.15771594643592834,
-0.08077961206436157,
-0.031236806884407997,
-0.011595608666539192,
-0.05022282153367996,
0.047523126006126404,
0.06264422833919525,
0.006788937374949455,
-0.11120638251304626,
0.05265801399946213,
-0.03781677410006523,
-0.03154580295085907,
-0.08177147805690765,
-0.060364820063114166,
0.05910094082355499,
-0.0520191490650177,
0.022135227918624878,
-0.12485349923372269,
-0.14817503094673157,
-0.0194524135440588,
0.00905279628932476,
0.017627842724323273,
-0.04404270648956299,
0.030841557309031487,
-0.021325333043932915,
-0.010873598977923393,
-0.03838267922401428,
0.04409907013177872,
-0.0735226571559906,
0.01654006540775299,
-0.05891093611717224,
0.006638679653406143,
-0.12875565886497498,
0.05859975516796112,
-0.07795718312263489,
-0.008538316003978252,
-0.12053868919610977,
0.06858694553375244,
-0.06013733893632889,
0.06287765502929688,
-0.06418386101722717,
-0.059423450380563736,
-0.05324282497167587,
0.03673676773905754,
-0.011127032339572906,
0.08703723549842834,
-0.11388625204563141,
-0.04371771216392517,
0.06918473541736603,
-0.1147933155298233,
-0.10631323605775833,
0.05931851267814636,
-0.015358645468950272,
0.03748219460248947,
0.07437919080257416,
0.06889817118644714,
0.13736900687217712,
-0.022680290043354034,
-0.030354108661413193,
0.0774422138929367,
-0.09487991780042648,
-0.0831356942653656,
0.043639007955789566,
-0.02724086493253708,
-0.027338385581970215,
0.010795310139656067,
-0.05432993173599243,
0.053616154938936234,
0.004504588432610035,
-0.041023075580596924,
-0.03911237046122551,
-0.03953332453966141,
-0.03608177229762077,
-0.007416178472340107,
0.04547516256570816,
-0.030613455921411514,
-0.04433886334300041,
0.014297567307949066,
0.05688025429844856,
0.0016403227346017957,
0.06881103664636612,
-0.06005219742655754,
0.05407850816845894,
-0.018144575878977776,
0.008445029146969318,
-0.16996923089027405,
-0.050326526165008545,
0.03246239572763443,
-0.04993841052055359,
0.021022705361247063,
0.02538926526904106,
0.02754078060388565,
-0.014172155410051346,
-0.03300439566373825,
-0.01649932749569416,
0.008937809616327286,
0.008251310326159,
-0.05732375383377075,
-0.12281204760074615,
-0.0057578012347221375,
-0.030820120126008987,
-0.01729525998234749,
-0.027541520074009895,
0.012479400262236595,
0.07977741956710815,
0.044011034071445465,
0.009504541754722595,
0.03403853625059128,
0.005829425062984228,
0.01828007400035858,
-0.025431593880057335,
-0.01253400556743145,
0.06733429431915283,
0.029115518555045128,
-0.08058406412601471,
0.13161158561706543,
-0.1250324547290802,
0.017595291137695312,
0.13202296197414398,
-0.10783318430185318,
0.04871136322617531,
-0.04314408451318741,
-0.031379710882902145,
-0.0057036676444113255,
-0.03291696310043335,
-0.022622399032115936,
0.12076297402381897,
-0.004013901576399803,
0.0929318368434906,
-0.07880045473575592,
-0.05966495722532272,
-0.0097917215898633,
0.0038842959329485893,
-0.010753002017736435,
0.0713651031255722,
0.12100948393344879,
-0.12231792509555817,
0.08771493285894394,
0.13989409804344177,
-0.003570202738046646,
0.12119580805301666,
-0.040826037526130676,
-0.07305974513292313,
0.03446398302912712,
0.004951390903443098,
-0.01554139144718647,
0.04345749691128731,
-0.12445223331451416,
0.015536606311798096,
0.05309225618839264,
0.08432695269584656,
0.11454753577709198,
-0.11917590349912643,
0.0067930202931165695,
0.015571664087474346,
-0.03209373727440834,
-0.09838879108428955,
-0.0021094102412462234,
-0.01198030449450016,
0.07585512846708298,
-0.010641621425747871,
0.034573666751384735,
0.06810279190540314,
-0.033153194934129715,
-0.1333116739988327,
0.1549171507358551,
-0.16806340217590332,
-0.2015051692724228,
-0.1637459099292755,
0.0050877295434474945,
-0.04799898713827133,
0.007370667066425085,
0.10294272005558014,
-0.07140845060348511,
-0.04824806749820709,
-0.030684374272823334,
0.044100407510995865,
-0.026180272921919823,
-0.004666412249207497,
-0.05338970944285393,
0.042327139526605606,
0.016776274889707565,
-0.12135301530361176,
-0.006609104573726654,
-0.0036472089122980833,
-0.017405901104211807,
0.041763320565223694,
-0.07078999280929565,
0.07727597653865814,
0.08807438611984253,
0.01986617222428322,
-0.01348377950489521,
-0.04507536441087723,
0.21933329105377197,
-0.05761788785457611,
0.04443945735692978,
0.2112521529197693,
-0.011195655912160873,
0.05879829078912735,
0.052676692605018616,
-0.004360849969089031,
-0.02312907949090004,
0.02751847915351391,
0.013980694115161896,
-0.0917525514960289,
-0.284024178981781,
-0.06044050678610802,
-0.09147080034017563,
-0.06417541950941086,
0.0023470930755138397,
0.06487169861793518,
0.03722584992647171,
0.050735119730234146,
-0.04127505049109459,
-0.014369443990290165,
0.041575632989406586,
0.05546524375677109,
0.09428055584430695,
0.013206657022237778,
0.06720979511737823,
-0.11070599406957626,
0.012433091178536415,
0.11600261181592941,
0.09805083274841309,
0.23822110891342163,
-0.026172611862421036,
0.04857305809855461,
0.11312685161828995,
0.11436714231967926,
0.08747904002666473,
0.10965031385421753,
-0.04313330352306366,
0.029491525143384933,
-0.041477788239717484,
-0.02848794311285019,
-0.03056623786687851,
0.08165007829666138,
-0.021050430834293365,
-0.0669928565621376,
-0.06534460186958313,
-0.021561436355113983,
0.06433352082967758,
0.21831148862838745,
-0.005303980782628059,
-0.07421030104160309,
-0.05741162598133087,
0.06067470833659172,
-0.010603219270706177,
-0.03260648995637894,
0.07761399447917938,
0.04806586354970932,
-0.07070834934711456,
0.08121111243963242,
-0.0008651819080114365,
0.11229313164949417,
0.014927586540579796,
0.018110742792487144,
0.03192968666553497,
-0.006562814116477966,
0.018823236227035522,
0.11194905638694763,
-0.26115158200263977,
0.12093576788902283,
-0.011933675035834312,
0.05421082302927971,
-0.0762992650270462,
0.022125624120235443,
0.04246993362903595,
0.1273622065782547,
0.09262282401323318,
0.043345239013433456,
-0.14447379112243652,
-0.0103498175740242,
0.0019879117608070374,
0.056262798607349396,
0.04705921933054924,
0.010509418323636055,
0.011808192357420921,
-0.011274687014520168,
0.013288512825965881,
0.015993541106581688,
0.08853989839553833,
-0.0517207607626915,
-0.16778728365898132,
0.04126080125570297,
-0.006679844111204147,
0.015364071354269981,
-0.07193623483181,
0.005432065576314926,
-0.039848387241363525,
0.08026738464832306,
-0.14921778440475464,
-0.0995694100856781,
-0.09588959813117981,
-0.08473988622426987,
0.08787038922309875,
-0.0654374435544014,
0.06406596302986145,
0.007063437253236771,
0.07009734213352203,
-0.03068877011537552,
-0.14400194585323334,
0.057342350482940674,
-0.11315649747848511,
-0.08588340878486633,
-0.0034783631563186646,
0.1125456690788269,
0.013006072491407394,
0.0326993428170681,
0.019293401390314102,
-0.010193015448749065,
-0.046051137149333954,
-0.11050929129123688,
0.01819794625043869,
0.07997186481952667,
0.05762843042612076,
0.05681586265563965,
-0.09417805820703506,
-0.029595911502838135,
-0.04396835342049599,
-0.06783974170684814,
0.1667174994945526,
0.10756810754537582,
-0.06620179116725922,
0.1170005351305008,
0.16132885217666626,
-0.1295406073331833,
-0.19067947566509247,
-0.043041959404945374,
-0.03716177120804787,
-0.01986772194504738,
-0.02807173877954483,
-0.1475318968296051,
0.07539469003677368,
0.09812235087156296,
-0.0024198442697525024,
0.04476727545261383,
-0.2080758810043335,
-0.08643566817045212,
0.04781382530927658,
0.0001132916659116745,
0.18663114309310913,
-0.10906235128641129,
-0.044622793793678284,
-0.052799079567193985,
-0.11352433264255524,
0.14947330951690674,
-0.06072360277175903,
0.05728284642100334,
0.00732648279517889,
0.0032600341364741325,
-0.006010896060615778,
-0.03202845901250839,
0.13251455128192902,
0.03674418851733208,
0.043941304087638855,
-0.028950639069080353,
0.055670365691185,
0.025221113115549088,
-0.04210628941655159,
0.044331781566143036,
-0.04143795743584633,
0.021573659032583237,
-0.14452756941318512,
-0.045783452689647675,
-0.08391499519348145,
0.04536336287856102,
-0.01962660811841488,
-0.039184920489788055,
-0.042702145874500275,
0.0021056830883026123,
0.08968086540699005,
-0.014609623700380325,
0.054970160126686096,
-0.05837482213973999,
0.059537388384342194,
0.07670764625072479,
0.15645602345466614,
-0.10332942008972168,
-0.1217908263206482,
-0.02813919633626938,
-0.021626751869916916,
0.08217503130435944,
-0.1376548409461975,
0.07488036155700684,
0.11647255718708038,
-0.04862956330180168,
0.08627215772867203,
0.055094070732593536,
-0.07522906363010406,
0.005458813160657883,
0.10446281731128693,
-0.07246147096157074,
-0.06008515506982803,
-0.0390506386756897,
0.08281482756137848,
0.029428621754050255,
0.03669393062591553,
0.12870027124881744,
-0.0009662613738328218,
-0.032006535679101944,
-0.02065277099609375,
0.06402802467346191,
-0.0347125381231308,
0.0842379555106163,
0.039138346910476685,
0.017695903778076172,
-0.08644309639930725,
0.1338755190372467,
0.01591717265546322,
-0.1885780692100525,
-0.009080829098820686,
0.11049890518188477,
-0.11632084846496582,
-0.0807068943977356,
-0.025578761473298073,
0.009648185223340988,
-0.14496804773807526,
-0.10269727557897568,
-0.032402172684669495,
-0.14044955372810364,
0.06719717383384705,
0.16770902276039124,
0.02116461656987667,
0.03302031010389328,
-0.006035410333424807,
-0.028474289923906326,
-0.03122306987643242,
-0.013931620866060257,
-0.0277186818420887,
0.017339259386062622,
-0.05742930248379707,
0.0389898419380188,
0.019052214920520782,
0.06222548335790634,
-0.016825642436742783,
0.015219002030789852,
-0.0635049119591713,
0.010224338620901108,
-0.15755128860473633,
0.024624407291412354,
-0.05316181108355522,
-0.017552291974425316,
-0.01309568714350462,
-0.014190688729286194,
-0.05401631444692612,
0.05547712743282318,
-0.09220357239246368,
-0.020396772772073746,
-0.03640173375606537,
0.027906954288482666,
-0.05868091061711311,
-0.02378096804022789,
0.050178419798612595,
-0.047896455973386765,
0.1031966358423233,
0.06362350285053253,
-0.05172618851065636,
0.0687813088297844,
-0.054818395525217056,
-0.03117077425122261,
0.04752342030405998,
0.04614630341529846,
0.0024338315706700087,
-0.06958136707544327,
0.053022030740976334,
0.0532219260931015,
-0.05545207858085632,
-0.00179347675293684,
-0.00773459579795599,
-0.106658935546875,
-0.026197271421551704,
0.026790166273713112,
-0.007194001227617264,
-0.06978121399879456,
-0.046734243631362915,
0.11590144783258438,
0.04049905017018318,
0.1221245527267456,
-0.006192055530846119,
0.039656929671764374,
-0.0669143944978714,
0.0076656281016767025,
-0.0022521058563143015,
-0.0823831707239151,
-0.09226924180984497,
-0.03779274597764015,
0.01616818644106388,
-0.032287146896123886,
0.20258086919784546,
-0.027631420642137527,
-0.001419171690940857,
0.05651244521141052,
0.0006612688302993774,
-0.019393347203731537,
0.012330395169556141,
0.1308664083480835,
0.009836757555603981,
0.00910251121968031,
-0.02165895141661167,
0.024475062265992165,
-0.016778085380792618,
-0.0041928887367248535,
0.13951833546161652,
0.09686414152383804,
0.16705356538295746,
0.1026294082403183,
0.002041283994913101,
-0.03730851784348488,
-0.01945573464035988,
-0.06291651725769043,
0.028470784425735474,
0.058277349919080734,
0.00701283710077405,
0.01671595126390457,
0.11799974739551544,
-0.09208066761493683,
0.09887351095676422,
-0.06913766264915466,
-0.05150070786476135,
-0.1154223382472992,
-0.09975318610668182,
-0.010319891385734081,
-0.08603835850954056,
-0.029209984466433525,
-0.13546273112297058,
-0.020355189219117165,
0.06884671002626419,
0.038902945816516876,
-0.05183970928192139,
0.007247757166624069,
-0.11964748054742813,
-0.10397026687860489,
0.04866647720336914,
-0.014942724257707596,
0.04801199212670326,
0.04992463439702988,
0.012288471683859825,
0.04020651429891586,
-0.03871965408325195,
-0.02611941285431385,
0.039849359542131424,
0.04274921864271164,
0.045678697526454926,
-0.03452969342470169,
-0.010724259540438652,
-0.028795840218663216,
0.012287003919482231,
0.02913379855453968,
0.22296985983848572,
0.010857714340090752,
-0.031127674505114555,
0.06575135886669159,
0.19841772317886353,
-0.04081185162067413,
-0.09934508055448532,
-0.09630105644464493,
0.1641465723514557,
0.045690394937992096,
0.029041193425655365,
0.016438156366348267,
-0.057314880192279816,
-0.00565880723297596,
0.22985684871673584,
0.17313337326049805,
-0.025465190410614014,
-0.009001774713397026,
-0.006234009750187397,
0.014965811744332314,
0.06765225529670715,
0.08408764004707336,
0.1142108142375946,
0.21731261909008026,
-0.06078370288014412,
0.0639389306306839,
-0.0446958988904953,
0.0698341652750969,
-0.034699127078056335,
0.17769397795200348,
0.001534847542643547,
-0.058601561933755875,
-0.0005737040191888809,
0.06745973229408264,
-0.07404687255620956,
-0.10022060573101044,
-0.03632500767707825,
-0.07546216249465942,
-0.05532532557845116,
-0.0036639878526329994,
0.05928308516740799,
0.022997058928012848,
0.041565168648958206,
-0.0059436894953250885,
-0.044010479003190994,
0.10706667602062225,
0.023562224581837654,
-0.17537671327590942,
-0.05920793116092682,
0.08894800394773483,
0.011057954281568527,
0.08248458802700043,
0.007951926440000534,
0.048754554241895676,
0.07727794349193573,
0.03654977306723595,
-0.11864081025123596,
0.07930269092321396,
-0.00997848343104124,
-0.05712908133864403,
0.007062149234116077,
0.014758620411157608,
-0.00663449801504612,
0.020782694220542908,
0.029319513589143753,
-0.03974604234099388,
0.05011465772986412,
0.03723395615816116,
-0.04944480583071709,
-0.00523018091917038,
0.03254415839910507,
-0.10453616082668304,
0.08378694206476212,
0.07704398781061172,
-0.022326018661260605,
-0.027006199583411217,
-0.05415914207696915,
0.02430926077067852,
0.012829716317355633,
-0.023813046514987946,
-0.024040259420871735,
-0.06499078124761581,
0.017759330570697784,
0.01358121633529663,
0.03267041966319084,
-0.13761897385120392,
-0.03263425827026367,
0.009745022282004356,
-0.028417814522981644,
-0.08249472081661224,
0.06735780090093613,
0.043854910880327225,
0.026215003803372383,
-0.03867888078093529,
0.0033451877534389496,
-0.004438048228621483,
0.04008375108242035,
-0.0645875409245491,
-0.08099328726530075
] |
null | null | transformers | # MetricX-23
*This is not an officially supported Google product.*
**GitHub repository: [https://github.com/google-research/metricx](https://github.com/google-research/metricx)**
This repository contains the MetricX-23 models,
a family of models for automatic evaluation of translations that were proposed
in the WMT'23 Metrics Shared Task submission
[MetricX-23: The Google Submission to the WMT 2023 Metrics Shared Task](https://aclanthology.org/2023.wmt-1.63/).
The models were trained in [T5X](https://github.com/google-research/t5x) and
then converted for use in PyTorch.
## Available Models
There are 6 models available on HuggingFace that vary in the number of
parameters and whether or not the model is reference-based or reference-free
(also known as quality estimation, or QE):
* [MetricX-23-XXL](https://huggingface.co/google/metricx-23-large-v2p0)
* [MetricX-23-XL](https://huggingface.co/google/metricx-23-xl-v2p0)
* [MetricX-23-Large](https://huggingface.co/google/metricx-23-xxl-v2p0)
* [MetricX-23-QE-XXL](https://huggingface.co/google/metricx-23-qe-large-v2p0)
* [MetricX-23-QE-XL](https://huggingface.co/google/metricx-23-qe-xl-v2p0)
* [MetricX-23-QE-Large](https://huggingface.co/google/metricx-23-qe-xxl-v2p0)
We recommend using the XXL model versions for the best agreement with human
judgments of translation quality, the Large versions for best speed, and the
XL for an intermediate use case.
## Changes to the WMT'23 Submission
These models available here are most similar to the primary submission to the WMT'23 Metrics
Shared Task. They are initialized with [mT5](https://aclanthology.org/2021.naacl-main.41/)
then fine-tuned on a combination of direct assessment and MQM data. However,
we made some changes that make these models different from the WMT'23 submissions.
First, the models are trained to regress the actual MQM score rather than a
normalized score between 0 and 1. **That means the output from the MetricX-23
models is a score in the range [0, 25] where lower is better (i.e., it predicts
an error score).**
Second, these models were trained with a larger variety of synthetic data that
makes them more robust to translation edge cases like over- and undertranslation,
described in more detail in the following section.
### Synthetic Data
In order for our MetricX models to learn to identify certain types of bad
translations that are not sufficiently (or at all) represented in the regular
training data, we created synthetic examples and mixed them in during training.
The synthetic training data was generated from the DA datasets ranging from
WMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have
the candidate translation manipulated so as to turn it into a bad translation
with a specific issue commonly unrecognized by learned metrics.
The table below provides an overview of the various failure modes that we
considered, including brief descriptions of how we prepared the synthetic data
to address them.
| Failure mode | Synthetic example description |
| ----------- | ----------- |
| Undertranslation | Candidate translation with an arbitrary sentence removed (if multi-sentence); alternatively, candidate with a certain proportion of words removed from the end. |
| Overtranslation | Candidate translation duplicated (with space in between). |
| Fluent but unrelated translation | Arbitrary reference of a similar length from the dataset. |
| Gibberish | Text of a similar length as the reference, generated by sampling words from the reference translation vocabulary (built from all references in the data). |
| Missing punctuation | Reference translation with the end punctuation removed (11 punctuation symbols considered). |
| Latin instead of Chinese/Japanese or Hindi/Bengali punctuation | Candidate translation with the language-specific punctuation symbol at the end replaced with the Latin equivalent (e.g., "." instead of "。" or "।"); alternatively, the punctuation symbol is replaced with the Latin equivalent in the reference, keeping the correct one in the candidate. |
| Reference-matching translation | Reference translation copied as the candidate translation (unlike the rest of the synthetic data, these examples are meant to train the metric to predict a perfect score for candidates matching the reference). |
Examples from the first 4 categories were assigned a label corresponding to the
worst score on the given rating scale (e.g., 25 when mixed with MQM training
data), whereas the reference-matching translation examples are assigned the best
score (e.g., 0 when used with MQM data). The missing/incorrect punctuation
examples were labeled with a score slightly worse than perfect.
Note that some of the synthetic datasets are only meaningful in the
reference-based scenario, and we thus excluded them when training a QE variant
of MetricX. These are the Latin-vs-special punctuation and the
reference-matching translation examples.
Most of the synthetic training sets were created using stratified sampling
across target languages, taking 500 examples per target language. One exception
is the missing punctuation set, which used a stratified sample across different
punctuation symbols instead.
When training MetricX, a small proportion of the synthetic examples was mixed
with the regular training examples. During the first-stage fine-tuning on DA
data, each synthetic training set constituted between 0.1% and 1% of all
training examples, whereas in the second-stage fine-tuning on MQM data we used
an even smaller proportion, around 0.05%.
As for evaluating the effect of the synthetic training data on the model's
performance, the DEMETR challenge set - which we originally used to evaluate the
models submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We
therefore created a new DEMETR-style test set based on the WMT22 DA data, with
examples constructed analogically to the synthetic training examples, as
described above. This test set helped us determine the right proportions of
synthetic data for fine-tuning in order to make MetricX robust for the failure
modes in consideration, without sacrificing the system- and segment-level
correlations with human ratings.
## Usage
The code for using MetricX models can be found at [https://github.com/google-research/metricx](https://github.com/google-research/metricx).
The repository contains example prediction scripts, described below.
The `metricx23/predict.py` script contains an example for how to run inference
on the models.
### Reference-Based
Example usage for a reference-based model:
```bash
python -m metricx23.predict \
--tokenizer google/mt5-xl \
--model_name_or_path google/metricx-23-xl-v2p0 \
--max_input_length 1024 \
--batch_size 1 \
--input_file input.jsonl \
--output_file output.jsonl
```
`input.jsonl` is expected to have 1 serialized JSON object per line with
`"reference"` and `"hypothesis"` fields. The output jsonl will be parallel
to `input.jsonl` but additionally contain a `"prediction"` field with the predicted score.
Note that the model was trained with a maximum input length of 1024 tokens, so
significantly increasing that value may lead to unpredictable behavior.
### Reference-Free
Example usage for a reference-free model:
```bash
python -m metricx23.predict \
--tokenizer google/mt5-xl \
--model_name_or_path google/metricx-23-qe-xl-v2p0 \
--max_input_length 1024 \
--batch_size 1 \
--input_file input.jsonl \
--output_file output.jsonl \
--qe
```
`input.jsonl` is expected to have 1 serialized JSON object per line with
`"source"` and `"hypothesis"` fields. The output jsonl will be parallel
to `input.jsonl` but additionally contain a `"prediction"` field with the predicted score.
## Meta-Evaluation
The `metricx23/evaluate.py` script contains code to calculate various correlations
between the MetricX-23 scores and MQM ratings of translation quality using the
[MT Metrics Eval](https://github.com/google-research/mt-metrics-eval) library.
Example usage:
```bash
python -m metricx23.evaluate \
--dataset wmt22 \
--lp en-de \
--input_file input.jsonl \
--output_file output.json
```
`input.jsonl` is expected to have one JSON object serialized per line.
Each JSON object is expected to contain 4 fields:
* `"system_id"`: The name of the system that generated the translation.
* `"segment_id"`: The 0-based index of the corresponding segment in the MT
Metrics Eval data.
* `"label"`: The ground-truth translation quality score (with higher is better).
* `"prediction"`: The model predicted translation quality score (with lower is
better; the script negates the scores so higher is better).
The script will calculate the 4 agreement/correlations that were used in the
WMT'23 Shared Task. Below are the results for the MetricX-23 models on the
WMT'22 Metrics Shared Task data:
English-German:
| Model | System-Level Accuracy | System-Level Pearson | Segment-Level Pearson | Segment-Level Pairwise Acc |
| ----------- | ----------- | ----------- | ----------- | ----------- |
| MetricX-23-XXL | 0.795 | 0.835 | 0.546 | 0.619 |
| MetricX-23-XL | 0.756 | 0.813 | 0.540 | 0.605 |
| MetricX-23-Large | 0.769 | 0.759 | 0.507 | 0.595 |
| MetricX-23-QE-XXL | 0.769 | 0.830 | 0.490 | 0.606 |
| MetricX-23-QE-XL | 0.718 | 0.684 | 0.421 | 0.594 |
| MetricX-23-QE-Large | 0.744 | 0.671 | 0.387 | 0.579 |
English-Russian:
| Model | System-Level Accuracy | System-Level Pearson | Segment-Level Pearson | Segment-Level Pairwise Acc |
| ----------- | ----------- | ----------- | ----------- | ----------- |
| MetricX-23-XXL | 0.905 | 0.943 | 0.477 | 0.609 |
| MetricX-23-XL | 0.876 | 0.906 | 0.498 | 0.589 |
| MetricX-23-Large | 0.876 | 0.841 | 0.474 | 0.569 |
| MetricX-23-QE-XXL | 0.895 | 0.940 | 0.470 | 0.602 |
| MetricX-23-QE-XL | 0.848 | 0.861 | 0.415 | 0.570 |
| MetricX-23-QE-Large | 0.819 | 0.778 | 0.411 | 0.551 |
Chinese-English:
| Model | System-Level Accuracy | System-Level Pearson | Segment-Level Pearson | Segment-Level Pairwise Acc |
| ----------- | ----------- | ----------- | ----------- | ----------- |
| MetricX-23-XXL | 0.868 | 0.919 | 0.605 | 0.551 |
| MetricX-23-XL | 0.868 | 0.924 | 0.584 | 0.543 |
| MetricX-23-Large | 0.857 | 0.919 | 0.555 | 0.539 |
| MetricX-23-QE-XXL | 0.857 | 0.928 | 0.573 | 0.544 |
| MetricX-23-QE-XL | 0.802 | 0.879 | 0.546 | 0.529 |
| MetricX-23-QE-Large | 0.758 | 0.904 | 0.522 | 0.529 |
The `metricx23/evaluate_wmt23.py` script re-calculates the average correlation
score that was used to rank submissions from the
[WMT'23 Shared Task](https://www2.statmt.org/wmt23/pdf/2023.wmt-1.51.pdf).
Example usage:
```bash
python -m metricx23.evaluate_wmt23 \
--en_de predictions_ende.jsonl \
--he_en predictions_heen.jsonl \
--zh_en predictions_zhen.jsonl \
--output_file output.json
```
Each of the 3 input files is expected to be in the same format as described
above. Each file should correspond to running inference on each of the language
pairs from the WMT'23 dataset.
The results for each of the models is the following:
| Model | Average Correlation |
| ----------- | ----------- |
| MetricX-23-XXL | 0.812 |
| MetricX-23-XL | 0.813 |
| MetricX-23-Large | 0.794 |
| MetricX-23-QE-XXL | 0.797 |
| MetricX-23-QE-XL | 0.767 |
| MetricX-23-QE-Large | 0.762 |
## Citation
If you use MetricX-23 in your research, please cite the following publication:
```bibtex
@inproceedings{juraska-etal-2023-metricx,
title = {{MetricX-23: The Google Submission to the WMT 2023 Metrics Shared Task}},
author = "Juraska, Juraj and
Finkelstein, Mara and
Deutsch, Daniel and
Siddhant, Aditya and
Mirzazadeh, Mehdi and
Freitag, Markus",
editor = "Koehn, Philipp and
Haddow, Barry and
Kocmi, Tom and
Monz, Christof",
booktitle = "Proceedings of the Eighth Conference on Machine Translation",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.wmt-1.63",
doi = "10.18653/v1/2023.wmt-1.63",
pages = "756--767",
}
``` | {"license": "apache-2.0"} | null | google/metricx-23-qe-xxl-v2p0 | [
"transformers",
"pytorch",
"mt5",
"license:apache-2.0",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T16:34:57+00:00 | [] | [] | TAGS
#transformers #pytorch #mt5 #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us
| MetricX-23
==========
*This is not an officially supported Google product.*
GitHub repository: URL
This repository contains the MetricX-23 models,
a family of models for automatic evaluation of translations that were proposed
in the WMT'23 Metrics Shared Task submission
MetricX-23: The Google Submission to the WMT 2023 Metrics Shared Task.
The models were trained in T5X and
then converted for use in PyTorch.
Available Models
----------------
There are 6 models available on HuggingFace that vary in the number of
parameters and whether or not the model is reference-based or reference-free
(also known as quality estimation, or QE):
* MetricX-23-XXL
* MetricX-23-XL
* MetricX-23-Large
* MetricX-23-QE-XXL
* MetricX-23-QE-XL
* MetricX-23-QE-Large
We recommend using the XXL model versions for the best agreement with human
judgments of translation quality, the Large versions for best speed, and the
XL for an intermediate use case.
Changes to the WMT'23 Submission
--------------------------------
These models available here are most similar to the primary submission to the WMT'23 Metrics
Shared Task. They are initialized with mT5
then fine-tuned on a combination of direct assessment and MQM data. However,
we made some changes that make these models different from the WMT'23 submissions.
First, the models are trained to regress the actual MQM score rather than a
normalized score between 0 and 1. That means the output from the MetricX-23
models is a score in the range [0, 25] where lower is better (i.e., it predicts
an error score).
Second, these models were trained with a larger variety of synthetic data that
makes them more robust to translation edge cases like over- and undertranslation,
described in more detail in the following section.
### Synthetic Data
In order for our MetricX models to learn to identify certain types of bad
translations that are not sufficiently (or at all) represented in the regular
training data, we created synthetic examples and mixed them in during training.
The synthetic training data was generated from the DA datasets ranging from
WMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have
the candidate translation manipulated so as to turn it into a bad translation
with a specific issue commonly unrecognized by learned metrics.
The table below provides an overview of the various failure modes that we
considered, including brief descriptions of how we prepared the synthetic data
to address them.
Examples from the first 4 categories were assigned a label corresponding to the
worst score on the given rating scale (e.g., 25 when mixed with MQM training
data), whereas the reference-matching translation examples are assigned the best
score (e.g., 0 when used with MQM data). The missing/incorrect punctuation
examples were labeled with a score slightly worse than perfect.
Note that some of the synthetic datasets are only meaningful in the
reference-based scenario, and we thus excluded them when training a QE variant
of MetricX. These are the Latin-vs-special punctuation and the
reference-matching translation examples.
Most of the synthetic training sets were created using stratified sampling
across target languages, taking 500 examples per target language. One exception
is the missing punctuation set, which used a stratified sample across different
punctuation symbols instead.
When training MetricX, a small proportion of the synthetic examples was mixed
with the regular training examples. During the first-stage fine-tuning on DA
data, each synthetic training set constituted between 0.1% and 1% of all
training examples, whereas in the second-stage fine-tuning on MQM data we used
an even smaller proportion, around 0.05%.
As for evaluating the effect of the synthetic training data on the model's
performance, the DEMETR challenge set - which we originally used to evaluate the
models submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We
therefore created a new DEMETR-style test set based on the WMT22 DA data, with
examples constructed analogically to the synthetic training examples, as
described above. This test set helped us determine the right proportions of
synthetic data for fine-tuning in order to make MetricX robust for the failure
modes in consideration, without sacrificing the system- and segment-level
correlations with human ratings.
Usage
-----
The code for using MetricX models can be found at URL
The repository contains example prediction scripts, described below.
The 'metricx23/URL' script contains an example for how to run inference
on the models.
### Reference-Based
Example usage for a reference-based model:
'URL' is expected to have 1 serialized JSON object per line with
'"reference"' and '"hypothesis"' fields. The output jsonl will be parallel
to 'URL' but additionally contain a '"prediction"' field with the predicted score.
Note that the model was trained with a maximum input length of 1024 tokens, so
significantly increasing that value may lead to unpredictable behavior.
### Reference-Free
Example usage for a reference-free model:
'URL' is expected to have 1 serialized JSON object per line with
'"source"' and '"hypothesis"' fields. The output jsonl will be parallel
to 'URL' but additionally contain a '"prediction"' field with the predicted score.
Meta-Evaluation
---------------
The 'metricx23/URL' script contains code to calculate various correlations
between the MetricX-23 scores and MQM ratings of translation quality using the
MT Metrics Eval library.
Example usage:
'URL' is expected to have one JSON object serialized per line.
Each JSON object is expected to contain 4 fields:
* '"system\_id"': The name of the system that generated the translation.
* '"segment\_id"': The 0-based index of the corresponding segment in the MT
Metrics Eval data.
* '"label"': The ground-truth translation quality score (with higher is better).
* '"prediction"': The model predicted translation quality score (with lower is
better; the script negates the scores so higher is better).
The script will calculate the 4 agreement/correlations that were used in the
WMT'23 Shared Task. Below are the results for the MetricX-23 models on the
WMT'22 Metrics Shared Task data:
English-German:
English-Russian:
Chinese-English:
The 'metricx23/evaluate\_wmt23.py' script re-calculates the average correlation
score that was used to rank submissions from the
WMT'23 Shared Task.
Example usage:
Each of the 3 input files is expected to be in the same format as described
above. Each file should correspond to running inference on each of the language
pairs from the WMT'23 dataset.
The results for each of the models is the following:
If you use MetricX-23 in your research, please cite the following publication:
| [
"### Synthetic Data\n\n\nIn order for our MetricX models to learn to identify certain types of bad\ntranslations that are not sufficiently (or at all) represented in the regular\ntraining data, we created synthetic examples and mixed them in during training.\nThe synthetic training data was generated from the DA datasets ranging from\nWMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have\nthe candidate translation manipulated so as to turn it into a bad translation\nwith a specific issue commonly unrecognized by learned metrics.\n\n\nThe table below provides an overview of the various failure modes that we\nconsidered, including brief descriptions of how we prepared the synthetic data\nto address them.\n\n\n\nExamples from the first 4 categories were assigned a label corresponding to the\nworst score on the given rating scale (e.g., 25 when mixed with MQM training\ndata), whereas the reference-matching translation examples are assigned the best\nscore (e.g., 0 when used with MQM data). The missing/incorrect punctuation\nexamples were labeled with a score slightly worse than perfect.\n\n\nNote that some of the synthetic datasets are only meaningful in the\nreference-based scenario, and we thus excluded them when training a QE variant\nof MetricX. These are the Latin-vs-special punctuation and the\nreference-matching translation examples.\n\n\nMost of the synthetic training sets were created using stratified sampling\nacross target languages, taking 500 examples per target language. One exception\nis the missing punctuation set, which used a stratified sample across different\npunctuation symbols instead.\n\n\nWhen training MetricX, a small proportion of the synthetic examples was mixed\nwith the regular training examples. During the first-stage fine-tuning on DA\ndata, each synthetic training set constituted between 0.1% and 1% of all\ntraining examples, whereas in the second-stage fine-tuning on MQM data we used\nan even smaller proportion, around 0.05%.\n\n\nAs for evaluating the effect of the synthetic training data on the model's\nperformance, the DEMETR challenge set - which we originally used to evaluate the\nmodels submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We\ntherefore created a new DEMETR-style test set based on the WMT22 DA data, with\nexamples constructed analogically to the synthetic training examples, as\ndescribed above. This test set helped us determine the right proportions of\nsynthetic data for fine-tuning in order to make MetricX robust for the failure\nmodes in consideration, without sacrificing the system- and segment-level\ncorrelations with human ratings.\n\n\nUsage\n-----\n\n\nThe code for using MetricX models can be found at URL\nThe repository contains example prediction scripts, described below.\n\n\nThe 'metricx23/URL' script contains an example for how to run inference\non the models.",
"### Reference-Based\n\n\nExample usage for a reference-based model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"reference\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nNote that the model was trained with a maximum input length of 1024 tokens, so\nsignificantly increasing that value may lead to unpredictable behavior.",
"### Reference-Free\n\n\nExample usage for a reference-free model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"source\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nMeta-Evaluation\n---------------\n\n\nThe 'metricx23/URL' script contains code to calculate various correlations\nbetween the MetricX-23 scores and MQM ratings of translation quality using the\nMT Metrics Eval library.\n\n\nExample usage:\n\n\n'URL' is expected to have one JSON object serialized per line.\nEach JSON object is expected to contain 4 fields:\n\n\n* '\"system\\_id\"': The name of the system that generated the translation.\n* '\"segment\\_id\"': The 0-based index of the corresponding segment in the MT\nMetrics Eval data.\n* '\"label\"': The ground-truth translation quality score (with higher is better).\n* '\"prediction\"': The model predicted translation quality score (with lower is\nbetter; the script negates the scores so higher is better).\n\n\nThe script will calculate the 4 agreement/correlations that were used in the\nWMT'23 Shared Task. Below are the results for the MetricX-23 models on the\nWMT'22 Metrics Shared Task data:\n\n\nEnglish-German:\n\n\n\nEnglish-Russian:\n\n\n\nChinese-English:\n\n\n\nThe 'metricx23/evaluate\\_wmt23.py' script re-calculates the average correlation\nscore that was used to rank submissions from the\nWMT'23 Shared Task.\n\n\nExample usage:\n\n\nEach of the 3 input files is expected to be in the same format as described\nabove. Each file should correspond to running inference on each of the language\npairs from the WMT'23 dataset.\n\n\nThe results for each of the models is the following:\n\n\n\nIf you use MetricX-23 in your research, please cite the following publication:"
] | [
"TAGS\n#transformers #pytorch #mt5 #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us \n",
"### Synthetic Data\n\n\nIn order for our MetricX models to learn to identify certain types of bad\ntranslations that are not sufficiently (or at all) represented in the regular\ntraining data, we created synthetic examples and mixed them in during training.\nThe synthetic training data was generated from the DA datasets ranging from\nWMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have\nthe candidate translation manipulated so as to turn it into a bad translation\nwith a specific issue commonly unrecognized by learned metrics.\n\n\nThe table below provides an overview of the various failure modes that we\nconsidered, including brief descriptions of how we prepared the synthetic data\nto address them.\n\n\n\nExamples from the first 4 categories were assigned a label corresponding to the\nworst score on the given rating scale (e.g., 25 when mixed with MQM training\ndata), whereas the reference-matching translation examples are assigned the best\nscore (e.g., 0 when used with MQM data). The missing/incorrect punctuation\nexamples were labeled with a score slightly worse than perfect.\n\n\nNote that some of the synthetic datasets are only meaningful in the\nreference-based scenario, and we thus excluded them when training a QE variant\nof MetricX. These are the Latin-vs-special punctuation and the\nreference-matching translation examples.\n\n\nMost of the synthetic training sets were created using stratified sampling\nacross target languages, taking 500 examples per target language. One exception\nis the missing punctuation set, which used a stratified sample across different\npunctuation symbols instead.\n\n\nWhen training MetricX, a small proportion of the synthetic examples was mixed\nwith the regular training examples. During the first-stage fine-tuning on DA\ndata, each synthetic training set constituted between 0.1% and 1% of all\ntraining examples, whereas in the second-stage fine-tuning on MQM data we used\nan even smaller proportion, around 0.05%.\n\n\nAs for evaluating the effect of the synthetic training data on the model's\nperformance, the DEMETR challenge set - which we originally used to evaluate the\nmodels submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We\ntherefore created a new DEMETR-style test set based on the WMT22 DA data, with\nexamples constructed analogically to the synthetic training examples, as\ndescribed above. This test set helped us determine the right proportions of\nsynthetic data for fine-tuning in order to make MetricX robust for the failure\nmodes in consideration, without sacrificing the system- and segment-level\ncorrelations with human ratings.\n\n\nUsage\n-----\n\n\nThe code for using MetricX models can be found at URL\nThe repository contains example prediction scripts, described below.\n\n\nThe 'metricx23/URL' script contains an example for how to run inference\non the models.",
"### Reference-Based\n\n\nExample usage for a reference-based model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"reference\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nNote that the model was trained with a maximum input length of 1024 tokens, so\nsignificantly increasing that value may lead to unpredictable behavior.",
"### Reference-Free\n\n\nExample usage for a reference-free model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"source\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nMeta-Evaluation\n---------------\n\n\nThe 'metricx23/URL' script contains code to calculate various correlations\nbetween the MetricX-23 scores and MQM ratings of translation quality using the\nMT Metrics Eval library.\n\n\nExample usage:\n\n\n'URL' is expected to have one JSON object serialized per line.\nEach JSON object is expected to contain 4 fields:\n\n\n* '\"system\\_id\"': The name of the system that generated the translation.\n* '\"segment\\_id\"': The 0-based index of the corresponding segment in the MT\nMetrics Eval data.\n* '\"label\"': The ground-truth translation quality score (with higher is better).\n* '\"prediction\"': The model predicted translation quality score (with lower is\nbetter; the script negates the scores so higher is better).\n\n\nThe script will calculate the 4 agreement/correlations that were used in the\nWMT'23 Shared Task. Below are the results for the MetricX-23 models on the\nWMT'22 Metrics Shared Task data:\n\n\nEnglish-German:\n\n\n\nEnglish-Russian:\n\n\n\nChinese-English:\n\n\n\nThe 'metricx23/evaluate\\_wmt23.py' script re-calculates the average correlation\nscore that was used to rank submissions from the\nWMT'23 Shared Task.\n\n\nExample usage:\n\n\nEach of the 3 input files is expected to be in the same format as described\nabove. Each file should correspond to running inference on each of the language\npairs from the WMT'23 dataset.\n\n\nThe results for each of the models is the following:\n\n\n\nIf you use MetricX-23 in your research, please cite the following publication:"
] | [
42,
666,
111,
457
] | [
"passage: TAGS\n#transformers #pytorch #mt5 #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us \n",
"passage: ### Synthetic Data\n\n\nIn order for our MetricX models to learn to identify certain types of bad\ntranslations that are not sufficiently (or at all) represented in the regular\ntraining data, we created synthetic examples and mixed them in during training.\nThe synthetic training data was generated from the DA datasets ranging from\nWMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have\nthe candidate translation manipulated so as to turn it into a bad translation\nwith a specific issue commonly unrecognized by learned metrics.\n\n\nThe table below provides an overview of the various failure modes that we\nconsidered, including brief descriptions of how we prepared the synthetic data\nto address them.\n\n\n\nExamples from the first 4 categories were assigned a label corresponding to the\nworst score on the given rating scale (e.g., 25 when mixed with MQM training\ndata), whereas the reference-matching translation examples are assigned the best\nscore (e.g., 0 when used with MQM data). The missing/incorrect punctuation\nexamples were labeled with a score slightly worse than perfect.\n\n\nNote that some of the synthetic datasets are only meaningful in the\nreference-based scenario, and we thus excluded them when training a QE variant\nof MetricX. These are the Latin-vs-special punctuation and the\nreference-matching translation examples.\n\n\nMost of the synthetic training sets were created using stratified sampling\nacross target languages, taking 500 examples per target language. One exception\nis the missing punctuation set, which used a stratified sample across different\npunctuation symbols instead.\n\n\nWhen training MetricX, a small proportion of the synthetic examples was mixed\nwith the regular training examples. During the first-stage fine-tuning on DA\ndata, each synthetic training set constituted between 0.1% and 1% of all\ntraining examples, whereas in the second-stage fine-tuning on MQM data we used\nan even smaller proportion, around 0.05%.\n\n\nAs for evaluating the effect of the synthetic training data on the model's\nperformance, the DEMETR challenge set - which we originally used to evaluate the\nmodels submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We\ntherefore created a new DEMETR-style test set based on the WMT22 DA data, with\nexamples constructed analogically to the synthetic training examples, as\ndescribed above. This test set helped us determine the right proportions of\nsynthetic data for fine-tuning in order to make MetricX robust for the failure\nmodes in consideration, without sacrificing the system- and segment-level\ncorrelations with human ratings.\n\n\nUsage\n-----\n\n\nThe code for using MetricX models can be found at URL\nThe repository contains example prediction scripts, described below.\n\n\nThe 'metricx23/URL' script contains an example for how to run inference\non the models.### Reference-Based\n\n\nExample usage for a reference-based model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"reference\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nNote that the model was trained with a maximum input length of 1024 tokens, so\nsignificantly increasing that value may lead to unpredictable behavior."
] | [
-0.08049466460943222,
-0.010928971692919731,
-0.004997325129806995,
0.02674533985555172,
0.10898007452487946,
0.017911424860358238,
0.030734051018953323,
0.10377532243728638,
-0.02402583509683609,
-0.03567541018128395,
0.03877764940261841,
0.08288230746984482,
0.015161662362515926,
0.022808298468589783,
0.01772439479827881,
-0.1944202184677124,
0.05506649613380432,
0.008813947439193726,
-0.07605475932359695,
0.06908859312534332,
0.11011916399002075,
-0.04383927211165428,
0.047993674874305725,
0.01802058517932892,
-0.046737369149923325,
0.03515304625034332,
0.02192578837275505,
-0.05999843776226044,
0.09158408641815186,
0.08443905413150787,
0.04427182674407959,
0.026185670867562294,
-0.03256708011031151,
-0.2039792835712433,
0.018693745136260986,
0.059222154319286346,
-0.04639671370387077,
0.032644130289554596,
0.06700465828180313,
-0.04408042132854462,
0.10856176912784576,
-0.03563929721713066,
-0.038342323154211044,
0.052534811198711395,
-0.08053379505872726,
-0.11484964191913605,
-0.06861758977174759,
0.005672447383403778,
0.05251080170273781,
0.11462049186229706,
0.016137007623910904,
0.08802412450313568,
-0.08221147954463959,
0.039690323173999786,
0.1186930239200592,
-0.24400970339775085,
0.022698264569044113,
0.031076578423380852,
0.0391852930188179,
0.059262461960315704,
-0.02663654275238514,
-0.011056511662900448,
0.04030945897102356,
0.036150794476270676,
0.04718852788209915,
-0.02858263999223709,
0.022121166810393333,
0.03753712773323059,
-0.11128073930740356,
-0.06610941141843796,
0.2194429337978363,
0.009360874071717262,
-0.039162129163742065,
-0.012831603176891804,
-0.040371835231781006,
-0.021052006632089615,
0.02562984824180603,
0.011642597615718842,
0.0126093290746212,
0.06509723514318466,
0.08087117969989777,
-0.06699725240468979,
-0.1469363272190094,
-0.02156376652419567,
-0.13039535284042358,
0.056314483284950256,
0.015906725078821182,
0.0635187178850174,
-0.13009536266326904,
0.06865113973617554,
-0.11303252726793289,
-0.037289686501026154,
-0.044206660240888596,
-0.09330593794584274,
0.027131132781505585,
0.0172458216547966,
-0.0671110600233078,
-0.061645977199077606,
0.01954670250415802,
0.09166590869426727,
-0.03881584480404854,
0.02113628201186657,
0.06317485123872757,
0.08239984512329102,
0.02081705816090107,
0.08037593215703964,
0.01793535053730011,
-0.011132435873150826,
0.08186338096857071,
-0.06666721403598785,
0.03507205471396446,
-0.00009720586240291595,
-0.05460785701870918,
-0.04377254843711853,
-0.02165292575955391,
0.07720592617988586,
0.024521537125110626,
0.024704361334443092,
-0.022436054423451424,
-0.014114882797002792,
0.15771594643592834,
-0.08077961206436157,
-0.031236806884407997,
-0.011595608666539192,
-0.05022282153367996,
0.047523126006126404,
0.06264422833919525,
0.006788937374949455,
-0.11120638251304626,
0.05265801399946213,
-0.03781677410006523,
-0.03154580295085907,
-0.08177147805690765,
-0.060364820063114166,
0.05910094082355499,
-0.0520191490650177,
0.022135227918624878,
-0.12485349923372269,
-0.14817503094673157,
-0.0194524135440588,
0.00905279628932476,
0.017627842724323273,
-0.04404270648956299,
0.030841557309031487,
-0.021325333043932915,
-0.010873598977923393,
-0.03838267922401428,
0.04409907013177872,
-0.0735226571559906,
0.01654006540775299,
-0.05891093611717224,
0.006638679653406143,
-0.12875565886497498,
0.05859975516796112,
-0.07795718312263489,
-0.008538316003978252,
-0.12053868919610977,
0.06858694553375244,
-0.06013733893632889,
0.06287765502929688,
-0.06418386101722717,
-0.059423450380563736,
-0.05324282497167587,
0.03673676773905754,
-0.011127032339572906,
0.08703723549842834,
-0.11388625204563141,
-0.04371771216392517,
0.06918473541736603,
-0.1147933155298233,
-0.10631323605775833,
0.05931851267814636,
-0.015358645468950272,
0.03748219460248947,
0.07437919080257416,
0.06889817118644714,
0.13736900687217712,
-0.022680290043354034,
-0.030354108661413193,
0.0774422138929367,
-0.09487991780042648,
-0.0831356942653656,
0.043639007955789566,
-0.02724086493253708,
-0.027338385581970215,
0.010795310139656067,
-0.05432993173599243,
0.053616154938936234,
0.004504588432610035,
-0.041023075580596924,
-0.03911237046122551,
-0.03953332453966141,
-0.03608177229762077,
-0.007416178472340107,
0.04547516256570816,
-0.030613455921411514,
-0.04433886334300041,
0.014297567307949066,
0.05688025429844856,
0.0016403227346017957,
0.06881103664636612,
-0.06005219742655754,
0.05407850816845894,
-0.018144575878977776,
0.008445029146969318,
-0.16996923089027405,
-0.050326526165008545,
0.03246239572763443,
-0.04993841052055359,
0.021022705361247063,
0.02538926526904106,
0.02754078060388565,
-0.014172155410051346,
-0.03300439566373825,
-0.01649932749569416,
0.008937809616327286,
0.008251310326159,
-0.05732375383377075,
-0.12281204760074615,
-0.0057578012347221375,
-0.030820120126008987,
-0.01729525998234749,
-0.027541520074009895,
0.012479400262236595,
0.07977741956710815,
0.044011034071445465,
0.009504541754722595,
0.03403853625059128,
0.005829425062984228,
0.01828007400035858,
-0.025431593880057335,
-0.01253400556743145,
0.06733429431915283,
0.029115518555045128,
-0.08058406412601471,
0.13161158561706543,
-0.1250324547290802,
0.017595291137695312,
0.13202296197414398,
-0.10783318430185318,
0.04871136322617531,
-0.04314408451318741,
-0.031379710882902145,
-0.0057036676444113255,
-0.03291696310043335,
-0.022622399032115936,
0.12076297402381897,
-0.004013901576399803,
0.0929318368434906,
-0.07880045473575592,
-0.05966495722532272,
-0.0097917215898633,
0.0038842959329485893,
-0.010753002017736435,
0.0713651031255722,
0.12100948393344879,
-0.12231792509555817,
0.08771493285894394,
0.13989409804344177,
-0.003570202738046646,
0.12119580805301666,
-0.040826037526130676,
-0.07305974513292313,
0.03446398302912712,
0.004951390903443098,
-0.01554139144718647,
0.04345749691128731,
-0.12445223331451416,
0.015536606311798096,
0.05309225618839264,
0.08432695269584656,
0.11454753577709198,
-0.11917590349912643,
0.0067930202931165695,
0.015571664087474346,
-0.03209373727440834,
-0.09838879108428955,
-0.0021094102412462234,
-0.01198030449450016,
0.07585512846708298,
-0.010641621425747871,
0.034573666751384735,
0.06810279190540314,
-0.033153194934129715,
-0.1333116739988327,
0.1549171507358551,
-0.16806340217590332,
-0.2015051692724228,
-0.1637459099292755,
0.0050877295434474945,
-0.04799898713827133,
0.007370667066425085,
0.10294272005558014,
-0.07140845060348511,
-0.04824806749820709,
-0.030684374272823334,
0.044100407510995865,
-0.026180272921919823,
-0.004666412249207497,
-0.05338970944285393,
0.042327139526605606,
0.016776274889707565,
-0.12135301530361176,
-0.006609104573726654,
-0.0036472089122980833,
-0.017405901104211807,
0.041763320565223694,
-0.07078999280929565,
0.07727597653865814,
0.08807438611984253,
0.01986617222428322,
-0.01348377950489521,
-0.04507536441087723,
0.21933329105377197,
-0.05761788785457611,
0.04443945735692978,
0.2112521529197693,
-0.011195655912160873,
0.05879829078912735,
0.052676692605018616,
-0.004360849969089031,
-0.02312907949090004,
0.02751847915351391,
0.013980694115161896,
-0.0917525514960289,
-0.284024178981781,
-0.06044050678610802,
-0.09147080034017563,
-0.06417541950941086,
0.0023470930755138397,
0.06487169861793518,
0.03722584992647171,
0.050735119730234146,
-0.04127505049109459,
-0.014369443990290165,
0.041575632989406586,
0.05546524375677109,
0.09428055584430695,
0.013206657022237778,
0.06720979511737823,
-0.11070599406957626,
0.012433091178536415,
0.11600261181592941,
0.09805083274841309,
0.23822110891342163,
-0.026172611862421036,
0.04857305809855461,
0.11312685161828995,
0.11436714231967926,
0.08747904002666473,
0.10965031385421753,
-0.04313330352306366,
0.029491525143384933,
-0.041477788239717484,
-0.02848794311285019,
-0.03056623786687851,
0.08165007829666138,
-0.021050430834293365,
-0.0669928565621376,
-0.06534460186958313,
-0.021561436355113983,
0.06433352082967758,
0.21831148862838745,
-0.005303980782628059,
-0.07421030104160309,
-0.05741162598133087,
0.06067470833659172,
-0.010603219270706177,
-0.03260648995637894,
0.07761399447917938,
0.04806586354970932,
-0.07070834934711456,
0.08121111243963242,
-0.0008651819080114365,
0.11229313164949417,
0.014927586540579796,
0.018110742792487144,
0.03192968666553497,
-0.006562814116477966,
0.018823236227035522,
0.11194905638694763,
-0.26115158200263977,
0.12093576788902283,
-0.011933675035834312,
0.05421082302927971,
-0.0762992650270462,
0.022125624120235443,
0.04246993362903595,
0.1273622065782547,
0.09262282401323318,
0.043345239013433456,
-0.14447379112243652,
-0.0103498175740242,
0.0019879117608070374,
0.056262798607349396,
0.04705921933054924,
0.010509418323636055,
0.011808192357420921,
-0.011274687014520168,
0.013288512825965881,
0.015993541106581688,
0.08853989839553833,
-0.0517207607626915,
-0.16778728365898132,
0.04126080125570297,
-0.006679844111204147,
0.015364071354269981,
-0.07193623483181,
0.005432065576314926,
-0.039848387241363525,
0.08026738464832306,
-0.14921778440475464,
-0.0995694100856781,
-0.09588959813117981,
-0.08473988622426987,
0.08787038922309875,
-0.0654374435544014,
0.06406596302986145,
0.007063437253236771,
0.07009734213352203,
-0.03068877011537552,
-0.14400194585323334,
0.057342350482940674,
-0.11315649747848511,
-0.08588340878486633,
-0.0034783631563186646,
0.1125456690788269,
0.013006072491407394,
0.0326993428170681,
0.019293401390314102,
-0.010193015448749065,
-0.046051137149333954,
-0.11050929129123688,
0.01819794625043869,
0.07997186481952667,
0.05762843042612076,
0.05681586265563965,
-0.09417805820703506,
-0.029595911502838135,
-0.04396835342049599,
-0.06783974170684814,
0.1667174994945526,
0.10756810754537582,
-0.06620179116725922,
0.1170005351305008,
0.16132885217666626,
-0.1295406073331833,
-0.19067947566509247,
-0.043041959404945374,
-0.03716177120804787,
-0.01986772194504738,
-0.02807173877954483,
-0.1475318968296051,
0.07539469003677368,
0.09812235087156296,
-0.0024198442697525024,
0.04476727545261383,
-0.2080758810043335,
-0.08643566817045212,
0.04781382530927658,
0.0001132916659116745,
0.18663114309310913,
-0.10906235128641129,
-0.044622793793678284,
-0.052799079567193985,
-0.11352433264255524,
0.14947330951690674,
-0.06072360277175903,
0.05728284642100334,
0.00732648279517889,
0.0032600341364741325,
-0.006010896060615778,
-0.03202845901250839,
0.13251455128192902,
0.03674418851733208,
0.043941304087638855,
-0.028950639069080353,
0.055670365691185,
0.025221113115549088,
-0.04210628941655159,
0.044331781566143036,
-0.04143795743584633,
0.021573659032583237,
-0.14452756941318512,
-0.045783452689647675,
-0.08391499519348145,
0.04536336287856102,
-0.01962660811841488,
-0.039184920489788055,
-0.042702145874500275,
0.0021056830883026123,
0.08968086540699005,
-0.014609623700380325,
0.054970160126686096,
-0.05837482213973999,
0.059537388384342194,
0.07670764625072479,
0.15645602345466614,
-0.10332942008972168,
-0.1217908263206482,
-0.02813919633626938,
-0.021626751869916916,
0.08217503130435944,
-0.1376548409461975,
0.07488036155700684,
0.11647255718708038,
-0.04862956330180168,
0.08627215772867203,
0.055094070732593536,
-0.07522906363010406,
0.005458813160657883,
0.10446281731128693,
-0.07246147096157074,
-0.06008515506982803,
-0.0390506386756897,
0.08281482756137848,
0.029428621754050255,
0.03669393062591553,
0.12870027124881744,
-0.0009662613738328218,
-0.032006535679101944,
-0.02065277099609375,
0.06402802467346191,
-0.0347125381231308,
0.0842379555106163,
0.039138346910476685,
0.017695903778076172,
-0.08644309639930725,
0.1338755190372467,
0.01591717265546322,
-0.1885780692100525,
-0.009080829098820686,
0.11049890518188477,
-0.11632084846496582,
-0.0807068943977356,
-0.025578761473298073,
0.009648185223340988,
-0.14496804773807526,
-0.10269727557897568,
-0.032402172684669495,
-0.14044955372810364,
0.06719717383384705,
0.16770902276039124,
0.02116461656987667,
0.03302031010389328,
-0.006035410333424807,
-0.028474289923906326,
-0.03122306987643242,
-0.013931620866060257,
-0.0277186818420887,
0.017339259386062622,
-0.05742930248379707,
0.0389898419380188,
0.019052214920520782,
0.06222548335790634,
-0.016825642436742783,
0.015219002030789852,
-0.0635049119591713,
0.010224338620901108,
-0.15755128860473633,
0.024624407291412354,
-0.05316181108355522,
-0.017552291974425316,
-0.01309568714350462,
-0.014190688729286194,
-0.05401631444692612,
0.05547712743282318,
-0.09220357239246368,
-0.020396772772073746,
-0.03640173375606537,
0.027906954288482666,
-0.05868091061711311,
-0.02378096804022789,
0.050178419798612595,
-0.047896455973386765,
0.1031966358423233,
0.06362350285053253,
-0.05172618851065636,
0.0687813088297844,
-0.054818395525217056,
-0.03117077425122261,
0.04752342030405998,
0.04614630341529846,
0.0024338315706700087,
-0.06958136707544327,
0.053022030740976334,
0.0532219260931015,
-0.05545207858085632,
-0.00179347675293684,
-0.00773459579795599,
-0.106658935546875,
-0.026197271421551704,
0.026790166273713112,
-0.007194001227617264,
-0.06978121399879456,
-0.046734243631362915,
0.11590144783258438,
0.04049905017018318,
0.1221245527267456,
-0.006192055530846119,
0.039656929671764374,
-0.0669143944978714,
0.0076656281016767025,
-0.0022521058563143015,
-0.0823831707239151,
-0.09226924180984497,
-0.03779274597764015,
0.01616818644106388,
-0.032287146896123886,
0.20258086919784546,
-0.027631420642137527,
-0.001419171690940857,
0.05651244521141052,
0.0006612688302993774,
-0.019393347203731537,
0.012330395169556141,
0.1308664083480835,
0.009836757555603981,
0.00910251121968031,
-0.02165895141661167,
0.024475062265992165,
-0.016778085380792618,
-0.0041928887367248535,
0.13951833546161652,
0.09686414152383804,
0.16705356538295746,
0.1026294082403183,
0.002041283994913101,
-0.03730851784348488,
-0.01945573464035988,
-0.06291651725769043,
0.028470784425735474,
0.058277349919080734,
0.00701283710077405,
0.01671595126390457,
0.11799974739551544,
-0.09208066761493683,
0.09887351095676422,
-0.06913766264915466,
-0.05150070786476135,
-0.1154223382472992,
-0.09975318610668182,
-0.010319891385734081,
-0.08603835850954056,
-0.029209984466433525,
-0.13546273112297058,
-0.020355189219117165,
0.06884671002626419,
0.038902945816516876,
-0.05183970928192139,
0.007247757166624069,
-0.11964748054742813,
-0.10397026687860489,
0.04866647720336914,
-0.014942724257707596,
0.04801199212670326,
0.04992463439702988,
0.012288471683859825,
0.04020651429891586,
-0.03871965408325195,
-0.02611941285431385,
0.039849359542131424,
0.04274921864271164,
0.045678697526454926,
-0.03452969342470169,
-0.010724259540438652,
-0.028795840218663216,
0.012287003919482231,
0.02913379855453968,
0.22296985983848572,
0.010857714340090752,
-0.031127674505114555,
0.06575135886669159,
0.19841772317886353,
-0.04081185162067413,
-0.09934508055448532,
-0.09630105644464493,
0.1641465723514557,
0.045690394937992096,
0.029041193425655365,
0.016438156366348267,
-0.057314880192279816,
-0.00565880723297596,
0.22985684871673584,
0.17313337326049805,
-0.025465190410614014,
-0.009001774713397026,
-0.006234009750187397,
0.014965811744332314,
0.06765225529670715,
0.08408764004707336,
0.1142108142375946,
0.21731261909008026,
-0.06078370288014412,
0.0639389306306839,
-0.0446958988904953,
0.0698341652750969,
-0.034699127078056335,
0.17769397795200348,
0.001534847542643547,
-0.058601561933755875,
-0.0005737040191888809,
0.06745973229408264,
-0.07404687255620956,
-0.10022060573101044,
-0.03632500767707825,
-0.07546216249465942,
-0.05532532557845116,
-0.0036639878526329994,
0.05928308516740799,
0.022997058928012848,
0.041565168648958206,
-0.0059436894953250885,
-0.044010479003190994,
0.10706667602062225,
0.023562224581837654,
-0.17537671327590942,
-0.05920793116092682,
0.08894800394773483,
0.011057954281568527,
0.08248458802700043,
0.007951926440000534,
0.048754554241895676,
0.07727794349193573,
0.03654977306723595,
-0.11864081025123596,
0.07930269092321396,
-0.00997848343104124,
-0.05712908133864403,
0.007062149234116077,
0.014758620411157608,
-0.00663449801504612,
0.020782694220542908,
0.029319513589143753,
-0.03974604234099388,
0.05011465772986412,
0.03723395615816116,
-0.04944480583071709,
-0.00523018091917038,
0.03254415839910507,
-0.10453616082668304,
0.08378694206476212,
0.07704398781061172,
-0.022326018661260605,
-0.027006199583411217,
-0.05415914207696915,
0.02430926077067852,
0.012829716317355633,
-0.023813046514987946,
-0.024040259420871735,
-0.06499078124761581,
0.017759330570697784,
0.01358121633529663,
0.03267041966319084,
-0.13761897385120392,
-0.03263425827026367,
0.009745022282004356,
-0.028417814522981644,
-0.08249472081661224,
0.06735780090093613,
0.043854910880327225,
0.026215003803372383,
-0.03867888078093529,
0.0033451877534389496,
-0.004438048228621483,
0.04008375108242035,
-0.0645875409245491,
-0.08099328726530075
] |
null | null | transformers |

---
Presenting the FP16 files for Psyonic-Cetacean-20B! This is an experimental Llama2-based stack merge based on the models and recipe below:
- [KoboldAI/PsyFighter-2-13b](https://huggingface.co/KoboldAI/LLaMA2-13B-Psyfighter2-GGUF)
- [microsoft/Orca-2-13b](https://huggingface.co/microsoft/Orca-2-13b)
```yaml
slices:
- sources:
- model: Orca2flat
layer_range: [0, 16]
- sources:
- model: LLaMA2-13B-Psyfighter2 (FP16 not yet available)
layer_range: [8, 24]
- sources:
- model: Orca2flat
layer_range: [17, 32]
- sources:
- model: LLaMA2-13B-Psyfighter2 (FP16 not yet available)
layer_range: [25, 40]
merge_method: passthrough
dtype: float16
```
Note: while we did run an inverted merge the output was not satisfactory and will not be released.
We first flatted the additional ChatML vocabulary tokens out of Orca-2-13B, then performed a stack merge with Psyfighter-2-13B. The results surprised us with their vividness, freshness of prose, obedience to instruction prompting, and formatting cohesion.
This model is focused on storywriting and text adventure, with a side order of Assistant and Chat functionality. Like its ancestor Psyfighter-2 this model will function better if you let it improvise and riff on your concepts rather than feeding it an excess of detail.
Additionally, either the removal of the ChatML vocab or the stack merging process itself has resulted in not only an uncensored model but an actively anti-censored model, so please be aware that this model can and will kill you during adventures or output NSFW material if prompted accordingly.
During testing, the model exhibited an especially strong affinity for science fiction and space opera writing, while handling fantasy elements quite well and horror elements slightly less so. Refer to the Psyfighter-2 model card for best prompting practices.
Despite that, we have tested the model out to 16000 context via Rope scaling and the model does not drive towards NSFW on its own. It will follow your tone and style very well.
Please enjoy, and if you encounter anything exciting or weird, please reach out to me at [[email protected]].
Special thanks as always to the KoboldAI crew who provided the mergebox, testing, and feedback on this model, and to gelukuMLG for the model mascot! | {"license": "other", "tags": ["storywriting", "text adventure", "not-for-all-audiences"], "license_name": "microsoft-research-license"} | text-generation | zaq-hack/psyonic-cetacean-20B-bpw350-h6-exl2-rpcal | [
"transformers",
"safetensors",
"llama",
"text-generation",
"storywriting",
"text adventure",
"not-for-all-audiences",
"license:other",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T16:35:02+00:00 | [] | [] | TAGS
#transformers #safetensors #llama #text-generation #storywriting #text adventure #not-for-all-audiences #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
!image/png
---
Presenting the FP16 files for Psyonic-Cetacean-20B! This is an experimental Llama2-based stack merge based on the models and recipe below:
- KoboldAI/PsyFighter-2-13b
- microsoft/Orca-2-13b
Note: while we did run an inverted merge the output was not satisfactory and will not be released.
We first flatted the additional ChatML vocabulary tokens out of Orca-2-13B, then performed a stack merge with Psyfighter-2-13B. The results surprised us with their vividness, freshness of prose, obedience to instruction prompting, and formatting cohesion.
This model is focused on storywriting and text adventure, with a side order of Assistant and Chat functionality. Like its ancestor Psyfighter-2 this model will function better if you let it improvise and riff on your concepts rather than feeding it an excess of detail.
Additionally, either the removal of the ChatML vocab or the stack merging process itself has resulted in not only an uncensored model but an actively anti-censored model, so please be aware that this model can and will kill you during adventures or output NSFW material if prompted accordingly.
During testing, the model exhibited an especially strong affinity for science fiction and space opera writing, while handling fantasy elements quite well and horror elements slightly less so. Refer to the Psyfighter-2 model card for best prompting practices.
Despite that, we have tested the model out to 16000 context via Rope scaling and the model does not drive towards NSFW on its own. It will follow your tone and style very well.
Please enjoy, and if you encounter anything exciting or weird, please reach out to me at [jebcarter@URL].
Special thanks as always to the KoboldAI crew who provided the mergebox, testing, and feedback on this model, and to gelukuMLG for the model mascot! | [] | [
"TAGS\n#transformers #safetensors #llama #text-generation #storywriting #text adventure #not-for-all-audiences #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
67
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #storywriting #text adventure #not-for-all-audiences #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
0.024552643299102783,
0.06800024956464767,
-0.006275040563195944,
0.025086428970098495,
0.07575658708810806,
0.03634592890739441,
0.25202906131744385,
0.08515089005231857,
-0.011156481690704823,
-0.05512445047497749,
0.17021185159683228,
0.13841918110847473,
-0.03512859344482422,
0.1285029947757721,
-0.11925232410430908,
-0.20949257910251617,
0.03797439858317375,
-0.015636201947927475,
0.13904894888401031,
0.0971633568406105,
0.11016163229942322,
-0.045477233827114105,
0.031189212575554848,
-0.029684241861104965,
-0.049808647483587265,
0.012434341013431549,
0.0614062175154686,
-0.1400814801454544,
0.12151455879211426,
0.03631461784243584,
0.09849084913730621,
0.057987798005342484,
0.01387382484972477,
-0.19102664291858673,
0.04213429614901543,
-0.026659240946173668,
-0.07191905379295349,
-0.011026190593838692,
0.049958232790231705,
-0.06349973380565643,
0.13821005821228027,
0.1146688461303711,
-0.003057145746424794,
0.0849781259894371,
-0.13439074158668518,
0.019003089517354965,
-0.005182833876460791,
0.01815907470881939,
0.0509924553334713,
0.12997621297836304,
-0.036264266818761826,
0.0909242257475853,
-0.043522339314222336,
0.07332784682512283,
0.12362038344144821,
-0.29529646039009094,
-0.04517381265759468,
0.07538816332817078,
0.14856989681720734,
-0.0027046469040215015,
-0.11599653214216232,
0.12709647417068481,
0.07879739999771118,
-0.000878071878105402,
0.012254495173692703,
-0.05533166229724884,
-0.02102483995258808,
-0.005984366871416569,
-0.048018649220466614,
-0.023642435669898987,
0.22733260691165924,
-0.04546692222356796,
0.014757025055587292,
-0.12295781075954437,
-0.04587619751691818,
0.014109168201684952,
-0.026046155020594597,
0.07250696420669556,
-0.02588612586259842,
0.07514931261539459,
-0.004969054367393255,
-0.02780364267528057,
-0.15780109167099,
-0.01355698425322771,
-0.0973406583070755,
0.06659654527902603,
-0.006334282457828522,
0.023584574460983276,
-0.18127739429473877,
0.04526813328266144,
0.07091961055994034,
-0.11227606236934662,
0.007962772622704506,
-0.09481605887413025,
0.08857917040586472,
-0.03652798384428024,
-0.013635757379233837,
-0.05076047033071518,
0.1406492292881012,
0.08988172560930252,
0.04714246094226837,
0.0011104901786893606,
-0.059307802468538284,
0.10416729003190994,
0.03554194048047066,
-0.03440386801958084,
0.11109348386526108,
-0.012477336451411247,
0.06036852300167084,
-0.04641164094209671,
0.08674035966396332,
-0.029425574466586113,
-0.12455346435308456,
0.040593862533569336,
-0.055759038776159286,
0.09845635294914246,
0.04699227958917618,
0.11172306537628174,
-0.03765106573700905,
0.0836324542760849,
-0.024948077276349068,
-0.019617557525634766,
-0.04803768917918205,
-0.005973572842776775,
0.047400686889886856,
0.045478954911231995,
-0.005817574914544821,
0.03922653570771217,
-0.023397700861096382,
0.12197400629520416,
-0.048147112131118774,
-0.058644529432058334,
-0.035749319940805435,
-0.012386869639158249,
0.10729731619358063,
-0.014203619211912155,
0.04573672637343407,
-0.14261916279792786,
-0.22342057526111603,
-0.020009182393550873,
0.021793736144900322,
-0.004237415734678507,
-0.05917438492178917,
-0.08977658301591873,
0.006093802861869335,
0.03232559561729431,
-0.05551949515938759,
-0.07079597562551498,
-0.10328826308250427,
0.09764797240495682,
-0.06331544369459152,
0.0504193939268589,
-0.08403073251247406,
0.0056554777547717094,
-0.10921580344438553,
0.013635451905429363,
0.023804308846592903,
0.011561618186533451,
-0.0031674180645495653,
0.21000196039676666,
0.08261462301015854,
0.06357459723949432,
-0.006601409986615181,
0.09485006332397461,
-0.11328576505184174,
0.2697242498397827,
-0.11523713171482086,
-0.023674432188272476,
0.23737502098083496,
-0.103493832051754,
-0.24764861166477203,
0.10483752191066742,
-0.027484051883220673,
0.17128965258598328,
0.1291683465242386,
0.21719159185886383,
0.018325531855225563,
-0.09560929238796234,
0.06064813956618309,
0.041477859020233154,
-0.09068485349416733,
-0.007705449126660824,
0.03005879372358322,
0.020218372344970703,
-0.04744383320212364,
0.0279146246612072,
0.13358613848686218,
0.0001401962072122842,
-0.015874184668064117,
-0.047281987965106964,
-0.01514775026589632,
-0.030335836112499237,
0.062154728919267654,
-0.02388199418783188,
0.03735813871026039,
-0.1371573954820633,
-0.031697358936071396,
-0.056911736726760864,
0.02699356898665428,
0.009886167012155056,
0.03861241042613983,
-0.06986287981271744,
0.03168988600373268,
0.10698442906141281,
0.061792951077222824,
-0.04639063775539398,
-0.02766101434826851,
-0.07580327242612839,
0.13158337771892548,
0.028735458850860596,
0.11764121800661087,
0.03723444417119026,
-0.020244674757122993,
-0.076547771692276,
0.004391200840473175,
0.14906659722328186,
-0.00561947375535965,
-0.03332475945353508,
-0.11564343422651291,
0.11601313948631287,
-0.06088661774992943,
0.026619553565979004,
-0.19724568724632263,
0.039403051137924194,
0.007507374510169029,
0.040395718067884445,
-0.00899811927229166,
0.04089222475886345,
0.008875788189470768,
-0.03261496126651764,
-0.09673244506120682,
-0.02243170700967312,
0.09705884754657745,
0.004469010978937149,
-0.1396852284669876,
0.17839373648166656,
-0.24855850636959076,
0.2090110033750534,
0.16184480488300323,
-0.27066341042518616,
-0.002455675508826971,
-0.08350001275539398,
0.0007730357465334237,
0.03313853219151497,
-0.04027766361832619,
0.02308301441371441,
-0.005815391428768635,
-0.009480372071266174,
0.16121219098567963,
-0.04701114445924759,
-0.07452880591154099,
-0.027887996286153793,
-0.0807064026594162,
-0.03967132046818733,
0.0888039767742157,
0.00938088446855545,
-0.1790393441915512,
0.17462293803691864,
0.27154117822647095,
0.04770695045590401,
0.26476940512657166,
0.0037540269549936056,
0.0604354627430439,
0.050089988857507706,
-0.020338168367743492,
0.015373273752629757,
-0.020795566961169243,
-0.14726382493972778,
0.013187559321522713,
0.04597891867160797,
-0.013240544125437737,
0.02476980723440647,
-0.09892520308494568,
-0.141377255320549,
-0.020208869129419327,
-0.02446736767888069,
-0.07478200644254684,
0.09729230403900146,
-0.04043428972363472,
0.09148257225751877,
-0.04148007929325104,
0.024069618433713913,
0.10977192223072052,
-0.03983813151717186,
-0.08527882397174835,
0.14562007784843445,
-0.06796591728925705,
-0.25092753767967224,
-0.20157302916049957,
-0.11487630754709244,
-0.009703908115625381,
0.052090249955654144,
0.13602185249328613,
-0.04231968894600868,
-0.038692548871040344,
-0.06209754943847656,
-0.007158034481108189,
-0.09475696831941605,
-0.034049250185489655,
-0.016210326924920082,
0.12361236661672592,
-0.06147293373942375,
-0.06062747910618782,
-0.022181125357747078,
0.00789758749306202,
-0.07334814220666885,
0.13340462744235992,
-0.10077392309904099,
0.10421361029148102,
0.08506505936384201,
0.041940346360206604,
-0.0021395133808255196,
-0.09493318945169449,
0.0940205529332161,
-0.14070361852645874,
0.001426322152838111,
0.20865778625011444,
-0.09155501425266266,
0.05357542261481285,
0.17959506809711456,
0.02837812341749668,
-0.10267776250839233,
0.09350911527872086,
-0.023243924602866173,
-0.08506158739328384,
-0.24538429081439972,
-0.040115781128406525,
-0.07560070604085922,
0.16982731223106384,
0.01556074433028698,
0.08657411485910416,
0.1438826024532318,
0.07791079580783844,
-0.06400014460086823,
-0.06258761882781982,
0.09431915730237961,
0.10531551390886307,
0.25717130303382874,
-0.011607469990849495,
0.08263862133026123,
-0.10608244687318802,
-0.08253690600395203,
0.0807003527879715,
0.002981617348268628,
0.06459205597639084,
0.12218087166547775,
0.09655627608299255,
0.0761043056845665,
0.004725807346403599,
0.09874024242162704,
0.11056267470121384,
0.09913510829210281,
-0.04518439620733261,
-0.03766189143061638,
-0.037798311561346054,
-0.03452032804489136,
0.06429759413003922,
-0.09028049558401108,
-0.11083921045064926,
-0.042543478310108185,
-0.04930552467703819,
0.06531030684709549,
0.10639654099941254,
0.050778668373823166,
-0.05764404311776161,
0.05022019147872925,
0.14868196845054626,
-0.04829220101237297,
-0.12058867514133453,
0.1110382154583931,
0.03420580178499222,
-0.11443531513214111,
0.14529183506965637,
-0.011771789751946926,
0.1382288634777069,
-0.05290690436959267,
0.0881744995713234,
-0.11976394057273865,
-0.15380679070949554,
-0.03562243655323982,
0.12125004827976227,
-0.3409517705440521,
0.2139718234539032,
0.022577650845050812,
0.04588121548295021,
-0.08311774581670761,
-0.06715845316648483,
0.024918219074606895,
0.16455893218517303,
0.1975151002407074,
-0.0301362257450819,
-0.05134611576795578,
-0.0412110760807991,
0.00003511626709951088,
0.019543228670954704,
0.08700890094041824,
-0.029512539505958557,
0.022741146385669708,
-0.08478443324565887,
0.018364109098911285,
-0.022697892040014267,
-0.008467350155115128,
-0.05338098108768463,
-0.15789566934108734,
0.02189684472978115,
0.1127079576253891,
0.13133300840854645,
-0.029481645673513412,
-0.0032808587420731783,
-0.08723343908786774,
0.07290764153003693,
-0.07489017397165298,
-0.027128152549266815,
-0.08750085532665253,
-0.10792524367570877,
-0.058988261967897415,
0.0026672598905861378,
0.010724034160375595,
-0.040624164044857025,
0.0680738240480423,
-0.14056387543678284,
-0.12983058393001556,
0.09150215238332748,
-0.07923561334609985,
-0.11579445004463196,
-0.05087887495756149,
0.12757492065429688,
-0.06683522462844849,
0.03861720487475395,
0.048804886639118195,
0.010099723003804684,
-0.08903469890356064,
-0.1087338775396347,
0.004107189830392599,
-0.0766809955239296,
-0.007311676628887653,
-0.05894060432910919,
-0.07814282178878784,
-0.06319831311702728,
-0.025133220478892326,
-0.1661362498998642,
0.22321660816669464,
0.34867510199546814,
-0.010433822870254517,
0.18686027824878693,
0.14680537581443787,
-0.12579557299613953,
-0.3470672369003296,
-0.13655027747154236,
-0.21006782352924347,
-0.07806874066591263,
0.04293923079967499,
-0.055371955037117004,
0.023506056517362595,
-0.01795108988881111,
-0.02330568991601467,
0.1137828528881073,
-0.17136253416538239,
-0.11587657034397125,
0.12197116017341614,
-0.010856055654585361,
0.32602739334106445,
-0.1982441544532776,
-0.0487438440322876,
-0.13322415947914124,
-0.07285565137863159,
0.055188070982694626,
-0.1204465925693512,
0.12010639160871506,
-0.007437644060701132,
0.09034236520528793,
0.014261064119637012,
-0.024851521477103233,
0.06800531595945358,
-0.13509199023246765,
0.05364105477929115,
-0.14163747429847717,
-0.0014003977412357926,
0.10801897943019867,
-0.03557075187563896,
-0.003281416604295373,
-0.22687503695487976,
-0.05253811553120613,
-0.060481809079647064,
-0.07031818479299545,
-0.03123307228088379,
0.08795113116502762,
-0.03486017882823944,
-0.05428633093833923,
-0.03695521131157875,
-0.04518020898103714,
-0.013758636079728603,
-0.013743762858211994,
0.26902616024017334,
-0.1369444578886032,
0.19221246242523193,
0.023955337703227997,
0.22811441123485565,
-0.07853924483060837,
0.08864311873912811,
-0.05528164282441139,
-0.08313252776861191,
0.00615558261051774,
-0.13969995081424713,
0.023851079866290092,
0.07635126262903214,
-0.05121266841888428,
0.062188442796468735,
0.0992860421538353,
0.01268355268985033,
0.0430474691092968,
0.15104900300502777,
-0.18661701679229736,
-0.08694476634263992,
-0.05470138043165207,
0.12216684967279434,
0.021877799183130264,
0.023732110857963562,
0.13673578202724457,
0.0038171184714883566,
0.028363779187202454,
0.0015041254227980971,
0.02920924685895443,
0.014596535824239254,
-0.003732886165380478,
-0.055541325360536575,
0.016229065135121346,
-0.09915612637996674,
0.0809159055352211,
0.0294171292334795,
-0.11212829500436783,
0.025413040071725845,
0.15533261001110077,
-0.0650755912065506,
-0.11768341064453125,
0.0057175676338374615,
0.148555189371109,
-0.10374071449041367,
-0.04402026906609535,
-0.062073465436697006,
-0.17184202373027802,
0.038068026304244995,
0.2248273491859436,
0.027189888060092926,
0.0705033466219902,
0.006424351595342159,
-0.008606072515249252,
-0.008470558561384678,
0.05588870123028755,
-0.016039615496993065,
-0.0008846343262121081,
-0.1332293599843979,
0.03644512966275215,
-0.022411677986383438,
0.053814224898815155,
-0.09663651138544083,
-0.03823291137814522,
-0.12467819452285767,
0.06714051961898804,
-0.1192275732755661,
0.00023566542949993163,
-0.12214960157871246,
-0.006123457103967667,
-0.0014943060232326388,
0.021250415593385696,
-0.062156952917575836,
-0.07691515982151031,
-0.08534861356019974,
-0.006057679653167725,
-0.002577902050688863,
0.04671122878789902,
-0.08457479625940323,
-0.020621569827198982,
0.06936324387788773,
-0.0422893688082695,
0.09338672459125519,
0.02773372270166874,
-0.0844649076461792,
0.05917053669691086,
-0.272941917181015,
0.0053605628199875355,
0.0876004695892334,
-0.012540568597614765,
-0.04511613771319389,
0.18080012500286102,
-0.04972167685627937,
0.0834890827536583,
0.013573896139860153,
0.08494513481855392,
-0.0717923641204834,
-0.10856453329324722,
0.00018396993982605636,
0.06715316325426102,
-0.1337595134973526,
0.01886873133480549,
-0.06925108283758163,
0.1079791858792305,
-0.10989644378423691,
0.1877453625202179,
-0.10827895998954773,
0.07824420928955078,
-0.02673555351793766,
0.05858476459980011,
0.04249152913689613,
-0.1291210651397705,
-0.08001314848661423,
-0.11095483601093292,
-0.021375169977545738,
-0.004540305119007826,
0.21591472625732422,
0.012018105946481228,
-0.05183045566082001,
0.09976179152727127,
0.0533432699739933,
0.01753421686589718,
0.010756858624517918,
0.25596606731414795,
0.05716362223029137,
-0.0705529972910881,
-0.14858059585094452,
0.0044577959924936295,
0.05593341588973999,
-0.12991869449615479,
0.08481264859437943,
0.03308482840657234,
-0.06647837162017822,
0.04766938090324402,
0.00456581823527813,
0.0951186940073967,
-0.022376136854290962,
-0.1433664709329605,
0.007966206409037113,
0.043242935091257095,
0.00026994472136721015,
0.025908365845680237,
0.17714416980743408,
0.03602594509720802,
-0.02349943108856678,
-0.11821579933166504,
-0.039522428065538406,
-0.1723768413066864,
-0.10499744117259979,
-0.07720645517110825,
-0.07751176506280899,
0.04657287895679474,
-0.06475445628166199,
0.08908320218324661,
0.02177288942039013,
0.07714133709669113,
-0.10031405091285706,
0.09067082405090332,
0.028232000768184662,
-0.04782482236623764,
0.0918969139456749,
-0.03279038146138191,
0.036186881363391876,
-0.07358130812644958,
0.02436944469809532,
-0.06267331540584564,
-0.04309646040201187,
-0.03310341015458107,
0.07883191108703613,
-0.02015214040875435,
0.01041588094085455,
-0.15917716920375824,
-0.07245973497629166,
-0.007560403551906347,
0.0715281292796135,
0.017882371321320534,
0.14909875392913818,
-0.0021014593075960875,
-0.030547933652997017,
0.07310996949672699,
0.2093002051115036,
-0.028740087524056435,
-0.12363199144601822,
-0.00708317244425416,
0.0667002946138382,
-0.040854401886463165,
0.08084486424922943,
-0.09793492406606674,
0.0012357545783743262,
0.016111919656395912,
0.33123570680618286,
0.25315988063812256,
-0.06639370322227478,
0.012642205692827702,
-0.09266163408756256,
0.042206183075904846,
0.03813379630446434,
0.07809721678495407,
0.049722738564014435,
0.18034042418003082,
-0.08328406512737274,
0.04224809631705284,
0.014901530928909779,
0.013792493380606174,
-0.1341090351343155,
0.11629787087440491,
0.015386324375867844,
0.03003045916557312,
-0.052709948271512985,
0.13859224319458008,
-0.20088759064674377,
0.1125604584813118,
-0.10647700726985931,
-0.06685298681259155,
0.01139877736568451,
-0.024373527616262436,
0.17648883163928986,
-0.0187532976269722,
0.04782719537615776,
0.02296639233827591,
-0.024096770212054253,
-0.06161698326468468,
0.004831473808735609,
-0.17059317231178284,
0.015253123827278614,
0.07425881922245026,
-0.09189016371965408,
0.1422913670539856,
-0.0476461723446846,
-0.054282136261463165,
0.06194471940398216,
-0.02206886000931263,
-0.03487440198659897,
0.13465729355812073,
0.02734525501728058,
-0.07007022947072983,
0.050422780215740204,
0.03805816173553467,
-0.017948413267731667,
-0.03648299723863602,
0.09214751422405243,
-0.0532359816133976,
0.06684883683919907,
0.027004847303032875,
-0.06449272483587265,
-0.017657961696386337,
0.11114069074392319,
-0.06183087080717087,
0.072293721139431,
0.02306167222559452,
-0.03551258519291878,
0.0022191223688423634,
-0.02228567749261856,
0.009448587894439697,
-0.02391415275633335,
-0.14522698521614075,
-0.013912297785282135,
-0.0681820958852768,
-0.07442320883274078,
0.12633441388607025,
0.00903337448835373,
-0.3678169548511505,
0.01395304687321186,
-0.11143926531076431,
0.03296716511249542,
-0.16528530418872833,
0.015266507863998413,
0.1719563752412796,
0.00485230190679431,
0.0060460832901299,
-0.15911318361759186,
0.059644103050231934,
0.06673851609230042,
-0.025990379974246025,
-0.0944560095667839
] |
null | null | diffusers |
MeinaMix Objective is to be able to do good art with little prompting.
For examples and prompts, please checkout: https://civitai.com/models/7240/meinamix
I have a discord server where you can post images that you generated, discuss prompt and/or ask for help.
https://discord.gg/XC9nGZNDUd If you like one of my models and want to support their updates
I've made a ko-fi page; https://ko-fi.com/meina where you can pay me a coffee <3
And a Patreon page; https://www.patreon.com/MeinaMix where you can support me and get acess to beta of my models!
You may also try this model using Sinkin.ai: https://sinkin.ai/m/vln8Nwr
MeinaMix and the other of Meinas will ALWAYS be FREE.
Recommendations of use: Enable Quantization in K samplers.
Hires.fix is needed for prompts where the character is far away in order to make decent images, it drastically improve the quality of face and eyes!
Recommended parameters:
Sampler: Euler a: 40 to 60 steps.
Sampler: DPM++ SDE Karras: 20 to 30 steps.
Sampler: DPM++ 2M Karras: 20 to 40 steps.
CFG Scale: 7.
Resolutions: 512x768, 512x1024 for Portrait!
Resolutions: 768x512, 1024x512, 1536x512 for Landscape!
Hires.fix: R-ESRGAN 4x+Anime6b, with 10 steps at 0.3 up to 0.5 denoising.
Clip Skip: 2.
Negatives: ' (worst quality, low quality:1.4), (zombie, sketch, interlocked fingers, comic) ' | {"language": ["en"], "license": "creativeml-openrail-m", "library_name": "diffusers", "tags": ["art", "anime", "stable diffusion"], "pipeline_tag": "text-to-image"} | text-to-image | roktimsardar123/MeinaMix_V11 | [
"diffusers",
"safetensors",
"art",
"anime",
"stable diffusion",
"text-to-image",
"en",
"license:creativeml-openrail-m",
"endpoints_compatible",
"diffusers:StableDiffusionPipeline",
"region:us"
] | 2024-02-07T16:35:05+00:00 | [] | [
"en"
] | TAGS
#diffusers #safetensors #art #anime #stable diffusion #text-to-image #en #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us
|
MeinaMix Objective is to be able to do good art with little prompting.
For examples and prompts, please checkout: URL
I have a discord server where you can post images that you generated, discuss prompt and/or ask for help.
URL If you like one of my models and want to support their updates
I've made a ko-fi page; URL where you can pay me a coffee <3
And a Patreon page; URL where you can support me and get acess to beta of my models!
You may also try this model using URL: URL
MeinaMix and the other of Meinas will ALWAYS be FREE.
Recommendations of use: Enable Quantization in K samplers.
URL is needed for prompts where the character is far away in order to make decent images, it drastically improve the quality of face and eyes!
Recommended parameters:
Sampler: Euler a: 40 to 60 steps.
Sampler: DPM++ SDE Karras: 20 to 30 steps.
Sampler: DPM++ 2M Karras: 20 to 40 steps.
CFG Scale: 7.
Resolutions: 512x768, 512x1024 for Portrait!
Resolutions: 768x512, 1024x512, 1536x512 for Landscape!
URL: R-ESRGAN 4x+Anime6b, with 10 steps at 0.3 up to 0.5 denoising.
Clip Skip: 2.
Negatives: ' (worst quality, low quality:1.4), (zombie, sketch, interlocked fingers, comic) ' | [] | [
"TAGS\n#diffusers #safetensors #art #anime #stable diffusion #text-to-image #en #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us \n"
] | [
66
] | [
"passage: TAGS\n#diffusers #safetensors #art #anime #stable diffusion #text-to-image #en #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us \n"
] | [
-0.02820507064461708,
-0.058283623307943344,
-0.006326538510620594,
-0.006846309173852205,
0.11472097039222717,
-0.00868819560855627,
0.21454261243343353,
0.023740345612168312,
-0.018035903573036194,
0.05042622610926628,
0.10414863377809525,
0.10362809151411057,
-0.018842702731490135,
0.1239745244383812,
-0.12414940446615219,
-0.283294141292572,
0.06358690559864044,
0.026866471394896507,
-0.026130767539143562,
0.08460679650306702,
0.10883961617946625,
-0.06357119232416153,
0.08958140760660172,
0.011383418925106525,
-0.09323050826787949,
-0.03266725316643715,
0.049566518515348434,
-0.12743155658245087,
0.08112233877182007,
0.02653445489704609,
0.08471253514289856,
0.10321523994207382,
0.025555722415447235,
-0.10654279589653015,
0.052243344485759735,
-0.0019154691835865378,
-0.06795305758714676,
0.030945448204874992,
0.059875283390283585,
-0.016013864427804947,
0.07179754972457886,
0.013182399794459343,
-0.011501157656311989,
0.062016889452934265,
-0.0870923176407814,
-0.05395730957388878,
0.03145739808678627,
0.012600375339388847,
0.1352943331003189,
0.013034838251769543,
0.018365545198321342,
0.00905368197709322,
-0.04753853753209114,
0.048652902245521545,
0.16454996168613434,
-0.3269755244255066,
-0.045287683606147766,
0.13772481679916382,
0.1763337254524231,
0.05367204546928406,
-0.1385420560836792,
0.11951465904712677,
0.02964889630675316,
-0.034651271998882294,
0.06590981781482697,
-0.06239113211631775,
0.1953936070203781,
-0.049309711903333664,
-0.04526752233505249,
0.10481920093297958,
0.20150163769721985,
0.016137028113007545,
0.015459693968296051,
-0.16644243896007538,
-0.05514901876449585,
0.11998439580202103,
-0.10450152307748795,
-0.000987434759736061,
0.009707792662084103,
0.030862171202898026,
-0.005982545670121908,
-0.03603753447532654,
-0.12718582153320312,
-0.023096228018403053,
-0.09000182896852493,
0.2155843824148178,
-0.02288622036576271,
0.04150072857737541,
-0.13542653620243073,
0.05934843420982361,
-0.10122828930616379,
-0.14320018887519836,
0.02739536575973034,
-0.13544927537441254,
0.09928041696548462,
0.06416179239749908,
0.006315093487501144,
-0.16482122242450714,
0.11217683553695679,
0.08581450581550598,
-0.035850994288921356,
0.012101784348487854,
0.008191988803446293,
0.15303249657154083,
0.040059421211481094,
-0.08875969797372818,
-0.0773395374417305,
0.0035663957241922617,
0.05596432462334633,
0.002699559321627021,
0.054195381700992584,
-0.0620984211564064,
-0.12645356357097626,
0.0020881600212305784,
-0.06674491614103317,
0.021568572148680687,
-0.00555651169270277,
0.05971710383892059,
-0.06850861012935638,
0.025301141664385796,
0.18277497589588165,
-0.0006971491384319961,
-0.013974103145301342,
-0.006471334025263786,
0.00261202291585505,
0.21208979189395905,
-0.008119647391140461,
0.016995079815387726,
0.08090165257453918,
0.21607205271720886,
-0.08635294437408447,
-0.023453859612345695,
0.006289010401815176,
-0.044185321778059006,
0.028844386339187622,
-0.14075517654418945,
0.05796967074275017,
-0.15296275913715363,
-0.09419441223144531,
0.0476517528295517,
0.03246799483895302,
-0.032041508704423904,
0.04078388214111328,
0.035015106201171875,
-0.09414485096931458,
0.08598925173282623,
0.010460025630891323,
-0.10175257176160812,
-0.05800669640302658,
0.09061025828123093,
-0.04425378888845444,
0.09610768407583237,
-0.11841090023517609,
0.002394696231931448,
-0.021007370203733444,
0.052953943610191345,
-0.17209342122077942,
-0.06617551296949387,
-0.05776246637105942,
0.11271350830793381,
0.022431960329413414,
-0.04625210165977478,
-0.04834744334220886,
0.025071144104003906,
-0.02249595895409584,
0.1979389637708664,
-0.098046213388443,
-0.03433656692504883,
0.16218620538711548,
-0.1399979442358017,
-0.13076376914978027,
0.08146616071462631,
0.009612279944121838,
0.016576047986745834,
-0.002730779116973281,
0.12814104557037354,
-0.10772868990898132,
-0.300600528717041,
0.08069222420454025,
0.06689001619815826,
-0.15254196524620056,
-0.028339972719550133,
0.07424405962228775,
0.07879152148962021,
0.07370205968618393,
0.04457157477736473,
-0.06887684017419815,
0.08720492571592331,
-0.08382967859506607,
0.014521469362080097,
0.0012835434172302485,
-0.03957662731409073,
0.058166489005088806,
0.017188778147101402,
0.08969943225383759,
-0.05725502967834473,
-0.05078854411840439,
0.031282637268304825,
-0.028104079887270927,
0.03468107059597969,
0.014359957538545132,
-0.06835239380598068,
0.18610435724258423,
-0.01133354939520359,
-0.0062066675163805485,
-0.06460081785917282,
-0.01802423596382141,
-0.015778690576553345,
0.2199852168560028,
-0.02659144438803196,
0.10782665759325027,
0.091596320271492,
0.02715029940009117,
-0.03636440634727478,
-0.018401168286800385,
0.05847003683447838,
0.036057207733392715,
0.021636787801980972,
-0.1754639893770218,
0.09399536997079849,
-0.11843959242105484,
-0.017371274530887604,
-0.09393323212862015,
0.021759912371635437,
0.09032539278268814,
0.14389857649803162,
0.05270470306277275,
0.029621737077832222,
0.012779796496033669,
-0.05217291787266731,
-0.10694984346628189,
0.004886162932962179,
0.09247860312461853,
0.0392053984105587,
-0.07344561070203781,
0.2429153174161911,
-0.1539742350578308,
0.3749220669269562,
0.16142360866069794,
-0.16853918135166168,
-0.07358089089393616,
-0.109664686024189,
-0.020192816853523254,
0.05443732440471649,
-0.021554604172706604,
-0.04397301748394966,
-0.09052130579948425,
-0.006736286450177431,
0.16477955877780914,
-0.06734474003314972,
0.01617255248129368,
0.06564770638942719,
-0.10316073894500732,
-0.0866166278719902,
0.02738744392991066,
0.05063702166080475,
-0.06465275585651398,
0.12539364397525787,
0.2140444964170456,
0.019163014367222786,
0.21421580016613007,
-0.008091333322227001,
-0.012544706463813782,
-0.014380528591573238,
0.07113803178071976,
0.03898117318749428,
0.14515264332294464,
-0.07583440840244293,
-0.01812269352376461,
0.03249369189143181,
-0.02390964888036251,
0.00225230329670012,
-0.11942113935947418,
-0.11533311754465103,
0.010184510610997677,
0.023660331964492798,
0.10014534741640091,
0.14324583113193512,
-0.05447402223944664,
0.15325191617012024,
-0.06514570862054825,
-0.08812233060598373,
0.03292315825819969,
-0.027745446190238,
-0.01974065974354744,
0.10413233935832977,
-0.10790741443634033,
-0.185425266623497,
-0.11274930089712143,
-0.10134655982255936,
-0.03587721660733223,
-0.0025958954356610775,
0.0758027657866478,
-0.06775093078613281,
-0.04944203048944473,
-0.057434625923633575,
-0.044506609439849854,
-0.07154087722301483,
-0.011949487961828709,
-0.006708352826535702,
0.065606489777565,
-0.08903741091489792,
-0.059557512402534485,
-0.01637915149331093,
-0.03315969929099083,
0.040957678109407425,
0.17889145016670227,
0.0027591241523623466,
0.14175358414649963,
0.11425875872373581,
-0.0013367502251639962,
0.004001591354608536,
-0.0033970538061112165,
0.13984210789203644,
-0.05535457655787468,
0.1270206719636917,
0.16058532893657684,
-0.002355957869440317,
0.0966029018163681,
0.12446972727775574,
0.0926327332854271,
-0.08875735104084015,
0.04771726205945015,
-0.080088309943676,
-0.10847443342208862,
-0.07684594392776489,
-0.05299539119005203,
-0.10725857317447662,
0.04670900106430054,
-0.0528884157538414,
0.04383443295955658,
0.06646887958049774,
0.08855974674224854,
0.054524198174476624,
-0.058213718235492706,
0.08521057665348053,
0.06988385319709778,
0.09371213614940643,
-0.04689433053135872,
0.07483857870101929,
-0.03581586480140686,
-0.0788566917181015,
0.09978710114955902,
-0.004101110156625509,
0.0666680708527565,
-0.012819302268326283,
-0.008647521957755089,
0.09679371863603592,
0.07480889558792114,
0.18815694749355316,
0.11213161051273346,
-0.015690838918089867,
-0.061862919479608536,
-0.024349909275770187,
-0.06747128814458847,
0.08900884538888931,
0.07533996552228928,
-0.026489704847335815,
-0.12983420491218567,
-0.01653718575835228,
-0.00941949337720871,
0.02564282715320587,
0.013237636536359787,
0.024824203923344612,
-0.142888143658638,
0.08150144666433334,
0.06753595173358917,
0.08604557812213898,
-0.05195193365216255,
0.06915167719125748,
0.2013007253408432,
-0.07440297305583954,
0.09348388016223907,
-0.05712272599339485,
0.0798632949590683,
0.12837685644626617,
-0.011249926872551441,
-0.007363179698586464,
-0.02453811652958393,
-0.01960393227636814,
-0.007802664302289486,
-0.17597244679927826,
0.1942506581544876,
-0.013460113666951656,
0.02620067074894905,
0.019808009266853333,
-0.03160301595926285,
0.06488662213087082,
0.18974661827087402,
0.22442111372947693,
-0.011978632770478725,
0.021644610911607742,
-0.04935308173298836,
-0.0600380003452301,
-0.029917512089014053,
0.14875413477420807,
0.04917881265282631,
-0.042179256677627563,
0.03251641243696213,
-0.04251578450202942,
0.017988473176956177,
0.05757947638630867,
-0.16465628147125244,
-0.14714349806308746,
0.036050938069820404,
0.08437100797891617,
0.0002911559131462127,
-0.049935489892959595,
-0.014863271266222,
-0.10755585879087448,
0.12300292402505875,
-0.07466596364974976,
-0.06729526072740555,
-0.13701435923576355,
-0.135432630777359,
-0.0067886472679674625,
-0.02233325131237507,
0.07730623334646225,
-0.11720035970211029,
0.06664668768644333,
-0.11803881824016571,
-0.11699385195970535,
0.09054198861122131,
-0.10439813137054443,
-0.10591629147529602,
-0.13335220515727997,
0.07388472557067871,
-0.040193647146224976,
-0.01348073035478592,
0.0106043117120862,
0.0026345031801611185,
0.0040216632187366486,
-0.11068378388881683,
0.046792566776275635,
0.015052994713187218,
0.014078031294047832,
-0.06728396564722061,
-0.13885298371315002,
-0.08681708574295044,
0.021191958338022232,
-0.026649238541722298,
0.08928153663873672,
0.3484558165073395,
-0.05576556921005249,
0.09770140051841736,
0.1870710551738739,
-0.01781589724123478,
-0.2476717233657837,
-0.08622656762599945,
-0.1448453962802887,
-0.023955941200256348,
0.021628843620419502,
-0.05581231787800789,
0.08549696207046509,
0.023831816390156746,
-0.0597965233027935,
0.1964922994375229,
-0.28889113664627075,
-0.11772166937589645,
0.01832554303109646,
0.10713851451873779,
0.3420526087284088,
-0.22825315594673157,
-0.04696284607052803,
-0.013759727589786053,
-0.2116715908050537,
0.08192048966884613,
0.015305798500776291,
0.056832119822502136,
-0.06090664118528366,
-0.0445222444832325,
-0.012532370164990425,
-0.08667603880167007,
0.13947530090808868,
-0.09380515664815903,
0.07240177690982819,
-0.11449038982391357,
-0.007138791959732771,
0.16634567081928253,
-0.019781123846769333,
0.020357375964522362,
-0.15701869130134583,
-0.002469881670549512,
-0.09568631649017334,
-0.01581808179616928,
-0.04291101545095444,
0.06958912312984467,
-0.02177705429494381,
-0.11310675740242004,
-0.05050261318683624,
0.044048208743333817,
-0.0304777342826128,
-0.00760832941159606,
0.139522984623909,
-0.022300884127616882,
0.11437598615884781,
0.14276346564292908,
0.01864476688206196,
-0.12619774043560028,
-0.028656942769885063,
-0.0765070766210556,
-0.03722181171178818,
0.08167730271816254,
-0.07224836200475693,
-0.02989819645881653,
0.1264374852180481,
0.03282175585627556,
0.09388689696788788,
0.07907622307538986,
-0.011709745973348618,
0.10644369572401047,
0.16091005504131317,
-0.16403056681156158,
-0.020595060661435127,
-0.021527975797653198,
0.18740089237689972,
0.1831270009279251,
0.028614817187190056,
0.09795915335416794,
-0.026666922494769096,
0.036407265812158585,
-0.02431231364607811,
0.0342552475631237,
-0.025306882336735725,
0.014191064052283764,
0.03461606800556183,
0.023752832785248756,
-0.04772140458226204,
0.03543448448181152,
-0.0696595087647438,
-0.19145508110523224,
-0.07687213271856308,
0.07326798886060715,
-0.11003707349300385,
-0.03452808037400246,
0.06223500892519951,
0.06566286832094193,
-0.1089978814125061,
-0.014967295341193676,
-0.04604736343026161,
-0.15779533982276917,
-0.011512409895658493,
0.17685472965240479,
0.03183295950293541,
0.01215519942343235,
0.03422841802239418,
-0.02514476329088211,
-0.02365834452211857,
0.011743689887225628,
0.04905753955245018,
0.0570218563079834,
-0.13854999840259552,
-0.10713621228933334,
0.010841481387615204,
0.022713901475071907,
-0.12786813080310822,
-0.03274788707494736,
-0.1538904458284378,
0.0033337040804326534,
-0.0580362044274807,
0.02720581740140915,
-0.13248470425605774,
-0.11277942359447479,
-0.05219123885035515,
-0.03189219906926155,
-0.045167289674282074,
-0.034130632877349854,
-0.05689875781536102,
0.024807224050164223,
0.010968714021146297,
-0.025298891589045525,
-0.09250132739543915,
-0.06315138936042786,
0.06408677995204926,
-0.0940525010228157,
0.07595968246459961,
0.029013514518737793,
-0.11167936772108078,
-0.04098730534315109,
-0.27663788199424744,
-0.057476144284009933,
0.1260214000940323,
-0.024641143158078194,
0.015837563201785088,
0.12419769912958145,
0.03672895208001137,
0.029365411028265953,
0.014125649817287922,
0.001150151714682579,
0.020122816786170006,
-0.10768842697143555,
0.04755343124270439,
-0.05024482682347298,
-0.05235891789197922,
-0.09369155019521713,
-0.05089555308222771,
0.1809169352054596,
-0.0022056519519537687,
0.12129867076873779,
-0.07545582205057144,
0.08327113837003708,
-0.018738627433776855,
0.017187148332595825,
0.0714467465877533,
-0.11841579526662827,
0.10743673890829086,
0.027149446308612823,
-0.04276224598288536,
-0.03546059876680374,
0.2378869205713272,
-0.00972604751586914,
-0.23676921427249908,
0.07679218053817749,
-0.018228737637400627,
-0.010922695510089397,
0.0428236685693264,
0.2155931442975998,
0.020899862051010132,
-0.008223720826208591,
-0.19607365131378174,
0.06590265035629272,
0.1021362841129303,
-0.11917341500520706,
0.08055084198713303,
0.1527959406375885,
-0.05910872668027878,
0.068827785551548,
0.02299199067056179,
0.05200960487127304,
-0.06181461364030838,
-0.008242378942668438,
-0.06999225914478302,
0.10399071872234344,
-0.0011607339838519692,
0.022208882495760918,
0.22634562849998474,
-0.010384191758930683,
0.002700550714507699,
-0.00027323001995682716,
-0.020436188206076622,
-0.09496096521615982,
-0.10465018451213837,
-0.045927900820970535,
-0.14395329356193542,
0.0695132464170456,
-0.027506023645401,
0.051147691905498505,
-0.03356548771262169,
0.10796993225812912,
0.0007889875560067594,
0.034548744559288025,
-0.05322596803307533,
-0.020562002435326576,
0.14612150192260742,
0.003977396059781313,
-0.06329815834760666,
0.0486043319106102,
0.007750121410936117,
-0.0751865953207016,
-0.07124892622232437,
-0.05385461077094078,
0.07112770527601242,
-0.05835118144750595,
0.03480793535709381,
-0.03281167149543762,
-0.07354390621185303,
-0.002189087448641658,
-0.0014843838289380074,
-0.02208641916513443,
0.14059840142726898,
0.005123488139361143,
0.008209768682718277,
-0.002713603898882866,
0.15987178683280945,
0.002720865188166499,
-0.016752000898122787,
-0.04033708572387695,
-0.07225028425455093,
-0.046501435339450836,
0.07154577970504761,
-0.05687171593308449,
-0.045385558158159256,
-0.02626967802643776,
0.3072011172771454,
0.19650593400001526,
-0.16214409470558167,
0.024596916511654854,
-0.06337564438581467,
0.0166885107755661,
0.03334623947739601,
0.06965331733226776,
0.028294378891587257,
0.35237380862236023,
-0.05401453748345375,
-0.07493606209754944,
-0.1262367218732834,
-0.026337232440710068,
-0.07854536175727844,
-0.0682651475071907,
0.07356158643960953,
-0.06104748696088791,
-0.11258260160684586,
0.09972461313009262,
-0.18545079231262207,
-0.003396396990865469,
0.056193504482507706,
-0.055307164788246155,
0.05919601395726204,
-0.03773713484406471,
0.1192188635468483,
-0.0012386019807308912,
0.052485715597867966,
-0.05601129308342934,
-0.0393960103392601,
-0.014223996549844742,
0.027480199933052063,
-0.07453663647174835,
0.09832841157913208,
-0.020830126479268074,
-0.10644389688968658,
0.09462806582450867,
-0.0165040772408247,
-0.06901651620864868,
0.04856652393937111,
0.017076551914215088,
-0.03289530798792839,
0.08881480991840363,
0.015855561941862106,
-0.06027144193649292,
-0.06564851105213165,
0.1389656811952591,
0.028971126303076744,
-0.036851782351732254,
0.012008200399577618,
-0.12049168348312378,
0.07936254143714905,
0.061296019703149796,
-0.10083691030740738,
-0.09989742189645767,
0.10611100494861603,
-0.07370901107788086,
0.0518813282251358,
-0.009520943276584148,
0.006222554016858339,
-0.055902957916259766,
0.0034284840803593397,
0.05074845999479294,
0.022343046963214874,
-0.12754754722118378,
0.010044286027550697,
-0.11339747160673141,
-0.04433836787939072,
0.05425040423870087,
0.020385922864079475,
-0.19518573582172394,
-0.028866833075881004,
-0.17353665828704834,
0.06527531147003174,
-0.12681053578853607,
0.005522825755178928,
0.204301118850708,
0.02914278581738472,
-0.005600803066045046,
-0.15863917768001556,
0.06946771591901779,
0.05609651282429695,
-0.028651108965277672,
-0.06897567212581635
] |
null | null | transformers | # MetricX-23
*This is not an officially supported Google product.*
**GitHub repository: [https://github.com/google-research/metricx](https://github.com/google-research/metricx)**
This repository contains the MetricX-23 models,
a family of models for automatic evaluation of translations that were proposed
in the WMT'23 Metrics Shared Task submission
[MetricX-23: The Google Submission to the WMT 2023 Metrics Shared Task](https://aclanthology.org/2023.wmt-1.63/).
The models were trained in [T5X](https://github.com/google-research/t5x) and
then converted for use in PyTorch.
## Available Models
There are 6 models available on HuggingFace that vary in the number of
parameters and whether or not the model is reference-based or reference-free
(also known as quality estimation, or QE):
* [MetricX-23-XXL](https://huggingface.co/google/metricx-23-large-v2p0)
* [MetricX-23-XL](https://huggingface.co/google/metricx-23-xl-v2p0)
* [MetricX-23-Large](https://huggingface.co/google/metricx-23-xxl-v2p0)
* [MetricX-23-QE-XXL](https://huggingface.co/google/metricx-23-qe-large-v2p0)
* [MetricX-23-QE-XL](https://huggingface.co/google/metricx-23-qe-xl-v2p0)
* [MetricX-23-QE-Large](https://huggingface.co/google/metricx-23-qe-xxl-v2p0)
We recommend using the XXL model versions for the best agreement with human
judgments of translation quality, the Large versions for best speed, and the
XL for an intermediate use case.
## Changes to the WMT'23 Submission
These models available here are most similar to the primary submission to the WMT'23 Metrics
Shared Task. They are initialized with [mT5](https://aclanthology.org/2021.naacl-main.41/)
then fine-tuned on a combination of direct assessment and MQM data. However,
we made some changes that make these models different from the WMT'23 submissions.
First, the models are trained to regress the actual MQM score rather than a
normalized score between 0 and 1. **That means the output from the MetricX-23
models is a score in the range [0, 25] where lower is better (i.e., it predicts
an error score).**
Second, these models were trained with a larger variety of synthetic data that
makes them more robust to translation edge cases like over- and undertranslation,
described in more detail in the following section.
### Synthetic Data
In order for our MetricX models to learn to identify certain types of bad
translations that are not sufficiently (or at all) represented in the regular
training data, we created synthetic examples and mixed them in during training.
The synthetic training data was generated from the DA datasets ranging from
WMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have
the candidate translation manipulated so as to turn it into a bad translation
with a specific issue commonly unrecognized by learned metrics.
The table below provides an overview of the various failure modes that we
considered, including brief descriptions of how we prepared the synthetic data
to address them.
| Failure mode | Synthetic example description |
| ----------- | ----------- |
| Undertranslation | Candidate translation with an arbitrary sentence removed (if multi-sentence); alternatively, candidate with a certain proportion of words removed from the end. |
| Overtranslation | Candidate translation duplicated (with space in between). |
| Fluent but unrelated translation | Arbitrary reference of a similar length from the dataset. |
| Gibberish | Text of a similar length as the reference, generated by sampling words from the reference translation vocabulary (built from all references in the data). |
| Missing punctuation | Reference translation with the end punctuation removed (11 punctuation symbols considered). |
| Latin instead of Chinese/Japanese or Hindi/Bengali punctuation | Candidate translation with the language-specific punctuation symbol at the end replaced with the Latin equivalent (e.g., "." instead of "。" or "।"); alternatively, the punctuation symbol is replaced with the Latin equivalent in the reference, keeping the correct one in the candidate. |
| Reference-matching translation | Reference translation copied as the candidate translation (unlike the rest of the synthetic data, these examples are meant to train the metric to predict a perfect score for candidates matching the reference). |
Examples from the first 4 categories were assigned a label corresponding to the
worst score on the given rating scale (e.g., 25 when mixed with MQM training
data), whereas the reference-matching translation examples are assigned the best
score (e.g., 0 when used with MQM data). The missing/incorrect punctuation
examples were labeled with a score slightly worse than perfect.
Note that some of the synthetic datasets are only meaningful in the
reference-based scenario, and we thus excluded them when training a QE variant
of MetricX. These are the Latin-vs-special punctuation and the
reference-matching translation examples.
Most of the synthetic training sets were created using stratified sampling
across target languages, taking 500 examples per target language. One exception
is the missing punctuation set, which used a stratified sample across different
punctuation symbols instead.
When training MetricX, a small proportion of the synthetic examples was mixed
with the regular training examples. During the first-stage fine-tuning on DA
data, each synthetic training set constituted between 0.1% and 1% of all
training examples, whereas in the second-stage fine-tuning on MQM data we used
an even smaller proportion, around 0.05%.
As for evaluating the effect of the synthetic training data on the model's
performance, the DEMETR challenge set - which we originally used to evaluate the
models submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We
therefore created a new DEMETR-style test set based on the WMT22 DA data, with
examples constructed analogically to the synthetic training examples, as
described above. This test set helped us determine the right proportions of
synthetic data for fine-tuning in order to make MetricX robust for the failure
modes in consideration, without sacrificing the system- and segment-level
correlations with human ratings.
## Usage
The code for using MetricX models can be found at [https://github.com/google-research/metricx](https://github.com/google-research/metricx).
The repository contains example prediction scripts, described below.
The `metricx23/predict.py` script contains an example for how to run inference
on the models.
### Reference-Based
Example usage for a reference-based model:
```bash
python -m metricx23.predict \
--tokenizer google/mt5-xl \
--model_name_or_path google/metricx-23-xl-v2p0 \
--max_input_length 1024 \
--batch_size 1 \
--input_file input.jsonl \
--output_file output.jsonl
```
`input.jsonl` is expected to have 1 serialized JSON object per line with
`"reference"` and `"hypothesis"` fields. The output jsonl will be parallel
to `input.jsonl` but additionally contain a `"prediction"` field with the predicted score.
Note that the model was trained with a maximum input length of 1024 tokens, so
significantly increasing that value may lead to unpredictable behavior.
### Reference-Free
Example usage for a reference-free model:
```bash
python -m metricx23.predict \
--tokenizer google/mt5-xl \
--model_name_or_path google/metricx-23-qe-xl-v2p0 \
--max_input_length 1024 \
--batch_size 1 \
--input_file input.jsonl \
--output_file output.jsonl \
--qe
```
`input.jsonl` is expected to have 1 serialized JSON object per line with
`"source"` and `"hypothesis"` fields. The output jsonl will be parallel
to `input.jsonl` but additionally contain a `"prediction"` field with the predicted score.
## Meta-Evaluation
The `metricx23/evaluate.py` script contains code to calculate various correlations
between the MetricX-23 scores and MQM ratings of translation quality using the
[MT Metrics Eval](https://github.com/google-research/mt-metrics-eval) library.
Example usage:
```bash
python -m metricx23.evaluate \
--dataset wmt22 \
--lp en-de \
--input_file input.jsonl \
--output_file output.json
```
`input.jsonl` is expected to have one JSON object serialized per line.
Each JSON object is expected to contain 4 fields:
* `"system_id"`: The name of the system that generated the translation.
* `"segment_id"`: The 0-based index of the corresponding segment in the MT
Metrics Eval data.
* `"label"`: The ground-truth translation quality score (with higher is better).
* `"prediction"`: The model predicted translation quality score (with lower is
better; the script negates the scores so higher is better).
The script will calculate the 4 agreement/correlations that were used in the
WMT'23 Shared Task. Below are the results for the MetricX-23 models on the
WMT'22 Metrics Shared Task data:
English-German:
| Model | System-Level Accuracy | System-Level Pearson | Segment-Level Pearson | Segment-Level Pairwise Acc |
| ----------- | ----------- | ----------- | ----------- | ----------- |
| MetricX-23-XXL | 0.795 | 0.835 | 0.546 | 0.619 |
| MetricX-23-XL | 0.756 | 0.813 | 0.540 | 0.605 |
| MetricX-23-Large | 0.769 | 0.759 | 0.507 | 0.595 |
| MetricX-23-QE-XXL | 0.769 | 0.830 | 0.490 | 0.606 |
| MetricX-23-QE-XL | 0.718 | 0.684 | 0.421 | 0.594 |
| MetricX-23-QE-Large | 0.744 | 0.671 | 0.387 | 0.579 |
English-Russian:
| Model | System-Level Accuracy | System-Level Pearson | Segment-Level Pearson | Segment-Level Pairwise Acc |
| ----------- | ----------- | ----------- | ----------- | ----------- |
| MetricX-23-XXL | 0.905 | 0.943 | 0.477 | 0.609 |
| MetricX-23-XL | 0.876 | 0.906 | 0.498 | 0.589 |
| MetricX-23-Large | 0.876 | 0.841 | 0.474 | 0.569 |
| MetricX-23-QE-XXL | 0.895 | 0.940 | 0.470 | 0.602 |
| MetricX-23-QE-XL | 0.848 | 0.861 | 0.415 | 0.570 |
| MetricX-23-QE-Large | 0.819 | 0.778 | 0.411 | 0.551 |
Chinese-English:
| Model | System-Level Accuracy | System-Level Pearson | Segment-Level Pearson | Segment-Level Pairwise Acc |
| ----------- | ----------- | ----------- | ----------- | ----------- |
| MetricX-23-XXL | 0.868 | 0.919 | 0.605 | 0.551 |
| MetricX-23-XL | 0.868 | 0.924 | 0.584 | 0.543 |
| MetricX-23-Large | 0.857 | 0.919 | 0.555 | 0.539 |
| MetricX-23-QE-XXL | 0.857 | 0.928 | 0.573 | 0.544 |
| MetricX-23-QE-XL | 0.802 | 0.879 | 0.546 | 0.529 |
| MetricX-23-QE-Large | 0.758 | 0.904 | 0.522 | 0.529 |
The `metricx23/evaluate_wmt23.py` script re-calculates the average correlation
score that was used to rank submissions from the
[WMT'23 Shared Task](https://www2.statmt.org/wmt23/pdf/2023.wmt-1.51.pdf).
Example usage:
```bash
python -m metricx23.evaluate_wmt23 \
--en_de predictions_ende.jsonl \
--he_en predictions_heen.jsonl \
--zh_en predictions_zhen.jsonl \
--output_file output.json
```
Each of the 3 input files is expected to be in the same format as described
above. Each file should correspond to running inference on each of the language
pairs from the WMT'23 dataset.
The results for each of the models is the following:
| Model | Average Correlation |
| ----------- | ----------- |
| MetricX-23-XXL | 0.812 |
| MetricX-23-XL | 0.813 |
| MetricX-23-Large | 0.794 |
| MetricX-23-QE-XXL | 0.797 |
| MetricX-23-QE-XL | 0.767 |
| MetricX-23-QE-Large | 0.762 |
## Citation
If you use MetricX-23 in your research, please cite the following publication:
```bibtex
@inproceedings{juraska-etal-2023-metricx,
title = {{MetricX-23: The Google Submission to the WMT 2023 Metrics Shared Task}},
author = "Juraska, Juraj and
Finkelstein, Mara and
Deutsch, Daniel and
Siddhant, Aditya and
Mirzazadeh, Mehdi and
Freitag, Markus",
editor = "Koehn, Philipp and
Haddow, Barry and
Kocmi, Tom and
Monz, Christof",
booktitle = "Proceedings of the Eighth Conference on Machine Translation",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.wmt-1.63",
doi = "10.18653/v1/2023.wmt-1.63",
pages = "756--767",
}
``` | {"license": "apache-2.0"} | null | google/metricx-23-qe-xl-v2p0 | [
"transformers",
"pytorch",
"mt5",
"license:apache-2.0",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T16:35:10+00:00 | [] | [] | TAGS
#transformers #pytorch #mt5 #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us
| MetricX-23
==========
*This is not an officially supported Google product.*
GitHub repository: URL
This repository contains the MetricX-23 models,
a family of models for automatic evaluation of translations that were proposed
in the WMT'23 Metrics Shared Task submission
MetricX-23: The Google Submission to the WMT 2023 Metrics Shared Task.
The models were trained in T5X and
then converted for use in PyTorch.
Available Models
----------------
There are 6 models available on HuggingFace that vary in the number of
parameters and whether or not the model is reference-based or reference-free
(also known as quality estimation, or QE):
* MetricX-23-XXL
* MetricX-23-XL
* MetricX-23-Large
* MetricX-23-QE-XXL
* MetricX-23-QE-XL
* MetricX-23-QE-Large
We recommend using the XXL model versions for the best agreement with human
judgments of translation quality, the Large versions for best speed, and the
XL for an intermediate use case.
Changes to the WMT'23 Submission
--------------------------------
These models available here are most similar to the primary submission to the WMT'23 Metrics
Shared Task. They are initialized with mT5
then fine-tuned on a combination of direct assessment and MQM data. However,
we made some changes that make these models different from the WMT'23 submissions.
First, the models are trained to regress the actual MQM score rather than a
normalized score between 0 and 1. That means the output from the MetricX-23
models is a score in the range [0, 25] where lower is better (i.e., it predicts
an error score).
Second, these models were trained with a larger variety of synthetic data that
makes them more robust to translation edge cases like over- and undertranslation,
described in more detail in the following section.
### Synthetic Data
In order for our MetricX models to learn to identify certain types of bad
translations that are not sufficiently (or at all) represented in the regular
training data, we created synthetic examples and mixed them in during training.
The synthetic training data was generated from the DA datasets ranging from
WMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have
the candidate translation manipulated so as to turn it into a bad translation
with a specific issue commonly unrecognized by learned metrics.
The table below provides an overview of the various failure modes that we
considered, including brief descriptions of how we prepared the synthetic data
to address them.
Examples from the first 4 categories were assigned a label corresponding to the
worst score on the given rating scale (e.g., 25 when mixed with MQM training
data), whereas the reference-matching translation examples are assigned the best
score (e.g., 0 when used with MQM data). The missing/incorrect punctuation
examples were labeled with a score slightly worse than perfect.
Note that some of the synthetic datasets are only meaningful in the
reference-based scenario, and we thus excluded them when training a QE variant
of MetricX. These are the Latin-vs-special punctuation and the
reference-matching translation examples.
Most of the synthetic training sets were created using stratified sampling
across target languages, taking 500 examples per target language. One exception
is the missing punctuation set, which used a stratified sample across different
punctuation symbols instead.
When training MetricX, a small proportion of the synthetic examples was mixed
with the regular training examples. During the first-stage fine-tuning on DA
data, each synthetic training set constituted between 0.1% and 1% of all
training examples, whereas in the second-stage fine-tuning on MQM data we used
an even smaller proportion, around 0.05%.
As for evaluating the effect of the synthetic training data on the model's
performance, the DEMETR challenge set - which we originally used to evaluate the
models submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We
therefore created a new DEMETR-style test set based on the WMT22 DA data, with
examples constructed analogically to the synthetic training examples, as
described above. This test set helped us determine the right proportions of
synthetic data for fine-tuning in order to make MetricX robust for the failure
modes in consideration, without sacrificing the system- and segment-level
correlations with human ratings.
Usage
-----
The code for using MetricX models can be found at URL
The repository contains example prediction scripts, described below.
The 'metricx23/URL' script contains an example for how to run inference
on the models.
### Reference-Based
Example usage for a reference-based model:
'URL' is expected to have 1 serialized JSON object per line with
'"reference"' and '"hypothesis"' fields. The output jsonl will be parallel
to 'URL' but additionally contain a '"prediction"' field with the predicted score.
Note that the model was trained with a maximum input length of 1024 tokens, so
significantly increasing that value may lead to unpredictable behavior.
### Reference-Free
Example usage for a reference-free model:
'URL' is expected to have 1 serialized JSON object per line with
'"source"' and '"hypothesis"' fields. The output jsonl will be parallel
to 'URL' but additionally contain a '"prediction"' field with the predicted score.
Meta-Evaluation
---------------
The 'metricx23/URL' script contains code to calculate various correlations
between the MetricX-23 scores and MQM ratings of translation quality using the
MT Metrics Eval library.
Example usage:
'URL' is expected to have one JSON object serialized per line.
Each JSON object is expected to contain 4 fields:
* '"system\_id"': The name of the system that generated the translation.
* '"segment\_id"': The 0-based index of the corresponding segment in the MT
Metrics Eval data.
* '"label"': The ground-truth translation quality score (with higher is better).
* '"prediction"': The model predicted translation quality score (with lower is
better; the script negates the scores so higher is better).
The script will calculate the 4 agreement/correlations that were used in the
WMT'23 Shared Task. Below are the results for the MetricX-23 models on the
WMT'22 Metrics Shared Task data:
English-German:
English-Russian:
Chinese-English:
The 'metricx23/evaluate\_wmt23.py' script re-calculates the average correlation
score that was used to rank submissions from the
WMT'23 Shared Task.
Example usage:
Each of the 3 input files is expected to be in the same format as described
above. Each file should correspond to running inference on each of the language
pairs from the WMT'23 dataset.
The results for each of the models is the following:
If you use MetricX-23 in your research, please cite the following publication:
| [
"### Synthetic Data\n\n\nIn order for our MetricX models to learn to identify certain types of bad\ntranslations that are not sufficiently (or at all) represented in the regular\ntraining data, we created synthetic examples and mixed them in during training.\nThe synthetic training data was generated from the DA datasets ranging from\nWMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have\nthe candidate translation manipulated so as to turn it into a bad translation\nwith a specific issue commonly unrecognized by learned metrics.\n\n\nThe table below provides an overview of the various failure modes that we\nconsidered, including brief descriptions of how we prepared the synthetic data\nto address them.\n\n\n\nExamples from the first 4 categories were assigned a label corresponding to the\nworst score on the given rating scale (e.g., 25 when mixed with MQM training\ndata), whereas the reference-matching translation examples are assigned the best\nscore (e.g., 0 when used with MQM data). The missing/incorrect punctuation\nexamples were labeled with a score slightly worse than perfect.\n\n\nNote that some of the synthetic datasets are only meaningful in the\nreference-based scenario, and we thus excluded them when training a QE variant\nof MetricX. These are the Latin-vs-special punctuation and the\nreference-matching translation examples.\n\n\nMost of the synthetic training sets were created using stratified sampling\nacross target languages, taking 500 examples per target language. One exception\nis the missing punctuation set, which used a stratified sample across different\npunctuation symbols instead.\n\n\nWhen training MetricX, a small proportion of the synthetic examples was mixed\nwith the regular training examples. During the first-stage fine-tuning on DA\ndata, each synthetic training set constituted between 0.1% and 1% of all\ntraining examples, whereas in the second-stage fine-tuning on MQM data we used\nan even smaller proportion, around 0.05%.\n\n\nAs for evaluating the effect of the synthetic training data on the model's\nperformance, the DEMETR challenge set - which we originally used to evaluate the\nmodels submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We\ntherefore created a new DEMETR-style test set based on the WMT22 DA data, with\nexamples constructed analogically to the synthetic training examples, as\ndescribed above. This test set helped us determine the right proportions of\nsynthetic data for fine-tuning in order to make MetricX robust for the failure\nmodes in consideration, without sacrificing the system- and segment-level\ncorrelations with human ratings.\n\n\nUsage\n-----\n\n\nThe code for using MetricX models can be found at URL\nThe repository contains example prediction scripts, described below.\n\n\nThe 'metricx23/URL' script contains an example for how to run inference\non the models.",
"### Reference-Based\n\n\nExample usage for a reference-based model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"reference\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nNote that the model was trained with a maximum input length of 1024 tokens, so\nsignificantly increasing that value may lead to unpredictable behavior.",
"### Reference-Free\n\n\nExample usage for a reference-free model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"source\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nMeta-Evaluation\n---------------\n\n\nThe 'metricx23/URL' script contains code to calculate various correlations\nbetween the MetricX-23 scores and MQM ratings of translation quality using the\nMT Metrics Eval library.\n\n\nExample usage:\n\n\n'URL' is expected to have one JSON object serialized per line.\nEach JSON object is expected to contain 4 fields:\n\n\n* '\"system\\_id\"': The name of the system that generated the translation.\n* '\"segment\\_id\"': The 0-based index of the corresponding segment in the MT\nMetrics Eval data.\n* '\"label\"': The ground-truth translation quality score (with higher is better).\n* '\"prediction\"': The model predicted translation quality score (with lower is\nbetter; the script negates the scores so higher is better).\n\n\nThe script will calculate the 4 agreement/correlations that were used in the\nWMT'23 Shared Task. Below are the results for the MetricX-23 models on the\nWMT'22 Metrics Shared Task data:\n\n\nEnglish-German:\n\n\n\nEnglish-Russian:\n\n\n\nChinese-English:\n\n\n\nThe 'metricx23/evaluate\\_wmt23.py' script re-calculates the average correlation\nscore that was used to rank submissions from the\nWMT'23 Shared Task.\n\n\nExample usage:\n\n\nEach of the 3 input files is expected to be in the same format as described\nabove. Each file should correspond to running inference on each of the language\npairs from the WMT'23 dataset.\n\n\nThe results for each of the models is the following:\n\n\n\nIf you use MetricX-23 in your research, please cite the following publication:"
] | [
"TAGS\n#transformers #pytorch #mt5 #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us \n",
"### Synthetic Data\n\n\nIn order for our MetricX models to learn to identify certain types of bad\ntranslations that are not sufficiently (or at all) represented in the regular\ntraining data, we created synthetic examples and mixed them in during training.\nThe synthetic training data was generated from the DA datasets ranging from\nWMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have\nthe candidate translation manipulated so as to turn it into a bad translation\nwith a specific issue commonly unrecognized by learned metrics.\n\n\nThe table below provides an overview of the various failure modes that we\nconsidered, including brief descriptions of how we prepared the synthetic data\nto address them.\n\n\n\nExamples from the first 4 categories were assigned a label corresponding to the\nworst score on the given rating scale (e.g., 25 when mixed with MQM training\ndata), whereas the reference-matching translation examples are assigned the best\nscore (e.g., 0 when used with MQM data). The missing/incorrect punctuation\nexamples were labeled with a score slightly worse than perfect.\n\n\nNote that some of the synthetic datasets are only meaningful in the\nreference-based scenario, and we thus excluded them when training a QE variant\nof MetricX. These are the Latin-vs-special punctuation and the\nreference-matching translation examples.\n\n\nMost of the synthetic training sets were created using stratified sampling\nacross target languages, taking 500 examples per target language. One exception\nis the missing punctuation set, which used a stratified sample across different\npunctuation symbols instead.\n\n\nWhen training MetricX, a small proportion of the synthetic examples was mixed\nwith the regular training examples. During the first-stage fine-tuning on DA\ndata, each synthetic training set constituted between 0.1% and 1% of all\ntraining examples, whereas in the second-stage fine-tuning on MQM data we used\nan even smaller proportion, around 0.05%.\n\n\nAs for evaluating the effect of the synthetic training data on the model's\nperformance, the DEMETR challenge set - which we originally used to evaluate the\nmodels submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We\ntherefore created a new DEMETR-style test set based on the WMT22 DA data, with\nexamples constructed analogically to the synthetic training examples, as\ndescribed above. This test set helped us determine the right proportions of\nsynthetic data for fine-tuning in order to make MetricX robust for the failure\nmodes in consideration, without sacrificing the system- and segment-level\ncorrelations with human ratings.\n\n\nUsage\n-----\n\n\nThe code for using MetricX models can be found at URL\nThe repository contains example prediction scripts, described below.\n\n\nThe 'metricx23/URL' script contains an example for how to run inference\non the models.",
"### Reference-Based\n\n\nExample usage for a reference-based model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"reference\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nNote that the model was trained with a maximum input length of 1024 tokens, so\nsignificantly increasing that value may lead to unpredictable behavior.",
"### Reference-Free\n\n\nExample usage for a reference-free model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"source\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nMeta-Evaluation\n---------------\n\n\nThe 'metricx23/URL' script contains code to calculate various correlations\nbetween the MetricX-23 scores and MQM ratings of translation quality using the\nMT Metrics Eval library.\n\n\nExample usage:\n\n\n'URL' is expected to have one JSON object serialized per line.\nEach JSON object is expected to contain 4 fields:\n\n\n* '\"system\\_id\"': The name of the system that generated the translation.\n* '\"segment\\_id\"': The 0-based index of the corresponding segment in the MT\nMetrics Eval data.\n* '\"label\"': The ground-truth translation quality score (with higher is better).\n* '\"prediction\"': The model predicted translation quality score (with lower is\nbetter; the script negates the scores so higher is better).\n\n\nThe script will calculate the 4 agreement/correlations that were used in the\nWMT'23 Shared Task. Below are the results for the MetricX-23 models on the\nWMT'22 Metrics Shared Task data:\n\n\nEnglish-German:\n\n\n\nEnglish-Russian:\n\n\n\nChinese-English:\n\n\n\nThe 'metricx23/evaluate\\_wmt23.py' script re-calculates the average correlation\nscore that was used to rank submissions from the\nWMT'23 Shared Task.\n\n\nExample usage:\n\n\nEach of the 3 input files is expected to be in the same format as described\nabove. Each file should correspond to running inference on each of the language\npairs from the WMT'23 dataset.\n\n\nThe results for each of the models is the following:\n\n\n\nIf you use MetricX-23 in your research, please cite the following publication:"
] | [
42,
666,
111,
457
] | [
"passage: TAGS\n#transformers #pytorch #mt5 #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us \n",
"passage: ### Synthetic Data\n\n\nIn order for our MetricX models to learn to identify certain types of bad\ntranslations that are not sufficiently (or at all) represented in the regular\ntraining data, we created synthetic examples and mixed them in during training.\nThe synthetic training data was generated from the DA datasets ranging from\nWMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have\nthe candidate translation manipulated so as to turn it into a bad translation\nwith a specific issue commonly unrecognized by learned metrics.\n\n\nThe table below provides an overview of the various failure modes that we\nconsidered, including brief descriptions of how we prepared the synthetic data\nto address them.\n\n\n\nExamples from the first 4 categories were assigned a label corresponding to the\nworst score on the given rating scale (e.g., 25 when mixed with MQM training\ndata), whereas the reference-matching translation examples are assigned the best\nscore (e.g., 0 when used with MQM data). The missing/incorrect punctuation\nexamples were labeled with a score slightly worse than perfect.\n\n\nNote that some of the synthetic datasets are only meaningful in the\nreference-based scenario, and we thus excluded them when training a QE variant\nof MetricX. These are the Latin-vs-special punctuation and the\nreference-matching translation examples.\n\n\nMost of the synthetic training sets were created using stratified sampling\nacross target languages, taking 500 examples per target language. One exception\nis the missing punctuation set, which used a stratified sample across different\npunctuation symbols instead.\n\n\nWhen training MetricX, a small proportion of the synthetic examples was mixed\nwith the regular training examples. During the first-stage fine-tuning on DA\ndata, each synthetic training set constituted between 0.1% and 1% of all\ntraining examples, whereas in the second-stage fine-tuning on MQM data we used\nan even smaller proportion, around 0.05%.\n\n\nAs for evaluating the effect of the synthetic training data on the model's\nperformance, the DEMETR challenge set - which we originally used to evaluate the\nmodels submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We\ntherefore created a new DEMETR-style test set based on the WMT22 DA data, with\nexamples constructed analogically to the synthetic training examples, as\ndescribed above. This test set helped us determine the right proportions of\nsynthetic data for fine-tuning in order to make MetricX robust for the failure\nmodes in consideration, without sacrificing the system- and segment-level\ncorrelations with human ratings.\n\n\nUsage\n-----\n\n\nThe code for using MetricX models can be found at URL\nThe repository contains example prediction scripts, described below.\n\n\nThe 'metricx23/URL' script contains an example for how to run inference\non the models.### Reference-Based\n\n\nExample usage for a reference-based model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"reference\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nNote that the model was trained with a maximum input length of 1024 tokens, so\nsignificantly increasing that value may lead to unpredictable behavior."
] | [
-0.08049466460943222,
-0.010928971692919731,
-0.004997325129806995,
0.02674533985555172,
0.10898007452487946,
0.017911424860358238,
0.030734051018953323,
0.10377532243728638,
-0.02402583509683609,
-0.03567541018128395,
0.03877764940261841,
0.08288230746984482,
0.015161662362515926,
0.022808298468589783,
0.01772439479827881,
-0.1944202184677124,
0.05506649613380432,
0.008813947439193726,
-0.07605475932359695,
0.06908859312534332,
0.11011916399002075,
-0.04383927211165428,
0.047993674874305725,
0.01802058517932892,
-0.046737369149923325,
0.03515304625034332,
0.02192578837275505,
-0.05999843776226044,
0.09158408641815186,
0.08443905413150787,
0.04427182674407959,
0.026185670867562294,
-0.03256708011031151,
-0.2039792835712433,
0.018693745136260986,
0.059222154319286346,
-0.04639671370387077,
0.032644130289554596,
0.06700465828180313,
-0.04408042132854462,
0.10856176912784576,
-0.03563929721713066,
-0.038342323154211044,
0.052534811198711395,
-0.08053379505872726,
-0.11484964191913605,
-0.06861758977174759,
0.005672447383403778,
0.05251080170273781,
0.11462049186229706,
0.016137007623910904,
0.08802412450313568,
-0.08221147954463959,
0.039690323173999786,
0.1186930239200592,
-0.24400970339775085,
0.022698264569044113,
0.031076578423380852,
0.0391852930188179,
0.059262461960315704,
-0.02663654275238514,
-0.011056511662900448,
0.04030945897102356,
0.036150794476270676,
0.04718852788209915,
-0.02858263999223709,
0.022121166810393333,
0.03753712773323059,
-0.11128073930740356,
-0.06610941141843796,
0.2194429337978363,
0.009360874071717262,
-0.039162129163742065,
-0.012831603176891804,
-0.040371835231781006,
-0.021052006632089615,
0.02562984824180603,
0.011642597615718842,
0.0126093290746212,
0.06509723514318466,
0.08087117969989777,
-0.06699725240468979,
-0.1469363272190094,
-0.02156376652419567,
-0.13039535284042358,
0.056314483284950256,
0.015906725078821182,
0.0635187178850174,
-0.13009536266326904,
0.06865113973617554,
-0.11303252726793289,
-0.037289686501026154,
-0.044206660240888596,
-0.09330593794584274,
0.027131132781505585,
0.0172458216547966,
-0.0671110600233078,
-0.061645977199077606,
0.01954670250415802,
0.09166590869426727,
-0.03881584480404854,
0.02113628201186657,
0.06317485123872757,
0.08239984512329102,
0.02081705816090107,
0.08037593215703964,
0.01793535053730011,
-0.011132435873150826,
0.08186338096857071,
-0.06666721403598785,
0.03507205471396446,
-0.00009720586240291595,
-0.05460785701870918,
-0.04377254843711853,
-0.02165292575955391,
0.07720592617988586,
0.024521537125110626,
0.024704361334443092,
-0.022436054423451424,
-0.014114882797002792,
0.15771594643592834,
-0.08077961206436157,
-0.031236806884407997,
-0.011595608666539192,
-0.05022282153367996,
0.047523126006126404,
0.06264422833919525,
0.006788937374949455,
-0.11120638251304626,
0.05265801399946213,
-0.03781677410006523,
-0.03154580295085907,
-0.08177147805690765,
-0.060364820063114166,
0.05910094082355499,
-0.0520191490650177,
0.022135227918624878,
-0.12485349923372269,
-0.14817503094673157,
-0.0194524135440588,
0.00905279628932476,
0.017627842724323273,
-0.04404270648956299,
0.030841557309031487,
-0.021325333043932915,
-0.010873598977923393,
-0.03838267922401428,
0.04409907013177872,
-0.0735226571559906,
0.01654006540775299,
-0.05891093611717224,
0.006638679653406143,
-0.12875565886497498,
0.05859975516796112,
-0.07795718312263489,
-0.008538316003978252,
-0.12053868919610977,
0.06858694553375244,
-0.06013733893632889,
0.06287765502929688,
-0.06418386101722717,
-0.059423450380563736,
-0.05324282497167587,
0.03673676773905754,
-0.011127032339572906,
0.08703723549842834,
-0.11388625204563141,
-0.04371771216392517,
0.06918473541736603,
-0.1147933155298233,
-0.10631323605775833,
0.05931851267814636,
-0.015358645468950272,
0.03748219460248947,
0.07437919080257416,
0.06889817118644714,
0.13736900687217712,
-0.022680290043354034,
-0.030354108661413193,
0.0774422138929367,
-0.09487991780042648,
-0.0831356942653656,
0.043639007955789566,
-0.02724086493253708,
-0.027338385581970215,
0.010795310139656067,
-0.05432993173599243,
0.053616154938936234,
0.004504588432610035,
-0.041023075580596924,
-0.03911237046122551,
-0.03953332453966141,
-0.03608177229762077,
-0.007416178472340107,
0.04547516256570816,
-0.030613455921411514,
-0.04433886334300041,
0.014297567307949066,
0.05688025429844856,
0.0016403227346017957,
0.06881103664636612,
-0.06005219742655754,
0.05407850816845894,
-0.018144575878977776,
0.008445029146969318,
-0.16996923089027405,
-0.050326526165008545,
0.03246239572763443,
-0.04993841052055359,
0.021022705361247063,
0.02538926526904106,
0.02754078060388565,
-0.014172155410051346,
-0.03300439566373825,
-0.01649932749569416,
0.008937809616327286,
0.008251310326159,
-0.05732375383377075,
-0.12281204760074615,
-0.0057578012347221375,
-0.030820120126008987,
-0.01729525998234749,
-0.027541520074009895,
0.012479400262236595,
0.07977741956710815,
0.044011034071445465,
0.009504541754722595,
0.03403853625059128,
0.005829425062984228,
0.01828007400035858,
-0.025431593880057335,
-0.01253400556743145,
0.06733429431915283,
0.029115518555045128,
-0.08058406412601471,
0.13161158561706543,
-0.1250324547290802,
0.017595291137695312,
0.13202296197414398,
-0.10783318430185318,
0.04871136322617531,
-0.04314408451318741,
-0.031379710882902145,
-0.0057036676444113255,
-0.03291696310043335,
-0.022622399032115936,
0.12076297402381897,
-0.004013901576399803,
0.0929318368434906,
-0.07880045473575592,
-0.05966495722532272,
-0.0097917215898633,
0.0038842959329485893,
-0.010753002017736435,
0.0713651031255722,
0.12100948393344879,
-0.12231792509555817,
0.08771493285894394,
0.13989409804344177,
-0.003570202738046646,
0.12119580805301666,
-0.040826037526130676,
-0.07305974513292313,
0.03446398302912712,
0.004951390903443098,
-0.01554139144718647,
0.04345749691128731,
-0.12445223331451416,
0.015536606311798096,
0.05309225618839264,
0.08432695269584656,
0.11454753577709198,
-0.11917590349912643,
0.0067930202931165695,
0.015571664087474346,
-0.03209373727440834,
-0.09838879108428955,
-0.0021094102412462234,
-0.01198030449450016,
0.07585512846708298,
-0.010641621425747871,
0.034573666751384735,
0.06810279190540314,
-0.033153194934129715,
-0.1333116739988327,
0.1549171507358551,
-0.16806340217590332,
-0.2015051692724228,
-0.1637459099292755,
0.0050877295434474945,
-0.04799898713827133,
0.007370667066425085,
0.10294272005558014,
-0.07140845060348511,
-0.04824806749820709,
-0.030684374272823334,
0.044100407510995865,
-0.026180272921919823,
-0.004666412249207497,
-0.05338970944285393,
0.042327139526605606,
0.016776274889707565,
-0.12135301530361176,
-0.006609104573726654,
-0.0036472089122980833,
-0.017405901104211807,
0.041763320565223694,
-0.07078999280929565,
0.07727597653865814,
0.08807438611984253,
0.01986617222428322,
-0.01348377950489521,
-0.04507536441087723,
0.21933329105377197,
-0.05761788785457611,
0.04443945735692978,
0.2112521529197693,
-0.011195655912160873,
0.05879829078912735,
0.052676692605018616,
-0.004360849969089031,
-0.02312907949090004,
0.02751847915351391,
0.013980694115161896,
-0.0917525514960289,
-0.284024178981781,
-0.06044050678610802,
-0.09147080034017563,
-0.06417541950941086,
0.0023470930755138397,
0.06487169861793518,
0.03722584992647171,
0.050735119730234146,
-0.04127505049109459,
-0.014369443990290165,
0.041575632989406586,
0.05546524375677109,
0.09428055584430695,
0.013206657022237778,
0.06720979511737823,
-0.11070599406957626,
0.012433091178536415,
0.11600261181592941,
0.09805083274841309,
0.23822110891342163,
-0.026172611862421036,
0.04857305809855461,
0.11312685161828995,
0.11436714231967926,
0.08747904002666473,
0.10965031385421753,
-0.04313330352306366,
0.029491525143384933,
-0.041477788239717484,
-0.02848794311285019,
-0.03056623786687851,
0.08165007829666138,
-0.021050430834293365,
-0.0669928565621376,
-0.06534460186958313,
-0.021561436355113983,
0.06433352082967758,
0.21831148862838745,
-0.005303980782628059,
-0.07421030104160309,
-0.05741162598133087,
0.06067470833659172,
-0.010603219270706177,
-0.03260648995637894,
0.07761399447917938,
0.04806586354970932,
-0.07070834934711456,
0.08121111243963242,
-0.0008651819080114365,
0.11229313164949417,
0.014927586540579796,
0.018110742792487144,
0.03192968666553497,
-0.006562814116477966,
0.018823236227035522,
0.11194905638694763,
-0.26115158200263977,
0.12093576788902283,
-0.011933675035834312,
0.05421082302927971,
-0.0762992650270462,
0.022125624120235443,
0.04246993362903595,
0.1273622065782547,
0.09262282401323318,
0.043345239013433456,
-0.14447379112243652,
-0.0103498175740242,
0.0019879117608070374,
0.056262798607349396,
0.04705921933054924,
0.010509418323636055,
0.011808192357420921,
-0.011274687014520168,
0.013288512825965881,
0.015993541106581688,
0.08853989839553833,
-0.0517207607626915,
-0.16778728365898132,
0.04126080125570297,
-0.006679844111204147,
0.015364071354269981,
-0.07193623483181,
0.005432065576314926,
-0.039848387241363525,
0.08026738464832306,
-0.14921778440475464,
-0.0995694100856781,
-0.09588959813117981,
-0.08473988622426987,
0.08787038922309875,
-0.0654374435544014,
0.06406596302986145,
0.007063437253236771,
0.07009734213352203,
-0.03068877011537552,
-0.14400194585323334,
0.057342350482940674,
-0.11315649747848511,
-0.08588340878486633,
-0.0034783631563186646,
0.1125456690788269,
0.013006072491407394,
0.0326993428170681,
0.019293401390314102,
-0.010193015448749065,
-0.046051137149333954,
-0.11050929129123688,
0.01819794625043869,
0.07997186481952667,
0.05762843042612076,
0.05681586265563965,
-0.09417805820703506,
-0.029595911502838135,
-0.04396835342049599,
-0.06783974170684814,
0.1667174994945526,
0.10756810754537582,
-0.06620179116725922,
0.1170005351305008,
0.16132885217666626,
-0.1295406073331833,
-0.19067947566509247,
-0.043041959404945374,
-0.03716177120804787,
-0.01986772194504738,
-0.02807173877954483,
-0.1475318968296051,
0.07539469003677368,
0.09812235087156296,
-0.0024198442697525024,
0.04476727545261383,
-0.2080758810043335,
-0.08643566817045212,
0.04781382530927658,
0.0001132916659116745,
0.18663114309310913,
-0.10906235128641129,
-0.044622793793678284,
-0.052799079567193985,
-0.11352433264255524,
0.14947330951690674,
-0.06072360277175903,
0.05728284642100334,
0.00732648279517889,
0.0032600341364741325,
-0.006010896060615778,
-0.03202845901250839,
0.13251455128192902,
0.03674418851733208,
0.043941304087638855,
-0.028950639069080353,
0.055670365691185,
0.025221113115549088,
-0.04210628941655159,
0.044331781566143036,
-0.04143795743584633,
0.021573659032583237,
-0.14452756941318512,
-0.045783452689647675,
-0.08391499519348145,
0.04536336287856102,
-0.01962660811841488,
-0.039184920489788055,
-0.042702145874500275,
0.0021056830883026123,
0.08968086540699005,
-0.014609623700380325,
0.054970160126686096,
-0.05837482213973999,
0.059537388384342194,
0.07670764625072479,
0.15645602345466614,
-0.10332942008972168,
-0.1217908263206482,
-0.02813919633626938,
-0.021626751869916916,
0.08217503130435944,
-0.1376548409461975,
0.07488036155700684,
0.11647255718708038,
-0.04862956330180168,
0.08627215772867203,
0.055094070732593536,
-0.07522906363010406,
0.005458813160657883,
0.10446281731128693,
-0.07246147096157074,
-0.06008515506982803,
-0.0390506386756897,
0.08281482756137848,
0.029428621754050255,
0.03669393062591553,
0.12870027124881744,
-0.0009662613738328218,
-0.032006535679101944,
-0.02065277099609375,
0.06402802467346191,
-0.0347125381231308,
0.0842379555106163,
0.039138346910476685,
0.017695903778076172,
-0.08644309639930725,
0.1338755190372467,
0.01591717265546322,
-0.1885780692100525,
-0.009080829098820686,
0.11049890518188477,
-0.11632084846496582,
-0.0807068943977356,
-0.025578761473298073,
0.009648185223340988,
-0.14496804773807526,
-0.10269727557897568,
-0.032402172684669495,
-0.14044955372810364,
0.06719717383384705,
0.16770902276039124,
0.02116461656987667,
0.03302031010389328,
-0.006035410333424807,
-0.028474289923906326,
-0.03122306987643242,
-0.013931620866060257,
-0.0277186818420887,
0.017339259386062622,
-0.05742930248379707,
0.0389898419380188,
0.019052214920520782,
0.06222548335790634,
-0.016825642436742783,
0.015219002030789852,
-0.0635049119591713,
0.010224338620901108,
-0.15755128860473633,
0.024624407291412354,
-0.05316181108355522,
-0.017552291974425316,
-0.01309568714350462,
-0.014190688729286194,
-0.05401631444692612,
0.05547712743282318,
-0.09220357239246368,
-0.020396772772073746,
-0.03640173375606537,
0.027906954288482666,
-0.05868091061711311,
-0.02378096804022789,
0.050178419798612595,
-0.047896455973386765,
0.1031966358423233,
0.06362350285053253,
-0.05172618851065636,
0.0687813088297844,
-0.054818395525217056,
-0.03117077425122261,
0.04752342030405998,
0.04614630341529846,
0.0024338315706700087,
-0.06958136707544327,
0.053022030740976334,
0.0532219260931015,
-0.05545207858085632,
-0.00179347675293684,
-0.00773459579795599,
-0.106658935546875,
-0.026197271421551704,
0.026790166273713112,
-0.007194001227617264,
-0.06978121399879456,
-0.046734243631362915,
0.11590144783258438,
0.04049905017018318,
0.1221245527267456,
-0.006192055530846119,
0.039656929671764374,
-0.0669143944978714,
0.0076656281016767025,
-0.0022521058563143015,
-0.0823831707239151,
-0.09226924180984497,
-0.03779274597764015,
0.01616818644106388,
-0.032287146896123886,
0.20258086919784546,
-0.027631420642137527,
-0.001419171690940857,
0.05651244521141052,
0.0006612688302993774,
-0.019393347203731537,
0.012330395169556141,
0.1308664083480835,
0.009836757555603981,
0.00910251121968031,
-0.02165895141661167,
0.024475062265992165,
-0.016778085380792618,
-0.0041928887367248535,
0.13951833546161652,
0.09686414152383804,
0.16705356538295746,
0.1026294082403183,
0.002041283994913101,
-0.03730851784348488,
-0.01945573464035988,
-0.06291651725769043,
0.028470784425735474,
0.058277349919080734,
0.00701283710077405,
0.01671595126390457,
0.11799974739551544,
-0.09208066761493683,
0.09887351095676422,
-0.06913766264915466,
-0.05150070786476135,
-0.1154223382472992,
-0.09975318610668182,
-0.010319891385734081,
-0.08603835850954056,
-0.029209984466433525,
-0.13546273112297058,
-0.020355189219117165,
0.06884671002626419,
0.038902945816516876,
-0.05183970928192139,
0.007247757166624069,
-0.11964748054742813,
-0.10397026687860489,
0.04866647720336914,
-0.014942724257707596,
0.04801199212670326,
0.04992463439702988,
0.012288471683859825,
0.04020651429891586,
-0.03871965408325195,
-0.02611941285431385,
0.039849359542131424,
0.04274921864271164,
0.045678697526454926,
-0.03452969342470169,
-0.010724259540438652,
-0.028795840218663216,
0.012287003919482231,
0.02913379855453968,
0.22296985983848572,
0.010857714340090752,
-0.031127674505114555,
0.06575135886669159,
0.19841772317886353,
-0.04081185162067413,
-0.09934508055448532,
-0.09630105644464493,
0.1641465723514557,
0.045690394937992096,
0.029041193425655365,
0.016438156366348267,
-0.057314880192279816,
-0.00565880723297596,
0.22985684871673584,
0.17313337326049805,
-0.025465190410614014,
-0.009001774713397026,
-0.006234009750187397,
0.014965811744332314,
0.06765225529670715,
0.08408764004707336,
0.1142108142375946,
0.21731261909008026,
-0.06078370288014412,
0.0639389306306839,
-0.0446958988904953,
0.0698341652750969,
-0.034699127078056335,
0.17769397795200348,
0.001534847542643547,
-0.058601561933755875,
-0.0005737040191888809,
0.06745973229408264,
-0.07404687255620956,
-0.10022060573101044,
-0.03632500767707825,
-0.07546216249465942,
-0.05532532557845116,
-0.0036639878526329994,
0.05928308516740799,
0.022997058928012848,
0.041565168648958206,
-0.0059436894953250885,
-0.044010479003190994,
0.10706667602062225,
0.023562224581837654,
-0.17537671327590942,
-0.05920793116092682,
0.08894800394773483,
0.011057954281568527,
0.08248458802700043,
0.007951926440000534,
0.048754554241895676,
0.07727794349193573,
0.03654977306723595,
-0.11864081025123596,
0.07930269092321396,
-0.00997848343104124,
-0.05712908133864403,
0.007062149234116077,
0.014758620411157608,
-0.00663449801504612,
0.020782694220542908,
0.029319513589143753,
-0.03974604234099388,
0.05011465772986412,
0.03723395615816116,
-0.04944480583071709,
-0.00523018091917038,
0.03254415839910507,
-0.10453616082668304,
0.08378694206476212,
0.07704398781061172,
-0.022326018661260605,
-0.027006199583411217,
-0.05415914207696915,
0.02430926077067852,
0.012829716317355633,
-0.023813046514987946,
-0.024040259420871735,
-0.06499078124761581,
0.017759330570697784,
0.01358121633529663,
0.03267041966319084,
-0.13761897385120392,
-0.03263425827026367,
0.009745022282004356,
-0.028417814522981644,
-0.08249472081661224,
0.06735780090093613,
0.043854910880327225,
0.026215003803372383,
-0.03867888078093529,
0.0033451877534389496,
-0.004438048228621483,
0.04008375108242035,
-0.0645875409245491,
-0.08099328726530075
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# videomae-base-finetuned-kinetics-finetuned-ucf101-subset
This model is a fine-tuned version of [MCG-NJU/videomae-base-finetuned-kinetics](https://huggingface.co/MCG-NJU/videomae-base-finetuned-kinetics) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 1.2309
- Accuracy: 0.9806
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- training_steps: 148
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 2.2587 | 0.13 | 19 | 1.2644 | 1.0 |
| 0.6711 | 1.13 | 38 | 0.2098 | 1.0 |
| 0.1355 | 2.13 | 57 | 0.0465 | 1.0 |
| 0.0295 | 3.13 | 76 | 0.0431 | 0.9857 |
| 0.0155 | 4.13 | 95 | 0.0226 | 1.0 |
| 0.0175 | 5.13 | 114 | 0.0178 | 1.0 |
| 0.0168 | 6.13 | 133 | 0.0180 | 1.0 |
| 0.008 | 7.1 | 148 | 0.0184 | 1.0 |
### Framework versions
- Transformers 4.38.0.dev0
- Pytorch 2.2.0+cu121
- Datasets 2.11.0
- Tokenizers 0.15.1
| {"license": "cc-by-nc-4.0", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "MCG-NJU/videomae-base-finetuned-kinetics", "model-index": [{"name": "videomae-base-finetuned-kinetics-finetuned-ucf101-subset", "results": []}]} | video-classification | Saini-Manisha/videomae-base-finetuned-kinetics-finetuned-ucf101-subset | [
"transformers",
"tensorboard",
"safetensors",
"videomae",
"video-classification",
"generated_from_trainer",
"base_model:MCG-NJU/videomae-base-finetuned-kinetics",
"license:cc-by-nc-4.0",
"endpoints_compatible",
"region:us"
] | 2024-02-07T16:35:16+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #videomae #video-classification #generated_from_trainer #base_model-MCG-NJU/videomae-base-finetuned-kinetics #license-cc-by-nc-4.0 #endpoints_compatible #region-us
| videomae-base-finetuned-kinetics-finetuned-ucf101-subset
========================================================
This model is a fine-tuned version of MCG-NJU/videomae-base-finetuned-kinetics on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 1.2309
* Accuracy: 0.9806
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* training\_steps: 148
### Training results
### Framework versions
* Transformers 4.38.0.dev0
* Pytorch 2.2.0+cu121
* Datasets 2.11.0
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 148",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.2.0+cu121\n* Datasets 2.11.0\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #videomae #video-classification #generated_from_trainer #base_model-MCG-NJU/videomae-base-finetuned-kinetics #license-cc-by-nc-4.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 148",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.2.0+cu121\n* Datasets 2.11.0\n* Tokenizers 0.15.1"
] | [
77,
115,
4,
38
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #videomae #video-classification #generated_from_trainer #base_model-MCG-NJU/videomae-base-finetuned-kinetics #license-cc-by-nc-4.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 148### Training results### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.2.0+cu121\n* Datasets 2.11.0\n* Tokenizers 0.15.1"
] | [
-0.14067408442497253,
0.13049742579460144,
-0.0006684681284241378,
0.0881192609667778,
0.11772530525922775,
0.018097219988703728,
0.1282297670841217,
0.14735157787799835,
-0.0849854126572609,
0.06018507108092308,
0.13015678524971008,
0.08501256257295609,
0.04157175123691559,
0.1958305388689041,
-0.03485799580812454,
-0.23080329596996307,
0.020348718389868736,
0.02824052795767784,
-0.06096375361084938,
0.12226862460374832,
0.09888572245836258,
-0.14025577902793884,
0.07830062508583069,
-0.018448606133461,
-0.20737385749816895,
-0.034432411193847656,
0.020543336868286133,
-0.0539497509598732,
0.12002922594547272,
0.014983358792960644,
0.09057852625846863,
0.042567938566207886,
0.12002113461494446,
-0.19211335480213165,
0.015917837619781494,
0.05376332253217697,
-0.0022722200956195593,
0.07303100824356079,
0.06342996656894684,
0.039087675511837006,
0.05234026908874512,
-0.1315518468618393,
0.055923715233802795,
0.019314872100949287,
-0.12532898783683777,
-0.2138097733259201,
-0.09037192910909653,
0.008923767134547234,
0.0959915742278099,
0.0641360804438591,
-0.01113113109022379,
0.0899103432893753,
-0.03747027739882469,
0.09340058267116547,
0.18814517557621002,
-0.2480284422636032,
-0.07839410752058029,
0.051359280943870544,
0.07061603665351868,
0.08502347767353058,
-0.12959888577461243,
0.008707619272172451,
0.05467330291867256,
0.008930491283535957,
0.1227605789899826,
-0.012343408539891243,
-0.02558991312980652,
-0.03260665759444237,
-0.13042689859867096,
-0.07544852048158646,
0.1178516373038292,
0.08644121140241623,
-0.039458248764276505,
-0.06324458867311478,
-0.03159990906715393,
-0.1707826852798462,
-0.08233102411031723,
0.0435677245259285,
0.035908784717321396,
-0.06157713383436203,
-0.11049041152000427,
-0.004999969620257616,
-0.09602393954992294,
-0.0922403484582901,
0.018123378977179527,
0.1151200607419014,
0.02763678878545761,
0.03168937563896179,
-0.04755963012576103,
0.10763691365718842,
0.023366501554846764,
-0.1662798374891281,
0.005249419249594212,
0.022999299690127373,
-0.02858167141675949,
-0.0348571240901947,
-0.03603449836373329,
-0.02497740462422371,
-0.005998498294502497,
0.10391648858785629,
-0.056142691522836685,
0.05425456166267395,
0.009628255851566792,
0.04976104572415352,
-0.09250131994485855,
0.16290953755378723,
-0.06878592818975449,
-0.03468257561326027,
0.02047041244804859,
0.15883634984493256,
0.027053700760006905,
-0.010381310246884823,
-0.10491899400949478,
0.02220495417714119,
0.08721409738063812,
0.03638143464922905,
0.001206211163662374,
0.048854511231184006,
-0.07273172587156296,
-0.01320872362703085,
0.07243660092353821,
-0.0926729217171669,
0.030055204406380653,
0.0064588007517158985,
-0.06476429849863052,
-0.06159970164299011,
0.010630075819790363,
0.015776587650179863,
0.03747326880693436,
0.06689665466547012,
-0.0849715918302536,
0.017216278240084648,
-0.08926385641098022,
-0.10575142502784729,
0.050340212881565094,
-0.11777295917272568,
-0.0011885890271514654,
-0.08587663620710373,
-0.11144808679819107,
-0.005499221850186586,
0.04281454533338547,
-0.02077217400074005,
-0.03172307461500168,
-0.07406511902809143,
-0.08351591974496841,
0.03717118129134178,
0.0028828775975853205,
0.11394907534122467,
-0.07256041467189789,
0.11324834078550339,
0.03438590466976166,
0.07995566725730896,
-0.03013612888753414,
0.02815224789083004,
-0.04767608270049095,
0.058263149112463,
-0.16992855072021484,
0.06659785658121109,
-0.0787903368473053,
0.03307756781578064,
-0.09261636435985565,
-0.061672113835811615,
0.047537077218294144,
-0.02476096712052822,
0.05047726258635521,
0.12673497200012207,
-0.22814132273197174,
-0.053500570356845856,
0.1553773730993271,
-0.07744207978248596,
-0.153091698884964,
0.1075049638748169,
-0.035960134118795395,
-0.027188342064619064,
0.034178607165813446,
0.1607249528169632,
0.09012421220541,
-0.16012559831142426,
-0.010979061014950275,
0.012257763184607029,
0.08342953771352768,
-0.04660770297050476,
0.12326165288686752,
0.03865659981966019,
0.06713896989822388,
-0.021084241569042206,
-0.1011606976389885,
0.06726216524839401,
-0.10788486897945404,
-0.09422174096107483,
-0.022555474191904068,
-0.09875858575105667,
0.028633328154683113,
0.05305412411689758,
0.03032688982784748,
-0.08097915351390839,
-0.10162666440010071,
0.025516143068671227,
0.10785278677940369,
-0.08193409442901611,
0.045552775263786316,
-0.10463758558034897,
0.09860710054636002,
-0.11288879811763763,
-0.029714424163103104,
-0.14954721927642822,
-0.09557193517684937,
0.009259525686502457,
-0.00007969753642100841,
-0.021631568670272827,
-0.04177950695157051,
0.06216295436024666,
0.0940709188580513,
-0.06011471152305603,
-0.047779470682144165,
-0.08289602398872375,
0.02288058027625084,
-0.07096036523580551,
-0.2207273542881012,
-0.06547114998102188,
-0.05977083370089531,
0.1760612279176712,
-0.19882315397262573,
-0.000722520228009671,
0.05334259569644928,
0.13329842686653137,
0.06402863562107086,
-0.058054111897945404,
0.005761889740824699,
0.05211537331342697,
-0.010805385187268257,
-0.08912115544080734,
0.058975689113140106,
0.01329054869711399,
-0.10082045197486877,
-0.002968474989756942,
-0.11990287899971008,
0.1003493219614029,
0.13193750381469727,
-0.05349205434322357,
-0.04137687012553215,
0.01624680869281292,
-0.05171443894505501,
-0.017612414434552193,
-0.022007806226611137,
0.008662767708301544,
0.07401081919670105,
0.022895388305187225,
0.1389944851398468,
-0.08643148839473724,
-0.03571237623691559,
0.05884448438882828,
-0.019574418663978577,
-0.02843605913221836,
0.08291950821876526,
0.0767742320895195,
-0.08574697375297546,
0.13219858705997467,
0.16581955552101135,
-0.04145370051264763,
0.1525193303823471,
-0.07297232747077942,
-0.10075026005506516,
-0.03188847377896309,
0.010267128236591816,
0.036697328090667725,
0.1688995063304901,
-0.05199746787548065,
-0.015655724331736565,
0.019182708114385605,
-0.011276328936219215,
-0.018506793305277824,
-0.22748234868049622,
-0.04107469692826271,
0.034854501485824585,
-0.07748465985059738,
-0.029309822246432304,
-0.02578636072576046,
-0.0147823765873909,
0.09721562266349792,
0.030791545286774635,
-0.06441047787666321,
0.036998454481363297,
-0.022968869656324387,
-0.06995464116334915,
0.17713932693004608,
-0.08163478970527649,
-0.15491588413715363,
-0.11720810830593109,
-0.053615033626556396,
-0.02906402200460434,
0.010173018090426922,
0.01925462856888771,
-0.08147457242012024,
-0.028378482908010483,
-0.08721676468849182,
-0.06249198690056801,
0.007516343146562576,
0.03281016647815704,
0.057691752910614014,
0.018659625202417374,
0.08640989661216736,
-0.07524815201759338,
0.010722218081355095,
-0.016633251681923866,
-0.052175529301166534,
0.040934767574071884,
0.0412503145635128,
0.14020505547523499,
0.11287728697061539,
-0.024317797273397446,
0.021720128133893013,
-0.04595404863357544,
0.2341282069683075,
-0.1137281209230423,
-0.013204601593315601,
0.1341903805732727,
-0.033824462443590164,
0.04840005934238434,
0.13874465227127075,
0.05999951809644699,
-0.09364397823810577,
-0.006264841184020042,
0.010270914062857628,
-0.030720233917236328,
-0.1674499213695526,
-0.024233803153038025,
-0.05533706769347191,
-0.00232850038446486,
0.11443793028593063,
0.022100387141108513,
-0.029866928234696388,
0.04811182618141174,
-0.005777512211352587,
0.04310259968042374,
0.001243877923116088,
0.08201077580451965,
0.06666816025972366,
0.05610709637403488,
0.10122837126255035,
-0.049920354038476944,
-0.0007180600659921765,
0.0359034389257431,
0.04076513275504112,
0.20884498953819275,
-0.005077298730611801,
0.18334178626537323,
0.061456356197595596,
0.15202203392982483,
0.02614891715347767,
0.02649620547890663,
0.0026448178105056286,
-0.029095491394400597,
0.001983578084036708,
-0.056763775646686554,
-0.03877193108201027,
0.016627229750156403,
-0.012278040871024132,
-0.0019864332862198353,
-0.1067550852894783,
0.06144680455327034,
0.043382808566093445,
0.2505447566509247,
0.07996466010808945,
-0.373210608959198,
-0.07670176029205322,
0.015152269043028355,
0.01040394976735115,
-0.038223881274461746,
0.003562919097021222,
0.15961791574954987,
-0.06484482437372208,
0.08050259202718735,
-0.07177574932575226,
0.06923197954893112,
-0.0742291659116745,
0.003346499986946583,
0.11012479662895203,
0.06689146906137466,
0.008055489510297775,
0.04374035447835922,
-0.2479795664548874,
0.2833007872104645,
0.019194355234503746,
0.06932079046964645,
-0.03520270437002182,
-0.020663518458604813,
0.024554306641221046,
0.05369638279080391,
0.13268834352493286,
-0.005500487517565489,
-0.06603018194437027,
-0.20388315618038177,
-0.09096333384513855,
0.028871089220046997,
0.10813656449317932,
-0.021704625338315964,
0.11480095237493515,
-0.023417005315423012,
-0.02990017645061016,
0.05332503840327263,
-0.07225026935338974,
-0.07154673337936401,
-0.09597498178482056,
-0.0009427718468941748,
0.03861585259437561,
-0.003106815507635474,
-0.09039510786533356,
-0.10228808969259262,
-0.07144681364297867,
0.12587182223796844,
-0.08261735737323761,
-0.022928012534976006,
-0.11930082738399506,
0.0733756273984909,
0.07941342145204544,
-0.07191600650548935,
0.07174335420131683,
-0.010467846877872944,
0.17081864178180695,
-0.01061234436929226,
-0.050173647701740265,
0.1038723960518837,
-0.08245942741632462,
-0.1832558959722519,
-0.07126166671514511,
0.12497056275606155,
0.004353191703557968,
0.05956636369228363,
-0.013111437670886517,
0.0378425233066082,
0.009691594168543816,
-0.05784818157553673,
0.025039251893758774,
0.0033213135320693254,
0.05026617646217346,
-0.02738855592906475,
-0.019798878580331802,
-0.008867443539202213,
-0.05351316183805466,
-0.012824621051549911,
0.15169228613376617,
0.3323821425437927,
-0.10680330544710159,
0.008984152227640152,
0.041946396231651306,
-0.030258361250162125,
-0.2058493196964264,
0.03627493605017662,
0.08693467825651169,
-0.005003742408007383,
0.01624279096722603,
-0.1384013444185257,
0.049992360174655914,
0.06980107724666595,
-0.029877863824367523,
0.1109836995601654,
-0.26589611172676086,
-0.12674736976623535,
0.055854327976703644,
0.1712208092212677,
0.043166384100914,
-0.1295974850654602,
-0.02871466986835003,
0.0003373011713847518,
-0.1622859537601471,
0.12672537565231323,
-0.05670083686709404,
0.11208314448595047,
-0.010498413816094398,
0.029110193252563477,
0.0010554963955655694,
-0.06115703284740448,
0.1462772786617279,
-0.019431838765740395,
0.08723695576190948,
-0.038536783307790756,
-0.045182131230831146,
0.10400955379009247,
-0.08412734419107437,
-0.0031572601292282343,
-0.06497523188591003,
0.025323674082756042,
-0.11344949156045914,
0.002409702632576227,
-0.08307812362909317,
-0.02255239710211754,
-0.0317709855735302,
-0.03395508974790573,
-0.03362559154629707,
0.06920686364173889,
0.0383707731962204,
-0.002238603075966239,
0.1859414279460907,
-0.0007993797771632671,
0.08960283547639847,
0.1318320482969284,
0.0690697729587555,
-0.05339718982577324,
-0.055618032813072205,
-0.018305350095033646,
-0.03326772525906563,
0.07491951435804367,
-0.14624352753162384,
0.03237660974264145,
0.1246565505862236,
0.03176059201359749,
0.135626420378685,
0.0525362528860569,
-0.03604275733232498,
0.04661093279719353,
0.09698931127786636,
-0.1255921721458435,
-0.10383638739585876,
0.007756377104669809,
-0.008139439858496189,
-0.13223206996917725,
-0.010696694254875183,
0.10125240683555603,
-0.07236015796661377,
0.014601417817175388,
-0.017529627308249474,
0.020383771508932114,
-0.05071290209889412,
0.16221043467521667,
0.07256736606359482,
0.06460659950971603,
-0.10709274560213089,
0.11174017935991287,
0.04592949151992798,
-0.10931921005249023,
0.0011368978302925825,
0.057194340974092484,
-0.0979316309094429,
-0.02053925395011902,
0.02839295007288456,
0.1251058429479599,
-0.04430895298719406,
-0.05412294343113899,
-0.1386171281337738,
-0.10362757742404938,
0.07107294350862503,
0.1444026082754135,
0.056905753910541534,
0.03044508956372738,
-0.0019627283327281475,
0.022151557728648186,
-0.13392557203769684,
0.11088737100362778,
0.025302628055214882,
0.07393442094326019,
-0.16707313060760498,
0.11830088496208191,
0.003298986703157425,
0.058514147996902466,
-0.02597605623304844,
0.042353555560112,
-0.05891042202711105,
0.01735652983188629,
-0.10033836960792542,
0.027778368443250656,
-0.04534980282187462,
0.017605861648917198,
-0.02435157634317875,
-0.055559322237968445,
-0.06121073663234711,
0.020354676991701126,
-0.08938519656658173,
-0.04482961818575859,
0.01727934740483761,
0.025314942002296448,
-0.14400027692317963,
-0.04895954206585884,
0.007934026420116425,
-0.09460011124610901,
0.05941488966345787,
0.011646939441561699,
0.038905102759599686,
0.028011629357933998,
-0.12542824447155,
-0.03233139216899872,
0.06348146498203278,
-0.0035605409648269415,
0.05105962976813316,
-0.08776895701885223,
-0.006232386454939842,
-0.014921259135007858,
0.001961773494258523,
-0.003685946110635996,
0.08292187005281448,
-0.11190996319055557,
0.004120910540223122,
-0.013686928898096085,
-0.015575302764773369,
-0.06178469583392143,
0.0705462172627449,
0.09792186319828033,
0.0178157277405262,
0.17052963376045227,
-0.07672397047281265,
0.0005363755044527352,
-0.21971313655376434,
-0.018255267292261124,
-0.013769497163593769,
-0.10029681026935577,
-0.11571461707353592,
-0.028045957908034325,
0.08426082879304886,
-0.06734500825405121,
0.09135314077138901,
-0.03571028262376785,
0.035142600536346436,
0.05347949266433716,
-0.049061715602874756,
0.005701563321053982,
0.045200787484645844,
0.18871217966079712,
0.03325149044394493,
-0.022469665855169296,
0.04957643523812294,
0.014990936033427715,
0.1029968187212944,
0.10045118629932404,
0.15403391420841217,
0.14542239904403687,
0.015520580112934113,
0.0942351445555687,
0.0705777257680893,
-0.03526295721530914,
-0.15347720682621002,
0.1048978790640831,
-0.08273059129714966,
0.11682755500078201,
-0.008213725872337818,
0.1492472141981125,
0.13224492967128754,
-0.19616730511188507,
0.004534064792096615,
-0.02682637982070446,
-0.0794418677687645,
-0.0663256123661995,
-0.0666605681180954,
-0.08596383035182953,
-0.13625304400920868,
0.02665090002119541,
-0.11835091561079025,
0.061233993619680405,
0.10388752818107605,
0.031245211139321327,
0.0044085136614739895,
0.19348518550395966,
0.02730143815279007,
0.025540200993418694,
0.09937304258346558,
0.029996100813150406,
-0.006444228813052177,
-0.0296743456274271,
-0.08375907689332962,
0.05927149951457977,
-0.03662831336259842,
0.03874830901622772,
-0.044008877128362656,
-0.04301232844591141,
0.09100517630577087,
0.031417086720466614,
-0.12702929973602295,
0.035120826214551926,
0.01684485375881195,
0.0743199959397316,
0.0649828240275383,
0.015148701146245003,
0.038967568427324295,
0.007830331102013588,
0.19539831578731537,
-0.06033894047141075,
-0.06703494489192963,
-0.10569752007722855,
0.15238888561725616,
-0.024900028482079506,
-0.013263518922030926,
0.016410011798143387,
-0.09253298491239548,
-0.0003183090011589229,
0.14547012746334076,
0.15694692730903625,
-0.11084409803152084,
-0.016518592834472656,
-0.02812315709888935,
-0.005822180304676294,
-0.05084666237235069,
0.11178966611623764,
0.06453780829906464,
-0.0005481139523908496,
-0.09116276353597641,
-0.08242805302143097,
-0.04688442498445511,
-0.017043795436620712,
0.007449000608175993,
0.010184607468545437,
0.013694974593818188,
0.009066061116755009,
-0.08111429214477539,
0.06056302785873413,
-0.02382509782910347,
-0.11883992701768875,
0.09834013879299164,
-0.19802634418010712,
-0.15404343605041504,
-0.023118620738387108,
0.09416410326957703,
0.00122396694496274,
0.030300624668598175,
-0.014661857858300209,
0.009048505686223507,
0.05257939174771309,
-0.015651684254407883,
-0.030615350231528282,
-0.09557074308395386,
0.07401274889707565,
-0.08103303611278534,
0.23884090781211853,
-0.04054170846939087,
0.05613091215491295,
0.09790061414241791,
0.03495083749294281,
-0.11597831547260284,
0.06320733577013016,
0.06995847076177597,
-0.06578560173511505,
0.011510677635669708,
0.15469586849212646,
-0.06267722696065903,
0.1364138275384903,
0.05300243943929672,
-0.08839091658592224,
0.013512391597032547,
-0.08188850432634354,
-0.06769725680351257,
-0.048391297459602356,
-0.048953138291835785,
-0.03195008262991905,
0.15360453724861145,
0.17207901179790497,
-0.05264612287282944,
0.008585131727159023,
-0.05858699977397919,
0.01883046142756939,
0.10641258209943771,
0.058880798518657684,
-0.0381380170583725,
-0.2117261439561844,
0.024763628840446472,
0.054360613226890564,
0.01153299119323492,
-0.19027848541736603,
-0.09127585589885712,
-0.009020330384373665,
-0.05569649487733841,
-0.056614067405462265,
0.10508375614881516,
0.06864593923091888,
0.04386597499251366,
-0.06802455335855484,
-0.06888433545827866,
-0.03132294863462448,
0.1504383236169815,
-0.147828608751297,
-0.0876506119966507
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.8.2 | {"library_name": "peft", "base_model": "meta-llama/Llama-2-7b-hf"} | null | noza-kit/Adapter_llama2_translate_Q_enpt_ex2-3epoch | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:meta-llama/Llama-2-7b-hf",
"region:us"
] | 2024-02-07T16:35:34+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #safetensors #arxiv-1910.09700 #base_model-meta-llama/Llama-2-7b-hf #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.8.2 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
"TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-meta-llama/Llama-2-7b-hf #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
41,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-meta-llama/Llama-2-7b-hf #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2"
] | [
-0.13237035274505615,
0.20393601059913635,
-0.002511046128347516,
0.02874687872827053,
0.07912357151508331,
0.019634027034044266,
0.04979075863957405,
0.13531364500522614,
0.020043307915329933,
0.10580451786518097,
0.0737132653594017,
0.11718367785215378,
0.10926163196563721,
0.20654499530792236,
0.003890186781063676,
-0.15440793335437775,
0.024214256554841995,
-0.08298544585704803,
0.006814117077738047,
0.1290476769208908,
0.14319083094596863,
-0.10468140989542007,
0.0831538662314415,
-0.014203370548784733,
0.0008161105797626078,
-0.03246506303548813,
-0.06674343347549438,
-0.015596466138958931,
0.04917285591363907,
0.02522817626595497,
0.05882670730352402,
-0.010089844465255737,
0.0929119735956192,
-0.26152917742729187,
0.018749000504612923,
0.04154228791594505,
0.009074261412024498,
0.08363344520330429,
0.0979103073477745,
-0.04074648395180702,
0.12078511714935303,
-0.024994686245918274,
0.13832204043865204,
0.09345067292451859,
-0.08226727694272995,
-0.233157217502594,
-0.06684722006320953,
0.07271547615528107,
0.18968668580055237,
0.08927863836288452,
-0.044125091284513474,
0.14097759127616882,
-0.07517150044441223,
0.02484818734228611,
0.04656748101115227,
-0.09290260076522827,
-0.06676048040390015,
0.0702265128493309,
0.13261590898036957,
0.0625041052699089,
-0.12113244831562042,
-0.03750992938876152,
0.03344248607754707,
0.044793009757995605,
0.06062353774905205,
0.005180627107620239,
0.16268815100193024,
0.034240271896123886,
-0.14592847228050232,
-0.05353321507573128,
0.14678435027599335,
0.01157673355191946,
-0.04636283218860626,
-0.21997328102588654,
-0.0027822081465274096,
-0.09489403665065765,
-0.022923149168491364,
-0.05228540673851967,
0.03324316069483757,
0.00603833794593811,
0.1196645051240921,
-0.042089227586984634,
-0.09635167568922043,
-0.029711460694670677,
0.0996040627360344,
0.05452839657664299,
0.02769845724105835,
-0.02099502831697464,
0.010653719305992126,
0.1290775090456009,
0.08296726644039154,
-0.1341402530670166,
-0.07021861523389816,
-0.0753326416015625,
-0.04316629841923714,
-0.03228989988565445,
0.03893959894776344,
0.019871119409799576,
0.07120058685541153,
0.2619621157646179,
-0.022196462377905846,
0.06401924788951874,
0.061033982783555984,
0.01709051802754402,
0.04062429443001747,
0.10795178264379501,
-0.03382651507854462,
-0.15705206990242004,
-0.007360270246863365,
0.10362072288990021,
-0.004135396331548691,
-0.02802850492298603,
-0.045986633747816086,
0.03152812272310257,
0.044165465980768204,
0.11501371115446091,
0.11203816533088684,
-0.019931387156248093,
-0.07717939466238022,
-0.05966082587838173,
0.19364216923713684,
-0.16149258613586426,
0.038572292774915695,
0.02467195875942707,
-0.006866174750030041,
-0.06484853476285934,
0.0073310090228915215,
0.016164373606443405,
-0.027354510501027107,
0.0603426918387413,
-0.0646006166934967,
-0.04179375246167183,
-0.1283673793077469,
-0.02387934736907482,
0.032629046589136124,
0.0170845165848732,
-0.0421639084815979,
-0.046661876142024994,
-0.08786044269800186,
-0.11000633984804153,
0.10926247388124466,
-0.05313732475042343,
-0.052913907915353775,
-0.02804330736398697,
-0.08941388875246048,
0.022293368354439735,
0.027490468695759773,
0.0755976140499115,
-0.02891632728278637,
0.052480049431324005,
0.003703000722452998,
0.059941843152046204,
0.0814133733510971,
0.027145687490701675,
-0.08097686618566513,
0.06685694307088852,
-0.19895170629024506,
0.07886288315057755,
-0.08557034283876419,
0.035526763647794724,
-0.16191443800926208,
-0.008882720954716206,
0.015485688112676144,
0.028551144525408745,
0.0418417863547802,
0.16628479957580566,
-0.21890771389007568,
-0.021091977134346962,
0.15901808440685272,
-0.10847076028585434,
-0.1374696046113968,
0.0436418242752552,
-0.04286689683794975,
0.18280568718910217,
0.028055870905518532,
0.010343263857066631,
0.09726855903863907,
-0.16840705275535583,
-0.02907063439488411,
-0.021288467571139336,
0.0036895605735480785,
0.07365763932466507,
0.09041544795036316,
-0.09089618921279907,
-0.0016403654590249062,
0.012144356034696102,
-0.06943254172801971,
-0.015110267326235771,
-0.04118245840072632,
-0.10628213733434677,
0.002018203027546406,
-0.09110194444656372,
0.023759065195918083,
0.0035124430432915688,
-0.09477277845144272,
-0.008542876690626144,
-0.1573835164308548,
-0.0652049109339714,
0.09409166127443314,
0.0002530320198275149,
-0.024702679365873337,
-0.10900412499904633,
0.06465248018503189,
-0.03883763402700424,
-0.026517964899539948,
-0.14125961065292358,
-0.023071611300110817,
0.01673055998980999,
-0.14134323596954346,
-0.01001854706555605,
-0.12183605134487152,
0.06567396223545074,
0.005137317348271608,
-0.0481104739010334,
-0.04708600044250488,
-0.004086394794285297,
0.0014921361580491066,
-0.05505292862653732,
-0.23444515466690063,
-0.028233496472239494,
-0.05085372179746628,
0.16539393365383148,
-0.2289838343858719,
0.044271692633628845,
0.014694449491798878,
0.11615854501724243,
-0.0018446118338033557,
-0.0661761611700058,
0.022094158455729485,
-0.07084274291992188,
-0.025033291429281235,
-0.07177132368087769,
-0.0071777342818677425,
0.00008959023398347199,
-0.029647991061210632,
0.015313859097659588,
-0.10952108353376389,
-0.053884293884038925,
0.100620798766613,
0.060472261160612106,
-0.14894865453243256,
0.008543584495782852,
-0.03779032453894615,
-0.06071627512574196,
-0.07427168637514114,
-0.0695083886384964,
0.0856412947177887,
0.052977994084358215,
0.03996400535106659,
-0.0812206119298935,
-0.07201940566301346,
0.005019875708967447,
-0.02742239646613598,
-0.005877636838704348,
0.11996077746152878,
0.07278608530759811,
-0.10015858709812164,
0.0890948474407196,
0.07567999511957169,
0.012905389070510864,
0.07863839715719223,
-0.028960783034563065,
-0.10615462064743042,
-0.03149069845676422,
0.05891314521431923,
0.0075002689845860004,
0.18196412920951843,
-0.07219336181879044,
0.05777830258011818,
0.046155888587236404,
-0.046635568141937256,
0.05089704319834709,
-0.09103982150554657,
0.0068960352800786495,
0.00045980032882653177,
-0.017081741243600845,
0.029599705711007118,
-0.020320137962698936,
0.006365274079144001,
0.07632698118686676,
0.05559656023979187,
0.02392573468387127,
0.023359429091215134,
-0.037590380758047104,
-0.1454712599515915,
0.18398217856884003,
-0.09283597022294998,
-0.235765740275383,
-0.15705986320972443,
0.0616452731192112,
0.049257904291152954,
-0.015689486637711525,
0.02697811834514141,
-0.055544715374708176,
-0.10059839487075806,
-0.08630408346652985,
-0.001965506933629513,
0.033574361354112625,
-0.05912783369421959,
-0.07473962754011154,
0.045523062348365784,
0.04523130878806114,
-0.11779510229825974,
0.02612960711121559,
0.06724361330270767,
-0.01014306303113699,
0.002122951438650489,
0.05421233922243118,
0.09625556319952011,
0.1871589571237564,
-0.0047584883868694305,
0.006493487861007452,
0.06463784724473953,
0.27302834391593933,
-0.16097134351730347,
0.10603976994752884,
0.1468280404806137,
-0.06509615480899811,
0.06928659975528717,
0.1811111718416214,
0.024897225201129913,
-0.0959320068359375,
0.024916043505072594,
0.02835996262729168,
-0.01960386149585247,
-0.2740720212459564,
-0.0512622706592083,
-0.015117009170353413,
-0.08622704446315765,
0.07128944247961044,
0.08718991279602051,
0.07891540229320526,
0.03938929736614227,
-0.05623466521501541,
-0.11011259257793427,
0.02521095983684063,
0.10682129859924316,
-0.01211885642260313,
0.003295447211712599,
0.08167944848537445,
-0.04613311216235161,
0.007927946746349335,
0.08699803054332733,
-0.01990879327058792,
0.1374768167734146,
0.04775961861014366,
0.09206060320138931,
0.08603846281766891,
0.10468525439500809,
-0.011216369457542896,
0.031460702419281006,
0.01713097095489502,
0.023083847016096115,
0.025577327236533165,
-0.0892123356461525,
0.00939508993178606,
0.11217135936021805,
0.02443520911037922,
0.02237142249941826,
0.016059260815382004,
-0.042084116488695145,
0.035355109721422195,
0.19778503477573395,
0.02863113395869732,
-0.21936152875423431,
-0.08315163850784302,
0.04950554668903351,
-0.07752750813961029,
-0.15846198797225952,
-0.0069001950323581696,
0.02585102617740631,
-0.16377925872802734,
0.015679948031902313,
-0.04114160314202309,
0.10047675669193268,
-0.07824478298425674,
-0.04026156663894653,
0.11029542237520218,
0.047400183975696564,
-0.01943347603082657,
0.05451195687055588,
-0.19536079466342926,
0.10843666642904282,
0.02992161363363266,
0.07536879926919937,
-0.08786998689174652,
0.09398660063743591,
0.006047630682587624,
-0.019160762429237366,
0.16931316256523132,
-0.0001144029592978768,
-0.049934081733226776,
-0.08560120314359665,
-0.09227954596281052,
0.0015766898868605494,
0.07818529009819031,
-0.12631447613239288,
0.0825691819190979,
-0.03569265082478523,
-0.024482207372784615,
-0.008127174340188503,
-0.08541606366634369,
-0.1325976550579071,
-0.14982733130455017,
0.05399367958307266,
-0.0976201519370079,
0.02554609440267086,
-0.08825770765542984,
-0.05347679927945137,
0.016768373548984528,
0.18224331736564636,
-0.21447692811489105,
-0.10864878445863724,
-0.14267513155937195,
-0.11213549226522446,
0.16079570353031158,
-0.042837124317884445,
0.08159231394529343,
0.00010400224709883332,
0.15704618394374847,
0.01110734511166811,
-0.015090357512235641,
0.08682332187891006,
-0.09437134861946106,
-0.19026298820972443,
-0.04887847229838371,
0.16311104595661163,
0.1444961428642273,
0.029530119150877,
-0.005065699107944965,
0.02549002133309841,
-0.06952440738677979,
-0.11216824501752853,
0.02609189972281456,
0.16361786425113678,
0.07300680130720139,
-0.012950204312801361,
-0.025871867313981056,
-0.0997539535164833,
-0.05963310971856117,
-0.04339827224612236,
-0.00898770522326231,
0.20425592362880707,
-0.06497634947299957,
0.14582973718643188,
0.10464579612016678,
-0.05606960505247116,
-0.21339629590511322,
0.03492094576358795,
0.04277806729078293,
0.026418045163154602,
0.04313372075557709,
-0.18166027963161469,
0.09741673618555069,
-0.014149999246001244,
-0.08650295436382294,
0.17498920857906342,
-0.17328102886676788,
-0.13439859449863434,
0.1159968227148056,
0.025544147938489914,
-0.21331895887851715,
-0.13972461223602295,
-0.10190334171056747,
-0.0198976993560791,
-0.126362144947052,
0.036111894994974136,
-0.0036879852414131165,
0.00850605871528387,
0.012948633171617985,
0.018173353746533394,
0.039593230932950974,
-0.05594787001609802,
0.21268853545188904,
-0.03937339782714844,
0.000047609177272534,
-0.050931964069604874,
-0.06770505011081696,
0.023772839456796646,
-0.0565045028924942,
0.12416863441467285,
-0.01210821233689785,
0.039195943623781204,
-0.17265570163726807,
-0.04285977780818939,
-0.058010976761579514,
0.03728554770350456,
-0.09242235124111176,
-0.0793662965297699,
-0.04483490809798241,
0.09155189245939255,
0.09041202813386917,
-0.018728721886873245,
0.0019666242878884077,
-0.09585212171077728,
0.07403325289487839,
0.20964933931827545,
0.20306745171546936,
0.0681707113981247,
-0.05247919633984566,
0.02836998738348484,
-0.03519117832183838,
0.04444263130426407,
-0.2148476094007492,
0.0430048331618309,
0.0631239265203476,
0.024400800466537476,
0.06267635524272919,
-0.01054441649466753,
-0.1590016484260559,
-0.07973737269639969,
0.08659059554338455,
-0.0608268640935421,
-0.16209019720554352,
-0.03262902423739433,
0.02129248157143593,
-0.2115628719329834,
-0.04105594381690025,
0.03599734604358673,
-0.014814808964729309,
-0.03840542584657669,
0.021407432854175568,
0.07970889657735825,
-0.028947602957487106,
0.1049608662724495,
0.09329938143491745,
0.09604475647211075,
-0.09774979948997498,
0.05453461781144142,
0.07179035246372223,
-0.031663764268159866,
0.03226640820503235,
0.1210775151848793,
-0.04315068572759628,
-0.046701591461896896,
0.08053972572088242,
0.11871292442083359,
-0.00035442441003397107,
-0.06335891038179398,
-0.0028557574842125177,
-0.0440225712954998,
0.054060470312833786,
0.10412941128015518,
0.036388467997312546,
0.0012024412862956524,
0.07687212526798248,
0.028011957183480263,
-0.09147296100854874,
0.12449978291988373,
0.06066809967160225,
0.02483541890978813,
-0.05523430183529854,
-0.038621995598077774,
-0.015819178894162178,
-0.0028008304070681334,
-0.01961326226592064,
-0.0014547118917107582,
-0.08309019356966019,
0.0061004795134067535,
-0.13227513432502747,
0.022323906421661377,
-0.07725922018289566,
0.00378548726439476,
0.036021001636981964,
-0.046576302498579025,
0.0013563713291659951,
-0.0008801636286079884,
-0.07430332899093628,
-0.05454954877495766,
-0.01629588007926941,
0.07790114730596542,
-0.13923588395118713,
0.03906119614839554,
0.07606222480535507,
-0.10726266354322433,
0.06878530234098434,
-0.007731399964541197,
0.008601504378020763,
0.0010856596054509282,
-0.13779860734939575,
0.05484551563858986,
-0.028775036334991455,
-0.006356567144393921,
0.005071246065199375,
-0.19585701823234558,
-0.00865773856639862,
-0.03182972967624664,
-0.0634872317314148,
0.019731810316443443,
-0.001073729363270104,
-0.11955288797616959,
0.1077868640422821,
0.004837313666939735,
-0.05712589994072914,
-0.0236744936555624,
0.042738161981105804,
0.0863419771194458,
-0.0053856209851801395,
0.12532570958137512,
-0.0293873380869627,
0.07612910121679306,
-0.17633569240570068,
-0.010070881806313992,
-0.015794692561030388,
0.05993741378188133,
-0.019834399223327637,
-0.03712667524814606,
0.06236843764781952,
-0.027145320549607277,
0.17265751957893372,
-0.004146610386669636,
0.07253459841012955,
0.0493277981877327,
0.008650471456348896,
0.04884583130478859,
0.07257263362407684,
0.06367837637662888,
-0.017801770940423012,
0.00016894470900297165,
0.04386947304010391,
-0.002970502246171236,
-0.051965516060590744,
-0.15762734413146973,
0.06277678161859512,
0.17842786014080048,
0.056998081505298615,
0.030175408348441124,
0.012138530611991882,
-0.12049488723278046,
-0.07329574972391129,
0.10845038294792175,
-0.021686408668756485,
-0.031095284968614578,
-0.06442723423242569,
0.21323516964912415,
0.1388614922761917,
-0.19825653731822968,
0.0702671930193901,
-0.06280558556318283,
-0.04658647999167442,
-0.14314492046833038,
-0.17366671562194824,
-0.059809304773807526,
-0.0547034814953804,
-0.026051264256238937,
-0.054700352251529694,
0.04570859298110008,
0.047346316277980804,
-0.0016739139100536704,
-0.02772514894604683,
0.1126171201467514,
0.02765420638024807,
-0.032165806740522385,
0.04451003298163414,
0.05619681254029274,
0.03682970255613327,
-0.09137814491987228,
0.007322985213249922,
0.0029695341363549232,
0.014342821203172207,
0.06777288764715195,
0.01613135077059269,
-0.06992621719837189,
0.02725713886320591,
-0.020467489957809448,
-0.12120343744754791,
0.042514219880104065,
-0.005491400603204966,
-0.02191038616001606,
0.14766326546669006,
0.039597559720277786,
0.008086306042969227,
-0.014769108034670353,
0.22978916764259338,
-0.079631008207798,
-0.08263124525547028,
-0.1393512636423111,
0.07894771546125412,
-0.07535439729690552,
0.020168637856841087,
0.02652786672115326,
-0.12502749264240265,
0.017455779016017914,
0.17437158524990082,
0.11967697739601135,
-0.01862110011279583,
0.005760727450251579,
0.04387581720948219,
0.003006097162142396,
-0.04732988774776459,
0.01692454144358635,
0.05290905013680458,
0.19558346271514893,
-0.0746847614645958,
0.054245725274086,
-0.01774757355451584,
-0.08059251308441162,
-0.020728278905153275,
0.09288354963064194,
-0.009933017194271088,
-0.004748775623738766,
-0.06074956804513931,
0.149005725979805,
-0.0759778842329979,
-0.20890262722969055,
0.06107410788536072,
-0.057474348694086075,
-0.13986754417419434,
-0.043588198721408844,
0.03270360454916954,
-0.02818191610276699,
-0.0004342520551290363,
0.05878293514251709,
-0.041880737990140915,
0.1787300854921341,
0.02775873802602291,
-0.04535049945116043,
-0.08805633336305618,
0.060195520520210266,
-0.15322564542293549,
0.28409940004348755,
0.02300625666975975,
0.06475372612476349,
0.11462150514125824,
-0.023716775700449944,
-0.14765876531600952,
0.016111766919493675,
0.11251717060804367,
-0.07146475464105606,
0.06923303008079529,
0.16616879403591156,
0.00888645276427269,
0.12871026992797852,
0.06517354398965836,
-0.04169101640582085,
0.03372213616967201,
-0.08477409183979034,
-0.04430316761136055,
-0.1301726996898651,
0.07585147768259048,
-0.09351208806037903,
0.15738072991371155,
0.11715016514062881,
-0.07169844210147858,
0.010452828370034695,
-0.02282477170228958,
0.09099912643432617,
0.012017005123198032,
0.10486294329166412,
0.01101954746991396,
-0.19380232691764832,
0.04388235881924629,
0.012521770782768726,
0.09230010956525803,
-0.21009819209575653,
-0.05027567222714424,
0.04558335989713669,
-0.022896859794855118,
-0.06855283677577972,
0.11809497326612473,
0.03357189893722534,
0.028112467378377914,
-0.037041857838630676,
-0.032784342765808105,
0.007307000923901796,
0.151776984333992,
-0.11639050394296646,
-0.019398227334022522
] |
null | null | transformers | # MetricX-23
*This is not an officially supported Google product.*
**GitHub repository: [https://github.com/google-research/metricx](https://github.com/google-research/metricx)**
This repository contains the MetricX-23 models,
a family of models for automatic evaluation of translations that were proposed
in the WMT'23 Metrics Shared Task submission
[MetricX-23: The Google Submission to the WMT 2023 Metrics Shared Task](https://aclanthology.org/2023.wmt-1.63/).
The models were trained in [T5X](https://github.com/google-research/t5x) and
then converted for use in PyTorch.
## Available Models
There are 6 models available on HuggingFace that vary in the number of
parameters and whether or not the model is reference-based or reference-free
(also known as quality estimation, or QE):
* [MetricX-23-XXL](https://huggingface.co/google/metricx-23-large-v2p0)
* [MetricX-23-XL](https://huggingface.co/google/metricx-23-xl-v2p0)
* [MetricX-23-Large](https://huggingface.co/google/metricx-23-xxl-v2p0)
* [MetricX-23-QE-XXL](https://huggingface.co/google/metricx-23-qe-large-v2p0)
* [MetricX-23-QE-XL](https://huggingface.co/google/metricx-23-qe-xl-v2p0)
* [MetricX-23-QE-Large](https://huggingface.co/google/metricx-23-qe-xxl-v2p0)
We recommend using the XXL model versions for the best agreement with human
judgments of translation quality, the Large versions for best speed, and the
XL for an intermediate use case.
## Changes to the WMT'23 Submission
These models available here are most similar to the primary submission to the WMT'23 Metrics
Shared Task. They are initialized with [mT5](https://aclanthology.org/2021.naacl-main.41/)
then fine-tuned on a combination of direct assessment and MQM data. However,
we made some changes that make these models different from the WMT'23 submissions.
First, the models are trained to regress the actual MQM score rather than a
normalized score between 0 and 1. **That means the output from the MetricX-23
models is a score in the range [0, 25] where lower is better (i.e., it predicts
an error score).**
Second, these models were trained with a larger variety of synthetic data that
makes them more robust to translation edge cases like over- and undertranslation,
described in more detail in the following section.
### Synthetic Data
In order for our MetricX models to learn to identify certain types of bad
translations that are not sufficiently (or at all) represented in the regular
training data, we created synthetic examples and mixed them in during training.
The synthetic training data was generated from the DA datasets ranging from
WMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have
the candidate translation manipulated so as to turn it into a bad translation
with a specific issue commonly unrecognized by learned metrics.
The table below provides an overview of the various failure modes that we
considered, including brief descriptions of how we prepared the synthetic data
to address them.
| Failure mode | Synthetic example description |
| ----------- | ----------- |
| Undertranslation | Candidate translation with an arbitrary sentence removed (if multi-sentence); alternatively, candidate with a certain proportion of words removed from the end. |
| Overtranslation | Candidate translation duplicated (with space in between). |
| Fluent but unrelated translation | Arbitrary reference of a similar length from the dataset. |
| Gibberish | Text of a similar length as the reference, generated by sampling words from the reference translation vocabulary (built from all references in the data). |
| Missing punctuation | Reference translation with the end punctuation removed (11 punctuation symbols considered). |
| Latin instead of Chinese/Japanese or Hindi/Bengali punctuation | Candidate translation with the language-specific punctuation symbol at the end replaced with the Latin equivalent (e.g., "." instead of "。" or "।"); alternatively, the punctuation symbol is replaced with the Latin equivalent in the reference, keeping the correct one in the candidate. |
| Reference-matching translation | Reference translation copied as the candidate translation (unlike the rest of the synthetic data, these examples are meant to train the metric to predict a perfect score for candidates matching the reference). |
Examples from the first 4 categories were assigned a label corresponding to the
worst score on the given rating scale (e.g., 25 when mixed with MQM training
data), whereas the reference-matching translation examples are assigned the best
score (e.g., 0 when used with MQM data). The missing/incorrect punctuation
examples were labeled with a score slightly worse than perfect.
Note that some of the synthetic datasets are only meaningful in the
reference-based scenario, and we thus excluded them when training a QE variant
of MetricX. These are the Latin-vs-special punctuation and the
reference-matching translation examples.
Most of the synthetic training sets were created using stratified sampling
across target languages, taking 500 examples per target language. One exception
is the missing punctuation set, which used a stratified sample across different
punctuation symbols instead.
When training MetricX, a small proportion of the synthetic examples was mixed
with the regular training examples. During the first-stage fine-tuning on DA
data, each synthetic training set constituted between 0.1% and 1% of all
training examples, whereas in the second-stage fine-tuning on MQM data we used
an even smaller proportion, around 0.05%.
As for evaluating the effect of the synthetic training data on the model's
performance, the DEMETR challenge set - which we originally used to evaluate the
models submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We
therefore created a new DEMETR-style test set based on the WMT22 DA data, with
examples constructed analogically to the synthetic training examples, as
described above. This test set helped us determine the right proportions of
synthetic data for fine-tuning in order to make MetricX robust for the failure
modes in consideration, without sacrificing the system- and segment-level
correlations with human ratings.
## Usage
The code for using MetricX models can be found at [https://github.com/google-research/metricx](https://github.com/google-research/metricx).
The repository contains example prediction scripts, described below.
The `metricx23/predict.py` script contains an example for how to run inference
on the models.
### Reference-Based
Example usage for a reference-based model:
```bash
python -m metricx23.predict \
--tokenizer google/mt5-xl \
--model_name_or_path google/metricx-23-xl-v2p0 \
--max_input_length 1024 \
--batch_size 1 \
--input_file input.jsonl \
--output_file output.jsonl
```
`input.jsonl` is expected to have 1 serialized JSON object per line with
`"reference"` and `"hypothesis"` fields. The output jsonl will be parallel
to `input.jsonl` but additionally contain a `"prediction"` field with the predicted score.
Note that the model was trained with a maximum input length of 1024 tokens, so
significantly increasing that value may lead to unpredictable behavior.
### Reference-Free
Example usage for a reference-free model:
```bash
python -m metricx23.predict \
--tokenizer google/mt5-xl \
--model_name_or_path google/metricx-23-qe-xl-v2p0 \
--max_input_length 1024 \
--batch_size 1 \
--input_file input.jsonl \
--output_file output.jsonl \
--qe
```
`input.jsonl` is expected to have 1 serialized JSON object per line with
`"source"` and `"hypothesis"` fields. The output jsonl will be parallel
to `input.jsonl` but additionally contain a `"prediction"` field with the predicted score.
## Meta-Evaluation
The `metricx23/evaluate.py` script contains code to calculate various correlations
between the MetricX-23 scores and MQM ratings of translation quality using the
[MT Metrics Eval](https://github.com/google-research/mt-metrics-eval) library.
Example usage:
```bash
python -m metricx23.evaluate \
--dataset wmt22 \
--lp en-de \
--input_file input.jsonl \
--output_file output.json
```
`input.jsonl` is expected to have one JSON object serialized per line.
Each JSON object is expected to contain 4 fields:
* `"system_id"`: The name of the system that generated the translation.
* `"segment_id"`: The 0-based index of the corresponding segment in the MT
Metrics Eval data.
* `"label"`: The ground-truth translation quality score (with higher is better).
* `"prediction"`: The model predicted translation quality score (with lower is
better; the script negates the scores so higher is better).
The script will calculate the 4 agreement/correlations that were used in the
WMT'23 Shared Task. Below are the results for the MetricX-23 models on the
WMT'22 Metrics Shared Task data:
English-German:
| Model | System-Level Accuracy | System-Level Pearson | Segment-Level Pearson | Segment-Level Pairwise Acc |
| ----------- | ----------- | ----------- | ----------- | ----------- |
| MetricX-23-XXL | 0.795 | 0.835 | 0.546 | 0.619 |
| MetricX-23-XL | 0.756 | 0.813 | 0.540 | 0.605 |
| MetricX-23-Large | 0.769 | 0.759 | 0.507 | 0.595 |
| MetricX-23-QE-XXL | 0.769 | 0.830 | 0.490 | 0.606 |
| MetricX-23-QE-XL | 0.718 | 0.684 | 0.421 | 0.594 |
| MetricX-23-QE-Large | 0.744 | 0.671 | 0.387 | 0.579 |
English-Russian:
| Model | System-Level Accuracy | System-Level Pearson | Segment-Level Pearson | Segment-Level Pairwise Acc |
| ----------- | ----------- | ----------- | ----------- | ----------- |
| MetricX-23-XXL | 0.905 | 0.943 | 0.477 | 0.609 |
| MetricX-23-XL | 0.876 | 0.906 | 0.498 | 0.589 |
| MetricX-23-Large | 0.876 | 0.841 | 0.474 | 0.569 |
| MetricX-23-QE-XXL | 0.895 | 0.940 | 0.470 | 0.602 |
| MetricX-23-QE-XL | 0.848 | 0.861 | 0.415 | 0.570 |
| MetricX-23-QE-Large | 0.819 | 0.778 | 0.411 | 0.551 |
Chinese-English:
| Model | System-Level Accuracy | System-Level Pearson | Segment-Level Pearson | Segment-Level Pairwise Acc |
| ----------- | ----------- | ----------- | ----------- | ----------- |
| MetricX-23-XXL | 0.868 | 0.919 | 0.605 | 0.551 |
| MetricX-23-XL | 0.868 | 0.924 | 0.584 | 0.543 |
| MetricX-23-Large | 0.857 | 0.919 | 0.555 | 0.539 |
| MetricX-23-QE-XXL | 0.857 | 0.928 | 0.573 | 0.544 |
| MetricX-23-QE-XL | 0.802 | 0.879 | 0.546 | 0.529 |
| MetricX-23-QE-Large | 0.758 | 0.904 | 0.522 | 0.529 |
The `metricx23/evaluate_wmt23.py` script re-calculates the average correlation
score that was used to rank submissions from the
[WMT'23 Shared Task](https://www2.statmt.org/wmt23/pdf/2023.wmt-1.51.pdf).
Example usage:
```bash
python -m metricx23.evaluate_wmt23 \
--en_de predictions_ende.jsonl \
--he_en predictions_heen.jsonl \
--zh_en predictions_zhen.jsonl \
--output_file output.json
```
Each of the 3 input files is expected to be in the same format as described
above. Each file should correspond to running inference on each of the language
pairs from the WMT'23 dataset.
The results for each of the models is the following:
| Model | Average Correlation |
| ----------- | ----------- |
| MetricX-23-XXL | 0.812 |
| MetricX-23-XL | 0.813 |
| MetricX-23-Large | 0.794 |
| MetricX-23-QE-XXL | 0.797 |
| MetricX-23-QE-XL | 0.767 |
| MetricX-23-QE-Large | 0.762 |
## Citation
If you use MetricX-23 in your research, please cite the following publication:
```bibtex
@inproceedings{juraska-etal-2023-metricx,
title = {{MetricX-23: The Google Submission to the WMT 2023 Metrics Shared Task}},
author = "Juraska, Juraj and
Finkelstein, Mara and
Deutsch, Daniel and
Siddhant, Aditya and
Mirzazadeh, Mehdi and
Freitag, Markus",
editor = "Koehn, Philipp and
Haddow, Barry and
Kocmi, Tom and
Monz, Christof",
booktitle = "Proceedings of the Eighth Conference on Machine Translation",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.wmt-1.63",
doi = "10.18653/v1/2023.wmt-1.63",
pages = "756--767",
}
``` | {"license": "apache-2.0"} | null | google/metricx-23-qe-large-v2p0 | [
"transformers",
"pytorch",
"mt5",
"license:apache-2.0",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T16:35:44+00:00 | [] | [] | TAGS
#transformers #pytorch #mt5 #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us
| MetricX-23
==========
*This is not an officially supported Google product.*
GitHub repository: URL
This repository contains the MetricX-23 models,
a family of models for automatic evaluation of translations that were proposed
in the WMT'23 Metrics Shared Task submission
MetricX-23: The Google Submission to the WMT 2023 Metrics Shared Task.
The models were trained in T5X and
then converted for use in PyTorch.
Available Models
----------------
There are 6 models available on HuggingFace that vary in the number of
parameters and whether or not the model is reference-based or reference-free
(also known as quality estimation, or QE):
* MetricX-23-XXL
* MetricX-23-XL
* MetricX-23-Large
* MetricX-23-QE-XXL
* MetricX-23-QE-XL
* MetricX-23-QE-Large
We recommend using the XXL model versions for the best agreement with human
judgments of translation quality, the Large versions for best speed, and the
XL for an intermediate use case.
Changes to the WMT'23 Submission
--------------------------------
These models available here are most similar to the primary submission to the WMT'23 Metrics
Shared Task. They are initialized with mT5
then fine-tuned on a combination of direct assessment and MQM data. However,
we made some changes that make these models different from the WMT'23 submissions.
First, the models are trained to regress the actual MQM score rather than a
normalized score between 0 and 1. That means the output from the MetricX-23
models is a score in the range [0, 25] where lower is better (i.e., it predicts
an error score).
Second, these models were trained with a larger variety of synthetic data that
makes them more robust to translation edge cases like over- and undertranslation,
described in more detail in the following section.
### Synthetic Data
In order for our MetricX models to learn to identify certain types of bad
translations that are not sufficiently (or at all) represented in the regular
training data, we created synthetic examples and mixed them in during training.
The synthetic training data was generated from the DA datasets ranging from
WMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have
the candidate translation manipulated so as to turn it into a bad translation
with a specific issue commonly unrecognized by learned metrics.
The table below provides an overview of the various failure modes that we
considered, including brief descriptions of how we prepared the synthetic data
to address them.
Examples from the first 4 categories were assigned a label corresponding to the
worst score on the given rating scale (e.g., 25 when mixed with MQM training
data), whereas the reference-matching translation examples are assigned the best
score (e.g., 0 when used with MQM data). The missing/incorrect punctuation
examples were labeled with a score slightly worse than perfect.
Note that some of the synthetic datasets are only meaningful in the
reference-based scenario, and we thus excluded them when training a QE variant
of MetricX. These are the Latin-vs-special punctuation and the
reference-matching translation examples.
Most of the synthetic training sets were created using stratified sampling
across target languages, taking 500 examples per target language. One exception
is the missing punctuation set, which used a stratified sample across different
punctuation symbols instead.
When training MetricX, a small proportion of the synthetic examples was mixed
with the regular training examples. During the first-stage fine-tuning on DA
data, each synthetic training set constituted between 0.1% and 1% of all
training examples, whereas in the second-stage fine-tuning on MQM data we used
an even smaller proportion, around 0.05%.
As for evaluating the effect of the synthetic training data on the model's
performance, the DEMETR challenge set - which we originally used to evaluate the
models submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We
therefore created a new DEMETR-style test set based on the WMT22 DA data, with
examples constructed analogically to the synthetic training examples, as
described above. This test set helped us determine the right proportions of
synthetic data for fine-tuning in order to make MetricX robust for the failure
modes in consideration, without sacrificing the system- and segment-level
correlations with human ratings.
Usage
-----
The code for using MetricX models can be found at URL
The repository contains example prediction scripts, described below.
The 'metricx23/URL' script contains an example for how to run inference
on the models.
### Reference-Based
Example usage for a reference-based model:
'URL' is expected to have 1 serialized JSON object per line with
'"reference"' and '"hypothesis"' fields. The output jsonl will be parallel
to 'URL' but additionally contain a '"prediction"' field with the predicted score.
Note that the model was trained with a maximum input length of 1024 tokens, so
significantly increasing that value may lead to unpredictable behavior.
### Reference-Free
Example usage for a reference-free model:
'URL' is expected to have 1 serialized JSON object per line with
'"source"' and '"hypothesis"' fields. The output jsonl will be parallel
to 'URL' but additionally contain a '"prediction"' field with the predicted score.
Meta-Evaluation
---------------
The 'metricx23/URL' script contains code to calculate various correlations
between the MetricX-23 scores and MQM ratings of translation quality using the
MT Metrics Eval library.
Example usage:
'URL' is expected to have one JSON object serialized per line.
Each JSON object is expected to contain 4 fields:
* '"system\_id"': The name of the system that generated the translation.
* '"segment\_id"': The 0-based index of the corresponding segment in the MT
Metrics Eval data.
* '"label"': The ground-truth translation quality score (with higher is better).
* '"prediction"': The model predicted translation quality score (with lower is
better; the script negates the scores so higher is better).
The script will calculate the 4 agreement/correlations that were used in the
WMT'23 Shared Task. Below are the results for the MetricX-23 models on the
WMT'22 Metrics Shared Task data:
English-German:
English-Russian:
Chinese-English:
The 'metricx23/evaluate\_wmt23.py' script re-calculates the average correlation
score that was used to rank submissions from the
WMT'23 Shared Task.
Example usage:
Each of the 3 input files is expected to be in the same format as described
above. Each file should correspond to running inference on each of the language
pairs from the WMT'23 dataset.
The results for each of the models is the following:
If you use MetricX-23 in your research, please cite the following publication:
| [
"### Synthetic Data\n\n\nIn order for our MetricX models to learn to identify certain types of bad\ntranslations that are not sufficiently (or at all) represented in the regular\ntraining data, we created synthetic examples and mixed them in during training.\nThe synthetic training data was generated from the DA datasets ranging from\nWMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have\nthe candidate translation manipulated so as to turn it into a bad translation\nwith a specific issue commonly unrecognized by learned metrics.\n\n\nThe table below provides an overview of the various failure modes that we\nconsidered, including brief descriptions of how we prepared the synthetic data\nto address them.\n\n\n\nExamples from the first 4 categories were assigned a label corresponding to the\nworst score on the given rating scale (e.g., 25 when mixed with MQM training\ndata), whereas the reference-matching translation examples are assigned the best\nscore (e.g., 0 when used with MQM data). The missing/incorrect punctuation\nexamples were labeled with a score slightly worse than perfect.\n\n\nNote that some of the synthetic datasets are only meaningful in the\nreference-based scenario, and we thus excluded them when training a QE variant\nof MetricX. These are the Latin-vs-special punctuation and the\nreference-matching translation examples.\n\n\nMost of the synthetic training sets were created using stratified sampling\nacross target languages, taking 500 examples per target language. One exception\nis the missing punctuation set, which used a stratified sample across different\npunctuation symbols instead.\n\n\nWhen training MetricX, a small proportion of the synthetic examples was mixed\nwith the regular training examples. During the first-stage fine-tuning on DA\ndata, each synthetic training set constituted between 0.1% and 1% of all\ntraining examples, whereas in the second-stage fine-tuning on MQM data we used\nan even smaller proportion, around 0.05%.\n\n\nAs for evaluating the effect of the synthetic training data on the model's\nperformance, the DEMETR challenge set - which we originally used to evaluate the\nmodels submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We\ntherefore created a new DEMETR-style test set based on the WMT22 DA data, with\nexamples constructed analogically to the synthetic training examples, as\ndescribed above. This test set helped us determine the right proportions of\nsynthetic data for fine-tuning in order to make MetricX robust for the failure\nmodes in consideration, without sacrificing the system- and segment-level\ncorrelations with human ratings.\n\n\nUsage\n-----\n\n\nThe code for using MetricX models can be found at URL\nThe repository contains example prediction scripts, described below.\n\n\nThe 'metricx23/URL' script contains an example for how to run inference\non the models.",
"### Reference-Based\n\n\nExample usage for a reference-based model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"reference\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nNote that the model was trained with a maximum input length of 1024 tokens, so\nsignificantly increasing that value may lead to unpredictable behavior.",
"### Reference-Free\n\n\nExample usage for a reference-free model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"source\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nMeta-Evaluation\n---------------\n\n\nThe 'metricx23/URL' script contains code to calculate various correlations\nbetween the MetricX-23 scores and MQM ratings of translation quality using the\nMT Metrics Eval library.\n\n\nExample usage:\n\n\n'URL' is expected to have one JSON object serialized per line.\nEach JSON object is expected to contain 4 fields:\n\n\n* '\"system\\_id\"': The name of the system that generated the translation.\n* '\"segment\\_id\"': The 0-based index of the corresponding segment in the MT\nMetrics Eval data.\n* '\"label\"': The ground-truth translation quality score (with higher is better).\n* '\"prediction\"': The model predicted translation quality score (with lower is\nbetter; the script negates the scores so higher is better).\n\n\nThe script will calculate the 4 agreement/correlations that were used in the\nWMT'23 Shared Task. Below are the results for the MetricX-23 models on the\nWMT'22 Metrics Shared Task data:\n\n\nEnglish-German:\n\n\n\nEnglish-Russian:\n\n\n\nChinese-English:\n\n\n\nThe 'metricx23/evaluate\\_wmt23.py' script re-calculates the average correlation\nscore that was used to rank submissions from the\nWMT'23 Shared Task.\n\n\nExample usage:\n\n\nEach of the 3 input files is expected to be in the same format as described\nabove. Each file should correspond to running inference on each of the language\npairs from the WMT'23 dataset.\n\n\nThe results for each of the models is the following:\n\n\n\nIf you use MetricX-23 in your research, please cite the following publication:"
] | [
"TAGS\n#transformers #pytorch #mt5 #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us \n",
"### Synthetic Data\n\n\nIn order for our MetricX models to learn to identify certain types of bad\ntranslations that are not sufficiently (or at all) represented in the regular\ntraining data, we created synthetic examples and mixed them in during training.\nThe synthetic training data was generated from the DA datasets ranging from\nWMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have\nthe candidate translation manipulated so as to turn it into a bad translation\nwith a specific issue commonly unrecognized by learned metrics.\n\n\nThe table below provides an overview of the various failure modes that we\nconsidered, including brief descriptions of how we prepared the synthetic data\nto address them.\n\n\n\nExamples from the first 4 categories were assigned a label corresponding to the\nworst score on the given rating scale (e.g., 25 when mixed with MQM training\ndata), whereas the reference-matching translation examples are assigned the best\nscore (e.g., 0 when used with MQM data). The missing/incorrect punctuation\nexamples were labeled with a score slightly worse than perfect.\n\n\nNote that some of the synthetic datasets are only meaningful in the\nreference-based scenario, and we thus excluded them when training a QE variant\nof MetricX. These are the Latin-vs-special punctuation and the\nreference-matching translation examples.\n\n\nMost of the synthetic training sets were created using stratified sampling\nacross target languages, taking 500 examples per target language. One exception\nis the missing punctuation set, which used a stratified sample across different\npunctuation symbols instead.\n\n\nWhen training MetricX, a small proportion of the synthetic examples was mixed\nwith the regular training examples. During the first-stage fine-tuning on DA\ndata, each synthetic training set constituted between 0.1% and 1% of all\ntraining examples, whereas in the second-stage fine-tuning on MQM data we used\nan even smaller proportion, around 0.05%.\n\n\nAs for evaluating the effect of the synthetic training data on the model's\nperformance, the DEMETR challenge set - which we originally used to evaluate the\nmodels submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We\ntherefore created a new DEMETR-style test set based on the WMT22 DA data, with\nexamples constructed analogically to the synthetic training examples, as\ndescribed above. This test set helped us determine the right proportions of\nsynthetic data for fine-tuning in order to make MetricX robust for the failure\nmodes in consideration, without sacrificing the system- and segment-level\ncorrelations with human ratings.\n\n\nUsage\n-----\n\n\nThe code for using MetricX models can be found at URL\nThe repository contains example prediction scripts, described below.\n\n\nThe 'metricx23/URL' script contains an example for how to run inference\non the models.",
"### Reference-Based\n\n\nExample usage for a reference-based model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"reference\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nNote that the model was trained with a maximum input length of 1024 tokens, so\nsignificantly increasing that value may lead to unpredictable behavior.",
"### Reference-Free\n\n\nExample usage for a reference-free model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"source\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nMeta-Evaluation\n---------------\n\n\nThe 'metricx23/URL' script contains code to calculate various correlations\nbetween the MetricX-23 scores and MQM ratings of translation quality using the\nMT Metrics Eval library.\n\n\nExample usage:\n\n\n'URL' is expected to have one JSON object serialized per line.\nEach JSON object is expected to contain 4 fields:\n\n\n* '\"system\\_id\"': The name of the system that generated the translation.\n* '\"segment\\_id\"': The 0-based index of the corresponding segment in the MT\nMetrics Eval data.\n* '\"label\"': The ground-truth translation quality score (with higher is better).\n* '\"prediction\"': The model predicted translation quality score (with lower is\nbetter; the script negates the scores so higher is better).\n\n\nThe script will calculate the 4 agreement/correlations that were used in the\nWMT'23 Shared Task. Below are the results for the MetricX-23 models on the\nWMT'22 Metrics Shared Task data:\n\n\nEnglish-German:\n\n\n\nEnglish-Russian:\n\n\n\nChinese-English:\n\n\n\nThe 'metricx23/evaluate\\_wmt23.py' script re-calculates the average correlation\nscore that was used to rank submissions from the\nWMT'23 Shared Task.\n\n\nExample usage:\n\n\nEach of the 3 input files is expected to be in the same format as described\nabove. Each file should correspond to running inference on each of the language\npairs from the WMT'23 dataset.\n\n\nThe results for each of the models is the following:\n\n\n\nIf you use MetricX-23 in your research, please cite the following publication:"
] | [
42,
666,
111,
457
] | [
"passage: TAGS\n#transformers #pytorch #mt5 #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us \n",
"passage: ### Synthetic Data\n\n\nIn order for our MetricX models to learn to identify certain types of bad\ntranslations that are not sufficiently (or at all) represented in the regular\ntraining data, we created synthetic examples and mixed them in during training.\nThe synthetic training data was generated from the DA datasets ranging from\nWMT15 to WMT21 (~ 43 language pairs). In most cases, the synthetic examples have\nthe candidate translation manipulated so as to turn it into a bad translation\nwith a specific issue commonly unrecognized by learned metrics.\n\n\nThe table below provides an overview of the various failure modes that we\nconsidered, including brief descriptions of how we prepared the synthetic data\nto address them.\n\n\n\nExamples from the first 4 categories were assigned a label corresponding to the\nworst score on the given rating scale (e.g., 25 when mixed with MQM training\ndata), whereas the reference-matching translation examples are assigned the best\nscore (e.g., 0 when used with MQM data). The missing/incorrect punctuation\nexamples were labeled with a score slightly worse than perfect.\n\n\nNote that some of the synthetic datasets are only meaningful in the\nreference-based scenario, and we thus excluded them when training a QE variant\nof MetricX. These are the Latin-vs-special punctuation and the\nreference-matching translation examples.\n\n\nMost of the synthetic training sets were created using stratified sampling\nacross target languages, taking 500 examples per target language. One exception\nis the missing punctuation set, which used a stratified sample across different\npunctuation symbols instead.\n\n\nWhen training MetricX, a small proportion of the synthetic examples was mixed\nwith the regular training examples. During the first-stage fine-tuning on DA\ndata, each synthetic training set constituted between 0.1% and 1% of all\ntraining examples, whereas in the second-stage fine-tuning on MQM data we used\nan even smaller proportion, around 0.05%.\n\n\nAs for evaluating the effect of the synthetic training data on the model's\nperformance, the DEMETR challenge set - which we originally used to evaluate the\nmodels submitted to the WMT23 Metrics Shared Task - was not adequate anymore. We\ntherefore created a new DEMETR-style test set based on the WMT22 DA data, with\nexamples constructed analogically to the synthetic training examples, as\ndescribed above. This test set helped us determine the right proportions of\nsynthetic data for fine-tuning in order to make MetricX robust for the failure\nmodes in consideration, without sacrificing the system- and segment-level\ncorrelations with human ratings.\n\n\nUsage\n-----\n\n\nThe code for using MetricX models can be found at URL\nThe repository contains example prediction scripts, described below.\n\n\nThe 'metricx23/URL' script contains an example for how to run inference\non the models.### Reference-Based\n\n\nExample usage for a reference-based model:\n\n\n'URL' is expected to have 1 serialized JSON object per line with\n'\"reference\"' and '\"hypothesis\"' fields. The output jsonl will be parallel\nto 'URL' but additionally contain a '\"prediction\"' field with the predicted score.\n\n\nNote that the model was trained with a maximum input length of 1024 tokens, so\nsignificantly increasing that value may lead to unpredictable behavior."
] | [
-0.08049466460943222,
-0.010928971692919731,
-0.004997325129806995,
0.02674533985555172,
0.10898007452487946,
0.017911424860358238,
0.030734051018953323,
0.10377532243728638,
-0.02402583509683609,
-0.03567541018128395,
0.03877764940261841,
0.08288230746984482,
0.015161662362515926,
0.022808298468589783,
0.01772439479827881,
-0.1944202184677124,
0.05506649613380432,
0.008813947439193726,
-0.07605475932359695,
0.06908859312534332,
0.11011916399002075,
-0.04383927211165428,
0.047993674874305725,
0.01802058517932892,
-0.046737369149923325,
0.03515304625034332,
0.02192578837275505,
-0.05999843776226044,
0.09158408641815186,
0.08443905413150787,
0.04427182674407959,
0.026185670867562294,
-0.03256708011031151,
-0.2039792835712433,
0.018693745136260986,
0.059222154319286346,
-0.04639671370387077,
0.032644130289554596,
0.06700465828180313,
-0.04408042132854462,
0.10856176912784576,
-0.03563929721713066,
-0.038342323154211044,
0.052534811198711395,
-0.08053379505872726,
-0.11484964191913605,
-0.06861758977174759,
0.005672447383403778,
0.05251080170273781,
0.11462049186229706,
0.016137007623910904,
0.08802412450313568,
-0.08221147954463959,
0.039690323173999786,
0.1186930239200592,
-0.24400970339775085,
0.022698264569044113,
0.031076578423380852,
0.0391852930188179,
0.059262461960315704,
-0.02663654275238514,
-0.011056511662900448,
0.04030945897102356,
0.036150794476270676,
0.04718852788209915,
-0.02858263999223709,
0.022121166810393333,
0.03753712773323059,
-0.11128073930740356,
-0.06610941141843796,
0.2194429337978363,
0.009360874071717262,
-0.039162129163742065,
-0.012831603176891804,
-0.040371835231781006,
-0.021052006632089615,
0.02562984824180603,
0.011642597615718842,
0.0126093290746212,
0.06509723514318466,
0.08087117969989777,
-0.06699725240468979,
-0.1469363272190094,
-0.02156376652419567,
-0.13039535284042358,
0.056314483284950256,
0.015906725078821182,
0.0635187178850174,
-0.13009536266326904,
0.06865113973617554,
-0.11303252726793289,
-0.037289686501026154,
-0.044206660240888596,
-0.09330593794584274,
0.027131132781505585,
0.0172458216547966,
-0.0671110600233078,
-0.061645977199077606,
0.01954670250415802,
0.09166590869426727,
-0.03881584480404854,
0.02113628201186657,
0.06317485123872757,
0.08239984512329102,
0.02081705816090107,
0.08037593215703964,
0.01793535053730011,
-0.011132435873150826,
0.08186338096857071,
-0.06666721403598785,
0.03507205471396446,
-0.00009720586240291595,
-0.05460785701870918,
-0.04377254843711853,
-0.02165292575955391,
0.07720592617988586,
0.024521537125110626,
0.024704361334443092,
-0.022436054423451424,
-0.014114882797002792,
0.15771594643592834,
-0.08077961206436157,
-0.031236806884407997,
-0.011595608666539192,
-0.05022282153367996,
0.047523126006126404,
0.06264422833919525,
0.006788937374949455,
-0.11120638251304626,
0.05265801399946213,
-0.03781677410006523,
-0.03154580295085907,
-0.08177147805690765,
-0.060364820063114166,
0.05910094082355499,
-0.0520191490650177,
0.022135227918624878,
-0.12485349923372269,
-0.14817503094673157,
-0.0194524135440588,
0.00905279628932476,
0.017627842724323273,
-0.04404270648956299,
0.030841557309031487,
-0.021325333043932915,
-0.010873598977923393,
-0.03838267922401428,
0.04409907013177872,
-0.0735226571559906,
0.01654006540775299,
-0.05891093611717224,
0.006638679653406143,
-0.12875565886497498,
0.05859975516796112,
-0.07795718312263489,
-0.008538316003978252,
-0.12053868919610977,
0.06858694553375244,
-0.06013733893632889,
0.06287765502929688,
-0.06418386101722717,
-0.059423450380563736,
-0.05324282497167587,
0.03673676773905754,
-0.011127032339572906,
0.08703723549842834,
-0.11388625204563141,
-0.04371771216392517,
0.06918473541736603,
-0.1147933155298233,
-0.10631323605775833,
0.05931851267814636,
-0.015358645468950272,
0.03748219460248947,
0.07437919080257416,
0.06889817118644714,
0.13736900687217712,
-0.022680290043354034,
-0.030354108661413193,
0.0774422138929367,
-0.09487991780042648,
-0.0831356942653656,
0.043639007955789566,
-0.02724086493253708,
-0.027338385581970215,
0.010795310139656067,
-0.05432993173599243,
0.053616154938936234,
0.004504588432610035,
-0.041023075580596924,
-0.03911237046122551,
-0.03953332453966141,
-0.03608177229762077,
-0.007416178472340107,
0.04547516256570816,
-0.030613455921411514,
-0.04433886334300041,
0.014297567307949066,
0.05688025429844856,
0.0016403227346017957,
0.06881103664636612,
-0.06005219742655754,
0.05407850816845894,
-0.018144575878977776,
0.008445029146969318,
-0.16996923089027405,
-0.050326526165008545,
0.03246239572763443,
-0.04993841052055359,
0.021022705361247063,
0.02538926526904106,
0.02754078060388565,
-0.014172155410051346,
-0.03300439566373825,
-0.01649932749569416,
0.008937809616327286,
0.008251310326159,
-0.05732375383377075,
-0.12281204760074615,
-0.0057578012347221375,
-0.030820120126008987,
-0.01729525998234749,
-0.027541520074009895,
0.012479400262236595,
0.07977741956710815,
0.044011034071445465,
0.009504541754722595,
0.03403853625059128,
0.005829425062984228,
0.01828007400035858,
-0.025431593880057335,
-0.01253400556743145,
0.06733429431915283,
0.029115518555045128,
-0.08058406412601471,
0.13161158561706543,
-0.1250324547290802,
0.017595291137695312,
0.13202296197414398,
-0.10783318430185318,
0.04871136322617531,
-0.04314408451318741,
-0.031379710882902145,
-0.0057036676444113255,
-0.03291696310043335,
-0.022622399032115936,
0.12076297402381897,
-0.004013901576399803,
0.0929318368434906,
-0.07880045473575592,
-0.05966495722532272,
-0.0097917215898633,
0.0038842959329485893,
-0.010753002017736435,
0.0713651031255722,
0.12100948393344879,
-0.12231792509555817,
0.08771493285894394,
0.13989409804344177,
-0.003570202738046646,
0.12119580805301666,
-0.040826037526130676,
-0.07305974513292313,
0.03446398302912712,
0.004951390903443098,
-0.01554139144718647,
0.04345749691128731,
-0.12445223331451416,
0.015536606311798096,
0.05309225618839264,
0.08432695269584656,
0.11454753577709198,
-0.11917590349912643,
0.0067930202931165695,
0.015571664087474346,
-0.03209373727440834,
-0.09838879108428955,
-0.0021094102412462234,
-0.01198030449450016,
0.07585512846708298,
-0.010641621425747871,
0.034573666751384735,
0.06810279190540314,
-0.033153194934129715,
-0.1333116739988327,
0.1549171507358551,
-0.16806340217590332,
-0.2015051692724228,
-0.1637459099292755,
0.0050877295434474945,
-0.04799898713827133,
0.007370667066425085,
0.10294272005558014,
-0.07140845060348511,
-0.04824806749820709,
-0.030684374272823334,
0.044100407510995865,
-0.026180272921919823,
-0.004666412249207497,
-0.05338970944285393,
0.042327139526605606,
0.016776274889707565,
-0.12135301530361176,
-0.006609104573726654,
-0.0036472089122980833,
-0.017405901104211807,
0.041763320565223694,
-0.07078999280929565,
0.07727597653865814,
0.08807438611984253,
0.01986617222428322,
-0.01348377950489521,
-0.04507536441087723,
0.21933329105377197,
-0.05761788785457611,
0.04443945735692978,
0.2112521529197693,
-0.011195655912160873,
0.05879829078912735,
0.052676692605018616,
-0.004360849969089031,
-0.02312907949090004,
0.02751847915351391,
0.013980694115161896,
-0.0917525514960289,
-0.284024178981781,
-0.06044050678610802,
-0.09147080034017563,
-0.06417541950941086,
0.0023470930755138397,
0.06487169861793518,
0.03722584992647171,
0.050735119730234146,
-0.04127505049109459,
-0.014369443990290165,
0.041575632989406586,
0.05546524375677109,
0.09428055584430695,
0.013206657022237778,
0.06720979511737823,
-0.11070599406957626,
0.012433091178536415,
0.11600261181592941,
0.09805083274841309,
0.23822110891342163,
-0.026172611862421036,
0.04857305809855461,
0.11312685161828995,
0.11436714231967926,
0.08747904002666473,
0.10965031385421753,
-0.04313330352306366,
0.029491525143384933,
-0.041477788239717484,
-0.02848794311285019,
-0.03056623786687851,
0.08165007829666138,
-0.021050430834293365,
-0.0669928565621376,
-0.06534460186958313,
-0.021561436355113983,
0.06433352082967758,
0.21831148862838745,
-0.005303980782628059,
-0.07421030104160309,
-0.05741162598133087,
0.06067470833659172,
-0.010603219270706177,
-0.03260648995637894,
0.07761399447917938,
0.04806586354970932,
-0.07070834934711456,
0.08121111243963242,
-0.0008651819080114365,
0.11229313164949417,
0.014927586540579796,
0.018110742792487144,
0.03192968666553497,
-0.006562814116477966,
0.018823236227035522,
0.11194905638694763,
-0.26115158200263977,
0.12093576788902283,
-0.011933675035834312,
0.05421082302927971,
-0.0762992650270462,
0.022125624120235443,
0.04246993362903595,
0.1273622065782547,
0.09262282401323318,
0.043345239013433456,
-0.14447379112243652,
-0.0103498175740242,
0.0019879117608070374,
0.056262798607349396,
0.04705921933054924,
0.010509418323636055,
0.011808192357420921,
-0.011274687014520168,
0.013288512825965881,
0.015993541106581688,
0.08853989839553833,
-0.0517207607626915,
-0.16778728365898132,
0.04126080125570297,
-0.006679844111204147,
0.015364071354269981,
-0.07193623483181,
0.005432065576314926,
-0.039848387241363525,
0.08026738464832306,
-0.14921778440475464,
-0.0995694100856781,
-0.09588959813117981,
-0.08473988622426987,
0.08787038922309875,
-0.0654374435544014,
0.06406596302986145,
0.007063437253236771,
0.07009734213352203,
-0.03068877011537552,
-0.14400194585323334,
0.057342350482940674,
-0.11315649747848511,
-0.08588340878486633,
-0.0034783631563186646,
0.1125456690788269,
0.013006072491407394,
0.0326993428170681,
0.019293401390314102,
-0.010193015448749065,
-0.046051137149333954,
-0.11050929129123688,
0.01819794625043869,
0.07997186481952667,
0.05762843042612076,
0.05681586265563965,
-0.09417805820703506,
-0.029595911502838135,
-0.04396835342049599,
-0.06783974170684814,
0.1667174994945526,
0.10756810754537582,
-0.06620179116725922,
0.1170005351305008,
0.16132885217666626,
-0.1295406073331833,
-0.19067947566509247,
-0.043041959404945374,
-0.03716177120804787,
-0.01986772194504738,
-0.02807173877954483,
-0.1475318968296051,
0.07539469003677368,
0.09812235087156296,
-0.0024198442697525024,
0.04476727545261383,
-0.2080758810043335,
-0.08643566817045212,
0.04781382530927658,
0.0001132916659116745,
0.18663114309310913,
-0.10906235128641129,
-0.044622793793678284,
-0.052799079567193985,
-0.11352433264255524,
0.14947330951690674,
-0.06072360277175903,
0.05728284642100334,
0.00732648279517889,
0.0032600341364741325,
-0.006010896060615778,
-0.03202845901250839,
0.13251455128192902,
0.03674418851733208,
0.043941304087638855,
-0.028950639069080353,
0.055670365691185,
0.025221113115549088,
-0.04210628941655159,
0.044331781566143036,
-0.04143795743584633,
0.021573659032583237,
-0.14452756941318512,
-0.045783452689647675,
-0.08391499519348145,
0.04536336287856102,
-0.01962660811841488,
-0.039184920489788055,
-0.042702145874500275,
0.0021056830883026123,
0.08968086540699005,
-0.014609623700380325,
0.054970160126686096,
-0.05837482213973999,
0.059537388384342194,
0.07670764625072479,
0.15645602345466614,
-0.10332942008972168,
-0.1217908263206482,
-0.02813919633626938,
-0.021626751869916916,
0.08217503130435944,
-0.1376548409461975,
0.07488036155700684,
0.11647255718708038,
-0.04862956330180168,
0.08627215772867203,
0.055094070732593536,
-0.07522906363010406,
0.005458813160657883,
0.10446281731128693,
-0.07246147096157074,
-0.06008515506982803,
-0.0390506386756897,
0.08281482756137848,
0.029428621754050255,
0.03669393062591553,
0.12870027124881744,
-0.0009662613738328218,
-0.032006535679101944,
-0.02065277099609375,
0.06402802467346191,
-0.0347125381231308,
0.0842379555106163,
0.039138346910476685,
0.017695903778076172,
-0.08644309639930725,
0.1338755190372467,
0.01591717265546322,
-0.1885780692100525,
-0.009080829098820686,
0.11049890518188477,
-0.11632084846496582,
-0.0807068943977356,
-0.025578761473298073,
0.009648185223340988,
-0.14496804773807526,
-0.10269727557897568,
-0.032402172684669495,
-0.14044955372810364,
0.06719717383384705,
0.16770902276039124,
0.02116461656987667,
0.03302031010389328,
-0.006035410333424807,
-0.028474289923906326,
-0.03122306987643242,
-0.013931620866060257,
-0.0277186818420887,
0.017339259386062622,
-0.05742930248379707,
0.0389898419380188,
0.019052214920520782,
0.06222548335790634,
-0.016825642436742783,
0.015219002030789852,
-0.0635049119591713,
0.010224338620901108,
-0.15755128860473633,
0.024624407291412354,
-0.05316181108355522,
-0.017552291974425316,
-0.01309568714350462,
-0.014190688729286194,
-0.05401631444692612,
0.05547712743282318,
-0.09220357239246368,
-0.020396772772073746,
-0.03640173375606537,
0.027906954288482666,
-0.05868091061711311,
-0.02378096804022789,
0.050178419798612595,
-0.047896455973386765,
0.1031966358423233,
0.06362350285053253,
-0.05172618851065636,
0.0687813088297844,
-0.054818395525217056,
-0.03117077425122261,
0.04752342030405998,
0.04614630341529846,
0.0024338315706700087,
-0.06958136707544327,
0.053022030740976334,
0.0532219260931015,
-0.05545207858085632,
-0.00179347675293684,
-0.00773459579795599,
-0.106658935546875,
-0.026197271421551704,
0.026790166273713112,
-0.007194001227617264,
-0.06978121399879456,
-0.046734243631362915,
0.11590144783258438,
0.04049905017018318,
0.1221245527267456,
-0.006192055530846119,
0.039656929671764374,
-0.0669143944978714,
0.0076656281016767025,
-0.0022521058563143015,
-0.0823831707239151,
-0.09226924180984497,
-0.03779274597764015,
0.01616818644106388,
-0.032287146896123886,
0.20258086919784546,
-0.027631420642137527,
-0.001419171690940857,
0.05651244521141052,
0.0006612688302993774,
-0.019393347203731537,
0.012330395169556141,
0.1308664083480835,
0.009836757555603981,
0.00910251121968031,
-0.02165895141661167,
0.024475062265992165,
-0.016778085380792618,
-0.0041928887367248535,
0.13951833546161652,
0.09686414152383804,
0.16705356538295746,
0.1026294082403183,
0.002041283994913101,
-0.03730851784348488,
-0.01945573464035988,
-0.06291651725769043,
0.028470784425735474,
0.058277349919080734,
0.00701283710077405,
0.01671595126390457,
0.11799974739551544,
-0.09208066761493683,
0.09887351095676422,
-0.06913766264915466,
-0.05150070786476135,
-0.1154223382472992,
-0.09975318610668182,
-0.010319891385734081,
-0.08603835850954056,
-0.029209984466433525,
-0.13546273112297058,
-0.020355189219117165,
0.06884671002626419,
0.038902945816516876,
-0.05183970928192139,
0.007247757166624069,
-0.11964748054742813,
-0.10397026687860489,
0.04866647720336914,
-0.014942724257707596,
0.04801199212670326,
0.04992463439702988,
0.012288471683859825,
0.04020651429891586,
-0.03871965408325195,
-0.02611941285431385,
0.039849359542131424,
0.04274921864271164,
0.045678697526454926,
-0.03452969342470169,
-0.010724259540438652,
-0.028795840218663216,
0.012287003919482231,
0.02913379855453968,
0.22296985983848572,
0.010857714340090752,
-0.031127674505114555,
0.06575135886669159,
0.19841772317886353,
-0.04081185162067413,
-0.09934508055448532,
-0.09630105644464493,
0.1641465723514557,
0.045690394937992096,
0.029041193425655365,
0.016438156366348267,
-0.057314880192279816,
-0.00565880723297596,
0.22985684871673584,
0.17313337326049805,
-0.025465190410614014,
-0.009001774713397026,
-0.006234009750187397,
0.014965811744332314,
0.06765225529670715,
0.08408764004707336,
0.1142108142375946,
0.21731261909008026,
-0.06078370288014412,
0.0639389306306839,
-0.0446958988904953,
0.0698341652750969,
-0.034699127078056335,
0.17769397795200348,
0.001534847542643547,
-0.058601561933755875,
-0.0005737040191888809,
0.06745973229408264,
-0.07404687255620956,
-0.10022060573101044,
-0.03632500767707825,
-0.07546216249465942,
-0.05532532557845116,
-0.0036639878526329994,
0.05928308516740799,
0.022997058928012848,
0.041565168648958206,
-0.0059436894953250885,
-0.044010479003190994,
0.10706667602062225,
0.023562224581837654,
-0.17537671327590942,
-0.05920793116092682,
0.08894800394773483,
0.011057954281568527,
0.08248458802700043,
0.007951926440000534,
0.048754554241895676,
0.07727794349193573,
0.03654977306723595,
-0.11864081025123596,
0.07930269092321396,
-0.00997848343104124,
-0.05712908133864403,
0.007062149234116077,
0.014758620411157608,
-0.00663449801504612,
0.020782694220542908,
0.029319513589143753,
-0.03974604234099388,
0.05011465772986412,
0.03723395615816116,
-0.04944480583071709,
-0.00523018091917038,
0.03254415839910507,
-0.10453616082668304,
0.08378694206476212,
0.07704398781061172,
-0.022326018661260605,
-0.027006199583411217,
-0.05415914207696915,
0.02430926077067852,
0.012829716317355633,
-0.023813046514987946,
-0.024040259420871735,
-0.06499078124761581,
0.017759330570697784,
0.01358121633529663,
0.03267041966319084,
-0.13761897385120392,
-0.03263425827026367,
0.009745022282004356,
-0.028417814522981644,
-0.08249472081661224,
0.06735780090093613,
0.043854910880327225,
0.026215003803372383,
-0.03867888078093529,
0.0033451877534389496,
-0.004438048228621483,
0.04008375108242035,
-0.0645875409245491,
-0.08099328726530075
] |
null | null | diffusers |
# DreamBooth - mustafakara/duck
This is a dreambooth model derived from CompVis/stable-diffusion-v1-4. The weights were trained on a photo of rsu monster toy using [DreamBooth](https://dreambooth.github.io/).
You can find some example images in the following.
DreamBooth for the text encoder was enabled: False.
| {"license": "creativeml-openrail-m", "tags": ["stable-diffusion", "stable-diffusion-diffusers", "text-to-image", "diffusers", "dreambooth"], "base_model": "CompVis/stable-diffusion-v1-4", "instance_prompt": "a photo of rsu monster toy", "inference": true} | text-to-image | mustafakara/duck | [
"diffusers",
"tensorboard",
"safetensors",
"stable-diffusion",
"stable-diffusion-diffusers",
"text-to-image",
"dreambooth",
"base_model:CompVis/stable-diffusion-v1-4",
"license:creativeml-openrail-m",
"endpoints_compatible",
"diffusers:StableDiffusionPipeline",
"region:us"
] | 2024-02-07T16:37:08+00:00 | [] | [] | TAGS
#diffusers #tensorboard #safetensors #stable-diffusion #stable-diffusion-diffusers #text-to-image #dreambooth #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us
|
# DreamBooth - mustafakara/duck
This is a dreambooth model derived from CompVis/stable-diffusion-v1-4. The weights were trained on a photo of rsu monster toy using DreamBooth.
You can find some example images in the following.
DreamBooth for the text encoder was enabled: False.
| [
"# DreamBooth - mustafakara/duck\n\nThis is a dreambooth model derived from CompVis/stable-diffusion-v1-4. The weights were trained on a photo of rsu monster toy using DreamBooth.\nYou can find some example images in the following. \n\n\n\nDreamBooth for the text encoder was enabled: False."
] | [
"TAGS\n#diffusers #tensorboard #safetensors #stable-diffusion #stable-diffusion-diffusers #text-to-image #dreambooth #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us \n",
"# DreamBooth - mustafakara/duck\n\nThis is a dreambooth model derived from CompVis/stable-diffusion-v1-4. The weights were trained on a photo of rsu monster toy using DreamBooth.\nYou can find some example images in the following. \n\n\n\nDreamBooth for the text encoder was enabled: False."
] | [
97,
78
] | [
"passage: TAGS\n#diffusers #tensorboard #safetensors #stable-diffusion #stable-diffusion-diffusers #text-to-image #dreambooth #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us \n# DreamBooth - mustafakara/duck\n\nThis is a dreambooth model derived from CompVis/stable-diffusion-v1-4. The weights were trained on a photo of rsu monster toy using DreamBooth.\nYou can find some example images in the following. \n\n\n\nDreamBooth for the text encoder was enabled: False."
] | [
-0.016848092898726463,
0.07732637226581573,
-0.0029867959674447775,
0.05986306816339493,
0.1297970414161682,
-0.008180400356650352,
0.10374586284160614,
0.003095132764428854,
-0.045823343098163605,
0.04523496702313423,
0.10247226059436798,
-0.016043057665228844,
0.034615203738212585,
0.09291303902864456,
0.044788505882024765,
-0.1972362995147705,
0.05109608918428421,
0.005485584493726492,
-0.0779714286327362,
0.04412805661559105,
0.03279309719800949,
-0.0661606565117836,
0.12367167323827744,
0.011217644438147545,
-0.17045196890830994,
-0.012493486516177654,
-0.025650428608059883,
-0.004274206236004829,
0.0791616290807724,
0.018406182527542114,
0.14658859372138977,
0.040980949997901917,
0.045259200036525726,
-0.11242329329252243,
0.039882756769657135,
0.0410088449716568,
-0.008804394863545895,
0.0487603060901165,
-0.017803752794861794,
0.01359639223664999,
0.04691983386874199,
0.017663460224866867,
0.05714977905154228,
0.07750322669744492,
-0.04806278645992279,
-0.02120893821120262,
0.04783930256962776,
0.08728673309087753,
0.10217709094285965,
0.049067482352256775,
-0.023101894184947014,
0.038581039756536484,
0.08363279700279236,
0.10033917427062988,
0.23160278797149658,
-0.16037224233150482,
-0.10128021240234375,
0.2545238137245178,
0.06833335012197495,
-0.020106075331568718,
-0.03473943471908569,
0.060672931373119354,
0.06540759652853012,
0.06260722875595093,
0.016576334834098816,
-0.0836653858423233,
0.06031731143593788,
-0.15375015139579773,
-0.0831744596362114,
0.06467484682798386,
0.00010067440598504618,
-0.03360694646835327,
-0.05065718665719032,
-0.1455385982990265,
-0.06948019564151764,
0.08612773567438126,
-0.030610758811235428,
-0.012980591505765915,
-0.016372477635741234,
0.002736551221460104,
-0.03135766461491585,
-0.027811488136649132,
-0.06769642978906631,
-0.07387089729309082,
0.03201870247721672,
0.01259428821504116,
-0.02098107524216175,
0.02262275107204914,
-0.06341444700956345,
0.10532126575708389,
-0.06112653389573097,
-0.13801437616348267,
0.02767033874988556,
-0.10899197310209274,
0.014515833929181099,
0.08047671616077423,
-0.04787582904100418,
-0.24168381094932556,
0.10325674712657928,
-0.07276545464992523,
0.027835732325911522,
0.011308667249977589,
-0.0030487135518342257,
0.0595528781414032,
0.03404640778899193,
0.004033102188259363,
-0.06678345799446106,
-0.04109690710902214,
0.008766061626374722,
0.0518857017159462,
0.04460913687944412,
-0.050589609891176224,
-0.09605857729911804,
0.05461294949054718,
-0.05140096694231033,
0.025504175573587418,
0.001422044006176293,
0.014011727645993233,
-0.0703425407409668,
-0.012122844345867634,
0.10771552473306656,
-0.0030545478221029043,
-0.005517392419278622,
-0.010570921935141087,
0.03640388697385788,
-0.014755524694919586,
0.15015855431556702,
0.01604508049786091,
-0.017804590985178947,
0.04683235287666321,
-0.09419914335012436,
0.015629056841135025,
-0.011589002795517445,
-0.08254390954971313,
-0.026406917721033096,
-0.16220644116401672,
0.01723041944205761,
-0.15580877661705017,
-0.06903625279664993,
-0.010424989275634289,
0.019443519413471222,
-0.011948919855058193,
0.027631517499685287,
-0.08048391342163086,
-0.1401362121105194,
-0.052753545343875885,
0.08329188823699951,
0.006779284682124853,
0.03685779497027397,
0.03835972398519516,
-0.03140929341316223,
0.10003698617219925,
-0.03472121059894562,
-0.04587160050868988,
-0.11431074142456055,
0.04410015046596527,
-0.09980732202529907,
0.10517246276140213,
-0.053219038993120193,
0.11193651705980301,
-0.034937534481287,
-0.04905570670962334,
0.004989512264728546,
0.025278475135564804,
0.005892224609851837,
0.1453162133693695,
-0.20242516696453094,
-0.055649884045124054,
0.15265701711177826,
-0.18649350106716156,
-0.12436997890472412,
0.05088351294398308,
-0.005775252357125282,
0.089248888194561,
0.04361202195286751,
0.08191432058811188,
0.06712839752435684,
-0.19381296634674072,
-0.026799052953720093,
-0.0931769534945488,
-0.07906176149845123,
0.06582582741975784,
-0.026433896273374557,
0.0764365941286087,
-0.03205517679452896,
0.0206486526876688,
-0.067408986389637,
0.09226281195878983,
-0.024521231651306152,
-0.018608305603265762,
-0.022801654413342476,
-0.09283322840929031,
0.045912664383649826,
0.008467626757919788,
0.02953721210360527,
-0.005572664085775614,
-0.030746446922421455,
0.008960937149822712,
-0.026931311935186386,
-0.04538304731249809,
-0.017768582329154015,
-0.04552889242768288,
0.049133844673633575,
-0.05015390366315842,
-0.00006472138193203136,
-0.12569567561149597,
-0.0661233589053154,
0.03507319092750549,
0.05654081329703331,
0.03695698082447052,
-0.002731392625719309,
0.08997521549463272,
0.13160419464111328,
0.016985319554805756,
-0.04685657471418381,
0.03961127623915672,
0.0444314107298851,
-0.008655666373670101,
-0.1053532212972641,
0.05457242578268051,
-0.08726809173822403,
-0.07718797773122787,
-0.11646117269992828,
0.07766567915678024,
0.038881268352270126,
0.25519078969955444,
0.0928206667304039,
-0.057509418576955795,
0.09159209579229355,
0.028027795255184174,
-0.019606199115514755,
-0.10235320776700974,
0.023747073486447334,
0.0051514809019863605,
-0.09817393124103546,
0.10873278975486755,
-0.14570178091526031,
0.09694773703813553,
0.11665070056915283,
0.09723199903964996,
-0.09364022314548492,
0.1169823557138443,
0.0026911627501249313,
0.007158753462135792,
-0.08507183194160461,
0.008695202879607677,
0.22277800738811493,
0.031669341027736664,
0.16942228376865387,
-0.029871866106987,
0.04145125299692154,
0.10814368724822998,
-0.041530586779117584,
-0.08053731918334961,
0.07062791287899017,
-0.08562074601650238,
-0.028527067974209785,
0.040492068976163864,
0.035678211599588394,
0.03952182084321976,
0.1952550709247589,
-0.01613931730389595,
0.024267151951789856,
-0.04490111768245697,
-0.024421585723757744,
0.02290481887757778,
0.15937523543834686,
-0.013156207278370857,
-0.02197244204580784,
-0.010151805356144905,
-0.07192292809486389,
-0.0044661276042461395,
-0.15202350914478302,
-0.020210476592183113,
0.042740266770124435,
-0.002604119013994932,
0.14306055009365082,
0.04921956732869148,
-0.0939338281750679,
0.02290363609790802,
-0.10766711086034775,
-0.07040737569332123,
-0.004349493887275457,
-0.03150489926338196,
-0.06490664184093475,
0.1373997926712036,
-0.09278054535388947,
-0.31245917081832886,
-0.1520344465970993,
0.040174275636672974,
-0.023660289123654366,
0.017584003508090973,
0.06389770656824112,
-0.10480441898107529,
-0.042089805006980896,
-0.11491334438323975,
0.07292192429304123,
0.0640702024102211,
0.04054222255945206,
0.06452157348394394,
0.025637609884142876,
0.014910691417753696,
-0.04642613232135773,
0.019143670797348022,
-0.03459310159087181,
0.030681980773806572,
0.06828319281339645,
0.0018342983676120639,
0.09433433413505554,
0.14132462441921234,
-0.008176993578672409,
-0.018314644694328308,
0.020379072055220604,
0.24808719754219055,
-0.006899423897266388,
0.08602821826934814,
0.1430688351392746,
0.010112962685525417,
0.07636833935976028,
0.17401225864887238,
0.030647972598671913,
-0.04216713458299637,
0.08573609590530396,
-0.059181347489356995,
-0.12901704013347626,
-0.00951792299747467,
-0.04823766276240349,
0.015478050336241722,
0.02872161753475666,
0.08190956711769104,
0.03689621388912201,
0.046419329941272736,
0.12505581974983215,
0.10093650966882706,
0.03326723352074623,
0.06479035317897797,
0.07147415727376938,
0.0066865477710962296,
-0.07208805531263351,
0.029475286602973938,
-0.053542375564575195,
-0.10322509706020355,
0.02701997198164463,
-0.08997468650341034,
0.10132958739995956,
-0.07206211239099503,
-0.09151912480592728,
0.024239609017968178,
-0.04122016206383705,
0.11136890202760696,
0.06849803775548935,
-0.04329754412174225,
-0.04396219551563263,
-0.024214794859290123,
-0.0864587277173996,
0.014827404171228409,
0.12644995748996735,
-0.01615576632320881,
0.0359509214758873,
-0.006111519876867533,
0.13491475582122803,
-0.011244328692555428,
0.0778893381357193,
0.11020734161138535,
-0.24326258897781372,
-0.05798975005745888,
-0.016251476481556892,
0.05134870857000351,
-0.049882564693689346,
0.028400886803865433,
0.3237946629524231,
-0.03125738725066185,
0.03563808277249336,
-0.09088657796382904,
0.03928317874670029,
0.09832820296287537,
0.01862616091966629,
-0.08022753149271011,
0.07314636558294296,
-0.037387341260910034,
-0.05258650332689285,
-0.1509789377450943,
0.013647561892867088,
-0.040412839502096176,
0.05407633259892464,
0.015536249615252018,
0.02462317794561386,
0.04036516696214676,
0.0975472629070282,
0.14231523871421814,
0.005891942419111729,
-0.027653297409415245,
-0.08637069910764694,
-0.11028841882944107,
0.015698209404945374,
0.02791559509932995,
-0.03183545544743538,
0.02375633828341961,
0.07996603846549988,
-0.041956979781389236,
-0.004583071917295456,
0.01048046350479126,
-0.12900970876216888,
-0.049580205231904984,
-0.022291414439678192,
0.09621786326169968,
0.06411796808242798,
-0.08026836067438126,
-0.0763002559542656,
0.09664932638406754,
0.0801108181476593,
-0.2081853151321411,
-0.11806107312440872,
-0.053234927356243134,
-0.09017284959554672,
0.05738312751054764,
-0.02975170873105526,
0.07773540169000626,
-0.06438416987657547,
0.1535917967557907,
-0.13403579592704773,
-0.09570961445569992,
0.05934137850999832,
-0.09869545698165894,
-0.17119789123535156,
-0.1327728033065796,
0.03470008820295334,
0.004410176072269678,
0.003701256588101387,
-0.0008098303223960102,
0.02704579383134842,
0.0018178216414526105,
-0.0435715988278389,
0.035203296691179276,
0.1823282092809677,
-0.11628393828868866,
0.05970924347639084,
0.017756834626197815,
-0.13955740630626678,
-0.10048386454582214,
0.06351516395807266,
0.06724744290113449,
0.19684310257434845,
-0.07217458635568619,
0.07616382837295532,
0.1076442152261734,
-0.057696450501680374,
-0.22602318227291107,
-0.1323460042476654,
0.04540173336863518,
0.015589228831231594,
0.009506655856966972,
-0.12037297338247299,
0.1821327805519104,
-0.03871222957968712,
-0.006466047838330269,
0.14425089955329895,
-0.31010931730270386,
-0.11570817232131958,
0.02747293747961521,
0.18402725458145142,
0.2090413123369217,
-0.12433411926031113,
-0.021967655047774315,
0.03346959874033928,
-0.11710863560438156,
0.20499315857887268,
-0.0238841213285923,
0.055248938500881195,
-0.0012803658610209823,
0.021347053349018097,
0.015283804386854172,
-0.039558690041303635,
0.06504256278276443,
-0.02764449641108513,
0.015787413343787193,
-0.046051543205976486,
-0.03864682465791702,
0.08568909019231796,
-0.03171773627400398,
0.014720160514116287,
-0.04963821917772293,
0.04379463568329811,
-0.013776247389614582,
0.002527260920032859,
-0.022094687446951866,
0.013056857511401176,
-0.015017298981547356,
-0.1222086250782013,
-0.059614673256874084,
0.046635132282972336,
0.03932362422347069,
0.007167390082031488,
-0.0277792327105999,
-0.025096260011196136,
0.03552043065428734,
0.20086705684661865,
-0.04504314437508583,
-0.04052756354212761,
-0.10303715616464615,
-0.04055628180503845,
-0.07045438885688782,
0.11101504415273666,
-0.0925922840833664,
-0.0276607908308506,
0.12714295089244843,
0.0622885562479496,
0.09085427224636078,
0.044933781027793884,
-0.0765291377902031,
0.07660729438066483,
0.08587639778852463,
-0.08997220546007156,
-0.08731617033481598,
-0.028989413753151894,
0.028046684339642525,
0.04622519016265869,
-0.0614769421517849,
0.16668911278247833,
-0.1314605474472046,
0.013750246725976467,
-0.024511007592082024,
0.0493900291621685,
0.004563299007713795,
0.05687924101948738,
0.02676895260810852,
0.06863055378198624,
-0.05439421162009239,
0.04567115008831024,
-0.05264198035001755,
-0.03918619081377983,
0.03881115838885307,
0.06924401968717575,
-0.10391975939273834,
-0.003415108425542712,
-0.05366441607475281,
0.1683056354522705,
-0.08175365626811981,
-0.013059412129223347,
-0.10837668925523758,
-0.13520176708698273,
-0.007103706244379282,
0.13019444048404694,
0.033883921802043915,
0.0338287428021431,
-0.02412625588476658,
-0.05945168435573578,
-0.06067493557929993,
0.06278927624225616,
0.02730346843600273,
0.07993481308221817,
-0.2053411304950714,
-0.003093170700594783,
0.06719513237476349,
0.003499552607536316,
-0.08615952730178833,
-0.01023923046886921,
-0.07076702266931534,
-0.03818688914179802,
-0.006335511337965727,
0.15281358361244202,
-0.08966096490621567,
-0.03972889110445976,
-0.007413058076053858,
-0.011417419649660587,
-0.011407422833144665,
0.053153958171606064,
-0.008649523369967937,
-0.01912117749452591,
-0.014776933006942272,
-0.04165341705083847,
-0.05977931246161461,
-0.061109937727451324,
0.011018129996955395,
-0.06931155920028687,
0.00548550533130765,
-0.09282092750072479,
-0.14002861082553864,
0.029807033017277718,
-0.2873762249946594,
-0.0032750361133366823,
0.14345578849315643,
-0.03316524252295494,
-0.012646201066672802,
0.013088198378682137,
-0.029285704717040062,
-0.04149137809872627,
0.01843675971031189,
-0.0005460764514282346,
0.0796373263001442,
-0.08358793705701828,
-0.032183993607759476,
-0.0002706493833102286,
0.05725669860839844,
-0.04759863764047623,
-0.024718711152672768,
0.11562251299619675,
0.09740475565195084,
0.15819182991981506,
-0.1534440815448761,
0.1499546468257904,
-0.09875164180994034,
-0.020040050148963928,
0.0048083350993692875,
-0.06377051770687103,
0.03848513960838318,
-0.0028991629369556904,
-0.040357641875743866,
-0.025678442791104317,
0.07345037907361984,
0.020234914496541023,
-0.0936468318104744,
0.013453319668769836,
-0.018113676458597183,
-0.007924213074147701,
0.06799835711717606,
0.1808679699897766,
0.007960006594657898,
-0.00414264015853405,
-0.11561879515647888,
0.06358464807271957,
0.14198921620845795,
0.1274683177471161,
0.0521477572619915,
-0.014520674012601376,
0.05880379304289818,
0.10727820545434952,
0.07265020906925201,
0.1411549597978592,
0.01456573698669672,
0.12498459964990616,
-0.08039674162864685,
0.11035384237766266,
-0.0377814956009388,
-0.030258269980549812,
0.08795050531625748,
-0.018091078847646713,
-0.05315808951854706,
0.056105077266693115,
-0.024051303043961525,
-0.01474879588931799,
0.010840652510523796,
-0.051335640251636505,
-0.10423846542835236,
0.012686450965702534,
-0.06492915004491806,
-0.09753318130970001,
0.012483312748372555,
-0.0025754161179065704,
0.035342633724212646,
0.14975711703300476,
0.043731629848480225,
0.02253447100520134,
0.11410633474588394,
0.0031505743972957134,
-0.08676587045192719,
0.10458943247795105,
0.011374454945325851,
0.06919684261083603,
0.0632171779870987,
-0.01846306212246418,
0.07042595744132996,
-0.022420722991228104,
0.018543832004070282,
0.02043391950428486,
-0.09591524302959442,
-0.0027740225195884705,
0.003949819132685661,
-0.08410673588514328,
0.06229012832045555,
0.13740280270576477,
-0.05431597679853439,
-0.03529895842075348,
0.07728344947099686,
-0.014544508419930935,
-0.06499968469142914,
-0.16310356557369232,
0.1218947172164917,
-0.12209998816251755,
0.0486428439617157,
-0.03695652261376381,
-0.106108158826828,
-0.07795634865760803,
0.12223746627569199,
0.13221555948257446,
-0.05084472522139549,
0.025299491360783577,
-0.07978108525276184,
-0.0005507302703335881,
-0.08690530061721802,
0.07913078367710114,
0.00554125988855958,
0.2440747767686844,
-0.06553584337234497,
0.024812716990709305,
-0.08839084208011627,
-0.1444178968667984,
-0.07989086210727692,
-0.2459752857685089,
0.0621824637055397,
0.0012793843634426594,
-0.08234404772520065,
0.07227634638547897,
-0.24859119951725006,
-0.09625987708568573,
0.1977764517068863,
-0.1352604627609253,
0.0014846023404970765,
-0.04126016050577164,
0.09243400394916534,
0.050342410802841187,
0.08514776825904846,
-0.03758855164051056,
0.028704075142741203,
0.11570392549037933,
-0.05086379498243332,
-0.09782077372074127,
0.049482882022857666,
-0.12017647176980972,
-0.2882535755634308,
0.19958873093128204,
-0.01750144548714161,
-0.032168999314308167,
0.034345123916864395,
-0.01497395895421505,
-0.08031610399484634,
0.041528210043907166,
-0.05105840787291527,
0.02302580513060093,
-0.0842040628194809,
0.18497829139232635,
0.008092289790511131,
0.05225646123290062,
0.010774603113532066,
-0.12354771047830582,
0.0022123863454908133,
0.04376606643199921,
-0.0031898345332592726,
-0.06987105309963226,
0.02078039012849331,
-0.027150338515639305,
0.07001940160989761,
0.08305739611387253,
-0.04141424968838692,
0.03778103366494179,
0.034425053745508194,
0.02179715409874916,
-0.0052713509649038315,
0.015195260755717754,
0.053341444581747055,
-0.10002252459526062,
-0.01589285582304001,
-0.008393602445721626,
-0.02232828177511692,
-0.305614709854126,
-0.08729568123817444,
-0.17380569875240326,
0.003164602443575859,
-0.002604131819680333,
0.050663307309150696,
0.15034113824367523,
0.036120589822530746,
0.030439216643571854,
-0.013055732473731041,
0.021242471411824226,
0.08435393869876862,
0.030975082889199257,
-0.124266117811203
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# spatio_temporal_vit-finetuned-ucf101-subset
This model is a fine-tuned version of [Tommidi/st_vit_untrained](https://huggingface.co/Tommidi/st_vit_untrained) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.1244
- Accuracy: 0.9
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- training_steps: 37
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.6013 | 1.0 | 37 | 0.1244 | 0.9 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "Tommidi/st_vit_untrained", "model-index": [{"name": "spatio_temporal_vit-finetuned-ucf101-subset", "results": []}]} | null | Tommidi/spatio_temporal_vit-finetuned-ucf101-subset | [
"transformers",
"tensorboard",
"safetensors",
"st_vit",
"generated_from_trainer",
"base_model:Tommidi/st_vit_untrained",
"endpoints_compatible",
"region:us"
] | 2024-02-07T16:39:37+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #st_vit #generated_from_trainer #base_model-Tommidi/st_vit_untrained #endpoints_compatible #region-us
| spatio\_temporal\_vit-finetuned-ucf101-subset
=============================================
This model is a fine-tuned version of Tommidi/st\_vit\_untrained on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 0.1244
* Accuracy: 0.9
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* training\_steps: 37
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 37",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #st_vit #generated_from_trainer #base_model-Tommidi/st_vit_untrained #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 37",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
52,
115,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #st_vit #generated_from_trainer #base_model-Tommidi/st_vit_untrained #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 37### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.12154345214366913,
0.027733275666832924,
-0.0016074137529358268,
0.09765663743019104,
0.1626262664794922,
0.02408507466316223,
0.13559022545814514,
0.1082008108496666,
-0.08322816342115402,
0.05402818322181702,
0.1430910974740982,
0.12281898409128189,
0.009775644168257713,
0.14909730851650238,
-0.04329872876405716,
-0.22571124136447906,
-0.0000609647722740192,
0.03482909873127937,
-0.10963454842567444,
0.13192324340343475,
0.0837164968252182,
-0.15183094143867493,
0.09070301055908203,
-0.015169448219239712,
-0.2032870352268219,
0.006259523332118988,
0.05006154254078865,
-0.030080417171120644,
0.15338358283042908,
-0.004305952228605747,
0.14833293855190277,
0.028821956366300583,
0.11928880214691162,
-0.16769346594810486,
0.018485046923160553,
0.08587243407964706,
0.019274890422821045,
0.0878673866391182,
0.05063018575310707,
0.008646149188280106,
0.06180416792631149,
-0.12256994843482971,
0.06244697421789169,
0.009303189814090729,
-0.14286689460277557,
-0.20243264734745026,
-0.08970928937196732,
-0.004832644015550613,
0.06739988178014755,
0.07701186090707779,
-0.01331160869449377,
0.16818593442440033,
-0.04818983003497124,
0.10662304610013962,
0.2379620522260666,
-0.2933730185031891,
-0.08598677814006805,
0.04382403939962387,
0.055456385016441345,
0.12388785928487778,
-0.10602647066116333,
0.009833274409174919,
0.05973118543624878,
0.02359345741569996,
0.15790300071239471,
-0.028418265283107758,
-0.06795275956392288,
0.040232252329587936,
-0.16126254200935364,
-0.008006886579096317,
0.10991333425045013,
0.04847400262951851,
-0.03538382053375244,
-0.005182510241866112,
-0.06726038455963135,
-0.16238832473754883,
-0.05662010237574577,
-0.045776017010211945,
0.023964617401361465,
-0.05036051198840141,
-0.10683994740247726,
-0.022082945331931114,
-0.1051752045750618,
-0.10862842947244644,
-0.003121131332591176,
0.12775647640228271,
0.03243427723646164,
-0.004519687034189701,
-0.03940511867403984,
0.11952296644449234,
-0.0816231444478035,
-0.12787237763404846,
0.03107186034321785,
0.022523978725075722,
-0.05508686602115631,
-0.07684894651174545,
-0.06086362525820732,
-0.13348063826560974,
-0.009204991161823273,
0.057164017111063004,
-0.036638207733631134,
0.050687797367572784,
0.03906450420618057,
0.024852773174643517,
-0.1183844804763794,
0.17947916686534882,
-0.0821148231625557,
-0.022024033591151237,
-0.01432726252824068,
0.08247159421443939,
-0.026867341250181198,
-0.008810950443148613,
-0.07928470522165298,
0.012164046056568623,
0.11246144026517868,
-0.010345611721277237,
-0.07202118635177612,
0.07094090431928635,
-0.030993958935141563,
-0.0037197177298367023,
-0.03551051765680313,
-0.0795891061425209,
0.04203275218605995,
0.0015790080651640892,
-0.04993871971964836,
-0.034570757299661636,
0.024054182693362236,
0.005812321323901415,
0.005900325253605843,
0.1345507800579071,
-0.11542203277349472,
0.037994883954524994,
-0.09446887671947479,
-0.13215401768684387,
-0.01827566884458065,
-0.1086685061454773,
0.010398143902420998,
-0.0700797364115715,
-0.12741008400917053,
-0.03791064769029617,
0.05872675031423569,
-0.036368176341056824,
0.013128881342709064,
-0.05775180459022522,
-0.0762481614947319,
0.01612882874906063,
0.006049350835382938,
0.10669124126434326,
-0.05373609811067581,
0.11944977194070816,
0.021248312667012215,
0.0669507160782814,
-0.004951520822942257,
0.029690535739064217,
-0.07301938533782959,
0.034048356115818024,
-0.20374898612499237,
0.014606595039367676,
-0.06065022200345993,
0.07753757387399673,
-0.09848300367593765,
-0.10573570430278778,
-0.005283128935843706,
0.006641948129981756,
0.10953126102685928,
0.1260734349489212,
-0.19534830749034882,
-0.0558844618499279,
0.1853986531496048,
-0.05997622758150101,
-0.10691019892692566,
0.11956615746021271,
-0.058164194226264954,
0.0699457898736,
0.06002971529960632,
0.1874569207429886,
0.08433649688959122,
-0.11517205089330673,
0.05219423025846481,
-0.019275549799203873,
0.05808171629905701,
-0.019965896382927895,
0.029252348467707634,
0.02128072828054428,
-0.007116896566003561,
0.0034678259398788214,
-0.08648316562175751,
0.02779081091284752,
-0.11541331559419632,
-0.07388411462306976,
-0.05623002350330353,
-0.11458130925893784,
0.06821400672197342,
0.05985666438937187,
0.07653170078992844,
-0.10670971125364304,
-0.07791320234537125,
0.0972360149025917,
0.07859708368778229,
-0.07822520285844803,
0.02759505994617939,
-0.0702991634607315,
0.04589587822556496,
-0.07362911850214005,
-0.029543746262788773,
-0.17648446559906006,
-0.0358395092189312,
-0.004535650368779898,
0.04623831436038017,
0.018083859235048294,
0.03104134276509285,
0.12165910005569458,
0.07088133692741394,
-0.06858935207128525,
-0.024114686995744705,
-0.03343033418059349,
0.015019821934401989,
-0.13191340863704681,
-0.23328931629657745,
-0.035523466765880585,
-0.03587343543767929,
0.08744975179433823,
-0.24515706300735474,
0.02964877337217331,
0.010401386767625809,
0.07793796807527542,
0.05189450457692146,
-0.03675446659326553,
-0.036535341292619705,
0.07360507547855377,
-0.015415867790579796,
-0.07054325193166733,
0.03484216332435608,
-0.016477271914482117,
-0.07664009928703308,
-0.04671833664178848,
-0.13557963073253632,
0.17485341429710388,
0.10272914171218872,
-0.0702999159693718,
-0.11712216585874557,
0.021162517368793488,
-0.0398983433842659,
-0.0240311361849308,
-0.029030390083789825,
0.0385863371193409,
0.08458315581083298,
0.00970809068530798,
0.12362968921661377,
-0.0580909289419651,
-0.038401857018470764,
0.031309302896261215,
-0.034704409539699554,
0.021784666925668716,
0.11105377227067947,
0.10467960685491562,
-0.11372993886470795,
0.11238507181406021,
0.14309315383434296,
-0.08290594816207886,
0.12606120109558105,
-0.034757424145936966,
-0.09252060949802399,
-0.027825355529785156,
0.015171597711741924,
0.022212671115994453,
0.15454186499118805,
-0.11787351220846176,
-0.016032014042139053,
0.0005292849964462221,
0.026177262887358665,
-0.012434172444045544,
-0.22830736637115479,
-0.049913179129362106,
0.0493755117058754,
-0.03160715475678444,
-0.03877609968185425,
-0.018246451392769814,
-0.0046470374800264835,
0.09825510531663895,
0.0013281535357236862,
-0.0610397532582283,
-0.005841927602887154,
0.006368687376379967,
-0.0746486708521843,
0.21307554841041565,
-0.07656940072774887,
-0.08219315111637115,
-0.10243819653987885,
-0.06006801128387451,
-0.008683992549777031,
0.019551502540707588,
0.03951994329690933,
-0.10794349759817123,
-0.02703676000237465,
-0.07966456562280655,
-0.0006104449275881052,
0.026079589501023293,
0.05307232588529587,
-0.008074745535850525,
-0.01710694283246994,
0.08787313848733902,
-0.09104883670806885,
0.020440956577658653,
-0.052627068012952805,
-0.036312002688646317,
0.0402803048491478,
0.05445951595902443,
0.12396623194217682,
0.1614622324705124,
-0.016428649425506592,
0.01637016423046589,
-0.029282063245773315,
0.2493453025817871,
-0.0884992927312851,
-0.0013040738413110375,
0.09700655937194824,
-0.02166897803544998,
0.0451095774769783,
0.1147027537226677,
0.06640470772981644,
-0.10053828358650208,
0.026602277532219887,
0.03279173746705055,
-0.031449172645807266,
-0.20127759873867035,
-0.0004297953855711967,
-0.05115818232297897,
-0.03908248618245125,
0.0906495675444603,
0.014974726364016533,
0.01982470601797104,
0.07196763157844543,
0.06263528764247894,
0.05717594176530838,
-0.030248720198869705,
0.059765301644802094,
0.03505809232592583,
0.034019019454717636,
0.10361683368682861,
-0.03948293998837471,
-0.057592932134866714,
0.026370104402303696,
-0.02260582149028778,
0.221738800406456,
-0.011030882596969604,
0.1183604896068573,
0.05377248674631119,
0.1920851767063141,
-0.02411598153412342,
0.09084624797105789,
-0.004195458255708218,
-0.07393915951251984,
-0.006952774245291948,
-0.045990124344825745,
-0.007895608432590961,
0.025052757933735847,
-0.06910975277423859,
0.034823816269636154,
-0.10185760259628296,
0.023207247257232666,
0.05526534840464592,
0.2474704533815384,
0.06905488669872284,
-0.3327542245388031,
-0.10116603970527649,
-0.005263855215162039,
-0.007099228445440531,
-0.04694712907075882,
-0.008031770586967468,
0.1805332899093628,
-0.0490042045712471,
0.0030866528395563364,
-0.09457936137914658,
0.07448890060186386,
0.013065272942185402,
0.041844189167022705,
0.06519210338592529,
0.12345550209283829,
-0.03456621617078781,
0.045253586024045944,
-0.26864728331565857,
0.31532055139541626,
0.015236143954098225,
0.08563963323831558,
-0.0407976396381855,
-0.03403569757938385,
0.010744179598987103,
0.12935109436511993,
0.0759548619389534,
-0.001301940530538559,
-0.06384270638227463,
-0.191789910197258,
-0.04081789776682854,
0.022255241870880127,
0.1493561714887619,
-0.030778225511312485,
0.11429955065250397,
-0.03474269434809685,
-0.015327810309827328,
0.08599425107240677,
-0.062212564051151276,
-0.06446070224046707,
-0.05373295024037361,
-0.006425937172025442,
-0.009262093342840672,
-0.010577021166682243,
-0.08086574077606201,
-0.10409128665924072,
-0.10229159891605377,
0.14268583059310913,
0.013977485708892345,
0.02724277600646019,
-0.137328639626503,
0.07319369912147522,
0.0971672311425209,
-0.06550631672143936,
0.044383980333805084,
0.04144437238574028,
0.05846607685089111,
0.04566051438450813,
-0.032816242426633835,
0.1351676732301712,
-0.06082116812467575,
-0.1583918184041977,
-0.06650881469249725,
0.09924185276031494,
0.04707227274775505,
0.04765806719660759,
-0.013196157291531563,
0.016927102580666542,
-0.006561078131198883,
-0.07386001944541931,
0.037285227328538895,
-0.03259403631091118,
0.03541012108325958,
-0.006410922855138779,
-0.019390450790524483,
0.05390424281358719,
-0.06229713186621666,
-0.01644160784780979,
0.14659051597118378,
0.24781224131584167,
-0.09931952506303787,
0.01428864523768425,
0.024746481329202652,
-0.05458926409482956,
-0.1897052526473999,
0.11425131559371948,
0.047519925981760025,
0.01833902671933174,
0.06992875784635544,
-0.13292302191257477,
0.08968930691480637,
0.04792408645153046,
-0.0012880455469712615,
0.10879327356815338,
-0.302095890045166,
-0.1402256190776825,
0.09033370763063431,
0.19507353007793427,
0.10586203634738922,
-0.12427451461553574,
-0.021374577656388283,
0.020491093397140503,
-0.10220322012901306,
0.05510517209768295,
-0.10400212556123734,
0.11214768141508102,
0.011400576680898666,
0.0787208080291748,
0.023924734443426132,
-0.07067246735095978,
0.1185920313000679,
-0.02429197169840336,
0.1194632276892662,
-0.05400698259472847,
-0.005986332893371582,
0.05965001508593559,
-0.06182026490569115,
-0.008796006441116333,
-0.04378896951675415,
0.03619547188282013,
-0.04703138396143913,
-0.02196306176483631,
-0.0844159722328186,
0.01428358256816864,
-0.030505016446113586,
-0.06170681118965149,
-0.01809915155172348,
0.03880888223648071,
0.028165295720100403,
-0.03483349084854126,
0.13462480902671814,
-0.00655349949374795,
0.16488739848136902,
0.07179054617881775,
0.08370582014322281,
-0.04748916253447533,
-0.044932667165994644,
0.03315889835357666,
-0.011367535218596458,
0.05370905250310898,
-0.159285306930542,
0.02815689705312252,
0.13349181413650513,
0.038340307772159576,
0.12552015483379364,
0.07226189970970154,
-0.03404010832309723,
0.0035376439336687326,
0.06990597397089005,
-0.17719827592372894,
-0.12119877338409424,
0.006026793271303177,
-0.08219641447067261,
-0.11921625584363937,
0.08326413482427597,
0.1244896724820137,
-0.07245498150587082,
0.004875750280916691,
-0.01766209863126278,
0.008597070351243019,
-0.06886380165815353,
0.23171836137771606,
0.0774206593632698,
0.05890079215168953,
-0.07557258754968643,
0.0368942953646183,
0.025452401489019394,
-0.0847269669175148,
0.014306255616247654,
0.054632414132356644,
-0.09646757692098618,
-0.020941561087965965,
0.12743733823299408,
0.14513084292411804,
-0.0017447681166231632,
-0.01062620710581541,
-0.15222692489624023,
-0.10435020923614502,
0.061698053032159805,
0.24135960638523102,
0.07076305896043777,
0.014806045219302177,
-0.05851557478308678,
0.025680415332317352,
-0.1392495334148407,
0.08932025730609894,
0.06843505054712296,
0.09563988447189331,
-0.18111880123615265,
0.20079025626182556,
-0.03131797909736633,
0.03683226555585861,
-0.029383698478341103,
0.04370664432644844,
-0.12107861787080765,
0.01627853512763977,
-0.1115887388586998,
-0.05062926560640335,
-0.009585138410329819,
-0.010009897872805595,
0.002040998777374625,
-0.10038124024868011,
-0.0642138198018074,
0.015078453347086906,
-0.12499882280826569,
-0.01077917218208313,
0.037849605083465576,
0.020712831988930702,
-0.1198616698384285,
-0.018999746069312096,
0.04019607976078987,
-0.07203944027423859,
0.0537039116024971,
0.03713363781571388,
0.0012328469892963767,
0.0803760215640068,
-0.16123215854167938,
0.01742725260555744,
0.048235829919576645,
-0.01008692104369402,
0.0708654522895813,
-0.07044364511966705,
-0.018242567777633667,
-0.04404476284980774,
0.07924414426088333,
0.0168907567858696,
0.07481622695922852,
-0.10483681410551071,
0.02152000553905964,
-0.03535223379731178,
-0.054112572222948074,
-0.05624309554696083,
0.033188462257385254,
0.05949019640684128,
0.01145323459059,
0.15473714470863342,
-0.07808459550142288,
0.036584172397851944,
-0.21431735157966614,
-0.013608423061668873,
-0.006106250453740358,
-0.1095370501279831,
-0.10385723412036896,
-0.066957987844944,
0.09349668025970459,
-0.05439385026693344,
0.057317737489938736,
-0.011475405655801296,
0.05249839276075363,
0.029340162873268127,
-0.06499398499727249,
0.011846697889268398,
0.04023313522338867,
0.201361283659935,
0.004832437727600336,
-0.043773166835308075,
0.033689387142658234,
0.06033988669514656,
0.09096278995275497,
0.10468112677335739,
0.21347954869270325,
0.1425418257713318,
-0.036183904856443405,
0.09377048909664154,
0.022708237171173096,
-0.04895677790045738,
-0.1518520712852478,
0.06024659797549248,
-0.08736241608858109,
0.09148788452148438,
-0.03924892097711563,
0.20568369328975677,
0.08634625375270844,
-0.16615769267082214,
0.022518211975693703,
-0.06635462492704391,
-0.1176205426454544,
-0.09531939774751663,
-0.037945959717035294,
-0.08687977492809296,
-0.17766073346138,
0.006581426132470369,
-0.10718993097543716,
0.03233443945646286,
0.07167643308639526,
0.02753513678908348,
0.009545711800456047,
0.24470727145671844,
0.06444646418094635,
0.050885092467069626,
0.0595010407269001,
0.006925911642611027,
-0.03135969489812851,
-0.06518688052892685,
-0.08046863228082657,
0.005034399684518576,
-0.040156517177820206,
0.024181216955184937,
-0.07832987606525421,
-0.055602800101041794,
0.03933033347129822,
0.01449610199779272,
-0.10177841037511826,
0.011194066144526005,
0.02955733612179756,
0.03810872137546539,
0.002570315496996045,
0.014248889870941639,
0.0013056498719379306,
-0.02803066372871399,
0.2674822509288788,
-0.09680777788162231,
-0.06130444258451462,
-0.12443690747022629,
0.2530182898044586,
0.03587054833769798,
0.014431609772145748,
0.011958405375480652,
-0.07531000673770905,
0.044903118163347244,
0.23212194442749023,
0.15623219311237335,
-0.09289176017045975,
0.0037583198864012957,
-0.007356765680015087,
-0.009697983972728252,
-0.054081473499536514,
0.11183726787567139,
0.0995035171508789,
-0.012003329582512379,
-0.09655686467885971,
-0.049014560878276825,
-0.04573297128081322,
-0.016184484586119652,
-0.016546623781323433,
0.019709236919879913,
0.0559205487370491,
0.027960414066910744,
-0.056161437183618546,
0.07902304083108902,
-0.02646501176059246,
-0.16689330339431763,
0.06431537121534348,
-0.19141395390033722,
-0.14470362663269043,
-0.03680518642067909,
0.11352751404047012,
-0.025834940373897552,
0.0544634647667408,
-0.030870191752910614,
-0.015548269264400005,
0.06395751237869263,
-0.0036987578496336937,
-0.022468289360404015,
-0.10268020629882812,
0.09609009325504303,
-0.14345867931842804,
0.19002340734004974,
-0.05174835026264191,
-0.001258806441910565,
0.11575952917337418,
0.027305563911795616,
-0.07218904793262482,
0.046575065702199936,
0.05013526231050491,
-0.10553284734487534,
-0.002486856421455741,
0.1298326849937439,
-0.05303152650594711,
0.1031450405716896,
0.021800825372338295,
-0.17464494705200195,
0.015251990407705307,
-0.08080717921257019,
-0.03813591226935387,
-0.032708361744880676,
-0.05746518820524216,
-0.04839780926704407,
0.13581174612045288,
0.21296410262584686,
-0.028908634558320045,
0.040783368051052094,
-0.0745386928319931,
0.03099348023533821,
0.052775003015995026,
0.06779231131076813,
-0.05519649013876915,
-0.2540982663631439,
0.01608046516776085,
0.058290526270866394,
-0.028136733919382095,
-0.22767150402069092,
-0.10762997716665268,
0.03340862691402435,
-0.0408804714679718,
-0.06540191173553467,
0.10031112283468246,
0.11526518315076828,
0.059211984276771545,
-0.06000267714262009,
-0.11270779371261597,
-0.06610222905874252,
0.1880737543106079,
-0.14519259333610535,
-0.07322249561548233
] |
null | null | stable-baselines3 |
# **PPO** Agent playing **LunarLander-v2**
This is a trained model of a **PPO** agent playing **LunarLander-v2**
using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3).
## Usage (with Stable-baselines3)
TODO: Add your code
```python
from stable_baselines3 import ...
from huggingface_sb3 import load_from_hub
...
```
| {"library_name": "stable-baselines3", "tags": ["LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "PPO", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "LunarLander-v2", "type": "LunarLander-v2"}, "metrics": [{"type": "mean_reward", "value": "169.08 +/- 100.02", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | scifarer/ppo-LunarLander-v2 | [
"stable-baselines3",
"LunarLander-v2",
"deep-reinforcement-learning",
"reinforcement-learning",
"model-index",
"region:us"
] | 2024-02-07T16:42:17+00:00 | [] | [] | TAGS
#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
|
# PPO Agent playing LunarLander-v2
This is a trained model of a PPO agent playing LunarLander-v2
using the stable-baselines3 library.
## Usage (with Stable-baselines3)
TODO: Add your code
| [
"# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
"TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n",
"# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
39,
41,
17
] | [
"passage: TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
0.03942384943366051,
0.04900386184453964,
-0.005304091144353151,
0.026427261531352997,
0.107408307492733,
-0.026511888951063156,
0.11188238859176636,
0.0814051404595375,
0.10722193866968155,
0.04762078449130058,
0.08338645845651627,
0.06030960753560066,
0.05080918222665787,
0.2571701407432556,
0.04754156619310379,
-0.22987541556358337,
0.036159250885248184,
-0.04869936779141426,
0.12395193427801132,
0.07178173214197159,
-0.0038484656251966953,
-0.06485428661108017,
0.020415637642145157,
-0.013290755450725555,
0.05367108806967735,
0.04282612353563309,
-0.01716216839849949,
-0.08207534998655319,
0.07169748842716217,
-0.06345846503973007,
0.06986866891384125,
0.07677983492612839,
0.13218913972377777,
-0.17832116782665253,
0.029566360637545586,
0.02571309357881546,
-0.07189024239778519,
0.01342033501714468,
0.008019951172173023,
0.05120139941573143,
0.17303818464279175,
0.019879888743162155,
0.07844575494527817,
-0.0025605305563658476,
-0.15412317216396332,
-0.018950799480080605,
0.0436202734708786,
0.12546207010746002,
0.08808347582817078,
0.04605821147561073,
0.01970590092241764,
0.17503218352794647,
-0.054352790117263794,
-0.028833400458097458,
0.21759237349033356,
-0.2881564497947693,
-0.031460098922252655,
0.321048766374588,
0.06997483223676682,
0.09725230932235718,
-0.07540661096572876,
-0.03619609400629997,
0.007783263456076384,
-0.013137873262166977,
-0.028666524216532707,
-0.07447073608636856,
0.17313385009765625,
0.05152064561843872,
-0.05057951435446739,
-0.09541505575180054,
0.16948209702968597,
0.006921638268977404,
0.0018855923553928733,
-0.019282981753349304,
0.009060598909854889,
0.07402525842189789,
-0.016097044572234154,
-0.07255112379789352,
0.057438433170318604,
0.05330665782094002,
0.019649166613817215,
-0.1435653269290924,
-0.10762494057416916,
-0.022740179672837257,
-0.008012006990611553,
0.17786912620067596,
-0.009255532175302505,
0.042902372777462006,
0.003065188182517886,
0.10384012013673782,
-0.12480384111404419,
-0.03354184702038765,
-0.0454259067773819,
-0.07565800100564957,
-0.0223417766392231,
-0.02058211714029312,
-0.03580251708626747,
0.07184842973947525,
0.11971849203109741,
0.027368178591132164,
0.09350208193063736,
0.047715865075588226,
-0.03206788748502731,
0.06343851238489151,
0.05555703118443489,
0.14222665131092072,
0.05807621404528618,
0.012854371219873428,
0.13179877400398254,
0.055213116109371185,
0.033023182302713394,
-0.0613492950797081,
-0.18252409994602203,
0.07489913702011108,
-0.07031869143247604,
0.007941240444779396,
0.12051256000995636,
-0.04480670019984245,
-0.1183447614312172,
-0.037500523030757904,
-0.017392054200172424,
-0.06224250793457031,
-0.025395862758159637,
0.0547584593296051,
-0.02883218228816986,
-0.03973718360066414,
0.0011496668448671699,
0.09384800493717194,
0.00953749567270279,
-0.1752052903175354,
0.03303423151373863,
-0.025042934343218803,
-0.10782608389854431,
0.009975161403417587,
0.0022444494534283876,
0.03394931182265282,
0.04408763721585274,
-0.11822668462991714,
-0.30899152159690857,
-0.07652641832828522,
0.05490870401263237,
-0.06516939401626587,
-0.18425025045871735,
-0.13193942606449127,
0.02454492449760437,
-0.09037084132432938,
-0.044885024428367615,
-0.12759265303611755,
-0.028549788519740105,
0.01743689924478531,
0.011519349180161953,
0.10758619755506516,
-0.0106219332665205,
-0.012188062071800232,
-0.1571401208639145,
0.008273907005786896,
-0.20951123535633087,
0.0890483483672142,
-0.019150104373693466,
0.037884220480918884,
-0.032381169497966766,
-0.07404014468193054,
0.030707746744155884,
0.052499737590551376,
-0.01474119070917368,
0.13510210812091827,
-0.15592676401138306,
-0.03691192343831062,
-0.007996266707777977,
-0.13611900806427002,
-0.04786273464560509,
-0.10358831286430359,
-0.04357128217816353,
0.13354332745075226,
0.018664736300706863,
0.15356586873531342,
-0.08709818124771118,
-0.0722038671374321,
0.20489206910133362,
-0.010411538183689117,
-0.12820468842983246,
-0.076752208173275,
0.10165707021951675,
0.021510310471057892,
-0.056606587022542953,
-0.02523270808160305,
-0.1839766949415207,
-0.0152357779443264,
-0.04550420492887497,
-0.047039128839969635,
0.01796751655638218,
-0.010888241231441498,
0.13837894797325134,
0.08494598418474197,
0.05018039792776108,
-0.06086122244596481,
-0.006730288732796907,
0.10779471695423126,
0.08823856711387634,
0.008680110797286034,
0.023406028747558594,
-0.05774238705635071,
0.09552932530641556,
-0.04003755748271942,
-0.0142367510125041,
-0.08283266425132751,
-0.036246106028556824,
-0.026256313547492027,
0.17507147789001465,
0.09440762549638748,
0.2257927656173706,
0.09567736834287643,
0.039160262793302536,
0.031270865350961685,
-0.13181598484516144,
-0.1425403207540512,
-0.0017254541162401438,
0.09020978957414627,
-0.14270411431789398,
-0.04119925573468208,
-0.08974775671958923,
-0.17768175899982452,
-0.12202505767345428,
0.0006432619411498308,
-0.17960017919540405,
0.06390921026468277,
0.05408334732055664,
-0.035177867859601974,
0.03272094577550888,
0.13032332062721252,
-0.011533179320394993,
-0.03967514634132385,
0.0831870287656784,
0.0379033200442791,
-0.041234664618968964,
-0.021742934361100197,
0.11885567009449005,
0.15673065185546875,
0.13124459981918335,
-0.03511447086930275,
0.004914294462651014,
0.07076404243707657,
-0.02309088408946991,
0.06539414077997208,
0.0558244064450264,
0.20973342657089233,
0.188301220536232,
0.038996949791908264,
0.008822928182780743,
-0.07048165798187256,
0.0855446457862854,
-0.0742373839020729,
-0.14302679896354675,
-0.05579735338687897,
0.08729292452335358,
0.016605578362941742,
0.023469142615795135,
0.08711627870798111,
0.024545932188630104,
0.09132762253284454,
0.15968108177185059,
0.01990218088030815,
-0.09659269452095032,
-0.050218869000673294,
0.01175848301500082,
0.027713103219866753,
0.04794301092624664,
-0.04514073207974434,
-0.00937939714640379,
0.017020760104060173,
-0.10303554683923721,
0.031789086759090424,
-0.1413339376449585,
-0.1358717679977417,
0.044326696544885635,
0.003906996920704842,
0.010907664895057678,
0.02786896750330925,
-0.0038291432429105043,
0.019039705395698547,
0.04351753741502762,
-0.06975466758012772,
0.047416772693395615,
-0.024745507165789604,
-0.020031947642564774,
0.03340689837932587,
-0.057257164269685745,
-0.205775648355484,
-0.17696654796600342,
0.00013708483311347663,
-0.09910997003316879,
0.10194740444421768,
0.018308809027075768,
-0.12373185902833939,
0.047737859189510345,
-0.05822649225592613,
0.027574289590120316,
-0.01875593699514866,
-0.049130141735076904,
0.10507171601057053,
0.1525275856256485,
-0.016146350651979446,
0.018018173053860664,
-0.04865182936191559,
-0.10157987475395203,
-0.19632206857204437,
0.0691583976149559,
0.04680244252085686,
0.014610917307436466,
0.10669491440057755,
0.018072687089443207,
0.02367905154824257,
-0.007674071006476879,
-0.016521066427230835,
-0.011659215204417706,
-0.08781040459871292,
0.31909599900245667,
0.04510033503174782,
-0.025173069909214973,
0.02041010931134224,
-0.0043001663871109486,
-0.028083480894565582,
0.03263787180185318,
-0.0985708013176918,
-0.07548979669809341,
-0.08774089068174362,
-0.04367410019040108,
-0.09784720093011856,
0.053299110382795334,
0.05916472524404526,
0.003188040340319276,
-0.07727594673633575,
0.04221395403146744,
0.11369874328374863,
-0.0923808291554451,
-0.07137343287467957,
0.07477962225675583,
0.0972946360707283,
-0.07331304252147675,
0.00012658814375754446,
0.00874367356300354,
0.023951783776283264,
0.037102166563272476,
0.06778035312891006,
-0.03966575115919113,
0.08589404821395874,
-0.19917890429496765,
0.0372927263379097,
0.106058269739151,
0.023754918947815895,
0.0638108178973198,
0.07643651217222214,
-0.1058402881026268,
-0.008500572293996811,
-0.032518330961465836,
-0.21341575682163239,
0.1668180525302887,
0.1355515867471695,
0.06788124144077301,
-0.025637222453951836,
-0.00461410591378808,
-0.0649740919470787,
0.05773647129535675,
0.02723747305572033,
-0.14758841693401337,
0.004883295856416225,
0.06064270809292793,
0.026899009943008423,
0.01614922471344471,
0.07971042394638062,
0.014697225764393806,
-0.1801026314496994,
-0.014406266622245312,
0.10730406641960144,
0.002390873385593295,
0.0053148469887673855,
-0.03175045922398567,
-0.1755964607000351,
0.0751047357916832,
0.004285442177206278,
0.07233936339616776,
-0.1676585078239441,
0.14297930896282196,
-0.10089799761772156,
0.07726949453353882,
-0.004285062663257122,
-0.021311495453119278,
0.02507244050502777,
-0.0541163794696331,
0.15163759887218475,
0.01058570109307766,
-0.021810131147503853,
-0.1200498715043068,
-0.1717042326927185,
-0.019227758049964905,
-0.11788936704397202,
-0.11679866164922714,
0.050424277782440186,
0.062185097485780716,
0.04923136904835701,
-0.061147067695856094,
0.1518532931804657,
-0.047422297298908234,
0.060713399201631546,
-0.06893875449895859,
-0.06755045056343079,
0.03764858841896057,
-0.12588608264923096,
-0.08176055550575256,
0.05573027580976486,
0.19166934490203857,
0.15833087265491486,
-0.02816431224346161,
-0.03472423925995827,
-0.047419581562280655,
-0.006212298292666674,
-0.007802055217325687,
0.0275666993111372,
0.023223137483000755,
0.07315318286418915,
-0.07681374251842499,
-0.11649256944656372,
0.033787861466407776,
-0.06713802367448807,
-0.055589709430933,
-0.015439179725944996,
0.1513158082962036,
0.04671623185276985,
0.07720734924077988,
-0.018946662545204163,
0.03887668624520302,
-0.001724981120787561,
-0.056474871933460236,
0.16197094321250916,
0.03885216265916824,
-0.05193585529923439,
0.06837689876556396,
0.053174007683992386,
0.043745119124650955,
0.03011113777756691,
-0.026783017441630363,
0.206032395362854,
0.1980147808790207,
0.014206883497536182,
0.2175983190536499,
0.03177616000175476,
-0.03772832080721855,
-0.1300560086965561,
-0.065880686044693,
-0.006372632458806038,
0.03559038043022156,
0.08070417493581772,
-0.18207235634326935,
-0.015011128038167953,
-0.05689644813537598,
-0.034518610686063766,
-0.15059494972229004,
-0.28553900122642517,
-0.05957856774330139,
0.20075850188732147,
0.14706264436244965,
0.27519428730010986,
-0.10432573407888412,
0.035197313874959946,
0.02663275972008705,
-0.04912831634283066,
-0.006501141935586929,
0.00018665487004909664,
0.10268618166446686,
-0.15421873331069946,
0.1176437959074974,
0.08486983180046082,
-0.019002694636583328,
0.01058861706405878,
-0.1619086116552353,
0.00936629343777895,
-0.12191236019134521,
0.05354422330856323,
0.1400289237499237,
-0.048128653317689896,
-0.054873593151569366,
0.14033560454845428,
-0.024562934413552284,
-0.22685599327087402,
-0.04648222774267197,
-0.043600670993328094,
-0.010640020482242107,
0.026607351377606392,
-0.1013401448726654,
0.04101909324526787,
0.1330099105834961,
0.009380043484270573,
0.1147187277674675,
0.11749245226383209,
-0.052566803991794586,
0.10792597383260727,
0.2257719188928604,
-0.018785694614052773,
0.04689010605216026,
-0.12743118405342102,
-0.0012336712097749114,
-0.028270328417420387,
0.013657891191542149,
-0.09504974633455276,
-0.09938385337591171,
0.02366873063147068,
0.02872389927506447,
0.009118586778640747,
0.0921793207526207,
-0.029922157526016235,
0.0759170651435852,
0.06817561388015747,
-0.13014446198940277,
-0.16288450360298157,
0.015828335657715797,
-0.007344507612287998,
0.08354310691356659,
0.00027861111448146403,
0.08878035843372345,
-0.11932205408811569,
-0.018093237653374672,
-0.03153328225016594,
-0.03319635987281799,
-0.130486860871315,
-0.07138993591070175,
0.06156524643301964,
0.028095467016100883,
-0.06602972000837326,
0.1398407518863678,
0.026440169662237167,
0.15942534804344177,
0.049197953194379807,
0.012499804608523846,
0.07227300107479095,
-0.05345509201288223,
0.1283530443906784,
0.13818155229091644,
-0.00868943240493536,
-0.05460423603653908,
-0.1013643890619278,
-0.10236792266368866,
0.08925779908895493,
-0.05773641914129257,
0.07476430386304855,
-0.14885357022285461,
-0.06675903499126434,
0.015772046521306038,
0.016141414642333984,
-0.09562095999717712,
0.02571965754032135,
-0.01625603251159191,
-0.18119946122169495,
0.056570518761873245,
-0.048285093158483505,
0.0440407395362854,
-0.06347788125276566,
-0.1110161691904068,
-0.17226378619670868,
0.06091433763504028,
0.08593481779098511,
-0.053876690566539764,
-0.12229149043560028,
0.011023230850696564,
-0.00012518465518951416,
-0.06341652572154999,
-0.05023367330431938,
0.09722746908664703,
-0.11020902544260025,
0.031452205032110214,
-0.012567701749503613,
0.08853451162576675,
-0.03510405123233795,
-0.011538895778357983,
0.044220831245183945,
-0.08039166033267975,
-0.009481523185968399,
0.03534642979502678,
-0.026372017338871956,
-0.04127239063382149,
-0.2689029574394226,
0.0036654395516961813,
0.0341104120016098,
0.02497158572077751,
0.07856601476669312,
0.011906822212040424,
0.021174922585487366,
0.03993808850646019,
-0.15396519005298615,
-0.013395369984209538,
0.14574195444583893,
-0.07689505815505981,
-0.022186370566487312,
0.05703273415565491,
-0.09054436534643173,
0.013882770203053951,
-0.030287226662039757,
0.1345842480659485,
0.023923413828015327,
0.06404478847980499,
-0.0851147472858429,
0.10106813907623291,
-0.1451139897108078,
-0.04998219385743141,
-0.01244612317532301,
0.09761348366737366,
0.07019034773111343,
-0.10272270441055298,
0.014697125181555748,
0.04210108891129494,
0.19416837394237518,
0.016384804621338844,
-0.0356343574821949,
-0.03396720811724663,
0.004015897400677204,
0.22076453268527985,
0.03044266067445278,
0.10457023978233337,
0.07281364500522614,
-0.026583973318338394,
0.12624378502368927,
0.09929762035608292,
0.11280370503664017,
-0.055645186454057693,
0.13904185593128204,
0.04667386785149574,
0.038641396909952164,
0.0614289753139019,
0.06836545467376709,
0.09098632633686066,
-0.0008288522367365658,
0.1138714924454689,
0.013811973854899406,
-0.02422109805047512,
-0.021335409954190254,
0.17759373784065247,
0.10501719266176224,
-0.14769648015499115,
0.029047364369034767,
-0.01258957851678133,
0.039933037012815475,
-0.014194529503583908,
-0.15634691715240479,
-0.07240267097949982,
-0.3315149247646332,
0.1226184144616127,
-0.07119352370500565,
0.019930170848965645,
0.007913772016763687,
-0.037425633519887924,
-0.03296699747443199,
-0.04477746784687042,
0.13151589035987854,
-0.013641550205647945,
-0.006079165264964104,
-0.04815853759646416,
-0.015360191464424133,
-0.11607866734266281,
-0.11200575530529022,
-0.013207737356424332,
-0.13671602308750153,
-0.010119039565324783,
0.05595948174595833,
0.003977729007601738,
0.01821410097181797,
-0.03142618387937546,
0.0024383175186812878,
0.06541839241981506,
-0.05751744285225868,
0.056182678788900375,
0.12097269296646118,
0.08766137808561325,
-0.1058853268623352,
0.031048951670527458,
0.2011747509241104,
0.04359564557671547,
-0.12483977526426315,
0.01449228823184967,
0.1819491684436798,
0.004885740112513304,
0.017068125307559967,
-0.006097703706473112,
-0.0540788508951664,
-0.07554277032613754,
0.1251034289598465,
0.08296554535627365,
-0.09985227137804031,
0.015833314508199692,
-0.0726347416639328,
-0.01594804972410202,
-0.06374675035476685,
0.10130585730075836,
0.09538925439119339,
0.04440245032310486,
-0.10621760785579681,
-0.08487539738416672,
-0.10891728103160858,
0.040588874369859695,
-0.08629853278398514,
-0.07311757653951645,
0.09629398584365845,
-0.07057105004787445,
-0.07029950618743896,
0.025521177798509598,
-0.17978744208812714,
-0.009467960335314274,
0.1711762249469757,
-0.24654000997543335,
-0.0916430801153183,
-0.10857923328876495,
0.14477859437465668,
0.016497576609253883,
0.1013975441455841,
-0.006207061931490898,
-0.007889035157859325,
-0.20577777922153473,
0.024890204891562462,
-0.05293011665344238,
-0.02073732763528824,
0.07814782857894897,
-0.09476397186517715,
0.22629831731319427,
-0.08276885002851486,
0.020940175279974937,
0.012659613974392414,
0.0870661810040474,
-0.030675338581204414,
0.09283176809549332,
-0.03660329803824425,
-0.12576518952846527,
-0.03620953485369682,
0.03001813031733036,
0.013904244638979435,
0.10071761906147003,
0.09772487729787827,
-0.03414725139737129,
0.03389119729399681,
0.09747414290904999,
0.04172342270612717,
-0.023843804374337196,
0.0360250361263752,
-0.17077107727527618,
0.02182629331946373,
-0.018498148769140244,
-0.06935930997133255,
0.03687669709324837,
-0.06603235751390457,
0.1639697551727295,
0.04022442549467087,
0.0670473501086235,
-0.036152735352516174,
0.0073931049555540085,
-0.014454689808189869,
-0.013775371946394444,
-0.026180334389209747,
-0.17259705066680908,
-0.10422050207853317,
-0.1347656100988388,
-0.012701659463346004,
-0.034971047192811966,
0.04591470584273338,
0.023234914988279343,
-0.0003200018545612693,
-0.014577031135559082,
-0.12090865522623062,
0.04360328987240791,
0.11146783083677292,
-0.04631396010518074,
-0.026193076744675636
] |
null | null | diffusers |
# SDXL LoRA DreamBooth - StorkelOpa/ancient-world
<Gallery />
## Model description
### These are StorkelOpa/ancient-world LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.
## Download model
### Use it with UIs such as AUTOMATIC1111, Comfy UI, SD.Next, Invoke
- **LoRA**: download **[`ancient-world.safetensors` here 💾](/StorkelOpa/ancient-world/blob/main/ancient-world.safetensors)**.
- Place it on your `models/Lora` folder.
- On AUTOMATIC1111, load the LoRA by adding `<lora:ancient-world:1>` to your prompt. On ComfyUI just [load it as a regular LoRA](https://comfyanonymous.github.io/ComfyUI_examples/lora/).
- *Embeddings*: download **[`ancient-world_emb.safetensors` here 💾](/StorkelOpa/ancient-world/blob/main/ancient-world_emb.safetensors)**.
- Place it on it on your `embeddings` folder
- Use it by adding `ancient-world_emb` to your prompt. For example, `ancient world painting`
(you need both the LoRA and the embeddings as they were trained together for this LoRA)
## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers)
```py
from diffusers import AutoPipelineForText2Image
import torch
from huggingface_hub import hf_hub_download
from safetensors.torch import load_file
pipeline = AutoPipelineForText2Image.from_pretrained('stabilityai/stable-diffusion-xl-base-1.0', torch_dtype=torch.float16).to('cuda')
pipeline.load_lora_weights('StorkelOpa/ancient-world', weight_name='pytorch_lora_weights.safetensors')
embedding_path = hf_hub_download(repo_id='StorkelOpa/ancient-world', filename='ancient-world_emb.safetensors' repo_type="model")
state_dict = load_file(embedding_path)
pipeline.load_textual_inversion(state_dict["clip_l"], token=[], text_encoder=pipeline.text_encoder, tokenizer=pipeline.tokenizer)
pipeline.load_textual_inversion(state_dict["clip_g"], token=[], text_encoder=pipeline.text_encoder_2, tokenizer=pipeline.tokenizer_2)
image = pipeline('ancient world painting').images[0]
```
For more details, including weighting, merging and fusing LoRAs, check the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters)
## Trigger words
To trigger image generation of trained concept(or concepts) replace each concept identifier in you prompt with the new inserted tokens:
to trigger concept `TOK` → use `<s0><s1>` in your prompt
## Details
All [Files & versions](/StorkelOpa/ancient-world/tree/main).
The weights were trained using [🧨 diffusers Advanced Dreambooth Training Script](https://github.com/huggingface/diffusers/blob/main/examples/advanced_diffusion_training/train_dreambooth_lora_sdxl_advanced.py).
LoRA for the text encoder was enabled. False.
Pivotal tuning was enabled: True.
Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
| {"license": "openrail++", "tags": ["stable-diffusion-xl", "stable-diffusion-xl-diffusers", "text-to-image", "diffusers", "lora", "template:sd-lora"], "widget": [{"text": "ancient world painting of Earth's Early Landscape, Showcasing Towering Mountains, Deep Valleys, and Volcanic Activity, Circa 4.5 Billion Years Ago.", "output": {"url": "image-0.png"}}, {"text": "ancient world painting of Earth's Early Ocean Floor, Alive with Primitive Plant Life Amidst Volcanic Rock Formations, Circa 3.5 Billion Years Ago.", "output": {"url": "image-1.png"}}, {"text": "ancient world painting of Cambrian Marine Life, Featuring Trilobites and Jellyfish Amidst Ocean Flora.", "output": {"url": "image-2.png"}}, {"text": "ancient world painting of the Cambrian Seabed, Featuring the Trilobites Paradoxides gracilis, Comocoryphe sulzeri, and Ptychoparia striata, with the Stalked Echinoderm Acadocrinus jani and the Algae Dalya, Set Against a Backdrop of Jellyfish in the Open Water.", "output": {"url": "image-3.png"}}, {"text": "ancient world painting of Upper Silurian Marine Life, with Predatory Nautiloids and Sea Lilies in a Coral Seabed Landscape.", "output": {"url": "image-4.png"}}, {"text": "ancient world painting of the Late Silurian Period, Depicting the First Land Plant Invasion with Primitive Psilophytes Colonizing Coastal Floodplains and Marshes.", "output": {"url": "image-5.png"}}, {"text": "ancient world painting of Middle Devonian Flora, Featuring True Horsetails, Clubmosses, and Ferns Amidst a Primitive Landscape with Waterfalls and Rocky Terrain.", "output": {"url": "image-6.png"}}, {"text": "ancient world painting", "output": {"url": "image-7.png"}}, {"text": "ancient world painting of Early Devonian Aquatic Life, Depicting Osteolepis Attacking Heterostracan Armored Fish with Primitive Plants in the Foreground.", "output": {"url": "image-8.png"}}, {"text": "ancient world painting of Devonian Aquatic Ecosystem, Illustrating Armored Placoderms Like Pterichthyodes and Bothrialepis Navigating the Ocean Floor.", "output": {"url": "image-9.png"}}, {"text": "ancient world painting of Devonian Sea Life, Showcasing the Arthrodira Placoderms in a Dynamic Underwater Scene.", "output": {"url": "image-10.png"}}, {"text": "ancient world painting of Silurian to Devonian Freshwater Fish, Depicting the Primitive Acanthodii Group with Climatius, Euthacanthus, and Parexus.", "output": {"url": "image-11.png"}}, {"text": "ancient world painting of Late Devonian Landscape, Featuring Ichthyostega and the Differentiated Archaeopteris Flora with Cyclostigma Trees and Sphenophyllum Plants.", "output": {"url": "image-12.png"}}, {"text": "ancient world painting", "output": {"url": "image-13.png"}}, {"text": "ancient world painting", "output": {"url": "image-14.png"}}, {"text": "ancient world painting", "output": {"url": "image-15.png"}}, {"text": "ancient world painting", "output": {"url": "image-16.png"}}, {"text": "ancient world painting", "output": {"url": "image-17.png"}}, {"text": "ancient world painting", "output": {"url": "image-18.png"}}, {"text": "ancient world painting", "output": {"url": "image-19.png"}}, {"text": "ancient world painting", "output": {"url": "image-20.png"}}, {"text": "ancient world painting", "output": {"url": "image-21.png"}}, {"text": "ancient world painting", "output": {"url": "image-22.png"}}, {"text": "ancient world painting", "output": {"url": "image-23.png"}}, {"text": "ancient world painting", "output": {"url": "image-24.png"}}, {"text": "ancient world painting", "output": {"url": "image-25.png"}}, {"text": "ancient world painting", "output": {"url": "image-26.png"}}, {"text": "ancient world painting", "output": {"url": "image-27.png"}}, {"text": "ancient world painting", "output": {"url": "image-28.png"}}, {"text": "ancient world painting", "output": {"url": "image-29.png"}}, {"text": "ancient world painting", "output": {"url": "image-30.png"}}, {"text": "ancient world painting", "output": {"url": "image-31.png"}}, {"text": "ancient world painting", "output": {"url": "image-32.png"}}, {"text": "ancient world painting", "output": {"url": "image-33.png"}}, {"text": "ancient world painting", "output": {"url": "image-34.png"}}, {"text": "ancient world painting", "output": {"url": "image-35.png"}}, {"text": "ancient world painting", "output": {"url": "image-36.png"}}, {"text": "ancient world painting", "output": {"url": "image-37.png"}}, {"text": "ancient world painting", "output": {"url": "image-38.png"}}, {"text": "ancient world painting", "output": {"url": "image-39.png"}}, {"text": "ancient world painting", "output": {"url": "image-40.png"}}, {"text": "ancient world painting", "output": {"url": "image-41.png"}}, {"text": "ancient world painting", "output": {"url": "image-42.png"}}, {"text": "ancient world painting", "output": {"url": "image-43.png"}}, {"text": "ancient world painting", "output": {"url": "image-44.png"}}, {"text": "ancient world painting", "output": {"url": "image-45.png"}}, {"text": "ancient world painting", "output": {"url": "image-46.png"}}, {"text": "ancient world painting", "output": {"url": "image-47.png"}}, {"text": "ancient world painting", "output": {"url": "image-48.png"}}, {"text": "ancient world painting", "output": {"url": "image-49.png"}}, {"text": "ancient world painting", "output": {"url": "image-50.png"}}, {"text": "ancient world painting", "output": {"url": "image-51.png"}}, {"text": "ancient world painting", "output": {"url": "image-52.png"}}, {"text": "ancient world painting", "output": {"url": "image-53.png"}}, {"text": "ancient world painting", "output": {"url": "image-54.png"}}, {"text": "ancient world painting", "output": {"url": "image-55.png"}}, {"text": "ancient world painting", "output": {"url": "image-56.png"}}, {"text": "ancient world painting", "output": {"url": "image-57.png"}}, {"text": "ancient world painting", "output": {"url": "image-58.png"}}, {"text": "ancient world painting", "output": {"url": "image-59.png"}}, {"text": "ancient world painting", "output": {"url": "image-60.png"}}, {"text": "ancient world painting", "output": {"url": "image-61.png"}}, {"text": "ancient world painting", "output": {"url": "image-62.png"}}, {"text": "ancient world painting", "output": {"url": "image-63.png"}}, {"text": "ancient world painting", "output": {"url": "image-64.png"}}, {"text": "ancient world painting", "output": {"url": "image-65.png"}}, {"text": "ancient world painting", "output": {"url": "image-66.png"}}, {"text": "ancient world painting", "output": {"url": "image-67.png"}}, {"text": "ancient world painting", "output": {"url": "image-68.png"}}, {"text": "ancient world painting", "output": {"url": "image-69.png"}}, {"text": "ancient world painting", "output": {"url": "image-70.png"}}, {"text": "ancient world painting", "output": {"url": "image-71.png"}}, {"text": "ancient world painting", "output": {"url": "image-72.png"}}, {"text": "ancient world painting", "output": {"url": "image-73.png"}}, {"text": "ancient world painting", "output": {"url": "image-74.png"}}, {"text": "ancient world painting", "output": {"url": "image-75.png"}}, {"text": "ancient world painting", "output": {"url": "image-76.png"}}, {"text": "ancient world painting", "output": {"url": "image-77.png"}}, {"text": "ancient world painting", "output": {"url": "image-78.png"}}, {"text": "ancient world painting", "output": {"url": "image-79.png"}}, {"text": "ancient world painting", "output": {"url": "image-80.png"}}, {"text": "ancient world painting", "output": {"url": "image-81.png"}}], "base_model": "stabilityai/stable-diffusion-xl-base-1.0", "instance_prompt": "ancient world painting"} | text-to-image | StorkelOpa/ancient-world | [
"diffusers",
"tensorboard",
"stable-diffusion-xl",
"stable-diffusion-xl-diffusers",
"text-to-image",
"lora",
"template:sd-lora",
"base_model:stabilityai/stable-diffusion-xl-base-1.0",
"license:openrail++",
"has_space",
"region:us"
] | 2024-02-07T16:43:00+00:00 | [] | [] | TAGS
#diffusers #tensorboard #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us
|
# SDXL LoRA DreamBooth - StorkelOpa/ancient-world
<Gallery />
## Model description
### These are StorkelOpa/ancient-world LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.
## Download model
### Use it with UIs such as AUTOMATIC1111, Comfy UI, SD.Next, Invoke
- LoRA: download 'ancient-world.safetensors' here .
- Place it on your 'models/Lora' folder.
- On AUTOMATIC1111, load the LoRA by adding '<lora:ancient-world:1>' to your prompt. On ComfyUI just load it as a regular LoRA.
- *Embeddings*: download 'ancient-world_emb.safetensors' here .
- Place it on it on your 'embeddings' folder
- Use it by adding 'ancient-world_emb' to your prompt. For example, 'ancient world painting'
(you need both the LoRA and the embeddings as they were trained together for this LoRA)
## Use it with the diffusers library
For more details, including weighting, merging and fusing LoRAs, check the documentation on loading LoRAs in diffusers
## Trigger words
To trigger image generation of trained concept(or concepts) replace each concept identifier in you prompt with the new inserted tokens:
to trigger concept 'TOK' → use '<s0><s1>' in your prompt
## Details
All Files & versions.
The weights were trained using diffusers Advanced Dreambooth Training Script.
LoRA for the text encoder was enabled. False.
Pivotal tuning was enabled: True.
Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
| [
"# SDXL LoRA DreamBooth - StorkelOpa/ancient-world\n\n<Gallery />",
"## Model description",
"### These are StorkelOpa/ancient-world LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.",
"## Download model",
"### Use it with UIs such as AUTOMATIC1111, Comfy UI, SD.Next, Invoke\n\n- LoRA: download 'ancient-world.safetensors' here .\n - Place it on your 'models/Lora' folder.\n - On AUTOMATIC1111, load the LoRA by adding '<lora:ancient-world:1>' to your prompt. On ComfyUI just load it as a regular LoRA.\n- *Embeddings*: download 'ancient-world_emb.safetensors' here .\n - Place it on it on your 'embeddings' folder\n - Use it by adding 'ancient-world_emb' to your prompt. For example, 'ancient world painting'\n (you need both the LoRA and the embeddings as they were trained together for this LoRA)",
"## Use it with the diffusers library\n\n\n\nFor more details, including weighting, merging and fusing LoRAs, check the documentation on loading LoRAs in diffusers",
"## Trigger words\n\nTo trigger image generation of trained concept(or concepts) replace each concept identifier in you prompt with the new inserted tokens:\n\nto trigger concept 'TOK' → use '<s0><s1>' in your prompt",
"## Details\nAll Files & versions.\n\nThe weights were trained using diffusers Advanced Dreambooth Training Script.\n\nLoRA for the text encoder was enabled. False.\n\nPivotal tuning was enabled: True.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix."
] | [
"TAGS\n#diffusers #tensorboard #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us \n",
"# SDXL LoRA DreamBooth - StorkelOpa/ancient-world\n\n<Gallery />",
"## Model description",
"### These are StorkelOpa/ancient-world LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.",
"## Download model",
"### Use it with UIs such as AUTOMATIC1111, Comfy UI, SD.Next, Invoke\n\n- LoRA: download 'ancient-world.safetensors' here .\n - Place it on your 'models/Lora' folder.\n - On AUTOMATIC1111, load the LoRA by adding '<lora:ancient-world:1>' to your prompt. On ComfyUI just load it as a regular LoRA.\n- *Embeddings*: download 'ancient-world_emb.safetensors' here .\n - Place it on it on your 'embeddings' folder\n - Use it by adding 'ancient-world_emb' to your prompt. For example, 'ancient world painting'\n (you need both the LoRA and the embeddings as they were trained together for this LoRA)",
"## Use it with the diffusers library\n\n\n\nFor more details, including weighting, merging and fusing LoRAs, check the documentation on loading LoRAs in diffusers",
"## Trigger words\n\nTo trigger image generation of trained concept(or concepts) replace each concept identifier in you prompt with the new inserted tokens:\n\nto trigger concept 'TOK' → use '<s0><s1>' in your prompt",
"## Details\nAll Files & versions.\n\nThe weights were trained using diffusers Advanced Dreambooth Training Script.\n\nLoRA for the text encoder was enabled. False.\n\nPivotal tuning was enabled: True.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix."
] | [
86,
23,
3,
37,
3,
190,
38,
54,
74
] | [
"passage: TAGS\n#diffusers #tensorboard #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us \n# SDXL LoRA DreamBooth - StorkelOpa/ancient-world\n\n<Gallery />## Model description### These are StorkelOpa/ancient-world LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.## Download model### Use it with UIs such as AUTOMATIC1111, Comfy UI, SD.Next, Invoke\n\n- LoRA: download 'ancient-world.safetensors' here .\n - Place it on your 'models/Lora' folder.\n - On AUTOMATIC1111, load the LoRA by adding '<lora:ancient-world:1>' to your prompt. On ComfyUI just load it as a regular LoRA.\n- *Embeddings*: download 'ancient-world_emb.safetensors' here .\n - Place it on it on your 'embeddings' folder\n - Use it by adding 'ancient-world_emb' to your prompt. For example, 'ancient world painting'\n (you need both the LoRA and the embeddings as they were trained together for this LoRA)## Use it with the diffusers library\n\n\n\nFor more details, including weighting, merging and fusing LoRAs, check the documentation on loading LoRAs in diffusers## Trigger words\n\nTo trigger image generation of trained concept(or concepts) replace each concept identifier in you prompt with the new inserted tokens:\n\nto trigger concept 'TOK' → use '<s0><s1>' in your prompt"
] | [
-0.019709482789039612,
0.055882569402456284,
-0.007858213968575,
0.028405578806996346,
0.11874446272850037,
0.014182920567691326,
0.1203756108880043,
0.06914161145687103,
0.1321174055337906,
0.11662759631872177,
-0.020571257919073105,
0.12772898375988007,
0.06633852422237396,
0.1515878438949585,
0.01747511886060238,
-0.17648041248321533,
0.020143406465649605,
-0.09241324663162231,
0.010286688804626465,
0.03348608314990997,
0.0609518326818943,
-0.0663989782333374,
0.11205621063709259,
-0.03148047253489494,
-0.041583504527807236,
0.038513801991939545,
-0.026901021599769592,
-0.02219696342945099,
0.04877869039773941,
0.07519818842411041,
0.07615137845277786,
-0.012311698868870735,
0.07319369167089462,
-0.17767468094825745,
0.02020064741373062,
0.08658705651760101,
-0.008935496211051941,
0.033891696482896805,
0.12012287229299545,
-0.08427233248949051,
0.11148428171873093,
-0.11802668124437332,
0.0474526584148407,
0.05988285690546036,
-0.10459965467453003,
-0.07336653769016266,
-0.117784284055233,
-0.023422133177518845,
0.13285456597805023,
0.05794612690806389,
0.010572080500423908,
0.022097423672676086,
0.004221814684569836,
0.050404660403728485,
0.1728801727294922,
-0.14561232924461365,
-0.007507815025746822,
0.04939818009734154,
0.03871937841176987,
0.029029661789536476,
-0.07936442643404007,
0.021438680589199066,
-0.007454596925526857,
-0.02045808546245098,
0.05893101543188095,
-0.06609278917312622,
-0.02069062739610672,
-0.045801930129528046,
-0.1110861748456955,
-0.03187434747815132,
0.18709592521190643,
0.01094641163945198,
-0.06737472116947174,
-0.1097104474902153,
-0.08553895354270935,
0.0585809201002121,
-0.01249318104237318,
-0.005279591307044029,
0.02184494026005268,
-0.006577827502042055,
0.11655254662036896,
-0.15974408388137817,
-0.08390166610479355,
-0.016457539051771164,
-0.04336383566260338,
0.11665382981300354,
0.007752361707389355,
0.014773462899029255,
0.034146811813116074,
0.10765884071588516,
-0.025574680417776108,
-0.10430154204368591,
-0.010773207060992718,
-0.020596889778971672,
-0.14563478529453278,
-0.01366202812641859,
0.017477523535490036,
-0.09379718452692032,
0.011868373490869999,
0.1973906308412552,
0.1031845435500145,
0.04216303303837776,
-0.12605398893356323,
0.06417599320411682,
0.009759759530425072,
0.13674476742744446,
-0.009222691878676414,
-0.09558247774839401,
0.026367109268903732,
-0.001879255985841155,
0.05528625100851059,
-0.041380345821380615,
-0.05379798635840416,
-0.04027172178030014,
-0.09694241732358932,
0.08809516578912735,
0.047002680599689484,
0.035421811044216156,
-0.06550576537847519,
-0.03827798739075661,
0.12087010592222214,
-0.15366710722446442,
0.0889829620718956,
0.007776631973683834,
-0.04629391431808472,
0.11319849640130997,
0.11879801005125046,
-0.01958499476313591,
-0.04620201885700226,
0.12113548815250397,
-0.04268250614404678,
0.047164589166641235,
-0.1206333339214325,
-0.06860823184251785,
0.022211959585547447,
-0.0986817330121994,
-0.03245391696691513,
-0.0946807861328125,
-0.17777422070503235,
-0.04378572851419449,
0.04185394197702408,
-0.028478119522333145,
0.03548736870288849,
-0.03580504655838013,
-0.029986564069986343,
0.0022427700459957123,
0.031024066731333733,
0.010347369126975536,
-0.017001541331410408,
0.025854874402284622,
-0.03949201852083206,
0.05718446150422096,
-0.008881454356014729,
0.019280074164271355,
-0.032016679644584656,
0.06239818409085274,
-0.24409469962120056,
0.11215421557426453,
-0.07537876814603806,
0.016806699335575104,
-0.10769078880548477,
0.05602166801691055,
0.004133312497287989,
0.01985803432762623,
0.00765205966308713,
0.10775149613618851,
-0.19683195650577545,
-0.019064469262957573,
0.10362323373556137,
-0.09599961340427399,
-0.019665999338030815,
0.06925424933433533,
-0.008473544381558895,
0.08612723648548126,
0.050094958394765854,
0.09162873029708862,
0.12459208816289902,
-0.19794203341007233,
-0.0454087071120739,
-0.051992516964673996,
-0.10024873912334442,
0.03196490556001663,
0.02458340674638748,
-0.014680281281471252,
0.054333653301000595,
0.0392395444214344,
-0.05799378827214241,
0.04832114279270172,
0.05311112850904465,
-0.011846168898046017,
-0.012637842446565628,
-0.02503674104809761,
-0.0285172238945961,
-0.059496402740478516,
-0.07545062899589539,
0.04460737854242325,
-0.05405681952834129,
0.020924538373947144,
0.10172441601753235,
0.0062209442257881165,
0.05054927617311478,
-0.050107523798942566,
0.09174633771181107,
-0.04032852500677109,
0.011209556832909584,
-0.11684025079011917,
-0.12207671254873276,
0.017510680481791496,
0.024462739005684853,
0.09329959005117416,
-0.038274649530649185,
0.05129194259643555,
0.08619768917560577,
0.01612020470201969,
-0.03836100921034813,
0.046809956431388855,
-0.0774117186665535,
-0.02677294611930847,
-0.07110924273729324,
-0.04554906487464905,
-0.05424109473824501,
0.09469890594482422,
-0.06604857742786407,
0.03538070619106293,
-0.0083256084471941,
0.09758123755455017,
0.05868149176239967,
-0.04841316491365433,
0.012902399525046349,
-0.05550849065184593,
-0.023790571838617325,
-0.045856960117816925,
-0.024062728509306908,
-0.012151148170232773,
-0.0539948046207428,
0.07416421920061111,
-0.209297314286232,
-0.07317931205034256,
0.0883268341422081,
0.06745769083499908,
-0.05663905665278435,
-0.11265553534030914,
0.014862214215099812,
-0.004199601709842682,
-0.10454679280519485,
-0.09188771992921829,
0.10390765219926834,
0.08070586621761322,
0.07481437921524048,
-0.030865458771586418,
-0.013321900740265846,
-0.040714021772146225,
-0.016416242346167564,
-0.04443703964352608,
0.05841660499572754,
0.09442794322967529,
-0.010181942954659462,
0.04727452993392944,
0.09780342876911163,
-0.05922186002135277,
0.05248069763183594,
0.04125203564763069,
-0.08579032868146896,
-0.03785662353038788,
0.05657335743308067,
0.05921311676502228,
0.014214928261935711,
0.09203735738992691,
0.07703504711389542,
0.0586177334189415,
-0.03095102682709694,
-0.024578986689448357,
-0.06756938248872757,
0.014949790202081203,
0.011952999047935009,
-0.030345184728503227,
0.11089527606964111,
0.07181482762098312,
0.039889413863420486,
0.07156001776456833,
-0.04635736346244812,
0.039069872349500656,
-0.046595681458711624,
-0.03954573720693588,
-0.05157208815217018,
0.05246783047914505,
-0.10983145982027054,
-0.16105538606643677,
-0.1905200332403183,
-0.07293573021888733,
-0.06438663601875305,
0.022158948704600334,
0.05040924623608589,
0.0009500632295385003,
-0.10047657787799835,
-0.07031354308128357,
0.10205119848251343,
0.06034514680504799,
-0.08271682262420654,
-0.07761724293231964,
0.037682004272937775,
0.003758935956284404,
-0.0755445808172226,
-0.020420877262949944,
0.043022286146879196,
-0.08666796237230301,
0.025023462250828743,
0.10737370699644089,
0.1149265319108963,
0.026785504072904587,
0.029426492750644684,
0.000317561614792794,
0.031016914173960686,
0.11273228377103806,
-0.08687388896942139,
0.15068601071834564,
0.18853846192359924,
-0.05265767499804497,
0.1174522116780281,
0.13182386755943298,
0.021708350628614426,
-0.058289676904678345,
0.030358487740159035,
0.08008687198162079,
-0.05066762864589691,
-0.16901086270809174,
-0.06498382985591888,
-0.050662145018577576,
-0.0312671922147274,
0.0932098850607872,
0.0621471181511879,
0.0660962238907814,
0.07507620006799698,
-0.05201442167162895,
-0.02815883606672287,
0.06476888060569763,
0.12423160672187805,
0.03426600620150566,
-0.03210708126425743,
0.031307220458984375,
-0.063263900578022,
-0.053399812430143356,
0.07200605422258377,
0.07104433327913284,
0.17055067420005798,
-0.020809125155210495,
0.06351769715547562,
-0.005963531322777271,
0.01587516814470291,
-0.02625643089413643,
0.10881874710321426,
0.016939058899879456,
0.0257396399974823,
-0.01918920874595642,
-0.08436287939548492,
0.0007151758181862533,
0.13084539771080017,
0.07678397744894028,
-0.026933034881949425,
0.03502617031335831,
0.011443089693784714,
0.07464457303285599,
0.11874176561832428,
-0.03975953161716461,
-0.19588761031627655,
0.08653990924358368,
0.06677757948637009,
0.020595403388142586,
-0.08894481509923935,
-0.014318258501589298,
0.05575292930006981,
-0.07966110110282898,
0.054520346224308014,
-0.03402669355273247,
0.07641509920358658,
-0.0814419761300087,
-0.057843685150146484,
-0.03148890286684036,
0.18738849461078644,
-0.003933690022677183,
0.03787858784198761,
-0.15800794959068298,
0.010618016123771667,
0.029413903132081032,
0.0207276102155447,
-0.05329674109816551,
0.06172284111380577,
0.05280272662639618,
0.004396532196551561,
0.15406739711761475,
-0.015981996431946754,
-0.14285367727279663,
-0.10637914389371872,
-0.0858454704284668,
-0.0471680723130703,
0.07576964795589447,
-0.09834659099578857,
0.06871338188648224,
-0.04015864059329033,
-0.0689634159207344,
-0.04397762939333916,
-0.010295741260051727,
-0.07573223859071732,
-0.18563783168792725,
0.024184295907616615,
0.04115573689341545,
0.036612555384635925,
-0.023262687027454376,
0.04212760925292969,
-0.06849133223295212,
0.10651952773332596,
-0.04292995110154152,
-0.05956284701824188,
-0.12022637575864792,
-0.09266270697116852,
0.13003501296043396,
-0.0622355081140995,
0.007745118346065283,
-0.030825255438685417,
0.09818769246339798,
-0.09450468420982361,
-0.08586426824331284,
0.006764120887964964,
-0.04906412586569786,
-0.03034093603491783,
-0.042638570070266724,
0.10390955209732056,
0.02152862586081028,
-0.004377192351967096,
-0.010863198898732662,
0.040791191160678864,
0.030436262488365173,
-0.09099731594324112,
0.025628099218010902,
0.22743991017341614,
0.029159318655729294,
0.0984969213604927,
-0.13391639292240143,
-0.0893687829375267,
-0.09136058390140533,
0.02761559560894966,
0.10946400463581085,
0.22884583473205566,
-0.05763029307126999,
0.09378892928361893,
0.007119255606085062,
-0.09335547685623169,
-0.18717627227306366,
-0.00811776239424944,
-0.00801448617130518,
-0.011991255916655064,
0.045937832444906235,
-0.15682338178157806,
0.10364892333745956,
0.08878399431705475,
-0.011030507273972034,
0.20169296860694885,
-0.2710387110710144,
-0.10737941414117813,
0.0162065289914608,
0.016908790916204453,
-0.11558017879724503,
-0.11989403516054153,
-0.05185171589255333,
-0.0884983092546463,
0.04085778445005417,
0.07390246540307999,
0.023000769317150116,
0.06971270591020584,
0.01976688578724861,
0.06215248256921768,
0.07876773923635483,
-0.05857888236641884,
0.1291547417640686,
-0.026552679017186165,
0.07797197997570038,
-0.08463000506162643,
-0.013646476902067661,
0.036964382976293564,
-0.08642285317182541,
0.1342594474554062,
-0.09940110892057419,
-0.02701047994196415,
-0.05861919745802879,
-0.007519999518990517,
-0.002898563863709569,
0.0996483862400055,
-0.023665398359298706,
-0.010604895651340485,
-0.058408014476299286,
0.029986266046762466,
0.07648951560258865,
-0.0022202511318027973,
-0.014136632904410362,
-0.04690565541386604,
-0.06818947196006775,
0.15641669929027557,
0.0034571324940770864,
0.14311940968036652,
-0.10143104940652847,
-0.004014040343463421,
0.0002927360183093697,
0.018993260338902473,
-0.06497155129909515,
0.04053986445069313,
0.07976699620485306,
-0.025766756385564804,
0.11432646214962006,
0.004454130306839943,
-0.10560137778520584,
0.01561546791344881,
0.07427762448787689,
-0.09485576301813126,
-0.12724818289279938,
-0.0019169036531820893,
0.09964379668235779,
-0.03784475848078728,
-0.08693820238113403,
0.14278006553649902,
0.008451512083411217,
-0.007197894621640444,
0.027842722833156586,
0.07682095468044281,
-0.019855236634612083,
0.03224979713559151,
-0.04419936612248421,
0.0001494035095674917,
-0.0473242849111557,
0.11181791126728058,
0.08197611570358276,
-0.08695071935653687,
-0.010959726758301258,
0.16382507979869843,
-0.02690376341342926,
-0.07144345343112946,
-0.04113141447305679,
0.10858894884586334,
-0.055670663714408875,
0.017721014097332954,
0.04890521615743637,
-0.045458775013685226,
0.004139580763876438,
0.10837974399328232,
0.00393542367964983,
-0.03147042542695999,
-0.004384622909128666,
-0.006180411204695702,
-0.05792541801929474,
0.07699652761220932,
0.023683389648795128,
0.059134289622306824,
-0.10327138751745224,
0.0426153689622879,
-0.003260191297158599,
-0.003369692014530301,
-0.005724222864955664,
-0.029732277616858482,
-0.06512143462896347,
-0.010365650989115238,
-0.03654976934194565,
0.062102191150188446,
-0.08711546659469604,
-0.01253355760127306,
0.012658285908401012,
0.005335309077054262,
0.031043697148561478,
-0.01784103363752365,
-0.05371743068099022,
-0.09521812200546265,
-0.03758952394127846,
0.07954943925142288,
-0.14216835796833038,
-0.08501891791820526,
0.08194135874509811,
-0.09995335340499878,
0.04961812123656273,
0.041951727122068405,
-0.0179172120988369,
-0.023118434473872185,
-0.10650364309549332,
-0.004468217492103577,
0.049350012093782425,
0.029854003340005875,
0.026338638737797737,
-0.092479407787323,
0.01624080166220665,
-0.07384631782770157,
-0.01794440671801567,
-0.06058540195226669,
0.040879830718040466,
-0.11772087961435318,
0.10209663212299347,
-0.0016913370927795768,
-0.09537435322999954,
-0.07580481469631195,
0.02215149998664856,
0.09392669051885605,
0.055090341717004776,
0.08492296189069748,
-0.03710323944687843,
0.10796266049146652,
-0.07286307960748672,
0.02378838323056698,
0.029350385069847107,
-0.04015035554766655,
0.05103355646133423,
-0.10631292313337326,
0.024002987891435623,
-0.013361848890781403,
0.025867337360978127,
-0.01863718032836914,
0.0015702456003054976,
-0.0005599294090643525,
0.006260199472308159,
-0.06262992322444916,
0.01670123264193535,
-0.00017988237959798425,
-0.0018109864322468638,
0.017200712114572525,
-0.01646582782268524,
-0.01884232461452484,
0.0333876796066761,
-0.026587411761283875,
0.06927601993083954,
0.13461872935295105,
0.0454208068549633,
0.029319237917661667,
0.01899116300046444,
-0.09344761818647385,
-0.06036893650889397,
0.003794824006035924,
-0.04776713252067566,
0.021098075434565544,
-0.08845667541027069,
0.10117610543966293,
0.0859050527215004,
-0.12254197150468826,
0.0896187573671341,
0.08607308566570282,
-0.010579156689345837,
-0.057937223464250565,
-0.19321443140506744,
-0.03148675709962845,
-0.020177258178591728,
0.024231644347310066,
-0.056233011186122894,
0.10730632394552231,
-0.038076002150774,
-0.01106460765004158,
0.03055969625711441,
0.11505499482154846,
-0.1259099692106247,
-0.049186527729034424,
0.017590247094631195,
0.018930036574602127,
0.012138346210122108,
0.0787782296538353,
-0.012873955070972443,
0.004354322329163551,
0.028974350541830063,
0.01873953826725483,
0.07287871092557907,
0.08337292820215225,
0.07804175466299057,
-0.08868949115276337,
-0.0400405153632164,
-0.0021518610883504152,
0.016510002315044403,
-0.034342825412750244,
0.12044551223516464,
0.0591125562787056,
-0.0327909030020237,
-0.05266047269105911,
0.14783252775669098,
-0.07506032288074493,
-0.054762136191129684,
-0.09413551539182663,
0.0750948116183281,
-0.0003890790685545653,
0.012136546894907951,
-0.046253543347120285,
-0.11010905355215073,
-0.042031463235616684,
0.1889137327671051,
0.11198967695236206,
-0.034931786358356476,
0.02317987009882927,
-0.01731117069721222,
0.015180760994553566,
-0.029800336807966232,
0.02338234707713127,
0.03604946658015251,
0.16423436999320984,
-0.04640456661581993,
0.09990953654050827,
-0.026695026084780693,
-0.027872735634446144,
-0.052312400192022324,
0.046600569039583206,
-0.13612975180149078,
-0.012338872067630291,
-0.009527784772217274,
0.041064731776714325,
-0.09202275425195694,
-0.16363553702831268,
0.051898304373025894,
-0.005946003366261721,
-0.061963457614183426,
0.019860224798321724,
0.05934380367398262,
0.008597223088145256,
0.033014215528964996,
-0.017189789563417435,
-0.05642525106668472,
0.12699957191944122,
-0.010960090905427933,
-0.10386423766613007,
-0.041229523718357086,
-0.006528965663164854,
-0.0806867927312851,
0.1661680042743683,
0.01396362204104662,
-0.022772416472434998,
0.05527596175670624,
-0.009760849177837372,
-0.12348735332489014,
0.046724170446395874,
0.027924641966819763,
-0.1438925862312317,
0.04945898801088333,
0.16551730036735535,
-0.03830012306571007,
0.09436856955289841,
0.034500859677791595,
-0.08393710851669312,
0.05144064500927925,
0.05240712687373161,
0.015317901037633419,
-0.05990447849035263,
0.05457807704806328,
-0.12353776395320892,
0.12261497974395752,
0.13778415322303772,
0.0396992526948452,
0.02758491225540638,
-0.04994559660553932,
0.00518313143402338,
0.03509657457470894,
0.03995925560593605,
-0.03193901851773262,
-0.03134959563612938,
-0.003738702042028308,
0.0017560977721586823,
0.07675987482070923,
-0.12115003168582916,
-0.07428059726953506,
-0.05169737711548805,
-0.05086319148540497,
-0.04806807264685631,
0.08772792667150497,
0.11732015013694763,
0.004303475841879845,
0.010110962204635143,
-0.2463890016078949,
0.05025471746921539,
0.10163865238428116,
-0.10895449668169022,
-0.018775029107928276
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | Adeptschneider/mistral_lora_instruct_model | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-07T16:43:03+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
31,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06646376848220825,
0.2168014943599701,
-0.00225935154594481,
0.023818302899599075,
0.1271018385887146,
-0.001635765191167593,
0.04218708351254463,
0.13324736058712006,
-0.020175931975245476,
0.11144465953111649,
0.046588581055402756,
0.09377603232860565,
0.09928803145885468,
0.18404334783554077,
0.04859916493296623,
-0.2059975117444992,
0.007056170143187046,
-0.09090408682823181,
0.014076028019189835,
0.1116579994559288,
0.13719257712364197,
-0.10291384905576706,
0.08272874355316162,
-0.04045208916068077,
-0.02019004337489605,
0.00012576708104461432,
-0.09259183704853058,
-0.07032395154237747,
0.06885425746440887,
0.06264153122901917,
0.051234472543001175,
0.001456156256608665,
0.09140396863222122,
-0.2864592671394348,
0.017265573143959045,
0.08406311273574829,
0.0027674848679453135,
0.06290827691555023,
0.07236549258232117,
-0.07389893382787704,
0.11328595131635666,
-0.08021481335163116,
0.13019037246704102,
0.08625296503305435,
-0.062064990401268005,
-0.23071379959583282,
-0.07525765895843506,
0.0963398814201355,
0.12251301854848862,
0.06215599179267883,
-0.022921854630112648,
0.15455181896686554,
-0.06248689442873001,
0.012971068732440472,
0.1294165402650833,
-0.11526761949062347,
-0.05572471022605896,
0.061741601675748825,
0.11775490641593933,
0.10740239918231964,
-0.14110268652439117,
-0.0017287094378843904,
0.04900608956813812,
0.029121357947587967,
0.08589313924312592,
0.022661056369543076,
0.12003941088914871,
0.04652795568108559,
-0.13695219159126282,
-0.04037507623434067,
0.12011898308992386,
0.038862764835357666,
-0.06446044892072678,
-0.2168138176202774,
-0.006778308190405369,
-0.0601806715130806,
-0.014732478186488152,
-0.07019448280334473,
0.039128515869379044,
-0.02470310963690281,
0.07317749410867691,
-0.04465159401297569,
-0.1063927412033081,
-0.0421026237308979,
0.0892222449183464,
0.07748593389987946,
0.011527054943144321,
-0.02519804798066616,
0.04627908393740654,
0.13455867767333984,
0.05402068421244621,
-0.10399353504180908,
-0.07017925381660461,
-0.06942764669656754,
-0.09420394152402878,
-0.04035796597599983,
0.056760527193546295,
0.031942449510097504,
0.02665667235851288,
0.22703726589679718,
0.016653569415211678,
0.04155244305729866,
0.0224777739495039,
0.01032855175435543,
0.043662428855895996,
0.0955500528216362,
-0.05303520709276199,
-0.15660029649734497,
-0.04072032496333122,
0.09077946096658707,
-0.0027527001220732927,
-0.036689214408397675,
-0.03966725245118141,
0.03849169611930847,
0.06843466311693192,
0.13122352957725525,
0.07552056759595871,
-0.017929591238498688,
-0.04813180863857269,
-0.030096933245658875,
0.23523783683776855,
-0.1493375599384308,
0.04426715523004532,
-0.02271856553852558,
-0.01804111897945404,
-0.03908449783921242,
0.03597262129187584,
0.022118929773569107,
-0.000004518366949923802,
0.09706240892410278,
-0.058981191366910934,
-0.05378659814596176,
-0.10168042778968811,
-0.03272576630115509,
0.04088849574327469,
-0.013975566253066063,
-0.010589460842311382,
-0.09025166928768158,
-0.09490354359149933,
-0.04766594246029854,
0.05537205561995506,
-0.05123869329690933,
-0.03770573064684868,
0.009465423412621021,
-0.08151785284280777,
-0.005444355774670839,
-0.005417742300778627,
0.10699385404586792,
-0.03222226724028587,
0.04445803165435791,
-0.027600755915045738,
0.05225523188710213,
0.09919606149196625,
0.031576547771692276,
-0.0773419588804245,
0.0561848059296608,
-0.22559374570846558,
0.07503069192171097,
-0.11481974273920059,
0.04335082694888115,
-0.1704932004213333,
-0.042439818382263184,
0.005444696638733149,
0.0139949731528759,
0.013206101022660732,
0.12720820307731628,
-0.19255615770816803,
-0.01654396951198578,
0.13260798156261444,
-0.09212633967399597,
-0.118110790848732,
0.07884611934423447,
-0.029701577499508858,
0.1624738723039627,
0.04682036489248276,
-0.027025915682315826,
0.09224298596382141,
-0.16434773802757263,
-0.07092688232660294,
-0.00949116237461567,
-0.01727987825870514,
0.12109188735485077,
0.07512219995260239,
-0.05991523340344429,
0.046571120619773865,
0.02832140028476715,
-0.038078423589468,
-0.04424772411584854,
-0.050857074558734894,
-0.10884185880422592,
-0.01070026308298111,
-0.08987759798765182,
0.04065500199794769,
-0.01250192429870367,
-0.07916021347045898,
-0.029885273426771164,
-0.18612512946128845,
-0.0030564051121473312,
0.10038342326879501,
0.0035033065360039473,
-0.005652366206049919,
-0.08666291832923889,
0.026358824223279953,
-0.03112892620265484,
-0.008404186926782131,
-0.16764774918556213,
-0.04399421438574791,
0.046902090311050415,
-0.16094985604286194,
0.020117372274398804,
-0.06413903087377548,
0.06334125250577927,
0.03641495108604431,
-0.05590536445379257,
-0.0248766727745533,
-0.01730942726135254,
0.011945613659918308,
-0.05083848536014557,
-0.18994836509227753,
-0.056277405470609665,
-0.037882111966609955,
0.149809330701828,
-0.25956398248672485,
0.032966937869787216,
0.051140617579221725,
0.14649195969104767,
0.00406361510977149,
-0.05115427449345589,
0.01429014839231968,
-0.05360214412212372,
-0.054652128368616104,
-0.06746816635131836,
-0.006135428790003061,
-0.027576493099331856,
-0.05147203803062439,
0.019243421033024788,
-0.1755700707435608,
-0.021410830318927765,
0.09424154460430145,
0.12876708805561066,
-0.1486445665359497,
-0.018640631809830666,
-0.048725154250860214,
-0.06339836865663528,
-0.0715010017156601,
-0.07038594037294388,
0.10712739825248718,
0.0513901449739933,
0.04796046018600464,
-0.07435787469148636,
-0.07092321664094925,
0.02726263552904129,
0.006906150374561548,
-0.03382374346256256,
0.08727246522903442,
0.05199531093239784,
-0.09209315478801727,
0.0756213590502739,
0.1092359870672226,
0.07177663594484329,
0.09363535046577454,
0.01574566215276718,
-0.11756632477045059,
-0.028492970392107964,
0.036266472190618515,
0.02740776725113392,
0.1465986967086792,
-0.05952361226081848,
0.04016614332795143,
0.04494241625070572,
-0.04170418903231621,
0.022319864481687546,
-0.08787637203931808,
0.024075502529740334,
0.025203049182891846,
-0.0034381982404738665,
0.06284574419260025,
-0.02525499276816845,
-0.0050758360885083675,
0.07016654312610626,
0.047779910266399384,
0.04621000960469246,
0.009655474685132504,
-0.01720241829752922,
-0.1047825813293457,
0.16950392723083496,
-0.0951867327094078,
-0.269941508769989,
-0.17632324993610382,
0.026197833940386772,
0.04035249724984169,
-0.022378476336598396,
0.031619444489479065,
-0.07056326419115067,
-0.10630585998296738,
-0.1060405746102333,
-0.002429972169920802,
0.01714223250746727,
-0.06364088505506516,
-0.0741225928068161,
0.07348573952913284,
0.04382912442088127,
-0.14902326464653015,
0.038552410900592804,
0.055694397538900375,
-0.057955220341682434,
-0.0233661737293005,
0.09118817001581192,
0.12397737801074982,
0.14583967626094818,
-0.021366750821471214,
-0.028626007959246635,
0.029004426673054695,
0.19620531797409058,
-0.13469526171684265,
0.10371150821447372,
0.13814030587673187,
-0.04545360431075096,
0.08360563963651657,
0.1560150384902954,
0.029186224564909935,
-0.08317049592733383,
0.05044832453131676,
0.04082648828625679,
-0.043159641325473785,
-0.2666129767894745,
-0.0534592866897583,
0.012832709588110447,
-0.06255637854337692,
0.09786593168973923,
0.10183793306350708,
0.11542957276105881,
0.034910861402750015,
-0.07166364789009094,
-0.043925940990448,
-0.0058974819257855415,
0.11737963557243347,
-0.05490213260054588,
-0.012639665976166725,
0.07686592638492584,
-0.05086168646812439,
0.005355054512619972,
0.10266812145709991,
0.02973790094256401,
0.17442677915096283,
0.020399179309606552,
0.11231429129838943,
0.06195578724145889,
0.08633565157651901,
0.0007386076031252742,
0.02951662428677082,
0.05147615820169449,
0.017203815281391144,
-0.002300140680745244,
-0.10421168059110641,
-0.006156572140753269,
0.1449710875749588,
0.028103826567530632,
0.029669636860489845,
-0.0018948549404740334,
-0.005003341939300299,
0.05121048167347908,
0.1746254414319992,
-0.011592294089496136,
-0.22072425484657288,
-0.0845772922039032,
0.06936841458082199,
-0.06218599155545235,
-0.12968985736370087,
-0.026130788028240204,
0.045467354357242584,
-0.17519839107990265,
0.026703642681241035,
-0.027433741837739944,
0.0919293761253357,
-0.09345759451389313,
-0.02221956104040146,
0.03687324374914169,
0.084866963326931,
-0.014529162086546421,
0.08703910559415817,
-0.14498743414878845,
0.11886418610811234,
0.02978132851421833,
0.09024628251791,
-0.11081171780824661,
0.07909037172794342,
-0.007550720125436783,
0.009180475026369095,
0.19379350543022156,
-0.011335089802742004,
-0.03514958545565605,
-0.08774717897176743,
-0.11210042238235474,
-0.013537433929741383,
0.12687496840953827,
-0.1243172138929367,
0.08773399889469147,
-0.015198243781924248,
-0.044079482555389404,
0.00937260314822197,
-0.12100647389888763,
-0.17273177206516266,
-0.19628387689590454,
0.05585884302854538,
-0.09575839340686798,
0.025643249973654747,
-0.11914430558681488,
-0.07089093327522278,
-0.02952558360993862,
0.241120383143425,
-0.1745356321334839,
-0.06510113179683685,
-0.1468164622783661,
-0.046294767409563065,
0.1662203073501587,
-0.04437198117375374,
0.0718095526099205,
-0.0208172257989645,
0.20345525443553925,
0.005988610442727804,
-0.004939318168908358,
0.06724198162555695,
-0.08892562240362167,
-0.16873881220817566,
-0.06771010160446167,
0.1510489284992218,
0.11680185794830322,
0.04907919466495514,
-0.002248800592496991,
0.0011772146681323647,
-0.016943959519267082,
-0.1137804463505745,
-0.0033210667315870523,
0.16037839651107788,
0.03878779336810112,
0.025986969470977783,
-0.05243593826889992,
-0.08797456324100494,
-0.06899320334196091,
-0.06853509694337845,
0.06221301481127739,
0.19590823352336884,
-0.10376439243555069,
0.1700313836336136,
0.147536963224411,
-0.07305635511875153,
-0.23175598680973053,
0.035342130810022354,
0.04983805492520332,
0.0014306638622656465,
0.04886869341135025,
-0.18252557516098022,
0.10521943867206573,
0.019543392583727837,
-0.05505957826972008,
0.13485197722911835,
-0.1557481735944748,
-0.1552847921848297,
0.0722852572798729,
0.03904085233807564,
-0.22423844039440155,
-0.1354004591703415,
-0.09622503817081451,
-0.05825018882751465,
-0.14065024256706238,
0.06054598465561867,
-0.002136280992999673,
0.015948504209518433,
0.03500790148973465,
-0.0015643214574083686,
0.027123261243104935,
-0.058935679495334625,
0.18609118461608887,
-0.004065449349582195,
0.020676052197813988,
-0.060264769941568375,
-0.0478842556476593,
0.09839435666799545,
-0.06130504235625267,
0.12208222597837448,
0.004057085141539574,
0.01594383642077446,
-0.10362856835126877,
-0.048314861953258514,
-0.04328322783112526,
0.05154227837920189,
-0.07548051327466965,
-0.10070807486772537,
-0.043625857681035995,
0.08841723203659058,
0.07005169242620468,
-0.03383097052574158,
0.00549331633374095,
-0.07189501076936722,
0.10019614547491074,
0.17795267701148987,
0.17573626339435577,
0.009926567785441875,
-0.07241068035364151,
0.01677953451871872,
-0.04142116755247116,
0.044231921434402466,
-0.2513144314289093,
0.03756171092391014,
0.06098250672221184,
0.029438555240631104,
0.09217222779989243,
-0.020435843616724014,
-0.1820858269929886,
-0.04050002992153168,
0.08094815909862518,
-0.05452597141265869,
-0.22617179155349731,
-0.019085140898823738,
0.0954197570681572,
-0.2020406424999237,
-0.007372708059847355,
0.03995226323604584,
-0.048725228756666183,
-0.023169852793216705,
0.00010950004070764408,
0.06317184865474701,
0.002471912419423461,
0.09773622453212738,
0.0735151618719101,
0.09715340286493301,
-0.08337292820215225,
0.10562895983457565,
0.10150538384914398,
-0.09572599828243256,
0.03605884686112404,
0.06754924356937408,
-0.05300498008728027,
-0.043293699622154236,
0.03665391728281975,
0.033023297786712646,
0.005234600510448217,
-0.060321882367134094,
0.013913018628954887,
-0.036497246474027634,
0.044923391193151474,
0.08326134830713272,
0.03754979372024536,
-0.013354414142668247,
0.06462216377258301,
0.03401726484298706,
-0.10898099094629288,
0.10366570204496384,
0.01731540448963642,
0.04105307161808014,
-0.08384523540735245,
-0.019968897104263306,
0.035425446927547455,
0.030576206743717194,
-0.01765924133360386,
-0.02306121215224266,
-0.02860277332365513,
-0.01614218018949032,
-0.14299540221691132,
-0.023106401786208153,
-0.07243485748767853,
0.006181265693157911,
0.014656842686235905,
-0.031884219497442245,
-0.011233693920075893,
0.02475680410861969,
-0.06979699432849884,
-0.07426341623067856,
-0.006949664559215307,
0.09833318740129471,
-0.15115703642368317,
0.008848577737808228,
0.06907843053340912,
-0.11088496446609497,
0.08190931379795074,
-0.008411259390413761,
0.016245156526565552,
0.022527478635311127,
-0.15448406338691711,
0.05601610988378525,
0.0008648968650959432,
0.01916889287531376,
0.025886621326208115,
-0.16471809148788452,
0.004104440100491047,
-0.04661374166607857,
-0.02149827405810356,
-0.00004464812809601426,
-0.02647159807384014,
-0.12325995415449142,
0.06858719140291214,
-0.015622655861079693,
-0.035931166261434555,
-0.02701525390148163,
0.0539589487016201,
0.07888586074113846,
-0.027474910020828247,
0.10445091128349304,
-0.008690856397151947,
0.04941811040043831,
-0.16801609098911285,
-0.02470702864229679,
-0.04982255399227142,
0.019377702847123146,
0.009884213097393513,
-0.007693959400057793,
0.04183054715394974,
-0.00976533442735672,
0.21883612871170044,
-0.05075952783226967,
0.1607085019350052,
0.05847611650824547,
-0.017352959141135216,
-0.0007513365126214921,
0.06180921941995621,
0.05997028574347496,
0.04658793285489082,
0.009480604901909828,
0.023740366101264954,
-0.022450892254710197,
-0.006695089396089315,
-0.15932634472846985,
0.01890849508345127,
0.14999441802501678,
0.06301083415746689,
0.024745315313339233,
0.05866100639104843,
-0.12775006890296936,
-0.12135478109121323,
0.09311001747846603,
-0.026755332946777344,
0.00928465835750103,
-0.08245618641376495,
0.1358020007610321,
0.14980104565620422,
-0.14000412821769714,
0.05256148427724838,
-0.06134212389588356,
-0.05217423290014267,
-0.10388828068971634,
-0.12032219022512436,
-0.05887215584516525,
-0.053666237741708755,
0.002330566756427288,
-0.03760887682437897,
0.054546963423490524,
0.03344334661960602,
-0.009351172484457493,
-0.00022941511997487396,
0.13597318530082703,
-0.019751882180571556,
-0.0028988157864660025,
0.048313532024621964,
0.03693558648228645,
0.02373051457107067,
-0.05275435373187065,
0.02940409444272518,
0.02539868652820587,
0.032232340425252914,
0.06546790152788162,
0.033412106335163116,
-0.047448933124542236,
0.03804153576493263,
-0.0025254099164158106,
-0.11207924783229828,
0.019641218706965446,
-0.00460948096588254,
-0.0742158442735672,
0.1268945336341858,
0.0407399944961071,
0.010224059224128723,
-0.03741471841931343,
0.24361543357372284,
-0.06653323769569397,
-0.06378097087144852,
-0.13251738250255585,
0.10491154342889786,
-0.0027236645109951496,
0.06476365029811859,
0.023412218317389488,
-0.1284150779247284,
0.005243356805294752,
0.13858191668987274,
0.12181595712900162,
0.0045748427510261536,
0.009228081442415714,
0.0518609918653965,
0.0025186820421367884,
-0.06998204439878464,
0.054019294679164886,
0.06992026418447495,
0.12919506430625916,
-0.07847554981708527,
0.07680778950452805,
0.0006860480643808842,
-0.08370215445756912,
-0.02947772853076458,
0.11312682181596756,
-0.0409729965031147,
0.03491825982928276,
-0.047444481402635574,
0.10916327685117722,
-0.05787910893559456,
-0.29412412643432617,
0.02350960113108158,
-0.09588567912578583,
-0.15202060341835022,
-0.018367812037467957,
0.05944539234042168,
-0.02624768204987049,
0.018029648810625076,
0.06971040368080139,
-0.06011629104614258,
0.20098382234573364,
0.0335683599114418,
-0.07864278554916382,
-0.0664360448718071,
0.04837050288915634,
-0.06564252078533173,
0.2949807047843933,
0.008418165147304535,
0.02863333560526371,
0.10770907253026962,
-0.03253700211644173,
-0.18271861970424652,
0.010723991319537163,
0.1133992001414299,
-0.08056149631738663,
0.08200647681951523,
0.19000613689422607,
-0.012578671798110008,
0.1209007054567337,
0.05294662341475487,
-0.047376248985528946,
0.04217283055186272,
-0.03389401361346245,
-0.051268599927425385,
-0.10752558708190918,
0.058453381061553955,
-0.05909625440835953,
0.15447644889354706,
0.10152646154165268,
-0.05671518296003342,
-0.004550917539745569,
-0.05555408447980881,
0.04875178262591362,
0.01804669201374054,
0.12263146042823792,
0.02951994352042675,
-0.1865430772304535,
0.032826557755470276,
-0.01144319772720337,
0.10186848044395447,
-0.25588861107826233,
-0.08421015739440918,
0.08833149075508118,
-0.011924264021217823,
-0.05105875805020332,
0.10560628771781921,
0.057650718837976456,
0.04243382066488266,
-0.043439045548439026,
-0.10480839014053345,
-0.02186836116015911,
0.14663739502429962,
-0.1469624787569046,
-0.025013303384184837
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# robust_llm_pythia-tt-1b-mz-v0
This model is a fine-tuned version of [EleutherAI/pythia-1b-deduped](https://huggingface.co/EleutherAI/pythia-1b-deduped) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 8
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 1
### Training results
### Framework versions
- Transformers 4.37.1
- Pytorch 2.1.2
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "EleutherAI/pythia-1b-deduped", "model-index": [{"name": "robust_llm_pythia-tt-1b-mz-v0", "results": []}]} | text-classification | AlignmentResearch/robust_llm_pythia-tt-1b-mz-v0 | [
"transformers",
"tensorboard",
"safetensors",
"gpt_neox",
"text-classification",
"generated_from_trainer",
"base_model:EleutherAI/pythia-1b-deduped",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T16:44:14+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #gpt_neox #text-classification #generated_from_trainer #base_model-EleutherAI/pythia-1b-deduped #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# robust_llm_pythia-tt-1b-mz-v0
This model is a fine-tuned version of EleutherAI/pythia-1b-deduped on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 8
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 1
### Training results
### Framework versions
- Transformers 4.37.1
- Pytorch 2.1.2
- Datasets 2.16.1
- Tokenizers 0.15.1
| [
"# robust_llm_pythia-tt-1b-mz-v0\n\nThis model is a fine-tuned version of EleutherAI/pythia-1b-deduped on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 8\n- eval_batch_size: 64\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.1\n- Pytorch 2.1.2\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #gpt_neox #text-classification #generated_from_trainer #base_model-EleutherAI/pythia-1b-deduped #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# robust_llm_pythia-tt-1b-mz-v0\n\nThis model is a fine-tuned version of EleutherAI/pythia-1b-deduped on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 8\n- eval_batch_size: 64\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.1\n- Pytorch 2.1.2\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
87,
49,
6,
12,
8,
3,
90,
4,
30
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #gpt_neox #text-classification #generated_from_trainer #base_model-EleutherAI/pythia-1b-deduped #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# robust_llm_pythia-tt-1b-mz-v0\n\nThis model is a fine-tuned version of EleutherAI/pythia-1b-deduped on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 8\n- eval_batch_size: 64\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1### Training results### Framework versions\n\n- Transformers 4.37.1\n- Pytorch 2.1.2\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.10270266234874725,
0.16874180734157562,
-0.0029664775356650352,
0.08463428914546967,
0.1430559903383255,
0.014525050297379494,
0.11809725314378738,
0.1328437626361847,
-0.07396194338798523,
0.06468554586172104,
0.10123640298843384,
0.08513332903385162,
0.06070046126842499,
0.15351441502571106,
-0.013439248315989971,
-0.23730027675628662,
0.017173849046230316,
0.01126418448984623,
-0.03425763547420502,
0.10775289684534073,
0.12379158288240433,
-0.08498255163431168,
0.08716442435979843,
0.03199028968811035,
-0.13101515173912048,
0.014501919038593769,
-0.033289507031440735,
-0.06500183790922165,
0.08782681822776794,
0.01558997668325901,
0.05989878624677658,
0.012111435644328594,
0.09241782873868942,
-0.19873221218585968,
-0.004826165735721588,
0.0515381284058094,
0.014292331412434578,
0.09204217046499252,
0.044398318976163864,
0.023280717432498932,
0.05846957862377167,
-0.13980086147785187,
0.10003362596035004,
0.026088044047355652,
-0.07849714159965515,
-0.1767137497663498,
-0.07944837957620621,
0.10220907628536224,
0.08604345470666885,
0.10278322547674179,
0.005833053961396217,
0.18155032396316528,
-0.005625489167869091,
0.07479090988636017,
0.1692560613155365,
-0.2627025842666626,
-0.045856062322854996,
0.017874909564852715,
0.06317578256130219,
0.09001582115888596,
-0.0880095511674881,
-0.0022700547706335783,
0.04826999455690384,
0.026599355041980743,
0.10212225466966629,
-0.0009534807177260518,
0.022518740966916084,
-0.034924816340208054,
-0.12295190989971161,
-0.06443926692008972,
0.19585107266902924,
0.06187467649579048,
-0.0694226399064064,
-0.12213287502527237,
-0.0478115938603878,
-0.1039276272058487,
-0.01832164078950882,
-0.05074326694011688,
0.018088029697537422,
-0.04679003730416298,
-0.03613143041729927,
-0.05981636047363281,
-0.080504409968853,
-0.037289053201675415,
0.03909432142972946,
0.13174298405647278,
0.04770854860544205,
0.016101939603686333,
-0.00023914904159028083,
0.08977748453617096,
-0.02188502624630928,
-0.14138706028461456,
-0.022018935531377792,
-0.0038402786012738943,
-0.03373437374830246,
-0.046414438635110855,
-0.02496710605919361,
-0.030199352651834488,
-0.0006354585057124496,
0.13597270846366882,
-0.039227236062288284,
0.05373309180140495,
0.03654204681515694,
-0.0025199581868946552,
-0.006674044765532017,
0.15188486874103546,
-0.028856193646788597,
-0.03942869231104851,
0.01723104901611805,
0.12149354070425034,
0.03378212824463844,
-0.027452021837234497,
-0.10596802830696106,
-0.01372514758259058,
0.1322353333234787,
0.06295840442180634,
-0.001966864336282015,
0.033806972205638885,
-0.06308358162641525,
-0.05785556882619858,
0.1179642602801323,
-0.12317174673080444,
0.019428370520472527,
-0.03723303973674774,
-0.07269167900085449,
-0.10749013721942902,
0.028093058615922928,
0.021267950534820557,
-0.039565250277519226,
0.016065578907728195,
-0.08124049007892609,
-0.0217118039727211,
-0.05755721777677536,
-0.04883448779582977,
-0.00820277538150549,
-0.05383693426847458,
0.026065504178404808,
-0.09956487268209457,
-0.19905072450637817,
-0.025644320994615555,
0.026542076840996742,
-0.056991513818502426,
-0.10288064926862717,
-0.007258227095007896,
-0.05007527396082878,
0.018297629430890083,
-0.01154283620417118,
0.07418404519557953,
-0.03399087116122246,
0.06560473889112473,
0.04466560110449791,
0.005191049538552761,
0.019275058060884476,
0.044719792902469635,
-0.08898064494132996,
0.04193703830242157,
-0.09841722995042801,
0.0830191820859909,
-0.07832720875740051,
0.04187479987740517,
-0.13412342965602875,
-0.10004778951406479,
0.03267308697104454,
-0.030320575460791588,
0.0655764788389206,
0.10817417502403259,
-0.10503976047039032,
-0.022769220173358917,
0.10449499636888504,
-0.07935401052236557,
-0.1285906732082367,
0.10359034687280655,
-0.0193795096129179,
0.047608308494091034,
0.05717873200774193,
0.15617255866527557,
0.13022629916667938,
-0.1089768335223198,
-0.026409590616822243,
0.027486853301525116,
0.06874291598796844,
0.00800400972366333,
0.09906629472970963,
0.005935306195169687,
0.015852652490139008,
0.011738787405192852,
-0.0728834718465805,
-0.016939004883170128,
-0.07350149005651474,
-0.09214742481708527,
-0.05025169625878334,
-0.08786796033382416,
0.05857887491583824,
0.025302495807409286,
0.03224852308630943,
-0.06085716933012009,
-0.12786518037319183,
0.06606104224920273,
0.13413196802139282,
-0.04670528322458267,
0.0070695276372134686,
-0.09136373549699783,
0.043500009924173355,
-0.06151002272963524,
-0.026014020666480064,
-0.18726390600204468,
-0.09134174883365631,
0.057947322726249695,
-0.08485125750303268,
0.03267265483736992,
-0.004281199537217617,
0.044203244149684906,
0.08350836485624313,
-0.036230988800525665,
-0.057515960186719894,
-0.10529042780399323,
0.0053194607608020306,
-0.11654404550790787,
-0.14790865778923035,
-0.06028023362159729,
-0.03502175956964493,
0.144606813788414,
-0.20987683534622192,
0.029279397800564766,
-0.0006380573031492531,
0.11333302408456802,
0.016106100752949715,
-0.057875510305166245,
0.005114114377647638,
0.015654589980840683,
-0.014581413008272648,
-0.09880195558071136,
0.03155449032783508,
0.022195519879460335,
-0.11554542928934097,
-0.03309289738535881,
-0.1574172079563141,
0.12510478496551514,
0.07290144264698029,
0.08415871113538742,
-0.0742364227771759,
-0.0037178611382842064,
-0.05746135488152504,
-0.049398861825466156,
-0.03863343968987465,
-0.022828547284007072,
0.19185274839401245,
-0.0013857265003025532,
0.134867325425148,
-0.064252570271492,
-0.06494492292404175,
0.013895450159907341,
-0.00897108856588602,
-0.041851844638586044,
0.07375900447368622,
-0.016681229695677757,
-0.18481718003749847,
0.10827518254518509,
0.10787811130285263,
-0.019127849489450455,
0.12399370968341827,
-0.04343444108963013,
-0.0771525502204895,
-0.04274412617087364,
0.00014925013238098472,
0.008228745311498642,
0.09844878315925598,
-0.09154286980628967,
0.004227132070809603,
0.050900839269161224,
0.005066118203103542,
0.025210581719875336,
-0.1402888000011444,
0.0009215839090757072,
0.04868634045124054,
-0.038218531757593155,
0.002842747140675783,
-0.0277363620698452,
-0.017497798427939415,
0.06561096012592316,
0.032556988298892975,
-0.014186030253767967,
0.03944319114089012,
-0.0008745408267714083,
-0.07613305002450943,
0.16902031004428864,
-0.10270246863365173,
-0.2030700445175171,
-0.13985520601272583,
0.03652132675051689,
-0.06334562599658966,
-0.0015506047056987882,
0.030569981783628464,
-0.05810100585222244,
-0.06511329114437103,
-0.08813945949077606,
-0.027047352865338326,
-0.0733569860458374,
-0.00048517368850298226,
0.08776890486478806,
0.0010555826593190432,
0.10515667498111725,
-0.12109791487455368,
0.0026043755933642387,
0.01543970126658678,
-0.05699045956134796,
-0.012240257114171982,
0.03977949917316437,
0.10499977320432663,
0.08143327385187149,
-0.01883314736187458,
0.008249621838331223,
-0.024321144446730614,
0.2691459059715271,
-0.08355075120925903,
-0.019160788506269455,
0.1695500910282135,
-0.0029823908116668463,
0.07592404633760452,
0.11272869259119034,
0.023157570511102676,
-0.08038212358951569,
0.028466159477829933,
0.018591688945889473,
-0.016887471079826355,
-0.2307988703250885,
-0.03963087871670723,
-0.011158764362335205,
-0.05204591155052185,
0.1179073229432106,
0.060640983283519745,
0.06731367111206055,
0.07613514363765717,
-0.03322335705161095,
0.05784502997994423,
-0.019441138952970505,
0.10757894814014435,
0.09813141077756882,
0.04321616142988205,
0.1023724228143692,
-0.02407846599817276,
-0.02622903324663639,
0.05445775017142296,
0.03420785814523697,
0.20841938257217407,
-0.021816181018948555,
0.19311197102069855,
0.02036091312766075,
0.14306433498859406,
-0.013364569284021854,
0.02024158276617527,
0.01945437118411064,
0.011165769770741463,
0.0030712648294866085,
-0.05366143584251404,
-0.04540542885661125,
0.03871443495154381,
-0.004191289655864239,
0.041697002947330475,
-0.07728754729032516,
0.05047864839434624,
0.003477053949609399,
0.2120600938796997,
0.03254234045743942,
-0.33274614810943604,
-0.10219252854585648,
0.020675385370850563,
-0.02397846058011055,
-0.08593863248825073,
-0.020453153178095818,
0.06364496052265167,
-0.140343576669693,
0.0689707100391388,
-0.05951521918177605,
0.0876287966966629,
-0.07652781903743744,
0.00756708113476634,
0.05390635132789612,
0.0903196781873703,
0.004256442654877901,
0.11204465478658676,
-0.20389914512634277,
0.19166138768196106,
0.021203070878982544,
0.09142898768186569,
-0.06806773692369461,
0.050015754997730255,
-0.004243485163897276,
0.09537971019744873,
0.11571414768695831,
-0.004534090403467417,
-0.024725845083594322,
-0.18102821707725525,
-0.11528880894184113,
0.009890151210129261,
0.08745688199996948,
-0.07753730565309525,
0.07411753386259079,
-0.044537320733070374,
0.011558073572814465,
0.029059916734695435,
-0.047375116497278214,
-0.1589600145816803,
-0.14438185095787048,
0.05320568010210991,
-0.00952189788222313,
-0.0007262508734129369,
-0.0885474681854248,
-0.10365280508995056,
-0.01570938713848591,
0.19644959270954132,
0.012212165631353855,
-0.0842258632183075,
-0.14504176378250122,
0.0796443298459053,
0.11172551661729813,
-0.08655772358179092,
0.02340887114405632,
0.0035303349141031504,
0.14993330836296082,
0.027522090822458267,
-0.06209792569279671,
0.056286975741386414,
-0.07633674889802933,
-0.1737402230501175,
-0.03528805822134018,
0.1262955665588379,
0.03132152557373047,
0.044529251754283905,
0.015708284452557564,
0.015004956163465977,
-0.014333968050777912,
-0.07729265838861465,
0.017514580860733986,
0.02022513374686241,
0.08826250582933426,
0.028411865234375,
-0.013812201097607613,
0.03906989097595215,
-0.07633158564567566,
-0.025459473952651024,
0.12449856102466583,
0.22796277701854706,
-0.06877218931913376,
0.058453891426324844,
0.07151654362678528,
-0.0705859437584877,
-0.1570013016462326,
-0.008786638267338276,
0.12869547307491302,
0.010764413513243198,
0.06798017770051956,
-0.16796040534973145,
0.0717378631234169,
0.07791312783956528,
-0.03898453712463379,
0.04473140090703964,
-0.282575398683548,
-0.13420076668262482,
0.08857543766498566,
0.11980879306793213,
0.021312594413757324,
-0.14218378067016602,
-0.06889162212610245,
-0.039159324020147324,
-0.11645139008760452,
0.08400484174489975,
-0.023486746475100517,
0.09570926427841187,
-0.016232673078775406,
0.05926522612571716,
0.02712203375995159,
-0.028170112520456314,
0.17520952224731445,
0.01586144231259823,
0.048582494258880615,
-0.07498310506343842,
0.04772850126028061,
0.11279761046171188,
-0.07670336961746216,
0.08803791552782059,
-0.05301816761493683,
0.07990814000368118,
-0.16324158012866974,
-0.02808978594839573,
-0.04854438826441765,
0.052684713155031204,
-0.054629404097795486,
-0.06426487118005753,
-0.024162691086530685,
0.041498176753520966,
0.049493640661239624,
-0.03528624027967453,
0.11948872357606888,
0.05222162976861,
0.06644876301288605,
0.14638452231884003,
0.08080877363681793,
0.018875090405344963,
-0.15403783321380615,
-0.03380339965224266,
-0.0244415532797575,
0.06078547611832619,
-0.11012988537549973,
0.02348010055720806,
0.09843690693378448,
0.022776268422603607,
0.12430283427238464,
0.012789566069841385,
-0.0727030336856842,
0.0014312989078462124,
0.030104339122772217,
-0.09188268333673477,
-0.16852131485939026,
-0.047760289162397385,
0.026048604398965836,
-0.15849952399730682,
0.026016436517238617,
0.13133805990219116,
-0.062644824385643,
-0.03155265748500824,
-0.019236575812101364,
0.02059292048215866,
-0.0021334246266633272,
0.17186713218688965,
0.057106874883174896,
0.07930057495832443,
-0.07170573621988297,
0.10432305932044983,
0.09961135685443878,
-0.04814169183373451,
0.054860763251781464,
0.023696087300777435,
-0.09018795192241669,
-0.033837106078863144,
0.05106664448976517,
0.11106032133102417,
-0.03880881145596504,
-0.04949856549501419,
-0.08229111135005951,
-0.061456549912691116,
0.041241634637117386,
0.017557138577103615,
0.05725282430648804,
-0.009219215251505375,
-0.027872461825609207,
0.009657464921474457,
-0.12121527642011642,
0.12111517041921616,
0.039653927087783813,
0.0891428291797638,
-0.1959625482559204,
0.03365766257047653,
0.013683971017599106,
0.0697857216000557,
-0.014991731382906437,
0.009885584935545921,
-0.059874989092350006,
-0.025177590548992157,
-0.10460713505744934,
0.01661013811826706,
-0.02413570135831833,
0.0072158160619437695,
-0.02401565946638584,
-0.07077469676733017,
-0.025755932554602623,
0.06438712030649185,
-0.05186127498745918,
-0.08022120594978333,
0.016646932810544968,
0.051620353013277054,
-0.11699030548334122,
-0.019925057888031006,
0.031077943742275238,
-0.0989261344075203,
0.09893514961004257,
0.03507973253726959,
0.025729484856128693,
0.0008266782970167696,
-0.007973370142281055,
0.04753711819648743,
0.03227017819881439,
0.026760606095194817,
0.04111489653587341,
-0.11020812392234802,
-0.01984482631087303,
-0.024211741983890533,
0.005627226084470749,
0.010178839787840843,
0.10276474058628082,
-0.1356809437274933,
-0.0478011779487133,
-0.033691518008708954,
-0.029196256771683693,
-0.06336336582899094,
0.061842333525419235,
0.07014171779155731,
0.015659401193261147,
0.16821108758449554,
-0.05482947453856468,
0.03505853936076164,
-0.19619008898735046,
-0.01802356168627739,
-0.0011967127211391926,
-0.043368011713027954,
-0.0420573428273201,
-0.03858419507741928,
0.06811036169528961,
-0.05248182266950607,
0.10868004709482193,
-0.0050301095470786095,
0.11455348879098892,
0.03606928884983063,
0.01396489143371582,
0.025147126987576485,
0.0027244111988693476,
0.17999020218849182,
0.046786174178123474,
0.0013089182320982218,
0.09400994330644608,
-0.02141028270125389,
0.0632278323173523,
0.01467794831842184,
0.09549760818481445,
0.1284891664981842,
-0.01644987426698208,
0.05263058841228485,
0.04502406343817711,
-0.06743081659078598,
-0.21114829182624817,
0.028478054329752922,
-0.025094421580433846,
0.112427718937397,
-0.0154119823127985,
0.10739672929048538,
0.11382858455181122,
-0.1494397521018982,
0.03770740330219269,
-0.04313240200281143,
-0.11113819479942322,
-0.07739704102277756,
-0.1071854680776596,
-0.08372669667005539,
-0.1070547103881836,
0.013342613354325294,
-0.11941526085138321,
0.00759840477257967,
0.08901376277208328,
-0.013182339258491993,
-0.013210956938564777,
0.15246045589447021,
0.0029203915037214756,
-0.005184302572160959,
0.04251137375831604,
-0.0012815268710255623,
-0.01800457388162613,
-0.03547406569123268,
-0.04682951793074608,
0.04197738319635391,
0.05395640805363655,
0.08222255855798721,
-0.045776329934597015,
-0.010781874880194664,
0.03929838910698891,
-0.0050467015244066715,
-0.09363290667533875,
0.015581349842250347,
0.02201797440648079,
0.017452286556363106,
0.05199722200632095,
0.02267368882894516,
0.007003049366176128,
-0.03528323024511337,
0.26897791028022766,
-0.04828987643122673,
-0.06664728373289108,
-0.11218958348035812,
0.1607970893383026,
0.03197452053427696,
-0.022113507613539696,
0.08127259463071823,
-0.10678350180387497,
0.006815172266215086,
0.12110878527164459,
0.13785097002983093,
-0.046660587191581726,
-0.020635385066270828,
-0.005083917640149593,
-0.01644916832447052,
-0.03582248464226723,
0.10186357796192169,
0.10104076564311981,
-0.007296425756067038,
-0.06594665348529816,
0.02223147265613079,
-0.014690998010337353,
-0.02944192849099636,
-0.09237265586853027,
0.07568122446537018,
0.008258368819952011,
0.018623357638716698,
-0.03418062627315521,
0.06936698406934738,
0.023122508078813553,
-0.1675555258989334,
0.003979284316301346,
-0.14672315120697021,
-0.1794782131910324,
-0.013658768497407436,
0.08857981115579605,
-0.007964780554175377,
0.047400880604982376,
0.019993625581264496,
0.009024309925734997,
0.09706911444664001,
-0.011673921719193459,
-0.06400605291128159,
-0.06223248317837715,
0.09451219439506531,
-0.08417164534330368,
0.23711369931697845,
-0.008227160200476646,
0.0834544375538826,
0.10485610365867615,
-0.008888295851647854,
-0.15066269040107727,
0.03994212672114372,
0.07112805545330048,
0.013810986652970314,
0.06656908243894577,
0.15939024090766907,
-0.018707890063524246,
0.05510655790567398,
0.037786420434713364,
-0.1330747753381729,
-0.05714531987905502,
-0.03353644534945488,
0.013705349527299404,
-0.06261057406663895,
-0.0004963777610100806,
-0.07264385372400284,
0.16452258825302124,
0.15890058875083923,
-0.06734610348939896,
-0.019927378743886948,
-0.05778143182396889,
0.02625160850584507,
0.06190874055027962,
0.08358198404312134,
-0.008416315540671349,
-0.2004786878824234,
-0.003973379265516996,
0.02264978736639023,
0.04426705092191696,
-0.24110552668571472,
-0.09965842217206955,
0.025396540760993958,
-0.05059634894132614,
-0.05921784043312073,
0.09928872436285019,
0.04917570948600769,
0.006980197969824076,
-0.045489661395549774,
-0.055279191583395004,
-0.07573115825653076,
0.12135764956474304,
-0.15435534715652466,
-0.060096967965364456
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | token-classification | MahtaFetrat/ner_transformer | [
"transformers",
"safetensors",
"bert",
"token-classification",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T16:44:45+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #bert #token-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #bert #token-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
47,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #bert #token-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.07082098722457886,
0.16636763513088226,
-0.0037270276807248592,
0.022060682997107506,
0.11734284460544586,
0.008460916578769684,
0.07778386771678925,
0.1078316792845726,
-0.02381012961268425,
0.12548619508743286,
0.03924554958939552,
0.10134156793355942,
0.10977756232023239,
0.19151130318641663,
0.002889276947826147,
-0.20817671716213226,
0.06115591153502464,
-0.1132790669798851,
0.009579826146364212,
0.12119550257921219,
0.14193499088287354,
-0.10718522220849991,
0.07162317633628845,
-0.038196589797735214,
-0.022544773295521736,
-0.031230665743350983,
-0.06254390627145767,
-0.06019896641373634,
0.06759835034608841,
0.060479529201984406,
0.0683075413107872,
0.021969657391309738,
0.0818997174501419,
-0.290944904088974,
0.019316690042614937,
0.07844683527946472,
0.004303961992263794,
0.06271221488714218,
0.07749470323324203,
-0.06878788024187088,
0.1121593713760376,
-0.05535507947206497,
0.15642251074314117,
0.07355572283267975,
-0.09398005902767181,
-0.18405179679393768,
-0.08185496181249619,
0.09494835883378983,
0.16383428871631622,
0.054388850927352905,
-0.03432890772819519,
0.14370861649513245,
-0.08036281913518906,
0.015431041829288006,
0.06882333010435104,
-0.07400276511907578,
-0.05343891680240631,
0.049595899879932404,
0.07492288202047348,
0.0931338220834732,
-0.13102465867996216,
-0.009580682963132858,
0.04242422804236412,
0.01917601190507412,
0.10738006234169006,
0.022504670545458794,
0.11513842642307281,
0.029516039416193962,
-0.1412474811077118,
-0.06101495027542114,
0.1207977756857872,
0.03242048993706703,
-0.05944650247693062,
-0.23808568716049194,
-0.005649200174957514,
-0.029130032286047935,
-0.02337740920484066,
-0.0440535806119442,
0.04105823487043381,
-0.031655408442020416,
0.08197494596242905,
0.006574035622179508,
-0.07103563845157623,
-0.05099502205848694,
0.09157908707857132,
0.059767186641693115,
0.025323325768113136,
-0.026500985026359558,
0.025504639372229576,
0.11718565225601196,
0.10022415220737457,
-0.11435261368751526,
-0.06446252018213272,
-0.06473218649625778,
-0.08744332939386368,
-0.04900897666811943,
0.03775005787611008,
0.0746912732720375,
0.04692034795880318,
0.19667848944664001,
0.005056249443441629,
0.05229094624519348,
0.030662083998322487,
0.014548277482390404,
0.06446343660354614,
0.07018620520830154,
-0.049711957573890686,
-0.12860210239887238,
-0.03947031870484352,
0.11897419393062592,
0.003330475650727749,
-0.033055614680051804,
-0.036121055483818054,
0.06198437139391899,
0.05603967607021332,
0.11939563602209091,
0.0618903674185276,
0.01788514479994774,
-0.06942655146121979,
-0.04313560202717781,
0.18259960412979126,
-0.1554141789674759,
0.022392934188246727,
0.015976034104824066,
-0.053847651928663254,
-0.042034100741147995,
0.01837879791855812,
0.008730842731893063,
-0.027687160298228264,
0.10565981268882751,
-0.06779567897319794,
-0.03990737348794937,
-0.10613231360912323,
-0.054058920592069626,
0.03368525952100754,
-0.019066810607910156,
-0.02883755788207054,
-0.04252570495009422,
-0.11520764976739883,
-0.07606863230466843,
0.06881143897771835,
-0.06148010492324829,
-0.0683162733912468,
-0.03660375624895096,
-0.05812487751245499,
0.012003005482256413,
0.0009617454488761723,
0.12335163354873657,
-0.02907939814031124,
0.04741102084517479,
-0.0517788864672184,
0.06723218411207199,
0.1344490498304367,
0.0335865393280983,
-0.0704164132475853,
0.06614815443754196,
-0.21231821179389954,
0.10163930058479309,
-0.09822800755500793,
0.031383901834487915,
-0.16311906278133392,
-0.0271411444991827,
0.032033201307058334,
0.036491744220256805,
-0.011380859650671482,
0.14042168855667114,
-0.1806737184524536,
-0.037693966180086136,
0.17895962297916412,
-0.1299365758895874,
-0.09427639842033386,
0.061841338872909546,
-0.06028208136558533,
0.13248777389526367,
0.053482506424188614,
-0.024331221356987953,
0.058300819247961044,
-0.1352057307958603,
-0.023484881967306137,
-0.057480666786432266,
-0.004754678346216679,
0.14702697098255157,
0.06274795532226562,
-0.05368298292160034,
0.02439025789499283,
0.018859071657061577,
-0.02367074228823185,
-0.04955499991774559,
-0.03540559858083725,
-0.09829455614089966,
0.007938322611153126,
-0.07976736128330231,
0.020528096705675125,
-0.01857740432024002,
-0.08529462665319443,
-0.03999420255422592,
-0.15634937584400177,
0.008778427727520466,
0.09782235324382782,
0.004635256715118885,
-0.02929295226931572,
-0.09311733394861221,
0.0021291705779731274,
0.014696736820042133,
-0.012728636153042316,
-0.14929775893688202,
-0.052893172949552536,
0.028985487297177315,
-0.16789935529232025,
0.03190700709819794,
-0.04834321513772011,
0.03575235232710838,
0.044794779270887375,
-0.04638027772307396,
-0.024846170097589493,
0.014177534729242325,
0.01915307715535164,
-0.024902109056711197,
-0.24534796178340912,
-0.016663696616888046,
-0.04930651932954788,
0.1770370900630951,
-0.24963445961475372,
0.04489697143435478,
0.062020692974328995,
0.1189032718539238,
0.0055616977624595165,
-0.0471452921628952,
0.03818141296505928,
-0.04912863299250603,
-0.040758419781923294,
-0.06522603332996368,
-0.0019246427109465003,
-0.033303845673799515,
-0.044918134808540344,
0.039850424975156784,
-0.1884368509054184,
-0.023593923076987267,
0.11044006049633026,
0.0722019225358963,
-0.17060783505439758,
-0.07832318544387817,
-0.032338351011276245,
-0.06064004451036453,
-0.08828365802764893,
-0.049234788864851,
0.09958060085773468,
0.04130701348185539,
0.05415206402540207,
-0.07162940502166748,
-0.05484432354569435,
0.013278530910611153,
-0.009936448186635971,
-0.034494977444410324,
0.09010061621665955,
0.08425097167491913,
-0.12193500250577927,
0.1044570803642273,
0.07009463757276535,
0.06440216302871704,
0.10408224165439606,
0.005283535458147526,
-0.09505786001682281,
-0.01272535603493452,
0.025559913367033005,
0.014358514919877052,
0.14346157014369965,
-0.07576300948858261,
0.02965460903942585,
0.042172882705926895,
-0.030027620494365692,
0.010098968632519245,
-0.10260028392076492,
0.019325416535139084,
0.03055759333074093,
-0.008464050479233265,
0.01970001310110092,
-0.05618233606219292,
0.013696934096515179,
0.10435303300619125,
0.0349164679646492,
0.026620987802743912,
0.017225060611963272,
-0.03990183025598526,
-0.1257268637418747,
0.17883455753326416,
-0.09718716144561768,
-0.2507709264755249,
-0.1324487328529358,
0.0005234793643467128,
0.04483891651034355,
-0.012933991849422455,
0.017141954973340034,
-0.05853249877691269,
-0.10673926025629044,
-0.10451403260231018,
0.02033991925418377,
0.054273948073387146,
-0.08803524821996689,
-0.06322101503610611,
0.0517018586397171,
0.03850249573588371,
-0.12421286106109619,
0.023155538365244865,
0.043988488614559174,
-0.07024580985307693,
0.00508910370990634,
0.05607360973954201,
0.08257793635129929,
0.17975331842899323,
0.011003134772181511,
-0.016949951648712158,
0.009263384155929089,
0.21750681102275848,
-0.14687077701091766,
0.0918775200843811,
0.13497301936149597,
-0.06259950995445251,
0.08381292968988419,
0.20346537232398987,
0.030857183039188385,
-0.09484723210334778,
0.03926195576786995,
0.03446268290281296,
-0.03740749508142471,
-0.24119141697883606,
-0.07486692816019058,
0.0031155261676758528,
-0.06816263496875763,
0.10543552786111832,
0.09081115573644638,
0.1144072636961937,
0.05188077315688133,
-0.1067470982670784,
-0.06758806109428406,
0.04753170907497406,
0.11911741644144058,
-0.027111025527119637,
0.003231929149478674,
0.09419949352741241,
-0.030448026955127716,
0.02105054259300232,
0.09140504896640778,
0.01745041273534298,
0.18582363426685333,
0.04117530956864357,
0.1312573403120041,
0.08528119325637817,
0.06527690589427948,
0.019173473119735718,
0.020444748923182487,
0.02246721275150776,
0.030073346570134163,
-0.020628679543733597,
-0.0852246806025505,
-0.012953821569681168,
0.14249984920024872,
0.02702030912041664,
0.032547831535339355,
0.004362224601209164,
-0.04016058146953583,
0.06746432930231094,
0.16617386043071747,
0.012980788946151733,
-0.22532860934734344,
-0.06538809835910797,
0.07354681193828583,
-0.07265309989452362,
-0.11321462690830231,
-0.01038071047514677,
0.030757596716284752,
-0.18158452212810516,
0.042576394975185394,
-0.02550625614821911,
0.10107572376728058,
-0.10972700268030167,
-0.02512514591217041,
0.042610276490449905,
0.06378325074911118,
-0.03664805367588997,
0.07849454134702682,
-0.20421163737773895,
0.14535386860370636,
0.006891076453030109,
0.06414555013179779,
-0.10753445327281952,
0.08170121163129807,
0.02090337499976158,
0.0046083019115030766,
0.16387850046157837,
-0.005854498129338026,
-0.0786028653383255,
-0.08882030844688416,
-0.07770101726055145,
-0.013747241348028183,
0.09857609122991562,
-0.10934799164533615,
0.08609026670455933,
-0.008221019990742207,
-0.032629311084747314,
-0.001329872291535139,
-0.11837238818407059,
-0.13177089393138885,
-0.18219637870788574,
0.051819708198308945,
-0.11911281198263168,
0.03897477313876152,
-0.11066468805074692,
-0.06379573792219162,
-0.036669451743364334,
0.19371679425239563,
-0.1956738978624344,
-0.08014166355133057,
-0.14646820724010468,
-0.07350575923919678,
0.11828155070543289,
-0.04158575087785721,
0.08056027442216873,
0.004819251596927643,
0.2022314816713333,
-0.0027081877924501896,
0.0012655918253585696,
0.08942532539367676,
-0.0949636846780777,
-0.20782062411308289,
-0.09535717219114304,
0.13889843225479126,
0.12820616364479065,
0.0447649285197258,
-0.0019121951190754771,
0.023472661152482033,
-0.002058375161141157,
-0.10908003151416779,
0.030727434903383255,
0.14770722389221191,
0.09537331014871597,
0.03949853777885437,
-0.028519228100776672,
-0.13996201753616333,
-0.10342669486999512,
-0.05459153279662132,
0.01654287800192833,
0.18560625612735748,
-0.07000812143087387,
0.16719648241996765,
0.15820586681365967,
-0.06586025655269623,
-0.20936474204063416,
0.03423137962818146,
0.03405798226594925,
-0.010427549481391907,
0.036926332861185074,
-0.20477096736431122,
0.07846766710281372,
0.016825877130031586,
-0.058902256190776825,
0.13370154798030853,
-0.16832934319972992,
-0.14904731512069702,
0.08974714577198029,
0.07688850909471512,
-0.2126045972108841,
-0.13182798027992249,
-0.09637613594532013,
-0.0503227598965168,
-0.1043887659907341,
0.09036606550216675,
0.006274270825088024,
0.00610304856672883,
0.03730666637420654,
0.021433580666780472,
0.0180149395018816,
-0.0519413948059082,
0.191897913813591,
-0.0013519321801140904,
0.0444704107940197,
-0.07892096042633057,
-0.0851464793086052,
0.03333723545074463,
-0.06510572135448456,
0.0794898197054863,
-0.02122758887708187,
0.0036784426774829626,
-0.11556956171989441,
-0.06427493691444397,
-0.04983310401439667,
0.03419099003076553,
-0.08840304613113403,
-0.0971493199467659,
-0.054171670228242874,
0.10596323013305664,
0.09103043377399445,
-0.035947684198617935,
-0.06095254793763161,
-0.09454575926065445,
0.07212961465120316,
0.2215559333562851,
0.1878495216369629,
0.07139308750629425,
-0.07100050896406174,
-0.002558534499257803,
-0.024434298276901245,
0.055652521550655365,
-0.20899704098701477,
0.046719521284103394,
0.040578074753284454,
0.03033704310655594,
0.13299931585788727,
-0.024206025525927544,
-0.15996594727039337,
-0.04795686900615692,
0.057683661580085754,
-0.06730669736862183,
-0.1570315808057785,
0.0025158768985420465,
0.08647031337022781,
-0.16013643145561218,
-0.051073893904685974,
0.02699451893568039,
-0.03499506786465645,
-0.028059793636202812,
0.002373971976339817,
0.08113706111907959,
0.025904107838869095,
0.11215173453092575,
0.07152648270130157,
0.11194757372140884,
-0.10030562430620193,
0.08277413994073868,
0.0892009288072586,
-0.10862851142883301,
0.03717753291130066,
0.06824232637882233,
-0.06286703795194626,
-0.03321940451860428,
0.030618587508797646,
0.08509371429681778,
0.029280737042427063,
-0.0731777623295784,
0.00199119676835835,
-0.10816600918769836,
0.06575141847133636,
0.14125216007232666,
0.0349125936627388,
0.004502575378865004,
0.04510723426938057,
0.031499505043029785,
-0.10004210472106934,
0.11529461294412613,
0.04151454567909241,
0.0373414121568203,
-0.051681190729141235,
0.0027241462375968695,
0.0408521331846714,
-0.01100252103060484,
-0.016924580559134483,
-0.03830192610621452,
-0.06845806539058685,
-0.010795616544783115,
-0.15674056112766266,
0.026378106325864792,
-0.06946871429681778,
0.009862695820629597,
0.0168803371489048,
-0.03252917155623436,
0.004998120479285717,
0.009924137964844704,
-0.07712483406066895,
-0.03860313072800636,
-0.004113807342946529,
0.10856198519468307,
-0.16059570014476776,
0.007967021316289902,
0.08694947510957718,
-0.12389717996120453,
0.07975487411022186,
-0.007367887068539858,
-0.008897624909877777,
0.018169350922107697,
-0.1400168091058731,
0.06403058022260666,
-0.009703104384243488,
0.005139497108757496,
0.024898670613765717,
-0.20380151271820068,
0.0032552045304328203,
-0.04942692071199417,
-0.05625125393271446,
-0.005749912466853857,
-0.03799205273389816,
-0.11167661845684052,
0.10115693509578705,
0.015623382292687893,
-0.08399864286184311,
-0.01796851120889187,
0.05034510791301727,
0.10852757841348648,
-0.056645460426807404,
0.13888587057590485,
-0.021414149552583694,
0.05864132568240166,
-0.17737187445163727,
-0.018431924283504486,
-0.01712135225534439,
0.012450824491679668,
-0.03453206643462181,
-0.008197006769478321,
0.052714504301548004,
-0.017661362886428833,
0.2243673950433731,
-0.022250786423683167,
0.02734900452196598,
0.065990149974823,
0.0005393415340222418,
-0.01577865332365036,
0.0916161760687828,
0.0463450625538826,
0.01792803406715393,
0.018617253750562668,
0.014746556989848614,
-0.04522430896759033,
-0.014044197276234627,
-0.13052548468112946,
0.08218254148960114,
0.16470149159431458,
0.08262880891561508,
-0.005870525259524584,
0.05077839270234108,
-0.11869116127490997,
-0.09111694246530533,
0.09609098732471466,
-0.03314165025949478,
-0.006128490902483463,
-0.05602835491299629,
0.14245474338531494,
0.15311889350414276,
-0.18047599494457245,
0.06635911017656326,
-0.07129369676113129,
-0.05865350365638733,
-0.10782642662525177,
-0.1736646145582199,
-0.06415880471467972,
-0.036063630133867264,
-0.007051798049360514,
-0.0602986179292202,
0.06560327112674713,
0.10788761079311371,
0.012162050232291222,
0.004834584891796112,
0.08647928386926651,
-0.03503880277276039,
0.0057860445231199265,
0.044105999171733856,
0.05439030006527901,
0.01840701512992382,
-0.06706250458955765,
0.00601657759398222,
0.0010539692593738437,
0.038941897451877594,
0.0557982474565506,
0.028882469981908798,
-0.012592652812600136,
0.008558751083910465,
-0.01646186038851738,
-0.10014970600605011,
0.039393551647663116,
-0.0273386612534523,
-0.046960845589637756,
0.14765694737434387,
0.018432708457112312,
-0.0001517597702331841,
-0.02122526988387108,
0.230571910738945,
-0.06766041368246078,
-0.07626979798078537,
-0.13866591453552246,
0.14627838134765625,
-0.0430525541305542,
0.050970423966646194,
0.05007150396704674,
-0.10352291166782379,
0.03566446155309677,
0.14551185071468353,
0.1465195268392563,
-0.028976434841752052,
0.008297421038150787,
0.012487477622926235,
0.004764636047184467,
-0.025961345061659813,
0.05680029094219208,
0.047920409590005875,
0.11735666543245316,
-0.0655166283249855,
0.09315122663974762,
-0.004929847549647093,
-0.08647413551807404,
-0.02131420560181141,
0.13454632461071014,
0.004034739453345537,
0.023980211466550827,
-0.0810064896941185,
0.11921875178813934,
-0.0646674707531929,
-0.2584960162639618,
0.0636199414730072,
-0.0677812248468399,
-0.15369856357574463,
-0.020554309710860252,
0.02254418656229973,
-0.0003442883607931435,
0.021893499419093132,
0.06417982280254364,
-0.06129874289035797,
0.15113437175750732,
0.03735675662755966,
-0.0744810625910759,
-0.07893598079681396,
0.0799177959561348,
-0.08204485476016998,
0.30506592988967896,
0.007174664177000523,
0.04969498887658119,
0.09485418349504471,
-0.03662348538637161,
-0.13314591348171234,
0.03679130598902702,
0.09527922421693802,
-0.059604041278362274,
0.06409622728824615,
0.20145608484745026,
-0.011349550448358059,
0.11938408017158508,
0.07237957417964935,
-0.08331728726625443,
0.05130653828382492,
-0.08248411118984222,
-0.09262096136808395,
-0.09028724581003189,
0.09243662655353546,
-0.06128277629613876,
0.15477608144283295,
0.13094381988048553,
-0.046571265906095505,
0.000665052211843431,
-0.028391189873218536,
0.05242398753762245,
-0.00248725269921124,
0.10953611135482788,
0.026073157787322998,
-0.19417516887187958,
0.031053941696882248,
-0.013271371833980083,
0.10035724937915802,
-0.25008654594421387,
-0.08122535794973373,
0.04168083891272545,
-0.009952341206371784,
-0.05798068270087242,
0.12185300886631012,
0.05373544245958328,
0.049118392169475555,
-0.05500508472323418,
-0.052748557180166245,
-0.005554255098104477,
0.16211286187171936,
-0.10710477083921432,
-0.0014328722609207034
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# wav2vec_RTSplit0208_3
This model is a fine-tuned version of [jonatasgrosman/wav2vec2-large-xlsr-53-japanese](https://huggingface.co/jonatasgrosman/wav2vec2-large-xlsr-53-japanese) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0234
- Wer: 0.2221
- Cer: 0.1631
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 32
- eval_batch_size: 32
- seed: 4
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 1000
- num_epochs: 10
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|
| 3.8578 | 1.0 | 120 | 3.5714 | 0.9484 | 0.9843 |
| 1.5722 | 2.0 | 240 | 1.3923 | 0.9926 | 0.7587 |
| 0.9411 | 3.0 | 360 | 0.7411 | 0.7542 | 0.4648 |
| 0.6235 | 4.0 | 480 | 0.4713 | 0.6691 | 0.3789 |
| 0.4954 | 5.0 | 600 | 0.3408 | 0.5381 | 0.3106 |
| 0.3909 | 6.0 | 720 | 0.2140 | 0.3727 | 0.2213 |
| 0.2891 | 7.0 | 840 | 0.1158 | 0.2806 | 0.1666 |
| 0.2193 | 8.0 | 960 | 0.0602 | 0.2556 | 0.1736 |
| 0.1925 | 9.0 | 1080 | 0.0287 | 0.2206 | 0.1586 |
| 0.1094 | 10.0 | 1200 | 0.0234 | 0.2221 | 0.1631 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.14.6
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["wer"], "base_model": "jonatasgrosman/wav2vec2-large-xlsr-53-japanese", "model-index": [{"name": "wav2vec_RTSplit0208_3", "results": []}]} | automatic-speech-recognition | tndklab/wav2vec_RTSplit0208_3 | [
"transformers",
"safetensors",
"wav2vec2",
"automatic-speech-recognition",
"generated_from_trainer",
"base_model:jonatasgrosman/wav2vec2-large-xlsr-53-japanese",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-07T16:45:55+00:00 | [] | [] | TAGS
#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-jonatasgrosman/wav2vec2-large-xlsr-53-japanese #license-apache-2.0 #endpoints_compatible #region-us
| wav2vec\_RTSplit0208\_3
=======================
This model is a fine-tuned version of jonatasgrosman/wav2vec2-large-xlsr-53-japanese on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0234
* Wer: 0.2221
* Cer: 0.1631
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 32
* eval\_batch\_size: 32
* seed: 4
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 1000
* num\_epochs: 10
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.14.6
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-jonatasgrosman/wav2vec2-large-xlsr-53-japanese #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
80,
116,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-jonatasgrosman/wav2vec2-large-xlsr-53-japanese #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
-0.14118388295173645,
0.1513376235961914,
-0.0005152393132448196,
0.09957455098628998,
0.11795935779809952,
0.008073536679148674,
0.17546747624874115,
0.14973945915699005,
-0.040967848151922226,
0.11099261045455933,
0.11360207200050354,
0.06234046816825867,
0.05541789159178734,
0.19745732843875885,
-0.08206405490636826,
-0.22053459286689758,
0.07701526582241058,
-0.0033247191458940506,
0.010011174716055393,
0.11214284598827362,
0.07103060185909271,
-0.11864254623651505,
0.09036611765623093,
-0.006963435560464859,
-0.1427508145570755,
-0.04181395843625069,
0.016152987256646156,
-0.11031178385019302,
0.10872969776391983,
0.009288477711379528,
0.06582261621952057,
0.03504345938563347,
0.08960989117622375,
-0.18766173720359802,
0.0022469565737992525,
0.017141321673989296,
0.01435944065451622,
0.07424193620681763,
0.04270157217979431,
-0.0012159834150224924,
0.0027513811364769936,
-0.11490482091903687,
0.03699565678834915,
0.015048346482217312,
-0.11695785820484161,
-0.19873808324337006,
-0.07786379754543304,
0.017535261809825897,
0.09932611882686615,
0.08391966670751572,
-0.02103434130549431,
0.12361589074134827,
0.0008332731667906046,
0.07967887073755264,
0.19855915009975433,
-0.3135106563568115,
-0.05437514930963516,
-0.017246698960661888,
0.03944491222500801,
0.08242533355951309,
-0.10150530934333801,
-0.018765252083539963,
0.05041765794157982,
0.021163977682590485,
0.0925753116607666,
-0.03130659833550453,
-0.0336025208234787,
-0.010997597128152847,
-0.12088784575462341,
-0.03848946467041969,
0.1899806410074234,
0.07315327972173691,
-0.06343334913253784,
-0.0812249630689621,
-0.06397605687379837,
-0.12176059931516647,
-0.054642703384160995,
-0.00814018677920103,
0.026548052206635475,
-0.03949858993291855,
-0.09949330240488052,
-0.00489323865622282,
-0.0798637866973877,
-0.09082527458667755,
-0.017095407471060753,
0.17612586915493011,
0.01114688627421856,
0.013611323200166225,
-0.012470031157135963,
0.054258283227682114,
-0.02387041598558426,
-0.18483866751194,
-0.022159598767757416,
0.026931334286928177,
-0.032688405364751816,
-0.014408618211746216,
-0.04372356832027435,
-0.033641211688518524,
0.04413202032446861,
0.1181967481970787,
-0.018560312688350677,
0.06547611951828003,
-0.024242987856268883,
0.001249976921826601,
-0.08463852852582932,
0.182333305478096,
-0.06463606655597687,
-0.06938374042510986,
0.019906673580408096,
0.1272294521331787,
0.06285319477319717,
-0.02319909818470478,
-0.09877349436283112,
-0.008964418433606625,
0.1465861052274704,
0.03547809645533562,
-0.04232742264866829,
0.05053495988249779,
-0.038675304502248764,
-0.014026766642928123,
0.05737104266881943,
-0.12155470997095108,
0.025580810382962227,
0.022164685651659966,
-0.06321313977241516,
-0.02220059186220169,
-0.012167129665613174,
0.013010796159505844,
0.01259292010217905,
0.05222075432538986,
-0.08240532130002975,
0.0036988360807299614,
-0.02336418628692627,
-0.09296302497386932,
0.02680950053036213,
-0.06885657459497452,
0.0002455086214467883,
-0.10807177424430847,
-0.1792655736207962,
-0.017911894246935844,
0.023634862154722214,
-0.048911672085523605,
-0.009945076890289783,
-0.11301852017641068,
-0.0975419208407402,
0.04686196520924568,
-0.022707432508468628,
0.03525571897625923,
-0.07948608696460724,
0.10767754912376404,
0.07949104905128479,
0.08755897730588913,
-0.04064375162124634,
0.026420699432492256,
-0.09513017535209656,
0.03219624236226082,
-0.17416203022003174,
0.07473883777856827,
-0.054269835352897644,
0.034595683217048645,
-0.12033510208129883,
-0.0673440620303154,
0.019730009138584137,
-0.022371632978320122,
0.06979899108409882,
0.14241422712802887,
-0.1906011998653412,
-0.057025082409381866,
0.19722899794578552,
-0.11986931413412094,
-0.14269396662712097,
0.1278074085712433,
-0.03618486225605011,
0.038308508694171906,
0.07103344798088074,
0.22304697334766388,
0.03084309957921505,
-0.10670869052410126,
-0.04001520201563835,
-0.06360205262899399,
0.08328385651111603,
-0.03686724230647087,
0.1115616038441658,
0.005013670772314072,
-0.0014164310414344072,
0.016217360273003578,
-0.08063652366399765,
0.03222700208425522,
-0.07089976221323013,
-0.10016977041959763,
-0.044187162071466446,
-0.10613103210926056,
0.028292791917920113,
0.015585470013320446,
0.05590164288878441,
-0.09868429601192474,
-0.07060223072767258,
0.011771205812692642,
0.10809079557657242,
-0.11718438565731049,
0.012904723174870014,
-0.10417094081640244,
0.09427313506603241,
-0.11367906630039215,
-0.020158156752586365,
-0.15430019795894623,
-0.00410782964900136,
0.053871724754571915,
0.01823066547513008,
0.014241056516766548,
-0.07536637037992477,
0.08248129487037659,
0.07677027583122253,
-0.048812806606292725,
-0.07412240654230118,
-0.00498126819729805,
0.01805686019361019,
-0.06261217594146729,
-0.17385663092136383,
-0.028737887740135193,
-0.05413864552974701,
0.1608741134405136,
-0.16503508388996124,
0.0014996977988630533,
0.008824673481285572,
0.09039325267076492,
0.044107191264629364,
-0.02336837165057659,
0.019647266715765,
0.04836883023381233,
-0.026001546531915665,
-0.07121788710355759,
0.02903486229479313,
0.015071400441229343,
-0.10376740992069244,
0.019600635394454002,
-0.16807952523231506,
0.15009671449661255,
0.1392962485551834,
0.041799139231443405,
-0.05244087427854538,
0.020045755431056023,
-0.014278299175202847,
-0.04192302003502846,
-0.054265860468149185,
-0.015523474663496017,
0.10089807957410812,
0.007953685708343983,
0.12119784951210022,
-0.10283432900905609,
0.015069880522787571,
0.06469476968050003,
-0.027236688882112503,
-0.028428832069039345,
0.08093540370464325,
0.010932014323771,
-0.13942010700702667,
0.13027305901050568,
0.11321879178285599,
-0.07171899825334549,
0.12632238864898682,
-0.061046771705150604,
-0.085269495844841,
-0.05007721483707428,
0.033539436757564545,
0.03383169323205948,
0.13762885332107544,
-0.08172392845153809,
-0.022573819383978844,
0.02135329321026802,
0.021954631432890892,
-0.01604454591870308,
-0.19322632253170013,
-0.01944556087255478,
0.014740300364792347,
-0.09451686590909958,
-0.009623108431696892,
0.005852391943335533,
-0.017728494480252266,
0.09443813562393188,
-0.0006852737278677523,
-0.11360033601522446,
0.02314845845103264,
-0.015061916783452034,
-0.0876469686627388,
0.17204688489437103,
-0.09222397953271866,
-0.17411498725414276,
-0.13528600335121155,
-0.07141708582639694,
-0.05600440874695778,
0.03755852207541466,
0.06046575680375099,
-0.06563454121351242,
-0.04083354398608208,
-0.11549428105354309,
-0.04752589762210846,
0.03249216452240944,
0.04562458395957947,
0.05096840113401413,
-0.008819636888802052,
0.06599970906972885,
-0.08180946111679077,
-0.005118743982166052,
-0.014260591939091682,
-0.007273803930729628,
0.028529318049550056,
0.0005596327246166766,
0.12637819349765778,
0.12142695486545563,
0.00641365023329854,
0.024710198864340782,
-0.03795645385980606,
0.22719408571720123,
-0.06981682777404785,
-0.019021375104784966,
0.12310279160737991,
-0.027643905952572823,
0.04546043276786804,
0.17790329456329346,
0.03071616031229496,
-0.10723342001438141,
0.001543888240121305,
-0.049799270927906036,
-0.015234281308948994,
-0.188801571726799,
-0.033299293369054794,
-0.04766744375228882,
0.013789367862045765,
0.10153107345104218,
0.029864661395549774,
0.01635241135954857,
0.04831988736987114,
0.02163584530353546,
0.045850567519664764,
0.004909783601760864,
0.08106916397809982,
0.09610755741596222,
0.07607916742563248,
0.10785672068595886,
-0.03219561278820038,
-0.04924085736274719,
0.03241001442074776,
0.02096749283373356,
0.2015942931175232,
0.03050840273499489,
0.19284749031066895,
0.0009453566162846982,
0.15461201965808868,
0.025513680651783943,
0.08029980212450027,
0.0196499302983284,
0.009997223503887653,
-0.020813269540667534,
-0.07838824391365051,
-0.053045131266117096,
0.05521805211901665,
-0.013811358250677586,
0.06097419187426567,
-0.10551154613494873,
0.02013915218412876,
0.050434961915016174,
0.2723030149936676,
0.08769560605287552,
-0.3685109615325928,
-0.08635305613279343,
0.02104947343468666,
-0.036972131580114365,
-0.01903359405696392,
0.016876420006155968,
0.1534101963043213,
-0.06155303493142128,
0.06892576813697815,
-0.07162059843540192,
0.06354092806577682,
-0.06427314132452011,
0.01977919228374958,
0.023796986788511276,
0.047146882861852646,
0.0028003251645714045,
0.030816223472356796,
-0.24248063564300537,
0.28648892045021057,
0.035792022943496704,
0.09525445103645325,
-0.05690842866897583,
-0.0036345715634524822,
0.03943183273077011,
-0.0068437037989497185,
0.11822443455457687,
-0.024794165045022964,
-0.11079125106334686,
-0.1794157326221466,
-0.13511526584625244,
0.048938214778900146,
0.10588322579860687,
-0.006933995056897402,
0.11613228172063828,
-0.013944382779300213,
-0.04456436634063721,
0.04456821084022522,
-0.02340116538107395,
-0.08081210404634476,
-0.0749657079577446,
0.009459982626140118,
0.11387364566326141,
0.044953037053346634,
-0.04946205019950867,
-0.09598170965909958,
-0.09000002592802048,
0.08921240270137787,
0.0022167686838656664,
-0.006378722842782736,
-0.10505329072475433,
0.018605045974254608,
0.1496697962284088,
-0.09152472764253616,
0.053013674914836884,
0.00948717724531889,
0.1103808656334877,
0.027091750875115395,
-0.04968217760324478,
0.09032276272773743,
-0.06209796667098999,
-0.17874783277511597,
-0.050470758229494095,
0.13884679973125458,
-0.007241794373840094,
0.04262242838740349,
0.021225502714514732,
0.05142869055271149,
-0.005594416055828333,
-0.06703758984804153,
0.03248269483447075,
0.0272601880133152,
0.04040838032960892,
0.020026197656989098,
-0.012260128743946552,
-0.09217273443937302,
-0.09288010746240616,
-0.022662481293082237,
0.15083232522010803,
0.2982642650604248,
-0.06597831845283508,
0.01851712167263031,
0.08686181157827377,
-0.017854949459433556,
-0.15149927139282227,
-0.004584764130413532,
0.04337412118911743,
0.044433437287807465,
-0.005071667488664389,
-0.12238527834415436,
0.04546479135751724,
0.061537906527519226,
-0.04462991654872894,
0.07658969610929489,
-0.24918095767498016,
-0.1279260665178299,
0.08928412944078445,
0.1333279311656952,
0.12402694672346115,
-0.15273624658584595,
-0.06697630137205124,
-0.024110818281769753,
-0.10794417560100555,
0.1045057401061058,
-0.07326747477054596,
0.132745623588562,
-0.0021595852449536324,
0.06386832147836685,
0.0073677487671375275,
-0.05131489410996437,
0.15022529661655426,
0.02315061166882515,
0.05349516496062279,
-0.02223295159637928,
-0.01721254177391529,
0.04753045365214348,
-0.07513923943042755,
0.06926781684160233,
-0.0858917385339737,
0.049377862364053726,
-0.05993237346410751,
-0.024863220751285553,
-0.06171049550175667,
-0.006019624415785074,
0.0036315324250608683,
-0.03500806540250778,
-0.010318802669644356,
0.03614702448248863,
0.05882398411631584,
0.003448877949267626,
0.13278301060199738,
0.01197106670588255,
0.08229319751262665,
0.14570775628089905,
0.0886906087398529,
-0.03906968981027603,
0.014958192594349384,
-0.006075490266084671,
-0.05644846335053444,
0.053838834166526794,
-0.13194115459918976,
0.04858151823282242,
0.09669843316078186,
0.018267197534441948,
0.16071775555610657,
0.04687272757291794,
-0.0492333360016346,
0.03808420151472092,
0.06964091956615448,
-0.15898656845092773,
-0.1112910732626915,
0.003452348755672574,
-0.01375506166368723,
-0.11069352924823761,
0.04991121590137482,
0.13865762948989868,
-0.0703156515955925,
-0.007155262865126133,
-0.01823554001748562,
0.021476034075021744,
-0.03947385773062706,
0.20047150552272797,
0.042284511029720306,
0.05118047818541527,
-0.10997112095355988,
0.08147826790809631,
0.05683886259794235,
-0.08807548880577087,
0.049443088471889496,
0.037657495588064194,
-0.11523519456386566,
-0.022784452885389328,
0.00005534075535251759,
0.14294663071632385,
0.005469065625220537,
-0.07627619802951813,
-0.13811787962913513,
-0.08861514180898666,
0.035080134868621826,
0.17778432369232178,
0.06808756291866302,
0.03646450862288475,
-0.01790330372750759,
-0.0017073694616556168,
-0.10371081531047821,
0.09405805170536041,
0.07413934916257858,
0.07491617649793625,
-0.14983569085597992,
0.08254153281450272,
-0.008294850587844849,
0.0264718197286129,
-0.02044260874390602,
0.016870073974132538,
-0.11009109020233154,
0.0050047170370817184,
-0.09842979907989502,
0.05644430220127106,
-0.07891111820936203,
-0.016069279983639717,
-0.0012212666915729642,
-0.08156505972146988,
-0.06149281933903694,
0.011379055678844452,
-0.08712979406118393,
-0.026130566373467445,
0.003458258928731084,
0.0433938093483448,
-0.13683374226093292,
-0.03750019147992134,
0.02246333472430706,
-0.0981111079454422,
0.08394405245780945,
0.0866144597530365,
-0.019790008664131165,
0.04684387892484665,
-0.09545118361711502,
-0.021657824516296387,
0.08283203840255737,
0.0022664524149149656,
0.05060122534632683,
-0.14470165967941284,
-0.013973113149404526,
0.03143203258514404,
0.05041022598743439,
0.021516086533665657,
0.14788220822811127,
-0.09649581462144852,
0.005971268285065889,
-0.06748102605342865,
-0.011066189967095852,
-0.05691782385110855,
0.021555546671152115,
0.14140185713768005,
0.002627637470141053,
0.18363331258296967,
-0.09465426206588745,
0.022600574418902397,
-0.19713501632213593,
0.0013647646410390735,
-0.03708446025848389,
-0.12615914642810822,
-0.14735087752342224,
-0.02729809284210205,
0.07842589169740677,
-0.062242452055215836,
0.09589572995901108,
-0.06168156489729881,
0.06943611055612564,
0.013083240017294884,
-0.05813242867588997,
-0.0013656788505613804,
0.04068639129400253,
0.24912409484386444,
0.057672739028930664,
-0.03611792251467705,
0.07662001997232437,
0.010403653606772423,
0.09453791379928589,
0.12455988675355911,
0.12363877147436142,
0.15776370465755463,
0.03212352469563484,
0.14337152242660522,
0.08356481045484543,
-0.024862807244062424,
-0.11890481412410736,
0.05916835740208626,
-0.06845055520534515,
0.0900888592004776,
0.025596898049116135,
0.2101191282272339,
0.09875553101301193,
-0.1637156754732132,
0.004310329910367727,
-0.03623693808913231,
-0.08481713384389877,
-0.09586827456951141,
-0.06154704466462135,
-0.13099995255470276,
-0.14432699978351593,
0.010924887843430042,
-0.1069321557879448,
0.03407773748040199,
0.06964482367038727,
0.014476785436272621,
0.00010840524919331074,
0.14145208895206451,
0.014773874543607235,
0.029025593772530556,
0.09571029245853424,
0.008392676711082458,
-0.04039377346634865,
-0.0008097448153421283,
-0.10300909727811813,
0.023898394778370857,
0.00501815602183342,
0.056743279099464417,
-0.021962566301226616,
-0.02582181803882122,
0.06900105625391006,
-0.025983864441514015,
-0.1252528429031372,
0.011205081827938557,
0.020153336226940155,
0.06013365089893341,
0.04367395490407944,
0.056713055819272995,
-0.017620805650949478,
0.025014393031597137,
0.20733173191547394,
-0.08926752954721451,
-0.07744365930557251,
-0.13346822559833527,
0.14780977368354797,
-0.0138862831518054,
-0.007439719047397375,
0.009504212997853756,
-0.10589496046304703,
0.00224756239913404,
0.1942233294248581,
0.1499839723110199,
-0.07326409965753555,
-0.0006150471162982285,
-0.02711077593266964,
-0.00640704995021224,
-0.03820818290114403,
0.0650281012058258,
0.07840298116207123,
0.034390028566122055,
-0.059590261429548264,
-0.06017826497554779,
-0.05757676437497139,
-0.041262898594141006,
-0.022971007972955704,
0.03991897031664848,
-0.0317038968205452,
-0.023118194192647934,
-0.04992280900478363,
0.07865259051322937,
-0.08221094310283661,
-0.0975438579916954,
0.00718031357973814,
-0.21697042882442474,
-0.17321449518203735,
-0.0021544182673096657,
0.07518170028924942,
0.03575673699378967,
0.026161469519138336,
-0.033357031643390656,
0.02538921684026718,
0.05649259686470032,
-0.013258103281259537,
-0.057701025158166885,
-0.05936339870095253,
0.04269890859723091,
-0.08285795897245407,
0.17427952587604523,
-0.00461580278351903,
0.06635239720344543,
0.1040571853518486,
0.08133987337350845,
-0.10723624378442764,
0.10374195873737335,
0.06103026121854782,
-0.07403004914522171,
0.055740538984537125,
0.15157853066921234,
-0.056069888174533844,
0.14287318289279938,
0.05193907395005226,
-0.10281984508037567,
0.001149781746789813,
0.010053437203168869,
-0.028709089383482933,
-0.07485763728618622,
-0.06516216695308685,
-0.045741770416498184,
0.14665868878364563,
0.13426938652992249,
-0.06621623039245605,
0.0017491946928203106,
-0.016887987032532692,
0.056069597601890564,
0.06269390136003494,
0.02119249664247036,
-0.06178002804517746,
-0.2836269438266754,
-0.016548514366149902,
0.03701779246330261,
0.022455256432294846,
-0.2418002039194107,
-0.08923479914665222,
-0.010862430557608604,
-0.04599393531680107,
-0.07464209198951721,
0.0931960716843605,
0.08006857335567474,
0.03134746104478836,
-0.054390132427215576,
-0.05154174938797951,
-0.028746988624334335,
0.17312726378440857,
-0.1632525771856308,
-0.11506713181734085
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# segformer-b5-finetuned-segments-instryde-foot-test
This model is a fine-tuned version of [nvidia/mit-b5](https://huggingface.co/nvidia/mit-b5) on the inStryde/inStrydeSegmentationFoot dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0149
- Mean Iou: 0.4800
- Mean Accuracy: 0.9599
- Overall Accuracy: 0.9599
- Per Category Iou: [0.0, 0.9599216842864238]
- Per Category Accuracy: [nan, 0.9599216842864238]
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 6e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 50
### Training results
| Training Loss | Epoch | Step | Validation Loss | Mean Iou | Mean Accuracy | Overall Accuracy | Per Category Iou | Per Category Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:-------------:|:----------------:|:-------------------------:|:-------------------------:|
| 0.1024 | 0.27 | 20 | 0.2085 | 0.4534 | 0.9067 | 0.9067 | [0.0, 0.9067344993758137] | [nan, 0.9067344993758137] |
| 0.0431 | 0.53 | 40 | 0.0487 | 0.4604 | 0.9207 | 0.9207 | [0.0, 0.9207331455341442] | [nan, 0.9207331455341442] |
| 0.0354 | 0.8 | 60 | 0.0319 | 0.4577 | 0.9155 | 0.9155 | [0.0, 0.9154662028576415] | [nan, 0.9154662028576415] |
| 0.0389 | 1.07 | 80 | 0.0276 | 0.4629 | 0.9257 | 0.9257 | [0.0, 0.9257162800419576] | [nan, 0.9257162800419576] |
| 0.0208 | 1.33 | 100 | 0.0244 | 0.4702 | 0.9404 | 0.9404 | [0.0, 0.9403945317069335] | [nan, 0.9403945317069335] |
| 0.0241 | 1.6 | 120 | 0.0212 | 0.4703 | 0.9406 | 0.9406 | [0.0, 0.9406131407017349] | [nan, 0.9406131407017349] |
| 0.0167 | 1.87 | 140 | 0.0208 | 0.4761 | 0.9521 | 0.9521 | [0.0, 0.9521215619420916] | [nan, 0.9521215619420916] |
| 0.0156 | 2.13 | 160 | 0.0205 | 0.4612 | 0.9224 | 0.9224 | [0.0, 0.9224359945462809] | [nan, 0.9224359945462809] |
| 0.0156 | 2.4 | 180 | 0.0208 | 0.4734 | 0.9468 | 0.9468 | [0.0, 0.9467575875538612] | [nan, 0.9467575875538612] |
| 0.0167 | 2.67 | 200 | 0.0182 | 0.4833 | 0.9667 | 0.9667 | [0.0, 0.9666659635383208] | [nan, 0.9666659635383208] |
| 0.0145 | 2.93 | 220 | 0.0243 | 0.4351 | 0.8702 | 0.8702 | [0.0, 0.8702122233110058] | [nan, 0.8702122233110058] |
| 0.0114 | 3.2 | 240 | 0.0176 | 0.4686 | 0.9373 | 0.9373 | [0.0, 0.93726765603217] | [nan, 0.93726765603217] |
| 0.0155 | 3.47 | 260 | 0.0161 | 0.4770 | 0.9541 | 0.9541 | [0.0, 0.9540767701096305] | [nan, 0.9540767701096305] |
| 0.0158 | 3.73 | 280 | 0.0169 | 0.4684 | 0.9368 | 0.9368 | [0.0, 0.9368239181251786] | [nan, 0.9368239181251786] |
| 0.0114 | 4.0 | 300 | 0.0162 | 0.4777 | 0.9554 | 0.9554 | [0.0, 0.9554348305492647] | [nan, 0.9554348305492647] |
| 0.0112 | 4.27 | 320 | 0.0159 | 0.4839 | 0.9678 | 0.9678 | [0.0, 0.9677532556440432] | [nan, 0.9677532556440432] |
| 0.0131 | 4.53 | 340 | 0.0154 | 0.4811 | 0.9622 | 0.9622 | [0.0, 0.9622032718479555] | [nan, 0.9622032718479555] |
| 0.0101 | 4.8 | 360 | 0.0156 | 0.4683 | 0.9367 | 0.9367 | [0.0, 0.9366846987126999] | [nan, 0.9366846987126999] |
| 0.0102 | 5.07 | 380 | 0.0152 | 0.4758 | 0.9517 | 0.9517 | [0.0, 0.9516509773164403] | [nan, 0.9516509773164403] |
| 0.0101 | 5.33 | 400 | 0.0169 | 0.4884 | 0.9768 | 0.9768 | [0.0, 0.9768393358121804] | [nan, 0.9768393358121804] |
| 0.0082 | 5.6 | 420 | 0.0150 | 0.4761 | 0.9522 | 0.9522 | [0.0, 0.9522462074215836] | [nan, 0.9522462074215836] |
| 0.01 | 5.87 | 440 | 0.0152 | 0.4788 | 0.9576 | 0.9576 | [0.0, 0.9575745140264517] | [nan, 0.9575745140264517] |
| 0.0098 | 6.13 | 460 | 0.0148 | 0.4783 | 0.9565 | 0.9565 | [0.0, 0.9565489693736469] | [nan, 0.9565489693736469] |
| 0.0088 | 6.4 | 480 | 0.0153 | 0.4795 | 0.9591 | 0.9591 | [0.0, 0.959051850601846] | [nan, 0.959051850601846] |
| 0.0091 | 6.67 | 500 | 0.0152 | 0.4828 | 0.9656 | 0.9656 | [0.0, 0.965590177169167] | [nan, 0.965590177169167] |
| 0.0102 | 6.93 | 520 | 0.0149 | 0.4800 | 0.9599 | 0.9599 | [0.0, 0.9599216842864238] | [nan, 0.9599216842864238] |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.0.1
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "other", "tags": ["vision", "image-segmentation", "generated_from_trainer"], "base_model": "nvidia/mit-b5", "model-index": [{"name": "segformer-b5-finetuned-segments-instryde-foot-test", "results": []}]} | image-segmentation | PostsDesert/segformer-b5-finetuned-segments-instryde-foot-test | [
"transformers",
"tensorboard",
"safetensors",
"segformer",
"vision",
"image-segmentation",
"generated_from_trainer",
"base_model:nvidia/mit-b5",
"license:other",
"endpoints_compatible",
"region:us"
] | 2024-02-07T16:47:37+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #segformer #vision #image-segmentation #generated_from_trainer #base_model-nvidia/mit-b5 #license-other #endpoints_compatible #region-us
| segformer-b5-finetuned-segments-instryde-foot-test
==================================================
This model is a fine-tuned version of nvidia/mit-b5 on the inStryde/inStrydeSegmentationFoot dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0149
* Mean Iou: 0.4800
* Mean Accuracy: 0.9599
* Overall Accuracy: 0.9599
* Per Category Iou: [0.0, 0.9599216842864238]
* Per Category Accuracy: [nan, 0.9599216842864238]
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 6e-05
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 50
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.0.1
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 6e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 50",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.0.1\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #segformer #vision #image-segmentation #generated_from_trainer #base_model-nvidia/mit-b5 #license-other #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 6e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 50",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.0.1\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
62,
98,
4,
30
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #segformer #vision #image-segmentation #generated_from_trainer #base_model-nvidia/mit-b5 #license-other #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 6e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 50### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.0.1\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.11522962898015976,
0.045119550079107285,
-0.0015287681017071009,
0.10776930302381516,
0.15040777623653412,
0.017196819186210632,
0.12917137145996094,
0.10620410740375519,
-0.08689790964126587,
0.0472620464861393,
0.12240622937679291,
0.12363925576210022,
0.008453181013464928,
0.10828991234302521,
-0.02343696542084217,
-0.2789453864097595,
-0.0005422983667813241,
0.048347312957048416,
-0.08657379448413849,
0.10327548533678055,
0.07919933646917343,
-0.1632670909166336,
0.09516878426074982,
-0.0015763052506372333,
-0.22293394804000854,
0.02841956354677677,
0.021568171679973602,
-0.038023266941308975,
0.14793531596660614,
0.04922318831086159,
0.1184476763010025,
0.015748394653201103,
0.10900945961475372,
-0.19905880093574524,
0.019163724035024643,
0.05856257677078247,
-0.009098980575799942,
0.06651351600885391,
0.05772549659013748,
0.01669321581721306,
0.10819203406572342,
-0.0972309410572052,
0.060222309082746506,
0.009440845809876919,
-0.12840867042541504,
-0.22478610277175903,
-0.047796376049518585,
0.04134181886911392,
0.10639340430498123,
0.09034362435340881,
-0.005513505078852177,
0.10077984631061554,
-0.052839282900094986,
0.10430456697940826,
0.24964497983455658,
-0.25648173689842224,
-0.07182144373655319,
0.05064673721790314,
0.015510090626776218,
0.0820796936750412,
-0.13057655096054077,
0.016233500093221664,
0.06950220465660095,
0.014581045135855675,
0.11119842529296875,
-0.02871013432741165,
-0.01963130012154579,
0.01724052242934704,
-0.1473914086818695,
-0.027879847213625908,
0.12210381031036377,
0.07574456930160522,
-0.021432653069496155,
-0.04207788035273552,
-0.07392657548189163,
-0.14889314770698547,
-0.06858900934457779,
-0.0032709527295082808,
0.05762477219104767,
-0.04853929206728935,
-0.10419248044490814,
-0.03146893531084061,
-0.13319730758666992,
-0.09992583841085434,
-0.03733306750655174,
0.12655974924564362,
0.04729008674621582,
0.025724641978740692,
-0.032020896673202515,
0.10799174755811691,
-0.05113956332206726,
-0.13004054129123688,
0.012772944755852222,
0.020613908767700195,
-0.022990427911281586,
-0.024355389177799225,
-0.05408762767910957,
-0.09551437199115753,
-0.002553701400756836,
0.1205466166138649,
-0.057850658893585205,
0.05448099970817566,
0.04759715870022774,
0.07241658121347427,
-0.11797910183668137,
0.15929526090621948,
-0.08609022200107574,
0.01013142243027687,
-0.009874711744487286,
0.06041751429438591,
0.029564110562205315,
0.0019934880547225475,
-0.1179974153637886,
0.0022776008117944,
0.08790061622858047,
-0.0221173707395792,
-0.0797700583934784,
0.07869765907526016,
-0.04217718169093132,
-0.005649321712553501,
-0.00897669792175293,
-0.06777915358543396,
0.028282323852181435,
-0.009886178188025951,
-0.06601089984178543,
-0.06121760979294777,
0.05857963487505913,
0.004772757180035114,
0.023268427699804306,
0.11157344281673431,
-0.09787540137767792,
0.03728562593460083,
-0.10533478111028671,
-0.09677588194608688,
0.007005583960562944,
-0.059697121381759644,
0.0473722442984581,
-0.10324554145336151,
-0.15087664127349854,
-0.006352177821099758,
0.05902674421668053,
-0.021341051906347275,
0.006394261494278908,
-0.04703628644347191,
-0.08299009501934052,
-0.00021471075888257474,
-0.007894876413047314,
0.0947788655757904,
-0.05050712078809738,
0.1123909279704094,
0.06118451803922653,
0.07845046371221542,
-0.07072953879833221,
0.02361133322119713,
-0.0957041010260582,
0.03623959422111511,
-0.1911376714706421,
0.02206907421350479,
-0.05513213947415352,
0.06997141242027283,
-0.05626712366938591,
-0.09146557003259659,
-0.04131743684411049,
0.01476714201271534,
0.07984724640846252,
0.10956771671772003,
-0.20024529099464417,
-0.07847573608160019,
0.16291022300720215,
-0.10281816124916077,
-0.146260067820549,
0.11827083677053452,
-0.054073888808488846,
0.023997820913791656,
0.06020919606089592,
0.20814163982868195,
0.056549977511167526,
-0.12217041850090027,
0.00997595489025116,
0.002785709220916033,
0.04490534961223602,
-0.05730598792433739,
0.06075204163789749,
0.026435431092977524,
0.05661935731768608,
0.021018093451857567,
-0.08272214978933334,
0.07894562929868698,
-0.11179323494434357,
-0.09016269445419312,
-0.03928397223353386,
-0.10055693984031677,
0.0345521904528141,
0.0921594649553299,
0.07555278390645981,
-0.10630271583795547,
-0.07589341700077057,
0.09587422013282776,
0.07888263463973999,
-0.0911208838224411,
0.03181803598999977,
-0.07311529666185379,
0.06374941021203995,
-0.07913199067115784,
-0.031027983874082565,
-0.14584845304489136,
-0.05711039900779724,
-0.020269213244318962,
0.009358920156955719,
0.013350924476981163,
0.02086537703871727,
0.08854938298463821,
0.08507268130779266,
-0.07631455361843109,
-0.04634253308176994,
-0.05173703283071518,
0.01374211348593235,
-0.1137223169207573,
-0.21027660369873047,
-0.028177812695503235,
-0.019302310422062874,
0.08478841930627823,
-0.24545449018478394,
0.03798803687095642,
0.01743573509156704,
0.10502021014690399,
0.041392914950847626,
-0.028780214488506317,
-0.05647515878081322,
0.0507737472653389,
-0.028263889253139496,
-0.06561795622110367,
0.05807628110051155,
-0.0034010440576821566,
-0.08310561627149582,
-0.04266468435525894,
-0.14023557305335999,
0.20842646062374115,
0.13086695969104767,
-0.16139455139636993,
-0.08707430958747864,
0.000019758730559260584,
-0.055418819189071655,
-0.03996799886226654,
-0.04225844889879227,
-0.004603296983987093,
0.1008872240781784,
-0.021571770310401917,
0.14053292572498322,
-0.07688967883586884,
-0.028997356072068214,
0.042631879448890686,
-0.053108394145965576,
-0.0011033174814656377,
0.11006389558315277,
0.11088067293167114,
-0.09318497776985168,
0.1451481133699417,
0.1548282951116562,
-0.10068725794553757,
0.1553608775138855,
-0.040087051689624786,
-0.08508210629224777,
-0.01219777762889862,
0.015346884727478027,
-0.0012885131873190403,
0.18971461057662964,
-0.1550634354352951,
-0.02042090706527233,
-0.007480934727936983,
0.004231064114719629,
0.028456546366214752,
-0.24848367273807526,
-0.05246448516845703,
0.043028514832258224,
-0.034298427402973175,
0.042410850524902344,
-0.026105990633368492,
-0.03190537542104721,
0.09688082337379456,
-0.015045274049043655,
-0.07718131691217422,
0.027462869882583618,
-0.008917608298361301,
-0.075106181204319,
0.21517036855220795,
-0.053094495087862015,
-0.14476536214351654,
-0.12529109418392181,
-0.029565168544650078,
-0.043670278042554855,
0.03100958652794361,
0.05965929105877876,
-0.08529380708932877,
-0.028072595596313477,
-0.08329986780881882,
0.007217762991786003,
0.014940904453396797,
0.04438621178269386,
0.00921480543911457,
0.005214906297624111,
0.08867418766021729,
-0.08883403241634369,
-0.000006311934157565702,
-0.06367773562669754,
-0.042524367570877075,
0.02637419104576111,
0.027422500774264336,
0.15554608404636383,
0.13742586970329285,
-0.029887765645980835,
0.012804741971194744,
-0.023321760818362236,
0.25248003005981445,
-0.08918806165456772,
-0.02456214651465416,
0.1292060762643814,
-0.012721525505185127,
0.03733988478779793,
0.12329593300819397,
0.07460089772939682,
-0.10408369451761246,
0.006441853474825621,
0.06098012998700142,
-0.04935694858431816,
-0.15393811464309692,
-0.01999957300722599,
-0.04724251478910446,
-0.013728850521147251,
0.08812342584133148,
0.04963665083050728,
0.01548822596669197,
0.07907280325889587,
0.02997702546417713,
0.05678590387105942,
-0.00561687583103776,
0.06880207359790802,
0.10176911950111389,
0.019228415563702583,
0.11142812669277191,
-0.052964866161346436,
-0.06965328007936478,
0.016450554132461548,
0.0241420716047287,
0.22844067215919495,
0.0039204880595207214,
0.12508584558963776,
0.07196608185768127,
0.15452714264392853,
0.000020497611330938525,
0.028087377548217773,
-0.01402156613767147,
-0.08112011849880219,
-0.001445673406124115,
-0.047173772007226944,
-0.003048887476325035,
0.03965003415942192,
-0.08014623820781708,
0.052596740424633026,
-0.11972271651029587,
0.022757546976208687,
0.06393890082836151,
0.19614169001579285,
0.053339194506406784,
-0.3395047187805176,
-0.07672445476055145,
0.009873734787106514,
-0.002745048375800252,
-0.02532857097685337,
0.024944636970758438,
0.1714961975812912,
-0.036541931331157684,
0.03483828902244568,
-0.08599067479372025,
0.07562639564275742,
-0.03878270462155342,
0.04287773370742798,
0.062032248824834824,
0.062205929309129715,
-0.0025613724719733,
0.0643613189458847,
-0.26349321007728577,
0.2909882366657257,
0.007608477026224136,
0.08095508068799973,
-0.03385179117321968,
-0.031393930315971375,
0.009257815778255463,
0.13140256702899933,
0.12343806028366089,
-0.021314382553100586,
-0.04753430560231209,
-0.2022184282541275,
-0.004028012976050377,
0.027332689613103867,
0.1251508891582489,
-0.017982887104153633,
0.09094899892807007,
-0.0187812689691782,
-0.0006124347564764321,
0.09337800741195679,
0.008549035526812077,
-0.06597258895635605,
-0.10241740196943283,
-0.030543651431798935,
-0.004499014467000961,
-0.046717263758182526,
-0.08383186161518097,
-0.09757868945598602,
-0.13643299043178558,
0.153356671333313,
-0.03651273250579834,
-0.0088682621717453,
-0.11082742363214493,
0.10944954305887222,
0.05286415293812752,
-0.08013872057199478,
0.05929797515273094,
0.03059956803917885,
0.09300214052200317,
0.027772197499871254,
-0.06309796124696732,
0.1273898184299469,
-0.06990378350019455,
-0.1470472663640976,
-0.06448645144701004,
0.09655141830444336,
0.020843196660280228,
0.024876683950424194,
-0.00773590337485075,
-0.002876593731343746,
-0.0073194801807403564,
-0.07961773872375488,
0.04987094923853874,
-0.0077717420645058155,
0.05038582161068916,
-0.011867749504745007,
-0.03923274949193001,
0.039005544036626816,
-0.0418127104640007,
-0.015828847885131836,
0.13650643825531006,
0.29614242911338806,
-0.08787977695465088,
-0.032928526401519775,
0.05437492951750755,
-0.07402503490447998,
-0.21734090149402618,
0.07803621888160706,
0.03690911829471588,
0.0017496817745268345,
0.07627788931131363,
-0.1300903558731079,
0.10052914172410965,
0.11430087685585022,
-0.023237314075231552,
0.09652674943208694,
-0.33694738149642944,
-0.1359233260154724,
0.10682126134634018,
0.20983366668224335,
0.11226800084114075,
-0.1701432466506958,
-0.019542725756764412,
-0.020931098610162735,
-0.11856095492839813,
0.08394411951303482,
-0.07366445660591125,
0.1193404495716095,
-0.012393295764923096,
0.0340598002076149,
0.01603214628994465,
-0.06980672478675842,
0.1291123479604721,
-0.02582281455397606,
0.14244480431079865,
-0.06745844334363937,
-0.011061036959290504,
0.08307024836540222,
-0.048958852887153625,
0.010604823008179665,
-0.044930748641490936,
0.04182254895567894,
-0.031054822728037834,
-0.02970724366605282,
-0.07672884315252304,
0.02984335646033287,
-0.02727937512099743,
-0.06713999807834625,
-0.05988522619009018,
0.0369013287127018,
0.025628114119172096,
0.00002335087992832996,
0.19948334991931915,
0.0034794427920132875,
0.10099322348833084,
0.08792778104543686,
0.0537552647292614,
-0.06668923050165176,
-0.09934936463832855,
-0.0207652784883976,
-0.015345143154263496,
0.0692514181137085,
-0.1529798060655594,
0.03241843730211258,
0.12959106266498566,
0.03658149763941765,
0.1355109065771103,
0.08520742505788803,
-0.0477730929851532,
0.025306159630417824,
0.08761946111917496,
-0.15233881771564484,
-0.16839656233787537,
-0.028071654960513115,
-0.057869844138622284,
-0.0817408561706543,
0.09749705344438553,
0.08038629591464996,
-0.09641619771718979,
0.025334222242236137,
-0.015588300302624702,
-0.000688795989844948,
-0.06125828996300697,
0.19724678993225098,
0.0914202407002449,
0.03131889924407005,
-0.0749015137553215,
0.08693955093622208,
0.009421834722161293,
-0.09283143281936646,
-0.007451636251062155,
0.051093053072690964,
-0.06788326799869537,
-0.03897227719426155,
0.06055005267262459,
0.17780353128910065,
-0.0748908668756485,
-0.049070265144109726,
-0.1618707776069641,
-0.0926733911037445,
0.04565086588263512,
0.1750318557024002,
0.10163482278585434,
0.0019091739086434245,
-0.018095992505550385,
0.03966885432600975,
-0.12290773540735245,
0.10672972351312637,
0.01035825815051794,
0.1056068018078804,
-0.1838977187871933,
0.126662939786911,
0.005799159873276949,
0.03931412845849991,
-0.033838506788015366,
0.043925780802965164,
-0.1241760179400444,
0.028034331277012825,
-0.11453301459550858,
-0.016275789588689804,
-0.02891785092651844,
-0.0025356165133416653,
0.0024606012739241123,
-0.05169818922877312,
-0.06701862812042236,
0.023650558665394783,
-0.11067581176757812,
-0.018746918067336082,
0.055393438786268234,
0.029346268624067307,
-0.11935145407915115,
-0.03209711238741875,
0.014940444380044937,
-0.0586451031267643,
0.04387231916189194,
0.0035160239785909653,
0.03285249322652817,
0.06342782080173492,
-0.20082175731658936,
0.017076902091503143,
0.0959489494562149,
-0.002734471345320344,
0.028625385835766792,
-0.021718861535191536,
-0.020124265924096107,
-0.009951062500476837,
0.05833122134208679,
-0.009564708918333054,
0.03607518970966339,
-0.13152740895748138,
-0.009935259819030762,
-0.049845945090055466,
-0.06404165178537369,
-0.060268376022577286,
0.043651968240737915,
0.049374788999557495,
0.03031422570347786,
0.16877295076847076,
-0.0869491845369339,
0.0065094116143882275,
-0.20914022624492645,
0.0012669810093939304,
0.011784862726926804,
-0.10595715790987015,
-0.055654432624578476,
-0.05651247128844261,
0.058103181421756744,
-0.06973996758460999,
0.12547199428081512,
0.011975418776273727,
0.034647006541490555,
0.032757531851530075,
-0.044236887246370316,
0.05886999890208244,
0.045193422585725784,
0.27444756031036377,
-0.000060630696680163965,
-0.030659295618534088,
0.030960842967033386,
0.08383867889642715,
0.12274773418903351,
0.13231079280376434,
0.14854660630226135,
0.1688086986541748,
-0.10275895893573761,
0.1171489730477333,
0.06760717183351517,
-0.037700217217206955,
-0.15402324497699738,
0.05316004529595375,
-0.053198765963315964,
0.08179549872875214,
-0.03991033136844635,
0.17540770769119263,
0.16212812066078186,
-0.16610680520534515,
0.025003070011734962,
-0.030938036739826202,
-0.09641914814710617,
-0.07907184958457947,
-0.09851536154747009,
-0.09245514124631882,
-0.167576864361763,
0.03181677311658859,
-0.10407406091690063,
0.008032633922994137,
0.12194829434156418,
0.008959691040217876,
-0.030161425471305847,
0.23312216997146606,
0.049652062356472015,
0.02848025970160961,
0.04203454777598381,
0.010940693318843842,
-0.04157126694917679,
-0.07822664082050323,
-0.06539949029684067,
0.04181288555264473,
-0.050846803933382034,
0.000491447513923049,
-0.07736951857805252,
-0.04869883507490158,
0.039715468883514404,
0.020289747044444084,
-0.08424773812294006,
0.009871022775769234,
0.025853777304291725,
0.0659082680940628,
0.021537622436881065,
0.0059014721773564816,
-0.0014714309945702553,
-0.02062542364001274,
0.23799599707126617,
-0.08251440525054932,
-0.054894108325242996,
-0.08823543787002563,
0.17607420682907104,
0.036255694925785065,
0.03336944431066513,
-0.009908913634717464,
-0.11172831803560257,
0.03734038397669792,
0.2278788834810257,
0.15781764686107635,
-0.09499973803758621,
0.003982928581535816,
-0.0009670348954387009,
-0.0050048548728227615,
-0.05921219661831856,
0.10076064616441727,
0.08840340375900269,
0.05422511696815491,
-0.08366434276103973,
-0.056273091584444046,
-0.03591115027666092,
-0.008362323977053165,
-0.04077908396720886,
0.058720964938402176,
0.06438767164945602,
0.02567806839942932,
-0.08672468364238739,
0.06279169768095016,
-0.030810004100203514,
-0.11008474975824356,
0.11288382112979889,
-0.18354251980781555,
-0.12688063085079193,
-0.013668786734342575,
0.13046544790267944,
-0.020097119733691216,
0.041055742651224136,
-0.05559615045785904,
0.006785855162888765,
0.02528642863035202,
0.0019792921375483274,
-0.09622114896774292,
-0.09969740360975266,
0.04384787753224373,
-0.08464968949556351,
0.24714992940425873,
-0.04585273936390877,
0.03660394623875618,
0.11297086626291275,
0.03903621807694435,
-0.06931804120540619,
0.08647772669792175,
0.02903922274708748,
-0.089903824031353,
0.009899980388581753,
0.12362182140350342,
-0.031779900193214417,
0.10007911175489426,
0.03343433886766434,
-0.14550448954105377,
0.018771560862660408,
-0.07946911454200745,
-0.05530785024166107,
-0.05174868926405907,
-0.037468064576387405,
-0.07001344859600067,
0.12067076563835144,
0.16163134574890137,
-0.017531869933009148,
0.006513224449008703,
-0.061974234879016876,
0.03183344006538391,
0.0931103304028511,
0.058987438678741455,
-0.023113243281841278,
-0.21622860431671143,
0.02184651419520378,
0.0662309005856514,
-0.03257593885064125,
-0.21849219501018524,
-0.10847365111112595,
0.014971698634326458,
-0.05789848789572716,
-0.06139020621776581,
0.07721524685621262,
0.15389758348464966,
0.05263340845704079,
-0.06908261775970459,
-0.1261717677116394,
-0.05722204968333244,
0.1543325036764145,
-0.12759827077388763,
-0.1095903292298317
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# mistral_yt_transcribe_classification
This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0363
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-07
- train_batch_size: 1
- eval_batch_size: 1
- seed: 42
- distributed_type: multi-GPU
- num_devices: 6
- gradient_accumulation_steps: 4
- total_train_batch_size: 24
- total_eval_batch_size: 6
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 4
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| 0.0398 | 1.0 | 427 | 0.0380 |
| 0.0268 | 2.0 | 854 | 0.0323 |
| 0.0188 | 3.0 | 1281 | 0.0334 |
| 0.0135 | 4.0 | 1708 | 0.0363 |
### Framework versions
- Transformers 4.36.0
- Pytorch 2.0.1+cu118
- Datasets 2.17.0
- Tokenizers 0.15.2
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "mistralai/Mistral-7B-Instruct-v0.2", "model-index": [{"name": "mistral_yt_transcribe_classification", "results": []}]} | text-generation | hiiamsid/mistral_yt_transcribe_classification | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"generated_from_trainer",
"conversational",
"base_model:mistralai/Mistral-7B-Instruct-v0.2",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T16:51:58+00:00 | [] | [] | TAGS
#transformers #safetensors #mistral #text-generation #generated_from_trainer #conversational #base_model-mistralai/Mistral-7B-Instruct-v0.2 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| mistral\_yt\_transcribe\_classification
=======================================
This model is a fine-tuned version of mistralai/Mistral-7B-Instruct-v0.2 on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0363
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-07
* train\_batch\_size: 1
* eval\_batch\_size: 1
* seed: 42
* distributed\_type: multi-GPU
* num\_devices: 6
* gradient\_accumulation\_steps: 4
* total\_train\_batch\_size: 24
* total\_eval\_batch\_size: 6
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: cosine
* lr\_scheduler\_warmup\_ratio: 0.1
* num\_epochs: 4
### Training results
### Framework versions
* Transformers 4.36.0
* Pytorch 2.0.1+cu118
* Datasets 2.17.0
* Tokenizers 0.15.2
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-07\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 1\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 6\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 24\n* total\\_eval\\_batch\\_size: 6\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 4",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.36.0\n* Pytorch 2.0.1+cu118\n* Datasets 2.17.0\n* Tokenizers 0.15.2"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #generated_from_trainer #conversational #base_model-mistralai/Mistral-7B-Instruct-v0.2 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-07\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 1\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 6\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 24\n* total\\_eval\\_batch\\_size: 6\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 4",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.36.0\n* Pytorch 2.0.1+cu118\n* Datasets 2.17.0\n* Tokenizers 0.15.2"
] | [
85,
180,
4,
35
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #generated_from_trainer #conversational #base_model-mistralai/Mistral-7B-Instruct-v0.2 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-07\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 1\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 6\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 24\n* total\\_eval\\_batch\\_size: 6\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 4### Training results### Framework versions\n\n\n* Transformers 4.36.0\n* Pytorch 2.0.1+cu118\n* Datasets 2.17.0\n* Tokenizers 0.15.2"
] | [
-0.1307399421930313,
0.13677912950515747,
-0.0037298770621418953,
0.07405771315097809,
0.10204345732927322,
0.051774635910987854,
0.13428661227226257,
0.13970041275024414,
-0.04776385426521301,
0.12916187942028046,
0.1003371849656105,
0.052705153822898865,
0.057674188166856766,
0.16420744359493256,
-0.048821792006492615,
-0.2759877145290375,
0.04287699609994888,
-0.0288854893296957,
-0.15990294516086578,
0.09898310899734497,
0.09579319506883621,
-0.09646277874708176,
0.07461854815483093,
-0.009970457293093204,
-0.11224387586116791,
-0.04147733375430107,
-0.027527859434485435,
-0.04176424816250801,
0.08738760650157928,
0.05351299047470093,
0.08128829300403595,
0.030123190954327583,
0.08116676658391953,
-0.2044382244348526,
0.003090400481596589,
0.06894811987876892,
0.012247723527252674,
0.08712676167488098,
0.09511347860097885,
0.005019961390644312,
0.10727574676275253,
-0.08001729846000671,
0.05251122638583183,
0.039077263325452805,
-0.12054139375686646,
-0.25933119654655457,
-0.10727515816688538,
0.08194909244775772,
0.1276680827140808,
0.04296074062585831,
-0.006913635414093733,
0.06822992116212845,
-0.01954694278538227,
0.07121207565069199,
0.24461187422275543,
-0.25471779704093933,
-0.07502580434083939,
0.052583131939172745,
0.02990684285759926,
0.06069401651620865,
-0.0906238779425621,
-0.01992211677134037,
0.041008882224559784,
0.009505226276814938,
0.07938577979803085,
0.01670680195093155,
0.03299357742071152,
-0.005468669347465038,
-0.14465290307998657,
-0.07278816401958466,
0.18805894255638123,
0.06312987953424454,
-0.017994146794080734,
-0.07949813455343246,
-0.06825494766235352,
-0.16037020087242126,
-0.0375671423971653,
-0.0050422376953065395,
0.023933187127113342,
-0.039877913892269135,
-0.03945842385292053,
0.021130753681063652,
-0.05685560032725334,
-0.09348724782466888,
0.027860011905431747,
0.16073274612426758,
0.06404540687799454,
0.003452081000432372,
-0.00014878968067932874,
0.11433766037225723,
0.011240503750741482,
-0.18073207139968872,
-0.03618396073579788,
0.009251363575458527,
-0.0661572590470314,
-0.020749012008309364,
-0.03068060614168644,
0.08011007308959961,
0.053779155015945435,
0.19742782413959503,
-0.09756997972726822,
0.05586812272667885,
0.05252046510577202,
0.001945781521499157,
-0.07316873222589493,
0.13204888999462128,
-0.07421943545341492,
-0.06758890300989151,
-0.016866592690348625,
0.11062831431627274,
0.0319894477725029,
0.0011676709400489926,
-0.06298130750656128,
0.010056378319859505,
0.09768348932266235,
0.057043012231588364,
0.005857916083186865,
0.04805542528629303,
-0.060717057436704636,
-0.03535708039999008,
0.10161758214235306,
-0.11445841193199158,
0.019283201545476913,
0.03468640148639679,
-0.0719948410987854,
-0.03578156605362892,
0.012519437819719315,
-0.012500280514359474,
0.003415168495848775,
0.06428065150976181,
-0.08820182085037231,
-0.045905955135822296,
-0.07664474844932556,
-0.0804823711514473,
0.04057895764708519,
-0.041013602167367935,
-0.005679862108081579,
-0.059689607471227646,
-0.16589559614658356,
-0.02716001868247986,
0.040720704942941666,
-0.06887279450893402,
-0.07992325723171234,
-0.026817617937922478,
-0.10987459868192673,
0.048713065683841705,
-0.012904183939099312,
0.1093430146574974,
-0.05541842058300972,
0.0678386464715004,
0.05058051273226738,
0.0695001408457756,
0.06352652609348297,
0.038464438170194626,
-0.04405129328370094,
0.08746085315942764,
-0.17190159857273102,
0.06790479272603989,
-0.10665928572416306,
0.03533525392413139,
-0.1161593422293663,
-0.09384655207395554,
0.017037859186530113,
-0.020830469205975533,
0.07244488596916199,
0.1399366408586502,
-0.11709950119256973,
-0.07014763355255127,
0.18900510668754578,
-0.12059126794338226,
-0.15836742520332336,
0.10692097991704941,
-0.000569898693356663,
-0.05081962049007416,
0.01301739364862442,
0.09579706937074661,
0.15049166977405548,
-0.09371453523635864,
-0.036687955260276794,
-0.011160912923514843,
0.12156425416469574,
0.035461388528347015,
0.11599301546812057,
-0.004024540074169636,
0.018837066367268562,
0.024449529126286507,
-0.09495604783296585,
0.011869068257510662,
-0.09256762266159058,
-0.08926476538181305,
-0.05396099388599396,
-0.07801490277051926,
0.023212673142552376,
0.04359809681773186,
0.026671288534998894,
-0.07495556771755219,
-0.11944807320833206,
0.024835782125592232,
0.13117021322250366,
-0.07925261557102203,
0.0018285244004800916,
-0.05550774559378624,
0.09533426910638809,
-0.026729561388492584,
0.0032268252689391375,
-0.16162307560443878,
-0.1408732682466507,
0.04611964151263237,
-0.06285091489553452,
-0.024469221010804176,
-0.009799393825232983,
0.06628482788801193,
0.0864616185426712,
-0.05992801487445831,
-0.0613560788333416,
-0.045765697956085205,
-0.00989443063735962,
-0.06081131845712662,
-0.22159487009048462,
-0.08306726813316345,
-0.036255158483982086,
0.16270649433135986,
-0.2137710452079773,
0.03546683490276337,
0.04649168998003006,
0.1472843438386917,
0.018973510712385178,
-0.037494316697120667,
-0.006144507322460413,
0.025585392490029335,
-0.054228294640779495,
-0.0872788354754448,
0.02001243270933628,
-0.0027737088967114687,
-0.08532211184501648,
-0.012481914833188057,
-0.15115727484226227,
0.1277027428150177,
0.08286592364311218,
0.04449525102972984,
-0.07902616262435913,
-0.02666611783206463,
-0.06692119687795639,
-0.05649615451693535,
-0.022486461326479912,
-0.0316585898399353,
0.09050213545560837,
0.003697049804031849,
0.10425405204296112,
-0.08533909916877747,
-0.06310587376356125,
0.03114696592092514,
-0.006537226960062981,
-0.02845892496407032,
0.1216185986995697,
0.025518931448459625,
-0.06060488522052765,
0.14348715543746948,
0.11735597252845764,
-0.06457154452800751,
0.11390715837478638,
-0.08887834846973419,
-0.07677759230136871,
-0.05040214583277702,
0.04057883843779564,
0.03671589493751526,
0.10524777323007584,
-0.0878758504986763,
0.00975453108549118,
0.030270058661699295,
0.02525770664215088,
0.01702887937426567,
-0.1697114109992981,
0.0002929862530436367,
0.02374742552638054,
-0.08440171927213669,
0.03299486264586449,
-0.008425354957580566,
-0.010583469644188881,
0.09793473780155182,
0.012048453092575073,
-0.04082391783595085,
-0.005461123771965504,
-0.019523223862051964,
-0.08692830055952072,
0.22184738516807556,
-0.09507706016302109,
-0.11331018805503845,
-0.15350697934627533,
0.02838989347219467,
-0.06002356857061386,
0.00027561670867726207,
0.02743029035627842,
-0.0628485307097435,
-0.05273215472698212,
-0.08018919080495834,
0.031328365206718445,
-0.005028684623539448,
0.03057451732456684,
0.052752669900655746,
0.011926758103072643,
0.0734352394938469,
-0.10064718127250671,
0.01456283126026392,
0.0034390336368232965,
-0.07931745797395706,
0.013816763646900654,
0.007892942987382412,
0.09212946891784668,
0.13661223649978638,
0.034529123455286026,
0.02104608155786991,
-0.01612774468958378,
0.19401264190673828,
-0.07874799519777298,
0.00897620152682066,
0.12036874890327454,
0.010905313305556774,
0.061310961842536926,
0.14921541512012482,
0.03755262494087219,
-0.0793040543794632,
0.0023513573687523603,
0.026270542293787003,
-0.02097972482442856,
-0.23519256711006165,
-0.04699395224452019,
-0.042785368859767914,
0.02740997076034546,
0.11151115596294403,
0.04014471545815468,
0.012134653516113758,
0.06265634298324585,
-0.0672195702791214,
0.015489251352846622,
0.027180200442671776,
0.08262285590171814,
0.07900682091712952,
0.05892069637775421,
0.11738134175539017,
-0.034363970160484314,
-0.0002828384458553046,
0.052305564284324646,
0.0242654699832201,
0.24778735637664795,
-0.021176287904381752,
0.19238220155239105,
0.04376709833741188,
0.13415250182151794,
-0.0021096584387123585,
0.06472650170326233,
0.007316801231354475,
0.005570380017161369,
0.011102755554020405,
-0.06260434538125992,
-0.006093320436775684,
0.046893563121557236,
-0.029688037931919098,
0.05228264629840851,
-0.09988602995872498,
0.028294729068875313,
0.045491307973861694,
0.2860247492790222,
0.07157725840806961,
-0.32454603910446167,
-0.11643794924020767,
0.042174771428108215,
-0.025102443993091583,
-0.029310544952750206,
0.017287224531173706,
0.12232407927513123,
-0.0725695788860321,
0.10016879439353943,
-0.06782632321119308,
0.0793490931391716,
-0.037774987518787384,
-0.010463859885931015,
0.10462719947099686,
0.09815790504217148,
0.0159713476896286,
0.06140698492527008,
-0.20411741733551025,
0.2756815254688263,
-0.004673033952713013,
0.03537425026297569,
-0.031790316104888916,
0.05941665172576904,
0.019608844071626663,
0.04985855519771576,
0.06931139528751373,
-0.005570358596742153,
-0.13121460378170013,
-0.15790598094463348,
-0.09697718173265457,
0.019372448325157166,
0.11897128075361252,
-0.10339085012674332,
0.13349641859531403,
-0.0369807705283165,
-0.03192278370261192,
0.050729647278785706,
-0.04902485013008118,
-0.0807386264204979,
-0.09466719627380371,
0.053513385355472565,
-0.025915438309311867,
0.03673844039440155,
-0.09176294505596161,
-0.10479767620563507,
-0.09597311913967133,
0.15501664578914642,
-0.13952802121639252,
-0.05812397971749306,
-0.1340859830379486,
0.06921891868114471,
0.17709489166736603,
-0.09927504509687424,
0.05297575891017914,
-0.012032909318804741,
0.13914813101291656,
0.03288191556930542,
-0.05293220281600952,
0.09040616452693939,
-0.09592417627573013,
-0.2665640711784363,
-0.03243456780910492,
0.1385432928800583,
0.015130058862268925,
0.03757677599787712,
-0.025432264432311058,
0.048975542187690735,
-0.0070062056183815,
-0.1063547432422638,
0.025535235181450844,
0.05347229912877083,
0.0777507945895195,
0.04393594339489937,
-0.057190489023923874,
-0.0173800066113472,
-0.02999059483408928,
-0.02792820893228054,
0.058063264936208725,
0.30697816610336304,
-0.10414528846740723,
0.00583687424659729,
0.0418412946164608,
-0.06353329867124557,
-0.16766858100891113,
-0.022302566096186638,
0.1103820651769638,
0.008401600643992424,
-0.0045231846161186695,
-0.18630863726139069,
0.0709521546959877,
0.12057016044855118,
-0.04290253669023514,
0.08939309418201447,
-0.2885008156299591,
-0.13996592164039612,
0.07722358405590057,
0.09265043586492538,
-0.00743511226028204,
-0.20204167068004608,
-0.07716190814971924,
-0.007379909977316856,
-0.1395692080259323,
0.10171961039304733,
-0.03435536101460457,
0.08667285740375519,
-0.02540610171854496,
0.015338738448917866,
0.009869443252682686,
-0.049685921519994736,
0.19396080076694489,
0.03141449764370918,
0.06167095899581909,
-0.024959558621048927,
0.009688127785921097,
0.08580152690410614,
-0.08789733797311783,
0.03647835552692413,
-0.06923762708902359,
0.06626588106155396,
-0.10939259082078934,
-0.013693120330572128,
-0.07056023180484772,
0.026675373315811157,
-0.05807318165898323,
-0.022710202261805534,
-0.03959435224533081,
0.0451999232172966,
0.058923590928316116,
-0.016886359080672264,
0.1557704657316208,
0.0223477091640234,
0.17222820222377777,
0.14912788569927216,
0.06473388522863388,
0.020107503980398178,
-0.07811731100082397,
-0.002630842849612236,
-0.00901767611503601,
0.04427403584122658,
-0.15856774151325226,
0.03706048056483269,
0.12349875271320343,
0.03704863414168358,
0.11509142071008682,
0.05992136523127556,
-0.05594252794981003,
-0.0010127893183380365,
0.07371281832456589,
-0.12662123143672943,
-0.142744779586792,
-0.01690744236111641,
0.011265371926128864,
-0.1678352952003479,
0.06306318938732147,
0.1128714308142662,
-0.04772872477769852,
-0.00606671255081892,
-0.004473301116377115,
0.055458687245845795,
-0.008525875397026539,
0.21266552805900574,
0.04129658639431,
0.10302398353815079,
-0.09149468690156937,
0.10194993764162064,
0.04833188280463219,
-0.10043686628341675,
0.007215934805572033,
0.08636560291051865,
-0.08251925557851791,
-0.02790910378098488,
0.028086485341191292,
0.059888098388910294,
-0.03068406879901886,
-0.04794127494096756,
-0.12332949042320251,
-0.131026029586792,
0.07728555053472519,
0.139677032828331,
0.06122307479381561,
0.06188139691948891,
0.01905401609838009,
0.03233843296766281,
-0.08666415512561798,
0.13932277262210846,
0.07795993238687515,
0.10251981019973755,
-0.1528051346540451,
0.08219235390424728,
-0.005448805633932352,
0.012747819535434246,
-0.014198021031916142,
0.024577302858233452,
-0.12682819366455078,
-0.034968599677085876,
-0.1240362673997879,
0.016396569088101387,
-0.06788241118192673,
0.001397299929521978,
-0.004240148235112429,
-0.064023457467556,
-0.0536666214466095,
0.029762791469693184,
-0.08929634839296341,
-0.042745791375637054,
-0.040037851780653,
0.06600061804056168,
-0.12994179129600525,
-0.012666109018027782,
0.045327432453632355,
-0.12894338369369507,
0.10758651793003082,
0.043970078229904175,
0.05513068288564682,
0.019431589171290398,
-0.05949530750513077,
0.034239646047353745,
0.025478271767497063,
0.014273498207330704,
0.029358340427279472,
-0.17571528255939484,
0.006321236025542021,
-0.03300038352608681,
0.009100369177758694,
-0.0008475108188576996,
0.04163893684744835,
-0.11939531564712524,
0.025923479348421097,
-0.026476988568902016,
-0.04022643715143204,
-0.05259987711906433,
0.03532835841178894,
0.10068228840827942,
-0.012359720654785633,
0.14619232714176178,
-0.0643266960978508,
0.02636025659739971,
-0.24642425775527954,
-0.008420583792030811,
-0.0016236045630648732,
-0.09011907875537872,
-0.09896901994943619,
-0.003885329933837056,
0.08108686655759811,
-0.04916676878929138,
0.11706152558326721,
-0.06205662339925766,
0.03981206193566322,
0.031631842255592346,
-0.03588152304291725,
0.05595667287707329,
0.06228245049715042,
0.17157013714313507,
0.05334294214844704,
-0.01311031635850668,
0.0483580119907856,
-0.00038902080268599093,
0.051913660019636154,
0.00038586644222959876,
0.17257060110569,
0.1136741116642952,
-0.03325905278325081,
0.07048417627811432,
0.07593828439712524,
-0.13051219284534454,
-0.12594103813171387,
0.07274270802736282,
-0.07181128859519958,
0.10661889612674713,
-0.0162676814943552,
0.12231045216321945,
0.12323440611362457,
-0.2150440663099289,
0.03318719193339348,
-0.04030872881412506,
-0.07348641008138657,
-0.11595616489648819,
-0.07215353101491928,
-0.09258571267127991,
-0.1775732785463333,
0.0020865791011601686,
-0.1430482566356659,
0.02634764090180397,
0.11509948968887329,
0.02227475307881832,
0.02948201820254326,
0.12476827949285507,
0.05618865787982941,
0.020206252112984657,
0.04192649573087692,
0.03842099383473396,
0.0005887573352083564,
-0.0072801681235432625,
-0.09751545637845993,
0.02496587485074997,
-0.01358101423829794,
0.04048651084303856,
-0.035479992628097534,
-0.03299444541335106,
0.060880959033966064,
0.011215483769774437,
-0.09600833803415298,
0.013645237311720848,
0.0008026601281017065,
0.027046749368309975,
0.04451540857553482,
0.008893206715583801,
0.0027822775300592184,
-0.013073018752038479,
0.19717171788215637,
-0.08669773489236832,
-0.05071432515978813,
-0.13057111203670502,
0.23585206270217896,
-0.019809149205684662,
-0.008030124008655548,
0.055149346590042114,
-0.06594560295343399,
-0.026852764189243317,
0.1058167815208435,
0.15307636559009552,
-0.012478108517825603,
-0.017276596277952194,
0.01930864527821541,
-0.014444788917899132,
-0.010855119675397873,
0.09005007892847061,
0.1044083833694458,
0.09122078120708466,
-0.05020984634757042,
-0.0405193492770195,
-0.003209995571523905,
-0.02973932959139347,
-0.04940890520811081,
0.06185515224933624,
-0.01334412582218647,
-0.006969884969294071,
-0.01804317906498909,
0.07634725421667099,
-0.044149380177259445,
-0.10144537687301636,
0.06697879731655121,
-0.19841913878917694,
-0.18329297006130219,
-0.04575148597359657,
0.062189456075429916,
-0.0007533298339694738,
0.060573358088731766,
0.0100222984328866,
-0.049397796392440796,
0.12770313024520874,
-0.010039118118584156,
-0.07385053485631943,
-0.11076629161834717,
0.0739782378077507,
-0.07677777856588364,
0.20245759189128876,
-0.03526753932237625,
0.042824190109968185,
0.12314067035913467,
-0.0054474505595862865,
-0.14024534821510315,
0.013575178571045399,
0.09545004367828369,
-0.10539508610963821,
0.03071085549890995,
0.13568730652332306,
-0.048125043511390686,
0.11128170043230057,
0.05663313344120979,
-0.10156615823507309,
-0.03105863556265831,
-0.002630749484524131,
-0.04461996629834175,
-0.06944101303815842,
-0.01515191514045,
-0.049028292298316956,
0.15526233613491058,
0.20672127604484558,
-0.07261480391025543,
-0.021299386397004128,
-0.02881431207060814,
0.049877554178237915,
0.04469089210033417,
0.14552688598632812,
-0.0011329902336001396,
-0.28303781151771545,
0.029156407341361046,
0.02952347882091999,
0.03610832989215851,
-0.1966117024421692,
-0.08321800082921982,
0.03752206265926361,
-0.03279799595475197,
-0.07271978259086609,
0.11570439487695694,
0.07374797761440277,
0.03692798689007759,
-0.05260317400097847,
-0.08248516917228699,
-0.06620084494352341,
0.16890022158622742,
-0.16957668960094452,
-0.08263429999351501
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# feb7th
This model is a fine-tuned version of [facebook/deit-base-distilled-patch16-224](https://huggingface.co/facebook/deit-base-distilled-patch16-224) on the imagefolder dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0464
- Accuracy: 0.9899
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 8
- seed: 1234
- gradient_accumulation_steps: 10
- total_train_batch_size: 160
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 5
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| No log | 0.97 | 12 | 0.0598 | 0.9798 |
| No log | 1.94 | 24 | 0.0480 | 0.9879 |
| No log | 2.98 | 37 | 0.0531 | 0.9838 |
| No log | 3.95 | 49 | 0.0456 | 0.9899 |
| No log | 4.84 | 60 | 0.0464 | 0.9899 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["imagefolder"], "metrics": ["accuracy"], "base_model": "facebook/deit-base-distilled-patch16-224", "model-index": [{"name": "feb7th", "results": [{"task": {"type": "image-classification", "name": "Image Classification"}, "dataset": {"name": "imagefolder", "type": "imagefolder", "config": "default", "split": "train", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.9898785425101214, "name": "Accuracy"}]}]}]} | image-classification | sruthis/feb7th | [
"transformers",
"safetensors",
"deit",
"image-classification",
"generated_from_trainer",
"dataset:imagefolder",
"base_model:facebook/deit-base-distilled-patch16-224",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T16:55:10+00:00 | [] | [] | TAGS
#transformers #safetensors #deit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-facebook/deit-base-distilled-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| feb7th
======
This model is a fine-tuned version of facebook/deit-base-distilled-patch16-224 on the imagefolder dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0464
* Accuracy: 0.9899
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 16
* eval\_batch\_size: 8
* seed: 1234
* gradient\_accumulation\_steps: 10
* total\_train\_batch\_size: 160
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 5
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 1234\n* gradient\\_accumulation\\_steps: 10\n* total\\_train\\_batch\\_size: 160\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #safetensors #deit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-facebook/deit-base-distilled-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 1234\n* gradient\\_accumulation\\_steps: 10\n* total\\_train\\_batch\\_size: 160\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
85,
127,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #deit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-facebook/deit-base-distilled-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 1234\n* gradient\\_accumulation\\_steps: 10\n* total\\_train\\_batch\\_size: 160\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.16963836550712585,
0.16255854070186615,
-0.001816028612665832,
0.08437006175518036,
0.153429314494133,
0.014639150351285934,
0.12409987300634384,
0.12575788795948029,
-0.10849817842245102,
0.08989442139863968,
0.12570831179618835,
0.09025009721517563,
0.05159824341535568,
0.18749752640724182,
-0.0309849102050066,
-0.2889218032360077,
0.02110927738249302,
0.04745285585522652,
-0.1560336947441101,
0.11976606398820877,
0.10957932472229004,
-0.13412903249263763,
0.07529515773057938,
0.026631219312548637,
-0.1913885474205017,
-0.017826996743679047,
-0.01503808330744505,
-0.07817541807889938,
0.10604389011859894,
0.032304201275110245,
0.11641640216112137,
0.02937593311071396,
0.08685502409934998,
-0.15689538419246674,
0.008181668817996979,
0.07504384219646454,
0.011941548436880112,
0.1035267785191536,
0.08538725227117538,
-0.0217935461550951,
0.08187823742628098,
-0.10188613831996918,
0.06605704873800278,
0.03540170192718506,
-0.10512737184762955,
-0.30997586250305176,
-0.09261031448841095,
0.1291607916355133,
0.13805529475212097,
0.06840499490499496,
-0.009604256600141525,
0.06810951977968216,
-0.06701445579528809,
0.0741877555847168,
0.23266352713108063,
-0.2360663264989853,
-0.08040942251682281,
0.0014828018611297011,
0.011312311515212059,
0.0028151054866611958,
-0.12229647487401962,
-0.04083598032593727,
0.06152952089905739,
0.02503015846014023,
0.1264374554157257,
0.01675955206155777,
0.045002423226833344,
-0.016305914148688316,
-0.14582811295986176,
-0.07145484536886215,
0.17096473276615143,
0.09866837412118912,
-0.05941618233919144,
-0.08155622333288193,
-0.0567617304623127,
-0.18505710363388062,
-0.0430840402841568,
0.0011264181230217218,
0.026362115517258644,
-0.0665440708398819,
-0.10301592200994492,
0.010042108595371246,
-0.0959419384598732,
-0.08252376317977905,
0.015526972711086273,
0.16648094356060028,
0.06671640276908875,
-0.0015431250212714076,
-0.0022606856655329466,
0.15082228183746338,
0.06808199733495712,
-0.16785253584384918,
-0.005578210577368736,
0.010642980225384235,
-0.02194649539887905,
-0.0178740955889225,
-0.02489664778113365,
-0.00003509366069920361,
0.005932788830250502,
0.17556317150592804,
-0.08425278216600418,
0.019854100421071053,
0.04833352565765381,
0.03515741229057312,
-0.08273035287857056,
0.17849849164485931,
-0.09368985146284103,
-0.029931871220469475,
0.024413660168647766,
0.12882839143276215,
0.03412991389632225,
0.005573905073106289,
-0.08339215070009232,
0.008971751667559147,
0.14379511773586273,
0.036109957844018936,
-0.011534457094967365,
0.04296206682920456,
-0.05794595181941986,
-0.04534827172756195,
0.13344191014766693,
-0.08196531981229782,
0.013752767816185951,
0.017809521406888962,
-0.09659307450056076,
-0.002740155905485153,
0.036083340644836426,
-0.008763076737523079,
0.005235749762505293,
0.09923918545246124,
-0.10011105239391327,
-0.02659081667661667,
-0.09758535772562027,
-0.08768372982740402,
0.018255870789289474,
-0.033352985978126526,
0.01772301271557808,
-0.10998900234699249,
-0.1769871860742569,
-0.02419852837920189,
0.03362632915377617,
-0.025690386071801186,
-0.08204320073127747,
-0.0322394073009491,
-0.10329551249742508,
0.028345663100481033,
-0.010236632078886032,
0.08987277746200562,
-0.043059080839157104,
0.1280525028705597,
0.05713140591979027,
0.06138397008180618,
0.03238491714000702,
0.046918995678424835,
-0.08418480306863785,
0.05429166555404663,
-0.2295490950345993,
0.05167853459715843,
-0.06423639506101608,
0.06785941123962402,
-0.1034015640616417,
-0.13179723918437958,
0.02215375192463398,
-0.019173579290509224,
0.08334557712078094,
0.14027643203735352,
-0.11595116555690765,
-0.09120864421129227,
0.14539361000061035,
-0.11052161455154419,
-0.13490672409534454,
0.11524619162082672,
-0.017290325835347176,
-0.026911024004220963,
0.036686260253190994,
0.10903867334127426,
0.10226739197969437,
-0.08101936429738998,
-0.04056591913104057,
-0.02115422673523426,
0.09558098763227463,
-0.02376023307442665,
0.10714898258447647,
-0.005946837365627289,
0.005107085220515728,
0.029233114793896675,
-0.11629217118024826,
0.07703741639852524,
-0.11484857648611069,
-0.08000880479812622,
-0.038425616919994354,
-0.09868695586919785,
0.08107796311378479,
0.07429289072751999,
0.05327543243765831,
-0.08108498901128769,
-0.12757964432239532,
0.06202191859483719,
0.14448392391204834,
-0.0866294726729393,
0.0007522808155044913,
-0.05402078106999397,
0.12552672624588013,
-0.08006551116704941,
-0.01823038049042225,
-0.1615990251302719,
-0.07373416423797607,
0.027386751025915146,
-0.05715659633278847,
-0.039375532418489456,
-0.056592684239149094,
0.05398911237716675,
0.10079189389944077,
-0.07087790220975876,
-0.11536262184381485,
-0.10146772116422653,
-0.010853174142539501,
-0.07143580168485641,
-0.20469580590724945,
-0.09931377321481705,
-0.01995461992919445,
0.17355877161026,
-0.22498774528503418,
0.04416519030928612,
0.02581086941063404,
0.15374267101287842,
0.043204981833696365,
-0.03028373047709465,
-0.0047775981947779655,
0.044255517423152924,
-0.040313467383384705,
-0.08979960530996323,
0.03165861591696739,
0.022547215223312378,
-0.0725720226764679,
-0.04543185234069824,
-0.08512959629297256,
0.18576399981975555,
0.12208686769008636,
-0.01371462456882,
-0.09445744007825851,
-0.03119046241044998,
-0.0951562449336052,
-0.043888457119464874,
-0.03529643639922142,
-0.013734456151723862,
0.07214943319559097,
0.006868221331387758,
0.14473740756511688,
-0.09422530233860016,
-0.04099792614579201,
0.03957158327102661,
-0.02082584798336029,
-0.014387326315045357,
0.10039550065994263,
0.052700940519571304,
-0.06213299185037613,
0.1523551195859909,
0.1353267878293991,
-0.08267839252948761,
0.12741217017173767,
-0.07714016735553741,
-0.09682294726371765,
-0.031047703698277473,
0.019408736377954483,
0.024547038599848747,
0.15973550081253052,
-0.09048692882061005,
-0.008454544469714165,
0.02981170080602169,
-0.002638463629409671,
0.013143941760063171,
-0.19773316383361816,
-0.00914021860808134,
0.02598162181675434,
-0.0427742600440979,
0.015376361086964607,
-0.008531112223863602,
0.0030533098615705967,
0.09658396989107132,
0.009691332466900349,
-0.041322797536849976,
0.03183336183428764,
0.013326510787010193,
-0.07079759240150452,
0.23208807408809662,
-0.09257741272449493,
-0.18240715563297272,
-0.13798032701015472,
0.04589056596159935,
-0.06968642771244049,
0.00313743413425982,
0.03633501008152962,
-0.08796383440494537,
-0.03314422816038132,
-0.04361449182033539,
0.04704074189066887,
0.010696050710976124,
0.03963487595319748,
0.002400670200586319,
0.019821349531412125,
0.09801855683326721,
-0.10083389282226562,
0.016846712678670883,
-0.017777154222130775,
-0.05029547959566116,
0.03160224109888077,
0.03807167708873749,
0.11413365602493286,
0.13355256617069244,
0.007623742800205946,
0.019286582246422768,
-0.033403199166059494,
0.1908070296049118,
-0.09904170781373978,
-0.010358689352869987,
0.16416075825691223,
0.026842717081308365,
0.05509908124804497,
0.10491766780614853,
0.04269848391413689,
-0.08046383410692215,
0.018242882564663887,
0.03249504417181015,
-0.005538108758628368,
-0.18799541890621185,
-0.03713010251522064,
-0.03353582322597504,
0.002773590851575136,
0.130792498588562,
0.03098081797361374,
0.059043847024440765,
0.08241087943315506,
-0.05049297586083412,
0.04713853821158409,
-0.03302677720785141,
0.08710294216871262,
0.054960742592811584,
0.049325332045555115,
0.12163075059652328,
-0.03343642130494118,
-0.03564213216304779,
0.029071010649204254,
-0.0016002055490389466,
0.22692237794399261,
-0.06334179639816284,
0.09274540841579437,
0.05585766211152077,
0.1928531378507614,
0.011882110498845577,
0.07764001190662384,
0.009413172490894794,
-0.031797412782907486,
0.012149741873145103,
-0.05585495010018349,
-0.03378381207585335,
0.025750646367669106,
-0.015009651891887188,
0.11568707972764969,
-0.17437413334846497,
0.010007655248045921,
0.04803915694355965,
0.31148412823677063,
0.07491454482078552,
-0.3712690472602844,
-0.1394759863615036,
0.003059888258576393,
-0.025909122079610825,
-0.05825544148683548,
0.009637950919568539,
0.09029939770698547,
-0.11257309466600418,
0.05751771107316017,
-0.09264767915010452,
0.07031773030757904,
-0.07052695751190186,
-0.0006214837194420397,
0.08828484266996384,
0.10841615498065948,
0.004335705656558275,
0.0630909726023674,
-0.24444139003753662,
0.275350958108902,
-0.003987464122474194,
0.07654760032892227,
-0.034123752266168594,
0.03418014198541641,
0.04437290132045746,
0.07884234189987183,
0.08719950914382935,
-0.012618477456271648,
-0.0064264568500220776,
-0.20837397873401642,
-0.09142110496759415,
0.009073828347027302,
0.0835033729672432,
-0.10125236213207245,
0.11619453877210617,
-0.012706529349088669,
-0.029157601296901703,
0.04160413146018982,
-0.05060238391160965,
-0.09757794439792633,
-0.0822966992855072,
0.00465439772233367,
-0.049844805151224136,
0.0735311433672905,
-0.13308170437812805,
-0.12348964810371399,
-0.05909545347094536,
0.13707882165908813,
-0.08386990427970886,
-0.06903908401727676,
-0.14326323568820953,
0.12773168087005615,
0.12100685387849808,
-0.09131458401679993,
0.08078477531671524,
-0.012286433950066566,
0.15373212099075317,
0.02342824824154377,
-0.04990512505173683,
0.08110732585191727,
-0.0970807820558548,
-0.24981817603111267,
-0.042293090373277664,
0.1399356722831726,
0.03414681553840637,
0.03450148180127144,
-0.012889948673546314,
0.011595447547733784,
-0.00897570513188839,
-0.0825890451669693,
0.027232512831687927,
0.014121043495833874,
0.08602599054574966,
0.04721985012292862,
-0.018729940056800842,
-0.008770431391894817,
-0.044431593269109726,
-0.011582549661397934,
0.09766892343759537,
0.25927743315696716,
-0.09643939882516861,
-0.03801371157169342,
0.07122514396905899,
-0.030292529612779617,
-0.17355351150035858,
0.043009135872125626,
0.12492929399013519,
0.028962746262550354,
-0.02282983809709549,
-0.18851961195468903,
0.08946985006332397,
0.10759735852479935,
-0.041573233902454376,
0.09757768362760544,
-0.2864927649497986,
-0.11841361224651337,
0.12305810302495956,
0.13702262938022614,
0.007499513681977987,
-0.1666376143693924,
-0.054283905774354935,
-0.009647300466895103,
-0.09402152895927429,
0.1029195487499237,
-0.034174758940935135,
0.0847739577293396,
-0.03409544751048088,
0.005252753850072622,
0.016401194036006927,
-0.060494329780340195,
0.1517314314842224,
-0.0032139455433934927,
0.09670520573854446,
-0.015375571325421333,
0.010626019909977913,
0.10014250874519348,
-0.07706525921821594,
0.04423975199460983,
-0.025090282782912254,
0.06857922673225403,
-0.11300723254680634,
-0.00880949292331934,
-0.11370790004730225,
0.05674345791339874,
-0.05306081101298332,
-0.05330008640885353,
-0.042116764932870865,
0.06194835901260376,
0.018852150067687035,
-0.00628373958170414,
0.1425030380487442,
0.053725846111774445,
0.15199454128742218,
0.08887504041194916,
0.03247765079140663,
-0.01606331206858158,
-0.11951426416635513,
-0.029040010645985603,
-0.028527991846203804,
0.07822464406490326,
-0.1590435951948166,
0.010833223350346088,
0.11640524864196777,
0.05645059421658516,
0.1306423544883728,
0.06399808824062347,
-0.05942843109369278,
0.021694833412766457,
0.08478373289108276,
-0.13205718994140625,
-0.08925923705101013,
-0.05193866416811943,
-0.006853868719190359,
-0.1531563103199005,
0.08456958830356598,
0.08911502361297607,
-0.0810321718454361,
-0.021166710183024406,
-0.0036134151741862297,
0.010294165462255478,
-0.01795646734535694,
0.19258558750152588,
0.10798081755638123,
0.08868986368179321,
-0.09754672646522522,
0.1059214398264885,
0.041866790503263474,
-0.1332659125328064,
-0.008896345272660255,
0.03772563114762306,
-0.0858631506562233,
-0.02423112280666828,
0.01929803192615509,
0.08799184858798981,
-0.051619578152894974,
-0.06164264306426048,
-0.16015833616256714,
-0.12123753875494003,
0.0636109709739685,
0.10152140259742737,
0.09145937114953995,
0.035466667264699936,
-0.011390501633286476,
0.044824015349149704,
-0.1359771192073822,
0.1297961324453354,
0.0822078213095665,
0.10821013897657394,
-0.1927167773246765,
0.13711971044540405,
0.015391132794320583,
0.04544707387685776,
-0.009189382195472717,
0.01592624932527542,
-0.10447865724563599,
-0.014764852821826935,
-0.09328998625278473,
-0.033473722636699677,
-0.06344771385192871,
-0.007489268202334642,
-0.014035306870937347,
-0.04754031449556351,
-0.039970654994249344,
0.0354972779750824,
-0.0964752584695816,
-0.04488528519868851,
0.0337880477309227,
0.04948236793279648,
-0.11619766801595688,
-0.025080638006329536,
0.02675650268793106,
-0.12746699154376984,
0.10461890697479248,
0.04512079432606697,
0.06359636783599854,
0.02667430229485035,
-0.04212738201022148,
0.016236545518040657,
0.06687132269144058,
-0.012678347527980804,
0.04041925072669983,
-0.12430807948112488,
0.018139421939849854,
-0.034463558346033096,
0.004094600677490234,
-0.012498955242335796,
0.050350122153759,
-0.14469848573207855,
-0.017506258562207222,
-0.02265634573996067,
-0.01035374216735363,
-0.07427772134542465,
0.06044521927833557,
0.08174949884414673,
0.01013888604938984,
0.16587035357952118,
-0.06957933306694031,
0.021965689957141876,
-0.23977528512477875,
-0.010578284971415997,
-0.04108681157231331,
-0.09883267432451248,
-0.11010388284921646,
-0.0013936813920736313,
0.07577656209468842,
-0.046865448355674744,
0.0726565271615982,
-0.044172775000333786,
0.09741181135177612,
0.030657853931188583,
-0.034669049084186554,
0.05893212556838989,
0.06269550323486328,
0.19260312616825104,
0.01291759591549635,
-0.015506230294704437,
0.05353419482707977,
0.039654865860939026,
0.08803411573171616,
0.05883614718914032,
0.14410202205181122,
0.13084033131599426,
-0.04077423736453056,
0.09713529050350189,
0.05021141096949577,
-0.07424843311309814,
-0.1811348795890808,
0.03751247376203537,
-0.03817771002650261,
0.11087067425251007,
-0.001840695389546454,
0.13789011538028717,
0.13742400705814362,
-0.18725088238716125,
0.02379711903631687,
-0.003807817120105028,
-0.07243199646472931,
-0.06592411547899246,
-0.055304188281297684,
-0.06961283832788467,
-0.19056689739227295,
0.010549894534051418,
-0.13035111129283905,
-0.024165259674191475,
0.07333757728338242,
-0.0020635584369301796,
-0.012282561510801315,
0.1831987053155899,
0.08956027776002884,
0.009594861418008804,
0.10321044921875,
0.03116491623222828,
-0.029619557783007622,
-0.032219741493463516,
-0.0905577763915062,
0.01962989754974842,
-0.016151325777173042,
0.041757792234420776,
-0.07355400919914246,
-0.0871632918715477,
0.07497144490480423,
0.03979463130235672,
-0.104082390666008,
0.033452991396188736,
-0.000325280474498868,
0.05199376121163368,
0.05892043560743332,
0.0015964783960953355,
0.02017602138221264,
-0.02507268451154232,
0.2215784788131714,
-0.09114864468574524,
-0.02449493110179901,
-0.13810794055461884,
0.23586821556091309,
0.026199214160442352,
-0.025720497593283653,
0.05962172523140907,
-0.10138313472270966,
-0.009832113981246948,
0.14054307341575623,
0.1324680596590042,
-0.005765127018094063,
-0.013100793585181236,
0.01476305816322565,
-0.03205151855945587,
-0.07229161262512207,
0.09363565593957901,
0.1212291568517685,
0.0746246948838234,
-0.06945806741714478,
-0.039877306669950485,
-0.04480942711234093,
-0.04109782353043556,
-0.029542362317442894,
0.07013219594955444,
0.009444575756788254,
-0.0009603126673027873,
-0.04335730895400047,
0.07868127524852753,
-0.019611291587352753,
-0.11103206872940063,
0.08517711609601974,
-0.1830914318561554,
-0.18363426625728607,
-0.036868736147880554,
0.029791295528411865,
0.02393578179180622,
0.06335525959730148,
-0.007305704522877932,
-0.01667427271604538,
0.13681486248970032,
-0.010277578607201576,
-0.04934113100171089,
-0.1298527717590332,
0.07794263958930969,
-0.05235839635133743,
0.2302209585905075,
-0.04357369244098663,
0.015508366748690605,
0.12783536314964294,
0.03178367391228676,
-0.13232076168060303,
0.009812857024371624,
0.06464987993240356,
-0.0632934719324112,
0.04042872041463852,
0.14985226094722748,
-0.012201238423585892,
0.08288273215293884,
0.026492193341255188,
-0.14779958128929138,
-0.014333926141262054,
-0.07926791906356812,
-0.051663804799318314,
-0.083852618932724,
0.009034711867570877,
-0.040168724954128265,
0.11356635391712189,
0.19280876219272614,
-0.06551357358694077,
-0.026840142905712128,
-0.057257845997810364,
0.04248068854212761,
0.09030203521251678,
0.0817141979932785,
0.013526054099202156,
-0.2439933717250824,
0.028006212785840034,
0.005128264427185059,
-0.001076607615686953,
-0.22822852432727814,
-0.08553300052881241,
0.03156502917408943,
-0.07660030573606491,
-0.08447063714265823,
0.09201256930828094,
0.047267716377973557,
0.052086398005485535,
-0.053321368992328644,
0.012147320434451103,
-0.11048789322376251,
0.16906380653381348,
-0.17123866081237793,
-0.0863979235291481
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# SMIDS_3x_beit_large_SGD_lr001_fold2
This model is a fine-tuned version of [microsoft/beit-large-patch16-224](https://huggingface.co/microsoft/beit-large-patch16-224) on the imagefolder dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2730
- Accuracy: 0.9101
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.001
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 50
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|
| 0.6183 | 1.0 | 450 | 0.5281 | 0.8103 |
| 0.4308 | 2.0 | 900 | 0.3746 | 0.8669 |
| 0.3321 | 3.0 | 1350 | 0.3194 | 0.8869 |
| 0.2741 | 4.0 | 1800 | 0.2941 | 0.8835 |
| 0.2911 | 5.0 | 2250 | 0.2779 | 0.8952 |
| 0.2345 | 6.0 | 2700 | 0.2676 | 0.8985 |
| 0.2656 | 7.0 | 3150 | 0.2640 | 0.8985 |
| 0.2454 | 8.0 | 3600 | 0.2617 | 0.8985 |
| 0.2325 | 9.0 | 4050 | 0.2651 | 0.8935 |
| 0.2736 | 10.0 | 4500 | 0.2583 | 0.8968 |
| 0.2631 | 11.0 | 4950 | 0.2630 | 0.8918 |
| 0.2185 | 12.0 | 5400 | 0.2609 | 0.8985 |
| 0.1998 | 13.0 | 5850 | 0.2581 | 0.8968 |
| 0.2041 | 14.0 | 6300 | 0.2537 | 0.9002 |
| 0.2148 | 15.0 | 6750 | 0.2607 | 0.9052 |
| 0.2184 | 16.0 | 7200 | 0.2551 | 0.9018 |
| 0.1852 | 17.0 | 7650 | 0.2565 | 0.9018 |
| 0.144 | 18.0 | 8100 | 0.2589 | 0.9068 |
| 0.2342 | 19.0 | 8550 | 0.2648 | 0.9052 |
| 0.1959 | 20.0 | 9000 | 0.2552 | 0.9068 |
| 0.1454 | 21.0 | 9450 | 0.2555 | 0.9085 |
| 0.2119 | 22.0 | 9900 | 0.2629 | 0.9101 |
| 0.2103 | 23.0 | 10350 | 0.2595 | 0.9085 |
| 0.1202 | 24.0 | 10800 | 0.2671 | 0.9085 |
| 0.1769 | 25.0 | 11250 | 0.2606 | 0.9101 |
| 0.1659 | 26.0 | 11700 | 0.2665 | 0.9101 |
| 0.1642 | 27.0 | 12150 | 0.2638 | 0.9101 |
| 0.159 | 28.0 | 12600 | 0.2681 | 0.9101 |
| 0.2289 | 29.0 | 13050 | 0.2645 | 0.9118 |
| 0.1249 | 30.0 | 13500 | 0.2685 | 0.9085 |
| 0.1195 | 31.0 | 13950 | 0.2692 | 0.9085 |
| 0.1041 | 32.0 | 14400 | 0.2692 | 0.9068 |
| 0.2053 | 33.0 | 14850 | 0.2639 | 0.9101 |
| 0.1366 | 34.0 | 15300 | 0.2708 | 0.9085 |
| 0.1378 | 35.0 | 15750 | 0.2715 | 0.9118 |
| 0.1913 | 36.0 | 16200 | 0.2686 | 0.9118 |
| 0.1193 | 37.0 | 16650 | 0.2681 | 0.9118 |
| 0.0953 | 38.0 | 17100 | 0.2718 | 0.9118 |
| 0.2328 | 39.0 | 17550 | 0.2715 | 0.9101 |
| 0.081 | 40.0 | 18000 | 0.2739 | 0.9101 |
| 0.1182 | 41.0 | 18450 | 0.2750 | 0.9118 |
| 0.1418 | 42.0 | 18900 | 0.2721 | 0.9101 |
| 0.142 | 43.0 | 19350 | 0.2739 | 0.9101 |
| 0.1749 | 44.0 | 19800 | 0.2729 | 0.9101 |
| 0.2226 | 45.0 | 20250 | 0.2735 | 0.9118 |
| 0.0963 | 46.0 | 20700 | 0.2738 | 0.9135 |
| 0.1257 | 47.0 | 21150 | 0.2731 | 0.9118 |
| 0.0797 | 48.0 | 21600 | 0.2732 | 0.9101 |
| 0.1681 | 49.0 | 22050 | 0.2730 | 0.9101 |
| 0.2027 | 50.0 | 22500 | 0.2730 | 0.9101 |
### Framework versions
- Transformers 4.32.1
- Pytorch 2.0.1
- Datasets 2.12.0
- Tokenizers 0.13.2
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["imagefolder"], "metrics": ["accuracy"], "base_model": "microsoft/beit-large-patch16-224", "model-index": [{"name": "SMIDS_3x_beit_large_SGD_lr001_fold2", "results": [{"task": {"type": "image-classification", "name": "Image Classification"}, "dataset": {"name": "imagefolder", "type": "imagefolder", "config": "default", "split": "test", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.9101497504159733, "name": "Accuracy"}]}]}]} | image-classification | onizukal/SMIDS_3x_beit_large_SGD_lr001_fold2 | [
"transformers",
"pytorch",
"beit",
"image-classification",
"generated_from_trainer",
"dataset:imagefolder",
"base_model:microsoft/beit-large-patch16-224",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T16:55:44+00:00 | [] | [] | TAGS
#transformers #pytorch #beit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/beit-large-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| SMIDS\_3x\_beit\_large\_SGD\_lr001\_fold2
=========================================
This model is a fine-tuned version of microsoft/beit-large-patch16-224 on the imagefolder dataset.
It achieves the following results on the evaluation set:
* Loss: 0.2730
* Accuracy: 0.9101
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.001
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* num\_epochs: 50
### Training results
### Framework versions
* Transformers 4.32.1
* Pytorch 2.0.1
* Datasets 2.12.0
* Tokenizers 0.13.2
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.32.1\n* Pytorch 2.0.1\n* Datasets 2.12.0\n* Tokenizers 0.13.2"
] | [
"TAGS\n#transformers #pytorch #beit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/beit-large-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.32.1\n* Pytorch 2.0.1\n* Datasets 2.12.0\n* Tokenizers 0.13.2"
] | [
81,
115,
4,
30
] | [
"passage: TAGS\n#transformers #pytorch #beit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/beit-large-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50### Training results### Framework versions\n\n\n* Transformers 4.32.1\n* Pytorch 2.0.1\n* Datasets 2.12.0\n* Tokenizers 0.13.2"
] | [
-0.1293599158525467,
0.1724882274866104,
-0.0023454553447663784,
0.13587836921215057,
0.11250235140323639,
0.015344180166721344,
0.13944171369075775,
0.16905122995376587,
-0.08231265097856522,
0.04725457355380058,
0.1399839073419571,
0.13659004867076874,
0.046719420701265335,
0.19427457451820374,
-0.05236957222223282,
-0.2601667046546936,
0.04119328409433365,
0.03234807401895523,
-0.02075078710913658,
0.12314869463443756,
0.09301083534955978,
-0.13055187463760376,
0.11641614139080048,
0.030263151973485947,
-0.1994711011648178,
-0.03690062463283539,
-0.00742433313280344,
-0.06729142367839813,
0.10523199290037155,
-0.003170925425365567,
0.06897550076246262,
0.03744976967573166,
0.0835329070687294,
-0.13024620711803436,
0.0019813377875834703,
0.043246712535619736,
0.0064185261726379395,
0.10353662818670273,
0.05471734330058098,
-0.015182994306087494,
0.07030405849218369,
-0.06875743716955185,
0.06702885776758194,
0.009385199286043644,
-0.11313743144273758,
-0.26980340480804443,
-0.10223843902349472,
0.07157823443412781,
0.08217991888523102,
0.068179190158844,
0.008332090452313423,
0.1646922081708908,
-0.015027978457510471,
0.10447341948747635,
0.23090173304080963,
-0.2640359699726105,
-0.055165741592645645,
0.0294360164552927,
0.014914325438439846,
0.06473758071660995,
-0.10603974759578705,
-0.018760167062282562,
0.02059783600270748,
0.044390130788087845,
0.1409236192703247,
-0.010635782033205032,
-0.02811739780008793,
-0.021928580477833748,
-0.10847067832946777,
-0.08875154703855515,
0.18579065799713135,
0.058072153478860855,
-0.04802494868636131,
-0.07736620306968689,
-0.07186304777860641,
-0.17165552079677582,
-0.041741833090782166,
0.009797174483537674,
0.04154014587402344,
-0.04674985632300377,
-0.10634559392929077,
-0.030910290777683258,
-0.07805538177490234,
-0.051461391150951385,
-0.023120464757084846,
0.1351369321346283,
0.03383360058069229,
0.05723920464515686,
-0.03597215935587883,
0.09929350018501282,
0.0073938071727752686,
-0.17543649673461914,
-0.028137801215052605,
-0.0017277015140280128,
0.015125435777008533,
-0.01991228759288788,
-0.030262885615229607,
-0.06523662805557251,
-0.001314454828388989,
0.1489848643541336,
-0.06082969531416893,
0.060913555324077606,
-0.007318461779505014,
0.04020942002534866,
-0.048562191426754,
0.1864238679409027,
-0.02870783396065235,
-0.01659870520234108,
0.0207351166754961,
0.08822518587112427,
0.06836054474115372,
-0.036532942205667496,
-0.12525734305381775,
0.03098621405661106,
0.12811045348644257,
0.0029374780133366585,
-0.021369412541389465,
0.05286439135670662,
-0.0643334686756134,
-0.0587083138525486,
0.09228596836328506,
-0.08908151835203171,
0.034838590770959854,
-0.010374085046350956,
-0.084370456635952,
-0.06779533624649048,
0.027354132384061813,
0.01850878819823265,
-0.0004323708708398044,
0.07165450602769852,
-0.09108522534370422,
0.014963540248572826,
-0.06533985584974289,
-0.10081785917282104,
0.016071073710918427,
-0.1107335016131401,
0.012424856424331665,
-0.09663169085979462,
-0.19710472226142883,
0.006891206838190556,
0.07721206545829773,
-0.05610070005059242,
-0.06814593821763992,
-0.03639180585741997,
-0.07652068138122559,
0.04154945909976959,
-0.011949662119150162,
0.07310567051172256,
-0.0747697651386261,
0.0913747176527977,
0.022405456751585007,
0.08734618872404099,
-0.05631003528833389,
0.0460691936314106,
-0.1024455726146698,
0.04984736442565918,
-0.19827407598495483,
0.0799676924943924,
-0.0493633896112442,
0.0617455318570137,
-0.09557844698429108,
-0.10542625933885574,
0.03370443359017372,
-0.05006959289312363,
0.06861566007137299,
0.0974634513258934,
-0.1729826033115387,
-0.057957619428634644,
0.1353495568037033,
-0.09663163125514984,
-0.14797286689281464,
0.10109983384609222,
-0.050693199038505554,
0.01928282529115677,
0.047161467373371124,
0.21422064304351807,
0.06321150809526443,
-0.09143578261137009,
-0.02580295503139496,
-0.03329068049788475,
0.04440530762076378,
-0.06494591385126114,
0.10175396502017975,
0.027680888772010803,
0.05365157872438431,
0.023984158411622047,
-0.032899804413318634,
0.03856229409575462,
-0.08387355506420135,
-0.10054522752761841,
-0.05070185661315918,
-0.08565592765808105,
0.039397966116666794,
0.05604296177625656,
0.05994046851992607,
-0.10856401175260544,
-0.09017187356948853,
0.04204317927360535,
0.0943065956234932,
-0.07395226508378983,
0.028962817043066025,
-0.09000826627016068,
0.11608705669641495,
-0.08325600624084473,
-0.02390553615987301,
-0.1791052222251892,
-0.04241684824228287,
0.040629271417856216,
-0.01629319041967392,
-0.006912850774824619,
-0.048891667276620865,
0.07074255496263504,
0.08783093094825745,
-0.05235742777585983,
-0.05203414335846901,
-0.05522594600915909,
0.008213330060243607,
-0.1105295866727829,
-0.1776295155286789,
-0.08015653491020203,
-0.0380605012178421,
0.14977632462978363,
-0.15268713235855103,
0.022225622087717056,
0.061116840690374374,
0.12500962615013123,
0.059797611087560654,
-0.04723487049341202,
-0.007436409126967192,
0.021452799439430237,
-0.05571167171001434,
-0.08678608387708664,
0.05719248577952385,
0.03528200834989548,
-0.07155010849237442,
-0.019102152436971664,
-0.10049699991941452,
0.1498662531375885,
0.13190734386444092,
-0.0015375686343759298,
-0.04512632265686989,
-0.01160994078963995,
-0.06610778719186783,
-0.030441991984844208,
-0.04081778973340988,
0.018804829567670822,
0.10142559558153152,
0.01744643971323967,
0.14419154822826385,
-0.09178037941455841,
-0.036961425095796585,
0.053544968366622925,
-0.028453968465328217,
-0.0331195667386055,
0.07361359149217606,
0.02190210297703743,
-0.14263916015625,
0.15015269815921783,
0.14882412552833557,
-0.04894813522696495,
0.12402692437171936,
-0.036747027188539505,
-0.0615357980132103,
-0.044876549392938614,
-0.037704430520534515,
0.014213677495718002,
0.1403394341468811,
-0.08333314955234528,
-0.005919712595641613,
0.05630137771368027,
0.019257593899965286,
-0.007085299585014582,
-0.18072617053985596,
0.0006808378966525197,
0.03521978110074997,
-0.04604950174689293,
-0.02278841845691204,
-0.014470276422798634,
0.0007941273506730795,
0.09172741323709488,
0.019804218783974648,
-0.07102026045322418,
0.05166372284293175,
0.010580740869045258,
-0.05623085796833038,
0.16415521502494812,
-0.07910753041505814,
-0.19727325439453125,
-0.11776646971702576,
-0.08754957467317581,
-0.10735819488763809,
0.013021474704146385,
0.06737184524536133,
-0.050448641180992126,
-0.04938974231481552,
-0.10206248611211777,
-0.04453543201088905,
0.021900271996855736,
0.02429220825433731,
0.05370878055691719,
-0.008031168952584267,
0.08405356109142303,
-0.09224440902471542,
-0.03291117399930954,
-0.014789600856602192,
0.018657125532627106,
0.06682770699262619,
0.018715238198637962,
0.11069032549858093,
0.08161229640245438,
-0.02844928205013275,
0.05646483600139618,
-0.01682325080037117,
0.2655041813850403,
-0.06765957176685333,
-0.006789656355977058,
0.13932959735393524,
-0.013368978165090084,
0.08428963273763657,
0.1268712729215622,
0.04151352122426033,
-0.09555158019065857,
-0.013173693791031837,
-0.00024822441628202796,
-0.05275752767920494,
-0.1537386178970337,
-0.04163756221532822,
-0.045641690492630005,
-0.0021682933438569307,
0.13930507004261017,
0.03818075731396675,
0.02474883571267128,
0.07807637751102448,
0.020041609182953835,
0.05664918199181557,
-0.017527885735034943,
0.10406769812107086,
0.08156019449234009,
0.06448414921760559,
0.13368317484855652,
-0.03653626888990402,
-0.019387291744351387,
0.05662747099995613,
0.04215037450194359,
0.20423758029937744,
-0.02541770040988922,
0.14701254665851593,
0.02641657367348671,
0.19307395815849304,
0.017521383240818977,
0.0728468969464302,
-0.014410126954317093,
0.0009393728105351329,
-0.019274147227406502,
-0.04702805355191231,
-0.06427313387393951,
0.03288881108164787,
-0.016649875789880753,
0.05632343888282776,
-0.09356046468019485,
0.039105307310819626,
0.059592608362436295,
0.30666422843933105,
0.06539998203516006,
-0.4122132360935211,
-0.09836560487747192,
0.012291035614907742,
0.0009865236934274435,
-0.055195607244968414,
-0.0072626820765435696,
0.0979013666510582,
-0.09949664771556854,
0.08215389400720596,
-0.09418605268001556,
0.08514873683452606,
-0.0845724418759346,
0.020298872143030167,
0.07689075917005539,
0.056060366332530975,
0.013226890936493874,
0.05964293330907822,
-0.21821673214435577,
0.24971400201320648,
0.018467964604496956,
0.04422129690647125,
-0.08908867090940475,
0.010060982778668404,
0.033364444971084595,
0.059161990880966187,
0.08554306626319885,
0.005977867171168327,
-0.09024009108543396,
-0.18880225718021393,
-0.1258762925863266,
0.0005427713040262461,
0.06169470399618149,
-0.036699384450912476,
0.09451829642057419,
-0.018175894394516945,
-0.012127134948968887,
0.021332256495952606,
0.0005201056483201683,
-0.03501477465033531,
-0.103630930185318,
0.02024604007601738,
0.034688886255025864,
-0.012138742953538895,
-0.06473075598478317,
-0.11475593596696854,
-0.03554871678352356,
0.16192500293254852,
0.05505121126770973,
-0.07524240761995316,
-0.1408705860376358,
0.07218684256076813,
0.07781627029180527,
-0.0855332687497139,
0.039305757731199265,
-0.016779718920588493,
0.14986851811408997,
0.020937321707606316,
-0.08943228423595428,
0.10178638249635696,
-0.05869165062904358,
-0.17860572040081024,
-0.041185978800058365,
0.09929849207401276,
0.007366738747805357,
0.05263189971446991,
0.004192214459180832,
0.06014186516404152,
-0.035002902150154114,
-0.0584394596517086,
0.06681792438030243,
-0.0073097143322229385,
0.10614755749702454,
-0.014883637428283691,
0.00864378735423088,
0.029195772483944893,
-0.04613848030567169,
0.00009839441918302327,
0.1684505194425583,
0.24079899489879608,
-0.10403203964233398,
0.060546230524778366,
0.03012177161872387,
-0.030879246070981026,
-0.18261685967445374,
0.010319743305444717,
0.07656802982091904,
-0.0001991603203350678,
0.04173794388771057,
-0.16060468554496765,
0.055176541209220886,
0.10514935851097107,
-0.043303944170475006,
0.08152011036872864,
-0.2768779397010803,
-0.11840421706438065,
0.0923023670911789,
0.138164222240448,
0.0691317543387413,
-0.13107311725616455,
-0.04327763617038727,
-0.041234806180000305,
-0.17335952818393707,
0.13665583729743958,
-0.05704028159379959,
0.11501350998878479,
-0.039327461272478104,
0.08051838725805283,
0.014901114627718925,
-0.056082114577293396,
0.14561402797698975,
0.005515002179890871,
0.08661133795976639,
-0.07185279577970505,
-0.0014093852369114757,
0.10643326491117477,
-0.10252601653337479,
0.07192501425743103,
-0.0869532898068428,
0.06187514215707779,
-0.10810889303684235,
-0.0037693935446441174,
-0.07425615191459656,
0.013987713493406773,
-0.013397954404354095,
-0.048907287418842316,
-0.0448833703994751,
0.03488645330071449,
0.06301422417163849,
-0.018155096098780632,
0.20988906919956207,
0.06445588916540146,
0.0862940326333046,
0.1728745847940445,
0.05397673323750496,
-0.10576145350933075,
-0.09408308565616608,
-0.04430058225989342,
-0.029343122616410255,
0.059755485504865646,
-0.13705183565616608,
0.053009506314992905,
0.12004052102565765,
0.013443393632769585,
0.1280696988105774,
0.05582417547702789,
-0.030783196911215782,
0.035687193274497986,
0.06206676363945007,
-0.1721130907535553,
-0.08640376478433609,
-0.010029762983322144,
0.030597826465964317,
-0.13003188371658325,
0.045725177973508835,
0.12137939780950546,
-0.0593545101583004,
-0.014887568540871143,
-0.004342919681221247,
0.03682979568839073,
-0.009421703405678272,
0.15946903824806213,
0.047883741557598114,
0.05509158596396446,
-0.11808934807777405,
0.11348052322864532,
0.057328153401613235,
-0.0728185623884201,
0.032391179352998734,
0.05030714347958565,
-0.10392948985099792,
-0.021465230733156204,
0.031419817358255386,
0.14932547509670258,
-0.06275127828121185,
-0.045640427619218826,
-0.13568063080310822,
-0.091814324259758,
0.06645428389310837,
0.07967224717140198,
0.0933644250035286,
0.01663324609398842,
-0.03539150580763817,
-0.013165266253054142,
-0.10855977237224579,
0.10982618480920792,
0.04324139654636383,
0.09105362743139267,
-0.17992232739925385,
0.054193608462810516,
-0.0015555275604128838,
0.07246194779872894,
-0.021836427971720695,
-0.00042325531831011176,
-0.08788467198610306,
0.003508437890559435,
-0.10813499987125397,
0.02464236691594124,
-0.052905477583408356,
0.006243168842047453,
-0.02064651995897293,
-0.0580705925822258,
-0.06364380568265915,
0.024784497916698456,
-0.11918067932128906,
-0.053243763744831085,
0.02146504819393158,
0.031834639608860016,
-0.12016978859901428,
-0.04392008110880852,
0.020345089957118034,
-0.08986733108758926,
0.09774119406938553,
0.06029992923140526,
-0.008077923208475113,
0.00773270707577467,
0.0036002967972308397,
-0.02274298295378685,
0.0666942149400711,
0.007561622653156519,
0.08597277849912643,
-0.1152612566947937,
-0.0221384409815073,
0.01634843461215496,
-0.004547150805592537,
0.017726117745041847,
0.15840598940849304,
-0.12086156010627747,
-0.0003179961640853435,
-0.014678256586194038,
-0.06600851565599442,
-0.06344839930534363,
0.06893838196992874,
0.10903503000736237,
0.02346671372652054,
0.21181334555149078,
-0.054371658712625504,
0.015811823308467865,
-0.20995409786701202,
-0.011581460013985634,
0.005185890011489391,
-0.1388559192419052,
-0.10497695952653885,
-0.03237957879900932,
0.06376256048679352,
-0.07031478732824326,
0.11765085160732269,
0.03525954857468605,
0.02161695808172226,
0.02906344085931778,
0.025029366835951805,
-0.0031726681627333164,
0.013450034894049168,
0.16309522092342377,
0.014403261244297028,
-0.028442582115530968,
0.12852592766284943,
0.028986822813749313,
0.09334488213062286,
0.11778779327869415,
0.17672526836395264,
0.11388354748487473,
0.04729508236050606,
0.09055530279874802,
0.05202596262097359,
-0.025968270376324654,
-0.22174733877182007,
0.03601896017789841,
-0.03978736698627472,
0.1488790065050125,
-0.0030294209718704224,
0.15902450680732727,
0.0920415073633194,
-0.18360793590545654,
0.040488436818122864,
-0.03700747340917587,
-0.0790853351354599,
-0.08454839885234833,
-0.12155362963676453,
-0.10311590880155563,
-0.15089921653270721,
0.002945262473076582,
-0.1040843203663826,
0.023338600993156433,
0.11202728003263474,
-0.008582104928791523,
-0.009919910691678524,
0.116677425801754,
-0.02631515823304653,
0.026041926816105843,
0.03836518153548241,
0.00608045794069767,
-0.059937771409749985,
-0.044151950627565384,
-0.08065995573997498,
0.014101422391831875,
0.032313644886016846,
0.05599058046936989,
-0.03235676884651184,
-0.007023791316896677,
0.03841041401028633,
-0.010091220960021019,
-0.12353866547346115,
0.01347822230309248,
0.005028906278312206,
0.05164548382163048,
0.0008541525457985699,
0.012780209071934223,
0.03201600909233093,
-0.015217483974993229,
0.19341084361076355,
-0.07325411587953568,
-0.027416478842496872,
-0.1228807121515274,
0.17896701395511627,
0.0026140701957046986,
-0.04994320869445801,
0.05295133590698242,
-0.09137362241744995,
-0.020702529698610306,
0.15485265851020813,
0.1892986297607422,
-0.07158271223306656,
-0.016520513221621513,
-0.017527583986520767,
-0.013897030614316463,
-0.022615507245063782,
0.09919055551290512,
0.0991419330239296,
-0.0069245584309101105,
-0.0751221776008606,
-0.028980256989598274,
-0.06606413424015045,
-0.034512959420681,
-0.03850788250565529,
0.06925404816865921,
-0.004570751916617155,
0.0070457919500768185,
-0.07483471930027008,
0.04310325160622597,
-0.02210995741188526,
-0.06085818260908127,
0.06226903945207596,
-0.21256737411022186,
-0.17790570855140686,
0.006773421075195074,
0.07538973540067673,
0.0015973751433193684,
0.0461571104824543,
-0.009913075715303421,
0.018662674352526665,
0.07594356685876846,
-0.02225665934383869,
-0.08672447502613068,
-0.09593749046325684,
0.10812120139598846,
-0.13375911116600037,
0.2528570294380188,
-0.03883460536599159,
0.03583916276693344,
0.12127543240785599,
0.041867125779390335,
-0.1335451751947403,
0.03351692110300064,
0.03981999680399895,
-0.032485269010066986,
0.00548918079584837,
0.14240407943725586,
-0.03740047290921211,
0.07958021014928818,
0.0458458811044693,
-0.1027912050485611,
-0.03964604437351227,
-0.04966754838824272,
-0.011354409158229828,
-0.024445757269859314,
-0.054610975086688995,
-0.036348532885313034,
0.13227923214435577,
0.17175258696079254,
-0.042096637189388275,
-0.023690558969974518,
-0.06475082784891129,
0.030860183760523796,
0.07729368656873703,
-0.03295742720365524,
-0.052064236253499985,
-0.23603148758411407,
0.0024359924718737602,
0.05229694023728371,
-0.013576737605035305,
-0.20701472461223602,
-0.110505111515522,
0.0060418094508349895,
-0.05801977962255478,
-0.07628542929887772,
0.09231390058994293,
0.06255589425563812,
0.035103797912597656,
-0.06320928037166595,
0.038133736699819565,
-0.07872021943330765,
0.14179112017154694,
-0.14508864283561707,
-0.07859515398740768
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# crossencoder-airline-refine
- This model is trained on open source airline related dataset.
- The base model is "cross-encoder/stsb-roberta-large"
## Model description
- Cross encoder is useful when we want to calculate the similarity between search query and data items.
- If a Cross-Encoder model is trained on a representative training set, it achieves higher accuracy than Bi-Encoders.
- A Cross-Encoder does not produce a sentence embedding. Also, we are not able to pass individual sentences to a Cross-Encoder.
## Intended uses & limitations
- The model is finetuned on limited data.
- It might not produce right result in airline related text.
- Model will be finetuned increamentally based on the availablity of the data.
## Training and evaluation data
- Below is the example of training data format for cross encoder.
- Training data has sentence1, sentence2 and the similarity score between the two sentence.

### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 500
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| No log | 1.0 | 2 | 2.0393 |
| No log | 2.0 | 4 | 1.3405 |
| No log | 3.0 | 6 | 0.9373 |
### Framework versions
- Transformers 4.35.0
- Pytorch 2.1.0
- Datasets 2.14.6
- Tokenizers 0.14.1
| {"tags": ["generated_from_trainer"], "model-index": [{"name": "crossencoder-airline-refine", "results": []}]} | text-classification | srmishra/crossencoder-airline-refine | [
"transformers",
"safetensors",
"roberta",
"text-classification",
"generated_from_trainer",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T16:57:17+00:00 | [] | [] | TAGS
#transformers #safetensors #roberta #text-classification #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us
| crossencoder-airline-refine
===========================
* This model is trained on open source airline related dataset.
* The base model is "cross-encoder/stsb-roberta-large"
Model description
-----------------
* Cross encoder is useful when we want to calculate the similarity between search query and data items.
* If a Cross-Encoder model is trained on a representative training set, it achieves higher accuracy than Bi-Encoders.
* A Cross-Encoder does not produce a sentence embedding. Also, we are not able to pass individual sentences to a Cross-Encoder.
Intended uses & limitations
---------------------------
* The model is finetuned on limited data.
* It might not produce right result in airline related text.
* Model will be finetuned increamentally based on the availablity of the data.
Training and evaluation data
----------------------------
* Below is the example of training data format for cross encoder.
* Training data has sentence1, sentence2 and the similarity score between the two sentence.
!image/png
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 500
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.35.0
* Pytorch 2.1.0
* Datasets 2.14.6
* Tokenizers 0.14.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.0\n* Pytorch 2.1.0\n* Datasets 2.14.6\n* Tokenizers 0.14.1"
] | [
"TAGS\n#transformers #safetensors #roberta #text-classification #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.0\n* Pytorch 2.1.0\n* Datasets 2.14.6\n* Tokenizers 0.14.1"
] | [
45,
116,
4,
30
] | [
"passage: TAGS\n#transformers #safetensors #roberta #text-classification #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.35.0\n* Pytorch 2.1.0\n* Datasets 2.14.6\n* Tokenizers 0.14.1"
] | [
-0.07682488858699799,
0.0360892079770565,
-0.0017966583836823702,
0.11062810570001602,
0.1907743215560913,
0.02049897238612175,
0.12489538639783859,
0.10732981562614441,
-0.11640490591526031,
0.032137494534254074,
0.11998828500509262,
0.17062926292419434,
-0.0033498844131827354,
0.1695004254579544,
-0.08555435389280319,
-0.2545020580291748,
0.0072898720391094685,
0.0009605194791220129,
-0.06829528510570526,
0.130793958902359,
0.08737646043300629,
-0.1552218347787857,
0.08709221333265305,
-0.031751006841659546,
-0.21597017347812653,
0.021870702505111694,
0.03522888198494911,
-0.06004450097680092,
0.14368653297424316,
0.010254189372062683,
0.14067722856998444,
0.0287252776324749,
0.11113322526216507,
-0.2000841647386551,
0.010769953951239586,
0.06296537816524506,
0.012911994010210037,
0.06936174631118774,
0.040079519152641296,
-0.04338259994983673,
0.08475040644407272,
-0.1145513504743576,
0.08018731325864792,
0.011072135530412197,
-0.14922857284545898,
-0.2083459049463272,
-0.07972685247659683,
-0.021666180342435837,
0.0853320062160492,
0.08331017196178436,
-0.019929958507418633,
0.1671018898487091,
-0.09860707074403763,
0.10844485461711884,
0.2513124942779541,
-0.2835536599159241,
-0.07506807148456573,
0.0409659817814827,
0.01075495034456253,
0.09415028989315033,
-0.12137888371944427,
0.004452258814126253,
0.06617481261491776,
0.03357185050845146,
0.14108037948608398,
-0.037061214447021484,
-0.13057726621627808,
0.010023420676589012,
-0.14353856444358826,
0.011533224023878574,
0.08085067570209503,
0.030872881412506104,
-0.03922789543867111,
-0.032849282026290894,
-0.055722448974847794,
-0.1509251892566681,
-0.05417095869779587,
-0.045066218823194504,
0.057825956493616104,
-0.06030178442597389,
-0.09595538675785065,
0.026088811457157135,
-0.08043605089187622,
-0.07684198021888733,
-0.04697752743959427,
0.20244011282920837,
0.03266533091664314,
-0.0006377751706168056,
-0.01924007013440132,
0.096457339823246,
-0.025956282392144203,
-0.1300939917564392,
0.01885039359331131,
0.016381561756134033,
-0.02833756059408188,
-0.08443892747163773,
-0.06709075719118118,
-0.059192951768636703,
0.017537551000714302,
0.13905148208141327,
-0.07610134780406952,
0.06634508073329926,
0.007249496877193451,
0.002622522646561265,
-0.0990205779671669,
0.17831291258335114,
-0.0253185722976923,
-0.044175226241350174,
0.005353986285626888,
0.06746149063110352,
0.004529768601059914,
-0.010948633775115013,
-0.09495365619659424,
0.020993413403630257,
0.11820989847183228,
0.02783861756324768,
-0.09509164839982986,
0.08449285477399826,
-0.04470442607998848,
0.007636512629687786,
-0.03184472396969795,
-0.10694032162427902,
0.04798363894224167,
-0.00857238844037056,
-0.07911906391382217,
-0.07274794578552246,
0.011966688558459282,
0.031941741704940796,
0.008153160102665424,
0.1530817300081253,
-0.08193284273147583,
0.044478315860033035,
-0.10250405967235565,
-0.13139677047729492,
-0.012418245896697044,
-0.058060526847839355,
0.023964880034327507,
-0.10742290318012238,
-0.15564800798892975,
-0.02288678288459778,
0.048869647085666656,
-0.03886231407523155,
-0.00736550148576498,
-0.07189124077558517,
-0.07744938880205154,
0.014779730699956417,
-0.02331540733575821,
0.14053867757320404,
-0.06943068653345108,
0.10142280161380768,
0.06089483201503754,
0.0888671800494194,
-0.027435483410954475,
0.033239684998989105,
-0.0993015244603157,
0.003160156076774001,
-0.24504850804805756,
0.04923940449953079,
-0.060437411069869995,
0.0810452550649643,
-0.07526004314422607,
-0.09494015574455261,
0.027221551164984703,
0.014530489221215248,
0.08676984161138535,
0.1227983757853508,
-0.1623970866203308,
-0.06655260920524597,
0.17590533196926117,
-0.08143211156129837,
-0.11331930011510849,
0.10291161388158798,
-0.07624337822198868,
0.0645604282617569,
0.08907100558280945,
0.1893850862979889,
0.032733358442783356,
-0.08982565253973007,
0.021844282746315002,
-0.055190227925777435,
0.0458105243742466,
-0.004241929855197668,
0.03146016597747803,
0.02367931418120861,
0.00012181926285848022,
0.02015010453760624,
-0.006710395682603121,
0.04345128685235977,
-0.11450941115617752,
-0.08140373229980469,
-0.03193073347210884,
-0.10634852200746536,
0.058048833161592484,
0.055977676063776016,
0.07556730508804321,
-0.137850821018219,
-0.0694037601351738,
0.07403817027807236,
0.056527186185121536,
-0.04910828173160553,
0.02786555327475071,
-0.07247095555067062,
0.023554310202598572,
-0.023238003253936768,
-0.019067158922553062,
-0.1792774349451065,
-0.04289683327078819,
0.006852228194475174,
0.044590823352336884,
0.01566435769200325,
-0.02875876985490322,
0.08863680809736252,
0.0715734139084816,
-0.08201085031032562,
-0.014909687452018261,
-0.0006278945365920663,
0.012881642207503319,
-0.139989972114563,
-0.21800673007965088,
0.00912525225430727,
-0.027873095124959946,
0.10938155651092529,
-0.22715602815151215,
0.03250448778271675,
-0.011328354477882385,
0.08125799894332886,
0.034782130271196365,
-0.004239604342728853,
-0.04254043102264404,
0.09137844294309616,
-0.028265386819839478,
-0.05634007602930069,
0.055764392018318176,
-0.023066340014338493,
-0.07466987520456314,
-0.06262807548046112,
-0.14095135033130646,
0.19455118477344513,
0.13218264281749725,
-0.11111675947904587,
-0.12082096934318542,
0.004609640687704086,
-0.04270339757204056,
-0.01294113788753748,
-0.06191377714276314,
0.05530651658773422,
0.14098776876926422,
-0.013484482653439045,
0.14230772852897644,
-0.060048993676900864,
-0.02870800532400608,
0.017214590683579445,
-0.04588483273983002,
0.04886932298541069,
0.10827159136533737,
0.08053799718618393,
-0.09185939282178879,
0.12700729072093964,
0.14701251685619354,
-0.08035613596439362,
0.12193499505519867,
-0.02195601351559162,
-0.04714813828468323,
-0.016009772196412086,
-0.016442960128188133,
0.007169570308178663,
0.08979522436857224,
-0.06446580588817596,
-0.01108564343303442,
-0.006852015852928162,
0.035496748983860016,
-0.0033284525852650404,
-0.21359014511108398,
-0.04355913773179054,
0.03245973587036133,
-0.032557178288698196,
-0.0023697461001574993,
-0.03212404623627663,
0.009879238903522491,
0.11951171606779099,
0.00998776126652956,
-0.06840343773365021,
0.021740572527050972,
-0.001610051840543747,
-0.07529650628566742,
0.2205691784620285,
-0.08862055093050003,
-0.11949581652879715,
-0.07699386775493622,
-0.1007273718714714,
-0.040252458304166794,
0.026367884129285812,
0.05946629121899605,
-0.1284945011138916,
-0.033084686845541,
-0.0694616511464119,
0.024802790954709053,
0.04326321929693222,
0.047772958874702454,
0.012666743248701096,
0.0015717739006504416,
0.0678388699889183,
-0.09064137190580368,
-0.01699104905128479,
-0.06419683992862701,
-0.07483242452144623,
0.07460271567106247,
0.054742004722356796,
0.12079132348299026,
0.13268741965293884,
-0.04286205768585205,
0.008183895610272884,
-0.03963044285774231,
0.24014659225940704,
-0.08033739030361176,
-0.031203504651784897,
0.09198997914791107,
-0.022548120468854904,
0.03602062538266182,
0.142562136054039,
0.05591815710067749,
-0.1340278536081314,
0.042604658752679825,
0.047591809183359146,
-0.030979957431554794,
-0.20227956771850586,
-0.027697743847966194,
-0.021399084478616714,
-0.03657754883170128,
0.080539770424366,
0.011845637112855911,
0.021769240498542786,
0.059210505336523056,
0.05361046642065048,
0.05034172534942627,
-0.013793925754725933,
0.06622382253408432,
0.07544273883104324,
0.05767607316374779,
0.1406608372926712,
-0.03974847123026848,
-0.08858466148376465,
0.024057583883404732,
-0.049865856766700745,
0.21521908044815063,
0.016961680725216866,
0.04793385788798332,
0.04440474137663841,
0.1487514227628708,
0.002132918918505311,
0.08277696371078491,
0.02750668302178383,
-0.07133277505636215,
-0.009022794663906097,
-0.042945101857185364,
-0.04058118164539337,
0.03359334170818329,
-0.0921512246131897,
0.04954272136092186,
-0.13070489466190338,
0.015003807842731476,
0.08214058727025986,
0.24844293296337128,
0.03795044496655464,
-0.32030272483825684,
-0.06642784178256989,
0.016746139153838158,
-0.036459118127822876,
-0.015232195146381855,
0.016733655706048012,
0.10040175914764404,
-0.08691836148500443,
0.065557561814785,
-0.06365717947483063,
0.0773104876279831,
-0.020545557141304016,
0.05832837522029877,
0.02597753144800663,
0.10479994863271713,
-0.031915806233882904,
0.04724252596497536,
-0.3425932228565216,
0.29131636023521423,
0.01932385563850403,
0.0957343652844429,
-0.05964261665940285,
-0.016643043607473373,
0.03817160800099373,
0.07852619141340256,
0.0595087967813015,
-0.027739370241761208,
-0.11054804921150208,
-0.21713605523109436,
-0.022802237421274185,
0.04340087249875069,
0.14659567177295685,
-0.014187193475663662,
0.12658019363880157,
-0.030487075448036194,
-0.0006146793020889163,
0.08453641831874847,
-0.04567848891019821,
-0.0952032133936882,
-0.058976661413908005,
-0.036885347217321396,
0.02007700316607952,
-0.00949364248663187,
-0.06192023307085037,
-0.10495369136333466,
-0.10793664306402206,
0.13795168697834015,
0.010264880955219269,
-0.011539842933416367,
-0.12787747383117676,
0.08981925249099731,
0.05509636178612709,
-0.07238873839378357,
0.03460347652435303,
0.029511909931898117,
0.06442044675350189,
0.033280596137046814,
-0.04402356222271919,
0.141699880361557,
-0.06684615463018417,
-0.17955255508422852,
-0.06099462881684303,
0.07523732632398605,
0.04177501052618027,
0.052651628851890564,
-0.0073569584637880325,
0.034363992512226105,
0.004543880466371775,
-0.08225049078464508,
0.042223669588565826,
-0.032563675194978714,
0.06094203516840935,
0.051353804767131805,
-0.07026699185371399,
-0.006103723309934139,
-0.07157876342535019,
-0.03291207179427147,
0.1637297123670578,
0.30928850173950195,
-0.0848713219165802,
0.009415550157427788,
0.04376363381743431,
-0.06217770650982857,
-0.21094641089439392,
0.09368106722831726,
0.04721198230981827,
-0.007267809007316828,
0.06702205538749695,
-0.14772361516952515,
0.12475959956645966,
0.09264451265335083,
-0.00955604761838913,
0.09540867805480957,
-0.2511541545391083,
-0.1420990377664566,
0.10445084422826767,
0.17636226117610931,
0.14860506355762482,
-0.15204401314258575,
-0.0010833307169377804,
-0.047481194138526917,
-0.08426839858293533,
0.08216365426778793,
-0.11336568742990494,
0.11581902951002121,
-0.009509352967143059,
0.0740760937333107,
0.015091451816260815,
-0.04802383482456207,
0.1026693657040596,
-0.003030539723113179,
0.13969686627388,
-0.07721041887998581,
-0.028492659330368042,
0.021153822541236877,
-0.05097278952598572,
-0.010352527722716331,
-0.07185210287570953,
0.023197587579488754,
-0.03265722841024399,
-0.023176824674010277,
-0.0763896107673645,
0.03202053904533386,
-0.032656632363796234,
-0.07288111001253128,
-0.030441123992204666,
0.036787353456020355,
0.0585755929350853,
-0.02538391761481762,
0.12687279284000397,
-0.023202262818813324,
0.1886776238679886,
0.09976620227098465,
0.09035109728574753,
-0.07278947532176971,
0.039004743099212646,
0.03052457608282566,
-0.02426324412226677,
0.04468553140759468,
-0.14617621898651123,
0.05004458129405975,
0.12882645428180695,
0.0023088727612048388,
0.13689278066158295,
0.08671880513429642,
-0.007914160378277302,
0.011036145500838757,
0.08621687442064285,
-0.17100553214550018,
-0.06898638606071472,
0.020138803869485855,
-0.08008634299039841,
-0.09434483200311661,
0.062202829867601395,
0.1100652888417244,
-0.06488718092441559,
-0.0045479461550712585,
-0.02501647174358368,
0.01178657729178667,
-0.05096783488988876,
0.2152719646692276,
0.05971099063754082,
0.054144568741321564,
-0.10000158101320267,
0.07587162405252457,
0.04028327018022537,
-0.059187836945056915,
0.028851671144366264,
0.0833515003323555,
-0.08772645145654678,
-0.02654366008937359,
0.11978667229413986,
0.2216269075870514,
-0.04289545863866806,
-0.02662707306444645,
-0.1611568182706833,
-0.12313926219940186,
0.05279488116502762,
0.2175792157649994,
0.09663613885641098,
-0.010117258876562119,
-0.014221632853150368,
0.0309744942933321,
-0.1482810527086258,
0.09306811541318893,
0.038808323442935944,
0.08289150893688202,
-0.14050285518169403,
0.1772104799747467,
-0.012948579154908657,
0.020834101364016533,
-0.03761231154203415,
0.048041488975286484,
-0.14211304485797882,
0.018495863303542137,
-0.12042806297540665,
-0.051289863884449005,
-0.002242805901914835,
0.003803281346336007,
0.005774381570518017,
-0.07058427482843399,
-0.0740547776222229,
0.00006347266753436998,
-0.12195280939340591,
-0.01410437747836113,
0.032950807362794876,
0.036070968955755234,
-0.11883839219808578,
-0.05600232258439064,
0.019589737057685852,
-0.06022990122437477,
0.05459563806653023,
0.03710402920842171,
0.007558661047369242,
0.07625075429677963,
-0.17974254488945007,
-0.008593399077653885,
0.07078316062688828,
-0.029608527198433876,
0.0756492167711258,
-0.08611763268709183,
-0.013870707713067532,
-0.011839802376925945,
0.10366872698068619,
0.03887728229165077,
0.088068388402462,
-0.1209431067109108,
0.03290373831987381,
-0.03725236654281616,
-0.09056083858013153,
-0.047132544219493866,
0.022519921883940697,
0.07812576740980148,
-0.01780959777534008,
0.18743029236793518,
-0.10966141521930695,
0.028643710538744926,
-0.20232616364955902,
-0.012715400196611881,
-0.01536631304770708,
-0.12528105080127716,
-0.13283617794513702,
-0.06390070915222168,
0.08395984768867493,
-0.05846128612756729,
0.13843218982219696,
0.038874879479408264,
0.0710468739271164,
0.04305899143218994,
-0.07266512513160706,
-0.003404550952836871,
0.04444229602813721,
0.17704463005065918,
0.04467366263270378,
-0.057141583412885666,
0.06868816167116165,
0.07093123346567154,
0.11866100877523422,
0.07631981372833252,
0.23551173508167267,
0.16397354006767273,
-0.028462806716561317,
0.09744308888912201,
0.010124091990292072,
-0.06849367171525955,
-0.11898193508386612,
0.0364358015358448,
-0.0670589953660965,
0.07220876216888428,
-0.03230396285653114,
0.17044518887996674,
0.07489653676748276,
-0.1558828204870224,
0.03634217381477356,
-0.08191264420747757,
-0.08568879961967468,
-0.1333719938993454,
0.03212924301624298,
-0.11573847383260727,
-0.1668701320886612,
0.009265286847949028,
-0.11468493938446045,
0.03245958313345909,
0.1084534302353859,
0.014341709204018116,
0.0015457018744200468,
0.1972513198852539,
0.01984511688351631,
0.05353185534477234,
0.05893949791789055,
-0.010484571568667889,
-0.027249468490481377,
-0.06782393157482147,
-0.09038379043340683,
-0.0148942144587636,
-0.017838656902313232,
0.021466879174113274,
-0.05871189758181572,
-0.09672670066356659,
0.04142056778073311,
-0.02320799045264721,
-0.11280298233032227,
0.03170935437083244,
0.041320983320474625,
0.06459657102823257,
0.017843518406152725,
0.008996465243399143,
-0.004629976116120815,
-0.008072104305028915,
0.25584912300109863,
-0.08384576439857483,
-0.09005899727344513,
-0.10071325302124023,
0.3032388985157013,
0.059616342186927795,
0.038083888590335846,
-0.0040626670233905315,
-0.07540474086999893,
0.013426226563751698,
0.21948371827602386,
0.17805205285549164,
-0.11264600604772568,
0.002036619232967496,
-0.03309449180960655,
-0.0082454988732934,
-0.020245982334017754,
0.12255646288394928,
0.12659384310245514,
-0.022431505843997,
-0.09036426246166229,
-0.02767249569296837,
-0.03187884762883186,
-0.02527254819869995,
-0.03788876533508301,
0.05258083716034889,
0.04956462234258652,
0.03121858462691307,
-0.0523727685213089,
0.06943540275096893,
-0.044521626085042953,
-0.11527295410633087,
0.06083358824253082,
-0.2222684621810913,
-0.16225393116474152,
-0.011417784728109837,
0.09212750196456909,
-0.006514462176710367,
0.06733875721693039,
-0.019250422716140747,
-0.022285031154751778,
0.021885475143790245,
-0.02396869659423828,
-0.02977832965552807,
-0.10740846395492554,
0.08123421669006348,
-0.13369221985340118,
0.2052876502275467,
-0.04592204838991165,
0.03153896704316139,
0.1341564655303955,
0.029055222868919373,
-0.061994243413209915,
0.09461741894483566,
0.03914087638258934,
-0.10746683180332184,
0.03395245969295502,
0.12361058592796326,
-0.06285907328128815,
0.07954709231853485,
0.053574442863464355,
-0.15618659555912018,
0.03429196774959564,
-0.07929681986570358,
-0.07842504233121872,
-0.0402042493224144,
-0.04922923818230629,
-0.051084619015455246,
0.12036466598510742,
0.22507242858409882,
-0.022518528625369072,
0.06694354116916656,
-0.07251438498497009,
0.029222480952739716,
0.04559464752674103,
0.05106866732239723,
-0.08216427266597748,
-0.2775473892688751,
0.028224332258105278,
0.13087525963783264,
-0.029224134981632233,
-0.2549545466899872,
-0.08030591905117035,
-0.000509970705024898,
-0.052769504487514496,
-0.1007470190525055,
0.11270082741975784,
0.09709520637989044,
0.05732458829879761,
-0.05543271824717522,
-0.1769789755344391,
-0.06387755274772644,
0.19097286462783813,
-0.13919052481651306,
-0.09544838219881058
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# SMIDS_3x_beit_large_RMSProp_lr00001_fold2
This model is a fine-tuned version of [microsoft/beit-large-patch16-224](https://huggingface.co/microsoft/beit-large-patch16-224) on the imagefolder dataset.
It achieves the following results on the evaluation set:
- Loss: 1.0857
- Accuracy: 0.9151
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 50
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|
| 0.2703 | 1.0 | 450 | 0.2711 | 0.9035 |
| 0.0491 | 2.0 | 900 | 0.4201 | 0.9018 |
| 0.0826 | 3.0 | 1350 | 0.4300 | 0.9118 |
| 0.0476 | 4.0 | 1800 | 0.6646 | 0.9085 |
| 0.0003 | 5.0 | 2250 | 0.6381 | 0.9135 |
| 0.1483 | 6.0 | 2700 | 0.7657 | 0.9035 |
| 0.0469 | 7.0 | 3150 | 0.7235 | 0.9068 |
| 0.0897 | 8.0 | 3600 | 0.9031 | 0.8985 |
| 0.1009 | 9.0 | 4050 | 0.7978 | 0.8968 |
| 0.058 | 10.0 | 4500 | 0.9975 | 0.8735 |
| 0.0 | 11.0 | 4950 | 0.9432 | 0.9035 |
| 0.035 | 12.0 | 5400 | 0.8407 | 0.8885 |
| 0.0 | 13.0 | 5850 | 0.9152 | 0.9035 |
| 0.0 | 14.0 | 6300 | 0.9215 | 0.9018 |
| 0.03 | 15.0 | 6750 | 0.9905 | 0.8852 |
| 0.0 | 16.0 | 7200 | 0.9325 | 0.8952 |
| 0.0524 | 17.0 | 7650 | 0.7955 | 0.9002 |
| 0.0 | 18.0 | 8100 | 0.8288 | 0.9168 |
| 0.0001 | 19.0 | 8550 | 0.9944 | 0.9101 |
| 0.0 | 20.0 | 9000 | 0.9683 | 0.9151 |
| 0.0301 | 21.0 | 9450 | 0.9011 | 0.9018 |
| 0.0 | 22.0 | 9900 | 0.9396 | 0.9068 |
| 0.0001 | 23.0 | 10350 | 1.0498 | 0.9168 |
| 0.0 | 24.0 | 10800 | 1.0341 | 0.9118 |
| 0.0001 | 25.0 | 11250 | 0.9397 | 0.9002 |
| 0.0 | 26.0 | 11700 | 0.8880 | 0.9135 |
| 0.0 | 27.0 | 12150 | 0.8988 | 0.9185 |
| 0.0 | 28.0 | 12600 | 0.9994 | 0.9201 |
| 0.006 | 29.0 | 13050 | 0.9581 | 0.9218 |
| 0.0 | 30.0 | 13500 | 0.9983 | 0.9018 |
| 0.0 | 31.0 | 13950 | 0.9678 | 0.9151 |
| 0.0 | 32.0 | 14400 | 0.9493 | 0.9085 |
| 0.0 | 33.0 | 14850 | 1.0032 | 0.9135 |
| 0.0 | 34.0 | 15300 | 0.9780 | 0.9185 |
| 0.0 | 35.0 | 15750 | 0.9913 | 0.9135 |
| 0.0 | 36.0 | 16200 | 1.0181 | 0.9201 |
| 0.0 | 37.0 | 16650 | 0.9777 | 0.9201 |
| 0.0 | 38.0 | 17100 | 1.0351 | 0.9151 |
| 0.0 | 39.0 | 17550 | 1.0920 | 0.9135 |
| 0.0052 | 40.0 | 18000 | 1.1361 | 0.9118 |
| 0.0 | 41.0 | 18450 | 1.1228 | 0.9052 |
| 0.0 | 42.0 | 18900 | 1.1117 | 0.9068 |
| 0.0 | 43.0 | 19350 | 1.0631 | 0.9135 |
| 0.0 | 44.0 | 19800 | 1.0763 | 0.9118 |
| 0.0 | 45.0 | 20250 | 1.0861 | 0.9118 |
| 0.0 | 46.0 | 20700 | 1.1097 | 0.9135 |
| 0.0 | 47.0 | 21150 | 1.0778 | 0.9151 |
| 0.005 | 48.0 | 21600 | 1.0803 | 0.9151 |
| 0.0 | 49.0 | 22050 | 1.0804 | 0.9151 |
| 0.0 | 50.0 | 22500 | 1.0857 | 0.9151 |
### Framework versions
- Transformers 4.32.1
- Pytorch 2.0.1
- Datasets 2.12.0
- Tokenizers 0.13.2
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["imagefolder"], "metrics": ["accuracy"], "base_model": "microsoft/beit-large-patch16-224", "model-index": [{"name": "SMIDS_3x_beit_large_RMSProp_lr00001_fold2", "results": [{"task": {"type": "image-classification", "name": "Image Classification"}, "dataset": {"name": "imagefolder", "type": "imagefolder", "config": "default", "split": "test", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.9151414309484193, "name": "Accuracy"}]}]}]} | image-classification | onizukal/SMIDS_3x_beit_large_RMSProp_lr00001_fold2 | [
"transformers",
"pytorch",
"beit",
"image-classification",
"generated_from_trainer",
"dataset:imagefolder",
"base_model:microsoft/beit-large-patch16-224",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T17:01:51+00:00 | [] | [] | TAGS
#transformers #pytorch #beit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/beit-large-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| SMIDS\_3x\_beit\_large\_RMSProp\_lr00001\_fold2
===============================================
This model is a fine-tuned version of microsoft/beit-large-patch16-224 on the imagefolder dataset.
It achieves the following results on the evaluation set:
* Loss: 1.0857
* Accuracy: 0.9151
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 1e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* num\_epochs: 50
### Training results
### Framework versions
* Transformers 4.32.1
* Pytorch 2.0.1
* Datasets 2.12.0
* Tokenizers 0.13.2
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.32.1\n* Pytorch 2.0.1\n* Datasets 2.12.0\n* Tokenizers 0.13.2"
] | [
"TAGS\n#transformers #pytorch #beit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/beit-large-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.32.1\n* Pytorch 2.0.1\n* Datasets 2.12.0\n* Tokenizers 0.13.2"
] | [
81,
116,
4,
30
] | [
"passage: TAGS\n#transformers #pytorch #beit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/beit-large-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50### Training results### Framework versions\n\n\n* Transformers 4.32.1\n* Pytorch 2.0.1\n* Datasets 2.12.0\n* Tokenizers 0.13.2"
] | [
-0.1292150914669037,
0.17132072150707245,
-0.002415567170828581,
0.13183215260505676,
0.11657863855361938,
0.020753253251314163,
0.1335890144109726,
0.16620413959026337,
-0.08238927274942398,
0.04929587244987488,
0.13697229325771332,
0.1357421576976776,
0.04955337569117546,
0.20790311694145203,
-0.053285520523786545,
-0.26080378890037537,
0.0391765721142292,
0.03443576768040657,
-0.020672276616096497,
0.12494900077581406,
0.09484300017356873,
-0.1312379240989685,
0.11272566765546799,
0.025938162580132484,
-0.20840293169021606,
-0.033587437123060226,
-0.01026944164186716,
-0.06854863464832306,
0.10221196711063385,
0.001568986801430583,
0.0741027221083641,
0.037979885935783386,
0.08491890877485275,
-0.12677186727523804,
0.000941311358474195,
0.04326357692480087,
0.0062435888685286045,
0.1065368577837944,
0.062226198613643646,
-0.008521218784153461,
0.06926212459802628,
-0.07453521341085434,
0.06115834787487984,
0.008060229010879993,
-0.11478453874588013,
-0.2692618668079376,
-0.09817449003458023,
0.07377522438764572,
0.08109822124242783,
0.06491127610206604,
0.006432840134948492,
0.16222304105758667,
-0.015434488654136658,
0.1024109497666359,
0.23076069355010986,
-0.2713507413864136,
-0.054792311042547226,
0.022649891674518585,
0.0155020197853446,
0.06252340972423553,
-0.10333037376403809,
-0.01993185468018055,
0.019141921773552895,
0.042880840599536896,
0.14450453221797943,
-0.012332411482930183,
-0.03331032395362854,
-0.02637922763824463,
-0.11139829456806183,
-0.08930420875549316,
0.18604889512062073,
0.06140090152621269,
-0.04917457327246666,
-0.07841385900974274,
-0.07612120360136032,
-0.17419220507144928,
-0.03924720734357834,
0.008911197073757648,
0.046679239720106125,
-0.04711441695690155,
-0.10239296406507492,
-0.03511375933885574,
-0.07504668086767197,
-0.05196268856525421,
-0.026160720735788345,
0.1420334428548813,
0.03879573196172714,
0.05471520125865936,
-0.027205273509025574,
0.10149593651294708,
0.010796112939715385,
-0.1717151701450348,
-0.02661297097802162,
0.0005703883362002671,
0.010487399995326996,
-0.01821139082312584,
-0.029929913580417633,
-0.06737607717514038,
-0.003975129686295986,
0.15347014367580414,
-0.07002666592597961,
0.058850113302469254,
-0.0054583000019192696,
0.041531506925821304,
-0.049319881945848465,
0.1874888390302658,
-0.029916515573859215,
-0.016198426485061646,
0.019476165995001793,
0.08928463608026505,
0.0656052976846695,
-0.030047036707401276,
-0.12371734529733658,
0.021691862493753433,
0.13241209089756012,
0.006458523217588663,
-0.022870952263474464,
0.054544735699892044,
-0.0711979940533638,
-0.0584990456700325,
0.09274657070636749,
-0.09275025129318237,
0.035496871918439865,
-0.011692462489008904,
-0.08981472253799438,
-0.06787234544754028,
0.029122935608029366,
0.011931490153074265,
-0.009771439246833324,
0.06940538436174393,
-0.09093258529901505,
0.01846885494887829,
-0.06650768965482712,
-0.09852384030818939,
0.01388985849916935,
-0.11549968272447586,
0.010918805375695229,
-0.10079170018434525,
-0.19154705107212067,
0.0032797311432659626,
0.07527101784944534,
-0.06246669217944145,
-0.06951755285263062,
-0.033377837389707565,
-0.07729615271091461,
0.03790769353508949,
-0.01523390132933855,
0.07408059388399124,
-0.07056254893541336,
0.09071778506040573,
0.02892814762890339,
0.09002465009689331,
-0.052364569157361984,
0.048610031604766846,
-0.09854818880558014,
0.05158581584692001,
-0.19896768033504486,
0.0824570581316948,
-0.04529954120516777,
0.05730293318629265,
-0.10005063563585281,
-0.10804302245378494,
0.029095064848661423,
-0.0466112419962883,
0.07224688678979874,
0.09985066950321198,
-0.16068536043167114,
-0.05396431311964989,
0.14283035695552826,
-0.09281232208013535,
-0.14269256591796875,
0.09829698503017426,
-0.045770496129989624,
0.014614340849220753,
0.04329100251197815,
0.2130173146724701,
0.04901750758290291,
-0.08417420834302902,
-0.023242823779582977,
-0.02969830296933651,
0.03785223513841629,
-0.0668954998254776,
0.10032020509243011,
0.025215676054358482,
0.05325069651007652,
0.02284027636051178,
-0.029413679614663124,
0.04126512631773949,
-0.08672589063644409,
-0.09880872070789337,
-0.053216658532619476,
-0.0853687971830368,
0.03892384096980095,
0.05334646999835968,
0.0614997074007988,
-0.10279879719018936,
-0.09344549477100372,
0.0453280434012413,
0.09495674818754196,
-0.07567895948886871,
0.02865210548043251,
-0.08989366888999939,
0.10926083475351334,
-0.08635354787111282,
-0.02427433431148529,
-0.18316780030727386,
-0.041861772537231445,
0.04194685444235802,
-0.025394707918167114,
-0.007599220145493746,
-0.05216266214847565,
0.06521623581647873,
0.0848059430718422,
-0.05379978567361832,
-0.05897609516978264,
-0.05670713260769844,
0.002749721286818385,
-0.10883764177560806,
-0.17341645061969757,
-0.08353621512651443,
-0.03381705656647682,
0.14265403151512146,
-0.15880316495895386,
0.019960513338446617,
0.05115775763988495,
0.12808771431446075,
0.060330405831336975,
-0.044940851628780365,
-0.0009795452933758497,
0.02373526245355606,
-0.05278978496789932,
-0.09012233465909958,
0.059676408767700195,
0.0331520177423954,
-0.07579167187213898,
-0.016548609361052513,
-0.09850107133388519,
0.1460651308298111,
0.1280234009027481,
-0.010448831133544445,
-0.04986010119318962,
-0.011923554353415966,
-0.06967874616384506,
-0.030430803075432777,
-0.036602724343538284,
0.019139016047120094,
0.09450183063745499,
0.012393946759402752,
0.14818525314331055,
-0.09332848340272903,
-0.034156475216150284,
0.05024607852101326,
-0.028047295287251472,
-0.03259625658392906,
0.0731319710612297,
0.025664178654551506,
-0.14941470324993134,
0.14837577939033508,
0.14845694601535797,
-0.04714515432715416,
0.12564225494861603,
-0.03889495134353638,
-0.06329566240310669,
-0.04632000997662544,
-0.02844901941716671,
0.013190032914280891,
0.13346467912197113,
-0.076783187687397,
-0.004412572830915451,
0.05686868354678154,
0.017921162769198418,
-0.004722983110696077,
-0.1827412098646164,
0.003951311111450195,
0.0321657620370388,
-0.05121494084596634,
-0.011695281602442265,
-0.017026077955961227,
0.003609517589211464,
0.09151934087276459,
0.02040533348917961,
-0.06441836804151535,
0.05384209007024765,
0.012033452279865742,
-0.05366513133049011,
0.1677880585193634,
-0.07823625206947327,
-0.20364677906036377,
-0.12268579006195068,
-0.06752478331327438,
-0.10258819162845612,
0.012170074507594109,
0.06315170973539352,
-0.04569438472390175,
-0.050954580307006836,
-0.0997823104262352,
-0.037851084023714066,
0.021281057968735695,
0.026625970378518105,
0.05139283835887909,
-0.005415658466517925,
0.09185726940631866,
-0.09241294115781784,
-0.030897676944732666,
-0.01631389558315277,
0.009287231601774693,
0.06772445887327194,
0.019780615344643593,
0.1102219671010971,
0.07713042199611664,
-0.029881305992603302,
0.05137522891163826,
-0.013354548253118992,
0.2620471715927124,
-0.06917091459035873,
-0.002909549279138446,
0.1375615894794464,
-0.015162656083703041,
0.08283410966396332,
0.1273423582315445,
0.041794080287218094,
-0.09746479243040085,
-0.011291430331766605,
-0.0008301159832626581,
-0.049490246921777725,
-0.16143162548542023,
-0.04317644611001015,
-0.0434197373688221,
-0.010716320015490055,
0.1416788250207901,
0.03848205506801605,
0.024626927450299263,
0.07702240347862244,
0.015813151374459267,
0.057987019419670105,
-0.02077260985970497,
0.1017511859536171,
0.0805719867348671,
0.06816057115793228,
0.13305824995040894,
-0.036980245262384415,
-0.02092074789106846,
0.057033997029066086,
0.04002218693494797,
0.21362732350826263,
-0.02804172970354557,
0.15433214604854584,
0.026679744943976402,
0.1909136176109314,
0.019870078191161156,
0.07247955352067947,
-0.010095180943608284,
0.0028269465547055006,
-0.018500015139579773,
-0.04554403945803642,
-0.05979170650243759,
0.03185109794139862,
-0.016015755012631416,
0.05207211896777153,
-0.09269700944423676,
0.028567379340529442,
0.06037893891334534,
0.3028397262096405,
0.061388690024614334,
-0.41139692068099976,
-0.09273239970207214,
0.009406263940036297,
-0.002105827210471034,
-0.06053102761507034,
-0.011343861930072308,
0.09683393687009811,
-0.09968853741884232,
0.08300996571779251,
-0.09414921700954437,
0.08760150521993637,
-0.08863518387079239,
0.016419410705566406,
0.07728815078735352,
0.06722814589738846,
0.01766069419682026,
0.057678405195474625,
-0.22131015360355377,
0.2517315745353699,
0.02006395347416401,
0.04867706075310707,
-0.08515261113643646,
0.013813616707921028,
0.029918700456619263,
0.058915551751852036,
0.08619558066129684,
0.0083828279748559,
-0.09208258241415024,
-0.19043345749378204,
-0.12182265520095825,
-0.0015020827995613217,
0.06677291542291641,
-0.03118232637643814,
0.0942893773317337,
-0.01760665327310562,
-0.012930129654705524,
0.019664883613586426,
0.00020212549134157598,
-0.039232417941093445,
-0.09916181117296219,
0.019594477489590645,
0.03770963475108147,
-0.0040510352700948715,
-0.06473120301961899,
-0.1088499054312706,
-0.027749689295887947,
0.1611177921295166,
0.0489477813243866,
-0.07595206052064896,
-0.14163517951965332,
0.0831608697772026,
0.0844789668917656,
-0.08478974550962448,
0.046326830983161926,
-0.015740465372800827,
0.14427345991134644,
0.02813553437590599,
-0.08791226893663406,
0.10567717254161835,
-0.05589807406067848,
-0.18345315754413605,
-0.035460758954286575,
0.09823724627494812,
0.006449915003031492,
0.047238387167453766,
0.0029976284131407738,
0.05834325775504112,
-0.03208146244287491,
-0.05784951522946358,
0.06896662712097168,
-0.0034485149662941694,
0.1075923964381218,
-0.0061480943113565445,
-0.0032397336326539516,
0.02182089537382126,
-0.04197082296013832,
-0.0014782516518607736,
0.1645156890153885,
0.23995232582092285,
-0.10496784001588821,
0.055536478757858276,
0.030249565839767456,
-0.03645236790180206,
-0.18277540802955627,
0.009984065778553486,
0.08414819091558456,
0.0021475672256201506,
0.040169790387153625,
-0.1663118302822113,
0.05386544391512871,
0.10983236879110336,
-0.04191310703754425,
0.07995743304491043,
-0.2803034782409668,
-0.1190505102276802,
0.08906996995210648,
0.13602600991725922,
0.06884066760540009,
-0.13274545967578888,
-0.045290667563676834,
-0.039063699543476105,
-0.16666166484355927,
0.1351267695426941,
-0.04754851385951042,
0.11997194588184357,
-0.040666740387678146,
0.06989686191082001,
0.015085658058524132,
-0.05448267608880997,
0.14587333798408508,
0.00877679605036974,
0.0857420563697815,
-0.07118549197912216,
0.0021252231672406197,
0.10074540972709656,
-0.0982399731874466,
0.07668103277683258,
-0.08308075368404388,
0.06399426609277725,
-0.11283876746892929,
-0.007322354707866907,
-0.07328318059444427,
0.015542288310825825,
-0.012007588520646095,
-0.043488435447216034,
-0.04113076627254486,
0.03472091257572174,
0.06403200328350067,
-0.015996064990758896,
0.20271754264831543,
0.0629286915063858,
0.08313194662332535,
0.17939580976963043,
0.04974674805998802,
-0.096995510160923,
-0.09814400225877762,
-0.04502987116575241,
-0.028452320024371147,
0.06312472373247147,
-0.13321243226528168,
0.05335186421871185,
0.1209464818239212,
0.008661448024213314,
0.12983813881874084,
0.054849762469530106,
-0.0316605418920517,
0.033173978328704834,
0.06366948038339615,
-0.16513317823410034,
-0.08843576163053513,
-0.011303714476525784,
0.01758752204477787,
-0.12545546889305115,
0.0447046272456646,
0.12079240381717682,
-0.057224519550800323,
-0.015418118797242641,
-0.0026640621945261955,
0.03586487099528313,
-0.00886022113263607,
0.16030296683311462,
0.05005719140172005,
0.05675157532095909,
-0.11541767418384552,
0.1181424930691719,
0.06067226454615593,
-0.0710521712899208,
0.031696248799562454,
0.05698402598500252,
-0.10586927086114883,
-0.022646361961960793,
0.03662630170583725,
0.14154238998889923,
-0.06414706259965897,
-0.04990902543067932,
-0.13196614384651184,
-0.0909038558602333,
0.07024894654750824,
0.0724560096859932,
0.09284354001283646,
0.016252439469099045,
-0.031063025817275047,
-0.014114780351519585,
-0.10623957961797714,
0.10545456409454346,
0.04753988981246948,
0.09451808035373688,
-0.17563696205615997,
0.06374634802341461,
0.0007657874375581741,
0.07206296175718307,
-0.024532334879040718,
0.005616967566311359,
-0.09020458161830902,
-0.0008940583793446422,
-0.10660925507545471,
0.025940274819731712,
-0.04968960955739021,
0.0027822551783174276,
-0.020955873653292656,
-0.058104176074266434,
-0.06385789811611176,
0.02704726532101631,
-0.11796805262565613,
-0.05728267878293991,
0.01832517236471176,
0.029680335894227028,
-0.11609132587909698,
-0.04758497327566147,
0.014494677074253559,
-0.09034118801355362,
0.09993617236614227,
0.05929066613316536,
-0.006737631745636463,
0.0029803363140672445,
0.011042662896215916,
-0.02363271825015545,
0.06827948242425919,
0.006517379079014063,
0.07795335352420807,
-0.11366859823465347,
-0.018052512779831886,
0.017967568710446358,
-0.002112566027790308,
0.011524608358740807,
0.15499049425125122,
-0.12699781358242035,
-0.0033930845092982054,
-0.022802060469985008,
-0.06095515564084053,
-0.06754840165376663,
0.06765563786029816,
0.10613249987363815,
0.0214694757014513,
0.2064255326986313,
-0.054858945310115814,
0.01148067507892847,
-0.21229742467403412,
-0.011367390863597393,
0.0014767643297091126,
-0.1394193321466446,
-0.10240225493907928,
-0.03432944789528847,
0.0646229088306427,
-0.07021024078130722,
0.1212792620062828,
0.036924295127391815,
0.015180133283138275,
0.028698688372969627,
0.025451842695474625,
-0.009322993457317352,
0.01828060857951641,
0.16467928886413574,
0.014544252306222916,
-0.030929861590266228,
0.12307319045066833,
0.026831358671188354,
0.0918813943862915,
0.11550118029117584,
0.17162561416625977,
0.1226300448179245,
0.042329173535108566,
0.09527058154344559,
0.05073356628417969,
-0.032373297959566116,
-0.2198440134525299,
0.04109371080994606,
-0.043747998774051666,
0.14987531304359436,
-0.0034218686632812023,
0.15886609256267548,
0.08696271479129791,
-0.1824999451637268,
0.04266338422894478,
-0.02988567017018795,
-0.08202743530273438,
-0.08238054066896439,
-0.1163601353764534,
-0.10495591163635254,
-0.15148837864398956,
0.0012598474277183414,
-0.10238117724657059,
0.02373862825334072,
0.11528778076171875,
-0.010980993509292603,
-0.00952758826315403,
0.1250862330198288,
-0.01644187793135643,
0.019042596220970154,
0.04508042708039284,
0.007425562012940645,
-0.05218745768070221,
-0.04613304138183594,
-0.08413935452699661,
0.015972480177879333,
0.0363130047917366,
0.05680973082780838,
-0.03208919242024422,
-0.008708061650395393,
0.03847881406545639,
-0.008026620373129845,
-0.12142552435398102,
0.013289375230669975,
0.007551861461251974,
0.04767835885286331,
-0.004989264067262411,
0.007813788950443268,
0.026865217834711075,
-0.01780105195939541,
0.195222407579422,
-0.06977689266204834,
-0.02860948257148266,
-0.12041912227869034,
0.17737813293933868,
0.00569287920370698,
-0.048185933381319046,
0.05394943431019783,
-0.09105358272790909,
-0.02213868498802185,
0.15108588337898254,
0.18787547945976257,
-0.06683575361967087,
-0.017941389232873917,
-0.014669668860733509,
-0.01477136928588152,
-0.01832989603281021,
0.10442051291465759,
0.09986825287342072,
-0.004740583244711161,
-0.07264549285173416,
-0.024389909580349922,
-0.06369390338659286,
-0.032235804945230484,
-0.04127946496009827,
0.07026855647563934,
-0.001124961650930345,
0.005972458980977535,
-0.07571399211883545,
0.03954308480024338,
-0.020357538014650345,
-0.06112333759665489,
0.07204564660787582,
-0.21083933115005493,
-0.1802441030740738,
0.0017737408634275198,
0.07683850824832916,
0.0021866720635443926,
0.04613208398222923,
-0.012570524588227272,
0.018509654328227043,
0.07427240163087845,
-0.02333001233637333,
-0.08794470131397247,
-0.09525144845247269,
0.1020299568772316,
-0.13951729238033295,
0.24700812995433807,
-0.03552914783358574,
0.0377071388065815,
0.1201176866889,
0.03583609312772751,
-0.13580889999866486,
0.03513867408037186,
0.03722600266337395,
-0.02918340638279915,
0.0181744247674942,
0.14616045355796814,
-0.03901152312755585,
0.07440102845430374,
0.04275068640708923,
-0.10678882896900177,
-0.04424819350242615,
-0.04619530588388443,
-0.015570126473903656,
-0.02712010033428669,
-0.05963090807199478,
-0.04089967906475067,
0.12949442863464355,
0.17410574853420258,
-0.04094170406460762,
-0.021948745474219322,
-0.06438223272562027,
0.035308949649333954,
0.08067496865987778,
-0.026465818285942078,
-0.04482371732592583,
-0.2364819198846817,
0.0028874515555799007,
0.050913918763399124,
-0.008316555991768837,
-0.19871793687343597,
-0.10607530176639557,
-0.00044736277777701616,
-0.05943094193935394,
-0.08227076381444931,
0.09325046092271805,
0.06211918964982033,
0.03563893958926201,
-0.06190048158168793,
0.02738066203892231,
-0.07750356942415237,
0.14178979396820068,
-0.14600589871406555,
-0.07656177133321762
] |
null | null | null |
A test quantization of OpenHermes-2.5-Mistral-7B by teknium using importance matrices computed on Ukrainian text, hopefully decreasing the coherence hit after quantization in Ukrainian at the cost of some performance in other languages.
Importance matrix was computed in roughly 20 minutes with a Ryzen 5 3550H and GTX 1650 with 8 layers offloaded.
Will be updated with perplexity testing later, probably? 😭 Haven't done proper tests quite yet, feels better than old quants when chatting in Ukrainian, hopefully I get around to actually benching it somehow | {"license": "apache-2.0"} | null | interrobang/OpenHermes-2.5-Mistral-7B-GGUF-ukrainian-imatrix | [
"gguf",
"license:apache-2.0",
"region:us"
] | 2024-02-07T17:04:31+00:00 | [] | [] | TAGS
#gguf #license-apache-2.0 #region-us
|
A test quantization of OpenHermes-2.5-Mistral-7B by teknium using importance matrices computed on Ukrainian text, hopefully decreasing the coherence hit after quantization in Ukrainian at the cost of some performance in other languages.
Importance matrix was computed in roughly 20 minutes with a Ryzen 5 3550H and GTX 1650 with 8 layers offloaded.
Will be updated with perplexity testing later, probably? Haven't done proper tests quite yet, feels better than old quants when chatting in Ukrainian, hopefully I get around to actually benching it somehow | [] | [
"TAGS\n#gguf #license-apache-2.0 #region-us \n"
] | [
17
] | [
"passage: TAGS\n#gguf #license-apache-2.0 #region-us \n"
] | [
-0.01330315787345171,
0.15578481554985046,
-0.008150381036102772,
0.0016573866596445441,
-0.04673202335834503,
0.030570833012461662,
0.13376784324645996,
0.09194369614124298,
0.128515362739563,
-0.07453092932701111,
0.17055568099021912,
0.028890712186694145,
0.03800550848245621,
0.03976481780409813,
0.025271227583289146,
-0.11722369492053986,
0.10813431441783905,
-0.03824392333626747,
-0.09146501868963242,
-0.0007550549926236272,
0.0461740680038929,
0.03938475623726845,
-0.020297255367040634,
-0.012304259464144707,
-0.04646717384457588,
-0.01807590387761593,
0.061380259692668915,
-0.021618852391839027,
0.0584920234978199,
0.044396135956048965,
-0.013440944254398346,
0.02849559485912323,
-0.044239118695259094,
-0.20778964459896088,
0.01814555749297142,
-0.030956419184803963,
-0.0963222086429596,
0.034455686807632446,
-0.002720483811572194,
0.03162013739347458,
0.061540063470602036,
0.11115614324808121,
-0.11918631196022034,
0.037214748561382294,
-0.13505050539970398,
-0.30473440885543823,
-0.18398606777191162,
0.06021247059106827,
0.0028733352664858103,
0.041688770055770874,
0.07521043717861176,
0.05276030674576759,
-0.15611688792705536,
-0.04681211709976196,
0.06740783900022507,
-0.35676148533821106,
0.026624629274010658,
0.1421503871679306,
-0.04767010360956192,
0.03409823030233383,
-0.016959721222519875,
0.10372579842805862,
0.08160287141799927,
-0.0309890303760767,
-0.11352469027042389,
-0.04248137027025223,
-0.07062457501888275,
0.11909499764442444,
-0.022388093173503876,
-0.10835286229848862,
0.299734890460968,
0.07829401642084122,
-0.03949820250272751,
0.13934776186943054,
-0.018232177942991257,
0.12180425226688385,
0.0037959686014801264,
0.07966145128011703,
0.07500946521759033,
0.2106771469116211,
0.1879701316356659,
-0.1285460740327835,
-0.15332761406898499,
-0.07658189535140991,
-0.18097953498363495,
0.09647365659475327,
0.00481637567281723,
0.15394099056720734,
-0.12812361121177673,
0.004832482896745205,
-0.1665661633014679,
-0.05767213553190231,
-0.09051498025655746,
-0.052037935703992844,
0.14505687355995178,
0.09770436584949493,
-0.07494958490133286,
0.17277710139751434,
0.21161453425884247,
0.26735973358154297,
-0.03192757070064545,
0.006477556191384792,
-0.06420958042144775,
0.15842780470848083,
-0.0583864264190197,
0.019410474225878716,
0.05191675201058388,
0.11087486892938614,
0.13493096828460693,
-0.1719009280204773,
0.09378605335950851,
-0.016259867697954178,
-0.11252851784229279,
-0.011826573871076107,
-0.1737414002418518,
0.1333753764629364,
0.046528834849596024,
-0.08763917535543442,
-0.02980431728065014,
0.06516948342323303,
0.13636615872383118,
-0.016305889934301376,
-0.012872343882918358,
-0.008937547914683819,
0.009359321556985378,
-0.06836733967065811,
0.0008180695003829896,
0.060471873730421066,
0.08052167296409607,
-0.010497825220227242,
-0.09346228837966919,
-0.02963828109204769,
0.03977816551923752,
0.13056598603725433,
0.12507236003875732,
-0.03587067127227783,
0.04867643862962723,
-0.07281927764415741,
-0.1537899523973465,
0.043609101325273514,
0.0720045417547226,
0.01405393797904253,
-0.02105209231376648,
0.12900052964687347,
0.02947195991873741,
-0.002579237800091505,
-0.07217522710561752,
-0.019955281168222427,
-0.08173951506614685,
0.05495830997824669,
-0.0830233246088028,
-0.021544981747865677,
-0.25846993923187256,
-0.004414012189954519,
-0.08588613569736481,
0.05437181144952774,
0.026437964290380478,
-0.07966095209121704,
-0.15488311648368835,
0.15681788325309753,
-0.04755254089832306,
0.0606415681540966,
-0.055992551147937775,
-0.019309088587760925,
-0.04433238506317139,
0.07082654535770416,
-0.06881338357925415,
-0.034905560314655304,
0.21830801665782928,
-0.14372503757476807,
-0.12107643485069275,
0.045915182679891586,
0.058830294758081436,
-0.05773618072271347,
0.03722799941897392,
0.3155798017978668,
-0.033125780522823334,
-0.02127721533179283,
0.10441244393587112,
0.17700493335723877,
-0.10252934694290161,
-0.19990848004817963,
0.1626713126897812,
-0.18464820086956024,
-0.21314117312431335,
0.03374022617936134,
-0.1354188770055771,
0.14721231162548065,
0.029108481481671333,
-0.10722507536411285,
-0.042635608464479446,
-0.08002658933401108,
-0.048498500138521194,
-0.0413840115070343,
0.047936778515577316,
-0.04613621532917023,
0.04448116570711136,
-0.18495601415634155,
0.06669430434703827,
0.12992608547210693,
0.04142272472381592,
-0.04430747032165527,
0.06424026191234589,
0.018889864906668663,
0.04718081280589104,
0.014743788167834282,
-0.021515699103474617,
0.02174583449959755,
-0.09805754572153091,
0.07063749432563782,
0.08627311885356903,
0.059764452278614044,
-0.06353570520877838,
0.012208250351250172,
0.06167324259877205,
0.0011235947022214532,
0.03216798976063728,
0.06799250841140747,
-0.09858707338571548,
0.1041717529296875,
0.003250909736379981,
0.08585058897733688,
0.03630579635500908,
-0.01942511461675167,
0.1006988137960434,
-0.05730016902089119,
-0.06386759132146835,
0.0249167513102293,
0.008679588325321674,
-0.1123015284538269,
0.05832165852189064,
-0.01720862090587616,
0.08887401223182678,
0.0644783154129982,
-0.10543974488973618,
0.169432133436203,
0.06818334013223648,
0.20611946284770966,
0.1413285881280899,
0.03921271115541458,
0.09898378700017929,
-0.015305405482649803,
-0.01739814504981041,
-0.0033933527301996946,
0.0786864310503006,
0.03553178906440735,
-0.018575791269540787,
-0.03419099375605583,
-0.01237060222774744,
-0.028527243062853813,
0.009019540622830391,
-0.028413871303200722,
-0.08427856862545013,
-0.038365963846445084,
-0.00293926103040576,
0.18400533497333527,
-0.1352609246969223,
0.15670859813690186,
0.39662665128707886,
0.05809931084513664,
0.10376974195241928,
-0.1573796272277832,
-0.04060497134923935,
-0.04825512692332268,
0.03319299966096878,
-0.02701469324529171,
0.184132382273674,
-0.10087321698665619,
0.03391242399811745,
0.07232333719730377,
0.057861048728227615,
0.06244037300348282,
-0.17538565397262573,
-0.14002028107643127,
-0.02046295255422592,
-0.09578375518321991,
-0.104468435049057,
0.024245386943221092,
-0.16045768558979034,
0.026515116915106773,
0.026555776596069336,
-0.06167523190379143,
0.1598798930644989,
-0.008611330762505531,
-0.04725697636604309,
0.07082630693912506,
-0.20357191562652588,
-0.12087341398000717,
-0.07626573741436005,
-0.0042744106613099575,
-0.05095795914530754,
0.023625193163752556,
0.08465500921010971,
-0.09592214971780777,
-0.05633152276277542,
0.07234445214271545,
-0.09036894887685776,
-0.06662289053201675,
0.015421921387314796,
0.07410863041877747,
0.01950029656291008,
0.031299129128456116,
-0.10658411681652069,
-0.05114024132490158,
-0.0207142923027277,
-0.06643637269735336,
0.043042417615652084,
-0.029125841334462166,
0.0803024172782898,
0.09105047583580017,
0.08921805769205093,
0.06849837303161621,
-0.0003459077561274171,
0.14692601561546326,
-0.04780511185526848,
-0.07088837027549744,
0.1597827970981598,
0.02298906072974205,
0.03533780947327614,
0.10251420736312866,
0.07653065770864487,
-0.10019218176603317,
-0.033925220370292664,
-0.03363695368170738,
-0.11724339425563812,
-0.2371799200773239,
-0.02427528239786625,
-0.08114678412675858,
0.09731483459472656,
-0.03697621077299118,
0.1421261876821518,
0.1297444850206375,
0.07038690894842148,
0.01488346979022026,
-0.02049473114311695,
-0.0003764019056688994,
-0.03656056523323059,
0.1047646775841713,
-0.040648601949214935,
-0.04150701314210892,
-0.13765324652194977,
0.05280669033527374,
0.2123951017856598,
0.139052614569664,
0.13875830173492432,
0.2556404173374176,
0.0865810289978981,
0.14902375638484955,
0.1700367033481598,
0.043575435876846313,
0.01834452524781227,
0.02130916342139244,
-0.025624319911003113,
-0.0733172670006752,
-0.021722665056586266,
-0.024601932615041733,
0.060184698551893234,
-0.011125342920422554,
-0.20776666700839996,
0.06618745625019073,
-0.23978964984416962,
0.06574483960866928,
0.0798097774386406,
0.06448644399642944,
0.07763633877038956,
0.05836908146739006,
0.08228960633277893,
0.07039085030555725,
0.030951116234064102,
0.1121925488114357,
-0.05280335620045662,
-0.07145904004573822,
0.08945973217487335,
0.033000752329826355,
0.055582351982593536,
0.05654873698949814,
0.018150946125388145,
-0.09060613065958023,
-0.1282111406326294,
0.06521598249673843,
0.14194566011428833,
-0.2339191883802414,
0.21457308530807495,
0.023072047159075737,
-0.07668514549732208,
-0.039042409509420395,
-0.00877209473401308,
0.10390420258045197,
0.15340037643909454,
0.12154663354158401,
0.0913349986076355,
-0.1597919762134552,
0.06428830325603485,
-0.07498005777597427,
0.06961594521999359,
-0.0024978057481348515,
-0.04789665341377258,
-0.1743926703929901,
-0.04370199516415596,
0.05610458180308342,
0.02999473176896572,
0.14242549240589142,
-0.15068703889846802,
-0.08502397686243057,
0.047799695283174515,
0.13189667463302612,
-0.012925185263156891,
-0.1351172924041748,
0.07789962738752365,
-0.0223079863935709,
0.13693971931934357,
-0.08600178360939026,
-0.009268487803637981,
-0.06393920630216599,
-0.11813131719827652,
0.03256971016526222,
-0.03487817198038101,
0.030751710757613182,
-0.08259879052639008,
-0.11836646497249603,
-0.0880444347858429,
-0.19962769746780396,
0.0970500111579895,
-0.0914505124092102,
0.011035526171326637,
-0.007812321186065674,
0.10870376229286194,
-0.04637550935149193,
0.011202405206859112,
-0.014477218501269817,
-0.005613986402750015,
-0.037054069340229034,
-0.18300330638885498,
0.12522146105766296,
-0.0354662649333477,
-0.038446735590696335,
0.010260424576699734,
0.029084086418151855,
0.08555833250284195,
0.07213917374610901,
-0.11324945837259293,
0.17003758251667023,
0.28164100646972656,
-0.06761249154806137,
0.21416988968849182,
0.31458553671836853,
-0.0837627723813057,
-0.2168940305709839,
-0.20412485301494598,
-0.243454247713089,
-0.1341201663017273,
0.013469581492245197,
-0.22552542388439178,
0.0766766369342804,
0.20877110958099365,
-0.1714908480644226,
0.2946220934391022,
-0.1911279857158661,
-0.002857143757864833,
0.18236051499843597,
-0.052634578198194504,
0.44611573219299316,
-0.171253964304924,
-0.12041395157575607,
-0.033438790589571,
-0.19569972157478333,
0.12554602324962616,
-0.16359169781208038,
0.05501075088977814,
0.024149509146809578,
-0.08415087312459946,
-0.054437655955553055,
-0.02173338644206524,
0.2504390478134155,
-0.003609958803281188,
0.04743576794862747,
-0.08139622211456299,
0.020269259810447693,
0.1750870943069458,
0.0035608301404863596,
0.017100313678383827,
-0.1639900505542755,
-0.01987898163497448,
-0.011092170141637325,
0.017119288444519043,
-0.04806634411215782,
0.1080654114484787,
0.021870044991374016,
-0.07784450799226761,
-0.11714762449264526,
0.001524197170510888,
-0.08141826093196869,
0.01126081496477127,
0.1802481710910797,
0.04831566661596298,
-0.022556783631443977,
0.03305482119321823,
-0.11502696573734283,
-0.21776944398880005,
-0.01617501676082611,
-0.11659006774425507,
-0.05422298610210419,
0.07242408394813538,
-0.22702434659004211,
0.00973254069685936,
0.03509848192334175,
-0.021992651745676994,
0.05734172463417053,
0.0652010440826416,
-0.10430524498224258,
0.01023867167532444,
0.1239415779709816,
-0.10563014447689056,
-0.1211216077208519,
0.024934910237789154,
0.025935843586921692,
0.17365825176239014,
0.050733763724565506,
0.07385742664337158,
0.06699050962924957,
0.004279726184904575,
0.016696322709321976,
0.07876995950937271,
-0.16484880447387695,
-0.029713543131947517,
0.07098285108804703,
-0.040593817830085754,
-0.1394682228565216,
0.19541186094284058,
0.028642958030104637,
0.029897678643465042,
-0.024543415755033493,
0.035570524632930756,
-0.05335932597517967,
-0.09447703510522842,
-0.13994671404361725,
-0.020488658919930458,
-0.11634241044521332,
-0.14416252076625824,
0.04803739860653877,
-0.08919020742177963,
-0.02546677179634571,
-0.03999754413962364,
0.07524143159389496,
0.13654081523418427,
0.058341559022665024,
0.013631465844810009,
0.16242444515228271,
-0.09581738710403442,
-0.18957501649856567,
-0.020761094987392426,
-0.06745804101228714,
-0.20285774767398834,
0.02592097781598568,
0.08869154006242752,
-0.020691489800810814,
-0.03635775297880173,
-0.09561870247125626,
0.04295523837208748,
-0.1443842053413391,
-0.008468776941299438,
-0.06995626538991928,
0.011855300515890121,
0.06764024496078491,
-0.09456440061330795,
0.010382445529103279,
0.04733350872993469,
-0.14076519012451172,
-0.057332105934619904,
0.0022140543442219496,
0.03670362010598183,
-0.10423135757446289,
-0.05571652576327324,
0.12314850091934204,
0.0703822523355484,
0.14754506945610046,
0.1352262943983078,
0.05553649738430977,
0.14190629124641418,
-0.22292354702949524,
-0.05461675673723221,
0.05658799037337303,
-0.004033025354146957,
-0.0409574992954731,
0.0034502120688557625,
-0.03508859872817993,
0.03941653296351433,
-0.07560205459594727,
0.042029887437820435,
-0.01114447321742773,
-0.131355419754982,
-0.16204427182674408,
-0.020497437566518784,
-0.1109241396188736,
0.033404845744371414,
-0.17885321378707886,
0.17627950012683868,
0.07166852056980133,
0.051701869815588,
0.07173995673656464,
-0.017255190759897232,
0.0028538221959024668,
0.030197259038686752,
-0.0401480607688427,
-0.08765622228384018,
-0.13222530484199524,
-0.008726547472178936,
-0.12079193443059921,
-0.03162060305476189,
0.33216890692710876,
-0.030980104580521584,
-0.1621956080198288,
0.027868445962667465,
0.07404996454715729,
0.12636078894138336,
0.0026254802942276,
0.2259112149477005,
0.054794542491436005,
0.014137940481305122,
-0.13172785937786102,
0.009069803170859814,
0.005791394039988518,
-0.27599358558654785,
0.0624869205057621,
0.0013532752636820078,
0.08246773481369019,
0.020798001438379288,
0.040949176996946335,
-0.08768724650144577,
-0.01853700540959835,
-0.03068496286869049,
0.07054902613162994,
0.011548550799489021,
0.041395153850317,
0.05476638674736023,
0.1752733737230301,
-0.024234607815742493,
0.014738751575350761,
-0.019994881004095078,
0.01285971887409687,
-0.1276385486125946,
-0.13978463411331177,
0.014189327135682106,
-0.15848460793495178,
0.045844435691833496,
-0.010509041137993336,
0.039326947182416916,
0.20746053755283356,
0.04208739846944809,
-0.06142117455601692,
-0.0455465130507946,
-0.08228625357151031,
-0.043509144335985184,
-0.009183176793158054,
-0.02066180668771267,
-0.06648684293031693,
-0.04331827908754349,
-0.09308581799268723,
-0.015449702739715576,
-0.06880850344896317,
-0.03242972493171692,
0.01768173836171627,
0.029128799214959145,
0.01990487426519394,
-0.08321307599544525,
0.002178385853767395,
-0.09582957625389099,
0.05271981284022331,
0.004241388291120529,
0.16314728558063507,
0.026698041707277298,
0.024437984451651573,
0.13244377076625824,
0.0729329064488411,
-0.013862117193639278,
-0.10025795549154282,
-0.05922839790582657,
0.03516887500882149,
-0.07846292853355408,
0.0765465646982193,
-0.0554979145526886,
0.0003111901751253754,
-0.007978007197380066,
0.23187364637851715,
0.18745596706867218,
-0.1058710440993309,
0.010778109543025494,
-0.024767739698290825,
0.011735105887055397,
0.06795872747898102,
0.13705003261566162,
0.05810157582163811,
0.1847129464149475,
-0.07218676060438156,
-0.043915461748838425,
0.00902556348592043,
0.017504174262285233,
-0.18057847023010254,
0.0833623856306076,
-0.00872169528156519,
-0.07224971055984497,
-0.03864730894565582,
0.10137778520584106,
-0.0695849135518074,
0.07668505609035492,
0.06028321757912636,
-0.028682032600045204,
0.027593085542321205,
-0.005067290738224983,
0.08621590584516525,
0.03740588203072548,
0.053074054419994354,
-0.1063728779554367,
-0.07684344053268433,
0.008703646250069141,
0.021847110241651535,
-0.3201918601989746,
-0.17543719708919525,
0.0938958078622818,
0.07460742443799973,
0.2910914719104767,
0.00119002943392843,
0.04769929125905037,
0.03156432509422302,
0.0401572659611702,
-0.12241633981466293,
0.12800614535808563,
0.012672079727053642,
-0.04014110192656517,
-0.1387740820646286,
-0.1924290657043457,
-0.032277900725603104,
-0.11549804359674454,
0.052928097546100616,
0.100664421916008,
0.012590800411999226,
0.21247832477092743,
-0.049709733575582504,
-0.022715643048286438,
-0.0057846843264997005,
-0.1684456169605255,
0.04534858837723732,
-0.063796766102314,
-0.03746533393859863,
-0.07755741477012634,
-0.06920577585697174,
0.006900185719132423,
0.07946354150772095,
-0.17020881175994873,
-0.052895352244377136,
0.2066379338502884,
0.04911750182509422,
0.1324104219675064,
0.011248554103076458,
0.05078674107789993,
-0.029439883306622505,
-0.07551411539316177,
0.09735040366649628,
-0.045069485902786255,
0.01129083801060915,
0.13425007462501526,
-0.018308015540242195,
0.02112884260714054,
-0.166777104139328,
0.02545231580734253,
-0.06915885210037231,
-0.0002544210001360625,
-0.06249799206852913
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# output
This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 64
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 120
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"tags": ["generated_from_trainer"], "model-index": [{"name": "output", "results": []}]} | null | io-roboto/decision-transformer | [
"transformers",
"tensorboard",
"safetensors",
"decision_transformer",
"generated_from_trainer",
"endpoints_compatible",
"region:us"
] | 2024-02-07T17:13:22+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #decision_transformer #generated_from_trainer #endpoints_compatible #region-us
|
# output
This model is a fine-tuned version of [](URL on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 64
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 120
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| [
"# output\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 64\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_ratio: 0.1\n- num_epochs: 120",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #decision_transformer #generated_from_trainer #endpoints_compatible #region-us \n",
"# output\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 64\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_ratio: 0.1\n- num_epochs: 120",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
40,
24,
6,
12,
8,
3,
104,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #decision_transformer #generated_from_trainer #endpoints_compatible #region-us \n# output\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 64\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_ratio: 0.1\n- num_epochs: 120### Training results### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.12064183503389359,
0.16336308419704437,
-0.001561130047775805,
0.08581382036209106,
0.15199384093284607,
0.018543168902397156,
0.11020784080028534,
0.09220647811889648,
-0.07194197922945023,
0.0997733548283577,
0.12699206173419952,
0.06754303723573685,
0.03785039111971855,
0.1269303411245346,
-0.039014533162117004,
-0.20895855128765106,
0.033722348511219025,
0.0161232128739357,
-0.05290130525827408,
0.10112975537776947,
0.08497485518455505,
-0.11477619409561157,
0.08488016575574875,
0.022822342813014984,
-0.1743943989276886,
-0.013540658168494701,
-0.02418050915002823,
-0.053646642714738846,
0.09334293752908707,
-0.0012894067913293839,
0.1277710348367691,
0.028329886496067047,
0.09717249125242233,
-0.18164756894111633,
0.007918886840343475,
0.0630919337272644,
0.016217024996876717,
0.10135666280984879,
0.044703394174575806,
-0.01725964993238449,
0.06182243302464485,
-0.1378507763147354,
0.06755281984806061,
0.01053515262901783,
-0.06831899285316467,
-0.1292576640844345,
-0.07948745787143707,
0.09430209547281265,
0.045570265501737595,
0.08554834872484207,
0.012419873848557472,
0.15399691462516785,
-0.08179890364408493,
0.06337037682533264,
0.20633210241794586,
-0.2863282859325409,
-0.03945416212081909,
0.013168955221772194,
0.032900430262088776,
0.06746326386928558,
-0.10925142467021942,
-0.032097309827804565,
0.0631929561495781,
0.02647322230041027,
0.0864517018198967,
0.01394999586045742,
-0.03994206711649895,
-0.016305692493915558,
-0.14441567659378052,
-0.03944665566086769,
0.1552119106054306,
0.05126706510782242,
-0.05777907371520996,
-0.07068060338497162,
-0.07799991220235825,
-0.1391940414905548,
-0.00727861700579524,
-0.02102447673678398,
0.015532554127275944,
-0.046469513326883316,
-0.08617677539587021,
-0.03728364035487175,
-0.10165620595216751,
-0.057551734149456024,
-0.024055911228060722,
0.17188075184822083,
0.05637001618742943,
0.038442909717559814,
-0.03645247593522072,
0.09621935337781906,
0.02792147919535637,
-0.11740662157535553,
-0.0017730359686538577,
-0.00726303132250905,
-0.030083440244197845,
-0.040878843516111374,
-0.05901319906115532,
-0.04338079318404198,
0.006667556241154671,
0.11711246520280838,
-0.06376184523105621,
0.04600534960627556,
0.04385007545351982,
0.029986895620822906,
-0.021778089925646782,
0.1375810205936432,
-0.07877253741025925,
0.0263202004134655,
0.029544023796916008,
0.10905422270298004,
0.058458052575588226,
-0.020230455324053764,
-0.10004136711359024,
-0.03463546186685562,
0.1264721155166626,
0.046167608350515366,
-0.05062570050358772,
0.04233524575829506,
-0.03749828040599823,
-0.026078565046191216,
0.05822031944990158,
-0.11020472645759583,
0.018601177260279655,
0.018490908667445183,
-0.07690596580505371,
-0.01738041080534458,
0.03632637485861778,
-0.012241080403327942,
-0.03231015428900719,
0.06688965857028961,
-0.12375061959028244,
-0.006321969907730818,
-0.06719031184911728,
-0.0745505839586258,
0.02913103625178337,
-0.10067398101091385,
0.004247085656970739,
-0.0935097485780716,
-0.1356326788663864,
-0.005733141675591469,
0.02730264514684677,
-0.03229675441980362,
-0.039578910917043686,
-0.038919221609830856,
-0.07245262712240219,
0.018479345366358757,
-0.0037461358588188887,
0.04657163843512535,
-0.04528189077973366,
0.09887462854385376,
0.021337777376174927,
0.045854635536670685,
-0.023768430575728416,
0.03096538968384266,
-0.10315605998039246,
0.04310756176710129,
-0.10921753942966461,
0.04907841235399246,
-0.06917537748813629,
0.0269148051738739,
-0.0896952897310257,
-0.09233228117227554,
-0.004877414554357529,
-0.03055913746356964,
0.09026900678873062,
0.12575894594192505,
-0.131570503115654,
-0.024854060262441635,
0.17140236496925354,
-0.09636989235877991,
-0.10848046839237213,
0.12784211337566376,
-0.03500847890973091,
-0.013414403423666954,
0.046101365238428116,
0.15419799089431763,
0.09133701771497726,
-0.12911590933799744,
-0.055456001311540604,
0.04140227660536766,
0.026349341496825218,
-0.08159229159355164,
0.07942692935466766,
0.009555350057780743,
-0.004011497832834721,
0.016516188159585,
-0.04075632616877556,
0.019934840500354767,
-0.09808362275362015,
-0.08084982633590698,
-0.06688471883535385,
-0.10124962031841278,
0.05849320441484451,
0.024293728172779083,
0.04166361317038536,
-0.07132300734519958,
-0.10873866081237793,
0.08542947471141815,
0.12173324078321457,
-0.07636170089244843,
0.01560372207313776,
-0.07747116684913635,
0.09523430466651917,
-0.10931096225976944,
-0.03754090517759323,
-0.18494530022144318,
-0.1258179247379303,
0.02578926645219326,
-0.04219690337777138,
0.015228926204144955,
-0.012890338897705078,
0.060754455626010895,
0.08880572766065598,
-0.019268548116087914,
-0.030553793534636497,
-0.06942852586507797,
-0.005458181723952293,
-0.12501570582389832,
-0.1575583517551422,
-0.05926143005490303,
-0.026968402788043022,
0.1568356156349182,
-0.22526778280735016,
0.011438241228461266,
-0.010806086473166943,
0.14098553359508514,
0.02401517890393734,
-0.045704782009124756,
0.001706932787783444,
0.0388580821454525,
-0.030163517221808434,
-0.08054584264755249,
0.0447760671377182,
0.024783482775092125,
-0.10533599555492401,
-0.032082945108413696,
-0.14880581200122833,
0.10491877049207687,
0.11292921751737595,
0.056867510080337524,
-0.06903376430273056,
-0.016283348202705383,
-0.07856898754835129,
-0.035672400146722794,
-0.056395191699266434,
0.009733769111335278,
0.13097095489501953,
0.011255154386162758,
0.13530468940734863,
-0.08257733285427094,
-0.04198714345693588,
0.0328369103372097,
-0.027737166732549667,
-0.04153703525662422,
0.08601866662502289,
0.08578880876302719,
-0.11668013036251068,
0.11309726536273956,
0.09026650339365005,
-0.08643104135990143,
0.10954505205154419,
-0.0487799234688282,
-0.10696890950202942,
-0.016115885227918625,
0.03142153471708298,
0.0015281859086826444,
0.13037817180156708,
-0.07524985820055008,
0.014546791091561317,
0.036199115216732025,
0.011954162269830704,
0.05204291269183159,
-0.1988159716129303,
-0.008397010155022144,
0.023358523845672607,
-0.028313249349594116,
-0.05480680987238884,
-0.01824703998863697,
0.017757104709744453,
0.08404842019081116,
0.023690596222877502,
-0.020812708884477615,
0.05483622848987579,
0.012086266651749611,
-0.07506263256072998,
0.20167778432369232,
-0.10298359394073486,
-0.15686511993408203,
-0.1563403308391571,
0.06354477256536484,
-0.06030108407139778,
-0.010331600904464722,
0.019501781091094017,
-0.06909119337797165,
-0.045531194657087326,
-0.09064449369907379,
-0.02982330694794655,
-0.04242556914687157,
0.02626579999923706,
0.06085911765694618,
0.005088335834443569,
0.07714565098285675,
-0.12056435644626617,
0.010687524452805519,
-0.045558661222457886,
-0.048524752259254456,
0.010153517127037048,
0.04519825428724289,
0.11623725295066833,
0.12023844569921494,
-0.044621530920267105,
0.017313331365585327,
-0.01209387555718422,
0.22711673378944397,
-0.04634816199541092,
-0.037654146552085876,
0.10946573317050934,
-0.018769653514027596,
0.07957769930362701,
0.09426061809062958,
0.03305387124419212,
-0.08282032608985901,
0.028185265138745308,
0.03937921300530434,
-0.027082866057753563,
-0.22490805387496948,
-0.04648453742265701,
-0.025648625567555428,
-0.03960046172142029,
0.08898571878671646,
0.04318378493189812,
-0.030987972393631935,
0.07924620062112808,
-0.011038721539080143,
0.04955218732357025,
-0.07967003434896469,
0.07592524588108063,
0.08074000477790833,
0.04781056568026543,
0.08574837446212769,
-0.04902808368206024,
-0.047567591071128845,
0.05347727611660957,
0.015370525419712067,
0.24318724870681763,
-0.049184177070856094,
0.16317902505397797,
0.02606845274567604,
0.19662940502166748,
-0.00039549460052512586,
0.07764894515275955,
-0.009482351131737232,
-0.008826720528304577,
0.010168898850679398,
-0.05271603539586067,
-0.0694541186094284,
0.012945631518959999,
-0.007237124256789684,
0.061562102288007736,
-0.11688786000013351,
0.08414194732904434,
0.015536966733634472,
0.22782497107982635,
0.04010510817170143,
-0.3581142723560333,
-0.10878162086009979,
-0.011744603514671326,
0.003178167389705777,
-0.05961792171001434,
0.01431175135076046,
0.15945394337177277,
-0.08169929683208466,
0.012693673372268677,
-0.06366143375635147,
0.07391924411058426,
-0.0588807575404644,
-0.008520062081515789,
0.014358299784362316,
0.14170749485492706,
-0.011228276416659355,
0.09693846106529236,
-0.16571125388145447,
0.2104417383670807,
0.0236362311989069,
0.09623392671346664,
-0.07421063631772995,
-0.004308064468204975,
-0.0046778479591012,
0.07987271994352341,
0.09768643975257874,
0.00007235953671624884,
-0.03333644941449165,
-0.18408317863941193,
-0.1354653686285019,
0.03896782919764519,
0.12127291411161423,
-0.0037237294018268585,
0.0997532308101654,
-0.03918235003948212,
0.014704869128763676,
0.05089527741074562,
-0.022809740155935287,
-0.114021435379982,
-0.13564735651016235,
0.010004815645515919,
0.046169113367795944,
-0.04434634745121002,
-0.07350989431142807,
-0.12340492755174637,
-0.011888372711837292,
0.1637178212404251,
-0.01630518212914467,
-0.057630572468042374,
-0.13286244869232178,
0.050757620483636856,
0.1272892951965332,
-0.09322446584701538,
0.008029387332499027,
-0.01266863290220499,
0.14958737790584564,
0.01581810973584652,
-0.10259240120649338,
0.06655313074588776,
-0.07205035537481308,
-0.1598248928785324,
-0.03983813896775246,
0.1351107954978943,
0.0467388853430748,
0.030273038893938065,
-0.007629455998539925,
-0.0020993498619645834,
-0.013944297097623348,
-0.09165722876787186,
0.027717959135770798,
0.052308015525341034,
0.0865619033575058,
0.037075188010931015,
-0.05977454409003258,
0.02776065282523632,
-0.015197197906672955,
0.006301636807620525,
0.15457484126091003,
0.19501590728759766,
-0.08023852109909058,
0.048005905002355576,
0.12395947426557541,
-0.04149499908089638,
-0.1973014622926712,
0.03132010996341705,
0.09198274463415146,
0.029679596424102783,
0.010016387328505516,
-0.1523994356393814,
0.13169406354427338,
0.10601571202278137,
-0.021587366238236427,
0.099202461540699,
-0.2896784842014313,
-0.11232217401266098,
0.09526222944259644,
0.1406431943178177,
0.12241464108228683,
-0.13049639761447906,
-0.04501917585730553,
-0.021254030987620354,
-0.15529634058475494,
0.16009697318077087,
-0.09328089654445648,
0.11704777926206589,
-0.0127867441624403,
0.10157297551631927,
0.0262129008769989,
-0.06533206254243851,
0.1261683851480484,
0.0210700836032629,
0.06599826365709305,
-0.04858291521668434,
-0.01356343924999237,
0.08360589295625687,
-0.06801435351371765,
0.07646545767784119,
-0.05540112778544426,
0.08067360520362854,
-0.0913744643330574,
-0.02624225802719593,
-0.06978084146976471,
0.06455454975366592,
-0.035514675080776215,
-0.048944324254989624,
-0.038812797516584396,
0.026631062850356102,
0.05644723400473595,
-0.019458016380667686,
0.1142960786819458,
0.07231162488460541,
0.07353170216083527,
0.10629347711801529,
0.07872290164232254,
-0.05263518542051315,
-0.08216548711061478,
-0.014373097568750381,
-0.0327855683863163,
0.07765544205904007,
-0.09395872056484222,
0.01881909742951393,
0.11464467644691467,
0.04271312057971954,
0.10325130820274353,
0.056296173483133316,
-0.06627672910690308,
0.005054390989243984,
0.048509761691093445,
-0.14667899906635284,
-0.1528060883283615,
-0.008107411675155163,
-0.02551410347223282,
-0.15091153979301453,
0.07680075615644455,
0.09560327976942062,
-0.07228776067495346,
-0.010516751557588577,
-0.03985428810119629,
0.010186904110014439,
-0.00020777180907316506,
0.1476297229528427,
0.08044823259115219,
0.0611327700316906,
-0.08539436012506485,
0.11880826205015182,
0.057572949677705765,
-0.1201091930270195,
0.056697774678468704,
0.02598349191248417,
-0.10794232040643692,
-0.012201094068586826,
0.05552913248538971,
0.126614511013031,
-0.06229954957962036,
-0.059912439435720444,
-0.11016786098480225,
-0.07634720951318741,
0.044285502284765244,
0.1359604001045227,
0.06485796719789505,
0.025828342884778976,
-0.022838879376649857,
0.024542544037103653,
-0.1807824671268463,
0.09599553793668747,
0.03216048330068588,
0.07617995887994766,
-0.16891703009605408,
0.11493846029043198,
-0.005415098275989294,
0.06260623037815094,
-0.017530808225274086,
0.023975083604454994,
-0.0987519845366478,
-0.02937024086713791,
-0.12061020731925964,
0.020347924903035164,
-0.050700657069683075,
0.003579816548153758,
-0.007255826611071825,
-0.05141409859061241,
-0.05185467749834061,
0.055522020906209946,
-0.06632240861654282,
-0.04410659894347191,
0.024381844326853752,
0.020843539386987686,
-0.1551266461610794,
-0.0007033621659502387,
0.02229125052690506,
-0.08633941411972046,
0.08593976497650146,
0.0803389698266983,
0.02294323593378067,
0.03032870963215828,
-0.1016039028763771,
-0.01102445088326931,
0.044860899448394775,
0.01012646034359932,
0.0739021971821785,
-0.06639959663152695,
-0.005011354107409716,
-0.003510534530505538,
0.02694639377295971,
0.001859375392086804,
0.09703446179628372,
-0.13023130595684052,
-0.04105639085173607,
-0.042852733284235,
-0.01535850577056408,
-0.08314608037471771,
0.0293822530657053,
0.08091322332620621,
0.028027368709445,
0.17658910155296326,
-0.06952610611915588,
0.02392192929983139,
-0.16063517332077026,
-0.024779848754405975,
-0.0013174135237932205,
-0.03733564168214798,
-0.06557289510965347,
0.00197206879965961,
0.06991427391767502,
-0.0571984127163887,
0.11471624672412872,
-0.043756864964962006,
0.04460199177265167,
0.018095429986715317,
0.0017980511765927076,
-0.01840616576373577,
0.004349546041339636,
0.2133983075618744,
0.04279649630188942,
-0.031118644401431084,
0.06340747326612473,
-0.00948619470000267,
0.08978378772735596,
0.02224445715546608,
0.19506581127643585,
0.11617358773946762,
-0.031268179416656494,
0.0940970629453659,
0.06504825502634048,
-0.07029459625482559,
-0.17299672961235046,
0.045065995305776596,
-0.05149642750620842,
0.08094330877065659,
-0.006809646729379892,
0.18625465035438538,
0.1190677285194397,
-0.11864253133535385,
0.008371364325284958,
-0.03195847198367119,
-0.1066591888666153,
-0.10280304402112961,
-0.07990918308496475,
-0.07940053939819336,
-0.10935476422309875,
0.024365274235606194,
-0.1220007836818695,
-0.006404742132872343,
0.10581959038972855,
0.0031728590838611126,
-0.006294839084148407,
0.18255764245986938,
0.009375615045428276,
0.01848417893052101,
0.027901839464902878,
0.021084235981106758,
-0.01666032150387764,
-0.011691175401210785,
-0.0814109593629837,
0.05728119984269142,
-0.006446356885135174,
0.08988147974014282,
-0.04339303448796272,
0.01744747720658779,
0.06785473227500916,
-0.0031893036793917418,
-0.07673856616020203,
0.006338100414723158,
0.020983576774597168,
0.04771256074309349,
0.009966705925762653,
0.029260581359267235,
0.001416503102518618,
-0.055208656936883926,
0.23466363549232483,
-0.05761164054274559,
-0.03244073688983917,
-0.11412733048200607,
0.18383467197418213,
0.03842847794294357,
-0.014947683550417423,
0.06944458931684494,
-0.13682980835437775,
-0.012314768508076668,
0.18340767920017242,
0.13122156262397766,
-0.06622841954231262,
-0.01485550869256258,
-0.018432604148983955,
-0.01060069352388382,
-0.0485796220600605,
0.11455205827951431,
0.10307303071022034,
0.07838927209377289,
-0.06224300339818001,
-0.053787633776664734,
-0.04533331096172333,
-0.00907314382493496,
-0.08604465425014496,
0.06086304038763046,
-0.0035964602138847113,
0.011785018257796764,
-0.07135683298110962,
0.03150082752108574,
-0.033214543014764786,
-0.1672530621290207,
0.027041884139180183,
-0.14586205780506134,
-0.15359680354595184,
-0.02953551523387432,
0.09591896831989288,
-0.022344376891851425,
0.06515657901763916,
-0.0140836201608181,
-0.008284272626042366,
0.11070149391889572,
-0.0037022375036031008,
-0.09882794320583344,
-0.03569570183753967,
0.04336448386311531,
-0.021533846855163574,
0.2249302864074707,
0.00975907314568758,
0.07722290605306625,
0.10916634649038315,
0.04959523305296898,
-0.14501963555812836,
0.04334338381886482,
0.05763192102313042,
-0.041293077170848846,
0.024994701147079468,
0.127200648188591,
-0.04235284775495529,
0.05675286427140236,
0.025229116901755333,
-0.14043892920017242,
-0.03639538958668709,
-0.04685539752244949,
-0.0038146572187542915,
-0.07001040130853653,
-0.029154114425182343,
-0.05880109220743179,
0.15198767185211182,
0.17687204480171204,
-0.04943940415978432,
-0.014285636134445667,
-0.06058914214372635,
0.02698039449751377,
0.0604596771299839,
0.04084368422627449,
-0.005687198601663113,
-0.2139522135257721,
0.018548406660556793,
0.017919205129146576,
0.018249312415719032,
-0.26750731468200684,
-0.06365074217319489,
0.030213311314582825,
-0.05181414261460304,
-0.06983125954866409,
0.06724666804075241,
0.0706612765789032,
0.052529819309711456,
-0.05235208570957184,
-0.00503175426274538,
-0.0650244951248169,
0.1256408542394638,
-0.15157383680343628,
-0.08451450616121292
] |
null | null | fastai |
# Amazing!
🥳 Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the [documentation here](https://huggingface.co/docs/hub/model-repos))!
2. Create a demo in Gradio or Streamlit using 🤗 Spaces ([documentation here](https://huggingface.co/docs/hub/spaces)).
3. Join the fastai community on the [Fastai Discord](https://discord.com/invite/YKrxeNn)!
Greetings fellow fastlearner 🤝! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| {"tags": ["fastai"]} | null | jegilj/intel-image-classification | [
"fastai",
"has_space",
"region:us"
] | 2024-02-07T17:13:42+00:00 | [] | [] | TAGS
#fastai #has_space #region-us
|
# Amazing!
Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the documentation here)!
2. Create a demo in Gradio or Streamlit using Spaces (documentation here).
3. Join the fastai community on the Fastai Discord!
Greetings fellow fastlearner ! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| [
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
"TAGS\n#fastai #has_space #region-us \n",
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
13,
20,
79,
3,
6,
12,
8
] | [
"passage: TAGS\n#fastai #has_space #region-us \n# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---# Model card## Model description\nMore information needed## Intended uses & limitations\nMore information needed## Training and evaluation data\nMore information needed"
] | [
-0.048121724277734756,
-0.024616125971078873,
0.002038001548498869,
0.10439170897006989,
0.135872021317482,
0.11887997388839722,
0.07405775785446167,
0.09980081021785736,
0.07783667743206024,
0.02590852417051792,
0.08961158245801926,
-0.08088712394237518,
0.08744348585605621,
0.271692156791687,
0.06988707184791565,
-0.22761479020118713,
0.04051019623875618,
-0.00024903909070417285,
0.08053462207317352,
0.06629016250371933,
0.13507555425167084,
-0.05464952811598778,
0.14010503888130188,
-0.004088983871042728,
-0.19050447642803192,
-0.042929794639348984,
-0.01773718371987343,
-0.02527874894440174,
0.12317648530006409,
-0.04744937643408775,
0.05381017178297043,
0.015037551522254944,
0.007565062493085861,
-0.07253646105527878,
0.0623294934630394,
0.040457066148519516,
0.01740180514752865,
0.059235580265522,
-0.07249044626951218,
0.08950132131576538,
0.08404164761304855,
-0.024370938539505005,
-0.1097978875041008,
0.07827875018119812,
-0.14424212276935577,
-0.21762843430042267,
-0.1253085881471634,
-0.09017651528120041,
0.028519365936517715,
0.004388005938380957,
-0.025051530450582504,
0.12801909446716309,
-0.13558274507522583,
-0.040698226541280746,
0.20124278962612152,
-0.17012301087379456,
-0.05505548417568207,
0.034343402832746506,
0.09226689487695694,
-0.05829555168747902,
-0.06347129493951797,
0.10614984482526779,
0.09640881419181824,
-0.019833475351333618,
0.05516824126243591,
0.002579754451289773,
0.021173657849431038,
0.01370104867964983,
-0.06150497496128082,
0.04717832803726196,
-0.010183089412748814,
0.048132527619600296,
-0.09465572983026505,
-0.1303568333387375,
-0.004072192590683699,
0.01214400865137577,
-0.048744890838861465,
-0.07019646465778351,
0.07833103090524673,
-0.011118141002953053,
-0.04357248544692993,
-0.13031910359859467,
-0.09131011366844177,
-0.12358787655830383,
0.008646543137729168,
0.09500427544116974,
0.003679296001791954,
0.07374339550733566,
-0.08258994668722153,
0.06774985045194626,
-0.17329485714435577,
-0.06484591960906982,
-0.08138520270586014,
-0.11546400189399719,
0.021133482456207275,
-0.0387684591114521,
0.02668963186442852,
0.15394504368305206,
0.12983950972557068,
0.023976242169737816,
0.04388163983821869,
-0.038937073200941086,
0.051190316677093506,
0.058571770787239075,
0.03395717963576317,
0.034934818744659424,
-0.036981891840696335,
-0.1793210655450821,
-0.016702448949217796,
-0.011550825089216232,
0.07954040914773941,
-0.07523109763860703,
-0.05632320046424866,
0.013454885222017765,
-0.11071494966745377,
0.07202339172363281,
-0.03576776012778282,
-0.0032025426626205444,
0.01168301422148943,
0.018371861428022385,
0.21271461248397827,
0.03955606371164322,
0.014191740192472935,
-0.008875265717506409,
-0.13453757762908936,
-0.06874168664216995,
-0.06896194815635681,
0.03361047804355621,
0.04448792710900307,
-0.0028071461711078882,
-0.07672245055437088,
0.04325154796242714,
-0.06045534089207649,
-0.03508453071117401,
0.008032378740608692,
-0.18221288919448853,
0.007458044681698084,
-0.10049355030059814,
-0.12126200646162033,
0.05306628718972206,
0.01695440337061882,
-0.08215925842523575,
0.08141279965639114,
0.02662261202931404,
0.020931517705321312,
-0.009988143108785152,
-0.005391082260757685,
0.06874798238277435,
-0.08508864045143127,
0.029901226982474327,
0.17170792818069458,
0.13024519383907318,
-0.08046911656856537,
-0.0006887061172164977,
-0.10965746641159058,
0.04426072910428047,
-0.13325683772563934,
0.02251482754945755,
-0.09062390774488449,
0.11723794043064117,
-0.042396437376737595,
0.002038756385445595,
-0.029030200093984604,
0.0960269495844841,
0.08189879357814789,
0.16663365066051483,
-0.2419009804725647,
-0.031095001846551895,
0.13240347802639008,
-0.10711425542831421,
-0.1807439625263214,
0.18486657738685608,
-0.012035200372338295,
0.11329247802495956,
-0.047014184296131134,
0.18334640562534332,
-0.02612062357366085,
-0.13582459092140198,
-0.058872904628515244,
0.005852419883012772,
-0.2269321084022522,
-0.06286033242940903,
0.09738040715456009,
0.13425657153129578,
-0.042984943836927414,
0.007112155202776194,
0.026316028088331223,
0.13609857857227325,
-0.06715573370456696,
-0.05195777863264084,
-0.012255736626684666,
-0.10902371257543564,
0.041914235800504684,
0.018215661868453026,
0.035408079624176025,
-0.059880174696445465,
-0.02931194379925728,
-0.053190283477306366,
0.13146710395812988,
0.09760832786560059,
-0.03670211136341095,
-0.049620725214481354,
0.1689043790102005,
-0.07763876020908356,
-0.033587727695703506,
0.07560533285140991,
-0.08268500119447708,
0.03266897425055504,
0.03090597130358219,
0.055881720036268234,
0.07766123116016388,
0.08522116392850876,
0.06057543307542801,
0.00819048099219799,
0.034654274582862854,
0.12095347046852112,
-0.013591280207037926,
-0.05039411783218384,
0.021508218720555305,
0.016904234886169434,
-0.019032588228583336,
0.29030677676200867,
-0.1951042115688324,
0.024724548682570457,
-0.06477324664592743,
0.07631538063287735,
0.06136792525649071,
0.003575638635084033,
0.08580143749713898,
-0.06023019179701805,
-0.019061198458075523,
-0.04803973436355591,
0.046805646270513535,
-0.0666879191994667,
-0.04162997007369995,
0.2621194124221802,
-0.05497581139206886,
0.044914912432432175,
0.12313763797283173,
-0.05873025581240654,
-0.07091446220874786,
0.01009807363152504,
-0.00793424155563116,
0.03249288722872734,
-0.04042816907167435,
0.043721720576286316,
-0.10840129852294922,
-0.06674089282751083,
0.1573198139667511,
-0.038477856665849686,
0.06786153465509415,
0.032288823276758194,
-0.04958454892039299,
-0.0648743286728859,
0.04650486260652542,
0.13598160445690155,
-0.0875244215130806,
0.07435166835784912,
0.17612984776496887,
-0.010562662966549397,
0.168031245470047,
0.08435525000095367,
-0.07075224816799164,
-0.09465329349040985,
-0.051014289259910583,
-0.021595727652311325,
0.21222901344299316,
-0.07084725052118301,
-0.054564714431762695,
0.05911700800061226,
-0.013703816570341587,
0.07196151465177536,
-0.06009222939610481,
-0.08332337439060211,
0.03227344527840614,
-0.04517695680260658,
0.011517706327140331,
0.13512636721134186,
-0.07090822607278824,
0.04681389778852463,
0.031489867717027664,
-0.0662703812122345,
0.02217509225010872,
0.033389873802661896,
0.0068921963684260845,
0.033959709107875824,
0.07332495599985123,
-0.20893315970897675,
-0.08408680558204651,
-0.13727638125419617,
0.037881869822740555,
0.021770721301436424,
0.045787326991558075,
-0.08602345734834671,
0.02231026627123356,
-0.08954031765460968,
-0.07987114042043686,
0.029592275619506836,
-0.026350297033786774,
-0.11349643021821976,
-0.03396226093173027,
-0.009560913778841496,
-0.06662604957818985,
-0.02250705659389496,
-0.05024505779147148,
0.03983384370803833,
0.04479299485683441,
0.058377087116241455,
0.12796473503112793,
-0.013808943331241608,
-0.03839317709207535,
0.000370211957488209,
-0.022712308913469315,
0.16396735608577728,
-0.14746315777301788,
0.07954913377761841,
0.19160102307796478,
0.11742953956127167,
0.028144672513008118,
0.028885571286082268,
0.03537585213780403,
-0.06289814412593842,
-0.000050317394197918475,
0.03226194158196449,
-0.09392514824867249,
-0.05801016092300415,
-0.020014392212033272,
-0.04031052812933922,
0.17134574055671692,
-0.12160717695951462,
0.03345204517245293,
0.04098419472575188,
0.09783966839313507,
0.10073629021644592,
-0.028829937800765038,
-0.1815856397151947,
0.038818612694740295,
-0.24060091376304626,
-0.05831146240234375,
0.027899866923689842,
-0.09110201895236969,
-0.06232144311070442,
0.17409387230873108,
0.013794700615108013,
0.011769929900765419,
-0.006736889015883207,
0.07983319461345673,
0.0110100656747818,
0.1217205822467804,
0.05947643890976906,
-0.05539114400744438,
0.025202350690960884,
-0.09962950646877289,
-0.07107596844434738,
-0.04035590961575508,
-0.05832801014184952,
0.07548832893371582,
0.1409129947423935,
-0.025475580245256424,
-0.020795362070202827,
0.023489827290177345,
0.08550169318914413,
0.0423230417072773,
0.16739299893379211,
-0.16016584634780884,
-0.026555389165878296,
0.04571257904171944,
-0.03384667634963989,
-0.05433850735425949,
-0.010291114449501038,
0.1137225553393364,
-0.02820689231157303,
-0.040318265557289124,
0.021242983639240265,
0.06503437459468842,
0.01481706090271473,
0.05012747645378113,
-0.04056356102228165,
0.14796851575374603,
-0.03461192920804024,
0.019330544397234917,
-0.12413888424634933,
0.13848772644996643,
0.021095896139740944,
-0.03901609033346176,
-0.06735876202583313,
-0.05808034539222717,
0.18150931596755981,
0.0025602965615689754,
0.10535930097103119,
0.012098877690732479,
-0.12160047143697739,
-0.1359938681125641,
-0.11211287975311279,
0.005111907608807087,
0.08330471813678741,
-0.023147236555814743,
-0.022247863933444023,
0.022165266796946526,
-0.036149751394987106,
-0.0530381016433239,
0.15749511122703552,
-0.1289154291152954,
-0.001082550617866218,
0.014728817157447338,
0.06971760839223862,
-0.08223173767328262,
0.026267826557159424,
0.014071501791477203,
-0.1119147390127182,
0.10590848326683044,
0.2521335482597351,
0.10338116437196732,
-0.09591643512248993,
-0.07697287201881409,
0.03418830782175064,
-0.012184361927211285,
-0.000774814048781991,
-0.006932659074664116,
0.0495428591966629,
-0.005566445179283619,
0.006762749515473843,
0.12971895933151245,
-0.07130889594554901,
0.011540771462023258,
-0.08449850976467133,
0.05566910281777382,
-0.05276734381914139,
0.01761564053595066,
-0.002672141883522272,
-0.008124710991978645,
-0.07340748608112335,
-0.061829522252082825,
0.1609770804643631,
-0.07277000695466995,
-0.06468547880649567,
0.05801168829202652,
0.03307786211371422,
0.01431563775986433,
-0.03584568202495575,
-0.04342148080468178,
0.18088261783123016,
0.29330700635910034,
-0.08191116154193878,
0.10001859813928604,
0.09677296131849289,
0.034820813685655594,
-0.23625829815864563,
0.029798466712236404,
-0.1455078274011612,
0.04449721798300743,
0.040447335690259933,
-0.0409548319876194,
0.04191497340798378,
0.10835777968168259,
-0.06094440817832947,
0.2048867791891098,
-0.03527235612273216,
-0.07983248680830002,
-0.01788630709052086,
0.03109324350953102,
0.29443636536598206,
-0.11833466589450836,
0.006058716680854559,
-0.10420958697795868,
-0.21566011011600494,
0.06983078271150589,
-0.18948867917060852,
0.13948246836662292,
-0.05087858438491821,
0.03576415032148361,
-0.01149723306298256,
-0.07561972737312317,
0.20518061518669128,
-0.15641045570373535,
0.05273103713989258,
-0.13722458481788635,
-0.1327189952135086,
0.01617460884153843,
-0.10048147290945053,
0.1545477658510208,
-0.11024226248264313,
-0.023215843364596367,
-0.2284185290336609,
0.012587235309183598,
-0.023200806230306625,
0.10030807554721832,
0.01800704374909401,
-0.07980740070343018,
-0.08767345547676086,
0.1316242516040802,
-0.06486566364765167,
0.034810543060302734,
-0.06996636837720871,
-0.050714004784822464,
-0.010929876938462257,
-0.045061707496643066,
0.03034941293299198,
-0.07934719324111938,
0.15192505717277527,
-0.016938980668783188,
-0.04507075995206833,
0.08636019378900528,
-0.2479533851146698,
0.023727843537926674,
0.025351112708449364,
-0.03495599329471588,
0.09001832455396652,
-0.025513244792819023,
-0.06256973743438721,
0.12282291799783707,
0.1402233988046646,
-0.07322840392589569,
-0.2460673749446869,
-0.06281693279743195,
0.0076784128323197365,
0.039165716618299484,
0.06561196595430374,
0.05125982314348221,
-0.07261458039283752,
-0.011131617240607738,
-0.026896944269537926,
0.030595947057008743,
-0.11692017316818237,
-0.03854857385158539,
0.07790639251470566,
0.017095070332288742,
-0.07846562564373016,
0.07280377298593521,
0.014225782826542854,
-0.021511616185307503,
0.007357571739703417,
0.148970365524292,
0.007519228849560022,
-0.14747941493988037,
-0.06656096875667572,
0.2007484883069992,
-0.01197928935289383,
-0.07260087132453918,
-0.05383119732141495,
-0.008990069851279259,
-0.0476234145462513,
0.05585788935422897,
0.05367223918437958,
-0.013585401698946953,
0.07708586007356644,
0.06263149529695511,
-0.10210110992193222,
-0.046256959438323975,
-0.066561758518219,
0.04169114679098129,
-0.10485753417015076,
0.060470130294561386,
0.009529483504593372,
0.12185006588697433,
-0.09983488917350769,
-0.01802929677069187,
-0.10810204595327377,
-0.06766588985919952,
-0.17349553108215332,
-0.05834362283349037,
-0.041105758398771286,
-0.015651104971766472,
0.03658895567059517,
0.010445823892951012,
-0.057867538183927536,
-0.0442853718996048,
-0.07536603510379791,
0.038444988429546356,
0.06147460639476776,
0.03932281583547592,
-0.03912714496254921,
0.04001858830451965,
0.05909334123134613,
0.013087345287203789,
0.17542624473571777,
0.038768354803323746,
0.05504675209522247,
-0.05045998468995094,
-0.16491834819316864,
-0.05276111513376236,
-0.0074316514655947685,
-0.07559102028608322,
0.1224973127245903,
-0.007679440546780825,
0.007880088873207569,
-0.08065467327833176,
0.03924860805273056,
0.028234204277396202,
0.10404064506292343,
-0.0028364830650389194,
0.10070426017045975,
0.019627176225185394,
-0.07226712256669998,
-0.025392837822437286,
0.021809715777635574,
0.12809939682483673,
0.01567147858440876,
0.026090998202562332,
0.033139873296022415,
0.016619985923171043,
-0.057361043989658356,
0.033977724611759186,
-0.04997231811285019,
-0.15123651921749115,
0.02628709189593792,
-0.05165188014507294,
0.005062380339950323,
-0.016889680176973343,
0.20362506806850433,
0.07867538928985596,
-0.06474173814058304,
-0.010664013214409351,
0.015816617757081985,
-0.0168940220028162,
-0.03121885471045971,
-0.012740966863930225,
0.04592578858137131,
-0.001151384087279439,
-0.04866636544466019,
0.11825273931026459,
0.05015748366713524,
0.05386412516236305,
0.0596686452627182,
0.12528513371944427,
0.016759619116783142,
0.13257254660129547,
0.061999931931495667,
-0.03403807803988457,
-0.13461735844612122,
-0.04495539888739586,
-0.1254577934741974,
0.04646851494908333,
-0.08697032928466797,
0.09941662102937698,
0.1144254133105278,
-0.05959030240774155,
-0.030464433133602142,
-0.08851305395364761,
-0.008356761187314987,
-0.06041252240538597,
0.039516255259513855,
-0.02262675203382969,
-0.0873224213719368,
0.0481097511947155,
0.05495472997426987,
-0.022752324119210243,
0.13218675553798676,
0.015727028250694275,
-0.036317698657512665,
0.13270340859889984,
-0.07583184540271759,
0.11758984625339508,
0.061510033905506134,
-0.043043944984674454,
-0.11560922116041183,
-0.020150646567344666,
-0.06641761213541031,
-0.10098972916603088,
-0.006782987620681524,
-0.005399650428444147,
-0.07349002361297607,
-0.059971679002046585,
0.08397487550973892,
-0.03124053031206131,
-0.09979676455259323,
-0.032152675092220306,
0.0038895104080438614,
0.06054706871509552,
-0.01686914451420307,
-0.0034020058810710907,
0.04728743061423302,
0.015076374635100365,
0.1653461456298828,
-0.02208263985812664,
0.06234867498278618,
-0.13855914771556854,
0.16070103645324707,
-0.14684462547302246,
-0.029404424130916595,
-0.1890171319246292,
-0.09729582816362381,
-0.05156542733311653,
0.20326784253120422,
0.2840938866138458,
-0.19109351933002472,
-0.010187864303588867,
0.020078664645552635,
-0.014484191313385963,
-0.08961770683526993,
0.12571553885936737,
0.029420215636491776,
-0.023631498217582703,
-0.07249019294977188,
-0.02037387527525425,
0.005258576478809118,
-0.06544211506843567,
-0.026979785412549973,
0.18310695886611938,
0.001496660872362554,
0.059546373784542084,
-0.09605178982019424,
0.01754261925816536,
-0.14839904010295868,
-0.10467469692230225,
-0.02111995778977871,
-0.16156397759914398,
-0.09646477550268173,
0.006635562051087618,
0.038640011101961136,
0.08000610023736954,
0.03268849849700928,
-0.015172510407865047,
0.06479045748710632,
-0.056333884596824646,
-0.0037216036580502987,
-0.1231912299990654,
0.00034658415825106204,
0.062129102647304535,
-0.07422006875276566,
0.2545335292816162,
-0.03070417232811451,
-0.12370815873146057,
0.09026903659105301,
-0.03299184888601303,
-0.12452623248100281,
0.07951879501342773,
-0.005700904875993729,
-0.11531132459640503,
-0.057989440858364105,
0.18941475450992584,
-0.012821312062442303,
-0.1364315301179886,
0.046368811279535294,
-0.17166484892368317,
0.031349923461675644,
0.0363016203045845,
-0.001313706859946251,
-0.04714022949337959,
0.024538639932870865,
-0.008008457720279694,
0.10724439471960068,
0.1382838785648346,
0.016739921644330025,
-0.011060068383812904,
-0.05056179314851761,
0.07912429422140121,
0.056927867233753204,
-0.05218246951699257,
-0.1282637119293213,
-0.08599764108657837,
0.03429819270968437,
0.04119478166103363,
-0.08113081753253937,
-0.16903182864189148,
-0.03668912500143051,
-0.10082915425300598,
-0.004939202684909105,
0.051785312592983246,
0.06585265696048737,
0.29044589400291443,
0.06326735019683838,
0.0016605621203780174,
-0.13649453222751617,
0.050569336861371994,
0.0868251696228981,
-0.04697931930422783,
-0.07670357078313828
] |
null | null | null |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# zephyr-support-chatbot-david-v4
This model is a fine-tuned version of [TheBloke/zephyr-7B-alpha-GPTQ](https://huggingface.co/TheBloke/zephyr-7B-alpha-GPTQ) on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- training_steps: 250
- mixed_precision_training: Native AMP
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.17.0
- Tokenizers 0.15.1
| {"license": "mit", "tags": ["trl", "sft", "generated_from_trainer"], "base_model": "TheBloke/zephyr-7B-alpha-GPTQ", "model-index": [{"name": "zephyr-support-chatbot-david-v4", "results": []}]} | null | David19930/zephyr-support-chatbot-david-v4 | [
"tensorboard",
"safetensors",
"trl",
"sft",
"generated_from_trainer",
"base_model:TheBloke/zephyr-7B-alpha-GPTQ",
"license:mit",
"region:us"
] | 2024-02-07T17:15:25+00:00 | [] | [] | TAGS
#tensorboard #safetensors #trl #sft #generated_from_trainer #base_model-TheBloke/zephyr-7B-alpha-GPTQ #license-mit #region-us
|
# zephyr-support-chatbot-david-v4
This model is a fine-tuned version of TheBloke/zephyr-7B-alpha-GPTQ on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- training_steps: 250
- mixed_precision_training: Native AMP
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.17.0
- Tokenizers 0.15.1
| [
"# zephyr-support-chatbot-david-v4\n\nThis model is a fine-tuned version of TheBloke/zephyr-7B-alpha-GPTQ on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- training_steps: 250\n- mixed_precision_training: Native AMP",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.1"
] | [
"TAGS\n#tensorboard #safetensors #trl #sft #generated_from_trainer #base_model-TheBloke/zephyr-7B-alpha-GPTQ #license-mit #region-us \n",
"# zephyr-support-chatbot-david-v4\n\nThis model is a fine-tuned version of TheBloke/zephyr-7B-alpha-GPTQ on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- training_steps: 250\n- mixed_precision_training: Native AMP",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.1"
] | [
53,
47,
6,
12,
8,
3,
102,
4,
33
] | [
"passage: TAGS\n#tensorboard #safetensors #trl #sft #generated_from_trainer #base_model-TheBloke/zephyr-7B-alpha-GPTQ #license-mit #region-us \n# zephyr-support-chatbot-david-v4\n\nThis model is a fine-tuned version of TheBloke/zephyr-7B-alpha-GPTQ on the None dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- training_steps: 250\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.1"
] | [
-0.11082810908555984,
0.005045703146606684,
-0.0009368295432068408,
0.06346536427736282,
0.11376094073057175,
0.030392322689294815,
0.12528647482395172,
0.12447568029165268,
-0.05949362739920616,
0.07176905125379562,
0.047219593077898026,
0.029702290892601013,
0.06739146262407303,
0.17940236628055573,
-0.02028045803308487,
-0.2051195353269577,
0.016965052112936974,
-0.016938524320721626,
-0.08592994511127472,
0.09743543714284897,
0.1056213453412056,
-0.12430789321660995,
0.04142119735479355,
0.0042041754350066185,
-0.1482614427804947,
0.0008090085466392338,
0.0022925541270524263,
-0.027071118354797363,
0.12369660288095474,
0.0019188643200322986,
0.11946819722652435,
0.0423843190073967,
0.1513470709323883,
-0.2325953245162964,
0.01540804747492075,
0.09470712393522263,
0.04507794231176376,
0.07400903850793839,
0.039315175265073776,
-0.0025048658717423677,
0.057410113513469696,
-0.14258182048797607,
0.09938006103038788,
0.021163728088140488,
-0.09591197967529297,
-0.17975342273712158,
-0.11891249567270279,
0.04762899503111839,
0.10548678785562515,
0.07857546210289001,
-0.0001565979910083115,
0.1432037502527237,
-0.09745236486196518,
0.03994350880384445,
0.21416543424129486,
-0.2440679967403412,
-0.07652988284826279,
0.10631723701953888,
0.07847514748573303,
0.07551597058773041,
-0.13076548278331757,
0.00003500713864923455,
0.03878024220466614,
0.02304098941385746,
0.09624957293272018,
-0.018415959551930428,
-0.09571796655654907,
-0.02364889346063137,
-0.1283389776945114,
-0.00030976085690781474,
0.08685634285211563,
0.04736979305744171,
-0.0431765541434288,
-0.09376591444015503,
-0.030768299475312233,
-0.07184112817049026,
-0.03657775744795799,
-0.0648079514503479,
0.03780151903629303,
-0.0331057608127594,
-0.056679584085941315,
-0.06477481126785278,
-0.10684894025325775,
-0.098975770175457,
0.012381908483803272,
0.10750430077314377,
0.02644839882850647,
0.00725127337500453,
-0.020480750128626823,
0.1407185196876526,
-0.04974452406167984,
-0.08800607919692993,
-0.01095494069159031,
-0.008014137856662273,
-0.11063705384731293,
-0.08915716409683228,
-0.030763063579797745,
-0.0683312863111496,
0.00888522993773222,
0.16021782159805298,
-0.0707755908370018,
0.07880911976099014,
-0.008701247163116932,
0.024692174047231674,
-0.05978918820619583,
0.08500665426254272,
-0.021251143887639046,
-0.023512279614806175,
0.007274908479303122,
0.11746446043252945,
-0.002117405878379941,
-0.009493541903793812,
-0.06808637082576752,
-0.02974013425409794,
0.046765804290771484,
0.05991748347878456,
-0.05996549129486084,
0.0348360575735569,
-0.06547991931438446,
-0.013332528993487358,
0.013364879414439201,
-0.09491658210754395,
0.05889052152633667,
0.0011878670193254948,
-0.05222582444548607,
-0.08603427559137344,
0.019809188321232796,
0.0362873338162899,
0.01934845745563507,
0.09028351306915283,
-0.06446237862110138,
0.020266247913241386,
-0.09432649612426758,
-0.07947367429733276,
0.018933262676000595,
-0.04038297384977341,
-0.03556188568472862,
-0.0792594701051712,
-0.17210108041763306,
-0.0599154569208622,
0.035569317638874054,
-0.039657872170209885,
0.016367116943001747,
-0.0475870817899704,
-0.022136511281132698,
0.02833271585404873,
-0.016926508396863937,
0.17383265495300293,
-0.05580170080065727,
0.07889176905155182,
-0.04150466248393059,
0.02446196973323822,
0.003673593746498227,
0.009346875362098217,
-0.08290864527225494,
0.02874860353767872,
-0.16175097227096558,
0.03415356203913689,
-0.10282687097787857,
-0.004755210597068071,
-0.1229313462972641,
-0.06006605550646782,
-0.05571616813540459,
-0.021639013662934303,
0.08202165365219116,
0.10692359507083893,
-0.22962641716003418,
-0.023708296939730644,
0.1910538375377655,
-0.10049671679735184,
-0.06073221564292908,
0.09226643294095993,
-0.06226439028978348,
0.06613456457853317,
0.06224001199007034,
0.15470658242702484,
0.08695328235626221,
-0.164031520485878,
0.009057082235813141,
-0.03637344762682915,
0.05875707045197487,
0.058017659932374954,
0.03530954197049141,
-0.041728612035512924,
0.02732572890818119,
-0.006626389920711517,
-0.05493541806936264,
0.016143999993801117,
-0.07556428015232086,
-0.06631549447774887,
-0.046447109431028366,
-0.07744654268026352,
0.07057572156190872,
0.02823948860168457,
0.019519485533237457,
-0.09659462422132492,
-0.09165121614933014,
0.07944367080926895,
0.12292933464050293,
-0.058087777346372604,
0.019131438806653023,
-0.06238958239555359,
-0.006828883662819862,
-0.008680351078510284,
-0.02366187795996666,
-0.16544997692108154,
-0.11497993767261505,
0.02771756611764431,
-0.05637148395180702,
0.04027946665883064,
0.03352763131260872,
0.07394808530807495,
0.06078816577792168,
-0.08700324594974518,
-0.004558565095067024,
-0.10155092179775238,
0.01299912016838789,
-0.11847715824842453,
-0.21524910628795624,
-0.032402317970991135,
-0.030146148055791855,
0.21043062210083008,
-0.25962501764297485,
0.021787412464618683,
0.01887776516377926,
0.1529577374458313,
0.03863853961229324,
-0.0539587065577507,
-0.004979467019438744,
0.06925738602876663,
0.01589491218328476,
-0.09460462629795074,
0.0398278534412384,
-0.0018969414522871375,
-0.10222242772579193,
-0.043969687074422836,
-0.15858818590641022,
-0.023489944636821747,
0.08529932796955109,
0.060723476111888885,
-0.1303626447916031,
-0.07125827670097351,
-0.060782331973314285,
-0.05155613645911217,
-0.10533449798822403,
0.013543365523219109,
0.1510242521762848,
0.0181371308863163,
0.10750439763069153,
-0.051251184195280075,
-0.05177995562553406,
0.0007337867864407599,
-0.0061612967401742935,
0.05046803876757622,
0.0637211725115776,
0.08758152276277542,
-0.1265701949596405,
0.07544733583927155,
0.10179401189088821,
-0.05559619516134262,
0.1516987383365631,
-0.040634479373693466,
-0.0767747238278389,
-0.01594139263033867,
0.027270324528217316,
0.015618483535945415,
0.16824759542942047,
-0.016159100458025932,
0.020457038655877113,
0.014037408865988255,
0.030764423310756683,
0.015547343529760838,
-0.20969133079051971,
-0.009558970108628273,
0.008767623454332352,
-0.04591822996735573,
0.019840506836771965,
-0.03446170315146446,
0.030728211626410484,
0.09239158779382706,
0.016134614124894142,
0.007250288967043161,
0.015578645281493664,
-0.003000755561515689,
-0.09722455590963364,
0.18046894669532776,
-0.12877385318279266,
-0.13812440633773804,
-0.0640849769115448,
0.09603656083345413,
-0.005591139197349548,
-0.020361624658107758,
0.006560600362718105,
-0.07824873924255371,
-0.02259909361600876,
-0.0824565589427948,
0.004224920645356178,
-0.012909408658742905,
-0.0029510250315070152,
-0.004762144759297371,
0.026396555826067924,
0.07263868302106857,
-0.12182436883449554,
0.011905104853212833,
-0.030905108898878098,
-0.10035354644060135,
0.021781539544463158,
0.05040046572685242,
0.07868187874555588,
0.14118658006191254,
-0.0267713014036417,
-0.012528305873274803,
-0.0710614025592804,
0.14621517062187195,
-0.10585924983024597,
0.02061944454908371,
0.11085119098424911,
0.017729956656694412,
0.03561629727482796,
0.10826840996742249,
0.04266886040568352,
-0.09656587243080139,
0.03518106788396835,
0.06560450792312622,
-0.031677160412073135,
-0.22237743437290192,
-0.03743806108832359,
-0.04013347625732422,
-0.05722709372639656,
0.09497636556625366,
0.04343150556087494,
0.032927095890045166,
0.051254481077194214,
-0.03832898661494255,
0.01385265402495861,
0.0267031230032444,
0.08406935632228851,
0.03374030813574791,
0.04736127704381943,
0.10638190060853958,
-0.02107835002243519,
-0.0037110471166670322,
0.06181095167994499,
0.046605732291936874,
0.24431733787059784,
0.006067665293812752,
0.09034153819084167,
0.06367859244346619,
0.15311269462108612,
-0.011934363283216953,
0.011215873062610626,
0.041330981999635696,
-0.04247623682022095,
0.0060002547688782215,
-0.06106097996234894,
-0.020431602373719215,
0.05665859207510948,
-0.004398041404783726,
0.02434820868074894,
-0.08266837894916534,
-0.01731758937239647,
0.02840256132185459,
0.237123504281044,
0.05336160957813263,
-0.2367260605096817,
-0.08643623441457748,
0.005785031709820032,
-0.019358692690730095,
-0.05659295991063118,
-0.022029047831892967,
0.141725093126297,
-0.14128565788269043,
0.06256218999624252,
-0.07187478244304657,
0.0746673196554184,
-0.02789328619837761,
-0.023241618648171425,
0.024904124438762665,
0.12206922471523285,
-0.024813394993543625,
0.0626336932182312,
-0.2476927638053894,
0.22265224158763885,
0.021938661113381386,
0.12584525346755981,
-0.05625392124056816,
0.014706971123814583,
0.03259091079235077,
0.05364791676402092,
0.08859197795391083,
-0.013615617528557777,
-0.11322203278541565,
-0.18747416138648987,
-0.060467906296253204,
0.030858535319566727,
0.1282520443201065,
-0.0525740310549736,
0.06735304743051529,
-0.029319871217012405,
0.01878049038350582,
0.05685288459062576,
-0.04725874215364456,
-0.22835326194763184,
-0.11086703836917877,
0.0050611235201358795,
0.04275865852832794,
-0.033377114683389664,
-0.11410112679004669,
-0.09728831797838211,
-0.01212230697274208,
0.1776013821363449,
-0.03553420677781105,
-0.008472724817693233,
-0.14429731667041779,
0.10067403316497803,
0.09861728549003601,
-0.03198206424713135,
0.031838446855545044,
0.038522835820913315,
0.13588403165340424,
-0.019030049443244934,
-0.05073564872145653,
0.07259215414524078,
-0.0807744637131691,
-0.20564669370651245,
-0.10282691568136215,
0.15749040246009827,
0.09544162452220917,
0.05452316626906395,
0.01542452909052372,
0.04338592290878296,
0.04321878403425217,
-0.08891534805297852,
0.040619201958179474,
0.11500230431556702,
0.09024659544229507,
0.03224669024348259,
-0.06741196662187576,
0.01730365864932537,
-0.029297735542058945,
-0.06579098850488663,
0.10597358644008636,
0.2453344762325287,
-0.09186743944883347,
0.089423768222332,
0.05024067312479019,
-0.0784895196557045,
-0.17650847136974335,
0.11579487472772598,
0.11934468895196915,
0.045145705342292786,
0.06026722863316536,
-0.17082546651363373,
0.06411704421043396,
0.13303495943546295,
-0.03018941916525364,
0.060603998601436615,
-0.3075500726699829,
-0.14933952689170837,
0.05573819577693939,
0.11057117581367493,
-0.0008044295245781541,
-0.11232173442840576,
-0.029370995238423347,
-0.03311120718717575,
-0.09948963671922684,
0.1223830059170723,
-0.15891918540000916,
0.10875076055526733,
0.003235618816688657,
0.04767845571041107,
0.022042980417609215,
-0.03783605247735977,
0.14942988753318787,
-0.012875959277153015,
0.09939584881067276,
-0.043567124754190445,
0.04605825990438461,
0.08320892602205276,
-0.07014576345682144,
0.0036283989902585745,
-0.027370048686861992,
0.041183676570653915,
-0.06231414154171944,
-0.018985774368047714,
-0.07972382754087448,
0.08044663071632385,
-0.043770723044872284,
-0.0717507004737854,
-0.06802728027105331,
0.08335193991661072,
0.017805304378271103,
-0.038220975548028946,
0.040864020586013794,
-0.02486412040889263,
0.15816913545131683,
0.06862102448940277,
0.1199558675289154,
0.006039049476385117,
-0.06310389935970306,
0.01553856860846281,
-0.04255174100399017,
0.06020557880401611,
-0.09934794157743454,
0.020862406119704247,
0.11757763475179672,
0.04497769474983215,
0.13373997807502747,
0.03577427938580513,
-0.07891059666872025,
0.019955886527895927,
0.054941460490226746,
-0.09454163163900375,
-0.1443546563386917,
0.034858062863349915,
0.04678432643413544,
-0.11184176802635193,
0.014850513078272343,
0.123371921479702,
-0.06349734216928482,
-0.021080929785966873,
-0.004415180999785662,
-0.0006133223068900406,
-0.04565805196762085,
0.19733034074306488,
0.03770311176776886,
0.06333749741315842,
-0.06117846071720123,
0.09718771278858185,
0.06353658437728882,
-0.060433000326156616,
0.03418257459998131,
0.0529300719499588,
-0.10534325242042542,
-0.011040138080716133,
0.07981164753437042,
0.1573062390089035,
0.0016219086246564984,
-0.049141108989715576,
-0.08023101091384888,
-0.08901331573724747,
0.00136184215079993,
0.13725923001766205,
0.022821476683020592,
0.00042244463111273944,
0.003013701643794775,
0.04829194024205208,
-0.1320822685956955,
0.056691549718379974,
0.010145350359380245,
0.08241492509841919,
-0.10766308009624481,
0.17192097008228302,
0.017390433698892593,
-0.00942031480371952,
-0.010219324380159378,
0.034925032407045364,
-0.08720380812883377,
-0.009556233882904053,
-0.10640623420476913,
-0.01988707296550274,
-0.0205236803740263,
-0.019751887768507004,
0.00029984163120388985,
-0.039686884731054306,
-0.007366885896772146,
0.034551385790109634,
-0.07536902278661728,
-0.04048429802060127,
-0.006763871293514967,
0.05792232230305672,
-0.1261519342660904,
-0.016649194061756134,
0.031692203134298325,
-0.0719858705997467,
0.049448780715465546,
0.035010918974876404,
0.06439702212810516,
0.044540587812662125,
-0.18679247796535492,
0.008971371687948704,
0.0275513157248497,
0.014112812466919422,
0.03279832750558853,
-0.08725914359092712,
-0.030983686447143555,
-0.04098550230264664,
0.040969062596559525,
0.03105185180902481,
0.02359052374958992,
-0.10757508873939514,
-0.0413975790143013,
-0.05932124704122543,
-0.053515899926424026,
-0.0410090871155262,
0.04188886284828186,
0.0767853781580925,
0.0575546957552433,
0.11816596239805222,
-0.10181606560945511,
0.04715298116207123,
-0.2144845873117447,
-0.04188564419746399,
-0.016494879499077797,
-0.012478012591600418,
-0.06562677770853043,
-0.03891737014055252,
0.09777384996414185,
-0.04130719229578972,
0.09007624536752701,
-0.017160192131996155,
0.09237886965274811,
0.03433840349316597,
-0.11565979570150375,
0.011665157973766327,
0.030189283192157745,
0.17014998197555542,
0.0626707673072815,
-0.018844639882445335,
0.06942576169967651,
0.013966848142445087,
0.04741515964269638,
0.05384426936507225,
0.214284747838974,
0.16114486753940582,
0.009689425118267536,
0.047239962965250015,
0.04974225535988808,
-0.12779083847999573,
-0.10178909450769424,
0.11015857011079788,
-0.034157413989305496,
0.06879041343927383,
-0.07347796112298965,
0.2004554122686386,
0.08911006152629852,
-0.19955094158649445,
0.035339921712875366,
-0.07653417438268661,
-0.10299184918403625,
-0.09034508466720581,
-0.05374748259782791,
-0.07905926555395126,
-0.14090238511562347,
0.012133656069636345,
-0.0906970277428627,
0.013487840071320534,
0.08755745738744736,
0.02250172570347786,
0.03478297218680382,
0.16056440770626068,
-0.012093540281057358,
0.017269954085350037,
0.05631832778453827,
0.023918392136693,
0.009964754804968834,
-0.09197462350130081,
-0.09876061975955963,
0.06727857887744904,
-0.03176131099462509,
0.06610079109668732,
-0.06864528357982635,
0.00007341938908211887,
0.02912735939025879,
0.03522973135113716,
-0.06641344726085663,
0.03985901549458504,
0.006010639946907759,
0.057265084236860275,
0.04684390127658844,
0.062337275594472885,
0.009281220845878124,
-0.03168058767914772,
0.2771972715854645,
-0.04945141077041626,
-0.10033810883760452,
-0.12332315742969513,
0.20492595434188843,
0.003021819284185767,
-0.0023338899482041597,
0.03457310423254967,
-0.12165538966655731,
0.033536870032548904,
0.12074341624975204,
0.13289107382297516,
-0.057964734733104706,
0.011081701144576073,
-0.017725631594657898,
-0.031124182045459747,
-0.07968726754188538,
0.11732977628707886,
0.05865952745079994,
-0.000914225762244314,
-0.07482168823480606,
-0.007122741546481848,
0.01934928633272648,
-0.024832261726260185,
-0.05330945551395416,
0.0629812628030777,
0.0031058359891176224,
0.017995547503232956,
-0.049349475651979446,
0.08629307150840759,
0.0828111544251442,
-0.19293245673179626,
0.07579030096530914,
-0.162678524851799,
-0.16305971145629883,
-0.02079819142818451,
0.05032980442047119,
-0.04329231008887291,
0.03268476575613022,
-0.03299041837453842,
0.012742800638079643,
0.08856261521577835,
-0.027118954807519913,
-0.002976200543344021,
-0.14020885527133942,
0.07472695410251617,
-0.05335187911987305,
0.21594774723052979,
-0.036426011472940445,
0.07795275747776031,
0.09649866074323654,
0.029648900032043457,
-0.09166081994771957,
0.0517578087747097,
0.05885219946503639,
-0.060818031430244446,
0.018727866932749748,
0.1530774086713791,
-0.05705437809228897,
0.11107555776834488,
0.061030082404613495,
-0.1758175641298294,
0.03773873299360275,
-0.06747309863567352,
-0.05355960130691528,
-0.09079821407794952,
0.03025662712752819,
-0.05995848774909973,
0.15616561472415924,
0.17020685970783234,
-0.03348405286669731,
0.022346820682287216,
-0.061738841235637665,
0.03525673598051071,
0.06439518928527832,
0.12799394130706787,
-0.027055773884058,
-0.21066096425056458,
0.036366041749715805,
0.07169993221759796,
0.008560249581933022,
-0.2009090632200241,
-0.1292116343975067,
0.03538886085152626,
-0.07182705402374268,
-0.03467327728867531,
0.11396495252847672,
0.04665976017713547,
0.04539494588971138,
-0.04029073566198349,
-0.22464193403720856,
-0.015314732678234577,
0.1352413445711136,
-0.12134748697280884,
-0.05669607222080231
] |
null | null | fastai |
# Amazing!
🥳 Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the [documentation here](https://huggingface.co/docs/hub/model-repos))!
2. Create a demo in Gradio or Streamlit using 🤗 Spaces ([documentation here](https://huggingface.co/docs/hub/spaces)).
3. Join the fastai community on the [Fastai Discord](https://discord.com/invite/YKrxeNn)!
Greetings fellow fastlearner 🤝! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| {"tags": ["fastai"]} | null | MiVaCod/intel-image-classification | [
"fastai",
"has_space",
"region:us"
] | 2024-02-07T17:15:35+00:00 | [] | [] | TAGS
#fastai #has_space #region-us
|
# Amazing!
Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the documentation here)!
2. Create a demo in Gradio or Streamlit using Spaces (documentation here).
3. Join the fastai community on the Fastai Discord!
Greetings fellow fastlearner ! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| [
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
"TAGS\n#fastai #has_space #region-us \n",
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
13,
20,
79,
3,
6,
12,
8
] | [
"passage: TAGS\n#fastai #has_space #region-us \n# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---# Model card## Model description\nMore information needed## Intended uses & limitations\nMore information needed## Training and evaluation data\nMore information needed"
] | [
-0.048121724277734756,
-0.024616125971078873,
0.002038001548498869,
0.10439170897006989,
0.135872021317482,
0.11887997388839722,
0.07405775785446167,
0.09980081021785736,
0.07783667743206024,
0.02590852417051792,
0.08961158245801926,
-0.08088712394237518,
0.08744348585605621,
0.271692156791687,
0.06988707184791565,
-0.22761479020118713,
0.04051019623875618,
-0.00024903909070417285,
0.08053462207317352,
0.06629016250371933,
0.13507555425167084,
-0.05464952811598778,
0.14010503888130188,
-0.004088983871042728,
-0.19050447642803192,
-0.042929794639348984,
-0.01773718371987343,
-0.02527874894440174,
0.12317648530006409,
-0.04744937643408775,
0.05381017178297043,
0.015037551522254944,
0.007565062493085861,
-0.07253646105527878,
0.0623294934630394,
0.040457066148519516,
0.01740180514752865,
0.059235580265522,
-0.07249044626951218,
0.08950132131576538,
0.08404164761304855,
-0.024370938539505005,
-0.1097978875041008,
0.07827875018119812,
-0.14424212276935577,
-0.21762843430042267,
-0.1253085881471634,
-0.09017651528120041,
0.028519365936517715,
0.004388005938380957,
-0.025051530450582504,
0.12801909446716309,
-0.13558274507522583,
-0.040698226541280746,
0.20124278962612152,
-0.17012301087379456,
-0.05505548417568207,
0.034343402832746506,
0.09226689487695694,
-0.05829555168747902,
-0.06347129493951797,
0.10614984482526779,
0.09640881419181824,
-0.019833475351333618,
0.05516824126243591,
0.002579754451289773,
0.021173657849431038,
0.01370104867964983,
-0.06150497496128082,
0.04717832803726196,
-0.010183089412748814,
0.048132527619600296,
-0.09465572983026505,
-0.1303568333387375,
-0.004072192590683699,
0.01214400865137577,
-0.048744890838861465,
-0.07019646465778351,
0.07833103090524673,
-0.011118141002953053,
-0.04357248544692993,
-0.13031910359859467,
-0.09131011366844177,
-0.12358787655830383,
0.008646543137729168,
0.09500427544116974,
0.003679296001791954,
0.07374339550733566,
-0.08258994668722153,
0.06774985045194626,
-0.17329485714435577,
-0.06484591960906982,
-0.08138520270586014,
-0.11546400189399719,
0.021133482456207275,
-0.0387684591114521,
0.02668963186442852,
0.15394504368305206,
0.12983950972557068,
0.023976242169737816,
0.04388163983821869,
-0.038937073200941086,
0.051190316677093506,
0.058571770787239075,
0.03395717963576317,
0.034934818744659424,
-0.036981891840696335,
-0.1793210655450821,
-0.016702448949217796,
-0.011550825089216232,
0.07954040914773941,
-0.07523109763860703,
-0.05632320046424866,
0.013454885222017765,
-0.11071494966745377,
0.07202339172363281,
-0.03576776012778282,
-0.0032025426626205444,
0.01168301422148943,
0.018371861428022385,
0.21271461248397827,
0.03955606371164322,
0.014191740192472935,
-0.008875265717506409,
-0.13453757762908936,
-0.06874168664216995,
-0.06896194815635681,
0.03361047804355621,
0.04448792710900307,
-0.0028071461711078882,
-0.07672245055437088,
0.04325154796242714,
-0.06045534089207649,
-0.03508453071117401,
0.008032378740608692,
-0.18221288919448853,
0.007458044681698084,
-0.10049355030059814,
-0.12126200646162033,
0.05306628718972206,
0.01695440337061882,
-0.08215925842523575,
0.08141279965639114,
0.02662261202931404,
0.020931517705321312,
-0.009988143108785152,
-0.005391082260757685,
0.06874798238277435,
-0.08508864045143127,
0.029901226982474327,
0.17170792818069458,
0.13024519383907318,
-0.08046911656856537,
-0.0006887061172164977,
-0.10965746641159058,
0.04426072910428047,
-0.13325683772563934,
0.02251482754945755,
-0.09062390774488449,
0.11723794043064117,
-0.042396437376737595,
0.002038756385445595,
-0.029030200093984604,
0.0960269495844841,
0.08189879357814789,
0.16663365066051483,
-0.2419009804725647,
-0.031095001846551895,
0.13240347802639008,
-0.10711425542831421,
-0.1807439625263214,
0.18486657738685608,
-0.012035200372338295,
0.11329247802495956,
-0.047014184296131134,
0.18334640562534332,
-0.02612062357366085,
-0.13582459092140198,
-0.058872904628515244,
0.005852419883012772,
-0.2269321084022522,
-0.06286033242940903,
0.09738040715456009,
0.13425657153129578,
-0.042984943836927414,
0.007112155202776194,
0.026316028088331223,
0.13609857857227325,
-0.06715573370456696,
-0.05195777863264084,
-0.012255736626684666,
-0.10902371257543564,
0.041914235800504684,
0.018215661868453026,
0.035408079624176025,
-0.059880174696445465,
-0.02931194379925728,
-0.053190283477306366,
0.13146710395812988,
0.09760832786560059,
-0.03670211136341095,
-0.049620725214481354,
0.1689043790102005,
-0.07763876020908356,
-0.033587727695703506,
0.07560533285140991,
-0.08268500119447708,
0.03266897425055504,
0.03090597130358219,
0.055881720036268234,
0.07766123116016388,
0.08522116392850876,
0.06057543307542801,
0.00819048099219799,
0.034654274582862854,
0.12095347046852112,
-0.013591280207037926,
-0.05039411783218384,
0.021508218720555305,
0.016904234886169434,
-0.019032588228583336,
0.29030677676200867,
-0.1951042115688324,
0.024724548682570457,
-0.06477324664592743,
0.07631538063287735,
0.06136792525649071,
0.003575638635084033,
0.08580143749713898,
-0.06023019179701805,
-0.019061198458075523,
-0.04803973436355591,
0.046805646270513535,
-0.0666879191994667,
-0.04162997007369995,
0.2621194124221802,
-0.05497581139206886,
0.044914912432432175,
0.12313763797283173,
-0.05873025581240654,
-0.07091446220874786,
0.01009807363152504,
-0.00793424155563116,
0.03249288722872734,
-0.04042816907167435,
0.043721720576286316,
-0.10840129852294922,
-0.06674089282751083,
0.1573198139667511,
-0.038477856665849686,
0.06786153465509415,
0.032288823276758194,
-0.04958454892039299,
-0.0648743286728859,
0.04650486260652542,
0.13598160445690155,
-0.0875244215130806,
0.07435166835784912,
0.17612984776496887,
-0.010562662966549397,
0.168031245470047,
0.08435525000095367,
-0.07075224816799164,
-0.09465329349040985,
-0.051014289259910583,
-0.021595727652311325,
0.21222901344299316,
-0.07084725052118301,
-0.054564714431762695,
0.05911700800061226,
-0.013703816570341587,
0.07196151465177536,
-0.06009222939610481,
-0.08332337439060211,
0.03227344527840614,
-0.04517695680260658,
0.011517706327140331,
0.13512636721134186,
-0.07090822607278824,
0.04681389778852463,
0.031489867717027664,
-0.0662703812122345,
0.02217509225010872,
0.033389873802661896,
0.0068921963684260845,
0.033959709107875824,
0.07332495599985123,
-0.20893315970897675,
-0.08408680558204651,
-0.13727638125419617,
0.037881869822740555,
0.021770721301436424,
0.045787326991558075,
-0.08602345734834671,
0.02231026627123356,
-0.08954031765460968,
-0.07987114042043686,
0.029592275619506836,
-0.026350297033786774,
-0.11349643021821976,
-0.03396226093173027,
-0.009560913778841496,
-0.06662604957818985,
-0.02250705659389496,
-0.05024505779147148,
0.03983384370803833,
0.04479299485683441,
0.058377087116241455,
0.12796473503112793,
-0.013808943331241608,
-0.03839317709207535,
0.000370211957488209,
-0.022712308913469315,
0.16396735608577728,
-0.14746315777301788,
0.07954913377761841,
0.19160102307796478,
0.11742953956127167,
0.028144672513008118,
0.028885571286082268,
0.03537585213780403,
-0.06289814412593842,
-0.000050317394197918475,
0.03226194158196449,
-0.09392514824867249,
-0.05801016092300415,
-0.020014392212033272,
-0.04031052812933922,
0.17134574055671692,
-0.12160717695951462,
0.03345204517245293,
0.04098419472575188,
0.09783966839313507,
0.10073629021644592,
-0.028829937800765038,
-0.1815856397151947,
0.038818612694740295,
-0.24060091376304626,
-0.05831146240234375,
0.027899866923689842,
-0.09110201895236969,
-0.06232144311070442,
0.17409387230873108,
0.013794700615108013,
0.011769929900765419,
-0.006736889015883207,
0.07983319461345673,
0.0110100656747818,
0.1217205822467804,
0.05947643890976906,
-0.05539114400744438,
0.025202350690960884,
-0.09962950646877289,
-0.07107596844434738,
-0.04035590961575508,
-0.05832801014184952,
0.07548832893371582,
0.1409129947423935,
-0.025475580245256424,
-0.020795362070202827,
0.023489827290177345,
0.08550169318914413,
0.0423230417072773,
0.16739299893379211,
-0.16016584634780884,
-0.026555389165878296,
0.04571257904171944,
-0.03384667634963989,
-0.05433850735425949,
-0.010291114449501038,
0.1137225553393364,
-0.02820689231157303,
-0.040318265557289124,
0.021242983639240265,
0.06503437459468842,
0.01481706090271473,
0.05012747645378113,
-0.04056356102228165,
0.14796851575374603,
-0.03461192920804024,
0.019330544397234917,
-0.12413888424634933,
0.13848772644996643,
0.021095896139740944,
-0.03901609033346176,
-0.06735876202583313,
-0.05808034539222717,
0.18150931596755981,
0.0025602965615689754,
0.10535930097103119,
0.012098877690732479,
-0.12160047143697739,
-0.1359938681125641,
-0.11211287975311279,
0.005111907608807087,
0.08330471813678741,
-0.023147236555814743,
-0.022247863933444023,
0.022165266796946526,
-0.036149751394987106,
-0.0530381016433239,
0.15749511122703552,
-0.1289154291152954,
-0.001082550617866218,
0.014728817157447338,
0.06971760839223862,
-0.08223173767328262,
0.026267826557159424,
0.014071501791477203,
-0.1119147390127182,
0.10590848326683044,
0.2521335482597351,
0.10338116437196732,
-0.09591643512248993,
-0.07697287201881409,
0.03418830782175064,
-0.012184361927211285,
-0.000774814048781991,
-0.006932659074664116,
0.0495428591966629,
-0.005566445179283619,
0.006762749515473843,
0.12971895933151245,
-0.07130889594554901,
0.011540771462023258,
-0.08449850976467133,
0.05566910281777382,
-0.05276734381914139,
0.01761564053595066,
-0.002672141883522272,
-0.008124710991978645,
-0.07340748608112335,
-0.061829522252082825,
0.1609770804643631,
-0.07277000695466995,
-0.06468547880649567,
0.05801168829202652,
0.03307786211371422,
0.01431563775986433,
-0.03584568202495575,
-0.04342148080468178,
0.18088261783123016,
0.29330700635910034,
-0.08191116154193878,
0.10001859813928604,
0.09677296131849289,
0.034820813685655594,
-0.23625829815864563,
0.029798466712236404,
-0.1455078274011612,
0.04449721798300743,
0.040447335690259933,
-0.0409548319876194,
0.04191497340798378,
0.10835777968168259,
-0.06094440817832947,
0.2048867791891098,
-0.03527235612273216,
-0.07983248680830002,
-0.01788630709052086,
0.03109324350953102,
0.29443636536598206,
-0.11833466589450836,
0.006058716680854559,
-0.10420958697795868,
-0.21566011011600494,
0.06983078271150589,
-0.18948867917060852,
0.13948246836662292,
-0.05087858438491821,
0.03576415032148361,
-0.01149723306298256,
-0.07561972737312317,
0.20518061518669128,
-0.15641045570373535,
0.05273103713989258,
-0.13722458481788635,
-0.1327189952135086,
0.01617460884153843,
-0.10048147290945053,
0.1545477658510208,
-0.11024226248264313,
-0.023215843364596367,
-0.2284185290336609,
0.012587235309183598,
-0.023200806230306625,
0.10030807554721832,
0.01800704374909401,
-0.07980740070343018,
-0.08767345547676086,
0.1316242516040802,
-0.06486566364765167,
0.034810543060302734,
-0.06996636837720871,
-0.050714004784822464,
-0.010929876938462257,
-0.045061707496643066,
0.03034941293299198,
-0.07934719324111938,
0.15192505717277527,
-0.016938980668783188,
-0.04507075995206833,
0.08636019378900528,
-0.2479533851146698,
0.023727843537926674,
0.025351112708449364,
-0.03495599329471588,
0.09001832455396652,
-0.025513244792819023,
-0.06256973743438721,
0.12282291799783707,
0.1402233988046646,
-0.07322840392589569,
-0.2460673749446869,
-0.06281693279743195,
0.0076784128323197365,
0.039165716618299484,
0.06561196595430374,
0.05125982314348221,
-0.07261458039283752,
-0.011131617240607738,
-0.026896944269537926,
0.030595947057008743,
-0.11692017316818237,
-0.03854857385158539,
0.07790639251470566,
0.017095070332288742,
-0.07846562564373016,
0.07280377298593521,
0.014225782826542854,
-0.021511616185307503,
0.007357571739703417,
0.148970365524292,
0.007519228849560022,
-0.14747941493988037,
-0.06656096875667572,
0.2007484883069992,
-0.01197928935289383,
-0.07260087132453918,
-0.05383119732141495,
-0.008990069851279259,
-0.0476234145462513,
0.05585788935422897,
0.05367223918437958,
-0.013585401698946953,
0.07708586007356644,
0.06263149529695511,
-0.10210110992193222,
-0.046256959438323975,
-0.066561758518219,
0.04169114679098129,
-0.10485753417015076,
0.060470130294561386,
0.009529483504593372,
0.12185006588697433,
-0.09983488917350769,
-0.01802929677069187,
-0.10810204595327377,
-0.06766588985919952,
-0.17349553108215332,
-0.05834362283349037,
-0.041105758398771286,
-0.015651104971766472,
0.03658895567059517,
0.010445823892951012,
-0.057867538183927536,
-0.0442853718996048,
-0.07536603510379791,
0.038444988429546356,
0.06147460639476776,
0.03932281583547592,
-0.03912714496254921,
0.04001858830451965,
0.05909334123134613,
0.013087345287203789,
0.17542624473571777,
0.038768354803323746,
0.05504675209522247,
-0.05045998468995094,
-0.16491834819316864,
-0.05276111513376236,
-0.0074316514655947685,
-0.07559102028608322,
0.1224973127245903,
-0.007679440546780825,
0.007880088873207569,
-0.08065467327833176,
0.03924860805273056,
0.028234204277396202,
0.10404064506292343,
-0.0028364830650389194,
0.10070426017045975,
0.019627176225185394,
-0.07226712256669998,
-0.025392837822437286,
0.021809715777635574,
0.12809939682483673,
0.01567147858440876,
0.026090998202562332,
0.033139873296022415,
0.016619985923171043,
-0.057361043989658356,
0.033977724611759186,
-0.04997231811285019,
-0.15123651921749115,
0.02628709189593792,
-0.05165188014507294,
0.005062380339950323,
-0.016889680176973343,
0.20362506806850433,
0.07867538928985596,
-0.06474173814058304,
-0.010664013214409351,
0.015816617757081985,
-0.0168940220028162,
-0.03121885471045971,
-0.012740966863930225,
0.04592578858137131,
-0.001151384087279439,
-0.04866636544466019,
0.11825273931026459,
0.05015748366713524,
0.05386412516236305,
0.0596686452627182,
0.12528513371944427,
0.016759619116783142,
0.13257254660129547,
0.061999931931495667,
-0.03403807803988457,
-0.13461735844612122,
-0.04495539888739586,
-0.1254577934741974,
0.04646851494908333,
-0.08697032928466797,
0.09941662102937698,
0.1144254133105278,
-0.05959030240774155,
-0.030464433133602142,
-0.08851305395364761,
-0.008356761187314987,
-0.06041252240538597,
0.039516255259513855,
-0.02262675203382969,
-0.0873224213719368,
0.0481097511947155,
0.05495472997426987,
-0.022752324119210243,
0.13218675553798676,
0.015727028250694275,
-0.036317698657512665,
0.13270340859889984,
-0.07583184540271759,
0.11758984625339508,
0.061510033905506134,
-0.043043944984674454,
-0.11560922116041183,
-0.020150646567344666,
-0.06641761213541031,
-0.10098972916603088,
-0.006782987620681524,
-0.005399650428444147,
-0.07349002361297607,
-0.059971679002046585,
0.08397487550973892,
-0.03124053031206131,
-0.09979676455259323,
-0.032152675092220306,
0.0038895104080438614,
0.06054706871509552,
-0.01686914451420307,
-0.0034020058810710907,
0.04728743061423302,
0.015076374635100365,
0.1653461456298828,
-0.02208263985812664,
0.06234867498278618,
-0.13855914771556854,
0.16070103645324707,
-0.14684462547302246,
-0.029404424130916595,
-0.1890171319246292,
-0.09729582816362381,
-0.05156542733311653,
0.20326784253120422,
0.2840938866138458,
-0.19109351933002472,
-0.010187864303588867,
0.020078664645552635,
-0.014484191313385963,
-0.08961770683526993,
0.12571553885936737,
0.029420215636491776,
-0.023631498217582703,
-0.07249019294977188,
-0.02037387527525425,
0.005258576478809118,
-0.06544211506843567,
-0.026979785412549973,
0.18310695886611938,
0.001496660872362554,
0.059546373784542084,
-0.09605178982019424,
0.01754261925816536,
-0.14839904010295868,
-0.10467469692230225,
-0.02111995778977871,
-0.16156397759914398,
-0.09646477550268173,
0.006635562051087618,
0.038640011101961136,
0.08000610023736954,
0.03268849849700928,
-0.015172510407865047,
0.06479045748710632,
-0.056333884596824646,
-0.0037216036580502987,
-0.1231912299990654,
0.00034658415825106204,
0.062129102647304535,
-0.07422006875276566,
0.2545335292816162,
-0.03070417232811451,
-0.12370815873146057,
0.09026903659105301,
-0.03299184888601303,
-0.12452623248100281,
0.07951879501342773,
-0.005700904875993729,
-0.11531132459640503,
-0.057989440858364105,
0.18941475450992584,
-0.012821312062442303,
-0.1364315301179886,
0.046368811279535294,
-0.17166484892368317,
0.031349923461675644,
0.0363016203045845,
-0.001313706859946251,
-0.04714022949337959,
0.024538639932870865,
-0.008008457720279694,
0.10724439471960068,
0.1382838785648346,
0.016739921644330025,
-0.011060068383812904,
-0.05056179314851761,
0.07912429422140121,
0.056927867233753204,
-0.05218246951699257,
-0.1282637119293213,
-0.08599764108657837,
0.03429819270968437,
0.04119478166103363,
-0.08113081753253937,
-0.16903182864189148,
-0.03668912500143051,
-0.10082915425300598,
-0.004939202684909105,
0.051785312592983246,
0.06585265696048737,
0.29044589400291443,
0.06326735019683838,
0.0016605621203780174,
-0.13649453222751617,
0.050569336861371994,
0.0868251696228981,
-0.04697931930422783,
-0.07670357078313828
] |
null | null | stable-baselines3 |
# **PPO** Agent playing **LunarLander-v2**
This is a trained model of a **PPO** agent playing **LunarLander-v2**
using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3).
## Usage (with Stable-baselines3)
TODO: Add your code
```python
from stable_baselines3 import ...
from huggingface_sb3 import load_from_hub
...
```
| {"library_name": "stable-baselines3", "tags": ["LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "PPO", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "LunarLander-v2", "type": "LunarLander-v2"}, "metrics": [{"type": "mean_reward", "value": "275.40 +/- 22.27", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | Poliuszko/ppo-LunarLander-v21-1 | [
"stable-baselines3",
"LunarLander-v2",
"deep-reinforcement-learning",
"reinforcement-learning",
"model-index",
"region:us"
] | 2024-02-07T17:16:49+00:00 | [] | [] | TAGS
#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
|
# PPO Agent playing LunarLander-v2
This is a trained model of a PPO agent playing LunarLander-v2
using the stable-baselines3 library.
## Usage (with Stable-baselines3)
TODO: Add your code
| [
"# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
"TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n",
"# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
39,
41,
17
] | [
"passage: TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
0.03942384943366051,
0.04900386184453964,
-0.005304091144353151,
0.026427261531352997,
0.107408307492733,
-0.026511888951063156,
0.11188238859176636,
0.0814051404595375,
0.10722193866968155,
0.04762078449130058,
0.08338645845651627,
0.06030960753560066,
0.05080918222665787,
0.2571701407432556,
0.04754156619310379,
-0.22987541556358337,
0.036159250885248184,
-0.04869936779141426,
0.12395193427801132,
0.07178173214197159,
-0.0038484656251966953,
-0.06485428661108017,
0.020415637642145157,
-0.013290755450725555,
0.05367108806967735,
0.04282612353563309,
-0.01716216839849949,
-0.08207534998655319,
0.07169748842716217,
-0.06345846503973007,
0.06986866891384125,
0.07677983492612839,
0.13218913972377777,
-0.17832116782665253,
0.029566360637545586,
0.02571309357881546,
-0.07189024239778519,
0.01342033501714468,
0.008019951172173023,
0.05120139941573143,
0.17303818464279175,
0.019879888743162155,
0.07844575494527817,
-0.0025605305563658476,
-0.15412317216396332,
-0.018950799480080605,
0.0436202734708786,
0.12546207010746002,
0.08808347582817078,
0.04605821147561073,
0.01970590092241764,
0.17503218352794647,
-0.054352790117263794,
-0.028833400458097458,
0.21759237349033356,
-0.2881564497947693,
-0.031460098922252655,
0.321048766374588,
0.06997483223676682,
0.09725230932235718,
-0.07540661096572876,
-0.03619609400629997,
0.007783263456076384,
-0.013137873262166977,
-0.028666524216532707,
-0.07447073608636856,
0.17313385009765625,
0.05152064561843872,
-0.05057951435446739,
-0.09541505575180054,
0.16948209702968597,
0.006921638268977404,
0.0018855923553928733,
-0.019282981753349304,
0.009060598909854889,
0.07402525842189789,
-0.016097044572234154,
-0.07255112379789352,
0.057438433170318604,
0.05330665782094002,
0.019649166613817215,
-0.1435653269290924,
-0.10762494057416916,
-0.022740179672837257,
-0.008012006990611553,
0.17786912620067596,
-0.009255532175302505,
0.042902372777462006,
0.003065188182517886,
0.10384012013673782,
-0.12480384111404419,
-0.03354184702038765,
-0.0454259067773819,
-0.07565800100564957,
-0.0223417766392231,
-0.02058211714029312,
-0.03580251708626747,
0.07184842973947525,
0.11971849203109741,
0.027368178591132164,
0.09350208193063736,
0.047715865075588226,
-0.03206788748502731,
0.06343851238489151,
0.05555703118443489,
0.14222665131092072,
0.05807621404528618,
0.012854371219873428,
0.13179877400398254,
0.055213116109371185,
0.033023182302713394,
-0.0613492950797081,
-0.18252409994602203,
0.07489913702011108,
-0.07031869143247604,
0.007941240444779396,
0.12051256000995636,
-0.04480670019984245,
-0.1183447614312172,
-0.037500523030757904,
-0.017392054200172424,
-0.06224250793457031,
-0.025395862758159637,
0.0547584593296051,
-0.02883218228816986,
-0.03973718360066414,
0.0011496668448671699,
0.09384800493717194,
0.00953749567270279,
-0.1752052903175354,
0.03303423151373863,
-0.025042934343218803,
-0.10782608389854431,
0.009975161403417587,
0.0022444494534283876,
0.03394931182265282,
0.04408763721585274,
-0.11822668462991714,
-0.30899152159690857,
-0.07652641832828522,
0.05490870401263237,
-0.06516939401626587,
-0.18425025045871735,
-0.13193942606449127,
0.02454492449760437,
-0.09037084132432938,
-0.044885024428367615,
-0.12759265303611755,
-0.028549788519740105,
0.01743689924478531,
0.011519349180161953,
0.10758619755506516,
-0.0106219332665205,
-0.012188062071800232,
-0.1571401208639145,
0.008273907005786896,
-0.20951123535633087,
0.0890483483672142,
-0.019150104373693466,
0.037884220480918884,
-0.032381169497966766,
-0.07404014468193054,
0.030707746744155884,
0.052499737590551376,
-0.01474119070917368,
0.13510210812091827,
-0.15592676401138306,
-0.03691192343831062,
-0.007996266707777977,
-0.13611900806427002,
-0.04786273464560509,
-0.10358831286430359,
-0.04357128217816353,
0.13354332745075226,
0.018664736300706863,
0.15356586873531342,
-0.08709818124771118,
-0.0722038671374321,
0.20489206910133362,
-0.010411538183689117,
-0.12820468842983246,
-0.076752208173275,
0.10165707021951675,
0.021510310471057892,
-0.056606587022542953,
-0.02523270808160305,
-0.1839766949415207,
-0.0152357779443264,
-0.04550420492887497,
-0.047039128839969635,
0.01796751655638218,
-0.010888241231441498,
0.13837894797325134,
0.08494598418474197,
0.05018039792776108,
-0.06086122244596481,
-0.006730288732796907,
0.10779471695423126,
0.08823856711387634,
0.008680110797286034,
0.023406028747558594,
-0.05774238705635071,
0.09552932530641556,
-0.04003755748271942,
-0.0142367510125041,
-0.08283266425132751,
-0.036246106028556824,
-0.026256313547492027,
0.17507147789001465,
0.09440762549638748,
0.2257927656173706,
0.09567736834287643,
0.039160262793302536,
0.031270865350961685,
-0.13181598484516144,
-0.1425403207540512,
-0.0017254541162401438,
0.09020978957414627,
-0.14270411431789398,
-0.04119925573468208,
-0.08974775671958923,
-0.17768175899982452,
-0.12202505767345428,
0.0006432619411498308,
-0.17960017919540405,
0.06390921026468277,
0.05408334732055664,
-0.035177867859601974,
0.03272094577550888,
0.13032332062721252,
-0.011533179320394993,
-0.03967514634132385,
0.0831870287656784,
0.0379033200442791,
-0.041234664618968964,
-0.021742934361100197,
0.11885567009449005,
0.15673065185546875,
0.13124459981918335,
-0.03511447086930275,
0.004914294462651014,
0.07076404243707657,
-0.02309088408946991,
0.06539414077997208,
0.0558244064450264,
0.20973342657089233,
0.188301220536232,
0.038996949791908264,
0.008822928182780743,
-0.07048165798187256,
0.0855446457862854,
-0.0742373839020729,
-0.14302679896354675,
-0.05579735338687897,
0.08729292452335358,
0.016605578362941742,
0.023469142615795135,
0.08711627870798111,
0.024545932188630104,
0.09132762253284454,
0.15968108177185059,
0.01990218088030815,
-0.09659269452095032,
-0.050218869000673294,
0.01175848301500082,
0.027713103219866753,
0.04794301092624664,
-0.04514073207974434,
-0.00937939714640379,
0.017020760104060173,
-0.10303554683923721,
0.031789086759090424,
-0.1413339376449585,
-0.1358717679977417,
0.044326696544885635,
0.003906996920704842,
0.010907664895057678,
0.02786896750330925,
-0.0038291432429105043,
0.019039705395698547,
0.04351753741502762,
-0.06975466758012772,
0.047416772693395615,
-0.024745507165789604,
-0.020031947642564774,
0.03340689837932587,
-0.057257164269685745,
-0.205775648355484,
-0.17696654796600342,
0.00013708483311347663,
-0.09910997003316879,
0.10194740444421768,
0.018308809027075768,
-0.12373185902833939,
0.047737859189510345,
-0.05822649225592613,
0.027574289590120316,
-0.01875593699514866,
-0.049130141735076904,
0.10507171601057053,
0.1525275856256485,
-0.016146350651979446,
0.018018173053860664,
-0.04865182936191559,
-0.10157987475395203,
-0.19632206857204437,
0.0691583976149559,
0.04680244252085686,
0.014610917307436466,
0.10669491440057755,
0.018072687089443207,
0.02367905154824257,
-0.007674071006476879,
-0.016521066427230835,
-0.011659215204417706,
-0.08781040459871292,
0.31909599900245667,
0.04510033503174782,
-0.025173069909214973,
0.02041010931134224,
-0.0043001663871109486,
-0.028083480894565582,
0.03263787180185318,
-0.0985708013176918,
-0.07548979669809341,
-0.08774089068174362,
-0.04367410019040108,
-0.09784720093011856,
0.053299110382795334,
0.05916472524404526,
0.003188040340319276,
-0.07727594673633575,
0.04221395403146744,
0.11369874328374863,
-0.0923808291554451,
-0.07137343287467957,
0.07477962225675583,
0.0972946360707283,
-0.07331304252147675,
0.00012658814375754446,
0.00874367356300354,
0.023951783776283264,
0.037102166563272476,
0.06778035312891006,
-0.03966575115919113,
0.08589404821395874,
-0.19917890429496765,
0.0372927263379097,
0.106058269739151,
0.023754918947815895,
0.0638108178973198,
0.07643651217222214,
-0.1058402881026268,
-0.008500572293996811,
-0.032518330961465836,
-0.21341575682163239,
0.1668180525302887,
0.1355515867471695,
0.06788124144077301,
-0.025637222453951836,
-0.00461410591378808,
-0.0649740919470787,
0.05773647129535675,
0.02723747305572033,
-0.14758841693401337,
0.004883295856416225,
0.06064270809292793,
0.026899009943008423,
0.01614922471344471,
0.07971042394638062,
0.014697225764393806,
-0.1801026314496994,
-0.014406266622245312,
0.10730406641960144,
0.002390873385593295,
0.0053148469887673855,
-0.03175045922398567,
-0.1755964607000351,
0.0751047357916832,
0.004285442177206278,
0.07233936339616776,
-0.1676585078239441,
0.14297930896282196,
-0.10089799761772156,
0.07726949453353882,
-0.004285062663257122,
-0.021311495453119278,
0.02507244050502777,
-0.0541163794696331,
0.15163759887218475,
0.01058570109307766,
-0.021810131147503853,
-0.1200498715043068,
-0.1717042326927185,
-0.019227758049964905,
-0.11788936704397202,
-0.11679866164922714,
0.050424277782440186,
0.062185097485780716,
0.04923136904835701,
-0.061147067695856094,
0.1518532931804657,
-0.047422297298908234,
0.060713399201631546,
-0.06893875449895859,
-0.06755045056343079,
0.03764858841896057,
-0.12588608264923096,
-0.08176055550575256,
0.05573027580976486,
0.19166934490203857,
0.15833087265491486,
-0.02816431224346161,
-0.03472423925995827,
-0.047419581562280655,
-0.006212298292666674,
-0.007802055217325687,
0.0275666993111372,
0.023223137483000755,
0.07315318286418915,
-0.07681374251842499,
-0.11649256944656372,
0.033787861466407776,
-0.06713802367448807,
-0.055589709430933,
-0.015439179725944996,
0.1513158082962036,
0.04671623185276985,
0.07720734924077988,
-0.018946662545204163,
0.03887668624520302,
-0.001724981120787561,
-0.056474871933460236,
0.16197094321250916,
0.03885216265916824,
-0.05193585529923439,
0.06837689876556396,
0.053174007683992386,
0.043745119124650955,
0.03011113777756691,
-0.026783017441630363,
0.206032395362854,
0.1980147808790207,
0.014206883497536182,
0.2175983190536499,
0.03177616000175476,
-0.03772832080721855,
-0.1300560086965561,
-0.065880686044693,
-0.006372632458806038,
0.03559038043022156,
0.08070417493581772,
-0.18207235634326935,
-0.015011128038167953,
-0.05689644813537598,
-0.034518610686063766,
-0.15059494972229004,
-0.28553900122642517,
-0.05957856774330139,
0.20075850188732147,
0.14706264436244965,
0.27519428730010986,
-0.10432573407888412,
0.035197313874959946,
0.02663275972008705,
-0.04912831634283066,
-0.006501141935586929,
0.00018665487004909664,
0.10268618166446686,
-0.15421873331069946,
0.1176437959074974,
0.08486983180046082,
-0.019002694636583328,
0.01058861706405878,
-0.1619086116552353,
0.00936629343777895,
-0.12191236019134521,
0.05354422330856323,
0.1400289237499237,
-0.048128653317689896,
-0.054873593151569366,
0.14033560454845428,
-0.024562934413552284,
-0.22685599327087402,
-0.04648222774267197,
-0.043600670993328094,
-0.010640020482242107,
0.026607351377606392,
-0.1013401448726654,
0.04101909324526787,
0.1330099105834961,
0.009380043484270573,
0.1147187277674675,
0.11749245226383209,
-0.052566803991794586,
0.10792597383260727,
0.2257719188928604,
-0.018785694614052773,
0.04689010605216026,
-0.12743118405342102,
-0.0012336712097749114,
-0.028270328417420387,
0.013657891191542149,
-0.09504974633455276,
-0.09938385337591171,
0.02366873063147068,
0.02872389927506447,
0.009118586778640747,
0.0921793207526207,
-0.029922157526016235,
0.0759170651435852,
0.06817561388015747,
-0.13014446198940277,
-0.16288450360298157,
0.015828335657715797,
-0.007344507612287998,
0.08354310691356659,
0.00027861111448146403,
0.08878035843372345,
-0.11932205408811569,
-0.018093237653374672,
-0.03153328225016594,
-0.03319635987281799,
-0.130486860871315,
-0.07138993591070175,
0.06156524643301964,
0.028095467016100883,
-0.06602972000837326,
0.1398407518863678,
0.026440169662237167,
0.15942534804344177,
0.049197953194379807,
0.012499804608523846,
0.07227300107479095,
-0.05345509201288223,
0.1283530443906784,
0.13818155229091644,
-0.00868943240493536,
-0.05460423603653908,
-0.1013643890619278,
-0.10236792266368866,
0.08925779908895493,
-0.05773641914129257,
0.07476430386304855,
-0.14885357022285461,
-0.06675903499126434,
0.015772046521306038,
0.016141414642333984,
-0.09562095999717712,
0.02571965754032135,
-0.01625603251159191,
-0.18119946122169495,
0.056570518761873245,
-0.048285093158483505,
0.0440407395362854,
-0.06347788125276566,
-0.1110161691904068,
-0.17226378619670868,
0.06091433763504028,
0.08593481779098511,
-0.053876690566539764,
-0.12229149043560028,
0.011023230850696564,
-0.00012518465518951416,
-0.06341652572154999,
-0.05023367330431938,
0.09722746908664703,
-0.11020902544260025,
0.031452205032110214,
-0.012567701749503613,
0.08853451162576675,
-0.03510405123233795,
-0.011538895778357983,
0.044220831245183945,
-0.08039166033267975,
-0.009481523185968399,
0.03534642979502678,
-0.026372017338871956,
-0.04127239063382149,
-0.2689029574394226,
0.0036654395516961813,
0.0341104120016098,
0.02497158572077751,
0.07856601476669312,
0.011906822212040424,
0.021174922585487366,
0.03993808850646019,
-0.15396519005298615,
-0.013395369984209538,
0.14574195444583893,
-0.07689505815505981,
-0.022186370566487312,
0.05703273415565491,
-0.09054436534643173,
0.013882770203053951,
-0.030287226662039757,
0.1345842480659485,
0.023923413828015327,
0.06404478847980499,
-0.0851147472858429,
0.10106813907623291,
-0.1451139897108078,
-0.04998219385743141,
-0.01244612317532301,
0.09761348366737366,
0.07019034773111343,
-0.10272270441055298,
0.014697125181555748,
0.04210108891129494,
0.19416837394237518,
0.016384804621338844,
-0.0356343574821949,
-0.03396720811724663,
0.004015897400677204,
0.22076453268527985,
0.03044266067445278,
0.10457023978233337,
0.07281364500522614,
-0.026583973318338394,
0.12624378502368927,
0.09929762035608292,
0.11280370503664017,
-0.055645186454057693,
0.13904185593128204,
0.04667386785149574,
0.038641396909952164,
0.0614289753139019,
0.06836545467376709,
0.09098632633686066,
-0.0008288522367365658,
0.1138714924454689,
0.013811973854899406,
-0.02422109805047512,
-0.021335409954190254,
0.17759373784065247,
0.10501719266176224,
-0.14769648015499115,
0.029047364369034767,
-0.01258957851678133,
0.039933037012815475,
-0.014194529503583908,
-0.15634691715240479,
-0.07240267097949982,
-0.3315149247646332,
0.1226184144616127,
-0.07119352370500565,
0.019930170848965645,
0.007913772016763687,
-0.037425633519887924,
-0.03296699747443199,
-0.04477746784687042,
0.13151589035987854,
-0.013641550205647945,
-0.006079165264964104,
-0.04815853759646416,
-0.015360191464424133,
-0.11607866734266281,
-0.11200575530529022,
-0.013207737356424332,
-0.13671602308750153,
-0.010119039565324783,
0.05595948174595833,
0.003977729007601738,
0.01821410097181797,
-0.03142618387937546,
0.0024383175186812878,
0.06541839241981506,
-0.05751744285225868,
0.056182678788900375,
0.12097269296646118,
0.08766137808561325,
-0.1058853268623352,
0.031048951670527458,
0.2011747509241104,
0.04359564557671547,
-0.12483977526426315,
0.01449228823184967,
0.1819491684436798,
0.004885740112513304,
0.017068125307559967,
-0.006097703706473112,
-0.0540788508951664,
-0.07554277032613754,
0.1251034289598465,
0.08296554535627365,
-0.09985227137804031,
0.015833314508199692,
-0.0726347416639328,
-0.01594804972410202,
-0.06374675035476685,
0.10130585730075836,
0.09538925439119339,
0.04440245032310486,
-0.10621760785579681,
-0.08487539738416672,
-0.10891728103160858,
0.040588874369859695,
-0.08629853278398514,
-0.07311757653951645,
0.09629398584365845,
-0.07057105004787445,
-0.07029950618743896,
0.025521177798509598,
-0.17978744208812714,
-0.009467960335314274,
0.1711762249469757,
-0.24654000997543335,
-0.0916430801153183,
-0.10857923328876495,
0.14477859437465668,
0.016497576609253883,
0.1013975441455841,
-0.006207061931490898,
-0.007889035157859325,
-0.20577777922153473,
0.024890204891562462,
-0.05293011665344238,
-0.02073732763528824,
0.07814782857894897,
-0.09476397186517715,
0.22629831731319427,
-0.08276885002851486,
0.020940175279974937,
0.012659613974392414,
0.0870661810040474,
-0.030675338581204414,
0.09283176809549332,
-0.03660329803824425,
-0.12576518952846527,
-0.03620953485369682,
0.03001813031733036,
0.013904244638979435,
0.10071761906147003,
0.09772487729787827,
-0.03414725139737129,
0.03389119729399681,
0.09747414290904999,
0.04172342270612717,
-0.023843804374337196,
0.0360250361263752,
-0.17077107727527618,
0.02182629331946373,
-0.018498148769140244,
-0.06935930997133255,
0.03687669709324837,
-0.06603235751390457,
0.1639697551727295,
0.04022442549467087,
0.0670473501086235,
-0.036152735352516174,
0.0073931049555540085,
-0.014454689808189869,
-0.013775371946394444,
-0.026180334389209747,
-0.17259705066680908,
-0.10422050207853317,
-0.1347656100988388,
-0.012701659463346004,
-0.034971047192811966,
0.04591470584273338,
0.023234914988279343,
-0.0003200018545612693,
-0.014577031135559082,
-0.12090865522623062,
0.04360328987240791,
0.11146783083677292,
-0.04631396010518074,
-0.026193076744675636
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | manche/gpt2-safeguard-sg1 | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T17:18:03+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #gpt2 #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #gpt2 #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
57,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #gpt2 #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.05622259899973869,
0.16002345085144043,
-0.004987028427422047,
0.023115945979952812,
0.0962471067905426,
0.011845538392663002,
0.06785304099321365,
0.11496778577566147,
-0.020396295934915543,
0.11142492294311523,
0.03292480856180191,
0.0972127765417099,
0.11474913358688354,
0.16215258836746216,
0.004439093638211489,
-0.23455148935317993,
0.04782992601394653,
-0.12695099413394928,
-0.033447545021772385,
0.11785799264907837,
0.14491069316864014,
-0.10402194410562515,
0.07766910642385483,
-0.030544815585017204,
-0.009361269883811474,
-0.03290390968322754,
-0.06365230679512024,
-0.05152205005288124,
0.05037128925323486,
0.06932847946882248,
0.06591591984033585,
0.007509593386203051,
0.09122733771800995,
-0.2655104100704193,
0.02280162274837494,
0.07630051672458649,
-0.0015554219717159867,
0.07497020810842514,
0.048351652920246124,
-0.08209776133298874,
0.0788840726017952,
-0.05696587264537811,
0.14718368649482727,
0.08216129243373871,
-0.08924587815999985,
-0.1965435892343521,
-0.08464295417070389,
0.10284840315580368,
0.18357418477535248,
0.05158785358071327,
-0.024141347035765648,
0.10476154088973999,
-0.08419200032949448,
0.008797040209174156,
0.06024181470274925,
-0.06443428993225098,
-0.05412506312131882,
0.06934051215648651,
0.07975570857524872,
0.07967228442430496,
-0.13025140762329102,
-0.014651902951300144,
0.011243549175560474,
0.007594773545861244,
0.08504551649093628,
0.022028017789125443,
0.14595499634742737,
0.04393624886870384,
-0.13030564785003662,
-0.044304780662059784,
0.09771761298179626,
0.04345165938138962,
-0.053857799619436264,
-0.2537047266960144,
-0.024983759969472885,
-0.03927002474665642,
-0.03094942681491375,
-0.038562554866075516,
0.04431856796145439,
-0.011080716736614704,
0.08032315224409103,
-0.01118796318769455,
-0.08149448037147522,
-0.041395120322704315,
0.06544242054224014,
0.062143467366695404,
0.026896316558122635,
-0.01158317644149065,
0.00973866879940033,
0.1224486380815506,
0.10907839238643646,
-0.12763150036334991,
-0.05768941715359688,
-0.06755511462688446,
-0.08307720720767975,
-0.04300352931022644,
0.03337155282497406,
0.044020529836416245,
0.04436098039150238,
0.2466370165348053,
0.01108562108129263,
0.05453123152256012,
0.045806169509887695,
0.010608446784317493,
0.06787561625242233,
0.11606968939304352,
-0.062306761741638184,
-0.09178462624549866,
-0.029058339074254036,
0.09215214103460312,
0.006741520017385483,
-0.042814407497644424,
-0.060904473066329956,
0.06479041278362274,
0.012608112767338753,
0.12110785394906998,
0.08444269746541977,
0.0026690615341067314,
-0.07305197417736053,
-0.06963318586349487,
0.18848419189453125,
-0.1598394364118576,
0.047875016927719116,
0.031182926148176193,
-0.038971830159425735,
-0.0014042917173355818,
0.008752269670367241,
0.02394084818661213,
-0.020246321335434914,
0.08923295140266418,
-0.05574449151754379,
-0.03784004598855972,
-0.11079790443181992,
-0.03252100944519043,
0.030985163524746895,
0.0051483530551195145,
-0.027043871581554413,
-0.033837489783763885,
-0.09040277451276779,
-0.059588029980659485,
0.0922931432723999,
-0.07471107691526413,
-0.04984431713819504,
-0.013726521283388138,
-0.07691634446382523,
0.023329194635152817,
0.016799474135041237,
0.08357251435518265,
-0.02157396264374256,
0.0384126678109169,
-0.0560205839574337,
0.0631464347243309,
0.11269522458314896,
0.029363946989178658,
-0.053069718182086945,
0.05750001594424248,
-0.24315528571605682,
0.10326608270406723,
-0.07320205867290497,
0.050549428910017014,
-0.15059062838554382,
-0.026000602170825005,
0.044471126049757004,
0.00805877335369587,
-0.013138634152710438,
0.14088952541351318,
-0.21621745824813843,
-0.0323486253619194,
0.16741067171096802,
-0.0939871072769165,
-0.07602590322494507,
0.059108685702085495,
-0.05233629792928696,
0.10869261622428894,
0.04351044446229935,
-0.02232111617922783,
0.060673557221889496,
-0.14475463330745697,
-0.01067100279033184,
-0.04139741137623787,
-0.02402937039732933,
0.16397778689861298,
0.07567544281482697,
-0.06286642700433731,
0.08052356541156769,
0.024165838956832886,
-0.017831770703196526,
-0.04484899342060089,
-0.023361295461654663,
-0.10819391161203384,
0.009856974706053734,
-0.06032416597008705,
0.02424289658665657,
-0.025761527940630913,
-0.09367526322603226,
-0.02868773601949215,
-0.1802000105381012,
-0.009223134256899357,
0.0881323292851448,
-0.011722641065716743,
-0.021903391927480698,
-0.12039245665073395,
0.011948852799832821,
0.031212422996759415,
0.002984174294397235,
-0.13029038906097412,
-0.05838731303811073,
0.027675874531269073,
-0.16422230005264282,
0.03272955119609833,
-0.05597274377942085,
0.05056252330541611,
0.03445037454366684,
-0.03187771514058113,
-0.033117350190877914,
0.009550533257424831,
0.006354342680424452,
-0.010578392073512077,
-0.2502359449863434,
-0.02440580166876316,
-0.0219739843159914,
0.17386503517627716,
-0.21793730556964874,
0.04213962331414223,
0.07686693966388702,
0.14929872751235962,
0.006240781396627426,
-0.038500864058732986,
0.010139784775674343,
-0.08222103863954544,
-0.030560437589883804,
-0.0643099993467331,
-0.012082485482096672,
-0.03717579320073128,
-0.05608142167329788,
0.05165567249059677,
-0.16133594512939453,
-0.028727244585752487,
0.1057019829750061,
0.06860516220331192,
-0.14001330733299255,
-0.019125886261463165,
-0.04171464592218399,
-0.043496038764715195,
-0.05877087265253067,
-0.0552728995680809,
0.1185101792216301,
0.05596614256501198,
0.04696191847324371,
-0.06956122815608978,
-0.07775315642356873,
0.007865429855883121,
-0.017090093344449997,
-0.017978519201278687,
0.08920905739068985,
0.07311701774597168,
-0.12023317068815231,
0.09247473627328873,
0.10194233059883118,
0.09365488588809967,
0.108615942299366,
-0.017981963232159615,
-0.08929306268692017,
-0.04584396257996559,
0.02045595459640026,
0.013332244008779526,
0.14797501266002655,
-0.01403066236525774,
0.056954506784677505,
0.03922648727893829,
-0.01123172789812088,
0.012020308524370193,
-0.09384570270776749,
0.027314940467476845,
0.034342724829912186,
-0.020308034494519234,
0.03796098753809929,
-0.04001156985759735,
0.019826533272862434,
0.08712323755025864,
0.04676510766148567,
0.04415108636021614,
0.011758276261389256,
-0.04233846068382263,
-0.10904491692781448,
0.173858180642128,
-0.12615609169006348,
-0.24583272635936737,
-0.14115718007087708,
0.0015609683468937874,
0.04152948409318924,
-0.009671499952673912,
0.003867273684591055,
-0.07054664939641953,
-0.11710625886917114,
-0.0934595838189125,
0.018713686615228653,
0.04491026699542999,
-0.07426843047142029,
-0.0596279613673687,
0.059872306883335114,
0.03894329443573952,
-0.14430272579193115,
0.022237464785575867,
0.047419775277376175,
-0.09032250195741653,
-0.006925572175532579,
0.08398029953241348,
0.06729988008737564,
0.17764869332313538,
0.009659109637141228,
-0.021044570952653885,
0.03080335259437561,
0.21258224546909332,
-0.14283664524555206,
0.11252175271511078,
0.14021345973014832,
-0.09024007618427277,
0.08099348843097687,
0.1948828399181366,
0.039186809211969376,
-0.10478170961141586,
0.03259138762950897,
0.02489176020026207,
-0.028939135372638702,
-0.25018003582954407,
-0.0680207833647728,
0.002590036718174815,
-0.04892077296972275,
0.07092583924531937,
0.0918794497847557,
0.09946957975625992,
0.015428726561367512,
-0.09732488542795181,
-0.08017807453870773,
0.0468163788318634,
0.10640767961740494,
0.0070237633772194386,
-0.01532268337905407,
0.08905128389596939,
-0.03260866180062294,
0.018378758803009987,
0.0954233929514885,
0.00412675691768527,
0.17459604144096375,
0.05586163327097893,
0.17767499387264252,
0.07751350849866867,
0.06634163856506348,
0.019167855381965637,
0.0069374511949718,
0.02067388966679573,
0.017508454620838165,
-0.004214957356452942,
-0.08522020280361176,
-0.00457410141825676,
0.12029227614402771,
0.06321834027767181,
0.024303704500198364,
0.0137604009360075,
-0.03941800817847252,
0.08438141644001007,
0.17332784831523895,
0.0020201504230499268,
-0.18486954271793365,
-0.07240456342697144,
0.07921045273542404,
-0.0910051167011261,
-0.10552998632192612,
-0.03353073075413704,
0.03346012532711029,
-0.1747758537530899,
0.02097497321665287,
-0.017018353566527367,
0.10809773951768875,
-0.13855572044849396,
-0.018670624122023582,
0.06328251957893372,
0.07232730835676193,
-0.0028869258239865303,
0.06308864802122116,
-0.153975248336792,
0.1050168052315712,
0.016289174556732178,
0.06754438579082489,
-0.09747608006000519,
0.10138221830129623,
-0.006303760688751936,
-0.007241528946906328,
0.13875643908977509,
0.010596190579235554,
-0.05694379657506943,
-0.08987913280725479,
-0.10555228590965271,
-0.008462639525532722,
0.12933635711669922,
-0.15157614648342133,
0.0847775787115097,
-0.028662750497460365,
-0.043171048164367676,
0.0024383023846894503,
-0.1199452206492424,
-0.1302652359008789,
-0.1875755488872528,
0.058235347270965576,
-0.1366453617811203,
0.039557021111249924,
-0.10582595318555832,
-0.04340389743447304,
-0.028466427698731422,
0.2041483372449875,
-0.2317875325679779,
-0.0682469978928566,
-0.1541893482208252,
-0.08429346233606339,
0.14446710050106049,
-0.04730919376015663,
0.08914490789175034,
-0.0013825427740812302,
0.19013537466526031,
0.024473950266838074,
-0.02387205697596073,
0.10308998823165894,
-0.09543927758932114,
-0.19450686872005463,
-0.08603953570127487,
0.15582145750522614,
0.13931062817573547,
0.03702725097537041,
-0.004593946039676666,
0.029260434210300446,
-0.020000332966446877,
-0.12535293400287628,
0.025526588782668114,
0.1793687790632248,
0.07859015464782715,
0.023437971249222755,
-0.025896867737174034,
-0.10993997752666473,
-0.06524094194173813,
-0.0335373692214489,
0.02718053013086319,
0.18264614045619965,
-0.07421271502971649,
0.1900695115327835,
0.13626199960708618,
-0.05445687845349312,
-0.1955246478319168,
0.018216576427221298,
0.040417760610580444,
0.010847307741641998,
0.03138056397438049,
-0.2078717201948166,
0.09027513861656189,
0.0014845491386950016,
-0.05172133818268776,
0.141556978225708,
-0.174949511885643,
-0.1512570083141327,
0.06491631269454956,
0.0364508256316185,
-0.19348180294036865,
-0.117862768471241,
-0.08817066252231598,
-0.046907443553209305,
-0.17498233914375305,
0.10519181191921234,
0.016932250931859016,
0.009516867808997631,
0.03492651879787445,
0.02640140987932682,
0.011080757714807987,
-0.03873949125409126,
0.19461296498775482,
-0.02505207620561123,
0.029532426968216896,
-0.08079101145267487,
-0.06136554479598999,
0.0607450045645237,
-0.05577658861875534,
0.07896649837493896,
-0.020188091322779655,
0.012835816480219364,
-0.1100873053073883,
-0.0468425452709198,
-0.027396185323596,
0.017321845516562462,
-0.09195652604103088,
-0.09473495930433273,
-0.05146971344947815,
0.09373841434717178,
0.08845265954732895,
-0.036603908985853195,
-0.04043547809123993,
-0.07348548620939255,
0.0325477197766304,
0.17183002829551697,
0.17659065127372742,
0.038550034165382385,
-0.08084331452846527,
-0.005880105309188366,
-0.01188716571778059,
0.04436201974749565,
-0.22519725561141968,
0.06208868324756622,
0.04557957127690315,
0.015879612416028976,
0.11362850666046143,
-0.018783990293741226,
-0.16298477351665497,
-0.06594224274158478,
0.06143777072429657,
-0.06664001196622849,
-0.18599680066108704,
0.0032026967965066433,
0.058006007224321365,
-0.1646854728460312,
-0.037671029567718506,
0.042260222136974335,
-0.0045668939128518105,
-0.04300284758210182,
0.01627597212791443,
0.08071378618478775,
0.005054219625890255,
0.07112491130828857,
0.05733523517847061,
0.0842885971069336,
-0.10417009145021439,
0.07519911974668503,
0.08007751405239105,
-0.08229218423366547,
0.031453702598810196,
0.08910130709409714,
-0.061817802488803864,
-0.03069761022925377,
0.032593827694654465,
0.07753410935401917,
0.019773589447140694,
-0.041717879474163055,
0.008655321784317493,
-0.09745000302791595,
0.06339588761329651,
0.09504765272140503,
0.03549657016992569,
0.014742289669811726,
0.034356739372015,
0.04988397657871246,
-0.07460241764783859,
0.11766603589057922,
0.022336218506097794,
0.01780087500810623,
-0.044981084764003754,
-0.05459042266011238,
0.032110098749399185,
-0.022974027320742607,
-0.010163158178329468,
-0.03885438293218613,
-0.07015778869390488,
-0.018130742013454437,
-0.15929651260375977,
-0.014899281784892082,
-0.04085385054349899,
0.007158880587667227,
0.02551902085542679,
-0.03834335505962372,
0.007963370531797409,
0.012195355258882046,
-0.07085035741329193,
-0.061454467475414276,
-0.022903166711330414,
0.09224231541156769,
-0.16436699032783508,
0.025155464187264442,
0.08285263180732727,
-0.12099926173686981,
0.09775067120790482,
0.021939631551504135,
0.0031351554207503796,
0.028338242322206497,
-0.1542527824640274,
0.04096807911992073,
-0.024365095421671867,
0.01272035762667656,
0.04409142583608627,
-0.22033950686454773,
0.001463581225834787,
-0.03818526118993759,
-0.05954346805810928,
-0.010227864608168602,
-0.033079732209444046,
-0.11291328817605972,
0.09883669763803482,
0.008058897219598293,
-0.08219768106937408,
-0.030809206888079643,
0.03451729565858841,
0.08243680745363235,
-0.02608415111899376,
0.15152283012866974,
0.0016822130419313908,
0.07172226905822754,
-0.17519205808639526,
-0.021702464669942856,
-0.011611736379563808,
0.02207101881504059,
-0.014536668546497822,
-0.015496513806283474,
0.042471300810575485,
-0.02421419881284237,
0.19108575582504272,
-0.026401294395327568,
0.038726791739463806,
0.06405707448720932,
0.01593620702624321,
-0.014801506884396076,
0.10957890748977661,
0.05975057929754257,
0.02399693801999092,
0.022115202620625496,
0.007329683285206556,
-0.039842452853918076,
-0.014149460941553116,
-0.19538825750350952,
0.06474217027425766,
0.1377464383840561,
0.08781574666500092,
-0.01322576031088829,
0.07683692127466202,
-0.10024392604827881,
-0.12397097796201706,
0.11215250939130783,
-0.06283260136842728,
-0.007701667957007885,
-0.06531554460525513,
0.13346771895885468,
0.14944057166576385,
-0.18992236256599426,
0.06835456937551498,
-0.06228158622980118,
-0.05332518368959427,
-0.11744599789381027,
-0.1957325041294098,
-0.055616896599531174,
-0.056456826627254486,
-0.014700124971568584,
-0.048795297741889954,
0.07307228446006775,
0.05693497136235237,
0.012962869368493557,
0.003600025549530983,
0.0766802653670311,
-0.015357231721282005,
0.0008028073934838176,
0.03077360987663269,
0.06600049883127213,
0.013312965631484985,
-0.02929985709488392,
0.020537450909614563,
-0.007275243755429983,
0.04005419462919235,
0.06378308683633804,
0.038119763135910034,
-0.02801438421010971,
0.01591232419013977,
-0.03770609200000763,
-0.10940317064523697,
0.0409080907702446,
-0.028551526367664337,
-0.08112191408872604,
0.13721226155757904,
0.02428387477993965,
0.005870606284588575,
-0.02180131897330284,
0.24582624435424805,
-0.07231455296278,
-0.09001907706260681,
-0.1473579704761505,
0.10211005061864853,
-0.04095151647925377,
0.06560079753398895,
0.04110138490796089,
-0.10732010751962662,
0.013498948886990547,
0.12688814103603363,
0.15896959602832794,
-0.044884394854307175,
0.020156091079115868,
0.03252736106514931,
0.003683826420456171,
-0.04006262496113777,
0.05253688618540764,
0.0694650411605835,
0.14883354306221008,
-0.04907030612230301,
0.08928520232439041,
0.005485867150127888,
-0.10256236046552658,
-0.03822692111134529,
0.11808354407548904,
-0.017866896465420723,
0.018703164532780647,
-0.057248231023550034,
0.11889533698558807,
-0.059861693531274796,
-0.23005777597427368,
0.06317704170942307,
-0.0720362737774849,
-0.14286935329437256,
-0.021647587418556213,
0.07456772774457932,
-0.017636949196457863,
0.02658887766301632,
0.07326807081699371,
-0.07681973278522491,
0.19899281859397888,
0.038975972682237625,
-0.05729197710752487,
-0.05658522993326187,
0.0789351835846901,
-0.114089734852314,
0.2792985737323761,
0.01164181251078844,
0.04984506592154503,
0.10365619510412216,
-0.016686614602804184,
-0.13768579065799713,
0.015234606340527534,
0.09244892746210098,
-0.09004336595535278,
0.03869183734059334,
0.2132277488708496,
-0.002569539239630103,
0.1152428612112999,
0.07714667171239853,
-0.07265080511569977,
0.04592108353972435,
-0.1130065843462944,
-0.0718315914273262,
-0.086885966360569,
0.09441597014665604,
-0.07240451127290726,
0.14123490452766418,
0.12318195402622223,
-0.053516924381256104,
0.010368985123932362,
-0.031209774315357208,
0.04651070013642311,
0.007842876948416233,
0.10365527868270874,
0.010769560933113098,
-0.18099099397659302,
0.022656621411442757,
0.018202748149633408,
0.10856854915618896,
-0.17241089046001434,
-0.09672945737838745,
0.04725200682878494,
0.001958663808181882,
-0.059874359518289566,
0.1282012164592743,
0.057909298688173294,
0.04923510178923607,
-0.043742597103118896,
-0.017267800867557526,
-0.009560109116137028,
0.13584671914577484,
-0.10737434774637222,
-0.0021453071385622025
] |
null | null | fastai |
# Amazing!
🥳 Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the [documentation here](https://huggingface.co/docs/hub/model-repos))!
2. Create a demo in Gradio or Streamlit using 🤗 Spaces ([documentation here](https://huggingface.co/docs/hub/spaces)).
3. Join the fastai community on the [Fastai Discord](https://discord.com/invite/YKrxeNn)!
Greetings fellow fastlearner 🤝! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| {"tags": ["fastai"]} | null | crrodrvi/intel-image-classification | [
"fastai",
"has_space",
"region:us"
] | 2024-02-07T17:20:34+00:00 | [] | [] | TAGS
#fastai #has_space #region-us
|
# Amazing!
Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the documentation here)!
2. Create a demo in Gradio or Streamlit using Spaces (documentation here).
3. Join the fastai community on the Fastai Discord!
Greetings fellow fastlearner ! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| [
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
"TAGS\n#fastai #has_space #region-us \n",
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
13,
20,
79,
3,
6,
12,
8
] | [
"passage: TAGS\n#fastai #has_space #region-us \n# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---# Model card## Model description\nMore information needed## Intended uses & limitations\nMore information needed## Training and evaluation data\nMore information needed"
] | [
-0.048121724277734756,
-0.024616125971078873,
0.002038001548498869,
0.10439170897006989,
0.135872021317482,
0.11887997388839722,
0.07405775785446167,
0.09980081021785736,
0.07783667743206024,
0.02590852417051792,
0.08961158245801926,
-0.08088712394237518,
0.08744348585605621,
0.271692156791687,
0.06988707184791565,
-0.22761479020118713,
0.04051019623875618,
-0.00024903909070417285,
0.08053462207317352,
0.06629016250371933,
0.13507555425167084,
-0.05464952811598778,
0.14010503888130188,
-0.004088983871042728,
-0.19050447642803192,
-0.042929794639348984,
-0.01773718371987343,
-0.02527874894440174,
0.12317648530006409,
-0.04744937643408775,
0.05381017178297043,
0.015037551522254944,
0.007565062493085861,
-0.07253646105527878,
0.0623294934630394,
0.040457066148519516,
0.01740180514752865,
0.059235580265522,
-0.07249044626951218,
0.08950132131576538,
0.08404164761304855,
-0.024370938539505005,
-0.1097978875041008,
0.07827875018119812,
-0.14424212276935577,
-0.21762843430042267,
-0.1253085881471634,
-0.09017651528120041,
0.028519365936517715,
0.004388005938380957,
-0.025051530450582504,
0.12801909446716309,
-0.13558274507522583,
-0.040698226541280746,
0.20124278962612152,
-0.17012301087379456,
-0.05505548417568207,
0.034343402832746506,
0.09226689487695694,
-0.05829555168747902,
-0.06347129493951797,
0.10614984482526779,
0.09640881419181824,
-0.019833475351333618,
0.05516824126243591,
0.002579754451289773,
0.021173657849431038,
0.01370104867964983,
-0.06150497496128082,
0.04717832803726196,
-0.010183089412748814,
0.048132527619600296,
-0.09465572983026505,
-0.1303568333387375,
-0.004072192590683699,
0.01214400865137577,
-0.048744890838861465,
-0.07019646465778351,
0.07833103090524673,
-0.011118141002953053,
-0.04357248544692993,
-0.13031910359859467,
-0.09131011366844177,
-0.12358787655830383,
0.008646543137729168,
0.09500427544116974,
0.003679296001791954,
0.07374339550733566,
-0.08258994668722153,
0.06774985045194626,
-0.17329485714435577,
-0.06484591960906982,
-0.08138520270586014,
-0.11546400189399719,
0.021133482456207275,
-0.0387684591114521,
0.02668963186442852,
0.15394504368305206,
0.12983950972557068,
0.023976242169737816,
0.04388163983821869,
-0.038937073200941086,
0.051190316677093506,
0.058571770787239075,
0.03395717963576317,
0.034934818744659424,
-0.036981891840696335,
-0.1793210655450821,
-0.016702448949217796,
-0.011550825089216232,
0.07954040914773941,
-0.07523109763860703,
-0.05632320046424866,
0.013454885222017765,
-0.11071494966745377,
0.07202339172363281,
-0.03576776012778282,
-0.0032025426626205444,
0.01168301422148943,
0.018371861428022385,
0.21271461248397827,
0.03955606371164322,
0.014191740192472935,
-0.008875265717506409,
-0.13453757762908936,
-0.06874168664216995,
-0.06896194815635681,
0.03361047804355621,
0.04448792710900307,
-0.0028071461711078882,
-0.07672245055437088,
0.04325154796242714,
-0.06045534089207649,
-0.03508453071117401,
0.008032378740608692,
-0.18221288919448853,
0.007458044681698084,
-0.10049355030059814,
-0.12126200646162033,
0.05306628718972206,
0.01695440337061882,
-0.08215925842523575,
0.08141279965639114,
0.02662261202931404,
0.020931517705321312,
-0.009988143108785152,
-0.005391082260757685,
0.06874798238277435,
-0.08508864045143127,
0.029901226982474327,
0.17170792818069458,
0.13024519383907318,
-0.08046911656856537,
-0.0006887061172164977,
-0.10965746641159058,
0.04426072910428047,
-0.13325683772563934,
0.02251482754945755,
-0.09062390774488449,
0.11723794043064117,
-0.042396437376737595,
0.002038756385445595,
-0.029030200093984604,
0.0960269495844841,
0.08189879357814789,
0.16663365066051483,
-0.2419009804725647,
-0.031095001846551895,
0.13240347802639008,
-0.10711425542831421,
-0.1807439625263214,
0.18486657738685608,
-0.012035200372338295,
0.11329247802495956,
-0.047014184296131134,
0.18334640562534332,
-0.02612062357366085,
-0.13582459092140198,
-0.058872904628515244,
0.005852419883012772,
-0.2269321084022522,
-0.06286033242940903,
0.09738040715456009,
0.13425657153129578,
-0.042984943836927414,
0.007112155202776194,
0.026316028088331223,
0.13609857857227325,
-0.06715573370456696,
-0.05195777863264084,
-0.012255736626684666,
-0.10902371257543564,
0.041914235800504684,
0.018215661868453026,
0.035408079624176025,
-0.059880174696445465,
-0.02931194379925728,
-0.053190283477306366,
0.13146710395812988,
0.09760832786560059,
-0.03670211136341095,
-0.049620725214481354,
0.1689043790102005,
-0.07763876020908356,
-0.033587727695703506,
0.07560533285140991,
-0.08268500119447708,
0.03266897425055504,
0.03090597130358219,
0.055881720036268234,
0.07766123116016388,
0.08522116392850876,
0.06057543307542801,
0.00819048099219799,
0.034654274582862854,
0.12095347046852112,
-0.013591280207037926,
-0.05039411783218384,
0.021508218720555305,
0.016904234886169434,
-0.019032588228583336,
0.29030677676200867,
-0.1951042115688324,
0.024724548682570457,
-0.06477324664592743,
0.07631538063287735,
0.06136792525649071,
0.003575638635084033,
0.08580143749713898,
-0.06023019179701805,
-0.019061198458075523,
-0.04803973436355591,
0.046805646270513535,
-0.0666879191994667,
-0.04162997007369995,
0.2621194124221802,
-0.05497581139206886,
0.044914912432432175,
0.12313763797283173,
-0.05873025581240654,
-0.07091446220874786,
0.01009807363152504,
-0.00793424155563116,
0.03249288722872734,
-0.04042816907167435,
0.043721720576286316,
-0.10840129852294922,
-0.06674089282751083,
0.1573198139667511,
-0.038477856665849686,
0.06786153465509415,
0.032288823276758194,
-0.04958454892039299,
-0.0648743286728859,
0.04650486260652542,
0.13598160445690155,
-0.0875244215130806,
0.07435166835784912,
0.17612984776496887,
-0.010562662966549397,
0.168031245470047,
0.08435525000095367,
-0.07075224816799164,
-0.09465329349040985,
-0.051014289259910583,
-0.021595727652311325,
0.21222901344299316,
-0.07084725052118301,
-0.054564714431762695,
0.05911700800061226,
-0.013703816570341587,
0.07196151465177536,
-0.06009222939610481,
-0.08332337439060211,
0.03227344527840614,
-0.04517695680260658,
0.011517706327140331,
0.13512636721134186,
-0.07090822607278824,
0.04681389778852463,
0.031489867717027664,
-0.0662703812122345,
0.02217509225010872,
0.033389873802661896,
0.0068921963684260845,
0.033959709107875824,
0.07332495599985123,
-0.20893315970897675,
-0.08408680558204651,
-0.13727638125419617,
0.037881869822740555,
0.021770721301436424,
0.045787326991558075,
-0.08602345734834671,
0.02231026627123356,
-0.08954031765460968,
-0.07987114042043686,
0.029592275619506836,
-0.026350297033786774,
-0.11349643021821976,
-0.03396226093173027,
-0.009560913778841496,
-0.06662604957818985,
-0.02250705659389496,
-0.05024505779147148,
0.03983384370803833,
0.04479299485683441,
0.058377087116241455,
0.12796473503112793,
-0.013808943331241608,
-0.03839317709207535,
0.000370211957488209,
-0.022712308913469315,
0.16396735608577728,
-0.14746315777301788,
0.07954913377761841,
0.19160102307796478,
0.11742953956127167,
0.028144672513008118,
0.028885571286082268,
0.03537585213780403,
-0.06289814412593842,
-0.000050317394197918475,
0.03226194158196449,
-0.09392514824867249,
-0.05801016092300415,
-0.020014392212033272,
-0.04031052812933922,
0.17134574055671692,
-0.12160717695951462,
0.03345204517245293,
0.04098419472575188,
0.09783966839313507,
0.10073629021644592,
-0.028829937800765038,
-0.1815856397151947,
0.038818612694740295,
-0.24060091376304626,
-0.05831146240234375,
0.027899866923689842,
-0.09110201895236969,
-0.06232144311070442,
0.17409387230873108,
0.013794700615108013,
0.011769929900765419,
-0.006736889015883207,
0.07983319461345673,
0.0110100656747818,
0.1217205822467804,
0.05947643890976906,
-0.05539114400744438,
0.025202350690960884,
-0.09962950646877289,
-0.07107596844434738,
-0.04035590961575508,
-0.05832801014184952,
0.07548832893371582,
0.1409129947423935,
-0.025475580245256424,
-0.020795362070202827,
0.023489827290177345,
0.08550169318914413,
0.0423230417072773,
0.16739299893379211,
-0.16016584634780884,
-0.026555389165878296,
0.04571257904171944,
-0.03384667634963989,
-0.05433850735425949,
-0.010291114449501038,
0.1137225553393364,
-0.02820689231157303,
-0.040318265557289124,
0.021242983639240265,
0.06503437459468842,
0.01481706090271473,
0.05012747645378113,
-0.04056356102228165,
0.14796851575374603,
-0.03461192920804024,
0.019330544397234917,
-0.12413888424634933,
0.13848772644996643,
0.021095896139740944,
-0.03901609033346176,
-0.06735876202583313,
-0.05808034539222717,
0.18150931596755981,
0.0025602965615689754,
0.10535930097103119,
0.012098877690732479,
-0.12160047143697739,
-0.1359938681125641,
-0.11211287975311279,
0.005111907608807087,
0.08330471813678741,
-0.023147236555814743,
-0.022247863933444023,
0.022165266796946526,
-0.036149751394987106,
-0.0530381016433239,
0.15749511122703552,
-0.1289154291152954,
-0.001082550617866218,
0.014728817157447338,
0.06971760839223862,
-0.08223173767328262,
0.026267826557159424,
0.014071501791477203,
-0.1119147390127182,
0.10590848326683044,
0.2521335482597351,
0.10338116437196732,
-0.09591643512248993,
-0.07697287201881409,
0.03418830782175064,
-0.012184361927211285,
-0.000774814048781991,
-0.006932659074664116,
0.0495428591966629,
-0.005566445179283619,
0.006762749515473843,
0.12971895933151245,
-0.07130889594554901,
0.011540771462023258,
-0.08449850976467133,
0.05566910281777382,
-0.05276734381914139,
0.01761564053595066,
-0.002672141883522272,
-0.008124710991978645,
-0.07340748608112335,
-0.061829522252082825,
0.1609770804643631,
-0.07277000695466995,
-0.06468547880649567,
0.05801168829202652,
0.03307786211371422,
0.01431563775986433,
-0.03584568202495575,
-0.04342148080468178,
0.18088261783123016,
0.29330700635910034,
-0.08191116154193878,
0.10001859813928604,
0.09677296131849289,
0.034820813685655594,
-0.23625829815864563,
0.029798466712236404,
-0.1455078274011612,
0.04449721798300743,
0.040447335690259933,
-0.0409548319876194,
0.04191497340798378,
0.10835777968168259,
-0.06094440817832947,
0.2048867791891098,
-0.03527235612273216,
-0.07983248680830002,
-0.01788630709052086,
0.03109324350953102,
0.29443636536598206,
-0.11833466589450836,
0.006058716680854559,
-0.10420958697795868,
-0.21566011011600494,
0.06983078271150589,
-0.18948867917060852,
0.13948246836662292,
-0.05087858438491821,
0.03576415032148361,
-0.01149723306298256,
-0.07561972737312317,
0.20518061518669128,
-0.15641045570373535,
0.05273103713989258,
-0.13722458481788635,
-0.1327189952135086,
0.01617460884153843,
-0.10048147290945053,
0.1545477658510208,
-0.11024226248264313,
-0.023215843364596367,
-0.2284185290336609,
0.012587235309183598,
-0.023200806230306625,
0.10030807554721832,
0.01800704374909401,
-0.07980740070343018,
-0.08767345547676086,
0.1316242516040802,
-0.06486566364765167,
0.034810543060302734,
-0.06996636837720871,
-0.050714004784822464,
-0.010929876938462257,
-0.045061707496643066,
0.03034941293299198,
-0.07934719324111938,
0.15192505717277527,
-0.016938980668783188,
-0.04507075995206833,
0.08636019378900528,
-0.2479533851146698,
0.023727843537926674,
0.025351112708449364,
-0.03495599329471588,
0.09001832455396652,
-0.025513244792819023,
-0.06256973743438721,
0.12282291799783707,
0.1402233988046646,
-0.07322840392589569,
-0.2460673749446869,
-0.06281693279743195,
0.0076784128323197365,
0.039165716618299484,
0.06561196595430374,
0.05125982314348221,
-0.07261458039283752,
-0.011131617240607738,
-0.026896944269537926,
0.030595947057008743,
-0.11692017316818237,
-0.03854857385158539,
0.07790639251470566,
0.017095070332288742,
-0.07846562564373016,
0.07280377298593521,
0.014225782826542854,
-0.021511616185307503,
0.007357571739703417,
0.148970365524292,
0.007519228849560022,
-0.14747941493988037,
-0.06656096875667572,
0.2007484883069992,
-0.01197928935289383,
-0.07260087132453918,
-0.05383119732141495,
-0.008990069851279259,
-0.0476234145462513,
0.05585788935422897,
0.05367223918437958,
-0.013585401698946953,
0.07708586007356644,
0.06263149529695511,
-0.10210110992193222,
-0.046256959438323975,
-0.066561758518219,
0.04169114679098129,
-0.10485753417015076,
0.060470130294561386,
0.009529483504593372,
0.12185006588697433,
-0.09983488917350769,
-0.01802929677069187,
-0.10810204595327377,
-0.06766588985919952,
-0.17349553108215332,
-0.05834362283349037,
-0.041105758398771286,
-0.015651104971766472,
0.03658895567059517,
0.010445823892951012,
-0.057867538183927536,
-0.0442853718996048,
-0.07536603510379791,
0.038444988429546356,
0.06147460639476776,
0.03932281583547592,
-0.03912714496254921,
0.04001858830451965,
0.05909334123134613,
0.013087345287203789,
0.17542624473571777,
0.038768354803323746,
0.05504675209522247,
-0.05045998468995094,
-0.16491834819316864,
-0.05276111513376236,
-0.0074316514655947685,
-0.07559102028608322,
0.1224973127245903,
-0.007679440546780825,
0.007880088873207569,
-0.08065467327833176,
0.03924860805273056,
0.028234204277396202,
0.10404064506292343,
-0.0028364830650389194,
0.10070426017045975,
0.019627176225185394,
-0.07226712256669998,
-0.025392837822437286,
0.021809715777635574,
0.12809939682483673,
0.01567147858440876,
0.026090998202562332,
0.033139873296022415,
0.016619985923171043,
-0.057361043989658356,
0.033977724611759186,
-0.04997231811285019,
-0.15123651921749115,
0.02628709189593792,
-0.05165188014507294,
0.005062380339950323,
-0.016889680176973343,
0.20362506806850433,
0.07867538928985596,
-0.06474173814058304,
-0.010664013214409351,
0.015816617757081985,
-0.0168940220028162,
-0.03121885471045971,
-0.012740966863930225,
0.04592578858137131,
-0.001151384087279439,
-0.04866636544466019,
0.11825273931026459,
0.05015748366713524,
0.05386412516236305,
0.0596686452627182,
0.12528513371944427,
0.016759619116783142,
0.13257254660129547,
0.061999931931495667,
-0.03403807803988457,
-0.13461735844612122,
-0.04495539888739586,
-0.1254577934741974,
0.04646851494908333,
-0.08697032928466797,
0.09941662102937698,
0.1144254133105278,
-0.05959030240774155,
-0.030464433133602142,
-0.08851305395364761,
-0.008356761187314987,
-0.06041252240538597,
0.039516255259513855,
-0.02262675203382969,
-0.0873224213719368,
0.0481097511947155,
0.05495472997426987,
-0.022752324119210243,
0.13218675553798676,
0.015727028250694275,
-0.036317698657512665,
0.13270340859889984,
-0.07583184540271759,
0.11758984625339508,
0.061510033905506134,
-0.043043944984674454,
-0.11560922116041183,
-0.020150646567344666,
-0.06641761213541031,
-0.10098972916603088,
-0.006782987620681524,
-0.005399650428444147,
-0.07349002361297607,
-0.059971679002046585,
0.08397487550973892,
-0.03124053031206131,
-0.09979676455259323,
-0.032152675092220306,
0.0038895104080438614,
0.06054706871509552,
-0.01686914451420307,
-0.0034020058810710907,
0.04728743061423302,
0.015076374635100365,
0.1653461456298828,
-0.02208263985812664,
0.06234867498278618,
-0.13855914771556854,
0.16070103645324707,
-0.14684462547302246,
-0.029404424130916595,
-0.1890171319246292,
-0.09729582816362381,
-0.05156542733311653,
0.20326784253120422,
0.2840938866138458,
-0.19109351933002472,
-0.010187864303588867,
0.020078664645552635,
-0.014484191313385963,
-0.08961770683526993,
0.12571553885936737,
0.029420215636491776,
-0.023631498217582703,
-0.07249019294977188,
-0.02037387527525425,
0.005258576478809118,
-0.06544211506843567,
-0.026979785412549973,
0.18310695886611938,
0.001496660872362554,
0.059546373784542084,
-0.09605178982019424,
0.01754261925816536,
-0.14839904010295868,
-0.10467469692230225,
-0.02111995778977871,
-0.16156397759914398,
-0.09646477550268173,
0.006635562051087618,
0.038640011101961136,
0.08000610023736954,
0.03268849849700928,
-0.015172510407865047,
0.06479045748710632,
-0.056333884596824646,
-0.0037216036580502987,
-0.1231912299990654,
0.00034658415825106204,
0.062129102647304535,
-0.07422006875276566,
0.2545335292816162,
-0.03070417232811451,
-0.12370815873146057,
0.09026903659105301,
-0.03299184888601303,
-0.12452623248100281,
0.07951879501342773,
-0.005700904875993729,
-0.11531132459640503,
-0.057989440858364105,
0.18941475450992584,
-0.012821312062442303,
-0.1364315301179886,
0.046368811279535294,
-0.17166484892368317,
0.031349923461675644,
0.0363016203045845,
-0.001313706859946251,
-0.04714022949337959,
0.024538639932870865,
-0.008008457720279694,
0.10724439471960068,
0.1382838785648346,
0.016739921644330025,
-0.011060068383812904,
-0.05056179314851761,
0.07912429422140121,
0.056927867233753204,
-0.05218246951699257,
-0.1282637119293213,
-0.08599764108657837,
0.03429819270968437,
0.04119478166103363,
-0.08113081753253937,
-0.16903182864189148,
-0.03668912500143051,
-0.10082915425300598,
-0.004939202684909105,
0.051785312592983246,
0.06585265696048737,
0.29044589400291443,
0.06326735019683838,
0.0016605621203780174,
-0.13649453222751617,
0.050569336861371994,
0.0868251696228981,
-0.04697931930422783,
-0.07670357078313828
] |
null | null | nemo |
# Canary 1B
<style>
img {
display: inline;
}
</style>
[](#model-architecture)
| [](#model-architecture)
| [](#datasets)
NVIDIA [NeMo Canary](https://nvidia.github.io/NeMo/blogs/2024/2024-02-canary/) is a family of multi-lingual multi-tasking models that achieves state-of-the art performance on multiple benchmarks. With 1 billion parameters, Canary-1B supports automatic speech-to-text recognition (ASR) in 4 languages (English, German, French, Spanish) and translation from English to German/French/Spanish and from German/French/Spanish to English with or without punctuation and capitalization (PnC).
## Model Architecture
Canary is an encoder-decoder model with FastConformer [1] encoder and Transformer Decoder [2].
With audio features extracted from the encoder, task tokens such as `<source language>`, `<target language>`, `<task>` and `<toggle PnC>`
are fed into the Transformer Decoder to trigger the text generation process. Canary uses a concatenated tokenizer from individual
SentencePiece [3] tokenizers of each language, which makes it easy to scale up to more languages.
The Canay-1B model has 24 encoder layers and 24 layers of decoder layers in total.
## NVIDIA NeMo
To train, fine-tune or play with the model you will need to install [NVIDIA NeMo](https://github.com/NVIDIA/NeMo). We recommend you install it after you've installed Cython and latest PyTorch version.
```
pip install git+https://github.com/NVIDIA/[email protected]#egg=nemo_toolkit[all]
```
## How to Use this Model
The model is available for use in the NeMo toolkit [4], and can be used as a pre-trained checkpoint for inference or for fine-tuning on another dataset.
### Loading the Model
```python
from nemo.collections.asr.models import EncDecMultiTaskModel
# load model
canary_model = EncDecMultiTaskModel.from_pretrained('nvidia/canary-1b')
# update dcode params
decode_cfg = canary_model.cfg.decoding
decode_cfg.beam.beam_size = 1
canary_model.change_decoding_strategy(decode_cfg)
```
### Input Format
The input to the model can be a directory containing audio files, in which case the model will perform ASR on English and produces text with punctuation and capitalization:
```python
predicted_text = canary_model.transcribe(
audio_dir="<path to directory containing audios>",
batch_size=16, # batch size to run the inference with
)
```
or use:
```bash
python [NEMO_GIT_FOLDER]/examples/asr/transcribe_speech.py
pretrained_name="nvidia/canary-1b"
audio_dir="<path to audio directory>"
```
Another recommended option is to use a json manifest as input, where each line in the file is a dictionary containing the following fields:
```yaml
# Example of a line in input_manifest.json
{
"audio_filepath": "/path/to/audio.wav", # path to the audio file
"duration": 1000, # duration of the audio
"taskname": "asr", # use "ast" for speech-to-text translation
"source_lang": "en", # language of the audio input, set `source_lang`==`target_lang` for ASR, choices=['en','de','es','fr']
"target_lang": "en", # language of the text output, choices=['en','de','es','fr']
"pnc": "yes", # whether to have PnC output, choices=['yes', 'no']
"answer": "na",
}
```
and then use:
```python
predicted_text = canary_model.transcribe(
"<path to input manifest file>",
batch_size=16, # batch size to run the inference with
)
```
or use:
```bash
python [NEMO_GIT_FOLDER]/examples/asr/transcribe_speech.py
pretrained_name="nvidia/canary-1b"
dataset_manifest="<path to manifest file>"
```
### Automatic Speech-to-text Recognition (ASR)
An example manifest for transcribing English audios can be:
```yaml
# Example of a line in input_manifest.json
{
"audio_filepath": "/path/to/audio.wav", # path to the audio file
"duration": 1000, # duration of the audio
"taskname": "asr",
"source_lang": "en", # language of the audio input, set `source_lang`==`target_lang` for ASR, choices=['en','de','es','fr']
"target_lang": "en", # language of the text output, choices=['en','de','es','fr']
"pnc": "yes", # whether to have PnC output, choices=['yes', 'no']
"answer": "na",
}
```
### Automatic Speech-to-text Translation (AST)
An example manifest for transcribing English audios into German text can be:
```yaml
# Example of a line in input_manifest.json
{
"audio_filepath": "/path/to/audio.wav", # path to the audio file
"duration": 1000, # duration of the audio
"taskname": "ast",
"source_lang": "en", # language of the audio input, choices=['en','de','es','fr']
"target_lang": "de", # language of the text output, choices=['en','de','es','fr']
"pnc": "yes", # whether to have PnC output, choices=['yes', 'no']
"answer": "na"
}
```
### Input
This model accepts single channel (mono) audio sampled at 16000 Hz, along with the task/languages/PnC tags as input.
### Output
The model outputs the transcribed/translated text corresponding to the input audio, in the specified target language and with or without punctuation and capitalization.
## Training
Canary-1B is trained using the NVIDIA NeMo toolkit [4] for 150k steps with dynamic bucketing and a batch duration of 360s per GPU on 128 NVIDIA A100 80GB GPUs.
The model can be trained using this [example script](https://github.com/NVIDIA/NeMo/blob/main/examples/asr/speech_multitask/speech_to_text_aed.py) and [base config](https://github.com/NVIDIA/NeMo/blob/main/examples/asr/conf/speech_multitask/fast-conformer_aed.yaml).
The tokenizers for these models were built using the text transcripts of the train set with this [script](https://github.com/NVIDIA/NeMo/blob/main/scripts/tokenizers/process_asr_text_tokenizer.py).
### Datasets
The Canary-1B model is trained on a total of 85k hrs of speech data. It consists of 31k hrs of public data, 20k hrs collected by [Suno](https://suno.ai/), and 34k hrs of in-house data.
The constituents of public data are as follows.
#### English (25.5k hours)
- Librispeech 960 hours
- Fisher Corpus
- Switchboard-1 Dataset
- WSJ-0 and WSJ-1
- National Speech Corpus (Part 1, Part 6)
- VCTK
- VoxPopuli (EN)
- Europarl-ASR (EN)
- Multilingual Librispeech (MLS EN) - 2,000 hour subset
- Mozilla Common Voice (v7.0)
- People's Speech - 12,000 hour subset
- Mozilla Common Voice (v11.0) - 1,474 hour subset
#### German (2.5k hours)
- Mozilla Common Voice (v12.0) - 800 hour subset
- Multilingual Librispeech (MLS DE) - 1,500 hour subset
- VoxPopuli (DE) - 200 hr subset
#### Spanish (1.4k hours)
- Mozilla Common Voice (v12.0) - 395 hour subset
- Multilingual Librispeech (MLS ES) - 780 hour subset
- VoxPopuli (ES) - 108 hour subset
- Fisher - 141 hour subset
#### French (1.8k hours)
- Mozilla Common Voice (v12.0) - 708 hour subset
- Multilingual Librispeech (MLS FR) - 926 hour subset
- VoxPopuli (FR) - 165 hour subset
## Performance
In both ASR and AST experiments, predictions were generated using beam search with width 5 and length penalty 1.0.
### ASR Performance (w/o PnC)
The ASR performance is measured with word error rate (WER), and we process the groundtruth and predicted text with [whisper-normalizer](https://pypi.org/project/whisper-normalizer/).
WER on [MCV-16.1](https://commonvoice.mozilla.org/en/datasets) test set:
| **Version** | **Model** | **En** | **De** | **Es** | **Fr** |
|:---------:|:-----------:|:------:|:------:|:------:|:------:|
| 1.23.0 | canary-1b | 7.97 | 4.61 | 3.99 | 6.53 |
WER on [MLS](https://huggingface.co/datasets/facebook/multilingual_librispeech) test set:
| **Version** | **Model** | **En** | **De** | **Es** | **Fr** |
|:---------:|:-----------:|:------:|:------:|:------:|:------:|
| 1.23.0 | canary-1b | 3.06 | 4.19 | 3.15 | 4.12 |
More details on evaluation can be found at [HuggingFace ASR Leaderboard](https://huggingface.co/spaces/hf-audio/open_asr_leaderboard)
### AST Performance
We evaluate AST performance with [BLEU score](https://lightning.ai/docs/torchmetrics/stable/text/sacre_bleu_score.html), and use native annotations with punctuation and capitalization in the datasets.
BLEU score on [FLEURS](https://huggingface.co/datasets/google/fleurs) test set:
| **Version** | **Model** | **En->De** | **En->Es** | **En->Fr** | **De->En** | **Es->En** | **Fr->En** |
|:-----------:|:---------:|:----------:|:----------:|:----------:|:----------:|:----------:|:----------:|
| 1.23.0 | canary-1b | 22.66 | 41.11 | 40.76 | 32.64 | 32.15 | 23.57 |
BLEU score on [COVOST-v2](https://github.com/facebookresearch/covost) test set:
| **Version** | **Model** | **De->En** | **Es->En** | **Fr->En** |
|:-----------:|:---------:|:----------:|:----------:|:----------:|
| 1.23.0 | canary-1b | 37.67 | 40.7 | 40.42 |
BLEU score on [mExpresso](https://huggingface.co/facebook/seamless-expressive#mexpresso-multilingual-expresso) test set:
| **Version** | **Model** | **En->De** | **En->Es** | **En->Fr** |
|:-----------:|:---------:|:----------:|:----------:|:----------:|
| 1.23.0 | canary-1b | 23.84 | 35.74 | 28.29 |
## NVIDIA Riva: Deployment
[NVIDIA Riva](https://developer.nvidia.com/riva), is an accelerated speech AI SDK deployable on-prem, in all clouds, multi-cloud, hybrid, on edge, and embedded.
Additionally, Riva provides:
* World-class out-of-the-box accuracy for the most common languages with model checkpoints trained on proprietary data with hundreds of thousands of GPU-compute hours
* Best in class accuracy with run-time word boosting (e.g., brand and product names) and customization of acoustic model, language model, and inverse text normalization
* Streaming speech recognition, Kubernetes compatible scaling, and enterprise-grade support
Although this model isn’t supported yet by Riva, the [list of supported models](https://huggingface.co/models?other=Riva) is here.
Check out [Riva live demo](https://developer.nvidia.com/riva#demos).
## References
[1] [Fast Conformer with Linearly Scalable Attention for Efficient Speech Recognition](https://arxiv.org/abs/2305.05084)
[2] [Attention is all you need](https://arxiv.org/abs/1706.03762)
[3] [Google Sentencepiece Tokenizer](https://github.com/google/sentencepiece)
[4] [NVIDIA NeMo Toolkit](https://github.com/NVIDIA/NeMo)
## Licence
License to use this model is covered by the [CC-BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/deed.en#:~:text=NonCommercial%20%E2%80%94%20You%20may%20not%20use,doing%20anything%20the%20license%20permits.). By downloading the public and release version of the model, you accept the terms and conditions of the CC-BY-NC-4.0 license. | {"language": ["en", "de", "es", "fr"], "license": "cc-by-nc-4.0", "library_name": "nemo", "tags": ["automatic-speech-recognition", "automatic-speech-translation", "speech", "audio", "Transformer", "FastConformer", "Conformer", "pytorch", "NeMo", "hf-asr-leaderboard"], "datasets": ["librispeech_asr", "fisher_corpus", "Switchboard-1", "WSJ-0", "WSJ-1", "National-Singapore-Corpus-Part-1", "National-Singapore-Corpus-Part-6", "vctk", "voxpopuli", "europarl", "multilingual_librispeech", "mozilla-foundation/common_voice_8_0", "MLCommons/peoples_speech"], "metrics": ["wer", "bleu"], "widget": [{"example_title": "Librispeech sample 1", "src": "https://cdn-media.huggingface.co/speech_samples/sample1.flac"}, {"example_title": "Librispeech sample 2", "src": "https://cdn-media.huggingface.co/speech_samples/sample2.flac"}], "pipeline_tag": "automatic-speech-recognition", "model-index": [{"name": "canary-1b", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "LibriSpeech (other)", "type": "librispeech_asr", "config": "other", "split": "test", "args": {"language": "en"}}, "metrics": [{"type": "wer", "value": 2.89, "name": "Test WER"}]}, {"task": {"type": "Automatic Speech Recognition", "name": "automatic-speech-recognition"}, "dataset": {"name": "SPGI Speech", "type": "kensho/spgispeech", "config": "test", "split": "test", "args": {"language": "en"}}, "metrics": [{"type": "wer", "value": 4.79, "name": "Test WER"}]}, {"task": {"type": "Automatic Speech Recognition", "name": "automatic-speech-recognition"}, "dataset": {"name": "Mozilla Common Voice 16.1", "type": "mozilla-foundation/common_voice_16_1", "config": "en", "split": "test", "args": {"language": "en"}}, "metrics": [{"type": "wer", "value": 7.97, "name": "Test WER (En)"}]}, {"task": {"type": "Automatic Speech Recognition", "name": "automatic-speech-recognition"}, "dataset": {"name": "Mozilla Common Voice 16.1", "type": "mozilla-foundation/common_voice_16_1", "config": "de", "split": "test", "args": {"language": "de"}}, "metrics": [{"type": "wer", "value": 4.61, "name": "Test WER (De)"}]}, {"task": {"type": "Automatic Speech Recognition", "name": "automatic-speech-recognition"}, "dataset": {"name": "Mozilla Common Voice 16.1", "type": "mozilla-foundation/common_voice_16_1", "config": "es", "split": "test", "args": {"language": "es"}}, "metrics": [{"type": "wer", "value": 3.99, "name": "Test WER (ES)"}]}, {"task": {"type": "Automatic Speech Recognition", "name": "automatic-speech-recognition"}, "dataset": {"name": "Mozilla Common Voice 16.1", "type": "mozilla-foundation/common_voice_16_1", "config": "fr", "split": "test", "args": {"language": "fr"}}, "metrics": [{"type": "wer", "value": 6.53, "name": "Test WER (Fr)"}]}, {"task": {"type": "Automatic Speech Translation", "name": "automatic-speech-translation"}, "dataset": {"name": "FLEURS", "type": "google/fleurs", "config": "en_us", "split": "test", "args": {"language": "en-de"}}, "metrics": [{"type": "bleu", "value": 22.66, "name": "Test BLEU (En->De)"}, {"type": "bleu", "value": 41.11, "name": "Test BLEU (En->Es)"}, {"type": "bleu", "value": 40.76, "name": "Test BLEU (En->Fr)"}]}, {"task": {"type": "Automatic Speech Translation", "name": "automatic-speech-translation"}, "dataset": {"name": "FLEURS", "type": "google/fleurs", "config": "de_de", "split": "test", "args": {"language": "de-en"}}, "metrics": [{"type": "bleu", "value": 32.64, "name": "Test BLEU (De->En)"}]}, {"task": {"type": "Automatic Speech Translation", "name": "automatic-speech-translation"}, "dataset": {"name": "FLEURS", "type": "google/fleurs", "config": "es_419", "split": "test", "args": {"language": "es-en"}}, "metrics": [{"type": "bleu", "value": 32.15, "name": "Test BLEU (Es->En)"}]}, {"task": {"type": "Automatic Speech Translation", "name": "automatic-speech-translation"}, "dataset": {"name": "FLEURS", "type": "google/fleurs", "config": "fr_fr", "split": "test", "args": {"language": "fr-en"}}, "metrics": [{"type": "bleu", "value": 23.57, "name": "Test BLEU (Fr->En)"}]}, {"task": {"type": "Automatic Speech Translation", "name": "automatic-speech-translation"}, "dataset": {"name": "COVOST", "type": "covost2", "config": "de_de", "split": "test", "args": {"language": "de-en"}}, "metrics": [{"type": "bleu", "value": 37.67, "name": "Test BLEU (De->En)"}]}, {"task": {"type": "Automatic Speech Translation", "name": "automatic-speech-translation"}, "dataset": {"name": "COVOST", "type": "covost2", "config": "es_419", "split": "test", "args": {"language": "es-en"}}, "metrics": [{"type": "bleu", "value": 40.7, "name": "Test BLEU (Es->En)"}]}, {"task": {"type": "Automatic Speech Translation", "name": "automatic-speech-translation"}, "dataset": {"name": "COVOST", "type": "covost2", "config": "fr_fr", "split": "test", "args": {"language": "fr-en"}}, "metrics": [{"type": "bleu", "value": 40.42, "name": "Test BLEU (Fr->En)"}]}]}]} | automatic-speech-recognition | nvidia/canary-1b | [
"nemo",
"automatic-speech-recognition",
"automatic-speech-translation",
"speech",
"audio",
"Transformer",
"FastConformer",
"Conformer",
"pytorch",
"NeMo",
"hf-asr-leaderboard",
"en",
"de",
"es",
"fr",
"dataset:librispeech_asr",
"dataset:fisher_corpus",
"dataset:Switchboard-1",
"dataset:WSJ-0",
"dataset:WSJ-1",
"dataset:National-Singapore-Corpus-Part-1",
"dataset:National-Singapore-Corpus-Part-6",
"dataset:vctk",
"dataset:voxpopuli",
"dataset:europarl",
"dataset:multilingual_librispeech",
"dataset:mozilla-foundation/common_voice_8_0",
"dataset:MLCommons/peoples_speech",
"arxiv:2305.05084",
"arxiv:1706.03762",
"license:cc-by-nc-4.0",
"model-index",
"has_space",
"region:us"
] | 2024-02-07T17:20:55+00:00 | [
"2305.05084",
"1706.03762"
] | [
"en",
"de",
"es",
"fr"
] | TAGS
#nemo #automatic-speech-recognition #automatic-speech-translation #speech #audio #Transformer #FastConformer #Conformer #pytorch #NeMo #hf-asr-leaderboard #en #de #es #fr #dataset-librispeech_asr #dataset-fisher_corpus #dataset-Switchboard-1 #dataset-WSJ-0 #dataset-WSJ-1 #dataset-National-Singapore-Corpus-Part-1 #dataset-National-Singapore-Corpus-Part-6 #dataset-vctk #dataset-voxpopuli #dataset-europarl #dataset-multilingual_librispeech #dataset-mozilla-foundation/common_voice_8_0 #dataset-MLCommons/peoples_speech #arxiv-2305.05084 #arxiv-1706.03762 #license-cc-by-nc-4.0 #model-index #has_space #region-us
| Canary 1B
=========
img {
display: inline;
}

| 
| 
NVIDIA NeMo Canary is a family of multi-lingual multi-tasking models that achieves state-of-the art performance on multiple benchmarks. With 1 billion parameters, Canary-1B supports automatic speech-to-text recognition (ASR) in 4 languages (English, German, French, Spanish) and translation from English to German/French/Spanish and from German/French/Spanish to English with or without punctuation and capitalization (PnC).
Model Architecture
------------------
Canary is an encoder-decoder model with FastConformer [1] encoder and Transformer Decoder [2].
With audio features extracted from the encoder, task tokens such as '', '', '' and ''
are fed into the Transformer Decoder to trigger the text generation process. Canary uses a concatenated tokenizer from individual
SentencePiece [3] tokenizers of each language, which makes it easy to scale up to more languages.
The Canay-1B model has 24 encoder layers and 24 layers of decoder layers in total.
NVIDIA NeMo
-----------
To train, fine-tune or play with the model you will need to install NVIDIA NeMo. We recommend you install it after you've installed Cython and latest PyTorch version.
How to Use this Model
---------------------
The model is available for use in the NeMo toolkit [4], and can be used as a pre-trained checkpoint for inference or for fine-tuning on another dataset.
### Loading the Model
### Input Format
The input to the model can be a directory containing audio files, in which case the model will perform ASR on English and produces text with punctuation and capitalization:
or use:
Another recommended option is to use a json manifest as input, where each line in the file is a dictionary containing the following fields:
and then use:
or use:
### Automatic Speech-to-text Recognition (ASR)
An example manifest for transcribing English audios can be:
### Automatic Speech-to-text Translation (AST)
An example manifest for transcribing English audios into German text can be:
### Input
This model accepts single channel (mono) audio sampled at 16000 Hz, along with the task/languages/PnC tags as input.
### Output
The model outputs the transcribed/translated text corresponding to the input audio, in the specified target language and with or without punctuation and capitalization.
Training
--------
Canary-1B is trained using the NVIDIA NeMo toolkit [4] for 150k steps with dynamic bucketing and a batch duration of 360s per GPU on 128 NVIDIA A100 80GB GPUs.
The model can be trained using this example script and base config.
The tokenizers for these models were built using the text transcripts of the train set with this script.
### Datasets
The Canary-1B model is trained on a total of 85k hrs of speech data. It consists of 31k hrs of public data, 20k hrs collected by Suno, and 34k hrs of in-house data.
The constituents of public data are as follows.
#### English (25.5k hours)
* Librispeech 960 hours
* Fisher Corpus
* Switchboard-1 Dataset
* WSJ-0 and WSJ-1
* National Speech Corpus (Part 1, Part 6)
* VCTK
* VoxPopuli (EN)
* Europarl-ASR (EN)
* Multilingual Librispeech (MLS EN) - 2,000 hour subset
* Mozilla Common Voice (v7.0)
* People's Speech - 12,000 hour subset
* Mozilla Common Voice (v11.0) - 1,474 hour subset
#### German (2.5k hours)
* Mozilla Common Voice (v12.0) - 800 hour subset
* Multilingual Librispeech (MLS DE) - 1,500 hour subset
* VoxPopuli (DE) - 200 hr subset
#### Spanish (1.4k hours)
* Mozilla Common Voice (v12.0) - 395 hour subset
* Multilingual Librispeech (MLS ES) - 780 hour subset
* VoxPopuli (ES) - 108 hour subset
* Fisher - 141 hour subset
#### French (1.8k hours)
* Mozilla Common Voice (v12.0) - 708 hour subset
* Multilingual Librispeech (MLS FR) - 926 hour subset
* VoxPopuli (FR) - 165 hour subset
Performance
-----------
In both ASR and AST experiments, predictions were generated using beam search with width 5 and length penalty 1.0.
### ASR Performance (w/o PnC)
The ASR performance is measured with word error rate (WER), and we process the groundtruth and predicted text with whisper-normalizer.
WER on MCV-16.1 test set:
WER on MLS test set:
More details on evaluation can be found at HuggingFace ASR Leaderboard
### AST Performance
We evaluate AST performance with BLEU score, and use native annotations with punctuation and capitalization in the datasets.
BLEU score on FLEURS test set:
BLEU score on COVOST-v2 test set:
BLEU score on mExpresso test set:
NVIDIA Riva: Deployment
-----------------------
NVIDIA Riva, is an accelerated speech AI SDK deployable on-prem, in all clouds, multi-cloud, hybrid, on edge, and embedded.
Additionally, Riva provides:
* World-class out-of-the-box accuracy for the most common languages with model checkpoints trained on proprietary data with hundreds of thousands of GPU-compute hours
* Best in class accuracy with run-time word boosting (e.g., brand and product names) and customization of acoustic model, language model, and inverse text normalization
* Streaming speech recognition, Kubernetes compatible scaling, and enterprise-grade support
Although this model isn’t supported yet by Riva, the list of supported models is here.
Check out Riva live demo.
References
----------
[1] Fast Conformer with Linearly Scalable Attention for Efficient Speech Recognition
[2] Attention is all you need
[3] Google Sentencepiece Tokenizer
[4] NVIDIA NeMo Toolkit
Licence
-------
License to use this model is covered by the CC-BY-NC-4.0. By downloading the public and release version of the model, you accept the terms and conditions of the CC-BY-NC-4.0 license.
| [
"### Loading the Model",
"### Input Format\n\n\nThe input to the model can be a directory containing audio files, in which case the model will perform ASR on English and produces text with punctuation and capitalization:\n\n\nor use:\n\n\nAnother recommended option is to use a json manifest as input, where each line in the file is a dictionary containing the following fields:\n\n\nand then use:\n\n\nor use:",
"### Automatic Speech-to-text Recognition (ASR)\n\n\nAn example manifest for transcribing English audios can be:",
"### Automatic Speech-to-text Translation (AST)\n\n\nAn example manifest for transcribing English audios into German text can be:",
"### Input\n\n\nThis model accepts single channel (mono) audio sampled at 16000 Hz, along with the task/languages/PnC tags as input.",
"### Output\n\n\nThe model outputs the transcribed/translated text corresponding to the input audio, in the specified target language and with or without punctuation and capitalization.\n\n\nTraining\n--------\n\n\nCanary-1B is trained using the NVIDIA NeMo toolkit [4] for 150k steps with dynamic bucketing and a batch duration of 360s per GPU on 128 NVIDIA A100 80GB GPUs.\nThe model can be trained using this example script and base config.\n\n\nThe tokenizers for these models were built using the text transcripts of the train set with this script.",
"### Datasets\n\n\nThe Canary-1B model is trained on a total of 85k hrs of speech data. It consists of 31k hrs of public data, 20k hrs collected by Suno, and 34k hrs of in-house data.\n\n\nThe constituents of public data are as follows.",
"#### English (25.5k hours)\n\n\n* Librispeech 960 hours\n* Fisher Corpus\n* Switchboard-1 Dataset\n* WSJ-0 and WSJ-1\n* National Speech Corpus (Part 1, Part 6)\n* VCTK\n* VoxPopuli (EN)\n* Europarl-ASR (EN)\n* Multilingual Librispeech (MLS EN) - 2,000 hour subset\n* Mozilla Common Voice (v7.0)\n* People's Speech - 12,000 hour subset\n* Mozilla Common Voice (v11.0) - 1,474 hour subset",
"#### German (2.5k hours)\n\n\n* Mozilla Common Voice (v12.0) - 800 hour subset\n* Multilingual Librispeech (MLS DE) - 1,500 hour subset\n* VoxPopuli (DE) - 200 hr subset",
"#### Spanish (1.4k hours)\n\n\n* Mozilla Common Voice (v12.0) - 395 hour subset\n* Multilingual Librispeech (MLS ES) - 780 hour subset\n* VoxPopuli (ES) - 108 hour subset\n* Fisher - 141 hour subset",
"#### French (1.8k hours)\n\n\n* Mozilla Common Voice (v12.0) - 708 hour subset\n* Multilingual Librispeech (MLS FR) - 926 hour subset\n* VoxPopuli (FR) - 165 hour subset\n\n\nPerformance\n-----------\n\n\nIn both ASR and AST experiments, predictions were generated using beam search with width 5 and length penalty 1.0.",
"### ASR Performance (w/o PnC)\n\n\nThe ASR performance is measured with word error rate (WER), and we process the groundtruth and predicted text with whisper-normalizer.\n\n\nWER on MCV-16.1 test set:\n\n\n\nWER on MLS test set:\n\n\n\nMore details on evaluation can be found at HuggingFace ASR Leaderboard",
"### AST Performance\n\n\nWe evaluate AST performance with BLEU score, and use native annotations with punctuation and capitalization in the datasets.\n\n\nBLEU score on FLEURS test set:\n\n\n\nBLEU score on COVOST-v2 test set:\n\n\n\nBLEU score on mExpresso test set:\n\n\n\nNVIDIA Riva: Deployment\n-----------------------\n\n\nNVIDIA Riva, is an accelerated speech AI SDK deployable on-prem, in all clouds, multi-cloud, hybrid, on edge, and embedded.\nAdditionally, Riva provides:\n\n\n* World-class out-of-the-box accuracy for the most common languages with model checkpoints trained on proprietary data with hundreds of thousands of GPU-compute hours\n* Best in class accuracy with run-time word boosting (e.g., brand and product names) and customization of acoustic model, language model, and inverse text normalization\n* Streaming speech recognition, Kubernetes compatible scaling, and enterprise-grade support\n\n\nAlthough this model isn’t supported yet by Riva, the list of supported models is here. \n\nCheck out Riva live demo.\n\n\nReferences\n----------\n\n\n[1] Fast Conformer with Linearly Scalable Attention for Efficient Speech Recognition\n\n\n[2] Attention is all you need\n\n\n[3] Google Sentencepiece Tokenizer\n\n\n[4] NVIDIA NeMo Toolkit\n\n\nLicence\n-------\n\n\nLicense to use this model is covered by the CC-BY-NC-4.0. By downloading the public and release version of the model, you accept the terms and conditions of the CC-BY-NC-4.0 license."
] | [
"TAGS\n#nemo #automatic-speech-recognition #automatic-speech-translation #speech #audio #Transformer #FastConformer #Conformer #pytorch #NeMo #hf-asr-leaderboard #en #de #es #fr #dataset-librispeech_asr #dataset-fisher_corpus #dataset-Switchboard-1 #dataset-WSJ-0 #dataset-WSJ-1 #dataset-National-Singapore-Corpus-Part-1 #dataset-National-Singapore-Corpus-Part-6 #dataset-vctk #dataset-voxpopuli #dataset-europarl #dataset-multilingual_librispeech #dataset-mozilla-foundation/common_voice_8_0 #dataset-MLCommons/peoples_speech #arxiv-2305.05084 #arxiv-1706.03762 #license-cc-by-nc-4.0 #model-index #has_space #region-us \n",
"### Loading the Model",
"### Input Format\n\n\nThe input to the model can be a directory containing audio files, in which case the model will perform ASR on English and produces text with punctuation and capitalization:\n\n\nor use:\n\n\nAnother recommended option is to use a json manifest as input, where each line in the file is a dictionary containing the following fields:\n\n\nand then use:\n\n\nor use:",
"### Automatic Speech-to-text Recognition (ASR)\n\n\nAn example manifest for transcribing English audios can be:",
"### Automatic Speech-to-text Translation (AST)\n\n\nAn example manifest for transcribing English audios into German text can be:",
"### Input\n\n\nThis model accepts single channel (mono) audio sampled at 16000 Hz, along with the task/languages/PnC tags as input.",
"### Output\n\n\nThe model outputs the transcribed/translated text corresponding to the input audio, in the specified target language and with or without punctuation and capitalization.\n\n\nTraining\n--------\n\n\nCanary-1B is trained using the NVIDIA NeMo toolkit [4] for 150k steps with dynamic bucketing and a batch duration of 360s per GPU on 128 NVIDIA A100 80GB GPUs.\nThe model can be trained using this example script and base config.\n\n\nThe tokenizers for these models were built using the text transcripts of the train set with this script.",
"### Datasets\n\n\nThe Canary-1B model is trained on a total of 85k hrs of speech data. It consists of 31k hrs of public data, 20k hrs collected by Suno, and 34k hrs of in-house data.\n\n\nThe constituents of public data are as follows.",
"#### English (25.5k hours)\n\n\n* Librispeech 960 hours\n* Fisher Corpus\n* Switchboard-1 Dataset\n* WSJ-0 and WSJ-1\n* National Speech Corpus (Part 1, Part 6)\n* VCTK\n* VoxPopuli (EN)\n* Europarl-ASR (EN)\n* Multilingual Librispeech (MLS EN) - 2,000 hour subset\n* Mozilla Common Voice (v7.0)\n* People's Speech - 12,000 hour subset\n* Mozilla Common Voice (v11.0) - 1,474 hour subset",
"#### German (2.5k hours)\n\n\n* Mozilla Common Voice (v12.0) - 800 hour subset\n* Multilingual Librispeech (MLS DE) - 1,500 hour subset\n* VoxPopuli (DE) - 200 hr subset",
"#### Spanish (1.4k hours)\n\n\n* Mozilla Common Voice (v12.0) - 395 hour subset\n* Multilingual Librispeech (MLS ES) - 780 hour subset\n* VoxPopuli (ES) - 108 hour subset\n* Fisher - 141 hour subset",
"#### French (1.8k hours)\n\n\n* Mozilla Common Voice (v12.0) - 708 hour subset\n* Multilingual Librispeech (MLS FR) - 926 hour subset\n* VoxPopuli (FR) - 165 hour subset\n\n\nPerformance\n-----------\n\n\nIn both ASR and AST experiments, predictions were generated using beam search with width 5 and length penalty 1.0.",
"### ASR Performance (w/o PnC)\n\n\nThe ASR performance is measured with word error rate (WER), and we process the groundtruth and predicted text with whisper-normalizer.\n\n\nWER on MCV-16.1 test set:\n\n\n\nWER on MLS test set:\n\n\n\nMore details on evaluation can be found at HuggingFace ASR Leaderboard",
"### AST Performance\n\n\nWe evaluate AST performance with BLEU score, and use native annotations with punctuation and capitalization in the datasets.\n\n\nBLEU score on FLEURS test set:\n\n\n\nBLEU score on COVOST-v2 test set:\n\n\n\nBLEU score on mExpresso test set:\n\n\n\nNVIDIA Riva: Deployment\n-----------------------\n\n\nNVIDIA Riva, is an accelerated speech AI SDK deployable on-prem, in all clouds, multi-cloud, hybrid, on edge, and embedded.\nAdditionally, Riva provides:\n\n\n* World-class out-of-the-box accuracy for the most common languages with model checkpoints trained on proprietary data with hundreds of thousands of GPU-compute hours\n* Best in class accuracy with run-time word boosting (e.g., brand and product names) and customization of acoustic model, language model, and inverse text normalization\n* Streaming speech recognition, Kubernetes compatible scaling, and enterprise-grade support\n\n\nAlthough this model isn’t supported yet by Riva, the list of supported models is here. \n\nCheck out Riva live demo.\n\n\nReferences\n----------\n\n\n[1] Fast Conformer with Linearly Scalable Attention for Efficient Speech Recognition\n\n\n[2] Attention is all you need\n\n\n[3] Google Sentencepiece Tokenizer\n\n\n[4] NVIDIA NeMo Toolkit\n\n\nLicence\n-------\n\n\nLicense to use this model is covered by the CC-BY-NC-4.0. By downloading the public and release version of the model, you accept the terms and conditions of the CC-BY-NC-4.0 license."
] | [
240,
5,
84,
29,
29,
36,
128,
70,
116,
54,
61,
85,
81,
352
] | [
"passage: TAGS\n#nemo #automatic-speech-recognition #automatic-speech-translation #speech #audio #Transformer #FastConformer #Conformer #pytorch #NeMo #hf-asr-leaderboard #en #de #es #fr #dataset-librispeech_asr #dataset-fisher_corpus #dataset-Switchboard-1 #dataset-WSJ-0 #dataset-WSJ-1 #dataset-National-Singapore-Corpus-Part-1 #dataset-National-Singapore-Corpus-Part-6 #dataset-vctk #dataset-voxpopuli #dataset-europarl #dataset-multilingual_librispeech #dataset-mozilla-foundation/common_voice_8_0 #dataset-MLCommons/peoples_speech #arxiv-2305.05084 #arxiv-1706.03762 #license-cc-by-nc-4.0 #model-index #has_space #region-us \n### Loading the Model### Input Format\n\n\nThe input to the model can be a directory containing audio files, in which case the model will perform ASR on English and produces text with punctuation and capitalization:\n\n\nor use:\n\n\nAnother recommended option is to use a json manifest as input, where each line in the file is a dictionary containing the following fields:\n\n\nand then use:\n\n\nor use:### Automatic Speech-to-text Recognition (ASR)\n\n\nAn example manifest for transcribing English audios can be:### Automatic Speech-to-text Translation (AST)\n\n\nAn example manifest for transcribing English audios into German text can be:### Input\n\n\nThis model accepts single channel (mono) audio sampled at 16000 Hz, along with the task/languages/PnC tags as input.",
"passage: ### Output\n\n\nThe model outputs the transcribed/translated text corresponding to the input audio, in the specified target language and with or without punctuation and capitalization.\n\n\nTraining\n--------\n\n\nCanary-1B is trained using the NVIDIA NeMo toolkit [4] for 150k steps with dynamic bucketing and a batch duration of 360s per GPU on 128 NVIDIA A100 80GB GPUs.\nThe model can be trained using this example script and base config.\n\n\nThe tokenizers for these models were built using the text transcripts of the train set with this script.### Datasets\n\n\nThe Canary-1B model is trained on a total of 85k hrs of speech data. It consists of 31k hrs of public data, 20k hrs collected by Suno, and 34k hrs of in-house data.\n\n\nThe constituents of public data are as follows.#### English (25.5k hours)\n\n\n* Librispeech 960 hours\n* Fisher Corpus\n* Switchboard-1 Dataset\n* WSJ-0 and WSJ-1\n* National Speech Corpus (Part 1, Part 6)\n* VCTK\n* VoxPopuli (EN)\n* Europarl-ASR (EN)\n* Multilingual Librispeech (MLS EN) - 2,000 hour subset\n* Mozilla Common Voice (v7.0)\n* People's Speech - 12,000 hour subset\n* Mozilla Common Voice (v11.0) - 1,474 hour subset#### German (2.5k hours)\n\n\n* Mozilla Common Voice (v12.0) - 800 hour subset\n* Multilingual Librispeech (MLS DE) - 1,500 hour subset\n* VoxPopuli (DE) - 200 hr subset#### Spanish (1.4k hours)\n\n\n* Mozilla Common Voice (v12.0) - 395 hour subset\n* Multilingual Librispeech (MLS ES) - 780 hour subset\n* VoxPopuli (ES) - 108 hour subset\n* Fisher - 141 hour subset#### French (1.8k hours)\n\n\n* Mozilla Common Voice (v12.0) - 708 hour subset\n* Multilingual Librispeech (MLS FR) - 926 hour subset\n* VoxPopuli (FR) - 165 hour subset\n\n\nPerformance\n-----------\n\n\nIn both ASR and AST experiments, predictions were generated using beam search with width 5 and length penalty 1.0.### ASR Performance (w/o PnC)\n\n\nThe ASR performance is measured with word error rate (WER), and we process the groundtruth and predicted text with whisper-normalizer.\n\n\nWER on MCV-16.1 test set:\n\n\n\nWER on MLS test set:\n\n\n\nMore details on evaluation can be found at HuggingFace ASR Leaderboard"
] | [
-0.06476311385631561,
0.09381026029586792,
-0.006047442555427551,
-0.0213244017213583,
0.04004813730716705,
-0.012204487808048725,
0.0902298092842102,
0.07827167958021164,
-0.03484522923827171,
0.06557691097259521,
-0.04266766458749771,
0.04375394806265831,
0.08726191520690918,
0.05530865490436554,
0.03828948363661766,
-0.19082435965538025,
0.03227783367037773,
-0.10078530013561249,
0.058741457760334015,
0.0582096129655838,
0.10338852554559708,
-0.04750174283981323,
0.04028838127851486,
0.019581452012062073,
0.04018738493323326,
-0.003100813366472721,
-0.006018103566020727,
-0.05120501667261124,
0.03259509801864624,
0.08199230581521988,
0.04117906466126442,
-0.013528215698897839,
0.07876380532979965,
-0.26066941022872925,
0.006411283742636442,
0.033405501395463943,
0.000058264704421162605,
0.02348361350595951,
0.1483229100704193,
-0.014387954026460648,
0.10818678885698318,
-0.030428767204284668,
-0.0059113577008247375,
0.09797106683254242,
-0.05452587082982063,
-0.11759517341852188,
-0.09078527987003326,
0.01200595311820507,
0.06205558404326439,
0.07803366333246231,
-0.05600334703922272,
0.015081074088811874,
-0.10799449682235718,
0.08810772001743317,
0.09629656374454498,
-0.08094561845064163,
-0.029659606516361237,
-0.03750631585717201,
0.06647417694330215,
0.08564936369657516,
-0.051700375974178314,
-0.013205280527472496,
-0.012413178570568562,
-0.01984545961022377,
-0.012544128112494946,
-0.0210565235465765,
-0.008047506213188171,
-0.04004514589905739,
-0.12678292393684387,
-0.049667052924633026,
0.1452258825302124,
0.03296809270977974,
-0.048469144850969315,
-0.1405639946460724,
0.012314023450016975,
0.0049162739887833595,
-0.034187234938144684,
0.022560765966773033,
0.015840008854866028,
0.000994737260043621,
0.11115366965532303,
-0.040895815938711166,
-0.06938352435827255,
-0.034513652324676514,
-0.024559929966926575,
0.06727125495672226,
0.017416737973690033,
0.0043328446336090565,
0.0027069204952567816,
0.05489272251725197,
-0.04345669597387314,
-0.08846975117921829,
-0.020877856761217117,
0.009394833818078041,
-0.12850767374038696,
-0.005311870947480202,
0.01777694560587406,
-0.12963038682937622,
0.0376480333507061,
0.07057574391365051,
0.03411737456917763,
0.02420107088983059,
-0.078172467648983,
0.008300233632326126,
0.035789474844932556,
0.10017287731170654,
-0.0715961679816246,
-0.07522433996200562,
-0.023233596235513687,
-0.0017740558832883835,
0.01604628935456276,
0.007548548746854067,
-0.05171108990907669,
0.007707088254392147,
-0.004527866840362549,
0.05312038213014603,
0.09320944547653198,
-0.013809184543788433,
-0.06403404474258423,
-0.024106960743665695,
0.1489875167608261,
-0.14271238446235657,
0.04495636373758316,
0.06379427015781403,
-0.049240592867136,
0.10130427032709122,
0.007075537461787462,
-0.0015974733978509903,
-0.11608059704303741,
0.0471264123916626,
-0.01205965131521225,
0.022474199533462524,
-0.09850995242595673,
-0.07260507345199585,
0.0639650747179985,
0.03615575283765793,
-0.06555262207984924,
-0.04402345046401024,
-0.0736696794629097,
-0.07644032686948776,
0.022958913818001747,
-0.07479910552501678,
-0.035286009311676025,
-0.07579834014177322,
-0.0483686663210392,
0.05400766432285309,
-0.053823843598365784,
0.09530901908874512,
-0.04618624225258827,
0.01642458513379097,
0.024278029799461365,
0.049641333520412445,
0.08575977385044098,
0.0419907420873642,
-0.004657992161810398,
-0.004967442713677883,
-0.12681855261325836,
0.16377875208854675,
-0.10487635433673859,
-0.013759264722466469,
-0.10312376171350479,
-0.026399433612823486,
-0.038871318101882935,
0.023069439455866814,
-0.003203967586159706,
0.08229872584342957,
-0.21157538890838623,
-0.04255202040076256,
0.19713938236236572,
-0.056604646146297455,
0.013636784628033638,
0.13851049542427063,
-0.003754344303160906,
-0.009155740961432457,
0.09184058010578156,
0.11136986315250397,
0.12908831238746643,
-0.17165496945381165,
-0.07711136341094971,
0.0068819597363471985,
0.01746433973312378,
0.08356698602437973,
0.0945783257484436,
-0.054508134722709656,
0.13406148552894592,
0.02083536982536316,
-0.015251720324158669,
0.020126981660723686,
0.010742094367742538,
-0.05796843767166138,
0.01904015988111496,
-0.045563943684101105,
0.009846651926636696,
-0.01098710298538208,
-0.050722699612379074,
-0.016614213585853577,
-0.11139574646949768,
0.0020935488864779472,
0.0797613263130188,
-0.02900877594947815,
0.048408642411231995,
-0.08109702169895172,
-0.03738431632518768,
0.008156081661581993,
0.01047047134488821,
-0.14237159490585327,
-0.01798081025481224,
0.025975853204727173,
-0.08112584799528122,
0.08644969761371613,
0.043639738112688065,
0.014188204891979694,
0.020877473056316376,
-0.03594512492418289,
0.008226053789258003,
0.03754214569926262,
-0.017891928553581238,
0.011993244290351868,
-0.12834982573986053,
-0.023972198367118835,
-0.035717956721782684,
0.0878414511680603,
-0.07163078337907791,
-0.01411096379160881,
0.05323326215147972,
0.10669701546430588,
0.007625236641615629,
-0.03865684196352959,
0.017407197505235672,
0.016608890146017075,
0.011810746043920517,
-0.007032428868114948,
-0.030111217871308327,
-0.0028742593713104725,
-0.020122632384300232,
0.1320236176252365,
-0.12952637672424316,
-0.09102845937013626,
0.0734851062297821,
0.023862455040216446,
-0.02540154941380024,
0.042049940675497055,
0.005663448944687843,
-0.04978853464126587,
-0.08817245066165924,
-0.11315728724002838,
0.19644224643707275,
0.042242638766765594,
0.06013001874089241,
-0.09782248735427856,
-0.06320013850927353,
-0.004618741571903229,
-0.04416795074939728,
-0.016365516930818558,
0.11459040641784668,
-0.01044833566993475,
-0.07732422649860382,
0.04539315775036812,
-0.10540645569562912,
-0.030916709452867508,
0.16925132274627686,
-0.017412645742297173,
-0.10008350014686584,
-0.03917720168828964,
0.03664201498031616,
0.03457210585474968,
0.01857488974928856,
-0.05100665241479874,
0.017498105764389038,
0.05205196142196655,
0.02528487890958786,
0.031098857522010803,
-0.09018132090568542,
0.06174424663186073,
0.0069848173297941685,
-0.10057638585567474,
-0.030631758272647858,
-0.019610440358519554,
0.0008966624736785889,
0.06689779460430145,
-0.020246166735887527,
0.013367884792387486,
-0.03140874207019806,
-0.07703046500682831,
-0.11766023933887482,
0.0732666403055191,
-0.12076626718044281,
-0.19761905074119568,
-0.18572327494621277,
-0.041729655116796494,
-0.029658177867531776,
0.00023602452711202204,
0.01316661387681961,
-0.057751767337322235,
-0.06586997210979462,
-0.07221006602048874,
0.030638089403510094,
0.00825701467692852,
-0.09633513540029526,
-0.03989425674080849,
0.04688912630081177,
0.03373884782195091,
-0.11286675930023193,
0.013109870254993439,
0.004704532213509083,
-0.045463383197784424,
-0.03354533016681671,
0.04091339558362961,
0.05024321749806404,
0.10467645525932312,
0.06685401499271393,
-0.021930847316980362,
0.018597807735204697,
0.1320599764585495,
-0.1118185818195343,
0.08204422891139984,
0.1293705403804779,
-0.002514820545911789,
0.03976885974407196,
0.135329931974411,
0.040961746126413345,
-0.011535518802702427,
-0.003519393503665924,
0.03932968154549599,
-0.027805456891655922,
-0.2755155861377716,
-0.10941699147224426,
-0.09669050574302673,
-0.05309561267495155,
0.03570372611284256,
0.007039000745862722,
0.03627626597881317,
-0.028584811836481094,
-0.11093307286500931,
0.00500890240073204,
0.09725387394428253,
0.012564551085233688,
0.11970300227403641,
0.0027380739338696003,
0.046638451516628265,
-0.04687195271253586,
-0.03700220212340355,
0.0765545517206192,
0.023710206151008606,
0.17651772499084473,
0.014006733894348145,
0.26920485496520996,
0.03451687842607498,
0.02735074609518051,
-0.0087069571018219,
0.0361286997795105,
-0.0010870397090911865,
0.03006349503993988,
0.011842763982713223,
-0.09177637100219727,
-0.019351014867424965,
0.07271161675453186,
0.1842672973871231,
-0.07681723684072495,
0.05994883179664612,
-0.04353085905313492,
0.042932868003845215,
0.114571213722229,
-0.005248600617051125,
-0.12007489800453186,
0.01793236844241619,
0.02265971712768078,
-0.08659252524375916,
-0.05626466125249863,
0.006725944112986326,
0.0976327508687973,
-0.0741705596446991,
0.09007339924573898,
0.02073628269135952,
0.10140097141265869,
-0.13623493909835815,
-0.04341311752796173,
-0.017307421192526817,
0.10442371666431427,
0.020013531669974327,
0.10123223066329956,
-0.15482327342033386,
0.09955764561891556,
0.015076426789164543,
0.10202832520008087,
-0.04126075655221939,
0.06022047623991966,
0.011065433733165264,
-0.046709559857845306,
0.10965073108673096,
0.026688434183597565,
-0.1272914707660675,
0.0065042427740991116,
-0.12397196888923645,
-0.04050549119710922,
0.07548017054796219,
-0.0066666379570961,
0.06851711124181747,
-0.008402963168919086,
-0.05754189193248749,
-0.03553394600749016,
-0.06307508796453476,
-0.15882796049118042,
-0.15968236327171326,
0.06048401817679405,
0.05557635799050331,
0.02758808061480522,
-0.045467935502529144,
-0.03285973519086838,
-0.08229905366897583,
0.07981914281845093,
-0.14626164734363556,
-0.047627173364162445,
-0.10472060739994049,
-0.07485812157392502,
0.18400755524635315,
-0.06039411202073097,
0.027448106557130814,
0.04162704199552536,
0.13203588128089905,
-0.03848708048462868,
-0.013142939656972885,
0.020447757095098495,
-0.08719654381275177,
-0.10459281504154205,
-0.0364668145775795,
0.15226790308952332,
0.09042277932167053,
0.08027333766222,
-0.007147949654608965,
0.027130216360092163,
0.001398689579218626,
-0.06827417016029358,
0.005510424729436636,
0.12659978866577148,
0.011922151781618595,
0.07494311034679413,
-0.03519463166594505,
-0.18380530178546906,
-0.10167177021503448,
-0.07700163871049881,
0.07080301642417908,
0.11304308474063873,
-0.027270499616861343,
0.12421731650829315,
0.13468924164772034,
-0.12308375537395477,
-0.18586811423301697,
-0.08594920486211777,
0.13380587100982666,
0.025038447231054306,
0.009951513260602951,
-0.19164986908435822,
0.031569186598062515,
0.02842964418232441,
-0.0353471003472805,
0.05286600440740585,
-0.23790019750595093,
-0.12891626358032227,
0.06372306495904922,
-0.049588363617658615,
-0.09409534931182861,
-0.0641968697309494,
-0.07842126488685608,
-0.08647570759057999,
-0.059450577944517136,
0.080140121281147,
-0.009426802396774292,
0.06331617385149002,
0.05824119597673416,
0.06005848944187164,
0.03382531926035881,
-0.03513474017381668,
0.10573389381170273,
0.06668103486299515,
-0.03349729999899864,
-0.05404432490468025,
0.006954676005989313,
0.04021277278661728,
-0.03128638118505478,
0.10512483865022659,
-0.03521905466914177,
-0.03064286708831787,
-0.06155148148536682,
-0.04221426323056221,
-0.08033274114131927,
0.02343650348484516,
-0.040250398218631744,
0.018269764259457588,
-0.05241861939430237,
0.03445048630237579,
0.08420075476169586,
0.009868800640106201,
0.02072293311357498,
-0.09931983053684235,
-0.07693551480770111,
0.22006069123744965,
0.1731102168560028,
0.059363748878240585,
-0.11867949366569519,
-0.04015867039561272,
0.013193395920097828,
0.02721608802676201,
-0.02744576521217823,
0.05655300244688988,
0.04717950522899628,
0.005803493782877922,
0.11601444333791733,
-0.03143304958939552,
-0.15894189476966858,
-0.010822468437254429,
0.05327396094799042,
-0.022872406989336014,
-0.12644463777542114,
-0.021028272807598114,
0.034455589950084686,
-0.10921014845371246,
-0.07695095241069794,
0.12539583444595337,
-0.002077743411064148,
-0.014230763539671898,
-0.00522853946313262,
0.09515316039323807,
-0.03973271697759628,
0.18732933700084686,
0.05263581871986389,
0.06064484640955925,
-0.08775493502616882,
0.048912402242422104,
0.08785261958837509,
-0.0059669129550457,
-0.010538456961512566,
0.1273706555366516,
-0.04718223959207535,
-0.049491435289382935,
-0.04072020575404167,
0.03963280841708183,
0.019834529608488083,
-0.06495915353298187,
-0.00856971088796854,
-0.08542156964540482,
0.06005106121301651,
0.030819997191429138,
0.0013975161127746105,
0.02072840929031372,
-0.014014857821166515,
-0.016674408689141273,
-0.03670301288366318,
0.1140996441245079,
0.028211262077093124,
0.016497474163770676,
-0.032664909958839417,
0.0561988539993763,
0.005782379303127527,
-0.02941913716495037,
0.0069782668724656105,
-0.024856194853782654,
-0.04446826130151749,
0.04446917399764061,
-0.03852961212396622,
0.06295622885227203,
-0.073919877409935,
0.01624966412782669,
-0.010277673602104187,
-0.01104864664375782,
-0.008797120302915573,
0.04799923300743103,
-0.07835233211517334,
-0.04859459772706032,
-0.05730173736810684,
0.10692618787288666,
-0.110406294465065,
0.008830118924379349,
0.05809502303600311,
-0.06284096837043762,
0.03988127410411835,
0.036363378167152405,
-0.04301770403981209,
0.05504700541496277,
-0.11543375253677368,
-0.02776729129254818,
0.012104641646146774,
0.09675362706184387,
0.007879262790083885,
-0.13949225842952728,
0.013787155970931053,
0.07418163865804672,
0.01248176395893097,
-0.03234804794192314,
-0.051467154175043106,
-0.1090562492609024,
0.05323476344347,
-0.043727487325668335,
-0.084197036921978,
-0.0437110960483551,
0.08360232412815094,
0.07856827229261398,
0.008143972605466843,
0.13175252079963684,
-0.04330942779779434,
0.055592477321624756,
-0.11783912777900696,
0.02919941209256649,
0.02787620946764946,
-0.03080156072974205,
0.03637697547674179,
-0.04199843108654022,
0.06884919106960297,
0.011639656499028206,
0.1383550763130188,
0.0400129035115242,
-0.014136086218059063,
0.012732798233628273,
-0.04671785980463028,
-0.0424206405878067,
0.04960869997739792,
0.07898101210594177,
0.009833220392465591,
-0.006119652185589075,
-0.01302887313067913,
-0.06398408114910126,
-0.018033467233181,
0.032782748341560364,
0.059263695031404495,
0.09701764583587646,
0.13405199348926544,
0.05218664929270744,
0.10489031672477722,
-0.11701887845993042,
-0.052450988441705704,
0.09707029163837433,
-0.10347827523946762,
0.045870304107666016,
-0.044849544763565063,
0.10089469701051712,
0.10379433631896973,
-0.14244051277637482,
0.13592322170734406,
0.012452082708477974,
-0.06916354596614838,
-0.08484883606433868,
-0.22087132930755615,
-0.05285631865262985,
-0.057944949716329575,
-0.0038984110578894615,
-0.09004497528076172,
0.10785659402608871,
0.040574315935373306,
0.02297690138220787,
-0.0016023050993680954,
0.08158425986766815,
-0.04730460047721863,
-0.11486845463514328,
0.07497184723615646,
-0.022462043911218643,
0.03114185482263565,
-0.001135098747909069,
0.015997348353266716,
0.07385575771331787,
0.03428395837545395,
0.07664768397808075,
0.061246272176504135,
-0.03300075605511665,
0.04013771191239357,
-0.019956786185503006,
-0.053049616515636444,
0.021267564967274666,
-0.03255699574947357,
0.040388401597738266,
0.1983587145805359,
0.06486108154058456,
-0.037725068628787994,
0.011001048609614372,
0.0782703384757042,
-0.03707171976566315,
-0.1461510807275772,
-0.12942180037498474,
-0.004864620044827461,
0.02402239479124546,
0.0324888601899147,
-0.00804170686751604,
-0.06905615329742432,
0.005033003631979227,
0.16680240631103516,
0.15827542543411255,
0.01849641092121601,
-0.010001054964959621,
-0.009223246946930885,
0.002093105111271143,
0.005517011508345604,
0.042578332126140594,
0.00888658594340086,
0.2709786295890808,
0.015348540619015694,
0.08019870519638062,
-0.002583680674433708,
-0.04657569155097008,
-0.022919902577996254,
0.07869716733694077,
-0.06722624599933624,
-0.009153377264738083,
-0.036762434989213943,
0.13944649696350098,
-0.06585344672203064,
-0.1705334335565567,
-0.024507679045200348,
0.01899750903248787,
-0.0797351822257042,
-0.002208167687058449,
0.10546642541885376,
0.05832463875412941,
0.04145490750670433,
0.009901905432343483,
-0.06310579925775528,
0.17097803950309753,
0.0070163654163479805,
-0.08623969554901123,
-0.05655965954065323,
-0.0026078345254063606,
-0.12487983703613281,
0.17159652709960938,
0.01366910245269537,
0.11992505192756653,
0.06414163112640381,
-0.0064267124980688095,
-0.07530003786087036,
0.08036072552204132,
0.022594600915908813,
-0.11578260362148285,
0.09093467891216278,
0.18329276144504547,
-0.04568835347890854,
0.1372137814760208,
0.04050713777542114,
-0.0032058730721473694,
0.08450247347354889,
-0.005537357181310654,
0.002401591744273901,
-0.09187792241573334,
0.08566030859947205,
-0.11472348123788834,
0.09500249475240707,
0.11728371679782867,
-0.017730534076690674,
-0.014741478487849236,
-0.02476350963115692,
0.01709589920938015,
-0.026058252900838852,
0.10731297731399536,
-0.024960998445749283,
-0.17285898327827454,
-0.024024607613682747,
0.010845367796719074,
0.0860218033194542,
-0.13050082325935364,
-0.03442680835723877,
0.004741913639008999,
-0.0658099353313446,
-0.03243153914809227,
0.05974970757961273,
0.06792937219142914,
-0.012156432494521141,
-0.029230652377009392,
-0.12605324387550354,
0.031042110174894333,
0.08357450366020203,
-0.10232216864824295,
-0.0372917503118515
] |
null | null | fastai |
# Amazing!
🥳 Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the [documentation here](https://huggingface.co/docs/hub/model-repos))!
2. Create a demo in Gradio or Streamlit using 🤗 Spaces ([documentation here](https://huggingface.co/docs/hub/spaces)).
3. Join the fastai community on the [Fastai Discord](https://discord.com/invite/YKrxeNn)!
Greetings fellow fastlearner 🤝! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| {"tags": ["fastai"]} | null | angela1996/intel-image-classification | [
"fastai",
"region:us"
] | 2024-02-07T17:21:03+00:00 | [] | [] | TAGS
#fastai #region-us
|
# Amazing!
Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the documentation here)!
2. Create a demo in Gradio or Streamlit using Spaces (documentation here).
3. Join the fastai community on the Fastai Discord!
Greetings fellow fastlearner ! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| [
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
"TAGS\n#fastai #region-us \n",
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
9,
20,
79,
3,
6,
12,
8
] | [
"passage: TAGS\n#fastai #region-us \n# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---# Model card## Model description\nMore information needed## Intended uses & limitations\nMore information needed## Training and evaluation data\nMore information needed"
] | [
-0.073318250477314,
-0.035918332636356354,
0.0016039619222283363,
0.09830865263938904,
0.16935402154922485,
0.11954792588949203,
0.06504721194505692,
0.08469552546739578,
0.09305626899003983,
0.008462822064757347,
0.08902737498283386,
-0.059808652848005295,
0.09601042419672012,
0.26935747265815735,
0.06010362133383751,
-0.24278773367404938,
0.02870224229991436,
-0.0036573195829987526,
0.08660013228654861,
0.06588653475046158,
0.12898924946784973,
-0.039593055844306946,
0.14736801385879517,
-0.018255524337291718,
-0.19320440292358398,
-0.054476846009492874,
-0.015185145661234856,
-0.019686169922351837,
0.12385433167219162,
-0.04793357476592064,
0.030790239572525024,
0.0026993011124432087,
-0.0015684126410633326,
-0.0995422899723053,
0.06401026993989944,
0.04089692234992981,
0.028817683458328247,
0.055760785937309265,
-0.04539911448955536,
0.08392030745744705,
0.054179996252059937,
-0.010920286178588867,
-0.12179892510175705,
0.09588204324245453,
-0.1474396139383316,
-0.2022949457168579,
-0.1278105229139328,
-0.11345728486776352,
0.047258179634809494,
0.01006549596786499,
-0.01907140202820301,
0.12847048044204712,
-0.14997079968452454,
-0.03727749362587929,
0.17807333171367645,
-0.15483331680297852,
-0.050517335534095764,
-0.0010879677720367908,
0.06801971048116684,
-0.06002732738852501,
-0.05137069150805473,
0.0968702957034111,
0.0906822457909584,
-0.019289257004857063,
0.015487968921661377,
0.0037353564985096455,
0.035032227635383606,
0.002429646672680974,
-0.0558350533246994,
0.06529499590396881,
-0.027788599953055382,
0.055927276611328125,
-0.1094130128622055,
-0.11809343844652176,
0.0010178228840231895,
0.03238791227340698,
-0.05549647659063339,
-0.06733305007219315,
0.0810781940817833,
0.007735111750662327,
-0.0603058859705925,
-0.11863275617361069,
-0.06696899980306625,
-0.12959590554237366,
0.00783742405474186,
0.09659197926521301,
0.0033950558863580227,
0.06878509372472763,
-0.09986882656812668,
0.06626693904399872,
-0.2048133760690689,
-0.04758621007204056,
-0.08781389147043228,
-0.1065201610326767,
0.02003002166748047,
-0.04773771017789841,
0.04778444394469261,
0.15393073856830597,
0.14042632281780243,
0.04171324521303177,
0.05645250529050827,
-0.029350629076361656,
0.038715146481990814,
0.04752078279852867,
0.018331103026866913,
0.03540196642279625,
-0.020549163222312927,
-0.18507646024227142,
0.0004176131042186171,
-0.04207618162035942,
0.08488372713327408,
-0.07463551312685013,
-0.05029602348804474,
0.01336510106921196,
-0.12160550057888031,
0.09655242413282394,
-0.05178983509540558,
-0.005084214266389608,
0.0036863412242382765,
0.008919943124055862,
0.20647431910037994,
0.04232640564441681,
0.004936119541525841,
-0.006976569537073374,
-0.1375076025724411,
-0.051532845944166183,
-0.09289269894361496,
0.034273598343133926,
0.02420172467827797,
0.01303885504603386,
-0.07711919397115707,
0.049177106469869614,
-0.046599894762039185,
-0.008231878280639648,
0.021442487835884094,
-0.20236440002918243,
0.010869519785046577,
-0.0969783291220665,
-0.1469350904226303,
0.06343341618776321,
0.0026821133214980364,
-0.07499043643474579,
0.08385025709867477,
-0.004780351184308529,
0.031972795724868774,
-0.030242523178458214,
-0.00177793821785599,
0.05239185318350792,
-0.08095952123403549,
0.023147141560912132,
0.1995297074317932,
0.10590710490942001,
-0.07641816139221191,
-0.0025978393387049437,
-0.12475098669528961,
0.04128078371286392,
-0.14157716929912567,
0.038516461849212646,
-0.08163458108901978,
0.15109841525554657,
-0.044047996401786804,
0.018007883802056313,
-0.0071970620192587376,
0.08468028157949448,
0.07606321573257446,
0.19981153309345245,
-0.23198086023330688,
-0.053279466927051544,
0.16512827575206757,
-0.11487894505262375,
-0.18565405905246735,
0.20080815255641937,
-0.00043150142300873995,
0.10752102732658386,
-0.010421866551041603,
0.17009462416172028,
-0.021746216341853142,
-0.14181379973888397,
-0.032203078269958496,
-0.0012119774473831058,
-0.24691128730773926,
-0.08980891108512878,
0.09945957362651825,
0.10481112450361252,
-0.059047527611255646,
0.029137471690773964,
0.012005627155303955,
0.15818172693252563,
-0.07679074257612228,
-0.04601999372243881,
-0.007829579524695873,
-0.10506698489189148,
0.022122014313936234,
0.01663162000477314,
0.034775324165821075,
-0.059334270656108856,
-0.00890427641570568,
-0.07678428292274475,
0.13092219829559326,
0.09849999099969864,
-0.03540538251399994,
-0.06064159423112869,
0.16454961895942688,
-0.0640924945473671,
-0.026323838159441948,
0.08331746608018875,
-0.08536569774150848,
0.047215063124895096,
0.04028964787721634,
0.05084947869181633,
0.009997997432947159,
0.09182237833738327,
0.0698544830083847,
0.006789602339267731,
0.03368524834513664,
0.13270887732505798,
-0.027426021173596382,
-0.05121328681707382,
0.01674247533082962,
0.04598715528845787,
-0.00979064591228962,
0.3169313669204712,
-0.19912512600421906,
0.018945744261145592,
-0.06457886099815369,
0.08035559207201004,
0.0660853385925293,
0.007019065320491791,
0.07570107281208038,
-0.05360652506351471,
-0.016966497525572777,
-0.045681122690439224,
0.06926878541707993,
-0.06979862600564957,
-0.054223138839006424,
0.2564660608768463,
-0.031106717884540558,
0.031359151005744934,
0.10653062164783478,
-0.06802138686180115,
-0.05823708325624466,
-0.02224794402718544,
-0.0014688228257000446,
0.023401014506816864,
-0.04168177396059036,
0.06067536398768425,
-0.08815024048089981,
-0.05285300314426422,
0.1703105866909027,
-0.038786694407463074,
0.07842917740345001,
0.035427022725343704,
-0.05379872769117355,
-0.04481838271021843,
0.061976201832294464,
0.14977918565273285,
-0.0965908095240593,
0.06779327243566513,
0.13305115699768066,
0.014980388805270195,
0.15411095321178436,
0.07098863273859024,
-0.07586279511451721,
-0.08855607360601425,
-0.018246978521347046,
-0.004062598571181297,
0.18133139610290527,
-0.07897800207138062,
-0.036732085049152374,
0.042683616280555725,
-0.011134039610624313,
0.06611642241477966,
-0.05846851319074631,
-0.0792742595076561,
0.01736506260931492,
-0.0582035630941391,
0.018060972914099693,
0.12486616522073746,
-0.08240851759910583,
0.04267239198088646,
0.03745635226368904,
-0.058472223579883575,
0.046025440096855164,
0.0389089435338974,
-0.01086228247731924,
0.05541912093758583,
0.06821268051862717,
-0.2134213149547577,
-0.10377796739339828,
-0.17595313489437103,
0.03000609390437603,
0.020109420642256737,
0.036413755267858505,
-0.10920769721269608,
0.02131613902747631,
-0.0651998370885849,
-0.07437032461166382,
0.04871295765042305,
-0.029500357806682587,
-0.10847225040197372,
-0.027001040056347847,
-0.024241603910923004,
-0.04816099628806114,
-0.021433888003230095,
-0.06250716745853424,
0.03129231557250023,
0.04526861384510994,
0.03191622346639633,
0.1321185976266861,
-0.010805734433233738,
-0.014524625614285469,
0.002761868294328451,
-0.017431288957595825,
0.1497519314289093,
-0.13988617062568665,
0.06941607594490051,
0.1812426596879959,
0.09771130234003067,
0.03844839334487915,
0.01466822624206543,
0.03106272965669632,
-0.07663184404373169,
0.005383877083659172,
0.034619297832250595,
-0.0891294777393341,
-0.08207139372825623,
-0.01874193549156189,
-0.03897557035088539,
0.21049608290195465,
-0.12441039085388184,
0.024025630205869675,
0.040357187390327454,
0.09686839580535889,
0.11187659204006195,
-0.04121972620487213,
-0.17262403666973114,
0.04177050292491913,
-0.2474004179239273,
-0.051238708198070526,
0.003026821883395314,
-0.09497712552547455,
-0.06320231407880783,
0.18337351083755493,
0.0052159554325044155,
0.0287664532661438,
0.00430127140134573,
0.12202860414981842,
-0.0009366215672343969,
0.12068869173526764,
0.0687243714928627,
-0.05316835641860962,
0.02255408652126789,
-0.09993521869182587,
-0.0696573555469513,
-0.03704388439655304,
-0.07047778367996216,
0.06136435270309448,
0.12800902128219604,
-0.024759603664278984,
-0.04259653389453888,
0.04763835668563843,
0.09553752839565277,
0.06145815551280975,
0.15860231220722198,
-0.16057826578617096,
-0.022865094244480133,
0.042546581476926804,
-0.029262376949191093,
-0.049140751361846924,
-0.009500340558588505,
0.08492209017276764,
-0.05378608778119087,
-0.02665375918149948,
0.003306680591776967,
0.07226359844207764,
-0.0019794153049588203,
0.0436936691403389,
-0.03244423121213913,
0.1845880150794983,
-0.029572106897830963,
0.023350762203335762,
-0.12604808807373047,
0.13696090877056122,
0.022422920912504196,
-0.015438690781593323,
-0.06568175554275513,
-0.05596291273832321,
0.18064838647842407,
0.02166406810283661,
0.11738308519124985,
0.011424299329519272,
-0.09442766010761261,
-0.1337079405784607,
-0.1388736516237259,
0.015837913379073143,
0.09729303419589996,
-0.01256689801812172,
-0.03353166952729225,
0.019608711823821068,
-0.04281611740589142,
-0.06777504086494446,
0.10452067106962204,
-0.11668688803911209,
-0.0018522912869229913,
0.005423946306109428,
0.0416572242975235,
-0.06085909157991409,
0.032720211893320084,
0.03296784311532974,
-0.0647648349404335,
0.121244877576828,
0.24137550592422485,
0.1064029112458229,
-0.09990023821592331,
-0.08652417361736298,
0.021780110895633698,
-0.034567005932331085,
-0.0014182132435962558,
-0.016133872792124748,
0.036385562270879745,
0.0019662054255604744,
0.003586959559470415,
0.13572031259536743,
-0.07582411170005798,
0.012567305937409401,
-0.08275366574525833,
0.07902812212705612,
-0.0409930944442749,
-0.0025117802433669567,
-0.003995150327682495,
-0.02950184792280197,
-0.03430648893117905,
-0.06180789694190025,
0.163230761885643,
-0.06168964132666588,
-0.08240502327680588,
0.07821446657180786,
0.01680770143866539,
0.017550375312566757,
-0.06227098032832146,
-0.054205916821956635,
0.1972212791442871,
0.31792324781417847,
-0.058273475617170334,
0.10361375659704208,
0.1383560746908188,
0.023166829720139503,
-0.22579050064086914,
0.036502011120319366,
-0.14466507732868195,
0.032058101147413254,
0.024782279506325722,
-0.06415819376707077,
0.05856261029839516,
0.1250556856393814,
-0.045668914914131165,
0.23617008328437805,
-0.03641456738114357,
-0.07633192092180252,
-0.013243574649095535,
0.043972890824079514,
0.3091393709182739,
-0.11325396597385406,
-0.02349173277616501,
-0.11636991053819656,
-0.21521669626235962,
0.06708590686321259,
-0.16208602488040924,
0.1406344771385193,
-0.05703224614262581,
0.023474344983696938,
-0.012111215852200985,
-0.07578689604997635,
0.19497497379779816,
-0.1371963620185852,
0.056931521743535995,
-0.1432308852672577,
-0.11647364497184753,
-0.005183211527764797,
-0.08439649641513824,
0.14731425046920776,
-0.08327576518058777,
-0.02632858417928219,
-0.2082071304321289,
0.001373599166981876,
-0.021641740575432777,
0.09738951921463013,
0.02311836928129196,
-0.07967846095561981,
-0.08035353571176529,
0.12579506635665894,
-0.07811200618743896,
0.036513522267341614,
-0.08704032748937607,
-0.03989429399371147,
-0.026884159073233604,
-0.08092786371707916,
0.06243825703859329,
-0.08906654268503189,
0.16072829067707062,
-0.049172405153512955,
-0.046159181743860245,
0.061650797724723816,
-0.20832203328609467,
0.026940656825900078,
0.036382775753736496,
-0.031731411814689636,
0.10237374156713486,
-0.029687397181987762,
-0.07129550725221634,
0.1133488118648529,
0.13133300840854645,
-0.07154961675405502,
-0.2563934028148651,
-0.0821671262383461,
-0.008923565037548542,
0.04608851298689842,
0.0829237625002861,
0.04836045205593109,
-0.05231332778930664,
-0.017525162547826767,
-0.031239798292517662,
0.03463910520076752,
-0.11768791079521179,
-0.02900020219385624,
0.06892099231481552,
0.0014350401470437646,
-0.09527117758989334,
0.0962897539138794,
-0.004287306685000658,
-0.02237984538078308,
-0.009249147027730942,
0.1892271637916565,
-0.014808090403676033,
-0.12871821224689484,
-0.057921428233385086,
0.24053727090358734,
-0.038428641855716705,
-0.07654319703578949,
-0.06858045607805252,
-0.011265470646321774,
-0.04038287326693535,
0.06209278851747513,
0.04795577749609947,
-0.01209679339081049,
0.08278531581163406,
0.06026776134967804,
-0.1221788227558136,
-0.060724351555109024,
-0.05533421039581299,
0.035240933299064636,
-0.09762322902679443,
0.04652146250009537,
0.016370195895433426,
0.12453475594520569,
-0.09184806793928146,
-0.03038635104894638,
-0.11205437779426575,
-0.059142544865608215,
-0.18314886093139648,
-0.0571221299469471,
-0.041237685829401016,
-0.008055833168327808,
0.03931373730301857,
0.02697678469121456,
-0.04493580758571625,
-0.048296377062797546,
-0.06704439222812653,
0.03899036720395088,
0.07422684133052826,
0.026717372238636017,
-0.03390409052371979,
0.05009619519114494,
0.06439550966024399,
0.008286280557513237,
0.1963774412870407,
0.06738202273845673,
0.061680130660533905,
-0.025940580293536186,
-0.19781054556369781,
-0.05686524137854576,
0.002742079785093665,
-0.09212438762187958,
0.12195391207933426,
-0.011633808724582195,
0.02040605992078781,
-0.06281229853630066,
0.03727225586771965,
0.026594331488013268,
0.10702691227197647,
-0.02029390074312687,
0.0958021730184555,
0.029817266389727592,
-0.08947111666202545,
-0.044351425021886826,
0.015944788232445717,
0.12201714515686035,
0.02899266965687275,
0.028689615428447723,
0.015606578439474106,
0.037100955843925476,
-0.03902486339211464,
0.0296308696269989,
-0.045808494091033936,
-0.14955224096775055,
0.01991276629269123,
-0.046732377260923386,
-0.006942411884665489,
-0.016697930172085762,
0.18722283840179443,
0.04047711566090584,
-0.046649303287267685,
-0.01265130564570427,
0.014551439322531223,
-0.004945865832269192,
-0.03270510211586952,
-0.004582806024700403,
0.06002182513475418,
-0.004176365211606026,
-0.047248490154743195,
0.13213102519512177,
0.046804413199424744,
0.04763852432370186,
0.0742364451289177,
0.09783162921667099,
-0.00930761732161045,
0.13372060656547546,
0.06815905123949051,
-0.01982966810464859,
-0.1131899505853653,
-0.05649255961179733,
-0.11679257452487946,
0.034573014825582504,
-0.05576380714774132,
0.12528598308563232,
0.11196581274271011,
-0.060735806822776794,
-0.03883470967411995,
-0.0771038830280304,
-0.03134944289922714,
-0.07594948261976242,
0.03614310547709465,
-0.0327751524746418,
-0.08104247599840164,
0.06421366333961487,
0.05536265671253204,
-0.036099426448345184,
0.11491319537162781,
0.020650042220950127,
-0.05702126771211624,
0.12617406249046326,
-0.07743373513221741,
0.10717736184597015,
0.07707828283309937,
-0.05362870916724205,
-0.12441752851009369,
0.011045942083001137,
-0.07996662706136703,
-0.11546584963798523,
-0.008837178349494934,
-0.011918267235159874,
-0.0746825784444809,
-0.05780024081468582,
0.10738345980644226,
-0.03462931141257286,
-0.09724929928779602,
-0.020749187096953392,
0.015756776556372643,
0.056543223559856415,
-0.019683608785271645,
0.0018315898487344384,
0.03772254288196564,
0.028699718415737152,
0.15574465692043304,
-0.0016714793164283037,
0.06267286092042923,
-0.1358945369720459,
0.18023191392421722,
-0.1432318240404129,
-0.027932528406381607,
-0.187766894698143,
-0.0886974111199379,
-0.025430310517549515,
0.22427266836166382,
0.26061514019966125,
-0.1923753172159195,
-0.03171071037650108,
0.004376344382762909,
-0.010204915888607502,
-0.07923580706119537,
0.14464490115642548,
0.02417137287557125,
-0.007147552911192179,
-0.06552806496620178,
-0.014752711169421673,
0.024085145443677902,
-0.07228498160839081,
-0.035760894417762756,
0.18496830761432648,
0.0086367791518569,
0.07214809954166412,
-0.09064984321594238,
0.03641578182578087,
-0.18433186411857605,
-0.0693570077419281,
-0.03508331999182701,
-0.138646200299263,
-0.09639570862054825,
-0.01481159869581461,
0.003136083483695984,
0.09603974968194962,
0.03350212052464485,
-0.01305394247174263,
0.06808507442474365,
-0.049502357840538025,
0.010726232081651688,
-0.16043636202812195,
-0.020468583330512047,
0.05376148223876953,
-0.052667658776044846,
0.23897892236709595,
-0.02351270616054535,
-0.12297288328409195,
0.08416848629713058,
-0.03519788756966591,
-0.12302011996507645,
0.0745280459523201,
-0.023310834541916847,
-0.10405170172452927,
-0.05555706471204758,
0.17993386089801788,
-0.01256539486348629,
-0.16247478127479553,
0.03247550129890442,
-0.15925332903862,
0.029797034338116646,
0.03576231747865677,
-0.011352102272212505,
-0.05518606677651405,
0.028951244428753853,
-0.027475930750370026,
0.10062393546104431,
0.14163273572921753,
0.017354421317577362,
-0.009662404656410217,
-0.06593839079141617,
0.09352979063987732,
0.06211914122104645,
-0.07753235101699829,
-0.11338558793067932,
-0.09994973242282867,
0.02616780437529087,
0.07790441066026688,
-0.08538854867219925,
-0.17278192937374115,
-0.029272083193063736,
-0.11865141987800598,
-0.002084053121507168,
0.0349934957921505,
0.06834512948989868,
0.2863384187221527,
0.06974043697118759,
0.004092831164598465,
-0.15255671739578247,
0.05762675032019615,
0.08219972252845764,
-0.02544020675122738,
-0.08790270239114761
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | SpideyDLK/wav2vec2-large-xls-r-300m-sinhala-test2-half-data | [
"transformers",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-07T17:21:51+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
26,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.08389580249786377,
0.19830818474292755,
-0.0013316317927092314,
0.02313883788883686,
0.11396584659814835,
0.01961737498641014,
0.053626976907253265,
0.14538456499576569,
0.0060051376931369305,
0.10656800121068954,
0.066679947078228,
0.09131570905447006,
0.09678101539611816,
0.20042605698108673,
0.04371999576687813,
-0.17659740149974823,
0.010636410675942898,
-0.06930278241634369,
-0.010073255747556686,
0.11651819199323654,
0.141214057803154,
-0.10151198506355286,
0.07627976685762405,
-0.03319970890879631,
-0.02870541252195835,
-0.0070160143077373505,
-0.07769215852022171,
-0.05755697935819626,
0.07573003321886063,
0.054863471537828445,
0.04207949340343475,
-0.0008347301045432687,
0.08447454124689102,
-0.2674994468688965,
0.013753628358244896,
0.07452993094921112,
0.010659529827535152,
0.05990942195057869,
0.07833302766084671,
-0.04036625102162361,
0.12881849706172943,
-0.06320446729660034,
0.13035163283348083,
0.0906217098236084,
-0.0681561604142189,
-0.24378153681755066,
-0.08239314705133438,
0.06505522131919861,
0.12533815205097198,
0.07694927603006363,
-0.02823091857135296,
0.16422191262245178,
-0.07247646898031235,
0.019290022552013397,
0.09481704235076904,
-0.1151006743311882,
-0.060644298791885376,
0.08318385481834412,
0.14101974666118622,
0.10340547561645508,
-0.1255619376897812,
-0.012289565056562424,
0.04275871813297272,
0.045979104936122894,
0.07389909774065018,
0.011339850723743439,
0.1143413558602333,
0.05629947781562805,
-0.13526225090026855,
-0.05700986459851265,
0.14547574520111084,
0.023872992023825645,
-0.057064127177000046,
-0.2138909548521042,
-0.002902575535699725,
-0.07730814069509506,
-0.011685127392411232,
-0.06846728920936584,
0.0291305985301733,
-0.01194276288151741,
0.060226380825042725,
-0.0496203787624836,
-0.09797755628824234,
-0.046314824372529984,
0.1015089675784111,
0.054820988327264786,
0.011354796588420868,
-0.01489334274083376,
0.03576440364122391,
0.13432876765727997,
0.04213530570268631,
-0.10012737661600113,
-0.07065672427415848,
-0.0701170489192009,
-0.09620913118124008,
-0.03947552293539047,
0.04272124543786049,
0.020167991518974304,
0.042202774435281754,
0.2283228635787964,
0.024096308276057243,
0.05459817871451378,
0.029667891561985016,
0.0026177873369306326,
0.03211980313062668,
0.1073630079627037,
-0.041210614144802094,
-0.188126802444458,
-0.03292805701494217,
0.0931866466999054,
-0.009821015410125256,
-0.028658604249358177,
-0.033444397151470184,
0.035014089196920395,
0.08379437029361725,
0.11821532249450684,
0.08875755965709686,
-0.012828069739043713,
-0.037612639367580414,
-0.03493109717965126,
0.2115669697523117,
-0.14141373336315155,
0.045799970626831055,
-0.022097334265708923,
-0.018195297569036484,
-0.06905751675367355,
0.030103791505098343,
0.01831657998263836,
-0.003142025787383318,
0.06966056674718857,
-0.061253178864717484,
-0.05794486775994301,
-0.11518853157758713,
-0.045523155480623245,
0.04711875319480896,
-0.024105608463287354,
-0.024469668045639992,
-0.07765042781829834,
-0.11219723522663116,
-0.06417357176542282,
0.06612563133239746,
-0.04156653955578804,
-0.03974827378988266,
0.005308232270181179,
-0.07131324708461761,
0.008387917652726173,
0.008993842639029026,
0.12122467905282974,
-0.030063031241297722,
0.05833350867033005,
-0.002476902212947607,
0.05916252359747887,
0.10643328726291656,
0.03227818012237549,
-0.08492200076580048,
0.057466037571430206,
-0.20633617043495178,
0.08371785283088684,
-0.11420095711946487,
0.034276340156793594,
-0.17048145830631256,
-0.024183684960007668,
0.008447963744401932,
0.023597201332449913,
0.023726604878902435,
0.1338067352771759,
-0.2097422182559967,
-0.016196569427847862,
0.14133213460445404,
-0.09649793803691864,
-0.12422871589660645,
0.07990546524524689,
-0.03459475561976433,
0.1747698187828064,
0.038475677371025085,
-0.019652999937534332,
0.09909367561340332,
-0.15559963881969452,
-0.05852397903800011,
-0.026064254343509674,
-0.008927824907004833,
0.08823978155851364,
0.07542291283607483,
-0.05844951793551445,
0.02285866066813469,
0.02562655322253704,
-0.04727208614349365,
-0.0268824752420187,
-0.05256075784564018,
-0.10127434879541397,
-0.023140445351600647,
-0.09642518311738968,
0.026515161618590355,
0.000058677000197349116,
-0.07310442626476288,
-0.028560271486639977,
-0.17347893118858337,
-0.02563360333442688,
0.10103316605091095,
0.004820956848561764,
-0.007559072691947222,
-0.08540112525224686,
0.022149885073304176,
-0.05362366884946823,
-0.006164622958749533,
-0.16996455192565918,
-0.03558015450835228,
0.051895126700401306,
-0.14917676150798798,
0.015460150316357613,
-0.07327745854854584,
0.07047311216592789,
0.02098717913031578,
-0.05859505757689476,
-0.03108096309006214,
0.0007694467785768211,
0.004292082041501999,
-0.06229274719953537,
-0.1903683841228485,
-0.058886781334877014,
-0.041500482708215714,
0.15720732510089874,
-0.24841000139713287,
0.0300158578902483,
0.03247617185115814,
0.13185922801494598,
0.007058668415993452,
-0.06344027817249298,
0.02096918225288391,
-0.04676475748419762,
-0.050621338188648224,
-0.06898977607488632,
-0.009901339188218117,
-0.014539826661348343,
-0.031393732875585556,
0.012980648316442966,
-0.14970256388187408,
-0.060514215379953384,
0.09452559798955917,
0.11224991828203201,
-0.14555825293064117,
0.00204002158716321,
-0.0460561066865921,
-0.07002599537372589,
-0.07487804442644119,
-0.0761631652712822,
0.07739497721195221,
0.044650159776210785,
0.049250341951847076,
-0.06317461282014847,
-0.06234706938266754,
0.023210179060697556,
0.005524294450879097,
-0.019023682922124863,
0.0948529988527298,
0.074309803545475,
-0.09122881293296814,
0.07973480224609375,
0.08461450785398483,
0.04414684325456619,
0.086973637342453,
0.005991141777485609,
-0.11396963149309158,
-0.03062884695827961,
0.037754856050014496,
0.024159027263522148,
0.15351562201976776,
-0.08692087233066559,
0.030462130904197693,
0.052177220582962036,
-0.03854219615459442,
0.03157065063714981,
-0.0923321321606636,
0.025362705811858177,
0.021495236083865166,
-0.006555700208991766,
0.05864228308200836,
-0.018769768998026848,
-0.01403577346354723,
0.06336429715156555,
0.05677810311317444,
0.044270504266023636,
0.02595379762351513,
-0.02093072421848774,
-0.1278371512889862,
0.16537296772003174,
-0.09028079360723495,
-0.2540280222892761,
-0.17074446380138397,
0.015454737469553947,
0.03706491366028786,
-0.021728800609707832,
0.039588842540979385,
-0.06286025792360306,
-0.10237989574670792,
-0.09417891502380371,
0.0029635571409016848,
0.023925531655550003,
-0.058347854763269424,
-0.0817074254155159,
0.060779985040426254,
0.04047083482146263,
-0.13689260184764862,
0.0349188968539238,
0.06170675903558731,
-0.03042641654610634,
0.0018567070364952087,
0.07321398705244064,
0.12743599712848663,
0.14838241040706635,
-0.006730219814926386,
-0.012446845881640911,
0.035035960376262665,
0.229813352227211,
-0.1490442156791687,
0.10630457103252411,
0.14053207635879517,
-0.021705523133277893,
0.06635113060474396,
0.1461038440465927,
0.023231739178299904,
-0.07546708732843399,
0.04147516191005707,
0.04027445614337921,
-0.04228919371962547,
-0.2589097023010254,
-0.05694316700100899,
-0.00946022942662239,
-0.07043391466140747,
0.09718906134366989,
0.09238530695438385,
0.11972260475158691,
0.0337289460003376,
-0.05568677559494972,
-0.025771914049983025,
-0.003401360474526882,
0.114128477871418,
-0.027640055865049362,
-0.004564122296869755,
0.07965842634439468,
-0.05878787487745285,
0.011684526689350605,
0.09941446036100388,
0.019347423687577248,
0.17601320147514343,
0.02533329278230667,
0.10681075602769852,
0.06725578010082245,
0.09347675740718842,
-0.0015635732561349869,
0.034774236381053925,
0.05337131395936012,
0.022044572979211807,
0.010453542694449425,
-0.09408048540353775,
-0.012431944720447063,
0.13713060319423676,
0.019816776737570763,
0.009031654335558414,
0.008926562033593655,
-0.01010479498654604,
0.03131420537829399,
0.20501568913459778,
0.0009575071162544191,
-0.22537250816822052,
-0.09500737488269806,
0.059459153562784195,
-0.06931101530790329,
-0.143676295876503,
-0.02094252221286297,
0.030270220711827278,
-0.17292405664920807,
0.016790566965937614,
-0.0316389761865139,
0.09112390875816345,
-0.07145322859287262,
-0.028050832450389862,
0.06891903281211853,
0.07569212466478348,
-0.012108199298381805,
0.07973295450210571,
-0.19069278240203857,
0.12254468351602554,
0.03037673607468605,
0.08605273067951202,
-0.11708726733922958,
0.07849059253931046,
-0.0019813794642686844,
-0.014807495288550854,
0.17999744415283203,
-0.014062200672924519,
-0.0586031936109066,
-0.08878950774669647,
-0.08704045414924622,
-0.011727320961654186,
0.10361312329769135,
-0.09322915226221085,
0.09586969763040543,
-0.02775636687874794,
-0.03705112263560295,
0.012418309226632118,
-0.10469507426023483,
-0.1636953055858612,
-0.18679304420948029,
0.06244563311338425,
-0.07802703976631165,
0.012347841635346413,
-0.11227322369813919,
-0.06334327906370163,
-0.01575082167983055,
0.23160123825073242,
-0.16648635268211365,
-0.07049825042486191,
-0.1498587429523468,
-0.03997112438082695,
0.17463743686676025,
-0.042160745710134506,
0.06849376112222672,
-0.021383514627814293,
0.1873992383480072,
-0.008081548847258091,
-0.013158116489648819,
0.06569221615791321,
-0.09637628495693207,
-0.16879262030124664,
-0.05748843029141426,
0.14160962402820587,
0.10863390564918518,
0.05731578543782234,
-0.0038195757661014795,
0.013171887956559658,
-0.03383830562233925,
-0.09896382689476013,
0.013824623078107834,
0.13817466795444489,
0.0034514935687184334,
0.00682973163202405,
-0.03995988517999649,
-0.07027145475149155,
-0.05825701728463173,
-0.07912654429674149,
0.057147104293107986,
0.187900573015213,
-0.09512355923652649,
0.1602867990732193,
0.12431421875953674,
-0.06468851119279861,
-0.2306901067495346,
0.03996593505144119,
0.04701630026102066,
0.007666614837944508,
0.022401191294193268,
-0.19138796627521515,
0.09788824617862701,
0.0009011493530124426,
-0.06807263940572739,
0.14616990089416504,
-0.16564498841762543,
-0.1461436152458191,
0.08002161979675293,
0.025075770914554596,
-0.22560662031173706,
-0.14821304380893707,
-0.1037549376487732,
-0.03735695406794548,
-0.13707835972309113,
0.048581719398498535,
0.02614329755306244,
0.019834673032164574,
0.025222565978765488,
0.005338077899068594,
0.029657263308763504,
-0.07272187620401382,
0.1870686560869217,
-0.020297454670071602,
0.0072362530045211315,
-0.050640691071748734,
-0.04617878794670105,
0.09227550774812698,
-0.06150037795305252,
0.11741586774587631,
0.018679620698094368,
0.018796883523464203,
-0.1431548148393631,
-0.049209367483854294,
-0.060803934931755066,
0.04456847906112671,
-0.07284719496965408,
-0.09393193572759628,
-0.04137463867664337,
0.08888561278581619,
0.07211937010288239,
-0.032792408019304276,
-0.0027768779546022415,
-0.07569456845521927,
0.09405932575464249,
0.184477761387825,
0.17357055842876434,
0.009977072477340698,
-0.07020942866802216,
0.024555526673793793,
-0.042279548943042755,
0.03349342197179794,
-0.24652716517448425,
0.03456863760948181,
0.066053606569767,
0.03803660348057747,
0.08509242534637451,
-0.016836483031511307,
-0.1781480610370636,
-0.04086102172732353,
0.08498652279376984,
-0.06206206604838371,
-0.19876568019390106,
-0.02703288197517395,
0.08424776047468185,
-0.20383712649345398,
-0.032998621463775635,
0.041543323546648026,
-0.03834589570760727,
-0.02396267279982567,
-0.002415500348433852,
0.06396626681089401,
-0.008327016606926918,
0.12156640738248825,
0.06747189164161682,
0.10266115516424179,
-0.09284433722496033,
0.08920657634735107,
0.10416955500841141,
-0.09140542894601822,
0.03545991703867912,
0.10264154523611069,
-0.05670900270342827,
-0.04460543021559715,
0.033935222774744034,
0.05925208330154419,
-0.028357384726405144,
-0.06409841030836105,
-0.000502707262057811,
-0.0359574519097805,
0.04993389546871185,
0.08058220148086548,
0.036113787442445755,
-0.01202210783958435,
0.06544706225395203,
0.028145326301455498,
-0.11693570017814636,
0.10949387401342392,
0.04405685141682625,
0.04509059712290764,
-0.07182393968105316,
-0.012280966155230999,
0.015999672934412956,
0.032540347427129745,
-0.019734015688300133,
-0.014576527290046215,
-0.03146412968635559,
-0.007561005651950836,
-0.1553635597229004,
-0.02064543403685093,
-0.06516171246767044,
0.006067827809602022,
0.022207623347640038,
-0.03830232471227646,
-0.012014663778245449,
0.01381110493093729,
-0.07979435473680496,
-0.07571027427911758,
-0.01700955256819725,
0.08539021760225296,
-0.1381402313709259,
0.006627439055591822,
0.07182712107896805,
-0.10980239510536194,
0.07347989827394485,
-0.0048679932951927185,
0.017079560086131096,
0.010923396795988083,
-0.11654401570558548,
0.04386281594634056,
-0.005810429807752371,
0.01551580335944891,
0.022556742653250694,
-0.171111062169075,
0.011553828604519367,
-0.038553636521101,
-0.03114982508122921,
0.011926400475203991,
-0.025060230866074562,
-0.11875922232866287,
0.08676479011774063,
-0.028097305446863174,
-0.037512701004743576,
-0.03292486071586609,
0.06296087801456451,
0.08736220002174377,
-0.011740099638700485,
0.09667140990495682,
-0.025766119360923767,
0.04818311333656311,
-0.1756584197282791,
-0.01910574547946453,
-0.050167568027973175,
0.02537350542843342,
-0.01759655587375164,
-0.0070639788173139095,
0.055272240191698074,
-0.004191063344478607,
0.20991376042366028,
-0.03921036794781685,
0.1548677533864975,
0.05199402943253517,
-0.009925156831741333,
0.010884369723498821,
0.05032730847597122,
0.06423956155776978,
0.031145188957452774,
0.00853167474269867,
0.04660189896821976,
-0.004552975296974182,
-0.020357951521873474,
-0.13699717819690704,
0.02791593410074711,
0.16117429733276367,
0.061918217688798904,
0.0392887257039547,
0.03704594820737839,
-0.1422400325536728,
-0.09538721293210983,
0.10306388139724731,
-0.0331864058971405,
0.014331420883536339,
-0.08317886292934418,
0.17621558904647827,
0.12328410148620605,
-0.1574767529964447,
0.0577850341796875,
-0.07234696298837662,
-0.05066767707467079,
-0.1024852767586708,
-0.11832084506750107,
-0.06293155997991562,
-0.06027044355869293,
-0.004747506696730852,
-0.042489297688007355,
0.05734556168317795,
0.026751231402158737,
-0.003270963439717889,
-0.006759525276720524,
0.12665949761867523,
-0.0249644722789526,
-0.004145825747400522,
0.04152364656329155,
0.0326087586581707,
0.019319625571370125,
-0.05872373282909393,
0.017997145652770996,
0.018602589145302773,
0.022180357947945595,
0.06835069507360458,
0.0260987039655447,
-0.059317342936992645,
0.044286735355854034,
0.00319746439345181,
-0.11313364654779434,
0.018146557733416557,
-0.00002245741598017048,
-0.05020225793123245,
0.13557326793670654,
0.04076748713850975,
0.01548024732619524,
-0.029270920902490616,
0.24342355132102966,
-0.07199113070964813,
-0.08681939542293549,
-0.13965600728988647,
0.11511493474245071,
-0.023563209921121597,
0.03755274787545204,
0.016542524099349976,
-0.12659503519535065,
0.011511262506246567,
0.18531471490859985,
0.12824349105358124,
0.012459068559110165,
-0.007656481582671404,
0.05736639350652695,
-0.0007639875984750688,
-0.05985576659440994,
0.05051197111606598,
0.0664999932050705,
0.16097788512706757,
-0.09069112688302994,
0.0652846097946167,
-0.008405503816902637,
-0.0831485390663147,
-0.027498632669448853,
0.11705785244703293,
-0.022675158455967903,
0.02148384228348732,
-0.03778035193681717,
0.11204422265291214,
-0.052532415837049484,
-0.2719486355781555,
0.02952493168413639,
-0.09503202140331268,
-0.13993041217327118,
-0.02591860294342041,
0.041448429226875305,
-0.03349510580301285,
0.01577647216618061,
0.06254769116640091,
-0.045389387756586075,
0.18837277591228485,
0.025987716391682625,
-0.08679025620222092,
-0.07755549252033234,
0.05874146893620491,
-0.08695939928293228,
0.2789687216281891,
0.003863075515255332,
0.04782010242342949,
0.12108923494815826,
-0.03053574077785015,
-0.18664880096912384,
0.014769754372537136,
0.11989909410476685,
-0.09114406257867813,
0.07780203968286514,
0.18139931559562683,
-0.005561648402363062,
0.12649618089199066,
0.04705416411161423,
-0.03877115994691849,
0.03976387158036232,
-0.02721380814909935,
-0.03821522742509842,
-0.12209630757570267,
0.05661242455244064,
-0.0612691193819046,
0.15957388281822205,
0.1158948540687561,
-0.05964287370443344,
0.001120698289014399,
-0.06126941740512848,
0.06300627440214157,
0.014774397015571594,
0.12115653604269028,
0.018452486023306847,
-0.2023056596517563,
0.05087360367178917,
-0.03283824771642685,
0.08166342973709106,
-0.254973828792572,
-0.08186668157577515,
0.07622263580560684,
-0.019022729247808456,
-0.04275642707943916,
0.12311509251594543,
0.06101066991686821,
0.03676839917898178,
-0.03853875398635864,
-0.08537755906581879,
-0.01412904355674982,
0.15376435220241547,
-0.14123432338237762,
-0.029574336484074593
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# vit-base-patch16-224-in21k-finetuned-mgasior-07-02-2024
This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the imagefolder dataset.
It achieves the following results on the evaluation set:
- Loss: 0.8842
- F1: 0.7717
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 10
### Training results
| Training Loss | Epoch | Step | Validation Loss | F1 |
|:-------------:|:-----:|:----:|:---------------:|:------:|
| 1.731 | 0.98 | 35 | 1.6748 | 0.3386 |
| 1.5196 | 1.99 | 71 | 1.4890 | 0.4173 |
| 1.3727 | 2.99 | 107 | 1.2938 | 0.5276 |
| 1.2194 | 4.0 | 143 | 1.1519 | 0.6457 |
| 1.1538 | 4.98 | 178 | 1.0544 | 0.6693 |
| 1.0379 | 5.99 | 214 | 0.9852 | 0.7165 |
| 1.0232 | 6.99 | 250 | 0.9439 | 0.7323 |
| 0.9586 | 8.0 | 286 | 0.9136 | 0.7480 |
| 0.9374 | 8.98 | 321 | 0.8946 | 0.7638 |
| 0.96 | 9.79 | 350 | 0.8842 | 0.7717 |
### Framework versions
- Transformers 4.36.1
- Pytorch 2.1.2+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["imagefolder"], "metrics": ["f1"], "base_model": "google/vit-base-patch16-224-in21k", "model-index": [{"name": "vit-base-patch16-224-in21k-finetuned-mgasior-07-02-2024", "results": [{"task": {"type": "image-classification", "name": "Image Classification"}, "dataset": {"name": "imagefolder", "type": "imagefolder", "config": "default", "split": "train", "args": "default"}, "metrics": [{"type": "f1", "value": 0.7716535433070866, "name": "F1"}]}]}]} | image-classification | MichalGas/vit-base-patch16-224-in21k-finetuned-mgasior-07-02-2024 | [
"transformers",
"safetensors",
"vit",
"image-classification",
"generated_from_trainer",
"dataset:imagefolder",
"base_model:google/vit-base-patch16-224-in21k",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T17:22:14+00:00 | [] | [] | TAGS
#transformers #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| vit-base-patch16-224-in21k-finetuned-mgasior-07-02-2024
=======================================================
This model is a fine-tuned version of google/vit-base-patch16-224-in21k on the imagefolder dataset.
It achieves the following results on the evaluation set:
* Loss: 0.8842
* F1: 0.7717
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* gradient\_accumulation\_steps: 4
* total\_train\_batch\_size: 32
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* num\_epochs: 10
### Training results
### Framework versions
* Transformers 4.36.1
* Pytorch 2.1.2+cu121
* Datasets 2.15.0
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.36.1\n* Pytorch 2.1.2+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.36.1\n* Pytorch 2.1.2+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
82,
144,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 10### Training results### Framework versions\n\n\n* Transformers 4.36.1\n* Pytorch 2.1.2+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
-0.11099163442850113,
0.14680173993110657,
-0.002388669177889824,
0.0891307145357132,
0.14104673266410828,
0.031817927956581116,
0.11645951122045517,
0.12510956823825836,
-0.06695397943258286,
0.11624473333358765,
0.12647736072540283,
0.10861365497112274,
0.06364066153764725,
0.15542729198932648,
-0.011984331533312798,
-0.2837429344654083,
0.019305652007460594,
-0.015301675535738468,
-0.10857073962688446,
0.12288821488618851,
0.0760834813117981,
-0.12411881983280182,
0.09241530299186707,
-0.0010382444597780704,
-0.14315961301326752,
-0.02809646539390087,
-0.015292729251086712,
-0.040818799287080765,
0.10720878094434738,
0.0297124981880188,
0.08825106173753738,
0.02777426689863205,
0.10231825709342957,
-0.23320406675338745,
0.008368182927370071,
0.07842575758695602,
0.0010983728570863605,
0.08412814140319824,
0.10523341596126556,
-0.0059121823869645596,
0.13031883537769318,
-0.10239294171333313,
0.062142785638570786,
0.046927716583013535,
-0.09473482519388199,
-0.22632630169391632,
-0.08344335854053497,
0.08172975480556488,
0.12844447791576385,
0.07103852927684784,
-0.018419889733195305,
0.07853987067937851,
-0.08618715405464172,
0.07593953609466553,
0.21624648571014404,
-0.27284857630729675,
-0.08085235953330994,
0.05025206506252289,
0.024011623114347458,
0.057976480573415756,
-0.12799769639968872,
-0.002227392513304949,
0.04253978654742241,
0.00018790802278090268,
0.12029154598712921,
0.029433725401759148,
0.051165636628866196,
0.0005456320359371603,
-0.15106570720672607,
-0.04843393713235855,
0.11481690406799316,
0.11878913640975952,
-0.022370386868715286,
-0.09051397442817688,
-0.05093376338481903,
-0.19975468516349792,
-0.04684748500585556,
0.007699128706008196,
0.0403042770922184,
-0.04701338708400726,
-0.07973885536193848,
0.02949800342321396,
-0.07425110787153244,
-0.08035310357809067,
0.040583983063697815,
0.10623103380203247,
0.06567848473787308,
-0.020341243594884872,
0.024974705651402473,
0.11948041617870331,
0.05537186935544014,
-0.15153717994689941,
-0.0008182500605471432,
0.009102714248001575,
-0.06782384216785431,
-0.019678035750985146,
0.0023815236054360867,
-0.001535150920972228,
0.030977299436926842,
0.1321055293083191,
-0.026814494282007217,
0.05872294679284096,
0.040131039917469025,
0.02916114777326584,
-0.09611589461565018,
0.15908589959144592,
-0.08196283131837845,
-0.08702980726957321,
-0.013129225932061672,
0.12683862447738647,
0.04033958539366722,
0.0028833053074777126,
-0.08194226771593094,
0.025794655084609985,
0.10941782593727112,
0.04156755283474922,
-0.0064295195043087006,
0.04131113365292549,
-0.0727660059928894,
-0.025606004521250725,
0.06945984810590744,
-0.07430903613567352,
0.04463515430688858,
0.02881617657840252,
-0.06377959251403809,
-0.023200880736112595,
0.02445306070148945,
-0.0002903015702031553,
0.0026049541775137186,
0.11544014513492584,
-0.10556725412607193,
-0.009329017251729965,
-0.07666745036840439,
-0.09000077098608017,
0.02848777361214161,
-0.08031561970710754,
0.007223078515380621,
-0.09711159020662308,
-0.10902424901723862,
-0.04063575714826584,
0.06626798957586288,
-0.05342884361743927,
-0.071344755589962,
-0.056363314390182495,
-0.09120190143585205,
0.044558264315128326,
0.010264821350574493,
0.09706567972898483,
-0.061623621731996536,
0.10803624987602234,
0.008502368815243244,
0.07431583851575851,
0.07377146929502487,
0.04242011904716492,
-0.07240095734596252,
0.05866685509681702,
-0.17214274406433105,
0.04446640610694885,
-0.08722440153360367,
0.07695168256759644,
-0.13110800087451935,
-0.10669280588626862,
-0.012251186184585094,
-0.02427753433585167,
0.070447638630867,
0.14622783660888672,
-0.1380927860736847,
-0.07313431054353714,
0.16639705002307892,
-0.08906084299087524,
-0.1347675323486328,
0.12507975101470947,
-0.006633400917053223,
-0.052889104932546616,
0.02698660083115101,
0.14208318293094635,
0.0850701853632927,
-0.10407929867506027,
-0.02967197448015213,
-0.033173419535160065,
0.0873388722538948,
0.004168146289885044,
0.1137678250670433,
-0.008898437954485416,
-0.0315680168569088,
0.009269576519727707,
-0.07887349277734756,
0.08374518901109695,
-0.10225139558315277,
-0.08874578028917313,
-0.03805190324783325,
-0.08897143602371216,
0.050288617610931396,
0.06111414358019829,
0.026781175285577774,
-0.0791238471865654,
-0.1349036693572998,
-0.012114878743886948,
0.11420916020870209,
-0.08454882353544235,
-0.0008916535298340023,
-0.05267678573727608,
0.10018383711576462,
-0.07060597836971283,
0.0005105487653054297,
-0.1145375445485115,
-0.0733432024717331,
0.03925085812807083,
-0.0705658346414566,
-0.010633752681314945,
-0.038757357746362686,
0.0636327937245369,
0.09135114401578903,
-0.06894964724779129,
-0.09264007955789566,
-0.0686381384730339,
-0.0011633981484919786,
-0.07489699125289917,
-0.25293684005737305,
-0.06725603342056274,
-0.019648080691695213,
0.16867081820964813,
-0.26469093561172485,
0.018909722566604614,
-0.008830269798636436,
0.12226960062980652,
0.04472707584500313,
-0.053802769631147385,
-0.0028959843330085278,
0.02552606165409088,
-0.04951531067490578,
-0.08986333012580872,
0.036915816366672516,
-0.006005244795233011,
-0.09027501195669174,
-0.028889544308185577,
-0.09393157064914703,
0.14405184984207153,
0.11304983496665955,
0.002175629371777177,
-0.09261512011289597,
-0.04470153525471687,
-0.07230323553085327,
-0.0491100549697876,
-0.040772803127765656,
0.027196161448955536,
0.052808910608291626,
0.012492022477090359,
0.10987932980060577,
-0.07789342850446701,
-0.026299238204956055,
0.054354313760995865,
-0.0017257372383028269,
-0.030935518443584442,
0.13170145452022552,
0.12083594501018524,
-0.09437016397714615,
0.13445992767810822,
0.14725370705127716,
-0.05907148867845535,
0.10629948228597641,
-0.04085874930024147,
-0.09327662736177444,
-0.029796581715345383,
0.037548523396253586,
0.028250209987163544,
0.1431284248828888,
-0.13378433883190155,
0.0034548889379948378,
0.019223561510443687,
0.008073687553405762,
-0.004715223331004381,
-0.18063673377037048,
-0.0259553249925375,
0.03654216602444649,
-0.05183875933289528,
-0.003923466429114342,
-0.025577226653695107,
-0.024964623153209686,
0.09480266273021698,
0.022810574620962143,
-0.042878907173871994,
-0.001296601607464254,
0.004241501912474632,
-0.0826677456498146,
0.20819368958473206,
-0.08580699563026428,
-0.1370285302400589,
-0.12337398529052734,
0.018122753128409386,
-0.057805754244327545,
0.001772116171196103,
0.03695359081029892,
-0.11833620071411133,
-0.04770127683877945,
-0.07785829901695251,
0.03220876306295395,
-0.009969709441065788,
0.04396718740463257,
0.005589146167039871,
0.02029959112405777,
0.07703692466020584,
-0.07834165543317795,
0.016981065273284912,
-0.010140893049538136,
-0.04137851297855377,
0.031490545719861984,
0.046275392174720764,
0.12454988062381744,
0.1382862776517868,
0.018618343397974968,
0.02217618003487587,
-0.018061837181448936,
0.1967037320137024,
-0.09360409528017044,
0.010485372506082058,
0.10945205390453339,
0.02944435365498066,
0.05076094716787338,
0.1581064909696579,
0.05148005485534668,
-0.084657222032547,
0.023484541103243828,
0.04731404781341553,
-0.011240300722420216,
-0.20404858887195587,
-0.037335652858018875,
-0.03837137669324875,
-0.0009401572751812637,
0.14014337956905365,
0.04393558204174042,
-0.005126736126840115,
0.07677857577800751,
-0.027657995000481606,
0.02605612762272358,
-0.004976293072104454,
0.08072619140148163,
0.03690352290868759,
0.04552891105413437,
0.10299507528543472,
-0.02822260744869709,
-0.016033314168453217,
0.035279273986816406,
-0.0051061552949249744,
0.22531935572624207,
-0.01660481095314026,
0.11191971600055695,
0.047737982124090195,
0.19014276564121246,
0.005536040756851435,
0.05289602279663086,
0.01978461630642414,
-0.036025792360305786,
0.011278828606009483,
-0.06101910024881363,
-0.028097301721572876,
0.05257004499435425,
0.02897782064974308,
0.06722159683704376,
-0.12175432592630386,
0.05326873064041138,
0.0555620938539505,
0.26852190494537354,
0.08983088284730911,
-0.34693068265914917,
-0.10154570639133453,
0.00888350885361433,
-0.03081003949046135,
-0.043432798236608505,
0.025158997625112534,
0.11976714432239532,
-0.0839310735464096,
0.0650630071759224,
-0.08948901295661926,
0.07211142033338547,
-0.06433640420436859,
-0.008767208084464073,
0.08410166203975677,
0.10418309271335602,
0.0011001832317560911,
0.07438094913959503,
-0.21285387873649597,
0.2770532965660095,
-0.008710571564733982,
0.0582263357937336,
-0.03983534872531891,
0.020173028111457825,
0.03413504362106323,
0.06264939159154892,
0.12023412436246872,
0.0036286981776356697,
-0.027721630409359932,
-0.1932068169116974,
-0.10893289744853973,
0.013858278281986713,
0.10754948109388351,
-0.09493522346019745,
0.11406251788139343,
-0.03688288852572441,
-0.046962130814790726,
0.04119940847158432,
-0.032492831349372864,
-0.11637858301401138,
-0.10660608112812042,
-0.0027956811245530844,
-0.03424980118870735,
0.0709882602095604,
-0.10074460506439209,
-0.10455404222011566,
-0.09991693496704102,
0.16107137501239777,
-0.08987632393836975,
-0.020316410809755325,
-0.13595546782016754,
0.10930123925209045,
0.12007834762334824,
-0.09025325626134872,
0.07050460577011108,
-0.020040027797222137,
0.12262368947267532,
0.03079339489340782,
-0.038488101214170456,
0.10664056241512299,
-0.09364580363035202,
-0.21307243406772614,
-0.0637563019990921,
0.13056206703186035,
0.022800015285611153,
0.04784345626831055,
-0.01838323473930359,
0.0184188149869442,
-0.006650642026215792,
-0.08767229318618774,
0.06214703246951103,
0.0455070361495018,
0.04776738956570625,
0.03506264463067055,
-0.04263097792863846,
0.010149550624191761,
-0.0541665256023407,
-0.04871978610754013,
0.10781064629554749,
0.28350168466567993,
-0.1079254299402237,
0.007757497951388359,
0.04614360257983208,
-0.03625434637069702,
-0.17461222410202026,
0.02030513621866703,
0.0971745029091835,
0.013882782310247421,
0.024105587974190712,
-0.18042491376399994,
0.08909256011247635,
0.09044869989156723,
-0.030348040163517,
0.09367502480745316,
-0.3052923083305359,
-0.1161341518163681,
0.07962033897638321,
0.15367962419986725,
-0.008705180138349533,
-0.16548623144626617,
-0.05601103976368904,
-0.00554960360750556,
-0.09212552756071091,
0.08973179757595062,
-0.009403272531926632,
0.1059274971485138,
-0.02896326594054699,
-0.020396709442138672,
0.02225113846361637,
-0.06913960725069046,
0.1439555287361145,
-0.02398141473531723,
0.07417996972799301,
-0.030520427972078323,
0.0031340341083705425,
0.004352340009063482,
-0.08151198923587799,
0.031007960438728333,
-0.08535720407962799,
0.05704532191157341,
-0.07971265912055969,
-0.014703999273478985,
-0.08375750482082367,
0.028481921181082726,
-0.04648309946060181,
-0.04136914014816284,
-0.03824235126376152,
0.057045310735702515,
0.08180145919322968,
-0.0019612167961895466,
0.16688351333141327,
0.02243894711136818,
0.1245027482509613,
0.07191624492406845,
0.03788413107395172,
-0.014556854963302612,
-0.09746120870113373,
-0.03955703601241112,
-0.018392618745565414,
0.0661277323961258,
-0.15854144096374512,
0.017241548746824265,
0.11838150769472122,
0.041382186114788055,
0.13787271082401276,
0.05646827816963196,
-0.047296348959207535,
0.0017419200157746673,
0.09100620448589325,
-0.11262565851211548,
-0.15695075690746307,
-0.03148537874221802,
-0.019874099642038345,
-0.14757463335990906,
0.03373774141073227,
0.07874227315187454,
-0.08141182363033295,
0.0013852780684828758,
-0.00010915321036009118,
0.05128360912203789,
-0.012059679254889488,
0.17652922868728638,
0.06957608461380005,
0.07303227484226227,
-0.08941919356584549,
0.10534349828958511,
0.05412605032324791,
-0.16092441976070404,
0.0053664445877075195,
0.06681333482265472,
-0.08615023642778397,
-0.0271084476262331,
0.0643877163529396,
0.09060133248567581,
-0.0074477125890553,
-0.03125441446900368,
-0.1143479198217392,
-0.11987661570310593,
0.06824829429388046,
0.07113276422023773,
0.06235497072339058,
0.02499077469110489,
-0.006466810591518879,
0.036620981991291046,
-0.11383239179849625,
0.127899631857872,
0.0691811591386795,
0.10289660096168518,
-0.21142703294754028,
0.07053983956575394,
0.01190478540956974,
0.01964174397289753,
-0.008595526218414307,
0.03767357021570206,
-0.1253463178873062,
-0.022078197449445724,
-0.06257499009370804,
-0.009812561795115471,
-0.07962306588888168,
0.007878676988184452,
-0.002198439557105303,
-0.05667885020375252,
-0.048415657132864,
0.014749700203537941,
-0.09547987580299377,
-0.0634436383843422,
0.005827364511787891,
0.08246766775846481,
-0.11210866272449493,
-0.00808731373399496,
0.03396213427186012,
-0.11801006644964218,
0.09699109196662903,
0.01959936134517193,
0.053957004100084305,
0.018545720726251602,
-0.11830097436904907,
0.024402130395174026,
0.06161430478096008,
-0.013412706553936005,
0.03031768836081028,
-0.14179523289203644,
0.007710699923336506,
-0.049470748752355576,
-0.010151103138923645,
-0.015369579195976257,
0.059584155678749084,
-0.139268696308136,
-0.01685589738190174,
-0.04652709141373634,
-0.04767191782593727,
-0.05146695300936699,
0.05589336156845093,
0.0642327219247818,
-0.028119787573814392,
0.17996661365032196,
-0.0775456354022026,
0.03432192653417587,
-0.2358083426952362,
-0.006337660830467939,
-0.023288333788514137,
-0.0692276880145073,
-0.10366875678300858,
-0.01699410006403923,
0.07000577449798584,
-0.05628800392150879,
0.06530912965536118,
-0.02154809981584549,
0.041209444403648376,
0.027525391429662704,
-0.02517496794462204,
0.03252524137496948,
0.043902620673179626,
0.1763312816619873,
0.02360369637608528,
-0.022307217121124268,
0.04486064985394478,
0.011678846552968025,
0.09102940559387207,
0.06448180228471756,
0.14517536759376526,
0.15017127990722656,
-0.03955008462071419,
0.09004101157188416,
0.045447882264852524,
-0.11689123511314392,
-0.1782682240009308,
0.1214674860239029,
-0.08227246254682541,
0.13953882455825806,
-0.015729788690805435,
0.1748497486114502,
0.11890687793493271,
-0.19003792107105255,
0.013045535422861576,
-0.02116197720170021,
-0.07602237164974213,
-0.08109234273433685,
-0.10310105234384537,
-0.09418823570013046,
-0.20281553268432617,
0.025305472314357758,
-0.10013182461261749,
0.016995307058095932,
0.05905737727880478,
0.019040415063500404,
0.011051778681576252,
0.17307285964488983,
0.07565123587846756,
0.015135074034333229,
0.0708034336566925,
0.040393535047769547,
-0.05161606892943382,
-0.020259564742445946,
-0.08469664305448532,
0.024091491475701332,
-0.053452812135219574,
0.03946103900671005,
-0.05409488081932068,
-0.08310635387897491,
0.09028694778680801,
0.040266118943691254,
-0.10136358439922333,
0.032576944679021835,
-0.016479412093758583,
0.0339084267616272,
0.07079895585775375,
0.003111958270892501,
0.009155635721981525,
-0.02783176302909851,
0.22383572161197662,
-0.08345537632703781,
-0.014198290184140205,
-0.12071472406387329,
0.21289274096488953,
0.0072785778902471066,
-0.008064284920692444,
0.026126787066459656,
-0.08835940808057785,
0.007184440270066261,
0.17422375082969666,
0.1460493952035904,
-0.02136056497693062,
-0.022993287071585655,
0.027370024472475052,
-0.02035476081073284,
-0.04752257466316223,
0.09093748778104782,
0.09851313382387161,
0.04551864415407181,
-0.06270401924848557,
-0.02021930180490017,
-0.04752349480986595,
-0.04736197367310524,
-0.02399955876171589,
0.05511214956641197,
0.029502563178539276,
-0.003686631331220269,
-0.0463624931871891,
0.09151767194271088,
-0.03409997373819351,
-0.09954833984375,
0.10671655088663101,
-0.18322306871414185,
-0.17222878336906433,
-0.052756015211343765,
0.09868869185447693,
0.009347602725028992,
0.04893624782562256,
-0.005409672856330872,
-0.031502991914749146,
0.09902653098106384,
-0.005075524095445871,
-0.07196159660816193,
-0.1153038963675499,
0.040558259934186935,
-0.03699779510498047,
0.25640326738357544,
-0.03602456673979759,
-0.01611071638762951,
0.1268184334039688,
0.035594649612903595,
-0.12878647446632385,
0.03407139703631401,
0.07367143779993057,
-0.09551205486059189,
0.049336522817611694,
0.1476576030254364,
-0.03350483253598213,
0.12256106734275818,
0.03891872614622116,
-0.09304796159267426,
-0.0002428438310744241,
-0.08209913969039917,
-0.055321384221315384,
-0.060214750468730927,
0.016584455966949463,
-0.03783170506358147,
0.14729441702365875,
0.20685076713562012,
-0.05974012240767479,
-0.022983478382229805,
-0.05785573273897171,
0.03829168528318405,
0.059940092265605927,
0.117233045399189,
0.006008483469486237,
-0.24066968262195587,
0.029709341004490852,
-0.015545625239610672,
0.024238737300038338,
-0.2112077921628952,
-0.09021497517824173,
0.01967574842274189,
-0.04183826595544815,
-0.10874681919813156,
0.10775049030780792,
0.09761664271354675,
0.04231175780296326,
-0.05898182839155197,
-0.09532817453145981,
-0.050129055976867676,
0.1748805046081543,
-0.1587899774312973,
-0.08369410783052444
] |
null | null | fastai |
# Amazing!
🥳 Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the [documentation here](https://huggingface.co/docs/hub/model-repos))!
2. Create a demo in Gradio or Streamlit using 🤗 Spaces ([documentation here](https://huggingface.co/docs/hub/spaces)).
3. Join the fastai community on the [Fastai Discord](https://discord.com/invite/YKrxeNn)!
Greetings fellow fastlearner 🤝! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| {"tags": ["fastai"]} | null | pamunarr/intel-image-classification | [
"fastai",
"has_space",
"region:us"
] | 2024-02-07T17:24:06+00:00 | [] | [] | TAGS
#fastai #has_space #region-us
|
# Amazing!
Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the documentation here)!
2. Create a demo in Gradio or Streamlit using Spaces (documentation here).
3. Join the fastai community on the Fastai Discord!
Greetings fellow fastlearner ! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| [
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
"TAGS\n#fastai #has_space #region-us \n",
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
13,
20,
79,
3,
6,
12,
8
] | [
"passage: TAGS\n#fastai #has_space #region-us \n# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---# Model card## Model description\nMore information needed## Intended uses & limitations\nMore information needed## Training and evaluation data\nMore information needed"
] | [
-0.048121724277734756,
-0.024616125971078873,
0.002038001548498869,
0.10439170897006989,
0.135872021317482,
0.11887997388839722,
0.07405775785446167,
0.09980081021785736,
0.07783667743206024,
0.02590852417051792,
0.08961158245801926,
-0.08088712394237518,
0.08744348585605621,
0.271692156791687,
0.06988707184791565,
-0.22761479020118713,
0.04051019623875618,
-0.00024903909070417285,
0.08053462207317352,
0.06629016250371933,
0.13507555425167084,
-0.05464952811598778,
0.14010503888130188,
-0.004088983871042728,
-0.19050447642803192,
-0.042929794639348984,
-0.01773718371987343,
-0.02527874894440174,
0.12317648530006409,
-0.04744937643408775,
0.05381017178297043,
0.015037551522254944,
0.007565062493085861,
-0.07253646105527878,
0.0623294934630394,
0.040457066148519516,
0.01740180514752865,
0.059235580265522,
-0.07249044626951218,
0.08950132131576538,
0.08404164761304855,
-0.024370938539505005,
-0.1097978875041008,
0.07827875018119812,
-0.14424212276935577,
-0.21762843430042267,
-0.1253085881471634,
-0.09017651528120041,
0.028519365936517715,
0.004388005938380957,
-0.025051530450582504,
0.12801909446716309,
-0.13558274507522583,
-0.040698226541280746,
0.20124278962612152,
-0.17012301087379456,
-0.05505548417568207,
0.034343402832746506,
0.09226689487695694,
-0.05829555168747902,
-0.06347129493951797,
0.10614984482526779,
0.09640881419181824,
-0.019833475351333618,
0.05516824126243591,
0.002579754451289773,
0.021173657849431038,
0.01370104867964983,
-0.06150497496128082,
0.04717832803726196,
-0.010183089412748814,
0.048132527619600296,
-0.09465572983026505,
-0.1303568333387375,
-0.004072192590683699,
0.01214400865137577,
-0.048744890838861465,
-0.07019646465778351,
0.07833103090524673,
-0.011118141002953053,
-0.04357248544692993,
-0.13031910359859467,
-0.09131011366844177,
-0.12358787655830383,
0.008646543137729168,
0.09500427544116974,
0.003679296001791954,
0.07374339550733566,
-0.08258994668722153,
0.06774985045194626,
-0.17329485714435577,
-0.06484591960906982,
-0.08138520270586014,
-0.11546400189399719,
0.021133482456207275,
-0.0387684591114521,
0.02668963186442852,
0.15394504368305206,
0.12983950972557068,
0.023976242169737816,
0.04388163983821869,
-0.038937073200941086,
0.051190316677093506,
0.058571770787239075,
0.03395717963576317,
0.034934818744659424,
-0.036981891840696335,
-0.1793210655450821,
-0.016702448949217796,
-0.011550825089216232,
0.07954040914773941,
-0.07523109763860703,
-0.05632320046424866,
0.013454885222017765,
-0.11071494966745377,
0.07202339172363281,
-0.03576776012778282,
-0.0032025426626205444,
0.01168301422148943,
0.018371861428022385,
0.21271461248397827,
0.03955606371164322,
0.014191740192472935,
-0.008875265717506409,
-0.13453757762908936,
-0.06874168664216995,
-0.06896194815635681,
0.03361047804355621,
0.04448792710900307,
-0.0028071461711078882,
-0.07672245055437088,
0.04325154796242714,
-0.06045534089207649,
-0.03508453071117401,
0.008032378740608692,
-0.18221288919448853,
0.007458044681698084,
-0.10049355030059814,
-0.12126200646162033,
0.05306628718972206,
0.01695440337061882,
-0.08215925842523575,
0.08141279965639114,
0.02662261202931404,
0.020931517705321312,
-0.009988143108785152,
-0.005391082260757685,
0.06874798238277435,
-0.08508864045143127,
0.029901226982474327,
0.17170792818069458,
0.13024519383907318,
-0.08046911656856537,
-0.0006887061172164977,
-0.10965746641159058,
0.04426072910428047,
-0.13325683772563934,
0.02251482754945755,
-0.09062390774488449,
0.11723794043064117,
-0.042396437376737595,
0.002038756385445595,
-0.029030200093984604,
0.0960269495844841,
0.08189879357814789,
0.16663365066051483,
-0.2419009804725647,
-0.031095001846551895,
0.13240347802639008,
-0.10711425542831421,
-0.1807439625263214,
0.18486657738685608,
-0.012035200372338295,
0.11329247802495956,
-0.047014184296131134,
0.18334640562534332,
-0.02612062357366085,
-0.13582459092140198,
-0.058872904628515244,
0.005852419883012772,
-0.2269321084022522,
-0.06286033242940903,
0.09738040715456009,
0.13425657153129578,
-0.042984943836927414,
0.007112155202776194,
0.026316028088331223,
0.13609857857227325,
-0.06715573370456696,
-0.05195777863264084,
-0.012255736626684666,
-0.10902371257543564,
0.041914235800504684,
0.018215661868453026,
0.035408079624176025,
-0.059880174696445465,
-0.02931194379925728,
-0.053190283477306366,
0.13146710395812988,
0.09760832786560059,
-0.03670211136341095,
-0.049620725214481354,
0.1689043790102005,
-0.07763876020908356,
-0.033587727695703506,
0.07560533285140991,
-0.08268500119447708,
0.03266897425055504,
0.03090597130358219,
0.055881720036268234,
0.07766123116016388,
0.08522116392850876,
0.06057543307542801,
0.00819048099219799,
0.034654274582862854,
0.12095347046852112,
-0.013591280207037926,
-0.05039411783218384,
0.021508218720555305,
0.016904234886169434,
-0.019032588228583336,
0.29030677676200867,
-0.1951042115688324,
0.024724548682570457,
-0.06477324664592743,
0.07631538063287735,
0.06136792525649071,
0.003575638635084033,
0.08580143749713898,
-0.06023019179701805,
-0.019061198458075523,
-0.04803973436355591,
0.046805646270513535,
-0.0666879191994667,
-0.04162997007369995,
0.2621194124221802,
-0.05497581139206886,
0.044914912432432175,
0.12313763797283173,
-0.05873025581240654,
-0.07091446220874786,
0.01009807363152504,
-0.00793424155563116,
0.03249288722872734,
-0.04042816907167435,
0.043721720576286316,
-0.10840129852294922,
-0.06674089282751083,
0.1573198139667511,
-0.038477856665849686,
0.06786153465509415,
0.032288823276758194,
-0.04958454892039299,
-0.0648743286728859,
0.04650486260652542,
0.13598160445690155,
-0.0875244215130806,
0.07435166835784912,
0.17612984776496887,
-0.010562662966549397,
0.168031245470047,
0.08435525000095367,
-0.07075224816799164,
-0.09465329349040985,
-0.051014289259910583,
-0.021595727652311325,
0.21222901344299316,
-0.07084725052118301,
-0.054564714431762695,
0.05911700800061226,
-0.013703816570341587,
0.07196151465177536,
-0.06009222939610481,
-0.08332337439060211,
0.03227344527840614,
-0.04517695680260658,
0.011517706327140331,
0.13512636721134186,
-0.07090822607278824,
0.04681389778852463,
0.031489867717027664,
-0.0662703812122345,
0.02217509225010872,
0.033389873802661896,
0.0068921963684260845,
0.033959709107875824,
0.07332495599985123,
-0.20893315970897675,
-0.08408680558204651,
-0.13727638125419617,
0.037881869822740555,
0.021770721301436424,
0.045787326991558075,
-0.08602345734834671,
0.02231026627123356,
-0.08954031765460968,
-0.07987114042043686,
0.029592275619506836,
-0.026350297033786774,
-0.11349643021821976,
-0.03396226093173027,
-0.009560913778841496,
-0.06662604957818985,
-0.02250705659389496,
-0.05024505779147148,
0.03983384370803833,
0.04479299485683441,
0.058377087116241455,
0.12796473503112793,
-0.013808943331241608,
-0.03839317709207535,
0.000370211957488209,
-0.022712308913469315,
0.16396735608577728,
-0.14746315777301788,
0.07954913377761841,
0.19160102307796478,
0.11742953956127167,
0.028144672513008118,
0.028885571286082268,
0.03537585213780403,
-0.06289814412593842,
-0.000050317394197918475,
0.03226194158196449,
-0.09392514824867249,
-0.05801016092300415,
-0.020014392212033272,
-0.04031052812933922,
0.17134574055671692,
-0.12160717695951462,
0.03345204517245293,
0.04098419472575188,
0.09783966839313507,
0.10073629021644592,
-0.028829937800765038,
-0.1815856397151947,
0.038818612694740295,
-0.24060091376304626,
-0.05831146240234375,
0.027899866923689842,
-0.09110201895236969,
-0.06232144311070442,
0.17409387230873108,
0.013794700615108013,
0.011769929900765419,
-0.006736889015883207,
0.07983319461345673,
0.0110100656747818,
0.1217205822467804,
0.05947643890976906,
-0.05539114400744438,
0.025202350690960884,
-0.09962950646877289,
-0.07107596844434738,
-0.04035590961575508,
-0.05832801014184952,
0.07548832893371582,
0.1409129947423935,
-0.025475580245256424,
-0.020795362070202827,
0.023489827290177345,
0.08550169318914413,
0.0423230417072773,
0.16739299893379211,
-0.16016584634780884,
-0.026555389165878296,
0.04571257904171944,
-0.03384667634963989,
-0.05433850735425949,
-0.010291114449501038,
0.1137225553393364,
-0.02820689231157303,
-0.040318265557289124,
0.021242983639240265,
0.06503437459468842,
0.01481706090271473,
0.05012747645378113,
-0.04056356102228165,
0.14796851575374603,
-0.03461192920804024,
0.019330544397234917,
-0.12413888424634933,
0.13848772644996643,
0.021095896139740944,
-0.03901609033346176,
-0.06735876202583313,
-0.05808034539222717,
0.18150931596755981,
0.0025602965615689754,
0.10535930097103119,
0.012098877690732479,
-0.12160047143697739,
-0.1359938681125641,
-0.11211287975311279,
0.005111907608807087,
0.08330471813678741,
-0.023147236555814743,
-0.022247863933444023,
0.022165266796946526,
-0.036149751394987106,
-0.0530381016433239,
0.15749511122703552,
-0.1289154291152954,
-0.001082550617866218,
0.014728817157447338,
0.06971760839223862,
-0.08223173767328262,
0.026267826557159424,
0.014071501791477203,
-0.1119147390127182,
0.10590848326683044,
0.2521335482597351,
0.10338116437196732,
-0.09591643512248993,
-0.07697287201881409,
0.03418830782175064,
-0.012184361927211285,
-0.000774814048781991,
-0.006932659074664116,
0.0495428591966629,
-0.005566445179283619,
0.006762749515473843,
0.12971895933151245,
-0.07130889594554901,
0.011540771462023258,
-0.08449850976467133,
0.05566910281777382,
-0.05276734381914139,
0.01761564053595066,
-0.002672141883522272,
-0.008124710991978645,
-0.07340748608112335,
-0.061829522252082825,
0.1609770804643631,
-0.07277000695466995,
-0.06468547880649567,
0.05801168829202652,
0.03307786211371422,
0.01431563775986433,
-0.03584568202495575,
-0.04342148080468178,
0.18088261783123016,
0.29330700635910034,
-0.08191116154193878,
0.10001859813928604,
0.09677296131849289,
0.034820813685655594,
-0.23625829815864563,
0.029798466712236404,
-0.1455078274011612,
0.04449721798300743,
0.040447335690259933,
-0.0409548319876194,
0.04191497340798378,
0.10835777968168259,
-0.06094440817832947,
0.2048867791891098,
-0.03527235612273216,
-0.07983248680830002,
-0.01788630709052086,
0.03109324350953102,
0.29443636536598206,
-0.11833466589450836,
0.006058716680854559,
-0.10420958697795868,
-0.21566011011600494,
0.06983078271150589,
-0.18948867917060852,
0.13948246836662292,
-0.05087858438491821,
0.03576415032148361,
-0.01149723306298256,
-0.07561972737312317,
0.20518061518669128,
-0.15641045570373535,
0.05273103713989258,
-0.13722458481788635,
-0.1327189952135086,
0.01617460884153843,
-0.10048147290945053,
0.1545477658510208,
-0.11024226248264313,
-0.023215843364596367,
-0.2284185290336609,
0.012587235309183598,
-0.023200806230306625,
0.10030807554721832,
0.01800704374909401,
-0.07980740070343018,
-0.08767345547676086,
0.1316242516040802,
-0.06486566364765167,
0.034810543060302734,
-0.06996636837720871,
-0.050714004784822464,
-0.010929876938462257,
-0.045061707496643066,
0.03034941293299198,
-0.07934719324111938,
0.15192505717277527,
-0.016938980668783188,
-0.04507075995206833,
0.08636019378900528,
-0.2479533851146698,
0.023727843537926674,
0.025351112708449364,
-0.03495599329471588,
0.09001832455396652,
-0.025513244792819023,
-0.06256973743438721,
0.12282291799783707,
0.1402233988046646,
-0.07322840392589569,
-0.2460673749446869,
-0.06281693279743195,
0.0076784128323197365,
0.039165716618299484,
0.06561196595430374,
0.05125982314348221,
-0.07261458039283752,
-0.011131617240607738,
-0.026896944269537926,
0.030595947057008743,
-0.11692017316818237,
-0.03854857385158539,
0.07790639251470566,
0.017095070332288742,
-0.07846562564373016,
0.07280377298593521,
0.014225782826542854,
-0.021511616185307503,
0.007357571739703417,
0.148970365524292,
0.007519228849560022,
-0.14747941493988037,
-0.06656096875667572,
0.2007484883069992,
-0.01197928935289383,
-0.07260087132453918,
-0.05383119732141495,
-0.008990069851279259,
-0.0476234145462513,
0.05585788935422897,
0.05367223918437958,
-0.013585401698946953,
0.07708586007356644,
0.06263149529695511,
-0.10210110992193222,
-0.046256959438323975,
-0.066561758518219,
0.04169114679098129,
-0.10485753417015076,
0.060470130294561386,
0.009529483504593372,
0.12185006588697433,
-0.09983488917350769,
-0.01802929677069187,
-0.10810204595327377,
-0.06766588985919952,
-0.17349553108215332,
-0.05834362283349037,
-0.041105758398771286,
-0.015651104971766472,
0.03658895567059517,
0.010445823892951012,
-0.057867538183927536,
-0.0442853718996048,
-0.07536603510379791,
0.038444988429546356,
0.06147460639476776,
0.03932281583547592,
-0.03912714496254921,
0.04001858830451965,
0.05909334123134613,
0.013087345287203789,
0.17542624473571777,
0.038768354803323746,
0.05504675209522247,
-0.05045998468995094,
-0.16491834819316864,
-0.05276111513376236,
-0.0074316514655947685,
-0.07559102028608322,
0.1224973127245903,
-0.007679440546780825,
0.007880088873207569,
-0.08065467327833176,
0.03924860805273056,
0.028234204277396202,
0.10404064506292343,
-0.0028364830650389194,
0.10070426017045975,
0.019627176225185394,
-0.07226712256669998,
-0.025392837822437286,
0.021809715777635574,
0.12809939682483673,
0.01567147858440876,
0.026090998202562332,
0.033139873296022415,
0.016619985923171043,
-0.057361043989658356,
0.033977724611759186,
-0.04997231811285019,
-0.15123651921749115,
0.02628709189593792,
-0.05165188014507294,
0.005062380339950323,
-0.016889680176973343,
0.20362506806850433,
0.07867538928985596,
-0.06474173814058304,
-0.010664013214409351,
0.015816617757081985,
-0.0168940220028162,
-0.03121885471045971,
-0.012740966863930225,
0.04592578858137131,
-0.001151384087279439,
-0.04866636544466019,
0.11825273931026459,
0.05015748366713524,
0.05386412516236305,
0.0596686452627182,
0.12528513371944427,
0.016759619116783142,
0.13257254660129547,
0.061999931931495667,
-0.03403807803988457,
-0.13461735844612122,
-0.04495539888739586,
-0.1254577934741974,
0.04646851494908333,
-0.08697032928466797,
0.09941662102937698,
0.1144254133105278,
-0.05959030240774155,
-0.030464433133602142,
-0.08851305395364761,
-0.008356761187314987,
-0.06041252240538597,
0.039516255259513855,
-0.02262675203382969,
-0.0873224213719368,
0.0481097511947155,
0.05495472997426987,
-0.022752324119210243,
0.13218675553798676,
0.015727028250694275,
-0.036317698657512665,
0.13270340859889984,
-0.07583184540271759,
0.11758984625339508,
0.061510033905506134,
-0.043043944984674454,
-0.11560922116041183,
-0.020150646567344666,
-0.06641761213541031,
-0.10098972916603088,
-0.006782987620681524,
-0.005399650428444147,
-0.07349002361297607,
-0.059971679002046585,
0.08397487550973892,
-0.03124053031206131,
-0.09979676455259323,
-0.032152675092220306,
0.0038895104080438614,
0.06054706871509552,
-0.01686914451420307,
-0.0034020058810710907,
0.04728743061423302,
0.015076374635100365,
0.1653461456298828,
-0.02208263985812664,
0.06234867498278618,
-0.13855914771556854,
0.16070103645324707,
-0.14684462547302246,
-0.029404424130916595,
-0.1890171319246292,
-0.09729582816362381,
-0.05156542733311653,
0.20326784253120422,
0.2840938866138458,
-0.19109351933002472,
-0.010187864303588867,
0.020078664645552635,
-0.014484191313385963,
-0.08961770683526993,
0.12571553885936737,
0.029420215636491776,
-0.023631498217582703,
-0.07249019294977188,
-0.02037387527525425,
0.005258576478809118,
-0.06544211506843567,
-0.026979785412549973,
0.18310695886611938,
0.001496660872362554,
0.059546373784542084,
-0.09605178982019424,
0.01754261925816536,
-0.14839904010295868,
-0.10467469692230225,
-0.02111995778977871,
-0.16156397759914398,
-0.09646477550268173,
0.006635562051087618,
0.038640011101961136,
0.08000610023736954,
0.03268849849700928,
-0.015172510407865047,
0.06479045748710632,
-0.056333884596824646,
-0.0037216036580502987,
-0.1231912299990654,
0.00034658415825106204,
0.062129102647304535,
-0.07422006875276566,
0.2545335292816162,
-0.03070417232811451,
-0.12370815873146057,
0.09026903659105301,
-0.03299184888601303,
-0.12452623248100281,
0.07951879501342773,
-0.005700904875993729,
-0.11531132459640503,
-0.057989440858364105,
0.18941475450992584,
-0.012821312062442303,
-0.1364315301179886,
0.046368811279535294,
-0.17166484892368317,
0.031349923461675644,
0.0363016203045845,
-0.001313706859946251,
-0.04714022949337959,
0.024538639932870865,
-0.008008457720279694,
0.10724439471960068,
0.1382838785648346,
0.016739921644330025,
-0.011060068383812904,
-0.05056179314851761,
0.07912429422140121,
0.056927867233753204,
-0.05218246951699257,
-0.1282637119293213,
-0.08599764108657837,
0.03429819270968437,
0.04119478166103363,
-0.08113081753253937,
-0.16903182864189148,
-0.03668912500143051,
-0.10082915425300598,
-0.004939202684909105,
0.051785312592983246,
0.06585265696048737,
0.29044589400291443,
0.06326735019683838,
0.0016605621203780174,
-0.13649453222751617,
0.050569336861371994,
0.0868251696228981,
-0.04697931930422783,
-0.07670357078313828
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | delli/mistral-7b-address-validator | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-07T17:25:39+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
31,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06646376848220825,
0.2168014943599701,
-0.00225935154594481,
0.023818302899599075,
0.1271018385887146,
-0.001635765191167593,
0.04218708351254463,
0.13324736058712006,
-0.020175931975245476,
0.11144465953111649,
0.046588581055402756,
0.09377603232860565,
0.09928803145885468,
0.18404334783554077,
0.04859916493296623,
-0.2059975117444992,
0.007056170143187046,
-0.09090408682823181,
0.014076028019189835,
0.1116579994559288,
0.13719257712364197,
-0.10291384905576706,
0.08272874355316162,
-0.04045208916068077,
-0.02019004337489605,
0.00012576708104461432,
-0.09259183704853058,
-0.07032395154237747,
0.06885425746440887,
0.06264153122901917,
0.051234472543001175,
0.001456156256608665,
0.09140396863222122,
-0.2864592671394348,
0.017265573143959045,
0.08406311273574829,
0.0027674848679453135,
0.06290827691555023,
0.07236549258232117,
-0.07389893382787704,
0.11328595131635666,
-0.08021481335163116,
0.13019037246704102,
0.08625296503305435,
-0.062064990401268005,
-0.23071379959583282,
-0.07525765895843506,
0.0963398814201355,
0.12251301854848862,
0.06215599179267883,
-0.022921854630112648,
0.15455181896686554,
-0.06248689442873001,
0.012971068732440472,
0.1294165402650833,
-0.11526761949062347,
-0.05572471022605896,
0.061741601675748825,
0.11775490641593933,
0.10740239918231964,
-0.14110268652439117,
-0.0017287094378843904,
0.04900608956813812,
0.029121357947587967,
0.08589313924312592,
0.022661056369543076,
0.12003941088914871,
0.04652795568108559,
-0.13695219159126282,
-0.04037507623434067,
0.12011898308992386,
0.038862764835357666,
-0.06446044892072678,
-0.2168138176202774,
-0.006778308190405369,
-0.0601806715130806,
-0.014732478186488152,
-0.07019448280334473,
0.039128515869379044,
-0.02470310963690281,
0.07317749410867691,
-0.04465159401297569,
-0.1063927412033081,
-0.0421026237308979,
0.0892222449183464,
0.07748593389987946,
0.011527054943144321,
-0.02519804798066616,
0.04627908393740654,
0.13455867767333984,
0.05402068421244621,
-0.10399353504180908,
-0.07017925381660461,
-0.06942764669656754,
-0.09420394152402878,
-0.04035796597599983,
0.056760527193546295,
0.031942449510097504,
0.02665667235851288,
0.22703726589679718,
0.016653569415211678,
0.04155244305729866,
0.0224777739495039,
0.01032855175435543,
0.043662428855895996,
0.0955500528216362,
-0.05303520709276199,
-0.15660029649734497,
-0.04072032496333122,
0.09077946096658707,
-0.0027527001220732927,
-0.036689214408397675,
-0.03966725245118141,
0.03849169611930847,
0.06843466311693192,
0.13122352957725525,
0.07552056759595871,
-0.017929591238498688,
-0.04813180863857269,
-0.030096933245658875,
0.23523783683776855,
-0.1493375599384308,
0.04426715523004532,
-0.02271856553852558,
-0.01804111897945404,
-0.03908449783921242,
0.03597262129187584,
0.022118929773569107,
-0.000004518366949923802,
0.09706240892410278,
-0.058981191366910934,
-0.05378659814596176,
-0.10168042778968811,
-0.03272576630115509,
0.04088849574327469,
-0.013975566253066063,
-0.010589460842311382,
-0.09025166928768158,
-0.09490354359149933,
-0.04766594246029854,
0.05537205561995506,
-0.05123869329690933,
-0.03770573064684868,
0.009465423412621021,
-0.08151785284280777,
-0.005444355774670839,
-0.005417742300778627,
0.10699385404586792,
-0.03222226724028587,
0.04445803165435791,
-0.027600755915045738,
0.05225523188710213,
0.09919606149196625,
0.031576547771692276,
-0.0773419588804245,
0.0561848059296608,
-0.22559374570846558,
0.07503069192171097,
-0.11481974273920059,
0.04335082694888115,
-0.1704932004213333,
-0.042439818382263184,
0.005444696638733149,
0.0139949731528759,
0.013206101022660732,
0.12720820307731628,
-0.19255615770816803,
-0.01654396951198578,
0.13260798156261444,
-0.09212633967399597,
-0.118110790848732,
0.07884611934423447,
-0.029701577499508858,
0.1624738723039627,
0.04682036489248276,
-0.027025915682315826,
0.09224298596382141,
-0.16434773802757263,
-0.07092688232660294,
-0.00949116237461567,
-0.01727987825870514,
0.12109188735485077,
0.07512219995260239,
-0.05991523340344429,
0.046571120619773865,
0.02832140028476715,
-0.038078423589468,
-0.04424772411584854,
-0.050857074558734894,
-0.10884185880422592,
-0.01070026308298111,
-0.08987759798765182,
0.04065500199794769,
-0.01250192429870367,
-0.07916021347045898,
-0.029885273426771164,
-0.18612512946128845,
-0.0030564051121473312,
0.10038342326879501,
0.0035033065360039473,
-0.005652366206049919,
-0.08666291832923889,
0.026358824223279953,
-0.03112892620265484,
-0.008404186926782131,
-0.16764774918556213,
-0.04399421438574791,
0.046902090311050415,
-0.16094985604286194,
0.020117372274398804,
-0.06413903087377548,
0.06334125250577927,
0.03641495108604431,
-0.05590536445379257,
-0.0248766727745533,
-0.01730942726135254,
0.011945613659918308,
-0.05083848536014557,
-0.18994836509227753,
-0.056277405470609665,
-0.037882111966609955,
0.149809330701828,
-0.25956398248672485,
0.032966937869787216,
0.051140617579221725,
0.14649195969104767,
0.00406361510977149,
-0.05115427449345589,
0.01429014839231968,
-0.05360214412212372,
-0.054652128368616104,
-0.06746816635131836,
-0.006135428790003061,
-0.027576493099331856,
-0.05147203803062439,
0.019243421033024788,
-0.1755700707435608,
-0.021410830318927765,
0.09424154460430145,
0.12876708805561066,
-0.1486445665359497,
-0.018640631809830666,
-0.048725154250860214,
-0.06339836865663528,
-0.0715010017156601,
-0.07038594037294388,
0.10712739825248718,
0.0513901449739933,
0.04796046018600464,
-0.07435787469148636,
-0.07092321664094925,
0.02726263552904129,
0.006906150374561548,
-0.03382374346256256,
0.08727246522903442,
0.05199531093239784,
-0.09209315478801727,
0.0756213590502739,
0.1092359870672226,
0.07177663594484329,
0.09363535046577454,
0.01574566215276718,
-0.11756632477045059,
-0.028492970392107964,
0.036266472190618515,
0.02740776725113392,
0.1465986967086792,
-0.05952361226081848,
0.04016614332795143,
0.04494241625070572,
-0.04170418903231621,
0.022319864481687546,
-0.08787637203931808,
0.024075502529740334,
0.025203049182891846,
-0.0034381982404738665,
0.06284574419260025,
-0.02525499276816845,
-0.0050758360885083675,
0.07016654312610626,
0.047779910266399384,
0.04621000960469246,
0.009655474685132504,
-0.01720241829752922,
-0.1047825813293457,
0.16950392723083496,
-0.0951867327094078,
-0.269941508769989,
-0.17632324993610382,
0.026197833940386772,
0.04035249724984169,
-0.022378476336598396,
0.031619444489479065,
-0.07056326419115067,
-0.10630585998296738,
-0.1060405746102333,
-0.002429972169920802,
0.01714223250746727,
-0.06364088505506516,
-0.0741225928068161,
0.07348573952913284,
0.04382912442088127,
-0.14902326464653015,
0.038552410900592804,
0.055694397538900375,
-0.057955220341682434,
-0.0233661737293005,
0.09118817001581192,
0.12397737801074982,
0.14583967626094818,
-0.021366750821471214,
-0.028626007959246635,
0.029004426673054695,
0.19620531797409058,
-0.13469526171684265,
0.10371150821447372,
0.13814030587673187,
-0.04545360431075096,
0.08360563963651657,
0.1560150384902954,
0.029186224564909935,
-0.08317049592733383,
0.05044832453131676,
0.04082648828625679,
-0.043159641325473785,
-0.2666129767894745,
-0.0534592866897583,
0.012832709588110447,
-0.06255637854337692,
0.09786593168973923,
0.10183793306350708,
0.11542957276105881,
0.034910861402750015,
-0.07166364789009094,
-0.043925940990448,
-0.0058974819257855415,
0.11737963557243347,
-0.05490213260054588,
-0.012639665976166725,
0.07686592638492584,
-0.05086168646812439,
0.005355054512619972,
0.10266812145709991,
0.02973790094256401,
0.17442677915096283,
0.020399179309606552,
0.11231429129838943,
0.06195578724145889,
0.08633565157651901,
0.0007386076031252742,
0.02951662428677082,
0.05147615820169449,
0.017203815281391144,
-0.002300140680745244,
-0.10421168059110641,
-0.006156572140753269,
0.1449710875749588,
0.028103826567530632,
0.029669636860489845,
-0.0018948549404740334,
-0.005003341939300299,
0.05121048167347908,
0.1746254414319992,
-0.011592294089496136,
-0.22072425484657288,
-0.0845772922039032,
0.06936841458082199,
-0.06218599155545235,
-0.12968985736370087,
-0.026130788028240204,
0.045467354357242584,
-0.17519839107990265,
0.026703642681241035,
-0.027433741837739944,
0.0919293761253357,
-0.09345759451389313,
-0.02221956104040146,
0.03687324374914169,
0.084866963326931,
-0.014529162086546421,
0.08703910559415817,
-0.14498743414878845,
0.11886418610811234,
0.02978132851421833,
0.09024628251791,
-0.11081171780824661,
0.07909037172794342,
-0.007550720125436783,
0.009180475026369095,
0.19379350543022156,
-0.011335089802742004,
-0.03514958545565605,
-0.08774717897176743,
-0.11210042238235474,
-0.013537433929741383,
0.12687496840953827,
-0.1243172138929367,
0.08773399889469147,
-0.015198243781924248,
-0.044079482555389404,
0.00937260314822197,
-0.12100647389888763,
-0.17273177206516266,
-0.19628387689590454,
0.05585884302854538,
-0.09575839340686798,
0.025643249973654747,
-0.11914430558681488,
-0.07089093327522278,
-0.02952558360993862,
0.241120383143425,
-0.1745356321334839,
-0.06510113179683685,
-0.1468164622783661,
-0.046294767409563065,
0.1662203073501587,
-0.04437198117375374,
0.0718095526099205,
-0.0208172257989645,
0.20345525443553925,
0.005988610442727804,
-0.004939318168908358,
0.06724198162555695,
-0.08892562240362167,
-0.16873881220817566,
-0.06771010160446167,
0.1510489284992218,
0.11680185794830322,
0.04907919466495514,
-0.002248800592496991,
0.0011772146681323647,
-0.016943959519267082,
-0.1137804463505745,
-0.0033210667315870523,
0.16037839651107788,
0.03878779336810112,
0.025986969470977783,
-0.05243593826889992,
-0.08797456324100494,
-0.06899320334196091,
-0.06853509694337845,
0.06221301481127739,
0.19590823352336884,
-0.10376439243555069,
0.1700313836336136,
0.147536963224411,
-0.07305635511875153,
-0.23175598680973053,
0.035342130810022354,
0.04983805492520332,
0.0014306638622656465,
0.04886869341135025,
-0.18252557516098022,
0.10521943867206573,
0.019543392583727837,
-0.05505957826972008,
0.13485197722911835,
-0.1557481735944748,
-0.1552847921848297,
0.0722852572798729,
0.03904085233807564,
-0.22423844039440155,
-0.1354004591703415,
-0.09622503817081451,
-0.05825018882751465,
-0.14065024256706238,
0.06054598465561867,
-0.002136280992999673,
0.015948504209518433,
0.03500790148973465,
-0.0015643214574083686,
0.027123261243104935,
-0.058935679495334625,
0.18609118461608887,
-0.004065449349582195,
0.020676052197813988,
-0.060264769941568375,
-0.0478842556476593,
0.09839435666799545,
-0.06130504235625267,
0.12208222597837448,
0.004057085141539574,
0.01594383642077446,
-0.10362856835126877,
-0.048314861953258514,
-0.04328322783112526,
0.05154227837920189,
-0.07548051327466965,
-0.10070807486772537,
-0.043625857681035995,
0.08841723203659058,
0.07005169242620468,
-0.03383097052574158,
0.00549331633374095,
-0.07189501076936722,
0.10019614547491074,
0.17795267701148987,
0.17573626339435577,
0.009926567785441875,
-0.07241068035364151,
0.01677953451871872,
-0.04142116755247116,
0.044231921434402466,
-0.2513144314289093,
0.03756171092391014,
0.06098250672221184,
0.029438555240631104,
0.09217222779989243,
-0.020435843616724014,
-0.1820858269929886,
-0.04050002992153168,
0.08094815909862518,
-0.05452597141265869,
-0.22617179155349731,
-0.019085140898823738,
0.0954197570681572,
-0.2020406424999237,
-0.007372708059847355,
0.03995226323604584,
-0.048725228756666183,
-0.023169852793216705,
0.00010950004070764408,
0.06317184865474701,
0.002471912419423461,
0.09773622453212738,
0.0735151618719101,
0.09715340286493301,
-0.08337292820215225,
0.10562895983457565,
0.10150538384914398,
-0.09572599828243256,
0.03605884686112404,
0.06754924356937408,
-0.05300498008728027,
-0.043293699622154236,
0.03665391728281975,
0.033023297786712646,
0.005234600510448217,
-0.060321882367134094,
0.013913018628954887,
-0.036497246474027634,
0.044923391193151474,
0.08326134830713272,
0.03754979372024536,
-0.013354414142668247,
0.06462216377258301,
0.03401726484298706,
-0.10898099094629288,
0.10366570204496384,
0.01731540448963642,
0.04105307161808014,
-0.08384523540735245,
-0.019968897104263306,
0.035425446927547455,
0.030576206743717194,
-0.01765924133360386,
-0.02306121215224266,
-0.02860277332365513,
-0.01614218018949032,
-0.14299540221691132,
-0.023106401786208153,
-0.07243485748767853,
0.006181265693157911,
0.014656842686235905,
-0.031884219497442245,
-0.011233693920075893,
0.02475680410861969,
-0.06979699432849884,
-0.07426341623067856,
-0.006949664559215307,
0.09833318740129471,
-0.15115703642368317,
0.008848577737808228,
0.06907843053340912,
-0.11088496446609497,
0.08190931379795074,
-0.008411259390413761,
0.016245156526565552,
0.022527478635311127,
-0.15448406338691711,
0.05601610988378525,
0.0008648968650959432,
0.01916889287531376,
0.025886621326208115,
-0.16471809148788452,
0.004104440100491047,
-0.04661374166607857,
-0.02149827405810356,
-0.00004464812809601426,
-0.02647159807384014,
-0.12325995415449142,
0.06858719140291214,
-0.015622655861079693,
-0.035931166261434555,
-0.02701525390148163,
0.0539589487016201,
0.07888586074113846,
-0.027474910020828247,
0.10445091128349304,
-0.008690856397151947,
0.04941811040043831,
-0.16801609098911285,
-0.02470702864229679,
-0.04982255399227142,
0.019377702847123146,
0.009884213097393513,
-0.007693959400057793,
0.04183054715394974,
-0.00976533442735672,
0.21883612871170044,
-0.05075952783226967,
0.1607085019350052,
0.05847611650824547,
-0.017352959141135216,
-0.0007513365126214921,
0.06180921941995621,
0.05997028574347496,
0.04658793285489082,
0.009480604901909828,
0.023740366101264954,
-0.022450892254710197,
-0.006695089396089315,
-0.15932634472846985,
0.01890849508345127,
0.14999441802501678,
0.06301083415746689,
0.024745315313339233,
0.05866100639104843,
-0.12775006890296936,
-0.12135478109121323,
0.09311001747846603,
-0.026755332946777344,
0.00928465835750103,
-0.08245618641376495,
0.1358020007610321,
0.14980104565620422,
-0.14000412821769714,
0.05256148427724838,
-0.06134212389588356,
-0.05217423290014267,
-0.10388828068971634,
-0.12032219022512436,
-0.05887215584516525,
-0.053666237741708755,
0.002330566756427288,
-0.03760887682437897,
0.054546963423490524,
0.03344334661960602,
-0.009351172484457493,
-0.00022941511997487396,
0.13597318530082703,
-0.019751882180571556,
-0.0028988157864660025,
0.048313532024621964,
0.03693558648228645,
0.02373051457107067,
-0.05275435373187065,
0.02940409444272518,
0.02539868652820587,
0.032232340425252914,
0.06546790152788162,
0.033412106335163116,
-0.047448933124542236,
0.03804153576493263,
-0.0025254099164158106,
-0.11207924783229828,
0.019641218706965446,
-0.00460948096588254,
-0.0742158442735672,
0.1268945336341858,
0.0407399944961071,
0.010224059224128723,
-0.03741471841931343,
0.24361543357372284,
-0.06653323769569397,
-0.06378097087144852,
-0.13251738250255585,
0.10491154342889786,
-0.0027236645109951496,
0.06476365029811859,
0.023412218317389488,
-0.1284150779247284,
0.005243356805294752,
0.13858191668987274,
0.12181595712900162,
0.0045748427510261536,
0.009228081442415714,
0.0518609918653965,
0.0025186820421367884,
-0.06998204439878464,
0.054019294679164886,
0.06992026418447495,
0.12919506430625916,
-0.07847554981708527,
0.07680778950452805,
0.0006860480643808842,
-0.08370215445756912,
-0.02947772853076458,
0.11312682181596756,
-0.0409729965031147,
0.03491825982928276,
-0.047444481402635574,
0.10916327685117722,
-0.05787910893559456,
-0.29412412643432617,
0.02350960113108158,
-0.09588567912578583,
-0.15202060341835022,
-0.018367812037467957,
0.05944539234042168,
-0.02624768204987049,
0.018029648810625076,
0.06971040368080139,
-0.06011629104614258,
0.20098382234573364,
0.0335683599114418,
-0.07864278554916382,
-0.0664360448718071,
0.04837050288915634,
-0.06564252078533173,
0.2949807047843933,
0.008418165147304535,
0.02863333560526371,
0.10770907253026962,
-0.03253700211644173,
-0.18271861970424652,
0.010723991319537163,
0.1133992001414299,
-0.08056149631738663,
0.08200647681951523,
0.19000613689422607,
-0.012578671798110008,
0.1209007054567337,
0.05294662341475487,
-0.047376248985528946,
0.04217283055186272,
-0.03389401361346245,
-0.051268599927425385,
-0.10752558708190918,
0.058453381061553955,
-0.05909625440835953,
0.15447644889354706,
0.10152646154165268,
-0.05671518296003342,
-0.004550917539745569,
-0.05555408447980881,
0.04875178262591362,
0.01804669201374054,
0.12263146042823792,
0.02951994352042675,
-0.1865430772304535,
0.032826557755470276,
-0.01144319772720337,
0.10186848044395447,
-0.25588861107826233,
-0.08421015739440918,
0.08833149075508118,
-0.011924264021217823,
-0.05105875805020332,
0.10560628771781921,
0.057650718837976456,
0.04243382066488266,
-0.043439045548439026,
-0.10480839014053345,
-0.02186836116015911,
0.14663739502429962,
-0.1469624787569046,
-0.025013303384184837
] |
null | null | diffusers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🧨 diffusers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "diffusers"} | null | a-r-r-o-w/motionctrl-svd | [
"diffusers",
"safetensors",
"arxiv:1910.09700",
"diffusers:StableVideoDiffusionPipeline",
"region:us"
] | 2024-02-07T17:26:35+00:00 | [
"1910.09700"
] | [] | TAGS
#diffusers #safetensors #arxiv-1910.09700 #diffusers-StableVideoDiffusionPipeline #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a diffusers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a diffusers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#diffusers #safetensors #arxiv-1910.09700 #diffusers-StableVideoDiffusionPipeline #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a diffusers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
38,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#diffusers #safetensors #arxiv-1910.09700 #diffusers-StableVideoDiffusionPipeline #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a diffusers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.05674594268202782,
0.19320698082447052,
-0.0040419804863631725,
0.025896959006786346,
0.09914743155241013,
0.003408581716939807,
0.06638630479574203,
0.11870917677879333,
0.005362680181860924,
0.12183193117380142,
0.029524236917495728,
0.07194967567920685,
0.12258999794721603,
0.16909313201904297,
0.004254455678164959,
-0.2344931811094284,
0.04453146457672119,
-0.09667602181434631,
0.020512275397777557,
0.11718295514583588,
0.14024171233177185,
-0.10610446333885193,
0.09178268909454346,
-0.01747998408973217,
-0.015762651339173317,
-0.022550584748387337,
-0.07601344585418701,
-0.06558284908533096,
0.059805355966091156,
0.06517806649208069,
0.065830759704113,
0.01280110701918602,
0.09234752506017685,
-0.27611440420150757,
0.017215430736541748,
0.08234287798404694,
0.004243253730237484,
0.06807784736156464,
0.07849449664354324,
-0.05862192064523697,
0.12848979234695435,
-0.07172422111034393,
0.14265665411949158,
0.07593835145235062,
-0.08792039752006531,
-0.1853395700454712,
-0.06975483894348145,
0.05489932373166084,
0.1312437206506729,
0.04966718703508377,
-0.029966438189148903,
0.13878244161605835,
-0.09652138501405716,
0.006849444471299648,
0.11528567224740982,
-0.06816258281469345,
-0.05466277897357941,
0.0520043671131134,
0.1096123531460762,
0.09323899447917938,
-0.12445498257875443,
-0.0062141865491867065,
0.03260517865419388,
0.015962868928909302,
0.09010785818099976,
0.01842493936419487,
0.1443631500005722,
0.03158879280090332,
-0.14082641899585724,
-0.04986229166388512,
0.09758704900741577,
0.03523777052760124,
-0.0527077279984951,
-0.23552405834197998,
-0.02061370573937893,
-0.004690903704613447,
-0.027290984988212585,
-0.046565599739551544,
0.05470653623342514,
-0.04044240713119507,
0.08644413948059082,
-0.010983049869537354,
-0.08135870844125748,
-0.0325210839509964,
0.060853634029626846,
0.05281313881278038,
0.02040923945605755,
-0.004407568834722042,
0.013970456086099148,
0.11820095032453537,
0.07054378092288971,
-0.13028840720653534,
-0.07297687232494354,
-0.07562507688999176,
-0.09336026012897491,
-0.04431890696287155,
0.0456673726439476,
0.07365239411592484,
0.03966696560382843,
0.19654138386249542,
-0.01801731437444687,
0.043932829052209854,
0.048194389790296555,
0.006270275916904211,
0.07013433426618576,
0.08589842170476913,
-0.06522876769304276,
-0.14858585596084595,
-0.05220165476202965,
0.09948329627513885,
-0.006978307850658894,
-0.03320886567234993,
-0.048547882586717606,
0.04047303646802902,
0.027367832139134407,
0.11545737832784653,
0.09053990989923477,
-0.007493921089917421,
-0.06119915470480919,
-0.03750481456518173,
0.2303643524646759,
-0.14374753832817078,
0.04203638806939125,
0.0025415210984647274,
-0.05042556673288345,
-0.00886561255902052,
0.005295928101986647,
0.016443179920315742,
-0.02792365290224552,
0.10685823857784271,
-0.07272970676422119,
-0.03262023255228996,
-0.1102341040968895,
-0.06020084023475647,
0.03262657672166824,
-0.010915471240878105,
-0.02317471243441105,
-0.040166035294532776,
-0.11240966618061066,
-0.05153106525540352,
0.06370186805725098,
-0.0744711235165596,
-0.05116258189082146,
0.0026214909739792347,
-0.04614520072937012,
0.004088256973773241,
0.004166212864220142,
0.11552682518959045,
-0.034895848482847214,
0.027485260739922523,
-0.047327082604169846,
0.07575970888137817,
0.10219335556030273,
0.03802846372127533,
-0.07281317561864853,
0.06922084093093872,
-0.2184373289346695,
0.09476417303085327,
-0.09057645499706268,
0.014035738073289394,
-0.14490769803524017,
-0.042879194021224976,
0.02939661778509617,
0.020843947306275368,
-0.007661222480237484,
0.13780128955841064,
-0.2172020822763443,
-0.02452842704951763,
0.15034019947052002,
-0.10961399227380753,
-0.09154536575078964,
0.06537376344203949,
-0.054606206715106964,
0.11492884159088135,
0.04523410275578499,
-0.020004326477646828,
0.05771946907043457,
-0.1568763107061386,
-0.04054718837141991,
-0.020383106544613838,
-0.012400434352457523,
0.14354920387268066,
0.0719374343752861,
-0.052248936146497726,
0.08744394034147263,
0.02082820050418377,
-0.03613397479057312,
-0.04026569053530693,
-0.04018837958574295,
-0.08996237814426422,
0.007775353267788887,
-0.07315173745155334,
0.03297838941216469,
-0.023467084392905235,
-0.09263875335454941,
-0.029797449707984924,
-0.1673215627670288,
0.02725343592464924,
0.0864863470196724,
0.01939108781516552,
-0.02247876115143299,
-0.09026284515857697,
0.023854082450270653,
-0.002074520569294691,
-0.021600104868412018,
-0.15696310997009277,
-0.04856028035283089,
0.0433291494846344,
-0.17877018451690674,
0.020030833780765533,
-0.03698635473847389,
0.051172900944948196,
0.03938166797161102,
-0.044451020658016205,
-0.0012733606854453683,
0.0009143375209532678,
0.014077886939048767,
-0.024111049249768257,
-0.19994795322418213,
-0.03063414990901947,
-0.03466328978538513,
0.1385105550289154,
-0.22502659261226654,
0.02710004895925522,
0.07198385894298553,
0.14177171885967255,
0.0034448676742613316,
-0.050858501344919205,
0.02201971784234047,
-0.060413870960474014,
-0.05123366415500641,
-0.06806281954050064,
-0.006435995455831289,
-0.03883291035890579,
-0.029664771631360054,
0.060538399964571,
-0.19227908551692963,
-0.046471331268548965,
0.10397804528474808,
0.11256171762943268,
-0.1470438539981842,
-0.03570089116692543,
-0.042537860572338104,
-0.05894942581653595,
-0.09126812219619751,
-0.056615691632032394,
0.133971706032753,
0.0464906319975853,
0.04847629740834236,
-0.08943909406661987,
-0.06460537016391754,
0.012348810210824013,
-0.001107867225073278,
-0.038208283483982086,
0.0801679715514183,
0.08502168953418732,
-0.09953686594963074,
0.08464240282773972,
0.07918588817119598,
0.0773317888379097,
0.10661150515079498,
0.0063849505968391895,
-0.10911723226308823,
-0.028368394821882248,
0.017203014343976974,
0.014260542578995228,
0.14833320677280426,
-0.03863676264882088,
0.047027092427015305,
0.052450764924287796,
-0.03058873862028122,
0.015676455572247505,
-0.1042872816324234,
0.02347022481262684,
0.04093635082244873,
-0.011974683031439781,
0.016118161380290985,
-0.03085828199982643,
0.022126318886876106,
0.09392853826284409,
0.03768773004412651,
0.03003782033920288,
-0.0034947379026561975,
-0.03616767376661301,
-0.1047595739364624,
0.16833209991455078,
-0.09066814929246902,
-0.2856373190879822,
-0.13294115662574768,
-0.0014021076494827867,
0.05396752804517746,
-0.022186707705259323,
0.00948950182646513,
-0.05215408280491829,
-0.10787417739629745,
-0.10347751528024673,
0.0007782797911204398,
0.04343613609671593,
-0.07510960102081299,
-0.07025954872369766,
0.05668036267161369,
0.0315823033452034,
-0.14023737609386444,
0.027269311249256134,
0.04734605923295021,
-0.03886457160115242,
-0.010189466178417206,
0.08049852401018143,
0.0972788855433464,
0.16299954056739807,
-0.009488953277468681,
-0.020279860123991966,
0.023009344935417175,
0.2400934100151062,
-0.14348365366458893,
0.12023822218179703,
0.15558595955371857,
-0.06169526278972626,
0.09990620613098145,
0.19119298458099365,
0.02278536930680275,
-0.08351574838161469,
0.032333407551050186,
0.037265412509441376,
-0.054970014840364456,
-0.2335030734539032,
-0.06122056767344475,
-0.0018651635618880391,
-0.07674646377563477,
0.09166789054870605,
0.08685286343097687,
0.11118897795677185,
0.04276535287499428,
-0.08866265416145325,
-0.07380890846252441,
0.025120751932263374,
0.10552673041820526,
-0.04635854810476303,
0.00036768201971426606,
0.08306246995925903,
-0.041425589472055435,
0.0024932054802775383,
0.11189714819192886,
0.012915363535284996,
0.19670119881629944,
0.022475428879261017,
0.13705596327781677,
0.07146791368722916,
0.03724724054336548,
0.02329368144273758,
0.023117151111364365,
0.02819833904504776,
0.009957139380276203,
-0.013827623799443245,
-0.0902317464351654,
0.02984015643596649,
0.1327328234910965,
0.06874900311231613,
0.01962335780262947,
0.020487558096647263,
-0.02709249220788479,
0.05300171300768852,
0.15444315969944,
0.007473839912563562,
-0.21567226946353912,
-0.03937254473567009,
0.08552859723567963,
-0.06802034378051758,
-0.12517599761486053,
-0.02326875738799572,
0.05492982268333435,
-0.17907613515853882,
0.04382708668708801,
-0.012155727483332157,
0.10706634074449539,
-0.12069477885961533,
-0.03262024372816086,
0.03448006883263588,
0.08662647008895874,
-0.03298177942633629,
0.07408178597688675,
-0.16679809987545013,
0.11876024305820465,
0.014608724042773247,
0.06661728024482727,
-0.11597619205713272,
0.09114355593919754,
0.011162713170051575,
-0.008390041999518871,
0.17627274990081787,
0.003577766939997673,
-0.06529130786657333,
-0.058780133724212646,
-0.08818133920431137,
-0.02593221142888069,
0.09716944396495819,
-0.10595813393592834,
0.07759737223386765,
-0.010262359865009785,
-0.037150487303733826,
0.001416870392858982,
-0.11905564367771149,
-0.1355631798505783,
-0.18985900282859802,
0.057730939239263535,
-0.10187991708517075,
-0.004942308645695448,
-0.10132411867380142,
-0.05576315522193909,
-0.03990165889263153,
0.18824586272239685,
-0.146890327334404,
-0.09656050056219101,
-0.1498064398765564,
-0.08979190140962601,
0.16771520674228668,
-0.040330760180950165,
0.09022361040115356,
-0.0033464664593338966,
0.2223995178937912,
-0.0018514374969527125,
-0.007574497722089291,
0.0686911940574646,
-0.08724970370531082,
-0.17773443460464478,
-0.08079950511455536,
0.12958717346191406,
0.12970267236232758,
0.04943448305130005,
-0.008907152339816093,
0.019145794212818146,
-0.03173605725169182,
-0.1113637238740921,
0.008365826681256294,
0.12975654006004333,
0.0649246871471405,
0.028938129544258118,
-0.0062340968288481236,
-0.11179567873477936,
-0.06962674111127853,
-0.04063126817345619,
0.024454647675156593,
0.2051597684621811,
-0.08268974721431732,
0.16732250154018402,
0.12882472574710846,
-0.04796557500958443,
-0.21108485758304596,
0.0383150577545166,
0.04693659394979477,
0.0012775019276887178,
0.05462753772735596,
-0.17691712081432343,
0.0653671994805336,
0.027575915679335594,
-0.049742136150598526,
0.16920346021652222,
-0.16409853100776672,
-0.15365298092365265,
0.06844818592071533,
0.05270837992429733,
-0.21750572323799133,
-0.1153065636754036,
-0.08087329566478729,
-0.055243685841560364,
-0.13725821673870087,
0.07866505533456802,
0.021718842908740044,
-0.0029149462934583426,
0.04147980362176895,
0.03633167967200279,
0.019756676629185677,
-0.05031364783644676,
0.21558766067028046,
-0.011432650499045849,
0.033435143530368805,
-0.08053123205900192,
-0.10425348579883575,
0.07429526001214981,
-0.05523476004600525,
0.09088967740535736,
-0.02924448810517788,
0.006353185977786779,
-0.08532547205686569,
-0.05416976287961006,
-0.05349917709827423,
0.03317053243517876,
-0.08206910640001297,
-0.10844283550977707,
-0.06620663404464722,
0.09630136936903,
0.08629259467124939,
-0.03269148990511894,
-0.0485510490834713,
-0.08688073605298996,
0.04010685905814171,
0.20614373683929443,
0.16804303228855133,
0.05157895013689995,
-0.08961613476276398,
0.006305147893726826,
-0.015561731532216072,
0.041285570710897446,
-0.21611550450325012,
0.03674960136413574,
0.046699944883584976,
0.02966972440481186,
0.11479803174734116,
-0.022887304425239563,
-0.16491393744945526,
-0.04500475153326988,
0.06122951954603195,
-0.03555853292346001,
-0.20142389833927155,
-0.00881927739828825,
0.06219223514199257,
-0.18501074612140656,
-0.0679524689912796,
0.013011729344725609,
-0.017488865181803703,
-0.026241520419716835,
0.0077438801527023315,
0.07417783886194229,
0.02572557143867016,
0.10115499049425125,
0.059680595993995667,
0.10158747434616089,
-0.11219438910484314,
0.08889372646808624,
0.09028034657239914,
-0.07631593197584152,
0.009741575457155704,
0.07778315246105194,
-0.05713494122028351,
-0.022860053926706314,
0.028393136337399483,
0.051142629235982895,
0.006548711564391851,
-0.056422896683216095,
-0.007778918836265802,
-0.09943147748708725,
0.06083812564611435,
0.11651549488306046,
0.03133244067430496,
-0.009220977313816547,
0.05764961242675781,
0.020565971732139587,
-0.08783701062202454,
0.10946150124073029,
0.02882060967385769,
0.032501958310604095,
-0.05826295167207718,
-0.02250773273408413,
0.04045774042606354,
0.004326104186475277,
-0.019964907318353653,
-0.02708086371421814,
-0.04757264256477356,
-0.014397348277270794,
-0.18062540888786316,
0.01630203053355217,
-0.07585856318473816,
0.005295482464134693,
0.010844874195754528,
-0.04103373363614082,
-0.01136869564652443,
0.01729384809732437,
-0.07505518943071365,
-0.048655614256858826,
-0.009715601801872253,
0.09759850054979324,
-0.1403069794178009,
0.006677433382719755,
0.08631838858127594,
-0.11630367487668991,
0.06600818037986755,
-0.020839398726820946,
-0.012428443878889084,
-0.005413975100964308,
-0.13784921169281006,
0.04300856217741966,
0.00022739957785233855,
0.021082205697894096,
0.04519442468881607,
-0.1675456464290619,
0.0037173249293118715,
-0.04325918108224869,
-0.0508633628487587,
-0.018124889582395554,
-0.06955049932003021,
-0.11158059537410736,
0.10950426012277603,
0.00096321408636868,
-0.07319780439138412,
-0.006825610995292664,
0.04777684062719345,
0.11161206662654877,
-0.03917236626148224,
0.11328131705522537,
0.004002503585070372,
0.06524818390607834,
-0.17989040911197662,
-0.028308264911174774,
-0.017615236341953278,
0.011486062780022621,
0.02895634062588215,
-0.009444717317819595,
0.045241426676511765,
-0.009537453763186932,
0.24638672173023224,
-0.02426966279745102,
0.06865056604146957,
0.06069047376513481,
0.03604987636208534,
0.00671251118183136,
0.0793164074420929,
0.05593963712453842,
0.012441196478903294,
0.0019929695408791304,
0.019091930240392685,
-0.031159263104200363,
-0.012995454482734203,
-0.16449643671512604,
0.07217858731746674,
0.15070576965808868,
0.07690038532018661,
0.0004793994012288749,
0.0624823272228241,
-0.11051314324140549,
-0.10127295553684235,
0.08856794983148575,
-0.038058552891016006,
-0.008159219287335873,
-0.0629461407661438,
0.14307990670204163,
0.15282998979091644,
-0.1750248670578003,
0.07731106877326965,
-0.039314527064561844,
-0.04717016965150833,
-0.10371298342943192,
-0.1580946296453476,
-0.060032788664102554,
-0.02451537922024727,
-0.0016253504436463118,
-0.05221286416053772,
0.07355072349309921,
0.09074901789426804,
0.005196854472160339,
-0.00005569026689045131,
0.10129555314779282,
-0.02776956558227539,
-0.017276663333177567,
0.036094073206186295,
0.05379357933998108,
0.03197963908314705,
-0.04958385229110718,
0.018751582130789757,
0.007575845345854759,
0.0323437824845314,
0.06061622127890587,
0.022524874657392502,
-0.03521033376455307,
0.018291110172867775,
-0.0019744473975151777,
-0.09930764138698578,
0.024647869169712067,
-0.028613561764359474,
-0.0658547431230545,
0.1395663470029831,
0.03314534202218056,
0.01651868224143982,
-0.03198994696140289,
0.20074114203453064,
-0.06141074001789093,
-0.07934345304965973,
-0.13193699717521667,
0.09542657434940338,
-0.03860177472233772,
0.05932838097214699,
0.04995577037334442,
-0.11536107212305069,
-0.0036864110734313726,
0.12807822227478027,
0.13597553968429565,
-0.026505660265684128,
0.0035947863943874836,
0.01666836254298687,
0.007101441267877817,
-0.04811312258243561,
0.04486067220568657,
0.03159695118665695,
0.15230244398117065,
-0.07086783647537231,
0.07857155054807663,
0.0002741335774771869,
-0.07888428121805191,
-0.03406864032149315,
0.12794262170791626,
0.002584414556622505,
0.02827957272529602,
-0.06852647662162781,
0.10345712304115295,
-0.07019247859716415,
-0.24566727876663208,
0.05149490386247635,
-0.08268437534570694,
-0.15144604444503784,
-0.014263256452977657,
0.02211342379450798,
-0.012918109074234962,
0.021799270063638687,
0.06599191576242447,
-0.05595657601952553,
0.1518101543188095,
0.03274545446038246,
-0.08406112343072891,
-0.061771564185619354,
0.07196340709924698,
-0.09331037104129791,
0.29569628834724426,
0.008130451664328575,
0.040586329996585846,
0.09572544693946838,
-0.02491658739745617,
-0.14421287178993225,
0.030478067696094513,
0.10085273534059525,
-0.0921834260225296,
0.0597606897354126,
0.18879646062850952,
-0.013558941893279552,
0.09987539798021317,
0.07060596346855164,
-0.05314319580793381,
0.05289614200592041,
-0.07387816160917282,
-0.06410772353410721,
-0.09685344994068146,
0.07143298536539078,
-0.05925896018743515,
0.1525508016347885,
0.11060643941164017,
-0.0457635298371315,
-0.002591128461062908,
-0.03172069787979126,
0.04553845897316933,
0.016335010528564453,
0.12024565041065216,
0.0142319705337286,
-0.16522495448589325,
0.031200766563415527,
-0.003748908406123519,
0.10541816800832748,
-0.20973658561706543,
-0.07993287593126297,
0.042268604040145874,
-0.032933104783296585,
-0.049715034663677216,
0.1034897193312645,
0.06719335913658142,
0.049855805933475494,
-0.04420776665210724,
-0.06823384761810303,
-0.00504080206155777,
0.1510942131280899,
-0.11425614356994629,
-0.007045311853289604
] |
null | null | fastai |
# Amazing!
🥳 Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the [documentation here](https://huggingface.co/docs/hub/model-repos))!
2. Create a demo in Gradio or Streamlit using 🤗 Spaces ([documentation here](https://huggingface.co/docs/hub/spaces)).
3. Join the fastai community on the [Fastai Discord](https://discord.com/invite/YKrxeNn)!
Greetings fellow fastlearner 🤝! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| {"tags": ["fastai"]} | null | luis56125/intel-image-classification | [
"fastai",
"has_space",
"region:us"
] | 2024-02-07T17:27:30+00:00 | [] | [] | TAGS
#fastai #has_space #region-us
|
# Amazing!
Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the documentation here)!
2. Create a demo in Gradio or Streamlit using Spaces (documentation here).
3. Join the fastai community on the Fastai Discord!
Greetings fellow fastlearner ! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| [
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
"TAGS\n#fastai #has_space #region-us \n",
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
13,
20,
79,
3,
6,
12,
8
] | [
"passage: TAGS\n#fastai #has_space #region-us \n# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---# Model card## Model description\nMore information needed## Intended uses & limitations\nMore information needed## Training and evaluation data\nMore information needed"
] | [
-0.048121724277734756,
-0.024616125971078873,
0.002038001548498869,
0.10439170897006989,
0.135872021317482,
0.11887997388839722,
0.07405775785446167,
0.09980081021785736,
0.07783667743206024,
0.02590852417051792,
0.08961158245801926,
-0.08088712394237518,
0.08744348585605621,
0.271692156791687,
0.06988707184791565,
-0.22761479020118713,
0.04051019623875618,
-0.00024903909070417285,
0.08053462207317352,
0.06629016250371933,
0.13507555425167084,
-0.05464952811598778,
0.14010503888130188,
-0.004088983871042728,
-0.19050447642803192,
-0.042929794639348984,
-0.01773718371987343,
-0.02527874894440174,
0.12317648530006409,
-0.04744937643408775,
0.05381017178297043,
0.015037551522254944,
0.007565062493085861,
-0.07253646105527878,
0.0623294934630394,
0.040457066148519516,
0.01740180514752865,
0.059235580265522,
-0.07249044626951218,
0.08950132131576538,
0.08404164761304855,
-0.024370938539505005,
-0.1097978875041008,
0.07827875018119812,
-0.14424212276935577,
-0.21762843430042267,
-0.1253085881471634,
-0.09017651528120041,
0.028519365936517715,
0.004388005938380957,
-0.025051530450582504,
0.12801909446716309,
-0.13558274507522583,
-0.040698226541280746,
0.20124278962612152,
-0.17012301087379456,
-0.05505548417568207,
0.034343402832746506,
0.09226689487695694,
-0.05829555168747902,
-0.06347129493951797,
0.10614984482526779,
0.09640881419181824,
-0.019833475351333618,
0.05516824126243591,
0.002579754451289773,
0.021173657849431038,
0.01370104867964983,
-0.06150497496128082,
0.04717832803726196,
-0.010183089412748814,
0.048132527619600296,
-0.09465572983026505,
-0.1303568333387375,
-0.004072192590683699,
0.01214400865137577,
-0.048744890838861465,
-0.07019646465778351,
0.07833103090524673,
-0.011118141002953053,
-0.04357248544692993,
-0.13031910359859467,
-0.09131011366844177,
-0.12358787655830383,
0.008646543137729168,
0.09500427544116974,
0.003679296001791954,
0.07374339550733566,
-0.08258994668722153,
0.06774985045194626,
-0.17329485714435577,
-0.06484591960906982,
-0.08138520270586014,
-0.11546400189399719,
0.021133482456207275,
-0.0387684591114521,
0.02668963186442852,
0.15394504368305206,
0.12983950972557068,
0.023976242169737816,
0.04388163983821869,
-0.038937073200941086,
0.051190316677093506,
0.058571770787239075,
0.03395717963576317,
0.034934818744659424,
-0.036981891840696335,
-0.1793210655450821,
-0.016702448949217796,
-0.011550825089216232,
0.07954040914773941,
-0.07523109763860703,
-0.05632320046424866,
0.013454885222017765,
-0.11071494966745377,
0.07202339172363281,
-0.03576776012778282,
-0.0032025426626205444,
0.01168301422148943,
0.018371861428022385,
0.21271461248397827,
0.03955606371164322,
0.014191740192472935,
-0.008875265717506409,
-0.13453757762908936,
-0.06874168664216995,
-0.06896194815635681,
0.03361047804355621,
0.04448792710900307,
-0.0028071461711078882,
-0.07672245055437088,
0.04325154796242714,
-0.06045534089207649,
-0.03508453071117401,
0.008032378740608692,
-0.18221288919448853,
0.007458044681698084,
-0.10049355030059814,
-0.12126200646162033,
0.05306628718972206,
0.01695440337061882,
-0.08215925842523575,
0.08141279965639114,
0.02662261202931404,
0.020931517705321312,
-0.009988143108785152,
-0.005391082260757685,
0.06874798238277435,
-0.08508864045143127,
0.029901226982474327,
0.17170792818069458,
0.13024519383907318,
-0.08046911656856537,
-0.0006887061172164977,
-0.10965746641159058,
0.04426072910428047,
-0.13325683772563934,
0.02251482754945755,
-0.09062390774488449,
0.11723794043064117,
-0.042396437376737595,
0.002038756385445595,
-0.029030200093984604,
0.0960269495844841,
0.08189879357814789,
0.16663365066051483,
-0.2419009804725647,
-0.031095001846551895,
0.13240347802639008,
-0.10711425542831421,
-0.1807439625263214,
0.18486657738685608,
-0.012035200372338295,
0.11329247802495956,
-0.047014184296131134,
0.18334640562534332,
-0.02612062357366085,
-0.13582459092140198,
-0.058872904628515244,
0.005852419883012772,
-0.2269321084022522,
-0.06286033242940903,
0.09738040715456009,
0.13425657153129578,
-0.042984943836927414,
0.007112155202776194,
0.026316028088331223,
0.13609857857227325,
-0.06715573370456696,
-0.05195777863264084,
-0.012255736626684666,
-0.10902371257543564,
0.041914235800504684,
0.018215661868453026,
0.035408079624176025,
-0.059880174696445465,
-0.02931194379925728,
-0.053190283477306366,
0.13146710395812988,
0.09760832786560059,
-0.03670211136341095,
-0.049620725214481354,
0.1689043790102005,
-0.07763876020908356,
-0.033587727695703506,
0.07560533285140991,
-0.08268500119447708,
0.03266897425055504,
0.03090597130358219,
0.055881720036268234,
0.07766123116016388,
0.08522116392850876,
0.06057543307542801,
0.00819048099219799,
0.034654274582862854,
0.12095347046852112,
-0.013591280207037926,
-0.05039411783218384,
0.021508218720555305,
0.016904234886169434,
-0.019032588228583336,
0.29030677676200867,
-0.1951042115688324,
0.024724548682570457,
-0.06477324664592743,
0.07631538063287735,
0.06136792525649071,
0.003575638635084033,
0.08580143749713898,
-0.06023019179701805,
-0.019061198458075523,
-0.04803973436355591,
0.046805646270513535,
-0.0666879191994667,
-0.04162997007369995,
0.2621194124221802,
-0.05497581139206886,
0.044914912432432175,
0.12313763797283173,
-0.05873025581240654,
-0.07091446220874786,
0.01009807363152504,
-0.00793424155563116,
0.03249288722872734,
-0.04042816907167435,
0.043721720576286316,
-0.10840129852294922,
-0.06674089282751083,
0.1573198139667511,
-0.038477856665849686,
0.06786153465509415,
0.032288823276758194,
-0.04958454892039299,
-0.0648743286728859,
0.04650486260652542,
0.13598160445690155,
-0.0875244215130806,
0.07435166835784912,
0.17612984776496887,
-0.010562662966549397,
0.168031245470047,
0.08435525000095367,
-0.07075224816799164,
-0.09465329349040985,
-0.051014289259910583,
-0.021595727652311325,
0.21222901344299316,
-0.07084725052118301,
-0.054564714431762695,
0.05911700800061226,
-0.013703816570341587,
0.07196151465177536,
-0.06009222939610481,
-0.08332337439060211,
0.03227344527840614,
-0.04517695680260658,
0.011517706327140331,
0.13512636721134186,
-0.07090822607278824,
0.04681389778852463,
0.031489867717027664,
-0.0662703812122345,
0.02217509225010872,
0.033389873802661896,
0.0068921963684260845,
0.033959709107875824,
0.07332495599985123,
-0.20893315970897675,
-0.08408680558204651,
-0.13727638125419617,
0.037881869822740555,
0.021770721301436424,
0.045787326991558075,
-0.08602345734834671,
0.02231026627123356,
-0.08954031765460968,
-0.07987114042043686,
0.029592275619506836,
-0.026350297033786774,
-0.11349643021821976,
-0.03396226093173027,
-0.009560913778841496,
-0.06662604957818985,
-0.02250705659389496,
-0.05024505779147148,
0.03983384370803833,
0.04479299485683441,
0.058377087116241455,
0.12796473503112793,
-0.013808943331241608,
-0.03839317709207535,
0.000370211957488209,
-0.022712308913469315,
0.16396735608577728,
-0.14746315777301788,
0.07954913377761841,
0.19160102307796478,
0.11742953956127167,
0.028144672513008118,
0.028885571286082268,
0.03537585213780403,
-0.06289814412593842,
-0.000050317394197918475,
0.03226194158196449,
-0.09392514824867249,
-0.05801016092300415,
-0.020014392212033272,
-0.04031052812933922,
0.17134574055671692,
-0.12160717695951462,
0.03345204517245293,
0.04098419472575188,
0.09783966839313507,
0.10073629021644592,
-0.028829937800765038,
-0.1815856397151947,
0.038818612694740295,
-0.24060091376304626,
-0.05831146240234375,
0.027899866923689842,
-0.09110201895236969,
-0.06232144311070442,
0.17409387230873108,
0.013794700615108013,
0.011769929900765419,
-0.006736889015883207,
0.07983319461345673,
0.0110100656747818,
0.1217205822467804,
0.05947643890976906,
-0.05539114400744438,
0.025202350690960884,
-0.09962950646877289,
-0.07107596844434738,
-0.04035590961575508,
-0.05832801014184952,
0.07548832893371582,
0.1409129947423935,
-0.025475580245256424,
-0.020795362070202827,
0.023489827290177345,
0.08550169318914413,
0.0423230417072773,
0.16739299893379211,
-0.16016584634780884,
-0.026555389165878296,
0.04571257904171944,
-0.03384667634963989,
-0.05433850735425949,
-0.010291114449501038,
0.1137225553393364,
-0.02820689231157303,
-0.040318265557289124,
0.021242983639240265,
0.06503437459468842,
0.01481706090271473,
0.05012747645378113,
-0.04056356102228165,
0.14796851575374603,
-0.03461192920804024,
0.019330544397234917,
-0.12413888424634933,
0.13848772644996643,
0.021095896139740944,
-0.03901609033346176,
-0.06735876202583313,
-0.05808034539222717,
0.18150931596755981,
0.0025602965615689754,
0.10535930097103119,
0.012098877690732479,
-0.12160047143697739,
-0.1359938681125641,
-0.11211287975311279,
0.005111907608807087,
0.08330471813678741,
-0.023147236555814743,
-0.022247863933444023,
0.022165266796946526,
-0.036149751394987106,
-0.0530381016433239,
0.15749511122703552,
-0.1289154291152954,
-0.001082550617866218,
0.014728817157447338,
0.06971760839223862,
-0.08223173767328262,
0.026267826557159424,
0.014071501791477203,
-0.1119147390127182,
0.10590848326683044,
0.2521335482597351,
0.10338116437196732,
-0.09591643512248993,
-0.07697287201881409,
0.03418830782175064,
-0.012184361927211285,
-0.000774814048781991,
-0.006932659074664116,
0.0495428591966629,
-0.005566445179283619,
0.006762749515473843,
0.12971895933151245,
-0.07130889594554901,
0.011540771462023258,
-0.08449850976467133,
0.05566910281777382,
-0.05276734381914139,
0.01761564053595066,
-0.002672141883522272,
-0.008124710991978645,
-0.07340748608112335,
-0.061829522252082825,
0.1609770804643631,
-0.07277000695466995,
-0.06468547880649567,
0.05801168829202652,
0.03307786211371422,
0.01431563775986433,
-0.03584568202495575,
-0.04342148080468178,
0.18088261783123016,
0.29330700635910034,
-0.08191116154193878,
0.10001859813928604,
0.09677296131849289,
0.034820813685655594,
-0.23625829815864563,
0.029798466712236404,
-0.1455078274011612,
0.04449721798300743,
0.040447335690259933,
-0.0409548319876194,
0.04191497340798378,
0.10835777968168259,
-0.06094440817832947,
0.2048867791891098,
-0.03527235612273216,
-0.07983248680830002,
-0.01788630709052086,
0.03109324350953102,
0.29443636536598206,
-0.11833466589450836,
0.006058716680854559,
-0.10420958697795868,
-0.21566011011600494,
0.06983078271150589,
-0.18948867917060852,
0.13948246836662292,
-0.05087858438491821,
0.03576415032148361,
-0.01149723306298256,
-0.07561972737312317,
0.20518061518669128,
-0.15641045570373535,
0.05273103713989258,
-0.13722458481788635,
-0.1327189952135086,
0.01617460884153843,
-0.10048147290945053,
0.1545477658510208,
-0.11024226248264313,
-0.023215843364596367,
-0.2284185290336609,
0.012587235309183598,
-0.023200806230306625,
0.10030807554721832,
0.01800704374909401,
-0.07980740070343018,
-0.08767345547676086,
0.1316242516040802,
-0.06486566364765167,
0.034810543060302734,
-0.06996636837720871,
-0.050714004784822464,
-0.010929876938462257,
-0.045061707496643066,
0.03034941293299198,
-0.07934719324111938,
0.15192505717277527,
-0.016938980668783188,
-0.04507075995206833,
0.08636019378900528,
-0.2479533851146698,
0.023727843537926674,
0.025351112708449364,
-0.03495599329471588,
0.09001832455396652,
-0.025513244792819023,
-0.06256973743438721,
0.12282291799783707,
0.1402233988046646,
-0.07322840392589569,
-0.2460673749446869,
-0.06281693279743195,
0.0076784128323197365,
0.039165716618299484,
0.06561196595430374,
0.05125982314348221,
-0.07261458039283752,
-0.011131617240607738,
-0.026896944269537926,
0.030595947057008743,
-0.11692017316818237,
-0.03854857385158539,
0.07790639251470566,
0.017095070332288742,
-0.07846562564373016,
0.07280377298593521,
0.014225782826542854,
-0.021511616185307503,
0.007357571739703417,
0.148970365524292,
0.007519228849560022,
-0.14747941493988037,
-0.06656096875667572,
0.2007484883069992,
-0.01197928935289383,
-0.07260087132453918,
-0.05383119732141495,
-0.008990069851279259,
-0.0476234145462513,
0.05585788935422897,
0.05367223918437958,
-0.013585401698946953,
0.07708586007356644,
0.06263149529695511,
-0.10210110992193222,
-0.046256959438323975,
-0.066561758518219,
0.04169114679098129,
-0.10485753417015076,
0.060470130294561386,
0.009529483504593372,
0.12185006588697433,
-0.09983488917350769,
-0.01802929677069187,
-0.10810204595327377,
-0.06766588985919952,
-0.17349553108215332,
-0.05834362283349037,
-0.041105758398771286,
-0.015651104971766472,
0.03658895567059517,
0.010445823892951012,
-0.057867538183927536,
-0.0442853718996048,
-0.07536603510379791,
0.038444988429546356,
0.06147460639476776,
0.03932281583547592,
-0.03912714496254921,
0.04001858830451965,
0.05909334123134613,
0.013087345287203789,
0.17542624473571777,
0.038768354803323746,
0.05504675209522247,
-0.05045998468995094,
-0.16491834819316864,
-0.05276111513376236,
-0.0074316514655947685,
-0.07559102028608322,
0.1224973127245903,
-0.007679440546780825,
0.007880088873207569,
-0.08065467327833176,
0.03924860805273056,
0.028234204277396202,
0.10404064506292343,
-0.0028364830650389194,
0.10070426017045975,
0.019627176225185394,
-0.07226712256669998,
-0.025392837822437286,
0.021809715777635574,
0.12809939682483673,
0.01567147858440876,
0.026090998202562332,
0.033139873296022415,
0.016619985923171043,
-0.057361043989658356,
0.033977724611759186,
-0.04997231811285019,
-0.15123651921749115,
0.02628709189593792,
-0.05165188014507294,
0.005062380339950323,
-0.016889680176973343,
0.20362506806850433,
0.07867538928985596,
-0.06474173814058304,
-0.010664013214409351,
0.015816617757081985,
-0.0168940220028162,
-0.03121885471045971,
-0.012740966863930225,
0.04592578858137131,
-0.001151384087279439,
-0.04866636544466019,
0.11825273931026459,
0.05015748366713524,
0.05386412516236305,
0.0596686452627182,
0.12528513371944427,
0.016759619116783142,
0.13257254660129547,
0.061999931931495667,
-0.03403807803988457,
-0.13461735844612122,
-0.04495539888739586,
-0.1254577934741974,
0.04646851494908333,
-0.08697032928466797,
0.09941662102937698,
0.1144254133105278,
-0.05959030240774155,
-0.030464433133602142,
-0.08851305395364761,
-0.008356761187314987,
-0.06041252240538597,
0.039516255259513855,
-0.02262675203382969,
-0.0873224213719368,
0.0481097511947155,
0.05495472997426987,
-0.022752324119210243,
0.13218675553798676,
0.015727028250694275,
-0.036317698657512665,
0.13270340859889984,
-0.07583184540271759,
0.11758984625339508,
0.061510033905506134,
-0.043043944984674454,
-0.11560922116041183,
-0.020150646567344666,
-0.06641761213541031,
-0.10098972916603088,
-0.006782987620681524,
-0.005399650428444147,
-0.07349002361297607,
-0.059971679002046585,
0.08397487550973892,
-0.03124053031206131,
-0.09979676455259323,
-0.032152675092220306,
0.0038895104080438614,
0.06054706871509552,
-0.01686914451420307,
-0.0034020058810710907,
0.04728743061423302,
0.015076374635100365,
0.1653461456298828,
-0.02208263985812664,
0.06234867498278618,
-0.13855914771556854,
0.16070103645324707,
-0.14684462547302246,
-0.029404424130916595,
-0.1890171319246292,
-0.09729582816362381,
-0.05156542733311653,
0.20326784253120422,
0.2840938866138458,
-0.19109351933002472,
-0.010187864303588867,
0.020078664645552635,
-0.014484191313385963,
-0.08961770683526993,
0.12571553885936737,
0.029420215636491776,
-0.023631498217582703,
-0.07249019294977188,
-0.02037387527525425,
0.005258576478809118,
-0.06544211506843567,
-0.026979785412549973,
0.18310695886611938,
0.001496660872362554,
0.059546373784542084,
-0.09605178982019424,
0.01754261925816536,
-0.14839904010295868,
-0.10467469692230225,
-0.02111995778977871,
-0.16156397759914398,
-0.09646477550268173,
0.006635562051087618,
0.038640011101961136,
0.08000610023736954,
0.03268849849700928,
-0.015172510407865047,
0.06479045748710632,
-0.056333884596824646,
-0.0037216036580502987,
-0.1231912299990654,
0.00034658415825106204,
0.062129102647304535,
-0.07422006875276566,
0.2545335292816162,
-0.03070417232811451,
-0.12370815873146057,
0.09026903659105301,
-0.03299184888601303,
-0.12452623248100281,
0.07951879501342773,
-0.005700904875993729,
-0.11531132459640503,
-0.057989440858364105,
0.18941475450992584,
-0.012821312062442303,
-0.1364315301179886,
0.046368811279535294,
-0.17166484892368317,
0.031349923461675644,
0.0363016203045845,
-0.001313706859946251,
-0.04714022949337959,
0.024538639932870865,
-0.008008457720279694,
0.10724439471960068,
0.1382838785648346,
0.016739921644330025,
-0.011060068383812904,
-0.05056179314851761,
0.07912429422140121,
0.056927867233753204,
-0.05218246951699257,
-0.1282637119293213,
-0.08599764108657837,
0.03429819270968437,
0.04119478166103363,
-0.08113081753253937,
-0.16903182864189148,
-0.03668912500143051,
-0.10082915425300598,
-0.004939202684909105,
0.051785312592983246,
0.06585265696048737,
0.29044589400291443,
0.06326735019683838,
0.0016605621203780174,
-0.13649453222751617,
0.050569336861371994,
0.0868251696228981,
-0.04697931930422783,
-0.07670357078313828
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.8.2 | {"library_name": "peft", "base_model": "Trelis/Llama-2-7b-chat-hf-sharded-bf16"} | null | SolaireOfTheSun/Llama-2-7b-chat-hf-sharded-bf16-feinabgestimmt-adapters-2 | [
"peft",
"arxiv:1910.09700",
"base_model:Trelis/Llama-2-7b-chat-hf-sharded-bf16",
"region:us"
] | 2024-02-07T17:27:54+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #arxiv-1910.09700 #base_model-Trelis/Llama-2-7b-chat-hf-sharded-bf16 #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.8.2 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
"TAGS\n#peft #arxiv-1910.09700 #base_model-Trelis/Llama-2-7b-chat-hf-sharded-bf16 #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
43,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#peft #arxiv-1910.09700 #base_model-Trelis/Llama-2-7b-chat-hf-sharded-bf16 #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2"
] | [
-0.1181381344795227,
0.19727149605751038,
-0.0028356341645121574,
0.029223458841443062,
0.07779452949762344,
0.015494933351874352,
0.05737197771668434,
0.1357649564743042,
0.036463662981987,
0.11323360353708267,
0.06885109841823578,
0.12088657170534134,
0.11562664806842804,
0.2212080955505371,
0.004279926419258118,
-0.1664223074913025,
0.01706680655479431,
-0.06928151845932007,
0.014364630915224552,
0.12016356736421585,
0.14479218423366547,
-0.09526396542787552,
0.08089473843574524,
-0.018188372254371643,
-0.004930650815367699,
-0.024937300011515617,
-0.07020042836666107,
-0.008950869552791119,
0.056061599403619766,
0.033739153295755386,
0.05408255010843277,
-0.01107756607234478,
0.08333621174097061,
-0.2709597647190094,
0.017724286764860153,
0.04098164662718773,
-0.004057694226503372,
0.08124442398548126,
0.09406667947769165,
-0.0441671647131443,
0.12708128988742828,
-0.015239475294947624,
0.13391275703907013,
0.0907587930560112,
-0.09883658587932587,
-0.22420738637447357,
-0.06269833445549011,
0.0812215730547905,
0.18298065662384033,
0.0842667743563652,
-0.04301773011684418,
0.12244658917188644,
-0.06498446315526962,
0.025797231122851372,
0.06983645260334015,
-0.10434993356466293,
-0.06664906442165375,
0.0714646652340889,
0.13343030214309692,
0.08002261817455292,
-0.1214556023478508,
-0.037892282009124756,
0.034509770572185516,
0.04869379103183746,
0.05023886263370514,
0.004202019423246384,
0.15324732661247253,
0.030881710350513458,
-0.14613425731658936,
-0.05120434612035751,
0.14337295293807983,
0.011956743896007538,
-0.03901150822639465,
-0.20988884568214417,
-0.0034170798026025295,
-0.10848358273506165,
-0.03974635899066925,
-0.0478997640311718,
0.036514703184366226,
0.013069549575448036,
0.12947554886341095,
-0.04857974499464035,
-0.08803272992372513,
-0.014932133257389069,
0.11144330352544785,
0.05725866183638573,
0.018973227590322495,
-0.020059572532773018,
0.0056142015382647514,
0.12330996245145798,
0.06388096511363983,
-0.1328437328338623,
-0.06580766290426254,
-0.06895118951797485,
-0.035474978387355804,
-0.02769492007791996,
0.03674232214689255,
0.02040569670498371,
0.06125297024846077,
0.2798122763633728,
-0.026357600465416908,
0.06613438576459885,
0.04405985400080681,
0.023886658251285553,
0.02874341793358326,
0.10989844053983688,
-0.031750261783599854,
-0.17080765962600708,
-0.008244126103818417,
0.0997881218791008,
-0.003357849782332778,
-0.035279981791973114,
-0.06533131748437881,
0.03614223003387451,
0.03597854822874069,
0.11551731079816818,
0.11004431545734406,
-0.026968713849782944,
-0.07493777573108673,
-0.05820320546627045,
0.18510743975639343,
-0.15498925745487213,
0.045349325984716415,
0.026603857055306435,
-0.0026552118360996246,
-0.0666942149400711,
0.007160454522818327,
0.019170569255948067,
-0.034147754311561584,
0.06864527612924576,
-0.0650380477309227,
-0.04231897369027138,
-0.12411431968212128,
-0.03357872739434242,
0.03640985116362572,
0.001136972801759839,
-0.041376009583473206,
-0.043346066027879715,
-0.07010353356599808,
-0.11157026141881943,
0.11146921664476395,
-0.05989838391542435,
-0.05995114892721176,
-0.02418203093111515,
-0.08280391246080399,
0.018977167084813118,
0.03798571228981018,
0.07484757155179977,
-0.024049602448940277,
0.045625265687704086,
-0.00583998765796423,
0.0690370500087738,
0.0666862279176712,
0.034300222992897034,
-0.07865653187036514,
0.06418787688016891,
-0.1941995471715927,
0.07840386033058167,
-0.07835167646408081,
0.04630007967352867,
-0.16043059527873993,
-0.004161621443927288,
-0.004898848477751017,
0.029608670622110367,
0.04850497841835022,
0.15709823369979858,
-0.21749383211135864,
-0.02971627376973629,
0.16169075667858124,
-0.10138624161481857,
-0.13351242244243622,
0.03961623087525368,
-0.03792359307408333,
0.18759804964065552,
0.024378320202231407,
0.03176095336675644,
0.08810292929410934,
-0.154221311211586,
-0.014327802695333958,
-0.018256189301609993,
0.01473134383559227,
0.06619387120008469,
0.08162213116884232,
-0.09280609339475632,
-0.003115785541012883,
0.01148303970694542,
-0.061079755425453186,
-0.01969340443611145,
-0.040286459028720856,
-0.10579638183116913,
0.0036032586358487606,
-0.08481569588184357,
0.006873821374028921,
0.004307607654482126,
-0.09461662918329239,
-0.008892207406461239,
-0.14766542613506317,
-0.047490525990724564,
0.08335020393133163,
0.0031538894400000572,
-0.015453570522367954,
-0.0972089022397995,
0.06403058767318726,
-0.03634766861796379,
-0.020803414285182953,
-0.1477097123861313,
-0.004365186206996441,
0.019695095717906952,
-0.13655759394168854,
0.0069341156631708145,
-0.11226584017276764,
0.06865353882312775,
-0.001955528510734439,
-0.04560066759586334,
-0.040206532925367355,
-0.007969454862177372,
-0.008147619664669037,
-0.06441042572259903,
-0.2355523705482483,
-0.029622018337249756,
-0.05054420605301857,
0.1726302057504654,
-0.2287760078907013,
0.04142492264509201,
0.005690731108188629,
0.11616000533103943,
0.001753757824189961,
-0.05837450921535492,
0.018159586936235428,
-0.060227371752262115,
-0.024702051654458046,
-0.07043436914682388,
-0.002803630894050002,
0.008455133996903896,
-0.023185569792985916,
0.010970372706651688,
-0.1153634786605835,
-0.06420443207025528,
0.09627197682857513,
0.058103349059820175,
-0.14625291526317596,
0.014798679389059544,
-0.040223196148872375,
-0.05807002633810043,
-0.06283935904502869,
-0.07185106724500656,
0.09177219867706299,
0.05021706596016884,
0.047123730182647705,
-0.08482160419225693,
-0.07033076882362366,
0.004973860457539558,
-0.022818956524133682,
-0.00970391370356083,
0.12907801568508148,
0.09714005887508392,
-0.10058607161045074,
0.08979696035385132,
0.0628291592001915,
0.021530071273446083,
0.08263126760721207,
-0.01864038035273552,
-0.10489299893379211,
-0.027758432552218437,
0.05735914036631584,
0.009980740025639534,
0.17240063846111298,
-0.08582990616559982,
0.05192724987864494,
0.04665563255548477,
-0.05618784576654434,
0.051453784108161926,
-0.09219805896282196,
0.007493637967854738,
0.0012070387601852417,
-0.01596822217106819,
0.03518155589699745,
-0.016257386654615402,
0.0009937105933204293,
0.08880914747714996,
0.0686771348118782,
0.01661018840968609,
0.011657055467367172,
-0.03642977029085159,
-0.14329618215560913,
0.17914502322673798,
-0.08981168270111084,
-0.2451286017894745,
-0.1502447873353958,
0.04489326849579811,
0.0559251569211483,
-0.013247373513877392,
0.03196219354867935,
-0.05284000560641289,
-0.09442916512489319,
-0.08512086421251297,
0.0060422602109611034,
0.026271410286426544,
-0.060462869703769684,
-0.06254339963197708,
0.03532658517360687,
0.03917548060417175,
-0.12261972576379776,
0.024169061332941055,
0.05751659348607063,
0.0021136715076863766,
-0.004555159714072943,
0.03897562250494957,
0.09354787319898605,
0.20794224739074707,
-0.005286749452352524,
0.008882980793714523,
0.061511434614658356,
0.28627923130989075,
-0.16131141781806946,
0.11507702618837357,
0.13694114983081818,
-0.06283509731292725,
0.07396627217531204,
0.19074928760528564,
0.030362091958522797,
-0.0978357344865799,
0.01998024620115757,
0.030792532488703728,
-0.025054074823856354,
-0.27338913083076477,
-0.05006987974047661,
-0.0272066630423069,
-0.07753065973520279,
0.08624901622533798,
0.0908370390534401,
0.09563709795475006,
0.028488392010331154,
-0.059524428099393845,
-0.08728070557117462,
0.021973803639411926,
0.11459164321422577,
-0.01424829289317131,
0.0019317283295094967,
0.08133579045534134,
-0.050357501953840256,
0.006600155029445887,
0.08700865507125854,
-0.015028851106762886,
0.11981251090765,
0.061104029417037964,
0.11078507453203201,
0.08402712643146515,
0.084307000041008,
-0.008380415849387646,
0.027836646884679794,
-0.00031975010642781854,
0.020215725526213646,
0.0203701164573431,
-0.0878191590309143,
0.016822397708892822,
0.1118163913488388,
0.015766069293022156,
0.018817709758877754,
0.01626560464501381,
-0.06387853622436523,
0.034121669828891754,
0.1956094354391098,
0.03129170462489128,
-0.20588234066963196,
-0.08010124415159225,
0.051518332213163376,
-0.0732668787240982,
-0.15834909677505493,
-0.01314424816519022,
0.007999151013791561,
-0.16007454693317413,
0.012169231660664082,
-0.036929916590452194,
0.11167705059051514,
-0.06867799907922745,
-0.04052245244383812,
0.1082296222448349,
0.050323616713285446,
-0.027475876733660698,
0.050317324697971344,
-0.2002214938402176,
0.10682982206344604,
0.028508713468909264,
0.06315074861049652,
-0.08971314877271652,
0.08875738829374313,
-0.006046023685485125,
-0.012159503996372223,
0.15731756389141083,
0.0007066592224873602,
-0.05479873716831207,
-0.07785545289516449,
-0.07410085201263428,
-0.0069300467148423195,
0.08276000618934631,
-0.1372804343700409,
0.07350901514291763,
-0.03518112376332283,
-0.028659584000706673,
-0.008439280092716217,
-0.08596987277269363,
-0.11594396084547043,
-0.16363799571990967,
0.06479094922542572,
-0.09006349742412567,
0.02223283424973488,
-0.07741783559322357,
-0.053138718008995056,
0.03444678336381912,
0.18598613142967224,
-0.19473934173583984,
-0.10642579942941666,
-0.14511141180992126,
-0.10035328567028046,
0.15426789224147797,
-0.045827437192201614,
0.08878437429666519,
-0.008907758630812168,
0.16149276494979858,
-0.002409412758424878,
-0.018442001193761826,
0.0869813784956932,
-0.09410133957862854,
-0.17934918403625488,
-0.0454990454018116,
0.18295595049858093,
0.13064441084861755,
0.030308052897453308,
-0.010929281823337078,
0.022723527625203133,
-0.07170780748128891,
-0.10858486592769623,
0.0286567322909832,
0.13643677532672882,
0.05812159553170204,
-0.02306309901177883,
-0.04135332256555557,
-0.07953198254108429,
-0.06566406786441803,
-0.04212135449051857,
-0.004481813870370388,
0.2014150470495224,
-0.07074250280857086,
0.1520845890045166,
0.10371026396751404,
-0.06049598753452301,
-0.20662494003772736,
0.03809158131480217,
0.04201696068048477,
0.019130051136016846,
0.024141104891896248,
-0.19706910848617554,
0.08071039617061615,
-0.028898410499095917,
-0.07990600168704987,
0.17875170707702637,
-0.19929231703281403,
-0.12851081788539886,
0.10677357763051987,
0.018770020455121994,
-0.19798976182937622,
-0.14952610433101654,
-0.10458961874246597,
-0.0204896479845047,
-0.12995094060897827,
0.041279539465904236,
0.014258908107876778,
0.014810405671596527,
0.010652083903551102,
0.02346709743142128,
0.03820135444402695,
-0.04403134435415268,
0.2022320032119751,
-0.040240850299596786,
-0.00677528977394104,
-0.05459889397025108,
-0.08097099512815475,
0.012206795625388622,
-0.05523540452122688,
0.12372337281703949,
-0.010677291080355644,
0.03454338386654854,
-0.17148974537849426,
-0.042799804359674454,
-0.06020277738571167,
0.035965804010629654,
-0.09800209105014801,
-0.08035019785165787,
-0.044318266212940216,
0.08121439814567566,
0.08592808991670609,
-0.011807112023234367,
0.004592899698764086,
-0.0995112806558609,
0.09020279347896576,
0.2008526772260666,
0.19356492161750793,
0.057227738201618195,
-0.056221771985292435,
0.033027902245521545,
-0.0363139733672142,
0.04097477346658707,
-0.2229323834180832,
0.039946265518665314,
0.0660935789346695,
0.027191683650016785,
0.07270630449056625,
-0.0050587123259902,
-0.16379666328430176,
-0.09244991093873978,
0.08992933481931686,
-0.05790415778756142,
-0.16807101666927338,
-0.03529549762606621,
0.04140728712081909,
-0.21035249531269073,
-0.04760543256998062,
0.037281136959791183,
-0.017871566116809845,
-0.04378291592001915,
0.0276334248483181,
0.0753527581691742,
-0.02573961578309536,
0.0857105553150177,
0.0968673974275589,
0.08900167047977448,
-0.09695399552583694,
0.051445744931697845,
0.07814038544893265,
-0.015816476196050644,
0.02846227027475834,
0.14087340235710144,
-0.03826410695910454,
-0.04601595178246498,
0.08259574323892593,
0.11946269869804382,
-0.011369331739842892,
-0.05124291777610779,
0.0039620790630578995,
-0.049147140234708786,
0.06518470495939255,
0.12247049808502197,
0.0250368844717741,
-0.014529009349644184,
0.07675154507160187,
0.02463647536933422,
-0.0901833325624466,
0.1191658079624176,
0.041540008038282394,
0.021193431690335274,
-0.03237847983837128,
-0.034603528678417206,
-0.012499326840043068,
0.0018930385122075677,
-0.013601796701550484,
-0.0026141954585909843,
-0.09225299209356308,
0.0024042355362325907,
-0.11352413147687912,
0.013482348993420601,
-0.060120537877082825,
0.0031534277368336916,
0.027116654440760612,
-0.051312822848558426,
-0.006352854426950216,
-0.0053253467194736,
-0.082282654941082,
-0.05532316118478775,
-0.02367786131799221,
0.07458903640508652,
-0.13407236337661743,
0.03929748386144638,
0.07778579741716385,
-0.10331465303897858,
0.06806895136833191,
-0.008187590166926384,
0.012664705514907837,
0.0053971088491380215,
-0.13933587074279785,
0.05808352306485176,
-0.03133996203541756,
-0.004783592652529478,
0.005858907010406256,
-0.1819247603416443,
-0.009001663886010647,
-0.04236048087477684,
-0.0687473714351654,
0.0123074259608984,
-0.01007341779768467,
-0.12471766024827957,
0.11227453500032425,
0.0002743391669355333,
-0.06740614026784897,
-0.014570803381502628,
0.04962038993835449,
0.07010367512702942,
-0.006095271557569504,
0.1029348373413086,
-0.02286364696919918,
0.08129947632551193,
-0.18399451673030853,
-0.0068764397874474525,
-0.01571904495358467,
0.05597268417477608,
-0.013975427486002445,
-0.05027436092495918,
0.05743827670812607,
-0.018061965703964233,
0.17213313281536102,
0.004430451430380344,
0.07709904760122299,
0.04961197450757027,
0.013601860031485558,
0.0427589975297451,
0.07069148123264313,
0.06631975620985031,
-0.017075147479772568,
-0.0007692971848882735,
0.03489156439900398,
0.003476202953606844,
-0.04633419215679169,
-0.13110819458961487,
0.07080189883708954,
0.17841176688671112,
0.07283827662467957,
0.022690467536449432,
0.013556991703808308,
-0.13281375169754028,
-0.07107050716876984,
0.10579551756381989,
-0.018156331032514572,
-0.028945455327630043,
-0.06893990933895111,
0.23089949786663055,
0.14948968589305878,
-0.19252033531665802,
0.07802820205688477,
-0.05396091192960739,
-0.039337433874607086,
-0.14227846264839172,
-0.16513042151927948,
-0.05926815792918205,
-0.05480305850505829,
-0.032690949738025665,
-0.06056531146168709,
0.05205392464995384,
0.03780041262507439,
-0.004041227512061596,
-0.02319422736763954,
0.10291451960802078,
0.02876470424234867,
-0.04133755341172218,
0.044333186000585556,
0.05758066475391388,
0.043520525097846985,
-0.10267458111047745,
0.012859337031841278,
0.00009975417924579233,
0.00781586766242981,
0.06644705682992935,
0.022875890135765076,
-0.068130262196064,
0.027877703309059143,
-0.0159730426967144,
-0.11902613937854767,
0.04861007258296013,
-0.008199073374271393,
-0.022566575556993484,
0.15131603181362152,
0.035203587263822556,
0.0075862049125134945,
-0.010280744172632694,
0.24109400808811188,
-0.07023292779922485,
-0.08434440195560455,
-0.133211150765419,
0.07812686264514923,
-0.06614357233047485,
0.023489415645599365,
0.012412266805768013,
-0.12309877574443817,
0.013406345620751381,
0.1877075582742691,
0.12149964272975922,
-0.018842259421944618,
0.010303139686584473,
0.051993947476148605,
0.010083645582199097,
-0.030714169144630432,
0.010844341479241848,
0.05824806168675423,
0.20381180942058563,
-0.08090908080339432,
0.05947291851043701,
-0.017558753490447998,
-0.07183664292097092,
-0.024221323430538177,
0.11179669946432114,
-0.0072897085919976234,
-0.014546004123985767,
-0.05833130329847336,
0.14228056371212006,
-0.07756908982992172,
-0.21425963938236237,
0.05329543352127075,
-0.0845317468047142,
-0.13932272791862488,
-0.05179408937692642,
0.02242196351289749,
-0.02796894498169422,
0.008448448032140732,
0.05870514735579491,
-0.05420953780412674,
0.1791611909866333,
0.02900891751050949,
-0.04865198954939842,
-0.10163167864084244,
0.0589936338365078,
-0.16139982640743256,
0.27064254879951477,
0.017819296568632126,
0.048935048282146454,
0.11234572529792786,
-0.015481040813028812,
-0.1309996396303177,
0.012740112841129303,
0.1132117360830307,
-0.060685716569423676,
0.06256001442670822,
0.15765917301177979,
0.0030845897272229195,
0.11834097653627396,
0.06628477573394775,
-0.056056614965200424,
0.03762415796518326,
-0.07457998394966125,
-0.04494589567184448,
-0.12201961129903793,
0.07539553195238113,
-0.09938636422157288,
0.15182992815971375,
0.12770362198352814,
-0.07337206602096558,
-0.005672953557223082,
-0.023329490795731544,
0.0787353366613388,
0.017383035272359848,
0.10956698656082153,
0.004856250248849392,
-0.18510037660598755,
0.04489986225962639,
0.004797583911567926,
0.09587015211582184,
-0.21170052886009216,
-0.05072372034192085,
0.04455697536468506,
-0.018744779750704765,
-0.08346759527921677,
0.1200728565454483,
0.04002266749739647,
0.020804665982723236,
-0.03638550639152527,
-0.048523325473070145,
0.016989044845104218,
0.1550002098083496,
-0.10584764182567596,
-0.014470396563410759
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bart-noised-with-gcd-dist-0.1
This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 10
- num_epochs: 3
- mixed_precision_training: Native AMP
### Training results
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.2+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "facebook/bart-base", "model-index": [{"name": "bart-noised-with-gcd-dist-0.1", "results": []}]} | text2text-generation | gayanin/bart-noised-with-gcd-dist-0.1 | [
"transformers",
"safetensors",
"bart",
"text2text-generation",
"generated_from_trainer",
"base_model:facebook/bart-base",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T17:28:08+00:00 | [] | [] | TAGS
#transformers #safetensors #bart #text2text-generation #generated_from_trainer #base_model-facebook/bart-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
# bart-noised-with-gcd-dist-0.1
This model is a fine-tuned version of facebook/bart-base on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 10
- num_epochs: 3
- mixed_precision_training: Native AMP
### Training results
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.2+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| [
"# bart-noised-with-gcd-dist-0.1\n\nThis model is a fine-tuned version of facebook/bart-base on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 10\n- num_epochs: 3\n- mixed_precision_training: Native AMP",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.1.2+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #safetensors #bart #text2text-generation #generated_from_trainer #base_model-facebook/bart-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# bart-noised-with-gcd-dist-0.1\n\nThis model is a fine-tuned version of facebook/bart-base on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 10\n- num_epochs: 3\n- mixed_precision_training: Native AMP",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.1.2+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
64,
38,
6,
12,
8,
3,
118,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #bart #text2text-generation #generated_from_trainer #base_model-facebook/bart-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# bart-noised-with-gcd-dist-0.1\n\nThis model is a fine-tuned version of facebook/bart-base on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 10\n- num_epochs: 3\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.1.2+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.09655893594026566,
0.10926298052072525,
-0.00282420520670712,
0.07439061254262924,
0.1303579956293106,
0.028945935890078545,
0.11453680694103241,
0.12269508093595505,
-0.0379435271024704,
0.06515540927648544,
0.06291167438030243,
0.029568517580628395,
0.04363338649272919,
0.18533754348754883,
-0.05316539108753204,
-0.19335491955280304,
0.022684963420033455,
-0.05543617531657219,
-0.08390999585390091,
0.10438170284032822,
0.08950456976890564,
-0.08538476377725601,
0.07464352995157242,
-0.012777931988239288,
-0.16501019895076752,
0.021246088668704033,
-0.00979650765657425,
-0.041754432022571564,
0.10406513512134552,
0.020389456301927567,
0.07797714322805405,
0.04267371818423271,
0.13829796016216278,
-0.20813925564289093,
0.0021960833109915257,
0.09393995255231857,
0.016054242849349976,
0.07706339657306671,
0.056525442749261856,
-0.025121740996837616,
0.09079672396183014,
-0.1312480866909027,
0.09947547316551208,
0.035151463001966476,
-0.09778351336717606,
-0.14330749213695526,
-0.09337221831083298,
0.04458147659897804,
0.09865394234657288,
0.0862344354391098,
0.0033270083367824554,
0.07174553722143173,
-0.11548040062189102,
0.06301147490739822,
0.19827422499656677,
-0.2591939866542816,
-0.06435108184814453,
0.02202998846769333,
0.04515222832560539,
0.052400168031454086,
-0.10748716443777084,
-0.005821059923619032,
0.03743714839220047,
0.03227647766470909,
0.1166016235947609,
-0.01307571493089199,
-0.07660557329654694,
-0.01849961094558239,
-0.11333653330802917,
-0.003145448165014386,
0.0900542214512825,
0.05177024006843567,
-0.053026460111141205,
-0.08765885978937149,
-0.059633661061525345,
-0.044276848435401917,
-0.018356038257479668,
-0.028156686574220657,
0.026948127895593643,
-0.03803634271025658,
-0.052788425236940384,
-0.042520347982645035,
-0.05789879336953163,
-0.07522998005151749,
0.009867528453469276,
0.1396535485982895,
0.02110404707491398,
0.019808243960142136,
-0.03254103660583496,
0.10451208055019379,
0.008392317220568657,
-0.12640994787216187,
0.005874264519661665,
-0.011203549802303314,
-0.12196871638298035,
-0.04185708239674568,
-0.04252638667821884,
0.01311800442636013,
0.01577550731599331,
0.15547746419906616,
-0.0483839176595211,
0.08178494870662689,
-0.0059797028079628944,
-0.013303438201546669,
-0.030677851289510727,
0.11788947135210037,
-0.03557619825005531,
-0.07471024990081787,
0.009272417053580284,
0.10550229996442795,
0.006140417885035276,
-0.018806295469403267,
-0.06899547576904297,
-0.022484151646494865,
0.06523013859987259,
0.058028172701597214,
-0.04323625564575195,
0.025709975510835648,
-0.047848477959632874,
-0.02208106219768524,
0.05614243820309639,
-0.13262690603733063,
0.05002753809094429,
0.022249286994338036,
-0.07685355097055435,
-0.06662937998771667,
0.03190866485238075,
0.0034475699067115784,
-0.005520170088857412,
0.10367302596569061,
-0.05825810879468918,
-0.009857245720922947,
-0.07527998089790344,
-0.06432042270898819,
0.009564869105815887,
-0.07207397371530533,
-0.017743688076734543,
-0.06094016879796982,
-0.2047811597585678,
-0.04491483047604561,
0.04767095297574997,
-0.07616198062896729,
-0.03536801412701607,
-0.04724312573671341,
-0.05783455818891525,
0.0327615886926651,
-0.02494416944682598,
0.1481310874223709,
-0.06282143294811249,
0.06610453128814697,
-0.011607399210333824,
0.044753435999155045,
0.022266866639256477,
0.04256180301308632,
-0.08126925677061081,
0.031463805586099625,
-0.17104437947273254,
0.09557372331619263,
-0.09036773443222046,
0.016505684703588486,
-0.10301651805639267,
-0.06559668481349945,
0.011259901337325573,
-0.013946550898253918,
0.06127103790640831,
0.13500942289829254,
-0.21241098642349243,
-0.034975018352270126,
0.1527608036994934,
-0.09233494102954865,
-0.057777222245931625,
0.07271767407655716,
-0.04783487319946289,
0.027743564918637276,
0.056430719792842865,
0.1703473925590515,
0.1015060618519783,
-0.13219420611858368,
0.004960808902978897,
-0.002783144824206829,
0.04745271056890488,
0.042154621332883835,
0.031727198511362076,
-0.0018171387491747737,
0.01232514064759016,
0.00847323052585125,
-0.06671886146068573,
0.006179068703204393,
-0.07404103875160217,
-0.06734251230955124,
-0.0504021979868412,
-0.08619940280914307,
0.053728070110082626,
0.004724117927253246,
0.01030578650534153,
-0.07287026196718216,
-0.10960710793733597,
0.08293086290359497,
0.11678597331047058,
-0.05970415472984314,
0.008721145801246166,
-0.06479398906230927,
0.01029970869421959,
0.018460016697645187,
-0.0218346007168293,
-0.1924084573984146,
-0.11944860965013504,
0.03661610186100006,
-0.07952654361724854,
0.034771110862493515,
0.010310395620763302,
0.06914428621530533,
0.06566745787858963,
-0.0462271049618721,
-0.021948453038930893,
-0.07822395116090775,
0.004882408306002617,
-0.08128977566957474,
-0.2278910130262375,
-0.05035370960831642,
-0.03777066990733147,
0.18481548130512238,
-0.22047662734985352,
0.010674619115889072,
-0.013755942694842815,
0.15297682583332062,
0.03187764436006546,
-0.04877949133515358,
-0.008668521419167519,
0.02273613214492798,
0.02124805748462677,
-0.09561184048652649,
0.026969412341713905,
0.008527098223567009,
-0.08411338925361633,
-0.02286565862596035,
-0.12642426788806915,
0.040774427354335785,
0.0656665563583374,
0.09982317686080933,
-0.10088586807250977,
-0.0224603358656168,
-0.06808637082576752,
-0.04003738984465599,
-0.08731681108474731,
0.012920912355184555,
0.1855381727218628,
0.029503123834729195,
0.10341078042984009,
-0.05666453391313553,
-0.06681101769208908,
-0.0031643006950616837,
0.007923653349280357,
0.01666964404284954,
0.07908041775226593,
0.06939078867435455,
-0.10582127422094345,
0.081393763422966,
0.11489321291446686,
-0.006430463865399361,
0.11944042891263962,
-0.03036586195230484,
-0.0704551711678505,
-0.01672724448144436,
0.00012063273607054725,
-0.026946019381284714,
0.13895927369594574,
-0.05285809561610222,
0.02412865310907364,
0.017486466094851494,
0.024097273126244545,
0.02660435065627098,
-0.1623501181602478,
-0.0038013551384210587,
0.004549257457256317,
-0.0699024572968483,
-0.024609729647636414,
-0.0332704596221447,
0.05247678980231285,
0.09558696299791336,
0.02103160321712494,
-0.03210268169641495,
0.02175399288535118,
-0.019542289897799492,
-0.08145685493946075,
0.17472544312477112,
-0.12085537612438202,
-0.16251826286315918,
-0.07567790150642395,
0.04670439288020134,
-0.040356334298849106,
-0.02947743982076645,
0.018838150426745415,
-0.08715908974409103,
-0.06401907652616501,
-0.08348392695188522,
0.003277840558439493,
0.03156930208206177,
-0.004993753042072058,
0.04641811549663544,
0.009027143940329552,
0.091402567923069,
-0.11275792866945267,
-0.00002294840305694379,
-0.024040857329964638,
-0.08310460299253464,
0.003383594797924161,
0.07314679771661758,
0.06380530446767807,
0.10228390246629715,
-0.005461286287754774,
0.01608460210263729,
-0.024355478584766388,
0.22026148438453674,
-0.07762793451547623,
0.02340124174952507,
0.10829848051071167,
-0.004319855477660894,
0.04687760770320892,
0.15210239589214325,
0.019901692867279053,
-0.11312136054039001,
0.04540896415710449,
0.08319826424121857,
-0.007318784482777119,
-0.22489401698112488,
-0.051208142191171646,
-0.015981461852788925,
-0.05691815912723541,
0.08955118060112,
0.041468288749456406,
-0.0298154354095459,
0.019755573943257332,
-0.01043040119111538,
0.0011278318706899881,
0.027102166786789894,
0.05287739261984825,
0.0630548819899559,
0.0497189536690712,
0.10317142307758331,
-0.014904525130987167,
0.004931039176881313,
0.07797543704509735,
-0.010011042468249798,
0.25185447931289673,
-0.03718306124210358,
0.050233494490385056,
0.041761670261621475,
0.139320507645607,
-0.020995626226067543,
0.02874370478093624,
0.027353106066584587,
-0.006044482346624136,
-0.007134999614208937,
-0.05614151060581207,
-0.02615256793797016,
0.02044936828315258,
-0.05815046653151512,
0.016725152730941772,
-0.10260818898677826,
0.04030008241534233,
0.03656083717942238,
0.28371742367744446,
0.048689037561416626,
-0.26262977719306946,
-0.07240821421146393,
0.006307387258857489,
-0.043051958084106445,
-0.06245075538754463,
0.005632386542856693,
0.1333683580160141,
-0.13110338151454926,
0.07807677239179611,
-0.057317882776260376,
0.08285189419984818,
-0.013653586618602276,
0.01816280372440815,
0.07747064530849457,
0.13233517110347748,
-0.007397810462862253,
0.06553039699792862,
-0.23802921175956726,
0.2099406123161316,
0.01604073867201805,
0.10580616444349289,
-0.0567886047065258,
0.02351115643978119,
0.017164818942546844,
0.05672040581703186,
0.0923035740852356,
0.0008702089544385672,
-0.06853681802749634,
-0.12296892702579498,
-0.10251093655824661,
0.04960210248827934,
0.10962124168872833,
-0.04570982605218887,
0.07249968498945236,
-0.04382145777344704,
-0.0075374869629740715,
0.04432709887623787,
-0.06118425726890564,
-0.18439260125160217,
-0.11354446411132812,
0.006908663548529148,
-0.007001516409218311,
-0.025537192821502686,
-0.0847509503364563,
-0.08842331171035767,
0.013887005858123302,
0.16916117072105408,
0.029441526159644127,
-0.03541595861315727,
-0.1476108729839325,
0.05904857814311981,
0.1272485852241516,
-0.056299492716789246,
0.01690174825489521,
0.027489565312862396,
0.12400515377521515,
0.03788118064403534,
-0.11135301738977432,
0.07806248217821121,
-0.08646563440561295,
-0.18797533214092255,
-0.05540568381547928,
0.12356209754943848,
0.09539294987916946,
0.04419751465320587,
0.008329033851623535,
0.016725897789001465,
0.02475805953145027,
-0.08977510780096054,
0.013283432461321354,
0.07307581603527069,
0.04043789580464363,
0.05466291308403015,
-0.06924883276224136,
-0.020918019115924835,
-0.02745526097714901,
-0.010378316976130009,
0.08077777922153473,
0.21592597663402557,
-0.08975259214639664,
0.12741512060165405,
0.0823938250541687,
-0.07252886146306992,
-0.1776410937309265,
0.07352955639362335,
0.11251313239336014,
0.023328686133027077,
0.05338885262608528,
-0.18453583121299744,
0.13372370600700378,
0.10972487926483154,
-0.029676342383027077,
0.03965480253100395,
-0.2996101379394531,
-0.13474823534488678,
0.07042238116264343,
0.0864967405796051,
0.03482581675052643,
-0.10954587906599045,
-0.03619740158319473,
-0.050302326679229736,
-0.1328980028629303,
0.14887215197086334,
-0.14331823587417603,
0.09551214426755905,
0.0017243963666260242,
0.0828206017613411,
0.016176270321011543,
-0.01867758296430111,
0.12267572432756424,
0.04587240144610405,
0.09559673070907593,
-0.04734497144818306,
0.05636044964194298,
0.020559724420309067,
-0.06462086737155914,
0.02209143154323101,
-0.06488902866840363,
0.06450366973876953,
-0.1104557141661644,
-0.008754842914640903,
-0.09530829638242722,
0.0722159892320633,
-0.06106874719262123,
-0.045946232974529266,
-0.026934420689940453,
0.05978516861796379,
0.06218801438808441,
-0.03225555270910263,
-0.007221200503408909,
-0.001159807899966836,
0.1068287268280983,
0.09200429916381836,
0.10156470537185669,
-0.03315272927284241,
-0.04468984156847,
0.010365026071667671,
-0.011636714451014996,
0.04371055215597153,
-0.10556032508611679,
0.046111270785331726,
0.11563849449157715,
0.03203338757157326,
0.13756774365901947,
0.0252789705991745,
-0.05548744648694992,
-0.011594646610319614,
0.0441247895359993,
-0.1369633972644806,
-0.1070425882935524,
0.01966378279030323,
-0.03816205635666847,
-0.10940408706665039,
0.0022170173469930887,
0.13673090934753418,
-0.03640732169151306,
-0.01792476698756218,
-0.020240135490894318,
0.03619251772761345,
-0.026622002944350243,
0.18005499243736267,
0.024849850684404373,
0.06155605986714363,
-0.09958803653717041,
0.13903798162937164,
0.0544009730219841,
-0.06774222105741501,
0.056597430258989334,
0.0934448093175888,
-0.09757852554321289,
-0.010785089805722237,
0.07151110470294952,
0.17830756306648254,
-0.021269671618938446,
-0.03633377328515053,
-0.08183151483535767,
-0.12272355705499649,
0.05139753594994545,
0.16533492505550385,
0.027593553066253662,
-0.003632887499406934,
-0.01594037376344204,
0.026734095066785812,
-0.11311328411102295,
0.07436702400445938,
0.06907026469707489,
0.0505676344037056,
-0.11239102482795715,
0.10581284016370773,
0.008532490581274033,
0.024310100823640823,
-0.025315893813967705,
0.04002420976758003,
-0.1076907068490982,
-0.035573218017816544,
-0.16495393216609955,
-0.002311424585059285,
-0.01746192015707493,
0.009824021719396114,
-0.016477705910801888,
-0.05369348078966141,
-0.0321318544447422,
0.028927795588970184,
-0.07443283498287201,
-0.0524035207927227,
0.0005912237684242427,
0.040445875376462936,
-0.17378674447536469,
-0.015482762828469276,
0.026291009038686752,
-0.09948597848415375,
0.08923426270484924,
0.0767262876033783,
0.026951197534799576,
0.03156464174389839,
-0.15207728743553162,
-0.03320015221834183,
0.019002893939614296,
0.009714464657008648,
0.07643424719572067,
-0.11502088606357574,
-0.012318115681409836,
-0.023532109335064888,
0.04238378256559372,
0.02108018472790718,
0.08045130968093872,
-0.10877873003482819,
0.00038359870086424053,
-0.05211808159947395,
-0.05432881787419319,
-0.0540030263364315,
0.035721033811569214,
0.11001531779766083,
0.03865519165992737,
0.14897651970386505,
-0.09616310894489288,
0.040392596274614334,
-0.20483200252056122,
-0.03918338567018509,
-0.0030050294008105993,
-0.013359973207116127,
-0.0799051895737648,
-0.02712416835129261,
0.09764178097248077,
-0.038455478847026825,
0.11228182911872864,
0.001958429114893079,
0.10787658393383026,
0.036189332604408264,
-0.07594801485538483,
-0.04190557077527046,
0.029085004702210426,
0.09817484021186829,
0.03983175754547119,
-0.009815393015742302,
0.10419081151485443,
-0.021850289776921272,
0.046687059104442596,
0.01589117757976055,
0.23888838291168213,
0.1598385125398636,
-0.007261485327035189,
0.04619481414556503,
0.07148031145334244,
-0.1361362487077713,
-0.11021307855844498,
0.12186292558908463,
-0.06865812093019485,
0.10882802307605743,
-0.06626539677381516,
0.1849967986345291,
0.04636760428547859,
-0.18636193871498108,
0.053957562893629074,
-0.06109408661723137,
-0.10618910193443298,
-0.12169431895017624,
-0.017022058367729187,
-0.07910960167646408,
-0.1355903148651123,
0.02341342903673649,
-0.12047040462493896,
0.05483577400445938,
0.07198087126016617,
0.00464055547490716,
0.020495031028985977,
0.13976271450519562,
-0.020206671208143234,
0.011101099662482738,
0.07200907915830612,
0.030326101928949356,
-0.0028007293585687876,
-0.037046439945697784,
-0.0662202537059784,
0.03968512639403343,
0.048251692205667496,
0.053836509585380554,
-0.04119066148996353,
-0.010289529338479042,
0.03385387733578682,
-0.009025021456182003,
-0.07200619578361511,
0.03631126508116722,
0.004890755750238895,
0.038347940891981125,
0.058433085680007935,
0.06111561879515648,
0.009341818280518055,
-0.028640737757086754,
0.31884798407554626,
-0.08407708257436752,
-0.08165699988603592,
-0.1465635895729065,
0.20096896588802338,
0.0051786466501653194,
-0.00670554768294096,
0.055940911173820496,
-0.09433909505605698,
-0.031018299981951714,
0.1403484046459198,
0.14009243249893188,
-0.1116802990436554,
-0.010907472111284733,
-0.04468550905585289,
-0.014198904857039452,
-0.03164459019899368,
0.13750018179416656,
0.10383610427379608,
0.0284221563488245,
-0.05941189080476761,
-0.013811292126774788,
0.00651969388127327,
-0.05013885349035263,
-0.0618051141500473,
0.08322784304618835,
-0.011016549542546272,
0.01377270556986332,
-0.03102080337703228,
0.07117236405611038,
0.016075685620307922,
-0.2165103405714035,
0.008293368853628635,
-0.16408413648605347,
-0.1817115694284439,
-0.03477444499731064,
0.061867062002420425,
-0.01132897287607193,
0.05693792179226875,
-0.0018975319107994437,
-0.011209681630134583,
0.15119269490242004,
-0.03070417419075966,
-0.021853337064385414,
-0.11637330800294876,
0.07710622996091843,
-0.1276211440563202,
0.22919049859046936,
-0.0061588897369802,
0.04533535987138748,
0.09368320554494858,
0.0254666805267334,
-0.13636235892772675,
0.029969841241836548,
0.05611412599682808,
-0.08008450269699097,
0.015618129633367062,
0.1630401909351349,
-0.05196545273065567,
0.08904904872179031,
0.06000591069459915,
-0.1305055320262909,
-0.0038072816096246243,
-0.03796996921300888,
-0.04221195727586746,
-0.06652497500181198,
-0.007120438851416111,
-0.05568786710500717,
0.15376171469688416,
0.21041233837604523,
-0.03268301486968994,
0.03467399254441261,
-0.08401463180780411,
0.0275947954505682,
0.03254563361406326,
0.10608094930648804,
-0.025909457355737686,
-0.20968690514564514,
0.033737003803253174,
0.0707363411784172,
0.027115697041153908,
-0.22968824207782745,
-0.08947282284498215,
0.020120596513152122,
-0.0442751869559288,
-0.0500609315931797,
0.13020192086696625,
0.021679416298866272,
0.040557511150836945,
-0.031101800501346588,
-0.1253862828016281,
-0.03409486636519432,
0.1606968343257904,
-0.1545112282037735,
-0.03817758709192276
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bart-noised-with-gcd-dist-0.2
This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 10
- num_epochs: 3
- mixed_precision_training: Native AMP
### Training results
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.2+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "facebook/bart-base", "model-index": [{"name": "bart-noised-with-gcd-dist-0.2", "results": []}]} | text2text-generation | gayanin/bart-noised-with-gcd-dist-0.2 | [
"transformers",
"safetensors",
"bart",
"text2text-generation",
"generated_from_trainer",
"base_model:facebook/bart-base",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T17:28:55+00:00 | [] | [] | TAGS
#transformers #safetensors #bart #text2text-generation #generated_from_trainer #base_model-facebook/bart-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
# bart-noised-with-gcd-dist-0.2
This model is a fine-tuned version of facebook/bart-base on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 10
- num_epochs: 3
- mixed_precision_training: Native AMP
### Training results
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.2+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| [
"# bart-noised-with-gcd-dist-0.2\n\nThis model is a fine-tuned version of facebook/bart-base on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 10\n- num_epochs: 3\n- mixed_precision_training: Native AMP",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.1.2+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #safetensors #bart #text2text-generation #generated_from_trainer #base_model-facebook/bart-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# bart-noised-with-gcd-dist-0.2\n\nThis model is a fine-tuned version of facebook/bart-base on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 10\n- num_epochs: 3\n- mixed_precision_training: Native AMP",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.1.2+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
64,
38,
6,
12,
8,
3,
118,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #bart #text2text-generation #generated_from_trainer #base_model-facebook/bart-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# bart-noised-with-gcd-dist-0.2\n\nThis model is a fine-tuned version of facebook/bart-base on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 10\n- num_epochs: 3\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.1.2+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.09646899253129959,
0.10994160920381546,
-0.0028676774818450212,
0.07422278076410294,
0.12993097305297852,
0.02918962761759758,
0.11443391442298889,
0.12261918187141418,
-0.03815293312072754,
0.06549184024333954,
0.0621463805437088,
0.029499882832169533,
0.043852515518665314,
0.1855185329914093,
-0.05328277498483658,
-0.19409357011318207,
0.022626886144280434,
-0.05556608736515045,
-0.08443257212638855,
0.10439693927764893,
0.0895853340625763,
-0.08548204600811005,
0.0746786817908287,
-0.012335622683167458,
-0.164951890707016,
0.020827967673540115,
-0.009238562546670437,
-0.041602738201618195,
0.10385110229253769,
0.020259743556380272,
0.07748426496982574,
0.042610619217157364,
0.1383470594882965,
-0.2079392820596695,
0.0021161693148314953,
0.09377768635749817,
0.015982098877429962,
0.0776590034365654,
0.056969404220581055,
-0.0248719509691,
0.09197411686182022,
-0.13077446818351746,
0.09919419139623642,
0.034907642751932144,
-0.097798191010952,
-0.14240960776805878,
-0.09380697458982468,
0.04453551769256592,
0.0995483472943306,
0.08545982092618942,
0.0033925508614629507,
0.07190603017807007,
-0.11579842865467072,
0.0627242773771286,
0.19800223410129547,
-0.2598258852958679,
-0.06450953334569931,
0.022137075662612915,
0.04531434550881386,
0.05205675587058067,
-0.10711279511451721,
-0.0057580675929784775,
0.03705943003296852,
0.03201070427894592,
0.11613591760396957,
-0.012865525670349598,
-0.07613261044025421,
-0.018237324431538582,
-0.11335873603820801,
-0.002997222589328885,
0.0898665338754654,
0.05183888599276543,
-0.0529085211455822,
-0.08760207146406174,
-0.05990776792168617,
-0.04332917928695679,
-0.01822333224117756,
-0.028425291180610657,
0.027265964075922966,
-0.03778046369552612,
-0.0528004989027977,
-0.04278949275612831,
-0.05771311745047569,
-0.07587343454360962,
0.009503223933279514,
0.13975940644741058,
0.021025467664003372,
0.0201199222356081,
-0.03307192400097847,
0.10451995581388474,
0.008402109146118164,
-0.1267067939043045,
0.005695766769349575,
-0.011657316237688065,
-0.12182600051164627,
-0.04166959598660469,
-0.042502254247665405,
0.013071312569081783,
0.01564512588083744,
0.1553325057029724,
-0.04817940294742584,
0.08180905133485794,
-0.006681832019239664,
-0.012998608872294426,
-0.030749985948204994,
0.11793272942304611,
-0.0352022759616375,
-0.07503903657197952,
0.008789786137640476,
0.10536984354257584,
0.006197887472808361,
-0.019085494801402092,
-0.06874505430459976,
-0.023072710260748863,
0.06490904837846756,
0.05787568911910057,
-0.04305358976125717,
0.02556477300822735,
-0.048174940049648285,
-0.022664785385131836,
0.05601776763796806,
-0.13262562453746796,
0.050174061208963394,
0.022366879507899284,
-0.07731349766254425,
-0.06641065329313278,
0.03169749304652214,
0.0040164426900446415,
-0.005560880061239004,
0.10394316166639328,
-0.05786428973078728,
-0.010247325524687767,
-0.07544907182455063,
-0.06390833854675293,
0.009392469190061092,
-0.07161146402359009,
-0.018260208889842033,
-0.06067315861582756,
-0.20451416075229645,
-0.04441317170858383,
0.04782751575112343,
-0.07629668712615967,
-0.03594333678483963,
-0.047533731907606125,
-0.057546090334653854,
0.03309616446495056,
-0.025501219555735588,
0.14806459844112396,
-0.06266357004642487,
0.06605300307273865,
-0.011938853189349174,
0.045088935643434525,
0.02220199443399906,
0.04235748574137688,
-0.08086538314819336,
0.03155360370874405,
-0.17097489535808563,
0.09558431804180145,
-0.0893668681383133,
0.015249563381075859,
-0.10344917327165604,
-0.06569898128509521,
0.011026115156710148,
-0.014130646362900734,
0.061319850385189056,
0.13497352600097656,
-0.2131568342447281,
-0.035057876259088516,
0.15292634069919586,
-0.09219323098659515,
-0.05762071534991264,
0.07254838943481445,
-0.04794174060225487,
0.028005380183458328,
0.056375544518232346,
0.17038081586360931,
0.10130217671394348,
-0.13187018036842346,
0.0057224612683057785,
-0.0027170483954250813,
0.04709414392709732,
0.04287989065051079,
0.03215490281581879,
-0.0022278244141489267,
0.012940849177539349,
0.008383459411561489,
-0.06743577122688293,
0.006126354448497295,
-0.07385021448135376,
-0.06734848767518997,
-0.04992061108350754,
-0.08615041524171829,
0.05466170981526375,
0.0043425667099654675,
0.010437382385134697,
-0.07243555784225464,
-0.1092485710978508,
0.08363991230726242,
0.11697788536548615,
-0.05963366851210594,
0.00884126964956522,
-0.06555468589067459,
0.010314530692994595,
0.019396789371967316,
-0.02202446386218071,
-0.19257324934005737,
-0.11919204145669937,
0.03677355498075485,
-0.0796055868268013,
0.03441789001226425,
0.010924408212304115,
0.06910497695207596,
0.06580086797475815,
-0.04616408422589302,
-0.022070229053497314,
-0.0782981812953949,
0.0045852260664105415,
-0.08133485168218613,
-0.22761079668998718,
-0.04994860664010048,
-0.03730778768658638,
0.18450260162353516,
-0.22011877596378326,
0.010476007126271725,
-0.013335514813661575,
0.1535625010728836,
0.03207328915596008,
-0.0487687885761261,
-0.009371727705001831,
0.022135986015200615,
0.020932136103510857,
-0.09528148174285889,
0.027263600379228592,
0.008524475619196892,
-0.08416431397199631,
-0.022438373416662216,
-0.126432403922081,
0.04083085060119629,
0.06571357697248459,
0.10019505023956299,
-0.09992567449808121,
-0.02272890880703926,
-0.06810703128576279,
-0.03991552069783211,
-0.0876387357711792,
0.013195192441344261,
0.1854504644870758,
0.02951836958527565,
0.1036411002278328,
-0.056917767971754074,
-0.06680720299482346,
-0.0035897165071219206,
0.007895432412624359,
0.0164580550044775,
0.0787537544965744,
0.0693284124135971,
-0.10585291683673859,
0.0807243287563324,
0.11521703004837036,
-0.006070963107049465,
0.1194186881184578,
-0.030175376683473587,
-0.07036745548248291,
-0.0172482430934906,
0.0001681081485003233,
-0.02645755559206009,
0.138290673494339,
-0.05195286497473717,
0.024635761976242065,
0.017748123034834862,
0.024609306827187538,
0.026553157716989517,
-0.16235420107841492,
-0.00366766145452857,
0.004315823782235384,
-0.07013100385665894,
-0.024469103664159775,
-0.03263499215245247,
0.052503958344459534,
0.09597626328468323,
0.0215695109218359,
-0.032097138464450836,
0.021544449031352997,
-0.020066341385245323,
-0.08174189925193787,
0.1742437332868576,
-0.12106466293334961,
-0.16266071796417236,
-0.07568269222974777,
0.04657876491546631,
-0.04018136486411095,
-0.029223255813121796,
0.018337441608309746,
-0.08657286316156387,
-0.0639253705739975,
-0.0832311287522316,
0.003424775553867221,
0.0311539676040411,
-0.004744587931782007,
0.04659596458077431,
0.009512333199381828,
0.0918722078204155,
-0.11273129284381866,
-0.00005335496462066658,
-0.024138811975717545,
-0.08323031663894653,
0.0030201219487935305,
0.07313765585422516,
0.06316307932138443,
0.10211669653654099,
-0.005237583536654711,
0.016269268468022346,
-0.02411535382270813,
0.2200406789779663,
-0.07771757245063782,
0.023279735818505287,
0.10879720747470856,
-0.00497437035664916,
0.04699545353651047,
0.15227514505386353,
0.019544506445527077,
-0.11335594207048416,
0.04569736123085022,
0.08302643150091171,
-0.007184482645243406,
-0.22470617294311523,
-0.05162913352251053,
-0.01617545820772648,
-0.0566776767373085,
0.08944128453731537,
0.04132774844765663,
-0.03051566705107689,
0.019285276532173157,
-0.010473117232322693,
0.0011860502418130636,
0.027446940541267395,
0.05277084559202194,
0.06276974827051163,
0.05030728504061699,
0.10316599905490875,
-0.014590899460017681,
0.0050676558166742325,
0.07821919023990631,
-0.010232623666524887,
0.25087210536003113,
-0.03727025166153908,
0.04937062785029411,
0.041941698640584946,
0.13967300951480865,
-0.020757248625159264,
0.028604982420802116,
0.027124159038066864,
-0.0056378659792244434,
-0.007449834141880274,
-0.05623925104737282,
-0.026818351820111275,
0.020709138363599777,
-0.057330358773469925,
0.016899337992072105,
-0.10254883021116257,
0.04060503840446472,
0.03675997257232666,
0.2838408052921295,
0.049060624092817307,
-0.2624892294406891,
-0.07236893475055695,
0.006980337202548981,
-0.04293734207749367,
-0.06231901794672012,
0.005683484952896833,
0.1336641013622284,
-0.13129766285419464,
0.07855615764856339,
-0.05748485401272774,
0.08276508003473282,
-0.013663173653185368,
0.0183672234416008,
0.07692524790763855,
0.13208602368831635,
-0.006930053234100342,
0.06555619835853577,
-0.23770615458488464,
0.20962673425674438,
0.01594211533665657,
0.10540817677974701,
-0.05642550811171532,
0.023540308699011803,
0.01734038069844246,
0.05628174915909767,
0.0920364111661911,
0.0008206625352613628,
-0.06884513795375824,
-0.12216119468212128,
-0.10212104022502899,
0.04932859539985657,
0.10996334999799728,
-0.04623446613550186,
0.0725221186876297,
-0.044062111526727676,
-0.007597993593662977,
0.04408213123679161,
-0.06129591912031174,
-0.1837555319070816,
-0.11315419524908066,
0.007119889836758375,
-0.007866958156228065,
-0.025140533223748207,
-0.08487066626548767,
-0.08830290287733078,
0.014348303899168968,
0.16995231807231903,
0.0280615221709013,
-0.03538046032190323,
-0.14728780090808868,
0.058726053684949875,
0.1272357851266861,
-0.05633847787976265,
0.01676703244447708,
0.027769774198532104,
0.12522441148757935,
0.03813108056783676,
-0.11150407046079636,
0.07810427993535995,
-0.08621501922607422,
-0.18871410191059113,
-0.0554010383784771,
0.12407732754945755,
0.09578624367713928,
0.04414225369691849,
0.008527861908078194,
0.017022524029016495,
0.025219647213816643,
-0.08972181379795074,
0.013341897167265415,
0.07413512468338013,
0.040969833731651306,
0.05458975210785866,
-0.06881734728813171,
-0.02088409662246704,
-0.027216879650950432,
-0.009785336442291737,
0.08022330701351166,
0.21675634384155273,
-0.08955641835927963,
0.12832516431808472,
0.08189424127340317,
-0.0726274847984314,
-0.1775261014699936,
0.07326734066009521,
0.11276417970657349,
0.02368355542421341,
0.05449383705854416,
-0.1841862052679062,
0.13348433375358582,
0.10972985625267029,
-0.029932832345366478,
0.03855714574456215,
-0.3000148832798004,
-0.1349380910396576,
0.06965405493974686,
0.08603206276893616,
0.03404786437749863,
-0.10878407955169678,
-0.03639965504407883,
-0.050981227308511734,
-0.1332293152809143,
0.14941957592964172,
-0.14391066133975983,
0.09489911049604416,
0.0013207305455580354,
0.0827355682849884,
0.01653423346579075,
-0.018402840942144394,
0.12332771718502045,
0.0458139143884182,
0.09483591467142105,
-0.047610197216272354,
0.055960409343242645,
0.020419662818312645,
-0.06500022858381271,
0.021640541031956673,
-0.06503583490848541,
0.0648580864071846,
-0.11069871485233307,
-0.008737929165363312,
-0.09500598162412643,
0.07165835797786713,
-0.061204563826322556,
-0.04578933119773865,
-0.027307894080877304,
0.06009860709309578,
0.06249937042593956,
-0.0316193588078022,
-0.007430369965732098,
-0.000709771178662777,
0.10627719759941101,
0.0920509472489357,
0.10249369591474533,
-0.03201664611697197,
-0.04431924968957901,
0.009975375607609749,
-0.011129648424685001,
0.043726760894060135,
-0.10571575164794922,
0.04643040895462036,
0.11564372479915619,
0.03173629567027092,
0.13737335801124573,
0.02537347935140133,
-0.05599045008420944,
-0.011228852905333042,
0.04442223161458969,
-0.13749514520168304,
-0.10720065981149673,
0.02007218264043331,
-0.03798208385705948,
-0.10969585925340652,
0.001857856521382928,
0.13694773614406586,
-0.03560911864042282,
-0.01776888221502304,
-0.01969311386346817,
0.03670022636651993,
-0.026884106919169426,
0.1798657774925232,
0.02426826022565365,
0.0615478940308094,
-0.09932199120521545,
0.13891386985778809,
0.05417069420218468,
-0.06697995215654373,
0.056384116411209106,
0.09335467219352722,
-0.09757038950920105,
-0.010043743997812271,
0.07122505456209183,
0.17721399664878845,
-0.02053302526473999,
-0.03641551360487938,
-0.08120250701904297,
-0.12247025221586227,
0.051252517849206924,
0.16447675228118896,
0.02768823318183422,
-0.0033539969008415937,
-0.016496585682034492,
0.02689446695148945,
-0.11237625032663345,
0.07405734807252884,
0.06925411522388458,
0.050343193113803864,
-0.11195023357868195,
0.106354258954525,
0.008428772911429405,
0.02424999140202999,
-0.025350850075483322,
0.03990298882126808,
-0.10757998377084732,
-0.03547874093055725,
-0.1647796779870987,
-0.002326301531866193,
-0.017396239563822746,
0.009531237185001373,
-0.016645845025777817,
-0.05400628596544266,
-0.03250596672296524,
0.028782209381461143,
-0.07433586567640305,
-0.052538398653268814,
0.0001819104072637856,
0.0403473898768425,
-0.17384831607341766,
-0.015298676677048206,
0.026723941788077354,
-0.09978752583265305,
0.08952180296182632,
0.07665358483791351,
0.027192248031497,
0.032295797020196915,
-0.15202614665031433,
-0.033856406807899475,
0.018947700038552284,
0.009889857843518257,
0.07657947391271591,
-0.11501604318618774,
-0.012491405010223389,
-0.02357003465294838,
0.04245233163237572,
0.020961089059710503,
0.08042895793914795,
-0.10865440964698792,
0.0006480339798144996,
-0.05206969007849693,
-0.05431989207863808,
-0.0542224757373333,
0.03539065271615982,
0.10984950512647629,
0.03923847898840904,
0.14930936694145203,
-0.0960095077753067,
0.040110692381858826,
-0.20455938577651978,
-0.039030771702528,
-0.002898634411394596,
-0.013428899459540844,
-0.07989588379859924,
-0.027097344398498535,
0.09772337228059769,
-0.03873024135828018,
0.11174684762954712,
0.0018089759396389127,
0.10762904584407806,
0.03627956658601761,
-0.07624594122171402,
-0.04170992970466614,
0.029190266504883766,
0.09797083586454391,
0.040215589106082916,
-0.009552567265927792,
0.10417576134204865,
-0.022077545523643494,
0.04646572470664978,
0.015614766627550125,
0.23979543149471283,
0.15933401882648468,
-0.00755910761654377,
0.0462346076965332,
0.07175552099943161,
-0.13560067117214203,
-0.11105229705572128,
0.12179955840110779,
-0.06969410926103592,
0.10856182873249054,
-0.06603722274303436,
0.18498286604881287,
0.046395450830459595,
-0.18691760301589966,
0.053874026983976364,
-0.06102440133690834,
-0.1062379777431488,
-0.12207567691802979,
-0.01695217750966549,
-0.07925564795732498,
-0.13519321382045746,
0.023304646834731102,
-0.12063947319984436,
0.0551987886428833,
0.07211779057979584,
0.004313046112656593,
0.020497463643550873,
0.13877366483211517,
-0.021177461370825768,
0.010826577432453632,
0.07203486561775208,
0.030355608090758324,
-0.002500347327440977,
-0.035907547920942307,
-0.06661171466112137,
0.03906676545739174,
0.04876654967665672,
0.054101768881082535,
-0.041264139115810394,
-0.009918978437781334,
0.033379364758729935,
-0.00900815986096859,
-0.07167993485927582,
0.036515455693006516,
0.004534170962870121,
0.03826911002397537,
0.057140421122312546,
0.06172791123390198,
0.009158456698060036,
-0.028572646901011467,
0.31859728693962097,
-0.08424301445484161,
-0.08150018006563187,
-0.1462700068950653,
0.20078659057617188,
0.0054731289856135845,
-0.0062992628663778305,
0.05631859228014946,
-0.09448057413101196,
-0.032047420740127563,
0.1398829072713852,
0.1398518681526184,
-0.11254679411649704,
-0.011001589708030224,
-0.04470442235469818,
-0.014074515551328659,
-0.031143372878432274,
0.1376822590827942,
0.10369208455085754,
0.028843220323324203,
-0.05891893059015274,
-0.014001254923641682,
0.006459439173340797,
-0.05017540976405144,
-0.061249107122421265,
0.08330980688333511,
-0.01138245314359665,
0.013811657205224037,
-0.031034141778945923,
0.07135666161775589,
0.016050340607762337,
-0.21635234355926514,
0.0077077243477106094,
-0.1641368418931961,
-0.1823236048221588,
-0.03418492153286934,
0.061387162655591965,
-0.011579230427742004,
0.05725553259253502,
-0.00218975730240345,
-0.011490521021187305,
0.1508149951696396,
-0.030486244708299637,
-0.02189464494585991,
-0.11583183705806732,
0.07727406173944473,
-0.12794704735279083,
0.22938703000545502,
-0.0059631140902638435,
0.04521644115447998,
0.09351834654808044,
0.02552981860935688,
-0.13564719259738922,
0.029724039137363434,
0.05630230903625488,
-0.0801960825920105,
0.015851495787501335,
0.16316169500350952,
-0.05194757133722305,
0.08894383907318115,
0.060818884521722794,
-0.13059665262699127,
-0.004009018652141094,
-0.03847980499267578,
-0.04213859140872955,
-0.06670872867107391,
-0.007070416118949652,
-0.05605953186750412,
0.15359196066856384,
0.21035979688167572,
-0.03279976546764374,
0.034711360931396484,
-0.08376278728246689,
0.02787143737077713,
0.03178268298506737,
0.10680664330720901,
-0.025679638609290123,
-0.20939014852046967,
0.033331383019685745,
0.07045701146125793,
0.02695292793214321,
-0.22920948266983032,
-0.08938366919755936,
0.019757041707634926,
-0.04422571137547493,
-0.049983903765678406,
0.13048750162124634,
0.021193796768784523,
0.03990986570715904,
-0.03103664517402649,
-0.1254841387271881,
-0.03410513699054718,
0.1606018990278244,
-0.15445038676261902,
-0.038422226905822754
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bart-noised-with-gcd-dist-0.3
This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 10
- num_epochs: 3
- mixed_precision_training: Native AMP
### Training results
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.2+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "facebook/bart-base", "model-index": [{"name": "bart-noised-with-gcd-dist-0.3", "results": []}]} | text2text-generation | gayanin/bart-noised-with-gcd-dist-0.3 | [
"transformers",
"safetensors",
"bart",
"text2text-generation",
"generated_from_trainer",
"base_model:facebook/bart-base",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T17:29:08+00:00 | [] | [] | TAGS
#transformers #safetensors #bart #text2text-generation #generated_from_trainer #base_model-facebook/bart-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
# bart-noised-with-gcd-dist-0.3
This model is a fine-tuned version of facebook/bart-base on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 10
- num_epochs: 3
- mixed_precision_training: Native AMP
### Training results
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.2+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| [
"# bart-noised-with-gcd-dist-0.3\n\nThis model is a fine-tuned version of facebook/bart-base on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 10\n- num_epochs: 3\n- mixed_precision_training: Native AMP",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.1.2+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #safetensors #bart #text2text-generation #generated_from_trainer #base_model-facebook/bart-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# bart-noised-with-gcd-dist-0.3\n\nThis model is a fine-tuned version of facebook/bart-base on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 10\n- num_epochs: 3\n- mixed_precision_training: Native AMP",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.1.2+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
64,
38,
6,
12,
8,
3,
118,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #bart #text2text-generation #generated_from_trainer #base_model-facebook/bart-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# bart-noised-with-gcd-dist-0.3\n\nThis model is a fine-tuned version of facebook/bart-base on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 10\n- num_epochs: 3\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.1.2+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.09675862640142441,
0.10854211449623108,
-0.0028542024083435535,
0.07414883375167847,
0.13016432523727417,
0.029152831062674522,
0.11474613100290298,
0.12253516912460327,
-0.03799143433570862,
0.06504390388727188,
0.06266549229621887,
0.030316047370433807,
0.04380203038454056,
0.18560463190078735,
-0.05326513573527336,
-0.1935899406671524,
0.023334411904215813,
-0.05580037087202072,
-0.08395068347454071,
0.10476104170084,
0.08957649022340775,
-0.08536185324192047,
0.07481911778450012,
-0.012624140828847885,
-0.16444425284862518,
0.021220428869128227,
-0.00993121974170208,
-0.04197339341044426,
0.10370573401451111,
0.0202606450766325,
0.07720019668340683,
0.04315978288650513,
0.13798020780086517,
-0.20819492638111115,
0.002235861262306571,
0.09361962229013443,
0.01583021692931652,
0.07704430818557739,
0.057094424962997437,
-0.024583997204899788,
0.09156819432973862,
-0.13043612241744995,
0.09972796589136124,
0.03515304997563362,
-0.09745238721370697,
-0.14168092608451843,
-0.09384550154209137,
0.044630810618400574,
0.09970396757125854,
0.08598054945468903,
0.0034320091363042593,
0.07195460796356201,
-0.11638695001602173,
0.06332642585039139,
0.1989641934633255,
-0.2598361372947693,
-0.06455380469560623,
0.022620925679802895,
0.04577210173010826,
0.052100904285907745,
-0.10705450922250748,
-0.005503957625478506,
0.03702050447463989,
0.03256811574101448,
0.11633312702178955,
-0.012916297651827335,
-0.07513494789600372,
-0.01853899471461773,
-0.11363953351974487,
-0.002607582835480571,
0.09049779921770096,
0.05198118835687637,
-0.05309439077973366,
-0.08866176754236221,
-0.059204842895269394,
-0.04436984658241272,
-0.017928825691342354,
-0.027967039495706558,
0.02693556435406208,
-0.03820747509598732,
-0.05292443931102753,
-0.04208865389227867,
-0.05757040157914162,
-0.07568331807851791,
0.00992058590054512,
0.13963380455970764,
0.02138112299144268,
0.01982063427567482,
-0.03258659318089485,
0.10476163774728775,
0.009123189374804497,
-0.12605375051498413,
0.005389892961829901,
-0.012401764281094074,
-0.12169398367404938,
-0.04164564609527588,
-0.04227491840720177,
0.012559845112264156,
0.01544236671179533,
0.15578614175319672,
-0.04824939742684364,
0.0822846069931984,
-0.005383966024965048,
-0.01327025517821312,
-0.03044864721596241,
0.11762377619743347,
-0.03563208505511284,
-0.07545556128025055,
0.008821127936244011,
0.10588228702545166,
0.00632026232779026,
-0.018880441784858704,
-0.06829608976840973,
-0.02243237942457199,
0.0654672384262085,
0.05842142552137375,
-0.04333250969648361,
0.02604658529162407,
-0.0478396937251091,
-0.022893685847520828,
0.05688288435339928,
-0.13231942057609558,
0.049648210406303406,
0.022439977154135704,
-0.07691185176372528,
-0.065320685505867,
0.03160153701901436,
0.003399532288312912,
-0.005471935961395502,
0.1037958562374115,
-0.05785524472594261,
-0.010458334349095821,
-0.074964739382267,
-0.06427820771932602,
0.009517410770058632,
-0.07185331732034683,
-0.018226943910121918,
-0.06122337281703949,
-0.204788476228714,
-0.04496549814939499,
0.047575585544109344,
-0.07611240446567535,
-0.035619478672742844,
-0.0472898930311203,
-0.05741938576102257,
0.03285262733697891,
-0.025636818259954453,
0.14786560833454132,
-0.06296951323747635,
0.06591939181089401,
-0.011780168861150742,
0.044540610164403915,
0.022439809516072273,
0.042457856237888336,
-0.0805603489279747,
0.03136855736374855,
-0.17096371948719025,
0.0951104462146759,
-0.089992456138134,
0.016013942658901215,
-0.10283161699771881,
-0.06617219746112823,
0.011757615953683853,
-0.01396387442946434,
0.06123509258031845,
0.13483603298664093,
-0.2122264951467514,
-0.034771475940942764,
0.15188969671726227,
-0.09233793616294861,
-0.057788390666246414,
0.0723995566368103,
-0.047539833933115005,
0.026717592030763626,
0.05652306601405144,
0.17000128328800201,
0.10177645832300186,
-0.13258375227451324,
0.005381601396948099,
-0.0030331520829349756,
0.04777635261416435,
0.041644200682640076,
0.03195584565401077,
-0.0018607140518724918,
0.012601739726960659,
0.00888180360198021,
-0.06775026768445969,
0.006580941379070282,
-0.0736716017127037,
-0.06710339337587357,
-0.04959560185670853,
-0.08642857521772385,
0.05369923263788223,
0.004457890056073666,
0.010172254405915737,
-0.07315327972173691,
-0.10891671478748322,
0.08251415938138962,
0.11700575798749924,
-0.06027310714125633,
0.008826537057757378,
-0.0649292841553688,
0.010921889916062355,
0.01928000897169113,
-0.02209402620792389,
-0.19179843366146088,
-0.11973364651203156,
0.036633968353271484,
-0.07847210764884949,
0.03370163217186928,
0.010632145218551159,
0.06915245205163956,
0.06598202884197235,
-0.04653973504900932,
-0.022791787981987,
-0.07777372747659683,
0.004836711101233959,
-0.0808321163058281,
-0.22755636274814606,
-0.05048704892396927,
-0.037521831691265106,
0.1833246648311615,
-0.2195167988538742,
0.010666374117136002,
-0.01409501489251852,
0.15294641256332397,
0.03148965910077095,
-0.04887454956769943,
-0.00845527183264494,
0.022597085684537888,
0.020907912403345108,
-0.09552083909511566,
0.027555063366889954,
0.00854615867137909,
-0.08384694904088974,
-0.02229534648358822,
-0.12642766535282135,
0.04174992814660072,
0.06548474729061127,
0.10026068240404129,
-0.10060008615255356,
-0.02281644381582737,
-0.06814419478178024,
-0.039979077875614166,
-0.08742029964923859,
0.012787813320755959,
0.18523009121418,
0.029838910326361656,
0.10344768315553665,
-0.0570872537791729,
-0.0671643316745758,
-0.0037712191697210073,
0.008145812898874283,
0.01600297912955284,
0.07886628061532974,
0.06932344287633896,
-0.10561690479516983,
0.0809381827712059,
0.1150638684630394,
-0.00601049792021513,
0.1199851930141449,
-0.029846113175153732,
-0.07007141411304474,
-0.016399594023823738,
0.0005014784401282668,
-0.026714978739619255,
0.1383371204137802,
-0.0514361709356308,
0.02404966577887535,
0.017322197556495667,
0.024596471339464188,
0.026518160477280617,
-0.16174061596393585,
-0.004228887148201466,
0.004375578835606575,
-0.0698724240064621,
-0.024040035903453827,
-0.03229592368006706,
0.05201730877161026,
0.0958583727478981,
0.02126518450677395,
-0.032809752970933914,
0.021637186408042908,
-0.019706400111317635,
-0.08163289725780487,
0.17483475804328918,
-0.12103618681430817,
-0.16346591711044312,
-0.07526478916406631,
0.046041298657655716,
-0.041210781782865524,
-0.02985408902168274,
0.018610527738928795,
-0.087277352809906,
-0.06422720849514008,
-0.08369436860084534,
0.00324089708738029,
0.03142184391617775,
-0.0047553302720189095,
0.04639928415417671,
0.009126494638621807,
0.09149666875600815,
-0.11295249313116074,
-0.0002273110148962587,
-0.02417815290391445,
-0.0828930214047432,
0.003363646101206541,
0.07253940403461456,
0.06356453895568848,
0.10180729627609253,
-0.004452444147318602,
0.016337810084223747,
-0.02446654438972473,
0.2200998216867447,
-0.07701433449983597,
0.023221513256430626,
0.108871690928936,
-0.0038987041916698217,
0.04674534499645233,
0.15205202996730804,
0.01942434534430504,
-0.11350462585687637,
0.04581144452095032,
0.08309933543205261,
-0.00733147282153368,
-0.22476527094841003,
-0.05097893998026848,
-0.01650857739150524,
-0.056619878858327866,
0.08917805552482605,
0.04158373177051544,
-0.030525051057338715,
0.0195102971047163,
-0.010708929039537907,
0.00044915633043274283,
0.02784026600420475,
0.053090885281562805,
0.06277509778738022,
0.04996713995933533,
0.10382521152496338,
-0.01438590046018362,
0.005214984528720379,
0.07829713821411133,
-0.00987199880182743,
0.2512741982936859,
-0.037428274750709534,
0.04894306883215904,
0.042312994599342346,
0.1393393576145172,
-0.020993752405047417,
0.0294911190867424,
0.02704138308763504,
-0.005976465065032244,
-0.007436051033437252,
-0.05603989586234093,
-0.026549041271209717,
0.019816862419247627,
-0.05716850981116295,
0.01694021001458168,
-0.10317148268222809,
0.04042133688926697,
0.03676486387848854,
0.2841353118419647,
0.04859793186187744,
-0.26280146837234497,
-0.0719703882932663,
0.006482207216322422,
-0.04338540509343147,
-0.06197690963745117,
0.0060487366281449795,
0.13310393691062927,
-0.13156069815158844,
0.07908566296100616,
-0.05725008621811867,
0.08217384666204453,
-0.013753414154052734,
0.018101677298545837,
0.07785937190055847,
0.13202117383480072,
-0.007254498079419136,
0.06507564336061478,
-0.2383323758840561,
0.20926547050476074,
0.015618430450558662,
0.1055588573217392,
-0.056119851768016815,
0.023384997621178627,
0.017782442271709442,
0.0563371405005455,
0.09222207963466644,
0.0006054377881810069,
-0.06781639158725739,
-0.12247447669506073,
-0.10233909636735916,
0.04952805116772652,
0.10932513326406479,
-0.04555606469511986,
0.07247481495141983,
-0.0440860241651535,
-0.007598027121275663,
0.04407869651913643,
-0.06124429032206535,
-0.18461096286773682,
-0.11332520097494125,
0.007157143671065569,
-0.007193204015493393,
-0.02562001720070839,
-0.08481479436159134,
-0.08872722089290619,
0.012856571935117245,
0.1702961027622223,
0.029046615585684776,
-0.03603124991059303,
-0.1477796733379364,
0.05771087855100632,
0.1273375153541565,
-0.05613914877176285,
0.016823938116431236,
0.027599191293120384,
0.12429855763912201,
0.03782108053565025,
-0.11175739020109177,
0.07749331742525101,
-0.08718247711658478,
-0.18826483190059662,
-0.055114977061748505,
0.12399221211671829,
0.09551995247602463,
0.044332198798656464,
0.008448160253465176,
0.01640424132347107,
0.02536500245332718,
-0.0900215208530426,
0.01293156761676073,
0.0738372728228569,
0.04082225635647774,
0.054417598992586136,
-0.06895022094249725,
-0.020747995004057884,
-0.027339844033122063,
-0.010130931623280048,
0.0806209146976471,
0.21655498445034027,
-0.08898334205150604,
0.12833137810230255,
0.08166138827800751,
-0.07222561538219452,
-0.17726373672485352,
0.07311230152845383,
0.11235084384679794,
0.023840416222810745,
0.05408202484250069,
-0.183912456035614,
0.13304105401039124,
0.10981371998786926,
-0.029688460752367973,
0.03879093751311302,
-0.29946017265319824,
-0.13463740050792694,
0.07029127329587936,
0.08728387951850891,
0.03464895486831665,
-0.1089748665690422,
-0.035833537578582764,
-0.05015788972377777,
-0.13262665271759033,
0.14882095158100128,
-0.14283005893230438,
0.09573499113321304,
0.000714586756657809,
0.0819745808839798,
0.01616896688938141,
-0.018644830211997032,
0.12293020635843277,
0.045216791331768036,
0.09539748728275299,
-0.04754019156098366,
0.0568382628262043,
0.020129498094320297,
-0.0650811567902565,
0.022076759487390518,
-0.0661165714263916,
0.06448987126350403,
-0.11021306365728378,
-0.008975468575954437,
-0.09597257524728775,
0.07171758264303207,
-0.06120525673031807,
-0.04578785225749016,
-0.02694564498960972,
0.060403838753700256,
0.06266378611326218,
-0.031847819685935974,
-0.007434582803398371,
-0.0011151296785101295,
0.10631170868873596,
0.09140126407146454,
0.10306377708911896,
-0.0327029675245285,
-0.0448124073445797,
0.00967575702816248,
-0.01159103587269783,
0.043234821408987045,
-0.10435603559017181,
0.045562226325273514,
0.11556024849414825,
0.031296201050281525,
0.1369781196117401,
0.02553226612508297,
-0.05571505054831505,
-0.011735524982213974,
0.04463972896337509,
-0.13747954368591309,
-0.10523132234811783,
0.01974904164671898,
-0.03840475156903267,
-0.10976957529783249,
0.0023217955604195595,
0.13681598007678986,
-0.035635534673929214,
-0.01806076243519783,
-0.019962996244430542,
0.0366675890982151,
-0.02633879892528057,
0.1807379275560379,
0.024494072422385216,
0.061523012816905975,
-0.09953366965055466,
0.13860051333904266,
0.05273555964231491,
-0.06636016815900803,
0.056774821132421494,
0.09381358325481415,
-0.09805835783481598,
-0.010384086519479752,
0.07079650461673737,
0.17663566768169403,
-0.021213311702013016,
-0.03643359988927841,
-0.08203092962503433,
-0.12308315932750702,
0.051281578838825226,
0.16434398293495178,
0.02765830233693123,
-0.003638296155259013,
-0.016016734763979912,
0.026320219039916992,
-0.11229871213436127,
0.07383722066879272,
0.06904397159814835,
0.05028945580124855,
-0.11278171837329865,
0.1054435670375824,
0.008456961251795292,
0.024154655635356903,
-0.02522014081478119,
0.040231894701719284,
-0.10806315392255783,
-0.03527640551328659,
-0.16479209065437317,
-0.00320522696711123,
-0.01782231591641903,
0.00984205212444067,
-0.016984811052680016,
-0.05323063209652901,
-0.03272169083356857,
0.029319344088435173,
-0.07416315376758575,
-0.052429210394620895,
0.0003585806116461754,
0.04038805514574051,
-0.17356593906879425,
-0.01541508175432682,
0.026193883270025253,
-0.0999198704957962,
0.08927561342716217,
0.0763254165649414,
0.026865651831030846,
0.03196045383810997,
-0.15224504470825195,
-0.03403990715742111,
0.01912839524447918,
0.009351375512778759,
0.07616915553808212,
-0.11546822637319565,
-0.01222936250269413,
-0.023020530119538307,
0.041985414922237396,
0.021410277113318443,
0.07986556738615036,
-0.10847198218107224,
0.0005730681587010622,
-0.05214760825037956,
-0.053984545171260834,
-0.05407478287816048,
0.03575441241264343,
0.11038553714752197,
0.03941517323255539,
0.1494007110595703,
-0.0965079590678215,
0.0398261621594429,
-0.20429746806621552,
-0.03891698643565178,
-0.002843592083081603,
-0.013646651990711689,
-0.0801529660820961,
-0.026397882029414177,
0.09778159111738205,
-0.03834071755409241,
0.11227539926767349,
0.0009679474751465023,
0.10761447250843048,
0.03620424121618271,
-0.0757548063993454,
-0.04113542661070824,
0.028922906145453453,
0.09796294569969177,
0.039254289120435715,
-0.009658695198595524,
0.10476455092430115,
-0.021680334582924843,
0.04665331542491913,
0.015460389666259289,
0.2398148626089096,
0.1591513603925705,
-0.007700532674789429,
0.04664795100688934,
0.07108742743730545,
-0.13580267131328583,
-0.11002320051193237,
0.1209167018532753,
-0.06903651356697083,
0.10876473039388657,
-0.06594906002283096,
0.18468201160430908,
0.04674797132611275,
-0.18686218559741974,
0.05358438566327095,
-0.0608564093708992,
-0.10606943815946579,
-0.12206261605024338,
-0.017207983881235123,
-0.07872255146503448,
-0.13532064855098724,
0.02366967871785164,
-0.1208142563700676,
0.05441594496369362,
0.07165517657995224,
0.0053205653093755245,
0.02067245915532112,
0.13895802199840546,
-0.0200955793261528,
0.010989407077431679,
0.07252629846334457,
0.030192818492650986,
-0.002827990800142288,
-0.03577451407909393,
-0.06614331156015396,
0.038556020706892014,
0.04805765300989151,
0.05402492359280586,
-0.041336096823215485,
-0.010534757748246193,
0.03379553556442261,
-0.008938748389482498,
-0.07205387949943542,
0.03622560575604439,
0.004508380778133869,
0.038503918796777725,
0.05878738686442375,
0.06157854199409485,
0.009047424420714378,
-0.028470508754253387,
0.3175913989543915,
-0.08396326005458832,
-0.08099719882011414,
-0.1465068757534027,
0.20023994147777557,
0.005714027211070061,
-0.006957343313843012,
0.05629326403141022,
-0.0940895825624466,
-0.0317683108150959,
0.1405334621667862,
0.14053985476493835,
-0.11198540776968002,
-0.010922570712864399,
-0.04479674622416496,
-0.014123726636171341,
-0.03110578842461109,
0.13765253126621246,
0.1035962849855423,
0.02879791334271431,
-0.05903426930308342,
-0.013759473338723183,
0.006497383117675781,
-0.05044357851147652,
-0.06090114265680313,
0.0827852115035057,
-0.01058497279882431,
0.013936219736933708,
-0.03077567368745804,
0.07128351181745529,
0.015536189079284668,
-0.21501997113227844,
0.007665734738111496,
-0.16346760094165802,
-0.18184401094913483,
-0.034550197422504425,
0.06058429554104805,
-0.011460185050964355,
0.05733371898531914,
-0.0019675714429467916,
-0.011124522425234318,
0.1509256660938263,
-0.030551820993423462,
-0.021745383739471436,
-0.11637650430202484,
0.07751626521348953,
-0.12640242278575897,
0.2287503331899643,
-0.005939659662544727,
0.04499482735991478,
0.09369949251413345,
0.025243505835533142,
-0.13607019186019897,
0.03024272993206978,
0.05630442127585411,
-0.08031783252954483,
0.01610121876001358,
0.16279706358909607,
-0.051899153739213943,
0.08962889760732651,
0.060026634484529495,
-0.12986627221107483,
-0.003485532943159342,
-0.038482416421175,
-0.04160880297422409,
-0.06660276651382446,
-0.005937970709055662,
-0.055796410888433456,
0.15414445102214813,
0.2103872448205948,
-0.03275122866034508,
0.03392726182937622,
-0.08384314924478531,
0.028022507205605507,
0.03204922378063202,
0.1057952269911766,
-0.02572346292436123,
-0.20963525772094727,
0.0335298627614975,
0.0703291967511177,
0.027089035138487816,
-0.22970423102378845,
-0.08970511704683304,
0.01976374350488186,
-0.04440520331263542,
-0.05112471804022789,
0.13042771816253662,
0.020764537155628204,
0.04047122597694397,
-0.03088553622364998,
-0.12347898632287979,
-0.03438107669353485,
0.16036267578601837,
-0.15494000911712646,
-0.03821001574397087
] |
null | null | fastai |
# Amazing!
🥳 Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the [documentation here](https://huggingface.co/docs/hub/model-repos))!
2. Create a demo in Gradio or Streamlit using 🤗 Spaces ([documentation here](https://huggingface.co/docs/hub/spaces)).
3. Join the fastai community on the [Fastai Discord](https://discord.com/invite/YKrxeNn)!
Greetings fellow fastlearner 🤝! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| {"tags": ["fastai"]} | null | luisvarona/intel-image-classification | [
"fastai",
"region:us"
] | 2024-02-07T17:29:27+00:00 | [] | [] | TAGS
#fastai #region-us
|
# Amazing!
Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the documentation here)!
2. Create a demo in Gradio or Streamlit using Spaces (documentation here).
3. Join the fastai community on the Fastai Discord!
Greetings fellow fastlearner ! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| [
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
"TAGS\n#fastai #region-us \n",
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
9,
20,
79,
3,
6,
12,
8
] | [
"passage: TAGS\n#fastai #region-us \n# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---# Model card## Model description\nMore information needed## Intended uses & limitations\nMore information needed## Training and evaluation data\nMore information needed"
] | [
-0.073318250477314,
-0.035918332636356354,
0.0016039619222283363,
0.09830865263938904,
0.16935402154922485,
0.11954792588949203,
0.06504721194505692,
0.08469552546739578,
0.09305626899003983,
0.008462822064757347,
0.08902737498283386,
-0.059808652848005295,
0.09601042419672012,
0.26935747265815735,
0.06010362133383751,
-0.24278773367404938,
0.02870224229991436,
-0.0036573195829987526,
0.08660013228654861,
0.06588653475046158,
0.12898924946784973,
-0.039593055844306946,
0.14736801385879517,
-0.018255524337291718,
-0.19320440292358398,
-0.054476846009492874,
-0.015185145661234856,
-0.019686169922351837,
0.12385433167219162,
-0.04793357476592064,
0.030790239572525024,
0.0026993011124432087,
-0.0015684126410633326,
-0.0995422899723053,
0.06401026993989944,
0.04089692234992981,
0.028817683458328247,
0.055760785937309265,
-0.04539911448955536,
0.08392030745744705,
0.054179996252059937,
-0.010920286178588867,
-0.12179892510175705,
0.09588204324245453,
-0.1474396139383316,
-0.2022949457168579,
-0.1278105229139328,
-0.11345728486776352,
0.047258179634809494,
0.01006549596786499,
-0.01907140202820301,
0.12847048044204712,
-0.14997079968452454,
-0.03727749362587929,
0.17807333171367645,
-0.15483331680297852,
-0.050517335534095764,
-0.0010879677720367908,
0.06801971048116684,
-0.06002732738852501,
-0.05137069150805473,
0.0968702957034111,
0.0906822457909584,
-0.019289257004857063,
0.015487968921661377,
0.0037353564985096455,
0.035032227635383606,
0.002429646672680974,
-0.0558350533246994,
0.06529499590396881,
-0.027788599953055382,
0.055927276611328125,
-0.1094130128622055,
-0.11809343844652176,
0.0010178228840231895,
0.03238791227340698,
-0.05549647659063339,
-0.06733305007219315,
0.0810781940817833,
0.007735111750662327,
-0.0603058859705925,
-0.11863275617361069,
-0.06696899980306625,
-0.12959590554237366,
0.00783742405474186,
0.09659197926521301,
0.0033950558863580227,
0.06878509372472763,
-0.09986882656812668,
0.06626693904399872,
-0.2048133760690689,
-0.04758621007204056,
-0.08781389147043228,
-0.1065201610326767,
0.02003002166748047,
-0.04773771017789841,
0.04778444394469261,
0.15393073856830597,
0.14042632281780243,
0.04171324521303177,
0.05645250529050827,
-0.029350629076361656,
0.038715146481990814,
0.04752078279852867,
0.018331103026866913,
0.03540196642279625,
-0.020549163222312927,
-0.18507646024227142,
0.0004176131042186171,
-0.04207618162035942,
0.08488372713327408,
-0.07463551312685013,
-0.05029602348804474,
0.01336510106921196,
-0.12160550057888031,
0.09655242413282394,
-0.05178983509540558,
-0.005084214266389608,
0.0036863412242382765,
0.008919943124055862,
0.20647431910037994,
0.04232640564441681,
0.004936119541525841,
-0.006976569537073374,
-0.1375076025724411,
-0.051532845944166183,
-0.09289269894361496,
0.034273598343133926,
0.02420172467827797,
0.01303885504603386,
-0.07711919397115707,
0.049177106469869614,
-0.046599894762039185,
-0.008231878280639648,
0.021442487835884094,
-0.20236440002918243,
0.010869519785046577,
-0.0969783291220665,
-0.1469350904226303,
0.06343341618776321,
0.0026821133214980364,
-0.07499043643474579,
0.08385025709867477,
-0.004780351184308529,
0.031972795724868774,
-0.030242523178458214,
-0.00177793821785599,
0.05239185318350792,
-0.08095952123403549,
0.023147141560912132,
0.1995297074317932,
0.10590710490942001,
-0.07641816139221191,
-0.0025978393387049437,
-0.12475098669528961,
0.04128078371286392,
-0.14157716929912567,
0.038516461849212646,
-0.08163458108901978,
0.15109841525554657,
-0.044047996401786804,
0.018007883802056313,
-0.0071970620192587376,
0.08468028157949448,
0.07606321573257446,
0.19981153309345245,
-0.23198086023330688,
-0.053279466927051544,
0.16512827575206757,
-0.11487894505262375,
-0.18565405905246735,
0.20080815255641937,
-0.00043150142300873995,
0.10752102732658386,
-0.010421866551041603,
0.17009462416172028,
-0.021746216341853142,
-0.14181379973888397,
-0.032203078269958496,
-0.0012119774473831058,
-0.24691128730773926,
-0.08980891108512878,
0.09945957362651825,
0.10481112450361252,
-0.059047527611255646,
0.029137471690773964,
0.012005627155303955,
0.15818172693252563,
-0.07679074257612228,
-0.04601999372243881,
-0.007829579524695873,
-0.10506698489189148,
0.022122014313936234,
0.01663162000477314,
0.034775324165821075,
-0.059334270656108856,
-0.00890427641570568,
-0.07678428292274475,
0.13092219829559326,
0.09849999099969864,
-0.03540538251399994,
-0.06064159423112869,
0.16454961895942688,
-0.0640924945473671,
-0.026323838159441948,
0.08331746608018875,
-0.08536569774150848,
0.047215063124895096,
0.04028964787721634,
0.05084947869181633,
0.009997997432947159,
0.09182237833738327,
0.0698544830083847,
0.006789602339267731,
0.03368524834513664,
0.13270887732505798,
-0.027426021173596382,
-0.05121328681707382,
0.01674247533082962,
0.04598715528845787,
-0.00979064591228962,
0.3169313669204712,
-0.19912512600421906,
0.018945744261145592,
-0.06457886099815369,
0.08035559207201004,
0.0660853385925293,
0.007019065320491791,
0.07570107281208038,
-0.05360652506351471,
-0.016966497525572777,
-0.045681122690439224,
0.06926878541707993,
-0.06979862600564957,
-0.054223138839006424,
0.2564660608768463,
-0.031106717884540558,
0.031359151005744934,
0.10653062164783478,
-0.06802138686180115,
-0.05823708325624466,
-0.02224794402718544,
-0.0014688228257000446,
0.023401014506816864,
-0.04168177396059036,
0.06067536398768425,
-0.08815024048089981,
-0.05285300314426422,
0.1703105866909027,
-0.038786694407463074,
0.07842917740345001,
0.035427022725343704,
-0.05379872769117355,
-0.04481838271021843,
0.061976201832294464,
0.14977918565273285,
-0.0965908095240593,
0.06779327243566513,
0.13305115699768066,
0.014980388805270195,
0.15411095321178436,
0.07098863273859024,
-0.07586279511451721,
-0.08855607360601425,
-0.018246978521347046,
-0.004062598571181297,
0.18133139610290527,
-0.07897800207138062,
-0.036732085049152374,
0.042683616280555725,
-0.011134039610624313,
0.06611642241477966,
-0.05846851319074631,
-0.0792742595076561,
0.01736506260931492,
-0.0582035630941391,
0.018060972914099693,
0.12486616522073746,
-0.08240851759910583,
0.04267239198088646,
0.03745635226368904,
-0.058472223579883575,
0.046025440096855164,
0.0389089435338974,
-0.01086228247731924,
0.05541912093758583,
0.06821268051862717,
-0.2134213149547577,
-0.10377796739339828,
-0.17595313489437103,
0.03000609390437603,
0.020109420642256737,
0.036413755267858505,
-0.10920769721269608,
0.02131613902747631,
-0.0651998370885849,
-0.07437032461166382,
0.04871295765042305,
-0.029500357806682587,
-0.10847225040197372,
-0.027001040056347847,
-0.024241603910923004,
-0.04816099628806114,
-0.021433888003230095,
-0.06250716745853424,
0.03129231557250023,
0.04526861384510994,
0.03191622346639633,
0.1321185976266861,
-0.010805734433233738,
-0.014524625614285469,
0.002761868294328451,
-0.017431288957595825,
0.1497519314289093,
-0.13988617062568665,
0.06941607594490051,
0.1812426596879959,
0.09771130234003067,
0.03844839334487915,
0.01466822624206543,
0.03106272965669632,
-0.07663184404373169,
0.005383877083659172,
0.034619297832250595,
-0.0891294777393341,
-0.08207139372825623,
-0.01874193549156189,
-0.03897557035088539,
0.21049608290195465,
-0.12441039085388184,
0.024025630205869675,
0.040357187390327454,
0.09686839580535889,
0.11187659204006195,
-0.04121972620487213,
-0.17262403666973114,
0.04177050292491913,
-0.2474004179239273,
-0.051238708198070526,
0.003026821883395314,
-0.09497712552547455,
-0.06320231407880783,
0.18337351083755493,
0.0052159554325044155,
0.0287664532661438,
0.00430127140134573,
0.12202860414981842,
-0.0009366215672343969,
0.12068869173526764,
0.0687243714928627,
-0.05316835641860962,
0.02255408652126789,
-0.09993521869182587,
-0.0696573555469513,
-0.03704388439655304,
-0.07047778367996216,
0.06136435270309448,
0.12800902128219604,
-0.024759603664278984,
-0.04259653389453888,
0.04763835668563843,
0.09553752839565277,
0.06145815551280975,
0.15860231220722198,
-0.16057826578617096,
-0.022865094244480133,
0.042546581476926804,
-0.029262376949191093,
-0.049140751361846924,
-0.009500340558588505,
0.08492209017276764,
-0.05378608778119087,
-0.02665375918149948,
0.003306680591776967,
0.07226359844207764,
-0.0019794153049588203,
0.0436936691403389,
-0.03244423121213913,
0.1845880150794983,
-0.029572106897830963,
0.023350762203335762,
-0.12604808807373047,
0.13696090877056122,
0.022422920912504196,
-0.015438690781593323,
-0.06568175554275513,
-0.05596291273832321,
0.18064838647842407,
0.02166406810283661,
0.11738308519124985,
0.011424299329519272,
-0.09442766010761261,
-0.1337079405784607,
-0.1388736516237259,
0.015837913379073143,
0.09729303419589996,
-0.01256689801812172,
-0.03353166952729225,
0.019608711823821068,
-0.04281611740589142,
-0.06777504086494446,
0.10452067106962204,
-0.11668688803911209,
-0.0018522912869229913,
0.005423946306109428,
0.0416572242975235,
-0.06085909157991409,
0.032720211893320084,
0.03296784311532974,
-0.0647648349404335,
0.121244877576828,
0.24137550592422485,
0.1064029112458229,
-0.09990023821592331,
-0.08652417361736298,
0.021780110895633698,
-0.034567005932331085,
-0.0014182132435962558,
-0.016133872792124748,
0.036385562270879745,
0.0019662054255604744,
0.003586959559470415,
0.13572031259536743,
-0.07582411170005798,
0.012567305937409401,
-0.08275366574525833,
0.07902812212705612,
-0.0409930944442749,
-0.0025117802433669567,
-0.003995150327682495,
-0.02950184792280197,
-0.03430648893117905,
-0.06180789694190025,
0.163230761885643,
-0.06168964132666588,
-0.08240502327680588,
0.07821446657180786,
0.01680770143866539,
0.017550375312566757,
-0.06227098032832146,
-0.054205916821956635,
0.1972212791442871,
0.31792324781417847,
-0.058273475617170334,
0.10361375659704208,
0.1383560746908188,
0.023166829720139503,
-0.22579050064086914,
0.036502011120319366,
-0.14466507732868195,
0.032058101147413254,
0.024782279506325722,
-0.06415819376707077,
0.05856261029839516,
0.1250556856393814,
-0.045668914914131165,
0.23617008328437805,
-0.03641456738114357,
-0.07633192092180252,
-0.013243574649095535,
0.043972890824079514,
0.3091393709182739,
-0.11325396597385406,
-0.02349173277616501,
-0.11636991053819656,
-0.21521669626235962,
0.06708590686321259,
-0.16208602488040924,
0.1406344771385193,
-0.05703224614262581,
0.023474344983696938,
-0.012111215852200985,
-0.07578689604997635,
0.19497497379779816,
-0.1371963620185852,
0.056931521743535995,
-0.1432308852672577,
-0.11647364497184753,
-0.005183211527764797,
-0.08439649641513824,
0.14731425046920776,
-0.08327576518058777,
-0.02632858417928219,
-0.2082071304321289,
0.001373599166981876,
-0.021641740575432777,
0.09738951921463013,
0.02311836928129196,
-0.07967846095561981,
-0.08035353571176529,
0.12579506635665894,
-0.07811200618743896,
0.036513522267341614,
-0.08704032748937607,
-0.03989429399371147,
-0.026884159073233604,
-0.08092786371707916,
0.06243825703859329,
-0.08906654268503189,
0.16072829067707062,
-0.049172405153512955,
-0.046159181743860245,
0.061650797724723816,
-0.20832203328609467,
0.026940656825900078,
0.036382775753736496,
-0.031731411814689636,
0.10237374156713486,
-0.029687397181987762,
-0.07129550725221634,
0.1133488118648529,
0.13133300840854645,
-0.07154961675405502,
-0.2563934028148651,
-0.0821671262383461,
-0.008923565037548542,
0.04608851298689842,
0.0829237625002861,
0.04836045205593109,
-0.05231332778930664,
-0.017525162547826767,
-0.031239798292517662,
0.03463910520076752,
-0.11768791079521179,
-0.02900020219385624,
0.06892099231481552,
0.0014350401470437646,
-0.09527117758989334,
0.0962897539138794,
-0.004287306685000658,
-0.02237984538078308,
-0.009249147027730942,
0.1892271637916565,
-0.014808090403676033,
-0.12871821224689484,
-0.057921428233385086,
0.24053727090358734,
-0.038428641855716705,
-0.07654319703578949,
-0.06858045607805252,
-0.011265470646321774,
-0.04038287326693535,
0.06209278851747513,
0.04795577749609947,
-0.01209679339081049,
0.08278531581163406,
0.06026776134967804,
-0.1221788227558136,
-0.060724351555109024,
-0.05533421039581299,
0.035240933299064636,
-0.09762322902679443,
0.04652146250009537,
0.016370195895433426,
0.12453475594520569,
-0.09184806793928146,
-0.03038635104894638,
-0.11205437779426575,
-0.059142544865608215,
-0.18314886093139648,
-0.0571221299469471,
-0.041237685829401016,
-0.008055833168327808,
0.03931373730301857,
0.02697678469121456,
-0.04493580758571625,
-0.048296377062797546,
-0.06704439222812653,
0.03899036720395088,
0.07422684133052826,
0.026717372238636017,
-0.03390409052371979,
0.05009619519114494,
0.06439550966024399,
0.008286280557513237,
0.1963774412870407,
0.06738202273845673,
0.061680130660533905,
-0.025940580293536186,
-0.19781054556369781,
-0.05686524137854576,
0.002742079785093665,
-0.09212438762187958,
0.12195391207933426,
-0.011633808724582195,
0.02040605992078781,
-0.06281229853630066,
0.03727225586771965,
0.026594331488013268,
0.10702691227197647,
-0.02029390074312687,
0.0958021730184555,
0.029817266389727592,
-0.08947111666202545,
-0.044351425021886826,
0.015944788232445717,
0.12201714515686035,
0.02899266965687275,
0.028689615428447723,
0.015606578439474106,
0.037100955843925476,
-0.03902486339211464,
0.0296308696269989,
-0.045808494091033936,
-0.14955224096775055,
0.01991276629269123,
-0.046732377260923386,
-0.006942411884665489,
-0.016697930172085762,
0.18722283840179443,
0.04047711566090584,
-0.046649303287267685,
-0.01265130564570427,
0.014551439322531223,
-0.004945865832269192,
-0.03270510211586952,
-0.004582806024700403,
0.06002182513475418,
-0.004176365211606026,
-0.047248490154743195,
0.13213102519512177,
0.046804413199424744,
0.04763852432370186,
0.0742364451289177,
0.09783162921667099,
-0.00930761732161045,
0.13372060656547546,
0.06815905123949051,
-0.01982966810464859,
-0.1131899505853653,
-0.05649255961179733,
-0.11679257452487946,
0.034573014825582504,
-0.05576380714774132,
0.12528598308563232,
0.11196581274271011,
-0.060735806822776794,
-0.03883470967411995,
-0.0771038830280304,
-0.03134944289922714,
-0.07594948261976242,
0.03614310547709465,
-0.0327751524746418,
-0.08104247599840164,
0.06421366333961487,
0.05536265671253204,
-0.036099426448345184,
0.11491319537162781,
0.020650042220950127,
-0.05702126771211624,
0.12617406249046326,
-0.07743373513221741,
0.10717736184597015,
0.07707828283309937,
-0.05362870916724205,
-0.12441752851009369,
0.011045942083001137,
-0.07996662706136703,
-0.11546584963798523,
-0.008837178349494934,
-0.011918267235159874,
-0.0746825784444809,
-0.05780024081468582,
0.10738345980644226,
-0.03462931141257286,
-0.09724929928779602,
-0.020749187096953392,
0.015756776556372643,
0.056543223559856415,
-0.019683608785271645,
0.0018315898487344384,
0.03772254288196564,
0.028699718415737152,
0.15574465692043304,
-0.0016714793164283037,
0.06267286092042923,
-0.1358945369720459,
0.18023191392421722,
-0.1432318240404129,
-0.027932528406381607,
-0.187766894698143,
-0.0886974111199379,
-0.025430310517549515,
0.22427266836166382,
0.26061514019966125,
-0.1923753172159195,
-0.03171071037650108,
0.004376344382762909,
-0.010204915888607502,
-0.07923580706119537,
0.14464490115642548,
0.02417137287557125,
-0.007147552911192179,
-0.06552806496620178,
-0.014752711169421673,
0.024085145443677902,
-0.07228498160839081,
-0.035760894417762756,
0.18496830761432648,
0.0086367791518569,
0.07214809954166412,
-0.09064984321594238,
0.03641578182578087,
-0.18433186411857605,
-0.0693570077419281,
-0.03508331999182701,
-0.138646200299263,
-0.09639570862054825,
-0.01481159869581461,
0.003136083483695984,
0.09603974968194962,
0.03350212052464485,
-0.01305394247174263,
0.06808507442474365,
-0.049502357840538025,
0.010726232081651688,
-0.16043636202812195,
-0.020468583330512047,
0.05376148223876953,
-0.052667658776044846,
0.23897892236709595,
-0.02351270616054535,
-0.12297288328409195,
0.08416848629713058,
-0.03519788756966591,
-0.12302011996507645,
0.0745280459523201,
-0.023310834541916847,
-0.10405170172452927,
-0.05555706471204758,
0.17993386089801788,
-0.01256539486348629,
-0.16247478127479553,
0.03247550129890442,
-0.15925332903862,
0.029797034338116646,
0.03576231747865677,
-0.011352102272212505,
-0.05518606677651405,
0.028951244428753853,
-0.027475930750370026,
0.10062393546104431,
0.14163273572921753,
0.017354421317577362,
-0.009662404656410217,
-0.06593839079141617,
0.09352979063987732,
0.06211914122104645,
-0.07753235101699829,
-0.11338558793067932,
-0.09994973242282867,
0.02616780437529087,
0.07790441066026688,
-0.08538854867219925,
-0.17278192937374115,
-0.029272083193063736,
-0.11865141987800598,
-0.002084053121507168,
0.0349934957921505,
0.06834512948989868,
0.2863384187221527,
0.06974043697118759,
0.004092831164598465,
-0.15255671739578247,
0.05762675032019615,
0.08219972252845764,
-0.02544020675122738,
-0.08790270239114761
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# results
This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: constant
- lr_scheduler_warmup_ratio: 0.3
- num_epochs: 2
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.38.0.dev0
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1 | {"license": "apache-2.0", "library_name": "peft", "tags": ["trl", "sft", "generated_from_trainer"], "base_model": "mistralai/Mistral-7B-v0.1", "model-index": [{"name": "results", "results": []}]} | null | prasadkanche/prasad_Mistral_FT_Model | [
"peft",
"tensorboard",
"safetensors",
"trl",
"sft",
"generated_from_trainer",
"base_model:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"region:us"
] | 2024-02-07T17:29:57+00:00 | [] | [] | TAGS
#peft #tensorboard #safetensors #trl #sft #generated_from_trainer #base_model-mistralai/Mistral-7B-v0.1 #license-apache-2.0 #region-us
|
# results
This model is a fine-tuned version of mistralai/Mistral-7B-v0.1 on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: constant
- lr_scheduler_warmup_ratio: 0.3
- num_epochs: 2
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.38.0.dev0
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1 | [
"# results\n\nThis model is a fine-tuned version of mistralai/Mistral-7B-v0.1 on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: constant\n- lr_scheduler_warmup_ratio: 0.3\n- num_epochs: 2",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#peft #tensorboard #safetensors #trl #sft #generated_from_trainer #base_model-mistralai/Mistral-7B-v0.1 #license-apache-2.0 #region-us \n",
"# results\n\nThis model is a fine-tuned version of mistralai/Mistral-7B-v0.1 on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: constant\n- lr_scheduler_warmup_ratio: 0.3\n- num_epochs: 2",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
55,
31,
6,
12,
8,
3,
127,
4,
44
] | [
"passage: TAGS\n#peft #tensorboard #safetensors #trl #sft #generated_from_trainer #base_model-mistralai/Mistral-7B-v0.1 #license-apache-2.0 #region-us \n# results\n\nThis model is a fine-tuned version of mistralai/Mistral-7B-v0.1 on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: constant\n- lr_scheduler_warmup_ratio: 0.3\n- num_epochs: 2### Training results### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.12576964497566223,
0.15023407340049744,
-0.003822293598204851,
0.07491406798362732,
0.13655750453472137,
0.028385527431964874,
0.082456074655056,
0.14355537295341492,
-0.03486507385969162,
0.11351215839385986,
0.10126354545354843,
0.026099152863025665,
0.06197613105177879,
0.1402137726545334,
-0.021329941228032112,
-0.20664197206497192,
0.018702244386076927,
-0.053165387362241745,
-0.03376200422644615,
0.08936028927564621,
0.09743154793977737,
-0.09529024362564087,
0.06058172509074211,
-0.002106641884893179,
-0.10591783374547958,
-0.011393184773623943,
-0.036996714770793915,
-0.0535283088684082,
0.08982942998409271,
-0.009859707206487656,
0.08581417798995972,
0.007852868176996708,
0.12449930608272552,
-0.21047499775886536,
0.0003712076286319643,
0.08301493525505066,
0.040730636566877365,
0.0896201953291893,
0.08530808240175247,
-0.011890435591340065,
0.0908140242099762,
-0.1398114562034607,
0.07309913635253906,
0.029951002448797226,
-0.08230707049369812,
-0.1631743609905243,
-0.10514160990715027,
0.08476162701845169,
0.08129172772169113,
0.08592437207698822,
0.01664014346897602,
0.11594019830226898,
-0.07316241413354874,
0.05997634679079056,
0.22115817666053772,
-0.2624547779560089,
-0.044493380934000015,
0.04854380339384079,
0.010709370486438274,
0.07480574399232864,
-0.10889896750450134,
-0.03208950161933899,
0.044049184769392014,
0.034071169793605804,
0.09042777121067047,
0.01968267187476158,
-0.031662434339523315,
0.011050385423004627,
-0.11723446846008301,
-0.045352473855018616,
0.13941501080989838,
0.02876155637204647,
-0.051583126187324524,
-0.11023309826850891,
-0.0426299087703228,
-0.1381848305463791,
-0.010701400227844715,
0.004855101462453604,
0.021120652556419373,
-0.033982690423727036,
-0.005938347894698381,
-0.028259361162781715,
-0.07163555920124054,
-0.060056112706661224,
0.012608434073626995,
0.046141669154167175,
0.05168420821428299,
0.03484224155545235,
-0.009410802274942398,
0.12252853810787201,
-0.01910991594195366,
-0.1301434487104416,
-0.023763209581375122,
-0.013173431158065796,
-0.08943697065114975,
-0.0383102148771286,
-0.030279673635959625,
-0.02291298843920231,
0.017020542174577713,
0.1330198049545288,
-0.0668986439704895,
0.0787559449672699,
0.04439067095518112,
0.00912918895483017,
-0.022991051897406578,
0.14622129499912262,
-0.05745744705200195,
-0.046643853187561035,
0.010610527358949184,
0.1191008910536766,
0.02624853514134884,
-0.012123891152441502,
-0.06543975323438644,
-0.0260340366512537,
0.09801598638296127,
0.06672504544258118,
-0.01820547878742218,
0.007646654266864061,
-0.05585334077477455,
-0.019253285601735115,
0.0640503391623497,
-0.12501290440559387,
0.033599816262722015,
0.024730997160077095,
-0.08851343393325806,
-0.01092314813286066,
0.03538455069065094,
-0.01638181321322918,
-0.05258304998278618,
0.09891360998153687,
-0.07801827788352966,
-0.015102283097803593,
-0.06800024211406708,
-0.04536944255232811,
0.0219425056129694,
-0.07325871288776398,
-0.0301931444555521,
-0.05822322145104408,
-0.20023567974567413,
-0.04145517945289612,
0.03422153741121292,
-0.07877494394779205,
-0.04411403462290764,
-0.026366474106907845,
-0.0879172757267952,
0.03550981357693672,
-0.004727946128696203,
0.09826085716485977,
-0.05765074864029884,
0.08410827815532684,
-0.02492346055805683,
0.0214762631803751,
0.05961369723081589,
0.038494840264320374,
-0.07151107490062714,
0.0502273365855217,
-0.11440084129571915,
0.07215309143066406,
-0.09540370851755142,
0.02189471572637558,
-0.1386670023202896,
-0.0908699706196785,
0.017904557287693024,
-0.04412199556827545,
0.08893248438835144,
0.14533092081546783,
-0.15865792334079742,
-0.004172493238002062,
0.14397139847278595,
-0.06233413144946098,
-0.07407954335212708,
0.10794860869646072,
-0.021065298467874527,
-0.04016369208693504,
0.04701083526015282,
0.13771602511405945,
0.11001667380332947,
-0.14167775213718414,
-0.011537264101207256,
0.02219555526971817,
0.07665697485208511,
0.000663955754134804,
0.06942849606275558,
-0.031626131385564804,
0.026780234649777412,
0.024282850325107574,
-0.06333112716674805,
0.007621925789862871,
-0.07930583506822586,
-0.07897327840328217,
-0.06913506239652634,
-0.09013412892818451,
0.030626045539975166,
0.01585783250629902,
0.012223096564412117,
-0.06492065638303757,
-0.11964312940835953,
0.04961840808391571,
0.1512640416622162,
-0.041798848658800125,
-0.0012188919354230165,
-0.061447713524103165,
0.09896020591259003,
-0.02986275590956211,
-0.023399924859404564,
-0.18591296672821045,
-0.11021637916564941,
0.04160512983798981,
-0.07484020292758942,
-0.011534084565937519,
0.009265500120818615,
0.0629759356379509,
0.0824190005660057,
-0.035228777676820755,
-0.041028451174497604,
-0.06843163073062897,
-0.008901440538465977,
-0.09267853200435638,
-0.16992345452308655,
-0.03997044265270233,
-0.034254804253578186,
0.16539454460144043,
-0.22356468439102173,
0.005591814871877432,
-0.015933290123939514,
0.1340695321559906,
0.030369102954864502,
-0.07408855855464935,
0.01345185935497284,
0.026515113189816475,
-0.0035992302000522614,
-0.10396789759397507,
0.02889036014676094,
-0.003337651025503874,
-0.08245081454515457,
-0.021489091217517853,
-0.12507547438144684,
0.04304127395153046,
0.06415843218564987,
0.14309659600257874,
-0.09956645965576172,
-0.04314263164997101,
-0.06279490143060684,
-0.04465928673744202,
-0.07473862171173096,
0.000465939607238397,
0.1535169780254364,
0.03862364590167999,
0.12164445966482162,
-0.061644311994314194,
-0.059794533997774124,
0.01088566705584526,
0.004006231669336557,
-0.013107724487781525,
0.10527490824460983,
0.04448806121945381,
-0.08805981278419495,
0.0885993167757988,
0.1010940819978714,
-0.0515303798019886,
0.09658385068178177,
-0.06852367520332336,
-0.10349363833665848,
-0.026600593701004982,
0.06086761876940727,
0.0016753876116126776,
0.12447734922170639,
-0.0246515404433012,
0.02130025438964367,
0.04497338458895683,
0.024113083258271217,
0.010685133747756481,
-0.16323800384998322,
-0.008733890019357204,
0.02520477958023548,
-0.048794928938150406,
-0.02568104863166809,
-0.004015076439827681,
0.028480254113674164,
0.08356335014104843,
0.019231200218200684,
-0.04327480494976044,
-0.0015179991023615003,
-0.01913074217736721,
-0.07201827317476273,
0.1746252477169037,
-0.12562568485736847,
-0.09707243740558624,
-0.1238183006644249,
0.07908917218446732,
-0.04160263016819954,
-0.03092551790177822,
0.010915706865489483,
-0.058739516884088516,
-0.06197626143693924,
-0.0999426543712616,
-0.04025585576891899,
-0.019050924107432365,
-0.009879578836262226,
0.038813505321741104,
0.0045794411562383175,
0.08650091290473938,
-0.12316595762968063,
0.001683423644863069,
-0.006199444178491831,
-0.06925441324710846,
0.008773748762905598,
0.06610587239265442,
0.0785493329167366,
0.11416315287351608,
0.0063051339238882065,
0.0029661478474736214,
-0.015534849837422371,
0.23046450316905975,
-0.05948779731988907,
0.004844718612730503,
0.12003899365663528,
0.022610343992710114,
0.07703487575054169,
0.1043298989534378,
0.028208674862980843,
-0.07329975813627243,
0.009912573732435703,
0.0541544109582901,
-0.016296876594424248,
-0.25475427508354187,
-0.05122263729572296,
-0.029255129396915436,
-0.04100474342703819,
0.09140654653310776,
0.062055736780166626,
-0.01730780489742756,
0.03713512048125267,
-0.025559715926647186,
-0.00030601993785239756,
-0.01149807870388031,
0.07431629300117493,
0.038937825709581375,
0.04282887279987335,
0.08007154613733292,
-0.033929355442523956,
0.012550556100904942,
0.06384069472551346,
0.02109440416097641,
0.2646034359931946,
-0.06290165334939957,
0.14222736656665802,
0.00996978860348463,
0.17196623980998993,
-0.041902199387550354,
0.03594660758972168,
-0.0004286204348318279,
0.007793030701577663,
0.004752740263938904,
-0.06888943910598755,
-0.03632652387022972,
0.027975166216492653,
0.008359890431165695,
0.05489018186926842,
-0.0836881622672081,
0.02956807054579258,
0.030582638457417488,
0.28463006019592285,
0.05948001891374588,
-0.2846415936946869,
-0.07755619287490845,
0.021412836387753487,
-0.007686345838010311,
-0.06667435169219971,
0.010504577308893204,
0.12873438000679016,
-0.1309429407119751,
0.04100466147065163,
-0.05555092170834541,
0.08699079602956772,
-0.048222724348306656,
-0.004075190052390099,
0.043861255049705505,
0.13638196885585785,
0.001580775249749422,
0.07954347878694534,
-0.16447149217128754,
0.19275999069213867,
0.021040678024291992,
0.09494705498218536,
-0.06507217884063721,
0.03664221242070198,
0.010847825556993484,
0.06765833497047424,
0.10540106147527695,
0.0105094313621521,
-0.038061995059251785,
-0.1670466810464859,
-0.12973549962043762,
0.005269316490739584,
0.10097882151603699,
-0.02385767363011837,
0.0666939988732338,
-0.04842197149991989,
-0.00229127355851233,
0.023978479206562042,
-0.07652583718299866,
-0.1216612234711647,
-0.12088125199079514,
0.03372059762477875,
0.007851590402424335,
-0.05208384618163109,
-0.08309289813041687,
-0.08854087442159653,
-0.039301469922065735,
0.14496254920959473,
-0.05098153278231621,
-0.06939622014760971,
-0.1459154486656189,
0.025262251496315002,
0.15415124595165253,
-0.06828991323709488,
0.01991243287920952,
-0.00319591979496181,
0.10601291060447693,
0.03384857624769211,
-0.07026402652263641,
0.04903724044561386,
-0.06960037350654602,
-0.18588879704475403,
-0.061879854649305344,
0.12081979215145111,
0.046947773545980453,
0.0436219684779644,
-0.02591400034725666,
0.026259884238243103,
-0.003568815067410469,
-0.09752136468887329,
0.022693106904625893,
0.12286468595266342,
0.07962530851364136,
0.0615602470934391,
-0.059433359652757645,
0.034064169973134995,
-0.025354579091072083,
-0.018368149176239967,
0.09557130187749863,
0.2535628378391266,
-0.09067733585834503,
0.11932878941297531,
0.06466995179653168,
-0.05235711857676506,
-0.18501389026641846,
0.015392079949378967,
0.11942482739686966,
0.02875692956149578,
0.05512334033846855,
-0.1605830043554306,
0.07688305526971817,
0.12229849398136139,
-0.034881480038166046,
0.10294710099697113,
-0.3687572479248047,
-0.11857379972934723,
0.0734788104891777,
0.08750714361667633,
0.024329589679837227,
-0.14486901462078094,
-0.05168737843632698,
-0.01761118322610855,
-0.13509942591190338,
0.04647838696837425,
-0.066677525639534,
0.101963572204113,
-0.030448192730545998,
0.07140915095806122,
0.043689291924238205,
-0.06180340424180031,
0.17680393159389496,
0.004762761294841766,
0.06159214675426483,
-0.058107949793338776,
0.050052303820848465,
0.05246300250291824,
-0.08232016861438751,
0.07683652639389038,
-0.09811914712190628,
0.08512929081916809,
-0.15011504292488098,
-0.002161588054150343,
-0.0667995885014534,
0.057884979993104935,
-0.0452846884727478,
-0.04653027281165123,
-0.04046687111258507,
0.05699911713600159,
0.07479837536811829,
-0.040680836886167526,
0.08372902125120163,
0.04219431430101395,
0.08289608359336853,
0.1406964510679245,
0.03673990070819855,
0.022341806441545486,
-0.1422736942768097,
-0.008918281644582748,
-0.020504789426922798,
0.05427805334329605,
-0.13440634310245514,
0.011095760390162468,
0.12156074494123459,
0.0379880890250206,
0.11678892374038696,
0.025120023638010025,
-0.08732657134532928,
-0.014615984633564949,
0.04624584689736366,
-0.10759323835372925,
-0.15931780636310577,
-0.00017430176376365125,
0.03450554609298706,
-0.1231289878487587,
0.004196560010313988,
0.10646133124828339,
-0.059807587414979935,
-0.014601977542042732,
-0.015213211067020893,
0.05574681609869003,
-0.0031455657444894314,
0.20008952915668488,
0.034920983016490936,
0.07139613479375839,
-0.07572044432163239,
0.14435872435569763,
0.07809019088745117,
-0.066984161734581,
0.054676152765750885,
0.06194183975458145,
-0.09500457346439362,
-0.018722904846072197,
0.10375333577394485,
0.10665059834718704,
-0.01801195554435253,
-0.03227658197283745,
-0.0593266636133194,
-0.08709704875946045,
0.058928992599248886,
0.05056692659854889,
0.0351557657122612,
0.004577572923153639,
-0.0214338731020689,
-0.009953432716429234,
-0.110891193151474,
0.10290079563856125,
0.07169955968856812,
0.06915508955717087,
-0.13877688348293304,
0.07437223196029663,
-0.013411694206297398,
0.035406310111284256,
-0.0017775822198018432,
0.018774135038256645,
-0.08853171765804291,
-0.019420593976974487,
-0.16753119230270386,
0.012323394417762756,
-0.026966003701090813,
0.0032243693713098764,
-0.008870957419276237,
-0.040425531566143036,
-0.008061474189162254,
0.05216274783015251,
-0.0722590759396553,
-0.07797027379274368,
-0.01588495261967182,
0.06254248321056366,
-0.14451611042022705,
-0.020101528614759445,
0.026471935212612152,
-0.10160090029239655,
0.08968929201364517,
0.06747492402791977,
0.02537056989967823,
0.02516043744981289,
-0.09408959001302719,
-0.006563157308846712,
0.017420975491404533,
0.02719797007739544,
0.055818989872932434,
-0.11671473830938339,
-0.014780642464756966,
-0.041139956563711166,
-0.01056119054555893,
0.008515072055161,
0.047110289335250854,
-0.13464686274528503,
-0.04443394020199776,
-0.042178165167570114,
-0.0370292067527771,
-0.07524342834949493,
0.03594683110713959,
0.10394205898046494,
0.03471766412258148,
0.14937537908554077,
-0.07123047858476639,
0.045091718435287476,
-0.17427369952201843,
-0.03695661574602127,
-0.007380584254860878,
-0.01302436925470829,
-0.07996344566345215,
0.0016282880678772926,
0.08656778186559677,
-0.05502249673008919,
0.06516709923744202,
-0.05006169155240059,
0.050103165209293365,
0.025002269074320793,
-0.027272524312138557,
-0.02598029188811779,
0.011646793223917484,
0.14299607276916504,
0.05383521318435669,
-0.0166146382689476,
0.06124584376811981,
-0.037474773824214935,
0.04531799256801605,
0.04980895295739174,
0.14999347925186157,
0.1421370953321457,
0.009125396609306335,
0.061880409717559814,
0.04498959705233574,
-0.13322970271110535,
-0.1335032880306244,
0.09291934967041016,
-0.04511841759085655,
0.12080632895231247,
-0.0563032440841198,
0.16328993439674377,
0.10698264837265015,
-0.17516957223415375,
0.045759860426187515,
-0.06815661489963531,
-0.10718697309494019,
-0.10752766579389572,
-0.07097230106592178,
-0.056762099266052246,
-0.09410716593265533,
0.019980769604444504,
-0.09312206506729126,
0.05540992692112923,
0.07038744539022446,
0.019650058820843697,
0.008039936423301697,
0.11914899200201035,
-0.00955080147832632,
0.0000062322210396814626,
0.0502830371260643,
0.03992583230137825,
-0.004231769125908613,
-0.014418384060263634,
-0.06776352226734161,
0.04834609478712082,
-0.009714932180941105,
0.07806558161973953,
-0.01652994379401207,
0.010498751886188984,
0.049668993800878525,
0.001512695336714387,
-0.07126972824335098,
0.0337657667696476,
0.00574380112811923,
0.014417113736271858,
0.08390142023563385,
0.062235020101070404,
0.01605810970067978,
-0.05378380045294762,
0.28649774193763733,
-0.08063609898090363,
-0.057639673352241516,
-0.12978114187717438,
0.20777936279773712,
0.042426276952028275,
-0.006880783475935459,
0.06600048393011093,
-0.12966419756412506,
-0.02468906342983246,
0.11810433119535446,
0.1452895849943161,
-0.07609807699918747,
-0.02572665922343731,
-0.02898857370018959,
-0.018963607028126717,
-0.06355434656143188,
0.11332658678293228,
0.09124727547168732,
0.06622438132762909,
-0.05959957465529442,
0.048944734036922455,
-0.008708410896360874,
-0.025430437177419662,
-0.11074201762676239,
0.07880508154630661,
-0.026914486661553383,
0.006548590492457151,
-0.04919706657528877,
0.05630849301815033,
-0.010257118381559849,
-0.19297680258750916,
0.046955447643995285,
-0.13670693337917328,
-0.19233688712120056,
-0.019888902083039284,
0.0882318764925003,
-0.003725677030161023,
0.05882072448730469,
-0.010462136939167976,
0.004777197726070881,
0.1437557339668274,
-0.03265107423067093,
-0.04529692232608795,
-0.10203810781240463,
0.05589834228157997,
-0.028731821104884148,
0.22112995386123657,
0.022291388362646103,
0.058078933507204056,
0.10336875915527344,
0.02400178276002407,
-0.17389120161533356,
0.03247763589024544,
0.08731371909379959,
-0.08194699883460999,
0.018490491434931755,
0.15885530412197113,
-0.047444745898246765,
0.0660964697599411,
0.04137030616402626,
-0.07256826013326645,
-0.036874741315841675,
-0.004997162148356438,
0.011117091402411461,
-0.07355010509490967,
-0.006009574979543686,
-0.03769799321889877,
0.1574784219264984,
0.1845914125442505,
-0.04456020146608353,
0.003518196754157543,
-0.058969102799892426,
0.02519279345870018,
0.033710408955812454,
0.05962371826171875,
-0.012116358615458012,
-0.17182046175003052,
0.04212207719683647,
0.011073189787566662,
0.05853581801056862,
-0.19495876133441925,
-0.09006242454051971,
0.032628562301397324,
-0.050380151718854904,
-0.056087344884872437,
0.12357896566390991,
0.03997708112001419,
0.016856027767062187,
-0.03443634510040283,
-0.06624665856361389,
-0.05078057572245598,
0.1306721419095993,
-0.13885261118412018,
-0.05760311707854271
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | Tommidi/st_vit_trained-1epoch-ucf101-subset | [
"transformers",
"safetensors",
"st_vit",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-07T17:30:24+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #st_vit #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #st_vit #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
35,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #st_vit #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.05570429936051369,
0.2137511670589447,
-0.003358065150678158,
0.024937782436609268,
0.12344643473625183,
0.0005687462398782372,
0.046563174575567245,
0.12071564793586731,
-0.025259099900722504,
0.1139182522892952,
0.030914440751075745,
0.08739195019006729,
0.108574777841568,
0.16494834423065186,
0.02581079676747322,
-0.21664361655712128,
0.021972263231873512,
-0.092513807117939,
0.01725189760327339,
0.1145021989941597,
0.13485705852508545,
-0.10691213607788086,
0.08287516981363297,
-0.030043086037039757,
-0.0121110575273633,
-0.0068214526399970055,
-0.08663208037614822,
-0.07515288889408112,
0.07015538215637207,
0.07068207859992981,
0.06604450196027756,
0.01837434247136116,
0.10832002013921738,
-0.28513917326927185,
0.01607416197657585,
0.08505406975746155,
-0.00419366080313921,
0.06454268097877502,
0.0650641918182373,
-0.07560091465711594,
0.10277936607599258,
-0.08335147053003311,
0.14047180116176605,
0.07896949350833893,
-0.07898496836423874,
-0.2099640965461731,
-0.07245686650276184,
0.08404058963060379,
0.12577512860298157,
0.06405868381261826,
-0.024201525375247,
0.15871097147464752,
-0.08194445073604584,
0.010673961602151394,
0.1275133192539215,
-0.08980897068977356,
-0.048496827483177185,
0.04347142577171326,
0.10769272595643997,
0.09592066705226898,
-0.1287527084350586,
0.006958120968192816,
0.04254455864429474,
0.016857653856277466,
0.09859035164117813,
0.021999230608344078,
0.09309759736061096,
0.048560939729213715,
-0.14204595983028412,
-0.04514462873339653,
0.11667134612798691,
0.03381441533565521,
-0.05610955134034157,
-0.20422019064426422,
-0.007625863887369633,
-0.021844346076250076,
-0.024577245116233826,
-0.0576767772436142,
0.04335090145468712,
-0.03588148206472397,
0.061597224324941635,
-0.030445009469985962,
-0.10099130868911743,
-0.03598058968782425,
0.07995391637086868,
0.05205940082669258,
0.016159117221832275,
-0.021451836451888084,
0.03278529644012451,
0.11880841106176376,
0.040246110409498215,
-0.11036428809165955,
-0.06811126321554184,
-0.06794390827417374,
-0.10133186727762222,
-0.05056769400835037,
0.04352125898003578,
0.02268749475479126,
0.02894686348736286,
0.1968018114566803,
-0.007479736115783453,
0.0383053682744503,
0.03028780221939087,
0.004508857615292072,
0.059042856097221375,
0.09156389534473419,
-0.061962686479091644,
-0.1466813087463379,
-0.048566561192274094,
0.08683482557535172,
-0.0066268728114664555,
-0.037792667746543884,
-0.05171079933643341,
0.04612499848008156,
0.047256212681531906,
0.11810140311717987,
0.08820898830890656,
-0.007267971057444811,
-0.050878267735242844,
-0.02596031129360199,
0.2245747298002243,
-0.1413850337266922,
0.04642539843916893,
-0.008974277414381504,
-0.030878139659762383,
-0.04591722413897514,
0.03382661193609238,
0.02838059887290001,
-0.015240989625453949,
0.10359866917133331,
-0.061273738741874695,
-0.03711912781000137,
-0.10017072409391403,
-0.04717188701033592,
0.0350082628428936,
-0.005749464500695467,
-0.01986204832792282,
-0.0679158866405487,
-0.10470074415206909,
-0.04338342323899269,
0.06313471496105194,
-0.06048855185508728,
-0.03847065195441246,
0.012152530252933502,
-0.0540895015001297,
-0.004443710669875145,
-0.002475028857588768,
0.09528233855962753,
-0.03426503390073776,
0.037511471658945084,
-0.034544337540864944,
0.056296344846487045,
0.10254496335983276,
0.03632688149809837,
-0.07030779123306274,
0.05436348915100098,
-0.2148793339729309,
0.0853784903883934,
-0.10443845391273499,
0.033942416310310364,
-0.1600191444158554,
-0.050949763506650925,
0.018983567133545876,
0.013327579945325851,
0.01016842108219862,
0.12698014080524445,
-0.18719986081123352,
-0.022464357316493988,
0.1394852250814438,
-0.09603378176689148,
-0.10146932303905487,
0.07693009078502655,
-0.046814609318971634,
0.13340498507022858,
0.044664476066827774,
-0.034117892384529114,
0.07576531171798706,
-0.15887397527694702,
-0.056561145931482315,
-0.015079798176884651,
-0.010562318377196789,
0.1349911093711853,
0.05987713858485222,
-0.05702650919556618,
0.06608377397060394,
0.02290940098464489,
-0.031488727778196335,
-0.04681078717112541,
-0.04875839129090309,
-0.1021764874458313,
-0.005642387550324202,
-0.08074445277452469,
0.05492212250828743,
-0.010906236246228218,
-0.08094646781682968,
-0.03165549784898758,
-0.17732281982898712,
0.051942698657512665,
0.0857093557715416,
0.016773676499724388,
-0.012266465462744236,
-0.08150290697813034,
0.010908380150794983,
-0.0242806114256382,
-0.015031070448458195,
-0.16976644098758698,
-0.04868513345718384,
0.0421152301132679,
-0.1604491025209427,
0.03876825422048569,
-0.043682049959897995,
0.05846671387553215,
0.04454082250595093,
-0.05563071742653847,
-0.006849009543657303,
-0.017990170046687126,
0.019307563081383705,
-0.029858946800231934,
-0.19623076915740967,
-0.04533543065190315,
-0.031218690797686577,
0.1587037295103073,
-0.2456575483083725,
0.03576651215553284,
0.05370330065488815,
0.14480885863304138,
-0.0017552556237205863,
-0.043860193341970444,
0.021135950461030006,
-0.05322791635990143,
-0.05387458577752113,
-0.06349681317806244,
-0.007854725234210491,
-0.031917497515678406,
-0.04265320673584938,
0.02610553801059723,
-0.18523219227790833,
-0.029340120032429695,
0.09736476093530655,
0.10079009085893631,
-0.1585850864648819,
-0.012431168928742409,
-0.0453631654381752,
-0.062189407646656036,
-0.09237845242023468,
-0.05810590833425522,
0.12623348832130432,
0.047401558607816696,
0.0468166284263134,
-0.0788840651512146,
-0.06419339776039124,
0.0179806686937809,
-0.00035360551555640996,
-0.036891721189022064,
0.07362664490938187,
0.07609687000513077,
-0.10450740903615952,
0.07835382968187332,
0.07581426203250885,
0.07763756066560745,
0.09502170979976654,
0.016131630167365074,
-0.10761359333992004,
-0.02340857870876789,
0.024034785106778145,
0.02486669272184372,
0.1478547602891922,
-0.05172751471400261,
0.03467090427875519,
0.05202454701066017,
-0.044535767287015915,
0.012908765114843845,
-0.10249797254800797,
0.028774211183190346,
0.037253886461257935,
-0.005231223069131374,
0.031543564051389694,
-0.043056584894657135,
0.010337475687265396,
0.08016949892044067,
0.04568897932767868,
0.0439731739461422,
0.002297354396432638,
-0.01454879716038704,
-0.10262338072061539,
0.1666969507932663,
-0.10020972788333893,
-0.297836035490036,
-0.1513657569885254,
0.017026960849761963,
0.04111466556787491,
-0.017570000141859055,
0.029580384492874146,
-0.06297087669372559,
-0.11030436307191849,
-0.10366018861532211,
-0.0063893962651491165,
0.02679377608001232,
-0.077422134578228,
-0.07610747218132019,
0.07304812222719193,
0.03659141808748245,
-0.14403004944324493,
0.040652085095644,
0.05314547196030617,
-0.048650361597537994,
-0.015032786875963211,
0.0859709158539772,
0.11319845169782639,
0.15890946984291077,
-0.01605132780969143,
-0.026822108775377274,
0.020166777074337006,
0.20049993693828583,
-0.13392098248004913,
0.1123659610748291,
0.13552233576774597,
-0.045246776193380356,
0.0912533551454544,
0.179486945271492,
0.024746287614107132,
-0.0813380554318428,
0.04188155755400658,
0.04037332162261009,
-0.04971021041274071,
-0.25642508268356323,
-0.059530653059482574,
0.006094237323850393,
-0.07979082316160202,
0.08929162472486496,
0.09343841671943665,
0.13655731081962585,
0.03974501043558121,
-0.0818399041891098,
-0.043714489787817,
0.009397866204380989,
0.11453833431005478,
-0.04530727490782738,
-0.0070200590416789055,
0.08340045809745789,
-0.043600402772426605,
-0.000855231424793601,
0.10623525083065033,
0.01234451960772276,
0.18994955718517303,
0.020909279584884644,
0.1330929547548294,
0.061901308596134186,
0.0745885893702507,
-0.0009371726191602647,
0.020029284060001373,
0.04781128466129303,
0.016355976462364197,
-0.006840861868113279,
-0.09894926846027374,
0.009572113864123821,
0.13868531584739685,
0.0529002845287323,
0.024806004017591476,
0.010628330521285534,
-0.021681886166334152,
0.0564131960272789,
0.1699296236038208,
-0.007705842610448599,
-0.21435433626174927,
-0.06889940053224564,
0.07485757023096085,
-0.052101362496614456,
-0.12418662756681442,
-0.03711807727813721,
0.03844699636101723,
-0.17647750675678253,
0.03583279997110367,
-0.018521131947636604,
0.10187669098377228,
-0.09582708775997162,
-0.025606056675314903,
0.022676410153508186,
0.08388198912143707,
-0.02404973842203617,
0.09180989116430283,
-0.15079450607299805,
0.12927617132663727,
0.026467958465218544,
0.08270661532878876,
-0.11270297318696976,
0.08497460931539536,
-0.0083125289529562,
0.028918158262968063,
0.17930249869823456,
-0.0033632630947977304,
-0.06019608676433563,
-0.072627492249012,
-0.08941643685102463,
-0.02500317618250847,
0.12369580566883087,
-0.11176468431949615,
0.08070283383131027,
-0.011035280302166939,
-0.05373460054397583,
0.010264890268445015,
-0.10793616622686386,
-0.16480356454849243,
-0.19261159002780914,
0.0684640109539032,
-0.09916727989912033,
0.016581548377871513,
-0.1082279309630394,
-0.06539474427700043,
-0.03187005966901779,
0.23353637754917145,
-0.13728411495685577,
-0.0746678039431572,
-0.156398743391037,
-0.06516322493553162,
0.1713014394044876,
-0.03856266289949417,
0.0808568149805069,
-0.007341065444052219,
0.21389321982860565,
0.005804437678307295,
-0.004157801158726215,
0.06728991121053696,
-0.0903351679444313,
-0.17630039155483246,
-0.07992105931043625,
0.1375371664762497,
0.11850666999816895,
0.053096313029527664,
-0.0016138690989464521,
0.012334010563790798,
-0.022664329037070274,
-0.10700617730617523,
-0.007034884765744209,
0.13501602411270142,
0.0746832937002182,
0.03249754384160042,
-0.03400249779224396,
-0.11801403760910034,
-0.064614437520504,
-0.052405767142772675,
0.045828431844711304,
0.17897291481494904,
-0.09827461838722229,
0.18075211346149445,
0.1508801132440567,
-0.0614202618598938,
-0.21255800127983093,
0.045685771852731705,
0.046330004930496216,
-0.009858465753495693,
0.04152912274003029,
-0.1801142692565918,
0.08006644248962402,
0.011026542633771896,
-0.05326437950134277,
0.13030751049518585,
-0.15676456689834595,
-0.15730518102645874,
0.06249556317925453,
0.04989376291632652,
-0.2271057665348053,
-0.13286885619163513,
-0.08708418160676956,
-0.06738945096731186,
-0.1584213525056839,
0.08001243323087692,
-0.011404044926166534,
0.0067445202730596066,
0.04740811139345169,
0.019548820331692696,
0.02197127789258957,
-0.05545957013964653,
0.1899607628583908,
-0.009789925999939442,
0.016534049063920975,
-0.0760115385055542,
-0.06803451478481293,
0.09808424115180969,
-0.05713300034403801,
0.11375405639410019,
0.00030867906752973795,
0.014927709475159645,
-0.08842936903238297,
-0.05445292592048645,
-0.04633399471640587,
0.05327881500124931,
-0.0801289975643158,
-0.11267786473035812,
-0.05046195909380913,
0.08835139125585556,
0.08229611068964005,
-0.03719346225261688,
-0.009481902234256268,
-0.07554474472999573,
0.09361272305250168,
0.18960008025169373,
0.17111411690711975,
0.028214775025844574,
-0.07702282071113586,
0.011984715238213539,
-0.034928806126117706,
0.03510679677128792,
-0.23739688098430634,
0.04045528918504715,
0.05100715160369873,
0.03777351230382919,
0.10942219197750092,
-0.025190284475684166,
-0.17138943076133728,
-0.044905710965394974,
0.06642034649848938,
-0.046001043170690536,
-0.21577958762645721,
-0.012075510807335377,
0.09204386174678802,
-0.19074086844921112,
-0.018072832375764847,
0.026728451251983643,
-0.0406242199242115,
-0.03064594231545925,
0.00013967703853268176,
0.06437265872955322,
0.021772800013422966,
0.09325477480888367,
0.07026443630456924,
0.0999721959233284,
-0.09489531069993973,
0.09713911265134811,
0.10735493898391724,
-0.08204701542854309,
0.02898799069225788,
0.06870359927415848,
-0.057291239500045776,
-0.04048014432191849,
0.053980425000190735,
0.039468247443437576,
0.006086519453674555,
-0.0581432469189167,
0.0027392355259507895,
-0.06397754698991776,
0.052448440343141556,
0.11998318880796432,
0.022666698321700096,
-0.02469021826982498,
0.0645255297422409,
0.029029250144958496,
-0.11026784032583237,
0.09418091922998428,
0.01740378886461258,
0.03674943372607231,
-0.06241896003484726,
-0.0039422158151865005,
0.0435040257871151,
0.022204797714948654,
-0.020824486389756203,
-0.02956131473183632,
-0.038571249693632126,
-0.01524095144122839,
-0.15963494777679443,
-0.006219600327312946,
-0.07046904414892197,
0.006098572164773941,
0.006757817696779966,
-0.04860519617795944,
-0.006638118997216225,
0.02923220954835415,
-0.07473722100257874,
-0.06071319431066513,
-0.006327567622065544,
0.0952065959572792,
-0.15823984146118164,
0.004469591658562422,
0.07923907041549683,
-0.10976911336183548,
0.06724712252616882,
-0.0050472114235162735,
-0.0031538018956780434,
0.022188758477568626,
-0.15520933270454407,
0.05041011795401573,
-0.007867993786931038,
0.023938659578561783,
0.04220309108495712,
-0.16217857599258423,
0.003241467522457242,
-0.04709721729159355,
-0.029618583619594574,
-0.00900796614587307,
-0.05794106796383858,
-0.11490039527416229,
0.08344589173793793,
-0.015472081489861012,
-0.05993541702628136,
-0.011983027681708336,
0.050589144229888916,
0.08906885981559753,
-0.04140951856970787,
0.09616050124168396,
-0.00018605781951919198,
0.06140374392271042,
-0.17175798118114471,
-0.030678842216730118,
-0.038107212632894516,
0.008101641200482845,
0.004325470421463251,
-0.011633587069809437,
0.043346602469682693,
-0.002451202366501093,
0.22120395302772522,
-0.04036574438214302,
0.148934468626976,
0.05550919473171234,
0.008950652554631233,
0.004798842128366232,
0.06785714626312256,
0.05609952658414841,
0.027364015579223633,
0.004905372858047485,
0.021905817091464996,
-0.02669665589928627,
-0.010301942005753517,
-0.17710481584072113,
0.04019173979759216,
0.1391494870185852,
0.07943812012672424,
0.007647035177797079,
0.0673217922449112,
-0.12486109137535095,
-0.11242853850126266,
0.08621212840080261,
-0.03235071897506714,
0.008325548842549324,
-0.07554933428764343,
0.1338898092508316,
0.15194785594940186,
-0.154417023062706,
0.06898906081914902,
-0.05426538735628128,
-0.05378818139433861,
-0.09343206137418747,
-0.11229638755321503,
-0.062257468700408936,
-0.04018183425068855,
0.003256029449403286,
-0.045899491757154465,
0.056370221078395844,
0.05029931664466858,
-0.010479401797056198,
0.008013274520635605,
0.11863664537668228,
-0.007783854845911264,
0.0000189884831343079,
0.03033854439854622,
0.039834123104810715,
0.02821636199951172,
-0.056421346962451935,
0.024834949523210526,
0.02278032898902893,
0.031680285930633545,
0.06319505721330643,
0.029422655701637268,
-0.03947543725371361,
0.027869224548339844,
0.00825512781739235,
-0.10685615986585617,
0.024968035519123077,
-0.017432047054171562,
-0.06924793124198914,
0.1221415251493454,
0.03398130461573601,
0.01257460005581379,
-0.03504735603928566,
0.2362840175628662,
-0.06781076639890671,
-0.076201431453228,
-0.13451990485191345,
0.10213013738393784,
-0.013717721216380596,
0.058102209120988846,
0.045565709471702576,
-0.12319894134998322,
0.004503061529248953,
0.12789539992809296,
0.11404299736022949,
0.0012703804532065988,
0.011767015792429447,
0.0357314832508564,
0.00519994692876935,
-0.06046826019883156,
0.04164519160985947,
0.05897338688373566,
0.13378684222698212,
-0.074750617146492,
0.06887663900852203,
0.008998668752610683,
-0.07617877423763275,
-0.03906102851033211,
0.11380910873413086,
-0.02562093548476696,
0.031789083033800125,
-0.043579090386629105,
0.10810410976409912,
-0.0631370022892952,
-0.300861656665802,
0.037658318877220154,
-0.09307476878166199,
-0.15385554730892181,
-0.016475550830364227,
0.05672220513224602,
-0.017636196687817574,
0.019389742985367775,
0.07200947403907776,
-0.05852338671684265,
0.17834995687007904,
0.03855089843273163,
-0.08383230865001678,
-0.058593153953552246,
0.058595769107341766,
-0.07618720084428787,
0.29584139585494995,
0.004314475692808628,
0.03156234323978424,
0.10291639715433121,
-0.02817497029900551,
-0.15865081548690796,
0.025919383391737938,
0.10971265286207199,
-0.08997935801744461,
0.08148243278265,
0.19308438897132874,
-0.02033587358891964,
0.11305531114339828,
0.05622059106826782,
-0.06627935916185379,
0.05504925176501274,
-0.042728643864393234,
-0.04827238246798515,
-0.09343753010034561,
0.06380754709243774,
-0.05933108925819397,
0.15512260794639587,
0.09560972452163696,
-0.04920487850904465,
-0.0063011967577040195,
-0.05258584022521973,
0.04541392996907234,
0.013106484897434711,
0.12142833322286606,
0.012441393919289112,
-0.17513152956962585,
0.03197946026921272,
-0.010420299135148525,
0.11354460567235947,
-0.23551952838897705,
-0.08292408287525177,
0.08480016142129898,
-0.02405981905758381,
-0.05182001739740372,
0.1004592552781105,
0.07474427670240402,
0.04224398359656334,
-0.04595436900854111,
-0.09319831430912018,
-0.014723841100931168,
0.15525862574577332,
-0.1369098573923111,
-0.011507726274430752
] |
null | null | fastai |
# Amazing!
🥳 Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the [documentation here](https://huggingface.co/docs/hub/model-repos))!
2. Create a demo in Gradio or Streamlit using 🤗 Spaces ([documentation here](https://huggingface.co/docs/hub/spaces)).
3. Join the fastai community on the [Fastai Discord](https://discord.com/invite/YKrxeNn)!
Greetings fellow fastlearner 🤝! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| {"tags": ["fastai"]} | null | valintea/primer-modelo | [
"fastai",
"has_space",
"region:us"
] | 2024-02-07T17:30:54+00:00 | [] | [] | TAGS
#fastai #has_space #region-us
|
# Amazing!
Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the documentation here)!
2. Create a demo in Gradio or Streamlit using Spaces (documentation here).
3. Join the fastai community on the Fastai Discord!
Greetings fellow fastlearner ! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| [
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
"TAGS\n#fastai #has_space #region-us \n",
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
13,
20,
79,
3,
6,
12,
8
] | [
"passage: TAGS\n#fastai #has_space #region-us \n# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---# Model card## Model description\nMore information needed## Intended uses & limitations\nMore information needed## Training and evaluation data\nMore information needed"
] | [
-0.048121724277734756,
-0.024616125971078873,
0.002038001548498869,
0.10439170897006989,
0.135872021317482,
0.11887997388839722,
0.07405775785446167,
0.09980081021785736,
0.07783667743206024,
0.02590852417051792,
0.08961158245801926,
-0.08088712394237518,
0.08744348585605621,
0.271692156791687,
0.06988707184791565,
-0.22761479020118713,
0.04051019623875618,
-0.00024903909070417285,
0.08053462207317352,
0.06629016250371933,
0.13507555425167084,
-0.05464952811598778,
0.14010503888130188,
-0.004088983871042728,
-0.19050447642803192,
-0.042929794639348984,
-0.01773718371987343,
-0.02527874894440174,
0.12317648530006409,
-0.04744937643408775,
0.05381017178297043,
0.015037551522254944,
0.007565062493085861,
-0.07253646105527878,
0.0623294934630394,
0.040457066148519516,
0.01740180514752865,
0.059235580265522,
-0.07249044626951218,
0.08950132131576538,
0.08404164761304855,
-0.024370938539505005,
-0.1097978875041008,
0.07827875018119812,
-0.14424212276935577,
-0.21762843430042267,
-0.1253085881471634,
-0.09017651528120041,
0.028519365936517715,
0.004388005938380957,
-0.025051530450582504,
0.12801909446716309,
-0.13558274507522583,
-0.040698226541280746,
0.20124278962612152,
-0.17012301087379456,
-0.05505548417568207,
0.034343402832746506,
0.09226689487695694,
-0.05829555168747902,
-0.06347129493951797,
0.10614984482526779,
0.09640881419181824,
-0.019833475351333618,
0.05516824126243591,
0.002579754451289773,
0.021173657849431038,
0.01370104867964983,
-0.06150497496128082,
0.04717832803726196,
-0.010183089412748814,
0.048132527619600296,
-0.09465572983026505,
-0.1303568333387375,
-0.004072192590683699,
0.01214400865137577,
-0.048744890838861465,
-0.07019646465778351,
0.07833103090524673,
-0.011118141002953053,
-0.04357248544692993,
-0.13031910359859467,
-0.09131011366844177,
-0.12358787655830383,
0.008646543137729168,
0.09500427544116974,
0.003679296001791954,
0.07374339550733566,
-0.08258994668722153,
0.06774985045194626,
-0.17329485714435577,
-0.06484591960906982,
-0.08138520270586014,
-0.11546400189399719,
0.021133482456207275,
-0.0387684591114521,
0.02668963186442852,
0.15394504368305206,
0.12983950972557068,
0.023976242169737816,
0.04388163983821869,
-0.038937073200941086,
0.051190316677093506,
0.058571770787239075,
0.03395717963576317,
0.034934818744659424,
-0.036981891840696335,
-0.1793210655450821,
-0.016702448949217796,
-0.011550825089216232,
0.07954040914773941,
-0.07523109763860703,
-0.05632320046424866,
0.013454885222017765,
-0.11071494966745377,
0.07202339172363281,
-0.03576776012778282,
-0.0032025426626205444,
0.01168301422148943,
0.018371861428022385,
0.21271461248397827,
0.03955606371164322,
0.014191740192472935,
-0.008875265717506409,
-0.13453757762908936,
-0.06874168664216995,
-0.06896194815635681,
0.03361047804355621,
0.04448792710900307,
-0.0028071461711078882,
-0.07672245055437088,
0.04325154796242714,
-0.06045534089207649,
-0.03508453071117401,
0.008032378740608692,
-0.18221288919448853,
0.007458044681698084,
-0.10049355030059814,
-0.12126200646162033,
0.05306628718972206,
0.01695440337061882,
-0.08215925842523575,
0.08141279965639114,
0.02662261202931404,
0.020931517705321312,
-0.009988143108785152,
-0.005391082260757685,
0.06874798238277435,
-0.08508864045143127,
0.029901226982474327,
0.17170792818069458,
0.13024519383907318,
-0.08046911656856537,
-0.0006887061172164977,
-0.10965746641159058,
0.04426072910428047,
-0.13325683772563934,
0.02251482754945755,
-0.09062390774488449,
0.11723794043064117,
-0.042396437376737595,
0.002038756385445595,
-0.029030200093984604,
0.0960269495844841,
0.08189879357814789,
0.16663365066051483,
-0.2419009804725647,
-0.031095001846551895,
0.13240347802639008,
-0.10711425542831421,
-0.1807439625263214,
0.18486657738685608,
-0.012035200372338295,
0.11329247802495956,
-0.047014184296131134,
0.18334640562534332,
-0.02612062357366085,
-0.13582459092140198,
-0.058872904628515244,
0.005852419883012772,
-0.2269321084022522,
-0.06286033242940903,
0.09738040715456009,
0.13425657153129578,
-0.042984943836927414,
0.007112155202776194,
0.026316028088331223,
0.13609857857227325,
-0.06715573370456696,
-0.05195777863264084,
-0.012255736626684666,
-0.10902371257543564,
0.041914235800504684,
0.018215661868453026,
0.035408079624176025,
-0.059880174696445465,
-0.02931194379925728,
-0.053190283477306366,
0.13146710395812988,
0.09760832786560059,
-0.03670211136341095,
-0.049620725214481354,
0.1689043790102005,
-0.07763876020908356,
-0.033587727695703506,
0.07560533285140991,
-0.08268500119447708,
0.03266897425055504,
0.03090597130358219,
0.055881720036268234,
0.07766123116016388,
0.08522116392850876,
0.06057543307542801,
0.00819048099219799,
0.034654274582862854,
0.12095347046852112,
-0.013591280207037926,
-0.05039411783218384,
0.021508218720555305,
0.016904234886169434,
-0.019032588228583336,
0.29030677676200867,
-0.1951042115688324,
0.024724548682570457,
-0.06477324664592743,
0.07631538063287735,
0.06136792525649071,
0.003575638635084033,
0.08580143749713898,
-0.06023019179701805,
-0.019061198458075523,
-0.04803973436355591,
0.046805646270513535,
-0.0666879191994667,
-0.04162997007369995,
0.2621194124221802,
-0.05497581139206886,
0.044914912432432175,
0.12313763797283173,
-0.05873025581240654,
-0.07091446220874786,
0.01009807363152504,
-0.00793424155563116,
0.03249288722872734,
-0.04042816907167435,
0.043721720576286316,
-0.10840129852294922,
-0.06674089282751083,
0.1573198139667511,
-0.038477856665849686,
0.06786153465509415,
0.032288823276758194,
-0.04958454892039299,
-0.0648743286728859,
0.04650486260652542,
0.13598160445690155,
-0.0875244215130806,
0.07435166835784912,
0.17612984776496887,
-0.010562662966549397,
0.168031245470047,
0.08435525000095367,
-0.07075224816799164,
-0.09465329349040985,
-0.051014289259910583,
-0.021595727652311325,
0.21222901344299316,
-0.07084725052118301,
-0.054564714431762695,
0.05911700800061226,
-0.013703816570341587,
0.07196151465177536,
-0.06009222939610481,
-0.08332337439060211,
0.03227344527840614,
-0.04517695680260658,
0.011517706327140331,
0.13512636721134186,
-0.07090822607278824,
0.04681389778852463,
0.031489867717027664,
-0.0662703812122345,
0.02217509225010872,
0.033389873802661896,
0.0068921963684260845,
0.033959709107875824,
0.07332495599985123,
-0.20893315970897675,
-0.08408680558204651,
-0.13727638125419617,
0.037881869822740555,
0.021770721301436424,
0.045787326991558075,
-0.08602345734834671,
0.02231026627123356,
-0.08954031765460968,
-0.07987114042043686,
0.029592275619506836,
-0.026350297033786774,
-0.11349643021821976,
-0.03396226093173027,
-0.009560913778841496,
-0.06662604957818985,
-0.02250705659389496,
-0.05024505779147148,
0.03983384370803833,
0.04479299485683441,
0.058377087116241455,
0.12796473503112793,
-0.013808943331241608,
-0.03839317709207535,
0.000370211957488209,
-0.022712308913469315,
0.16396735608577728,
-0.14746315777301788,
0.07954913377761841,
0.19160102307796478,
0.11742953956127167,
0.028144672513008118,
0.028885571286082268,
0.03537585213780403,
-0.06289814412593842,
-0.000050317394197918475,
0.03226194158196449,
-0.09392514824867249,
-0.05801016092300415,
-0.020014392212033272,
-0.04031052812933922,
0.17134574055671692,
-0.12160717695951462,
0.03345204517245293,
0.04098419472575188,
0.09783966839313507,
0.10073629021644592,
-0.028829937800765038,
-0.1815856397151947,
0.038818612694740295,
-0.24060091376304626,
-0.05831146240234375,
0.027899866923689842,
-0.09110201895236969,
-0.06232144311070442,
0.17409387230873108,
0.013794700615108013,
0.011769929900765419,
-0.006736889015883207,
0.07983319461345673,
0.0110100656747818,
0.1217205822467804,
0.05947643890976906,
-0.05539114400744438,
0.025202350690960884,
-0.09962950646877289,
-0.07107596844434738,
-0.04035590961575508,
-0.05832801014184952,
0.07548832893371582,
0.1409129947423935,
-0.025475580245256424,
-0.020795362070202827,
0.023489827290177345,
0.08550169318914413,
0.0423230417072773,
0.16739299893379211,
-0.16016584634780884,
-0.026555389165878296,
0.04571257904171944,
-0.03384667634963989,
-0.05433850735425949,
-0.010291114449501038,
0.1137225553393364,
-0.02820689231157303,
-0.040318265557289124,
0.021242983639240265,
0.06503437459468842,
0.01481706090271473,
0.05012747645378113,
-0.04056356102228165,
0.14796851575374603,
-0.03461192920804024,
0.019330544397234917,
-0.12413888424634933,
0.13848772644996643,
0.021095896139740944,
-0.03901609033346176,
-0.06735876202583313,
-0.05808034539222717,
0.18150931596755981,
0.0025602965615689754,
0.10535930097103119,
0.012098877690732479,
-0.12160047143697739,
-0.1359938681125641,
-0.11211287975311279,
0.005111907608807087,
0.08330471813678741,
-0.023147236555814743,
-0.022247863933444023,
0.022165266796946526,
-0.036149751394987106,
-0.0530381016433239,
0.15749511122703552,
-0.1289154291152954,
-0.001082550617866218,
0.014728817157447338,
0.06971760839223862,
-0.08223173767328262,
0.026267826557159424,
0.014071501791477203,
-0.1119147390127182,
0.10590848326683044,
0.2521335482597351,
0.10338116437196732,
-0.09591643512248993,
-0.07697287201881409,
0.03418830782175064,
-0.012184361927211285,
-0.000774814048781991,
-0.006932659074664116,
0.0495428591966629,
-0.005566445179283619,
0.006762749515473843,
0.12971895933151245,
-0.07130889594554901,
0.011540771462023258,
-0.08449850976467133,
0.05566910281777382,
-0.05276734381914139,
0.01761564053595066,
-0.002672141883522272,
-0.008124710991978645,
-0.07340748608112335,
-0.061829522252082825,
0.1609770804643631,
-0.07277000695466995,
-0.06468547880649567,
0.05801168829202652,
0.03307786211371422,
0.01431563775986433,
-0.03584568202495575,
-0.04342148080468178,
0.18088261783123016,
0.29330700635910034,
-0.08191116154193878,
0.10001859813928604,
0.09677296131849289,
0.034820813685655594,
-0.23625829815864563,
0.029798466712236404,
-0.1455078274011612,
0.04449721798300743,
0.040447335690259933,
-0.0409548319876194,
0.04191497340798378,
0.10835777968168259,
-0.06094440817832947,
0.2048867791891098,
-0.03527235612273216,
-0.07983248680830002,
-0.01788630709052086,
0.03109324350953102,
0.29443636536598206,
-0.11833466589450836,
0.006058716680854559,
-0.10420958697795868,
-0.21566011011600494,
0.06983078271150589,
-0.18948867917060852,
0.13948246836662292,
-0.05087858438491821,
0.03576415032148361,
-0.01149723306298256,
-0.07561972737312317,
0.20518061518669128,
-0.15641045570373535,
0.05273103713989258,
-0.13722458481788635,
-0.1327189952135086,
0.01617460884153843,
-0.10048147290945053,
0.1545477658510208,
-0.11024226248264313,
-0.023215843364596367,
-0.2284185290336609,
0.012587235309183598,
-0.023200806230306625,
0.10030807554721832,
0.01800704374909401,
-0.07980740070343018,
-0.08767345547676086,
0.1316242516040802,
-0.06486566364765167,
0.034810543060302734,
-0.06996636837720871,
-0.050714004784822464,
-0.010929876938462257,
-0.045061707496643066,
0.03034941293299198,
-0.07934719324111938,
0.15192505717277527,
-0.016938980668783188,
-0.04507075995206833,
0.08636019378900528,
-0.2479533851146698,
0.023727843537926674,
0.025351112708449364,
-0.03495599329471588,
0.09001832455396652,
-0.025513244792819023,
-0.06256973743438721,
0.12282291799783707,
0.1402233988046646,
-0.07322840392589569,
-0.2460673749446869,
-0.06281693279743195,
0.0076784128323197365,
0.039165716618299484,
0.06561196595430374,
0.05125982314348221,
-0.07261458039283752,
-0.011131617240607738,
-0.026896944269537926,
0.030595947057008743,
-0.11692017316818237,
-0.03854857385158539,
0.07790639251470566,
0.017095070332288742,
-0.07846562564373016,
0.07280377298593521,
0.014225782826542854,
-0.021511616185307503,
0.007357571739703417,
0.148970365524292,
0.007519228849560022,
-0.14747941493988037,
-0.06656096875667572,
0.2007484883069992,
-0.01197928935289383,
-0.07260087132453918,
-0.05383119732141495,
-0.008990069851279259,
-0.0476234145462513,
0.05585788935422897,
0.05367223918437958,
-0.013585401698946953,
0.07708586007356644,
0.06263149529695511,
-0.10210110992193222,
-0.046256959438323975,
-0.066561758518219,
0.04169114679098129,
-0.10485753417015076,
0.060470130294561386,
0.009529483504593372,
0.12185006588697433,
-0.09983488917350769,
-0.01802929677069187,
-0.10810204595327377,
-0.06766588985919952,
-0.17349553108215332,
-0.05834362283349037,
-0.041105758398771286,
-0.015651104971766472,
0.03658895567059517,
0.010445823892951012,
-0.057867538183927536,
-0.0442853718996048,
-0.07536603510379791,
0.038444988429546356,
0.06147460639476776,
0.03932281583547592,
-0.03912714496254921,
0.04001858830451965,
0.05909334123134613,
0.013087345287203789,
0.17542624473571777,
0.038768354803323746,
0.05504675209522247,
-0.05045998468995094,
-0.16491834819316864,
-0.05276111513376236,
-0.0074316514655947685,
-0.07559102028608322,
0.1224973127245903,
-0.007679440546780825,
0.007880088873207569,
-0.08065467327833176,
0.03924860805273056,
0.028234204277396202,
0.10404064506292343,
-0.0028364830650389194,
0.10070426017045975,
0.019627176225185394,
-0.07226712256669998,
-0.025392837822437286,
0.021809715777635574,
0.12809939682483673,
0.01567147858440876,
0.026090998202562332,
0.033139873296022415,
0.016619985923171043,
-0.057361043989658356,
0.033977724611759186,
-0.04997231811285019,
-0.15123651921749115,
0.02628709189593792,
-0.05165188014507294,
0.005062380339950323,
-0.016889680176973343,
0.20362506806850433,
0.07867538928985596,
-0.06474173814058304,
-0.010664013214409351,
0.015816617757081985,
-0.0168940220028162,
-0.03121885471045971,
-0.012740966863930225,
0.04592578858137131,
-0.001151384087279439,
-0.04866636544466019,
0.11825273931026459,
0.05015748366713524,
0.05386412516236305,
0.0596686452627182,
0.12528513371944427,
0.016759619116783142,
0.13257254660129547,
0.061999931931495667,
-0.03403807803988457,
-0.13461735844612122,
-0.04495539888739586,
-0.1254577934741974,
0.04646851494908333,
-0.08697032928466797,
0.09941662102937698,
0.1144254133105278,
-0.05959030240774155,
-0.030464433133602142,
-0.08851305395364761,
-0.008356761187314987,
-0.06041252240538597,
0.039516255259513855,
-0.02262675203382969,
-0.0873224213719368,
0.0481097511947155,
0.05495472997426987,
-0.022752324119210243,
0.13218675553798676,
0.015727028250694275,
-0.036317698657512665,
0.13270340859889984,
-0.07583184540271759,
0.11758984625339508,
0.061510033905506134,
-0.043043944984674454,
-0.11560922116041183,
-0.020150646567344666,
-0.06641761213541031,
-0.10098972916603088,
-0.006782987620681524,
-0.005399650428444147,
-0.07349002361297607,
-0.059971679002046585,
0.08397487550973892,
-0.03124053031206131,
-0.09979676455259323,
-0.032152675092220306,
0.0038895104080438614,
0.06054706871509552,
-0.01686914451420307,
-0.0034020058810710907,
0.04728743061423302,
0.015076374635100365,
0.1653461456298828,
-0.02208263985812664,
0.06234867498278618,
-0.13855914771556854,
0.16070103645324707,
-0.14684462547302246,
-0.029404424130916595,
-0.1890171319246292,
-0.09729582816362381,
-0.05156542733311653,
0.20326784253120422,
0.2840938866138458,
-0.19109351933002472,
-0.010187864303588867,
0.020078664645552635,
-0.014484191313385963,
-0.08961770683526993,
0.12571553885936737,
0.029420215636491776,
-0.023631498217582703,
-0.07249019294977188,
-0.02037387527525425,
0.005258576478809118,
-0.06544211506843567,
-0.026979785412549973,
0.18310695886611938,
0.001496660872362554,
0.059546373784542084,
-0.09605178982019424,
0.01754261925816536,
-0.14839904010295868,
-0.10467469692230225,
-0.02111995778977871,
-0.16156397759914398,
-0.09646477550268173,
0.006635562051087618,
0.038640011101961136,
0.08000610023736954,
0.03268849849700928,
-0.015172510407865047,
0.06479045748710632,
-0.056333884596824646,
-0.0037216036580502987,
-0.1231912299990654,
0.00034658415825106204,
0.062129102647304535,
-0.07422006875276566,
0.2545335292816162,
-0.03070417232811451,
-0.12370815873146057,
0.09026903659105301,
-0.03299184888601303,
-0.12452623248100281,
0.07951879501342773,
-0.005700904875993729,
-0.11531132459640503,
-0.057989440858364105,
0.18941475450992584,
-0.012821312062442303,
-0.1364315301179886,
0.046368811279535294,
-0.17166484892368317,
0.031349923461675644,
0.0363016203045845,
-0.001313706859946251,
-0.04714022949337959,
0.024538639932870865,
-0.008008457720279694,
0.10724439471960068,
0.1382838785648346,
0.016739921644330025,
-0.011060068383812904,
-0.05056179314851761,
0.07912429422140121,
0.056927867233753204,
-0.05218246951699257,
-0.1282637119293213,
-0.08599764108657837,
0.03429819270968437,
0.04119478166103363,
-0.08113081753253937,
-0.16903182864189148,
-0.03668912500143051,
-0.10082915425300598,
-0.004939202684909105,
0.051785312592983246,
0.06585265696048737,
0.29044589400291443,
0.06326735019683838,
0.0016605621203780174,
-0.13649453222751617,
0.050569336861371994,
0.0868251696228981,
-0.04697931930422783,
-0.07670357078313828
] |
null | null | transformers | # original model repo :
📖 this is a cutomized version of the following model [aaronespasa/deepfake-detection-resnetinceptionv1](https://huggingface.co/aaronespasa/deepfake-detection-resnetinceptionv1)
# how to use
```python
from transformers import pipeline
pipe = pipeline(model="not-lain/deepfake",trust_remote_code=True)
pipe.predict("img_path.jpg")
```
```python
>> {"confidences":confidences,"face_with_mask": face_with_mask}
```
# dependencies
to install related dependencies simply use the command
```
!wget https://huggingface.co/not-lain/deepfake/resolve/main/requirements.txt && pip install -r requirements.txt
``` | {"license": "apache-2.0", "library_name": "transformers", "base_model": "aaronespasa/deepfake-detection-resnetinceptionv1"} | image-classification | wyyadd/fork-detect-fake | [
"transformers",
"pytorch",
"safetensors",
"ResNet",
"image-classification",
"custom_code",
"base_model:aaronespasa/deepfake-detection-resnetinceptionv1",
"license:apache-2.0",
"autotrain_compatible",
"region:us"
] | 2024-02-07T17:31:03+00:00 | [] | [] | TAGS
#transformers #pytorch #safetensors #ResNet #image-classification #custom_code #base_model-aaronespasa/deepfake-detection-resnetinceptionv1 #license-apache-2.0 #autotrain_compatible #region-us
| # original model repo :
this is a cutomized version of the following model aaronespasa/deepfake-detection-resnetinceptionv1
# how to use
# dependencies
to install related dependencies simply use the command
| [
"# original model repo : \n this is a cutomized version of the following model aaronespasa/deepfake-detection-resnetinceptionv1",
"# how to use",
"# dependencies\nto install related dependencies simply use the command"
] | [
"TAGS\n#transformers #pytorch #safetensors #ResNet #image-classification #custom_code #base_model-aaronespasa/deepfake-detection-resnetinceptionv1 #license-apache-2.0 #autotrain_compatible #region-us \n",
"# original model repo : \n this is a cutomized version of the following model aaronespasa/deepfake-detection-resnetinceptionv1",
"# how to use",
"# dependencies\nto install related dependencies simply use the command"
] | [
70,
34,
4,
12
] | [
"passage: TAGS\n#transformers #pytorch #safetensors #ResNet #image-classification #custom_code #base_model-aaronespasa/deepfake-detection-resnetinceptionv1 #license-apache-2.0 #autotrain_compatible #region-us \n# original model repo : \n this is a cutomized version of the following model aaronespasa/deepfake-detection-resnetinceptionv1# how to use# dependencies\nto install related dependencies simply use the command"
] | [
-0.0821857824921608,
0.09073251485824585,
-0.003327552927657962,
0.057437993586063385,
0.09513994306325912,
-0.015139881521463394,
0.08901015669107437,
0.07307354360818863,
0.005902146454900503,
-0.032572973519563675,
0.07852710038423538,
0.12695467472076416,
0.05402353033423424,
0.07660137116909027,
-0.013868422247469425,
-0.096404068171978,
0.07684991508722305,
-0.003277468727901578,
-0.024924444034695625,
0.07795963436365128,
0.1334812492132187,
-0.01716664619743824,
0.10647532343864441,
0.008296513929963112,
-0.12830857932567596,
0.04882693290710449,
0.029733650386333466,
-0.0855836570262909,
0.09356942027807236,
0.02029428631067276,
0.09037183225154877,
0.020520050078630447,
0.09922314435243607,
-0.0781237930059433,
0.03051513433456421,
0.0854342058300972,
-0.011784381233155727,
0.08342382311820984,
0.03918388485908508,
0.02815544418990612,
-0.16262978315353394,
0.005775626748800278,
0.006681632250547409,
0.015921225771307945,
-0.003361885203048587,
-0.15829648077487946,
-0.045936219394207,
0.11622656136751175,
0.11614193767309189,
0.07191202044487,
0.06762975454330444,
0.24710553884506226,
-0.0471305288374424,
0.03627876192331314,
0.10149581730365753,
-0.2868134379386902,
-0.03456132858991623,
0.11336155235767365,
-0.08626419305801392,
-0.011654728092253208,
0.028371477499604225,
0.014999286271631718,
0.0630570650100708,
0.025662295520305634,
-0.025381632149219513,
-0.03753872215747833,
-0.09080503135919571,
-0.05268348753452301,
-0.07357119768857956,
-0.025975922122597694,
0.1669788807630539,
0.07589886337518692,
-0.05419408529996872,
0.05196031183004379,
-0.1471119225025177,
0.11718402057886124,
-0.056199051439762115,
-0.014252731576561928,
0.04423447325825691,
0.05881650000810623,
0.03828977793455124,
-0.024760261178016663,
-0.11855591833591461,
-0.05734599009156227,
-0.1238628476858139,
0.06808428466320038,
0.04972338676452637,
0.11689294874668121,
-0.0143497409299016,
0.0411817729473114,
-0.05340276286005974,
-0.1358586996793747,
0.04099247604608536,
-0.10411647707223892,
0.03709612414240837,
0.023517658933997154,
-0.05689191818237305,
0.014054952189326286,
0.13958220183849335,
0.30800601840019226,
-0.07901567220687866,
-0.008438359946012497,
-0.0781666561961174,
0.07485076040029526,
-0.04733339324593544,
0.06462441384792328,
-0.12948496639728546,
-0.02018708921968937,
0.1640353500843048,
0.01237957738339901,
0.1539800465106964,
-0.0012751949252560735,
-0.1146383136510849,
-0.027293620631098747,
0.04942908510565758,
0.05686356872320175,
0.08738937228918076,
-0.002130416687577963,
-0.03653643652796745,
-0.07027840614318848,
0.16960257291793823,
-0.10613352060317993,
-0.03848729655146599,
0.020556557923555374,
-0.04293985664844513,
0.07246298342943192,
0.13476470112800598,
-0.011364852078258991,
-0.018830398097634315,
-0.07837647944688797,
-0.00909055583178997,
-0.08797111362218857,
-0.007009400520473719,
-0.07272914052009583,
0.04550635442137718,
-0.05199752375483513,
0.0771973729133606,
-0.2230752557516098,
-0.23955556750297546,
-0.004956162068992853,
0.09409154206514359,
-0.01649467647075653,
-0.10009057819843292,
0.06132924184203148,
-0.03658362105488777,
-0.07412939518690109,
0.002114561153575778,
-0.029309269040822983,
-0.04024256765842438,
-0.004898799117654562,
-0.024604735895991325,
0.04844847321510315,
-0.10457386076450348,
0.029033655300736427,
-0.15885409712791443,
-0.00582502456381917,
-0.13658837974071503,
0.006493853870779276,
-0.05050891265273094,
0.1570044606924057,
-0.101222924888134,
0.003203334752470255,
0.004798675421625376,
0.007973508909344673,
0.015659943222999573,
0.13641174137592316,
-0.0956651121377945,
-0.041487645357847214,
0.04465220123529434,
-0.210333913564682,
-0.15558764338493347,
0.08137048035860062,
0.03802895173430443,
0.08956792205572128,
0.054712940007448196,
0.06528882682323456,
0.06184124946594238,
-0.02953450381755829,
0.025955550372600555,
0.03469463437795639,
-0.04829246178269386,
-0.1448013037443161,
0.02144770510494709,
0.07215819507837296,
-0.13802553713321686,
0.03226830065250397,
-0.1890907734632492,
0.13762739300727844,
-0.005813779309391975,
-0.0904192179441452,
-0.05362432822585106,
-0.14903192222118378,
-0.07265523076057434,
-0.015037288889288902,
0.042018525302410126,
0.012577448971569538,
-0.037138745188713074,
-0.1370372772216797,
0.1276087462902069,
0.027479056268930435,
0.018172789365053177,
-0.10677413642406464,
0.16086365282535553,
0.008499739691615105,
0.0497557632625103,
-0.11188334971666336,
0.006871803663671017,
0.05073575675487518,
-0.008367142640054226,
0.0001784947671694681,
-0.09490754455327988,
0.05553441122174263,
-0.005580285098403692,
0.015721146017313004,
-0.08141446858644485,
0.07001017779111862,
0.007329127751290798,
0.06089453771710396,
-0.18691834807395935,
0.006388251669704914,
-0.039660584181547165,
0.19684046506881714,
-0.05536830425262451,
0.02128606289625168,
-0.1321045607328415,
0.15029625594615936,
0.013321107253432274,
0.10030236095190048,
-0.009248307906091213,
-0.051755502820014954,
-0.004252140410244465,
-0.0321526974439621,
0.03309564292430878,
0.04730316996574402,
-0.04353295639157295,
0.06700726598501205,
0.003205703105777502,
0.15417438745498657,
0.13961908221244812,
-0.18689097464084625,
-0.02026565931737423,
0.13136188685894012,
-0.003933676052838564,
-0.03403211757540703,
0.034707698971033096,
0.008743833750486374,
-0.050410039722919464,
0.02518080174922943,
0.06267000734806061,
-0.05814129486680031,
0.08035413920879364,
-0.0321701280772686,
-0.10214400291442871,
-0.0021503798197954893,
-0.002488094847649336,
0.07694582641124725,
-0.1926824450492859,
0.10281211137771606,
0.2598821818828583,
-0.14672063291072845,
-0.03242487087845802,
-0.08232582360506058,
-0.03944619372487068,
0.01835792511701584,
0.019249657168984413,
0.01675988920032978,
0.03840178996324539,
0.05518379062414169,
0.03898971527814865,
0.0651518777012825,
-0.12318170815706253,
0.0012565783690661192,
-0.061973679810762405,
0.011850772425532341,
0.00567912682890892,
0.01766272634267807,
0.05190570279955864,
-0.027114354074001312,
-0.015527977608144283,
0.04762808233499527,
-0.03176324814558029,
-0.1759686917066574,
0.04754052311182022,
-0.023646032437682152,
-0.043461184948682785,
0.15604284405708313,
-0.11285343021154404,
-0.20484189689159393,
-0.13405004143714905,
-0.06341080367565155,
-0.06687386333942413,
0.010422141291201115,
0.033364083617925644,
-0.1257910132408142,
-0.05903162062168121,
-0.03864246979355812,
-0.07969935983419418,
0.19248469173908234,
0.0031336257234215736,
0.08492235839366913,
0.05921584740281105,
0.08893121033906937,
-0.06772472709417343,
0.005413881968706846,
-0.016857856884598732,
-0.13795456290245056,
0.08996716886758804,
-0.09219572693109512,
-0.009799446910619736,
0.10659348219633102,
0.03045148029923439,
-0.031048288568854332,
0.044962361454963684,
0.11127908527851105,
0.013819851912558079,
0.061917033046483994,
0.28598251938819885,
-0.013226913288235664,
0.037649061530828476,
0.15188443660736084,
0.011150488629937172,
-0.011582646518945694,
-0.014188764616847038,
-0.029681816697120667,
-0.026050085201859474,
-0.18485604226589203,
-0.10118332505226135,
0.01846865378320217,
0.040243975818157196,
0.07411982864141464,
0.07559601962566376,
0.08223491907119751,
0.1421099305152893,
-0.030131248757243156,
-0.00196092645637691,
0.007605814840644598,
0.06074081361293793,
0.2022341638803482,
-0.003736026817932725,
0.0979413241147995,
-0.08912720531225204,
-0.017764862626791,
0.09582844376564026,
0.05786572024226189,
0.04707694426178932,
0.044007107615470886,
-0.014811128377914429,
0.027730057016015053,
0.19485560059547424,
0.060977913439273834,
0.14559553563594818,
0.005665602162480354,
-0.015484006144106388,
-0.024693647399544716,
-0.061584118753671646,
-0.07908444851636887,
0.05534324422478676,
-0.14332008361816406,
0.028740761801600456,
0.06659367680549622,
0.12361686676740646,
0.0035874447785317898,
0.2505042254924774,
0.04498753324151039,
-0.24995951354503632,
-0.05301840975880623,
0.015413165092468262,
0.03225933387875557,
-0.001731321681290865,
0.023619990795850754,
-0.09994308650493622,
-0.053721290081739426,
0.07489991933107376,
-0.018403081223368645,
0.1036502867937088,
-0.0024301493540406227,
0.05242454260587692,
-0.03314604610204697,
0.07423672825098038,
0.07966343313455582,
0.04869059473276138,
-0.12279977649450302,
0.10882264375686646,
0.026434237137436867,
-0.024669621139764786,
-0.0425863191485405,
0.026920562610030174,
0.12363099306821823,
0.17603974044322968,
0.08388333767652512,
0.009735285304486752,
-0.055667899549007416,
-0.0012510096421465278,
-0.1349080502986908,
0.04771513491868973,
-0.08994009345769882,
0.0018917336128652096,
-0.03441794589161873,
-0.03326823562383652,
-0.038632672280073166,
0.024883385747671127,
0.09892905503511429,
-0.10283055901527405,
-0.04033360257744789,
-0.03193370997905731,
0.051528431475162506,
-0.06041179224848747,
-0.10603878647089005,
-0.06661678105592728,
-0.019510220736265182,
0.11657596379518509,
0.029435239732265472,
-0.06145678833127022,
-0.0401623472571373,
0.011990549974143505,
0.07388874143362045,
-0.08417100459337234,
0.0949496328830719,
-0.08376152068376541,
0.029276607558131218,
-0.00570311164483428,
-0.1634843498468399,
0.09205751121044159,
-0.16348059475421906,
-0.0074727036990225315,
0.0026699842419475317,
0.026632852852344513,
-0.027401339262723923,
-0.006082902662456036,
0.0575912706553936,
-0.010968673042953014,
-0.062990702688694,
-0.03233415633440018,
0.010349239222705364,
0.054522983729839325,
-0.0885251984000206,
0.0999782383441925,
-0.12526245415210724,
-0.16789250075817108,
-0.023026246577501297,
0.04775885492563248,
0.13083752989768982,
0.1538574993610382,
-0.046456966549158096,
-0.01384952012449503,
0.24923710525035858,
0.012421432882547379,
-0.2807198166847229,
-0.12145092338323593,
-0.02252928540110588,
-0.012250532396137714,
0.017505142837762833,
-0.18976043164730072,
0.22416427731513977,
0.04261040687561035,
-0.10305910557508469,
-0.011230194009840488,
-0.05873824283480644,
-0.009867810644209385,
0.29530656337738037,
0.010320042259991169,
0.2462279349565506,
-0.07721256464719772,
-0.028070027008652687,
-0.0636172965168953,
-0.04800903797149658,
0.08417264372110367,
-0.19027850031852722,
0.0324920155107975,
-0.005835827440023422,
0.014750799164175987,
-0.015738891437649727,
0.01045325119048357,
0.05473189800977707,
-0.0725516825914383,
0.06672566384077072,
-0.07605006545782089,
0.11862056702375412,
0.09104690700769424,
-0.03509419411420822,
0.08277416229248047,
-0.008481624536216259,
0.07371986657381058,
0.053694456815719604,
0.026478314772248268,
-0.07235508412122726,
0.10455189645290375,
0.014799459837377071,
-0.06485532224178314,
0.0037096268497407436,
0.021761944517493248,
-0.02458662912249565,
0.039080727845430374,
0.13036487996578217,
0.09800010919570923,
0.10322877764701843,
0.2917703092098236,
-0.005132749211043119,
-0.07911164313554764,
0.08527503162622452,
-0.042303573340177536,
-0.08146153390407562,
0.13077087700366974,
-0.10379572212696075,
0.061561618000268936,
0.015266395173966885,
-0.05023915320634842,
-0.030095160007476807,
-0.00023998752294573933,
-0.10926515609025955,
-0.08747867494821548,
0.08647390455007553,
-0.09973011910915375,
0.00547014782205224,
0.0020915595814585686,
0.14592209458351135,
0.022904641926288605,
0.0633501335978508,
0.17289431393146515,
-0.04080702364444733,
-0.0431339293718338,
0.050512999296188354,
0.03198036551475525,
-0.09023558348417282,
0.08075928688049316,
0.12495961040258408,
-0.004292165860533714,
-0.09966070204973221,
0.18322527408599854,
0.03526833653450012,
0.03592035174369812,
-0.013333024457097054,
0.03006594069302082,
-0.11995302885770798,
-0.12087695300579071,
-0.03367232903838158,
0.07136405259370804,
-0.17955276370048523,
-0.13298723101615906,
-0.11317133158445358,
-0.09099780768156052,
-0.016429202631115913,
0.0062309447675943375,
0.12182211875915527,
0.002880294341593981,
0.037544503808021545,
-0.06581400334835052,
-0.00176126672886312,
0.056430671364068985,
0.04700781777501106,
0.04796209931373596,
-0.19939522445201874,
-0.17486350238323212,
0.009278029203414917,
0.10294287651777267,
-0.05175457522273064,
0.01749267615377903,
-0.07546444237232208,
0.018683336675167084,
-0.22806432843208313,
0.02820867858827114,
-0.011645592749118805,
0.013781865127384663,
0.0242715235799551,
-0.017587844282388687,
-0.049620673060417175,
0.06472749263048172,
-0.0767144039273262,
-0.024377277120947838,
0.002834462793543935,
0.039877310395240784,
-0.10677298903465271,
-0.02638464793562889,
0.04653555899858475,
-0.014812630601227283,
0.09438659250736237,
0.05824596434831619,
-0.059907712042331696,
0.05581849068403244,
-0.14848431944847107,
-0.1726132035255432,
0.06673270463943481,
0.05818978697061539,
0.01784510165452957,
-0.06187532842159271,
0.011564923450350761,
0.06667491793632507,
-0.10473186522722244,
-0.046121034771203995,
0.16809070110321045,
-0.03550000488758087,
0.0308191180229187,
-0.03399660438299179,
-0.029747674241662025,
-0.01418844610452652,
-0.06975177675485611,
0.11180686950683594,
0.047892939299345016,
0.16319529712200165,
-0.037947386503219604,
0.021544059738516808,
-0.11223138123750687,
0.024491433054208755,
-0.07480534166097641,
-0.066502645611763,
-0.09064903110265732,
-0.038826532661914825,
-0.025450369343161583,
-0.01805008016526699,
0.1412961333990097,
-0.07201315462589264,
-0.0015379786491394043,
0.05138600245118141,
0.15235628187656403,
0.11952504515647888,
0.0063703469932079315,
0.15410679578781128,
0.07567081600427628,
0.029199253767728806,
-0.03191820904612541,
-0.03716351091861725,
0.0655868798494339,
-0.10837139934301376,
-0.0995233952999115,
0.13578960299491882,
0.06607639044523239,
0.06939388066530228,
0.13545967638492584,
0.019705437123775482,
0.08298906683921814,
-0.026867596432566643,
-0.013838415034115314,
0.013230433687567711,
0.014424161985516548,
0.08462651818990707,
0.2206573337316513,
0.023001113906502724,
-0.03740658238530159,
-0.005903477314859629,
0.034402135759592056,
-0.15853896737098694,
-0.17428499460220337,
-0.13027025759220123,
-0.19297227263450623,
-0.02267940528690815,
-0.0058236876502633095,
-0.152353897690773,
0.12310869991779327,
0.01791050285100937,
-0.05388542264699936,
0.04082827270030975,
-0.027663270011544228,
-0.038548793643713,
0.041446395218372345,
0.006026725750416517,
-0.08320140838623047,
0.09257976710796356,
-0.07298094034194946,
0.0628400593996048,
0.12791050970554352,
-0.008018956519663334,
-0.004175703041255474,
0.096101313829422,
0.16028043627738953,
-0.040321122854948044,
-0.07979941368103027,
-0.03019103780388832,
-0.026657452806830406,
-0.07202878594398499,
0.07912173867225647,
-0.036800842732191086,
0.03513322398066521,
0.0543462410569191,
0.19786979258060455,
-0.08829203248023987,
-0.1681956648826599,
-0.1290712207555771,
0.08744603395462036,
-0.0637667328119278,
0.03936136141419411,
-0.03179653733968735,
-0.06296063214540482,
-0.058089934289455414,
0.15358801186084747,
0.280547559261322,
-0.08690332621335983,
0.022735172882676125,
-0.000723190838471055,
0.0010438392637297511,
-0.17190296947956085,
0.03588708117604256,
0.14106814563274384,
-0.023496318608522415,
-0.050636354833841324,
-0.02921648882329464,
-0.05201315879821777,
-0.03296828269958496,
-0.2019488662481308,
-0.06587851792573929,
-0.0021033540833741426,
-0.06392442435026169,
-0.030688809230923653,
0.08505146950483322,
-0.062310103327035904,
-0.007597627118229866,
-0.023528190329670906,
-0.08608158677816391,
-0.12271232157945633,
-0.08607103675603867,
0.09629552811384201,
0.09587076306343079,
0.02027096413075924,
-0.07670897990465164,
0.0482676699757576,
0.08255139738321304,
-0.0343502014875412,
-0.06312180310487747,
-0.15803100168704987,
0.045369502156972885,
-0.014883832074701786,
0.22315585613250732,
0.032991692423820496,
0.0034469948150217533,
0.06029887497425079,
0.003568923333659768,
-0.11082755029201508,
0.051528919488191605,
-0.016281118616461754,
-0.0829077884554863,
-0.009843621402978897,
-0.042564403265714645,
-0.04045909270644188,
0.08527685701847076,
0.02599724940955639,
-0.1051960438489914,
-0.037680648267269135,
0.053729113191366196,
-0.05038798600435257,
-0.08872011303901672,
0.03653779625892639,
-0.11796560883522034,
0.07544676214456558,
0.0406917966902256,
-0.05572319030761719,
-0.02355813793838024,
-0.09067802876234055,
0.05113966017961502,
0.15140098333358765,
-0.17492341995239258,
-0.0008544575539417565,
-0.0642637386918068,
0.043545372784137726,
-0.05202244594693184,
0.054946623742580414,
-0.00531458854675293,
-0.02233988605439663,
-0.11078803241252899,
-0.03889643773436546,
-0.1327047199010849,
0.01719895750284195,
0.09156892448663712,
0.04315459728240967,
0.018738161772489548,
-0.06048810854554176,
0.005218614358454943,
0.013825726695358753,
-0.08034247905015945,
-0.14803546667099
] |
null | null | fastai |
# Amazing!
🥳 Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the [documentation here](https://huggingface.co/docs/hub/model-repos))!
2. Create a demo in Gradio or Streamlit using 🤗 Spaces ([documentation here](https://huggingface.co/docs/hub/spaces)).
3. Join the fastai community on the [Fastai Discord](https://discord.com/invite/YKrxeNn)!
Greetings fellow fastlearner 🤝! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| {"tags": ["fastai"]} | null | jonruida/intel-image-classification | [
"fastai",
"has_space",
"region:us"
] | 2024-02-07T17:31:23+00:00 | [] | [] | TAGS
#fastai #has_space #region-us
|
# Amazing!
Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the documentation here)!
2. Create a demo in Gradio or Streamlit using Spaces (documentation here).
3. Join the fastai community on the Fastai Discord!
Greetings fellow fastlearner ! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| [
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
"TAGS\n#fastai #has_space #region-us \n",
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
13,
20,
79,
3,
6,
12,
8
] | [
"passage: TAGS\n#fastai #has_space #region-us \n# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---# Model card## Model description\nMore information needed## Intended uses & limitations\nMore information needed## Training and evaluation data\nMore information needed"
] | [
-0.048121724277734756,
-0.024616125971078873,
0.002038001548498869,
0.10439170897006989,
0.135872021317482,
0.11887997388839722,
0.07405775785446167,
0.09980081021785736,
0.07783667743206024,
0.02590852417051792,
0.08961158245801926,
-0.08088712394237518,
0.08744348585605621,
0.271692156791687,
0.06988707184791565,
-0.22761479020118713,
0.04051019623875618,
-0.00024903909070417285,
0.08053462207317352,
0.06629016250371933,
0.13507555425167084,
-0.05464952811598778,
0.14010503888130188,
-0.004088983871042728,
-0.19050447642803192,
-0.042929794639348984,
-0.01773718371987343,
-0.02527874894440174,
0.12317648530006409,
-0.04744937643408775,
0.05381017178297043,
0.015037551522254944,
0.007565062493085861,
-0.07253646105527878,
0.0623294934630394,
0.040457066148519516,
0.01740180514752865,
0.059235580265522,
-0.07249044626951218,
0.08950132131576538,
0.08404164761304855,
-0.024370938539505005,
-0.1097978875041008,
0.07827875018119812,
-0.14424212276935577,
-0.21762843430042267,
-0.1253085881471634,
-0.09017651528120041,
0.028519365936517715,
0.004388005938380957,
-0.025051530450582504,
0.12801909446716309,
-0.13558274507522583,
-0.040698226541280746,
0.20124278962612152,
-0.17012301087379456,
-0.05505548417568207,
0.034343402832746506,
0.09226689487695694,
-0.05829555168747902,
-0.06347129493951797,
0.10614984482526779,
0.09640881419181824,
-0.019833475351333618,
0.05516824126243591,
0.002579754451289773,
0.021173657849431038,
0.01370104867964983,
-0.06150497496128082,
0.04717832803726196,
-0.010183089412748814,
0.048132527619600296,
-0.09465572983026505,
-0.1303568333387375,
-0.004072192590683699,
0.01214400865137577,
-0.048744890838861465,
-0.07019646465778351,
0.07833103090524673,
-0.011118141002953053,
-0.04357248544692993,
-0.13031910359859467,
-0.09131011366844177,
-0.12358787655830383,
0.008646543137729168,
0.09500427544116974,
0.003679296001791954,
0.07374339550733566,
-0.08258994668722153,
0.06774985045194626,
-0.17329485714435577,
-0.06484591960906982,
-0.08138520270586014,
-0.11546400189399719,
0.021133482456207275,
-0.0387684591114521,
0.02668963186442852,
0.15394504368305206,
0.12983950972557068,
0.023976242169737816,
0.04388163983821869,
-0.038937073200941086,
0.051190316677093506,
0.058571770787239075,
0.03395717963576317,
0.034934818744659424,
-0.036981891840696335,
-0.1793210655450821,
-0.016702448949217796,
-0.011550825089216232,
0.07954040914773941,
-0.07523109763860703,
-0.05632320046424866,
0.013454885222017765,
-0.11071494966745377,
0.07202339172363281,
-0.03576776012778282,
-0.0032025426626205444,
0.01168301422148943,
0.018371861428022385,
0.21271461248397827,
0.03955606371164322,
0.014191740192472935,
-0.008875265717506409,
-0.13453757762908936,
-0.06874168664216995,
-0.06896194815635681,
0.03361047804355621,
0.04448792710900307,
-0.0028071461711078882,
-0.07672245055437088,
0.04325154796242714,
-0.06045534089207649,
-0.03508453071117401,
0.008032378740608692,
-0.18221288919448853,
0.007458044681698084,
-0.10049355030059814,
-0.12126200646162033,
0.05306628718972206,
0.01695440337061882,
-0.08215925842523575,
0.08141279965639114,
0.02662261202931404,
0.020931517705321312,
-0.009988143108785152,
-0.005391082260757685,
0.06874798238277435,
-0.08508864045143127,
0.029901226982474327,
0.17170792818069458,
0.13024519383907318,
-0.08046911656856537,
-0.0006887061172164977,
-0.10965746641159058,
0.04426072910428047,
-0.13325683772563934,
0.02251482754945755,
-0.09062390774488449,
0.11723794043064117,
-0.042396437376737595,
0.002038756385445595,
-0.029030200093984604,
0.0960269495844841,
0.08189879357814789,
0.16663365066051483,
-0.2419009804725647,
-0.031095001846551895,
0.13240347802639008,
-0.10711425542831421,
-0.1807439625263214,
0.18486657738685608,
-0.012035200372338295,
0.11329247802495956,
-0.047014184296131134,
0.18334640562534332,
-0.02612062357366085,
-0.13582459092140198,
-0.058872904628515244,
0.005852419883012772,
-0.2269321084022522,
-0.06286033242940903,
0.09738040715456009,
0.13425657153129578,
-0.042984943836927414,
0.007112155202776194,
0.026316028088331223,
0.13609857857227325,
-0.06715573370456696,
-0.05195777863264084,
-0.012255736626684666,
-0.10902371257543564,
0.041914235800504684,
0.018215661868453026,
0.035408079624176025,
-0.059880174696445465,
-0.02931194379925728,
-0.053190283477306366,
0.13146710395812988,
0.09760832786560059,
-0.03670211136341095,
-0.049620725214481354,
0.1689043790102005,
-0.07763876020908356,
-0.033587727695703506,
0.07560533285140991,
-0.08268500119447708,
0.03266897425055504,
0.03090597130358219,
0.055881720036268234,
0.07766123116016388,
0.08522116392850876,
0.06057543307542801,
0.00819048099219799,
0.034654274582862854,
0.12095347046852112,
-0.013591280207037926,
-0.05039411783218384,
0.021508218720555305,
0.016904234886169434,
-0.019032588228583336,
0.29030677676200867,
-0.1951042115688324,
0.024724548682570457,
-0.06477324664592743,
0.07631538063287735,
0.06136792525649071,
0.003575638635084033,
0.08580143749713898,
-0.06023019179701805,
-0.019061198458075523,
-0.04803973436355591,
0.046805646270513535,
-0.0666879191994667,
-0.04162997007369995,
0.2621194124221802,
-0.05497581139206886,
0.044914912432432175,
0.12313763797283173,
-0.05873025581240654,
-0.07091446220874786,
0.01009807363152504,
-0.00793424155563116,
0.03249288722872734,
-0.04042816907167435,
0.043721720576286316,
-0.10840129852294922,
-0.06674089282751083,
0.1573198139667511,
-0.038477856665849686,
0.06786153465509415,
0.032288823276758194,
-0.04958454892039299,
-0.0648743286728859,
0.04650486260652542,
0.13598160445690155,
-0.0875244215130806,
0.07435166835784912,
0.17612984776496887,
-0.010562662966549397,
0.168031245470047,
0.08435525000095367,
-0.07075224816799164,
-0.09465329349040985,
-0.051014289259910583,
-0.021595727652311325,
0.21222901344299316,
-0.07084725052118301,
-0.054564714431762695,
0.05911700800061226,
-0.013703816570341587,
0.07196151465177536,
-0.06009222939610481,
-0.08332337439060211,
0.03227344527840614,
-0.04517695680260658,
0.011517706327140331,
0.13512636721134186,
-0.07090822607278824,
0.04681389778852463,
0.031489867717027664,
-0.0662703812122345,
0.02217509225010872,
0.033389873802661896,
0.0068921963684260845,
0.033959709107875824,
0.07332495599985123,
-0.20893315970897675,
-0.08408680558204651,
-0.13727638125419617,
0.037881869822740555,
0.021770721301436424,
0.045787326991558075,
-0.08602345734834671,
0.02231026627123356,
-0.08954031765460968,
-0.07987114042043686,
0.029592275619506836,
-0.026350297033786774,
-0.11349643021821976,
-0.03396226093173027,
-0.009560913778841496,
-0.06662604957818985,
-0.02250705659389496,
-0.05024505779147148,
0.03983384370803833,
0.04479299485683441,
0.058377087116241455,
0.12796473503112793,
-0.013808943331241608,
-0.03839317709207535,
0.000370211957488209,
-0.022712308913469315,
0.16396735608577728,
-0.14746315777301788,
0.07954913377761841,
0.19160102307796478,
0.11742953956127167,
0.028144672513008118,
0.028885571286082268,
0.03537585213780403,
-0.06289814412593842,
-0.000050317394197918475,
0.03226194158196449,
-0.09392514824867249,
-0.05801016092300415,
-0.020014392212033272,
-0.04031052812933922,
0.17134574055671692,
-0.12160717695951462,
0.03345204517245293,
0.04098419472575188,
0.09783966839313507,
0.10073629021644592,
-0.028829937800765038,
-0.1815856397151947,
0.038818612694740295,
-0.24060091376304626,
-0.05831146240234375,
0.027899866923689842,
-0.09110201895236969,
-0.06232144311070442,
0.17409387230873108,
0.013794700615108013,
0.011769929900765419,
-0.006736889015883207,
0.07983319461345673,
0.0110100656747818,
0.1217205822467804,
0.05947643890976906,
-0.05539114400744438,
0.025202350690960884,
-0.09962950646877289,
-0.07107596844434738,
-0.04035590961575508,
-0.05832801014184952,
0.07548832893371582,
0.1409129947423935,
-0.025475580245256424,
-0.020795362070202827,
0.023489827290177345,
0.08550169318914413,
0.0423230417072773,
0.16739299893379211,
-0.16016584634780884,
-0.026555389165878296,
0.04571257904171944,
-0.03384667634963989,
-0.05433850735425949,
-0.010291114449501038,
0.1137225553393364,
-0.02820689231157303,
-0.040318265557289124,
0.021242983639240265,
0.06503437459468842,
0.01481706090271473,
0.05012747645378113,
-0.04056356102228165,
0.14796851575374603,
-0.03461192920804024,
0.019330544397234917,
-0.12413888424634933,
0.13848772644996643,
0.021095896139740944,
-0.03901609033346176,
-0.06735876202583313,
-0.05808034539222717,
0.18150931596755981,
0.0025602965615689754,
0.10535930097103119,
0.012098877690732479,
-0.12160047143697739,
-0.1359938681125641,
-0.11211287975311279,
0.005111907608807087,
0.08330471813678741,
-0.023147236555814743,
-0.022247863933444023,
0.022165266796946526,
-0.036149751394987106,
-0.0530381016433239,
0.15749511122703552,
-0.1289154291152954,
-0.001082550617866218,
0.014728817157447338,
0.06971760839223862,
-0.08223173767328262,
0.026267826557159424,
0.014071501791477203,
-0.1119147390127182,
0.10590848326683044,
0.2521335482597351,
0.10338116437196732,
-0.09591643512248993,
-0.07697287201881409,
0.03418830782175064,
-0.012184361927211285,
-0.000774814048781991,
-0.006932659074664116,
0.0495428591966629,
-0.005566445179283619,
0.006762749515473843,
0.12971895933151245,
-0.07130889594554901,
0.011540771462023258,
-0.08449850976467133,
0.05566910281777382,
-0.05276734381914139,
0.01761564053595066,
-0.002672141883522272,
-0.008124710991978645,
-0.07340748608112335,
-0.061829522252082825,
0.1609770804643631,
-0.07277000695466995,
-0.06468547880649567,
0.05801168829202652,
0.03307786211371422,
0.01431563775986433,
-0.03584568202495575,
-0.04342148080468178,
0.18088261783123016,
0.29330700635910034,
-0.08191116154193878,
0.10001859813928604,
0.09677296131849289,
0.034820813685655594,
-0.23625829815864563,
0.029798466712236404,
-0.1455078274011612,
0.04449721798300743,
0.040447335690259933,
-0.0409548319876194,
0.04191497340798378,
0.10835777968168259,
-0.06094440817832947,
0.2048867791891098,
-0.03527235612273216,
-0.07983248680830002,
-0.01788630709052086,
0.03109324350953102,
0.29443636536598206,
-0.11833466589450836,
0.006058716680854559,
-0.10420958697795868,
-0.21566011011600494,
0.06983078271150589,
-0.18948867917060852,
0.13948246836662292,
-0.05087858438491821,
0.03576415032148361,
-0.01149723306298256,
-0.07561972737312317,
0.20518061518669128,
-0.15641045570373535,
0.05273103713989258,
-0.13722458481788635,
-0.1327189952135086,
0.01617460884153843,
-0.10048147290945053,
0.1545477658510208,
-0.11024226248264313,
-0.023215843364596367,
-0.2284185290336609,
0.012587235309183598,
-0.023200806230306625,
0.10030807554721832,
0.01800704374909401,
-0.07980740070343018,
-0.08767345547676086,
0.1316242516040802,
-0.06486566364765167,
0.034810543060302734,
-0.06996636837720871,
-0.050714004784822464,
-0.010929876938462257,
-0.045061707496643066,
0.03034941293299198,
-0.07934719324111938,
0.15192505717277527,
-0.016938980668783188,
-0.04507075995206833,
0.08636019378900528,
-0.2479533851146698,
0.023727843537926674,
0.025351112708449364,
-0.03495599329471588,
0.09001832455396652,
-0.025513244792819023,
-0.06256973743438721,
0.12282291799783707,
0.1402233988046646,
-0.07322840392589569,
-0.2460673749446869,
-0.06281693279743195,
0.0076784128323197365,
0.039165716618299484,
0.06561196595430374,
0.05125982314348221,
-0.07261458039283752,
-0.011131617240607738,
-0.026896944269537926,
0.030595947057008743,
-0.11692017316818237,
-0.03854857385158539,
0.07790639251470566,
0.017095070332288742,
-0.07846562564373016,
0.07280377298593521,
0.014225782826542854,
-0.021511616185307503,
0.007357571739703417,
0.148970365524292,
0.007519228849560022,
-0.14747941493988037,
-0.06656096875667572,
0.2007484883069992,
-0.01197928935289383,
-0.07260087132453918,
-0.05383119732141495,
-0.008990069851279259,
-0.0476234145462513,
0.05585788935422897,
0.05367223918437958,
-0.013585401698946953,
0.07708586007356644,
0.06263149529695511,
-0.10210110992193222,
-0.046256959438323975,
-0.066561758518219,
0.04169114679098129,
-0.10485753417015076,
0.060470130294561386,
0.009529483504593372,
0.12185006588697433,
-0.09983488917350769,
-0.01802929677069187,
-0.10810204595327377,
-0.06766588985919952,
-0.17349553108215332,
-0.05834362283349037,
-0.041105758398771286,
-0.015651104971766472,
0.03658895567059517,
0.010445823892951012,
-0.057867538183927536,
-0.0442853718996048,
-0.07536603510379791,
0.038444988429546356,
0.06147460639476776,
0.03932281583547592,
-0.03912714496254921,
0.04001858830451965,
0.05909334123134613,
0.013087345287203789,
0.17542624473571777,
0.038768354803323746,
0.05504675209522247,
-0.05045998468995094,
-0.16491834819316864,
-0.05276111513376236,
-0.0074316514655947685,
-0.07559102028608322,
0.1224973127245903,
-0.007679440546780825,
0.007880088873207569,
-0.08065467327833176,
0.03924860805273056,
0.028234204277396202,
0.10404064506292343,
-0.0028364830650389194,
0.10070426017045975,
0.019627176225185394,
-0.07226712256669998,
-0.025392837822437286,
0.021809715777635574,
0.12809939682483673,
0.01567147858440876,
0.026090998202562332,
0.033139873296022415,
0.016619985923171043,
-0.057361043989658356,
0.033977724611759186,
-0.04997231811285019,
-0.15123651921749115,
0.02628709189593792,
-0.05165188014507294,
0.005062380339950323,
-0.016889680176973343,
0.20362506806850433,
0.07867538928985596,
-0.06474173814058304,
-0.010664013214409351,
0.015816617757081985,
-0.0168940220028162,
-0.03121885471045971,
-0.012740966863930225,
0.04592578858137131,
-0.001151384087279439,
-0.04866636544466019,
0.11825273931026459,
0.05015748366713524,
0.05386412516236305,
0.0596686452627182,
0.12528513371944427,
0.016759619116783142,
0.13257254660129547,
0.061999931931495667,
-0.03403807803988457,
-0.13461735844612122,
-0.04495539888739586,
-0.1254577934741974,
0.04646851494908333,
-0.08697032928466797,
0.09941662102937698,
0.1144254133105278,
-0.05959030240774155,
-0.030464433133602142,
-0.08851305395364761,
-0.008356761187314987,
-0.06041252240538597,
0.039516255259513855,
-0.02262675203382969,
-0.0873224213719368,
0.0481097511947155,
0.05495472997426987,
-0.022752324119210243,
0.13218675553798676,
0.015727028250694275,
-0.036317698657512665,
0.13270340859889984,
-0.07583184540271759,
0.11758984625339508,
0.061510033905506134,
-0.043043944984674454,
-0.11560922116041183,
-0.020150646567344666,
-0.06641761213541031,
-0.10098972916603088,
-0.006782987620681524,
-0.005399650428444147,
-0.07349002361297607,
-0.059971679002046585,
0.08397487550973892,
-0.03124053031206131,
-0.09979676455259323,
-0.032152675092220306,
0.0038895104080438614,
0.06054706871509552,
-0.01686914451420307,
-0.0034020058810710907,
0.04728743061423302,
0.015076374635100365,
0.1653461456298828,
-0.02208263985812664,
0.06234867498278618,
-0.13855914771556854,
0.16070103645324707,
-0.14684462547302246,
-0.029404424130916595,
-0.1890171319246292,
-0.09729582816362381,
-0.05156542733311653,
0.20326784253120422,
0.2840938866138458,
-0.19109351933002472,
-0.010187864303588867,
0.020078664645552635,
-0.014484191313385963,
-0.08961770683526993,
0.12571553885936737,
0.029420215636491776,
-0.023631498217582703,
-0.07249019294977188,
-0.02037387527525425,
0.005258576478809118,
-0.06544211506843567,
-0.026979785412549973,
0.18310695886611938,
0.001496660872362554,
0.059546373784542084,
-0.09605178982019424,
0.01754261925816536,
-0.14839904010295868,
-0.10467469692230225,
-0.02111995778977871,
-0.16156397759914398,
-0.09646477550268173,
0.006635562051087618,
0.038640011101961136,
0.08000610023736954,
0.03268849849700928,
-0.015172510407865047,
0.06479045748710632,
-0.056333884596824646,
-0.0037216036580502987,
-0.1231912299990654,
0.00034658415825106204,
0.062129102647304535,
-0.07422006875276566,
0.2545335292816162,
-0.03070417232811451,
-0.12370815873146057,
0.09026903659105301,
-0.03299184888601303,
-0.12452623248100281,
0.07951879501342773,
-0.005700904875993729,
-0.11531132459640503,
-0.057989440858364105,
0.18941475450992584,
-0.012821312062442303,
-0.1364315301179886,
0.046368811279535294,
-0.17166484892368317,
0.031349923461675644,
0.0363016203045845,
-0.001313706859946251,
-0.04714022949337959,
0.024538639932870865,
-0.008008457720279694,
0.10724439471960068,
0.1382838785648346,
0.016739921644330025,
-0.011060068383812904,
-0.05056179314851761,
0.07912429422140121,
0.056927867233753204,
-0.05218246951699257,
-0.1282637119293213,
-0.08599764108657837,
0.03429819270968437,
0.04119478166103363,
-0.08113081753253937,
-0.16903182864189148,
-0.03668912500143051,
-0.10082915425300598,
-0.004939202684909105,
0.051785312592983246,
0.06585265696048737,
0.29044589400291443,
0.06326735019683838,
0.0016605621203780174,
-0.13649453222751617,
0.050569336861371994,
0.0868251696228981,
-0.04697931930422783,
-0.07670357078313828
] |
null | null | transformers |
Machine-generated text-detection by fine-tuning of language models
===
This project is related to a bachelor's thesis with the title "*Turning Poachers into Gamekeepers: Detecting Machine-Generated Text in Academia using Large Language Models*" (see [here](https://ntnuopen.ntnu.no/ntnu-xmlui/handle/11250/3078096)) written by *Nicolai Thorer Sivesind* and *Andreas Bentzen Winje* at the *Department of Computer Science* at the *Norwegian University of Science and Technology*.
It contains text classification models trained to distinguish human-written text from text generated by language models like ChatGPT and GPT-3. The best models were able to achieve an accuracy of 100% on real and *GPT-3*-generated wikipedia articles (4500 samples), and an accuracy of 98.4% on real and *ChatGPT*-generated research abstracts (3000 samples).
The dataset card for the dataset that was created in relation to this project can be found [here](https://huggingface.co/datasets/NicolaiSivesind/human-vs-machine).
**NOTE**: the hosted inference on this site only works for the RoBERTa-models, and not for the Bloomz-models. The Bloomz-models otherwise can produce wrong predictions when not explicitly providing the attention mask from the tokenizer to the model for inference. To be sure, the [pipeline](https://huggingface.co/docs/transformers/main_classes/pipelines)-library seems to produce the most consistent results.
## Fine-tuned detectors
This project includes 12 fine-tuned models based on the RoBERTa-base model, and three sizes of the bloomz-models.
| Base-model | RoBERTa-base | Bloomz-560m | Bloomz-1b7 | Bloomz-3b |
|------------|--------------------------------------------------------------------------------|--------------------------------------------------------------------------------------------|------------------------------------------------------------------------------------------|----------------------------------------------------------------------------------------|
| Wiki | [roberta-wiki](https://huggingface.co/andreas122001/roberta-wiki-detector) | [Bloomz-560m-wiki](https://huggingface.co/andreas122001/bloomz-560m-wiki-detector) | [Bloomz-1b7-wiki](https://huggingface.co/andreas122001/bloomz-1b7-wiki-detector) | [Bloomz-3b-wiki](https://huggingface.co/andreas122001/bloomz-3b-wiki-detector) |
| Academic | [roberta-academic](https://huggingface.co/andreas122001/roberta-academic-detector) | [Bloomz-560m-academic](https://huggingface.co/andreas122001/bloomz-560m-academic-detector) | [Bloomz-1b7-academic](https://huggingface.co/andreas122001/bloomz-1b7-academic-detector) | [Bloomz-3b-academic](https://huggingface.co/andreas122001/bloomz-3b-academic-detector) |
| Mixed | [roberta-mixed](https://huggingface.co/andreas122001/roberta-mixed-detector) | [Bloomz-560m-mixed](https://huggingface.co/andreas122001/bloomz-560m-mixed-detector) | [Bloomz-1b7-mixed](https://huggingface.co/andreas122001/bloomz-1b7-mixed-detector) | [Bloomz-3b-mixed](https://huggingface.co/andreas122001/bloomz-3b-mixed-detector) |
### Datasets
The models were trained on selections from the [GPT-wiki-intros]() and [ChatGPT-Research-Abstracts](), and are separated into three types, **wiki**-detectors, **academic**-detectors and **mixed**-detectors, respectively.
- **Wiki-detectors**:
- Trained on 30'000 datapoints (10%) of GPT-wiki-intros.
- Best model (in-domain) is Bloomz-3b-wiki, with an accuracy of 100%.
- **Academic-detectors**:
- Trained on 20'000 datapoints (100%) of ChatGPT-Research-Abstracts.
- Best model (in-domain) is Bloomz-3b-academic, with an accuracy of 98.4%
- **Mixed-detectors**:
- Trained on 15'000 datapoints (5%) of GPT-wiki-intros and 10'000 datapoints (50%) of ChatGPT-Research-Abstracts.
- Best model (in-domain) is RoBERTa-mixed, with an F1-score of 99.3%.
### Hyperparameters
All models were trained using the same hyperparameters:
```python
{
"num_train_epochs": 1,
"adam_beta1": 0.9,
"adam_beta2": 0.999,
"batch_size": 8,
"adam_epsilon": 1e-08
"optim": "adamw_torch" # the optimizer (AdamW)
"learning_rate": 5e-05, # (LR)
"lr_scheduler_type": "linear", # scheduler type for LR
"seed": 42, # seed for PyTorch RNG-generator.
}
```
### Metrics
Metrics can be found at https://wandb.ai/idatt2900-072/IDATT2900-072.
In-domain performance of wiki-detectors:
| Base model | Accuracy | Precision | Recall | F1-score |
|-------------|----------|-----------|--------|----------|
| Bloomz-560m | 0.973 | *1.000 | 0.945 | 0.972 |
| Bloomz-1b7 | 0.972 | *1.000 | 0.945 | 0.972 |
| Bloomz-3b | *1.000 | *1.000 | *1.000 | *1.000 |
| RoBERTa | 0.998 | 0.999 | 0.997 | 0.998 |
In-domain peformance of academic-detectors:
| Base model | Accuracy | Precision | Recall | F1-score |
|-------------|----------|-----------|--------|----------|
| Bloomz-560m | 0.964 | 0.963 | 0.965 | 0.964 |
| Bloomz-1b7 | 0.946 | 0.941 | 0.951 | 0.946 |
| Bloomz-3b | *0.984 | *0.983 | 0.985 | *0.984 |
| RoBERTa | 0.982 | 0.968 | *0.997 | 0.982 |
F1-scores of the mixed-detectors on all three datasets:
| Base model | Mixed | Wiki | CRA |
|-------------|--------|--------|--------|
| Bloomz-560m | 0.948 | 0.972 | *0.848 |
| Bloomz-1b7 | 0.929 | 0.964 | 0.816 |
| Bloomz-3b | 0.988 | 0.996 | 0.772 |
| RoBERTa | *0.993 | *0.997 | 0.829 |
## Credits
- [GPT-wiki-intro](https://huggingface.co/datasets/aadityaubhat/GPT-wiki-intro), by Aaditya Bhat
- [arxiv-abstracts-2021](https://huggingface.co/datasets/gfissore/arxiv-abstracts-2021), by Giancarlo
- [Bloomz](bigscience/bloomz), by BigScience
- [RoBERTa](https://huggingface.co/roberta-base), by Liu et. al.
## Citation
Please use the following citation:
```
@misc {sivesind_2023,
author = { {Nicolai Thorer Sivesind} and {Andreas Bentzen Winje} },
title = { Machine-generated text-detection by fine-tuning of language models },
url = { https://huggingface.co/andreas122001/roberta-academic-detector }
year = 2023,
publisher = { Hugging Face }
}
``` | {"language": ["en"], "license": "openrail", "tags": ["mgt-detection", "ai-detection"], "datasets": ["NicolaiSivesind/human-vs-machine", "gfissore/arxiv-abstracts-2021"], "widget": [{"text": "I am totally a human, trust me bro.", "example_title": "default"}, {"text": "In Finnish folklore, all places and things, and also human beings, have a haltija (a genius, guardian spirit) of their own. One such haltija is called eti\u00e4inen\u2014an image, doppelg\u00e4nger, or just an impression that goes ahead of a person, doing things the person in question later does. For example, people waiting at home might hear the door close or even see a shadow or a silhouette, only to realize that no one has yet arrived. Eti\u00e4inen can also refer to some kind of a feeling that something is going to happen. Sometimes it could, for example, warn of a bad year coming. In modern Finnish, the term has detached from its shamanistic origins and refers to premonition. Unlike clairvoyance, divination, and similar practices, eti\u00e4iset (plural) are spontaneous and can't be induced. Quite the opposite, they may be unwanted and cause anxiety, like ghosts. Eti\u00e4iset need not be too dramatic and may concern everyday events, although ones related to e.g. deaths are common. As these phenomena are still reported today, they can be considered a living tradition, as a way to explain the psychological experience of premonition.", "example_title": "real wikipedia"}, {"text": "In Finnish folklore, all places and things, animate or inanimate, have a spirit or \"eti\u00e4inen\" that lives there. Eti\u00e4inen can manifest in many forms, but is usually described as a kind, elderly woman with white hair. She is the guardian of natural places and often helps people in need. Eti\u00e4inen has been a part of Finnish culture for centuries and is still widely believed in today. Folklorists study eti\u00e4inen to understand Finnish traditions and how they have changed over time.", "example_title": "generated wikipedia"}, {"text": "This paper presents a novel framework for sparsity-certifying graph decompositions, which are important tools in various areas of computer science, including algorithm design, complexity theory, and optimization. Our approach is based on the concept of \"cut sparsifiers,\" which are sparse graphs that preserve the cut structure of the original graph up to a certain error bound. We show that cut sparsifiers can be efficiently constructed using a combination of spectral techniques and random sampling, and we use them to develop new algorithms for decomposing graphs into sparse subgraphs.", "example_title": "from ChatGPT"}, {"text": "Recent work has demonstrated substantial gains on many NLP tasks and benchmarks by pre-training on a large corpus of text followed by fine-tuning on a specific task. While typically task-agnostic in architecture, this method still requires task-specific fine-tuning datasets of thousands or tens of thousands of examples. By contrast, humans can generally perform a new language task from only a few examples or from simple instructions - something which current NLP systems still largely struggle to do. Here we show that scaling up language models greatly improves task-agnostic, few-shot performance, sometimes even reaching competitiveness with prior state-of-the-art fine-tuning approaches. Specifically, we train GPT-3, an autoregressive language model with 175 billion parameters, 10x more than any previous non-sparse language model, and test its performance in the few-shot setting. For all tasks, GPT-3 is applied without any gradient updates or fine-tuning, with tasks and few-shot demonstrations specified purely via text interaction with the model. GPT-3 achieves strong performance on many NLP datasets, including translation, question-answering, and cloze tasks, as well as several tasks that require on-the-fly reasoning or domain adaptation, such as unscrambling words, using a novel word in a sentence, or performing 3-digit arithmetic. At the same time, we also identify some datasets where GPT-3's few-shot learning still struggles, as well as some datasets where GPT-3 faces methodological issues related to training on large web corpora. Finally, we find that GPT-3 can generate samples of news articles which human evaluators have difficulty distinguishing from articles written by humans. We discuss broader societal impacts of this finding and of GPT-3 in general.", "example_title": "GPT-3 paper"}], "pipeline_tag": "text-classification"} | text-classification | hossamdaoud/bloomz-1b7-academic-detector | [
"transformers",
"pytorch",
"safetensors",
"bloom",
"text-classification",
"mgt-detection",
"ai-detection",
"en",
"dataset:NicolaiSivesind/human-vs-machine",
"dataset:gfissore/arxiv-abstracts-2021",
"license:openrail",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-07T17:36:09+00:00 | [] | [
"en"
] | TAGS
#transformers #pytorch #safetensors #bloom #text-classification #mgt-detection #ai-detection #en #dataset-NicolaiSivesind/human-vs-machine #dataset-gfissore/arxiv-abstracts-2021 #license-openrail #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| Machine-generated text-detection by fine-tuning of language models
==================================================================
This project is related to a bachelor's thesis with the title "*Turning Poachers into Gamekeepers: Detecting Machine-Generated Text in Academia using Large Language Models*" (see here) written by *Nicolai Thorer Sivesind* and *Andreas Bentzen Winje* at the *Department of Computer Science* at the *Norwegian University of Science and Technology*.
It contains text classification models trained to distinguish human-written text from text generated by language models like ChatGPT and GPT-3. The best models were able to achieve an accuracy of 100% on real and *GPT-3*-generated wikipedia articles (4500 samples), and an accuracy of 98.4% on real and *ChatGPT*-generated research abstracts (3000 samples).
The dataset card for the dataset that was created in relation to this project can be found here.
NOTE: the hosted inference on this site only works for the RoBERTa-models, and not for the Bloomz-models. The Bloomz-models otherwise can produce wrong predictions when not explicitly providing the attention mask from the tokenizer to the model for inference. To be sure, the pipeline-library seems to produce the most consistent results.
Fine-tuned detectors
--------------------
This project includes 12 fine-tuned models based on the RoBERTa-base model, and three sizes of the bloomz-models.
### Datasets
The models were trained on selections from the GPT-wiki-intros and ChatGPT-Research-Abstracts, and are separated into three types, wiki-detectors, academic-detectors and mixed-detectors, respectively.
* Wiki-detectors:
+ Trained on 30'000 datapoints (10%) of GPT-wiki-intros.
+ Best model (in-domain) is Bloomz-3b-wiki, with an accuracy of 100%.
* Academic-detectors:
+ Trained on 20'000 datapoints (100%) of ChatGPT-Research-Abstracts.
+ Best model (in-domain) is Bloomz-3b-academic, with an accuracy of 98.4%
* Mixed-detectors:
+ Trained on 15'000 datapoints (5%) of GPT-wiki-intros and 10'000 datapoints (50%) of ChatGPT-Research-Abstracts.
+ Best model (in-domain) is RoBERTa-mixed, with an F1-score of 99.3%.
### Hyperparameters
All models were trained using the same hyperparameters:
### Metrics
Metrics can be found at URL
In-domain performance of wiki-detectors:
In-domain peformance of academic-detectors:
F1-scores of the mixed-detectors on all three datasets:
Credits
-------
* GPT-wiki-intro, by Aaditya Bhat
* arxiv-abstracts-2021, by Giancarlo
* Bloomz, by BigScience
* RoBERTa, by Liu et. al.
Please use the following citation:
| [
"### Datasets\n\n\nThe models were trained on selections from the GPT-wiki-intros and ChatGPT-Research-Abstracts, and are separated into three types, wiki-detectors, academic-detectors and mixed-detectors, respectively.\n\n\n* Wiki-detectors:\n\t+ Trained on 30'000 datapoints (10%) of GPT-wiki-intros.\n\t+ Best model (in-domain) is Bloomz-3b-wiki, with an accuracy of 100%.\n* Academic-detectors:\n\t+ Trained on 20'000 datapoints (100%) of ChatGPT-Research-Abstracts.\n\t+ Best model (in-domain) is Bloomz-3b-academic, with an accuracy of 98.4%\n* Mixed-detectors:\n\t+ Trained on 15'000 datapoints (5%) of GPT-wiki-intros and 10'000 datapoints (50%) of ChatGPT-Research-Abstracts.\n\t+ Best model (in-domain) is RoBERTa-mixed, with an F1-score of 99.3%.",
"### Hyperparameters\n\n\nAll models were trained using the same hyperparameters:",
"### Metrics\n\n\nMetrics can be found at URL\n\n\nIn-domain performance of wiki-detectors:\n\n\n\nIn-domain peformance of academic-detectors:\n\n\n\nF1-scores of the mixed-detectors on all three datasets:\n\n\n\nCredits\n-------\n\n\n* GPT-wiki-intro, by Aaditya Bhat\n* arxiv-abstracts-2021, by Giancarlo\n* Bloomz, by BigScience\n* RoBERTa, by Liu et. al.\n\n\nPlease use the following citation:"
] | [
"TAGS\n#transformers #pytorch #safetensors #bloom #text-classification #mgt-detection #ai-detection #en #dataset-NicolaiSivesind/human-vs-machine #dataset-gfissore/arxiv-abstracts-2021 #license-openrail #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Datasets\n\n\nThe models were trained on selections from the GPT-wiki-intros and ChatGPT-Research-Abstracts, and are separated into three types, wiki-detectors, academic-detectors and mixed-detectors, respectively.\n\n\n* Wiki-detectors:\n\t+ Trained on 30'000 datapoints (10%) of GPT-wiki-intros.\n\t+ Best model (in-domain) is Bloomz-3b-wiki, with an accuracy of 100%.\n* Academic-detectors:\n\t+ Trained on 20'000 datapoints (100%) of ChatGPT-Research-Abstracts.\n\t+ Best model (in-domain) is Bloomz-3b-academic, with an accuracy of 98.4%\n* Mixed-detectors:\n\t+ Trained on 15'000 datapoints (5%) of GPT-wiki-intros and 10'000 datapoints (50%) of ChatGPT-Research-Abstracts.\n\t+ Best model (in-domain) is RoBERTa-mixed, with an F1-score of 99.3%.",
"### Hyperparameters\n\n\nAll models were trained using the same hyperparameters:",
"### Metrics\n\n\nMetrics can be found at URL\n\n\nIn-domain performance of wiki-detectors:\n\n\n\nIn-domain peformance of academic-detectors:\n\n\n\nF1-scores of the mixed-detectors on all three datasets:\n\n\n\nCredits\n-------\n\n\n* GPT-wiki-intro, by Aaditya Bhat\n* arxiv-abstracts-2021, by Giancarlo\n* Bloomz, by BigScience\n* RoBERTa, by Liu et. al.\n\n\nPlease use the following citation:"
] | [
104,
259,
19,
116
] | [
"passage: TAGS\n#transformers #pytorch #safetensors #bloom #text-classification #mgt-detection #ai-detection #en #dataset-NicolaiSivesind/human-vs-machine #dataset-gfissore/arxiv-abstracts-2021 #license-openrail #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Datasets\n\n\nThe models were trained on selections from the GPT-wiki-intros and ChatGPT-Research-Abstracts, and are separated into three types, wiki-detectors, academic-detectors and mixed-detectors, respectively.\n\n\n* Wiki-detectors:\n\t+ Trained on 30'000 datapoints (10%) of GPT-wiki-intros.\n\t+ Best model (in-domain) is Bloomz-3b-wiki, with an accuracy of 100%.\n* Academic-detectors:\n\t+ Trained on 20'000 datapoints (100%) of ChatGPT-Research-Abstracts.\n\t+ Best model (in-domain) is Bloomz-3b-academic, with an accuracy of 98.4%\n* Mixed-detectors:\n\t+ Trained on 15'000 datapoints (5%) of GPT-wiki-intros and 10'000 datapoints (50%) of ChatGPT-Research-Abstracts.\n\t+ Best model (in-domain) is RoBERTa-mixed, with an F1-score of 99.3%.### Hyperparameters\n\n\nAll models were trained using the same hyperparameters:### Metrics\n\n\nMetrics can be found at URL\n\n\nIn-domain performance of wiki-detectors:\n\n\n\nIn-domain peformance of academic-detectors:\n\n\n\nF1-scores of the mixed-detectors on all three datasets:\n\n\n\nCredits\n-------\n\n\n* GPT-wiki-intro, by Aaditya Bhat\n* arxiv-abstracts-2021, by Giancarlo\n* Bloomz, by BigScience\n* RoBERTa, by Liu et. al.\n\n\nPlease use the following citation:"
] | [
-0.058261554688215256,
0.07812947779893875,
-0.008540788665413857,
0.09696026891469955,
0.050334632396698,
-0.003122422145679593,
0.08197064697742462,
0.07499383389949799,
0.07698243856430054,
0.12387566268444061,
-0.0355805978178978,
0.07263767719268799,
0.05357050150632858,
0.14380301535129547,
-0.016693193465471268,
-0.16814114153385162,
0.04725952073931694,
-0.03734816610813141,
-0.02316729724407196,
0.0825607106089592,
0.12249801307916641,
-0.07916221022605896,
0.022898118942975998,
-0.023357024416327477,
-0.04011571407318115,
-0.017615599557757378,
0.01932782493531704,
-0.06758521497249603,
0.09468640387058258,
0.060365814715623856,
0.12263192236423492,
0.0014817173359915614,
0.01785627193748951,
-0.18764035403728485,
0.01942959427833557,
0.07348792999982834,
-0.0040369899943470955,
0.0673820972442627,
0.08292112499475479,
-0.024567538872361183,
0.06854498386383057,
-0.10227259248495102,
0.11192473024129868,
0.03823835402727127,
-0.13566915690898895,
-0.12570948898792267,
-0.1734941005706787,
0.06273075193166733,
0.1414591521024704,
0.05129270255565643,
-0.01985282450914383,
0.03582354635000229,
-0.03432517126202583,
0.06915237754583359,
0.09611190110445023,
-0.2582515776157379,
-0.04122629389166832,
-0.04686801880598068,
0.01015329547226429,
0.031168287619948387,
-0.0957658514380455,
-0.017719531431794167,
-0.046356555074453354,
0.029447974637150764,
0.11337887495756149,
-0.023471588268876076,
0.10782880336046219,
-0.030742812901735306,
-0.1141481027007103,
-0.019408591091632843,
0.06300696730613708,
0.029490478336811066,
-0.06347768753767014,
-0.16467873752117157,
-0.045756708830595016,
-0.003195079741999507,
-0.03421369194984436,
-0.06397085636854172,
0.025986839085817337,
-0.010957041755318642,
0.06907393038272858,
-0.02597317472100258,
-0.0735013410449028,
0.03779199719429016,
-0.10700717568397522,
0.1189081072807312,
0.033880483359098434,
0.005613246001303196,
0.047730278223752975,
0.0654856264591217,
0.06233806535601616,
-0.07445831596851349,
-0.022477293387055397,
-0.048291221261024475,
-0.13704614341259003,
-0.05937817320227623,
-0.003975706174969673,
-0.029758570715785027,
0.024625949561595917,
0.20530523359775543,
-0.10192732512950897,
0.027024725452065468,
-0.0333988219499588,
-0.015585397370159626,
0.045615583658218384,
0.11263400316238403,
-0.03832829371094704,
-0.07155071198940277,
0.05313587933778763,
0.0009824682492762804,
0.017963606864213943,
0.01631232723593712,
0.03369894623756409,
0.0010544476099312305,
0.0421021543443203,
0.054196421056985855,
0.024033382534980774,
0.004289626143872738,
-0.07761633396148682,
-0.04049242287874222,
0.09365987032651901,
-0.16777190566062927,
0.04491211846470833,
0.02603060193359852,
0.0027275229804217815,
0.07451362907886505,
0.012128427624702454,
-0.014834254048764706,
-0.05590634047985077,
0.1088402271270752,
-0.0343468151986599,
-0.010169458575546741,
-0.008336120285093784,
-0.07967136055231094,
0.031051337718963623,
-0.03668126091361046,
-0.07140816003084183,
-0.0822317972779274,
-0.05203233286738396,
-0.07762402296066284,
0.04021947830915451,
-0.04726225137710571,
0.012207777239382267,
0.011994178406894207,
0.02004551887512207,
0.029652686789631844,
0.0006227658013813198,
-0.010534921661019325,
-0.03113233484327793,
0.02107935957610607,
-0.03910308703780174,
0.03347575291991234,
0.01037472952157259,
0.03220606967806816,
-0.03283972665667534,
0.05129576474428177,
-0.2377428263425827,
0.10269156098365784,
-0.10700882226228714,
0.0472087636590004,
-0.1552683413028717,
-0.02450580894947052,
0.03516402840614319,
0.0006996203446760774,
0.0346454493701458,
0.07797648757696152,
-0.11439543217420578,
-0.058323830366134644,
0.17010849714279175,
-0.09707130491733551,
-0.06039394065737724,
0.10114649683237076,
-0.05789351835846901,
0.02593848668038845,
0.09724055975675583,
0.09119877964258194,
0.019265174865722656,
-0.07820066809654236,
-0.07737869024276733,
-0.10715878009796143,
0.05722995847463608,
0.12335854768753052,
0.049881234765052795,
-0.056006066501140594,
0.07486404478549957,
0.008346528746187687,
-0.009989168494939804,
0.014922600239515305,
0.03156377002596855,
-0.06816354393959045,
-0.009963703341782093,
-0.04254854470491409,
0.024491824209690094,
-0.03389105945825577,
0.026495549827814102,
-0.026530776172876358,
-0.09974447637796402,
0.016653934493660927,
0.0653197392821312,
-0.02159830369055271,
0.05420151352882385,
-0.1032964289188385,
0.0762147381901741,
0.0331907644867897,
0.010715574957430363,
-0.16247671842575073,
-0.10910527408123016,
0.06272463500499725,
-0.06878071278333664,
0.020719097927212715,
-0.09307441115379333,
0.04478500410914421,
0.03702213987708092,
-0.062291186302900314,
-0.0020278547890484333,
0.011217725463211536,
0.009114999324083328,
-0.06741761416196823,
-0.12571106851100922,
-0.02489415556192398,
-0.05304010957479477,
0.08761415630578995,
-0.10773984342813492,
-0.017436034977436066,
0.0681130513548851,
0.16341263055801392,
0.06287631392478943,
-0.04984134063124657,
0.002221314935013652,
0.026421938091516495,
0.008161403238773346,
-0.04941367730498314,
0.008662884123623371,
-0.02173667587339878,
-0.02137572504580021,
-0.005690211430191994,
-0.059113021939992905,
-0.03451356664299965,
0.06054465472698212,
0.12008108198642731,
-0.09468982368707657,
-0.08554869145154953,
-0.04531971737742424,
0.006396172568202019,
-0.0212192814797163,
-0.04608398675918579,
0.1603315770626068,
0.0740920901298523,
0.07035791128873825,
-0.027805229648947716,
-0.06876184046268463,
-0.03054843097925186,
0.05143822729587555,
-0.043433044105768204,
0.09948034584522247,
-0.03899887949228287,
-0.1845807284116745,
0.04281060770153999,
0.0923507809638977,
0.05466549098491669,
0.023478670045733452,
-0.05749793350696564,
-0.07515688240528107,
-0.06913135200738907,
0.010027719661593437,
0.039956122636795044,
0.04276342689990997,
0.02492458000779152,
0.019040588289499283,
0.060238067060709,
-0.0037996647879481316,
-0.017759693786501884,
-0.07057925313711166,
0.04213150963187218,
0.012739074416458607,
-0.0004198667302262038,
0.08139539510011673,
-0.016750263050198555,
0.02302594855427742,
0.10727895051240921,
0.028506027534604073,
-0.020189151167869568,
-0.015742607414722443,
-0.0365668423473835,
-0.07011942565441132,
0.11314474791288376,
-0.1158333569765091,
-0.1861162781715393,
-0.10530151426792145,
-0.06105195730924606,
-0.0339529812335968,
-0.020179182291030884,
-0.0014321889029815793,
-0.030320441350340843,
-0.09580041468143463,
-0.09669698029756546,
0.006490661762654781,
0.08941657841205597,
-0.04730803146958351,
0.037884462624788284,
0.02069113403558731,
0.007060764357447624,
-0.09717143326997757,
-0.019743872806429863,
-0.005528729408979416,
-0.03240179270505905,
0.024540111422538757,
0.06875451654195786,
0.03406738117337227,
0.08499214053153992,
0.05629146099090576,
-0.004302854184061289,
-0.016727212816476822,
0.21829211711883545,
-0.1259523332118988,
0.11964550614356995,
0.10803613066673279,
-0.030457554385066032,
0.043724656105041504,
0.1545272171497345,
0.04344582185149193,
-0.054195426404476166,
0.0067183771170675755,
0.04933018609881401,
0.018833229318261147,
-0.284661203622818,
-0.07225503027439117,
-0.052705757319927216,
-0.08635859936475754,
0.06993182003498077,
0.06387511640787125,
-0.021888481453061104,
0.0015029044589027762,
-0.04076799005270004,
0.00012239694478921592,
0.08569502085447311,
0.03706800192594528,
0.1829826533794403,
0.016569538041949272,
0.08796166628599167,
-0.04601119086146355,
-0.021733751520514488,
0.07629834115505219,
0.009644940495491028,
0.19068440794944763,
-0.008577942848205566,
0.13315349817276,
0.04768243432044983,
0.0367136150598526,
0.005257289856672287,
0.03598759323358536,
-0.0054725222289562225,
0.04857519641518593,
0.004311883822083473,
-0.04599509388208389,
0.03947838395833969,
0.05490579456090927,
0.010045748203992844,
-0.07640989124774933,
0.01478290930390358,
-0.07239588350057602,
0.10014065355062485,
0.27074286341667175,
0.03730392828583717,
-0.11313904076814651,
-0.09699524194002151,
0.04875345155596733,
-0.12025537341833115,
-0.056710921227931976,
-0.02614540420472622,
0.00017006922280415893,
-0.1154787614941597,
0.04289235547184944,
-0.03414914011955261,
0.04870924353599548,
-0.12849842011928558,
-0.02965211309492588,
0.03789253160357475,
0.04626151919364929,
-0.017657531425356865,
0.0034931290429085493,
-0.20930634438991547,
0.09490315616130829,
0.0056856474839150906,
0.07345825433731079,
-0.009072483517229557,
0.08173614740371704,
0.027893757447600365,
-0.021751083433628082,
0.1149904876947403,
0.01963273249566555,
-0.12137490510940552,
-0.09019256383180618,
-0.096014603972435,
-0.002905911998823285,
0.07473692297935486,
-0.10148022323846817,
0.08126746863126755,
-0.009273174218833447,
-0.02671331726014614,
-0.009430671110749245,
0.0008966111927293241,
-0.09458307921886444,
-0.17582793533802032,
0.08577436208724976,
-0.07460708916187286,
0.028564536944031715,
-0.054650694131851196,
-0.048526741564273834,
-0.14671827852725983,
0.14769402146339417,
-0.11135910451412201,
-0.08198672533035278,
-0.12838518619537354,
-0.05532609671354294,
0.14435525238513947,
-0.06825923174619675,
0.10240627825260162,
0.006454591639339924,
0.08381710946559906,
-0.03467358648777008,
-0.056518614292144775,
0.0715118870139122,
-0.03401348739862442,
-0.15315814316272736,
-0.05347488075494766,
0.1317123919725418,
0.03237871825695038,
0.0345938615500927,
0.017470333725214005,
0.09522463381290436,
0.052913978695869446,
-0.05859273672103882,
0.06728214770555496,
0.13522088527679443,
0.007637408096343279,
0.02776084840297699,
-0.11252302676439285,
-0.08364002406597137,
-0.11344104260206223,
-0.05169656500220299,
0.16293638944625854,
0.24247653782367706,
-0.06757788360118866,
0.13048958778381348,
0.08221063017845154,
-0.09282049536705017,
-0.15548668801784515,
-0.01263671275228262,
0.018931619822978973,
-0.010580907575786114,
-0.026493048295378685,
-0.2283434271812439,
0.11886834353208542,
0.12073919177055359,
-0.02054607681930065,
0.08202549070119858,
-0.15092052519321442,
-0.1277332603931427,
0.06941965967416763,
0.04677227884531021,
-0.04317835345864296,
-0.12526459991931915,
-0.07978115975856781,
-0.05811312049627304,
-0.09959529340267181,
0.13576237857341766,
0.035478103905916214,
0.07080800086259842,
-0.011384570971131325,
-0.017303211614489555,
0.010228076949715614,
-0.016268784180283546,
0.09972041100263596,
0.003544009756296873,
0.05328020080924034,
-0.05364116653800011,
-0.010459140874445438,
0.05168687552213669,
-0.04208802431821823,
0.040118325501680374,
0.05196800082921982,
-0.0035598473623394966,
-0.07334060966968536,
-0.04248390346765518,
-0.06965382397174835,
0.02527456171810627,
-0.07068205624818802,
-0.03297138959169388,
-0.07373266667127609,
0.12831038236618042,
0.10263919085264206,
-0.03512393310666084,
0.006549341604113579,
0.018841948360204697,
0.05522095412015915,
0.07382123917341232,
0.07371112704277039,
0.0687384232878685,
-0.11639852821826935,
-0.004392826464027166,
-0.033130429685115814,
0.010835225693881512,
-0.09349078685045242,
0.06739895790815353,
0.12801098823547363,
0.0010056180180981755,
0.13500411808490753,
-0.0011441911337897182,
-0.12380610406398773,
-0.011173215694725513,
0.06472114473581314,
-0.10083562880754471,
-0.15857672691345215,
0.030150726437568665,
-0.027510249987244606,
-0.08078013360500336,
-0.00490347295999527,
0.17009875178337097,
0.03989333659410477,
-0.06184983626008034,
0.03207237273454666,
0.08625836670398712,
0.012578763999044895,
0.11793800443410873,
-0.0037471484392881393,
0.016149727627635002,
-0.06788402050733566,
0.08374006301164627,
0.14255794882774353,
-0.09155368059873581,
0.01209244504570961,
0.1348160058259964,
-0.06298432499170303,
-0.05680864304304123,
-0.03627822548151016,
0.00786649901419878,
0.004410441964864731,
0.054279010742902756,
-0.02059040404856205,
-0.14872460067272186,
0.04734698683023453,
0.14513100683689117,
-0.020666692405939102,
0.048526275902986526,
-0.011607414111495018,
-0.02617420256137848,
-0.03101413883268833,
0.10255225002765656,
0.01745743118226528,
0.06130106374621391,
-0.025219207629561424,
0.05260312557220459,
-0.061443280428647995,
-0.04658392071723938,
0.006399169564247131,
-0.01629401184618473,
-0.10553522408008575,
-0.030160995200276375,
-0.14518967270851135,
0.013134163804352283,
-0.06817875802516937,
0.00007559260848211125,
-0.03533431887626648,
0.0033801135141402483,
0.00964908953756094,
0.014278757385909557,
-0.052094314247369766,
-0.02651781216263771,
-0.028883343562483788,
0.09483631700277328,
-0.13149037957191467,
-0.01856689155101776,
0.05921003594994545,
-0.10238145291805267,
0.15304818749427795,
-0.009397361427545547,
0.0016719838604331017,
-0.0009626516257412732,
-0.1290140300989151,
0.005114714149385691,
-0.02857062965631485,
-0.0006797083769924939,
0.037338584661483765,
-0.20646460354328156,
0.030837567523121834,
-0.02013975940644741,
-0.04893556237220764,
0.04548493027687073,
0.06129549443721771,
-0.09308120608329773,
-0.008244593627750874,
0.0010771609377115965,
-0.04736626520752907,
-0.06415347009897232,
0.0036933491937816143,
0.12885834276676178,
0.008290405385196209,
0.13004155457019806,
-0.055606428533792496,
0.06085892394185066,
-0.14899487793445587,
-0.013975000008940697,
-0.0007419667090289295,
-0.05131550133228302,
0.025280682370066643,
-0.05140925571322441,
0.049827031791210175,
-0.011931461282074451,
0.12934288382530212,
-0.04052938148379326,
-0.07486113160848618,
0.06706642359495163,
-0.0693713054060936,
-0.09025219827890396,
0.07888058573007584,
0.06250916421413422,
0.027734458446502686,
-0.05212179571390152,
0.013051965273916721,
-0.06873738765716553,
-0.02443874068558216,
0.11906541138887405,
0.09169655293226242,
0.2360939085483551,
0.09356957674026489,
0.002318494487553835,
0.04848000034689903,
-0.06495226919651031,
-0.009127762168645859,
0.06765298545360565,
-0.0844615027308464,
0.0711221918463707,
-0.0497896634042263,
0.11570167541503906,
-0.007006573490798473,
-0.19418743252754211,
0.10930517315864563,
-0.05682427063584328,
-0.0738101527094841,
-0.039179544895887375,
-0.11177653819322586,
-0.0551009364426136,
0.02714560739696026,
0.002939541358500719,
-0.10414346307516098,
0.002883970271795988,
0.0466245673596859,
0.03358565643429756,
-0.022351544350385666,
0.06470885872840881,
-0.03136609122157097,
-0.042063720524311066,
0.16364258527755737,
0.03861882910132408,
-0.04256686940789223,
0.04215884581208229,
-0.03330887481570244,
-0.05170321837067604,
0.03534764051437378,
0.031987264752388,
0.046776946634054184,
0.001868014340288937,
0.0012394103687256575,
-0.024894660338759422,
-0.07502297312021255,
0.022308720275759697,
-0.005351828411221504,
-0.019262563437223434,
0.14991624653339386,
0.06299008429050446,
-0.016374096274375916,
0.001192821771837771,
0.15746277570724487,
0.00910487025976181,
-0.033674098551273346,
-0.1466401219367981,
0.12259120494127274,
0.0016424726927652955,
-0.0013404380297288299,
0.03524361550807953,
-0.10229983180761337,
0.04501164332032204,
0.10996119678020477,
0.1076415553689003,
-0.050491880625486374,
0.0233446191996336,
0.016276879236102104,
0.0035255905240774155,
0.02927689626812935,
0.08411353826522827,
0.04604591429233551,
0.14682820439338684,
-0.044750016182661057,
0.07672437280416489,
-0.011277999728918076,
-0.05834679305553436,
-0.056405846029520035,
0.14994366466999054,
-0.020561490207910538,
0.033637311309576035,
-0.0706593319773674,
0.08169067651033401,
-0.06340986490249634,
-0.20185521245002747,
0.06287238746881485,
-0.12348153442144394,
-0.16703513264656067,
-0.016027385368943214,
-0.004243459086865187,
-0.00023535841319244355,
0.04829766973853111,
0.05847389996051788,
-0.03508121520280838,
0.12144415080547333,
0.009342717006802559,
0.02030782774090767,
-0.02419992908835411,
0.12594620883464813,
-0.07363760471343994,
0.17077520489692688,
0.030474776402115822,
0.07763376832008362,
0.10653343796730042,
0.016900302842259407,
-0.11623100936412811,
0.026664089411497116,
0.03347984328866005,
-0.07425456494092941,
0.03841988369822502,
0.1413557082414627,
0.01612410694360733,
0.09316198527812958,
0.10587458312511444,
0.040057890117168427,
0.07655647397041321,
-0.0003595459565985948,
-0.09278818219900131,
-0.07487861067056656,
0.04550402984023094,
-0.10775436460971832,
0.11948978900909424,
0.20363128185272217,
-0.04412936046719551,
0.03486788272857666,
-0.05432967469096184,
0.0011084751458838582,
0.006401201710104942,
0.05943961441516876,
-0.06235413998365402,
-0.19171427190303802,
0.08181170374155045,
0.052387114614248276,
0.09804556518793106,
-0.17326857149600983,
-0.1207183226943016,
-0.007472981233149767,
-0.02031242661178112,
-0.0851549506187439,
0.14553315937519073,
0.1178513616323471,
0.031115368008613586,
-0.019028808921575546,
-0.07590816915035248,
0.0010407204972580075,
0.06882724165916443,
-0.06201077625155449,
0.019374124705791473
] |
null | null | fastai |
# Amazing!
🥳 Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the [documentation here](https://huggingface.co/docs/hub/model-repos))!
2. Create a demo in Gradio or Streamlit using 🤗 Spaces ([documentation here](https://huggingface.co/docs/hub/spaces)).
3. Join the fastai community on the [Fastai Discord](https://discord.com/invite/YKrxeNn)!
Greetings fellow fastlearner 🤝! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| {"tags": ["fastai"]} | null | Hitomiblood/intel-image-classification | [
"fastai",
"has_space",
"region:us"
] | 2024-02-07T17:37:52+00:00 | [] | [] | TAGS
#fastai #has_space #region-us
|
# Amazing!
Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the documentation here)!
2. Create a demo in Gradio or Streamlit using Spaces (documentation here).
3. Join the fastai community on the Fastai Discord!
Greetings fellow fastlearner ! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| [
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
"TAGS\n#fastai #has_space #region-us \n",
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
13,
20,
79,
3,
6,
12,
8
] | [
"passage: TAGS\n#fastai #has_space #region-us \n# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---# Model card## Model description\nMore information needed## Intended uses & limitations\nMore information needed## Training and evaluation data\nMore information needed"
] | [
-0.048121724277734756,
-0.024616125971078873,
0.002038001548498869,
0.10439170897006989,
0.135872021317482,
0.11887997388839722,
0.07405775785446167,
0.09980081021785736,
0.07783667743206024,
0.02590852417051792,
0.08961158245801926,
-0.08088712394237518,
0.08744348585605621,
0.271692156791687,
0.06988707184791565,
-0.22761479020118713,
0.04051019623875618,
-0.00024903909070417285,
0.08053462207317352,
0.06629016250371933,
0.13507555425167084,
-0.05464952811598778,
0.14010503888130188,
-0.004088983871042728,
-0.19050447642803192,
-0.042929794639348984,
-0.01773718371987343,
-0.02527874894440174,
0.12317648530006409,
-0.04744937643408775,
0.05381017178297043,
0.015037551522254944,
0.007565062493085861,
-0.07253646105527878,
0.0623294934630394,
0.040457066148519516,
0.01740180514752865,
0.059235580265522,
-0.07249044626951218,
0.08950132131576538,
0.08404164761304855,
-0.024370938539505005,
-0.1097978875041008,
0.07827875018119812,
-0.14424212276935577,
-0.21762843430042267,
-0.1253085881471634,
-0.09017651528120041,
0.028519365936517715,
0.004388005938380957,
-0.025051530450582504,
0.12801909446716309,
-0.13558274507522583,
-0.040698226541280746,
0.20124278962612152,
-0.17012301087379456,
-0.05505548417568207,
0.034343402832746506,
0.09226689487695694,
-0.05829555168747902,
-0.06347129493951797,
0.10614984482526779,
0.09640881419181824,
-0.019833475351333618,
0.05516824126243591,
0.002579754451289773,
0.021173657849431038,
0.01370104867964983,
-0.06150497496128082,
0.04717832803726196,
-0.010183089412748814,
0.048132527619600296,
-0.09465572983026505,
-0.1303568333387375,
-0.004072192590683699,
0.01214400865137577,
-0.048744890838861465,
-0.07019646465778351,
0.07833103090524673,
-0.011118141002953053,
-0.04357248544692993,
-0.13031910359859467,
-0.09131011366844177,
-0.12358787655830383,
0.008646543137729168,
0.09500427544116974,
0.003679296001791954,
0.07374339550733566,
-0.08258994668722153,
0.06774985045194626,
-0.17329485714435577,
-0.06484591960906982,
-0.08138520270586014,
-0.11546400189399719,
0.021133482456207275,
-0.0387684591114521,
0.02668963186442852,
0.15394504368305206,
0.12983950972557068,
0.023976242169737816,
0.04388163983821869,
-0.038937073200941086,
0.051190316677093506,
0.058571770787239075,
0.03395717963576317,
0.034934818744659424,
-0.036981891840696335,
-0.1793210655450821,
-0.016702448949217796,
-0.011550825089216232,
0.07954040914773941,
-0.07523109763860703,
-0.05632320046424866,
0.013454885222017765,
-0.11071494966745377,
0.07202339172363281,
-0.03576776012778282,
-0.0032025426626205444,
0.01168301422148943,
0.018371861428022385,
0.21271461248397827,
0.03955606371164322,
0.014191740192472935,
-0.008875265717506409,
-0.13453757762908936,
-0.06874168664216995,
-0.06896194815635681,
0.03361047804355621,
0.04448792710900307,
-0.0028071461711078882,
-0.07672245055437088,
0.04325154796242714,
-0.06045534089207649,
-0.03508453071117401,
0.008032378740608692,
-0.18221288919448853,
0.007458044681698084,
-0.10049355030059814,
-0.12126200646162033,
0.05306628718972206,
0.01695440337061882,
-0.08215925842523575,
0.08141279965639114,
0.02662261202931404,
0.020931517705321312,
-0.009988143108785152,
-0.005391082260757685,
0.06874798238277435,
-0.08508864045143127,
0.029901226982474327,
0.17170792818069458,
0.13024519383907318,
-0.08046911656856537,
-0.0006887061172164977,
-0.10965746641159058,
0.04426072910428047,
-0.13325683772563934,
0.02251482754945755,
-0.09062390774488449,
0.11723794043064117,
-0.042396437376737595,
0.002038756385445595,
-0.029030200093984604,
0.0960269495844841,
0.08189879357814789,
0.16663365066051483,
-0.2419009804725647,
-0.031095001846551895,
0.13240347802639008,
-0.10711425542831421,
-0.1807439625263214,
0.18486657738685608,
-0.012035200372338295,
0.11329247802495956,
-0.047014184296131134,
0.18334640562534332,
-0.02612062357366085,
-0.13582459092140198,
-0.058872904628515244,
0.005852419883012772,
-0.2269321084022522,
-0.06286033242940903,
0.09738040715456009,
0.13425657153129578,
-0.042984943836927414,
0.007112155202776194,
0.026316028088331223,
0.13609857857227325,
-0.06715573370456696,
-0.05195777863264084,
-0.012255736626684666,
-0.10902371257543564,
0.041914235800504684,
0.018215661868453026,
0.035408079624176025,
-0.059880174696445465,
-0.02931194379925728,
-0.053190283477306366,
0.13146710395812988,
0.09760832786560059,
-0.03670211136341095,
-0.049620725214481354,
0.1689043790102005,
-0.07763876020908356,
-0.033587727695703506,
0.07560533285140991,
-0.08268500119447708,
0.03266897425055504,
0.03090597130358219,
0.055881720036268234,
0.07766123116016388,
0.08522116392850876,
0.06057543307542801,
0.00819048099219799,
0.034654274582862854,
0.12095347046852112,
-0.013591280207037926,
-0.05039411783218384,
0.021508218720555305,
0.016904234886169434,
-0.019032588228583336,
0.29030677676200867,
-0.1951042115688324,
0.024724548682570457,
-0.06477324664592743,
0.07631538063287735,
0.06136792525649071,
0.003575638635084033,
0.08580143749713898,
-0.06023019179701805,
-0.019061198458075523,
-0.04803973436355591,
0.046805646270513535,
-0.0666879191994667,
-0.04162997007369995,
0.2621194124221802,
-0.05497581139206886,
0.044914912432432175,
0.12313763797283173,
-0.05873025581240654,
-0.07091446220874786,
0.01009807363152504,
-0.00793424155563116,
0.03249288722872734,
-0.04042816907167435,
0.043721720576286316,
-0.10840129852294922,
-0.06674089282751083,
0.1573198139667511,
-0.038477856665849686,
0.06786153465509415,
0.032288823276758194,
-0.04958454892039299,
-0.0648743286728859,
0.04650486260652542,
0.13598160445690155,
-0.0875244215130806,
0.07435166835784912,
0.17612984776496887,
-0.010562662966549397,
0.168031245470047,
0.08435525000095367,
-0.07075224816799164,
-0.09465329349040985,
-0.051014289259910583,
-0.021595727652311325,
0.21222901344299316,
-0.07084725052118301,
-0.054564714431762695,
0.05911700800061226,
-0.013703816570341587,
0.07196151465177536,
-0.06009222939610481,
-0.08332337439060211,
0.03227344527840614,
-0.04517695680260658,
0.011517706327140331,
0.13512636721134186,
-0.07090822607278824,
0.04681389778852463,
0.031489867717027664,
-0.0662703812122345,
0.02217509225010872,
0.033389873802661896,
0.0068921963684260845,
0.033959709107875824,
0.07332495599985123,
-0.20893315970897675,
-0.08408680558204651,
-0.13727638125419617,
0.037881869822740555,
0.021770721301436424,
0.045787326991558075,
-0.08602345734834671,
0.02231026627123356,
-0.08954031765460968,
-0.07987114042043686,
0.029592275619506836,
-0.026350297033786774,
-0.11349643021821976,
-0.03396226093173027,
-0.009560913778841496,
-0.06662604957818985,
-0.02250705659389496,
-0.05024505779147148,
0.03983384370803833,
0.04479299485683441,
0.058377087116241455,
0.12796473503112793,
-0.013808943331241608,
-0.03839317709207535,
0.000370211957488209,
-0.022712308913469315,
0.16396735608577728,
-0.14746315777301788,
0.07954913377761841,
0.19160102307796478,
0.11742953956127167,
0.028144672513008118,
0.028885571286082268,
0.03537585213780403,
-0.06289814412593842,
-0.000050317394197918475,
0.03226194158196449,
-0.09392514824867249,
-0.05801016092300415,
-0.020014392212033272,
-0.04031052812933922,
0.17134574055671692,
-0.12160717695951462,
0.03345204517245293,
0.04098419472575188,
0.09783966839313507,
0.10073629021644592,
-0.028829937800765038,
-0.1815856397151947,
0.038818612694740295,
-0.24060091376304626,
-0.05831146240234375,
0.027899866923689842,
-0.09110201895236969,
-0.06232144311070442,
0.17409387230873108,
0.013794700615108013,
0.011769929900765419,
-0.006736889015883207,
0.07983319461345673,
0.0110100656747818,
0.1217205822467804,
0.05947643890976906,
-0.05539114400744438,
0.025202350690960884,
-0.09962950646877289,
-0.07107596844434738,
-0.04035590961575508,
-0.05832801014184952,
0.07548832893371582,
0.1409129947423935,
-0.025475580245256424,
-0.020795362070202827,
0.023489827290177345,
0.08550169318914413,
0.0423230417072773,
0.16739299893379211,
-0.16016584634780884,
-0.026555389165878296,
0.04571257904171944,
-0.03384667634963989,
-0.05433850735425949,
-0.010291114449501038,
0.1137225553393364,
-0.02820689231157303,
-0.040318265557289124,
0.021242983639240265,
0.06503437459468842,
0.01481706090271473,
0.05012747645378113,
-0.04056356102228165,
0.14796851575374603,
-0.03461192920804024,
0.019330544397234917,
-0.12413888424634933,
0.13848772644996643,
0.021095896139740944,
-0.03901609033346176,
-0.06735876202583313,
-0.05808034539222717,
0.18150931596755981,
0.0025602965615689754,
0.10535930097103119,
0.012098877690732479,
-0.12160047143697739,
-0.1359938681125641,
-0.11211287975311279,
0.005111907608807087,
0.08330471813678741,
-0.023147236555814743,
-0.022247863933444023,
0.022165266796946526,
-0.036149751394987106,
-0.0530381016433239,
0.15749511122703552,
-0.1289154291152954,
-0.001082550617866218,
0.014728817157447338,
0.06971760839223862,
-0.08223173767328262,
0.026267826557159424,
0.014071501791477203,
-0.1119147390127182,
0.10590848326683044,
0.2521335482597351,
0.10338116437196732,
-0.09591643512248993,
-0.07697287201881409,
0.03418830782175064,
-0.012184361927211285,
-0.000774814048781991,
-0.006932659074664116,
0.0495428591966629,
-0.005566445179283619,
0.006762749515473843,
0.12971895933151245,
-0.07130889594554901,
0.011540771462023258,
-0.08449850976467133,
0.05566910281777382,
-0.05276734381914139,
0.01761564053595066,
-0.002672141883522272,
-0.008124710991978645,
-0.07340748608112335,
-0.061829522252082825,
0.1609770804643631,
-0.07277000695466995,
-0.06468547880649567,
0.05801168829202652,
0.03307786211371422,
0.01431563775986433,
-0.03584568202495575,
-0.04342148080468178,
0.18088261783123016,
0.29330700635910034,
-0.08191116154193878,
0.10001859813928604,
0.09677296131849289,
0.034820813685655594,
-0.23625829815864563,
0.029798466712236404,
-0.1455078274011612,
0.04449721798300743,
0.040447335690259933,
-0.0409548319876194,
0.04191497340798378,
0.10835777968168259,
-0.06094440817832947,
0.2048867791891098,
-0.03527235612273216,
-0.07983248680830002,
-0.01788630709052086,
0.03109324350953102,
0.29443636536598206,
-0.11833466589450836,
0.006058716680854559,
-0.10420958697795868,
-0.21566011011600494,
0.06983078271150589,
-0.18948867917060852,
0.13948246836662292,
-0.05087858438491821,
0.03576415032148361,
-0.01149723306298256,
-0.07561972737312317,
0.20518061518669128,
-0.15641045570373535,
0.05273103713989258,
-0.13722458481788635,
-0.1327189952135086,
0.01617460884153843,
-0.10048147290945053,
0.1545477658510208,
-0.11024226248264313,
-0.023215843364596367,
-0.2284185290336609,
0.012587235309183598,
-0.023200806230306625,
0.10030807554721832,
0.01800704374909401,
-0.07980740070343018,
-0.08767345547676086,
0.1316242516040802,
-0.06486566364765167,
0.034810543060302734,
-0.06996636837720871,
-0.050714004784822464,
-0.010929876938462257,
-0.045061707496643066,
0.03034941293299198,
-0.07934719324111938,
0.15192505717277527,
-0.016938980668783188,
-0.04507075995206833,
0.08636019378900528,
-0.2479533851146698,
0.023727843537926674,
0.025351112708449364,
-0.03495599329471588,
0.09001832455396652,
-0.025513244792819023,
-0.06256973743438721,
0.12282291799783707,
0.1402233988046646,
-0.07322840392589569,
-0.2460673749446869,
-0.06281693279743195,
0.0076784128323197365,
0.039165716618299484,
0.06561196595430374,
0.05125982314348221,
-0.07261458039283752,
-0.011131617240607738,
-0.026896944269537926,
0.030595947057008743,
-0.11692017316818237,
-0.03854857385158539,
0.07790639251470566,
0.017095070332288742,
-0.07846562564373016,
0.07280377298593521,
0.014225782826542854,
-0.021511616185307503,
0.007357571739703417,
0.148970365524292,
0.007519228849560022,
-0.14747941493988037,
-0.06656096875667572,
0.2007484883069992,
-0.01197928935289383,
-0.07260087132453918,
-0.05383119732141495,
-0.008990069851279259,
-0.0476234145462513,
0.05585788935422897,
0.05367223918437958,
-0.013585401698946953,
0.07708586007356644,
0.06263149529695511,
-0.10210110992193222,
-0.046256959438323975,
-0.066561758518219,
0.04169114679098129,
-0.10485753417015076,
0.060470130294561386,
0.009529483504593372,
0.12185006588697433,
-0.09983488917350769,
-0.01802929677069187,
-0.10810204595327377,
-0.06766588985919952,
-0.17349553108215332,
-0.05834362283349037,
-0.041105758398771286,
-0.015651104971766472,
0.03658895567059517,
0.010445823892951012,
-0.057867538183927536,
-0.0442853718996048,
-0.07536603510379791,
0.038444988429546356,
0.06147460639476776,
0.03932281583547592,
-0.03912714496254921,
0.04001858830451965,
0.05909334123134613,
0.013087345287203789,
0.17542624473571777,
0.038768354803323746,
0.05504675209522247,
-0.05045998468995094,
-0.16491834819316864,
-0.05276111513376236,
-0.0074316514655947685,
-0.07559102028608322,
0.1224973127245903,
-0.007679440546780825,
0.007880088873207569,
-0.08065467327833176,
0.03924860805273056,
0.028234204277396202,
0.10404064506292343,
-0.0028364830650389194,
0.10070426017045975,
0.019627176225185394,
-0.07226712256669998,
-0.025392837822437286,
0.021809715777635574,
0.12809939682483673,
0.01567147858440876,
0.026090998202562332,
0.033139873296022415,
0.016619985923171043,
-0.057361043989658356,
0.033977724611759186,
-0.04997231811285019,
-0.15123651921749115,
0.02628709189593792,
-0.05165188014507294,
0.005062380339950323,
-0.016889680176973343,
0.20362506806850433,
0.07867538928985596,
-0.06474173814058304,
-0.010664013214409351,
0.015816617757081985,
-0.0168940220028162,
-0.03121885471045971,
-0.012740966863930225,
0.04592578858137131,
-0.001151384087279439,
-0.04866636544466019,
0.11825273931026459,
0.05015748366713524,
0.05386412516236305,
0.0596686452627182,
0.12528513371944427,
0.016759619116783142,
0.13257254660129547,
0.061999931931495667,
-0.03403807803988457,
-0.13461735844612122,
-0.04495539888739586,
-0.1254577934741974,
0.04646851494908333,
-0.08697032928466797,
0.09941662102937698,
0.1144254133105278,
-0.05959030240774155,
-0.030464433133602142,
-0.08851305395364761,
-0.008356761187314987,
-0.06041252240538597,
0.039516255259513855,
-0.02262675203382969,
-0.0873224213719368,
0.0481097511947155,
0.05495472997426987,
-0.022752324119210243,
0.13218675553798676,
0.015727028250694275,
-0.036317698657512665,
0.13270340859889984,
-0.07583184540271759,
0.11758984625339508,
0.061510033905506134,
-0.043043944984674454,
-0.11560922116041183,
-0.020150646567344666,
-0.06641761213541031,
-0.10098972916603088,
-0.006782987620681524,
-0.005399650428444147,
-0.07349002361297607,
-0.059971679002046585,
0.08397487550973892,
-0.03124053031206131,
-0.09979676455259323,
-0.032152675092220306,
0.0038895104080438614,
0.06054706871509552,
-0.01686914451420307,
-0.0034020058810710907,
0.04728743061423302,
0.015076374635100365,
0.1653461456298828,
-0.02208263985812664,
0.06234867498278618,
-0.13855914771556854,
0.16070103645324707,
-0.14684462547302246,
-0.029404424130916595,
-0.1890171319246292,
-0.09729582816362381,
-0.05156542733311653,
0.20326784253120422,
0.2840938866138458,
-0.19109351933002472,
-0.010187864303588867,
0.020078664645552635,
-0.014484191313385963,
-0.08961770683526993,
0.12571553885936737,
0.029420215636491776,
-0.023631498217582703,
-0.07249019294977188,
-0.02037387527525425,
0.005258576478809118,
-0.06544211506843567,
-0.026979785412549973,
0.18310695886611938,
0.001496660872362554,
0.059546373784542084,
-0.09605178982019424,
0.01754261925816536,
-0.14839904010295868,
-0.10467469692230225,
-0.02111995778977871,
-0.16156397759914398,
-0.09646477550268173,
0.006635562051087618,
0.038640011101961136,
0.08000610023736954,
0.03268849849700928,
-0.015172510407865047,
0.06479045748710632,
-0.056333884596824646,
-0.0037216036580502987,
-0.1231912299990654,
0.00034658415825106204,
0.062129102647304535,
-0.07422006875276566,
0.2545335292816162,
-0.03070417232811451,
-0.12370815873146057,
0.09026903659105301,
-0.03299184888601303,
-0.12452623248100281,
0.07951879501342773,
-0.005700904875993729,
-0.11531132459640503,
-0.057989440858364105,
0.18941475450992584,
-0.012821312062442303,
-0.1364315301179886,
0.046368811279535294,
-0.17166484892368317,
0.031349923461675644,
0.0363016203045845,
-0.001313706859946251,
-0.04714022949337959,
0.024538639932870865,
-0.008008457720279694,
0.10724439471960068,
0.1382838785648346,
0.016739921644330025,
-0.011060068383812904,
-0.05056179314851761,
0.07912429422140121,
0.056927867233753204,
-0.05218246951699257,
-0.1282637119293213,
-0.08599764108657837,
0.03429819270968437,
0.04119478166103363,
-0.08113081753253937,
-0.16903182864189148,
-0.03668912500143051,
-0.10082915425300598,
-0.004939202684909105,
0.051785312592983246,
0.06585265696048737,
0.29044589400291443,
0.06326735019683838,
0.0016605621203780174,
-0.13649453222751617,
0.050569336861371994,
0.0868251696228981,
-0.04697931930422783,
-0.07670357078313828
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# Mistral-7B-Instruct-v0.2-atc
This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on the generator dataset.
It achieves the following results on the evaluation set:
- Loss: 0.1517
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 3
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_ratio: 0.03
- training_steps: 100
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| 0.13 | 0.04 | 100 | 0.1517 |
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1 | {"license": "apache-2.0", "library_name": "peft", "tags": ["trl", "sft", "generated_from_trainer"], "datasets": ["generator"], "base_model": "mistralai/Mistral-7B-Instruct-v0.2", "model-index": [{"name": "Mistral-7B-Instruct-v0.2-atc", "results": []}]} | null | atlaspilotpuppy/Mistral-7B-Instruct-v0.2-atc | [
"peft",
"safetensors",
"trl",
"sft",
"generated_from_trainer",
"dataset:generator",
"base_model:mistralai/Mistral-7B-Instruct-v0.2",
"license:apache-2.0",
"region:us"
] | 2024-02-07T17:38:29+00:00 | [] | [] | TAGS
#peft #safetensors #trl #sft #generated_from_trainer #dataset-generator #base_model-mistralai/Mistral-7B-Instruct-v0.2 #license-apache-2.0 #region-us
| Mistral-7B-Instruct-v0.2-atc
============================
This model is a fine-tuned version of mistralai/Mistral-7B-Instruct-v0.2 on the generator dataset.
It achieves the following results on the evaluation set:
* Loss: 0.1517
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 3
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: cosine
* lr\_scheduler\_warmup\_ratio: 0.03
* training\_steps: 100
### Training results
### Framework versions
* PEFT 0.8.2
* Transformers 4.37.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 3\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_ratio: 0.03\n* training\\_steps: 100",
"### Training results",
"### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#peft #safetensors #trl #sft #generated_from_trainer #dataset-generator #base_model-mistralai/Mistral-7B-Instruct-v0.2 #license-apache-2.0 #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 3\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_ratio: 0.03\n* training\\_steps: 100",
"### Training results",
"### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
60,
117,
4,
39
] | [
"passage: TAGS\n#peft #safetensors #trl #sft #generated_from_trainer #dataset-generator #base_model-mistralai/Mistral-7B-Instruct-v0.2 #license-apache-2.0 #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 3\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_ratio: 0.03\n* training\\_steps: 100### Training results### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.12526676058769226,
0.06797836720943451,
-0.00267362711019814,
0.1236623227596283,
0.1171383187174797,
0.013476521708071232,
0.11753801256418228,
0.11609800159931183,
-0.059750694781541824,
0.09444279968738556,
0.136331245303154,
0.09556011110544205,
0.03373301774263382,
0.2053685188293457,
-0.05791088938713074,
-0.19587449729442596,
0.03561590984463692,
-0.03188083693385124,
-0.03200340270996094,
0.12737064063549042,
0.07988570630550385,
-0.12789857387542725,
0.10071219503879547,
-0.03684290498495102,
-0.16131886839866638,
-0.02294338494539261,
0.00955196563154459,
-0.023235373198986053,
0.12191743403673172,
0.016658170148730278,
0.11395593732595444,
0.03939702734351158,
0.11284638196229935,
-0.19978177547454834,
0.012611850164830685,
0.0722251906991005,
0.013998285867273808,
0.08081571757793427,
0.04399913176894188,
0.004523843992501497,
0.09388890862464905,
-0.08019813895225525,
0.05548829957842827,
0.02874678187072277,
-0.15415823459625244,
-0.1743178814649582,
-0.13542440533638,
0.04366809129714966,
0.08777102828025818,
0.07908637076616287,
-0.0011758374748751521,
0.14703305065631866,
-0.060410384088754654,
0.08192338049411774,
0.2800082862377167,
-0.28164181113243103,
-0.07966902107000351,
0.058129798620939255,
0.04295889288187027,
0.09951326996088028,
-0.11398724466562271,
-0.02573573961853981,
0.06193094700574875,
0.033896464854478836,
0.1348579078912735,
-0.02634401060640812,
-0.08883011341094971,
-0.0034073463175445795,
-0.14862850308418274,
-0.04488854482769966,
0.1213119700551033,
0.04273764416575432,
-0.05517205968499184,
-0.025007959455251694,
-0.07464154064655304,
-0.17562752962112427,
-0.048273809254169464,
-0.021433910354971886,
0.05108188837766647,
-0.024383127689361572,
-0.02083372138440609,
-0.002277473919093609,
-0.0935821682214737,
-0.09031874686479568,
-0.03151024878025055,
0.10778331756591797,
0.03018367663025856,
0.01082424446940422,
-0.011572610586881638,
0.10963831096887589,
-0.03749454393982887,
-0.14609715342521667,
-0.017552979290485382,
0.011683382093906403,
-0.027353793382644653,
-0.05510059744119644,
-0.017256123945116997,
0.006312435492873192,
0.04782736301422119,
0.17226581275463104,
-0.13127806782722473,
0.06110392138361931,
0.004609873052686453,
0.03328670561313629,
-0.12009938061237335,
0.1231827363371849,
-0.05035001039505005,
-0.031093178316950798,
0.014817518182098866,
0.11049960553646088,
0.043958112597465515,
-0.012447291053831577,
-0.0728951171040535,
0.025086313486099243,
0.06364776939153671,
0.047917112708091736,
-0.05062421038746834,
0.03273530304431915,
-0.06246443837881088,
0.01713467203080654,
0.08327827602624893,
-0.09036282449960709,
0.02872476540505886,
0.0248574111610651,
-0.05843891575932503,
-0.08019871264696121,
0.004612007178366184,
0.01144344825297594,
0.014047399163246155,
0.07451865822076797,
-0.09994643181562424,
0.022886503487825394,
-0.07273617386817932,
-0.11926297098398209,
0.030667895451188087,
-0.09739194810390472,
-0.008572590537369251,
-0.0994289293885231,
-0.16751033067703247,
-0.04079333692789078,
0.03865792229771614,
-0.05309155583381653,
-0.008663591928780079,
-0.06467733532190323,
-0.1090976893901825,
0.022341685369610786,
-0.009220562875270844,
0.11357121914625168,
-0.08993393927812576,
0.09586526453495026,
-0.006753346882760525,
0.06143943965435028,
-0.04070547595620155,
0.019450481981039047,
-0.0950477346777916,
0.048750344663858414,
-0.19344890117645264,
0.015638696029782295,
-0.07151299715042114,
0.07352103292942047,
-0.12231074273586273,
-0.07545600831508636,
0.0029328998643904924,
-0.022154437378048897,
0.10405103862285614,
0.14379632472991943,
-0.20254793763160706,
-0.01946188136935234,
0.17714257538318634,
-0.09800615906715393,
-0.11696581542491913,
0.11736398190259933,
-0.04285454377532005,
0.031183606013655663,
0.04964084178209305,
0.1974824070930481,
0.07428493350744247,
-0.1441761553287506,
0.0019365482730790973,
-0.020735539495944977,
0.06009698286652565,
-0.011253931559622288,
0.06941261887550354,
-0.014860938303172588,
0.013469291850924492,
0.007277554366737604,
-0.0586584210395813,
0.03676098585128784,
-0.0963558629155159,
-0.08138497918844223,
-0.04665574058890343,
-0.09905026108026505,
0.021286960691213608,
0.040201447904109955,
0.015106025151908398,
-0.12112296372652054,
-0.07450861483812332,
0.0008044641581363976,
0.104946568608284,
-0.057565610855817795,
0.030339758843183517,
-0.06907152384519577,
0.0981060266494751,
-0.015929361805319786,
-0.02460409700870514,
-0.16076046228408813,
-0.06045025214552879,
0.030292555689811707,
0.013631277717649937,
-0.01105534192174673,
-0.03304186090826988,
0.07583148777484894,
0.08745154738426208,
-0.05387858301401138,
-0.02708662673830986,
-0.02282649278640747,
0.0009242218220606446,
-0.12260991334915161,
-0.23847971856594086,
-0.023897184059023857,
-0.04643336683511734,
0.1272965520620346,
-0.219757080078125,
0.03716060891747475,
0.027231886982917786,
0.08546455204486847,
0.03815311938524246,
-0.03956955671310425,
-0.015548388473689556,
0.06669136136770248,
-0.028572045266628265,
-0.0864182859659195,
0.04729468375444412,
0.0014740489423274994,
-0.054169099777936935,
-0.013517194427549839,
-0.16219334304332733,
0.12335650622844696,
0.1043502688407898,
0.05698004364967346,
-0.11130448430776596,
-0.027928214520215988,
-0.052500225603580475,
-0.02368253841996193,
-0.044188227504491806,
0.0353752002120018,
0.10732194036245346,
0.005098593886941671,
0.11620337516069412,
-0.0897139385342598,
-0.04305558279156685,
0.04508904367685318,
-0.015582886524498463,
0.036769427359104156,
0.12620283663272858,
0.08145873248577118,
-0.09541108459234238,
0.13817982375621796,
0.1605697125196457,
-0.05156869441270828,
0.1008005291223526,
-0.05393778905272484,
-0.06878609210252762,
-0.03211896866559982,
0.049577079713344574,
0.018020063638687134,
0.1464563012123108,
-0.01801600307226181,
0.03171447291970253,
0.005902462173253298,
0.030018558725714684,
-0.01586090214550495,
-0.21287764608860016,
-0.038537174463272095,
0.014836471527814865,
-0.08101285248994827,
-0.04505201801657677,
-0.017411397770047188,
-0.011675357818603516,
0.10302706062793732,
-0.002326607471331954,
-0.10107143223285675,
-0.0028584140818566084,
0.010716281831264496,
-0.07934588938951492,
0.2131984978914261,
-0.10869179666042328,
-0.08587411791086197,
-0.08560455590486526,
-0.006569364108145237,
-0.011742950417101383,
0.0037887482903897762,
0.06693830341100693,
-0.06736330687999725,
-0.03412028029561043,
-0.1313476264476776,
-0.03271590545773506,
0.04867810755968094,
0.015428910031914711,
0.004330422263592482,
-0.018495948985219002,
0.06144779548048973,
-0.10987892746925354,
-0.00851465854793787,
-0.033533744513988495,
-0.04619654640555382,
0.04312115162611008,
0.028042200952768326,
0.12101226300001144,
0.13968053460121155,
0.007285234052687883,
0.0020362732466310263,
-0.03962274640798569,
0.2425917088985443,
-0.07736490666866302,
0.012732228264212608,
0.10404002666473389,
0.01574595458805561,
0.06139445677399635,
0.1630762666463852,
0.049169883131980896,
-0.11652656644582748,
0.007841002196073532,
0.026111995801329613,
-0.042701687663793564,
-0.21478863060474396,
-0.03281783312559128,
-0.04121559485793114,
-0.005329071078449488,
0.08380985260009766,
0.03284092992544174,
0.001688405522145331,
0.04168115556240082,
0.0015192541759461164,
0.004081935156136751,
0.01416720263659954,
0.06526050716638565,
0.02842184528708458,
0.04927095025777817,
0.10114310681819916,
-0.04891953989863396,
0.0012082798639312387,
0.05551066994667053,
-0.00197887746617198,
0.24289990961551666,
-0.008554847911000252,
0.13349051773548126,
0.05774495750665665,
0.17458392679691315,
-0.019660374149680138,
0.06575431674718857,
0.007305930368602276,
-0.04041421785950661,
-0.017706438899040222,
-0.061392106115818024,
-0.015602131374180317,
0.02416856959462166,
-0.08436544239521027,
0.05095130577683449,
-0.09703340381383896,
0.035176750272512436,
0.06453382968902588,
0.28659698367118835,
0.05595821514725685,
-0.34337562322616577,
-0.09329626709222794,
0.0032338269520550966,
-0.005525948945432901,
-0.02871532551944256,
0.025786038488149643,
0.17957137525081635,
-0.05297466740012169,
0.046127282083034515,
-0.053395409137010574,
0.08431944251060486,
0.017071010544896126,
0.03209304064512253,
0.05008421838283539,
0.11575637757778168,
-0.025285879150032997,
0.02629116177558899,
-0.2579517662525177,
0.30231332778930664,
0.024954697117209435,
0.09191518276929855,
-0.02519874833524227,
-0.013296732679009438,
0.03461957722902298,
0.10272172093391418,
0.10462162643671036,
-0.005117109045386314,
-0.08953218162059784,
-0.20461605489253998,
-0.10413536429405212,
0.02726241946220398,
0.09910435974597931,
-0.02827371098101139,
0.09868323802947998,
-0.023190120235085487,
-0.005556303542107344,
0.05055822432041168,
-0.058204520493745804,
-0.09307321161031723,
-0.06162400171160698,
-0.019753960892558098,
0.025500288233160973,
-0.016069594770669937,
-0.0809527114033699,
-0.07937972247600555,
-0.07420818507671356,
0.075888991355896,
-0.07255471497774124,
-0.01872876100242138,
-0.12500490248203278,
0.037035781890153885,
0.10639436542987823,
-0.0741887167096138,
0.028526946902275085,
0.01269610971212387,
0.06338667869567871,
0.010307221673429012,
-0.030312785878777504,
0.11454331874847412,
-0.07434038072824478,
-0.2047300934791565,
-0.06506877392530441,
0.11303994804620743,
0.0593467652797699,
0.061511848121881485,
0.003149618860334158,
0.039346788078546524,
0.0016576532507315278,
-0.09814628213644028,
0.020850546658039093,
0.03107079118490219,
0.07592166215181351,
-0.0058036926202476025,
-0.03580642491579056,
0.02296357974410057,
-0.060276348143815994,
-0.037296853959560394,
0.11005209386348724,
0.32824742794036865,
-0.10296275466680527,
0.062276262789964676,
0.05426808074116707,
-0.06404165178537369,
-0.19491690397262573,
0.0468033142387867,
0.04771128669381142,
0.0035003332886844873,
0.042027391493320465,
-0.13604067265987396,
0.05083116888999939,
0.13581719994544983,
-0.018552135676145554,
0.12546221911907196,
-0.33665570616722107,
-0.12662501633167267,
0.07504395395517349,
0.12442242354154587,
0.08048463612794876,
-0.15803106129169464,
-0.02583405375480652,
0.0005851338501088321,
-0.12398020923137665,
0.07157707959413528,
-0.142708420753479,
0.09863117337226868,
-0.015301515348255634,
0.02068013697862625,
0.0051749106496572495,
-0.05541511997580528,
0.15057462453842163,
0.008117436431348324,
0.13048934936523438,
-0.044049397110939026,
-0.00012060086737619713,
0.06358332931995392,
-0.07060030102729797,
0.03289833664894104,
-0.09628519415855408,
0.04642149433493614,
-0.04217294603586197,
0.002108841435983777,
-0.07986412197351456,
0.025191785767674446,
-0.04057110846042633,
-0.04141771048307419,
-0.05284431204199791,
0.036997679620981216,
0.05917204171419144,
-0.02075878344476223,
0.1234152540564537,
-0.0030251150019466877,
0.16456682980060577,
0.12531551718711853,
0.06435571610927582,
-0.08059702068567276,
-0.03199801221489906,
0.023959048092365265,
-0.029081538319587708,
0.05288820341229439,
-0.16559769213199615,
0.025256343185901642,
0.12969960272312164,
0.024803634732961655,
0.1142287403345108,
0.0589175708591938,
-0.04125230014324188,
0.020014161244034767,
0.06425998359918594,
-0.15021266043186188,
-0.11718741059303284,
0.03303398936986923,
0.00015510551747865975,
-0.11582884192466736,
0.0351656936109066,
0.12155624479055405,
-0.07202732563018799,
-0.001151758013293147,
-0.011693273670971394,
0.04480275139212608,
-0.037725284695625305,
0.2207309901714325,
0.042834196239709854,
0.06180936098098755,
-0.1061636358499527,
0.06693314015865326,
0.034704215824604034,
-0.0799495279788971,
0.020412201061844826,
0.0858473926782608,
-0.1148100271821022,
-0.03949284926056862,
0.10509742796421051,
0.13651247322559357,
-0.03457887843251228,
-0.04072435200214386,
-0.12278436869382858,
-0.10741052031517029,
0.07056599855422974,
0.23128727078437805,
0.07616361975669861,
0.03234067186713219,
0.009671882726252079,
-0.0002715056762099266,
-0.10546145588159561,
0.08364435285329819,
0.032360270619392395,
0.07662811130285263,
-0.12600257992744446,
0.13812871277332306,
-0.014773880131542683,
-0.006289573851972818,
-0.017572157084941864,
0.047304995357990265,
-0.12131708115339279,
0.0014152919175103307,
-0.12725046277046204,
0.020717499777674675,
-0.0418081097304821,
0.007639962248504162,
0.0015393960056826472,
-0.056083958595991135,
-0.060013726353645325,
0.02759207971394062,
-0.10869400948286057,
-0.024073833599686623,
0.003247254528105259,
0.057542186230421066,
-0.13728687167167664,
-0.030940668657422066,
0.02484525553882122,
-0.07489757984876633,
0.07220171391963959,
0.03025209531188011,
0.016304759308695793,
0.046633265912532806,
-0.1758882701396942,
0.010871194303035736,
0.05636061728000641,
-0.004590928088873625,
0.05260268598794937,
-0.10319800674915314,
-0.025759728625416756,
-0.025757672265172005,
0.02445487678050995,
0.028419842943549156,
0.09530196338891983,
-0.11396078020334244,
0.021881822496652603,
-0.017903631553053856,
-0.05997403338551521,
-0.048710744827985764,
0.024957578629255295,
0.10659336298704147,
0.003704681294038892,
0.1587652564048767,
-0.09248416125774384,
0.01590590924024582,
-0.20195139944553375,
-0.017618345096707344,
-0.003891837550327182,
-0.09341752529144287,
-0.13629411160945892,
-0.030775699764490128,
0.08783639222383499,
-0.05033846199512482,
0.07332367449998856,
-0.023159263655543327,
0.05255895107984543,
0.0339977853000164,
-0.05463036894798279,
-0.007598035037517548,
0.036138225346803665,
0.1497218906879425,
0.026734517887234688,
-0.0422285795211792,
0.05150098353624344,
0.02289615012705326,
0.0938766673207283,
0.07514296472072601,
0.234437957406044,
0.15544727444648743,
0.029922811314463615,
0.09707805514335632,
0.025717850774526596,
-0.10634280741214752,
-0.10519273579120636,
0.0845823884010315,
-0.06732913106679916,
0.08455467969179153,
-0.033323418349027634,
0.17905989289283752,
0.11208495497703552,
-0.1804923564195633,
0.02254018746316433,
-0.06978665292263031,
-0.082025907933712,
-0.11402978748083115,
-0.024131450802087784,
-0.08716298639774323,
-0.15420442819595337,
-0.004988263826817274,
-0.10988809913396835,
0.053049810230731964,
0.12349224090576172,
0.014446856454014778,
0.03500937670469284,
0.15694481134414673,
0.02717120200395584,
0.039811693131923676,
0.028349706903100014,
0.017834631726145744,
-0.028575142845511436,
-0.07050773501396179,
-0.10889264941215515,
0.04220903664827347,
-0.0711798295378685,
0.038062915205955505,
-0.0323510468006134,
-0.0037991574499756098,
0.05180508270859718,
-0.016385424882173538,
-0.09049256891012192,
0.025119248777627945,
0.028742868453264236,
0.042531512677669525,
0.04398186504840851,
0.04310457035899162,
0.01030095387250185,
0.008261351846158504,
0.22665348649024963,
-0.0686192587018013,
-0.07545353472232819,
-0.12601138651371002,
0.2666912376880646,
0.034074246883392334,
0.006011601537466049,
0.01620216853916645,
-0.07676123827695847,
0.004546897020190954,
0.16611236333847046,
0.16977445781230927,
-0.07946094125509262,
-0.002029828727245331,
-0.028157679364085197,
-0.007557480130344629,
-0.047243501991033554,
0.10876491665840149,
0.11101316660642624,
0.024906069040298462,
-0.09731552004814148,
-0.04014606401324272,
-0.03691625967621803,
-0.01671581156551838,
-0.049355387687683105,
0.030355332419276237,
-0.006100883241742849,
0.010360655374825,
-0.04974440112709999,
0.06600639969110489,
-0.028505200520157814,
-0.1091730073094368,
0.08059686422348022,
-0.1938794106245041,
-0.14466121792793274,
-0.016382327303290367,
0.0695415809750557,
0.0025371641386300325,
0.046574611216783524,
-0.03138384222984314,
-0.011653205379843712,
0.07855592668056488,
-0.04132525995373726,
-0.06224910542368889,
-0.13504524528980255,
0.0694727823138237,
-0.09517920017242432,
0.23255585134029388,
-0.038092292845249176,
0.04555477946996689,
0.12016322463750839,
0.029657859355211258,
-0.11853156238794327,
0.07748693972826004,
0.06768269091844559,
-0.08087214827537537,
-0.00013937498442828655,
0.08102183789014816,
-0.055716127157211304,
0.10840652883052826,
0.04791708663105965,
-0.11739282310009003,
0.006557777523994446,
-0.03501003235578537,
-0.047242023050785065,
-0.04070868715643883,
-0.04457199200987816,
-0.047358438372612,
0.1429380178451538,
0.17020775377750397,
-0.03615422546863556,
0.029066862538456917,
-0.0444464348256588,
0.04104035347700119,
0.062209710478782654,
0.05742210894823074,
-0.03741670027375221,
-0.2434021234512329,
0.03944828733801842,
0.049669649451971054,
-0.008986246772110462,
-0.21161861717700958,
-0.08894523978233337,
-0.0035263069439679384,
-0.03633502498269081,
-0.082382433116436,
0.10134238004684448,
0.0742262452840805,
0.05156368389725685,
-0.05290893465280533,
-0.11572622507810593,
-0.05379835516214371,
0.17347775399684906,
-0.11375828832387924,
-0.08804848790168762
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# train_2024-02-07-03-18-19
This model is a fine-tuned version of [Qwen/Qwen1.5-7B](https://huggingface.co/Qwen/Qwen1.5-7B) on the openorca dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 1
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 16
- total_train_batch_size: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- num_epochs: 1.5
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.1.1+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1 | {"license": "other", "library_name": "peft", "tags": ["llama-factory", "lora", "generated_from_trainer"], "base_model": "Qwen/Qwen1.5-7B", "model-index": [{"name": "train_2024-02-07-03-18-19", "results": []}]} | null | Crystalcareai/CrystalQwen-1.5-7B-Alpha-Lora | [
"peft",
"safetensors",
"llama-factory",
"lora",
"generated_from_trainer",
"base_model:Qwen/Qwen1.5-7B",
"license:other",
"region:us"
] | 2024-02-07T17:39:24+00:00 | [] | [] | TAGS
#peft #safetensors #llama-factory #lora #generated_from_trainer #base_model-Qwen/Qwen1.5-7B #license-other #region-us
|
# train_2024-02-07-03-18-19
This model is a fine-tuned version of Qwen/Qwen1.5-7B on the openorca dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 1
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 16
- total_train_batch_size: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- num_epochs: 1.5
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.1.1+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1 | [
"# train_2024-02-07-03-18-19\n\nThis model is a fine-tuned version of Qwen/Qwen1.5-7B on the openorca dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 1\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 16\n- total_train_batch_size: 16\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- num_epochs: 1.5",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#peft #safetensors #llama-factory #lora #generated_from_trainer #base_model-Qwen/Qwen1.5-7B #license-other #region-us \n",
"# train_2024-02-07-03-18-19\n\nThis model is a fine-tuned version of Qwen/Qwen1.5-7B on the openorca dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 1\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 16\n- total_train_batch_size: 16\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- num_epochs: 1.5",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
48,
36,
6,
12,
8,
3,
114,
4,
39
] | [
"passage: TAGS\n#peft #safetensors #llama-factory #lora #generated_from_trainer #base_model-Qwen/Qwen1.5-7B #license-other #region-us \n# train_2024-02-07-03-18-19\n\nThis model is a fine-tuned version of Qwen/Qwen1.5-7B on the openorca dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 1\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 16\n- total_train_batch_size: 16\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- num_epochs: 1.5### Training results### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.14526879787445068,
0.13578520715236664,
-0.001567199593409896,
0.11850582808256149,
0.10730592161417007,
0.011245585978031158,
0.099681057035923,
0.1319352090358734,
-0.0746554285287857,
0.1077953651547432,
0.0975528135895729,
0.016452157869935036,
0.0545944981276989,
0.18430395424365997,
-0.03146540746092796,
-0.23163306713104248,
0.02734270878136158,
-0.04308157414197922,
-0.06414633989334106,
0.08016136288642883,
0.07127220928668976,
-0.11489356309175491,
0.08447999507188797,
0.01312529481947422,
-0.14222143590450287,
-0.007418876048177481,
-0.051844220608472824,
-0.04009191319346428,
0.0923578217625618,
0.03843254968523979,
0.10114217549562454,
0.02470128983259201,
0.14057543873786926,
-0.21962003409862518,
0.007735773455351591,
0.04615805670619011,
0.01965468004345894,
0.08000858873128891,
0.06306435912847519,
0.057438384741544724,
0.04688984900712967,
-0.1440654695034027,
0.08488937467336655,
0.01580796018242836,
-0.08000761270523071,
-0.10040730237960815,
-0.08616656064987183,
0.0745416134595871,
0.06546714901924133,
0.09920921176671982,
0.015953311696648598,
0.1678989976644516,
-0.0909452959895134,
0.06009848788380623,
0.23106542229652405,
-0.32287171483039856,
-0.08110561221837997,
0.06403212994337082,
0.062454015016555786,
0.0782868042588234,
-0.1448037177324295,
-0.04375430569052696,
0.06337690353393555,
0.006533531937748194,
0.06886892765760422,
-0.01575307548046112,
-0.030500955879688263,
-0.01778108812868595,
-0.1274474561214447,
-0.023740965873003006,
0.15151788294315338,
0.0700162723660469,
-0.04786047339439392,
-0.10180656611919403,
-0.06000969558954239,
-0.14816296100616455,
-0.012124067172408104,
-0.03469430282711983,
0.0309160016477108,
-0.031081033870577812,
-0.049829427152872086,
-0.0290537029504776,
-0.09087982773780823,
-0.05466596037149429,
0.0038188141770660877,
0.09873003512620926,
0.03851944953203201,
0.0094026243314147,
0.003099990775808692,
0.10620871186256409,
-0.006357708014547825,
-0.1445264369249344,
-0.047964658588171005,
-0.011667733080685139,
-0.08393870294094086,
-0.07373757660388947,
-0.018442882224917412,
0.009993556886911392,
0.045274022966623306,
0.13339661061763763,
-0.11067315936088562,
0.06927476823329926,
0.026123661547899246,
0.02148466370999813,
-0.02050890401005745,
0.15723764896392822,
-0.04900764301419258,
-0.04245143011212349,
0.01678593084216118,
0.11475194990634918,
-0.0027997572906315327,
-0.01468642707914114,
-0.10380566120147705,
-0.05020808055996895,
0.08808644860982895,
0.05630853772163391,
-0.06209111213684082,
0.007149846293032169,
-0.03557576239109039,
-0.02815261483192444,
0.09140008687973022,
-0.10146627575159073,
0.04952345788478851,
0.0024371682666242123,
-0.0744587779045105,
-0.03087250515818596,
0.004532684106379747,
0.022863082587718964,
-0.00928102433681488,
0.08041848987340927,
-0.09846578538417816,
-0.009084110148251057,
-0.051478538662195206,
-0.032476890832185745,
0.013316421769559383,
-0.059328190982341766,
0.013438168913125992,
-0.10420877486467361,
-0.16428762674331665,
-0.05928300321102142,
0.00852777436375618,
-0.063310407102108,
-0.0745491087436676,
-0.061362963169813156,
-0.07934921234846115,
0.010221587494015694,
-0.007916565984487534,
0.0944344624876976,
-0.05774914473295212,
0.0750216618180275,
-0.022105716168880463,
0.008799021132290363,
-0.0035837991163134575,
0.012344605289399624,
-0.07343367487192154,
0.04378321394324303,
-0.11682590842247009,
0.052443113178014755,
-0.07406715303659439,
0.045124005526304245,
-0.11075073480606079,
-0.09990984946489334,
-0.02800629660487175,
-0.04690885916352272,
0.06168930232524872,
0.1149374470114708,
-0.14344075322151184,
-0.006041030865162611,
0.19672055542469025,
-0.08269122987985611,
-0.08746405690908432,
0.10545387119054794,
-0.0617903396487236,
-0.014954599551856518,
0.02293327823281288,
0.13531212508678436,
0.15372300148010254,
-0.1691664159297943,
-0.025643404573202133,
0.010528589598834515,
0.06732582300901413,
0.0012495981063693762,
0.08230271190404892,
-0.004412952810525894,
0.05293819680809975,
0.004025722853839397,
-0.020452281460165977,
-0.0020821287762373686,
-0.07818495482206345,
-0.09281515330076218,
-0.05093448981642723,
-0.08876040577888489,
0.007112293038517237,
0.007451099809259176,
0.04103071615099907,
-0.04574389010667801,
-0.08433768898248672,
0.004771067760884762,
0.1593051254749298,
-0.03553032502532005,
-0.0018214988522231579,
-0.08140294253826141,
0.08639615774154663,
-0.0598137304186821,
-0.045407041907310486,
-0.1762632578611374,
-0.11543986201286316,
0.08205511420965195,
-0.07176317274570465,
0.026111388579010963,
0.058400124311447144,
0.06214681267738342,
0.09620514512062073,
-0.03191704303026199,
-0.04316588491201401,
-0.09686799347400665,
-0.025485120713710785,
-0.11786971986293793,
-0.18976977467536926,
-0.031206054612994194,
-0.048234689980745316,
0.14030808210372925,
-0.2435038685798645,
-0.0050690858624875546,
-0.0018339571543037891,
0.12040705978870392,
0.05005622282624245,
-0.04282794147729874,
0.0047528548166155815,
0.059275005012750626,
-0.0006201894721016288,
-0.09334395825862885,
0.019142931327223778,
0.02285495400428772,
-0.06767115741968155,
-0.06718466430902481,
-0.14999620616436005,
0.11207439005374908,
0.05482543632388115,
0.07244107872247696,
-0.09859004616737366,
-0.11685846745967865,
-0.07872961461544037,
-0.04225475341081619,
-0.09538133442401886,
0.02579692006111145,
0.23113562166690826,
0.006906007882207632,
0.12542101740837097,
-0.10057748854160309,
-0.07102528214454651,
0.028964757919311523,
0.019881699234247208,
0.014807933010160923,
0.08642033487558365,
0.09646710008382797,
-0.1342141479253769,
0.07453260570764542,
0.09406249970197678,
-0.04842193424701691,
0.1545167863368988,
-0.06177651137113571,
-0.08813992142677307,
-0.04747647047042847,
0.06785979866981506,
-0.01045341044664383,
0.14180094003677368,
-0.051231127232313156,
0.014956817962229252,
0.016108661890029907,
0.031351715326309204,
0.02238507755100727,
-0.1610296368598938,
-0.030874082818627357,
0.0216178297996521,
-0.03571191802620888,
0.034744247794151306,
-0.018727287650108337,
0.01870475895702839,
0.08639253675937653,
0.03813450410962105,
-0.05920463427901268,
-0.006629794370383024,
-0.013644876889884472,
-0.08047693222761154,
0.19868776202201843,
-0.09926537424325943,
-0.08088208734989166,
-0.11446140706539154,
0.09020991623401642,
-0.039276815950870514,
-0.021997621282935143,
0.018695970997214317,
-0.047973014414310455,
-0.03140226751565933,
-0.10349730402231216,
-0.07078937441110611,
-0.03295125812292099,
0.019523419439792633,
0.04409708455204964,
0.02050880342721939,
0.06895940005779266,
-0.12023409456014633,
0.01297500915825367,
-0.016964983195066452,
-0.07982742786407471,
0.0037544285878539085,
0.0234879981726408,
0.11616308987140656,
0.1102568581700325,
-0.004400226287543774,
0.009255019016563892,
-0.04333553835749626,
0.25116342306137085,
-0.07711756974458694,
-0.021149517968297005,
0.05847568064928055,
0.0028797131963074207,
0.05686988681554794,
0.1330300122499466,
0.037267427891492844,
-0.10165815055370331,
0.007219402585178614,
0.0344076007604599,
-0.006140604615211487,
-0.21719703078269958,
-0.04145950451493263,
-0.035043325275182724,
-0.02805640734732151,
0.1070278063416481,
0.06806600093841553,
0.029160423204302788,
0.05646649748086929,
-0.00720200315117836,
0.055536262691020966,
-0.04205010086297989,
0.08709213137626648,
0.02686491422355175,
0.06791242212057114,
0.08208058774471283,
-0.017089232802391052,
-0.027455570176243782,
0.06587463617324829,
0.02396601252257824,
0.27344197034835815,
-0.014632046222686768,
0.11028584092855453,
0.020896868780255318,
0.18713240325450897,
-0.0006968080997467041,
0.030179638415575027,
0.02804465778172016,
-0.020298413932323456,
-0.0038476940244436264,
-0.06826970726251602,
-0.025659987702965736,
0.050361115485429764,
-0.00780597236007452,
0.07121344655752182,
-0.10650676488876343,
0.0826740637421608,
0.014940720051527023,
0.26985397934913635,
0.02802998013794422,
-0.26302245259284973,
-0.05193368345499039,
-0.020092418417334557,
-0.0430472269654274,
-0.045474372804164886,
0.030389519408345222,
0.16938205063343048,
-0.13972920179367065,
0.017670536413788795,
-0.04874897375702858,
0.08911003172397614,
-0.06991560012102127,
-0.014018901623785496,
0.040300168097019196,
0.09477834403514862,
-0.01794675923883915,
0.07403242588043213,
-0.1696084588766098,
0.25578004121780396,
0.025729257613420486,
0.10247324407100677,
-0.03564347326755524,
0.022761359810829163,
0.009567742235958576,
0.08032581955194473,
0.1283835619688034,
-0.018418725579977036,
-0.0488055981695652,
-0.20952455699443817,
-0.12616829574108124,
0.04593003913760185,
0.11108995974063873,
-0.049487072974443436,
0.09411190450191498,
-0.02473752386868,
0.006897589657455683,
0.025446146726608276,
0.01070439349859953,
-0.12830430269241333,
-0.11492769420146942,
0.00448194844648242,
0.02574795112013817,
-0.04101615771651268,
-0.08376621454954147,
-0.07922707498073578,
-0.07197989523410797,
0.10246813297271729,
-0.05382680147886276,
-0.02398117259144783,
-0.13594862818717957,
0.0716022327542305,
0.10417389869689941,
-0.08502703905105591,
0.018200183287262917,
0.015334040857851505,
0.11327902227640152,
0.007576433476060629,
-0.02584686316549778,
0.05955207720398903,
-0.05025801062583923,
-0.21441315114498138,
-0.078639455139637,
0.12145785242319107,
0.05077574402093887,
0.049506284296512604,
0.021660268306732178,
0.005003620870411396,
0.01393148209899664,
-0.11914604902267456,
-0.004373129922896624,
0.07313905656337738,
0.0707373395562172,
0.02639951929450035,
-0.0713052749633789,
0.06947933882474899,
-0.02958081290125847,
-0.009733251295983791,
0.127067431807518,
0.27137500047683716,
-0.1081656888127327,
0.04086501523852348,
0.0681261196732521,
-0.06964728981256485,
-0.1225711777806282,
0.037437766790390015,
0.1242854967713356,
0.031790245324373245,
0.008809599094092846,
-0.16411609947681427,
0.09112632274627686,
0.12886695563793182,
-0.022901620715856552,
0.03659575805068016,
-0.324215292930603,
-0.12166178226470947,
0.06277717649936676,
0.086313895881176,
0.04513758048415184,
-0.1402135044336319,
-0.041460003703832626,
0.016291610896587372,
-0.09766270965337753,
0.05607768893241882,
-0.09649743139743805,
0.0996042937040329,
-0.014975161291658878,
0.05919773504137993,
0.02868380770087242,
-0.035305287688970566,
0.1433388590812683,
0.02494925446808338,
0.10255435109138489,
-0.0476534329354763,
-0.00857640989124775,
-0.004418237134814262,
-0.07624795287847519,
0.0299568809568882,
-0.04327166825532913,
0.07228253781795502,
-0.15833203494548798,
-0.02043853886425495,
-0.036819394677877426,
0.06049841269850731,
-0.04530872404575348,
-0.0645887702703476,
-0.059088245034217834,
0.06462043523788452,
0.06818887591362,
-0.009032847359776497,
0.07417424768209457,
0.028251847252249718,
0.07767916470766068,
0.1319052129983902,
0.08818976581096649,
-0.0034972636494785547,
-0.07733622193336487,
-0.007017507217824459,
-0.01433807797729969,
0.0743110403418541,
-0.105800099670887,
0.009030194021761417,
0.1452801525592804,
0.027631325647234917,
0.125649094581604,
0.028685778379440308,
-0.0510428249835968,
0.023418931290507317,
0.03628723695874214,
-0.11388441175222397,
-0.13182546198368073,
-0.011000528000295162,
0.037649013102054596,
-0.13836726546287537,
-0.0103519381955266,
0.10891781747341156,
-0.071938157081604,
-0.014149206690490246,
-0.012667167000472546,
-0.006792561151087284,
-0.006030458956956863,
0.18643847107887268,
0.04578469693660736,
0.06190340220928192,
-0.06864001601934433,
0.09670089185237885,
0.08634766936302185,
-0.04634945094585419,
0.04323285073041916,
0.07395307719707489,
-0.0976388230919838,
-0.016653932631015778,
0.09215976297855377,
0.20405182242393494,
-0.04201094061136246,
-0.036612462252378464,
-0.06125776097178459,
-0.06422661244869232,
0.04388093203306198,
0.1466982662677765,
0.04709906503558159,
-0.0018074254039674997,
0.002738520037382841,
0.02165418304502964,
-0.124982550740242,
0.09444228559732437,
0.025105593726038933,
0.07421112805604935,
-0.1608010083436966,
0.08904211223125458,
-0.026714375242590904,
0.039663635194301605,
-0.01332762185484171,
0.02194739691913128,
-0.10067107528448105,
-0.038561828434467316,
-0.10847984254360199,
0.058449845761060715,
-0.02218806929886341,
0.025086969137191772,
-0.023567406460642815,
-0.05341991409659386,
-0.03668580576777458,
0.03622475266456604,
-0.06457103043794632,
-0.03590691462159157,
0.008787702769041061,
0.063688263297081,
-0.09928658604621887,
-0.0190727636218071,
0.012223468162119389,
-0.07129979133605957,
0.09023091197013855,
0.016448702663183212,
0.02889920212328434,
0.007835257798433304,
-0.0826832726597786,
0.04194904491305351,
0.027752507477998734,
0.03134531155228615,
0.06168365478515625,
-0.06839776039123535,
-0.00584951089695096,
-0.04658255726099014,
0.021985167637467384,
0.015099352225661278,
0.038710977882146835,
-0.13552464544773102,
-0.015931187197566032,
-0.03629397600889206,
-0.065134696662426,
-0.06244336813688278,
0.021442685276269913,
0.09991911053657532,
-0.009404763579368591,
0.1324874311685562,
-0.0813743844628334,
0.062442246824502945,
-0.2113547921180725,
-0.03348087519407272,
0.037013307213783264,
0.008172934874892235,
-0.0712713748216629,
-0.04724664241075516,
0.07936245203018188,
-0.06691355258226395,
0.10303715616464615,
-0.038522761315107346,
0.09702137112617493,
0.01578487642109394,
-0.0890214741230011,
-0.003006091807037592,
0.012910811230540276,
0.16434982419013977,
0.09370767325162888,
-0.021838011220097542,
0.09522369503974915,
-0.013699091039597988,
0.07482385635375977,
0.053467102348804474,
0.16217149794101715,
0.1431228220462799,
-0.0015718326903879642,
0.07410860806703568,
0.080409936606884,
-0.12294532358646393,
-0.11770663410425186,
0.08551568537950516,
-0.04065560922026634,
0.07538275420665741,
-0.05705997720360756,
0.09859564900398254,
0.12227794528007507,
-0.16583454608917236,
0.011867017485201359,
-0.014935547485947609,
-0.09806567430496216,
-0.12021152675151825,
-0.04034262150526047,
-0.10347623378038406,
-0.11846307665109634,
0.021805863827466965,
-0.12797857820987701,
0.0020869618747383356,
0.01939079351723194,
0.007996923290193081,
0.012837862595915794,
0.15586905181407928,
-0.022910723462700844,
0.02694166824221611,
0.06186140701174736,
0.03327251225709915,
0.029269259423017502,
-0.08735823631286621,
-0.06708154082298279,
0.03892188146710396,
-0.04463454335927963,
0.05958925560116768,
-0.044520121067762375,
0.027658870443701744,
0.03348734229803085,
0.005442212335765362,
-0.052549444139003754,
0.03025786578655243,
-0.001195670454762876,
0.015656424686312675,
0.04306845739483833,
0.047488611191511154,
0.0049926177598536015,
-0.03178083524107933,
0.248941570520401,
-0.05409665405750275,
-0.04320492595434189,
-0.14606042206287384,
0.11993115395307541,
0.030078735202550888,
0.0009349387255497277,
0.04741201177239418,
-0.09526234120130539,
0.015283551998436451,
0.17234280705451965,
0.11623644828796387,
-0.08855102211236954,
0.006488134618848562,
0.017556767910718918,
-0.014543995261192322,
-0.06530995666980743,
0.08742551505565643,
0.08389583230018616,
0.056095197796821594,
-0.06378012895584106,
-0.030603498220443726,
-0.0043318807147443295,
-0.006468652747571468,
-0.0583374910056591,
0.026097046211361885,
0.022504789754748344,
0.01670732907950878,
-0.07052605599164963,
0.05635831877589226,
-0.03804640844464302,
-0.10562966763973236,
0.05908762663602829,
-0.17072361707687378,
-0.18315348029136658,
-0.014045651070773602,
0.03209351748228073,
0.0029567775782197714,
0.05038401484489441,
-0.024210592731833458,
-0.026906801387667656,
0.11168011277914047,
-0.03705639764666557,
-0.06041654571890831,
-0.09458474069833755,
0.046876758337020874,
-0.04672444984316826,
0.22662337124347687,
-0.00003821492282440886,
0.06535458564758301,
0.11104850471019745,
0.028619980439543724,
-0.1364317685365677,
0.0209357887506485,
0.09436088800430298,
-0.08556383848190308,
0.0033052857033908367,
0.12444685399532318,
-0.02995114028453827,
0.12037185579538345,
0.06763004511594772,
-0.08657879382371902,
-0.00957456510514021,
-0.04590962454676628,
0.024290546774864197,
-0.0954446792602539,
0.02298823371529579,
-0.049704987555742264,
0.16888025403022766,
0.1686149388551712,
-0.033884067088365555,
-0.010172797366976738,
-0.025858571752905846,
0.052608806639909744,
0.04071356728672981,
0.08047952502965927,
-0.032119497656822205,
-0.20647764205932617,
0.02868354134261608,
-0.0019778984133154154,
0.0263749361038208,
-0.19165624678134918,
-0.09919197112321854,
0.05283776670694351,
-0.047229889780282974,
-0.04929573833942413,
0.12260603159666061,
0.04678849130868912,
0.03069675713777542,
-0.03963256999850273,
-0.1386907547712326,
-0.049698274582624435,
0.12458249181509018,
-0.13726668059825897,
-0.07298815995454788
] |
null | null | null |
# **Reinforce** Agent playing **CartPole-v1**
This is a trained model of a **Reinforce** agent playing **CartPole-v1** .
To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: https://huggingface.co/deep-rl-course/unit4/introduction
| {"tags": ["CartPole-v1", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class"], "model-index": [{"name": "cartPole8", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "CartPole-v1", "type": "CartPole-v1"}, "metrics": [{"type": "mean_reward", "value": "500.00 +/- 0.00", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | turgutburak01/cartPole8 | [
"CartPole-v1",
"reinforce",
"reinforcement-learning",
"custom-implementation",
"deep-rl-class",
"model-index",
"region:us"
] | 2024-02-07T17:39:35+00:00 | [] | [] | TAGS
#CartPole-v1 #reinforce #reinforcement-learning #custom-implementation #deep-rl-class #model-index #region-us
|
# Reinforce Agent playing CartPole-v1
This is a trained model of a Reinforce agent playing CartPole-v1 .
To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL
| [
"# Reinforce Agent playing CartPole-v1\n This is a trained model of a Reinforce agent playing CartPole-v1 .\n To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL"
] | [
"TAGS\n#CartPole-v1 #reinforce #reinforcement-learning #custom-implementation #deep-rl-class #model-index #region-us \n",
"# Reinforce Agent playing CartPole-v1\n This is a trained model of a Reinforce agent playing CartPole-v1 .\n To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL"
] | [
39,
54
] | [
"passage: TAGS\n#CartPole-v1 #reinforce #reinforcement-learning #custom-implementation #deep-rl-class #model-index #region-us \n# Reinforce Agent playing CartPole-v1\n This is a trained model of a Reinforce agent playing CartPole-v1 .\n To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL"
] | [
0.007526164408773184,
-0.12498430907726288,
-0.0013541718944907188,
0.09601131081581116,
0.11848696321249008,
-0.04186001420021057,
0.11405468732118607,
0.05624859035015106,
0.09539441019296646,
0.04239490255713463,
0.13636724650859833,
0.06906966865062714,
-0.004102868959307671,
0.12412862479686737,
0.09840741008520126,
-0.26058563590049744,
0.07420794665813446,
-0.04403980076313019,
-0.009944677352905273,
0.10139261186122894,
0.07836852967739105,
-0.08325441926717758,
0.051592715084552765,
0.00009572553972247988,
-0.044259943068027496,
0.0321260429918766,
0.013628939166665077,
-0.053157225251197815,
0.1606452465057373,
-0.07313758134841919,
0.10494591295719147,
-0.03843724727630615,
0.14574295282363892,
-0.1126825287938118,
0.04758213832974434,
0.05111503228545189,
-0.04548581689596176,
0.03848232328891754,
-0.12538743019104004,
-0.06033875793218613,
0.026815801858901978,
-0.015865681692957878,
0.12249194830656052,
0.03647647053003311,
-0.1777559220790863,
-0.13461355865001678,
-0.0165896974503994,
0.12325166910886765,
0.1627800315618515,
0.00512364786118269,
0.014270431362092495,
0.16791965067386627,
-0.1761058121919632,
0.025937072932720184,
0.11400806158781052,
-0.37275227904319763,
-0.00034436015994288027,
0.2240462601184845,
0.06164427846670151,
0.1252165287733078,
-0.12646614015102386,
0.010440526530146599,
0.07403992861509323,
0.04368630796670914,
0.049784936010837555,
-0.015430688858032227,
-0.12260042130947113,
0.08455035835504532,
-0.1383819431066513,
-0.058066487312316895,
0.1495426446199417,
-0.019741326570510864,
-0.009476418606936932,
-0.016515808179974556,
-0.009238536469638348,
-0.050979889929294586,
-0.03430935740470886,
-0.11778499186038971,
0.10755524039268494,
0.04975730925798416,
0.0038771627005189657,
-0.04602450504899025,
-0.05612579360604286,
-0.09815777093172073,
-0.03123871050775051,
0.0372777059674263,
-0.013706400990486145,
0.01091629359871149,
0.027692900970578194,
0.09935613721609116,
-0.13446329534053802,
0.01825822703540325,
-0.028096558526158333,
-0.028040969744324684,
-0.1316804438829422,
-0.11984307318925858,
-0.026084421202540398,
0.004223645199090242,
0.03029833547770977,
0.20433813333511353,
0.020139509811997414,
0.059011414647102356,
-0.0022708347532898188,
0.09776382148265839,
0.029780851677060127,
0.13517548143863678,
-0.04466623440384865,
0.19488364458084106,
0.07711011171340942,
0.05364556983113289,
0.03204274922609329,
-0.05344729498028755,
-0.19369827210903168,
0.04861246794462204,
0.06659778952598572,
0.08274952322244644,
-0.1178959533572197,
0.0059632807970047,
-0.10316018015146255,
0.0028950648847967386,
-0.10474003106355667,
-0.0642905905842781,
-0.02892979420721531,
0.031841445714235306,
-0.10535725951194763,
0.028785312548279762,
0.025052599608898163,
0.04140377417206764,
0.0676041767001152,
-0.12253966927528381,
-0.07404746115207672,
-0.021733485162258148,
-0.12817098200321198,
-0.09923440217971802,
0.08802318572998047,
-0.026199497282505035,
-0.005110981408506632,
-0.1253623217344284,
-0.2661486268043518,
-0.05670225992798805,
0.06396034359931946,
-0.03231031447649002,
-0.08589376509189606,
-0.1633463054895401,
0.026403428986668587,
-0.07700273394584656,
0.05221332609653473,
0.04776721075177193,
-0.03665859252214432,
0.02023705095052719,
-0.07958202809095383,
0.12739010155200958,
0.049698662012815475,
0.00541001046076417,
-0.09916839748620987,
0.07882837951183319,
-0.3034103214740753,
-0.02581131085753441,
-0.15228183567523956,
0.0772043839097023,
-0.07893010973930359,
0.01308529730886221,
0.05044940114021301,
0.043790437281131744,
-0.016942394897341728,
0.16269747912883759,
-0.17043575644493103,
-0.05301272124052048,
0.026445282623171806,
-0.09261117875576019,
-0.09916394203901291,
0.07275339215993881,
-0.06339669227600098,
0.21263530850410461,
0.08751397579908371,
0.17006252706050873,
-0.011036526411771774,
-0.16256992518901825,
0.1207515075802803,
0.07522942125797272,
-0.1639646589756012,
0.004287737421691418,
0.061784300953149796,
-0.0016935690073296428,
0.02746843732893467,
-0.01872866041958332,
-0.07289361208677292,
0.06302516162395477,
-0.07825060933828354,
0.022581040859222412,
0.06258945167064667,
-0.09531243145465851,
0.23986859619617462,
-0.005434412509202957,
0.0862451046705246,
-0.025957979261875153,
-0.09802921861410141,
0.00908072479069233,
0.07164718210697174,
-0.0014321404742076993,
0.01703714393079281,
-0.14553219079971313,
0.23044352233409882,
-0.07965081930160522,
0.011176814325153828,
-0.11607582122087479,
-0.1256982982158661,
0.011873425915837288,
0.13336114585399628,
0.059921663254499435,
0.16569606959819794,
0.09518871456384659,
-0.032197169959545135,
0.017584815621376038,
-0.0023385772947221994,
-0.09040450304746628,
0.01580043137073517,
-0.0021571461111307144,
-0.12167251110076904,
-0.07353103160858154,
-0.08134473115205765,
0.12585052847862244,
-0.20988115668296814,
0.015492538921535015,
0.04099845886230469,
0.008103687316179276,
0.04467369243502617,
0.023746047168970108,
-0.013269703835248947,
-0.00007021807687124237,
0.03244573250412941,
-0.10098352283239365,
0.12937165796756744,
0.013381263241171837,
0.014676140621304512,
-0.006365173030644655,
-0.05572463944554329,
0.03720450773835182,
0.040439579635858536,
-0.11237845569849014,
-0.11330515146255493,
-0.009658765979111195,
-0.0015364213613793254,
0.02637762948870659,
-0.022321155294775963,
0.052120618522167206,
0.27587956190109253,
0.05387469753623009,
0.10401033610105515,
-0.05769326910376549,
0.015315087512135506,
-0.015322818420827389,
-0.07135670632123947,
0.06358719617128372,
0.025013601407408714,
0.08050397783517838,
-0.03531401976943016,
0.03759452700614929,
0.1675453782081604,
-0.015888912603259087,
0.11127935349941254,
-0.06545067578554153,
-0.03844274953007698,
-0.043109722435474396,
0.05627678707242012,
0.015021559782326221,
0.04564907029271126,
0.0000015355876712419558,
-0.08444724231958389,
-0.03503387048840523,
-0.03988509997725487,
-0.010637006722390652,
-0.12273643165826797,
-0.00499896751716733,
0.01265440508723259,
-0.021940499544143677,
0.04488934203982353,
0.07375624030828476,
-0.04849626496434212,
0.025821007788181305,
0.06070821359753609,
-0.10193055868148804,
0.08957115560770035,
0.015067169442772865,
-0.06946801394224167,
0.13769419491291046,
-0.07484805583953857,
-0.045293889939785004,
-0.1025395318865776,
-0.1568877100944519,
0.09384927153587341,
0.06704871356487274,
-0.05427970737218857,
-0.1503879576921463,
-0.0016851738328114152,
-0.008973666466772556,
0.09206123650074005,
-0.006399387493729591,
-0.12621140480041504,
0.01989075168967247,
0.08295059949159622,
-0.05633419007062912,
-0.09804849326610565,
-0.0075809285044670105,
-0.05280788615345955,
-0.17707788944244385,
-0.03888550028204918,
-0.06398582458496094,
-0.06734282523393631,
0.23586803674697876,
0.02017230913043022,
0.08274748176336288,
-0.044721852988004684,
0.04250151664018631,
-0.012231717817485332,
0.0006326579605229199,
0.10689259320497513,
-0.09043551236391068,
-0.017900818958878517,
-0.001320177922025323,
-0.024820495396852493,
-0.07327181100845337,
0.029733488336205482,
-0.04272191599011421,
-0.08249637484550476,
-0.1415451467037201,
-0.04993678629398346,
-0.011005163192749023,
0.10754310339689255,
0.07337497919797897,
0.0048001972027122974,
-0.11733713001012802,
0.062058478593826294,
0.13692134618759155,
0.031207585707306862,
0.004062763415277004,
0.028157465159893036,
0.14977529644966125,
-0.10706274956464767,
-0.022463621571660042,
-0.038119975477457047,
-0.054863203316926956,
0.004114252515137196,
0.016883620992302895,
0.08840765058994293,
0.1410384476184845,
0.11468084901571274,
0.047563645988702774,
0.0464191697537899,
0.06561273336410522,
0.1694946140050888,
0.059157438576221466,
-0.10448314249515533,
-0.044678982347249985,
-0.0040070898830890656,
-0.10903503000736237,
0.057307638227939606,
0.16030821204185486,
0.06326017528772354,
-0.14463356137275696,
0.021787412464618683,
-0.038982175290584564,
0.13649246096611023,
0.020638149231672287,
-0.2677258849143982,
-0.008139112964272499,
0.023630544543266296,
-0.0010347915813326836,
-0.012379839085042477,
0.10821118950843811,
-0.040134772658348083,
-0.233198344707489,
-0.12299054861068726,
0.010077533312141895,
0.031144635751843452,
-0.1509784311056137,
0.015542911365628242,
-0.14036494493484497,
0.08027976751327515,
-0.007007129956036806,
0.07418135553598404,
-0.025149788707494736,
0.15060245990753174,
-0.028731435537338257,
0.01628703810274601,
-0.07902143895626068,
-0.047717493027448654,
0.09898673743009567,
-0.0046631391160190105,
0.1931537538766861,
0.005480166990309954,
-0.023713182657957077,
-0.12098433077335358,
-0.05229806900024414,
-0.04967813938856125,
0.010598190128803253,
-0.05373382940888405,
0.0765683576464653,
-0.02441473677754402,
-0.0039579677395522594,
-0.010900177992880344,
0.08942947536706924,
-0.05291692912578583,
0.03636563941836357,
-0.11246588081121445,
-0.05034820735454559,
0.14550213515758514,
-0.09163831174373627,
-0.10174685716629028,
-0.16205860674381256,
0.14137998223304749,
0.15070600807666779,
0.058216437697410583,
-0.04001476243138313,
0.03867831453680992,
-0.019183965399861336,
-0.024241572245955467,
0.07880574464797974,
0.009653856977820396,
0.1324782371520996,
-0.08983246237039566,
0.014327390119433403,
0.14589735865592957,
-0.05275948345661163,
0.016191845759749413,
-0.02304735779762268,
0.12202176451683044,
0.04650457948446274,
0.06189403310418129,
0.018547222018241882,
0.06655703485012054,
0.06466961652040482,
-0.02262885868549347,
0.08456692099571228,
0.030712679028511047,
-0.18644161522388458,
0.058530256152153015,
-0.09805119782686234,
0.22581584751605988,
0.05066308751702309,
0.06047345697879791,
0.2993181645870209,
0.21986234188079834,
-0.05372472479939461,
0.1669820249080658,
0.044286344200372696,
-0.05891284719109535,
-0.21245966851711273,
-0.03684934973716736,
-0.030655447393655777,
0.09436552971601486,
0.15607263147830963,
-0.0981721356511116,
-0.04201313853263855,
-0.00972361396998167,
-0.032264553010463715,
0.020120708271861076,
-0.24663487076759338,
-0.01734781451523304,
0.14379777014255524,
0.10629188269376755,
0.2451348900794983,
-0.006132842972874641,
0.023609744384884834,
0.049030207097530365,
0.018605992197990417,
-0.02483358606696129,
-0.21013511717319489,
0.09079083055257797,
0.006071676965802908,
0.04935038834810257,
0.022885039448738098,
-0.006052911281585693,
0.04500092566013336,
-0.073696069419384,
0.08904470503330231,
-0.08561883866786957,
-0.08341272175312042,
0.2185351401567459,
-0.03945168852806091,
-0.00661163916811347,
0.12917985022068024,
-0.011526807211339474,
-0.1097102016210556,
-0.015364703722298145,
0.027403371408581734,
0.030678823590278625,
-0.030246863141655922,
-0.03609466925263405,
0.024012766778469086,
0.10202405601739883,
-0.04282205551862717,
0.04565315693616867,
0.10240072011947632,
-0.020902957767248154,
0.15945613384246826,
0.13205459713935852,
0.10420060157775879,
0.002927543595433235,
-0.06464727967977524,
0.014349685050547123,
-0.055471502244472504,
0.02962767891585827,
-0.17038846015930176,
-0.0070191239938139915,
0.055695805698633194,
0.04772466421127319,
0.0945243164896965,
0.11333164572715759,
-0.127106174826622,
0.0300484336912632,
0.028996523469686508,
-0.06286120414733887,
-0.06029998138546944,
-0.002275418024510145,
-0.016458535566926003,
-0.008173024281859398,
-0.09947093576192856,
0.07884971052408218,
-0.10555081814527512,
-0.03306307643651962,
0.05025126785039902,
-0.0607193186879158,
-0.12852220237255096,
-0.010904680006206036,
0.1252979338169098,
0.061709314584732056,
-0.05078592896461487,
0.14939077198505402,
0.06109785661101341,
-0.08055379986763,
0.037185851484537125,
0.027442200109362602,
-0.08008874952793121,
-0.10198270529508591,
-0.0004569833690766245,
0.31761088967323303,
0.06076094135642052,
-0.0329466350376606,
-0.11946453154087067,
-0.15002015233039856,
0.04840146750211716,
0.1035679280757904,
0.12359631806612015,
0.011757869273424149,
-0.05322748050093651,
0.02236519381403923,
-0.05275069922208786,
0.03814244270324707,
0.06910209357738495,
-0.03928454965353012,
-0.13761694729328156,
0.0077122850343585014,
0.026647454127669334,
0.10174071043729782,
-0.06771174818277359,
-0.09184598177671432,
-0.18085066974163055,
0.09208621084690094,
-0.03432070091366768,
-0.10890032351016998,
0.027215104550123215,
-0.017406610772013664,
0.014248576015233994,
0.07639352232217789,
-0.047281619161367416,
0.01244808267802,
-0.1517520695924759,
0.07082249224185944,
0.05706808716058731,
0.08926787972450256,
0.000014311663107946515,
-0.054843269288539886,
0.07618319988250732,
-0.05763502046465874,
0.06680037826299667,
-0.053477559238672256,
0.005539732985198498,
0.10781200975179672,
-0.23264040052890778,
-0.021164139732718468,
0.009476077742874622,
-0.04681631922721863,
0.08765807747840881,
-0.19047698378562927,
0.024190550670027733,
-0.08897756040096283,
-0.024605726823210716,
0.01802127994596958,
-0.1086471825838089,
-0.04306677728891373,
0.08475461602210999,
0.037119291722774506,
-0.031288959085941315,
-0.04612116143107414,
-0.019314980134367943,
-0.0914498046040535,
0.053634315729141235,
0.07442525774240494,
-0.0687926784157753,
0.08314394950866699,
-0.05507456883788109,
0.00841207429766655,
-0.052043743431568146,
0.06760627031326294,
-0.012366239912807941,
-0.12672528624534607,
-0.02123171091079712,
-0.044928714632987976,
0.11662110686302185,
-0.023402327671647072,
0.022080281749367714,
0.014599837362766266,
0.0323631577193737,
-0.012065601535141468,
0.05028461292386055,
0.1019197478890419,
0.05136820673942566,
0.014879679307341576,
0.02292765863239765,
0.055746350437402725,
0.0757644772529602,
-0.1134679913520813,
0.06457309424877167,
-0.02098844014108181,
-0.08620109409093857,
0.1013324111700058,
0.06909440457820892,
0.037490107119083405,
0.15593400597572327,
0.22674402594566345,
0.10539932548999786,
-0.03564648702740669,
-0.03126971051096916,
0.12967991828918457,
0.17799612879753113,
-0.07682197540998459,
0.015780627727508545,
-0.0020607721526175737,
-0.017265556380152702,
-0.09849067777395248,
-0.13722245395183563,
-0.060460351407527924,
-0.2453264594078064,
0.1078341007232666,
-0.03288164362311363,
-0.04169659689068794,
0.128489688038826,
0.027952738106250763,
0.03724630922079086,
0.08183616399765015,
-0.12909026443958282,
-0.013460557907819748,
0.07749562710523605,
-0.08914026618003845,
-0.033571500331163406,
-0.17521262168884277,
-0.06771576404571533,
-0.08741120994091034,
-0.15989220142364502,
-0.06844990700483322,
0.029948782175779343,
0.035394806414842606,
0.010386589914560318,
-0.039711855351924896,
-0.01962728053331375,
0.011063394136726856,
-0.0025537724141031504,
-0.04985455423593521,
-0.01753084547817707,
0.021317757666110992,
-0.11333847790956497,
-0.024336790665984154,
0.16320326924324036,
-0.03297848999500275,
-0.18396754562854767,
-0.0405106395483017,
0.2157316505908966,
0.025046708062291145,
0.0590171180665493,
-0.073721744120121,
-0.016323629766702652,
0.021523483097553253,
0.20813441276550293,
0.10171995311975479,
-0.10821312665939331,
0.015457749366760254,
-0.03655189648270607,
0.0013793212128803134,
-0.061893612146377563,
0.10775819420814514,
0.06519263982772827,
-0.07549984753131866,
-0.17567221820354462,
-0.04389495030045509,
-0.08628730475902557,
0.03370477631688118,
-0.14383791387081146,
-0.03786516562104225,
0.1168690100312233,
0.004516853019595146,
-0.053927481174468994,
0.07883694022893906,
-0.17713546752929688,
0.03441957011818886,
-0.04880853369832039,
-0.13215437531471252,
-0.09491758048534393,
-0.10123858600854874,
0.0027463934384286404,
0.08913854509592056,
0.15567956864833832,
-0.06151591241359711,
-0.07471925020217896,
-0.009579092264175415,
-0.028091613203287125,
-0.052700337022542953,
-0.07900123298168182,
0.059512585401535034,
0.0007560851518064737,
0.16147300601005554,
-0.07439453154802322,
0.09558981657028198,
0.09099138528108597,
-0.021246420219540596,
-0.00915549136698246,
0.032866667956113815,
-0.003863809397444129,
-0.07436864078044891,
-0.04970616102218628,
0.02312966249883175,
0.027639856562018394,
0.10846075415611267,
-0.030836544930934906,
-0.1934703141450882,
0.11230092495679855,
0.09140218049287796,
-0.04296138137578964,
-0.046487610787153244,
0.05351927503943443,
-0.07097935676574707,
0.1252279132604599,
0.03444884717464447,
-0.02163051813840866,
0.013762647286057472,
-0.06370721012353897,
0.08370721340179443,
0.11594565212726593,
-0.048265840858221054,
-0.08278503268957138,
-0.06164652109146118,
0.012770666740834713,
0.02961382456123829,
-0.13650155067443848,
-0.21160630881786346,
-0.10802312940359116,
-0.1383298933506012,
0.004740108735859394,
-0.04703504592180252,
0.08498300611972809,
0.12991970777511597,
0.09780163317918777,
-0.011416295543313026,
-0.004867587238550186,
0.018085451796650887,
0.13192623853683472,
-0.11232008039951324,
-0.08192373812198639
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# mixtral-nek-finetune_0.3_all_data_4_lines
This model is a fine-tuned version of [mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.8051
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 1
- num_epochs: 1.0
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:-----:|:---------------:|
| 0.8456 | 0.09 | 1000 | 0.8573 |
| 0.838 | 0.18 | 2000 | 0.8426 |
| 0.8373 | 0.27 | 3000 | 0.8341 |
| 0.8168 | 0.36 | 4000 | 0.8274 |
| 0.8163 | 0.44 | 5000 | 0.8222 |
| 0.8079 | 0.53 | 6000 | 0.8181 |
| 0.8089 | 0.62 | 7000 | 0.8140 |
| 0.8119 | 0.71 | 8000 | 0.8108 |
| 0.8007 | 0.8 | 9000 | 0.8082 |
| 0.809 | 0.89 | 10000 | 0.8062 |
| 0.8084 | 0.98 | 11000 | 0.8051 |
### Framework versions
- PEFT 0.8.2.dev0
- Transformers 4.38.0.dev0
- Pytorch 2.1.2+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1 | {"license": "apache-2.0", "library_name": "peft", "tags": ["generated_from_trainer"], "base_model": "mistralai/Mixtral-8x7B-Instruct-v0.1", "model-index": [{"name": "mixtral-nek-finetune_0.3_all_data_4_lines", "results": []}]} | null | POLYQ/mixtral-nek-finetune_0.3_all_data_4_lines | [
"peft",
"safetensors",
"generated_from_trainer",
"base_model:mistralai/Mixtral-8x7B-Instruct-v0.1",
"license:apache-2.0",
"region:us"
] | 2024-02-07T17:40:12+00:00 | [] | [] | TAGS
#peft #safetensors #generated_from_trainer #base_model-mistralai/Mixtral-8x7B-Instruct-v0.1 #license-apache-2.0 #region-us
| mixtral-nek-finetune\_0.3\_all\_data\_4\_lines
==============================================
This model is a fine-tuned version of mistralai/Mixtral-8x7B-Instruct-v0.1 on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 0.8051
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 16
* eval\_batch\_size: 8
* seed: 42
* gradient\_accumulation\_steps: 2
* total\_train\_batch\_size: 32
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 1
* num\_epochs: 1.0
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* PEFT 0.8.2.dev0
* Transformers 4.38.0.dev0
* Pytorch 2.1.2+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1\n* num\\_epochs: 1.0\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* PEFT 0.8.2.dev0\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.2+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#peft #safetensors #generated_from_trainer #base_model-mistralai/Mixtral-8x7B-Instruct-v0.1 #license-apache-2.0 #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1\n* num\\_epochs: 1.0\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* PEFT 0.8.2.dev0\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.2+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
51,
159,
4,
47
] | [
"passage: TAGS\n#peft #safetensors #generated_from_trainer #base_model-mistralai/Mixtral-8x7B-Instruct-v0.1 #license-apache-2.0 #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1\n* num\\_epochs: 1.0\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* PEFT 0.8.2.dev0\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.2+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.14227882027626038,
0.08426709473133087,
-0.0027504805475473404,
0.07438094168901443,
0.12235257029533386,
0.022700078785419464,
0.10354270786046982,
0.12962345778942108,
-0.056764934211969376,
0.1104850172996521,
0.11824646592140198,
0.06281991302967072,
0.055038873106241226,
0.15144801139831543,
-0.036017369478940964,
-0.28221726417541504,
0.026535360142588615,
-0.011685479432344437,
-0.1013123095035553,
0.11673221737146378,
0.10328413546085358,
-0.09950561821460724,
0.0558135062456131,
-0.005855117924511433,
-0.09582848846912384,
-0.006857587955892086,
-0.014514683745801449,
-0.038468677550554276,
0.11228462308645248,
0.03810110688209534,
0.11262209713459015,
0.010440838523209095,
0.09940049797296524,
-0.24732330441474915,
0.01156590972095728,
0.0840897411108017,
0.026928462088108063,
0.08016689866781235,
0.10326986759901047,
0.013402198441326618,
0.13061873614788055,
-0.07790522277355194,
0.07251665741205215,
0.042243458330631256,
-0.13490062952041626,
-0.3025970757007599,
-0.11645662784576416,
0.07612217217683792,
0.14105547964572906,
0.06891955435276031,
-0.00950002484023571,
0.0878816545009613,
-0.02885015681385994,
0.07566269487142563,
0.24812369048595428,
-0.26511672139167786,
-0.10077182948589325,
0.008073545061051846,
0.043274570256471634,
0.08835968375205994,
-0.106965571641922,
-0.04459739103913307,
0.04811197891831398,
0.049418941140174866,
0.10133723169565201,
0.00582285039126873,
0.04443855211138725,
-0.003196127014234662,
-0.15219399333000183,
-0.04937935248017311,
0.1400367021560669,
0.06593088060617447,
-0.0506872795522213,
-0.05302194505929947,
-0.05412527546286583,
-0.19777627289295197,
-0.04835521802306175,
-0.011621804907917976,
0.020068813115358353,
-0.029869444668293,
-0.00029600050766021013,
-0.015472332015633583,
-0.07095516473054886,
-0.10645526647567749,
0.043446674942970276,
0.19040215015411377,
0.057338353246450424,
-0.012407390400767326,
0.009639226831495762,
0.11889110505580902,
0.014253076165914536,
-0.15291930735111237,
-0.014159459620714188,
0.005670974962413311,
-0.04169956594705582,
-0.03427828848361969,
-0.04638831689953804,
0.011835698038339615,
0.0038055444601923227,
0.18281954526901245,
-0.12097160518169403,
0.08611803501844406,
0.031169017776846886,
0.02312908135354519,
-0.11568480730056763,
0.11397779732942581,
-0.05332028865814209,
-0.02277638390660286,
-0.016238326206803322,
0.12772196531295776,
0.02701376937329769,
-0.0018410120392218232,
-0.06232292205095291,
0.0198687557131052,
0.09436044096946716,
0.0476263128221035,
-0.04198453575372696,
-0.001783824642188847,
-0.06249548867344856,
-0.020394060760736465,
0.06442815810441971,
-0.10167564451694489,
0.03678616136312485,
0.020436478778719902,
-0.05593511089682579,
-0.04569213464856148,
0.011917399242520332,
0.01219048909842968,
-0.0026591659989207983,
0.11865393817424774,
-0.08211024850606918,
-0.002134918700903654,
-0.08919733017683029,
-0.09538788348436356,
0.023363022133708,
-0.01760886237025261,
0.00288461335003376,
-0.08843361586332321,
-0.14019562304019928,
-0.04509498551487923,
0.054691266268491745,
-0.06325346976518631,
-0.059570953249931335,
-0.023648027330636978,
-0.08658076822757721,
0.03723600134253502,
-0.014185615815222263,
0.12631317973136902,
-0.065907321870327,
0.11230222880840302,
0.025053536519408226,
0.039768725633621216,
0.0033517477568238974,
0.03805955871939659,
-0.045074064284563065,
0.0703323483467102,
-0.18734373152256012,
0.02562209963798523,
-0.08650211244821548,
0.040663834661245346,
-0.1387854516506195,
-0.1068539246916771,
-0.039202459156513214,
-0.011968404054641724,
0.12231874465942383,
0.13463623821735382,
-0.1318283975124359,
-0.05301458016037941,
0.17087401449680328,
-0.10933518409729004,
-0.1422116905450821,
0.10834866762161255,
-0.013762886635959148,
-0.004830801393836737,
0.017936697229743004,
0.12811465561389923,
0.11887750774621964,
-0.12646286189556122,
0.01934828609228134,
-0.03939921036362648,
0.15295931696891785,
0.015560449101030827,
0.10981203615665436,
-0.037802498787641525,
-0.031649213284254074,
-0.004761900752782822,
-0.08398774266242981,
0.05417449027299881,
-0.09539848566055298,
-0.0752171128988266,
-0.036767035722732544,
-0.0851326584815979,
0.026838399469852448,
0.04979676380753517,
0.024749046191573143,
-0.09131192415952682,
-0.1076047345995903,
0.04753486439585686,
0.1254579722881317,
-0.0586639828979969,
0.021824272349476814,
-0.03665460646152496,
0.09473390132188797,
-0.0504530631005764,
-0.023615123704075813,
-0.16222316026687622,
-0.06093680113554001,
0.024137891829013824,
-0.046132538467645645,
-0.03079778514802456,
-0.052434880286455154,
0.08271501958370209,
0.09521389752626419,
-0.07537499815225601,
-0.06829122453927994,
-0.10793577879667282,
-0.000686601793859154,
-0.09850587695837021,
-0.213290736079216,
-0.0825963243842125,
-0.03318554535508156,
0.14688117802143097,
-0.24995826184749603,
0.029135307297110558,
0.0005317823961377144,
0.1201062798500061,
0.03678367659449577,
-0.05064128711819649,
-0.02420220710337162,
0.0727555900812149,
-0.021367037668824196,
-0.08016146719455719,
0.039459243416786194,
0.007515698671340942,
-0.08332381397485733,
-0.031048379838466644,
-0.13065040111541748,
0.14065715670585632,
0.08629725873470306,
0.016656452789902687,
-0.09770318865776062,
-0.06539936363697052,
-0.07477772980928421,
-0.046766869723796844,
-0.03393811360001564,
0.008094976656138897,
0.10091307759284973,
0.027543170377612114,
0.1237616017460823,
-0.0902995690703392,
-0.03902619332075119,
0.047113317996263504,
-0.009273845702409744,
-0.003533961484208703,
0.12460215389728546,
0.0930711179971695,
-0.036150529980659485,
0.11699363589286804,
0.16105791926383972,
-0.05355723202228546,
0.09883367270231247,
-0.06430321931838989,
-0.10957951098680496,
-0.024514028802514076,
0.06470489501953125,
0.027585741132497787,
0.14166302978992462,
-0.05590823292732239,
0.027867969125509262,
0.005493908654898405,
0.03354460746049881,
0.013815329410135746,
-0.21121163666248322,
-0.027542216703295708,
0.03783481568098068,
-0.04854390025138855,
-0.01123786810785532,
-0.018583688884973526,
-0.014376765117049217,
0.0941818505525589,
0.019336897879838943,
-0.04983367398381233,
-0.010929720476269722,
-0.0033716473262757063,
-0.09018396586179733,
0.2046278417110443,
-0.10820646584033966,
-0.08926888555288315,
-0.1396166980266571,
0.00573516171425581,
-0.037612199783325195,
-0.004749917890876532,
0.047869954258203506,
-0.07340025156736374,
-0.030142564326524734,
-0.08662153780460358,
0.014944415539503098,
-0.021182600408792496,
0.017751844599843025,
0.004892151802778244,
-0.012622763402760029,
0.0961766242980957,
-0.08504556119441986,
0.013049214147031307,
0.0026558469980955124,
-0.036510124802589417,
0.011343204416334629,
0.02803211659193039,
0.10131218284368515,
0.13652198016643524,
0.03919559717178345,
0.013563461601734161,
-0.02644660696387291,
0.2323852926492691,
-0.09224622696638107,
0.00027481955476105213,
0.13295000791549683,
0.0000142695898830425,
0.05767538398504257,
0.11760411411523819,
0.05881338194012642,
-0.10878767818212509,
0.02553052455186844,
0.04339032247662544,
-0.017344463616609573,
-0.21268673241138458,
-0.039361655712127686,
-0.05362488701939583,
-0.03936992213129997,
0.13358335196971893,
0.04833560809493065,
-0.006998543161898851,
0.04406868293881416,
-0.036540571600198746,
0.013196676038205624,
0.001915721339173615,
0.09174279123544693,
0.050073519349098206,
0.04100256413221359,
0.09796224534511566,
-0.024618886411190033,
0.0003006088372785598,
0.04634102061390877,
0.014460605569183826,
0.22261033952236176,
0.001888084807433188,
0.09561628848314285,
0.050160784274339676,
0.17173494398593903,
-0.023480795323848724,
0.07365275919437408,
0.007209460716694593,
-0.028037838637828827,
0.004618334583938122,
-0.06907856464385986,
-0.033962298184633255,
0.05345942825078964,
-0.05023505166172981,
0.07196123152971268,
-0.13590823113918304,
-0.021381139755249023,
0.025128155946731567,
0.3434113562107086,
0.06769829988479614,
-0.3110734522342682,
-0.10930074751377106,
0.012176750227808952,
-0.03175279125571251,
-0.0660204291343689,
0.03420635312795639,
0.12696969509124756,
-0.06019289791584015,
0.06810139864683151,
-0.08233080059289932,
0.0868823230266571,
0.0014767332468181849,
0.0047014374285936356,
0.10261095315217972,
0.0966355949640274,
-0.01171025913208723,
0.03889986872673035,
-0.23441649973392487,
0.28711190819740295,
0.0066982745192945,
0.07771392166614532,
-0.01576952636241913,
0.02352076955139637,
0.0341668576002121,
0.045831937342882156,
0.06788533180952072,
-0.0010254119988530874,
-0.0917505994439125,
-0.20369084179401398,
-0.08464577049016953,
0.02676606923341751,
0.11682067066431046,
-0.07239339500665665,
0.12132712453603745,
-0.04154379665851593,
0.003956569358706474,
0.04720776528120041,
-0.03749915584921837,
-0.1037038266658783,
-0.07263593375682831,
0.023165883496403694,
-0.0006809298065491021,
0.028495289385318756,
-0.11598755419254303,
-0.09627602249383926,
-0.039407193660736084,
0.08704651892185211,
-0.10250388085842133,
-0.027528781443834305,
-0.1410076916217804,
0.06888724118471146,
0.1374056041240692,
-0.0699128806591034,
0.0541691780090332,
0.0099501246586442,
0.10054432600736618,
0.016652550548315048,
-0.01216775830835104,
0.10637680441141129,
-0.08573053777217865,
-0.22487814724445343,
-0.039040904492139816,
0.15438106656074524,
0.034332070499658585,
0.04223396256566048,
-0.03687620162963867,
0.04469182714819908,
0.0034564861562103033,
-0.09867176413536072,
0.02303399331867695,
0.010914009064435959,
0.03988845273852348,
0.0289829783141613,
-0.06210951507091522,
0.0779493898153305,
-0.050844766199588776,
-0.04759776592254639,
0.0729508250951767,
0.3328075408935547,
-0.08823473751544952,
0.0002870790776796639,
0.03217385336756706,
-0.054196394979953766,
-0.14762941002845764,
0.014437118545174599,
0.10615731030702591,
-0.0028167322743684053,
0.042828477919101715,
-0.20276136696338654,
0.054575853049755096,
0.12824378907680511,
-0.036094747483730316,
0.1487349569797516,
-0.3163071572780609,
-0.12368916720151901,
0.08340585231781006,
0.14537379145622253,
0.017447134479880333,
-0.1927555501461029,
-0.06618154048919678,
0.0030703707598149776,
-0.12172193080186844,
0.045667171478271484,
-0.07335976511240005,
0.08292921632528305,
-0.01806638389825821,
0.01508155558258295,
0.019569311290979385,
-0.04521487280726433,
0.1909055858850479,
-0.03597262129187584,
0.08478544652462006,
-0.022672023624181747,
0.05586029589176178,
-0.017565494403243065,
-0.0716090276837349,
0.014138882048428059,
-0.07419830560684204,
0.038843706250190735,
-0.13293862342834473,
-0.011101090349256992,
-0.08688993006944656,
0.02099018543958664,
-0.05936156585812569,
-0.0370611809194088,
-0.019423607736825943,
0.051293838769197464,
0.046570271253585815,
-0.010596854612231255,
0.16167116165161133,
-0.013008841313421726,
0.20962224900722504,
0.11844242364168167,
0.04031899943947792,
-0.0181977991014719,
-0.06739232689142227,
0.0019802055321633816,
-0.02468009479343891,
0.06737459450960159,
-0.16322806477546692,
0.022025708109140396,
0.12569423019886017,
0.035491134971380234,
0.13371959328651428,
0.058304447680711746,
-0.06415367126464844,
-0.0023611870128661394,
0.07490396499633789,
-0.11972308903932571,
-0.1350090056657791,
-0.021931931376457214,
0.07489626854658127,
-0.1441006064414978,
0.015319942496716976,
0.12248574942350388,
-0.07738599926233292,
-0.008732412941753864,
0.003331169718876481,
0.046418577432632446,
-0.045893363654613495,
0.2086840569972992,
0.05420477315783501,
0.07874390482902527,
-0.06887012720108032,
0.08949421346187592,
0.049764540046453476,
-0.1164446696639061,
0.01407236885279417,
0.09736822545528412,
-0.06429964303970337,
-0.030724311247467995,
0.04131266847252846,
0.05997828021645546,
-0.01062041986733675,
-0.05512094870209694,
-0.11838402599096298,
-0.12488394975662231,
0.07715438306331635,
0.11332172155380249,
0.04639185220003128,
0.013261700980365276,
0.008341382257640362,
0.03533955290913582,
-0.10462042689323425,
0.10442891716957092,
0.045149195939302444,
0.0898793488740921,
-0.1460028737783432,
0.10807973146438599,
0.000015754436390125193,
0.0025205195415765047,
-0.000713719695340842,
0.02981197088956833,
-0.11135667562484741,
0.0026762315537780523,
-0.14620481431484222,
-0.02309229224920273,
-0.041618961840867996,
0.002025892259553075,
-0.004281666129827499,
-0.06099717691540718,
-0.06200948357582092,
0.02479523792862892,
-0.10825817286968231,
-0.04809340834617615,
-0.017667243257164955,
0.06070266664028168,
-0.11952350288629532,
-0.015607175417244434,
0.04326789826154709,
-0.11778213828802109,
0.07612370699644089,
0.024888819083571434,
0.04869198054075241,
0.044411804527044296,
-0.08612747490406036,
0.051845185458660126,
0.03135671466588974,
-0.013677950017154217,
0.01786397024989128,
-0.16435258090496063,
-0.00954335555434227,
-0.03941831737756729,
0.024911431595683098,
0.0013990928418934345,
0.05610637739300728,
-0.1367209255695343,
-0.020938504487276077,
-0.02651718445122242,
-0.05354807525873184,
-0.04079410061240196,
0.026800526306033134,
0.06701589375734329,
0.026485159993171692,
0.13700829446315765,
-0.0870833545923233,
0.033356621861457825,
-0.23059143126010895,
-0.012689007446169853,
-0.030522583052515984,
-0.05674735829234123,
-0.07793804258108139,
-0.022970043122768402,
0.07288434356451035,
-0.047269292175769806,
0.05884162709116936,
-0.06155974417924881,
0.08657599240541458,
0.0477902926504612,
-0.08919211477041245,
0.015593825839459896,
0.03814325109124184,
0.20162628591060638,
0.041188180446624756,
-0.02325418032705784,
0.04345778748393059,
0.02582808956503868,
0.059680044651031494,
0.11662363260984421,
0.17946413159370422,
0.17638254165649414,
0.009033276699483395,
0.07903485000133514,
0.03339286521077156,
-0.12270796298980713,
-0.10837206989526749,
0.07411307841539383,
-0.0027841473929584026,
0.1100393608212471,
-0.015008046291768551,
0.19543828070163727,
0.12407863885164261,
-0.20393605530261993,
0.02815297059714794,
-0.04404791072010994,
-0.06793125718832016,
-0.11855863034725189,
-0.010054809972643852,
-0.06379012763500214,
-0.18312132358551025,
-0.007464445196092129,
-0.11318328976631165,
0.02679147571325302,
0.07433368265628815,
0.01955266296863556,
0.024765701964497566,
0.16026219725608826,
0.058970559388399124,
0.013421080075204372,
0.07295146584510803,
0.031238051131367683,
0.007810460403561592,
-0.04968805983662605,
-0.12381330877542496,
0.0352444164454937,
-0.0649692490696907,
0.0394277423620224,
-0.05039917677640915,
-0.08248069882392883,
0.05474778637290001,
0.02984834462404251,
-0.10051573812961578,
0.036924876272678375,
0.010649984702467918,
0.04378984123468399,
0.10480853915214539,
0.018092770129442215,
0.00880747102200985,
-0.027120383456349373,
0.25160732865333557,
-0.08666540682315826,
-0.028564954176545143,
-0.12753111124038696,
0.2690596580505371,
0.022392671555280685,
-0.007694949861615896,
0.017202790826559067,
-0.08825205266475677,
-0.004359856713563204,
0.12467300891876221,
0.10004625469446182,
-0.054463036358356476,
-0.011071664281189442,
0.018094174563884735,
-0.009025133214890957,
-0.04858848825097084,
0.102284736931324,
0.12968610227108002,
0.04170233756303787,
-0.06864640861749649,
-0.0213323924690485,
-0.04805884510278702,
-0.016164183616638184,
-0.03842661902308464,
0.05655167996883392,
0.015285196714103222,
-0.0033498380798846483,
-0.03488777205348015,
0.10844354331493378,
-0.014612352475523949,
-0.11920379847288132,
0.07546166330575943,
-0.1731872409582138,
-0.1853262335062027,
-0.037425048649311066,
0.040120720863342285,
0.0018246592953801155,
0.0567171536386013,
-0.0187975000590086,
-0.034486956894397736,
0.12936633825302124,
-0.018277762457728386,
-0.016598733142018318,
-0.1693487912416458,
0.08971842378377914,
-0.06835634261369705,
0.2308371216058731,
-0.03214993700385094,
-0.001094588777050376,
0.11789605021476746,
0.032990626990795135,
-0.11956154555082321,
0.04186449944972992,
0.09254786372184753,
-0.0975499376654625,
0.011001212522387505,
0.13199998438358307,
-0.05169421061873436,
0.11690205335617065,
0.043252088129520416,
-0.13175687193870544,
-0.013482592068612576,
-0.029458383098244667,
-0.07670052349567413,
-0.04746509715914726,
-0.014161692932248116,
-0.05345863848924637,
0.14165249466896057,
0.20635274052619934,
-0.06482476741075516,
-0.00491930078715086,
-0.04820197448134422,
0.05209419131278992,
0.0539492703974247,
0.13282573223114014,
-0.011376094073057175,
-0.25035494565963745,
0.03039364516735077,
0.0267924927175045,
0.0025108051486313343,
-0.2572990655899048,
-0.08459319174289703,
0.037519022822380066,
-0.04807547852396965,
-0.06824588775634766,
0.11748648434877396,
0.04896634444594383,
0.0451655387878418,
-0.058528583496809006,
-0.14280067384243011,
-0.0660630464553833,
0.18745668232440948,
-0.13451462984085083,
-0.06850792467594147
] |
null | null | fastai |
# Amazing!
🥳 Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the [documentation here](https://huggingface.co/docs/hub/model-repos))!
2. Create a demo in Gradio or Streamlit using 🤗 Spaces ([documentation here](https://huggingface.co/docs/hub/spaces)).
3. Join the fastai community on the [Fastai Discord](https://discord.com/invite/YKrxeNn)!
Greetings fellow fastlearner 🤝! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| {"tags": ["fastai"]} | null | crrodrvi/Practica1 | [
"fastai",
"region:us"
] | 2024-02-07T17:40:29+00:00 | [] | [] | TAGS
#fastai #region-us
|
# Amazing!
Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the documentation here)!
2. Create a demo in Gradio or Streamlit using Spaces (documentation here).
3. Join the fastai community on the Fastai Discord!
Greetings fellow fastlearner ! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| [
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
"TAGS\n#fastai #region-us \n",
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
9,
20,
79,
3,
6,
12,
8
] | [
"passage: TAGS\n#fastai #region-us \n# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---# Model card## Model description\nMore information needed## Intended uses & limitations\nMore information needed## Training and evaluation data\nMore information needed"
] | [
-0.073318250477314,
-0.035918332636356354,
0.0016039619222283363,
0.09830865263938904,
0.16935402154922485,
0.11954792588949203,
0.06504721194505692,
0.08469552546739578,
0.09305626899003983,
0.008462822064757347,
0.08902737498283386,
-0.059808652848005295,
0.09601042419672012,
0.26935747265815735,
0.06010362133383751,
-0.24278773367404938,
0.02870224229991436,
-0.0036573195829987526,
0.08660013228654861,
0.06588653475046158,
0.12898924946784973,
-0.039593055844306946,
0.14736801385879517,
-0.018255524337291718,
-0.19320440292358398,
-0.054476846009492874,
-0.015185145661234856,
-0.019686169922351837,
0.12385433167219162,
-0.04793357476592064,
0.030790239572525024,
0.0026993011124432087,
-0.0015684126410633326,
-0.0995422899723053,
0.06401026993989944,
0.04089692234992981,
0.028817683458328247,
0.055760785937309265,
-0.04539911448955536,
0.08392030745744705,
0.054179996252059937,
-0.010920286178588867,
-0.12179892510175705,
0.09588204324245453,
-0.1474396139383316,
-0.2022949457168579,
-0.1278105229139328,
-0.11345728486776352,
0.047258179634809494,
0.01006549596786499,
-0.01907140202820301,
0.12847048044204712,
-0.14997079968452454,
-0.03727749362587929,
0.17807333171367645,
-0.15483331680297852,
-0.050517335534095764,
-0.0010879677720367908,
0.06801971048116684,
-0.06002732738852501,
-0.05137069150805473,
0.0968702957034111,
0.0906822457909584,
-0.019289257004857063,
0.015487968921661377,
0.0037353564985096455,
0.035032227635383606,
0.002429646672680974,
-0.0558350533246994,
0.06529499590396881,
-0.027788599953055382,
0.055927276611328125,
-0.1094130128622055,
-0.11809343844652176,
0.0010178228840231895,
0.03238791227340698,
-0.05549647659063339,
-0.06733305007219315,
0.0810781940817833,
0.007735111750662327,
-0.0603058859705925,
-0.11863275617361069,
-0.06696899980306625,
-0.12959590554237366,
0.00783742405474186,
0.09659197926521301,
0.0033950558863580227,
0.06878509372472763,
-0.09986882656812668,
0.06626693904399872,
-0.2048133760690689,
-0.04758621007204056,
-0.08781389147043228,
-0.1065201610326767,
0.02003002166748047,
-0.04773771017789841,
0.04778444394469261,
0.15393073856830597,
0.14042632281780243,
0.04171324521303177,
0.05645250529050827,
-0.029350629076361656,
0.038715146481990814,
0.04752078279852867,
0.018331103026866913,
0.03540196642279625,
-0.020549163222312927,
-0.18507646024227142,
0.0004176131042186171,
-0.04207618162035942,
0.08488372713327408,
-0.07463551312685013,
-0.05029602348804474,
0.01336510106921196,
-0.12160550057888031,
0.09655242413282394,
-0.05178983509540558,
-0.005084214266389608,
0.0036863412242382765,
0.008919943124055862,
0.20647431910037994,
0.04232640564441681,
0.004936119541525841,
-0.006976569537073374,
-0.1375076025724411,
-0.051532845944166183,
-0.09289269894361496,
0.034273598343133926,
0.02420172467827797,
0.01303885504603386,
-0.07711919397115707,
0.049177106469869614,
-0.046599894762039185,
-0.008231878280639648,
0.021442487835884094,
-0.20236440002918243,
0.010869519785046577,
-0.0969783291220665,
-0.1469350904226303,
0.06343341618776321,
0.0026821133214980364,
-0.07499043643474579,
0.08385025709867477,
-0.004780351184308529,
0.031972795724868774,
-0.030242523178458214,
-0.00177793821785599,
0.05239185318350792,
-0.08095952123403549,
0.023147141560912132,
0.1995297074317932,
0.10590710490942001,
-0.07641816139221191,
-0.0025978393387049437,
-0.12475098669528961,
0.04128078371286392,
-0.14157716929912567,
0.038516461849212646,
-0.08163458108901978,
0.15109841525554657,
-0.044047996401786804,
0.018007883802056313,
-0.0071970620192587376,
0.08468028157949448,
0.07606321573257446,
0.19981153309345245,
-0.23198086023330688,
-0.053279466927051544,
0.16512827575206757,
-0.11487894505262375,
-0.18565405905246735,
0.20080815255641937,
-0.00043150142300873995,
0.10752102732658386,
-0.010421866551041603,
0.17009462416172028,
-0.021746216341853142,
-0.14181379973888397,
-0.032203078269958496,
-0.0012119774473831058,
-0.24691128730773926,
-0.08980891108512878,
0.09945957362651825,
0.10481112450361252,
-0.059047527611255646,
0.029137471690773964,
0.012005627155303955,
0.15818172693252563,
-0.07679074257612228,
-0.04601999372243881,
-0.007829579524695873,
-0.10506698489189148,
0.022122014313936234,
0.01663162000477314,
0.034775324165821075,
-0.059334270656108856,
-0.00890427641570568,
-0.07678428292274475,
0.13092219829559326,
0.09849999099969864,
-0.03540538251399994,
-0.06064159423112869,
0.16454961895942688,
-0.0640924945473671,
-0.026323838159441948,
0.08331746608018875,
-0.08536569774150848,
0.047215063124895096,
0.04028964787721634,
0.05084947869181633,
0.009997997432947159,
0.09182237833738327,
0.0698544830083847,
0.006789602339267731,
0.03368524834513664,
0.13270887732505798,
-0.027426021173596382,
-0.05121328681707382,
0.01674247533082962,
0.04598715528845787,
-0.00979064591228962,
0.3169313669204712,
-0.19912512600421906,
0.018945744261145592,
-0.06457886099815369,
0.08035559207201004,
0.0660853385925293,
0.007019065320491791,
0.07570107281208038,
-0.05360652506351471,
-0.016966497525572777,
-0.045681122690439224,
0.06926878541707993,
-0.06979862600564957,
-0.054223138839006424,
0.2564660608768463,
-0.031106717884540558,
0.031359151005744934,
0.10653062164783478,
-0.06802138686180115,
-0.05823708325624466,
-0.02224794402718544,
-0.0014688228257000446,
0.023401014506816864,
-0.04168177396059036,
0.06067536398768425,
-0.08815024048089981,
-0.05285300314426422,
0.1703105866909027,
-0.038786694407463074,
0.07842917740345001,
0.035427022725343704,
-0.05379872769117355,
-0.04481838271021843,
0.061976201832294464,
0.14977918565273285,
-0.0965908095240593,
0.06779327243566513,
0.13305115699768066,
0.014980388805270195,
0.15411095321178436,
0.07098863273859024,
-0.07586279511451721,
-0.08855607360601425,
-0.018246978521347046,
-0.004062598571181297,
0.18133139610290527,
-0.07897800207138062,
-0.036732085049152374,
0.042683616280555725,
-0.011134039610624313,
0.06611642241477966,
-0.05846851319074631,
-0.0792742595076561,
0.01736506260931492,
-0.0582035630941391,
0.018060972914099693,
0.12486616522073746,
-0.08240851759910583,
0.04267239198088646,
0.03745635226368904,
-0.058472223579883575,
0.046025440096855164,
0.0389089435338974,
-0.01086228247731924,
0.05541912093758583,
0.06821268051862717,
-0.2134213149547577,
-0.10377796739339828,
-0.17595313489437103,
0.03000609390437603,
0.020109420642256737,
0.036413755267858505,
-0.10920769721269608,
0.02131613902747631,
-0.0651998370885849,
-0.07437032461166382,
0.04871295765042305,
-0.029500357806682587,
-0.10847225040197372,
-0.027001040056347847,
-0.024241603910923004,
-0.04816099628806114,
-0.021433888003230095,
-0.06250716745853424,
0.03129231557250023,
0.04526861384510994,
0.03191622346639633,
0.1321185976266861,
-0.010805734433233738,
-0.014524625614285469,
0.002761868294328451,
-0.017431288957595825,
0.1497519314289093,
-0.13988617062568665,
0.06941607594490051,
0.1812426596879959,
0.09771130234003067,
0.03844839334487915,
0.01466822624206543,
0.03106272965669632,
-0.07663184404373169,
0.005383877083659172,
0.034619297832250595,
-0.0891294777393341,
-0.08207139372825623,
-0.01874193549156189,
-0.03897557035088539,
0.21049608290195465,
-0.12441039085388184,
0.024025630205869675,
0.040357187390327454,
0.09686839580535889,
0.11187659204006195,
-0.04121972620487213,
-0.17262403666973114,
0.04177050292491913,
-0.2474004179239273,
-0.051238708198070526,
0.003026821883395314,
-0.09497712552547455,
-0.06320231407880783,
0.18337351083755493,
0.0052159554325044155,
0.0287664532661438,
0.00430127140134573,
0.12202860414981842,
-0.0009366215672343969,
0.12068869173526764,
0.0687243714928627,
-0.05316835641860962,
0.02255408652126789,
-0.09993521869182587,
-0.0696573555469513,
-0.03704388439655304,
-0.07047778367996216,
0.06136435270309448,
0.12800902128219604,
-0.024759603664278984,
-0.04259653389453888,
0.04763835668563843,
0.09553752839565277,
0.06145815551280975,
0.15860231220722198,
-0.16057826578617096,
-0.022865094244480133,
0.042546581476926804,
-0.029262376949191093,
-0.049140751361846924,
-0.009500340558588505,
0.08492209017276764,
-0.05378608778119087,
-0.02665375918149948,
0.003306680591776967,
0.07226359844207764,
-0.0019794153049588203,
0.0436936691403389,
-0.03244423121213913,
0.1845880150794983,
-0.029572106897830963,
0.023350762203335762,
-0.12604808807373047,
0.13696090877056122,
0.022422920912504196,
-0.015438690781593323,
-0.06568175554275513,
-0.05596291273832321,
0.18064838647842407,
0.02166406810283661,
0.11738308519124985,
0.011424299329519272,
-0.09442766010761261,
-0.1337079405784607,
-0.1388736516237259,
0.015837913379073143,
0.09729303419589996,
-0.01256689801812172,
-0.03353166952729225,
0.019608711823821068,
-0.04281611740589142,
-0.06777504086494446,
0.10452067106962204,
-0.11668688803911209,
-0.0018522912869229913,
0.005423946306109428,
0.0416572242975235,
-0.06085909157991409,
0.032720211893320084,
0.03296784311532974,
-0.0647648349404335,
0.121244877576828,
0.24137550592422485,
0.1064029112458229,
-0.09990023821592331,
-0.08652417361736298,
0.021780110895633698,
-0.034567005932331085,
-0.0014182132435962558,
-0.016133872792124748,
0.036385562270879745,
0.0019662054255604744,
0.003586959559470415,
0.13572031259536743,
-0.07582411170005798,
0.012567305937409401,
-0.08275366574525833,
0.07902812212705612,
-0.0409930944442749,
-0.0025117802433669567,
-0.003995150327682495,
-0.02950184792280197,
-0.03430648893117905,
-0.06180789694190025,
0.163230761885643,
-0.06168964132666588,
-0.08240502327680588,
0.07821446657180786,
0.01680770143866539,
0.017550375312566757,
-0.06227098032832146,
-0.054205916821956635,
0.1972212791442871,
0.31792324781417847,
-0.058273475617170334,
0.10361375659704208,
0.1383560746908188,
0.023166829720139503,
-0.22579050064086914,
0.036502011120319366,
-0.14466507732868195,
0.032058101147413254,
0.024782279506325722,
-0.06415819376707077,
0.05856261029839516,
0.1250556856393814,
-0.045668914914131165,
0.23617008328437805,
-0.03641456738114357,
-0.07633192092180252,
-0.013243574649095535,
0.043972890824079514,
0.3091393709182739,
-0.11325396597385406,
-0.02349173277616501,
-0.11636991053819656,
-0.21521669626235962,
0.06708590686321259,
-0.16208602488040924,
0.1406344771385193,
-0.05703224614262581,
0.023474344983696938,
-0.012111215852200985,
-0.07578689604997635,
0.19497497379779816,
-0.1371963620185852,
0.056931521743535995,
-0.1432308852672577,
-0.11647364497184753,
-0.005183211527764797,
-0.08439649641513824,
0.14731425046920776,
-0.08327576518058777,
-0.02632858417928219,
-0.2082071304321289,
0.001373599166981876,
-0.021641740575432777,
0.09738951921463013,
0.02311836928129196,
-0.07967846095561981,
-0.08035353571176529,
0.12579506635665894,
-0.07811200618743896,
0.036513522267341614,
-0.08704032748937607,
-0.03989429399371147,
-0.026884159073233604,
-0.08092786371707916,
0.06243825703859329,
-0.08906654268503189,
0.16072829067707062,
-0.049172405153512955,
-0.046159181743860245,
0.061650797724723816,
-0.20832203328609467,
0.026940656825900078,
0.036382775753736496,
-0.031731411814689636,
0.10237374156713486,
-0.029687397181987762,
-0.07129550725221634,
0.1133488118648529,
0.13133300840854645,
-0.07154961675405502,
-0.2563934028148651,
-0.0821671262383461,
-0.008923565037548542,
0.04608851298689842,
0.0829237625002861,
0.04836045205593109,
-0.05231332778930664,
-0.017525162547826767,
-0.031239798292517662,
0.03463910520076752,
-0.11768791079521179,
-0.02900020219385624,
0.06892099231481552,
0.0014350401470437646,
-0.09527117758989334,
0.0962897539138794,
-0.004287306685000658,
-0.02237984538078308,
-0.009249147027730942,
0.1892271637916565,
-0.014808090403676033,
-0.12871821224689484,
-0.057921428233385086,
0.24053727090358734,
-0.038428641855716705,
-0.07654319703578949,
-0.06858045607805252,
-0.011265470646321774,
-0.04038287326693535,
0.06209278851747513,
0.04795577749609947,
-0.01209679339081049,
0.08278531581163406,
0.06026776134967804,
-0.1221788227558136,
-0.060724351555109024,
-0.05533421039581299,
0.035240933299064636,
-0.09762322902679443,
0.04652146250009537,
0.016370195895433426,
0.12453475594520569,
-0.09184806793928146,
-0.03038635104894638,
-0.11205437779426575,
-0.059142544865608215,
-0.18314886093139648,
-0.0571221299469471,
-0.041237685829401016,
-0.008055833168327808,
0.03931373730301857,
0.02697678469121456,
-0.04493580758571625,
-0.048296377062797546,
-0.06704439222812653,
0.03899036720395088,
0.07422684133052826,
0.026717372238636017,
-0.03390409052371979,
0.05009619519114494,
0.06439550966024399,
0.008286280557513237,
0.1963774412870407,
0.06738202273845673,
0.061680130660533905,
-0.025940580293536186,
-0.19781054556369781,
-0.05686524137854576,
0.002742079785093665,
-0.09212438762187958,
0.12195391207933426,
-0.011633808724582195,
0.02040605992078781,
-0.06281229853630066,
0.03727225586771965,
0.026594331488013268,
0.10702691227197647,
-0.02029390074312687,
0.0958021730184555,
0.029817266389727592,
-0.08947111666202545,
-0.044351425021886826,
0.015944788232445717,
0.12201714515686035,
0.02899266965687275,
0.028689615428447723,
0.015606578439474106,
0.037100955843925476,
-0.03902486339211464,
0.0296308696269989,
-0.045808494091033936,
-0.14955224096775055,
0.01991276629269123,
-0.046732377260923386,
-0.006942411884665489,
-0.016697930172085762,
0.18722283840179443,
0.04047711566090584,
-0.046649303287267685,
-0.01265130564570427,
0.014551439322531223,
-0.004945865832269192,
-0.03270510211586952,
-0.004582806024700403,
0.06002182513475418,
-0.004176365211606026,
-0.047248490154743195,
0.13213102519512177,
0.046804413199424744,
0.04763852432370186,
0.0742364451289177,
0.09783162921667099,
-0.00930761732161045,
0.13372060656547546,
0.06815905123949051,
-0.01982966810464859,
-0.1131899505853653,
-0.05649255961179733,
-0.11679257452487946,
0.034573014825582504,
-0.05576380714774132,
0.12528598308563232,
0.11196581274271011,
-0.060735806822776794,
-0.03883470967411995,
-0.0771038830280304,
-0.03134944289922714,
-0.07594948261976242,
0.03614310547709465,
-0.0327751524746418,
-0.08104247599840164,
0.06421366333961487,
0.05536265671253204,
-0.036099426448345184,
0.11491319537162781,
0.020650042220950127,
-0.05702126771211624,
0.12617406249046326,
-0.07743373513221741,
0.10717736184597015,
0.07707828283309937,
-0.05362870916724205,
-0.12441752851009369,
0.011045942083001137,
-0.07996662706136703,
-0.11546584963798523,
-0.008837178349494934,
-0.011918267235159874,
-0.0746825784444809,
-0.05780024081468582,
0.10738345980644226,
-0.03462931141257286,
-0.09724929928779602,
-0.020749187096953392,
0.015756776556372643,
0.056543223559856415,
-0.019683608785271645,
0.0018315898487344384,
0.03772254288196564,
0.028699718415737152,
0.15574465692043304,
-0.0016714793164283037,
0.06267286092042923,
-0.1358945369720459,
0.18023191392421722,
-0.1432318240404129,
-0.027932528406381607,
-0.187766894698143,
-0.0886974111199379,
-0.025430310517549515,
0.22427266836166382,
0.26061514019966125,
-0.1923753172159195,
-0.03171071037650108,
0.004376344382762909,
-0.010204915888607502,
-0.07923580706119537,
0.14464490115642548,
0.02417137287557125,
-0.007147552911192179,
-0.06552806496620178,
-0.014752711169421673,
0.024085145443677902,
-0.07228498160839081,
-0.035760894417762756,
0.18496830761432648,
0.0086367791518569,
0.07214809954166412,
-0.09064984321594238,
0.03641578182578087,
-0.18433186411857605,
-0.0693570077419281,
-0.03508331999182701,
-0.138646200299263,
-0.09639570862054825,
-0.01481159869581461,
0.003136083483695984,
0.09603974968194962,
0.03350212052464485,
-0.01305394247174263,
0.06808507442474365,
-0.049502357840538025,
0.010726232081651688,
-0.16043636202812195,
-0.020468583330512047,
0.05376148223876953,
-0.052667658776044846,
0.23897892236709595,
-0.02351270616054535,
-0.12297288328409195,
0.08416848629713058,
-0.03519788756966591,
-0.12302011996507645,
0.0745280459523201,
-0.023310834541916847,
-0.10405170172452927,
-0.05555706471204758,
0.17993386089801788,
-0.01256539486348629,
-0.16247478127479553,
0.03247550129890442,
-0.15925332903862,
0.029797034338116646,
0.03576231747865677,
-0.011352102272212505,
-0.05518606677651405,
0.028951244428753853,
-0.027475930750370026,
0.10062393546104431,
0.14163273572921753,
0.017354421317577362,
-0.009662404656410217,
-0.06593839079141617,
0.09352979063987732,
0.06211914122104645,
-0.07753235101699829,
-0.11338558793067932,
-0.09994973242282867,
0.02616780437529087,
0.07790441066026688,
-0.08538854867219925,
-0.17278192937374115,
-0.029272083193063736,
-0.11865141987800598,
-0.002084053121507168,
0.0349934957921505,
0.06834512948989868,
0.2863384187221527,
0.06974043697118759,
0.004092831164598465,
-0.15255671739578247,
0.05762675032019615,
0.08219972252845764,
-0.02544020675122738,
-0.08790270239114761
] |
null | null | fastai |
# Amazing!
🥳 Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the [documentation here](https://huggingface.co/docs/hub/model-repos))!
2. Create a demo in Gradio or Streamlit using 🤗 Spaces ([documentation here](https://huggingface.co/docs/hub/spaces)).
3. Join the fastai community on the [Fastai Discord](https://discord.com/invite/YKrxeNn)!
Greetings fellow fastlearner 🤝! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| {"tags": ["fastai"]} | null | angela1996/Practica1 | [
"fastai",
"has_space",
"region:us"
] | 2024-02-07T17:40:55+00:00 | [] | [] | TAGS
#fastai #has_space #region-us
|
# Amazing!
Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the documentation here)!
2. Create a demo in Gradio or Streamlit using Spaces (documentation here).
3. Join the fastai community on the Fastai Discord!
Greetings fellow fastlearner ! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| [
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
"TAGS\n#fastai #has_space #region-us \n",
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
13,
20,
79,
3,
6,
12,
8
] | [
"passage: TAGS\n#fastai #has_space #region-us \n# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---# Model card## Model description\nMore information needed## Intended uses & limitations\nMore information needed## Training and evaluation data\nMore information needed"
] | [
-0.048121724277734756,
-0.024616125971078873,
0.002038001548498869,
0.10439170897006989,
0.135872021317482,
0.11887997388839722,
0.07405775785446167,
0.09980081021785736,
0.07783667743206024,
0.02590852417051792,
0.08961158245801926,
-0.08088712394237518,
0.08744348585605621,
0.271692156791687,
0.06988707184791565,
-0.22761479020118713,
0.04051019623875618,
-0.00024903909070417285,
0.08053462207317352,
0.06629016250371933,
0.13507555425167084,
-0.05464952811598778,
0.14010503888130188,
-0.004088983871042728,
-0.19050447642803192,
-0.042929794639348984,
-0.01773718371987343,
-0.02527874894440174,
0.12317648530006409,
-0.04744937643408775,
0.05381017178297043,
0.015037551522254944,
0.007565062493085861,
-0.07253646105527878,
0.0623294934630394,
0.040457066148519516,
0.01740180514752865,
0.059235580265522,
-0.07249044626951218,
0.08950132131576538,
0.08404164761304855,
-0.024370938539505005,
-0.1097978875041008,
0.07827875018119812,
-0.14424212276935577,
-0.21762843430042267,
-0.1253085881471634,
-0.09017651528120041,
0.028519365936517715,
0.004388005938380957,
-0.025051530450582504,
0.12801909446716309,
-0.13558274507522583,
-0.040698226541280746,
0.20124278962612152,
-0.17012301087379456,
-0.05505548417568207,
0.034343402832746506,
0.09226689487695694,
-0.05829555168747902,
-0.06347129493951797,
0.10614984482526779,
0.09640881419181824,
-0.019833475351333618,
0.05516824126243591,
0.002579754451289773,
0.021173657849431038,
0.01370104867964983,
-0.06150497496128082,
0.04717832803726196,
-0.010183089412748814,
0.048132527619600296,
-0.09465572983026505,
-0.1303568333387375,
-0.004072192590683699,
0.01214400865137577,
-0.048744890838861465,
-0.07019646465778351,
0.07833103090524673,
-0.011118141002953053,
-0.04357248544692993,
-0.13031910359859467,
-0.09131011366844177,
-0.12358787655830383,
0.008646543137729168,
0.09500427544116974,
0.003679296001791954,
0.07374339550733566,
-0.08258994668722153,
0.06774985045194626,
-0.17329485714435577,
-0.06484591960906982,
-0.08138520270586014,
-0.11546400189399719,
0.021133482456207275,
-0.0387684591114521,
0.02668963186442852,
0.15394504368305206,
0.12983950972557068,
0.023976242169737816,
0.04388163983821869,
-0.038937073200941086,
0.051190316677093506,
0.058571770787239075,
0.03395717963576317,
0.034934818744659424,
-0.036981891840696335,
-0.1793210655450821,
-0.016702448949217796,
-0.011550825089216232,
0.07954040914773941,
-0.07523109763860703,
-0.05632320046424866,
0.013454885222017765,
-0.11071494966745377,
0.07202339172363281,
-0.03576776012778282,
-0.0032025426626205444,
0.01168301422148943,
0.018371861428022385,
0.21271461248397827,
0.03955606371164322,
0.014191740192472935,
-0.008875265717506409,
-0.13453757762908936,
-0.06874168664216995,
-0.06896194815635681,
0.03361047804355621,
0.04448792710900307,
-0.0028071461711078882,
-0.07672245055437088,
0.04325154796242714,
-0.06045534089207649,
-0.03508453071117401,
0.008032378740608692,
-0.18221288919448853,
0.007458044681698084,
-0.10049355030059814,
-0.12126200646162033,
0.05306628718972206,
0.01695440337061882,
-0.08215925842523575,
0.08141279965639114,
0.02662261202931404,
0.020931517705321312,
-0.009988143108785152,
-0.005391082260757685,
0.06874798238277435,
-0.08508864045143127,
0.029901226982474327,
0.17170792818069458,
0.13024519383907318,
-0.08046911656856537,
-0.0006887061172164977,
-0.10965746641159058,
0.04426072910428047,
-0.13325683772563934,
0.02251482754945755,
-0.09062390774488449,
0.11723794043064117,
-0.042396437376737595,
0.002038756385445595,
-0.029030200093984604,
0.0960269495844841,
0.08189879357814789,
0.16663365066051483,
-0.2419009804725647,
-0.031095001846551895,
0.13240347802639008,
-0.10711425542831421,
-0.1807439625263214,
0.18486657738685608,
-0.012035200372338295,
0.11329247802495956,
-0.047014184296131134,
0.18334640562534332,
-0.02612062357366085,
-0.13582459092140198,
-0.058872904628515244,
0.005852419883012772,
-0.2269321084022522,
-0.06286033242940903,
0.09738040715456009,
0.13425657153129578,
-0.042984943836927414,
0.007112155202776194,
0.026316028088331223,
0.13609857857227325,
-0.06715573370456696,
-0.05195777863264084,
-0.012255736626684666,
-0.10902371257543564,
0.041914235800504684,
0.018215661868453026,
0.035408079624176025,
-0.059880174696445465,
-0.02931194379925728,
-0.053190283477306366,
0.13146710395812988,
0.09760832786560059,
-0.03670211136341095,
-0.049620725214481354,
0.1689043790102005,
-0.07763876020908356,
-0.033587727695703506,
0.07560533285140991,
-0.08268500119447708,
0.03266897425055504,
0.03090597130358219,
0.055881720036268234,
0.07766123116016388,
0.08522116392850876,
0.06057543307542801,
0.00819048099219799,
0.034654274582862854,
0.12095347046852112,
-0.013591280207037926,
-0.05039411783218384,
0.021508218720555305,
0.016904234886169434,
-0.019032588228583336,
0.29030677676200867,
-0.1951042115688324,
0.024724548682570457,
-0.06477324664592743,
0.07631538063287735,
0.06136792525649071,
0.003575638635084033,
0.08580143749713898,
-0.06023019179701805,
-0.019061198458075523,
-0.04803973436355591,
0.046805646270513535,
-0.0666879191994667,
-0.04162997007369995,
0.2621194124221802,
-0.05497581139206886,
0.044914912432432175,
0.12313763797283173,
-0.05873025581240654,
-0.07091446220874786,
0.01009807363152504,
-0.00793424155563116,
0.03249288722872734,
-0.04042816907167435,
0.043721720576286316,
-0.10840129852294922,
-0.06674089282751083,
0.1573198139667511,
-0.038477856665849686,
0.06786153465509415,
0.032288823276758194,
-0.04958454892039299,
-0.0648743286728859,
0.04650486260652542,
0.13598160445690155,
-0.0875244215130806,
0.07435166835784912,
0.17612984776496887,
-0.010562662966549397,
0.168031245470047,
0.08435525000095367,
-0.07075224816799164,
-0.09465329349040985,
-0.051014289259910583,
-0.021595727652311325,
0.21222901344299316,
-0.07084725052118301,
-0.054564714431762695,
0.05911700800061226,
-0.013703816570341587,
0.07196151465177536,
-0.06009222939610481,
-0.08332337439060211,
0.03227344527840614,
-0.04517695680260658,
0.011517706327140331,
0.13512636721134186,
-0.07090822607278824,
0.04681389778852463,
0.031489867717027664,
-0.0662703812122345,
0.02217509225010872,
0.033389873802661896,
0.0068921963684260845,
0.033959709107875824,
0.07332495599985123,
-0.20893315970897675,
-0.08408680558204651,
-0.13727638125419617,
0.037881869822740555,
0.021770721301436424,
0.045787326991558075,
-0.08602345734834671,
0.02231026627123356,
-0.08954031765460968,
-0.07987114042043686,
0.029592275619506836,
-0.026350297033786774,
-0.11349643021821976,
-0.03396226093173027,
-0.009560913778841496,
-0.06662604957818985,
-0.02250705659389496,
-0.05024505779147148,
0.03983384370803833,
0.04479299485683441,
0.058377087116241455,
0.12796473503112793,
-0.013808943331241608,
-0.03839317709207535,
0.000370211957488209,
-0.022712308913469315,
0.16396735608577728,
-0.14746315777301788,
0.07954913377761841,
0.19160102307796478,
0.11742953956127167,
0.028144672513008118,
0.028885571286082268,
0.03537585213780403,
-0.06289814412593842,
-0.000050317394197918475,
0.03226194158196449,
-0.09392514824867249,
-0.05801016092300415,
-0.020014392212033272,
-0.04031052812933922,
0.17134574055671692,
-0.12160717695951462,
0.03345204517245293,
0.04098419472575188,
0.09783966839313507,
0.10073629021644592,
-0.028829937800765038,
-0.1815856397151947,
0.038818612694740295,
-0.24060091376304626,
-0.05831146240234375,
0.027899866923689842,
-0.09110201895236969,
-0.06232144311070442,
0.17409387230873108,
0.013794700615108013,
0.011769929900765419,
-0.006736889015883207,
0.07983319461345673,
0.0110100656747818,
0.1217205822467804,
0.05947643890976906,
-0.05539114400744438,
0.025202350690960884,
-0.09962950646877289,
-0.07107596844434738,
-0.04035590961575508,
-0.05832801014184952,
0.07548832893371582,
0.1409129947423935,
-0.025475580245256424,
-0.020795362070202827,
0.023489827290177345,
0.08550169318914413,
0.0423230417072773,
0.16739299893379211,
-0.16016584634780884,
-0.026555389165878296,
0.04571257904171944,
-0.03384667634963989,
-0.05433850735425949,
-0.010291114449501038,
0.1137225553393364,
-0.02820689231157303,
-0.040318265557289124,
0.021242983639240265,
0.06503437459468842,
0.01481706090271473,
0.05012747645378113,
-0.04056356102228165,
0.14796851575374603,
-0.03461192920804024,
0.019330544397234917,
-0.12413888424634933,
0.13848772644996643,
0.021095896139740944,
-0.03901609033346176,
-0.06735876202583313,
-0.05808034539222717,
0.18150931596755981,
0.0025602965615689754,
0.10535930097103119,
0.012098877690732479,
-0.12160047143697739,
-0.1359938681125641,
-0.11211287975311279,
0.005111907608807087,
0.08330471813678741,
-0.023147236555814743,
-0.022247863933444023,
0.022165266796946526,
-0.036149751394987106,
-0.0530381016433239,
0.15749511122703552,
-0.1289154291152954,
-0.001082550617866218,
0.014728817157447338,
0.06971760839223862,
-0.08223173767328262,
0.026267826557159424,
0.014071501791477203,
-0.1119147390127182,
0.10590848326683044,
0.2521335482597351,
0.10338116437196732,
-0.09591643512248993,
-0.07697287201881409,
0.03418830782175064,
-0.012184361927211285,
-0.000774814048781991,
-0.006932659074664116,
0.0495428591966629,
-0.005566445179283619,
0.006762749515473843,
0.12971895933151245,
-0.07130889594554901,
0.011540771462023258,
-0.08449850976467133,
0.05566910281777382,
-0.05276734381914139,
0.01761564053595066,
-0.002672141883522272,
-0.008124710991978645,
-0.07340748608112335,
-0.061829522252082825,
0.1609770804643631,
-0.07277000695466995,
-0.06468547880649567,
0.05801168829202652,
0.03307786211371422,
0.01431563775986433,
-0.03584568202495575,
-0.04342148080468178,
0.18088261783123016,
0.29330700635910034,
-0.08191116154193878,
0.10001859813928604,
0.09677296131849289,
0.034820813685655594,
-0.23625829815864563,
0.029798466712236404,
-0.1455078274011612,
0.04449721798300743,
0.040447335690259933,
-0.0409548319876194,
0.04191497340798378,
0.10835777968168259,
-0.06094440817832947,
0.2048867791891098,
-0.03527235612273216,
-0.07983248680830002,
-0.01788630709052086,
0.03109324350953102,
0.29443636536598206,
-0.11833466589450836,
0.006058716680854559,
-0.10420958697795868,
-0.21566011011600494,
0.06983078271150589,
-0.18948867917060852,
0.13948246836662292,
-0.05087858438491821,
0.03576415032148361,
-0.01149723306298256,
-0.07561972737312317,
0.20518061518669128,
-0.15641045570373535,
0.05273103713989258,
-0.13722458481788635,
-0.1327189952135086,
0.01617460884153843,
-0.10048147290945053,
0.1545477658510208,
-0.11024226248264313,
-0.023215843364596367,
-0.2284185290336609,
0.012587235309183598,
-0.023200806230306625,
0.10030807554721832,
0.01800704374909401,
-0.07980740070343018,
-0.08767345547676086,
0.1316242516040802,
-0.06486566364765167,
0.034810543060302734,
-0.06996636837720871,
-0.050714004784822464,
-0.010929876938462257,
-0.045061707496643066,
0.03034941293299198,
-0.07934719324111938,
0.15192505717277527,
-0.016938980668783188,
-0.04507075995206833,
0.08636019378900528,
-0.2479533851146698,
0.023727843537926674,
0.025351112708449364,
-0.03495599329471588,
0.09001832455396652,
-0.025513244792819023,
-0.06256973743438721,
0.12282291799783707,
0.1402233988046646,
-0.07322840392589569,
-0.2460673749446869,
-0.06281693279743195,
0.0076784128323197365,
0.039165716618299484,
0.06561196595430374,
0.05125982314348221,
-0.07261458039283752,
-0.011131617240607738,
-0.026896944269537926,
0.030595947057008743,
-0.11692017316818237,
-0.03854857385158539,
0.07790639251470566,
0.017095070332288742,
-0.07846562564373016,
0.07280377298593521,
0.014225782826542854,
-0.021511616185307503,
0.007357571739703417,
0.148970365524292,
0.007519228849560022,
-0.14747941493988037,
-0.06656096875667572,
0.2007484883069992,
-0.01197928935289383,
-0.07260087132453918,
-0.05383119732141495,
-0.008990069851279259,
-0.0476234145462513,
0.05585788935422897,
0.05367223918437958,
-0.013585401698946953,
0.07708586007356644,
0.06263149529695511,
-0.10210110992193222,
-0.046256959438323975,
-0.066561758518219,
0.04169114679098129,
-0.10485753417015076,
0.060470130294561386,
0.009529483504593372,
0.12185006588697433,
-0.09983488917350769,
-0.01802929677069187,
-0.10810204595327377,
-0.06766588985919952,
-0.17349553108215332,
-0.05834362283349037,
-0.041105758398771286,
-0.015651104971766472,
0.03658895567059517,
0.010445823892951012,
-0.057867538183927536,
-0.0442853718996048,
-0.07536603510379791,
0.038444988429546356,
0.06147460639476776,
0.03932281583547592,
-0.03912714496254921,
0.04001858830451965,
0.05909334123134613,
0.013087345287203789,
0.17542624473571777,
0.038768354803323746,
0.05504675209522247,
-0.05045998468995094,
-0.16491834819316864,
-0.05276111513376236,
-0.0074316514655947685,
-0.07559102028608322,
0.1224973127245903,
-0.007679440546780825,
0.007880088873207569,
-0.08065467327833176,
0.03924860805273056,
0.028234204277396202,
0.10404064506292343,
-0.0028364830650389194,
0.10070426017045975,
0.019627176225185394,
-0.07226712256669998,
-0.025392837822437286,
0.021809715777635574,
0.12809939682483673,
0.01567147858440876,
0.026090998202562332,
0.033139873296022415,
0.016619985923171043,
-0.057361043989658356,
0.033977724611759186,
-0.04997231811285019,
-0.15123651921749115,
0.02628709189593792,
-0.05165188014507294,
0.005062380339950323,
-0.016889680176973343,
0.20362506806850433,
0.07867538928985596,
-0.06474173814058304,
-0.010664013214409351,
0.015816617757081985,
-0.0168940220028162,
-0.03121885471045971,
-0.012740966863930225,
0.04592578858137131,
-0.001151384087279439,
-0.04866636544466019,
0.11825273931026459,
0.05015748366713524,
0.05386412516236305,
0.0596686452627182,
0.12528513371944427,
0.016759619116783142,
0.13257254660129547,
0.061999931931495667,
-0.03403807803988457,
-0.13461735844612122,
-0.04495539888739586,
-0.1254577934741974,
0.04646851494908333,
-0.08697032928466797,
0.09941662102937698,
0.1144254133105278,
-0.05959030240774155,
-0.030464433133602142,
-0.08851305395364761,
-0.008356761187314987,
-0.06041252240538597,
0.039516255259513855,
-0.02262675203382969,
-0.0873224213719368,
0.0481097511947155,
0.05495472997426987,
-0.022752324119210243,
0.13218675553798676,
0.015727028250694275,
-0.036317698657512665,
0.13270340859889984,
-0.07583184540271759,
0.11758984625339508,
0.061510033905506134,
-0.043043944984674454,
-0.11560922116041183,
-0.020150646567344666,
-0.06641761213541031,
-0.10098972916603088,
-0.006782987620681524,
-0.005399650428444147,
-0.07349002361297607,
-0.059971679002046585,
0.08397487550973892,
-0.03124053031206131,
-0.09979676455259323,
-0.032152675092220306,
0.0038895104080438614,
0.06054706871509552,
-0.01686914451420307,
-0.0034020058810710907,
0.04728743061423302,
0.015076374635100365,
0.1653461456298828,
-0.02208263985812664,
0.06234867498278618,
-0.13855914771556854,
0.16070103645324707,
-0.14684462547302246,
-0.029404424130916595,
-0.1890171319246292,
-0.09729582816362381,
-0.05156542733311653,
0.20326784253120422,
0.2840938866138458,
-0.19109351933002472,
-0.010187864303588867,
0.020078664645552635,
-0.014484191313385963,
-0.08961770683526993,
0.12571553885936737,
0.029420215636491776,
-0.023631498217582703,
-0.07249019294977188,
-0.02037387527525425,
0.005258576478809118,
-0.06544211506843567,
-0.026979785412549973,
0.18310695886611938,
0.001496660872362554,
0.059546373784542084,
-0.09605178982019424,
0.01754261925816536,
-0.14839904010295868,
-0.10467469692230225,
-0.02111995778977871,
-0.16156397759914398,
-0.09646477550268173,
0.006635562051087618,
0.038640011101961136,
0.08000610023736954,
0.03268849849700928,
-0.015172510407865047,
0.06479045748710632,
-0.056333884596824646,
-0.0037216036580502987,
-0.1231912299990654,
0.00034658415825106204,
0.062129102647304535,
-0.07422006875276566,
0.2545335292816162,
-0.03070417232811451,
-0.12370815873146057,
0.09026903659105301,
-0.03299184888601303,
-0.12452623248100281,
0.07951879501342773,
-0.005700904875993729,
-0.11531132459640503,
-0.057989440858364105,
0.18941475450992584,
-0.012821312062442303,
-0.1364315301179886,
0.046368811279535294,
-0.17166484892368317,
0.031349923461675644,
0.0363016203045845,
-0.001313706859946251,
-0.04714022949337959,
0.024538639932870865,
-0.008008457720279694,
0.10724439471960068,
0.1382838785648346,
0.016739921644330025,
-0.011060068383812904,
-0.05056179314851761,
0.07912429422140121,
0.056927867233753204,
-0.05218246951699257,
-0.1282637119293213,
-0.08599764108657837,
0.03429819270968437,
0.04119478166103363,
-0.08113081753253937,
-0.16903182864189148,
-0.03668912500143051,
-0.10082915425300598,
-0.004939202684909105,
0.051785312592983246,
0.06585265696048737,
0.29044589400291443,
0.06326735019683838,
0.0016605621203780174,
-0.13649453222751617,
0.050569336861371994,
0.0868251696228981,
-0.04697931930422783,
-0.07670357078313828
] |
null | null | stable-baselines3 |
# **DQN** Agent playing **SpaceInvadersNoFrameskip-v4**
This is a trained model of a **DQN** agent playing **SpaceInvadersNoFrameskip-v4**
using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3)
and the [RL Zoo](https://github.com/DLR-RM/rl-baselines3-zoo).
The RL Zoo is a training framework for Stable Baselines3
reinforcement learning agents,
with hyperparameter optimization and pre-trained agents included.
## Usage (with SB3 RL Zoo)
RL Zoo: https://github.com/DLR-RM/rl-baselines3-zoo<br/>
SB3: https://github.com/DLR-RM/stable-baselines3<br/>
SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib
Install the RL Zoo (with SB3 and SB3-Contrib):
```bash
pip install rl_zoo3
```
```
# Download model and save it into the logs/ folder
python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga Paquique -f logs/
python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/
```
If you installed the RL Zoo3 via pip (`pip install rl_zoo3`), from anywhere you can do:
```
python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga Paquique -f logs/
python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/
```
## Training (with the RL Zoo)
```
python -m rl_zoo3.train --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/
# Upload the model and generate video (when possible)
python -m rl_zoo3.push_to_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ -orga Paquique
```
## Hyperparameters
```python
OrderedDict([('batch_size', 64),
('buffer_size', 100000),
('env_wrapper',
['stable_baselines3.common.atari_wrappers.AtariWrapper']),
('exploration_final_eps', 0.01),
('exploration_fraction', 0.1),
('frame_stack', 4),
('gradient_steps', 1),
('learning_rate', 0.0001),
('learning_starts', 100000),
('n_timesteps', 1000000.0),
('optimize_memory_usage', False),
('policy', 'CnnPolicy'),
('target_update_interval', 1000),
('train_freq', 4),
('normalize', False)])
```
# Environment Arguments
```python
{'render_mode': 'rgb_array'}
```
| {"library_name": "stable-baselines3", "tags": ["SpaceInvadersNoFrameskip-v4", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "DQN", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "SpaceInvadersNoFrameskip-v4", "type": "SpaceInvadersNoFrameskip-v4"}, "metrics": [{"type": "mean_reward", "value": "548.00 +/- 276.52", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | Paquique/dqn-SpaceInvadersNoFrameskip-v4 | [
"stable-baselines3",
"SpaceInvadersNoFrameskip-v4",
"deep-reinforcement-learning",
"reinforcement-learning",
"model-index",
"region:us"
] | 2024-02-07T17:40:57+00:00 | [] | [] | TAGS
#stable-baselines3 #SpaceInvadersNoFrameskip-v4 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
|
# DQN Agent playing SpaceInvadersNoFrameskip-v4
This is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4
using the stable-baselines3 library
and the RL Zoo.
The RL Zoo is a training framework for Stable Baselines3
reinforcement learning agents,
with hyperparameter optimization and pre-trained agents included.
## Usage (with SB3 RL Zoo)
RL Zoo: URL
SB3: URL
SB3 Contrib: URL
Install the RL Zoo (with SB3 and SB3-Contrib):
If you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do:
## Training (with the RL Zoo)
## Hyperparameters
# Environment Arguments
| [
"# DQN Agent playing SpaceInvadersNoFrameskip-v4\nThis is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4\nusing the stable-baselines3 library\nand the RL Zoo.\n\nThe RL Zoo is a training framework for Stable Baselines3\nreinforcement learning agents,\nwith hyperparameter optimization and pre-trained agents included.",
"## Usage (with SB3 RL Zoo)\n\nRL Zoo: URL\nSB3: URL\nSB3 Contrib: URL\n\nInstall the RL Zoo (with SB3 and SB3-Contrib):\n\n\n\n\nIf you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do:",
"## Training (with the RL Zoo)",
"## Hyperparameters",
"# Environment Arguments"
] | [
"TAGS\n#stable-baselines3 #SpaceInvadersNoFrameskip-v4 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n",
"# DQN Agent playing SpaceInvadersNoFrameskip-v4\nThis is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4\nusing the stable-baselines3 library\nand the RL Zoo.\n\nThe RL Zoo is a training framework for Stable Baselines3\nreinforcement learning agents,\nwith hyperparameter optimization and pre-trained agents included.",
"## Usage (with SB3 RL Zoo)\n\nRL Zoo: URL\nSB3: URL\nSB3 Contrib: URL\n\nInstall the RL Zoo (with SB3 and SB3-Contrib):\n\n\n\n\nIf you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do:",
"## Training (with the RL Zoo)",
"## Hyperparameters",
"# Environment Arguments"
] | [
43,
90,
73,
9,
5,
7
] | [
"passage: TAGS\n#stable-baselines3 #SpaceInvadersNoFrameskip-v4 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# DQN Agent playing SpaceInvadersNoFrameskip-v4\nThis is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4\nusing the stable-baselines3 library\nand the RL Zoo.\n\nThe RL Zoo is a training framework for Stable Baselines3\nreinforcement learning agents,\nwith hyperparameter optimization and pre-trained agents included.## Usage (with SB3 RL Zoo)\n\nRL Zoo: URL\nSB3: URL\nSB3 Contrib: URL\n\nInstall the RL Zoo (with SB3 and SB3-Contrib):\n\n\n\n\nIf you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do:## Training (with the RL Zoo)## Hyperparameters# Environment Arguments"
] | [
0.043572068214416504,
0.2414778620004654,
-0.0026879787910729647,
0.012635791674256325,
0.05784223601222038,
0.0030472534708678722,
0.08585051447153091,
0.10650663822889328,
0.024212315678596497,
-0.001382096204906702,
0.003954293206334114,
0.17533031105995178,
0.03632635250687599,
0.13125447928905487,
-0.018073517829179764,
-0.2066594809293747,
-0.013479253277182579,
-0.06247470900416374,
-0.07153085619211197,
0.036099132150411606,
0.07206681370735168,
-0.030116932466626167,
0.036061208695173264,
-0.051406677812337875,
-0.057161085307598114,
0.036824777722358704,
-0.03157254680991173,
0.007067287806421518,
0.15158706903457642,
-0.1222257912158966,
0.12329676002264023,
0.020955175161361694,
0.1896144151687622,
-0.12332789599895477,
0.0339222252368927,
0.08982209116220474,
-0.036988191306591034,
0.013221588917076588,
0.00975361280143261,
-0.052562564611434937,
0.1590864509344101,
-0.09371145814657211,
0.07146181166172028,
0.010926910676062107,
-0.07592244446277618,
-0.1774153709411621,
-0.09356249868869781,
0.07947742193937302,
0.0617753230035305,
0.005319166928529739,
0.03726791962981224,
0.11306490749120712,
-0.020991774275898933,
0.06488905102014542,
0.11562903225421906,
-0.17549200356006622,
0.013578375801444054,
0.17859570682048798,
0.003242473118007183,
0.15767055749893188,
-0.05546637624502182,
0.019877681508660316,
0.02752300351858139,
0.04758313298225403,
0.06873945891857147,
-0.08186400681734085,
-0.1364826112985611,
-0.056155186146497726,
-0.15456219017505646,
-0.03352400287985802,
0.05195203423500061,
-0.011860138736665249,
-0.05783402919769287,
-0.010724928230047226,
-0.04010869935154915,
0.0008851495804265141,
-0.028637725859880447,
0.01805497519671917,
0.07031578570604324,
-0.01226285845041275,
0.02092539705336094,
-0.08391954004764557,
-0.0390290804207325,
-0.038563769310712814,
-0.018022390082478523,
0.12054917961359024,
0.08285853266716003,
0.0266572255641222,
-0.04135355353355408,
0.10274127870798111,
-0.07091585546731949,
-0.05454207584261894,
0.04555258899927139,
-0.03786851093173027,
-0.10615779459476471,
0.02120024710893631,
-0.05905991420149803,
0.026879185810685158,
0.09943640232086182,
0.18048083782196045,
-0.09862488508224487,
0.012620617635548115,
-0.03430783003568649,
0.08121664822101593,
-0.03196052461862564,
0.03197542577981949,
-0.0840383991599083,
-0.016251085326075554,
0.17835216224193573,
0.0030782297253608704,
0.022272996604442596,
0.002074616262689233,
-0.049819961190223694,
-0.02881433069705963,
-0.017756454646587372,
0.06631895154714584,
0.07032092660665512,
0.010587303899228573,
-0.0037596761249005795,
-0.027667716145515442,
-0.036921944469213486,
-0.05629328638315201,
-0.04952820762991905,
0.018803736194968224,
-0.04712437093257904,
-0.047942135483026505,
0.06027210131287575,
-0.005624116864055395,
0.11337806284427643,
-0.025607796385884285,
0.026316547766327858,
-0.019410157576203346,
-0.07494441419839859,
-0.13221681118011475,
-0.0304415225982666,
0.0691632330417633,
0.04371757060289383,
-0.22497159242630005,
-0.16994807124137878,
-0.008539012633264065,
0.017946386709809303,
-0.018741264939308167,
-0.11334165185689926,
0.02453240379691124,
-0.007166135590523481,
-0.049758363515138626,
-0.01601579785346985,
0.10474669933319092,
-0.020438622683286667,
0.018010856583714485,
-0.05593825876712799,
0.16603368520736694,
-0.14290283620357513,
0.031004127115011215,
-0.08706212788820267,
0.023509707301855087,
-0.21286657452583313,
0.041208744049072266,
-0.177636057138443,
0.04863585904240608,
-0.08500861376523972,
0.02327173389494419,
0.021320728585124016,
0.01968831568956375,
0.08580207824707031,
0.10143322497606277,
-0.23631145060062408,
0.05405791476368904,
0.07900930196046829,
-0.022739801555871964,
-0.04218491166830063,
0.06798892468214035,
-0.06558530032634735,
0.1382148116827011,
0.046505436301231384,
0.24831900000572205,
0.10361487418413162,
-0.2036508023738861,
0.061786454170942307,
0.0578593946993351,
-0.08880111575126648,
-0.004730981774628162,
-0.020022382959723473,
0.11598580330610275,
-0.01114928349852562,
0.03338807821273804,
-0.12186288088560104,
0.1456439197063446,
0.02738998830318451,
-0.0165485180914402,
-0.04454165697097778,
-0.1614885926246643,
0.10309953987598419,
-0.015504824928939342,
0.09532155096530914,
-0.042415786534547806,
0.0001161050095106475,
-0.011168917641043663,
0.18012429773807526,
-0.043841805309057236,
0.0007168867159634829,
0.07871408760547638,
0.10895700752735138,
0.028009075671434402,
-0.020230965688824654,
-0.20380273461341858,
-0.0423048660159111,
0.02367858961224556,
0.044489551335573196,
0.2190362960100174,
0.19936694204807281,
0.07770156860351562,
-0.022313760593533516,
-0.025487221777439117,
-0.003248062450438738,
-0.05106664076447487,
0.03467361256480217,
-0.027858436107635498,
-0.024532482028007507,
0.06065356358885765,
-0.09305168688297272,
0.02817818708717823,
-0.13112716376781464,
0.06307920068502426,
-0.17345242202281952,
0.06863926351070404,
0.021998396143317223,
-0.005436043255031109,
0.024577690288424492,
-0.011292695067822933,
-0.034188106656074524,
-0.06233125180006027,
0.07110602408647537,
0.06098933145403862,
0.014702376909554005,
0.0021991983521729708,
-0.0683600977063179,
-0.13828523457050323,
0.08231553435325623,
-0.04042381793260574,
-0.14305958151817322,
0.06392676383256912,
0.011172642931342125,
0.04875864461064339,
-0.05975872278213501,
0.016254881396889687,
0.22900153696537018,
0.05321883037686348,
0.09785865992307663,
-0.04092191904783249,
-0.022525805979967117,
-0.06617844104766846,
-0.06677833944559097,
0.09694591909646988,
0.10812206566333771,
0.060318704694509506,
-0.0030071530491113663,
0.07626225054264069,
0.10942911356687546,
-0.1035122498869896,
-0.0651884600520134,
0.03220061957836151,
-0.05973697826266289,
0.019652515649795532,
0.049140311777591705,
0.02971293032169342,
0.08619047701358795,
0.1833551675081253,
0.008245792239904404,
0.0386311337351799,
-0.025997694581747055,
0.026109617203474045,
-0.15547916293144226,
-0.03145433962345123,
0.04308181628584862,
0.00886955764144659,
-0.07408110797405243,
0.04994636029005051,
0.051439400762319565,
0.13607151806354523,
-0.08217083662748337,
-0.13170577585697174,
-0.059745315462350845,
-0.03804200142621994,
-0.04239124804735184,
0.14975430071353912,
-0.08507520705461502,
-0.19221234321594238,
-0.017164425924420357,
-0.15751953423023224,
-0.02518727444112301,
-0.005179801490157843,
0.002318724524229765,
-0.08325926214456558,
0.017780914902687073,
0.010001576505601406,
-0.03129372000694275,
-0.0684933215379715,
-0.06596160680055618,
-0.05786636844277382,
0.09124112874269485,
0.06932931393384933,
-0.12240120023488998,
-0.00961651187390089,
-0.03742414712905884,
-0.020465577021241188,
0.04516167193651199,
0.08452648669481277,
-0.007267598994076252,
0.07773483544588089,
-0.13209199905395508,
-0.06962883472442627,
0.02834828943014145,
0.2766247093677521,
0.02882981114089489,
0.004668009467422962,
0.17051753401756287,
-0.03629542142152786,
0.04912714660167694,
0.16181479394435883,
0.030781643465161324,
-0.14196757972240448,
0.07090470939874649,
-0.011341600678861141,
-0.09542687982320786,
-0.1706860214471817,
-0.10215658694505692,
-0.037867411971092224,
-0.05015881359577179,
0.05638284236192703,
0.004951419774442911,
-0.04476970434188843,
0.05910305306315422,
0.08782228082418442,
-0.017004497349262238,
-0.06151578947901726,
0.11129767447710037,
0.032263003289699554,
-0.030136963352560997,
0.08078382909297943,
-0.042354047298431396,
-0.04206389561295509,
0.0032403599470853806,
0.22643887996673584,
0.0937788337469101,
-0.01775507442653179,
-0.042567066848278046,
0.019317636266350746,
0.05095715448260307,
0.03613382205367088,
0.11312435567378998,
-0.06975842267274857,
-0.06826137751340866,
-0.035185977816581726,
0.027829548344016075,
-0.02945687249302864,
0.08205190300941467,
0.0630207508802414,
0.005563626065850258,
-0.04653681069612503,
-0.07972332090139389,
-0.04849022626876831,
0.08408913016319275,
-0.027642227709293365,
-0.10093270242214203,
0.09321888536214828,
0.048575710505247116,
0.0016974330646917224,
0.03055831417441368,
0.027994604781270027,
0.01462269201874733,
-0.07982148975133896,
-0.06775744259357452,
0.011468625627458096,
0.07076629996299744,
-0.06822766363620758,
-0.027886953204870224,
-0.19817815721035004,
0.14578363299369812,
0.010630400851368904,
0.04118429124355316,
-0.13048617541790009,
0.1209396943449974,
-0.023116756230592728,
-0.026430301368236542,
0.013811616227030754,
0.0014643745962530375,
0.08203291147947311,
-0.04806509613990784,
0.15762180089950562,
0.009528410620987415,
-0.28092408180236816,
-0.1418946087360382,
-0.08416824042797089,
-0.051183976233005524,
-0.022873088717460632,
0.014752174727618694,
0.0642135739326477,
0.01516205258667469,
0.003868846921250224,
-0.013076163828372955,
0.03185269236564636,
-0.09826882928609848,
-0.06493937969207764,
-0.04839126765727997,
-0.02250157669186592,
-0.06525848805904388,
-0.05647949501872063,
-0.0006809153710491955,
-0.17226077616214752,
0.12522587180137634,
0.11787347495555878,
-0.06451737880706787,
-0.041814323514699936,
-0.06554657220840454,
0.046191465109586716,
-0.07571537792682648,
0.0469326451420784,
0.003414976177737117,
0.019198855385184288,
-0.06806991249322891,
-0.17922484874725342,
0.016097763553261757,
-0.10899919271469116,
0.03772687539458275,
-0.05070559307932854,
0.020257100462913513,
0.08594245463609695,
0.17520126700401306,
0.05856714025139809,
0.01460097823292017,
-0.07239776104688644,
-0.07543374598026276,
-0.0017121878918260336,
-0.06344114243984222,
0.05762333422899246,
-0.009151889942586422,
-0.20333483815193176,
0.02763226442039013,
-0.11414948850870132,
0.06860900670289993,
0.3310066759586334,
0.3324824273586273,
-0.10698744654655457,
0.1177443116903305,
0.04819539934396744,
-0.042202454060316086,
-0.21051374077796936,
-0.002244179602712393,
0.012272895313799381,
0.024992236867547035,
0.13725964725017548,
-0.12924811244010925,
0.05453680083155632,
0.0794181227684021,
-0.024458877742290497,
0.01456840243190527,
-0.09078162908554077,
-0.10816970467567444,
0.20847418904304504,
0.14226987957954407,
0.04421741142868996,
-0.09421348571777344,
0.08391669392585754,
0.004295284394174814,
0.08375877887010574,
0.2107764035463333,
-0.052112679928541183,
0.10695768147706985,
0.005195184610784054,
0.19852910935878754,
0.0328996516764164,
-0.023768596351146698,
0.10834760218858719,
-0.009801650419831276,
0.07911337912082672,
0.03985166177153587,
-0.007676942739635706,
0.010487722232937813,
-0.04522453248500824,
0.014148596674203873,
-0.028376007452607155,
0.010284217074513435,
-0.2274095118045807,
0.0582297146320343,
-0.06368855386972427,
0.04604509472846985,
0.008256820961833,
-0.0999874547123909,
-0.03583388403058052,
0.06431841105222702,
0.08014573156833649,
0.01975327916443348,
0.0436067171394825,
-0.03867863491177559,
0.11051398515701294,
0.20660489797592163,
-0.009811338968575,
0.17751595377922058,
-0.0615963339805603,
0.01464168168604374,
-0.023011628538370132,
-0.04223164543509483,
-0.1462583988904953,
-0.035259708762168884,
0.03498423472046852,
0.057734888046979904,
0.015203364193439484,
0.049647457897663116,
-0.05656236410140991,
0.08498423546552658,
0.021687336266040802,
-0.041541360318660736,
0.033579520881175995,
0.08835696429014206,
0.12415177375078201,
0.010754258371889591,
-0.030121933668851852,
0.06147436052560806,
-0.08128108084201813,
-0.09446098655462265,
-0.004497923422604799,
-0.029991207644343376,
-0.1083834245800972,
0.11353230476379395,
0.16914646327495575,
0.039594944566488266,
-0.057076629251241684,
0.10688766092061996,
-0.02768099494278431,
0.10047874599695206,
0.009198128245770931,
0.06507332623004913,
-0.014091075398027897,
-0.03691792115569115,
0.10611724853515625,
-0.05442855879664421,
-0.01637818105518818,
0.07645545154809952,
-0.06522727757692337,
-0.023877469822764397,
-0.0801999643445015,
0.06034626066684723,
0.09222240000963211,
-0.16854619979858398,
-0.0639432892203331,
-0.032122284173965454,
-0.08628080040216446,
0.013965039514005184,
0.012447911314666271,
0.0710059329867363,
-0.08589600026607513,
0.06316167116165161,
-0.024337708950042725,
0.015639442950487137,
-0.03689891844987869,
0.019222697243094444,
-0.19525384902954102,
-0.002140450058504939,
-0.11280795186758041,
-0.00348020251840353,
-0.002931603929027915,
0.04463808611035347,
-0.04961875081062317,
-0.029358822852373123,
-0.0030675032176077366,
0.044366419315338135,
-0.16609135270118713,
0.002798673929646611,
-0.011639905162155628,
0.03210212290287018,
-0.0002893915225286037,
-0.0983390137553215,
0.014195028692483902,
-0.04294256120920181,
-0.04198618605732918,
0.04925514757633209,
0.009436776861548424,
0.06470516324043274,
-0.2795179784297943,
-0.14905457198619843,
0.030816160142421722,
0.0683867484331131,
0.05483196675777435,
-0.1830425262451172,
0.03568267077207565,
-0.08042316138744354,
-0.02253127470612526,
-0.037770628929138184,
0.018491698428988457,
-0.0539514496922493,
0.0018174031283706427,
-0.04225044324994087,
-0.023033907637000084,
-0.028055014088749886,
-0.07556360960006714,
0.0826747715473175,
0.12462522834539413,
0.07555580884218216,
-0.03807181864976883,
0.09595896303653717,
-0.10009756684303284,
-0.04657831788063049,
-0.04052736237645149,
-0.036951083689928055,
0.017965637147426605,
-0.0870552659034729,
0.048530060797929764,
0.05188591405749321,
0.18719671666622162,
-0.08520494401454926,
-0.058800119906663895,
-0.014255574904382229,
0.0746525228023529,
0.07849094271659851,
0.005095830652862787,
0.17779210209846497,
-0.045693784952163696,
0.05693846940994263,
0.021304311230778694,
0.046699028462171555,
0.10497613251209259,
-0.023569339886307716,
0.14490213990211487,
0.21171095967292786,
-0.037196725606918335,
-0.11048602312803268,
0.043668005615472794,
0.01745123788714409,
-0.002401199424639344,
0.05968761444091797,
0.11983796209096909,
-0.050589341670274734,
-0.10903856158256531,
0.23442286252975464,
0.054169271141290665,
-0.11218088120222092,
0.09546315670013428,
0.039532262831926346,
-0.015890996903181076,
-0.1301896870136261,
0.010444961488246918,
-0.0013640925753861666,
-0.11233190447092056,
0.03386834263801575,
-0.06087532266974449,
-0.025547027587890625,
0.11809267848730087,
0.008789865300059319,
0.03317064419388771,
-0.04139537364244461,
-0.03756232187151909,
-0.04352104663848877,
-0.04273213446140289,
-0.012549578212201595,
-0.02991986647248268,
-0.030186517164111137,
-0.07621737569570541,
-0.007770835887640715,
-0.012012424878776073,
0.030795488506555557,
-0.015285328030586243,
-0.02503054589033127,
-0.021192016080021858,
-0.06697061657905579,
-0.0026312144473195076,
-0.008178025484085083,
0.015549594536423683,
0.010121971368789673,
0.2358063906431198,
0.07042546570301056,
-0.10260069370269775,
-0.01036880537867546,
0.22197756171226501,
-0.03853277862071991,
-0.06528383493423462,
-0.07849395275115967,
0.25128230452537537,
-0.10482002794742584,
0.051095426082611084,
-0.005819917656481266,
-0.06550488620996475,
-0.07153836637735367,
0.2309868484735489,
0.13502730429172516,
-0.1677926480770111,
0.06329060345888138,
-0.0368385910987854,
-0.009490780532360077,
-0.14286863803863525,
0.16013580560684204,
0.1865294873714447,
0.09480160474777222,
-0.12259847670793533,
0.0023130534682422876,
-0.03518044203519821,
-0.018328361213207245,
-0.1660851687192917,
-0.004593863617628813,
-0.029364850372076035,
-0.0427238829433918,
-0.050771355628967285,
0.029773715883493423,
-0.15205919742584229,
-0.0927426889538765,
-0.1916799396276474,
-0.11482496559619904,
-0.12386849522590637,
-0.04549141973257065,
-0.11142764985561371,
-0.0019938007462769747,
0.02257080189883709,
-0.0641874223947525,
0.021061956882476807,
-0.0212461706250906,
-0.05887424945831299,
0.015386379323899746,
-0.08395619690418243,
0.0674985870718956,
0.06488548219203949,
0.15327942371368408,
-0.0790991559624672,
0.025424562394618988,
0.07090727984905243,
-0.057595450431108475,
-0.10164349526166916,
0.06067253649234772,
0.015708057209849358,
-0.1972588747739792,
0.007548294495791197,
0.17712996900081635,
-0.10420889407396317,
0.09745754301548004,
0.048501528799533844,
-0.012951982207596302,
0.0867827981710434,
-0.024721821770071983,
-0.016682926565408707,
-0.04852180927991867,
-0.011212974786758423,
-0.10143939405679703,
0.09892100840806961,
0.0876845121383667,
-0.0517118014395237,
0.07436849176883698,
-0.09508965909481049,
-0.04068392515182495,
0.13103286921977997,
-0.010057874955236912,
-0.08450483530759811,
-0.11667824536561966,
-0.04081142693758011,
0.09684515744447708,
-0.018041390925645828,
-0.20185889303684235,
-0.11639472097158432,
-0.11752668023109436,
-0.00014377340266946703,
-0.03563340753316879,
0.061800602823495865,
0.02430674433708191,
-0.02556120604276657,
-0.008150683715939522,
-0.17615078389644623,
-0.06614746153354645,
0.13479791581630707,
-0.10176112502813339,
-0.07456064969301224
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# wav2vec_RTSplit0208_4
This model is a fine-tuned version of [jonatasgrosman/wav2vec2-large-xlsr-53-japanese](https://huggingface.co/jonatasgrosman/wav2vec2-large-xlsr-53-japanese) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0286
- Wer: 0.2110
- Cer: 0.2340
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5.5e-05
- train_batch_size: 32
- eval_batch_size: 32
- seed: 4
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 1000
- num_epochs: 8
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|
| 3.5617 | 1.0 | 120 | 3.4716 | 1.0 | 0.9474 |
| 1.5803 | 2.0 | 240 | 1.3647 | 1.0 | 0.7373 |
| 0.9177 | 3.0 | 360 | 0.7341 | 0.8214 | 0.6103 |
| 0.5589 | 4.0 | 480 | 0.4265 | 0.7531 | 0.5284 |
| 0.3676 | 5.0 | 600 | 0.2065 | 0.4244 | 0.3298 |
| 0.2415 | 6.0 | 720 | 0.0940 | 0.2638 | 0.2046 |
| 0.2053 | 7.0 | 840 | 0.0500 | 0.2361 | 0.1960 |
| 0.1712 | 8.0 | 960 | 0.0286 | 0.2110 | 0.2340 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.14.6
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["wer"], "base_model": "jonatasgrosman/wav2vec2-large-xlsr-53-japanese", "model-index": [{"name": "wav2vec_RTSplit0208_4", "results": []}]} | automatic-speech-recognition | tndklab/wav2vec_RTSplit0208_4 | [
"transformers",
"safetensors",
"wav2vec2",
"automatic-speech-recognition",
"generated_from_trainer",
"base_model:jonatasgrosman/wav2vec2-large-xlsr-53-japanese",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-07T17:41:34+00:00 | [] | [] | TAGS
#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-jonatasgrosman/wav2vec2-large-xlsr-53-japanese #license-apache-2.0 #endpoints_compatible #region-us
| wav2vec\_RTSplit0208\_4
=======================
This model is a fine-tuned version of jonatasgrosman/wav2vec2-large-xlsr-53-japanese on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0286
* Wer: 0.2110
* Cer: 0.2340
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5.5e-05
* train\_batch\_size: 32
* eval\_batch\_size: 32
* seed: 4
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 1000
* num\_epochs: 8
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.14.6
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5.5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 8",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-jonatasgrosman/wav2vec2-large-xlsr-53-japanese #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5.5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 8",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
80,
116,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-jonatasgrosman/wav2vec2-large-xlsr-53-japanese #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5.5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 8### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
-0.1416521668434143,
0.15231576561927795,
-0.0005603038007393479,
0.10063644498586655,
0.11849697679281235,
0.008984331041574478,
0.1743982881307602,
0.1505867838859558,
-0.04132495075464249,
0.11120389401912689,
0.11344815790653229,
0.060427792370319366,
0.054913755506277084,
0.1962769329547882,
-0.0819840207695961,
-0.22068257629871368,
0.07666398584842682,
-0.003938599955290556,
0.009955009445548058,
0.11176874488592148,
0.07094193994998932,
-0.11867162585258484,
0.0897790789604187,
-0.007773552555590868,
-0.143516406416893,
-0.042054396122694016,
0.01707952842116356,
-0.10967446863651276,
0.10822425037622452,
0.009785709902644157,
0.06561127305030823,
0.035010457038879395,
0.08909126371145248,
-0.18818454444408417,
0.0020208521746098995,
0.01803073100745678,
0.014820248819887638,
0.07439117878675461,
0.04391889646649361,
0.00017175775428768247,
0.0031204598490148783,
-0.11392830312252045,
0.036929208785295486,
0.01579548791050911,
-0.11667035520076752,
-0.19877471029758453,
-0.07784965634346008,
0.016412805765867233,
0.09913831949234009,
0.08371337503194809,
-0.020506637170910835,
0.12272398173809052,
0.00035763258347287774,
0.08002498000860214,
0.19646312296390533,
-0.31315991282463074,
-0.05485222116112709,
-0.01637648232281208,
0.03843139111995697,
0.08388074487447739,
-0.10232923924922943,
-0.017992103472352028,
0.04975762963294983,
0.021735871210694313,
0.09219805151224136,
-0.0312325619161129,
-0.034768421202898026,
-0.011004535481333733,
-0.12001021951436996,
-0.03902936354279518,
0.19005079567432404,
0.07318458706140518,
-0.06370215117931366,
-0.08091289550065994,
-0.0639798641204834,
-0.12154904007911682,
-0.05428833141922951,
-0.007392325438559055,
0.026221267879009247,
-0.03905731067061424,
-0.09971770644187927,
-0.004373535979539156,
-0.07992051541805267,
-0.09107047319412231,
-0.017052462324500084,
0.17492707073688507,
0.010855483822524548,
0.014318611472845078,
-0.011713307350873947,
0.054962847381830215,
-0.024142764508724213,
-0.18482787907123566,
-0.022869868203997612,
0.026672441512346268,
-0.033219609409570694,
-0.013979192823171616,
-0.04406684637069702,
-0.034891098737716675,
0.04358501732349396,
0.11606881767511368,
-0.01989586278796196,
0.06639353930950165,
-0.024975808337330818,
0.001574174384586513,
-0.08518648892641068,
0.1826273649930954,
-0.06345509737730026,
-0.06869041919708252,
0.01973448321223259,
0.12760382890701294,
0.06217905133962631,
-0.023536039516329765,
-0.09854742139577866,
-0.009291871450841427,
0.14756102859973907,
0.035263534635305405,
-0.04283446818590164,
0.04984349012374878,
-0.039159346371889114,
-0.014491844922304153,
0.05640894174575806,
-0.1209423765540123,
0.026020150631666183,
0.021960625424981117,
-0.06292986124753952,
-0.0227982010692358,
-0.011182006448507309,
0.012097802944481373,
0.012511351145803928,
0.05239659547805786,
-0.08222419023513794,
0.004087643697857857,
-0.023971159011125565,
-0.09210819005966187,
0.026488710194826126,
-0.06779901683330536,
-0.00003616911271819845,
-0.10790187865495682,
-0.17777115106582642,
-0.017982322722673416,
0.024202367290854454,
-0.04997178167104721,
-0.010579396039247513,
-0.11217106878757477,
-0.0969209372997284,
0.04816645383834839,
-0.023094015195965767,
0.03563275188207626,
-0.07954283058643341,
0.10825987160205841,
0.07975974678993225,
0.08701836317777634,
-0.03987785056233406,
0.026628922671079636,
-0.09402554482221603,
0.032031070441007614,
-0.17547975480556488,
0.07519315928220749,
-0.05403360724449158,
0.033848997205495834,
-0.11996884644031525,
-0.06739335507154465,
0.02059425413608551,
-0.022680077701807022,
0.07002267241477966,
0.1423128843307495,
-0.19017033278942108,
-0.056031059473752975,
0.19586409628391266,
-0.12027470767498016,
-0.14221099019050598,
0.12898172438144684,
-0.03614957258105278,
0.03756829351186752,
0.07013069838285446,
0.22297145426273346,
0.03284775838255882,
-0.10563252121210098,
-0.03851517289876938,
-0.06331296265125275,
0.0851268544793129,
-0.037616536021232605,
0.1110086739063263,
0.004914650693535805,
-0.0018710278673097491,
0.01674548350274563,
-0.08060979843139648,
0.032695986330509186,
-0.07096053659915924,
-0.09992992877960205,
-0.04475810006260872,
-0.10594291985034943,
0.027823492884635925,
0.016577500849962234,
0.05604519695043564,
-0.098753921687603,
-0.07039985805749893,
0.011080390773713589,
0.108296699821949,
-0.11699311435222626,
0.012669777497649193,
-0.10356700420379639,
0.09398839622735977,
-0.11380679160356522,
-0.020340487360954285,
-0.15481537580490112,
-0.004622957669198513,
0.05435006693005562,
0.020086025819182396,
0.013744096271693707,
-0.07509239763021469,
0.08213549107313156,
0.0774499922990799,
-0.04991897940635681,
-0.07361490279436111,
-0.004902792163193226,
0.017748335376381874,
-0.06340279430150986,
-0.17395353317260742,
-0.028873080387711525,
-0.05375329777598381,
0.15960189700126648,
-0.16576892137527466,
0.0007898173644207418,
0.00985600333660841,
0.09055764973163605,
0.04418988898396492,
-0.023415589705109596,
0.020038722082972527,
0.04883560910820961,
-0.02583049237728119,
-0.07130507379770279,
0.029749196022748947,
0.01543353870511055,
-0.10305999964475632,
0.019544966518878937,
-0.16675932705402374,
0.15245139598846436,
0.13804586231708527,
0.04138542711734772,
-0.0528542585670948,
0.02008969336748123,
-0.01355352159589529,
-0.04280754178762436,
-0.05480244755744934,
-0.014847248792648315,
0.10164619237184525,
0.007989929057657719,
0.1220778226852417,
-0.1028716117143631,
0.01645234227180481,
0.06500449031591415,
-0.0268496572971344,
-0.02799966372549534,
0.08064081519842148,
0.010382457636296749,
-0.13875392079353333,
0.12979555130004883,
0.11175474524497986,
-0.07268914580345154,
0.1266240030527115,
-0.061948955059051514,
-0.08509109169244766,
-0.049796272069215775,
0.03429628908634186,
0.03415772691369057,
0.13763904571533203,
-0.08183307200670242,
-0.02226833440363407,
0.021038377657532692,
0.02304365485906601,
-0.015965880826115608,
-0.19314135611057281,
-0.019839487969875336,
0.014704600907862186,
-0.09494518488645554,
-0.008283012546598911,
0.005615855101495981,
-0.017601581290364265,
0.09457366913557053,
-0.000883720291312784,
-0.11425274610519409,
0.023895123973488808,
-0.015262739732861519,
-0.08755475282669067,
0.17213934659957886,
-0.09207827597856522,
-0.17479614913463593,
-0.1365613341331482,
-0.0711062103509903,
-0.05714680254459381,
0.03693916276097298,
0.06066662445664406,
-0.06600242853164673,
-0.040584124624729156,
-0.11569657176733017,
-0.048396069556474686,
0.032208461314439774,
0.04574784263968468,
0.04982975125312805,
-0.008772005327045918,
0.067414790391922,
-0.0813615545630455,
-0.00472258822992444,
-0.014015201479196548,
-0.0076734693720936775,
0.029594827443361282,
0.0006293753394857049,
0.126233771443367,
0.12172220647335052,
0.006236088462173939,
0.024821242317557335,
-0.03748829662799835,
0.2264401763677597,
-0.06926576793193817,
-0.019435277208685875,
0.12340638786554337,
-0.027483809739351273,
0.045787911862134933,
0.17751628160476685,
0.03075435943901539,
-0.10705778002738953,
0.0017856821650639176,
-0.04998410493135452,
-0.015200897119939327,
-0.18865841627120972,
-0.03303465619683266,
-0.0486757829785347,
0.013971392996609211,
0.10143491625785828,
0.030121197924017906,
0.014428957365453243,
0.04798053950071335,
0.022219162434339523,
0.045422859489917755,
0.00430995412170887,
0.08098297566175461,
0.09676896035671234,
0.07662659138441086,
0.10846129059791565,
-0.03190721571445465,
-0.04817984253168106,
0.03271808102726936,
0.020907824859023094,
0.2028363198041916,
0.0291864275932312,
0.1921800971031189,
0.000349435635143891,
0.15605568885803223,
0.02624501846730709,
0.07961162179708481,
0.018472304567694664,
0.00992793869227171,
-0.020957110449671745,
-0.0779685452580452,
-0.05427340418100357,
0.054767679423093796,
-0.014759950339794159,
0.06113133952021599,
-0.10674641281366348,
0.021154502406716347,
0.04954751580953598,
0.2737855017185211,
0.08787969499826431,
-0.36818042397499084,
-0.08654342591762543,
0.02062247321009636,
-0.036866605281829834,
-0.019731566309928894,
0.016786161810159683,
0.15560324490070343,
-0.06151933968067169,
0.06805241107940674,
-0.07225997000932693,
0.06330756843090057,
-0.06420214474201202,
0.019307507202029228,
0.025530628859996796,
0.04757826402783394,
0.0029408466070890427,
0.031052490696310997,
-0.24230308830738068,
0.2861466407775879,
0.03598778322339058,
0.09501178562641144,
-0.056578379124403,
-0.00356899481266737,
0.039379410445690155,
-0.005160613916814327,
0.11663217842578888,
-0.02469443529844284,
-0.111576609313488,
-0.17986367642879486,
-0.13459579646587372,
0.049224402755498886,
0.10536382347345352,
-0.006649328861385584,
0.1156100183725357,
-0.014204435981810093,
-0.04427343234419823,
0.04500559717416763,
-0.02329474873840809,
-0.08001597970724106,
-0.07572807371616364,
0.009524318389594555,
0.11416337639093399,
0.044665709137916565,
-0.04997749999165535,
-0.09591855108737946,
-0.08735961467027664,
0.09040538966655731,
0.003137752879410982,
-0.006660701707005501,
-0.1054760068655014,
0.01851012371480465,
0.14962713420391083,
-0.09137416630983353,
0.05292701721191406,
0.009245136752724648,
0.11035801470279694,
0.028003035113215446,
-0.04933254420757294,
0.09060999006032944,
-0.06226043775677681,
-0.1787278950214386,
-0.051083534955978394,
0.13831165432929993,
-0.007863691076636314,
0.04305732250213623,
0.02076517790555954,
0.05136464163661003,
-0.005119224078953266,
-0.06748228520154953,
0.031936775892972946,
0.026990210637450218,
0.04103126749396324,
0.02063833922147751,
-0.012296152301132679,
-0.09058242291212082,
-0.09235145151615143,
-0.023640768602490425,
0.15033818781375885,
0.2975933253765106,
-0.06646092236042023,
0.018357181921601295,
0.08651206642389297,
-0.018205052241683006,
-0.15108747780323029,
-0.004305514972656965,
0.0445161871612072,
0.044636283069849014,
-0.004072494804859161,
-0.12263128906488419,
0.04505032300949097,
0.061492063105106354,
-0.04486415535211563,
0.07715693861246109,
-0.2489887773990631,
-0.12737591564655304,
0.08829639852046967,
0.13310593366622925,
0.12607449293136597,
-0.15335559844970703,
-0.06704343855381012,
-0.022875970229506493,
-0.10748536884784698,
0.10360315442085266,
-0.07281716167926788,
0.1335403323173523,
-0.00312104937620461,
0.06376536190509796,
0.0076929558999836445,
-0.05116251856088638,
0.15059663355350494,
0.0220597255975008,
0.053580161184072495,
-0.022253816947340965,
-0.015353093855082989,
0.045860372483730316,
-0.07609099894762039,
0.0688006803393364,
-0.08713462203741074,
0.05023961141705513,
-0.061373453587293625,
-0.024799318984150887,
-0.061800144612789154,
-0.0070847864262759686,
0.003488904098048806,
-0.0343170203268528,
-0.010318437591195107,
0.03606942668557167,
0.0585615448653698,
0.0032982409466058016,
0.1327333301305771,
0.012194394133985043,
0.08270338177680969,
0.14977239072322845,
0.08813391625881195,
-0.04014086723327637,
0.013446941040456295,
-0.00663809385150671,
-0.056290093809366226,
0.05374028906226158,
-0.132358118891716,
0.04883670434355736,
0.09682066738605499,
0.018109053373336792,
0.1607360988855362,
0.04666454344987869,
-0.04892454668879509,
0.03831847012042999,
0.06941424310207367,
-0.15927885472774506,
-0.11118414252996445,
0.0031255423091351986,
-0.012310276739299297,
-0.11201860010623932,
0.04805121570825577,
0.13912077248096466,
-0.07058347016572952,
-0.005982224829494953,
-0.017488976940512657,
0.021889301016926765,
-0.03901759162545204,
0.20138752460479736,
0.041609711945056915,
0.0518956296145916,
-0.10926301777362823,
0.08106806874275208,
0.05655520781874657,
-0.08764895051717758,
0.04899587482213974,
0.037728291004896164,
-0.11543168872594833,
-0.022817697376012802,
-0.00008412777242483571,
0.14217811822891235,
0.005306280218064785,
-0.07609368860721588,
-0.1386883705854416,
-0.08924244344234467,
0.03448374196887016,
0.17552168667316437,
0.0682656317949295,
0.036522019654512405,
-0.01856006495654583,
-0.0025720084086060524,
-0.10348781943321228,
0.09442557394504547,
0.07426261901855469,
0.07521814107894897,
-0.15098612010478973,
0.08222828060388565,
-0.008154675364494324,
0.02670569345355034,
-0.02037554606795311,
0.0165905449539423,
-0.10963503271341324,
0.005446241237223148,
-0.09949247539043427,
0.057657331228256226,
-0.0772402361035347,
-0.015851961448788643,
-0.001594056375324726,
-0.08186923712491989,
-0.061409346759319305,
0.012022759765386581,
-0.0870140865445137,
-0.026207465678453445,
0.0032590795308351517,
0.0435623936355114,
-0.1352715790271759,
-0.03800482302904129,
0.022525111213326454,
-0.09848035126924515,
0.08381954580545425,
0.08575370907783508,
-0.019917329773306847,
0.04638371989130974,
-0.09426191449165344,
-0.021776843816041946,
0.08226069808006287,
0.0021264252718538046,
0.05044756457209587,
-0.14400367438793182,
-0.013879073783755302,
0.031359270215034485,
0.050214968621730804,
0.02181389182806015,
0.14856772124767303,
-0.0973929837346077,
0.005486358422785997,
-0.06730689853429794,
-0.011177713982760906,
-0.05682748183608055,
0.02177783101797104,
0.14174701273441315,
0.0035442886874079704,
0.18419149518013,
-0.0955568253993988,
0.022457418963313103,
-0.19827735424041748,
0.0013734949752688408,
-0.03692111000418663,
-0.12555257976055145,
-0.1475173830986023,
-0.026949433609843254,
0.0786232054233551,
-0.06240832805633545,
0.09476541727781296,
-0.061278849840164185,
0.06961840391159058,
0.012964576482772827,
-0.05965392664074898,
-0.00039188977098092437,
0.04041041061282158,
0.2485661506652832,
0.057954657822847366,
-0.03563062474131584,
0.07826772332191467,
0.010085067711770535,
0.09473174065351486,
0.1261102706193924,
0.12425408512353897,
0.1579289734363556,
0.03184063732624054,
0.1436096876859665,
0.08334680646657944,
-0.02455662004649639,
-0.11893065273761749,
0.06034472584724426,
-0.0682448297739029,
0.09096178412437439,
0.02469257265329361,
0.20791736245155334,
0.09908261895179749,
-0.16395096480846405,
0.0043943943455815315,
-0.03697655349969864,
-0.08503299206495285,
-0.09561478346586227,
-0.05958867073059082,
-0.13126176595687866,
-0.146140456199646,
0.01031696517020464,
-0.106942318379879,
0.03492067754268646,
0.0701318234205246,
0.014409836381673813,
0.00022174170590005815,
0.14051245152950287,
0.01553852204233408,
0.029242079704999924,
0.0966508761048317,
0.008369714953005314,
-0.04007099196314812,
0.0006236056215129793,
-0.10348640382289886,
0.02355637401342392,
0.005408172495663166,
0.056937042623758316,
-0.021044857800006866,
-0.024828292429447174,
0.06884340196847916,
-0.025652971118688583,
-0.12562932074069977,
0.01088319718837738,
0.01964438706636429,
0.060035668313503265,
0.04402228444814682,
0.05651693791151047,
-0.01734256185591221,
0.0248918104916811,
0.20764942467212677,
-0.0891345664858818,
-0.07548970729112625,
-0.13299985229969025,
0.148090198636055,
-0.013996648602187634,
-0.007559389341622591,
0.009712876752018929,
-0.1060309037566185,
0.001951043144799769,
0.1929313987493515,
0.14812956750392914,
-0.07468932867050171,
-0.0011283630738034844,
-0.02684847079217434,
-0.006728011183440685,
-0.037687480449676514,
0.0669892430305481,
0.07768827676773071,
0.0342918299138546,
-0.059824999421834946,
-0.060962822288274765,
-0.05751146748661995,
-0.04114268720149994,
-0.022563505917787552,
0.03889298066496849,
-0.031851742416620255,
-0.02217292971909046,
-0.05023286119103432,
0.07799110561609268,
-0.0824526697397232,
-0.09598786383867264,
0.007823686115443707,
-0.21760009229183197,
-0.17412254214286804,
-0.0020312522538006306,
0.0752948448061943,
0.034618690609931946,
0.02598434127867222,
-0.03410730138421059,
0.026231758296489716,
0.056471534073352814,
-0.014275571331381798,
-0.05652080848813057,
-0.06083887070417404,
0.042322780936956406,
-0.08195296674966812,
0.17480778694152832,
-0.004077015910297632,
0.06578070670366287,
0.10385921597480774,
0.08150412142276764,
-0.10792244225740433,
0.10291001945734024,
0.06052424758672714,
-0.07460279017686844,
0.0556335411965847,
0.15228016674518585,
-0.056208327412605286,
0.14373594522476196,
0.05159805715084076,
-0.10221272706985474,
-0.00005992828664602712,
0.007738456130027771,
-0.02835691347718239,
-0.07475341856479645,
-0.06627774238586426,
-0.045850396156311035,
0.14703334867954254,
0.13515812158584595,
-0.06621544063091278,
0.0010646163718774915,
-0.016657905653119087,
0.05637483671307564,
0.06231863796710968,
0.022338880226016045,
-0.06155761331319809,
-0.2838592827320099,
-0.01603774167597294,
0.03917286917567253,
0.02194291166961193,
-0.2428244948387146,
-0.09001222252845764,
-0.009078788571059704,
-0.046210940927267075,
-0.07433000206947327,
0.09481046348810196,
0.0794455036520958,
0.031459469348192215,
-0.05482996627688408,
-0.0526672825217247,
-0.029080521315336227,
0.1730322390794754,
-0.16342733800411224,
-0.1152462363243103
] |
null | null | fastai |
# Amazing!
🥳 Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the [documentation here](https://huggingface.co/docs/hub/model-repos))!
2. Create a demo in Gradio or Streamlit using 🤗 Spaces ([documentation here](https://huggingface.co/docs/hub/spaces)).
3. Join the fastai community on the [Fastai Discord](https://discord.com/invite/YKrxeNn)!
Greetings fellow fastlearner 🤝! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| {"tags": ["fastai"]} | null | islasher/intel-image-classification | [
"fastai",
"has_space",
"region:us"
] | 2024-02-07T17:51:13+00:00 | [] | [] | TAGS
#fastai #has_space #region-us
|
# Amazing!
Congratulations on hosting your fastai model on the Hugging Face Hub!
# Some next steps
1. Fill out this model card with more information (see the template below and the documentation here)!
2. Create a demo in Gradio or Streamlit using Spaces (documentation here).
3. Join the fastai community on the Fastai Discord!
Greetings fellow fastlearner ! Don't forget to delete this content from your model card.
---
# Model card
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
| [
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
"TAGS\n#fastai #has_space #region-us \n",
"# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!",
"# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---",
"# Model card",
"## Model description\nMore information needed",
"## Intended uses & limitations\nMore information needed",
"## Training and evaluation data\nMore information needed"
] | [
13,
20,
79,
3,
6,
12,
8
] | [
"passage: TAGS\n#fastai #has_space #region-us \n# Amazing!\n\n Congratulations on hosting your fastai model on the Hugging Face Hub!# Some next steps\n1. Fill out this model card with more information (see the template below and the documentation here)!\n\n2. Create a demo in Gradio or Streamlit using Spaces (documentation here).\n\n3. Join the fastai community on the Fastai Discord!\n\nGreetings fellow fastlearner ! Don't forget to delete this content from your model card.\n\n\n---# Model card## Model description\nMore information needed## Intended uses & limitations\nMore information needed## Training and evaluation data\nMore information needed"
] | [
-0.048121724277734756,
-0.024616125971078873,
0.002038001548498869,
0.10439170897006989,
0.135872021317482,
0.11887997388839722,
0.07405775785446167,
0.09980081021785736,
0.07783667743206024,
0.02590852417051792,
0.08961158245801926,
-0.08088712394237518,
0.08744348585605621,
0.271692156791687,
0.06988707184791565,
-0.22761479020118713,
0.04051019623875618,
-0.00024903909070417285,
0.08053462207317352,
0.06629016250371933,
0.13507555425167084,
-0.05464952811598778,
0.14010503888130188,
-0.004088983871042728,
-0.19050447642803192,
-0.042929794639348984,
-0.01773718371987343,
-0.02527874894440174,
0.12317648530006409,
-0.04744937643408775,
0.05381017178297043,
0.015037551522254944,
0.007565062493085861,
-0.07253646105527878,
0.0623294934630394,
0.040457066148519516,
0.01740180514752865,
0.059235580265522,
-0.07249044626951218,
0.08950132131576538,
0.08404164761304855,
-0.024370938539505005,
-0.1097978875041008,
0.07827875018119812,
-0.14424212276935577,
-0.21762843430042267,
-0.1253085881471634,
-0.09017651528120041,
0.028519365936517715,
0.004388005938380957,
-0.025051530450582504,
0.12801909446716309,
-0.13558274507522583,
-0.040698226541280746,
0.20124278962612152,
-0.17012301087379456,
-0.05505548417568207,
0.034343402832746506,
0.09226689487695694,
-0.05829555168747902,
-0.06347129493951797,
0.10614984482526779,
0.09640881419181824,
-0.019833475351333618,
0.05516824126243591,
0.002579754451289773,
0.021173657849431038,
0.01370104867964983,
-0.06150497496128082,
0.04717832803726196,
-0.010183089412748814,
0.048132527619600296,
-0.09465572983026505,
-0.1303568333387375,
-0.004072192590683699,
0.01214400865137577,
-0.048744890838861465,
-0.07019646465778351,
0.07833103090524673,
-0.011118141002953053,
-0.04357248544692993,
-0.13031910359859467,
-0.09131011366844177,
-0.12358787655830383,
0.008646543137729168,
0.09500427544116974,
0.003679296001791954,
0.07374339550733566,
-0.08258994668722153,
0.06774985045194626,
-0.17329485714435577,
-0.06484591960906982,
-0.08138520270586014,
-0.11546400189399719,
0.021133482456207275,
-0.0387684591114521,
0.02668963186442852,
0.15394504368305206,
0.12983950972557068,
0.023976242169737816,
0.04388163983821869,
-0.038937073200941086,
0.051190316677093506,
0.058571770787239075,
0.03395717963576317,
0.034934818744659424,
-0.036981891840696335,
-0.1793210655450821,
-0.016702448949217796,
-0.011550825089216232,
0.07954040914773941,
-0.07523109763860703,
-0.05632320046424866,
0.013454885222017765,
-0.11071494966745377,
0.07202339172363281,
-0.03576776012778282,
-0.0032025426626205444,
0.01168301422148943,
0.018371861428022385,
0.21271461248397827,
0.03955606371164322,
0.014191740192472935,
-0.008875265717506409,
-0.13453757762908936,
-0.06874168664216995,
-0.06896194815635681,
0.03361047804355621,
0.04448792710900307,
-0.0028071461711078882,
-0.07672245055437088,
0.04325154796242714,
-0.06045534089207649,
-0.03508453071117401,
0.008032378740608692,
-0.18221288919448853,
0.007458044681698084,
-0.10049355030059814,
-0.12126200646162033,
0.05306628718972206,
0.01695440337061882,
-0.08215925842523575,
0.08141279965639114,
0.02662261202931404,
0.020931517705321312,
-0.009988143108785152,
-0.005391082260757685,
0.06874798238277435,
-0.08508864045143127,
0.029901226982474327,
0.17170792818069458,
0.13024519383907318,
-0.08046911656856537,
-0.0006887061172164977,
-0.10965746641159058,
0.04426072910428047,
-0.13325683772563934,
0.02251482754945755,
-0.09062390774488449,
0.11723794043064117,
-0.042396437376737595,
0.002038756385445595,
-0.029030200093984604,
0.0960269495844841,
0.08189879357814789,
0.16663365066051483,
-0.2419009804725647,
-0.031095001846551895,
0.13240347802639008,
-0.10711425542831421,
-0.1807439625263214,
0.18486657738685608,
-0.012035200372338295,
0.11329247802495956,
-0.047014184296131134,
0.18334640562534332,
-0.02612062357366085,
-0.13582459092140198,
-0.058872904628515244,
0.005852419883012772,
-0.2269321084022522,
-0.06286033242940903,
0.09738040715456009,
0.13425657153129578,
-0.042984943836927414,
0.007112155202776194,
0.026316028088331223,
0.13609857857227325,
-0.06715573370456696,
-0.05195777863264084,
-0.012255736626684666,
-0.10902371257543564,
0.041914235800504684,
0.018215661868453026,
0.035408079624176025,
-0.059880174696445465,
-0.02931194379925728,
-0.053190283477306366,
0.13146710395812988,
0.09760832786560059,
-0.03670211136341095,
-0.049620725214481354,
0.1689043790102005,
-0.07763876020908356,
-0.033587727695703506,
0.07560533285140991,
-0.08268500119447708,
0.03266897425055504,
0.03090597130358219,
0.055881720036268234,
0.07766123116016388,
0.08522116392850876,
0.06057543307542801,
0.00819048099219799,
0.034654274582862854,
0.12095347046852112,
-0.013591280207037926,
-0.05039411783218384,
0.021508218720555305,
0.016904234886169434,
-0.019032588228583336,
0.29030677676200867,
-0.1951042115688324,
0.024724548682570457,
-0.06477324664592743,
0.07631538063287735,
0.06136792525649071,
0.003575638635084033,
0.08580143749713898,
-0.06023019179701805,
-0.019061198458075523,
-0.04803973436355591,
0.046805646270513535,
-0.0666879191994667,
-0.04162997007369995,
0.2621194124221802,
-0.05497581139206886,
0.044914912432432175,
0.12313763797283173,
-0.05873025581240654,
-0.07091446220874786,
0.01009807363152504,
-0.00793424155563116,
0.03249288722872734,
-0.04042816907167435,
0.043721720576286316,
-0.10840129852294922,
-0.06674089282751083,
0.1573198139667511,
-0.038477856665849686,
0.06786153465509415,
0.032288823276758194,
-0.04958454892039299,
-0.0648743286728859,
0.04650486260652542,
0.13598160445690155,
-0.0875244215130806,
0.07435166835784912,
0.17612984776496887,
-0.010562662966549397,
0.168031245470047,
0.08435525000095367,
-0.07075224816799164,
-0.09465329349040985,
-0.051014289259910583,
-0.021595727652311325,
0.21222901344299316,
-0.07084725052118301,
-0.054564714431762695,
0.05911700800061226,
-0.013703816570341587,
0.07196151465177536,
-0.06009222939610481,
-0.08332337439060211,
0.03227344527840614,
-0.04517695680260658,
0.011517706327140331,
0.13512636721134186,
-0.07090822607278824,
0.04681389778852463,
0.031489867717027664,
-0.0662703812122345,
0.02217509225010872,
0.033389873802661896,
0.0068921963684260845,
0.033959709107875824,
0.07332495599985123,
-0.20893315970897675,
-0.08408680558204651,
-0.13727638125419617,
0.037881869822740555,
0.021770721301436424,
0.045787326991558075,
-0.08602345734834671,
0.02231026627123356,
-0.08954031765460968,
-0.07987114042043686,
0.029592275619506836,
-0.026350297033786774,
-0.11349643021821976,
-0.03396226093173027,
-0.009560913778841496,
-0.06662604957818985,
-0.02250705659389496,
-0.05024505779147148,
0.03983384370803833,
0.04479299485683441,
0.058377087116241455,
0.12796473503112793,
-0.013808943331241608,
-0.03839317709207535,
0.000370211957488209,
-0.022712308913469315,
0.16396735608577728,
-0.14746315777301788,
0.07954913377761841,
0.19160102307796478,
0.11742953956127167,
0.028144672513008118,
0.028885571286082268,
0.03537585213780403,
-0.06289814412593842,
-0.000050317394197918475,
0.03226194158196449,
-0.09392514824867249,
-0.05801016092300415,
-0.020014392212033272,
-0.04031052812933922,
0.17134574055671692,
-0.12160717695951462,
0.03345204517245293,
0.04098419472575188,
0.09783966839313507,
0.10073629021644592,
-0.028829937800765038,
-0.1815856397151947,
0.038818612694740295,
-0.24060091376304626,
-0.05831146240234375,
0.027899866923689842,
-0.09110201895236969,
-0.06232144311070442,
0.17409387230873108,
0.013794700615108013,
0.011769929900765419,
-0.006736889015883207,
0.07983319461345673,
0.0110100656747818,
0.1217205822467804,
0.05947643890976906,
-0.05539114400744438,
0.025202350690960884,
-0.09962950646877289,
-0.07107596844434738,
-0.04035590961575508,
-0.05832801014184952,
0.07548832893371582,
0.1409129947423935,
-0.025475580245256424,
-0.020795362070202827,
0.023489827290177345,
0.08550169318914413,
0.0423230417072773,
0.16739299893379211,
-0.16016584634780884,
-0.026555389165878296,
0.04571257904171944,
-0.03384667634963989,
-0.05433850735425949,
-0.010291114449501038,
0.1137225553393364,
-0.02820689231157303,
-0.040318265557289124,
0.021242983639240265,
0.06503437459468842,
0.01481706090271473,
0.05012747645378113,
-0.04056356102228165,
0.14796851575374603,
-0.03461192920804024,
0.019330544397234917,
-0.12413888424634933,
0.13848772644996643,
0.021095896139740944,
-0.03901609033346176,
-0.06735876202583313,
-0.05808034539222717,
0.18150931596755981,
0.0025602965615689754,
0.10535930097103119,
0.012098877690732479,
-0.12160047143697739,
-0.1359938681125641,
-0.11211287975311279,
0.005111907608807087,
0.08330471813678741,
-0.023147236555814743,
-0.022247863933444023,
0.022165266796946526,
-0.036149751394987106,
-0.0530381016433239,
0.15749511122703552,
-0.1289154291152954,
-0.001082550617866218,
0.014728817157447338,
0.06971760839223862,
-0.08223173767328262,
0.026267826557159424,
0.014071501791477203,
-0.1119147390127182,
0.10590848326683044,
0.2521335482597351,
0.10338116437196732,
-0.09591643512248993,
-0.07697287201881409,
0.03418830782175064,
-0.012184361927211285,
-0.000774814048781991,
-0.006932659074664116,
0.0495428591966629,
-0.005566445179283619,
0.006762749515473843,
0.12971895933151245,
-0.07130889594554901,
0.011540771462023258,
-0.08449850976467133,
0.05566910281777382,
-0.05276734381914139,
0.01761564053595066,
-0.002672141883522272,
-0.008124710991978645,
-0.07340748608112335,
-0.061829522252082825,
0.1609770804643631,
-0.07277000695466995,
-0.06468547880649567,
0.05801168829202652,
0.03307786211371422,
0.01431563775986433,
-0.03584568202495575,
-0.04342148080468178,
0.18088261783123016,
0.29330700635910034,
-0.08191116154193878,
0.10001859813928604,
0.09677296131849289,
0.034820813685655594,
-0.23625829815864563,
0.029798466712236404,
-0.1455078274011612,
0.04449721798300743,
0.040447335690259933,
-0.0409548319876194,
0.04191497340798378,
0.10835777968168259,
-0.06094440817832947,
0.2048867791891098,
-0.03527235612273216,
-0.07983248680830002,
-0.01788630709052086,
0.03109324350953102,
0.29443636536598206,
-0.11833466589450836,
0.006058716680854559,
-0.10420958697795868,
-0.21566011011600494,
0.06983078271150589,
-0.18948867917060852,
0.13948246836662292,
-0.05087858438491821,
0.03576415032148361,
-0.01149723306298256,
-0.07561972737312317,
0.20518061518669128,
-0.15641045570373535,
0.05273103713989258,
-0.13722458481788635,
-0.1327189952135086,
0.01617460884153843,
-0.10048147290945053,
0.1545477658510208,
-0.11024226248264313,
-0.023215843364596367,
-0.2284185290336609,
0.012587235309183598,
-0.023200806230306625,
0.10030807554721832,
0.01800704374909401,
-0.07980740070343018,
-0.08767345547676086,
0.1316242516040802,
-0.06486566364765167,
0.034810543060302734,
-0.06996636837720871,
-0.050714004784822464,
-0.010929876938462257,
-0.045061707496643066,
0.03034941293299198,
-0.07934719324111938,
0.15192505717277527,
-0.016938980668783188,
-0.04507075995206833,
0.08636019378900528,
-0.2479533851146698,
0.023727843537926674,
0.025351112708449364,
-0.03495599329471588,
0.09001832455396652,
-0.025513244792819023,
-0.06256973743438721,
0.12282291799783707,
0.1402233988046646,
-0.07322840392589569,
-0.2460673749446869,
-0.06281693279743195,
0.0076784128323197365,
0.039165716618299484,
0.06561196595430374,
0.05125982314348221,
-0.07261458039283752,
-0.011131617240607738,
-0.026896944269537926,
0.030595947057008743,
-0.11692017316818237,
-0.03854857385158539,
0.07790639251470566,
0.017095070332288742,
-0.07846562564373016,
0.07280377298593521,
0.014225782826542854,
-0.021511616185307503,
0.007357571739703417,
0.148970365524292,
0.007519228849560022,
-0.14747941493988037,
-0.06656096875667572,
0.2007484883069992,
-0.01197928935289383,
-0.07260087132453918,
-0.05383119732141495,
-0.008990069851279259,
-0.0476234145462513,
0.05585788935422897,
0.05367223918437958,
-0.013585401698946953,
0.07708586007356644,
0.06263149529695511,
-0.10210110992193222,
-0.046256959438323975,
-0.066561758518219,
0.04169114679098129,
-0.10485753417015076,
0.060470130294561386,
0.009529483504593372,
0.12185006588697433,
-0.09983488917350769,
-0.01802929677069187,
-0.10810204595327377,
-0.06766588985919952,
-0.17349553108215332,
-0.05834362283349037,
-0.041105758398771286,
-0.015651104971766472,
0.03658895567059517,
0.010445823892951012,
-0.057867538183927536,
-0.0442853718996048,
-0.07536603510379791,
0.038444988429546356,
0.06147460639476776,
0.03932281583547592,
-0.03912714496254921,
0.04001858830451965,
0.05909334123134613,
0.013087345287203789,
0.17542624473571777,
0.038768354803323746,
0.05504675209522247,
-0.05045998468995094,
-0.16491834819316864,
-0.05276111513376236,
-0.0074316514655947685,
-0.07559102028608322,
0.1224973127245903,
-0.007679440546780825,
0.007880088873207569,
-0.08065467327833176,
0.03924860805273056,
0.028234204277396202,
0.10404064506292343,
-0.0028364830650389194,
0.10070426017045975,
0.019627176225185394,
-0.07226712256669998,
-0.025392837822437286,
0.021809715777635574,
0.12809939682483673,
0.01567147858440876,
0.026090998202562332,
0.033139873296022415,
0.016619985923171043,
-0.057361043989658356,
0.033977724611759186,
-0.04997231811285019,
-0.15123651921749115,
0.02628709189593792,
-0.05165188014507294,
0.005062380339950323,
-0.016889680176973343,
0.20362506806850433,
0.07867538928985596,
-0.06474173814058304,
-0.010664013214409351,
0.015816617757081985,
-0.0168940220028162,
-0.03121885471045971,
-0.012740966863930225,
0.04592578858137131,
-0.001151384087279439,
-0.04866636544466019,
0.11825273931026459,
0.05015748366713524,
0.05386412516236305,
0.0596686452627182,
0.12528513371944427,
0.016759619116783142,
0.13257254660129547,
0.061999931931495667,
-0.03403807803988457,
-0.13461735844612122,
-0.04495539888739586,
-0.1254577934741974,
0.04646851494908333,
-0.08697032928466797,
0.09941662102937698,
0.1144254133105278,
-0.05959030240774155,
-0.030464433133602142,
-0.08851305395364761,
-0.008356761187314987,
-0.06041252240538597,
0.039516255259513855,
-0.02262675203382969,
-0.0873224213719368,
0.0481097511947155,
0.05495472997426987,
-0.022752324119210243,
0.13218675553798676,
0.015727028250694275,
-0.036317698657512665,
0.13270340859889984,
-0.07583184540271759,
0.11758984625339508,
0.061510033905506134,
-0.043043944984674454,
-0.11560922116041183,
-0.020150646567344666,
-0.06641761213541031,
-0.10098972916603088,
-0.006782987620681524,
-0.005399650428444147,
-0.07349002361297607,
-0.059971679002046585,
0.08397487550973892,
-0.03124053031206131,
-0.09979676455259323,
-0.032152675092220306,
0.0038895104080438614,
0.06054706871509552,
-0.01686914451420307,
-0.0034020058810710907,
0.04728743061423302,
0.015076374635100365,
0.1653461456298828,
-0.02208263985812664,
0.06234867498278618,
-0.13855914771556854,
0.16070103645324707,
-0.14684462547302246,
-0.029404424130916595,
-0.1890171319246292,
-0.09729582816362381,
-0.05156542733311653,
0.20326784253120422,
0.2840938866138458,
-0.19109351933002472,
-0.010187864303588867,
0.020078664645552635,
-0.014484191313385963,
-0.08961770683526993,
0.12571553885936737,
0.029420215636491776,
-0.023631498217582703,
-0.07249019294977188,
-0.02037387527525425,
0.005258576478809118,
-0.06544211506843567,
-0.026979785412549973,
0.18310695886611938,
0.001496660872362554,
0.059546373784542084,
-0.09605178982019424,
0.01754261925816536,
-0.14839904010295868,
-0.10467469692230225,
-0.02111995778977871,
-0.16156397759914398,
-0.09646477550268173,
0.006635562051087618,
0.038640011101961136,
0.08000610023736954,
0.03268849849700928,
-0.015172510407865047,
0.06479045748710632,
-0.056333884596824646,
-0.0037216036580502987,
-0.1231912299990654,
0.00034658415825106204,
0.062129102647304535,
-0.07422006875276566,
0.2545335292816162,
-0.03070417232811451,
-0.12370815873146057,
0.09026903659105301,
-0.03299184888601303,
-0.12452623248100281,
0.07951879501342773,
-0.005700904875993729,
-0.11531132459640503,
-0.057989440858364105,
0.18941475450992584,
-0.012821312062442303,
-0.1364315301179886,
0.046368811279535294,
-0.17166484892368317,
0.031349923461675644,
0.0363016203045845,
-0.001313706859946251,
-0.04714022949337959,
0.024538639932870865,
-0.008008457720279694,
0.10724439471960068,
0.1382838785648346,
0.016739921644330025,
-0.011060068383812904,
-0.05056179314851761,
0.07912429422140121,
0.056927867233753204,
-0.05218246951699257,
-0.1282637119293213,
-0.08599764108657837,
0.03429819270968437,
0.04119478166103363,
-0.08113081753253937,
-0.16903182864189148,
-0.03668912500143051,
-0.10082915425300598,
-0.004939202684909105,
0.051785312592983246,
0.06585265696048737,
0.29044589400291443,
0.06326735019683838,
0.0016605621203780174,
-0.13649453222751617,
0.050569336861371994,
0.0868251696228981,
-0.04697931930422783,
-0.07670357078313828
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# swin-tiny-patch4-window7-224-finetuned-eurosat
This model is a fine-tuned version of [microsoft/swin-tiny-patch4-window7-224](https://huggingface.co/microsoft/swin-tiny-patch4-window7-224) on the imagefolder dataset.
It achieves the following results on the evaluation set:
- Loss: 0.1135
- Accuracy: 0.9703
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 32
- eval_batch_size: 32
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 128
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 5
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.1503 | 1.0 | 57 | 0.2204 | 0.9381 |
| 0.1349 | 2.0 | 114 | 0.1394 | 0.9567 |
| 0.0552 | 3.0 | 171 | 0.1430 | 0.9678 |
| 0.0722 | 4.0 | 228 | 0.1568 | 0.9629 |
| 0.0523 | 5.0 | 285 | 0.1135 | 0.9703 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["imagefolder"], "metrics": ["accuracy"], "base_model": "microsoft/swin-tiny-patch4-window7-224", "model-index": [{"name": "swin-tiny-patch4-window7-224-finetuned-eurosat", "results": [{"task": {"type": "image-classification", "name": "Image Classification"}, "dataset": {"name": "imagefolder", "type": "imagefolder", "config": "default", "split": "train", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.9702970297029703, "name": "Accuracy"}]}]}]} | image-classification | jvbjkbjkbfjis/swin-tiny-patch4-window7-224-finetuned-eurosat | [
"transformers",
"tensorboard",
"safetensors",
"swin",
"image-classification",
"generated_from_trainer",
"dataset:imagefolder",
"base_model:microsoft/swin-tiny-patch4-window7-224",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-07T17:51:48+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #swin #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/swin-tiny-patch4-window7-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| swin-tiny-patch4-window7-224-finetuned-eurosat
==============================================
This model is a fine-tuned version of microsoft/swin-tiny-patch4-window7-224 on the imagefolder dataset.
It achieves the following results on the evaluation set:
* Loss: 0.1135
* Accuracy: 0.9703
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 32
* eval\_batch\_size: 32
* seed: 42
* gradient\_accumulation\_steps: 4
* total\_train\_batch\_size: 128
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* num\_epochs: 5
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #swin #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/swin-tiny-patch4-window7-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
88,
144,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #swin #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/swin-tiny-patch4-window7-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 5### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.11950050294399261,
0.16994433104991913,
-0.0024043042212724686,
0.09070786088705063,
0.11077117919921875,
0.02888409234583378,
0.1078600138425827,
0.13683843612670898,
-0.0632045567035675,
0.1170341968536377,
0.1388625204563141,
0.08470466732978821,
0.07028268277645111,
0.15406827628612518,
-0.005242266226559877,
-0.2928418517112732,
0.018512075766921043,
-0.01286622229963541,
-0.1416057050228119,
0.10939979553222656,
0.06758690625429153,
-0.1260313093662262,
0.09106457978487015,
0.004312277305871248,
-0.14306758344173431,
-0.02729647047817707,
-0.039466340094804764,
-0.046252354979515076,
0.09939402341842651,
0.036541897803545,
0.0819014459848404,
0.03265363723039627,
0.1136382594704628,
-0.2321285456418991,
0.007382033858448267,
0.07358430325984955,
0.012512996792793274,
0.09743890911340714,
0.11765105277299881,
0.016987374052405357,
0.13907840847969055,
-0.11078713089227676,
0.06568101048469543,
0.041207753121852875,
-0.08268461376428604,
-0.23632708191871643,
-0.061157338321208954,
0.09002139419317245,
0.12318366765975952,
0.05324169248342514,
-0.009605195373296738,
0.07989849150180817,
-0.06777787208557129,
0.08403339236974716,
0.22544069588184357,
-0.24414610862731934,
-0.07347176969051361,
0.04207862168550491,
0.029460439458489418,
0.03760809078812599,
-0.1328730285167694,
-0.007643865887075663,
0.03938794136047363,
0.0019549953285604715,
0.11146232485771179,
0.026491722092032433,
0.06161850318312645,
0.008195897564291954,
-0.14075787365436554,
-0.04594893753528595,
0.09538660198450089,
0.10818411409854889,
-0.01845346763730049,
-0.12089847028255463,
-0.054673753678798676,
-0.1943042278289795,
-0.044403888285160065,
0.0125008225440979,
0.040390659123659134,
-0.05661248788237572,
-0.07803214341402054,
0.030570290982723236,
-0.07247446477413177,
-0.07880302518606186,
0.04599347338080406,
0.12437744438648224,
0.060991499572992325,
-0.00408996781334281,
0.02093009650707245,
0.11760847270488739,
0.09421813488006592,
-0.16259217262268066,
-0.00004423544305609539,
0.005923346616327763,
-0.07294879108667374,
-0.0005708226817660034,
-0.00815062876790762,
0.02537432126700878,
0.04288133606314659,
0.14030762016773224,
-0.02654362842440605,
0.08017711341381073,
0.08491623401641846,
0.020853087306022644,
-0.07924521714448929,
0.14527471363544464,
-0.08038394898176193,
-0.08875207602977753,
-0.028287511318922043,
0.1195329874753952,
0.03180062025785446,
-0.008725588209927082,
-0.08532034605741501,
0.022093672305345535,
0.10758572816848755,
0.02816654182970524,
-0.001469830283895135,
0.042382609099149704,
-0.05636901780962944,
-0.03136960417032242,
0.08470446616411209,
-0.08506551384925842,
0.04424486309289932,
0.0352679118514061,
-0.06620503216981888,
-0.011191361583769321,
0.02751638926565647,
-0.014266911894083023,
0.006648977752774954,
0.10702378302812576,
-0.09861622005701065,
-0.028851326555013657,
-0.07986565679311752,
-0.07760193943977356,
0.03044244460761547,
-0.09168349206447601,
0.01670023612678051,
-0.08502736687660217,
-0.11222101747989655,
-0.038456786423921585,
0.06433568894863129,
-0.06233419477939606,
-0.07303836196660995,
-0.05183456838130951,
-0.10063394904136658,
0.05975358560681343,
0.006967143155634403,
0.13018356263637543,
-0.05210058018565178,
0.0946609377861023,
0.00431168032810092,
0.07898398488759995,
0.06657607108354568,
0.03675604984164238,
-0.06349555402994156,
0.06671591848134995,
-0.16197559237480164,
0.05189656466245651,
-0.08722544461488724,
0.06786655634641647,
-0.12130050361156464,
-0.1033872663974762,
-0.0081481309607625,
-0.013023391366004944,
0.06549853086471558,
0.14419953525066376,
-0.15232224762439728,
-0.06823685765266418,
0.14789272844791412,
-0.08846236020326614,
-0.11842364817857742,
0.10507562011480331,
-0.013635094277560711,
-0.06015276536345482,
0.011256725527346134,
0.16566094756126404,
0.08404062688350677,
-0.08553722500801086,
-0.03551160544157028,
0.004692078568041325,
0.09748972207307816,
-0.0029218532145023346,
0.10211949795484543,
-0.0007343683973886073,
0.011453190818428993,
0.01789722964167595,
-0.0763038694858551,
0.07689081877470016,
-0.0899737998843193,
-0.07893429696559906,
-0.03931069001555443,
-0.08503801375627518,
0.02809973992407322,
0.06308432668447495,
0.02488207072019577,
-0.07859069108963013,
-0.1359645277261734,
0.014937418513000011,
0.12012416869401932,
-0.09603691101074219,
-0.006339706480503082,
-0.05454953387379646,
0.07169663906097412,
-0.05562763288617134,
-0.010570480488240719,
-0.12704473733901978,
-0.07442354410886765,
0.03377588838338852,
-0.0839192196726799,
-0.015520067885518074,
-0.008330265060067177,
0.07430456578731537,
0.08914850652217865,
-0.05584233999252319,
-0.08939135819673538,
-0.05659959837794304,
0.010936996899545193,
-0.07871481776237488,
-0.25899389386177063,
-0.08048082143068314,
-0.027924545109272003,
0.1444798856973648,
-0.2558952569961548,
0.016200633719563484,
0.013488425873219967,
0.14574411511421204,
0.04524271562695503,
-0.058805227279663086,
0.00411091698333621,
0.012090793810784817,
-0.04465070739388466,
-0.09927614778280258,
0.033087026327848434,
0.0021102039609104395,
-0.10776393860578537,
-0.024244094267487526,
-0.1188773438334465,
0.12201521545648575,
0.10383828729391098,
0.013092180714011192,
-0.093838632106781,
-0.04232233390212059,
-0.0762806087732315,
-0.05552899092435837,
-0.022646982222795486,
0.017645403742790222,
0.08303063362836838,
0.010630463249981403,
0.10753629356622696,
-0.08041738718748093,
-0.05653155595064163,
0.042364250868558884,
-0.0024351580068469048,
-0.02858918532729149,
0.14119809865951538,
0.1093788743019104,
-0.07587239146232605,
0.13383986055850983,
0.13027672469615936,
-0.053452715277671814,
0.12906904518604279,
-0.05853378027677536,
-0.09874875098466873,
-0.032599762082099915,
0.023279482498764992,
0.01867278292775154,
0.1547902375459671,
-0.09362509101629257,
0.00957106426358223,
0.02713697962462902,
0.010420500300824642,
0.011649789288640022,
-0.1736517995595932,
-0.015446819365024567,
0.046254713088274,
-0.04940564185380936,
0.017073538154363632,
-0.031026357784867287,
-0.024483781307935715,
0.09221348166465759,
0.003182237269356847,
-0.04947255179286003,
-0.005072079133242369,
-0.0060454243794083595,
-0.08078867197036743,
0.2097596377134323,
-0.07531048357486725,
-0.14715003967285156,
-0.12436766177415848,
0.03937062621116638,
-0.04030890390276909,
-0.004724894650280476,
0.01863381825387478,
-0.10655622184276581,
-0.05315682291984558,
-0.08650362491607666,
0.002804737538099289,
-0.012968841008841991,
0.052109163254499435,
0.010501560755074024,
0.016577037051320076,
0.08061236143112183,
-0.08427710831165314,
0.022501081228256226,
-0.008124095387756824,
-0.010881118476390839,
0.03141538053750992,
0.04434659332036972,
0.119411401450634,
0.12910343706607819,
0.016705160960555077,
0.017432838678359985,
-0.00872090645134449,
0.19080547988414764,
-0.09374217689037323,
0.030144335702061653,
0.10056699812412262,
-0.00311025301925838,
0.050383564084768295,
0.13380637764930725,
0.044719550758600235,
-0.0735592171549797,
0.014838800765573978,
0.03402961790561676,
-0.01665291003882885,
-0.1919606626033783,
-0.02998001128435135,
-0.027816979214549065,
0.007187872193753719,
0.13367004692554474,
0.04616563022136688,
-0.03301175311207771,
0.06924106180667877,
-0.017359089106321335,
0.009281757287681103,
-0.018469950184226036,
0.07115042954683304,
0.023051133379340172,
0.04823097959160805,
0.10704479366540909,
-0.038076549768447876,
-0.02328427881002426,
0.03912581875920296,
-0.0038811424747109413,
0.21581578254699707,
-0.030109068378806114,
0.1444968432188034,
0.02386939898133278,
0.16672350466251373,
0.004386989865452051,
0.06324020028114319,
0.016281787306070328,
-0.0328756719827652,
0.005287684965878725,
-0.05349832400679588,
-0.024165237322449684,
0.05411134287714958,
0.020157840102910995,
0.05866680666804314,
-0.10918577015399933,
0.06854833662509918,
0.045613475143909454,
0.26513129472732544,
0.07564223557710648,
-0.3359636068344116,
-0.09144284576177597,
0.015733208507299423,
-0.03469613194465637,
-0.045324165374040604,
0.023864958435297012,
0.15672637522220612,
-0.08493366837501526,
0.07361623644828796,
-0.08649015426635742,
0.0677950531244278,
-0.07103931158781052,
-0.0038066068664193153,
0.08761058747768402,
0.10813698172569275,
0.004094721283763647,
0.0749434158205986,
-0.19476599991321564,
0.2546394467353821,
-0.007903964258730412,
0.04451516643166542,
-0.059066835790872574,
0.03170173987746239,
0.027517855167388916,
0.0209338441491127,
0.11232953518629074,
-0.0036870113108307123,
-0.10501988232135773,
-0.18415123224258423,
-0.12253081798553467,
0.021683892235159874,
0.11624240130186081,
-0.08178006857633591,
0.11325155943632126,
-0.03272707015275955,
-0.038958434015512466,
0.04861238971352577,
-0.06277395039796829,
-0.08184365183115005,
-0.12542986869812012,
0.000960197183303535,
-0.03678678348660469,
0.006017159204930067,
-0.09529890865087509,
-0.10430572926998138,
-0.0953797772526741,
0.14844128489494324,
-0.11337782442569733,
-0.0392225906252861,
-0.1559082418680191,
0.10419976711273193,
0.1431107074022293,
-0.08315407484769821,
0.06201969459652901,
-0.00918690674006939,
0.12680982053279877,
0.038079436868429184,
-0.04849713668227196,
0.11457640677690506,
-0.0969407930970192,
-0.22941599786281586,
-0.05639585852622986,
0.11244680732488632,
0.03961538150906563,
0.05957407131791115,
-0.026297034695744514,
0.021516749635338783,
-0.014326500706374645,
-0.09667489677667618,
0.06002586707472801,
0.04046012461185455,
0.036001838743686676,
0.01801140233874321,
-0.03817407414317131,
0.02920815534889698,
-0.028255078941583633,
-0.03360797464847565,
0.105784110724926,
0.2793390154838562,
-0.11845361441373825,
0.023951314389705658,
0.02863282710313797,
-0.04780091345310211,
-0.18068718910217285,
0.016104746609926224,
0.10343805700540543,
0.024521535262465477,
0.034613557159900665,
-0.1751137375831604,
0.10656893253326416,
0.08605173230171204,
-0.024016855284571648,
0.10203297436237335,
-0.2901950478553772,
-0.12156055122613907,
0.0927579253911972,
0.13462647795677185,
-0.03578878194093704,
-0.16806933283805847,
-0.05478183180093765,
-0.0037065150681883097,
-0.07336048036813736,
0.08677427470684052,
0.0037975222803652287,
0.09822305291891098,
-0.030320527032017708,
-0.017211761325597763,
0.023378882557153702,
-0.07288448512554169,
0.1592586487531662,
-0.013518577441573143,
0.0868251770734787,
-0.03213617950677872,
0.015376942232251167,
-0.004471026826649904,
-0.07758694142103195,
0.035730354487895966,
-0.11660291999578476,
0.05552012845873833,
-0.10184155404567719,
-0.014926769770681858,
-0.07859154790639877,
0.027983304113149643,
-0.051641810685396194,
-0.04203711822628975,
-0.038760166615247726,
0.04867914691567421,
0.0768495425581932,
-0.0017842553788796067,
0.14007334411144257,
0.01396908424794674,
0.09840934723615646,
0.11570145934820175,
0.05573178082704544,
0.006528177298605442,
-0.10155165195465088,
-0.03668413311243057,
-0.009497709572315216,
0.0487414225935936,
-0.15242204070091248,
0.013252504169940948,
0.13000977039337158,
0.040250957012176514,
0.11608676612377167,
0.051124222576618195,
-0.052723247557878494,
-0.01728632301092148,
0.08339253813028336,
-0.1104716956615448,
-0.1377519965171814,
-0.025719277560710907,
0.005164583679288626,
-0.1618347465991974,
0.018058698624372482,
0.07504202425479889,
-0.06854446232318878,
0.005250264424830675,
0.0009384627919644117,
0.05041106045246124,
0.004460912197828293,
0.19176894426345825,
0.08239445835351944,
0.08067938685417175,
-0.08650229871273041,
0.10857374221086502,
0.031158102676272392,
-0.14169184863567352,
0.025002675130963326,
0.06693225353956223,
-0.08038315922021866,
-0.011664786376059055,
0.0889279767870903,
0.09677787125110626,
-0.020468536764383316,
-0.044290393590927124,
-0.1267208755016327,
-0.11951704323291779,
0.06958357244729996,
0.0651080384850502,
0.06624583899974823,
0.02064638026058674,
-0.004233045969158411,
0.030909201130270958,
-0.10943175107240677,
0.1408100724220276,
0.07691428065299988,
0.10009953379631042,
-0.19225816428661346,
0.08819769322872162,
0.010070092976093292,
0.007894113659858704,
-0.014841558411717415,
0.05150880664587021,
-0.12395233660936356,
-0.030303681269288063,
-0.06921202689409256,
0.009988794103264809,
-0.07002849876880646,
0.008074463345110416,
0.00209794077090919,
-0.05013373866677284,
-0.03931139409542084,
0.007122091483324766,
-0.09397649765014648,
-0.060595571994781494,
0.00010608405864331871,
0.06329943239688873,
-0.09978330135345459,
-0.01711118593811989,
0.03747502341866493,
-0.12110644578933716,
0.09187658131122589,
0.013313835486769676,
0.044154368340969086,
0.012743385508656502,
-0.08616949617862701,
0.03046351484954357,
0.04840420186519623,
-0.0024554701521992683,
0.025584537535905838,
-0.13373804092407227,
-0.004638702608644962,
-0.049549322575330734,
-0.007380565628409386,
-0.021161118522286415,
0.0469607375562191,
-0.13595595955848694,
0.0010805778438225389,
-0.06016739085316658,
-0.05026865378022194,
-0.06151561066508293,
0.04923491179943085,
0.068328358232975,
-0.016896545886993408,
0.16867835819721222,
-0.07501363009214401,
0.040658704936504364,
-0.23758703470230103,
-0.0017444806871935725,
-0.013450528495013714,
-0.06353842467069626,
-0.08612404018640518,
-0.009639346972107887,
0.07726234197616577,
-0.05196730047464371,
0.09326780587434769,
-0.03524046763777733,
0.019063053652644157,
0.027385523542761803,
-0.034646764397621155,
0.0451643280684948,
0.05181615427136421,
0.1959172636270523,
0.01719692349433899,
-0.014387324452400208,
0.06618281453847885,
0.016496503725647926,
0.08457614481449127,
0.06003660708665848,
0.1583457589149475,
0.15318629145622253,
-0.052015580236911774,
0.10785988718271255,
0.047366250306367874,
-0.12235783785581589,
-0.15889906883239746,
0.1492328941822052,
-0.06982064247131348,
0.1339222937822342,
-0.02282589301466942,
0.1734936535358429,
0.11680872738361359,
-0.2019118368625641,
0.005020570009946823,
-0.015457574278116226,
-0.08098114281892776,
-0.09372816979885101,
-0.09724362194538116,
-0.09039358049631119,
-0.17153076827526093,
0.015656784176826477,
-0.10397074371576309,
0.00653931824490428,
0.07303358614444733,
0.024682870134711266,
0.022854294627904892,
0.16144578158855438,
0.06075438857078552,
0.02569766715168953,
0.06354031711816788,
0.04933059960603714,
-0.04299988970160484,
-0.029326343908905983,
-0.08425577729940414,
0.019112613052129745,
-0.021780336275696754,
0.03778412193059921,
-0.06515851616859436,
-0.06637177616357803,
0.08749748766422272,
0.04378746077418327,
-0.09936872869729996,
0.022462869063019753,
-0.020785557106137276,
0.04105434939265251,
0.06295428425073624,
0.010827952064573765,
0.008439349941909313,
-0.046072229743003845,
0.20181116461753845,
-0.09381504356861115,
-0.007584931794553995,
-0.11416667699813843,
0.169452503323555,
-0.014617937617003918,
-0.008079899474978447,
0.033542171120643616,
-0.08924201875925064,
-0.0057778810150921345,
0.1526118665933609,
0.15614941716194153,
-0.04361100494861603,
-0.02271106466650963,
0.018802890554070473,
-0.014967059716582298,
-0.04122757539153099,
0.08954808861017227,
0.09307464957237244,
0.053692132234573364,
-0.07041924446821213,
-0.047087863087654114,
-0.04043756425380707,
-0.057600751519203186,
-0.030671115964651108,
0.05657384544610977,
0.036198414862155914,
-0.007856232114136219,
-0.04469958320260048,
0.07532007992267609,
-0.04235990345478058,
-0.11655454337596893,
0.09029693156480789,
-0.1784372180700302,
-0.174116313457489,
-0.03356719762086868,
0.08763529360294342,
0.015613706782460213,
0.046098493039608,
-0.003447588998824358,
-0.021992117166519165,
0.09871625155210495,
-0.0038798744790256023,
-0.08398259431123734,
-0.08548378199338913,
0.04383382946252823,
-0.04949594661593437,
0.2364988476037979,
-0.028420036658644676,
0.006922836881130934,
0.12642517685890198,
0.03592256084084511,
-0.13941173255443573,
0.009902333840727806,
0.07589764893054962,
-0.09859250485897064,
0.046901099383831024,
0.15189321339130402,
-0.026494286954402924,
0.12035191059112549,
0.04131514951586723,
-0.09660965204238892,
-0.009537807665765285,
-0.07724933326244354,
-0.056377798318862915,
-0.0560447983443737,
0.004346002824604511,
-0.037486180663108826,
0.15664106607437134,
0.19945600628852844,
-0.06032969802618027,
-0.03547060117125511,
-0.04595637321472168,
0.0386468879878521,
0.04874350503087044,
0.09521229565143585,
0.0069207618944346905,
-0.2309902012348175,
0.02954651229083538,
-0.030066384002566338,
0.022579103708267212,
-0.19475337862968445,
-0.09133192151784897,
0.016461897641420364,
-0.04952580854296684,
-0.09912458807229996,
0.1050848588347435,
0.07906346023082733,
0.04947519302368164,
-0.06031835824251175,
-0.04223792627453804,
-0.05330721661448479,
0.15738320350646973,
-0.16528093814849854,
-0.07788796722888947
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.