diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed1/README.md b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed1/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed1/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_config.json b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..4f2872e95fbd512736a55ac435be3f4604c3a96a --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "lm_head", + "v_proj", + "q_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_model.safetensors b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b3d0e8b1a5a71696ee9900bdfaf1fae24c06d49a --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:294dd86276f680c2b80116b0efc4a6e7801581e5318853bf95a85bff6c88f419 +size 1077088768 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed1/log.txt b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..eb89973ec527bdee7e8485ebe2926c37d48c929b --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed1/log.txt @@ -0,0 +1,3 @@ +2024-12-06 12:52:55,895 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8074324131011963, val_ece: 0.05116993561387062, val_nll: 0.4652288854122162, val_brier: 0.25549647212028503 +2024-12-07 02:45:53,800 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8277027010917664, val_ece: 0.05604162439703941, val_nll: 0.5442683100700378, val_brier: 0.26357054710388184 +2025-01-24 12:27:11,349 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:268] - INFO: val_acc: 0.8243243098258972, val_ece: 0.029743729159235954, val_nll: 0.46511635184288025, val_brier: 0.25662732124328613, time: 685.5795176029205 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed2/README.md b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed2/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed2/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_config.json b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..3843c19e946b66379b8b0b485ae63237912dea68 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "lm_head", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_model.safetensors b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..54cf1354c3806c79f10c6eed5828106dc3eb923c --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d70a13979293ef9ec37c35f844d1f290e663e8864e47d66626b5fe320d72732 +size 1077088768 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed2/log.txt b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..dc3f0e0b14092eabe25f4e07342c5fe69e1aafe5 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed2/log.txt @@ -0,0 +1,2 @@ +2024-12-06 12:50:56,517 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8412162065505981, val_ece: 0.0546087920665741, val_nll: 0.4233132600784302, val_brier: 0.22746728360652924 +2024-12-07 02:54:35,176 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8074324131011963, val_ece: 0.06733152270317078, val_nll: 0.5256535410881042, val_brier: 0.2567933201789856 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed3/README.md b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed3/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed3/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_config.json b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..9ec29a0d908cd4be72a55962f40f4d633b638da3 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "v_proj", + "lm_head", + "q_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_model.safetensors b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e23ee286f3221608fd31c0e845f4bc2202669f98 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5adfb353d4f796a6f553dd4c9769d2c9760b6ca5297c0eaf3042e7416424f267 +size 1077088768 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed3/log.txt b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..42325d40eadf31b49687a87987ddfbeb8b486b98 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/blob-ARC-Challenge-sample10-eps0.05-kllr0.0075-beta0.15-seed3/log.txt @@ -0,0 +1 @@ +2024-12-07 03:03:21,631 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8344594836235046, val_ece: 0.03894827514886856, val_nll: 0.4676978290081024, val_brier: 0.24005942046642303 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed1/README.md b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed1/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed1/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_config.json b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..3843c19e946b66379b8b0b485ae63237912dea68 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "lm_head", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_model.safetensors b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4afce998d19b04cee70f6999e1b3ef012f628f8b --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:771c4882c1ab55ecbf57b6d54e1e02e47e8a2ba21dadd1302842e4955ab19fb4 +size 1077088768 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed1/log.txt b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..77522688194c46818a5970e03f6a19b4356dce5a --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed1/log.txt @@ -0,0 +1,2 @@ +2024-12-07 03:19:06,978 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.908450722694397, val_ece: 0.019797053188085556, val_nll: 0.2379530519247055, val_brier: 0.12433462589979172 +2025-01-24 12:54:53,294 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:268] - INFO: val_acc: 0.9313380122184753, val_ece: 0.024036165326833725, val_nll: 0.2061406373977661, val_brier: 0.0999348983168602, time: 1360.183313369751 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed2/README.md b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed2/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed2/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_config.json b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..5e381b71e3419d8943c0849058376362acf12bca --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "v_proj", + "q_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_model.safetensors b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..92a86496b587a2d776a488c479fbeb7bb28bdc51 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68ae1cb92aefc086c1321daf0eab3954b73171308a392f5803cd2de9e6646acc +size 1077088768 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed2/log.txt b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..0903b15eccdf29955d5dbf74e0384817cdd436a5 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed2/log.txt @@ -0,0 +1 @@ +2024-12-07 08:15:51,621 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.9066901206970215, val_ece: 0.03922111541032791, val_nll: 0.2310458868741989, val_brier: 0.1204155907034874 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed3/README.md b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed3/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed3/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_config.json b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..9ec29a0d908cd4be72a55962f40f4d633b638da3 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "v_proj", + "lm_head", + "q_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_model.safetensors b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..02576eb243b42ce2573fef4d7e51357c0c8e1020 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9b11c066df3c658f9df0dd0b3bfd7b244c018641dffaccbd23c2376ff8d05b10 +size 1077088768 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed3/log.txt b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..0787cffc0f13f1d77fec56ae85d94f4b86785e37 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/blob-ARC-Easy-sample10-eps0.05-kllr0.0075-beta0.15-seed3/log.txt @@ -0,0 +1 @@ +2024-12-07 08:44:38,779 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.9190140962600708, val_ece: 0.02209547534584999, val_nll: 0.2080429345369339, val_brier: 0.11617469787597656 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed1/README.md b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed1/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed1/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_config.json b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..9ec29a0d908cd4be72a55962f40f4d633b638da3 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "v_proj", + "lm_head", + "q_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_model.safetensors b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..84dac8e41703d5fbc05653d30f222a38441ededf --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2b281d044b81e3290b1f4f13b92cfc120516f163eae088a8eda5ba08f8d3cd44 +size 1077088768 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed1/log.txt b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..e4fbe30505554e13e30303920c69f5f2a5463ba0 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed1/log.txt @@ -0,0 +1,2 @@ +2024-12-07 07:38:35,003 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8968788385391235, val_ece: 0.028964433819055557, val_nll: 0.2614209055900574, val_brier: 0.15382617712020874 +2025-01-24 17:33:42,396 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:268] - INFO: val_acc: 0.8950428366661072, val_ece: 0.0258608628064394, val_nll: 0.26247158646583557, val_brier: 0.15412576496601105, time: 9015.174057483673 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed2/README.md b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed2/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed2/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_config.json b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..3843c19e946b66379b8b0b485ae63237912dea68 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "lm_head", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_model.safetensors b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..530191f8ae8699df644e5c51133160f194938681 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a22be27bbddf203ca9642a7b75b0fd67b78cec283c4d518ced9d85753eef2a01 +size 1077088768 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed2/log.txt b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..b71af20d2364e0353537b85968fe584f7d39d071 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed2/log.txt @@ -0,0 +1 @@ +2024-12-07 07:13:10,781 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8981028199195862, val_ece: 0.0229333508759737, val_nll: 0.25092485547065735, val_brier: 0.14934252202510834 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed3/README.md b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed3/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed3/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_config.json b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..9ec29a0d908cd4be72a55962f40f4d633b638da3 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "v_proj", + "lm_head", + "q_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_model.safetensors b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..524698519282d32d9a65d2fafaa7e8b0fb614c6b --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:abb9becfd9b45aa52a7c8ccd26ee5524520cccc20e83b1f729f264e47acf9a72 +size 1077088768 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed3/log.txt b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..20e5af353376b261647b7d538a9ca4b75a6857fd --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/boolq/blob-boolq-sample10-eps0.05-kllr0.0075-beta0.15-seed3/log.txt @@ -0,0 +1 @@ +2024-12-07 22:42:12,436 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8944308161735535, val_ece: 0.025517355650663376, val_nll: 0.2617236077785492, val_brier: 0.15499068796634674 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed1/README.md b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed1/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed1/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_config.json b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..4f2872e95fbd512736a55ac435be3f4604c3a96a --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "lm_head", + "v_proj", + "q_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_model.safetensors b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..984d8c3084a4a9ed8ba43f7abd328857b6ad430d --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0f750b3df8118da1a811d75a177c683d630917c44c45121e16de9a8d92f08b0d +size 1077088768 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed1/log.txt b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..d121fd2df0f59fd5bb50179a9dc73825f07f631b --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed1/log.txt @@ -0,0 +1,2 @@ +2024-12-07 03:58:15,132 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8729838728904724, val_ece: 0.040628835558891296, val_nll: 0.3389940559864044, val_brier: 0.17259451746940613 +2025-01-24 14:30:25,362 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:268] - INFO: val_acc: 0.878000020980835, val_ece: 0.028683193027973175, val_nll: 0.3263632655143738, val_brier: 0.17146383225917816, time: 3257.9740467071533 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed2/README.md b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed2/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed2/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_config.json b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..65116e896a3afb6e2418900d9b81a63abdfc26bc --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "lm_head", + "q_proj", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_model.safetensors b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..aec21bcf73c9fd59f10b1f0d0bdc37065dc3fed6 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:de2282ea399fcc20826e5afcebe8e2973bb7982fe6f21e4fab4209745270bec5 +size 1077088768 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed2/log.txt b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..240613a91ee79077c58e875beb7bb3cd32a5f416 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed2/log.txt @@ -0,0 +1 @@ +2024-12-07 12:42:31,595 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.878000020980835, val_ece: 0.028614133596420288, val_nll: 0.35127460956573486, val_brier: 0.18491266667842865 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed3/README.md b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed3/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed3/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_config.json b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..3843c19e946b66379b8b0b485ae63237912dea68 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "lm_head", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_model.safetensors b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7cdef505de24418621dd1634402b7e44ea14320f --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e9af818b6267dc463acb6f278e5163cd2272df566315e7e92206ab8ba4ba57e +size 1077088768 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed3/log.txt b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..dbaff24590deb7bf211e6cec2220f6c97d8befc7 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/obqa/blob-obqa-sample10-eps0.05-kllr0.0075-beta0.15-seed3/log.txt @@ -0,0 +1 @@ +2024-12-07 19:51:47,238 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8759999871253967, val_ece: 0.01807345449924469, val_nll: 0.33220216631889343, val_brier: 0.1741209477186203 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed1/README.md b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed1/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed1/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_config.json b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..9ec29a0d908cd4be72a55962f40f4d633b638da3 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "v_proj", + "lm_head", + "q_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_model.safetensors b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5713de28ef37f5ebb9ec9fb4285c0231b62354f3 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b219a22b78efa221a32817fa6988bfb7cf7a709c4c85a5f9f379816fbb209ed +size 1077088768 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed1/log.txt b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..c427ecf5fceac5e03b6c9bd9941c6dcbeee9f739 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed1/log.txt @@ -0,0 +1,2 @@ +2024-12-06 12:08:18,715 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8188291192054749, val_ece: 0.05160905048251152, val_nll: 0.4427221417427063, val_brier: 0.26996511220932007 +2025-01-24 13:30:25,135 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:268] - INFO: val_acc: 0.8188291192054749, val_ece: 0.05160902813076973, val_nll: 0.4427221417427063, val_brier: 0.26996511220932007, time: 1476.4500977993011 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed2/README.md b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed2/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed2/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_config.json b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..4f2872e95fbd512736a55ac435be3f4604c3a96a --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "lm_head", + "v_proj", + "q_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_model.safetensors b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9f4a466ac3a17a271f8e9e8c72f82bf9e8b6ce19 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae0747607df4fedfb90b87ced64d269fcf4e5511c47839897b9d14d5bc7085d2 +size 1077088768 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed2/log.txt b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..e0d03547ead0e5a84e61309aebeb96597a4da486 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed2/log.txt @@ -0,0 +1 @@ +2024-12-06 12:07:24,003 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.827531635761261, val_ece: 0.03678793087601662, val_nll: 0.4201059937477112, val_brier: 0.2542714774608612 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed3/README.md b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed3/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed3/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_config.json b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..5e381b71e3419d8943c0849058376362acf12bca --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "v_proj", + "q_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_model.safetensors b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f13d51c215be6c96a175e112abf854b2750d96e7 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:73310a557a948111e735db6286772cc4d944c5895ce49f096b5b9429c411741f +size 1077088768 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed3/log.txt b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..90039241f5047ad32ff06626b17c8ba8276679e1 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_m/blob-winogrande_m-sample10-eps0.05-kllr0.0075-beta0.15-seed3/log.txt @@ -0,0 +1 @@ +2024-12-06 12:04:01,497 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8140822649002075, val_ece: 0.039884716272354126, val_nll: 0.4330524504184723, val_brier: 0.26635611057281494 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed1/README.md b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed1/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed1/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_config.json b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..cd6c33663b0838ef49aa4e588cd99ded0e842cb2 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "v_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_model.safetensors b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..756da00e1955064bc6d0b0a0a1b10fd46abcb767 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed1/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2dfa91efe321249120b15bb078665befc0bf7f40917e134db5f32be48661ff00 +size 1077088768 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed1/log.txt b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..f50a97285cc4d4f8ae8872bad8a7504d5c7e8b41 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed1/log.txt @@ -0,0 +1,2 @@ +2024-12-05 13:11:28,788 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.761867105960846, val_ece: 0.09859101474285126, val_nll: 0.5755568742752075, val_brier: 0.34843841195106506 +2025-01-24 12:12:48,120 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:268] - INFO: val_acc: 0.761867105960846, val_ece: 0.09859099984169006, val_nll: 0.5755568742752075, val_brier: 0.34843841195106506, time: 371.84587121009827 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed2/README.md b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed2/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed2/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_config.json b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..9ec29a0d908cd4be72a55962f40f4d633b638da3 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "v_proj", + "lm_head", + "q_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_model.safetensors b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..47966231addc4893b969ae6f9b2e6ae63f6aa596 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed2/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:14bc555058b232b4dc5c6bba474ba29c3dd1cb9c2b75557e558dd90c43dbe808 +size 1077088768 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed2/log.txt b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..8eb6c387ba875942de1ce1ff11a616eccc4321a3 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed2/log.txt @@ -0,0 +1 @@ +2024-12-05 13:10:55,986 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.7697784900665283, val_ece: 0.09702765941619873, val_nll: 0.5830958485603333, val_brier: 0.343814879655838 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed3/README.md b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed3/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed3/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_config.json b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..3843c19e946b66379b8b0b485ae63237912dea68 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "lm_head", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_model.safetensors b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..48d0413fadfee16c798485874b799ea6df70d7a5 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed3/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:79b8bdf87b5866cb86329b32a890268b280bdb2b67a9d78f01c0b9fe2695c499 +size 1077088768 diff --git a/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed3/log.txt b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..2cdf59744d17d30f1bd6b281874ee539f4c98b49 --- /dev/null +++ b/blob/meta-llama/Meta-Llama-3.1-8B/winogrande_s/blob-winogrande_s-sample10-eps0.05-kllr0.0075-beta0.15-seed3/log.txt @@ -0,0 +1 @@ +2024-12-05 13:11:35,187 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.761867105960846, val_ece: 0.10229812562465668, val_nll: 0.5742840766906738, val_brier: 0.3398875296115875 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed1/README.md b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed1/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed1/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed1/adapter_config.json b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed1/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..9ec29a0d908cd4be72a55962f40f4d633b638da3 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed1/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "v_proj", + "lm_head", + "q_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed1/adapter_model.safetensors b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed1/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b19185150579989dbbcdfb82f2b3b2713d3698cd --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed1/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cd5ac33c36119a711e58da495ba675a4ca8b1e7e36a05e5c9328122cc66df9cd +size 1068559504 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed1/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..14a85deb76d9e6f4381cc0a06311975ba7f40ebf --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed1/log.txt @@ -0,0 +1,2 @@ +2024-12-14 15:04:09,337 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 1.0, val_ece: 0.00203323969617486, val_nll: 0.0022251326590776443, val_brier: 0.0005589975626207888 +2024-12-14 22:07:08,671 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.7804054021835327, val_ece: 0.1783970296382904, val_nll: 1.44715416431427, val_brier: 0.37119361758232117 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed2/README.md b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed2/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed2/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed2/adapter_config.json b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed2/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..4f2872e95fbd512736a55ac435be3f4604c3a96a --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed2/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "lm_head", + "v_proj", + "q_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed2/adapter_model.safetensors b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed2/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6c153f76a7d33b53511c7d20ede54de9addadb3d --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed2/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:04dd7de01ca1eac82602d3c97b7490341b88bf8f5bbb3d999142c4e4c81f4847 +size 1068559504 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed2/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..140eddb334c7db08b9d4f9e381883bf5b7713b68 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed2/log.txt @@ -0,0 +1,2 @@ +2024-12-14 15:04:12,212 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 1.0, val_ece: 0.0024595835711807013, val_nll: 0.0027070085052400827, val_brier: 0.0007683130679652095 +2024-12-14 22:07:50,749 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8412162065505981, val_ece: 0.13930147886276245, val_nll: 0.9248173832893372, val_brier: 0.29149678349494934 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed3/README.md b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed3/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed3/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed3/adapter_config.json b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed3/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..cd6c33663b0838ef49aa4e588cd99ded0e842cb2 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed3/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "v_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed3/adapter_model.safetensors b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed3/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c2e9057e95f209dd92443ea9951ff6d2db6a0746 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed3/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c6eb1809f3ed6b83ed7103eeafeb892719a774847b0401262d64c9b58d19e11f +size 1068559504 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed3/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..f0bc091d6cad36b16a59a34af62f7a140c4c41c8 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ARC-Challenge-seed3/log.txt @@ -0,0 +1,2 @@ +2024-12-14 15:04:00,953 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.9991039633750916, val_ece: 0.004154346417635679, val_nll: 0.005711093544960022, val_brier: 0.002725856378674507 +2024-12-14 22:08:27,656 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8108108043670654, val_ece: 0.1553792655467987, val_nll: 0.9894828796386719, val_brier: 0.33308807015419006 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ood-ARC-Challenge-seed1/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ood-ARC-Challenge-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..50d0b5b30e0974b1e0610fa9672716b338ff1639 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ood-ARC-Challenge-seed1/log.txt @@ -0,0 +1 @@ +2024-12-17 09:41:11,440 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8090277910232544, val_ece: 0.15864937007427216, val_nll: 1.1362065076828003, val_brier: 0.34106728434562683 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ood-ARC-Challenge-seed2/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ood-ARC-Challenge-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..91638591b5135d3f8f1691044bd5e3735fb7b18b --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ood-ARC-Challenge-seed2/log.txt @@ -0,0 +1 @@ +2024-12-17 09:41:11,978 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.7881944179534912, val_ece: 0.1693383753299713, val_nll: 1.23374605178833, val_brier: 0.3671594262123108 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ood-ARC-Challenge-seed3/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ood-ARC-Challenge-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..cab33ed2936ae9c77e3348163c7f79cbc3e8066e --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/map-ood-ARC-Challenge-seed3/log.txt @@ -0,0 +1 @@ +2024-12-17 09:41:12,867 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8020833134651184, val_ece: 0.1657986342906952, val_nll: 1.2022100687026978, val_brier: 0.3432585299015045 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed1/README.md b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed1/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed1/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed1/adapter_config.json b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed1/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..65116e896a3afb6e2418900d9b81a63abdfc26bc --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed1/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "lm_head", + "q_proj", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed1/adapter_model.safetensors b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed1/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9ab5ff14d72770404288e4363c53308f1cefe33c --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed1/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f6cc76e9f40872d066730880204ec51802c91bac09db5d8e1767680f80514b4 +size 1068559504 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed1/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..a91d6507aaa9035133ffe84b35a24d29295d7eee --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed1/log.txt @@ -0,0 +1,2 @@ +2024-12-14 15:27:49,374 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 1.0, val_ece: 0.0007599767413921654, val_nll: 0.0007948501734063029, val_brier: 0.00010431231930851936 +2024-12-14 22:09:16,242 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.9207746386528015, val_ece: 0.06141234561800957, val_nll: 0.418718546628952, val_brier: 0.13715893030166626 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed2/README.md b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed2/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed2/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed2/adapter_config.json b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed2/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..65116e896a3afb6e2418900d9b81a63abdfc26bc --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed2/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "lm_head", + "q_proj", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed2/adapter_model.safetensors b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed2/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..17e1c3610141f68c74b11999055a3e4e53255466 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed2/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b34e637ae33feb1d5706d6fcf79c4b8186a926abfcefe5f2f0ad825451dc0088 +size 1068559504 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed2/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..a5c31f75b2d6f79244cf2dd993ce188a8a31451e --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed2/log.txt @@ -0,0 +1,2 @@ +2024-12-14 15:28:09,870 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 1.0, val_ece: 0.0005745090311393142, val_nll: 0.0006096586002968252, val_brier: 9.143204079009593e-05 +2024-12-14 22:10:05,428 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.9102112650871277, val_ece: 0.0753006562590599, val_nll: 0.4827510714530945, val_brier: 0.1565469652414322 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed3/README.md b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed3/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed3/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed3/adapter_config.json b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed3/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..65116e896a3afb6e2418900d9b81a63abdfc26bc --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed3/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "lm_head", + "q_proj", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed3/adapter_model.safetensors b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed3/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d4d60950ffeb6a9f6510486f7210fa73178f9237 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed3/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ad06369468a9759f2de752c65b0b835ff3769903d91b3e03963fa9c30d4cf428 +size 1068559504 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed3/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..9bbbab3b18355a8a61399cfaada2efcbf689f372 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ARC-Easy-seed3/log.txt @@ -0,0 +1,2 @@ +2024-12-14 15:27:53,972 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 1.0, val_ece: 0.000411855784477666, val_nll: 0.0004264691669959575, val_brier: 2.610845876915846e-05 +2024-12-14 22:10:55,079 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.9172534942626953, val_ece: 0.06193051114678383, val_nll: 0.4692331552505493, val_brier: 0.13936258852481842 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ood-ARC-Easy-seed1/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ood-ARC-Easy-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..1dd3123b2fda10699f3f42c2cdfbd700d1d7dd1a --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ood-ARC-Easy-seed1/log.txt @@ -0,0 +1 @@ +2024-12-17 09:41:45,335 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.876838207244873, val_ece: 0.10581789165735245, val_nll: 0.7706588506698608, val_brier: 0.22390379011631012 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ood-ARC-Easy-seed2/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ood-ARC-Easy-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..d560fcb493cc13b1bce9ccecc2152e379c735bcd --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ood-ARC-Easy-seed2/log.txt @@ -0,0 +1 @@ +2024-12-17 09:41:46,610 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8584558963775635, val_ece: 0.11981205642223358, val_nll: 0.7963311076164246, val_brier: 0.256119966506958 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ood-ARC-Easy-seed3/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ood-ARC-Easy-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..64aed58565c4ecde4f27a3b7d3089f2d91d8a79b --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/map-ood-ARC-Easy-seed3/log.txt @@ -0,0 +1 @@ +2024-12-17 09:41:46,816 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8621323704719543, val_ece: 0.1152179166674614, val_nll: 0.9169273972511292, val_brier: 0.2554650604724884 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/MMLU-chem/map-ood-MMLU-chem-seed1/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/MMLU-chem/map-ood-MMLU-chem-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..d5181ffd25ab210b06a4dbdbc2ae9281f919465f --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/MMLU-chem/map-ood-MMLU-chem-seed1/log.txt @@ -0,0 +1 @@ +2024-12-17 09:42:15,208 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.4791666567325592, val_ece: 0.31576216220855713, val_nll: 1.899208903312683, val_brier: 0.7759158611297607 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/MMLU-chem/map-ood-MMLU-chem-seed2/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/MMLU-chem/map-ood-MMLU-chem-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..558c7c8f76f8805ecf5c7f293718f4cdb0b7f403 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/MMLU-chem/map-ood-MMLU-chem-seed2/log.txt @@ -0,0 +1 @@ +2024-12-17 09:42:15,007 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.4583333432674408, val_ece: 0.32721197605133057, val_nll: 1.8577340841293335, val_brier: 0.7985126972198486 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/MMLU-chem/map-ood-MMLU-chem-seed3/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/MMLU-chem/map-ood-MMLU-chem-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..beaa53ed20a43eaec6f4962606df78c30d4cf41f --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/MMLU-chem/map-ood-MMLU-chem-seed3/log.txt @@ -0,0 +1 @@ +2024-12-17 09:42:16,121 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.3645833432674408, val_ece: 0.40059468150138855, val_nll: 2.144956588745117, val_brier: 0.9017971754074097 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/MMLU-phy/map-ood-MMLU-phy-seed1/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/MMLU-phy/map-ood-MMLU-phy-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..a081e6bbc84cb3b4e5e1e854dba99808073db0e1 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/MMLU-phy/map-ood-MMLU-phy-seed1/log.txt @@ -0,0 +1 @@ +2024-12-17 09:42:43,029 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.4270833432674408, val_ece: 0.36374998092651367, val_nll: 2.4146342277526855, val_brier: 0.8907133340835571 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/MMLU-phy/map-ood-MMLU-phy-seed2/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/MMLU-phy/map-ood-MMLU-phy-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..6b3862e7f3088164140c99dfba460a830bb6b101 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/MMLU-phy/map-ood-MMLU-phy-seed2/log.txt @@ -0,0 +1 @@ +2024-12-17 09:42:42,530 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.3854166567325592, val_ece: 0.37612661719322205, val_nll: 2.1860013008117676, val_brier: 0.9116983413696289 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/MMLU-phy/map-ood-MMLU-phy-seed3/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/MMLU-phy/map-ood-MMLU-phy-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..cd9511f259b78664c6b98d1946d59e8a6d3219d5 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/MMLU-phy/map-ood-MMLU-phy-seed3/log.txt @@ -0,0 +1 @@ +2024-12-17 09:42:41,335 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.34375, val_ece: 0.4149787425994873, val_nll: 2.368589162826538, val_brier: 0.9601256251335144 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed1/README.md b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed1/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed1/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed1/adapter_config.json b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed1/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..cd6c33663b0838ef49aa4e588cd99ded0e842cb2 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed1/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "v_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed1/adapter_model.safetensors b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed1/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3225e063d2e27f6f37ee81a2c14651887b25f7d8 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed1/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:edd46a234a27dcb9e2b09e6fd2bd3f4b8e26c7d3f7da21df496285adabd42b79 +size 1068559504 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed1/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..1f006043f488682daf986c79da2f7b8a604fabbc --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed1/log.txt @@ -0,0 +1 @@ +2024-12-14 18:30:14,816 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8968788385391235, val_ece: 0.08753526210784912, val_nll: 0.5325602293014526, val_brier: 0.1863720864057541 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed2/README.md b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed2/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed2/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed2/adapter_config.json b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed2/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..cd6c33663b0838ef49aa4e588cd99ded0e842cb2 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed2/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "v_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed2/adapter_model.safetensors b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed2/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..81d715f15bc0f396eb85e59a1fc6d65ac99c4a7f --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed2/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb1238d6b0e71af20f2b66a8c456258d0a7be4d631bf4e7826919a988d7f8db9 +size 1068559504 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed2/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..4a08193479de42858a986b2ab0af8dd9f104ee50 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed2/log.txt @@ -0,0 +1 @@ +2024-12-14 18:29:37,085 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.9029988050460815, val_ece: 0.08216545730829239, val_nll: 0.5193498134613037, val_brier: 0.17758621275424957 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed3/README.md b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed3/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed3/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed3/adapter_config.json b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed3/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..cd6c33663b0838ef49aa4e588cd99ded0e842cb2 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed3/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "v_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed3/adapter_model.safetensors b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed3/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..17f9965aa58e8d76dcaf0e5a889611efd3a98ac1 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed3/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc35dcc835e0bc1fc7c0f34203c8fda2def00d04e5ab113f0f037389b1a62b34 +size 1068559504 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed3/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..906491fa717f1bb68085e69ae60145df6171fddf --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/boolq/map-boolq-seed3/log.txt @@ -0,0 +1 @@ +2024-12-14 18:30:51,846 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.9026927947998047, val_ece: 0.08229164034128189, val_nll: 0.5037733316421509, val_brier: 0.17754466831684113 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed1/README.md b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed1/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed1/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed1/adapter_config.json b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed1/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..5e381b71e3419d8943c0849058376362acf12bca --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed1/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "v_proj", + "q_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed1/adapter_model.safetensors b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed1/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6855c0d4a7b5fc00cab1855361c709a1461976af --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed1/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d42e35593ea23e96e43bdfea10280b0cf7ac136ddccb4293816d814cd01be1f1 +size 1068559504 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed1/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..d33cc1a0843a1ebddda7dc6d105bfdb8281cb45d --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed1/log.txt @@ -0,0 +1 @@ +2024-12-14 16:37:17,731 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8560000061988831, val_ece: 0.1184162050485611, val_nll: 0.7734560966491699, val_brier: 0.2501131296157837 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed2/README.md b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed2/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed2/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed2/adapter_config.json b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed2/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..3843c19e946b66379b8b0b485ae63237912dea68 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed2/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "lm_head", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed2/adapter_model.safetensors b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed2/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cffc3994b473827d3ad0340805b8d2fb3016197d --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed2/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:79434057c9b47ad428359a2249bf981786745f3719cf7bc06e179d57ab7b32bd +size 1068559504 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed2/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..957ea33bf420a4acd2a53a46dbec801152873192 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed2/log.txt @@ -0,0 +1 @@ +2024-12-14 16:37:06,062 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8600000143051147, val_ece: 0.11762800067663193, val_nll: 0.7909122109413147, val_brier: 0.25003135204315186 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed3/README.md b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed3/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed3/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed3/adapter_config.json b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed3/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..5e381b71e3419d8943c0849058376362acf12bca --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed3/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "v_proj", + "q_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed3/adapter_model.safetensors b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed3/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..17ad474687caa7038835b6b2b7262b5494e6770d --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed3/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e3a22f3bd8543da3e306d5db1321ee1dbdb8882a0d4ee4caac3ef947c8d1864b +size 1068559504 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed3/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..c365cf01056ec31cd6e987f61e41d2d88c6c0a30 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/obqa/map-obqa-seed3/log.txt @@ -0,0 +1 @@ +2024-12-14 16:36:35,188 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8560000061988831, val_ece: 0.12962956726551056, val_nll: 0.8149111866950989, val_brier: 0.2606203556060791 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed1/README.md b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed1/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed1/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed1/adapter_config.json b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed1/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..65116e896a3afb6e2418900d9b81a63abdfc26bc --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed1/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "lm_head", + "q_proj", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed1/adapter_model.safetensors b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed1/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fd6c3e54721d599a6d8339d29e66f17c269bb2fd --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed1/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7e81a184ba683e195946f2917432e1df22bcc21c712de784794320effd206579 +size 1068559504 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed1/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..79658200361ec2ddaf6ebb9e776314f2d4c58282 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed1/log.txt @@ -0,0 +1 @@ +2024-12-14 15:51:22,379 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.827531635761261, val_ece: 0.1323654055595398, val_nll: 0.658226490020752, val_brier: 0.29541531205177307 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed2/README.md b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed2/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed2/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed2/adapter_config.json b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed2/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..5e381b71e3419d8943c0849058376362acf12bca --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed2/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "v_proj", + "q_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed2/adapter_model.safetensors b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed2/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..003ef8966c7724a58026e111ae43cd5a325e9318 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed2/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4bd36f2b04f861f9632d766e8a832c96661223eaeacee07f71ae8043a2314f42 +size 1068559504 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed2/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..5b24547a49723276a96a7622e001ed6c3758ab7f --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed2/log.txt @@ -0,0 +1 @@ +2024-12-14 15:51:44,720 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8283227682113647, val_ece: 0.14087456464767456, val_nll: 0.7319655418395996, val_brier: 0.3087533414363861 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed3/README.md b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed3/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed3/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed3/adapter_config.json b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed3/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..65116e896a3afb6e2418900d9b81a63abdfc26bc --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed3/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "lm_head", + "q_proj", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed3/adapter_model.safetensors b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed3/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f950081fa21c556d2f912dfa6394b556b3b954e5 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed3/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3df3595c203c73d21dab9afa1d89064129e0736c94b1e5801cdcc7624a0528f8 +size 1068559504 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed3/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..f0bf3f18f134c525dd0c48707ca02d7d53a1f74b --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_m/map-winogrande_m-seed3/log.txt @@ -0,0 +1 @@ +2024-12-14 15:51:35,666 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8219936490058899, val_ece: 0.15460968017578125, val_nll: 0.8187390565872192, val_brier: 0.32534125447273254 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed1/README.md b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed1/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed1/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed1/adapter_config.json b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed1/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..9ec29a0d908cd4be72a55962f40f4d633b638da3 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed1/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "v_proj", + "lm_head", + "q_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed1/adapter_model.safetensors b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed1/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d42cde10395aff8a992ebbf450489d6232061cc2 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed1/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e444016675e1711c787b4c4fa2ee33cc1ee63ba02e6928046c71778c1a56e864 +size 1068559504 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed1/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..501d37a7f859a37f08d76128869a8d973eb73f20 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed1/log.txt @@ -0,0 +1,2 @@ +2024-12-14 14:52:23,752 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.995312511920929, val_ece: 0.008510075509548187, val_nll: 0.019496619701385498, val_brier: 0.007887888699769974 +2024-12-14 22:02:56,835 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.7689873576164246, val_ece: 0.17675691843032837, val_nll: 0.9400362372398376, val_brier: 0.3966866731643677 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed2/README.md b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed2/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed2/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed2/adapter_config.json b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed2/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..3843c19e946b66379b8b0b485ae63237912dea68 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed2/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "lm_head", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed2/adapter_model.safetensors b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed2/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cd6b053f5e0e5b9bd95935fa526b73343071266e --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed2/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:843595cf3da2e7d8b9a156850feee371a8b5d87d51b97a13d81107734071f1dc +size 1068559504 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed2/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..b806e94f7843aea4bdd978b6ed04d32eba619c54 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed2/log.txt @@ -0,0 +1,2 @@ +2024-12-14 14:52:22,713 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.995312511920929, val_ece: 0.005974077619612217, val_nll: 0.02973599173128605, val_brier: 0.011292606592178345 +2024-12-14 22:05:08,962 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.7571202516555786, val_ece: 0.19361382722854614, val_nll: 0.957730770111084, val_brier: 0.421448677778244 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed3/README.md b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed3/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed3/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed3/adapter_config.json b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed3/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..3843c19e946b66379b8b0b485ae63237912dea68 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed3/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "lm_head", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed3/adapter_model.safetensors b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed3/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f4b696419d0ee39faa66050e0b1b93abda5dff96 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed3/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e1aba0fa171a438788b701ba3e2c360e404a262f42c27fb5e27a7dce535193ee +size 1068559504 diff --git a/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed3/log.txt b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..b26b53b25d384b63f09136bb4b563c3cbc2b0522 --- /dev/null +++ b/map/meta-llama/Meta-Llama-3.1-8B/winogrande_s/map-winogrande_s-seed3/log.txt @@ -0,0 +1,2 @@ +2024-12-14 14:52:15,068 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.9937499761581421, val_ece: 0.0023702143225818872, val_nll: 0.022181442007422447, val_brier: 0.010977118276059628 +2024-12-14 22:06:28,864 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.7808544039726257, val_ece: 0.1908925324678421, val_nll: 1.0868216753005981, val_brier: 0.400711327791214 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed1/README.md b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed1/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed1/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed1/adapter_config.json b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed1/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..5e381b71e3419d8943c0849058376362acf12bca --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed1/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "v_proj", + "q_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed1/adapter_model.safetensors b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed1/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..97c008186d888761f68bc77b5a899f43da89891e --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed1/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a38d797a33dda29ed7377996f2033ad62c9fdbcb9e504511ebeade575da63852 +size 1068559504 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed1/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..52c87b0bfb9205320b996da4df670f691c8d66cb --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed1/log.txt @@ -0,0 +1,2 @@ +2024-12-05 11:43:08,190 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8074324131011963, val_ece: 0.17170952260494232, val_nll: 1.271486759185791, val_brier: 0.3462175726890564 +2024-12-05 11:59:04,051 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8074324131011963, val_ece: 0.17170952260494232, val_nll: 1.271486759185791, val_brier: 0.3462175726890564 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed2/README.md b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed2/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed2/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed2/adapter_config.json b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed2/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..9ec29a0d908cd4be72a55962f40f4d633b638da3 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed2/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "v_proj", + "lm_head", + "q_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed2/adapter_model.safetensors b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed2/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..505513aa8b30296095c5333989f8f9584d0a4211 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed2/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf3e73b813d6dd71ba6f2d128848df221da636bdd8ea789d24d9de47b5dc2fa5 +size 1068559504 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed2/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..a93601448be385931d5d19232c341dceb16121ee --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed2/log.txt @@ -0,0 +1 @@ +2024-12-05 12:05:19,579 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8074324131011963, val_ece: 0.16361668705940247, val_nll: 1.274039387702942, val_brier: 0.34218987822532654 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed3/README.md b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed3/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed3/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed3/adapter_config.json b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed3/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..65116e896a3afb6e2418900d9b81a63abdfc26bc --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed3/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "lm_head", + "q_proj", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed3/adapter_model.safetensors b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed3/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2db5f4e6ae83295c0f97ae7fd93daa7d50031725 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed3/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a6f69ef8d23cc4145f27ae5424543ffd2fbd8b86d7053070b1a1ccf67afeb243 +size 1068559504 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed3/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..128771afb1ace5eb77cb8a183c18463328b29134 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ARC-Challenge-seed3/log.txt @@ -0,0 +1 @@ +2024-12-05 12:12:24,494 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8175675868988037, val_ece: 0.15508592128753662, val_nll: 1.0410358905792236, val_brier: 0.3130238354206085 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ood-ARC-Challenge-seed1/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ood-ARC-Challenge-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..08a6807eb8262cea86c01971e082addc922414da --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ood-ARC-Challenge-seed1/log.txt @@ -0,0 +1,2 @@ +2024-12-07 21:02:44,122 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8090277910232544, val_ece: 0.17311592400074005, val_nll: 1.3518962860107422, val_brier: 0.36044809222221375 +2024-12-07 21:09:50,006 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8263888955116272, val_ece: 0.13394713401794434, val_nll: 1.0342274904251099, val_brier: 0.3036152422428131 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ood-ARC-Challenge-seed2/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ood-ARC-Challenge-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..e192b47c24905e291cc54fdd927273327c877ed9 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ood-ARC-Challenge-seed2/log.txt @@ -0,0 +1,2 @@ +2024-12-07 21:03:11,333 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8194444179534912, val_ece: 0.15276409685611725, val_nll: 1.2564178705215454, val_brier: 0.3297855257987976 +2024-12-07 21:10:22,363 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8368055820465088, val_ece: 0.12457814812660217, val_nll: 1.0218181610107422, val_brier: 0.2817496359348297 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ood-ARC-Challenge-seed3/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ood-ARC-Challenge-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..c6bbd4f4a06eb94a541fa287979d312c8e832909 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Challenge/mle-ood-ARC-Challenge-seed3/log.txt @@ -0,0 +1,2 @@ +2024-12-07 21:03:38,017 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8333333134651184, val_ece: 0.14142078161239624, val_nll: 1.0520117282867432, val_brier: 0.29338598251342773 +2024-12-07 21:10:52,911 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.78125, val_ece: 0.1750248819589615, val_nll: 1.1634546518325806, val_brier: 0.367326557636261 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed1/README.md b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed1/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed1/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed1/adapter_config.json b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed1/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..5e381b71e3419d8943c0849058376362acf12bca --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed1/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "v_proj", + "q_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed1/adapter_model.safetensors b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed1/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..71a9872a4b82faf3859a43d45b5794f27bf6484c --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed1/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:facd5b7d2b4c52bf4c2344d016ff21255e1f6ca146ee27de16f01a8d0c8646c9 +size 1068559504 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed1/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..77d3b43323a9e88c63d96f461f294f61707ab3e6 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed1/log.txt @@ -0,0 +1 @@ +2024-12-05 12:24:02,005 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.9119718074798584, val_ece: 0.07547640800476074, val_nll: 0.5243775248527527, val_brier: 0.16016806662082672 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed2/README.md b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed2/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed2/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed2/adapter_config.json b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed2/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..5e381b71e3419d8943c0849058376362acf12bca --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed2/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "v_proj", + "q_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed2/adapter_model.safetensors b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed2/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..63fc854ae662e5bab819d63f7d6213347b2ce098 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed2/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b2f62e3178e6109f9d14a563209f735af076d4d0ffcdcd097b69c096e40ace11 +size 1068559504 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed2/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..2f74bfc912de9aa8664757dbebda0d33c0461ffa --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed2/log.txt @@ -0,0 +1 @@ +2024-12-05 12:36:23,747 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.9172534942626953, val_ece: 0.07168968766927719, val_nll: 0.4214722216129303, val_brier: 0.14623402059078217 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed3/README.md b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed3/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed3/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed3/adapter_config.json b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed3/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..4f2872e95fbd512736a55ac435be3f4604c3a96a --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed3/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "lm_head", + "v_proj", + "q_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed3/adapter_model.safetensors b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed3/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..07c7c3df7622805e3cf65b53beedc7a80323d180 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed3/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c0fadb4171bcaaac4f5ecca2b7c9002ff9edc82c7f27bfd3078e34dc65464ac6 +size 1068559504 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed3/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..656abba10f8caf8d1b4fdd7b8d6dac8cda633595 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ARC-Easy-seed3/log.txt @@ -0,0 +1 @@ +2024-12-05 12:49:10,187 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.9207746386528015, val_ece: 0.06293165683746338, val_nll: 0.44233497977256775, val_brier: 0.1450883150100708 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ood-ARC-Easy-seed1/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ood-ARC-Easy-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..a0ad7130abe764cb7242da7e26973b270cfac090 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ood-ARC-Easy-seed1/log.txt @@ -0,0 +1,2 @@ +2024-12-07 21:04:09,437 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.9025735259056091, val_ece: 0.08729612827301025, val_nll: 0.547069787979126, val_brier: 0.17183753848075867 +2024-12-07 21:11:23,567 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8621323704719543, val_ece: 0.11024198681116104, val_nll: 0.7480144500732422, val_brier: 0.24288974702358246 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ood-ARC-Easy-seed2/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ood-ARC-Easy-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..7592f3ca279a81df5b2892f1a2e204a7742b5e6f --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ood-ARC-Easy-seed2/log.txt @@ -0,0 +1,2 @@ +2024-12-07 21:05:22,337 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.9246323704719543, val_ece: 0.06578487902879715, val_nll: 0.4203348457813263, val_brier: 0.14054684340953827 +2024-12-07 21:11:52,582 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8805146813392639, val_ece: 0.10087592154741287, val_nll: 0.6398301720619202, val_brier: 0.20662707090377808 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ood-ARC-Easy-seed3/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ood-ARC-Easy-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..d6454e54c3d4341e1e3a137977fb4a9d0c47811d --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/ARC-Easy/mle-ood-ARC-Easy-seed3/log.txt @@ -0,0 +1,2 @@ +2024-12-07 21:05:58,773 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.9172794222831726, val_ece: 0.0684446468949318, val_nll: 0.4791075885295868, val_brier: 0.1531638503074646 +2024-12-07 21:12:23,092 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8621323704719543, val_ece: 0.11229818314313889, val_nll: 0.7657393217086792, val_brier: 0.24372056126594543 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/MMLU-chem/mle-ood-MMLU-chem-seed1/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/MMLU-chem/mle-ood-MMLU-chem-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..4545f4ea343f3c967cc863f026b6ec447146a01c --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/MMLU-chem/mle-ood-MMLU-chem-seed1/log.txt @@ -0,0 +1 @@ +2024-12-07 21:12:47,516 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.4479166567325592, val_ece: 0.30881139636039734, val_nll: 1.7797354459762573, val_brier: 0.7533901929855347 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/MMLU-chem/mle-ood-MMLU-chem-seed2/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/MMLU-chem/mle-ood-MMLU-chem-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..4cef7adec8ea4ac7211455116f36c3576103ce6a --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/MMLU-chem/mle-ood-MMLU-chem-seed2/log.txt @@ -0,0 +1 @@ +2024-12-07 21:13:10,226 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.46875, val_ece: 0.3037206530570984, val_nll: 1.802008867263794, val_brier: 0.7715045809745789 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/MMLU-chem/mle-ood-MMLU-chem-seed3/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/MMLU-chem/mle-ood-MMLU-chem-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..7ce675baa33d7500b7deca481cd3f521114f75d5 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/MMLU-chem/mle-ood-MMLU-chem-seed3/log.txt @@ -0,0 +1 @@ +2024-12-07 21:13:31,947 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.4583333432674408, val_ece: 0.3612423837184906, val_nll: 2.139535427093506, val_brier: 0.796871542930603 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/MMLU-phy/mle-ood-MMLU-phy-seed1/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/MMLU-phy/mle-ood-MMLU-phy-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..bb2676d00ee7b7b2d07f00c143e5fd347f67efa3 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/MMLU-phy/mle-ood-MMLU-phy-seed1/log.txt @@ -0,0 +1 @@ +2024-12-07 21:13:54,191 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.4479166567325592, val_ece: 0.34233608841896057, val_nll: 1.9788306951522827, val_brier: 0.8110661506652832 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/MMLU-phy/mle-ood-MMLU-phy-seed2/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/MMLU-phy/mle-ood-MMLU-phy-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..c429fa2cce3a0e5e78dd02ca7d5ad39d3261df32 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/MMLU-phy/mle-ood-MMLU-phy-seed2/log.txt @@ -0,0 +1 @@ +2024-12-07 21:14:16,408 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.4166666567325592, val_ece: 0.36432090401649475, val_nll: 2.2815535068511963, val_brier: 0.882830023765564 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/MMLU-phy/mle-ood-MMLU-phy-seed3/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/MMLU-phy/mle-ood-MMLU-phy-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..4e3b6f58435008f2d1dbb6d49ab3b331dc1b6607 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/MMLU-phy/mle-ood-MMLU-phy-seed3/log.txt @@ -0,0 +1 @@ +2024-12-07 21:14:37,065 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.40625, val_ece: 0.44562259316444397, val_nll: 2.4975829124450684, val_brier: 0.9568134546279907 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed1/README.md b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed1/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed1/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed1/adapter_config.json b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed1/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..65116e896a3afb6e2418900d9b81a63abdfc26bc --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed1/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "lm_head", + "q_proj", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed1/adapter_model.safetensors b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed1/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8eca344f1ce478b2bcda75b40bc5ccadd91fb5ff --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed1/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5e2c170f629834ed7c77e542dbb39eeacdbe02d26f27b5e7b8d15fb60d4eff6e +size 1068559504 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed1/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..da34a35e0ddb266ffa7b8362cdc747c89d074281 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed1/log.txt @@ -0,0 +1 @@ +2024-12-05 15:40:32,686 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8995097875595093, val_ece: 0.08392772078514099, val_nll: 0.5090630054473877, val_brier: 0.18148060142993927 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed2/README.md b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed2/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed2/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed2/adapter_config.json b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed2/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..4f2872e95fbd512736a55ac435be3f4604c3a96a --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed2/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "lm_head", + "v_proj", + "q_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed2/adapter_model.safetensors b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed2/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a0f05c4fe0d4e16ab9ba66a986120cc27308438b --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed2/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:37fa1fc5f25e59115b17c4e574101d9841e32ebe5d6c605c68aec66c53b21fbe +size 1068559504 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed2/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..a0ca0a8a1971f022841049685ce82771abf84a36 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed2/log.txt @@ -0,0 +1 @@ +2024-12-05 16:40:13,716 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8939951062202454, val_ece: 0.08866064995527267, val_nll: 0.5245975255966187, val_brier: 0.19038984179496765 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed3/README.md b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed3/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed3/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed3/adapter_config.json b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed3/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..3843c19e946b66379b8b0b485ae63237912dea68 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed3/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "lm_head", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed3/adapter_model.safetensors b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed3/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..16f653a04e61aa89ebcefa01a9a030ea42d48e37 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed3/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6083ee2402e190e5d3a048047002d240a06f18a56ef67cbc2a3f213b4d51bd3d +size 1068559504 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed3/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..0867cc28f2dc5b4613bd85b8d6103e3ce44f6919 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/boolq/mle-boolq-seed3/log.txt @@ -0,0 +1 @@ +2024-12-05 17:39:29,380 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8939951062202454, val_ece: 0.08807679265737534, val_nll: 0.5311131477355957, val_brier: 0.18754567205905914 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed1/README.md b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed1/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed1/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed1/adapter_config.json b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed1/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..65116e896a3afb6e2418900d9b81a63abdfc26bc --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed1/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "lm_head", + "q_proj", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed1/adapter_model.safetensors b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed1/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c178bb8d661608af6c1ec0b16710c3328ef5df8e --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed1/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:de18fe0526a718efe4fcb95e31993b286f481587e4bdeb027b9aa215760eefa4 +size 1068559504 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed1/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..067f4147b5afde9ca06e6d1a67044e33f9019cbb --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed1/log.txt @@ -0,0 +1 @@ +2024-12-05 13:51:46,658 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8911290168762207, val_ece: 0.08631911873817444, val_nll: 0.5434727668762207, val_brier: 0.18999190628528595 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed2/README.md b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed2/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed2/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed2/adapter_config.json b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed2/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..5e381b71e3419d8943c0849058376362acf12bca --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed2/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "v_proj", + "q_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed2/adapter_model.safetensors b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed2/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..55cac884fbf77ff602c0d6ab5d7d1175199da716 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed2/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:edd5950b78892f7b20fca23ac72935923b38f23d170c805efc755616e1dc939a +size 1068559504 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed2/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..c35da0ed2f8807acaddd7b35ccfbd94f93389b15 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed2/log.txt @@ -0,0 +1 @@ +2024-12-05 14:15:37,205 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8709677457809448, val_ece: 0.10438328236341476, val_nll: 0.6008764505386353, val_brier: 0.21831800043582916 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed3/README.md b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed3/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed3/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed3/adapter_config.json b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed3/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..5e381b71e3419d8943c0849058376362acf12bca --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed3/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "v_proj", + "q_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed3/adapter_model.safetensors b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed3/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..07be970ed181b7e1e0a228116c048ac4f7fa0e97 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed3/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aac794c157de2e796fda687e439d44924b52e0e5000ad9805b3324faa426671f +size 1068559504 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed3/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..4daed84abeb3b4c7cec880caa23979447890190b --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/obqa/mle-obqa-seed3/log.txt @@ -0,0 +1 @@ +2024-12-05 14:40:35,314 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.875, val_ece: 0.10245662182569504, val_nll: 0.6832837462425232, val_brier: 0.2182440608739853 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed1/README.md b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed1/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed1/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed1/adapter_config.json b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed1/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..cd6c33663b0838ef49aa4e588cd99ded0e842cb2 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed1/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "v_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed1/adapter_model.safetensors b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed1/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c1cd5f3682c8b1c287084426d656c5e269b1ed08 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed1/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8ec6b6014403e4b90ad60ec079c9678098338417351999d13955b66088345d0f +size 1068559504 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed1/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..4aecd531b4871833b06bd0084acf4a0870ab8df4 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed1/log.txt @@ -0,0 +1 @@ +2024-12-05 13:01:56,905 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.8259493708610535, val_ece: 0.13681189715862274, val_nll: 0.6712241768836975, val_brier: 0.3030058741569519 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed2/README.md b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed2/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed2/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed2/adapter_config.json b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed2/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..5e381b71e3419d8943c0849058376362acf12bca --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed2/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "v_proj", + "q_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed2/adapter_model.safetensors b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed2/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bf09aa5e1f63494e47dfc1d4d4917a1da9039fa2 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed2/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71c455cc1f28a23fcf9c451e280eb80161624a9eda7e2826807259105e49e873 +size 1068559504 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed2/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..515c74b328d16d36f1dd198fd50b61049554f938 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed2/log.txt @@ -0,0 +1 @@ +2024-12-05 13:14:34,143 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.815664529800415, val_ece: 0.14684313535690308, val_nll: 0.705989420413971, val_brier: 0.31897345185279846 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed3/README.md b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed3/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed3/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed3/adapter_config.json b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed3/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..cd6c33663b0838ef49aa4e588cd99ded0e842cb2 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed3/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "v_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed3/adapter_model.safetensors b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed3/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..48c413ca51cfcd52ab7b467cd1f22509fe05ccfe --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed3/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:76a656251214dfbec09dd65241efb9d4d856f16e6d940f1ef742c181a748bccb +size 1068559504 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed3/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..56dab3c36d3fb79e6d40fe468d079c3301578fcc --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_m/mle-winogrande_m-seed3/log.txt @@ -0,0 +1 @@ +2024-12-05 13:27:59,570 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.827531635761261, val_ece: 0.13118334114551544, val_nll: 0.648461103439331, val_brier: 0.29343584179878235 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed1/README.md b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed1/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed1/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed1/adapter_config.json b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed1/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..cd6c33663b0838ef49aa4e588cd99ded0e842cb2 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed1/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "v_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed1/adapter_model.safetensors b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed1/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f3ff12b8bf14852b686e870774915ebf029c96ca --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed1/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f3964847c9edac4c29ee1c288b5c6753e6f511d30f3b2678f538c4496698a0e9 +size 1068559504 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed1/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed1/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..e80bc5a251d84e273ee1255b2aa13e23cb208cca --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed1/log.txt @@ -0,0 +1 @@ +2024-12-05 11:27:47,105 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.7713607549667358, val_ece: 0.175684854388237, val_nll: 0.9175033569335938, val_brier: 0.3921114206314087 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed2/README.md b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed2/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed2/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed2/adapter_config.json b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed2/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..3843c19e946b66379b8b0b485ae63237912dea68 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed2/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "lm_head", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed2/adapter_model.safetensors b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed2/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8ead342b05bd8d0a55c4c9e184aca24a1ae3b866 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed2/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:547923201b48e924e82296283e47a3534ee949ff4c5abc1d1710b1b117ceb102 +size 1068559504 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed2/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed2/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..ebc086032928142e7004ae531b997a1b6d7ee10b --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed2/log.txt @@ -0,0 +1 @@ +2024-12-05 11:31:56,216 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.7808544039726257, val_ece: 0.17064127326011658, val_nll: 0.9140081405639648, val_brier: 0.3841215670108795 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed3/README.md b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed3/README.md new file mode 100644 index 0000000000000000000000000000000000000000..65ed539d6149793da1dfe58d2057481617bbd9ee --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed3/README.md @@ -0,0 +1,202 @@ +--- +base_model: /data/local/public_llms/llamas/Meta-Llama-3.1-8B +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.0 \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed3/adapter_config.json b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed3/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..cd6c33663b0838ef49aa4e588cd99ded0e842cb2 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed3/adapter_config.json @@ -0,0 +1,30 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/data/local/public_llms/llamas/Meta-Llama-3.1-8B", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "v_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed3/adapter_model.safetensors b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed3/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..70caf4dfad9b4b64175ff49a77621d13285ce693 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed3/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:acdc246de3b5941b98acfd77009f5987723e1bc0af90279af2cc308034b2a3f1 +size 1068559504 diff --git a/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed3/log.txt b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed3/log.txt new file mode 100644 index 0000000000000000000000000000000000000000..5f27180e01fe2d571b5c4601b875f9dd183a0a26 --- /dev/null +++ b/mle/meta-llama/Meta-Llama-3.1-8B/winogrande_s/mle-winogrande_s-seed3/log.txt @@ -0,0 +1 @@ +2024-12-05 11:35:58,939 - /common/home/yw1131/repos/bayesian-peft-dev/modelwrappers/wrapperbase.py[line:264] - INFO: val_acc: 0.7840189933776855, val_ece: 0.1643400490283966, val_nll: 0.822324275970459, val_brier: 0.3753136396408081