type
stringclasses 1
value | id
stringlengths 5
122
| num_branches
int64 1
1.76k
| branches
sequencelengths 1
1.76k
| main_branch_size
int64 0
32,943B
|
---|---|---|---|---|
model | Niveen/beit-base-patch16-224-pt22k-finetuned-amd_v1 | 1 | [
"main"
] | 1,519 |
model | thanhduc1180/Vistral_Abmusu_400news_adapter | 1 | [
"main"
] | 330,031,271 |
model | AnhDuc2507/model_weight_1 | 1 | [
"main"
] | 377,899,695 |
model | ykuzmenko73/whisper-medium-ru | 1 | [
"main"
] | 3,057,767,394 |
model | Litzy619/Phi0503HMA15 | 1 | [
"main"
] | 8,352,589,468 |
model | optimum-intel/fastrag-retriever | 1 | [
"main"
] | 2,963 |
model | jackswie/r | 1 | [
"main"
] | 115,107,791 |
model | cycy233/misllm1.1 | 1 | [
"main"
] | 14,484,084,900 |
model | rhye/cat_lora | 1 | [
"main"
] | 53,839,837 |
model | optimum-intel/fastrag-ranker | 1 | [
"main"
] | 2,248 |
model | Akish/gpt2-finance-finetuned | 1 | [
"main"
] | 12,810,349 |
model | ptrdung/viT5naive_VLSP2018 | 1 | [
"main"
] | 903,842,023 |
model | virtualemployee/VirtualEmployee | 1 | [
"main"
] | 1,519 |
model | TomEijkelenkamp/sd-onepiece-model-3 | 1 | [
"main"
] | 1,519 |
model | hdanie/Bitnet-Nous-Llama2-70M | 1 | [
"main"
] | 312,239,103 |
model | Gloxiderri/Sovereign_The_Reaper | 1 | [
"main"
] | 63,688,978 |
model | rombodawg/Llama-3-8B-Base-Coder-v3.5-10k | 1 | [
"main"
] | 16,069,791,052 |
model | patriotyk/vocos-mel-hifigan-compat-44100khz | 1 | [
"main"
] | 1,324,613,822 |
model | ada-nai/bert-finetuned-ner | 1 | [
"main"
] | 431,822,747 |
model | Malkith99/Mistral-7b-absa-semeval-2014-restaurants | 1 | [
"main"
] | 14,485,814,022 |
model | RichardErkhov/namespace-Pt_-_activation-beacon-llama2-7b-chat-4bits | 1 | [
"main"
] | 4,170,067,361 |
model | RichardErkhov/lex-hue_-_Delexa-7b-4bits | 1 | [
"main"
] | 4,453,130,218 |
model | srivardhinireddy/mistral_QLORA_model_emotion_detection | 1 | [
"main"
] | 1,342,426,333 |
model | NereuS28/BERT1distilbert-base-uncased-finetuned-cola | 1 | [
"main"
] | 268,805,909 |
model | rnaveensrinivas/Mistral-7B-Instruct-v0.2-GPTQ_retrained_IoV | 1 | [
"main"
] | 10,731,425 |
model | Peacoc/peacoc-4 | 1 | [
"main"
] | 3,295,853,076 |
model | PD98/Chatbot2 | 1 | [
"main"
] | 1,543 |
model | 0x0son0/ft_341 | 1 | [
"main"
] | 6,782,636 |
model | RichardErkhov/namespace-Pt_-_activation-beacon-llama2-7b-chat-8bits | 1 | [
"main"
] | 7,008,759,970 |
model | RichardErkhov/lex-hue_-_Delexa-7b-8bits | 1 | [
"main"
] | 7,512,089,387 |
model | ricar0/imp-v1-3b-196-q4f16_1-mlc | 1 | [
"main"
] | 1,799,483,362 |
model | quangtqv/crypto_Gliner_10_5_2024 | 1 | [
"main"
] | 1,782,045,113 |
model | dapsvi/APKs | 1 | [
"main"
] | 212,850,938 |
model | afrideva/gemma-1.1-2b-it_orpo_traslate_en_es_V1-GGUF | 1 | [
"main"
] | 15,761,381,304 |
model | Qamar110/peft-dialogue-summary-training-1715340654 | 1 | [
"main"
] | 1,519 |
model | fragata/PULI-GPT-2-small-instruct | 1 | [
"main"
] | 447,305,059 |
model | mradermacher/Llama-3-6B-Granite-v0.1-GGUF | 1 | [
"main"
] | 65,223,597,832 |
model | HarshilPatel1905/Emotion_Repo | 1 | [
"main"
] | 1,519 |
model | takumi12/calm-text-e3 | 1 | [
"main"
] | 4,558,636 |
model | RichardErkhov/h2oai_-_h2o-danube2-1.8b-sft-4bits | 1 | [
"main"
] | 1,268,156,147 |
model | RichardErkhov/lex-hue_-_Delexa-7b-gguf | 1 | [
"main"
] | 87,511,277,449 |
model | alijawad07/multilingual-Llama-3-8B-AWQ-GEMVFast-4bit | 1 | [
"main"
] | 5,818,930,207 |
model | vilmay/haskell-4 | 1 | [
"main"
] | 16,801,297 |
model | Onoliya/ol_chiki | 1 | [
"main"
] | 497,782,899 |
model | RichardErkhov/h2oai_-_h2o-danube2-1.8b-sft-8bits | 1 | [
"main"
] | 1,999,714,459 |
model | lielbin/babyberta-Wikipedia_2.5-0.1-finetuned-QAMR | 1 | [
"main"
] | 34,527,270 |
model | Iniquitous1982/test | 1 | [
"main"
] | 1,519 |
model | rajuvk/Image-Text-to-Text | 1 | [
"main"
] | 1,543 |
model | dgdgafrg/videomae-base-finetuned-ucf101-subset | 1 | [
"main"
] | 345,027,531 |
model | NikolayKozloff/Nxcode-CQ-7B-orpo-Q6_K-GGUF | 1 | [
"main"
] | 6,377,591,366 |
model | Litzy619/Phi0503HMA14 | 1 | [
"main"
] | 8,352,589,468 |
model | aloobun/Meta-Llama-3-7B-28Layers | 1 | [
"main"
] | 14,324,816,271 |
model | RichardErkhov/h2oai_-_h2o-danube2-1.8b-sft-gguf | 1 | [
"main"
] | 22,389,031,712 |
model | RakhaMB/DialoGPT-small-Saori | 1 | [
"main"
] | 1,519 |
model | antony-pk/Llama-2-7b-chat-finetune | 1 | [
"main"
] | 13,479,328,627 |
model | Kurkur99/mistral_categorization3_new_jummat | 1 | [
"main"
] | 1,519 |
model | farhananis005/Phi-3-mini-128k-instruct-MoE-v1 | 1 | [
"main"
] | 7,644,536,507 |
model | Qamar110/peft-dialogue-summary-training-1715341443 | 1 | [
"main"
] | 1,519 |
model | muthuramkumar/roberta-base-conversation-classification-2 | 1 | [
"main"
] | 502,017,825 |
model | AlignmentResearch/robust_llm_pythia-410m_niki-044_imdb_random-token-1280_30-rounds_seed-4 | 1 | [
"main"
] | 1,417,757,357 |
model | pEpOo/setfit-model-24-3 | 1 | [
"main"
] | 502,308,625 |
model | Terminator-of-AI/Qwen1.5-0.5B-finetuning-by-super-battleship | 1 | [
"main"
] | 1,861,403,928 |
model | DUAL-GPO-2/zephyr-7b-gpo-v13-i1 | 1 | [
"main"
] | 673,798,774 |
model | mudogruer/mixtral-7x8b-SciQ | 1 | [
"main"
] | 13,656,394 |
model | oleshy/ontochem_biobert_1e-5_10epochs | 1 | [
"main"
] | 431,804,102 |
model | Pragash-Mohanarajah/xlm-roberta-base-finetuned-bible | 1 | [
"main"
] | 1,135,374,310 |
model | jonruida/Spotify | 1 | [
"main"
] | 136,916,008 |
model | lielbin/XLMR-finetuned-SQuAD1 | 1 | [
"main"
] | 1,132,020,019 |
model | RichardErkhov/mindy-labs_-_mindy-7b-v2-4bits | 1 | [
"main"
] | 4,453,124,241 |
model | sanya94/Instance_Segmentation | 1 | [
"main"
] | 351,152,079 |
model | DeepIQ-Inc/drilling_drilling_prompt_set | 1 | [
"main"
] | 1,519 |
model | AlignmentResearch/robust_llm_pythia-14m_niki-045_pm_random-token-1280_seed-3 | 31 | [
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model | akil-17-11/mistral_7b_hf | 1 | [
"main"
] | 1,519 |
model | onsba/my_model | 1 | [
"main"
] | 266,593,713 |
model | am-azadi/EXIST2024_Task1_xlmRoberta_large_3_16_merged_unshuffled | 1 | [
"main"
] | 2,261,775,395 |
model | aloobun/Meta-Llama-3-7B-29Layers | 1 | [
"main"
] | 14,761,042,016 |
model | oleshy/ontochem_biobert_1e-5_10epochs_v2 | 1 | [
"main"
] | 431,804,108 |
model | Aimodelbyharis/scriptgenerationmodel-GGUF | 1 | [
"main"
] | 4,112,182 |
model | LujainAbdulrahman/llama-2-7b-AE-FTuned3 | 1 | [
"main"
] | 4,829,227,871 |
model | am-azadi/EXIST2024_Task2_xlmRoberta_large_3_16 | 1 | [
"main"
] | 2,261,783,660 |
model | kunnisingh/xlsr_hindi_LMless_300m_finetuned | 1 | [
"main"
] | 1,262,368,835 |
model | RichardErkhov/mindy-labs_-_mindy-7b-v2-8bits | 1 | [
"main"
] | 7,512,083,410 |
model | MaziyarPanahi/calme-2.2-phi3-4b | 1 | [
"main"
] | 7,644,710,882 |
model | hanseungwook/vas-llama-2-7b-hh-sft | 1 | [
"main"
] | 26,956,066,278 |
model | oleshy/ontochem_biobert_1e-5_10epochs_v3 | 1 | [
"main"
] | 431,804,108 |
model | samolego/slovene-phi2 | 1 | [
"main"
] | 2,145,499,198 |
model | MaziyarPanahi/calme-2.3-phi3-4b | 1 | [
"main"
] | 7,644,711,167 |
model | fakezeta/Llama3-Aloe-8B-Alpha-ov-int8 | 1 | [
"main"
] | 8,048,877,797 |
model | zhanjun/lora-trained-xl-notion | 1 | [
"main"
] | 1,442,099,823 |
model | newbienewbie/Llama-2-7b-chat-finetune | 1 | [
"main"
] | 13,479,328,407 |
model | StrangeSX/Saraa-8B-LoRA | 1 | [
"main"
] | 176,971,843 |
model | MCZK/ArrowPro-7B-RobinHood-GGUF | 1 | [
"main"
] | 62,986,713,786 |
model | RichardErkhov/mindy-labs_-_mindy-7b-v2-gguf | 1 | [
"main"
] | 87,511,268,096 |
model | M4NT/flower-recognition-classification | 1 | [
"main"
] | 1,520 |
model | abc88767/2c7 | 1 | [
"main"
] | 3,295,853,120 |
model | johnmisquitta00/Intro | 1 | [
"main"
] | 1,550 |
model | lc1290/test | 1 | [
"main"
] | 1,550 |
model | acecalisto3/InstructiPhi | 1 | [
"main"
] | 2,393,237,364 |
model | Pragash-Mohanarajah/bert-base-multilingual-cased-finetuned-bible | 1 | [
"main"
] | 1,519 |
model | sezinarseven/mbti-classification-4 | 1 | [
"main"
] | 741,976,559 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.