text
stringlengths
15
129
model:GetialJuan/bert-base-uncased-finetuned-ner-biobert [42]
model:MayBashendy/Arabic_FineTuningAraBERT_AugV5_k40_task1_organization_fold0 [32]
model:matrixportal/Llama-3.1-8B-Instruct-IQ3_XXS-GGUF [33]
model:tensorblock/illuni-llama-2-ko-7b-GGUF [22]
model:GetialJuan/bert-base-spanish-wwm-cased-finetuned-ner-biobert [42]
model:davidrd123/Mary-Cassatt-Oil-Crops-Phase-2-After-Slow-Beta-2-2-4-SS2_0-1536Flux-LoKr [37]
model:valenmejia/valentinaperez [37]
model:getad72493/assexpose [37]
model:MayBashendy/Arabic_FineTuningAraBERT_AugV5_k40_task1_organization_fold1 [32]
model:koshirowada/pythia_14m_RM [32]
model:getad72493/assupfbh [37]
model:John6666/koronemixillustrious-v30-sdxl [37]
model:John6666/noobmerge-v04-sdxl [37]
model:John6666/pornmaster-anime-noobxlv1-sdxl [37]
model:John6666/solmeleon-illustriousxl-v10-sdxl [37]
model:esslushy/bert-classification [32]
model:tensorblock/SOLAR-10.7B-Instruct-STOCK_SOLAR-ties-GGUF [33]
model:CamiloRojas/camilor [37]
model:nteku1/test [32]
model:wamegabe/llama-3-2-1B-wame-4bit-curi [33]
model:Hanisnabila/Malay-Sentiment-6 [32]
model:Robertsowasp/bert-finetuned-squad [22]
model:MayBashendy/Arabic_FineTuningAraBERT_AugV5_k60_task1_organization_fold0 [32]
model:vonjack/Qwen2.5-Coder-0.5B-Merged [33]
model:abarelka/model_16bit_run6 [33]
model:crestf411/L3.1-8B-Dark-Planet-Slush [33]
model:vonjack/Qwen2.5-Coder-1.5B-Merged [33]
model:osmanh/en-lt-t5-small [40]
model:tensorblock/SOLAR-10.7B-v1.2-GGUF [33]
model:cmaytadatag1/cmaytadatag1-distilroberta-base-mrpc-glue-cmayta [32]
model:Kacharuk/cnn_news_summary_model_trained_on_reduced_data [40]
model:kdrianm/emotion_classification [10]
model:Yin2610/autotrain2 [10]
model:ngmediastudio89/gaby [37]
model:chanchal1987/me-flux-lora [37]
model:c01zaut/gemma-2-9b-it-rk3588-1.1.1 [33]
model:tensorblock/pygmalion-2-7b-GGUF [33]
model:Marty-17/emotion-classification-model [32]
model:HK9872/test_trainer [32]
model:MayBashendy/Arabic_FineTuningAraBERT_AugV5_k60_task1_organization_fold1 [32]
model:win10/falcon-mamba-7b-instruct-Q8_0-GGUF [33]
model:xfuu11ecaf/my_awesome_opus_books_model [40]
model:2121-8/japanese-parler-tts-mini-bate [36, 40, 38]
model:MinaMila/llama_3_instruct_adult_epochs [33]
model:maybethisismary/Letopis [37]
model:eligapris/Qwen2.5-Coder-32B-Instruct-Q2_K-GGUF [33]
model:outlookAi/VWNi1xEAsu [37]
model:tamertheai/tims23 [37]
model:outlookAi/uFPk146rMN [37]
model:eulerevangelista/bert-base-uncased-issues-128 [8]
model:huTG00/lora_model [33]
model:charith78/shihan_test_1 [37]
model:huTG00/Llama-3.1-8B-kowiki-alpaca-4bit [33]
model:mini1013/master_item_ap [32]
model:do1232/xlm-roberta-base-finetuned-panx-de-fr [42]
model:mipat12/st-main-cassatt-oils-phase2-slowbake-highres-nocrops-ss-0_7 [37]
model:mini1013/master_cate_ap0 [32]
model:LucAI12/9kbap [37]
model:do1232/xlm-roberta-base-finetuned-panx-fr [42]
model:qkdaf/xlm-roberta-base-finetuned-panx-de-fr [42]
model:aachalgupta/saffola_oats [37]
model:do1232/xlm-roberta-base-finetuned-panx-it [42]
model:do1232/xlm-roberta-base-finetuned-panx-en [42]
model:mini1013/master_cate_ap1 [32]
model:do1232/xlm-roberta-base-finetuned-panx-all [42]
model:John6666/muonxlmixx-v301il-sdxl [37]
model:yakazimir/llama3instruct_-cpo-10-0_4-9e-7-1_best [33]
model:yakazimir/llama3instruct_-orpo-10-0_5-1e-6-1_best [33]
model:yakazimir/llama3instruct_-qfUNL-10-0_3-1e-6-1_best [33]
model:yakazimir/llama3instruct_-l5-10-0_3-1e-6-2_best [33]
model:qkdaf/xlm-roberta-base-finetuned-panx-fr [42]
model:mini1013/master_cate_ap2 [32]
model:Shuhaib73/LLama_3.1_8B_FineTuned [33]
model:qkdaf/xlm-roberta-base-finetuned-panx-it [42]
model:yakazimir/llama3_qfUNL_best_entropy [33]
model:yakazimir/llama3_l5_best_entropy [33]
model:yakazimir/llama3_cpo_best_entropy [33]
model:yakazimir/llama3_orpo_best_entropy [33]
model:qkdaf/xlm-roberta-base-finetuned-panx-en [42]
model:qkdaf/xlm-roberta-base-finetuned-panx-all [42]
model:GoldenLlama/krx_sg_qwen2.5_7b_it_v5 [33]
model:mini1013/master_cate_ap3 [32]
model:SunTzu77/PYlora [37]
model:vasukumarp/speecht5_tts_tamil [36]
model:real-jiakai/bert-base-uncased-finetuned-squad [22]
model:kakarooky/roberta-base-naver-review-classification [32]
model:yakazimir/qwen_uncCPO_entropy [33]
model:yakazimir/qwen_unl_entropy [33]
model:yakazimir/qwen_ce_entropy [33]
model:yakazimir/qwen_cfUNL_entropy [33]
model:yakazimir/qwen_fUNL_entropy [33]
model:yakazimir/qwen_cUNL_entropy [33]
model:yakazimir/qwen_orpo_entropy [33]
model:yakazimir/qwen_cCPO_entropy [33]
model:yakazimir/qwen_qfUNL_entropy [33]
model:sab-in-scienceland/gpt2-finetuned-wikitext2 [33]
model:suehyunpark/potpourri-8b-inst-fft-induction-bc-trajectory-max10-per-task-tool-token-fix [33]
model:cems-official/panels_detection_rtdetr [21]
model:charisgao/finetuned-roberta-bias-classification [32]
model:wndlek3/xlm-roberta-base-finetuned-panx-de-fr [42]