index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
0 | HuggingFaceTB/SmolLM3-3B-Base | smollm3 | ["SmolLM3ForCausalLM"] | 128256 | bfloat16 | 4.53.0.dev0 | 2,048 | 11,008 | 36 | 16 | 4 | silu | 0 | True | 65,536 | 5,000,000 | 0.000001 | 0.02 | null | 128001 | true | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
1 | ChatDOC/OCRFlux-3B | qwen2_5_vl | ["Qwen2_5_VLForConditionalGeneration"] | 151680 | bfloat16 | 4.49.0 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | False | 128,000 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
2 | jinaai/jina-embeddings-v4 | null | ["JinaEmbeddingsV4Model"] | null | bfloat16 | 4.52.0 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 128,000 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
3 | HuggingFaceTB/SmolLM3-3B | smollm3 | ["SmolLM3ForCausalLM"] | 128256 | bfloat16 | 4.54.0.dev0 | 2,048 | 11,008 | 36 | 16 | 4 | silu | 0 | False | 65,536 | 5,000,000 | 0.000001 | 0.02 | 128000 | 128012 | true | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
4 | KurmaAI/AQUA-7B | mistral | ["MistralForCausalLM"] | 32768 | bfloat16 | 4.52.4 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
5 | LiquidAI/LFM2-1.2B | lfm2 | ["Lfm2ForCausalLM"] | 65536 | bfloat16 | 4.54.0.dev0 | 2,048 | null | 16 | 32 | 8 | null | null | True | 128,000 | 1,000,000 | null | 0.02 | 1 | 7 | null | 64 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.805306 | small | very_long |
6 | LiquidAI/LFM2-350M | lfm2 | ["Lfm2ForCausalLM"] | 65536 | bfloat16 | 4.54.0.dev0 | 1,024 | null | 16 | 16 | 8 | null | null | True | 128,000 | 1,000,000 | null | 0.02 | 1 | 7 | null | 64 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.201327 | small | very_long |
7 | mistralai/Devstral-Small-2507 | mistral | ["MistralForCausalLM"] | 131072 | bfloat16 | 4.53.1 | 5,120 | 32,768 | 40 | 32 | 8 | silu | 0 | True | 131,072 | 1,000,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
8 | kyutai/tts-1.6b-en_fr | tts | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
9 | apple/DiffuCoder-7B-cpGRPO | Dream | ["DreamModel"] | 152064 | bfloat16 | 4.52.0.dev0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 131,072 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
10 | microsoft/Phi-4-mini-flash-reasoning | phi4flash | ["Phi4FlashForCausalLM"] | 200064 | bfloat16 | 4.46.1 | 2,560 | 10,240 | 32 | 40 | 20 | silu | 0 | True | 262,144 | null | null | 0.02 | 199999 | 199999 | true | 64 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | false | 2.516582 | medium | very_long |
11 | agentica-org/DeepSWE-Preview | qwen3 | ["Qwen3ForCausalLM"] | 151936 | float32 | 4.51.3 | 5,120 | 25,600 | 64 | 64 | 8 | silu | 0 | True | 40,960 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | false | 80 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
12 | moonshotai/Kimi-K2-Instruct | kimi_k2 | ["DeepseekV3ForCausalLM"] | 163840 | bfloat16 | 4.48.3 | 7,168 | 18,432 | 61 | 64 | 64 | silu | 0 | True | 131,072 | 50,000 | 0.000001 | 0.02 | 163584 | 163585 | false | 112 | 1 | false | 384 | 8 | false | false | false | false | false | true | false | true | false | true | 37.610324 | large | very_long |
13 | nanonets/Nanonets-OCR-s | qwen2_5_vl | ["Qwen2_5_VLForConditionalGeneration"] | 151936 | bfloat16 | 4.52.4 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 128,000 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | null | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
14 | tencent/Hunyuan-A13B-Instruct | hunyuan_v1_moe | ["HunYuanMoEV1ForCausalLM"] | 128167 | bfloat16 | 4.41.2 | 4,096 | 3,072 | 32 | 32 | 8 | silu | 0.1 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 127960 | true | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
15 | openai/whisper-large-v3 | whisper | ["WhisperForConditionalGeneration"] | 51866 | float16 | 4.36.0.dev0 | null | null | 32 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
16 | K-intelligence/Midm-2.0-Base-Instruct | llama | ["LlamaForCausalLM"] | 131384 | bfloat16 | 4.51.3 | 4,096 | 14,336 | 48 | 32 | 8 | silu | 0 | True | 32,768 | 8,000,000 | 0.00001 | 0.02 | 0 | 2 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 9.663676 | large | very_long |
17 | Qwen/Qwen2.5-VL-7B-Instruct | qwen2_5_vl | ["Qwen2_5_VLForConditionalGeneration"] | 152064 | bfloat16 | 4.41.2 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 128,000 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
18 | LiquidAI/LFM2-700M | lfm2 | ["Lfm2ForCausalLM"] | 65536 | bfloat16 | 4.54.0.dev0 | 1,536 | null | 16 | 24 | 8 | null | null | True | 128,000 | 1,000,000 | null | 0.02 | 1 | 7 | null | 64 | 3 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.452985 | small | very_long |
19 | sentence-transformers/all-MiniLM-L6-v2 | bert | ["BertModel"] | 30522 | null | 4.8.2 | 384 | 1,536 | 6 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.010617 | small | short |
20 | tngtech/DeepSeek-TNG-R1T2-Chimera | deepseek_v3 | ["DeepseekV3ForCausalLM"] | 129280 | bfloat16 | 4.46.3 | 7,168 | 18,432 | 61 | 128 | 128 | silu | 0 | True | 163,840 | 10,000 | 0.000001 | 0.02 | 0 | 1 | false | 56 | 1 | false | 256 | 8 | false | false | false | false | false | true | false | true | false | true | 37.610324 | large | very_long |
21 | MiniMaxAI/MiniMax-M1-80k | minimax_m1 | ["MiniMaxM1ForCausalLM"] | 200064 | null | 4.45.2 | 6,144 | 9,216 | 80 | 64 | 8 | silu | 0 | True | 10,240,000 | 10,000,000 | 0.00001 | 0.02 | null | null | false | 96 | 8 | false | 32 | 2 | false | false | false | false | false | false | false | true | true | true | 36.238787 | large | very_long |
22 | unsloth/DeepSeek-R1-0528-Qwen3-8B-GGUF | qwen3 | ["Qwen3ForCausalLM"] | 151936 | bfloat16 | 4.52.4 | 4,096 | 12,288 | 36 | 32 | 8 | silu | 0 | True | 131,072 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 7.247757 | large | very_long |
23 | deepseek-ai/DeepSeek-R1-0528-Qwen3-8B | qwen3 | ["Qwen3ForCausalLM"] | 151936 | bfloat16 | 4.51.0 | 4,096 | 12,288 | 36 | 32 | 8 | silu | 0 | True | 131,072 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 7.247757 | large | very_long |
24 | mistralai/Magistral-Small-2506 | mistral | ["MistralForCausalLM"] | 131072 | bfloat16 | 4.52.4 | 5,120 | 32,768 | 40 | 32 | 8 | silu | 0 | True | 40,960 | 1,000,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
25 | ByteDance-Seed/Tar-7B | qwen2 | ["LlavaQwenForCausalLM"] | 217207 | bfloat16 | 4.50.0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
26 | THUDM/GLM-4.1V-9B-Thinking | glm4v | ["Glm4vForConditionalGeneration"] | null | null | 4.57.1 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
27 | microsoft/NextCoder-7B | qwen2 | ["Qwen2ForCausalLM"] | 152064 | float32 | 4.46.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
28 | unsloth/Devstral-Small-2507-GGUF | mistral3 | ["Mistral3ForConditionalGeneration"] | null | bfloat16 | 4.52.0.dev0 | null | null | null | null | null | null | null | null | null | null | null | null | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | true | false | false | true | null | null | null |
29 | mistralai/Mistral-Small-3.2-24B-Instruct-2506 | mistral3 | ["Mistral3ForConditionalGeneration"] | null | bfloat16 | 4.52.4 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | true | false | false | true | null | null | null |
30 | deepseek-ai/DeepSeek-R1 | deepseek_v3 | ["DeepseekV3ForCausalLM"] | 129280 | bfloat16 | 4.46.3 | 7,168 | 18,432 | 61 | 128 | 128 | silu | 0 | True | 163,840 | 10,000 | 0.000001 | 0.02 | 0 | 1 | false | 56 | 1 | false | 256 | 8 | false | false | false | false | false | true | false | true | false | true | 37.610324 | large | very_long |
31 | deepseek-ai/DeepSeek-R1-0528 | deepseek_v3 | ["DeepseekV3ForCausalLM"] | 129280 | bfloat16 | 4.46.3 | 7,168 | 18,432 | 61 | 128 | 128 | silu | 0 | True | 163,840 | 10,000 | 0.000001 | 0.02 | 0 | 1 | false | 56 | 1 | false | 256 | 8 | false | false | false | false | false | true | false | true | false | true | 37.610324 | large | very_long |
32 | RekaAI/reka-flash-3.1 | llama | ["LlamaForCausalLM"] | 100352 | bfloat16 | 4.50.3 | 6,144 | 19,648 | 44 | 64 | 8 | silu | 0 | True | 98,304 | 8,000,000 | 0.00001 | 0.006 | 100257 | 100257 | false | 96 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 19.931333 | large | very_long |
33 | skt/A.X-4.0 | qwen2 | ["Qwen2ForCausalLM"] | 102400 | bfloat16 | 4.46.0 | 8,192 | 29,568 | 80 | 64 | 8 | silu | 0 | False | 131,072 | 5,000,000 | 0.00001 | 0.02 | 0 | 0 | false | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
34 | Qwen/Qwen3-0.6B | qwen3 | ["Qwen3ForCausalLM"] | 151936 | bfloat16 | 4.51.0 | 1,024 | 3,072 | 28 | 16 | 8 | silu | 0 | True | 40,960 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.352322 | small | very_long |
35 | mistralai/Devstral-Small-2505 | mistral | ["MistralForCausalLM"] | 131072 | bfloat16 | 4.51.3 | 5,120 | 32,768 | 40 | 32 | 8 | silu | 0 | True | 131,072 | 1,000,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
36 | kyutai/stt-2.6b-en | stt | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
37 | mistralai/Mistral-7B-Instruct-v0.3 | mistral | ["MistralForCausalLM"] | 32768 | bfloat16 | 4.42.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
38 | baidu/ERNIE-4.5-21B-A3B-PT | ernie4_5_moe | ["Ernie4_5_MoeForCausalLM"] | 103424 | bfloat16 | 4.54.0.dev0 | 2,560 | 12,288 | 28 | 20 | 4 | silu | null | True | 131,072 | 500,000 | 0.00001 | 0.02 | 1 | 2 | true | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 2.20201 | medium | very_long |
39 | openai/whisper-large-v3-turbo | whisper | ["WhisperForConditionalGeneration"] | 51866 | float16 | 4.46.0.dev0 | null | null | 32 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
40 | katanemo/Arch-Router-1.5B | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float16 | 4.51.3 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
41 | Menlo/Jan-nano | qwen3 | ["Qwen3ForCausalLM"] | 151936 | bfloat16 | 4.52.4 | 2,560 | 9,728 | 36 | 32 | 8 | silu | 0 | True | 40,960 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 80 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 2.831155 | medium | very_long |
42 | Qwen/Qwen3-Embedding-0.6B | qwen3 | ["Qwen3ForCausalLM"] | 151669 | bfloat16 | 4.51.3 | 1,024 | 3,072 | 28 | 16 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | true | 64 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.352322 | small | very_long |
43 | SicariusSicariiStuff/Impish_LLAMA_4B | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.51.3 | 3,072 | 9,216 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128019 | false | 96 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.623879 | medium | very_long |
44 | zeroentropy/zerank-1-small | qwen3 | ["Qwen3ForCausalLM"] | 151936 | bfloat16 | 4.51.3 | 2,048 | 6,144 | 28 | 16 | 8 | silu | 0 | True | 40,960 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.409286 | medium | very_long |
45 | skt/A.X-4.0-Light | qwen2 | ["Qwen2ForCausalLM"] | 102400 | bfloat16 | 4.51.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 16,384 | 1,000,000 | 0.00001 | 0.02 | 0 | 0 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | long |
46 | Intelligent-Internet/II-Medical-32B-Preview | qwen3 | ["Qwen3ForCausalLM"] | 151936 | bfloat16 | 4.52.1 | 5,120 | 25,600 | 64 | 64 | 8 | silu | 0 | True | 40,960 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | false | 80 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
47 | nvidia/OpenCodeReasoning-Nemotron-1.1-32B | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.47.1 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 65,536 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
48 | internlm/POLAR-7B | internlm2 | ["InternLM2ForRewardModel"] | 92544 | float16 | 4.49.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | True | 262,144 | 50,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
49 | skt/A.X-3.1-Light | llama | ["LlamaForCausalLM"] | 102400 | bfloat16 | 4.51.3 | 4,096 | 10,880 | 32 | 32 | 32 | silu | 0.1 | False | 32,768 | 500,000 | 0.00001 | 0.02 | 0 | 0 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | very_long |
50 | microsoft/NextCoder-32B | qwen2 | ["Qwen2ForCausalLM"] | 152064 | float32 | 4.46.0 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
51 | KurmaAI/AQUA-1B | gemma3_text | ["Gemma3ForCausalLM"] | 262144 | bfloat16 | 4.52.4 | 1,152 | 6,912 | 26 | 4 | 1 | null | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 2 | [1, 106] | null | 288 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.414056 | small | very_long |
52 | naver/provence-reranker-debertav3-v1 | Provence | ["Provence"] | 128100 | float32 | 4.45.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | short |
53 | cognitivecomputations/Devstral-Vision-Small-2507 | mistral3 | ["Mistral3ForConditionalGeneration"] | null | bfloat16 | 4.52.4 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | true | false | false | true | null | null | null |
54 | moonshotai/Kimi-K2-Base | kimi_k2 | ["DeepseekV3ForCausalLM"] | 163840 | bfloat16 | 4.48.3 | 7,168 | 18,432 | 61 | 64 | 64 | silu | 0 | True | 131,072 | 50,000 | 0.000001 | 0.02 | 163584 | 163585 | false | 112 | 1 | false | 384 | 8 | false | false | false | false | false | true | false | true | false | true | 37.610324 | large | very_long |
55 | marcelbinz/Llama-3.1-Centaur-70B | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.43.3 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
56 | 0-hero/Matter-0.1-7B-DPO-preview | mistral | ["MistralForCausalLM"] | 32006 | bfloat16 | 4.38.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
57 | 0-hero/Matter-0.1-7B-boost | mistral | ["MistralForCausalLM"] | 32006 | bfloat16 | 4.38.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
58 | 0-hero/Matter-0.2-32B | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.40.0.dev0 | 5,120 | 27,392 | 64 | 40 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | 128245 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
59 | 0-hero/Matter-0.2-8x22B | mixtral | ["MixtralForCausalLM"] | 32006 | bfloat16 | 4.40.0.dev0 | 6,144 | 16,384 | 56 | 48 | 8 | silu | 0 | False | 65,536 | 1,000,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 6 | false | 8 | 2 | false | false | false | false | false | false | false | true | true | true | 25.367151 | large | very_long |
60 | 0-hero/flan-OIG-base | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.27.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
61 | 0-hero/flan-OIG-small | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.27.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
62 | 0-hero/flan-OIG-ul2 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.27.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
63 | 0-hero/flan-OIG-xl | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.27.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
64 | 0-hero/flan-alpaca-ul2 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.27.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
65 | 0-hero/r1-7B-grpo-v3.1-epoch-2 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.50.0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 131,072 | 10,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
66 | 0-hero/r1-7b-grpo-full | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.49.0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 131,072 | 10,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
67 | 0-ma/beit-geometric-shapes-base | beit | ["BeitForImageClassification"] | 8192 | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
68 | 0-ma/efficientnet-b2-geometric-shapes | efficientnet | ["EfficientNetForImageClassification"] | null | float32 | 4.44.2 | null | null | 64 | null | null | swish | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
69 | 0-ma/focalnet-geometric-shapes-tiny | focalnet | ["FocalNetForImageClassification"] | null | float32 | 4.44.2 | null | null | null | null | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
70 | 0-ma/mobilenet-v2-geometric-shapes | mobilenet_v2 | ["MobileNetV2ForImageClassification"] | null | float32 | 4.44.2 | null | null | null | null | null | relu6 | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
71 | 0-ma/swin-geometric-shapes-tiny | swin | ["SwinForImageClassification"] | null | float32 | 4.44.2 | 768 | null | null | null | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
72 | 00000-X/Dolphin-2.6-FC_Hermes-2-DPO | mistral | ["MistralForCausalLM"] | 32001 | bfloat16 | 4.38.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
73 | 00000-X/Dolphin-2.6-FC_Hermes-2-Pro | mistral | ["MistralForCausalLM"] | 32001 | bfloat16 | 4.38.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
74 | 00000-X/Hermes-2-DPO_WestLake-7B-v2 | mistral | ["MistralForCausalLM"] | 32002 | bfloat16 | 4.38.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
75 | 00000-X/Nous-Hermes-2-DPO_into_Dolphin_Mistral_2.8_v02 | mistral | ["MistralForCausalLM"] | 32002 | bfloat16 | 4.38.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
76 | 00000-X/Nous-Hermes-2-DPO_into_Nous_Hermes-2-Pro | mistral | ["MistralForCausalLM"] | 32002 | bfloat16 | 4.38.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
77 | 00000-X/Nous-Hermes-2-Pro_into_Nous_Hermes-2-DPO | mistral | ["MistralForCausalLM"] | 32032 | bfloat16 | 4.38.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
78 | 000Yash000/crickaichat-llama-3.2-7b | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
79 | 002311-A/RoBERTwito | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 250002 | float32 | 4.32.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
80 | 002311-A/robertinho | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 250002 | float32 | 4.32.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
81 | 007ankit/Phi-Marketing-EMO-phi-128k-3.8b | phi3 | ["Phi3ForCausalLM"] | 32011 | float16 | 4.48.2 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 131,072 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.623879 | medium | very_long |
82 | 007ankit/mergekit-della-cczeply | phi3 | ["Phi3ForCausalLM"] | 32011 | float16 | 4.48.2 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 131,072 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.623879 | medium | very_long |
83 | 007ankit/mergekit-della-igvenqz | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.48.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
84 | 007ankit/mergekit-ties-kjfuaal | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.48.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
85 | 007ankit/mergekit-ties-xbmrcyj | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.48.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
86 | 007ankit/mergekit-ties-xxazzps | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.48.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
87 | 007imvishesh/gemma-Code-Instruct-Finetune-test | gemma | ["GemmaForCausalLM"] | 256000 | float16 | 4.38.0 | 2,048 | 16,384 | 18 | 8 | 1 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
88 | 00K4M1/ppo_LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
89 | 01-ai/Yi-1.5-34B-32K | llama | ["LlamaForCausalLM"] | 64000 | bfloat16 | 4.37.1 | 7,168 | 20,480 | 60 | 56 | 8 | silu | 0 | True | 32,768 | 5,000,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 7 | false | null | null | true | false | false | false | false | false | false | false | true | true | 36.993761 | large | very_long |
90 | 01-ai/Yi-1.5-34B-Chat | llama | ["LlamaForCausalLM"] | 64000 | bfloat16 | 4.40.0 | 7,168 | 20,480 | 60 | 56 | 8 | silu | 0 | False | 4,096 | 5,000,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 7 | false | null | null | true | false | false | false | false | false | false | false | true | true | 36.993761 | large | medium |
91 | 01-ai/Yi-1.5-34B-Chat-16K | llama | ["LlamaForCausalLM"] | 64000 | bfloat16 | 4.40.0 | 7,168 | 20,480 | 60 | 56 | 8 | silu | 0 | False | 16,384 | 5,000,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 7 | false | null | null | true | false | false | false | false | false | false | false | true | true | 36.993761 | large | long |
92 | 01-ai/Yi-1.5-9B-32K | llama | ["LlamaForCausalLM"] | 64000 | bfloat16 | 4.37.1 | 4,096 | 11,008 | 48 | 32 | 4 | silu | 0 | True | 32,768 | 5,000,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 9.663676 | large | very_long |
93 | 01-ai/Yi-1.5-9B-Chat | llama | ["LlamaForCausalLM"] | 64000 | bfloat16 | 4.40.0 | 4,096 | 11,008 | 48 | 32 | 4 | silu | 0 | False | 4,096 | 5,000,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 9.663676 | large | medium |
94 | 01-ai/Yi-1.5-9B-Chat-16K | llama | ["LlamaForCausalLM"] | 64000 | bfloat16 | 4.40.0 | 4,096 | 11,008 | 48 | 32 | 4 | silu | 0 | False | 16,384 | 5,000,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 9.663676 | large | long |
95 | 01-ai/Yi-34B-200K | llama | ["LlamaForCausalLM"] | 64000 | bfloat16 | 4.34.0 | 7,168 | 20,480 | 60 | 56 | 8 | silu | null | True | 200,000 | 10,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 7 | false | null | null | true | false | false | false | false | false | false | false | true | true | 36.993761 | large | very_long |
96 | 01-ai/Yi-Coder-1.5B-Chat | llama | ["LlamaForCausalLM"] | 64000 | bfloat16 | 4.44.0 | 2,048 | 5,504 | 24 | 16 | 16 | silu | 0 | False | 131,072 | 10,000,000 | 0.00001 | 0.02 | 1 | 7 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | very_long |
97 | 019100385b/SitiosTuristicosyArqueologicosCusco | vit | ["ViTForImageClassification"] | null | float32 | 4.40.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
98 | 01GangaPutraBheeshma/ut_imdb_movies_classification | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.35.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
99 | 01PrathamS/mistral-v2_finetune_unsloth_train_merged | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.51.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.