Itay Levy
itlevy
AI & ML interests
None yet
Organizations
Llama-3_3-Nemotron-Super-49B-v1_5-NVFP4/llama_nemotron_toolcall_parser_no_streaming.py missing
1
#1 opened 3 months ago
by
SuperbEmphasis
Update README and toolcall_parser
#5 opened 2 months ago
by
itlevy
_prepare_generation_config bugfix (failed due to version update in transformers)
#14 opened 7 months ago
by
ishahaf
_prepare_generation_config bugfix (failed due to version update in transformers)
#25 opened 7 months ago
by
ishahaf
_prepare_generation_config bugfix (failed due to version update in transformers)
#2 opened 7 months ago
by
ishahaf
Nemotron 253B?
2
#10 opened 10 months ago
by
BoshiAI
How come this pruned model has 162 layers
5
#3 opened 10 months ago
by
ymcki
add model card
#1 opened 10 months ago
by
itlevy
Patching hf bug that creates wrong cache length if only inputs_embeds are passed to the model
#19 opened over 1 year ago
by
tomer-nv
DeciLMForCausalLM(DeciLMPreTrainedModel, GenerationMixin) for v4.50
#16 opened over 1 year ago
by
itlevy
add batch_size attribute to VariableCache
#15 opened over 1 year ago
by
itlevy
nvidia-open-model-license
#14 opened over 1 year ago
by
itlevy
nvidia-open-model-license
#13 opened over 1 year ago
by
itlevy
nvidia-open-model-license
#12 opened over 1 year ago
by
itlevy
v4.46 support
#7 opened over 1 year ago
by
itlevy
loading as llama model
1
#4 opened over 1 year ago
by
KnutJaegersberg
v4.45 support
#6 opened over 1 year ago
by
itlevy
fixed flash_attention backward_compat
#3 opened over 1 year ago
by
itlevy
flash_attention_utils_backward_compat
#2 opened over 1 year ago
by
itlevy