CompactAI commited on
Commit
5ca1b9e
·
verified ·
1 Parent(s): fc89df5

Upload 2 files

Browse files
Files changed (1) hide show
  1. app.py +8 -2
app.py CHANGED
@@ -2862,6 +2862,9 @@ def _chat_stream(history, version, ckpt_label, mode_key, use_custom,
2862
  }
2863
  else:
2864
  cfg = dict(MODES[mode_key])
 
 
 
2865
 
2866
  if prior_msgs:
2867
  prompt = _build_conversation_prompt(prior_msgs, new_msg)
@@ -2903,6 +2906,9 @@ def _compare_fn(prompt, selected_versions, mode_key, use_custom,
2903
  }
2904
  else:
2905
  cfg = dict(MODES[mode_key])
 
 
 
2906
 
2907
  all_versions = _collection_versions()
2908
  selected = set(selected_versions or [])
@@ -3047,8 +3053,8 @@ def _advanced_block():
3047
  soft_cap = gr.Slider(0.0, 50.0, value=20.0, step=0.5, label="Logit soft cap")
3048
  loop_pen = gr.Slider(0.0, 50.0, value=15.0, step=0.5, label="Loop penalty")
3049
  with gr.Row():
3050
- max_tokens = gr.Slider(16, 2048, value=512, step=16, label="Max new tokens")
3051
- ctx_win = gr.Slider(128, 2048, value=2048, step=128, label="Context window")
3052
  return use_custom, temperature, top_k, top_p, min_p, rep_penalty, ngram_size, soft_cap, loop_pen, max_tokens, ctx_win, raw_mode
3053
 
3054
 
 
2862
  }
2863
  else:
2864
  cfg = dict(MODES[mode_key])
2865
+ # Max new tokens slider always applies (independent of preset override)
2866
+ cfg["max_new_tokens"] = int(max_tokens)
2867
+ cfg["context_window"] = int(ctx_win)
2868
 
2869
  if prior_msgs:
2870
  prompt = _build_conversation_prompt(prior_msgs, new_msg)
 
2906
  }
2907
  else:
2908
  cfg = dict(MODES[mode_key])
2909
+ # Max new tokens slider always applies (independent of preset override)
2910
+ cfg["max_new_tokens"] = int(max_tokens)
2911
+ cfg["context_window"] = int(ctx_win)
2912
 
2913
  all_versions = _collection_versions()
2914
  selected = set(selected_versions or [])
 
3053
  soft_cap = gr.Slider(0.0, 50.0, value=20.0, step=0.5, label="Logit soft cap")
3054
  loop_pen = gr.Slider(0.0, 50.0, value=15.0, step=0.5, label="Loop penalty")
3055
  with gr.Row():
3056
+ max_tokens = gr.Slider(16, 4096, value=512, step=16, label="Max new tokens")
3057
+ ctx_win = gr.Slider(128, 4096, value=2048, step=128, label="Context window")
3058
  return use_custom, temperature, top_k, top_p, min_p, rep_penalty, ngram_size, soft_cap, loop_pen, max_tokens, ctx_win, raw_mode
3059
 
3060