burman-ai commited on
Commit
a00ad13
·
verified ·
1 Parent(s): 544cdb6

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +17 -18
app.py CHANGED
@@ -15,18 +15,12 @@ def respond(
15
  message,
16
  history: list[tuple[str, str]],
17
  system_message,
18
- max_tokens= 2048,
19
- temperature= 0.7,
20
- top_p= 0.95,
21
- frequency_penalty= 0,
22
- seed= -1,
23
  max_tokens,
24
  temperature,
25
  top_p,
26
  frequency_penalty,
27
  seed,
28
- custom_model_box,
29
-
30
  ):
31
  print(f"Received message: {message}")
32
  print(f"History: {history}")
@@ -51,11 +45,12 @@ def respond(
51
 
52
  for message_chunk in client.chat.completions.create(
53
  model=model_to_use,
54
- max_tokens= 2048,
55
- temperature= 0.7,
56
- top_p= 0.95,
57
- frequency_penalty= 0,
58
- seed= -1,
 
59
  messages=messages,
60
  ):
61
  token_text = message_chunk.choices[0].delta.content
@@ -66,7 +61,11 @@ chatbot = gr.Chatbot(height=600, show_copy_button=True, placeholder="ChatGPT is
66
 
67
  system_message_box = gr.Label(value="You can select Max Tokens, Temperature, Top-P, Seed")
68
 
69
-
 
 
 
 
70
 
71
  custom_model_box = gr.Textbox(value="meta-llama/Llama-3.2-3B-Instruct", label="AI Mode is ")
72
 
@@ -74,11 +73,11 @@ demo = gr.ChatInterface(
74
  fn=respond,
75
  additional_inputs=[
76
  system_message_box,
77
- max_tokens,
78
- temperature,
79
- top_p,
80
- frequency_penalty,
81
- seed,
82
  custom_model_box,
83
  ],
84
  fill_height=True,
 
15
  message,
16
  history: list[tuple[str, str]],
17
  system_message,
 
 
 
 
 
18
  max_tokens,
19
  temperature,
20
  top_p,
21
  frequency_penalty,
22
  seed,
23
+ custom_model
 
24
  ):
25
  print(f"Received message: {message}")
26
  print(f"History: {history}")
 
45
 
46
  for message_chunk in client.chat.completions.create(
47
  model=model_to_use,
48
+ max_tokens=max_tokens,
49
+ stream=True,
50
+ temperature=temperature,
51
+ top_p=top_p,
52
+ frequency_penalty=frequency_penalty,
53
+ seed=seed,
54
  messages=messages,
55
  ):
56
  token_text = message_chunk.choices[0].delta.content
 
61
 
62
  system_message_box = gr.Label(value="You can select Max Tokens, Temperature, Top-P, Seed")
63
 
64
+ max_tokens_slider = gr.Slider(1024, 2048, value=1024, step=100, label="Max new tokens")
65
+ temperature_slider = gr.Slider(0.1, 1.0, value=0.7, step=0.1, label="Temperature")
66
+ top_p_slider = gr.Slider(0.1, 1.0, value=0.95, step=0.05, label="Top-P")
67
+ frequency_penalty_slider = gr.Slider(-2.0, 2.0, value=0.0, step=0.1, label="Frequency Penalty")
68
+ seed_slider = gr.Slider(-1, 65535, value=-1, step=1, label="Seed (-1 for random)")
69
 
70
  custom_model_box = gr.Textbox(value="meta-llama/Llama-3.2-3B-Instruct", label="AI Mode is ")
71
 
 
73
  fn=respond,
74
  additional_inputs=[
75
  system_message_box,
76
+ max_tokens_slider,
77
+ temperature_slider,
78
+ top_p_slider,
79
+ frequency_penalty_slider,
80
+ seed_slider,
81
  custom_model_box,
82
  ],
83
  fill_height=True,