Jerrycool commited on
Commit
cc4532a
Β·
verified Β·
1 Parent(s): be52959

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +0 -87
app.py CHANGED
@@ -138,93 +138,6 @@ with demo:
138
  # (Content unchanged)
139
  gr.Markdown(LLM_BENCHMARKS_TEXT, elem_classes="markdown-text")
140
 
141
- with gr.TabItem("πŸš€ Submit here! ", elem_id="llm-benchmark-tab-table", id=3):
142
- # (Content unchanged, still uses potentially empty/mock queue data)
143
- with gr.Column():
144
- with gr.Row():
145
- gr.Markdown(EVALUATION_QUEUE_TEXT, elem_classes="markdown-text")
146
- with gr.Column():
147
- with gr.Accordion(
148
- f"βœ… Finished Evaluations ({len(finished_eval_queue_df)})",
149
- open=False,
150
- ):
151
- with gr.Row():
152
- finished_eval_table = gr.components.Dataframe(
153
- value=finished_eval_queue_df,
154
- headers=EVAL_COLS,
155
- datatype=EVAL_TYPES,
156
- row_count=5,
157
- )
158
- with gr.Accordion(
159
- f"πŸ”„ Running Evaluation Queue ({len(running_eval_queue_df)})",
160
- open=False,
161
- ):
162
- with gr.Row():
163
- running_eval_table = gr.components.Dataframe(
164
- value=running_eval_queue_df,
165
- headers=EVAL_COLS,
166
- datatype=EVAL_TYPES,
167
- row_count=5,
168
- )
169
- with gr.Accordion(
170
- f"⏳ Pending Evaluation Queue ({len(pending_eval_queue_df)})",
171
- open=False,
172
- ):
173
- with gr.Row():
174
- pending_eval_table = gr.components.Dataframe(
175
- value=pending_eval_queue_df,
176
- headers=EVAL_COLS,
177
- datatype=EVAL_TYPES,
178
- row_count=5,
179
- )
180
-
181
- with gr.Row():
182
- gr.Markdown("# βœ‰οΈβœ¨ Submit your model here!", elem_classes="markdown-text")
183
- with gr.Row():
184
- # Submission form - kept as is
185
- with gr.Column():
186
- model_name_textbox = gr.Textbox(label="Model name")
187
- revision_name_textbox = gr.Textbox(label="Revision commit", placeholder="main")
188
- model_type = gr.Dropdown(
189
- choices=["Type A", "Type B", "Type C"], # Example choices
190
- label="Model type",
191
- multiselect=False,
192
- value=None,
193
- interactive=True,
194
- )
195
- with gr.Column():
196
- precision = gr.Dropdown(
197
- choices=["float16", "bfloat16", "float32", "int8"], # Example choices
198
- label="Precision",
199
- multiselect=False,
200
- value="float16",
201
- interactive=True,
202
- )
203
- weight_type = gr.Dropdown(
204
- choices=["Original", "Adapter", "Delta"], # Example choices
205
- label="Weights type",
206
- multiselect=False,
207
- value="Original",
208
- interactive=True,
209
- )
210
- base_model_name_textbox = gr.Textbox(label="Base model (for delta or adapter weights)")
211
-
212
- submit_button = gr.Button("Submit Eval")
213
- submission_result = gr.Markdown()
214
-
215
- submit_button.click(
216
- add_new_eval,
217
- [
218
- model_name_textbox,
219
- base_model_name_textbox,
220
- revision_name_textbox,
221
- precision,
222
- weight_type,
223
- model_type,
224
- ],
225
- submission_result,
226
- )
227
-
228
  with gr.Row():
229
  with gr.Accordion("πŸ“™ Citation", open=False):
230
  # (Content unchanged)
 
138
  # (Content unchanged)
139
  gr.Markdown(LLM_BENCHMARKS_TEXT, elem_classes="markdown-text")
140
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
141
  with gr.Row():
142
  with gr.Accordion("πŸ“™ Citation", open=False):
143
  # (Content unchanged)