Clémentine commited on
Commit
06fe586
·
1 Parent(s): 676e2ea

added debug mode locally, fixed a couple typos

Browse files
Files changed (1) hide show
  1. app.py +36 -22
app.py CHANGED
@@ -33,6 +33,8 @@ ref_level_len = {"validation": {1: 53, 2: 86, 3: 26}, "test": {1: 93, 2: 159, 3:
33
 
34
  os.makedirs("scored", exist_ok=True)
35
 
 
 
36
  # Display the results
37
  eval_results = load_dataset(RESULTS_DATASET, YEAR_VERSION, token=TOKEN, download_mode="force_redownload", verification_mode=VerificationMode.NO_CHECKS, trust_remote_code=True)
38
  contact_infos = load_dataset(CONTACT_DATASET, YEAR_VERSION, token=TOKEN, download_mode="force_redownload", verification_mode=VerificationMode.NO_CHECKS, trust_remote_code=True)
@@ -81,8 +83,8 @@ def add_new_eval(
81
  profile: gr.OAuthProfile,
82
  ):
83
  contact_infos = load_dataset(CONTACT_DATASET, YEAR_VERSION, token=TOKEN, download_mode="force_redownload", verification_mode=VerificationMode.NO_CHECKS, trust_remote_code=True)
84
- user_submission_dates = sorted(row["date"] for row in contact_info[val_or_test] if row["username"] == profile.username)
85
- if user_submission_dates[-1] == datetime.datetime.today().strftime('%Y-%m-%d'):
86
  return format_error("You already submitted once today, please try again tomorrow.")
87
 
88
 
@@ -102,13 +104,16 @@ def add_new_eval(
102
  return format_warning("Please attach a file.")
103
 
104
  # Save submitted file
105
- api.upload_file(
106
- repo_id=SUBMISSION_DATASET,
107
- path_or_fileobj=path_to_file.name,
108
- path_in_repo=f"{organisation}/{model}/{YEAR_VERSION}_{val_or_test}_raw_{datetime.datetime.today()}.jsonl",
109
- repo_type="dataset",
110
- token=TOKEN
111
- )
 
 
 
112
 
113
  # Compute score
114
  file_path = path_to_file.name
@@ -157,24 +162,27 @@ def add_new_eval(
157
  return format_error(f"Your submission has {num_questions[1]} questions for level 1, {num_questions[2]} for level 2, and {num_questions[3]} for level 3, but it should have {ref_level_len[val_or_test][1]}, {ref_level_len[val_or_test][2]}, and {ref_level_len[val_or_test][3]} respectively. Please check your submission.")
158
 
159
  # Save scored file
160
- api.upload_file(
161
- repo_id=SUBMISSION_DATASET,
162
- path_or_fileobj=f"scored/{organisation}_{model}.jsonl",
163
- path_in_repo=f"{organisation}/{model}/{YEAR_VERSION}_{val_or_test}_scored_{datetime.datetime.today()}.jsonl",
164
- repo_type="dataset",
165
- token=TOKEN
166
- )
167
-
168
- # Save scored file
169
- if is_validation:
170
  api.upload_file(
171
- repo_id=SUBMISSION_DATASET_PUBLIC,
172
  path_or_fileobj=f"scored/{organisation}_{model}.jsonl",
173
  path_in_repo=f"{organisation}/{model}/{YEAR_VERSION}_{val_or_test}_scored_{datetime.datetime.today()}.jsonl",
174
  repo_type="dataset",
175
  token=TOKEN
176
  )
177
 
 
 
 
 
 
 
 
 
 
 
178
  # Actual submission
179
  eval_entry = {
180
  "model": model,
@@ -193,7 +201,10 @@ def add_new_eval(
193
 
194
  eval_results[val_or_test] = eval_results[val_or_test].add_item(eval_entry)
195
  print(eval_results)
196
- eval_results.push_to_hub(RESULTS_DATASET, config_name = YEAR_VERSION, token=TOKEN)
 
 
 
197
 
198
  contact_info = {
199
  "model": model,
@@ -205,7 +216,10 @@ def add_new_eval(
205
  "date": datetime.datetime.today().strftime('%Y-%m-%d')
206
  }
207
  contact_infos[val_or_test]= contact_infos[val_or_test].add_item(contact_info)
208
- contact_infos.push_to_hub(CONTACT_DATASET, config_name = YEAR_VERSION, token=TOKEN)
 
 
 
209
 
210
  return format_log(f"Model {model} submitted by {organisation} successfully.\nPlease wait a few hours and refresh the leaderboard to see your score displayed.")
211
 
 
33
 
34
  os.makedirs("scored", exist_ok=True)
35
 
36
+ LOCAL_DEBUG = False
37
+
38
  # Display the results
39
  eval_results = load_dataset(RESULTS_DATASET, YEAR_VERSION, token=TOKEN, download_mode="force_redownload", verification_mode=VerificationMode.NO_CHECKS, trust_remote_code=True)
40
  contact_infos = load_dataset(CONTACT_DATASET, YEAR_VERSION, token=TOKEN, download_mode="force_redownload", verification_mode=VerificationMode.NO_CHECKS, trust_remote_code=True)
 
83
  profile: gr.OAuthProfile,
84
  ):
85
  contact_infos = load_dataset(CONTACT_DATASET, YEAR_VERSION, token=TOKEN, download_mode="force_redownload", verification_mode=VerificationMode.NO_CHECKS, trust_remote_code=True)
86
+ user_submission_dates = sorted(row["date"] for row in contact_infos[val_or_test] if row["username"] == profile.username)
87
+ if len(user_submission_dates) > 0 and user_submission_dates[-1] == datetime.datetime.today().strftime('%Y-%m-%d'):
88
  return format_error("You already submitted once today, please try again tomorrow.")
89
 
90
 
 
104
  return format_warning("Please attach a file.")
105
 
106
  # Save submitted file
107
+ if LOCAL_DEBUG:
108
+ print("mock uploaded submission")
109
+ else:
110
+ api.upload_file(
111
+ repo_id=SUBMISSION_DATASET,
112
+ path_or_fileobj=path_to_file.name,
113
+ path_in_repo=f"{organisation}/{model}/{YEAR_VERSION}_{val_or_test}_raw_{datetime.datetime.today()}.jsonl",
114
+ repo_type="dataset",
115
+ token=TOKEN
116
+ )
117
 
118
  # Compute score
119
  file_path = path_to_file.name
 
162
  return format_error(f"Your submission has {num_questions[1]} questions for level 1, {num_questions[2]} for level 2, and {num_questions[3]} for level 3, but it should have {ref_level_len[val_or_test][1]}, {ref_level_len[val_or_test][2]}, and {ref_level_len[val_or_test][3]} respectively. Please check your submission.")
163
 
164
  # Save scored file
165
+ if LOCAL_DEBUG:
166
+ print("mock uploaded scored submission")
167
+ else:
 
 
 
 
 
 
 
168
  api.upload_file(
169
+ repo_id=SUBMISSION_DATASET,
170
  path_or_fileobj=f"scored/{organisation}_{model}.jsonl",
171
  path_in_repo=f"{organisation}/{model}/{YEAR_VERSION}_{val_or_test}_scored_{datetime.datetime.today()}.jsonl",
172
  repo_type="dataset",
173
  token=TOKEN
174
  )
175
 
176
+ # Save scored file
177
+ if is_validation:
178
+ api.upload_file(
179
+ repo_id=SUBMISSION_DATASET_PUBLIC,
180
+ path_or_fileobj=f"scored/{organisation}_{model}.jsonl",
181
+ path_in_repo=f"{organisation}/{model}/{YEAR_VERSION}_{val_or_test}_scored_{datetime.datetime.today()}.jsonl",
182
+ repo_type="dataset",
183
+ token=TOKEN
184
+ )
185
+
186
  # Actual submission
187
  eval_entry = {
188
  "model": model,
 
201
 
202
  eval_results[val_or_test] = eval_results[val_or_test].add_item(eval_entry)
203
  print(eval_results)
204
+ if LOCAL_DEBUG:
205
+ print("mock uploaded results to lb")
206
+ else:
207
+ eval_results.push_to_hub(RESULTS_DATASET, config_name = YEAR_VERSION, token=TOKEN)
208
 
209
  contact_info = {
210
  "model": model,
 
216
  "date": datetime.datetime.today().strftime('%Y-%m-%d')
217
  }
218
  contact_infos[val_or_test]= contact_infos[val_or_test].add_item(contact_info)
219
+ if LOCAL_DEBUG:
220
+ print("mock uploaded contact info")
221
+ else:
222
+ contact_infos.push_to_hub(CONTACT_DATASET, config_name = YEAR_VERSION, token=TOKEN)
223
 
224
  return format_log(f"Model {model} submitted by {organisation} successfully.\nPlease wait a few hours and refresh the leaderboard to see your score displayed.")
225