Spaces:
Sleeping
Sleeping
Fix requirements
Browse files- app.py +5 -2
- requirements.txt +1 -1
- utilities.py +1 -0
app.py
CHANGED
|
@@ -734,6 +734,8 @@ with gr.Blocks(theme='shivi/calm_seafoam', css_paths='style.css', js=js_func) as
|
|
| 734 |
metrics_conc = target_df
|
| 735 |
if 'valid_efficency_score' not in metrics_conc.columns:
|
| 736 |
metrics_conc['valid_efficency_score'] = metrics_conc['VES']
|
|
|
|
|
|
|
| 737 |
eval_text = generate_eval_text("End evaluation")
|
| 738 |
yield gr.Markdown(eval_text, visible=True), gr.Image(), gr.Markdown(), gr.Markdown(), gr.Markdown(), metrics_conc, *[predictions_dict[model][columns_to_visulize] for model in model_list]
|
| 739 |
|
|
@@ -868,14 +870,15 @@ with gr.Blocks(theme='shivi/calm_seafoam', css_paths='style.css', js=js_func) as
|
|
| 868 |
db_path_name="db_path"
|
| 869 |
)
|
| 870 |
else:
|
|
|
|
| 871 |
metrics_df_model = us.evaluate_answer(predictions_dict[model])
|
| 872 |
metrics_df_model['model'] = model
|
| 873 |
metrics_conc = pd.concat([metrics_conc, metrics_df_model], ignore_index=True)
|
| 874 |
|
| 875 |
-
if 'valid_efficency_score' not in metrics_conc.columns
|
| 876 |
metrics_conc['valid_efficency_score'] = metrics_conc['VES']
|
| 877 |
|
| 878 |
-
if 'VES' not in metrics_conc.columns
|
| 879 |
metrics_conc['VES'] = metrics_conc['valid_efficency_score']
|
| 880 |
|
| 881 |
eval_text = generate_eval_text("End evaluation")
|
|
|
|
| 734 |
metrics_conc = target_df
|
| 735 |
if 'valid_efficency_score' not in metrics_conc.columns:
|
| 736 |
metrics_conc['valid_efficency_score'] = metrics_conc['VES']
|
| 737 |
+
if 'VES' not in metrics_conc.columns:
|
| 738 |
+
metrics_conc['VES'] = metrics_conc['valid_efficency_score']
|
| 739 |
eval_text = generate_eval_text("End evaluation")
|
| 740 |
yield gr.Markdown(eval_text, visible=True), gr.Image(), gr.Markdown(), gr.Markdown(), gr.Markdown(), metrics_conc, *[predictions_dict[model][columns_to_visulize] for model in model_list]
|
| 741 |
|
|
|
|
| 870 |
db_path_name="db_path"
|
| 871 |
)
|
| 872 |
else:
|
| 873 |
+
|
| 874 |
metrics_df_model = us.evaluate_answer(predictions_dict[model])
|
| 875 |
metrics_df_model['model'] = model
|
| 876 |
metrics_conc = pd.concat([metrics_conc, metrics_df_model], ignore_index=True)
|
| 877 |
|
| 878 |
+
if 'valid_efficency_score' not in metrics_conc.columns:
|
| 879 |
metrics_conc['valid_efficency_score'] = metrics_conc['VES']
|
| 880 |
|
| 881 |
+
if 'VES' not in metrics_conc.columns:
|
| 882 |
metrics_conc['VES'] = metrics_conc['valid_efficency_score']
|
| 883 |
|
| 884 |
eval_text = generate_eval_text("End evaluation")
|
requirements.txt
CHANGED
|
@@ -22,7 +22,7 @@ statistics>=1.0.3.5
|
|
| 22 |
transformers>=4.49.0
|
| 23 |
fsspec==2024.12.0
|
| 24 |
pytest
|
| 25 |
-
|
| 26 |
# Docs dependencies
|
| 27 |
mkdocs>=1.5.3
|
| 28 |
mkdocs-material>=9.4.7
|
|
|
|
| 22 |
transformers>=4.49.0
|
| 23 |
fsspec==2024.12.0
|
| 24 |
pytest
|
| 25 |
+
json
|
| 26 |
# Docs dependencies
|
| 27 |
mkdocs>=1.5.3
|
| 28 |
mkdocs-material>=9.4.7
|
utilities.py
CHANGED
|
@@ -10,6 +10,7 @@ from qatch.evaluate_dataset.metrics_evaluators import CellPrecision, CellRecall,
|
|
| 10 |
import qatch.evaluate_dataset.orchestrator_evaluator as eva
|
| 11 |
import utils_get_db_tables_info
|
| 12 |
#import tiktoken
|
|
|
|
| 13 |
from transformers import AutoTokenizer
|
| 14 |
|
| 15 |
def extract_tables(file_path):
|
|
|
|
| 10 |
import qatch.evaluate_dataset.orchestrator_evaluator as eva
|
| 11 |
import utils_get_db_tables_info
|
| 12 |
#import tiktoken
|
| 13 |
+
import json
|
| 14 |
from transformers import AutoTokenizer
|
| 15 |
|
| 16 |
def extract_tables(file_path):
|