From 4d8a7663d3a2fb82f88a59cc3489f1ccac9675b3 Mon Sep 17 00:00:00 2001 From: IlyasMoutawwakil Date: Thu, 4 Jul 2024 10:56:39 +0200 Subject: [PATCH] fix --- examples/onnxruntime/optimization/multiple-choice/run_swag.py | 2 +- examples/onnxruntime/quantization/question-answering/run_qa.py | 1 - 2 files changed, 1 insertion(+), 2 deletions(-) diff --git a/examples/onnxruntime/optimization/multiple-choice/run_swag.py b/examples/onnxruntime/optimization/multiple-choice/run_swag.py index b2a9398d94..bcddc97590 100644 --- a/examples/onnxruntime/optimization/multiple-choice/run_swag.py +++ b/examples/onnxruntime/optimization/multiple-choice/run_swag.py @@ -346,7 +346,7 @@ def compute_metrics(eval_predictions): outputs = evaluation_loop( model=model, dataset=eval_dataset, - label_names=["labels"], + label_names=["label"], compute_metrics=compute_metrics, ) diff --git a/examples/onnxruntime/quantization/question-answering/run_qa.py b/examples/onnxruntime/quantization/question-answering/run_qa.py index 4b5648d70d..50661b7b42 100644 --- a/examples/onnxruntime/quantization/question-answering/run_qa.py +++ b/examples/onnxruntime/quantization/question-answering/run_qa.py @@ -683,7 +683,6 @@ def compute_metrics(p: EvalPrediction): outputs = evaluation_loop( model=model, dataset=predict_dataset, - compute_metrics=compute_metrics, label_names=["start_positions", "end_positions"], ) predictions = post_processing_function(predict_examples, predict_dataset, outputs.predictions)