diff --git a/models/nlp/language_model/bert_large_squad/igie/Int8QAT/run_qat.py b/models/nlp/language_model/bert_large_squad/igie/Int8QAT/run_qat.py index 943d04f4856446fb90fb80f4a269a7dd465b40c4..e53b85b377103cf9864074e5c8a43cdd0fb1fcad 100644 --- a/models/nlp/language_model/bert_large_squad/igie/Int8QAT/run_qat.py +++ b/models/nlp/language_model/bert_large_squad/igie/Int8QAT/run_qat.py @@ -350,6 +350,7 @@ def main(): os.path.join(cur_dir,'squad_download.py'), data_args.dataset_config_name, cache_dir=model_args.cache_dir, + trust_remote_code=True, ) else: data_files = {} @@ -685,7 +686,7 @@ def main(): ] return EvalPrediction(predictions=formatted_predictions, label_ids=references) - metric = load_metric(os.path.join(cur_dir,'squad_data.py')) + metric = load_metric(os.path.join(cur_dir,'squad_data.py'), trust_remote_code=True) def compute_metrics(p: EvalPrediction): return metric.compute(predictions=p.predictions, references=p.label_ids) diff --git a/models/nlp/language_model/bert_large_squad/igie/Int8QAT/squad_download.py b/models/nlp/language_model/bert_large_squad/igie/Int8QAT/squad_download.py index 400b626ab1788a73696a8ae6d2fcad887158433c..3a4bedf8219b4054ddc2055a2eeea07dd55dff85 100644 --- a/models/nlp/language_model/bert_large_squad/igie/Int8QAT/squad_download.py +++ b/models/nlp/language_model/bert_large_squad/igie/Int8QAT/squad_download.py @@ -121,7 +121,8 @@ class Squad(datasets.GeneratorBasedBuilder): ) def _split_generators(self, dl_manager): - downloaded_files = dl_manager.download_and_extract(_URLS) + # downloaded_files = dl_manager.download_and_extract(_URLS) + downloaded_files = { "train": "./data/train-v1.1.json", "dev": "./data/dev-v1.1.json" } return [ datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"filepath": downloaded_files["train"]}), diff --git a/models/nlp/language_model/bert_large_squad/igie/README.md b/models/nlp/language_model/bert_large_squad/igie/README.md index f7bbf87aa2fca31318ac8e9e597ba228fb196ef2..7302c3f0a0aaa637eb765f4e4dc7223a0ea5e926 100644 --- a/models/nlp/language_model/bert_large_squad/igie/README.md +++ b/models/nlp/language_model/bert_large_squad/igie/README.md @@ -25,11 +25,21 @@ Dataset: # Get FP16 Onnx Model python3 export.py --output bert-large-uncased-squad-v1.onnx -# Do QAT for INT8 test, will take a long time +# Do QAT for INT8 test, will take a long time (16 gpus need 1h) cd Int8QAT + +# prepare dataset +mkdir -p data +cp /path/to/SQuAD/train-v1.1.json /path/to/SQuAD/dev-v1.1.json data/ + +# prepare model into bert-large-uncased from +mkdir -p bert-large-uncased + bash run_qat.sh +# model: quant_bert_large/pytorch_model.bin or quant_bert_large/model.safetensors python3 export_hdf5.py --model quant_bert_large/pytorch_model.bin --model_name large + cd .. ``` diff --git a/models/nlp/language_model/bert_large_squad/igie/requirements.txt b/models/nlp/language_model/bert_large_squad/igie/requirements.txt index 21949dbe7c407f3412a9e4abe9444a820e3d60ac..2b340d5815b964d8b6d495ed3d3169765e3be650 100644 --- a/models/nlp/language_model/bert_large_squad/igie/requirements.txt +++ b/models/nlp/language_model/bert_large_squad/igie/requirements.txt @@ -1,4 +1,4 @@ onnx tqdm -numpy==1.23.5 -transformers +numpy==1.23.2 +transformers==4.37.1