Skip to content

Commit 08ec908

Browse files
Remove itrex dependency for 2x example (#2024)
Signed-off-by: Kaihui-intel <[email protected]> Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
1 parent d9377b8 commit 08ec908

File tree

14 files changed

+44
-25
lines changed

14 files changed

+44
-25
lines changed

examples/onnxrt/nlp/huggingface_model/text_generation/llama/quantization/ptq_static/main.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -197,7 +197,7 @@ def replace_architectures(json_path):
197197
json.dump(data, file, indent=4)
198198

199199
def eval_func(model):
200-
from intel_extension_for_transformers.transformers.llm.evaluation.lm_eval import evaluate, LMEvalParser
200+
from neural_compressor.evaluation.lm_eval import evaluate, LMEvalParser
201201

202202
model_dir = model
203203
if isinstance(model, str) and model.endswith(".onnx"):

examples/onnxrt/nlp/huggingface_model/text_generation/llama/quantization/ptq_static/requirements.txt

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,6 +7,7 @@ onnxruntime-extensions; python_version < '3.11'
77
datasets
88
optimum
99
evaluate
10-
intel-extension-for-transformers >= 1.4.1
1110
peft
12-
lm-eval==0.4.2
11+
lm-eval==0.4.3
12+
numba
13+
pydantic

examples/onnxrt/nlp/huggingface_model/text_generation/llama/quantization/weight_only/main.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -134,7 +134,7 @@ def replace_architectures(json_path):
134134
json.dump(data, file, indent=4)
135135

136136
def eval_func(model):
137-
from intel_extension_for_transformers.transformers.llm.evaluation.lm_eval import evaluate, LMEvalParser
137+
from neural_compressor.evaluation.lm_eval import evaluate, LMEvalParser
138138

139139
model_dir = model
140140
if isinstance(model, str) and model.endswith(".onnx"):

examples/onnxrt/nlp/huggingface_model/text_generation/llama/quantization/weight_only/requirements.txt

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,6 +7,7 @@ onnxruntime-extensions; python_version < '3.11'
77
datasets
88
optimum
99
evaluate
10-
intel-extension-for-transformers >= 1.4.1
1110
peft
12-
lm-eval==0.4.2
11+
lm-eval==0.4.3
12+
numba
13+
pydantic
Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,6 @@
11
accelerate
22
datasets
33
einops
4-
intel-extension-for-transformers
54
optimum
65
peft
76
sentencepiece
@@ -10,4 +9,6 @@ torch
109
tqdm
1110
tiktoken
1211
transformers_stream_generator
13-
lm_eval==0.4.2
12+
lm_eval==0.4.3
13+
numba
14+
pydantic

examples/pytorch/nlp/huggingface_models/language-modeling/pruning/eager/run_clm_sparsegpt.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -588,7 +588,7 @@ def group_texts(examples):
588588
eval_batch = args.per_device_eval_batch_size
589589
user_model = None if args.use_accelerate else model
590590

591-
from intel_extension_for_transformers.transformers.llm.evaluation.lm_eval import evaluate, LMEvalParser
591+
from neural_compressor.evaluation.lm_eval import evaluate, LMEvalParser
592592
eval_args = LMEvalParser(
593593
model="hf",
594594
user_model=user_model,

examples/pytorch/nlp/huggingface_models/language-modeling/quantization/llm/requirements.txt

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -8,6 +8,6 @@ transformers
88
pytest
99
wandb
1010
einops
11-
neural-compressor
12-
intel-extension-for-transformers
13-
lm_eval==0.4.2
11+
lm_eval==0.4.3
12+
numba
13+
pydantic

examples/pytorch/nlp/huggingface_models/language-modeling/quantization/llm/run_benchmark.sh

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -89,10 +89,10 @@ function run_benchmark {
8989
extra_cmd=$extra_cmd" --woq_algo TEQ"
9090
elif [ "${topology}" = "opt_125m_ipex" ]; then
9191
model_name_or_path="facebook/opt-125m"
92-
extra_cmd=$extra_cmd" --ipex --int8_bf16_mixed"
92+
extra_cmd=$extra_cmd" --ipex"
9393
elif [ "${topology}" = "opt_125m_ipex_sq" ]; then
9494
model_name_or_path="facebook/opt-125m"
95-
extra_cmd=$extra_cmd" --ipex --int8_bf16_mixed --sq --alpha 0.5"
95+
extra_cmd=$extra_cmd" --ipex --sq --alpha 0.5"
9696
elif [ "${topology}" = "bloom_560m_ipex_sq" ]; then
9797
model_name_or_path="bigscience/bloom-560m"
9898
extra_cmd=$extra_cmd" --ipex --sq --alpha 0.5"

examples/pytorch/nlp/huggingface_models/language-modeling/quantization/llm/run_clm_no_trainer.py

Lines changed: 7 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -343,7 +343,10 @@ def eval_func(model):
343343

344344
if args.ipex:
345345
user_model = load(os.path.abspath(os.path.expanduser(args.output_dir)))
346+
from transformers import AutoTokenizer, AutoConfig
346347
tokenizer = AutoTokenizer.from_pretrained(args.model, trust_remote_code=args.trust_remote_code)
348+
config = AutoConfig.from_pretrained(args.model, trust_remote_code=args.trust_remote_code)
349+
setattr(user_model, "config", config)
347350
else:
348351
user_model, tokenizer = get_user_model()
349352
kwargs = {'weight_only': True} if args.approach == 'weight_only' else {}
@@ -354,7 +357,7 @@ def eval_func(model):
354357
if args.accuracy:
355358
user_model.eval()
356359
if args.code_generation:
357-
from intel_extension_for_transformers.transformers.llm.evaluation.bigcode_eval import evaluate
360+
from neural_compressor.evaluation.bigcode_eval import evaluate
358361
from transformers import AutoTokenizer
359362
tokenizer = AutoTokenizer.from_pretrained(args.model, trust_remote_code=args.trust_remote_code)
360363
results = evaluate(
@@ -370,7 +373,7 @@ def eval_func(model):
370373
else:
371374
acc = results["results"][task_name]["acc"]
372375
else:
373-
from intel_extension_for_transformers.transformers.llm.evaluation.lm_eval import evaluate, LMEvalParser
376+
from neural_compressor.evaluation.lm_eval import evaluate, LMEvalParser
374377
eval_args = LMEvalParser(
375378
model="hf",
376379
user_model=user_model,
@@ -395,7 +398,7 @@ def eval_func(model):
395398
samples = args.iters * args.batch_size
396399

397400
if args.code_generation:
398-
from intel_extension_for_transformers.transformers.llm.evaluation.bigcode_eval import evaluate
401+
from neural_compressor.evaluation.bigcode_eval import evaluate
399402
from transformers import AutoTokenizer
400403
tokenizer = AutoTokenizer.from_pretrained(args.model, trust_remote_code=args.trust_remote_code)
401404
start = time.time()
@@ -413,7 +416,7 @@ def eval_func(model):
413416
else:
414417
acc = results["results"][task_name]["acc"]
415418
else:
416-
from intel_extension_for_transformers.transformers.llm.evaluation.lm_eval import evaluate, LMEvalParser
419+
from neural_compressor.evaluation.lm_eval import evaluate, LMEvalParser
417420
eval_args = LMEvalParser(
418421
model="hf",
419422
user_model=user_model,

examples/pytorch/speech_recognition/whisper_large/quantization/ptq_dynamic/fx/run_quant.sh

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -45,7 +45,8 @@ function run_tuning {
4545
--tune \
4646
--batch_size $batch_size \
4747
--output_dir ${output_model} \
48-
--cache_dir ${dataset_location}
48+
--cache_dir ${dataset_location} \
49+
--trust_remote_code
4950

5051
}
5152

0 commit comments

Comments
 (0)