Skip to content
This repository has been archived by the owner on Oct 25, 2024. It is now read-only.

Commit

Permalink
Merge branch 'main' into wangchang/inc3.x
Browse files Browse the repository at this point in the history
  • Loading branch information
changwangss committed Jul 3, 2024
2 parents 81d0b35 + 86087dc commit 35b24f1
Show file tree
Hide file tree
Showing 37 changed files with 783 additions and 409 deletions.
2 changes: 1 addition & 1 deletion .github/workflows/script/formatScan/pylint.sh
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@ else
echo "Not found requirements.txt file."
fi
# install packages
pip install lm-eval
pip install lm-eval==0.4.2
pip install accelerate nlpaug nltk schema optimum-intel optimum peft
pip install --upgrade --force-reinstall transformers==4.36.2
pip install optimum-habana
Expand Down
1 change: 1 addition & 0 deletions .github/workflows/script/install_binary.sh
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@ source /intel-extension-for-transformers/.github/workflows/script/change_color.s
cd /intel-extension-for-transformers
export CMAKE_ARGS="-DNE_DNNL_CACHE_DIR=/cache"
pip install -U pip
pip install -r requirements.txt
$BOLD_YELLOW && echo "---------------- git submodule update --init --recursive -------------" && $RESET
git config --global --add safe.directory "*"
git submodule update --init --recursive
Expand Down
5 changes: 1 addition & 4 deletions docs/api_doc/optimization/config.rst
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,4 @@ Config
.. autoapisummary::
intel_extension_for_transformers.transformers.utils.metrics
intel_extension_for_transformers.transformers.utils.objectives
intel_extension_for_transformers.transformers.config
intel_extension_for_transformers.transformers.quantization
intel_extension_for_transformers.transformers.distillation
intel_extension_for_transformers.transformers.pruning
intel_extension_for_transformers.transformers.utils.config
6 changes: 0 additions & 6 deletions docs/api_doc/optimization/tf_optimization.rst

This file was deleted.

2 changes: 0 additions & 2 deletions docs/api_doc/user_api.rst
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,5 @@ The following Python API information is available:
:maxdepth: 1

optimization/trainer.rst
optimization/optimizer.rst
optimization/model.rst
optimization/tf_optimization.rst
optimization/config.rst
52 changes: 52 additions & 0 deletions examples/.config/pytorch_optimize.json
Original file line number Diff line number Diff line change
Expand Up @@ -2450,6 +2450,58 @@
}
}
},
"phi_2b_gen_ipex_static": {
"working_dir": "huggingface/pytorch/text-generation/quantization",
"tune": {
"cmd": "bash run_tuning.sh",
"params": {
"topology": "phi_2b",
"task": "generation",
"approach": "static",
"output_model": "saved_results"
}
},
"benchmark": {
"cmd": "bash run_benchmark.sh",
"params": {
"topology": "phi_2b",
"task": "generation",
"approach": "static",
"backend": "ipex",
"mode": "benchmark",
"batch_size": "112",
"iters": "100",
"int8": "false",
"config": "saved_results"
}
}
},
"phi_3b_gen_ipex_static": {
"working_dir": "huggingface/pytorch/text-generation/quantization",
"tune": {
"cmd": "bash run_tuning.sh",
"params": {
"topology": "phi_3b",
"task": "generation",
"approach": "static",
"output_model": "saved_results"
}
},
"benchmark": {
"cmd": "bash run_benchmark.sh",
"params": {
"topology": "phi_3b",
"task": "generation",
"approach": "static",
"backend": "ipex",
"mode": "benchmark",
"batch_size": "112",
"iters": "100",
"int8": "false",
"config": "saved_results"
}
}
},
"flan-t5-large_gen_ipex_static": {
"working_dir": "huggingface/pytorch/text2text-generation",
"tune": {
Expand Down
255 changes: 0 additions & 255 deletions examples/.config/tensorflow_optimize.json

This file was deleted.

5 changes: 2 additions & 3 deletions examples/huggingface/neural_speed/requirements.txt
Original file line number Diff line number Diff line change
@@ -1,12 +1,11 @@
intel_extension_for_transformers
neural-speed
lm-eval
lm-eval==0.4.2
sentencepiece
gguf
--extra-index-url https://download.pytorch.org/whl/cpu
torch==2.3.0+cpu
transformers
intel_extension_for_pytorch==2.3.0
tiktoken
transformers_stream_generator
zipfile38
zipfile38
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
transformers
accelerate
sentencepiece != 0.1.92
lm-eval
lm-eval==0.4.2
Original file line number Diff line number Diff line change
Expand Up @@ -7,5 +7,5 @@ transformers
torch==2.0.1
tqdm
neural_compressor
lm-eval
lm-eval==0.4.2

Original file line number Diff line number Diff line change
Expand Up @@ -9,5 +9,5 @@ wandb
einops
neural-compressor
pytest==8.0.0
lm-eval
lm-eval==0.4.2
git+https://github.com/huggingface/peft.git@6c44096c7b8d55a2ecf24be9bc68393467e1584a
Original file line number Diff line number Diff line change
@@ -1,4 +1,5 @@
neural-compressor
intel-extension-for-transformers==1.4.2
neural-compressor==2.6
transformers
accelerate
datasets >= 1.8.0
Expand Down
Loading

0 comments on commit 35b24f1

Please sign in to comment.