You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Add IPEX-LLM inference jupyter notebook and instructions (#3310)
* Initial commit for IPEX LLM notebook for BF16
* update notebook with commands for quantization
* update READMEs with instructions, fix quantization commands
* minor updates
## 2.3 [Optional] Setup for Running Jupyter Notebooks
61
+
62
+
After setting up your docker or conda environment, you may follow these additional steps to setup and run Jupyter Notebooks. The port number can be changed.
63
+
64
+
### 2.3.1 Jupyter Notebooks for Docker-based Environments
1. Open up a web browser with the given URL and token.
92
+
2. Open the notebook.
93
+
3. Change your Jupyter Notebook kernel to IPEX-LLM.
94
+
4. Run all cells.
95
+
60
96
<br>
61
97
62
98
*Note*: In `env_setup.sh` script a `prompt.json` file is downloaded, which provides prompt samples with pre-defined input token lengths for benchmarking.
Copy file name to clipboardExpand all lines: examples/cpu/llm/inference/README.md
+4Lines changed: 4 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -114,6 +114,10 @@ python run.py --help # for more detailed usages
114
114
115
115
*Note:* You may need to log in your HuggingFace account to access the model files. Please refer to [HuggingFace login](https://huggingface.co/docs/huggingface_hub/quick-start#login).
116
116
117
+
**Alternatively, you can run the Jupyter Notebook to see ipex.llm with BF16 and various other quick start examples.**
118
+
119
+
Additional setup instructions for running the notebook can be found [here](../README.md#23-optional-setup-for-running-jupyter-notebooks).
120
+
117
121
## 2.1 Quick example for running Llama2-7b
118
122
119
123
### 2.1.1 To run generation task and benchmark performance
0 commit comments