You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardexpand all lines: site-src/guides/index.md
+1-3
Original file line number
Diff line number
Diff line change
@@ -18,7 +18,7 @@ This quickstart guide is intended for engineers familiar with k8s and model serv
18
18
Requirements: a Hugging Face access token that grants access to the model [meta-llama/Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b-hf).
19
19
20
20
1. CPU-based model server (not using GPUs).
21
-
Requirements: a Hugging Face access token that grants access to the model [Qwen/Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct).
21
+
The sample usees the model [Qwen/Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct).
22
22
23
23
Choose one of these options and follow the steps below. Please do not deploy both, as the deployments have the same name and will override each other.
24
24
@@ -34,10 +34,8 @@ This quickstart guide is intended for engineers familiar with k8s and model serv
34
34
35
35
#### CPU-Based Model Server
36
36
37
-
Create a Hugging Face secret to download the model [Qwen/Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct). Ensure that the token grants access to this model.
38
37
Deploy a sample vLLM deployment with the proper protocol to work with the LLM Instance Gateway.
39
38
```bash
40
-
kubectl create secret generic hf-token --from-literal=token=$HF_TOKEN# Your Hugging Face Token with access to Qwen
0 commit comments