You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/stable/cli/cli_api.md
+37-4Lines changed: 37 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -43,7 +43,9 @@ After setting up the ServerlessLLM cluster, you can use the commands listed belo
43
43
```
44
44
45
45
### sllm-cli deploy
46
-
Deploy a model using a configuration file or model name.
46
+
Deploy a model using a configuration file or model name, with options to overwrite default configurations. The configuration file requires minimal specifications, as sensible defaults are provided for advanced configuration options.
47
+
48
+
For more details on the advanced configuration options and their default values, please refer to the [Example Configuration File](#example-configuration-file-configjson) section.
47
49
48
50
##### Usage
49
51
```bash
@@ -52,18 +54,49 @@ sllm-cli deploy [OPTIONS]
52
54
53
55
##### Options
54
56
- `--model <model_name>`
55
-
- Model name to deploy with default configuration. The model name must be a huggingface pretrained model name. You can find the list of available models [here](https://huggingface.co/models).
57
+
- Model name to deploy with default configuration. The model name must be a Hugging Face pretrained model name. You can find the list of available models [here](https://huggingface.co/models).
56
58
57
59
- `--config <config_path>`
58
-
- Path to the JSON configuration file.
60
+
- Path to the JSON configuration file. The configuration file can be incomplete, and missing sections will be filled in by the default configuration.
59
61
60
-
##### Example
62
+
- `--backend <backend_name>`
63
+
- Overwrite the backend in the default configuration.
64
+
65
+
- `--num_gpus <number>`
66
+
- Overwrite the number of GPUs in the default configuration.
67
+
68
+
- `--target <number>`
69
+
- Overwrite the target concurrency in the default configuration.
70
+
71
+
- `--min_instances <number>`
72
+
- Overwrite the minimum instances in the default configuration.
73
+
74
+
- `--max_instances <number>`
75
+
- Overwrite the maximum instances in the default configuration.
76
+
77
+
##### Examples
78
+
Deploy using a model name with default configuration:
61
79
```bash
62
80
sllm-cli deploy --model facebook/opt-1.3b
81
+
```
82
+
83
+
Deploy using a configuration file:
84
+
```bash
63
85
sllm-cli deploy --config /path/to/config.json
64
86
```
65
87
88
+
Deploy using a model name and overwrite the backend:
Copy file name to clipboardExpand all lines: docs/stable/getting_started/docker_quickstart.md
+4-4Lines changed: 4 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -6,9 +6,9 @@ sidebar_position: 2
6
6
7
7
This guide will help you get started with the basics of using ServerlessLLM with Docker. Please make sure you have Docker installed on your system and have installed ServerlessLLM CLI following the [installation guide](./installation.md).
8
8
9
-
## Pre-requirements
9
+
## Pre-requisites
10
10
11
-
Ensure you have the following pre-requirements installed:
11
+
Ensure you have the following pre-requisites:
12
12
13
13
1.**GPUs**: Ensure you have at least 2 GPUs available. If more GPUs are provided, you can adjust the number of workers and the number of devices assigned to each worker.
14
14
2.**NVIDIA Docker Toolkit**: This allows Docker to use NVIDIA GPUs. You can find the installation guide [here](https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html).
0 commit comments