You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+11-2Lines changed: 11 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -38,6 +38,15 @@ There are 5 possible states:
38
38
39
39
Note that the base URL is only available when model is in `READY` state, and if you've changed the Slurm log directory path, you also need to specify it when using the `status` command.
40
40
41
+
Once your server is ready, you can check performance metrics by providing the Slurm job ID to the `metrics` command:
42
+
```bash
43
+
vec-inf metrics
44
+
```
45
+
46
+
And you will see the performance metrics streamed to your console, note that the metrics are updated with a 10-second interval.
Copy file name to clipboardExpand all lines: vec_inf/README.md
+1Lines changed: 1 addition & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -2,6 +2,7 @@
2
2
3
3
*`launch`: Specify a model family and other optional parameters to launch an OpenAI compatible inference server, `--json-mode` supported. Check [`here`](./models/README.md) for complete list of available options.
4
4
*`list`: List all available model names, `--json-mode` supported.
5
+
*`metrics`: Streams performance metrics to the console.
5
6
*`status`: Check the model status by providing its Slurm job ID, `--json-mode` supported.
6
7
*`shutdown`: Shutdown a model by providing its Slurm job ID.
0 commit comments