1+ # ruff: noqa: D301, D411
12"""Command line interface for Vector Inference.
23
34This module provides the command-line interface for interacting with Vector
@@ -135,10 +136,12 @@ def launch(
135136) -> None :
136137 """Launch a model on the cluster.
137138
139+ \b
138140 Parameters
139141 ----------
140142 model_name : str
141143 Name of the model to launch
144+ \f
142145 **cli_kwargs : dict
143146 Additional launch options including:
144147 - model_family : str, optional
@@ -229,10 +232,12 @@ def batch_launch(
229232) -> None :
230233 """Launch multiple models in a batch.
231234
235+ \b
232236 Parameters
233237 ----------
234238 model_names : tuple[str, ...]
235239 Names of the models to launch
240+ \f
236241 batch_config : str
237242 Model configuration for batch launch
238243 json_mode : bool, default=False
@@ -276,10 +281,12 @@ def batch_launch(
276281def status (slurm_job_id : str , json_mode : bool = False ) -> None :
277282 """Get the status of a running model on the cluster.
278283
284+ \b
279285 Parameters
280286 ----------
281287 slurm_job_id : str
282288 ID of the SLURM job to check
289+ \f
283290 json_mode : bool, default=False
284291 Whether to output in JSON format
285292
@@ -311,10 +318,12 @@ def status(slurm_job_id: str, json_mode: bool = False) -> None:
311318def shutdown (slurm_job_id : str ) -> None :
312319 """Shutdown a running model on the cluster.
313320
321+ \b
314322 Parameters
315323 ----------
316324 slurm_job_id : str
317325 ID of the SLURM job to shut down
326+ \f
318327
319328 Raises
320329 ------
@@ -339,10 +348,12 @@ def shutdown(slurm_job_id: str) -> None:
339348def list_models (model_name : Optional [str ] = None , json_mode : bool = False ) -> None :
340349 """List all available models, or get default setup of a specific model.
341350
351+ \b
342352 Parameters
343353 ----------
344354 model_name : str, optional
345355 Name of specific model to get information for
356+ \f
346357 json_mode : bool, default=False
347358 Whether to output in JSON format
348359
@@ -372,10 +383,12 @@ def list_models(model_name: Optional[str] = None, json_mode: bool = False) -> No
372383def metrics (slurm_job_id : str ) -> None :
373384 """Stream real-time performance metrics from the model endpoint.
374385
386+ \b
375387 Parameters
376388 ----------
377389 slurm_job_id : str
378390 ID of the SLURM job to monitor
391+ \f
379392
380393 Raises
381394 ------
@@ -442,6 +455,8 @@ def cleanup_logs_cli(
442455) -> None :
443456 """Clean up log files based on optional filters.
444457
458+ \f
459+
445460 Parameters
446461 ----------
447462 log_dir : str or Path, optional
@@ -456,7 +471,7 @@ def cleanup_logs_cli(
456471 If provided, only delete logs with job ID less than this value.
457472 dry_run : bool
458473 If True, return matching files without deleting them.
459- """
474+ """ # NOQA: D301, the \f prevents click from printing options twice.
460475 try :
461476 client = VecInfClient ()
462477 matched = client .cleanup_logs (
0 commit comments