Skip to content

Commit 3905449

Browse files
llama-server: update test models to new HF repos
1 parent 53352d0 commit 3905449

File tree

2 files changed

+6
-6
lines changed

2 files changed

+6
-6
lines changed

tools/server/tests/unit/test_router.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -104,8 +104,8 @@ def test_router_models_max_evicts_lru():
104104

105105
candidate_models = [
106106
"ggml-org/tinygemma3-GGUF:Q8_0",
107-
"ggml-org/models/tinyllamas/stories260K.gguf",
108-
"ggml-org/models/bert-bge-small/ggml-model-f16.gguf",
107+
"ggml-org/test-model-stories260K",
108+
"ggml-org/test-model-stories260K-infill",
109109
]
110110

111111
loaded_models: list[str] = []

tools/server/tests/utils.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -450,8 +450,8 @@ def load_all() -> None:
450450
@staticmethod
451451
def tinyllama2() -> ServerProcess:
452452
server = ServerProcess()
453-
server.model_hf_repo = "ggml-org/models"
454-
server.model_hf_file = "tinyllamas/stories260K.gguf"
453+
server.model_hf_repo = "ggml-org/test-model-stories260K"
454+
server.model_hf_file = None
455455
server.model_alias = "tinyllama-2"
456456
server.n_ctx = 512
457457
server.n_batch = 32
@@ -495,8 +495,8 @@ def bert_bge_small_with_fa() -> ServerProcess:
495495
def tinyllama_infill() -> ServerProcess:
496496
server = ServerProcess()
497497
server.offline = True # will be downloaded by load_all()
498-
server.model_hf_repo = "ggml-org/models"
499-
server.model_hf_file = "tinyllamas/stories260K-infill.gguf"
498+
server.model_hf_repo = "ggml-org/test-model-stories260K-infill"
499+
server.model_hf_file = None
500500
server.model_alias = "tinyllama-infill"
501501
server.n_ctx = 2048
502502
server.n_batch = 1024

0 commit comments

Comments
 (0)