File tree Expand file tree Collapse file tree 2 files changed +6
-6
lines changed
Expand file tree Collapse file tree 2 files changed +6
-6
lines changed Original file line number Diff line number Diff line change @@ -104,8 +104,8 @@ def test_router_models_max_evicts_lru():
104104
105105 candidate_models = [
106106 "ggml-org/tinygemma3-GGUF:Q8_0" ,
107- "ggml-org/models/tinyllamas/ stories260K.gguf " ,
108- "ggml-org/models/bert-bge-small/ggml- model-f16.gguf " ,
107+ "ggml-org/test-model- stories260K" ,
108+ "ggml-org/test- model-stories260K-infill " ,
109109 ]
110110
111111 loaded_models : list [str ] = []
Original file line number Diff line number Diff line change @@ -450,8 +450,8 @@ def load_all() -> None:
450450 @staticmethod
451451 def tinyllama2 () -> ServerProcess :
452452 server = ServerProcess ()
453- server .model_hf_repo = "ggml-org/models "
454- server .model_hf_file = "tinyllamas/stories260K.gguf"
453+ server .model_hf_repo = "ggml-org/test-model-stories260K "
454+ server .model_hf_file = None
455455 server .model_alias = "tinyllama-2"
456456 server .n_ctx = 512
457457 server .n_batch = 32
@@ -495,8 +495,8 @@ def bert_bge_small_with_fa() -> ServerProcess:
495495 def tinyllama_infill () -> ServerProcess :
496496 server = ServerProcess ()
497497 server .offline = True # will be downloaded by load_all()
498- server .model_hf_repo = "ggml-org/models "
499- server .model_hf_file = "tinyllamas/stories260K-infill.gguf"
498+ server .model_hf_repo = "ggml-org/test-model-stories260K-infill "
499+ server .model_hf_file = None
500500 server .model_alias = "tinyllama-infill"
501501 server .n_ctx = 2048
502502 server .n_batch = 1024
You can’t perform that action at this time.
0 commit comments