@@ -33,6 +33,10 @@ def init_cli():
33
33
parser .add_argument ("--dryrun" ,
34
34
action = 'store_true' ,
35
35
help = "show container runtime command without executing it" )
36
+ parser .add_argument ("--runtime" ,
37
+ default = "llama.cpp" ,
38
+ choices = ["llama.cpp" , "vllm" ],
39
+ help = "specify the runtime to use (default: llama.cpp). Valid options are 'llama.cpp' and 'vllm'." )
36
40
parser .add_argument ("--nocontainer" ,
37
41
default = not use_container (),
38
42
action = 'store_true' ,
@@ -163,16 +167,20 @@ def list_files_by_modification():
163
167
reverse = True )
164
168
165
169
166
- def list_parser (subparsers ):
167
- parser = subparsers .add_parser (
168
- 'list' , help = 'List all downloaded AI Models' )
170
+ def add_list_parser (subparsers , name , func ):
171
+ parser = subparsers .add_parser (name , help = 'List all downloaded AI Models' )
169
172
parser .add_argument ("-n" , "--noheading" , dest = "noheading" ,
170
173
action = 'store_true' ,
171
174
help = "do not display heading" )
172
175
parser .add_argument ("--json" , dest = "json" ,
173
176
action = 'store_true' ,
174
177
help = "print using json" )
175
- parser .set_defaults (func = list_cli )
178
+ parser .set_defaults (func = func )
179
+
180
+
181
+ def list_parser (subparsers ):
182
+ add_list_parser (subparsers , 'list' , list_cli )
183
+ add_list_parser (subparsers , 'ls' , list_cli )
176
184
177
185
178
186
def list_cli (args ):
@@ -247,15 +255,9 @@ def run_cli(args):
247
255
248
256
249
257
def serve_parser (subparsers ):
250
- port = "8080"
251
- host = os .getenv ('RAMALAMA_HOST' , port )
252
- split = host .rsplit (':' , 1 )
253
- if len (split ) > 1 :
254
- port = split [1 ]
255
-
256
258
parser = subparsers .add_parser (
257
- 'serve' , help = 'Serve RESTAPI on specified AI Model' )
258
- parser .add_argument ("--port" , default = port ,
259
+ 'serve' , help = 'Serve REST API on specified AI Model' )
260
+ parser .add_argument ("--port" , default = "8080" ,
259
261
help = "port for AI Model server to listen on" )
260
262
parser .add_argument ('model' ) # positional argument
261
263
parser .set_defaults (func = serve_cli )
@@ -299,9 +301,6 @@ def run_container(args):
299
301
wd = target
300
302
break
301
303
302
- port = "8080"
303
- host = os .getenv ('RAMALAMA_HOST' , port )
304
-
305
304
conman_args = [conman , "run" ,
306
305
"--rm" ,
307
306
"-it" ,
@@ -310,9 +309,8 @@ def run_container(args):
310
309
f"-v{ home } :{ home } " ,
311
310
"-v/tmp:/tmp" ,
312
311
f"-v{ sys .argv [0 ]} :/usr/bin/ramalama:ro" ,
313
- "-e" , "RAMALAMA_HOST" ,
314
312
"-e" , "RAMALAMA_TRANSPORT" ,
315
- "-p" , f"{ host } :{ port } " ,
313
+ "-p" , f"{ args . port } :{ args . port } " ,
316
314
f"-v{ wd } :/usr/share/ramalama/ramalama:ro" ]
317
315
if os .path .exists ("/dev/dri" ):
318
316
conman_args += ["--device" , "/dev/dri" ]
0 commit comments