Skip to content

Commit

Permalink
fixed llamafile args quoting
Browse files Browse the repository at this point in the history
  • Loading branch information
sariola committed Oct 8, 2024
1 parent 1b56a93 commit 2d7f16b
Showing 1 changed file with 3 additions and 1 deletion.
4 changes: 3 additions & 1 deletion flow_judge/models/llamafile.py
Original file line number Diff line number Diff line change
Expand Up @@ -240,7 +240,7 @@ def start_llamafile_server(self):
f"--threads {self.generation_params.get('thread_count', os.cpu_count() or 1)} " \
f"--nobrowser -b {self.generation_params.get('batch_size', 32)} " \
f"--parallel {self.generation_params.get('max_concurrent_requests', 1)} " \
f"--cont-batching'"
f"--cont-batching"

if self.disable_kv_offload:
command += " -nkvo"
Expand All @@ -262,6 +262,8 @@ def start_llamafile_server(self):
command += f" --{key} {value}"
logging.info(f"Additional server argument added: --{key} {value}")

Check warning on line 263 in flow_judge/models/llamafile.py

View check run for this annotation

Codecov / codecov/patch

flow_judge/models/llamafile.py#L261-L263

Added lines #L261 - L263 were not covered by tests

command += "'"

Check warning on line 265 in flow_judge/models/llamafile.py

View check run for this annotation

Codecov / codecov/patch

flow_judge/models/llamafile.py#L265

Added line #L265 was not covered by tests

logging.info(f"Starting llamafile server with command: {command}")

Check warning on line 267 in flow_judge/models/llamafile.py

View check run for this annotation

Codecov / codecov/patch

flow_judge/models/llamafile.py#L267

Added line #L267 was not covered by tests

def log_output(pipe, log_func):
Expand Down

0 comments on commit 2d7f16b

Please sign in to comment.