bitllm / app.py
ar08's picture
Update app.py
9fa612b verified
raw
history blame
418 Bytes
import subprocess
import os
# Prepare the command
command = [
"python",
"run_inference_server.py",
"-m", "./models/Llama3-8B-1.58-100B-tokens-TQ2_0.gguf",
"--host", "0.0.0.0",
"--port", "7860"
]
# Run it as a subprocess
try:
subprocess.run("ls",check=True)
subprocess.run(command, check=True)
except subprocess.CalledProcessError as e:
print(f"Error running inference server: {e}")