Tooling to run Llama Stack as a separate service.
Llama Stack Runner is a utility designed to run Llama Stack as a standalone service, enabling streamlined deployment, monitoring and management. This tool simplifies the process of integrating Llama Stack into existing infrastructure by providing a dedicated service environment. It offers a robust and efficient way to execute Llama Stack, ensuring optimal performance and reliability.
Export your OpenAI API key first:
export OPENAI_API_KEY="your_key"Then it is possible to run the Llama Stack or test the runner:
- run:
make run - test:
make test
curl localhost:8321/openapi.json | jq .curl localhost:8321/v1/models | jq ..
├── LICENSE
├── Makefile
├── pyproject.toml
├── run.yaml
├── test_client.py
└── uv.lock
Contributions are welcome! Here's how you can help:
- Fork the repository
- Clone your fork:
git clone https://github.com/lightspeed-core/llama-stack-runner.git - Create a new branch:
git checkout -b feature/your-feature - Commit your changes:
git commit -am 'Add some feature' - Push to your branch:
git push origin feature/your-feature - Open a pull request
Please ensure your code follows the project's style guidelines and includes tests where applicable.
This project is licensed under the Apache license.