[CLI] bake run.yaml file inside docker container #122
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Changes
Motivation: We should not need to install llama CLI & run
llama stack configure
/llama stack run
outside of docker containers. Downloading docker image should be sufficient to start Llama Stack server.[RFC] Only use docker commands for running w/ docker container. New developer flow for interaction with docker image.
Developer Flow
podman run --network host -it -p 5000:5000 -v ~/.llama:/root/.llama --gpus=all llamastack-local-gpu --port 5000
podman run --network host -it \ -p 5000:5000 \ -v path/to/run.yaml:/app/run.yaml \ -v ~/.llama:/root/.llama \ --gpus=all \ llamastack-d1 \ --port 5000 --config /app/config.yaml \
where
path/to/run.yaml
is absolute path to config outside container.podman run --network host -it --entrypoint "/bin/bash" llamastack-d0 $ (inside container) llama stack configure llamastack-build.yaml ... Run configuration saved to d0-run.yaml
podman run --network host -it -p 5001:5001 -v ~/.llama:/root/.llama --gpus=all llamastack-d0 --port 5001 --config /app/d0-run.yaml
Distribution Owner: Building Docker