-
Following examples from here we have build a logic around deploying a single workflow through llama deploy. We have 3 key elements all being invoked from a bash script with 10 seconds of sleep time between:
This was working fine up till recently, now on the first start it works fine, however if I decide to kill the workflow and redeploy it, then I get: |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 13 replies
-
@hristogg have you considered using the apiserver, or even deploy the llama_deploy official Docker image directly to Cloud Run? Example: Step 1: run the LlamaDeploy container, mounting the folder containing your workflow: At this point the control plane and the message queue are up and running inside the docker container Step 2: define your deployment in a
See the quick start example for more details about this Step 3: deploy your workflow
At this point the workflow is deployed and waiting for requests, no need to add a custom Flask app in front of it Step 4: run your deployed workflow
Let me know if this is helpful for your use case and I can follow up with more details about how you could pack all of this into a Cloud Run instance, I'm working on something similar right now! |
Beta Was this translation helpful? Give feedback.
@hristogg have you considered using the apiserver, or even deploy the llama_deploy official Docker image directly to Cloud Run?
Example:
Step 1: run the LlamaDeploy container, mounting the folder containing your workflow:
docker run -p 4501:4501 -v ./my-app:/opt/my-app -w /opt/my-app llamaindex/llama-deploy:main
At this point the control plane and the message queue are up and running inside the docker container
Step 2: define your deployment in a
deployment.yml
file, something like