Hey Kedroids! :kedro:
(Apologies in advance for the long message but would really really appreciate a good discussion on below from the kedro community! π )
I have a usecase of deploying kedro pipelines using VertexAI SDK.
https://my.web.app/api/v1/some-task
RUN mkdir /usr/kedro WORKDIR /usr/kedro/ COPY . .
src/entrypoint.py
,from kedro.framework import cli import os os.chdir("/usr/kedro") cli.main()
cli.main()
picks up sys.argv to infer pipeline name and parameters so one could thatkedro run --pipeline <my_pipeline> --params=param_key1=value1,param_key2=2.0
Hi Abhishek, Based on my understanding here are some suggestions -
COPY ..
should work fine.whl
should be your optionfrom my_project.__main__ import main main([ "--pipeline", "<my_pipeline>", "--params", "param_key1=value1,param_key2=2.0" ])I would also wait for the community to respond if someone has tried this and have any recommendations
Thanks for the response π
On packaging as a whl file and submitting it to dataproc cluster, my main question is, if I do not include the kedro project folder in the dockerfile, then how would kedro find the conf
folder?
I guess the kedro project folder structure has to be brought in somehow so as to execute the project (cloning it or packaging it with the docker image)
And thanks for the 4th point π
Also, hopefully this discussion helps to guide towards creating a kedro gcp development / deployment guide. I can contribute too! π