Episode 8: Autopilot
Scheduling Compute in the Cloud.
This example revisits Episode 06-statistics-redux: Computing in the
Cloud. Wi th Metaflow, you don't need to make any code changes to
schedule your flow in the cloud. In this example, we will schedule the stats.py
workflow using the argo-workflows create command-line argument. This instructs
Metaflow to schedule your flow on Argo
Workflows without changing any code. You
can execute your flow on Argo Workflows by using the argo-workflows trigger
command-line argument. You can use a notebook to set up a simple dashboard to monitor
all of your Metaflow flows.
You can find the tutorial code on GitHub
Showcasing:
argo-workflows createecommand-line optionargo-workflows triggercommand-line option- Accessing data locally or remotely through the Metaflow Client API
Before playing this episode:
python -m pip install notebookpython -m pip install plotly- This tutorial requires access to compute and storage resources on AWS, which can be configured by
To play this episode:
cd metaflow-tutorialspython 02-statistics/stats.py argo-workflows create --max-workers 4python 02-statistics/stats.py argo-workflows triggerjupyter-notebook 08-autopilot/autopilot.ipynb- Open autopilot.ipynb in your notebook