Skip to main content

Uploading Job to Scheduler

Compile and upload all jobs in the project by using this command:

$ optimus scheduler upload-all

Note: add --config flag if you are not in the same directory with your client configuration (optimus.yaml).

This command will compile all of the jobs in the project to Airflow DAG files and will store the result to the path that has been set as STORAGE_PATH in the project configuration. Do note that STORAGE secret might be needed if the storage requires a credential.

Once you have the DAG files in the storage, you can sync the files to Airflow as you’d like.