You can schedule Domino Jobs in advance and set them to execute on a regular cadence. These can be useful when you have a data source that is updated regularly.
-
Go to a project.
-
From the navigation pane, click Scheduled Jobs.
-
Click New Scheduled Job.
-
In the Create a Scheduled Job page, define the Job:
- Scheduled Job Name
-
Enter the name of the Job. The Jobs Dashboard lists each Job by this name.
- File Name
-
Enter the name of the file to execute. Include any optional arguments to pass to the file.
- Hardware tier
-
Select the hardware tier used by the Job.
- Environment
-
Select the compute environment used by the Job.
- Data
-
Click to expand the section to see the Datasets configuration used by the Job.
-
Select and define the compute cluster.
The configuration options for each cluster type are explained here:
-
Set up the schedule.
- Use custom expression
-
Enter a custom Quartz CronTrigger expression. For example, if you want to run the Job on the 5th minute of every day, enter the following:
`0 5 * ? * *`
NoteTo learn more about these expressions, see CronTrigger expressions. - Repeat every
-
Set the frequency at which you want the Job to repeat.
- Run sequentially
-
When you set a Job to run sequentially, the scheduler always waits for the last Job it started to complete before starting the next one. For example, if you set up a scheduled Job to run once per hour, and one of the Jobs launched by the scheduler takes 90 minutes to complete, the next hourly Job does not start until the previous one has finished. Otherwise, multiple Jobs from this scheduler are allowed to run simultaneously. The scheduler does not wait for the previous Job to finish if it’s still running. This mode should be used when your Job doesn’t depend on output from the previous Job.
-
Set up any additional actions:
- Notify emails
-
Enter a list of email addresses to notify when the Job is completed.
- Update Model API
-
If a Model API has been published from the project, the selected Model API will be republished after the Job has been completed. Use this for retraining and updating a Model API regularly.
-
Automate complex pipelines with Apache Airflow or Kubeflow.