This guide covers how to integrate the CAD Job Manager with the most popular job schedulers. After following these steps, running a simulation on a cluster will be just as seamless as running it on your local computer. You will be able to just click “Run” and the job will automatically be submitted to the job schedulers queue (the file will be transferred if necessary), the Lumerical Job Manager will periodically update with the simulations current state and progress, and upon completion, the results will be automatically loaded in your current CAD session.
See also: Job scheduler submission scripts (SGE, Slurm, Torque) for additional details and sample submission scripts for the supported job schedulers.
The job manager options: Quit, Quit & Save, and Force Quit, are currently not supported.
- Lumerical Products 2020a R3 (or newer)
- Ansys Lumerical 2023 R2.2 (or newer) for user based configuration
- Cluster with either: Slurm, Torque, LSF, SGE, X11-Display, Lumerical products installed with license server configured. If using AWS ParallelCluster, see AWS-ParallelCluster documentation for details.
Add/edit a new resource, set the capacity to either 0 (inf) if the queue is already configured to handle the number of available licenses, or to the maximum number of available engine licenses on the license server:
- Edit the advanced resource settings, select your job scheduler from the "Job Launching Presets":
- The preset is automatically populated with the defaults for AWS ParallelCluster.
- You may need to modify the submission command with the proper number of nodes and processes per node, as well as update the submission script to set a Lumerical compute environment appropriate for your cluster.
(Beta) Launching Lumerical Jobs on a Remote Cluster From Your Local Computer
If you would like to launch jobs from your local computer for a more seamless experience, you can configure Lumerical on your local computer following the same steps as above and enable ssh by setting up the file job_scheduler_input.json. A template can be found in the Lumerical installation folder:
The configuration with a json file was introduced with the 2023 R2.2 release. With previous versions, you will have to modify the .py file corresponding to your job scheduler.
Copy this file in your home folder:
- On Linux
- On Windows
Then edit the file with your job scheduler settings:
Your user name on the master node. If left empty, the user name will be dynamically assigned using
|use_ssh||If set to 1, ssh* is used to run the submission command on the master node.|
|use_scp||If set to 1, the simulation files will be copied to the master node using scp*.|
|cluster_cwd||Shared folder where the simulation files will be copied when use_scp=1.|
|master_node_ip||IP address or hostname of the master node used for the remote connection and job submission.|
|ssh_key||Location of your private ssh key for passwordless connection.|
Can be used to translate the paths when there is a shared file system on Windows to a Linux cluster. Use unix-style path delimiters '/':
"path_translation": ["X:/Working Dir/", "/share/working/dir/"]
To submit jobs from your local computer to a remote Linux cluster, set user_ssh and use_scp to 1. Set master_node_ip to the This method will copy your simulation file to the remote server using scp*, launch the job with ssh*, and copy all the generated files back to your local computer at the end.
You can now run any Lumerical simulation (single, sweeps, optimizations, etc.) directly from the CAD's Job Manager.