$ cryosparcm cli 'get_scheduler_targets()'
[{'cache_path': '/scratch', 'cache_quota_mb': None, 'cache_reserve_mb': 10000, 'custom_var_names': [], 'custom_vars': {}, 'desc': None, 'hostname': 'cryosparc', 'lane': 'cryosparc', 'name': 'cryosparc', 'qdel_cmd_tpl': 'scancel {{ cluster_job_id }}', 'qinfo_cmd_tpl': "sinfo --format='%.8N %.6D %.10P %.6T %.14C %.5c %.6z %.7m %.7G %.9d %20E'", 'qstat_cmd_tpl': 'squeue -j {{ cluster_job_id }}', 'qstat_code_cmd_tpl': None, 'qsub_cmd_tpl': 'sbatch {{ script_path_abs }}', 'script_tpl': '#!/bin/bash\n#SBATCH --job-name {{ project_uid }}_{{ job_uid }}\n#SBATCH --partition=cryosparc\n#SBATCH --output={{ job_log_path_abs }}\n#SBATCH --error={{ job_log_path_abs }}\n#SBATCH --nodes=1\n#SBATCH --mem={{ (ram_gb*1000)|int }}M\n#SBATCH --ntasks-per-node=1\n#SBATCH --cpus-per-task={{ num_cpu }}\n#SBATCH --gres=gpu:{{ num_gpu }}\n#SBATCH --gres-flags=enforce-binding\n##SBATCH --exclusive\n\nsrun {{ run_cmd }}\n\n\n', 'send_cmd_tpl': '{{ command }}', 'title': 'cryosparc', 'tpl_vars': ['command', 'run_cmd', 'num_gpu', 'project_uid', 'ram_gb', 'job_uid', 'num_cpu', 'job_log_path_abs', 'cluster_job_id'], 'type': 'cluster', 'worker_bin_path': '/home_local/hpc/cryosparc2/cryosparc2_worker/bin/cryosparcw'}, {'cache_path': '/scratch', 'cache_quota_mb': None, 'cache_reserve_mb': 10000, 'custom_var_names': [], 'custom_vars': {}, 'desc': None, 'hostname': 'cryosparc1', 'lane': 'cryosparc1', 'name': 'cryosparc1', 'qdel_cmd_tpl': 'scancel {{ cluster_job_id }}', 'qinfo_cmd_tpl': "sinfo --format='%.8N %.6D %.10P %.6T %.14C %.5c %.6z %.7m %.7G %.9d %20E'", 'qstat_cmd_tpl': 'squeue -j {{ cluster_job_id }}', 'qstat_code_cmd_tpl': None, 'qsub_cmd_tpl': 'sbatch {{ script_path_abs }}', 'script_tpl': '#!/bin/bash\n#SBATCH --job-name {{ project_uid }}_{{ job_uid }}\n#SBATCH --partition=cryosparc1\n#SBATCH --output={{ job_log_path_abs }}\n#SBATCH --error={{ job_log_path_abs }}\n#SBATCH --nodes=1\n#SBATCH --mem={{ (ram_gb*1000)|int }}M\n#SBATCH --ntasks-per-node=1\n#SBATCH --cpus-per-task={{ num_cpu }}\n#SBATCH --gres=gpu:{{ num_gpu }}\n#SBATCH --gres-flags=enforce-binding\n##SBATCH --exclusive\n\nsrun {{ run_cmd }}\n\n\n\n', 'send_cmd_tpl': '{{ command }}', 'title': 'cryosparc1', 'tpl_vars': ['command', 'run_cmd', 'num_gpu', 'project_uid', 'ram_gb', 'job_uid', 'num_cpu', 'job_log_path_abs', 'cluster_job_id'], 'type': 'cluster', 'worker_bin_path': '/home_local/hpc/cryosparc2/cryosparc2_worker/bin/cryosparcw'}, {'cache_path': '/scratch', 'cache_quota_mb': None, 'cache_reserve_mb': 10000, 'custom_var_names': [], 'custom_vars': {}, 'desc': None, 'hostname': 'cryosparc2', 'lane': 'cryosparc2', 'name': 'cryosparc2', 'qdel_cmd_tpl': 'scancel {{ cluster_job_id }}', 'qinfo_cmd_tpl': "sinfo --format='%.8N %.6D %.10P %.6T %.14C %.5c %.6z %.7m %.7G %.9d %20E'", 'qstat_cmd_tpl': 'squeue -j {{ cluster_job_id }}', 'qstat_code_cmd_tpl': None, 'qsub_cmd_tpl': 'sbatch {{ script_path_abs }}', 'script_tpl': '#!/bin/bash\n#SBATCH --job-name {{ project_uid }}_{{ job_uid }}\n#SBATCH --partition=cryosparc2\n#SBATCH --output={{ job_log_path_abs }}\n#SBATCH --error={{ job_log_path_abs }}\n#SBATCH --nodes=1\n#SBATCH --mem={{ (ram_gb*1000)|int }}M\n#SBATCH --ntasks-per-node=1\n#SBATCH --cpus-per-task={{ num_cpu }}\n#SBATCH --gres=gpu:{{ num_gpu }}\n#SBATCH --gres-flags=enforce-binding\n##SBATCH --exclusive\n\nsrun {{ run_cmd }}\n\n\n\n', 'send_cmd_tpl': '{{ command }}', 'title': 'cryosparc2', 'tpl_vars': ['command', 'run_cmd', 'num_gpu', 'project_uid', 'ram_gb', 'job_uid', 'num_cpu', 'job_log_path_abs', 'cluster_job_id'], 'type': 'cluster', 'worker_bin_path': '/home_local/hpc/cryosparc2/cryosparc2_worker/bin/cryosparcw'}, {'cache_path': '/scratch', 'cache_quota_mb': None, 'cache_reserve_mb': 10000, 'custom_var_names': [], 'custom_vars': {}, 'desc': None, 'hostname': 'cryosparc3', 'lane': 'cryosparc3', 'name': 'cryosparc3', 'qdel_cmd_tpl': 'scancel {{ cluster_job_id }}', 'qinfo_cmd_tpl': "sinfo --format='%.8N %.6D %.10P %.6T %.14C %.5c %.6z %.7m %.7G %.9d %20E'", 'qstat_cmd_tpl': 'squeue -j {{ cluster_job_id }}', 'qstat_code_cmd_tpl': None, 'qsub_cmd_tpl': 'sbatch {{ script_path_abs }}', 'script_tpl': '#!/bin/bash\n#SBATCH --job-name {{ project_uid }}_{{ job_uid }}\n#SBATCH --partition=cryosparc3\n#SBATCH --output={{ job_log_path_abs }}\n#SBATCH --error={{ job_log_path_abs }}\n#SBATCH --nodes=1\n#SBATCH --mem={{ (ram_gb*1000)|int }}M\n#SBATCH --ntasks-per-node=1\n#SBATCH --cpus-per-task={{ num_cpu }}\n#SBATCH --gres=gpu:{{ num_gpu }}\n#SBATCH --gres-flags=enforce-binding\n##SBATCH --exclusive\n\nsrun {{ run_cmd }}\n\n\n\n', 'send_cmd_tpl': '{{ command }}', 'title': 'cryosparc3', 'tpl_vars': ['command', 'run_cmd', 'num_gpu', 'project_uid', 'ram_gb', 'job_uid', 'num_cpu', 'job_log_path_abs', 'cluster_job_id'], 'type': 'cluster', 'worker_bin_path': '/home_local/hpc/cryosparc2/cryosparc2_worker/bin/cryosparcw'}, {'cache_path': '/scratch', 'cache_quota_mb': None, 'cache_reserve_mb': 10000, 'custom_var_names': [], 'desc': None, 'hostname': 'cryosparc4', 'lane': 'cryosparc4', 'name': 'cryosparc4', 'qdel_cmd_tpl': 'scancel {{ cluster_job_id }}', 'qinfo_cmd_tpl': "sinfo --format='%.8N %.6D %.10P %.6T %.14C %.5c %.6z %.7m %.7G %.9d %20E'", 'qstat_cmd_tpl': 'squeue -j {{ cluster_job_id }}', 'qsub_cmd_tpl': 'sbatch {{ script_path_abs }}', 'script_tpl': '#!/bin/bash\n#SBATCH --job-name cryosparc_{{ project_uid }}_{{ job_uid }}\n#SBATCH --partition=cryosparc4\n#SBATCH --output={{ job_log_path_abs }}\n#SBATCH --error={{ job_log_path_abs }}\n#SBATCH --nodes=1\n#SBATCH --mem={{ (ram_gb*1000)|int }}M\n#SBATCH --ntasks-per-node=1\n#SBATCH --cpus-per-task={{ num_cpu }}\n#SBATCH --gres=gpu:{{ num_gpu }}\n#SBATCH --gres-flags=enforce-binding\n##SBATCH --exclusive\n\nsrun {{ run_cmd }}\n\n\n', 'send_cmd_tpl': '{{ command }}', 'title': 'cryosparc4', 'tpl_vars': ['command', 'run_cmd', 'num_gpu', 'project_uid', 'ram_gb', 'job_uid', 'num_cpu', 'job_log_path_abs', 'cluster_job_id'], 'type': 'cluster', 'worker_bin_path': '/home_local/hpc/cryosparc2/cryosparc2_worker/bin/cryosparcw'}]