Search 61 Results

Can't Find What You Are Looking For? Sign In Now
You are currently not logged in and some search results only show up if you login.

Tools for Researchers

, data storage, Caligari Web and database servers Accounts for Researchers - For general research computing, data storage, and Discovery cluster accounts Software Resources - Applications in

Single-Core Job Example

sample script are exampled in a comment line that precedes the directive. The full list of available directives is explained in the man page for the sbatch command which is available on discovery

Scheduling Jobs

The Batch System The batch system used on Discovery is Slurm. Users login, via ssh,  to one of the submit nodes and submit jobs to be run on the compute nodes by writing a script file that

GPU Job Example

#!/bin/bash # Name of the job #SBATCH --job-name=gpu_job # Number of compute nodes #SBATCH --nodes=1 # Number of cores, in this case one #SBATCH --ntasks-per-node=1 # Request the GPU

Multi-Core Job Example

Below is an example script which will submit for 4 cores on a single compute node. Feel free to copy and paste it as a job template.   #!/bin/bash #SBATCH --job-name=multicore_job #SBATCH

Leaving Dartmouth - Research Computing resources

When you leave Dartmouth you will no longer be able to login to the multi-user Linux systems such as Polaris, Andes, and the Discovery cluster.    You will also lose access to DartFS network

When do I need to use the VPN client?

solution has enough capacity by only using it when you need to. Requires VPN DartFS-Networked Attached Storage used in labs, HPC and web applications Discovery (HPC) MyFiles/OurFiles (File

Job array example

JOB ARRAYS Job arrays are multiple jobs to be executed with identical or related parameters. Job arrays are submitted with -a <indices> or --array=<indices>. The indices specification identifies

Graduating Students - Research Computing Resources

Discovery cluster You will also lose access to DartFS network storage. Deadline Approximately 60 days after graduating, any DartFS storage that is in your name will be deleted. After 60

User commands

Option MOAB Torque command (mksub) Slurm command (sbatch) Job Submission mksub <file name> sbatch <file name> Job Deletion qdel / mjobctl -c scancel

Job directives

Option MOAB Torque command (mksub) Slurm command (sbatch) Job Name -N <name> --job-name=<name> Account -a <account> --account=<account

Research Computing Bill of Rights

Performance Computing (HPC), GIS, data visualization, statistical analysis, software development and more access to HPC resources (Discovery, Andes, and Polaris) DartFS: personal and shared data storage

Recover files/folders from a DartFS snapshot

From a Mac or Linux The easiest way to recover these files is to: Log into Andes/Polaris/Discovery from a Terminal application Go to the hidden .snapshot folder – note, you cannot see

Important update regarding our service charges collection policy, effective from July 1, 2023.

financial year or at the beginning of service. Specifically, data storage and compute charges, including but not limited to DartFS, virtual servers, containers, and Discovery shares, will be

Example: Interactive job (srun)

An interactive job is launched on a compute node and provides you with a command line prompt. Interactive jobs are useful when debugging or interacting with an application. You will use thesruncommand