To request a priority account or a condo, use the account form on the CCV homepage. For more information on resources available to priority accounts and costs, visit the CCV Rates page.
What username and password should I be using?
If you are at Brown and have requested a regular CCV account, your Oscar login can be authenticated using your Brown credentials itself, i.e. the same username and password that you use to login to any Brown service such as "canvas".
If you are an external user, you will have to get a sponsored ID at Brown through the department with which you are associated, before requesting an account on Oscar. Once you have the sponsored ID at Brown, you can on Oscar and use your Brown username and password to login.
Changing Passwords
Oscar users should use their Brown passwords to log into Oscar. Users should change their Brown passwords at .
Exploratory Account
Exploratory accounts are available to all members of the Brown community for free.
See thefor detailed description of the resources
Jobs are submitted to the batch partition. See the page for available hardware
Priority Accounts
The following accounts are billed quarterly and offer more computational resources than the exploratory accounts. See thefor pricing and detailed description of the resources
HPC Priority
Intended for users running CPU-intensive jobs. These offer more CPU and memory resources than an exploratory account
Two types of accounts:
HPC Priority
Standard GPU Priority
Intended for users running GPU intensive jobs. These accounts offer fewer CPU and memory resources but more GPU resources than an exploratory account.
Two types of accounts:
Standard GPU Priority
High End GPU Priority
Intended for GPU jobs required high-end gpus. These offer the same number of CPUS as Standard GPU priority accounts
High end GPUS like A40, v100 and a6000 are available
See thefor pricing and detailed description of the resources
Large Memory Priority
Intended for jobs requiring large amounts of memory.
These accounts offer 2TB of memory and twice the wall-time of exploratory accounts.
See thefor pricing and detailed description of the resources
Condo
PIs who purchase hardware (compute nodes) for the CCV machine get a Condo account. Condo account users have the highest priority on the number of cores equivalent to the hardware they purchased. Condo accounts last for five years and give their owners access to 25% more CPU cores than they purchase for the first three years of their lifespan. GPU resources do not decrease over the lifetime of the condo.
Investigators may also purchase condos to grant access to computing resources for others working with them. After a condo is purchased, they can have users request to join the condo group through the "Request Access to Existing Condo" option on the on the CCV homepage.
Oscar - Brown University's Cluster
Oscar - Ocean State Center for Advanced Resources - is Brown University's high performance computing cluster for both research and classes. Oscar is maintained and supported by.
Please contact if there are any questions on Oscar.
Accounts
If you do not have an Oscar account, you can request one by clicking the following link
HPC Priority+ (Twice the resources of HPC Priority)
See the CCV Rates page for pricing and detailed description of the resources.
Jobs are submitted to the batch partition. See the System Hardware page for available hardware
Standard GPU Priority+ (Twice the resources of Standard GPU Priority)
See the CCV Rates page for pricing and detailed description of the resources.
Jobs are submitted to the gpu partition. See the System Hardware page for available GPU hardware
Jobs are submitted to the gpu-he partition. See the System Hardware page for available GPU hardware
Jobs are submitted to the bigmem partition. See the System Hardware page for available hardware
Individuals external to Brown can get access to Oscar by having a sponsored Brown account. Please work with your department to get sponsored Brown accounts for any external collaborators.
Authorized users must comply with the following Brown University policies:
Users can run their computing-intensive and/or long runtime jobs/program in Oscar to take advantage of high performance computing resources there, as highlighted below:
Users should not run computations or simulations on the login nodes, because they are shared with other users. You can use the login nodes to compile your codes, manage files, and launch jobs on the compute nodes.
To allow users sharing access to Oscar, there are limits on the maximum number of pending and running jobs a user account may have/submit:
CCV Staff install software upon user requests or help users on software installation
Storage
Oscar has 3.27PB storage from the General Parallel File System (GPFS) of IBM, which provides high performance access of storage. Users have Home, Scratch, and Data directories as their storage with quota in Oscar. Please refer to the details at Oscar's filesystem.
Access and User Accounts - User accounts are controlled via central authentication and directories on HPC are only deleted on the request of the user, PI, or departmental chair.
Files not accessed for 30 days will be deleted from your scratch directory. Use ~/data for files you wish to keep long term.
non-disruptive work, including software changes, maintenance, and testing
may occur at any time
no notification provided
Monthly Scheduled Maintenance:
no downtime expected, but there may be limited degradation of performance
first Tuesday of the month, 8:00 am - 12:00 noon
Unscheduled Maintenance:
maximum 1 day downtime
occurs very rarely and includes any unplanned emergency issues that arise
Major Upgrade Maintenance:
service may be brought down for 3-5 days
occurs annually
Unplanned Outage
During Business Hours:
Send email to [email protected]. A ticket will get created and CCV staff will attempt to address the issue as soon as possible.
During Non-Business Hours:
Send email to .
Call CIS Operations Center at (401) 863-7562. A ticket will get created and CCV staff will be contacted to address the issue.
User and Research Support
CCV staff support for researchers seeking help with statistical modeling, machine learning, data mining, data visualization, computational biology, high-performance computing, and software engineering.
CCV staff provides tutorials on using Oscar for classes, groups and individual. Please check CCV Events for upcoming trainings and office hours.
CCV provides short videos (coming soon) for users to learn as well.
If you publish research that benefited from the use of CCV services or resources, we would greatly appreciate an acknowledgment that states:
Best Practices for I/O
Efficient I/O is essential for good performance in data-intensive applications. Often, the file system is a substantial bottleneck on HPC systems, because CPU and memory technology has improved much more drastically in the last few decades than I/O technology.
Parallel I/O libraries such as MPI-IO, HDF5 and netCDF can help parallelize, aggregate and efficiently manage I/O operations. HDF5 and netCDF also have the benefit of using self-describing binary file formats that support complex data models and provide system portability. However, some simple guidelines can be used for almost any type of I/O on Oscar:
Try to aggregate small chunks of data into larger reads and writes.
For the GPFS file systems, reads and writes in multiples of 512KB
provide the highest bandwidth.
Avoid using ASCII representations of your data. They will usually
require much more space to store, and require conversion to/from
binary when reading/writing.
Avoid creating directory hierarchies with thousands or millions of
files in a directory. This causes a significant overhead in managing
file metadata.
While it may seem convenient to use a directory hierarchy for managing large sets of very small files, this causes severe performance problems due to the large amount of file metadata. A better approach might be to implement the data hierarchy inside a single HDF5 file using HDF5's grouping and dataset mechanisms. This single data file would exhibit better I/O performance and would also be more portable than the directory approach.
Software on Oscar
Many scientific and HPC software packages are already installed on Oscar, and additional packages can be requested by submitting a ticket to [email protected]. If you want a particular version of the software, do mention it in the email along with a link to the web page from where it can be downloaded. You can also install your own software on Oscar.
CCV cannot, however, supply funding for the purchase of commercial software. This is normally attributed as a direct cost of research, and should be purchased with research funding. CCV can help in identifying other potential users of the software to potentially share the cost of purchase and maintenance. Several commercial software products that are licensed campus-wide at Brown are available on Oscar.
For software that requires a Graphical User Interface (GUI) we recommend using CCV's VNC Client rather than X-Forwarding.
SSH Agent Forwarding
How to forward local ssh keys to Oscar
SSH provides a method of sharing the ssh keys on your local machine with Oscar. This feature is called Agent Forwarding and can be useful for instance when working with version control or other services that authenticate via ssh keys. Below are instructions on how to configure your SSH connection to forward ssh-agent for diffeent operating systems
This research [Part of this research] was conducted using [computational/visualization]
resources and services at the Center for Computation and Visualization, Brown University.
Matlab Batch Jobs
Matlab can be used within a batch script. Here is an example batch script for running a serial Matlab program on an Oscar compute node:
#!/bin/bash
# Request an hour of runtime:
#SBATCH --time=1:00:00
# Default resources are 1 core with 2.8GB of memory.
# Use more memory (4GB):
#SBATCH --mem=4G
# Specify a job name:
#SBATCH -J MyMatlabJob
# Specify an output file
#SBATCH -o MyMatlabJob-%j.out
#SBATCH -e MyMatlabJob-%j.out
# Run a matlab function called 'foo.m' in the same directory as this batch script.
matlab -r "run foo.m; exit"
This is also available in your home directory as the file:
~/batch_scripts/matlab-serial.sh
Note the exit command at the end which is very important to include either there or in the Matlab function/script itself. If you don't make Matlab exit the interpreter, it will keep waiting for the next command until SLURM cancels the job after running out of requested walltime. So for example, if you requested 4 hours of walltime and your actual program completes in 1 hour, the SLURM job will not complete until the designated 4 hours which results in idle cores and wastage of resources and also blocks up your other jobs.
If the name of your batch script file is matlab-serial.sh, the batch job can be submitted using the following command:
sbatch matlab-serial.sh
no notification provided
Prior notification provided (depending on the issue, 1 day to 4 weeks advance notice provided)
SSH Agent Forwarding on a Windows system using PuTTY, with an example application to git.
Agent Forwarding with PuTTY
Once adding your private key to Pageant, open PuTTY and navigate to the Auth menu.
2. Check the 'Allow agent forwarding' checkbox, and return to the Session menu.
3. Enter the Host Name you usually use to connect to Oscar, and click 'Open'.
4. Entering your password. If you have ssh keys setup on your local computer to connect to GitHub, you can confirm your ssh-agent was properly forwarded by checking GitHub . If the ssh command fails, your agent has not been properly forwarded.
Open OnDemand
Open OnDemand (OOD) is a web portal to the Oscar computing cluster. An Oscar account is required to access Open OnDemand. Visit this link in a web browser and sign in with your Brown username and password to access this portal.
OOD provides with a several resources for interacting with Oscar.
Use the in the portal to view, copy, download or delete files on Oscar.
Launch , like Matlab and Jupyter Notebook, inside your web browser.
with your browser without needing a separate terminal emulator. This is especially handy for Windows users, since you do not need to install a separate program.
Features:
No installation needed. Just use your favorite browser!
No need to enter your password again. in seconds!
No need to use two-factor authentication multiple times. Just do it once, when you log into OOD.
Getting Help
Here are some ways to get help with using OSCAR
Filing a Support Ticket
Filing a good support ticket makes it much easier for CCV staff to deal with your request
State the problem/request in the subject of the email
Describe which software and with version you are using
Error message (if there was one)
The job number
How you were running, e.g. batch, interactively, vnc
Give as as small an example as possible that reproduces the problem
Q&A Forum
Ask questions and search for previous problems at our .
Office Hours
CCV holds weekly office hours. These are drop in sessions where we'll have one or more CCV staff members available to answer questions and help with any problems you have. Please visit for upcoming office hours and events.
Arrange a Meeting
You can arrange to meet with a CCV staff member in person to go over difficult problems, or to discuss how best to use Oscar. Email [email protected] to arrange a consultation.
Interactive Apps on OOD
You can launch several different apps on the Open OnDemand (OOD) interface. All of these apps start of a Slurm batch job on the Oscar cluster with the requested amount of resources. These jobs can access the filesystem on Oscar and all output files are written to the Oscar's file system.
Launching an App on OOD
Open on any browser of the your choice
If prompted, enter your Brown username and password.
Click on the "Interactive Apps" tab at the top of the screen to see the list of available apps. This will open the form to enter the details of the job.
Follow the instructions on the form to complete it. Some of fields can be left blank and OOD will choose the default option for you.
Click Launch to submit an OOD job. This will open a new tab on the browser It may take a few minutes for this job to start.
Click "Launch <APP>" again if prompted in the next tab.
SLURM limits on resources such CPUs, memory, GPUs or time for each partition still applies for OOD jobs. Please keep these in mind before choosing these options on the OOD form.
When submit a batch job from a terminal of the Desktop app or the Advanced Desktop app, users need to
run "unset SLURM_MEM_PER_NODE"before submitting a job if the job needs to specify --mem-per-cpu
Managing Modules
Modules
Command
Description
module list
Using RStudio
RStudio is an IDE for R that can be run on Oscar.
Launching RStudio
Open the Open On Demand Dashboard by following this link. Select RStudio (under "Default GUI's"). Fill in the form to allocate the required resources, and optionally select your R modules. Finally, click the "Launch Session" button.
Known Issues
Plotting figures may not work within RStudio. If this is the case, save the plots to a file, and view them through the Open On Demand Desktop App. If plots are required for your task, launch RStudio through the Desktop App.
To learn about using the Open On Demand Desktop App, look .
Miniconda
The Miniconda modules include only conda, mamba, python, and a few other packages. Users can use either mamba (preferred) or condo to install packages in their own conda environment.
Mamba is a drop-in replacement of conda, and is faster than conda.
Only activating and deactivating a conda environment still requires conda
For all other commands, conda can be replaced with mamba.
Setting Job Submission Settings
We have provided templates for you to use for job submission settings. These templates are in/gpfs/runtime/opt/forge/19.1.2/templates
Click Run and debug a program to open the following menu
Click Configure next to Submit to Queue and enter /gpfs/runtime/opt/forge/19.1.2/templates/slurm-ccv.qtf as the Submission template file
slurm-ccv-qtf lets you specify the total number of tasks. The number of tasks may not be equal for each node. This option will be the shortest time in the queue, but may not give you consistent run times.
slurm-ccv-mpi.qtf is for MPI jobs where you want to specify number of nodes and tasks per node
slurm-ccv-threaded.qtf is for threaded (single node) jobs
Mac/Linux/Windows(PowerShell)
Step 1 : Check for existing SSH key pair
Before generating new SSH key pair first check if you have an SSH key on your local machine.
If there are existing keys, please move to Step 3
Using Python or Conda environments in the Jupyter App
We recommend all users to install Python packages within an environment. This can be a Conda to a python virtual environment. More information can be found . Follow these steps to use such environments in the .
Python Environments:
Managing Jobs
Listing running and queued jobs
The squeue command will list all jobs scheduled in the cluster. We have also written wrappers for squeue on Oscar that you may find more convenient:
Installing JAX
This page describes how to install JAX with Python virtual environments
In this example, we will install Jax.
Step 1: Request an interactive session on a GPU node with Ampere architecture GPUs
Here, -f = feature. We only need to build on Ampere once.
Step 2: Once your session has started on a compute node, run nvidia-smi to verify the GPU and then load the appropriate modules
Step 3: Create and activate the virtual environment
Setup virtual environment and debugger
If you have an existing virtual environment, proceed to step 2. Otherwise, to create a new virtual environment:
2. Search for Python.VenvPath as shown in the picture below:
3. VSCode expects you to have multiple virtual environments for each of your different python projects, and it expects you to put them all in the same directory. Pointing to the parent directory lets it scan and find all expected virtual environments, and then you can easily toggle between them in interface.
Compiling CUDA
Compiling with CUDA
To compile a CUDA program on Oscar, first load the CUDA with:
The CUDA compiler is called nvcc, and for compiling a simple CUDA program it uses syntax simlar to gcc:
Python on Oscar
Several versions of Python are available on Oscar as modules. For Python 2, we recommend using the python/2.7.16 module. For Python 3, we recommend using the python/3.7.4 module. These modules include the pip and virtualenv commands, but do not include other common Python packages (e.g., SciPy, NumPy). This affords individual users complete control over the packages they are using, thereby avoiding issues that can arise when code written in Python requires specific versions of Python packages.
To use the recommend Python modules, use the following commands to load the relevant module:
module load python/2.7.16
module load python/3.7.4
Restoring Deleted Files
Nightly snaphots of the file system are available for the last 30 days.
CCV does not guarantee that each of the last 30 days will be available in snapshots because occasionally the snapshot process does not complete within 24 hours.
Installing Frameworks (PyTorch, TensorFlow, Jax)
This page describes installing popular frameworks like TensorFlow, PyTorch & JAX, etc. on your Oscar account.
Preface: Oscar is a heterogeneous cluster meaning we have nodes with different architecture GPUs (Pascal, Volta, Turing, and Ampere). We recommend building the environment first time on Ampere GPUs with the latest CUDA11 modules so it's backward compatible with older architecture GPUs.
In this example, we will install PyTorch (refer to sub-pages for TensorFlow and Jax).
Step 1: Request an interactive session on a GPU node with Ampere architecture GPUs
Oscar's Filesystem
CCV uses IBM's General Parallel File System (GPFS). Users have a home, data, and scratch space.
home~
20GB of space
Optimized for many small files
Installing TensorFlow
In this example, we will install TensorFlow.
Step 1: Request an interactive session on a GPU node with Ampere architecture GPUs
Here, -f = feature. We only need to build on Ampere once.
Step 2: Once your session has started on a compute node, run nvidia-smi to verify the GPU and then load the appropriate modules
Step 3: Create and activate the virtual environment
Dependent Jobs
Here is an example script for running dependent jobs on Oscar.
There are 3 batch jobs. Each job has it's own batch script: job1.sh, job2,sh, jobs.sh. The script above (script.sh) submits the three jobs.
line 4: job1 is submitted.
line 7: job2 depends on job1 finishing successfully.
line 10: job3 depends on job2 finishing successfully.
Gaussian
Gaussian is a general purpose computational chemistry package. Oscar uses the Gaussian 9 package.
Setting Up Gaussian
In order to use Gaussian on Oscar, you must be a part of the g09 group. To check your groups, run the groups command in the terminal.
You must first choose a Gaussian module to load. To see available Gaussian modules, run module avail gauss
Transferring Files between Oscar and Campus File Storage (Replicated and Non-Replicated)
You may use either Globus (recommended) or smbclient to transfer data between Oscar and Campus File Storage.
Globus
Follow .
Web-based Terminal App
Open OnDemand offers a browser-based terminal app to access Oscar. Windows users who do not want to install an SSH client like Putty will find this app very useful.
Accessing the terminal
Windows(PuTTY)
Key Generation & Setup
Open PuTTYgen (this comes as part of the package), change the 'Number of bits in a generated key:' to 4096 (recommended), then click 'Generate'
Parallel Matlab
You can explore GPU computing through Matlab if you think your program can benefit from massively parallel computations:
Finally, parallel computing features like parfor and spmd can be used by launching a pool of workers on a node.
Improving Performance and Memory Management
Matlab programs often suffer from poor performance and running out of memory. Among other things, you can refer the following web pages for best practices for an efficient code:
The first step to speeding up Matlab applications is identifying the part which takes up most of the run time. Matlab's "Profiling" tool can be very helpful in doing that:
SSH Key Login (Passwordless SSH)
How to set up SSH key authentication.
When connecting from a campus network to sshcampus.ccv.brown.edu you can set up SSH keys as a form of authentication instead of having to enter your password interactively. Follow the insctructions below that correspond to your operating system/connection method.
Users can install any Python package they require by following the instructions given on the Installing Python Packages page.
Intel provides optimized packages for numerical and scientific work that you can install through pip or anaconda.
Python 2 will enter End-of-Life (EOL) status and will receive no further official support as of January 2020. As a consequence, you may see the following message when using pip with Python 2.
DEPRECATION: Python 2.7 will reach the end of its life on January 1st, 2020. Please upgrade your Python as Python 2.7 won't be maintained after that date. A future version of pip will drop support for Python 2.7.
Lists all modules that are currently loaded in your software environment.
module avail
Lists all available modules on the system. Note that a module can have multiple versions.
module help <name>
Prints additional information about the given software.
module load <name>
Adds a module to your current environment. If you load using just the name of a module, you will get the default version. To load a specific version, load the module using its full name with the version: "module load gcc/6.2"
Load the relevant python module and create and/or activate the environment. See this page for more information about creating virtual environments.
Run pip install notebook to install Jupyter notebook, if not already installed.
Run pip install ipykernel to install ipykernel in this environment.
Run python -m ipykernel install --user --name=<myenv> where <myenv> is the name of the environment.
Launching Jupyter Notebook
Open the "Basic Jupyter Notebook for Python Environments" app on the Open OnDemand interface
Under "Python Module on Oscar", choose the python module you loaded when the environment was created.
Under "Python Virtual Environment", add the name of the Virtual Environment you created. Note: If your virtual environment is not at the top level of your home directory, you should input the absolute path to the environment directory.
Under the "Modules" , enter the name of the python module used to create the environment. Add any additional modules you may need separated with a space.
Choose the other options as required.
Click "Launch" to start the job
Click "Connect to Jupyter" on the next screen.
To start a new notebook, click "New" -> <myenv> where <myenv> is the environment.
For starting a pre-existing notebook, open the notebook. In the Jupyter interface, click "Kernel" -> "Change Kernel" -> <myenv> where myenv is the name of the environment.
Conda Environments
One Time Setup:
Open a terminal on Oscar.
Activate the conda environment.
Run pip install notebook to install Jupyter notebook, if not already installed.
Run pip install ipykernel to install ipykernel in this environment.
Run python -m ipykernel install --user --name=<myenv> where <myenv> is the name of the environment.
Launching Jupyter Notebook
Open the "Basic Jupyter Notebook with Anaconda" app on the Open OnDemand interface
Under "Oscar Anaconda module", choose "anaconda/2020.02"
Enter the name of the conda environment in "Conda Env"
Choose the other options as required.
Click "Launch" to start the job
Click "Connect to Jupyter" on the next screen.
To start a new notebook, click "New" -> <myenv> where <myenv> is the environment.
For starting a pre-existing notebook, open the notebook. In the Jupyter interface, click "Kernel" -> "Change Kernel" -> <myenv> where myenv is the name of the environment.
Oscar is a shared machine used by hundreds of users at once. User requests are called jobs. A job is the combination of the resource requested and the program you want to run on the compute nodes of the Oscar cluster. On Oscar, Slurm is used to schedule and manage jobs.
Jobs can be run on Oscar in two different ways:
Interactivejobs allow the user to interact with programs (e.g., by entering input manually, using a GUI) while they are running. However, if your connection to the system is interrupted, the job will abort. Small jobs with short run times and jobs that require the use of a GUI are best-suited for running interactively.
Batchjobs allow you to submit a script that tells the cluster how to run your program. Your program can run for long periods of time in the background, so you don't need to be connected to Oscar. The output of your program is continuously written to an output file that you can view both during and after your program runs.
Jobs are scheduled to run on the cluster according to your account priority and the resources you request (i.e., cores, memory, and runtime). In general, the fewer resources you request, the less time your job will spend waiting in the queue.
Please do not run CPU-intense or long-running programs directly on the login nodes! The login nodes are shared by many users, and you will interrupt other users' work.
To use the above script to submit the 3 jobs, run the script as follows:
./script.sh
For details on the types of dependencies you can use in slurm see the sbatch manual page.
#!/bin/bash
# first job - no dependencies
jobID_1=$(sbatch job1.sh | cut -f 4 -d' ')
# second job - depends on job1
jobID_2=$(sbatch --dependency=afterok:$jobID_1 job2.sh | cut -f 4 -d' ')
# third job - depends on job2
sbatch --dependency=afterany:$jobID_2 job3.sh
Step 2 : Generate a new SSH Keypair
ssh-keygen -t rsa
ssh-keygen.exe
Press Enter to accept the default file location and file name.
The ssh-keygen will ask you to type a secure passphrase. This is optional. If you don't want to use a passphrase just press Enter
Verify the SSH keys are generated correctly, you should see two files id_rsa and id_rsa.pub under ~/.ssh directory.
DO NOT upload or send the private key.
Step 3 : Copy the public key to Oscar
You will now need to copy your public key to Oscar. There are two ways to acomplish this.
With ssh-copy-id
If your OS comes with the ssh-copy-id utility, then you'll be able to copy your public key into Oscar as follows:
You will be prompted for a Password. The public key will be appended to the authorized_keys file on Oscar.
If you used a custom name for your key instead of the default id_rsa then you'll need pass the name of your key to ssh-copy-id i.e.,
Without ssh-copy-id
If your system does not come with the ssh-copy-id utility installed, then you'll need to copy your public key by hand.
Get the contents of id_rsa.pub file. One option is to use cat in your teminal
cat id_rsa.pub.
Copy the contents of this file to your clipboard, as we need to upload it to Oscar.
Login into Oscar via regular ssh ssh <username>@ssh.ccv.brown.edu. Once you are on the login node, open the authorized_keys file with your text editor of choice e.g.,
vim ~/.ssh/authorized_keys
or
nano ~/.ssh/authorized_keys
Add your public keys to end of this file. Save and exit.
Step 4 : Login to Oscar using your SSH keys
If everything went well, you will be logged in immediately withouth prompting you for a password.
Viewing estimated time until completion for pending jobs
This command will list all of your pending jobs and the estimated time until completion.
Canceling jobs
View details about completed jobs
The sacct command will list all of your running, queued and completed jobs since midnight of the previous day. To pick an earlier start date, specify it with the -S option:
To find out more information about a specific job, such as its exit status or the amount of runtime or memory it used, specify the -l ("long" format) and -j options with the job ID:
The myjobinfo command uses the sacct command to display "Elapsed Time", "Requested Memory" and "Maximum Memory used on any one Node" for your jobs. This can be used to optimize the requested time and memory to have the job started as early as possible. Make sure you request a conservative amount based on how much was used.
ReqMem shows the requested memory: A c at the end of number represents Memory Per CPU, a n represents Memory Per Node. MaxRSS is the maximum memory used on any one node. Note that memory specified to sbatch using --mem is Per Node.
Step 4: Install the required packages
Step 5: Test that PyTorch is able to detect GPUs
If the above function returns gpu, then it's working correctly. You are all set, now you can install other necessary packages.
The Oscar GPU nodes feature NVIDIA M2050 cards with the Fermi architecture, which supports CUDA's "compute capability" 2.0. To fully utilize the hardware optimizations available in this architecture, add the -arch=sm_20 flag to your compile line:
This means that the resulting executable will not be backwards-compatible with earlier GPU architectures, but this should not be a problem since CCV nodes only use the M2050.
Memory caching
The Fermi architecture has two levels of memory cache similar to the L1 and L2 caches of a CPU. The 768KB L2 cache is shared by all multiprocessors, while the L1 cache by default uses only 16KB of the available 64KB shared memory on each multiprocessor.
You can increase the amount of L1 cache to 48KB at compile time by adding the flags -Xptxas -dlcm=ca to your compile line:
If your kernel primarily accesses global memory and uses less than 16KB of shared memory, you may see a benefit by increasing the L1 cache size.
If your kernel has a simple memory access pattern, you may have better results by explicitly caching global memory into shared memory from within your kernel. You can turn off the L1 cache using the flags –Xptxas –dlcm=cg.
Nightly snapshots of the file system are available for the last 5-7 days can be found in the following directories.
Home directory snapshot
Data directory snapshot
Scratch directory snapshot
To restore a file, copy the file from the snapshot to your directory.
Do not use the links in your home directory snapshot to try and retrieve snapshots of data and scratch. The links will always point to the current versions of these files. An easy way to check what a link is pointing to is to use ls -l
e.g.:
Restore a file from a snapshot taken between 8 and 30 days ago
If you need one or more files from a nightly snapshot taken between 8 and 30 days ago, please contact [email protected] for help.
512GB of space - contact [email protected] increase on a temporary basis
Optimized for reading/writing large files
NO BACKUPS
Purging: files not accessed for 30 days may be deleted
Quota is per individual user
Grace period of 21 days
Files not accessed for 30 days will be deleted from your scratch directory. This is because scratch is high performance space. The fuller scratch is the worse the read/write performance. Use ~/data for files you need to keep long term.
The scratch purge is on individual files. It is by 'atime' which is when the file was last read. You can use 'find' to find files that are at risk of being purged, e.g. to find files in the current directory that have not been accessed in the last 25 days:
find . -atime +25
A good practice is to configure your application to read any initial input data from ~/data and write all output into ~/scratch. Then, when the application has finished, move or copy data you would like to save from ~/scratch to ~/data.
Note: class or temporary accounts may not have a ~/data directory!
To see how much space you have on Oscar you can use the command myquota. Below is an example output
You can go over your quota up to the hard limit for a grace period. This grace period is to give you time to manage your files. When the grace period expires you will be unable to write any files until you are back under quota.
There is a quota for space used and for number of files. If you hit the hard limit on either of these you will be unable to write any more files until you are back under quota.
Step 4: Install the required packages
Step 5: Test that PyTorch is able to detect GPUs
If the above function returns True, then you are all set.
. You can load a Gaussian module using the command
module load <module-name>
.
Available Versions
Gaussian 9 (g09)
Running Gaussian
Gaussian can be run either interactively or within a batch script using one of two command styles:
g09 job-name
g09 <input-file >output-file
In the first form, the program reads input from job-name.gjf and writes its output to job-name.log. When no job-name has been specified, the program will read from standard input and write to standard output
Given a valid .gjf file (we'll call it test-file.gjf), we can use the following simple batch script to run Gaussian:
g09-test.sh
Then queue the script using
Once the job has been completed, you should have a g09-test.out, a g09-test.err, and a test-file.out.
smbclient
You can transfer files between Campus File Storage and Oscar using smbclient.
Transfer Instructions
1) Log into Oscar:
2) Start a screen session. This will allow you to reattach to your terminal window if you disconnect.
3) To use Oscar's high-speed connection to Campus File Storage - Replicated:
Replace SHARE_NAME, DIRECTORY_NAME, and BROWN_ID. DIRECTORY_NAME is an optional parameter. The password required is your Brown password.
4) Upload/download your data using the FTP "put"/"get" commands. Replace DIRECTORY_NAME with the folder you'd like to upload.
5) You can detach from the screen session with a "CTRL+A D" keypress. To reattach to your session:
smbclient basics
put is upload to Campus File Storage
Usage: put <local_file> [remote file name]
Copy <local_file> from Oscar to Campus File Storage. The remote file name is optional (use if you want to rename the file)
get is download to Oscar
Usage: get <remote_file> [local file name] Copy <remote_file> from the Campus File Storage to Oscar. The local file name is optional (use if you want to rename the file)
Moving more than one file:
To move more than one file at once use mput or mget. By default:
recurse is OFF. smbclient will not recurse into any subdirectories when copying files
prompt is ON. smbclient will ask for confirmation for each file in the subdirectories
myq List only your own jobs.
allq List all jobs, but organized by partition, and a summary of the nodes in use in the
partition.
allq <partition> List all jobs in a single partition.
myjobinfo Get the time and memory used for your jobs.
squeue -u <your-username> -t PENDING --start
scancel <jobid>
sacct -S 2012-01-01
sacct -lj <jobid>
myjobinfo
Info about jobs for user 'mdave' submitted since 2017-05-19T00:00:00
Use option '-S' for a different date or option '-j' for a specific Job ID.
JobID JobName Submit State Elapsed ReqMem MaxRSS
1861 ior 2017-05-19T08:31:01 COMPLETED 00:00:09 2800Mc 1744K
1862 ior 2017-05-19T08:31:11 COMPLETED 00:00:54 2800Mc 22908K
1911 ior 2017-05-19T15:02:01 COMPLETED 00:00:06 2800Mc 1748K
1912 ior 2017-05-19T15:02:07 COMPLETED 00:00:21 2800Mc 1744K
ls -l /gpfs/.snapshots/2020-07-15/home/ghopper/data
lrwxrwxrwx 1 ghopper navy 22 Mar 1 2016 /gpfs/.snapshots/2020-07-15/ghopper/scratch -> /gpfs/data/navy
In the top menu, click Clusters -> >_OSCAR Shell Access
A new tab will open and the web-based terminal app will be launched in it. The shell will be launched on one of the login nodes.
The shell DOES NOT start on a compute node. Please do not run computations or simulations on the login nodes, because they are shared with other users. You can use the login nodes to compile your code, manage files, and launch jobs on the compute nodes.
No installation needed. Just use your favorite browser!
No need to enter your password again. SSH into Oscar in seconds!
No need to use two factor authentication again. Just do it once, when you log into OOD.
Use it with, or without, VPN. Your workflow remains the same.
2. Move your cursor around randomly in order to "salt" your key, while the key is being generated. Once the key is generated, you should see something like this:
3. Replace the text in the 'Key comment:' field with something recognizable and enter a passphrase in the two fields below.
4. Copy the text in the 'Public key for pasting...' field (the text continues past what is displayed) and paste it wherever the public key is needed. If you are using GitHub, you can now create a new SSH key in your Personal Settings and paste this text into the 'Key' field.
5. Click on 'Save private key' and select a logical/recognizable name and directory for the file. Your private key is saved in the selected file.
6. Open Pageant (also part of the PuTTY package). If a message saying "Pageant is already running" is displayed, open your system tray and double click on the Pageant icon.
To open your system tray, click on the up arrow (looks like: ^ ) icon at the bottom right of your screen (assuming your taskbar is at the bottom of your screen).
7. Click on 'Add Key' and select the file you saved when generating your key earlier (Step 5). If it is requested, enter the passphrase you created at Step 3 to complete the process.
In order to not have to add the key to Pageant after every time your machine reboots, you can add the key file(s) to your Windows startup folder (the directory for the current user is C:\Users\[User Name]\AppData\Roaming\Microsoft\Windows\Start Menu\Programs\Startup). You may still be prompted to enter the passphrase after a reboot, but you will not have to find and add the key to Pageant every time.
Hi JaneDoe! You've successfully authenticated, but GitHub does not provide shell access.
Connection to github.com closed.
Version Control
Git Overview
Version Control refers to the management of changes made to source code or any such large amount of information in a robust manner by multiple collaborators. Git is by far the most popular version control system.
Git enables effective collaboration among developers. In a team setting, multiple developers often work on the same project simultaneously. With Git, each developer can work on their own local copy of the project, making changes and experimenting freely without affecting the main codebase. Git allows developers to merge their changes seamlessly, ensuring that modifications made by different individuals can be consolidated efficiently. It provides mechanisms to track who made specific changes, making it easier to understand the evolution of the project and identify potential issues.
Git Workflow
Nearly all operations that are performed by Git are in you local computing environment, for the exception of few used purely to synchronize with a remote. Some of the most common git operations are depicted below. In summary a typical flow consists of making changes to your files, staging them via git add, marking a save point via git commit, then finally syncing to your remote (e.g., GitHub) via git push. If you are pushing changes to your remote from multiple places, you can bring changes your most recent version using git pull, which is the equivalent of doing git fetch followed by a git merge operation
Cheatsheet
Below are some of the most commonly used Git commands. You can also get much more information by running git --help. And if you'd like to learn more there is an
Command
Summary
Git Configuration
While using Git on Oscar, make sure that you to have your correct Name and Email ID to avoid confusion while working with remote repositories (e.g., GitHub, GitLab, BitBucket).
Getting Out of Trouble
Git can sometimes be a bit tricky. And we all eventually find ourselves in a place where we want to undo something or fix a mistake we made with Git. (pardon the profanity) has a bunch of really excellent solutions to common problems we sometimes run in to with Git.
Job Arrays
A job array is a collection of jobs that all run the same program, but on different values of a parameter. It is very useful for running parameter sweeps, since you don't have to write a separate batch script for each parameter setting.
To use a job array, add the option:
#SBATCH --array=<range>
in your batch script. The range can be a comma separated list of integers, along with ranges separated by a dash. For example:
1-20
1-10,12,14,16-20
A job will be submitted for each value in the range. The values in the range will be substituted for the variable $SLURM_ARRAY_TASK_ID in the remainder of the script. Here is an example of a script for running a serial Matlab script on 16 different parameters by submitting 16 different jobs as an array:
#!/bin/bash
#SBATCH -J MATLAB
#SBATCH -t 1:00:00
#SBATCH --array=1-16
# Use '%A' for array-job ID, '%J' for job ID and '%a' for task ID
#SBATCH -e arrayjob-%a.err
#SBATCH -o arrayjob-%a.out
echo "Starting job $SLURM_ARRAY_TASK_ID on $HOSTNAME"
matlab -r "MyMatlabFunction($SLURM_ARRAY_TASK_ID); quit;"
You can then submit the multiple jobs using a single sbatch command:
The $SLURM_ARRAY_TASK_ID can be manipulated as needed. For example, you can generate a fixed length number form it. The following example generates a number of length of 3 from $SLURM_ARRAY_TASK_ID.
For more info:
Installing R Packages
Installing R packages
Users should install R packages for themselves locally. This documentation shows you how to install R packages locally (without root access) on Oscar.
If the package you want to install has operating-system-level dependencies (i.e. the package depends on core libraries), then we can install it as a module.
Installing an R package
First load the R version that you want to use the package with:
Start an R session
Note some packages will require code to be compiled so it is best to do R packages installs on the login node.
To install the package 'wordcloud':
You will see a warning:
Answer y . If you have not installed any R packages before you will see the following message:
Answer y . The package will then be installed. If the install is successful you will see a message like:
If the installation was not successful you will see a message like:
There is normally information in the message that gives the reason why the install failed. Look for the word ERROR in the message.
Possible reasons for an installation failing include:
Other software is needed to build the R package, e.g. the R package rgdal needs gdal so you have to do module load gdal
A directory needs deleting from a previous failed installation.
Removing an R package
Start an R session:
To remove the 'wordcloud' package:
Ampere Architecture GPUs
The new Ampere architecture GPUs on Oscar (A6000's and RTX 3090's)
The new Ampere architecture GPUs do not support older CUDA modules. Users must re-compile their applications with the newer CUDA/11 or older modules. Here are detailed instructions to compile major frameworks such as PyTorch, and TensorFlow.
PyTorch
Users can install PyTorch from a pip virtual environment or use pre-built singularity containers provided by Nvidia NGC.
To install via virtual environment:
To use NGC containers via Singularity :
Pull the image from NGC
Export PATHs to mount the Oscar file system
To use the image interactively
To submit batch jobs
Screen
screen is a "terminal multiplexer", it enables a number of terminals (or windows) to be accessed and controlled from a single terminal. screen is a great way to save an interactive session between connections to oscar. You can reconnect to the session from anywhere!
Screen commands
Common commands are:
start a new screen session with session name: screen -S <name>
list running sessions/screens: screen -ls
attach to session by name: screen -r <name>
detach: Ctrl+a d
detach and logout (quick exit): Ctrl+a d d
kill a screen session: screen -XS session_name quit
Reconnecting to your screen session
There are several login nodes in Oscar, and the node from where you launched screen matters! That is, you can only reconnect from the login node in which you launched screen from
In order to reconnect to a running screen session, you need to be connected to the same login node that you launched your screen session from. In order to locate and identify your screen sessions correctly, we recommed the following:
Create a directory to store the information of your screen sessions. You only need do this once.
Put the following line into your /.bashrc. This tells the screen program to save the information of your screen sessions in the directory created in the previous step . This allows you to query your screen sessions across different login nodes. To make this change effective in your current sessions, you need run 'source /.bashrc' in each of your current session . However, you do not need to run 'source /bashrc' in your new sessions.
Name your new screen session using the name of the login node. For instance, start your screen with a commnd similar to
Matlab GUI
On Oscar, the command matlab is actually a wrapper that sets up MATLAB to run as a single-threaded, command-line program, which is the optimal way to pack multiple Matlab scripts onto the Oscar compute nodes.
To run the actual multi-threaded version with JVM and Display enabled, use:
Similarly, to run this without the display enabled:
Do not run Matlab on the Oscar login nodes. Request a compute node either with an interactive session, a batch script, or using the VNC.
VNC
The client provided by CCV is the best way to launch GUI applications on Oscar, including Matlab. From the terminal emulator in VNC use the matlab-threaded command to launch the Matlab GUI. For example,
Here is a snapshot of what it looks like:
X11 Forwarding
You can also run the MATLAB GUI in an X-forwarded interactive session. This requires installing an X server on your workstation/PC and logging in to Oscar with enabled. Use the interact command to get interactive access to a compute node. Again, for launching the GUI, you need to use the matlab-threaded command, which enables the display and JVM. You may however experience a lag in response from the Matlab GUI in an X forwarded session. Note that if Matlab does not find the X window system available, it will launch in command line mode (next section).
CIFS
A workaround in some situations may be to use to mount the Oscar filesystem on your PC and using the Matlab installation on your computer. For example, if you have your simulation results residing on Oscar, this might be a quick way to do post-processing on the data instead of having to move the data to your computer or using the Matlab GUI on Oscar. Note that users can connect to CIFS only from Brown computers or on Brown WiFi.
Mac/Linux
Agent Forwarding in Mac and Linux Systems
Start the SSH-Agent
First, start your ssh-agent with the command below.
You should see an output similar to this:
Add Key(s)
Next, add your ssh private keys to the running agent (using the ssh-add command on line 1). This step may be repeated for every key pair you use to connect to different git servers. For most, this file is called id_rsa and will live in ~/.ssh/id_rsa. If you set a password for your ssh keys, the agent will prompt you to enter them.
Confirm the ssh keys have been loaded into the agent with ssh-add -L:
Connect to Oscar
Now ssh into Oscar with the -A option as shown on the first line below (replace username with your Oscar username). -A will forward your ssh-agent to Oscar, enabling you to use the ssh keys on your laptop while logged into Oscar.
If you have ssh keys setup on your local computer to connect to GitHub, you can confirm your ssh-agent was properly forwarded by checking GitHub . If the ssh command fails, your agent has not been properly forwarded.
Always connecting with Agent Forwarding
To make these changes permanent, you can add the ForwardAgent yes option to your ssh configuration file. To learn more about configuring your ssh connections, visit
Mixing MPI and CUDA
Combining CUDA and MPI
Mixing MPI (C) and CUDA (C++) code requires some care during linking because of differences between the C and C++ calling conventions and runtimes. One option is to compile and link all source files with a C++ compiler, which will enforce additional restrictions on C code. Alternatively, if you wish to compile your MPI/C code with a C compiler and call CUDA kernels from within an MPI task, you can wrap the appropriate CUDA-compiled functions with the extern keyword, as in the following example.
These two source files can be compiled and linked with both a C and C++ compiler into a single executable on Oscar using:
Arm Forge
is available on Oscar. There are two products, DDT (debugger) and MAP (performance reports).
We recommend you use the Arm Forge remote client to launch your debugging jobs on Oscar. The first time you set up Arm Forge you will need to configure the client with the following steps:
Download the on your machine.
matlab-threaded
matlab-threaded -nodisplay
$ eval $(ssh-agent)
Agent pid 48792
Get remote repo's commits and download (try and resolve conflicts)
The CUDA/C++ compiler nvcc is used only to compile the CUDA source file, and the MPI C compiler mpicc is used to compile the C code and to perform the linking. / multiply.cu /
include
global void multiply (const float a, float b) { const int i = threadIdx.x + blockIdx.x blockDim.x; b[i] = a[i]; }
extern "C" void launch_multiply(const float a, const b) { / ... load CPU data into GPU buffers a_gpu and b_gpu /
Note the use of extern "C" around the function launch_multiply, which instructs the C++ compiler (nvcc in this case) to make that function callable from the C runtime. The following C code shows how the function could be called from an MPI task.
/ main.c /
include
void launch_multiply(const float a, float b);
int main (int argc, char **argv) { int rank, nprocs; MPI_Init (&argc, &argv); MPI_Comm_rank (MPI_COMM_WORLD, &rank); MPI_Comm_size (MPI_COMM_WORLD, &nprocs);
Mixing MPI and CUDA
Mixing MPI (C) and CUDA (C++) code requires some care during linking because of differences between the C and C++ calling conventions and runtimes. One option is to compile and link all source files with a C++ compiler, which will enforce additional restrictions on C code. Alternatively, if you wish to compile your MPI/C code with a C compiler and call CUDA kernels from within an MPI task, you can wrap the appropriate CUDA-compiled functions with the extern keyword, as in the following example.
These two source files can be compiled and linked with both a C and C++ compiler into a single executable on Oscar using:
The CUDA/C++ compiler nvcc is used only to compile the CUDA source file, and the MPI C compiler mpicc is used to compile the C code and to perform the linking.
Note the use of extern "C" around the function launch_multiply, which instructs the C++ compiler (nvcc in this case) to make that function callable from the C runtime. The following C code shows how the function could be called from an MPI task.
Warning in install.packages("wordcloud", repos = "http://cran.r-project.org") :
'lib = "/gpfs/runtime/opt/R/3.4.2/lib64/R/library"' is not writable
Would you like to use a personal library instead? (y/n)
Would you like to create a personal library
~/R/x86_64-pc-linux-gnu-library/3.4
to install packages into? (y/n)
** R
** data
** preparing package for lazy loading
** help
*** installing help indices
** building package indices
** testing if installed package can be loaded
* DONE (wordcloud)
Warning message:
In install.packages("wordcloud", repos = "http://cran.r-project.org") :
installation of package ‘wordcloud’ had non-zero exit status
__multiply__ <<< ...block configuration... >>> (a_gpu, b_gpu);
safecall(cudaThreadSynchronize());
safecall(cudaGetLastError());
/* ... transfer data from GPU to CPU */
/* ... prepare arrays a and b */
launch_multiply (a, b);
MPI_Finalize();
return 1;
Compile your code with -g so you can see the source code in your debugging session
Arm DDT
Arm DDT is a powerful graphical debugger suitable for many different development environments, including:
Single process and multithreaded software.
OpenMP.
Parallel (MPI) software.
Arm MAP
Arm MAP is a parallel profiler that shows you which lines of code took the most time to run, and why. Arm MAP does not require any complicated configuration, and you do not need to have experience with profiling tools to use it.
Arm MAP supports:
MPI, OpenMP and single-threaded programs.
Small data files. All data is aggregated on the cluster and only a few megabytes written to disk, regardless of the size or duration of the run.
Sophisticated source code view, enabling you to analyze performance across individual functions.
Both interactive and batch modes for gathering profile data.
A rich set of metrics, that show memory usage, floating-point calculations and MPI usage across processes, including:
Percentage of vectorized instructions, including AVX extensions, used in each part of the code.
Time spent in memory operations, and how it varies over time and processes, to verify if there are any cache bottlenecks.
Oscar has compute nodes in the partitions listed below.
batch - The batch partition is for programs/jobs which need neither GPUs nor large memory.
bigmem - The bigmem partition is for programs/jobs which require large memory.
debug - The debug partition is for users to debug programs/jobs.
Below are node details including cores and memory for all partitions.
Hardware details
Hardware details for all partitions. The Features column shows the features available for the --constraint option for SLURM. This includes the available CPU types as well GPUs.
Partition
CPUs/ Node
Nodes
Total CPUs
GPUs/ Node
Total GPUs
Memory (GB)
Features
Remote IDE (VSCode)
Access Oscar's file-system remotely from VSCode.
VSCode one-time setup
To use VSCode you must be on a Brown compliant network or connected to the VPN. Please install the Brown VPN client before proceeding.
To use VSCode you will need to be connected to the VPN. Please install the before proceeding.
Install the for VSCode:
2. Open VSCode settings and uncheck symlink:
Code > Preferences > Settings
File > Preferences > Settings
Search for symlink and make sure the symlink searching is unchecked
3. Make sure you have set up passwordless SSH authentication to Oscar. If you haven't, please refer to this .
If you have Windows Subsystem for Linux (WSL) installed in your computer, you need to follow the instructions for Windows (PowerShell).
4. Edit the config file:
The config file is located at:
~/.ssh/config
The config file is located at:
If you have Windows Subsystem for Linux (WSL) installed in your computer, you need to follow the instructions for Windows (PowerShell).
Edit the config file on your local machine, add the following lines. Replace <username> with your Oscar username.
5. In VSCode, select Remote-SSH: Connect to Host… and after the list populates select ccv-vscode-node
6. Install and set up of VSCode
After a moment, VS Code will connect to the SSH server and set itself up.
After a moment, VS Code will connect to the SSH server and set itself up. You might see the Firewall prompt, please click allow.
Configure VSCode
Important: Please run the following to add a settings.json file to your config. This is because the filewatcher and file searcher (rg) indexes all the files you have access to in your workspace. If you have a large dataset (e.g. machine learning) this can take a lot of resources on the vscode node.
Connect to VSCode first.
You can either create a symlink via the ln command below,
or manually create /users/$USER/.vscode-server/data/Machine/settings.json file with following contents
Reconnect to VSCode
Click the green icon "Open a Remote Window" in the bottom left corner of VSCode Window. Then click "Connect to Host" in the drop down list.
2. Select the ccv-vscode-node option to connect to Oscar.
Common Linux Commands
Common Linux Commands
Command
Related Word/Phrase
Description
Using File Explorer on OOD
The filesystem on Oscar can be accessed through the file explorer on this web portal. The file explorer allows you
List files
Create a directory
Rename files
Copy/Move files
To access the file explorer, click "Files" -> "Home Directory" at the top of the screen.
Check the documentation below for some of these services below:
Changing directories on File explorer
To access a directory, click "Change directory" and enter the path name
Do not use "~" in your directory path name. The path should start with "/users" or "/gpfs/"
To access your home directory, click the "Home Directory" link on the left. The path name at the top of the page should change to "/users/<username>"
To access your scratch directory, click the "scratch" directory in your home directory OR click "Change directory" and enter "/users/<username>/scratch"
To access your data directory, click the "data" directory in your home directory OR click "Change directory" and enter "/users/<username>/data"
Edit plain-text files
that contains the plain-text file.
Click the icon with the three dots -> Edit
The file will open in a text editor in a new tab
Download files or directories
that contains the file or directory.
Click the icon with the three dots -> Download
To download multiples files:
Click the check-box to the left of the file name.
Scroll to the top of the page and click "Download"
Directories are downloaded as zipped files on your computer.
Upload files or directories
where you need to upload the files.
Click the "Upload" button.
Follow the instructions on the screen. You can click the "Browse" buttons or drag and drop files.
Launch a terminal
where you would like to open the terminal
Click "Open in Terminal" at the top of the page.
A web-based terminal will open in a new tab of your browser. You will be logged into one of the login nodes.
Slurm Partitions
Partition Overview
Oscar has the following slurm partitions. The number and size of jobs allowed on Oscar vary with both partition and type of user account. You can email [email protected] if you need advice on which partitions to use.
To list partitions on Oscar available to your account, run the following command:
To view all partitions (including ones you don't have access to), replace the -O in the command above with -aO.
batch is the default partition.
Partition Details
Below are brief summary of partitions. For the details of nodes in partitions, please see .
batch
General purpose computing
Priority is determined by account type (from highest
to lowest: condo, priority, exploratory)
Condo limits apply to the group (i.e., they reflect the sum of all users on the condo). Condo users can check the limits on their condo with the command condos.
There is no limit on the time for condo jobs, but users should be aware that planned maintenance on the machine may occur (one month’s notice is given prior to any planned maintenance).
debug
Short wait time, short run time access for debugging
All users have the same limits and priority on the debug partition
vnc
These nodes are for running VNC sessions/jobs
Account type may affect Priority
gpu
For GPU-based jobs
GPU Priority users get higher priority and more resources than free users on the GPU partition
Condo users submit to the gpu partition with normal or priority access (if they have a priority account in addition to their condo)
gpu-he
For GPU-based jobs
Uses Tesla V100 GPUs
Restricted to High End GPU Priority users
gpu-debug
Short wait time, short run time gpu access for debugging
All users have the same limits and priority on the gpu-debug partition
bigmem
For jobs requiring large amounts of memory
Priority users get higher priority and more resources than free users on the bigmem partition
Condo users submit to the bigmem partition with normal or priority access (if they have a priority account in addition to their condo)
Using CCMake
Guide to build and compile software using CCMake.
Open-source software refers to any program whose source code is available for use or modification as users or other developers see fit. This is usually developed as a public collaboration and made freely available.
CMake and CCMake
Due to the complexity of some software, we often have to link to third party or external libraries. When working with software that has complicated building and linking steps, it is often impractical to use GCC (or your favorite compiler) directly. GNU Make is a build system that can simplify things somewhat, but "makefiles" can become unwieldy in their own way. Thankfully for us, there is a tool that simplifies this process.
CMake is a build system generator that one can use to facilitate the software build process. CMake allows one to specify—at a higher level than GNU Make—the instructions for compiling and linking our software. Additionally, CMake comes packaged with CCMake, which is an easy-to-use interactive tool that will let us provide build instructions to the compiler and the linker for projects written in C, Fortran, or C++. For more information about CMake and CCMake, please click .
Make sure the source code has a CMakeLists.txt file in the root folder
Getting the source code from a Git Repository
Much of the time, source code is available on platforms such as GitHub, GitLab or BitBucket. Cloning (or downloading) the project from any of those is the same process. First, you need to get the URL from the repository. It usually looks like this:
GitHub repository
Bitbucket repository
Where username indicates the GitHub (or BitBucket, etc) account of the owner of the project, and project_name indicates, well, try to guess.
GitHub and BitBucket have a button at the top right side of the repository web page labeled "clone". Copy that URL
Clone The Repository
Create a new folder on a path with the necessary read/write permissions
Go inside inside that folder
Clone the repository:
URL is the repository's link mentioned above.
Getting the source code from a .tar or .zip file
In case you downloaded the project from a different source, and it is contained in a .tar or .zip file. Just extract the source code on a folder with the necessary read/write permissions.
Build the Project
Create a new folder and name it build
Go inside inside that folder
Execute CCMake pointing to the root folder which has a CMakeLists.txt file
In this example, let's assume the build folder is at the same level as the CMakeLists.txt file.
The CCMake text interface will pop up with all the necessary attributes to build the software.
Set up the paths to the required libraries and press "c" to configure the project. Some errors might come up about CMake unable to find some specific libraries. This could be because that library does not exist in the system or you have not loaded the right module. Please contact CCV staff on how to fix this type of errors.
Make sure the attribute CMAKE_INSTALL_PREFIX points to a path with the necessary read/write permissions. By default it is set to the folder /usr/bin/ , which most of the users have no access to.
Once the configuration process has ended successfully, press "g" to generate the project. Generate the project does not mean compile or execute the program, please continue reading.
Compile the Project
Compile the project using the command make
You might want to increase the number of jobs compiling the software.
To speed up the compilation process, add the parameter "-j 8" to parallelize the job.
Once it is done, your project will be installed in the path set in the CMAKE_INSTALL_PREFIX attributeas explained above.
If you have any questions or need help please email [email protected].
Desktop App (VNC)
The Desktop app on Open OnDemand is a replacement for the older VNC Java client. This app allows you to launch a Desktop GUI on Oscar.
Advanced users looking for more resources can try the .
SSH Configuration File
How to save ssh configurations to a configuration file
When regularly connecting to multiple remote systems over SSH, you’ll find that remembering all the hosts and various command-line options becomes tedious. OpenSSH allows setting up a configuration file to store different SSH options for each remote machine you connect t.
SSH Config File Location
OpenSSH client-side (in this case your personal computer) configuration file is named config, and it is stored in the hidded .ssh
GPUs on Oscar
To view the various GPUs available on Oscar, use the command
nodes gpu
SMB (Local Mount)
CCV users can access their home, data, and scratch directories as a local mount on their own Windows, Mac, or Linux system using the Common Internet File System (CIFS) protocol (also called Samba). This allows you to use applications on your machine to open files stored on Oscar. It is also a convenient way to move files between Oscar and your own machine, as you can drag and drop files.
To use SMB you will need to be connected to the VPN. Please install the before proceeding.
Installing Python Packages
For Python 2, we recommend using the python/2.7.16 module. For Python 3, we recommend using the python/3.7.4 module.
Both these modules include the pipand virtualenvcommands, but do not include other common Python packages (e.g., SciPy, NumPy). This affords individual users complete control over the packages they are using.
There are several ways for users to install python packages on Oscar
Using Modules
CCV uses the package for managing the software environment on OSCAR. The advantage of the modules approach is that it allows multiple versions of the same software to be installed at the same time. With the modules approach, you can "load'' and "unload'' modules to dynamically control your environment.
Check out our !
MPI4PY
This page documents how to use the MPI for Python package within a Conda environment.
Using MPI4PY in a Python Script
The installation of mpi4py will be discussed in the following sections. This section provides an example of how mpi4py would be used in a python script after such an installation.
Users can check the nodes in a partition using this command nodes . As of May 2023 the Oscar cluster currently has following nodes in bigmem partition.
NODES CORES CPU/NODE MEM Features PARTITION
2 64 32-cores 2095GB 32core,intel,scalable,cascade,edr bigmem
2 64 32-cores 753GB 32core,intel,scalable,cascade,edr bigmem
All Oscar users have access to this partition, and can submit jobs to it. To submit batch jobs to large memory nodes, include the following in your batch script:
#SBATCH -p bigmem
To run an interactive job on large memory node, launch the interact command with the following flag:
$ interact -q bigmem
directory under your user’s home directory (i.e.,
~/.ssh
)
When you use the sshcommand for the first time. The ~/.ssh directory is automatically created. If the directory doesn’t exist on your system, create it using the command below:
By default, the SSH configuration file may not exist, so you may need to create it using the touch command :
This file must be readable and writable only by the user and not accessible by others:
SSH Config File Structure Basics
The SSH Config File takes the following structure:
The contents of the SSH config file is organized into sections. Each section starts with the Host directive and contains specific SSH options used when establishing a connection with the remote SSH server.
Oscar Hosts
Here we peovide a list of Oscar hosts and typical SSH configuration options. You have two options
Copy the list of hosts below directly into your SSH Config File (i.e., ~/.ssh/config)
Keep this content in a spearate file for Oscar hosts, lets say ~/.ssh/config.oscar and include that file in your main configuration file. In this case, the first line of ~/.ssh/config will be
Include "~/.ssh/config.oscar"
Don't forget to replace <username> with your user. Also the configuration assumes your identity key is ~/.ssh/id_rsa - if you named it anything else, please update the value. If you need to generate a key got here
Connecting to your preconfigured host
You may now connect using the shortchut notation provided by your configuration file. That is, all you need to type is:
According to the configuration above, this is equivalent to
We recommend using a virtual environment for your workflow if you preferpip. If you are a conda user we recommend managing your workflow with conda environments . You can load an anaconda module and then use conda.
In this document, we use angular brackets <> to denote command line options that you should replace with an appropriate value
Note for python3 packages replace python with python3and pip with pip3
Intel provides optimized packages for numerical and scientific work that you can install through pip or anaconda.
Using virtualenv
Virtual environments are a cleaner way to install python packages for a specific workflow. This webpage gives a good explanation of the use cases. In the example below, a virtual environment called 'my_cool_science' is set up in your home directory:
line 1: load the version of python you want to use
line 2: change directory to home
line 3: create the virtual environment
line 4: activate the virtual environment
line 5: install any packages you need for the virtual environment
line 6: deactivate the environment
When you want to use the environment, e.g. in a batch script or an interactive session
source ~/my_cool_science/bin/activate
When your work is finished, deactivate the environment with
deactivate
Install into your home directory
The --user flag will instruct pip to install to you home directory
This will install the package under the following path in user's HOME directory:
If you omit the --user flag you will see
This is because users do not have access to the default locations where software is installed.
Python packages can often have conflicting dependencies. For workflows that require a lot of python packages, we recommend using virtual environments.
Install at custom location
Users have a limit of 20GB for their home directories on Oscar. Hence, users might want to use their data directory instead for installing software. Another motivation to do that is to have shared access to the software among the whole research group.
This path to install location will have to be added to the PYTHONPATH environment variable so that python can find the python modules to be used. This is not necessary for software installed using the --user option.
This can be added at the end of your .bashrc file in your home directory. This will update the PYTHONPATH environment variable each time during startup. Alternatively, you can update PYTHONPATH in your batch script as required. This can be cleaner as compared to the former method. If you have a lot of python installs at different locations, adding everything to PYTHONPATH can create conflicts and other issues.
A caveat of using this method is that pip will install the packages (along with its requirements) even if the package required is already installed under the global install or the default local install location. Hence, this is more of a brute force method and not the most efficient one.
For example, if your package depends on numpy or scipy, you might want to use the numpy and scipy under our global install as those have been compiled with MKL support. Using the --target option will reinstall numpy with default optimizations and without MKL support at the specified location.
Installing from source
Sometimes, python software is not packaged by the developers to be installed by pip. Or, you may want to use the development version which has not been packaged. In this case, the python package can be installed by downloading the source code itself. Most python packages can be installed by running the setup.py script that should be included in the downloaded files.
You will need to provide a "prefix path" for the install location
This will create the sub-directories bin, lib, etc. at the location provided above and install the packages there. The environment will have to be set up accordingly to use the package:
To use MPI in a python script through mpi4py, you must first import it using the following code:
Example Script
Here is an example python script mpi4pytest.py that uses MPI:
The file mpi4pytest.pycan be found at /gpfs/runtime/softwareexamples/mpi4py/
Once you have activated your conda environment, run the following commands to install mpi4py:
You may change the python version in the pip command.
To check that the installation process was a success you can run
If no errors result from running the command, the installation has worked correctly.
Here is an example batch job script mpi4pytest_conda.sh that uses mpi4pytest.py and the conda environment setup:
The example script above runs the python script on two nodes by using the #SBATCH -N 2 command. For more information on #SBATCH options, see our documentation.
Python Virtual Environment
Start by creating and activating a python virtual environment:
Once you have activated your conda environment, run the following command to install mpi4py:
Here is an example batch job script mpi4pytest_virtualenv.sh and the python virtual environment setup:
mkdir -p ~/.ssh && chmod 700 ~/.ssh
touch ~/.ssh/config
chmod 600 ~/.ssh/config
Host hostname1
SSH_OPTION value
SSH_OPTION value
Host hostname2
SSH_OPTION value
Host *
SSH_OPTION value
# Oscar Hosts. Any hosts with the -campus suffix can be accessed
# only whithin Brown network i.e. campus or vpn
# Hosts without -campus sufix can be accessed from outside Brown
# but will requiere 2FA
# Hosts to connect to login nodes
Host oscar
HostName ssh.ccv.brown.edu
User <username>
IdentityFile ~/.ssh/id_rsa
ForwardAgent yes
ForwardX11 yes
TCPKeepAlive yes
ServerAliveCountMax 20
ServerAliveInterval 15
Host oscar-campus
HostName sshcampus.ccv.brown.edu
User <username>
IdentityFile ~/.ssh/id_rsa
ForwardAgent yes
ForwardX11 yes
TCPKeepAlive yes
ServerAliveCountMax 20
ServerAliveInterval 15
# These are jumphosts for vscode, we don't use them directly.
# For connecting your IDE see the next section
Host desktop-oscar-campus
HostName desktop.ccv.brown.edu
IdentityFile ~/.ssh/id_rsa
User <username>
ForwardAgent yes
Host desktop-oscar
HostName ssh8.ccv.brown.edu
IdentityFile ~/.ssh/id_rsa
User <username>
# When connecting from VSCODE use the following hosts
Host vscode-oscar-campus
HostName oscar2
User <username>
ProxyCommand ssh -q -W %h:%p desktop-oscar-campus
Host vscode-oscar
HostName oscar2
User <username>
ProxyCommand ssh -q -W %h:%p desktop-oscar
Do not load any anaconda module in your .modules or .bashrc file. These modules prevent Desktop sessions from starting correctly. You may load them inside the Desktop session.
Launching Desktop App (VNC)
0. Launch Open OnDemand
Click here to launch Open OnDemand (OOD) and log in with you Brown Credentials.
1. Select the Desktop option in Interactive Apps dropdown list:
2. Choose the resource option:
3. Wait and Launch!
You may change the Image Quality if your internet connection is bad. Image quality can be changed in the middle of the session.
Reconnecting to session
A session may get disconnected if it is not active for a while:
If the session disconnects as shown above, please don't click the "Connect" button on the screen. You may go to Open OnDemand page and click “My Interactive Sessions” to find the session again:
Please don’t launch a new session if you have an existing session. You cannot launch two desktop sessions at the same time.
Sometimes, the “My interactive Sessions” button is shortened to look like:
Copying and pasting text
If you are using Google Chrome, switch on the "Clipboard" permission and you can directly copy and paste text into the OOD Desktop from any other program.
Click the Lock icon to the left of the URL
Switch on the "Clipboard" permission
Click the side panel button on the extreme left hand side of the screen.
To copy text into the Desktop session, paste the data into the Clipboard. It will be available to paste inside the Desktop session.
To copy text from the Desktop session, open the Clipboard. The copied text will be displayed inside it. You can select and copy the text inside the Clipboard and paste it to an external program.
Desktop (Advanced)
If you need more or differentresources than those available from the default Desktop session, you should use the Advanced Desktop app. Resources requested here count against the resources allowed for your Oscar account.
1. Select the Desktop (Advanced) app under Interactive Apps.
2. Choose required resources
Fill out the form with your required resources.
Account: Enter your condo account name. If you are not a member of a condo, leave this blank
Desktop Environment: Choose XFCE. KDE works for CPU jobs, but may not be able to use GPU acceleration correctly.
Number of hours: Choose appropriately. Your Desktop session will end abruptly after this time has lapsed. Requesting a very long session will result in a lower job priority.
Partition: Equivalent to #SBATCH -p option. The desktop session will run on this partition.
Num Cores: Equivalent to the #SBATCH -n option.
Num GPUs: Equivalent to the #SBATCH --gres=gpu: option. This field is ignored if the partition does not have any GPU nodes, e.g. batch
Memory (GB): Equivalent to the#SBATCH --mem= option.
Reservation: Equivalent to the #SBATCH --reservation= option. Leave blank if you are not using a reservation.
3. Wait and Launch!
Wait and launch this session like the regular Desktop session.
To start an interactive session on a GPU node, use the interact command and specify the gpu partition. You also need to specify the requested number of GPUs using the -g option:
To start an interactive session on a particular GPU type (QuadroRTX, 1080ti, p100 etc) use the feature -f option:
GPU Batch Job
For production runs, please submit a batch job to the gpu partition. E.g. for using 1 GPU:
This can also be mentioned inside the batch script:
You can view the status of the gpu partition with:
Sample batch script for CUDA program:
Getting started with GPUs
While you can program GPUs directly with CUDA, a language and runtime library from NVIDIA, this can be daunting for programmers who do not have experience with C or with the details of computer architecture.
You may find the easiest way to tap the computation power of GPUs is to link your existing CPU program against numerical libraries that target the GPU:
CUBLAS is a drop-in replacement for BLAS libraries that runs BLAS routines on the GPU instead of the CPU.
CUFFT, CUSPARSE, and CURAND provide FFT, sparse matrix, and random number generation routines that run on the GPU.
combines custom GPU kernels, CUBLAS, and a CPU BLAS library to use both the GPU and CPU to simultaneously use both the GPU and CPU; it is available in the 'magma' module on Oscar.
Matlab has a feature, available through the Parallel Computing Toolkit, for creating arrays on the GPU and operating on them with many built-in Matlab functions. The PCT toolkit is licensed by CIS and is available to any Matlab session running on Oscar or workstations on the Brown campus network.
is an interface to CUDA from Python. It also has a feature and is available in the cuda module on Oscar.
OpenACC
OpenACC is a portable, directive-based parallel programming construct. You can parallelize loops and code segments simply by inserting directives - which are ignored as comments if OpenACC is not enabled while compiling. It works on CPUs as well as GPUs. We have the PGI compiler suite installed on Oscar which has support for compiling OpenACC directives. To get you started with OpenACC:
To submit interactive job to NVLink Enabled GPU nodes:
To submit batch job(s) add following line to your batch script.
A user's Windows machine is required to have Crowdstrike Home installed on the Windows machine to use SMB.
Users should ensure that the date and time are set correctly on their machine. Now you are ready to mount your CCV directories locally. Instructions for each of the various operating systems are given below.
macOS
In the Finder, press "Command + K" or select "Connect to Server..."
from the "Go" menu.
For "Server Address", enter smb://smb.ccv.brown.edu/<volume>/<user>
and click "Connect".
To access your Home directory, entersmb://smb.ccv.brown.edu/home/<user>
To access your Scratch space, entersmb://smb.ccv.brown.edu/scratch/<user>
To access your Data directory, entersmb://smb.ccv.brown.edu/data/<pi_group>/<user>
To check your PI group run 'groups' command.
Enter your AD username and password.
You may choose to add your login credentials to your keychain so you will not need to enter this again.
Optional. If you would like to automatically connect to the share at startup:
Open "System Preferences" (leave the Finder window open).
Go to "Accounts" > "(your account name)".
Select "Login Items".
Drag your data share from the "Finder" window to the "Login Items" window.
Linux
Install the cifs-utils package:
Make a directory to mount the share into:
Create a credentials file and add your AD account information:
Allow only root access to the credentials files:
Add an entry to thefstab:
Thefstabentry should be following:
Replace<localUser>to the login used on your Linux workstation, and replace <user> and <pi_group> with your Oscar username and PI group, respectively.
Mount the share:
Windows
Right-click "Computer" and select "Map Network Drive"
Select an unassigned drive letter
To mount specific volumes:
For Home directory, enter\\smb.ccv.brown.edu\home\<user>
For Scratch space, enter\\smb.ccv.brown.edu\scratch\<user>
For Data directory, enter\\smb.ccv.brown.edu\data\<pi_group>\<user>
To check your<pi_group>run 'groups' command.
Check "Connect using different credentials"
Click "Finish"
Enter your AD user name. If your computer is not in Active Directory (AD), you should enter your username in the format ad\username
Enter your AD password and click "OK"
You can now access your home directory through Windows Explorer with the assigned drive letter. Your data and scratch directories are available as the subdirectories (~/data and ~/scratch) of your home directory.
Lists all modules that are currently loaded in your software environment.
module avail
Lists all available modules on the system. Note that a module can have multiple versions.
module help <name>
Prints additional information about the given software.
Finding modules
The module avail command allows searching modules based on partial names. For example:
will list all available modules whose name starts with "bo".
Output:
This feature can be used for finding what versions of a module are available.
Auto-completion using tab key
The module load command supports auto-completion of the module name using the "tab" key. For example, writing module load bo"on the shell prompt and hitting "tab" key a couple of times will show results similar to that shown above. Similarly, the module unload command also auto completes using the names of modules which are loaded.
Modules loaded at startup
You can customize the default environment that is loaded when you login by putting the appropriate module commands in the .modules file in your home directory. For instance, if you edited your .modules file to contain
python/3.5.2 will be loaded every time you log in.
If you have not yet created a .modulesfile in your home directory, you will have to create one and then add the commands you wish to have as defaults to that file.
What modules actually do...
Loading a module sets the relevant environment variables like PATH, LD_LIBRARY_PATH and CPATH. For example, PATH contains all the directory paths (colon separated) where executable programs are searched for. So, by setting PATH through a module, now you can execute a program from anywhere in the file-system. Otherwise, you would have to mention the full path to the executable program file to run it which is very inconvenient. Similarly, LD_LIBRARY_PATH has all the directory paths where the run time linker searches for libraries while running a program, and so on. To see the values in an environment variable, use the echo command. For instance, to see what's in PATH:
MPI is a standard that dictates the semantics and features of "message passing". There are different implementations of MPI. Those installed on Oscar are
MVAPICH2
OpenMPI
We recommend using MVAPICH2 as it is integrated with the SLURM scheduler and optimized for the Infiniband network.
MPI modules on Oscar
The MPI module is called "mpi". The different implementations (mvapich2, openmpi, different base compilers) are in the form of versions of the module "mpi". This is to make sure that no two implementations can be loaded simultaneously, which is a common source of errors and confusion.
You can just use "module load mpi" to load the default version which is mpi/openmpi_4.0.7_gcc_10.2_slurm22. This is the recommended version.
The module naming format is
srun instead of mpirun
Use srun --mpi=pmix or srun --mpi=pmi2 to run MPI programs. All MPI implementations listed with suffix _slurm22 are built with SLURM support. Hence, the programs need to be run using SLURM's srun command, except if you are using the above mentioned legacy versions.
The --mpi=pmix flag is also required to match the configuration with which MPI is installed on Oscar.
Running MPI programs - Interactive
To run an MPI program interactively, first create an allocation from the login nodes using the salloc command:
For example, to request 4 cores to run 4 tasks (MPI processes):
Once the allocation is fulfilled, you can run MPI programs with the srun command:
When you are finished running MPI commands, you can release the allocation by exiting the shell:
Also, if you only need to run a single MPI program, you can skip the salloc command and specify the resources in a single sruncommand:
This will create the allocation, run the MPI program, and release the allocation.
Note: It is not possible to run MPI programs on compute nodes by using the interact command.
salloc documentation:
srun documentation:
Running MPI programs - Batch Jobs
Here is a sample batch script to run an MPI program:
Hybrid MPI+OpenMP
If your program has multi-threading capability using OpenMP, you can have several cores attached with a single MPI task using the --cpus-per-task or -c option with sbatch or salloc. The environment variable OMP_NUM_THREADS governs the number of threads that will be used.
The above batch script will launch 4 MPI tasks - 2 on each node - and allocate 4 CPUs for each task (total 16 cores for the job). Setting OMP_NUM_THREADS governs the number of threads to be used, although this can also be set in the program.
Performance Scaling
The maximum theoretical speedup that can be achieved by a parallel program is governed by the proportion of sequential part in the program (Amdahl's law). Moreover, as the number of MPI processes increases, the communication overhead increases i.e. the amount of time spent in sending and receiving messages among the processes increases. For more than a certain number of processes, this increase starts dominating over the decrease in computational run time. This results in the overall program slowing down instead of speeding up as number of processes are increased.
Hence, MPI programs (or any parallel program) do not run faster as the number of processes are increased beyond a certain point.
If you intend to carry out a lot of runs for a program, the correct approach would be to find out the optimum number of processes which will result in the least run time or a reasonably less run time. Start with a small number of processes like 2 or 4 and first verify the correctness of the results by comparing them with the sequential runs. Then increase the number of processes gradually to find the optimum number beyond which the run time flattens out or starts increasing.
Maximum Number of Nodes for MPI Programs
An MPI program is allowed to run on at most 32 nodes. When a user requests more than 32 nodes for an MPI program/job, the user will receive the following error:
Batch job submission failed: Requested node configuration is not available
X-Forwarding
Instructions to forward X11 applications from Oscar to local computer
If you have an installation of X11 on your local system, you can access Oscar with X forwarding enabled, so that the windows, menus, cursor, etc. of any X applications running on Oscar are all forwarded to your local X11 server. Here are some resources for setting up X11:
One limitation of X forwarding is its sensitivity to your network connection's latency. We advise against using X forwarding from a connection outside of the Brown campus network, since you will likely experience lag between your actions and their response in the GUI.
Mac/Linux
Once your X11 server is running locally, open a terminal and use
to establish the X forwarding connection. Then, you can launch GUI applications from Oscar and they will be displayed locally on your X11 server.
Windows (PuTTY)
For Windows users using PuTTY, enable X forwarding under Connections->SSH->X11:
Submitting GPU Jobs
The Oscar GPUs are in a separate partition to the regular compute nodes. The partition is called gpu. To see how many jobs are running and pending in the gpu partition, use
allq gpu
Interactive use
To start an session on a GPU node, use the interact command and specify the gpu partition. You also need to specify the requested number of GPUs using the -g option:
Batch jobs
Here is an example batch script for a cuda job that uses 1 gpu and 1 cpu for 5 minutes
To submit this script:
DGX GPU Nodes in the GPU-HE Partition
All the nodes in the gpu-he partition have V100 GPUs. However, two of them are DGX nodes (gpu1404/1405) which have 8 GPUs. When a gpu-he job requests for more than 4 GPUs, the job will automatically be allocated to the DGX nodes.
The other non-DGX nodes actually have a better NVLink interconnect topology as all of them have direct links to the other. So the non-DGX nodes are better for a gpu-he job if the job does not require more than 4 GPUs.
Python in batch jobs
By default, print in Python is buffered. When running Python in a batch job in SLURM you may see output less often than you would when running interactively. This is because the output is being buffered - the print statements are collected until there is a large amount to print, then the messages are all printed at once. For debugging or checking that a Python script is producing the correct output, you may want to switch off buffering.
Switch off buffering
For a single python script you can use the -u option, e.g.
python -u my_script.py
The -u stands for "unbuffered". You can use the environment variable PYTHONUNBUFFERED to set unbuffered I/O for your whole batch script.
There is some performance penalty for having unbuffered print statements, so you may want to reduce the number of print statements, or run buffered for production runs.
VASP
The Vienna Ab initio Simulation Package (VASP) is a package for performing advanced mechanical computations. This page will explain how VASP can be accessed and used on Oscar.
Setting up VASP
In order to use VASP, you must be a part of the vasp group on Oscar. To check your groups, run the groups command in the terminal.
First, you must choose which VASP module to load. You can see the available modules using module avail vasp. You can load your preferred VASP module using module load <module-name>.
Available Versions
VASP 5.4.1
VASP 5.4.4
VASP 6.1.1
Running VASP
Within a batch job, you should specify the number of MPI tasks as
If you would like 40 cores for your calculation, you would include the following in your batch script:
If you're not sure how many cores you should include in your calculation, refer to
Configuring Remote Launch
Configuring Remote Launch from the client
You will need to configure remote launch for Oscar
Add /gpfs/runtime/opt/forge/19.1.2 as the Remote Installation Directory
Test Remote Launch. You should enter the password used for Oscar. If successful you should see the message Remote Launch test completed successfully
If you have a mismatch between your client version on the version of Forge on Oscar you will see an error message. To fix this make sure you are using compatible client and remote versions
Once you are connected you will see a Licence checked out and "Connected to [email protected]' on the client.
Interactive Jobs
To start an interactive session for running serial or threaded programs on an Oscar compute node, simply run the command interact from the login node:
interact
By default, this will create an interactive session that reserves 1 core and 4GB of memory for a period of 30 minutes. You can change the resources reserved for the session from these default limits by modifying the interact command:
usage: interact [-n cores] [-t walltime] [-m memory] [-q queue]
[-o outfile] [-X] [-f featurelist] [-h hostname] [-g ngpus]
Starts an interactive job by wrapping the SLURM 'salloc' and 'srun' commands.
options:
-n cores (default: 1)
-t walltime as hh:mm:ss (default: 30:00)
-m memory as #[k|m|g] (default: 4g)
-q queue (default: 'batch')
-o outfile save a copy of the sessions output to outfile (default: off)
-X enable X forwarding (default: no)
-f featurelist CCV-defined node features (e.g., 'e5-2600'),
combined with '&' and '|' (default: none)
-h hostname only run on the specific node 'hostname'
(default: none, use any available node)
-a account user SLURM accounting account name
-g ngpus number of GPUs
For example, the command
requests an interactive session with 20 cores and 10 GB of memory (per node) for a period of 1 hour.
Keeping Interactive Jobs Alive:
If you lose connectivity to your login node, you lose access to your interactive job. To mitigate this issue you can use screen to keep your connection alive. For more information on using screen on the login nodes, see the
IDL
Interactive Data Language (IDL) is a programming language used for data analysis and is popular in several scientific fields. This page explains how to use the IDL module on Oscar run IDL programs.
Setting Up IDL
First load the IDL module that you want to use with module load idl/version_number:
From Non-compliant Networks (2-FA)
Accessing VSCode from Non-Brown compliant networks
This guide is only for users connecting from Non-Brown Compliant Networks. 2-FA is mandatory.
Install the for VSCode
Intro to Parallel Programming
This page serves as a guide for application developers getting started with parallel programming, or users wanting to know more about the working of parallel programs/software they are using.
Although there are several ways to classify parallel programming models, a basic classification is:
Distributed Memory Programming
Condo/Priority Jobs
Note: we do not provide users condo access by default if their group/PI has a condo on the system. You will have to explicitly request a condo access and we will ask for approval from the PI.
To use your condo account to submit jobs, include the following line in your batch script:
You can also provide this option on the command line while submitting the job using sbatch:
# Jump box with public IP address
Host jump-box
HostName poodcit4.services.brown.edu
User <username>
# Target machine with private IP address
Host ccv-vscode-node
HostName node1103
User <username>
ProxyCommand ssh -q -W %h:%p jump-box
Adds a module to your current environment. If you load using just the name of a module, you will get the default version. To load a specific version, load the module using its full name with the version: "module load gcc/6.2"
This model is useful when all threads/processes have access to a common memory space. The most basic form of shared memory parallelism is Multithreading. According to Wikipedia, a thread of execution is the smallest sequence of programmed instructions that can be managed independently by a scheduler (Operating System).
Note that most compilers have inherent support for multithreading up to some level. Multithreading comes into play when the compiler converts your code to a set of instructions such that they are divided into several independent instruction sequences (threads) which can be executed in parallel by the Operating System. Apart from multithreading, there are other features like "vectorized instructions" which the compiler uses to optimize the use of compute resources. In some programming languages, the way of writing the sequential code can significantly affect the level of optimization the compiler can induce. However, this is not the focus here.
Multithreading can also be induced at code level by the application developer and this is what we are interested in. If programmed correctly, it can also be the most "efficient" way of parallel programming as it is managed at the Operating System level and ensures optimum use of "available" resources. Here too, there are different parallel programming constructs which support multithreading.
Pthreads
POSIX threads is a standardized C language threads programming interface. It is a widely accepted standard because of being lightweight, highly efficient and portable. The routine to create Pthreads in a C program is called pthread_create and an "entry point" function is defined which is to be executed by the threads created. There are mechanisms to synchronize the threads, create "locks and mutexes", etc. Help pages:
OpenMP is a popular directive based construct for shared memory programming. Like POSIX threads, OpenMP is also just a "standard" interface which can be implemented in different ways by different vendors.
Compiler directives appear as comments in your source code and are ignored by compilers unless you tell them otherwise - usually by specifying the appropriate compiler flag (https://computing.llnl.gov/tutorials/openMP). This makes the code more portable and easier to parallelize. you can parallelize loop iterations and code segments by inserting these directives. OpenMP also makes it simpler to tune the application during run time using environment variables. for example, you can set the number of threads to be used by setting the environment variable OMP_NUM_THREADS before running the program. Help pages:
This policy establishes requirements for the use and management of Brown University's Center for Computation and Visualization resources.
Policy Purpose
This policy establishes requirements for the use and management of Brown University’s Center for Computation and Visualization resources to ensure their integrity, confidentiality and availability in support of appropriate education, research, outreach, and administrative objectives.
Account Usage
HPC users are not permitted to:
Share their accounts or passwords with others or enable unauthorized users to access Center for Computation and Visualization resources.
Use Center for Computation and Visualization resources for personal economic gain.
Engage in unauthorized activity (Mining etc.) that intentionally impacts integrity of resources.
Storage
Each user (premium or exploratory) gets 20GB Home Directory, 512GB short-term Scratch, and 256G Data directory (shared amongst the members of group)
Files in Scratch Directory not accessed for last 30 days are automatically purged. CCV only stores snapshots for 7 days after that files will be automatically deleted.
PI has the ultimate access to Data Directory - if a student leaves Brown the files in Data directory will be owned by the PI.
Software and Data
All software and data stored or used on Center hosted systems must be appropriately and legally acquired and must be used in compliance with applicable licensing terms. Unauthorized misuse or copying of copyrighted materials is prohibited.
Data Retention
CCV reserves the right to remove any data at any time and/or transfer data or other individuals (such as Principal Investigators working on a same or similar project) after a user account is deleted is no longer affiliated with Brown University.
Accounts Validity
Accounts are valid for two years or validity of Brown AD credentials.
You can use the command module load idl to simply load the default version. This is demonstrated in the following command followed by system dialogue.
As indicated by the system dialogue, you will need to enter the following command to set up the environment for IDL:
IDL Command Line
Once you've set up IDL in the way outlined above, you can open the IDL command line by simply using the command idl:
Note: To exit this environment, simply use the command exit
As is stated in the IDL Documentation, IDL in command-line mode "uses a text-only interface and sends output to your terminal screen or shell window." Thus, this is a mode in which you can enter commands and see their results in real time, but it is not where one should write full IDL programs.
IDL Programs
To write anIDL program, you can use any of the text editors on Oscar (such as vim, emacs, and nano) or you can create the program in a file on your own computer and then copy that file to Oscar when you are finished. Here is an example (hello world) IDL program idl_hello_world.pro:
This file and the batch file below can be found at /gpfs/runtime/software_examples/idl/8.5.1 if you wish to copy them and test the process yourself.
Once you have the .pro file on Oscar, you can then run this file using a batch script. Here is a bare bones version of a batch script (called idl_hello_world.sh)that will run the script idl_hello_world.pro (note that the .pro is omitted in the script).
We can then run the batch file by using the sbatch command:
Similarly, you can change the account while asking for interactive access too:
Condo account names are typically <groupname>-condo, and you can view a full list with the condos command on Oscar.
To see the running and pending jobs in a condo:
condo <condo-name>
Premium Account (priority) jobs
If you have a premium account, that should be your default QOS for submitting jobs. You can check if you have a premium account with the command groups. If you have a priority account you will see priority in your the output form groups.
You can check the qos for a running job by running the command myq. The QOS column should show "pri-<username>"
If you are interested in seeing all your accounts and associations, you can use the following command:
#!/bin/bash
# Request an hour of runtime:
#SBATCH --time=1:00:00
# Use 2 nodes with 8 tasks each, for 16 MPI tasks:
#SBATCH --nodes=2
#SBATCH --tasks-per-node=8
# Specify a job name:
#SBATCH -J MyMPIJob
# Specify an output file
#SBATCH -o MyMPIJob-%j.out
#SBATCH -e MyMPIJob-%j.err
# Load required modules
module load mpi
srun --mpi=pmix MyMPIProgram
#!/bin/bash
# Use 2 nodes with 2 tasks each (4 MPI tasks)
# And allocate 4 CPUs to each task for multi-threading
#SBATCH --nodes=2
#SBATCH --tasks-per-node=2
#SBATCH --cpus-per-task=4
# Load required modules
module load mpi/openmpi_4.0.7_gcc_10.2_slurm22
export OMP_NUM_THREADS=4
srun --mpi=pmix ./MyMPIProgram
$ ssh -X <user>@ssh.ccv.brown.edu
#!/bin/bash
# Request a GPU partition node and access to 1 GPU
#SBATCH -p gpu --gres=gpu:1
# Request 1 CPU core
#SBATCH -n 1
#SBATCH -t 00:05:00
# Load a CUDA module
module load cuda
# Run program
./my_cuda_program
$ module load idl
module: loading 'idl/8.5.1'
module: idl: License owned by Jonathan Pober. Set up the environment for IDL by running: "shopt -s expand_aliases; source $IDL/envi53/bin/envi_setup.bash".
$ idl
IDL Version 8.5.1 (linux x86_64 m64). (c) 2015, Exelis Visual Information Solutions, Inc., a subsidiary of Harris Corporation.
Installation number: 5501393-2.
Licensed for use by: Brown University
IDL>
Search for symlink and make sure the symlink searching is unchecked
3. Under VSCode settings, search for remote ssh timeout and manually enter a timeout value i.e. 50s. It should give you enough time to complete 2-Factor Authentication.
4. Edit the ~/.ssh/config file on your local machine, add the following lines. Replace <username> with your Oscar username.
6. In VSCode, select Remote-SSH: Connect to Host… and after the list populates select login-node
Most inquiries can be directed to CCV’s support address, [email protected], which will create a support ticket with one of our staff.
What are the fees for CCV services?
All CCV services are billed quarterly, and rates can be found (requires a Brown authentication to view). Questions about rates should be directed to [email protected].
How do I acknowledge CCV in a research publication?
We greatly appreciate acknowledgements in research publications that benefited from the use of CCV services or resources.
Oscar
What is Oscar?
Oscar is our primary research computing cluster with several hundred multi-core nodes sharing a high-performance interconnect and file system. Applications can be run interactively or scheduled as batch jobs.
How do I request an account on Oscar?
To request an account, please fill out a All accounts are subject to our .
How do I run a job on Oscar?
Sample batch scripts are available in your home directory at ~/batch_scripts and can be run with the sbatch <jobscript> command. For more information, visit our manual page on .
Can I use Oscar for teaching?
See our page on
How do I find out when the system is down?
We post updates to our user mailing list, [email protected] which you are automatically subscribed to when setting up an account with CCV. If you need to be added to the mailing list, please submit a support ticket to [email protected]. We also have an announcement mailing list for office hours, workshops and other events relevant to CCV users, ccv-announce.listserve.brown.edu.
How do I run a job array on Oscar?
A job array is a special type of job submission that allows you to submit many related batch jobs with a single command. This makes it easy to do parameter sweeps or other schemes where the submitted jobs are all the same except for a single parameter such as a filename or input variable. Job arrays require special syntax in your job script. Sample batch scripts for job arrays are available in your home directory at ~/batch_scripts and can be run with the sbatch <jobscript> command. For more information, visit our manual page on .
How do I run a MPI job on Oscar?
MPI is a type of programming interface. Programs written with MPI can run on and communicate across multiple nodes. You can run MPI-capable programs by calling srun --mpi=pmix <program> in your batch script. For more detailed info, visit our manual page on .
I have some MPI-enabled source code. How can I compile it on Oscar?
Load an mpi module module load mpi. For a list of mpi modules available, module avail mpi
What applications are available on Oscar?
Many scientific and HPC software packages are already installed on Oscar, including python, perl, R, Matlab, Mathematica, and Maple. Use the module avail command on Oscar to view the whole list or search for packages. See our manual page on to understand how software modules work. Additional packages can be requested by submitting a support ticket to [email protected].
What compilers are available on Oscar?
By default, the gcc compiler is available when you login to Oscar, providing the GNU compiler suite of gcc (C), g++ (C++), and gfortran. We also provide compilers from Intel (intel module) and the Portland Group (pgi module). For more information, visit our manual page on .
How do I get information about finished jobs?
The sacct command will list all of your completed jobs since midnight of the previous day (as well as running and queued jobs). You can pick an earlier start date with the -S option, e.g. sacct -S 2012-01-01.
How much storage am I using?
The myquota command on Oscar will print a summary of your usage on the home, data, and scratch file systems. For more information, see our manual page on .
My job keeps terminating unexpectedly with a "Killed" message, or without any errors. What happened?
These are symptoms of not requesting enough memory for your job. The default memory allocation is about 3 GB. If your job is resource-intensive, you may need to specifically allocate more. See the for instructions on requesting memory and other resources.
How do I request a certain amount of memory per CPU?
Specify the SLURM option --mem-per-cpu= in your script.
How do I link against a BLAS and LAPACK library?
We recommend linking against the Intel Math Kernels Library (MKL) which provides both BLAS and LAPACK. The easiest way to do this on Oscar is to include the special environment variable $MKL at the end of your link line, e.g. gcc -o blas-app blas-app.c $MKL. For more complicated build systems, you may want to consult the .
RUNNING JOBS
How is a job identified?
By a unique JobID, e.g. 13180139
Which of my jobs are running/pending?
Use the command myq
How do I check the progress of my running job?
You can look at the output file. The default output file is slurm-%j.out" where %j is the JobID. If you specified and output file using #SBATCH -o output_filename and/or an error file #SBATCH -e error_filename you can check these files for any output from your job. You can view the contents of a text file using the program less , e.g.
Use the spacebar to move down the file, b to move back up the file, and q to quit.
My job is not running how I indented it too. How do I cancel the job?
scancel <JobID> where <JobID> is the job allocation number, e.g. 13180139
How do I save a copy of an interactive session?
You can use interact -o outfile to save a copy of the session's output to "outfile"
I've submitted a bunch of jobs. How do I tell which one is which? myq will list the running and pending jobs with their JobID and the name of the job. The name of the job is set in the batch script with #SBATCH -J jobname. For jobs that are in the queue (running or pending) you can use the command scontrol show job <JobID> where <JobID> is the job allocation number, e.g.13180139 to give you more detail about what was submitted.
How do I ask for a haswell node?
Use the --constraint (or -C) option:
You can use the --constraint option restrict your allocation according to other features too. The nodes command provides a list of "features" for each type of node.
Why won't my job start?
When your job is pending (PD) in the queue, SLURM will display a reason why your job is pending. The table below shows some common reasons for which jobs are kept pending.
Reason
Meaning
Why is my job taking so long to start? Just waiting in (Priority) or (Resources)
Overall system busy: when tens of thousands of jobs are submitted it total by all users, the time it takes SLURM to process these into the system may increase from the normal almost instantly to a half-hour or more.
Specific resource busy: if you request very specific resources (e.g., a specific processor) you then have to wait for that specific resource to become available while other similar resources may be going unused.
Specified resource not available: if you request something that is not or may never be available, your job will simply wait in the queue. E.g., requesting 64 GB of RAM on a 64 GB node will never run because the system needs at least 1 GB for itself so you should reduce your request to less than 64.
TRANSFERRING FILES
How do I transfer big files to/from Oscar?
Please use the server transfer.ccv.brown.edu
Transfer local file to Oscar:
2. Transfer remote file on Oscar to the local system:
Alternatively, Oscar has an endpoint for "Globusonline" () that you can use to more effectively transfer files. See our manual page on how to use to transfer files.
Cloud HPC Options
The use of cloud resources for HPC varies according to your demands and circumstances. Cloud options are changing rapidly both in service providers and various services being offered. For those who have short-term needs that don't demand the highest of computational performance, a cloud option might be appropriate. For others, a local option customized to individual needs may be better. The cost of cloud services also varies quite a bit and includes not only compute time but data transfer charges. Other issues involved licensing, file synchronization, etc.
We are actively investigating a number of options to connect Brown users seamlessly to suitable cloud options. We are collecting such information for publishing on the CIS website as part of research services available. At this point, the best course of action is to request an individual consultation to help address your specific needs. Please send email to support@ccv. brown.edu.
Arbiter2
Arbiter2 is a cgroups-based mechanism that is designed to prevent the misuse of login nodes and VSCode node, which are scarce, shared resources. It is installed on shared nodes listed below:
login005
login006
node1103 (VSCode)
Status and Limits
Arbiter2 applies different limits to a user's processes depending on the user's status: normal, penalty1, and penalty2.
Arbiter2 limits apply only to the shared nodes, not compute nodes.
Normal Status and Limits
Upon first log in, the user is in the normal status. These normal limits apply to all the user's processes on the node:
1/3 of the total CPU time. For example, a user's processes can use up to 1/3 of the total CPU time of the 24 cores on a login node.
40GB
Penalty1 Status and Limits
When a user's processes consume CPU time more than the default CPU time limit for a period of time, the user's status is changed to the penalty1 status. These penalty1 limits are applied:
80% of the normal limit.
0.8 * 40GB = 32GB (80% of the normal limit)
While a user is in penalty1 status, their processes are throttled if they consume more CPU time than penalty1 limit. However, if a user's processes exceed penalty1 memory limit, the processes (PIDs) will be terminated by cgroups.
The user's status returns to the normal status after a user's processes consume CPU time less than the penalty1 limit for 30 minutes.
Penalty restrictions are enforced independently for each shared node, and the penalty status does not carry over between these nodes.
Penalty2 Status and Limits
When a user's processes consume more CPU time than the penalty1 limit for a period of time, the user is put in the penalty2 status, and the penalty2 limits apply to the user's processes.
50% of the normal limit
20GB (50% of the normal limit)
In penalty2 status, the user's processes will be throttled if they consume more CPU time than penalty2 limit. However, if a user's processes exceed penalty2 memory limit, the processes (PIDs) will be terminated by cgroups.
The user's status returns to the normal status after a user's processes consume CPU time less than the penalty2 limit for one hour.
Penalty3 Status and Limits
When a user's processes consume more CPU time than the penalty2 limit for a period of time, the user is put in the penalty3 status. These penalty3 limits apply to the user's processes.
30% of the normal limit
12GB (30% of the normal limit)
In penalty3 status, the user's processes will be throttled if they consume more CPU time than penalty3 limit. If a user's processes exceed penalty3 memory limit, the processes (PIDs) will be terminated by cgroups
The user's status returns to the normal status after a user's processes consume CPU time less than the penalty3 limit for two hours.
Email Notification
A user receives an email notification upon each violation. Below is a example email:
Violation of usage policy
A violation of the usage policy by ccvdemo (CCV Demo,,,,ccvdemo) on login006 was automatically detected starting at 08:53 on 04/25.
This may indicate that you are running computationally-intensive work on the interactive/login node (when it should be run on compute nodes instead). Please utilize the 'interact' command to initiate a SLURM session on a compute node and run your workloads there.
You now have the status penalty1 because your usage has exceeded the thresholds for appropriate usage on the node. Your CPU usage is now limited to 80% of your original limit (8.0 cores) for the next 30 minutes. In addition, your memory limit is 80% of your original limit (40.0 GB) for the same period of time.
These limits will apply on login006.
High-impact processes
Usage values are recent averages. Instantaneous usage metrics may differ. The processes listed are probable suspects, but there may be some variation in the processes responsible for your impact on the node. Memory usage is expressed in GB and CPU usage is relative to one core (and may exceed 100% as a result).
Process
Average core usage (%)
Average memory usage (GB)
Recent system usage
*This process is generally permitted on interactive nodes and is only counted against you when considering memory usage (regardless of the process, too much memory usage is still considered bad; it cannot be throttled like CPU). The process is included in this report to show usage holistically.
**This accounts for the difference between the overall usage and the collected PID usage (which can be less accurate). This may be large if there are a lot of short-lived processes (such as compilers or quick commands) that account for a significant fraction of the total usage. These processes are whitelisted as defined above.
Required User Actions
When a user receives an alert email that the user is put in a penalty status, the user should
kill the processes that use too much resources on the shared node listed in the alert email, and/or reduce the resources used by the processes
submit an, a , or an to run computational intensive programs including but not limited to Python, R and Matlab
consider attending to learn more about correctly using Oscar.
CCV reserves the right to suspend a user's access to Oscar, if the user repeatedly violates the limits, and the user is not able to work with CCV to find a solution.
Exempt Processes
Essential Linux utilities, such as rsync, cp, scp, SLURM commands, creating Singularity images, and code compilation, are exempt. To obtain a comprehensive list, please get in touch with us
The CPU resources used by exempt programs are not count against the CPU limits. However, the memory resources used by exempt program still counted against the memory limits.
Transferring Files to and from Oscar
There are several ways to move files between your machine and Oscar. Which method you choose will depend on how much data you need to move and your personal preference for each method.
You can drag and drop files from your machine to the Oscar filesystem via SMB. This is an easy method for a small number of files. Please refer to this for mounting filesystem via SMB.
2. Command line
Mac and Linux
SCP
You can use scp to transfer files. For example to copy a file from your computer to Oscar:
To copy a file from Oscar to your computer:
RSYNC
You can use rsync to sync files across your local computer to Oscar:
Windows On Windows, if you have PuTTY installed, you can use it's pscp function from the terminal.
3. GUI programs for transferring files using the sftp protocol and transfer.ccv.brown.edu hostname
DUO is required if you are not connected to approved networks, e.g., home network
There is no interactive terminal message but your Phone will get a prompt automatically
DUO is NOT required if you are connected to approved Brown networks
In general, you can specify the following for your GUI programs:
Protocol: SFTP
Host: transfer.ccv.brown.edu
User: your Oscar username
3.1 for Windows
3.1.1 Limit Concurrent Transfer and Change Reconnect Options
Click the Optionsand then Preferences menu in WinsCP. In the poped up window, click Transfer and then Background to (Figure 1)
change Maximal number of transfers at the same time to 1
uncheck Use multiple connections for single transfer
click Endurace to (Figure 2)
set Automatically reconnect session to 5 seconds
uncheck Automatically reconnect session, if it stalls
set Keep reconnection for to 10 seconds
3.1.2 Add a New Site
3.2 FileZilla
3.2.1. Disable Timeout
Click the Edit menu and then select the Settings submenu, and then change the Timeout in seconds to 0 to disable, as shown in Figure 2
3.2.2 Add a New Site
Open the Site Manager as show in Figure 5.
Click the 'New Site' button to add a new site, as shown in Figure 4:
Limit the number of simultaneous connections to 1, as shown in Figure 5.
Click the 'Connect' button to connect to Oscar and transfer files.
3.3 Cyberduck
You may see a popup window on 'Unknown Fingerprint'. You just need to check the 'Always' option and click 'Allow'. This is windows should not pop up again unless the transfer server is changed again.
4. Globus online
is a secure, reliable research data management service. You can move data directly to Oscar from anothe Globus endpoint. Oscar has one Globus endpoint:
If you want to use Globus Online to move data to/from you own machine, you can install Globus Connect Personal. For more instructions on how to use Globus, see the in the Globus documentation.
5. LFTP
is a sophisticated file transfer program supporting a number of network protocols (ftp, http, sftp, fish, torrent). It has bookmarks, a built-in mirror command, can transfer several files in parallel and was designed with reliability in mind. You can use the LFTP module from Oscar to transfer data from any (S)FTP server you have access to directly to Oscar. Below are the main LFTP commands to get you started:
Anaconda
Anaconda provides Python, R and other packages for scientific computing including data sciences, machine learning, etc.
Anaconda Modules
There are several anaconda modules available on Oscar. To list all anaconda modules, run module avail anaconda. The anaconda/2022.5 module is recommended.
Do not activate a conda environment before submitting a batch job if the batch job activates a conda environment. Otherwise, the batch job will not be able to activate the conda environment and hence fail.
anaconda/2022.05
This is the newest anaconda module on Oscar. The first time you load the anaconda/2022.05 module, you need initialize the environment by running the following command:
Do not load the module in your .modules or .bashrc file. Otherwise, your VNC session cannot start.
anaconda/3-5.2.0
If you load the module in your .modules or .bashrc file, you may need to have the following lines in your .bashrc as well to be able to start a VNC session:
unset SESSION_MANAGER
unset DBUS_SESSION_BUS_ADDRESS
Conda
Anaconda uses conda to install packages and manage their depenencies. You can use conda to manage . To access conda, you need to load an Anaconda module. For example,
or
Conda Environment
A user may install all needed software packages for a project in a conda environment. A conda environment can be
shared among all users if the environment is installed in a shared directory
private to one user if the environment is installed in a user's private directory
The command 'conda info' shows important configurations for conda environment.
Below are some important configurations:
envs directories: a list of directories where a conda environment is installed by default. In the output of 'conda info' above, the first default directory to install a conda environment is a $HOME/anaconda.
package cache: a list of directories where downloaded packages are stored.
Create a New Conda Environment
To create a new conda environment in a default directory, run the following command:
To create a new conda environment in a different directory, run the following command:
Activate a Conda Environment
After creating a conda environment, users can activate a conda environment to install or access packages in the environment. The command is slightly different for different anaconda modules.
For the anaoconda/2022.05 module, users can activate an environment with the following command:
For the anaconda/3-5.2.0 module, users can activate an environment with the following command:
The commands above will only work if:
A conda environment with the specified name (conda_environment_name in the example) exists
If you need to activate a conda environment in a bash script, you need to source the conda.sh as shown in the following example bash script:
If you are using anaconda/3-5.2.0, replace the instances of 2022.05 with 3-5.2.0.
After installing packages in an active environment (instructions below), you do not need to load or install those packages in the bash script; any packages installed in the conda environment (before the script even starts) will be available through the environment after it is activated (line 4 in the code above).
To deactivate a conda environment, simply use the following command:
Install Packages in an Active Conda Environment
To install a package, we need to first activate a conda environment, and then run
The "=version" is optional. By default, conda install a package from the anaconda channel. To install a package from a different channel, run conda install with the -c option. For example, to install a package from the conda_forge channel, run
Delete a Conda Environment
To delete a conda environment, run
Remove Caches
Conda may download lots of additional packages when installing a package. A user may use up all quota due to these downloaded packages. To remove the downloaded packges, run
To run a batch job on Oscar, you first have to write a script that describes what resources you need and how your program will run. Some example batch scripts are available in your home directory on Oscar, in the directory:
A batch script starts by specifying the bash shell as its interpreter with the line:
By default, a batch job will reserve 1 core and 2.8GB of memory per core for your job. You can customize the amount of resources allocated for your job by explicitly requesting them in your batch script with a series of lines starting with #SBATCH, e.g.,
The above lines request 4 cores (-n), 16GB of memory per node (--mem), and one hour of runtime (-t). After you have described the resources you want allocated for the job, you then give the commands that you want to be executed.
All of the #SBATCH instructions in your batch script must appear before the commands you want to run.
Once you have your batch script, you can submit a batch job to the queue using the sbatch command:
Submitting jobs from the command line
As an alternative to requesting resources within your batch script, it is possible to define the resources requested as command-line options to sbatch. For example, the command below requests 4 cores (-n), 16GB of memory per node (--mem), and one hour of runtime (-t) to run the job defined in the batch script.
Note that command-line options passed to sbatch will override the resources specified in the script, so this is a handy way to reuse an existing batch script when you just want to change a few of the resource values.
Output from batch jobs
The sbatch command will return a number, which is your Job ID. You can view the output of your job in the file slurm-<jobid>.out in the directory where you invoked the sbatch command. For instance, you can view the last 10 lines of output with:
Alternatively, you can mention the file names where you want to dump the standard output and errors using the -o and -e flags. You can use %j within the output/error filenames to add the id of the job. If you would like to change your output file to be MyOutput-<job-id>, you can add the following line to your batch job:
sbatch command options
A full description of all of the options forsbatch can be found or by using the following command on Oscar:
The table below summarizes some of the more useful options forsbatch .
Passing environment variables to a batch job
When a user logs into Oscar, there are pre-set environment variables such as HOME, which are the user's login environment variables. A user may modify an existing enviornmet variable, or add a new environment variable. So when a user submits a slurm batch job, the user's current environment variables may differ from the user's login environment. By default, a user's current environment variables, instead of the user's login environment variables, are accessible to the user's batch jobs on Oscar.
To modify or add an environment variable, run the following command:
run the following command in your shell
or have the following line in your batch script
After the step above to modify or add an environment variable, your batch job can access the environment variable my_variable whose value is my_value.
To export more than one environment variables, just list all the name=value pairs separated by commas:
Here is an example that a batch script loops over an input file and submits a job for each directory in the input file, where a directory is passed to a batch job for processing.
The input file test.txt has multiple lines where each line is a directory:
The loop.sh script reads each line (directory) from the input file and passes the directory as an environment variable to a batch job:
The test.job is a job script, which runs the test.sh to process the directory passed as an environment variable:
The test.sh is a bash script which simply echoes the directory:
If you run ./loop.sh, then three jobs are submitted. Each job generates an output like the following:
Using variables to set slurm job name, output filename, and error filename
Variables can be passed at the sbatch command line to set the job name, output and error file names, as shown in the following example:
Jupyter Notebooks on Oscar
Installing Jupyter Notebook
The anaconda/3-5.2.0 module provides jupyter-notebook. Users can also use pip or anaconda to install jupyter notebook.
Running Jupyter Notebook on Oscar
There are a couple of ways to use Notebook on Oscar. You can run Jupyter Notebook
in a VNC session
using a batch job
in an interactive session
With the batch job or interactive session method, you use a browser on your machine to connect to your Jupyter Notebook server on Oscar.
Start by going to the directory you want to access when using Jupyter Notebook, and then start Jupyter Notebook. The directory where a Jupyter Notebook is started is the working directory for the Notebook.
Do not run Jupyter Notebook on login nodes.
In a VNC Session
Start a , and open up a terminal in the VNC session . To start a Jupyter Notebook, enter
This will start the Jupyter Notebook server and open up a browser with the notebook.
If you installed Jupyter Notebook with pip, you may need to give the full path:
~/.local/bin/jupyter-notebook
Using a Batch Job
Submit an ssh tunnel to the server.
Set up an ssh tunnel to the server.
Open a browser to view the notebook.
1. Submit batch script
Here is an example batch script to start a Jupyter notebook server on an Oscar compute node
If you installed Jupyter notebook with pip you may need to give the full path:
This script can be found in ~/batch_scripts. Copy this example and submit this script with
sbatch jupyter.sh
Once your batch job is running there will be a file named jupyter-log-{jobid}.txtcontaining the information you need to connect to your jupyter notebook server on Oscar. To check if your job is running, use myq.
The output from myq will look something like this:
2. Set up an ssh tunnel to the notebook server
In this example the jobID is 7239096. To view the notebook server information, use cat. For this example:
cat jupyter-log-7239096.txt
Open a terminal on your machine and copy and paste the ssh -N -L ........ line into the terminal.
If you are using Windows, follow the documentation to complete this step.
Enter your Oscar password. Note it will appear that nothing has happened.
3. Open a browser to view the notebook
Open a browser on your local machine to the address given in cat jupyter-log-{jobid}.txt.
The notebook will ask for a token. Copy the token from jupyter-log-{jobid}.txt. Then your notebook will start.
Remember to scancel {jobid} when you are done with your notebook session.
In an Interactive Session
Start Jupyter Notebook in an interactive job.
Set up an ssh tunnel to the server.
Open a browser to view the notebook.
1. Start a Jupyter Notebook in an interactive job
Start an and then in your interactive session enter the following:
An output similar to the one below indicates that Jupyter Notebook has started:
Open a terminal on your machine and enter the following line (replace $ipnip and $ipnport with the values from the two echo commands in the previous step).
If you are using Windows, follow the documentation to complete this step.
Enter your Oscar password. Note it will appear that nothing has happened.
3. Open a browser to view the notebook
Open a browser on your local machine to the address:
Again, you need to replace $ipnport with the value from the first echo command in Step 1. The notebook will ask for a token. You can copy the token from the output from Step 2.
4. Press Ctrl+C twice to kill your Jupyter Notebook server
Once you finish and no longer need the Jupyter Notebook server, you can kill the server by pressing Ctrl+C twice in your interactive session.
Tunneling into Jupyter with Windows
This page is for users trying to open Jupyter Notebooks/Labs through Oscar with Windows.
Software that makes it easy
If you are using Windows, you can use any of the following options to open a terminal on your machine (ranked in order of least difficult to set up and use):
(we recommend Ubuntu as your Linux distribution)
After opening a terminal using any of these programs, simply enter the ssh command provided by the jupyter-log-{jobid}.txt file. Then continue with the steps given by the documentation that led you to this page.
If you have PuTTY and would prefer to not download any additional software, there are steps (explained below) that you can take to use PuTTY to tunnel into a Jupyter Notebook/Lab.
Using PuTTY
These instructions will use ssh -N -L 9283:172.20.209.14:9283 [email protected] as an example command that could be found in the jupyter-log-{jobid}.txt file.
Open PuTTY and enter your host name ([email protected]) in the textbox.
Next, navigate to the 'Tunnels' Menu (click the '+' next to SSH in order to have it displayed).
Enter the source port (9283 in the example) and destination (172.20.209.14:9283 in the example). Click 'Add'. The source port and destination should show up as a pair in the box above. Then click 'Open'. A new window should open requesting your password.
After entering your password, you should be able to access the notebook/lab in a browser using localhost:ipnport (see the documentation that led you here for details).
Inspecting Disk Usage (Ncdu)
To determine the sizes of files and discover the largest files in a directory, one can use the Ncdu module.
To get started with NCDU, load the module using the following command:
module load ncdu/1.14
Once the module has been loaded, it can be used to easily show the size of all files within a directory:
ncdu my_directory
To view options you can use with the ncdu command, simply use the command ncdu --help
The line above uses Ncdu to rank all of the files within the my_directory directory. Your window should change to show a loading screen (if the directory doesn't have a lot in it, you may not even see this screen):
Once Ncdu has finished loading, you will see a result like this:
The files will be ordered with the largest file at the top and the smallest file at the bottom. The bottom left corner shows the Total disk usage (which in this case is 25.5 KiB). To quit out of this display, simply press q on your keyboard.
If there is a subdirectory within the directory you're inspecting, the files and directories within that subdirectory can be viewed by selecting the directory with the gray bar (using up and down arrow keys as needed) and then using the right arrow key.
Common Acronyms and Terms
Common Acronyms and Terms
Quickstart Guide
This guide assumes you have an Oscar account. To request an account see .
If you're confused about any acronyms or terms throughout the guide, check out our page to see definitions of commonly used terms
DMTCP
This page is under construction!
(DMTCP) checkpoints a running program on Linux with no modifications to the program or OS. It allows to restart running the program from a checkpoint.
Jupyter Labs on Oscar
Installing Jupyter Lab
The anaconda/3-5.2.0 module provides jupyter-lab. Users can also use pip or anaconda to .
Associations & Quality of Service (QOS)
Associations
Oscar uses associations to control job submissions from users. An association refers to a combination of four factors: Cluster, Account, User, and Partition. For a user to submit jobs to a partition, an association for the user and partition is required in Oscar.
To view a table of association data for a specific user (thegrouch in the example), enter the following command in Oscar:
Intro to CUDA
Introduction to CUDA
is an extension of the C language, as well as a runtime library, to facilitate general-purpose programming of NVIDIA GPUs. If you already program in C, you will probably find the syntax of CUDA programs familiar. If you are more comfortable with C++, you may consider instead using the higher-level library, which resembles the Standard Template Library and is included with CUDA.
In either case, you will probably find that because of the differences between GPU and CPU architectures, there are several new concepts you will encounter that do not arise when programming serial or threaded programs for CPUs. These are mainly to do with how CUDA uses threads and how memory is arranged on the GPU, both described in more detail below.
# Jump box with public IP address
Host jump-box
HostName ssh8.ccv.brown.edu
User <username>
# Target machine with private IP address
Host ccv-vscode-node
HostName node1103
User <username>
ProxyCommand ssh -q -W %h:%p jump-box
CUDA uses a data-parallel programming model, which allows you to program at the level of what operations an individual thread performs on the data that it owns. This model works best for problems that can be expressed as a few operations that all threads apply in parallel to an array of data. CUDA allows you to define a thread-level function, then execute this function by mapping threads to the elements of your data array.
A thread-level function in CUDA is called a kernel. To launch a kernel on the GPU, you must specify a grid, and a decomposition of the grid into smaller thread blocks. A thread block usually has around 32 to 512 threads, and the grid may have many thread blocks totalling thousands of threads. The GPU uses this high thread count to help it hide the latency of memory references, which can take 100s of clock cycles.
Conceptually, it can be useful to map the grid onto the data you are processing in some meaningful way. For instance, if you have a 2D image, you can create a 2D grid where each thread in the grid corresponds to a pixel in the image. For example, you may have a 512x512 pixel image, on which you impose a grid of 512x512 threads that are subdivided into thread blocks with 8x8 threads each, for a total of 64x64 thread blocks. If your data does not allow for a clean mapping like this, you can always use a flat 1D array for the grid.
The CUDA runtime dynamically schedules the thread blocks to run on the multiprocessors of the GPU. The M2050 GPUs available on Oscar each have 14 multiprocessors. By adjusting the size of the thread block, you can control how much work is done concurrently on each multiprocessor.
Memory on the GPU
The GPU has a separate memory subsystem from the CPU. The M2050 GPUs have GDDR5 memory, which is a higher bandwidth memory than the DDR2 or DDR3 memory used by the CPU. The M2050 can deliver a peak memory bandwidth of almost 150 GB/sec, while a multi-core Nehalem CPU is limited to more like 25 GB/sec.
The trade-off is that there is usually less memory available on a GPU. For instance, on the Oscar GPU nodes, each M2050 has only 3 GB of memory shared by 14 multiprocessors (219 MB per multiprocessor), while the dual quad-core Nehalem CPUs have 24 GB shared by 8 cores (3 GB per core).
Another bottleneck is transferring data between the GPU and CPU, which happens over the PCI Express bus. For a CUDA program that must process a large dataset residing in CPU memory, it may take longer to transfer that data to the GPU than to perform the actual computation. The GPU offers the largest benefit over the CPU for programs where the input data is small, or there is a large amount of computation relative to the size of the input data.
CUDA kernels can access memory from three different locations with very different latencies: global GDDR5 memory (100s of cycles), shared memory (1-2 cycles), and constant memory (1 cycle). Global memory is available to all threads across all thread blocks, and can be transferred to and from CPU memory. Shared memory can only be shared by threads within a thread block and is only accessible on the GPU. Constant memory is accessible to all threads and the CPU, but is limited in size (64KB).
There are not enough free resources to fulfill your request
(JobHeldUser)
You have put a hold on the job. The job will not run until you lift the hold.
(ReqNodeNotAvail)
The resources you have requested are not available. Note this normally means you have requested something impossible, e.g. 100 cores on 1 node, or a 24 core sandy bridge node. Double check your batch script for any errors. Your job will never run if you are requesting something that does not exist on Oscar.
(PartitionNodeLimit)
You have asked for more nodes than exist in the partition. For example if you make a typo and have specified -N (nodes) but meant -n (tasks) and have asked for more than 64 nodes. Your job will never run. Double check your batch script.
(None)
You may see this for a short time when you first submit a job
(QOSGrpCpuLimit)
All your condo cores are currently in use
(QOSGrpMemLimit)
The total memory of your running jobs and this pending job is more than the limit for your account.
Add a feature constraint (a tag that describes a type of node).
Note: you can view the available features on Oscar with the nodescommand or sinfo -o "%20N %10c %10m %25f %10G "
You can also select multiple feature constraints using '|', i.e. #SBATCH -C quadrortx|intel
--mail-type=
Specify the events that you should be notified of by email: BEGIN, END, FAIL, REQUEUE, and ALL
--mail-user=
Email ID where you should be notified
option
purpose
-J
Specify the job name that will be displayed when when listing the job
-n
Number of tasks (= number of cores, if "--cpus-per-task" or "-c" option is not mentioned)
-c
Number of CPUs or cores per task (on the same node)
To access dmtcp, load a dmtcp module. For example:
module load dmtcp/2.6.0
Example Programs
The following example programs can be copied from /gpfs/runtime/oftware_examples/dmtcp/
dmtcp_serial
dmtcp_serial.c
dmtcp_serial_job.sh
Basic Usage
Launch a Program
The dmtcp_lanuch command launches a program, and automatically checkpoints the program. To specify the interval (seconds) for checkpoints, add the "-i num_seconds" option to the dmtcp_lauch command.
Example: the following command launches the program dmtcp_serial and checkpoints every 8 seconds.
As shown in the example above, a checkpoint file (ckpt_dmtcp_serial_24f183c2194a7dc4-40000-42af86bb59385.dmtcpp) is created, and can be used to restart the program
Restart from a checkpoint
The dmtcp_resart command restarts a program from a checkpoint, and also automatically checkpoints the program. To specify the interval (seconds) for checkpoints, add the "-i num_seconds" option to the dmtcp_restart command.
Example: the following command restarts the dmtcp_serial program from a checkpoint, and checkpoints every 12 seconds
Batch Jobs
It is desirable goal that single job script can
launch a program if there is checkpoints, or
automatically restarts from a checkpoint if there is one or more checkpoints
The job script dmtcp_serial_job.sh below is an example which shows how to achieve the goal:
If there is no checkpoint in the current directory, launch the program dmtcp_serial
If one or more checkpoints exist in the current directory, restart the program dmtcp_serial from the latest checkpoint
First Submission - Launch a Program
Submit dmtcp_serial_job.sh and then wait for the job to run until time out. Below shows the beginning and end of the job output file
Later Submissions - Restart from a Checkpoint
Submit dmtcp_serial_job.sh and then wait for the job to run until time out. Below shows the beginning of the job output file, which demonstrate that the job restarts from the checkpoint of the previous job.
Job Array
The following example script
creates a sub directory for each task of a job array, and then saves a task's checkpoint in the task's own sub directory when the job script is submitted for the first time
restarts checkpoints in task subdirectories when the job script is submitted for the second time or later
This script can be found in ~/batch_scripts. Copy this example and submit this script with
sbatch jupyter.sh
Once your batch job is running there will be a file named jupyter-log-{jobid}.txtcontaining the information you need to connect to your Jupyter lab server on Oscar. To check if your job is running, use myq.
The output from myq will look something like this:
2. Setup an ssh tunnel to the notebook server
In this example the jobID is 7239096. To view the lab server information, use cat. For this example:
cat jupyter-log-7239096.txt
Open a terminal on your machine and copy and paste the ssh -N -L ........ line into the terminal.
[I 13:12:03.410 LabApp] Use Control-C to stop this server and shut down all kernels (twice to skip confirmation).
[C 13:12:03.411 LabApp]
2. Setup an ssh tunnel to the server
Open a terminal on your machine and enter the following line (replace $ipnip and $ipnport with the values from the two echo commands in the previous step).
Enter your Oscar password. Note it will appear that nothing has happened.
3. Open a browser to view the notebook
Open a browser on your local machine to the address:
Again, you need to replace $ipnport with the value from the first echo command in Step 1. The notebook will ask for a token. You can copy the token from the output from Step 2.
4. Press Ctrl+C twice to kill your Jupyter Lab server
Once you finish and no longer need the Jupyter Lab server, you can kill the server by pressing Ctrl+C twice in your interactive session.
If thegrouch has an exploratory account, you should see an output similar to this:
Note that the first four columns correspond to the four factors that form an association. Each row of the table corresponds to a unique association (i.e., a unique combination of Cluster, Account, User, and Partition values). Each association is assigned a Quality of Service (see QOS section below for more details).
Some associations have a value for GrpTRESRunMins. This value indicates a limit on the total number of Trackable RESource (TRES) minutes that can be used by jobs running with this association at any given time. The cpu=110000 for the association with the batch partition indicates that all of the jobs running with this association can have at most an accumulated 110,000 core-minute cost. If this limit is reached, new jobs will be delayed until other jobs have completed and freed up resources.
Example of GrpTRESRunMins Limit
Here is an example file that incurs a significant core-minute cost:
If this file is named too_many_cpu_minutes.sh, a user withthegrouch's QOS might experience something like this:
The REASON field will be (None) at first, but after a minute or so, it should resemble the output above (after another myq command).
Note that the REASON the job is pending and not yet running is AssocGrpCPURunMinutesLimit. This is because the program requests 30 cores for 90 hours, which is more than the oscar/default/thegrouch/batch association allows (30 cores * 90 hours * 60 minutes/hour = 162,000 core-minutes > 110,000 core-minutes). In fact, this job could be pending indefinitely, so it would be a good idea for thegrouch to run scancel 12345678 and make a less demanding job request (or use an association that allows for that amount of resources).
Quality of Service
An association's QOS is used for job scheduling when a user requests that a job be run. Every QOS is linked to a set of job limits that reflect the limits of the cluster/account/user/partition of the association(s) that has/have that QOS. QOS's can also have information on GrpTRESRunMins limits for their corresponding associations. For example, HPC Priority accounts have job limits of 1,198,080 core-minutes per job, which are associated with those accounts' QOS's. Whenever a job request is made (necessarily through a specific association), the job will only be queued if it meets the requirements of the association's QOS. In some cases, a QOS can be defined to have limits that differ from its corresponding association. In such cases, the limits of the QOS override the limits of the corresponding association.
sftp <username>@transfer.ccv.brown.edu
put /path/local_file
sftp <username>@transfer.ccv.brown.edu
get -r filename.txt
$conda init bash
no change /gpfs/runtime/opt/anaconda/2022.05/condabin/conda
no change /gpfs/runtime/opt/anaconda/2022.05/bin/conda
no change /gpfs/runtime/opt/anaconda/2022.05/bin/conda-env
no change /gpfs/runtime/opt/anaconda/2022.05/bin/activate
no change /gpfs/runtime/opt/anaconda/2022.05/bin/deactivate
no change /gpfs/runtime/opt/anaconda/2022.05/etc/profile.d/conda.sh
no change /gpfs/runtime/opt/anaconda/2022.05/etc/fish/conf.d/conda.fish
no change /gpfs/runtime/opt/anaconda/2022.05/shell/condabin/Conda.psm1
no change /gpfs/runtime/opt/anaconda/2022.05/shell/condabin/conda-hook.ps1
no change /gpfs/runtime/opt/anaconda/2022.05/lib/python3.7/site-packages/xontrib/conda.xsh
no change /gpfs/runtime/opt/anaconda/2022.05/etc/profile.d/conda.csh
modified /users/yliu385/.bashrc
==> For changes to take effect, close and re-open your current shell. <==
#!/bin/bash
#SBATCH --nodes 1
#SBATCH -c 6
#SBATCH --time 04:00:00
#SBATCH --mem-per-cpu 3G
#SBATCH --job-name tunnel
#SBATCH --output jupyter-log-%J.txt
## get tunneling info
XDG_RUNTIME_DIR=""
ipnport=$(shuf -i8000-9999 -n1)
ipnip=$(hostname -i)
## print tunneling instructions to jupyter-log-{jobid}.txt
echo -e "
Copy/Paste this in your local terminal to ssh tunnel with remote
-----------------------------------------------------------------
ssh -N -L $ipnport:$ipnip:$ipnport [email protected] -----------------------------------------------------------------
Then open a browser on your local machine to the following address
------------------------------------------------------------------
localhost:$ipnport (prefix w/ https:// if using password)
------------------------------------------------------------------
"
## start an ipcluster instance and launch jupyter server
module load anaconda/3-5.2.0
jupyter-notebook --no-browser --port=$ipnport --ip=$ipnip
Jobs for user mhamilton
Running:
ID NAME PART. QOS CPU WALLTIME REMAIN NODES
7239096 tunnel batch pri-mhamilt 6 4:00:00 3:57:33 node1036
Pending:
(none)
#!/bin/bash
#SBATCH -n 1
#SBATCH --array=1-4
#SBATCH -t 5:00
#SBATCH -J dmtcp_job_array
checkpoint_interval=8
port=$((SLURM_JOB_ID %20000 + 40000))
task_dir=jobtask_$SLURM_ARRAY_TASK_ID
if [ ! -d $task_dir ]; then
mkdir $task_dir
cd $task_dir
dmtcp_launch -p $port -i $checkpoint_interval ../dmtcp_serial
else
cd $task_dir
checkpoint_file=`ls ckpt_*.dmtcp -t|head -n 1`
if [ -z $checkpoint_file ]; then
dmtcp_launch -p $port -i $checkpoint_interval ../dmtcp_serial
else
dmtcp_restart -p $port -i $checkpoint_interval $checkpoint_file
fi
fi
jupyter-lab
#!/bin/bash
#SBATCH --nodes 1
#SBATCH -c 6
#SBATCH --time 04:00:00
#SBATCH --mem-per-cpu 3G
#SBATCH --job-name tunnel
#SBATCH --output jupyter-log-%J.txt
## get tunneling info
XDG_RUNTIME_DIR=""
ipnport=$(shuf -i8000-9999 -n1)
ipnip=$(hostname -i)
## print tunneling instructions to jupyter-log-{jobid}.txt
echo -e "
Copy/Paste this in your local terminal to ssh tunnel with remote
-----------------------------------------------------------------
ssh -N -L $ipnport:$ipnip:$ipnport [email protected] -----------------------------------------------------------------
Then open a browser on your local machine to the following address
------------------------------------------------------------------
localhost:$ipnport (prefix w/ https:// if using password)
------------------------------------------------------------------
"
## start an ipcluster instance and launch jupyter server
module load anaconda/3-5.2.0
jupyter-lab --no-browser --port=$ipnport --ip=$ipnip
Jobs for user mhamilton
Running:
ID NAME PART. QOS CPU WALLTIME REMAIN NODES
7239096 tunnel batch pri-mhamilt 6 4:00:00 3:57:33 node1036
Pending:
(none)
localhost:$ipnport (prefix w/ http:// if using password)
(sacctmgr list assoc | head -2; sacctmgr list assoc | grep thegrouch) | cat
Cluster Account User Partition Share GrpJobs GrpTRES GrpSubmit GrpWall GrpTRESMins MaxJobs MaxTRES MaxTRESPerNode MaxSubmit MaxWall MaxTRESMins QOS Def QOS GrpTRESRunMin
---------- ---------- ---------- ---------- --------- ------- ------------- --------- ----------- ------------- ------- ------------- -------------- --------- ----------- ------------- -------------------- --------- -------------
oscar default thegrouch gpu-debug 1 gpu-debug gpu-debug
oscar default thegrouch bigmem 1 norm-bigmem norm-big+
oscar default thegrouch smp 1 norm-smp norm-smp
oscar default thegrouch gpu 1 norm-gpu norm-gpu cpu=34560,gr+
oscar default thegrouch batch 1 normal normal cpu=110000
oscar default thegrouch vnc 1 vnc vnc
oscar default thegrouch debug 1 debug debug
#!/bin/bash
#SBATCH -n 30
#SBATCH --mem=32G
#SBATCH -t 90:00:00
echo "Is this too much to ask? (Hint: What is the GrpTRESRunMins limit for batch?)"
$ sbatch too_many_cpu_minutes.sh
Submitted batch job 12345678
$ myq
Jobs for user thegrouch
Running:
(none)
Pending:
ID NAME PART. QOS CPU WALLTIME EST.START REASON
15726799 too_many_cpu_minutes.sh batch normal 30 3-18:00:00 N/A (AssocGrpCPURunMinutesLimit)
A personal Windows computer must have CrowdStrike installed in order to be on approved Brown networks.
Within Oscar, an association refers to a combination of four factors: Cluster, Account, User, and Partition. Associations are used to control job submissions for users.
[]
Batch Jobs
Put simply, batch jobs are scheduled programs that are assigned to run on a computer without further user interaction.
[]
CCV
Brown University's Center for Computation and Visualization. Provides software, expertise, and other services for Brown's research community. See for more information.
CESM
Stands for Community Earth System Model. "CESM is a fully-coupled, community, global climate model that provides state-of-the-art computer simulations of the Earth's past, present, and future climate states." ()
[]
CIFS
Stands for Common Internet File System. CIFS is a network protocol used for providing shared access to files and devices on the same network. Users who are unable to use SMB for file-sharing should try CIFS as an alternative.
[]
Condo
PIs can purchase condos that have a significant amount of computing resources which can be shared with others.
[]
CUDA
" is an extension of the C language, as well as a runtime library, to facilitate general-purpose programming of NVIDIA GPUs." ()
[]
HPC
Stands for High Performance Computing. HPC is the ability to process data and perform highly complex calculations at an accelerated rate. Oscar is the service that CCV offers to the Brown community for their High Performance Computing needs.
Job Array
A job array is a collection of jobs that all run the same program but on different values of a parameter.
[]
Jupyter Notebook
"The Jupyter Notebook is an open-source web application that allows you to create and share documents that contain live code, equations, visualizations and narrative text."
[]
Interactive Jobs
Jobs that allow the user to interact in real time with applications within Oscar, often from the command line. This differs from batch jobs in that each command to be run must be put in one at a time.
[]
Modules
Modules are software components that can easily be loaded or unloaded into Oscar. For instance, a user can load the Python 3 module using a module load command.
[]
MPI
Stands for . MPI is a system that aims to be the standard for portable and efficient message passing. is a technique often used in object-oriented programming and parallel programming
[]
Partition
Partitions are essentially groupings of nodes that allocate resources for specific types of tasks. On Oscar, partitions are based on job submissions through the Slurm workload manager.
[]
PI
Stands for Principal Investigator. Mainly used to refer to the individual responsible for conducting and administrating a research grant. Within Oscar, PIs have their own data directories that can be shared to students. PIs may also purchase condos.
[]
PuTTY
A client for SSH for Windows and Unix that emulates a terminal
[]
Python
An object-oriented, high-level, and popular programming language
[]
Slurm
A workload manager used within Oscar to schedule jobs
[]
SSH
Stands for Secure Shell Protocol. Used to communicate securely between computers and often used within a command-line interface (CLI) for connections to remote servers
[]
SMB
The Server Message Block (SMB) protocol is a network protocol that allows users to communicate with remote computers for file-sharing and other uses. It is one of the versions of the Common Internet File System (CIFS). Within Oscar, SMB is mainly used for file transfer.
[]
VNC
Stands for Virtual Network Computing. VNC is a desktop sharing system that allows you to remotely control another desktop. On Oscar, it is used to allow a desktop interface for applications.
[]
Quality of Service (QOS)
The job limits that are linked to a given association. For instance, Priority Accounts will generally have a higher quality of service than Exploratory Accounts.
[]
Term
Description
Anaconda / Conda
A distribution of Python and R used for scientific computing that is meant to simplify package management and deployment. Conda is used for installing packages and managing their dependencies.
[]
OSCAR
Oscar is the shared compute cluster operated by CCV.
Oscar runs the Linux RedHat7 operating system. General Linux documentation is available from The Linux Documentation Project. We recommend you read up on basic Linux commands before using Oscar. Some of the most common commands you'll be using in Oscar can also be found on our Quick Reference page.
Oscar has two login nodes and several hundred compute nodes. When users log in through Secure Shell (SSH), they are first put on one of the login nodes which are shared among several users at a time. You can use the login nodes to compile your code, manage files, and launch jobs on the compute nodes from your own computer. Running computationally intensive or memory intensive programs on the login node slows down the system for all users. Any processes taking up too much CPU or memory on a login node will be killed. Please do not run Matlab on the login nodes.
What username and password should I be using?
If you are at Brown and have requested a regular CCV account, your Oscar login will be authenticated using your Brown credentials, i.e. the same username and password that you use to log into any Brown service such as "canvas". We have seen login problems with the Brown credentials for some users so accounts moved to the RedHat7 system after September 1st 2018 can also log into RedHat7 with their CCV password.
If you are an external user, you will have to get a sponsored ID at Brown through the department with which you are associated before requesting an account on Oscar. Once you have the sponsored ID at Brown, you can request an account on Oscar and use your Brown username and password to log in.
Connecting to Oscar for the first time
To log in to Oscar you need Secure Shell (SSH) on your computer. Mac and Linux machines normally have SSH available. To login in to Oscar, open a terminal and type
Windows users need to install an SSH client. We recommend PuTTY, a free SSH client for Windows. Once you've installed PuTTY, open the client and use <username>@ssh.ccv.brown.edufor the Host Name and click Open. The configuration should look similar to the screenshot below.
The first time you connect to Oscar you will see a message like:
You can type yes . You will be prompted for your password. Note that nothing will show up on the screen when you type in your password; just type it in and press enter. You will now be in your home directory on Oscar. In your terminal you will see a prompt like this:
Congratulations, you are now on one of the Oscar login nodes.
Note: Please do not run computations or simulations on the login nodes, because they are shared with other users. You can use the login nodes to compile your code, manage files, and launch jobs on the compute nodes.
File system
Users on Oscar have three places to store files:
home
scratch
data
Note that class accounts may not have a data directory. Users who are members of more than one research group may have access to multiple data directories.
From the home directory, you can use the command ls to see your scratch directory and your data directory (if you have one) and use cd to navigate into them if needed.
To see how much space you have, use the command myquota. Below is an example output:
Files not accessed for 30 days may be deleted from your scratch directory. This is because scratch is high performance space. The fuller scratch is, the worse the read/write performance. Use ~/data for files you need to keep long term.
A good practice is to configure your application to read any initial input data from ~/data and write all output into ~/scratch. Then, when the application has finished, move or copy data you would like to save from ~/scratch to ~/data. For more information on which directories are backed up and best practices for reading/writing files, see Oscar's Filesystem and Best Practices. You can go over your quota up to the hard limit for a grace period. This grace period is to give you time to manage your files. When the grace period expires you will be unable to write any files until you are back under quota.
CCV uses the PyModules package for managing the software environment on OSCAR. To see the software available on Oscar, use the command module avail. You can load any one of these software modules using module load <module>. The command module list shows what modules you have loaded. Below is an example of checking which versions of the module 'workshop' are available and loading a given version.
For a list of all PyModule commands, see Software Modules. If you have a request for software to be installed on Oscar, email [email protected].
Using a Desktop on Oscar
You can connect remotely to a graphical desktop environment on Oscar using CCV's OpenOnDemand. The OOD Desktop integrates with the scheduling system on Oscar to create dedicated, persistent VNC sessions that are tied to a single user.
Using VNC, you can run graphical user interface (GUI) applications like Matlab, Mathematica, etc. while having access to Oscar's compute power and file system.
Choose a session that suits your needs
Running Jobs
You are on Oscar's login nodes when you log in through SSH. You should not (and would not want to) run your programs on these nodes as these are shared by all active users to perform tasks like managing files and compiling programs.
With so many active users, a shared cluster has to use a "job scheduler" to assign compute resources to users for running programs. When you submit a job (a set of commands) to the scheduler along with the resources you need, it puts your job in a queue. The job is run when the required resources (cores, memory, etc.) become available. Note that since Oscar is a shared resource, you must be prepared to wait for your job to start running, and it can't be expected to start running straight away.
Oscar uses the SLURM job scheduler. Batch jobs are the preferred mode of running programs, where all commands are mentioned in a "batch script" along with the required resources (number of cores, wall-time, etc.). However, there is also a way to run programs interactively.
For information on how to submit jobs on Oscar, see Running Jobs.
There is also extensive documentation on the web on using SLURM (quick start guide).
An example loading screen for Ncdu (the full directory for the "Current item" has been obscured)
A list of files and their sizes as displayed in the output of a call to Ncdu
ParaView Remote Rendering
Running Paraview Remote Rendering in Oscar
This service is new and is in Beta
The Center for Computation and Visualization (CCV) offers to the academic community a way to visualize large datasets using Oscar and its powerful GPUs as a rendering server. The current GPU hardware and available memory on Oscar surpasses the common desktop models, offering a modern and robust solution to display large datasets in parallel jobs using the widely used opensource software Paraview. It is a simple two-steps process. Start the server and connect the client.
Who benefits from this service?
The target audience for this service are members of the academic community that interact, and analyze large 3D datasets, i.e., point clouds, volumetric data, tiff-stacks and mesh-data. This includes groups working with microscopy data, MRI images, structural analysis, fluid dynamics, climate sciences, astrophysics and more. In fact, ParaView can handle over 100 different file formats. The remote rendering service is targeted to scenarios where the personal/lab computer setup may not have the resources to handle the size of the underlying datasets. Common obstacles are older GPU technology or low RAM availability which may cause performance issues.
Workflow Overview
Above is a graphical representation on how the parallel render server works using Oscar. The user logins to Oscar either via SSH or VNC session. From the terminal, the user loads the Paraview module and executes the convinience script called run-remote-serverto start the Paraview server session and allocate the memory and walltime limit. Once the server starts, the user receives an email with the information needed to access the server. Lastly, the user connects the Paraview client (i.e., desktop application) to the server that is running in Oscar. The client displays images that are processed by the server (on Oscar) which reconstructs the information computed by the nodes.
0. Prerequisites
Paraview Desktop:
You can either download the Paraview Desktop App to your presonal computer or access the desktop application already installed in Oscar's VNC. Installing in your local computer may give you better interactivity.
Download Paraview Desktop to your desktop computer (Recommended)
Go to the official . Select your Operational system (Linux, Windows or Mac), get the ParaView-5.11.0--Python3.9 version that suit your local machine and run the Installer wizard.
Not all versions will work. You must select 5.11.0
Using Paraview UI installed in Oscar
Connect to
Open terminal: Applications - > Utilities -> Terminal (this might differ depending on the Operating System UI)
Run the commands
If this is your first time opening paraview, it will take a few minutes.
1. Start the Server
You need to allocate the resources via SLURM indicating the amount of memory you want to reserve, as well as a few optional parameters to configure your session. We have created a convinience script for you to do so called run-remote-server
In order to have the run-remote-server be found we need to load the Paraview module that supports this service (this appends the correct path to out PATH environment variable)
The flag -uindicates where the confirmation email will be sent. Technically it could be any email address, but the remote render session can only be used by existing Oscar users.
The only mandatory parameter is -u <user-email>.
Memory Request
The number of CPU cores and GPUs are determined by the memory request.
By default, the run-remote-server script's minimum memory request is 45 GB (1 CPU/GPU ) and the maximum is 180 GB (4 CPU/GPU ). You can add more resources to your session using the -m flag. Every multiplier of 45GB adds a CPU core and a GPU. i.e :
The following is the description of the command and the available configuration settings.
usage: run-remote-render [-n cores] [-t walltime] [-m memory] [-q queue] [-o outfile] [-g ngpus] [-u user brown email]
Allocates resources, start up the render server and send and email to the user requesting the service
options:
-t walltime as hh:mm:ss (default: 1:30:00)
-m memory as #[k|m|g] (default: 45G)
-o outfile save a copy of the session's output to outfile (default: off)
-q slurm partition (gpu (default)| gpu-he)
-u brown email of the user requesting the service
After executing the command, the system will allocate resources, and it will send a confirmation email indicating that the service is ready; the email includes additional instruction on how to connect to the server using Paraview UI.
NOTE: You might not receive the email instantly. Sometimes it may take a while before there are sufficient resources (i.e., GPUs) available. You will get the notification as soon as they are available for your use. You can also use the myq command to see the status of your job
2. Wait for the confirmation Email
In the email sent by the system has important information such as :
How to create an SSH tunnel
The IP address and port where the service is deployed
How to connect to the server from multiple systems
Please read it and get familiar on how the process works.
Please, read the email carefully. It contains information about your connection such as server IP address and port. You need them in order to proceed with the next steps.
3. Connect to the Server
There are two options to connect to the remote server:
Your personal computer
VNC
3.1 Setting up SSH Tunneling
This step is only needed if you are using your personal computer and not VNC
Open a terminal and execute the command:
<SERVER_IP> This is the ip of the compute node in Oscar. Replace with the value sent in the confirmation email
<port-number>This is the port exposed to access the rendering server. Replace with the value sent in the confirmation email
The IP address and the Port might vary from use to user. Check the confirmation email for the correct details for your connection.
NOTE: After entering your credentials, you will notice the terminal command line appears to hang. That is normal, it indicates you are connected and the SSH tunneling is set up.
3.2 Connect the client/desktop application to the remote server
This step will reset the scene, so before doing it make sure to save all your data.
Open Paraview Desktop Application (see the )
In paraview UI go to menu bar File -> Connect ..
Add Server:
After a few seconds, you get connected to the HPC automatically.
In subsequent connections you can reuse the Server, but you will need to update the host and port values each time you launch a new server session
Verifying the connection is set up correctly.
In Paraview UI go to the menu bar View and select Memory Inspector. You will notice a list of servers indicating the number of processes running on them
Summary
Open a terminal an connect to Oscar (Follow to know how to do it)
Load the Paraview module module load paraview/5.11.0_openmpi_4.0.7_intel_2020.2_slurm22
If you find any issues following this guide or require additional help, do not hesitate contacting CCV services at [email protected]
SSH (Terminal)
To log in to Oscar you need Secure Shell (SSH) on your computer.
You need log in using your Brown password. Old Oscar password can not be used for ssh any more.
There are two options for signing into Oscar: with or without VPN.
If you are connected to the Brown VPN, you have the option of to connect to Oscar without having to enter your password.
Summary of SSH Hosts
ssh.ccv.brown.edu You can connect from anywhere. You will need Two Factor Authentication
sshcampus.ccv.brown.edu You can connect when whithin Brown Wifi, Network or VPN. You will need Two Factor Authentication unless you .
macOS and Linux
To log in to Oscar, open a terminal and
If you are not connected to the Brown VPN, use the following command:
If you are connected to the Brown VPN, use the following command:
The -X allows Oscar to display windows on your machine. This allows you to open and use GUI-based applications, such as the text editor gedit.
Watch our videos on and .
Windows
Windows users need to install an SSH client. We recommend , a free SSH client for Windows.
If you are not connected to the Brown VPN, use [email protected] as the Host Name and click Open.
If you are connected to the Brown VPN, use [email protected] as the Host Name and click Open.
Confused? Watch our or .
Connecting to Oscar for the First Time
The first time you connect to Oscar you will see a message about the authenticity of the host:
You can type yes and press return. On subsequent logins you should not see this message.
You will then be prompted for your password.
Nothing will show up on the screen as you type in your password. Just type it in and press enter.
You will now be in your home directory on Oscar. In your terminal you will see a prompt like this:
Congratulations, you are now on one of the Oscar login nodes! The login nodes are for administrative tasks such as editing files and compiling code. To use Oscar for computation you will need to use the compute nodes. To get to the compute nodes from the login nodes you can either start an interactive session on a compute node, or submit a batch job.
Please do not run CPU-intense or long-running programs directly on the login nodes! The login nodes are shared by many users, and you will interrupt other users' work.
module load lftp # To load the LFTP module from Oscar
lftp -u login,passwd MyAwesomeUrl # To connect to your (S)FTP server
ls # To list files on the (S)FTP server
!ls # To list files in your directory on Oscar
get MyAwesomeFile # To download a single file
mirror # To download everything as is from the server
mirror --directory=/name_of_directory/ # To download a specific directory
ssh <username>@ssh.ccv.brown.edu
The authenticity of host 'ssh.ccv.brown.edu (138.16.172.8)' can't be established.
RSA key fingerprint is SHA256:Nt***************vL3cH7A.
Are you sure you want to continue connecting (yes/no)?
The authenticity of host 'ssh.ccv.brown.edu (138.16.172.8)' can't be established.
RSA key fingerprint is SHA256:Nt***************vL3cH7A.
Are you sure you want to continue connecting (yes/no)?
This page contains Linux commands commonly used on Oscar, basic module commands, and definitions for common terms used within this documentation.
Common Linux Commands
Command
Related Word/Phrase
Modules
Common Acronyms and Terms
module load <name>
Adds a module to your current environment. If you load using just the name of a module, you will get the default version. To load a specific version, load the module using its full name with the version: "module load gcc/6.2"
module unload <name>
Removes a module from your current environment.
CCV
Brown University's Center for Computation and Visualization. Provides software, expertise, and other services for Brown's research community. See for more information.
CESM
Stands for Community Earth System Model. "CESM is a fully-coupled, community, global climate model that provides state-of-the-art computer simulations of the Earth's past, present, and future climate states." ()
[]
CIFS
Stands for Common Internet File System. CIFS is a network protocol used for providing shared access to files and devices on the same network. Users who are unable to use SMB for file-sharing should try CIFS as an alternative.
[]
Condo
PIs can purchase condos that have a significant amount of computing resources which can be shared with others.
[]
CUDA
" is an extension of the C language, as well as a runtime library, to facilitate general-purpose programming of NVIDIA GPUs." ()
[]
HPC
Stands for High Performance Computing. HPC is the ability to process data and perform highly complex calculations at an accelerated rate. Oscar is the service that CCV offers to the Brown community for their High Performance Computing needs.
Job Array
A job array is a collection of jobs that all run the same program but on different values of a parameter.
[]
Jupyter Notebook
"The Jupyter Notebook is an open-source web application that allows you to create and share documents that contain live code, equations, visualizations and narrative text."
[]
Interactive Jobs
Jobs that allow the user to interact in real time with applications within Oscar, often from the command line. This differs from batch jobs in that each command to be run must be put in one at a time.
[]
Modules
Modules are software components that can easily be loaded or unloaded into Oscar. For instance, a user can load the Python 3 module using a module load command.
[]
MPI
Stands for . MPI is a system that aims to be the standard for portable and efficient message passing. is a technique often used in object-oriented programming and parallel programming
[]
Partition
Partitions are essentially groupings of nodes that allocate resources for specific types of tasks. On Oscar, partitions are based on job submissions through the Slurm workload manager.
[]
PI
Stands for Principal Investigator. Mainly used to refer to the individual responsible for conducting and administrating a research grant. Within Oscar, PIs have their own data directories that can be shared to students. PIs may also purchase condos.
[]
PuTTY
A client for SSH for Windows and Unix that emulates a terminal
[]
Python
An object-oriented, high-level, and popular programming language
[]
Slurm
A workload manager used within Oscar to schedule jobs
[]
SSH
Stands for Secure Shell Protocol. Used to communicate securely between computers and often used within a command-line interface (CLI) for connections to remote servers
[]
SMB
The Server Message Block (SMB) protocol is a network protocol that allows users to communicate with remote computers for file-sharing and other uses. It is one of the versions of the Common Internet File System (CIFS). Within Oscar, SMB is mainly used for file transfer.
[]
VNC
Stands for Virtual Network Computing. VNC is a desktop sharing system that allows you to remotely control another desktop. On Oscar, it is used to allow a desktop interface for applications.
[]
Quality of Service (QOS)
The job limits that are linked to a given association. For instance, Priority Accounts will generally have a higher quality of service than Exploratory Accounts.
[]
Description
cd
Change Directory
Moves the user into the specified directory
cd .. to move one directory up
cd by itself to move to home directory
cd- to move to previous directory
cd <directory-path> to move to a directory (can be an absolute path or relative path)
cp <old_file_path> <new_directory_path>
Copy
Copies the file into the specified directory
clear
Clear
Clears the terminal
cat <filename>
Concatenate
Lists the contents of a file
ls
List
Lists contents within the current directory
grep <string_to_match> <filename>
Globally Search for a Regular Expression and Print Matching Lines
Searches for the string / regular expression within the specified file and prints the line(s) with the result
pwd
Present Working Directory
Displays the path of the current directory that you are in
man <command>
Manual
Displays the manual instruction for the given command
mv <file_name> <new_directory>
Move
Moves a file into a new directory
mv <old_file_name> <new_file_name> to rename a file
mkdir <directory_name>
Make Directory
Creates a new directory
rm <directory_name>
Remove
Deletes directories and the contents within them
rmdir <directory_name>
Remove Directory
Removes the specified directory (must be empty)
touch
Touch
Creates a blank new file
Command
Description
module list
Lists all modules that are currently loaded in your software environment.
module avail
Lists all available modules on the system. Note that a module can have multiple versions.
module help <name>
Prints additional information about the given software.
Term
Description
Anaconda / Conda
A distribution of Python and R used for scientific computing that is meant to simplify package management and deployment. Conda is used for installing packages and managing their dependencies.
Within Oscar, an association refers to a combination of four factors: Cluster, Account, User, and Partition. Associations are used to control job submissions for users.
In January 2023, Oscar will be migrating to use Slurm version 22.05.7.
Slurm version 22.05.7
improves security and speed,
supports boths PMI2 and PMIX, and
provides REST APIs
allows users to prioritize their jobs via scontrol top <job_id>
While most applications will be unaffected by these changes, applications built to make use of MPI may need to be rebuilt to work properly. To help facilitate this, we are providing users who use MPI-based applications (either through Oscar's module system or built by users) with advanced access to a test cluster running the new version of Slurm. Instructions for accessing the test cluster, building MPI-based applications, and submitting MPI jobs using the new Slurm, are provided below.
Please note - some existing modules of MPI-based applications will be deprecated and removed from the system as part of this upgrade. A list of modules that will no longer be available to users following the upgrade is given at the bottom of the page.
Instructions for Testing Applications with Slurm 22.05.7
Request access to the Slurm 22.05.7 test cluster (email [email protected])
Connect to Oscar via either SSH or Open OnDemand (instructions below)
Build your application using the new MPI applications listed below
Users must contact [email protected] to obtain access to the test cluster in order to submit jobs using Slurm 22.05.7.
Connecting via SSH
Connect to Oscar using the ssh command in a terminal window
From Oscar's command line, connect to the test cluster using the command ssh node1947
From the node1947 command line, submit your jobs (either interactive or batch) as follows:
For CPU-only jobs: interact -q image-test
For GPU jobs: interact -q gpu
Include the following line within your batch script and then submit using the
Connecting via Open OnDemand
Open a web browser and connect to poodcit2.services.brown.edu
Login with your Oscar username and password
Start a session using the Advanced Desktop App
Only the Advanced Desktop App will connect to the test cluster
The Advanced Desktop App must connect to the gpu partition
MPI Applications
Migrated or New Modules
If the "Current Module Version" for an application is blank, a new version is built for the application.
Application
Current Module Version
Migrated or New Module Version
To build custom applications:
We recommend using following MPI modules to build your custom applications:
A new module might be available for a deprecated application module. Please search the table above to check if a new module is available for an application.
Application
Deprecated Module
Submit your job
sbatch
command, as usual
For CPU-only jobs: #SBATCH -p image-test
For GPU jobs: #SBATCH -p gpu
Select the gpu partition and click the launch button.