Job and File Storage Space
Disk Space Organization Overview
Below is a list of of different forms of disk space available on the cluster to the user for reading and writing. (Unless noted otherwise, there is no backup and thus no warranty about the security of the data.)
/home/username | user home directory with regular backups; 90 MB default SHARED quota. |
/home/username/jobs | only available on stats.phys.uconn.edu ideal for launching Condor jobs. No backups, no per-user quota; ~150 GB space total |
/scratch | temporary space shared by all users; ~100 GB total; NO BACKUP, DO NOT PERMANENTLY STORE RESULTS HERE |
Results
If you have files that you will not be using and do not want to delete, please move them to your own personal storage off of the cluster. The jobs directory is NOT backed up so any results that are important should be moved off of the cluster. The cluster is where you stage your work but it is not a recommended permanent home for your results.
/home/username
Upon logging in, the user lands in his or her home directory: /home/username. This modest amount of space is intended as private space for development, testing of code etc. Results may safely be kept here for some time as this space is backed up. It is not intended for large data files or the many files generated by Condor jobs.
The quota for this directory is shared among all statistics users. If one person fills their home directory it will prevent other users from being able to write to theirs. For large files, please use the jobs directory.
/home/username/jobs
A special directory /home/username/jobs is a more efficient space for launching jobs. It is actually a link to local space on stats.phys.uconn.edu (and is not accessible from processing nodes.) It is not bound by quotas.
This is the recommended location for storing files that are in use.
/scratch
For this and other purposes a temporary directory /scratch is also available from all nodes. This is a larger working space suitable for collaborative work and launching of jobs. Big files that need to be transfered to the cluster can be copied directly to the scratch space by specifying /scratch as a destination in your scp client. For example, compression-enabled transfer of a large file directly to a working directory under the cluster's scratch space with the console-based scp client looks as follows:
scp -C WarAndPeace.txt Intolstoy@stats.phys.uconn.edu:/scratch/LeosSpace/novel_parse
scp | -C | WarAndPeace.txt | lntolstoy@stats.phys.uconn.edu:/scratch/LeosSpace/novel_parse |
secure copy command | specific option for scp | local file in present directory | remote file location on stats server |