SP2 Usage Notes
The main SP2 node is called 鈥sp2.dal.ca鈥. This is the name to which you should login via telnet (except before the first time you use the SP2; see 鈥淧assword changes鈥 below). Each node has an 鈥渋nternal鈥 name of the form
sp2-eN.ucis.dal.ca
where 鈥淣鈥 is the SP2 node number, 1, 3, 5, or 7. While it is possible to do so, usually there is no need to login to a particular node.
The 鈥vi鈥, 鈥jove鈥, and 鈥pico鈥 text editors are available.
The command to run the Fortran 90 compiler is 鈥xlf鈥. Fortran options can be looked up with the command 鈥man xlf鈥. It is recommended that you use at least 鈥-O鈥 optimization because this can achieve a substantial improvement in execution time.
The command to run the High Performance Fortran compiler, with which it is possible to automatically parallelize a program, is 鈥xlhpf鈥. Command options can be looked up with 鈥man xlhpf鈥.
Password changes
All the SP2 nodes share the same password file. The master copy of the password file is stored on the control workstation and periodically is copied to the nodes. Consequently, to change your password, and, especially, before you use the SP2 for the first time, you must login to the SP2 control workstation 鈥sp2c.ucis.dal.ca鈥 and use the 鈥passwd鈥 command there.
Note: If you change your password on 鈥sp2.dal.ca鈥 or one of the other nodes the change WILL NOT LAST because it will be overwritten by the password from the control workstation.
After you have changed your password on the control workstation it will take until about 10 past the hour for the change to get to 鈥sp2.dal.ca鈥 and the other nodes.
Mail Forwarding
The first time you login you should probably set up an e-mail forwarding to the place where you regularly read e-mail (create a file 鈥.forward鈥 in your home directory containing the address to which e-mail sent to you on the SP2 should be redirected; test this by sending mail to 鈥username@sp2.dal.ca鈥, where 鈥渦sername鈥 is your SP2 username).
Disk space
User permanent disk space is allocated in the /home file system. At this time there are no disk space quotas, so please be conservative in your disk space usage. /home is locally mounted on SP2 node 1 and NFS mounted on the other nodes so that files in /home can be accessed from any node. Please do not use the /tmp file system for temporary files: it is not very large and in the future may be even smaller. Each node has its own two gigabyte /usertmp file system for temporary files.
Temporary files that need to be shared across nodes can be put in the two gigabyte /globaltmp file system which is locally mounted on node 1 and NFS mounted on the other nodes.
There is currently no regular removal of files from /tmp or /usertmp (files stored in these file systems are kept across system reboots), so please clean up files you don鈥檛 need any more.
Shell
The standard system shell is the Korn Shell, ksh. Please don鈥檛 change your login shell, as this may cause difficulties with system login procedures. If you prefer to use a different shell, 鈥渆xec鈥 it in your .profile file. Note: The current directory is not in the PATH environment variable. Thus to execute a program from your current directory you must type 鈥./program_file鈥, not 鈥program_file鈥.
LoadLeveler
Please run anything which needs a non-trivial amount of CPU time via LoadLeveler. This is the only way in which machine resources can be used efficiently and fairly. If you run a program interactively, e.g.,
$ ./program
or
$ ./program &
it runs on the node to which you have logged in (i.e., node 1). If several people all do this they all run on the same node and compete with each other for CPU time.
LoadLeveler automatically runs your program on the least loaded SP2 node. To use LoadLeveler make a LoadLeveler command file (e.g., 鈥job.cmd鈥) that looks like
`# @ class = short
# @ error = name_of_file_to_hold_error_output
# @ output = name_of_file_to_hold_regular_output
# @ queue
./program`
and then use the 鈥llsubmit鈥 command:
$ llsubmit job.cmd
The 鈥llstatus鈥 and 鈥llq鈥 commands can be used to track running of your job. A mail message is sent to you when a job completes (this can be controlled with the 鈥渘otification鈥 and 鈥渘otify_user鈥 LoadLeveler statements).
Three LoadLeveler job classes have been defined:
- short - Maximum 15 minutes CPU time.
- medium - Maximum one hour CPU time.
- long - No CPU time limit
All the job classes have the same base 鈥渜ueue priority鈥, so scheduling of jobs to run is first come, first served. You may submit any number of jobs, but only two at a time may be running or under consideration to be run. This prevents a user submitting numerous jobs to 鈥渞eserve鈥 blocks of time by getting in ahead of everyone else.
Each SP2 node is allowed to run just one LoadLeveler job at a time, so the job has exclusive use of that node.
There is a minimally site-specific, from the Maui High Performance Computing Center.
Reporting problems
Send mail to cfo-systems@dal.ca.