Resource Requests
Aside from the time, RAM, and CPU requirements listed on the SGEBasics page, we have several other requestable resources. Generally, if you don't know if you need a particular resource, you should use the default. These can be generated with the command
qconf -sc | awk '{ if ($5 != "NO") { print }}'
name | shortcut | type | relop | requestable | consumable | default | urgency |
---|---|---|---|---|---|---|---|
arch | a | RESTRING | == | YES | NO | NONE | 0 |
avx | avx | BOOL | == | YES | NO | FALSE | 0 |
calendar | c | RESTRING | == | YES | NO | NONE | 0 |
cpu | cpu | DOUBLE | >= | YES | NO | 0 | 0 |
cpu_flags | c_f | STRING | == | YES | NO | NONE | 0 |
cuda | cuda | INT | <= | YES | JOB | 0 | 0 |
display_win_gui | dwg | BOOL | == | YES | NO | 0 | 0 |
exclusive | excl | BOOL | EXCL | YES | YES | 0 | 1000 |
h_core | h_core | MEMORY | <= | YES | NO | 0 | 0 |
h_cpu | h_cpu | TIME | <= | YES | NO | 0:0:0 | 0 |
h_data | h_data | MEMORY | <= | YES | NO | 0 | 0 |
h_fsize | h_fsize | MEMORY | <= | YES | NO | 0 | 0 |
h_rss | h_rss | MEMORY | <= | YES | NO | 0 | 0 |
h_rt | h_rt | TIME | <= | FORCED | NO | 0:0:0 | 0 |
h_stack | h_stack | MEMORY | <= | YES | NO | 0 | 0 |
h_vmem | h_vmem | MEMORY | <= | YES | NO | 0 | 0 |
hostname | h | HOST | == | YES | NO | NONE | 0 |
infiniband | ib | BOOL | == | YES | NO | FALSE | 0 |
m_core | core | INT | <= | YES | NO | 0 | 0 |
m_socket | socket | INT | <= | YES | NO | 0 | 0 |
m_thread | thread | INT | <= | YES | NO | 0 | 0 |
m_topology | topo | RESTRING | == | YES | NO | NONE | 0 |
m_topology_inuse | utopo | RESTRING | == | YES | NO | NONE | 0 |
mem_free | mf | MEMORY | <= | YES | NO | 0 | 0 |
mem_total | mt | MEMORY | <= | YES | NO | 0 | 0 |
mem_used | mu | MEMORY | >= | YES | NO | 0 | 0 |
memory | mem | MEMORY | <= | FORCED | YES | 0 | 0 |
num_proc | p | INT | == | YES | NO | 0 | 0 |
qname | q | RESTRING | == | YES | NO | NONE | 0 |
s_core | s_core | MEMORY | <= | YES | NO | 0 | 0 |
s_cpu | s_cpu | TIME | <= | YES | NO | 0:0:0 | 0 |
s_data | s_data | MEMORY | <= | YES | NO | 0 | 0 |
s_fsize | s_fsize | MEMORY | <= | YES | NO | 0 | 0 |
s_rss | s_rss | MEMORY | <= | YES | NO | 0 | 0 |
s_rt | s_rt | TIME | <= | YES | NO | 0:0:0 | 0 |
s_stack | s_stack | MEMORY | <= | YES | NO | 0 | 0 |
s_vmem | s_vmem | MEMORY | <= | YES | NO | 0 | 0 |
slots | s | INT | <= | YES | YES | 1 | 1000 |
swap_free | sf | MEMORY | <= | YES | NO | 0 | 0 |
swap_rate | sr | MEMORY | >= | YES | NO | 0 | 0 |
swap_rsvd | srsv | MEMORY | >= | YES | NO | 0 | 0 |
swap_total | st | MEMORY | <= | YES | NO | 0 | 0 |
swap_used | su | MEMORY | >= | YES | NO | 0 | 0 |
virtual_free | vf | MEMORY | <= | YES | NO | 0 | 0 |
virtual_total | vt | MEMORY | <= | YES | NO | 0 | 0 |
virtual_used | vu | MEMORY | >= | YES | NO | 0 | 0 |
The good news is that most of these nobody ever uses. There are a couple of exceptions, though:
Infiniband
First of all, let me state that just because it sounds "cool" doesn't mean you need it or even want it. Infiniband does absolutely no good if running in a 'single' parallel environment. Infiniband is a high-speed host-to-host communication fabric. It is used in conjunction with MPI jobs (discussed below). Several times we have had jobs which could run just fine, except that the submitter requested Infiniband, and all the nodes with Infiniband were currently busy. In fact, some of our fastest nodes do not have Infiniband, so by requesting it when you don't need it, you are actually slowing down your job. To request Infiniband, add -l ib=true to your qsub command-line.
CUDA
CUDA is the resource required for GPU computing. We have a very small number of nodes which have GPUs installed. To request one of these nodes, add -l cuda=true to your qsub command-line.
Exclusive
Some programs just don't play nicely with others. They will attempt to use all available memory or will try to use all the cores it can use. The way to be a nice neighbor if your program has this problem is to request exclusive use of a node with -l excl=true. This can also be useful for benchmarking, where you can be sure that no other jobs are interfering with yours.
Parallel Jobs
There are two ways jobs can run in parallel, intranode and internode. Note: Beocat will not automatically make a job run in parallel. Have I said that enough? It's a common misperception.
Intranode jobs
Intranode jobs are easier to code and can take advantage of many common libraries, such as OpenMP, or Java's threads. Many times, your program will need to know how many cores you want it to use. Many will use all available cores if not told explicitly otherwise. This can be a problem when you are sharing resources, as Beocat does. To request multiple cores, use the qsub directive '-pe single n', where n is the number of cores you wish to use. If your command can take an environment variable, you can use $nslots to tell how many cores you've been allocated.
Internode (MPI) jobs
"Talking" between nodes is trickier that talking between cores on the same node. The specification for doing so is called "Message Passing Interface", or MPI. We have OpenMPI installed on Beocat for this purpose. Most programs written to take advantage of large multi-node systems will use MPI. You can tell if you have an MPI-enabled program because its directions will tell you to run 'mpirun program'. Requesting MPI resources is only mildly more difficult than requesting single-node jobs. Instead of using '-pe single n' for your qsub request, you will use one of the following:
Parallel Environment | Description |
---|---|
mpi-fill | This environment will use as many slots on each node as it can until it reaches the number of cores you have requested. |
mpi-spread | This environment will spread itself out over as many nodes as possible until it reaches the number of cores you have requested. |
mpi-1 | This environment will allocate the slots you've requested 1 per node. |
mpi-2 | This environment will allocate the slots you've requested 2 per node. You must request cores as a multiple of 2 |
mpi-4 | This environment will allocate the slots you've requested 4 per node. You must request cores as a multiple of 4 |
mpi-8 | This environment will allocate the slots you've requested 8 per node. You must request cores as a multiple of 8 |
mpi-10 | This environment will allocate the slots you've requested 10 per node. You must request cores as a multiple of 10 |
mpi-12 | This environment will allocate the slots you've requested 12 per node. You must request cores as a multiple of 12 |
mpi-16 | This environment will allocate the slots you've requested 16 per node. You must request cores as a multiple of 16 |
mpi-20 | This environment will allocate the slots you've requested 20 per node. You must request cores as a multiple of 20 |
mpi-80 | This environment will allocate the slots you've requested 80 per node. You must request cores as a multiple of 80 |
Some quick examples:
-pe mpi-4 16 will give you 4 chunks of 4 cores apiece. They might all happen to be allocated on the same node (16 cores), on 4 different nodes (4 cores each), on 3 nodes (8 cores on one and 4 cores on the other two), or on 2 nodes (8 cores each).
-pe mpi-fill 40 will give you 40 cores, but will attempt to get them all on the same node.
-pe mpi-fill 100 will give you 100 cores, and place them on as few nodes as possible. In this case it's likely you would get a full mage (80 cores) and either part of another mage (the remaining 20 cores) or one of the 20-core elves.
-pe mpi-spread 40 will give you 40 cores, and will attempt to place each on a separate node.
Requesting memory for multi-core jobs
All memory requests are per core. One of the more common scenarios is where somebody will need, say 20 cores and 400 GB of memory. So they will make a request like '-pe single 20, -l mem=400G' This will never run, because what you are really requesting is 20 cores and 8000GB of memory (20 * 400). Since we have no nodes with 8000 terabytes of memory, the job will never run. In this case, you will divide the 400GB total memory request by the number of cores (20), so the correct command would be '-pe single 20, -l mem=20G'.
Other Handy SGE Features
Email status changes
One of the most commonly used options when submitting jobs not related to resource requests is to have have SGE email you when a job changes its status. This takes two directives to qsub: '-M someone@somewhere.com' will give the email address to which to send status updates. '-m abe' is probably the most common directive given for when to send updates. This will send email messages when a job (a)borts, (b)egins, or (e)nds. Other possibilities are (s)uspended and (n)ever.
Job Naming
If you have several jobs in the queue, running the same script with different parameters, it's handy to have a different name for each job as it shows up in the queue. This is accomplished with the '-N JobName' qsub directive.
Combining Output Streams
Normally, SGE will create two files for output. One will be .ejobnumber and the other .ojobnumber. If you want both of these to be combined into a single file, you can use the qsub directive '-j y'.
Running from the Current Directory
By default, jobs run from your home directory. Many programs incorrectly assume that you are running the script from the current directory. You can use the '-cwd' directive to change to the "current working directory" you used when submitting the job.
Running in a specific class of machine
If you want to run on a specific class of machines, e.g., the Dwarves, you can add the flag "-q \*@@dwarves" to select that queue.
SGE Environment Variables
Within an actual job, sometimes you need to know specific things about the running environment to setup your scripts correctly. Here is a listing of environment variables that SGE makes available to you. Of course the value of these variables will be different based on many different factors.
HOSTNAME=titan1.beocat
SGE_TASK_STEPSIZE=undefined
SGE_INFOTEXT_MAX_COLUMN=5000
SHELL=/usr/local/bin/sh
NHOSTS=2
SGE_O_WORKDIR=/homes/mozes
TMPDIR=/tmp/105.1.batch.q
SGE_O_HOME=/homes/mozes
SGE_ARCH=lx24-amd64
SGE_CELL=default
RESTARTED=0
ARC=lx24-amd64
USER=mozes
QUEUE=batch.q
PVM_ARCH=LINUX64
SGE_TASK_ID=undefined
SGE_BINARY_PATH=/opt/sge/bin/lx24-amd64
SGE_STDERR_PATH=/homes/mozes/sge_test.sub.e105
SGE_STDOUT_PATH=/homes/mozes/sge_test.sub.o105
SGE_ACCOUNT=sge
SGE_RSH_COMMAND=builtin
JOB_SCRIPT=/opt/sge/default/spool/titan1/job_scripts/105
JOB_NAME=sge_test.sub
SGE_NOMSG=1
SGE_ROOT=/opt/sge
REQNAME=sge_test.sub
SGE_JOB_SPOOL_DIR=/opt/sge/default/spool/titan1/active_jobs/105.1
ENVIRONMENT=BATCH
PE_HOSTFILE=/opt/sge/default/spool/titan1/active_jobs/105.1/pe_hostfile
SGE_CWD_PATH=/homes/mozes
NQUEUES=2
SGE_O_LOGNAME=mozes
SGE_O_MAIL=/var/mail/mozes
TMP=/tmp/105.1.batch.q
JOB_ID=105
LOGNAME=mozes
PE=mpi-fill
SGE_TASK_FIRST=undefined
SGE_O_HOST=loki
SGE_O_SHELL=/bin/bash
SGE_CLUSTER_NAME=beocat
REQUEST=sge_test.sub
NSLOTS=32
SGE_STDIN_PATH=/dev/null
Sometimes it is nice to know what hosts you have access to during a PE job. You would checkout the PE_HOSTFILE to know that. If your job has been restarted, it is nice to be able to change what happens rather than redoing all of your work. If this is the case, RESTARTED would equal 1. There are lots of useful Environment Variables there, I will leave it to you to identify the ones you want.
Some of the most commonly-used variables we see used are $NSLOTS, $HOSTNAME, and $SGE_TASK_ID (used for array jobs, discussed below).
Running from a qsub Submit Script
No doubt after you've run a few jobs you get tired of typing something like 'qsub -l mem=2G,h_rt=10:00 -pe single 8 -n MyJobTitle MyScript.sh'. How are you supposed to remember all of these every time? The answer is to create a 'submit script', which outlines all of these for you. Below is a sample submit script, which you can modify and use for your own purposes.
#!/bin/bash
## A Sample qsub script created by Kyle Hutson
##
## Note: Usually a '#" at the beginning of the line is ignored. However, in
## the case of qsub, lines beginning with #$ are commands for qsub itself, so
## I have taken the convention here of starting *every* line with a '#', just
## Delete the first one if you want to use that line, and then modify it to
## your own purposes. The only exception here is the first line, which *must*
## be #!/bin/bash (or another valid shell).
## Specify the amount of RAM needed _per_core_. Default is 1G
##$ -l mem=1G
## Specify the maximum runtime. Default is 1 hour (1:00:00)
##$ -l h_rt=1:00:00
## Require the use of infiniband. If you don't know what this is, you probably
## don't need it. Default is "FALSE"
##$ -l ib=TRUE
## CUDA directive. If You don't know what this is, you probably don't need it
## Default is "FALSE"
##$ -l cuda=TRUE
## Parallel environment. Syntax is '-pe Environment NumberOfCores' A list of
## valid environments can be found at
## https://support.beocat.ksu.edu/BeocatDocs/index.php/AdvancedSGE (section 2). One
## quick note here. Jobs requesting 16 or fewer cores tend to get scheduled
## fairly quickly. If you need a job that requires more than that, you might
## benefit from emailing us at beocat@cs.ksu.edu to see how we can assist in
## getting your job scheduled in a reasonable amount of time. Default is
## "single 1"
##$ -pe single 12
##$ -pe mpi-1 2
##$ -pe mpi-fill 20
##$ -pe mpi-spread 16
## Checkpointing. Options are BLCR or dmtcp. Default is no checkpointing.
##$ -ckpt dmtcp
## Use the current working directory instead of your home directory
##$ -cwd
## Merge output and error text streams into a single stream
##$ -j y
## Name my job, to make it easier to find in the queue
##$ -N MyJobTitle
## And finally, we run the job we came here to do.
## $HOME/ProgramDir/ProgramName ProgramArguments
## OR, for the case of MPI-capable jobs
## mpirun $HOME/path/MpiJobName
## Send email when a job is aborted (a), begins (b), and/or ends (e)
##$ -m abe
## Email address to send the email to based on the above line.
##$ -M myemail@ksu.edu
File Access
Beocat has a variety of options for storing and accessing your files. Every user has a home directory for general use which is limited in size, has decent file access performance, and will soon be backed up nightly. Larger files should be stored in the /bulk subdirectories which have the same decent performance but are not backed up. The /scratch file system will soon be implemented on a Lustre file system that will provide very fast temporary file access. When fast IO is critical to the application performance, access to the local disk on each node or to a RAM disk are the best options.
Home directory
Every user has a /homes/username directory that they drop into when they log into Beocat. The home directory is for general use and provides decent performance for most file IO. Disk space in each home directory is limited to 1 TB, so larger files should be kept in the /bulk directory, and there is a limit of 100,000 files in each subdirectory in your account. This file system is fully redundant, so 3 specific hard disks would need to fail before any data was lost. All files will soon be backed up nightly to a separate file server in Nichols Hall, so if you do accidentally delete something it can be recovered.
Bulk directory
Each user also has a /bulk/username directory where large files should be stored. File access is the same speed as for the home directories, and the same limit of 100,000 files per subdirectory applies. There is no limit to the disk space you can use in your bulk directory, but the files there will not be backed up. They are still redundantly stored so you don't need to worry about losing data to hardware failures, just don't delete something by accident. Unused files are automatically removed after two years. If you need to back up large files in the bulk directory, talk to Dan Andresen (dan@ksu.edu) about purchasing some hard disks for archival storage.
Scratch file system
The /scratch file system will soon be using the Lustre software which is much faster than the speed of the file access on /homes or /bulk. In order to use scratch, you first need to make a directory for yourself. Scratch offers greater speed, no limit to the size of files nor the number of files in each subdirectory. It is meant as temporary space for prepositioning files and accessing them during runs. Once runs are completed, any files that need to be kept should be moved to your home or bulk directories since files on the scratch file system get purged after 30 days. Lustre is faster than the home and bulk file systems in part because it does not redundantly store files by striping them across multiple disks, so if a hard disk fails data will be lost. When we get scratch set up to use Lustre we will post the difference in file access rates.
mkdir /scratch/username
Local disk
If you are running on a single node, it may also be faster to access your files from the local disk on that node. This can be done conveniently using the environment variable $TMPDIR which is set to point to a subdirectory on /tmp set up for each job. You may need to copy files to local disk at the start of your script, or set the output directory for your application to point to a file on the local disk, then you'll need to copy any files you want off the local disk before the job finishes since SGE will remove all files in your job's directory on /tmp on completion of the job or when it aborts. When we get the scratch file system working with Lustre, it may end up being faster than accessing local disk so we will post the access rates for each. Most nodes have around 600 GB of file space accessible on the local disk.
Copy input files to the tmp directory if needed
cp input_files $TMPDIR
Make an 'out' directory to pass to the app if needed
mkdir $TMPDIR/out
Example of running an app and passing the tmp directory in/out
app -input_directory $TMPDIR -output_directory $TMPDIR/out
Copy the 'out' directory back to the current working directory after the run
cp -rp $TMPDIR/out .
RAM disk
If you need ultrafast access to files, you can use a RAM disk which is a file system set up in the memory of the compute node you are running on. The RAM disk is limited to a maximum of half of the physical memory on that node, and you should account for this usage when you request memory for your job. You'll need to set up a directory on the RAM disk manually at the start of your script, you may need to copy files over before your script runs the application, then you'll need to copy data off the RAM disk and you absolutely must remove all traces of the RAM disk at the end of your job or it will remain in memory and affect future jobs. If your job crashes during the run, you'll need to email beocat@cs.ksu.edu to let us know which node needs the RAM disk cleaned up. Below is an example of how to use the RAM disk.
Make the RAM disk and copy input files over if necessary
mkdir /dev/shm/username-$JOB_ID
cp any_input_files /dev/shm/username-$JOB_ID
Run the application, possibly giving it the path to the RAM disk to use for output files
app -output_directory /dev/shm/username-$JOB_ID
Copy files from the RAM disk to the current working directory and clean it up
cp /dev/shm/username-$JOB_ID/* .
rm -fr /dev/shm/username-$JOB_ID
When you leave KSU
If you are done with your account and leaving KSU, please clean up your directory, move any files to your supervisor's account that need to be kept after you leave, and notify us so that we can disable your account. The easiest way to move your files to your supervisor's account is for them to set up a subdirectory for you with the appropriate write permissions. The example below shows moving just a user's 'data' subdirectory to their supervisor. The 'nohup' command is used so that the move will continue even if the window you are doing the move from gets disconnected.
Supervisor:
mkdir /bulk/supervisorsname/username
chmod ugo+w /bulk/supervisorsname/username
Student:
nohup mv /homes/username/data /bulk/supervisorsname/username &
Array Jobs
One of SGE's useful options is the ability to run "Array Jobs"
It can be used with the following option to qsub.
-t n[-m[:s]] Submits a so called Array Job, i.e. an array of identical tasks being differentiated only by an index number and being treated by Grid Engine almost like a series of jobs. The option argument to -t specifies the number of array job tasks and the index number which will be associated with the tasks. The index numbers will be exported to the job tasks via the environment variable SGE_TASK_ID. The option arguments n, m and s will be available through the environment variables SGE_TASK_FIRST, SGE_TASK_LAST and SGE_TASK_STEPSIZE. Following restrictions apply to the values n and m: 1 <= n <= 1,000,000 1 <= m <= 1,000,000 n <= m The task id range specified in the option argument may be a single number, a simple range of the form n-m or a range with a step size. Hence, the task id range specified by 2-10:2 would result in the task id indexes 2, 4, 6, 8, and 10, for a total of 5 identical tasks, each with the environment variable SGE_TASK_ID containing one of the 5 index numbers. Array jobs are commonly used to execute the same type of operation on varying input data sets correlated with the task index number. The number of tasks in a array job is unlimited. STDOUT and STDERR of array job tasks will be written into different files with the default location <jobname>.['e'|'o']<job_id>'.'<task_id>
Examples
Change the Size of the Run
Array Jobs have a variety of uses, one of the easiest to comprehend is the following:
I have an application, app1 I need to run the exact same way, on the same data set, with only the size of the run changing.
My original script looks like this:
#!/bin/bash
RUNSIZE=50
#RUNSIZE=100
#RUNSIZE=150
#RUNSIZE=200
app1 $RUNSIZE dataset.txt
For every run of that job I have to change the RUNSIZE variable, and submit each script. This gets tedious.
With Array Jobs the script can be written like so:
#!/bin/bash
#$ -t 50:200:50
RUNSIZE=$SGE_TASK_ID
app1 $RUNSIZE dataset.txt
I then submit that job, and SGE understands that it needs to run it 4 times, once for each task. It also knows that it can and should run these tasks in parallel.
Choosing a Dataset
A slightly more complex use of Array Jobs is the following:
I have an application, app2, that needs to be run against every line of my dataset. Every line changes how app2 runs slightly, but I need to compare the runs against each other.
Originally I had to take each line of my dataset and generate a new submit script and submit the job. This was done with yet another script:
#!/bin/bash
DATASET=dataset.txt
scriptnum=0
while read LINE
do
echo "app2 $LINE" > ${scriptnum}.sh
qsub ${scriptnum}.sh
scriptnum=$(( $scriptnum + 1 ))
done < $DATASET
Not only is this needlessly complex, it is also slow, as qsub has to verify each job as it is submitted. This can be done easily with array jobs, as long as you know the number of lines in the dataset. This number can be obtained like so: wc -l dataset.txt in this case lets call it 5000.
#!/bin/bash
#$ -t 1:5000
app2 `sed -n "${SGE_TASK_ID}p" dataset.txt`
This uses a subshell via `, and has the sed command print out only the line number $SGE_TASK_ID out of the file dataset.txt.
Not only is this a smaller script, it is also faster to submit because it is one job instead of 5000, so qsub doesn't have to verify as many.
To give you an idea about time saved: submitting 1 job takes 1-2 seconds. by extension if you are submitting 5000, that is 5,000-10,000 seconds, or 1.5-3 hours.
Running jobs interactively
Some jobs just don't behave like we think they should, or need to be run with somebody sitting at the keyboard and typing in response to the output the computers are generating. Beocat has a facility for this, called 'qrsh'. qrsh uses the exact same command-line arguments as qsub. If no node is available with your resource requirements, qrsh will tell you
Your "qrsh" request could not be scheduled, try again later.
Note that, like qsub, your interactive job will timeout after your allotted time has passed.
Altering Job Requests
We generally do not support users to modify job parameters once the job has been submitted. It can be done, but there are numerous catches, and all of the variations can be a bit problematic; it is normally easier to simply delete the job and resubmit it with the right parameters. If your job doesn't start after modifying such parameters (after a reasonable amount of time), delete the job and resubmit it.
qalter
qalter is the command that can be used to modify parameters of the job after it has been submitted. Note: resource requests (memory, runtime, et. al.) can only be modified on jobs that have yet to start running.
Changing resource requests
Syntax:
qalter -l $all_resources $jobid
When modifying resource requests, you must specify all of the resources your job needs, not just the one you plan to change. If you just specify h_rt, it will drop the memory request. If you just specify memory, it will drop the h_rt. And so on. This leads to jobs failing to start.
Changing core requests
Syntax:
qalter -pe $pe_name $number_of_cores $jobid
If you request more cores than are available in the parallel environment that you need, the job may fail to start.
- i.e. requesting 400 cores in the single environment will fail due to the fact that we have no machines with 400 cores.
Determining why a job is not running
Syntax:
qalter -w v $jobid
This will output the scheduler's reasoning as to why the job has not started. Note that lines like:
Job 1122334455 cannot run in PE "single" because it only offers 0 slots
Are usually red herrings. Sometimes they are indicative that the scheduler cannot meet the resources requests for that job at this moment in time.
Sometimes you will see output like this:
Job 1122334455 does not request 'forced' resource "memory" of queue instance batch.q@elf73.beocat
In this case the user performed a qalter and forgot to specify the memory request. The job will never run in this state.
Other times it will have lots of lines like this:
verification: found possible assignment with 1 slots
This indicates that the job should be scheduled shortly.
Killable jobs
There are a growing number of machines within Beocat that are owned by a particular person or group. Normally jobs from users that aren't in the group designated by the owner of these machines cannot use them. This is because we have guaranteed that the nodes will be accessible and available to the owner at any given time. We will allow others to use these nodes if they designate their job as "killable." If your job is designated as killable, your job will be able to use these nodes, but can (and will) be killed off at any point in time to make way for the designated owner's jobs. Jobs that are marked killable will be re-queued and may restart on another node.
The way you would designate your job as killable is to add -l killable to the qsub or qrsh arguments. This could be either on the command-line or in your script file.
Note: This is a submit-time only request, it cannot be added by a normal user after the job has been submitted. If you would like jobs modified to be killable after the jobs have been submitted (and it is too much work to qdel the jobs and re-submit), send an e-mail to the administrators detailing the job ids and what you would like done.
Scheduling Priority
The scheduler uses a complex formula to determine the order that jobs get scheduled in. Jobs in general get run in the order that they are submitted to the queue with the following exceptions. Jobs for users in a group that owns nodes will immediately get scheduled on those nodes even if that means bumping existing jobs off. Users in groups that have contributed funds to Beocat may have higher scheduling priority. You can check the base scheduling priority of each group using qconf -sst. If you do not have a group your jobs are scheduled using BEODEFAULT. The higher the priority, the faster your job will be moved to the front of the queue. A fair scheduling algorithm adjusts this scheduling priority down as users in that group submit more jobs.
Since all users not in a group having higher priority get put into BEODEFAULT, the priority is always very low and each job gets scheduled in the order it was submitted. Groups with a higher priority may jump ahead of the BEODEFAULT jobs, but if these groups are submitting lots of jobs their priority will become low as well. Groups with the highest priority that are submitting the fewest jobs may see those jobs moved to the front of the queue quickly.
When processing cores become available, the scheduler looks at the head of the queue to find jobs that will fit within the resources available. Shorter jobs of 12 hours or less get marked as killable and will be run on nodes owned by other groups. These jobs will jump past longer jobs when resources become available on owned nodes. Many jobs in the queue may require more memory than is available on some nodes, so smaller memory jobs will be scheduled ahead of larger memory jobs on hosts with more limited memory. kstat -q will show you the order in the queue and allow you to see jobs marked as "killable" and those that require large memory.
Job Accounting
Some people may find it useful to know what their job did during its run. The qacct tool will read SGE's accounting file and give you summarized or detailed views on jobs that have run within Beocat.
qacct
This data can usually be used to diagnose two very common job failures.
Job debugging
It is simplest if you know the job number of the job you are trying to get information on.
# if you know the jobid, put it here:
qacct -j 1122334455
# if you don't know the job id, you can look at your jobs over some number of days in this case the past 14 days:
qacct -o $USER -d 14 -j
My job didn't do anything when it ran!
qname batch.q hostname mage07.beocat group some_user_users owner some_user project BEODEFAULT department defaultdepartment jobname my_job_script.sh jobnumber 1122334455 ... snipped to save space ... exit_status 1 ru_wallclock 1s ru_utime 0.030s ru_stime 0.030s ... snipped to save space ... arid undefined category -u some_user -q batch.q,long.q -l h_rt=604800,mem_free=1024.0M,memory=2G
If you look at the line showing ru_wallclock. You can see that it shows 1s. This means that the job started and then promptly ended. This points to something being wrong with your submission script. Perhaps there is a typo somewhere in it.
My job ran but didn't finish!
qname batch.q hostname scout59.beocat group some_user_users owner some_user project BEODEFAULT department defaultdepartment jobname my_job_script.sh jobnumber 1122334455 ... snipped to save space ... slots 1 failed 37 : qmaster enforced h_rt, h_cpu, or h_vmem limit exit_status 0 ru_wallclock 21600s ru_utime 0.130s ru_stime 0.020s ... snipped to save space ... arid undefined category -u some_user -q batch.q,long.q -l h_rt=21600,mem_free=512.0M,memory=1G
If you look at the lines showing failed, ru_wallclock and category we can see some pointers to the issue. It didn't finish because the scheduler (qmaster) enforced some limit. If you look at the category line, the only limit requested was h_rt. So it was a runtime (wallclock) limit. Comparing ru_wallclock and the h_rt request, we can see that it ran until the h_rt time was hit, and then the scheduler enforce the limit and killed the job. You will need to resubmit the job and ask for more time next time.