User Tools

Site Tools


hpc_concepts

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revisionPrevious revision
Next revisionBoth sides next revision
hpc_concepts [2009/11/05 09:18] 172.26.0.166hpc_concepts [2009/11/05 09:19] 172.26.0.166
Line 22: Line 22:
  
 SGE is typically used on a computer farm or high-performance computing (HPC) cluster and is responsible for **accepting, scheduling, dispatching, and managing the remote and distributed execution of large numbers of standalone, parallel or interactive user jobs**. It also manages and schedules the allocation of distributed resources such as processors, memory, disk space, and software licenses. SGE is typically used on a computer farm or high-performance computing (HPC) cluster and is responsible for **accepting, scheduling, dispatching, and managing the remote and distributed execution of large numbers of standalone, parallel or interactive user jobs**. It also manages and schedules the allocation of distributed resources such as processors, memory, disk space, and software licenses.
-http://wiki.gridengine.info/wiki/index.php/Main_Page+http://wiki.gridengine.info/wiki/index.php/Main_Page [[Basic usage of Grid Engine (commands)]]
  
-[[Basic usage of Grid Engine (commands)]] + 
-** + 
-SLURM: A Highly Scalable Resource Manager**+====SLURM: A Highly Scalable Resource Manager====
  
 SLURM is an open-source resource manager designed for Linux clusters of all sizes. It provides three key functions. First it allocates exclusive and/or non-exclusive access to resources (computer nodes) to users for some duration of time so they can perform work. Second, it provides a framework for starting, executing, and monitoring work (typically a parallel job) on a set of allocated nodes. Finally, it arbitrates contention for resources by managing a queue of pending work.  SLURM is an open-source resource manager designed for Linux clusters of all sizes. It provides three key functions. First it allocates exclusive and/or non-exclusive access to resources (computer nodes) to users for some duration of time so they can perform work. Second, it provides a framework for starting, executing, and monitoring work (typically a parallel job) on a set of allocated nodes. Finally, it arbitrates contention for resources by managing a queue of pending work. 
hpc_concepts.txt · Last modified: 2010/05/22 14:19 by 127.0.0.1