[lug] Managing work queue for scientific processing

Vince Dean vincedean at frii.com
Tue May 3 17:50:20 MDT 2011


Dear BLUG folks:

I'm looking for tools to run the data processing for
an atmospheric research instrument: perhaps a "batch
queue manager", "resource manager" or "scientific
workflow system."

One recommendation has been:
- the Maui scheduler, which runs on the
- Torque resource manager, a fork of the
- Portable Batch Manager (PBM)

Do you have thoughts or personal experience with this
or any competing systems?

We have four Linux boxes--dual quad-core Xeon systems
running Scientific Linux.  The team is small: a few
developers and scientists.

We use cron and Perl scripts to launch the jobs today.
This works for us, but the CPU utilization is not as
high as I would like and the system is hard to understand
and manage. The system is big enough that is has grown
complex, but this is not what I would call high-performance
computing.

There ought to be a better way.  This is a well-studied
problem and I expect there are some standard solutions.

Desirable:
- cheap/free
- relatively simple to install and experiment with; we
  cannot afford to make this a large project
- scheduling based on priority and resource availability:
  - CPUs
  - memory
- ability to monitor and manage the queue
- distribute jobs to multiple compute servers
- command-line interface for ease of integration
- modular/light-weight  enough to be adapted to our existing
  structure

I'd be grateful for any comments or suggestions.

Thanks,
Vince Dean




More information about the LUG mailing list