Job scheduler

From Wikipedia, the free encyclopedia

A job scheduler is an enterprise software application that is in charge of unattended background executions, commonly known for historical reasons as batch processing.

Synonyms are batch system, Distributed Resource Management System (DRMS), and Distributed Resource Manager (DRM). Today's job schedulers typically provide a graphical user interface and a single point of control for definition and monitoring of background executions in a distributed network of computers. Increasingly job schedulers are required to orchestrate the integration of real-time business activities with traditional background IT processing, across different operating system platforms and business application environments.

Contents

[edit] Overview

Basic features expected of job scheduler software are:

  • Interfaces to define workflows and/or job dependencies
  • Automatic submission of executions
  • Interfaces to monitor the executions
  • Priorities and/or queues to control the execution order of unrelated jobs

If software from a completely different area includes all or some of those features, this software is considered to have job scheduling capabilities.

Most operating system platforms such as Unix and Windows provide basic job scheduling capabilities, for example Cron. Many programs such as DBMS, backup, ERPs, and BPM also include relevant job scheduling capabilities. Operating System (OS) or point program supplied job scheduling will not usually provide the ability to schedule beyond a single OS instance or outside the remit of the specific program. Organizations needing to automate highly complex related and un-related IT workload will also be expecting more advanced features from a job scheduler, such as:

  • Real-time scheduling based on external, un-predictable events
  • Automatic restart and recovery in event of failures
  • Alerting and notification to operations personnel
  • Generation of incident reports
  • Audit trails for regulatory compliance purposes

These advanced capabilities can be written by in-house developers but are more often provided by solutions from suppliers that specialize in systems management software.

[edit] Main concepts

There are many concepts that are central to almost every job scheduler implementation and that are widely recognized with minimal variations:

  • Jobs
  • Dependencies
  • Job Streams
  • Users

Beyond the basic, single OS instance scheduling tools there are two major architectures that exist for Job Scheduling software.

  • Master/Agent architecture — the historic architecture for Job scheduling software. The Job Scheduling software is installed on a single machine (Master) while on production machines only a very small component (Agent) is installed that awaits commands from the Master, executes them, and returns the exit code back to the Master.
  • Cooperative architecture — a decentralized model where each machine is capable of helping with scheduling and can offload locally scheduled jobs to other cooperating machines. This enables dynamic workload balancing to maximize hardware resource utilization and high availability to ensure service delivery.

[edit] History

Job Scheduling has a long history. Job Schedulers are one of the major components of the IT infrastructure since the early mainframe systems. At first, stacks of punch cards were processed one after the other, hence the term “batch processing.”

From a historical point of view, we can distinguish two main eras about Job Schedulers:

  1. The mainframe era
    • Job Control Language (JCL) on IBM mainframes. Initially based on JCL functionality to handle dependencies this era is typified by the development of sophisticated scheduling solutions forming part of the systems management and automation toolset on the mainframe.
  2. The open systems era
    • Modern schedulers on a variety of architectures and operating systems. With standard scheduling tools limited to such as Cron, the need for mainframe standard job schedulers has grown with the increased adoption of distributed computing environments.

In terms of the type of scheduling there are also distinct eras:

  1. Batch processing - the traditional date and time based execution of background tasks based on a defined period during which resources were available for batch processing (the batch window). In effect the original mainframe approach transposed onto the open systems environment.
  2. Event-driven process automation - where background processes cannot be simply run at a defined time, either because the nature of the business demands that workload is based on the occurrence of external events (such as the arrival of an order from a customer or a stock update from a store branch) or because there is no / insufficient batch window.
  3. Service Oriented job scheduling - recent developments in Service Oriented Architecture (SOA) have seen a move towards deploying job scheduling as a reusable IT infrastructure service that can play a role in the integration of existing business application workload with new Web Services based real-time applications.

[edit] Scheduling

Various schemes are used to decide which particular job to run. Parameters that might be considered include:

  • Job priority
  • Compute resource availability
  • License key if job is using licensed software
  • Execution time allocated to user
  • Number of simultaneous jobs allowed for a user
  • Estimated execution time
  • Elapsed execution time
  • Availability of peripheral devices
  • Occurrence of prescribed events

[edit] Implementations

Other distributed command execution software: Tentakel (python), ClusterSSH, dsh, fanout, ghost macro language, mussh shell script, p-run utility, pconsole shell tool, Taktuk scheduller, smux shell multiplexer, rshall, rgang and distribulator among others.

[edit] References

[edit] See also

Languages