Job Scheduling Strategies for Parallel Processing [electronic resource] : IPPS '97 Processing Workshop Geneva, Switzerland, April 5, 1997 Proceedings / edited by Dror G. Feitelson, Larry Rudolph.Material type: TextLanguage: English Series: Lecture Notes in Computer Science: 1291Publisher: Berlin, Heidelberg : Springer Berlin Heidelberg, 1997Description: VII, 305 p. online resourceContent type: text Media type: computer Carrier type: online resourceISBN: 9783540695998Subject(s): Computer science | Logic design | Operating systems (Computers) | Computer software | Computer Science | Operating Systems | Programming Techniques | Algorithm Analysis and Problem Complexity | Processor Architectures | Logic DesignAdditional physical formats: Printed edition:: No titleDDC classification: 005.43 LOC classification: QA76.76.O63Online resources: Click here to access online
Theory and practice in parallel job scheduling -- Using queue time predictions for processor allocation -- A historical application profiler for use by parallel schedulers -- Memory usage in the LANL CM-5 workload -- Modeling of workload in MPPs -- PScheD Political scheduling on the CRAY T3E -- An experimental evaluation of processor pool-based scheduling for shared-memory NUMA multiprocessors -- Implementing multiprocessor scheduling disciplines -- Objective-oriented algorithm for job scheduling in parallel heterogeneous systems -- Implications of I/O for gang scheduled workloads -- Improved utilization and responsiveness with gang scheduling -- Global state detection using network preemption -- Performance evaluation of gang scheduling for parallel and distributed multiprogramming.
This book constitutes the strictly refereed post-workshop proceedings of the 1997 IPPS Workshop on Job Scheduling Strategies for Parallel Processing held in Geneva, Switzerland, in April 1997, as a satelite meeting of the IEEE/CS International Parallel Processing Symposium. The 12 revised full papers presented were carefully reviewed and revised for inclusion in the book. Also included is a detailed introduction surveying the state of the art in the area. Among the topics covered are processor allocation, parallel scheduling, massively parallel processing, shared-memory architectures, gang scheduling, etc.