An introduction to parallel programming (Record no. 6835)

MARC details
008 - FIXED-LENGTH DATA ELEMENTS--GENERAL INFORMATION
fixed length control field 111011s2011 ne a b 001 0 eng
010 ## - LIBRARY OF CONGRESS CONTROL NUMBER
LC control number 2010039584
020 ## - INTERNATIONAL STANDARD BOOK NUMBER
International Standard Book Number
035 ## - SYSTEM CONTROL NUMBER
System control number (Sirsi) u7910
040 ## - CATALOGING SOURCE
Original cataloging agency EG-CaNU
Transcribing agency EG-CaNU
Modifying agency EG-CaNU
042 ## - AUTHENTICATION CODE
Authentication code ncode
082 ## - DEWEY DECIMAL CLASSIFICATION NUMBER
Classification number 005.275
Edition number 22
100 1# - MAIN ENTRY--PERSONAL NAME
Personal name Pacheco, Peter S.
9 (RLIN) 14283
245 13 - TITLE STATEMENT
Title An introduction to parallel programming
260 ## - PUBLICATION, DISTRIBUTION, ETC.
Place of publication, distribution, etc. Amsterdam :
-- Boston :
Name of publisher, distributor, etc. Morgan Kaufmann,
Date of publication, distribution, etc. c2011.
300 ## - PHYSICAL DESCRIPTION
Extent xix, 370 p. :
Other physical details ill. ;
Dimensions 25 cm.
504 ## - BIBLIOGRAPHY, ETC. NOTE
Bibliography, etc. note Includes bibliographical references (p. 357-359) and index.
505 ## - FORMATTED CONTENTS NOTE
Formatted contents note Machine generated contents note: 1 Why Parallel Computing1.1 Why We Need Ever-Increasing Performance 1.2 Why We’re Building Parallel Systems 1.3 Why We Need to Write Parallel Programs 1.4 How Do We Write Parallel Programs? 1.5 What We’ll Be Doing 1.6 Concurrent, Parallel, Distributed 1.7 The Rest of the Book 1.8 A Word of Warning 1.9 Typographical Conventions 1.10 Summary 1.11 Exercises 2 Parallel Hardware and Parallel Software2.1 Some Background 2.2 Modifications to the von Neumann Model 2.3 Parallel Hardware 2.4 Parallel Software 2.5 Input and Output 2.6 Performance 2.7 Parallel Program Design 2.8 Writing and Running Parallel Programs 2.9 Assumptions 2.10 Summary 2.11 Exercises 3 Distributed Memory Programming with MPI3.1 Getting Started 3.2 The Trapezoidal Rule in MPI 3.3 Dealing with I/O 3.4 Collective Communication 3.5 MPI Derived Datatypes 3.7 A Parallel Sorting Algorithm 3.8 Summary3.9 Exercises 3.10 Programming Assignments 4 Shared Memory Programming with Pthreads4.1 Processes, Threads and Pthreads 4.2 Hello, World4.3 Matrix-Vector Multiplication 4.4 Critical Sections 4.5 Busy-Waiting 4.6 Mutexes 4.7 Producer-Consumer Synchronization and Semaphores 4.8 Barriers and Condition Variables 4.9 Read-Write Locks 4.10 Caches, Cache-Coherence, and False Sharing 4.11 Thread-Safety 4.12 Summary 4.13 Exercises4.14 Programming Assignments 5 Shared Memory Programming with OpenMP5.1 Getting Started 5.2 The Trapezoidal Rule 5.3 Scope of Variables 5.4 The Reduction Clause 5.5 The Parallel For Directive 5.6 More About Loops in OpenMP: Sorting 5.7 Scheduling Loops 5.8 Producers and Consumers 5.9 Caches, Cache-Coherence, and False Sharing 5.10 Thread-Safety 5.11 Summary 5.12 Exercises 5.13 Programming Assignments 6 Parallel Program Development6.1 Two N-Body Solvers 6.2 Tree Search 6.3 A Word of Caution 6.4 Which API? 6.5 Summary 6.6 Exercises 6.7 Programming Assignments 7 Where to Go from Here .
596 ## -
-- 1
650 ## - SUBJECT ADDED ENTRY--TOPICAL TERM
Topical term or geographic name entry element Pacheco, Peter S.
9 (RLIN) 14284
Holdings
Withdrawn status Lost status Source of classification or shelving scheme Damaged status Not for loan Home library Current library Shelving location Date acquired Source of acquisition Total Checkouts Total Renewals Full call number Barcode Date last seen Date last checked out Copy number Price effective from Koha item type
    Dewey Decimal Classification     Main library Main library General Stacks 10/11/2011 PURCHASE 1 2 005.275 / PA. I 2011 011513 04/22/2024 03/13/2024 1 11/24/2019 Books