MARC details
000 -LEADER |
fixed length control field |
03581cam a22003375i 4500 |
003 - CONTROL NUMBER IDENTIFIER |
control field |
EG-CaNGU |
005 - DATE AND TIME OF LATEST TRANSACTION |
control field |
20240131132159.0 |
008 - FIXED-LENGTH DATA ELEMENTS--GENERAL INFORMATION |
fixed length control field |
231214t2023 maua frb 001 0 eng d |
020 ## - INTERNATIONAL STANDARD BOOK NUMBER |
International Standard Book Number |
9780323912310 |
040 ## - CATALOGING SOURCE |
Original cataloging agency |
DLC |
Language of cataloging |
eng |
Description conventions |
rda |
Transcribing agency |
DLC |
Modifying agency |
DLC |
-- |
EG-CaNGU |
082 04 - DEWEY DECIMAL CLASSIFICATION NUMBER |
Classification number |
004.35 |
Item number |
HWP |
Edition number |
23 |
100 1# - MAIN ENTRY--PERSONAL NAME |
Personal name |
Hwu, Wen-mei, |
Relator term |
author. |
9 (RLIN) |
5975 |
245 10 - TITLE STATEMENT |
Title |
Programming massively parallel processors : |
Remainder of title |
a hands-on approach / |
250 ## - EDITION STATEMENT |
Edition statement |
Fourth edition. |
264 #1 - PRODUCTION, PUBLICATION, DISTRIBUTION, MANUFACTURE, AND COPYRIGHT NOTICE |
Place of production, publication, distribution, manufacture |
Cambridge, Massachusetts : |
Name of producer, publisher, distributor, manufacturer |
Morgan Kauffmann, Elsevier Inc., |
Date of production, publication, distribution, manufacture, or copyright notice |
[2023] |
264 #4 - PRODUCTION, PUBLICATION, DISTRIBUTION, MANUFACTURE, AND COPYRIGHT NOTICE |
Date of production, publication, distribution, manufacture, or copyright notice |
©2023 |
300 ## - PHYSICAL DESCRIPTION |
Extent |
xxviii, 551 pages : |
Other physical details |
illustrations ; |
Dimensions |
24 cm |
336 ## - CONTENT TYPE |
Content type term |
text |
Content type code |
txt |
Source |
rdacontent |
337 ## - MEDIA TYPE |
Media type term |
unmediated |
Media type code |
n |
Source |
rdamedia |
338 ## - CARRIER TYPE |
Carrier type term |
volume |
Carrier type code |
nc |
Source |
rdacarrier |
504 ## - BIBLIOGRAPHY, ETC. NOTE |
Bibliography, etc. note |
Includes bibliographical references and index. |
505 00 - FORMATTED CONTENTS NOTE |
Formatted contents note |
Chapter 1. Introduction -- Part I: Fundamental Concepts -- Chapter 2. Heterogeneous data parallel computing -- Chapter 3. Multidimensional grids and data -- Chapter 4. Compute architecture and scheduling -- Chapter 5. Memory architecture and data locality -- Chapter 6. Performance considerations -- Part II: Parallel Patterns -- Chapter 7. Convolution: An introduction to constant memory and caching -- Chapter 8. Stencil -- Chapter 9. Parallel histogram: An introduction to atomic operations and privatization -- Chapter 10. Reduction: And minimizing divergence -- Chapter 11. Prefix sum (scan): An introduction to work efficiency in parallel algorithms -- Chapter 12. Merge: An introduction to dynamic input data identification -- Part III: Advanced Patterns and Applications -- Chapter 13. Sorting -- Chapter 14. Sparse matrix computation -- Chapter 15. Graph traversal -- Chapter 16. Deep learning -- Chapter 17. Iterative magnetic resonance imaging reconstruction -- Chapter 18. Electrostatic potential map -- Chapter 19. Parallel programming and computational thinking -- Chapter 20. Programming a heterogeneous computing cluster: An introduction to CUDA streams -- Chapter 21. CUDA dynamic parallelism -- Chapter 22. Advanced practices and future evolution -- Chapter 23. Conclusion and outlook. |
520 ## - SUMMARY, ETC. |
Summary, etc. |
Programming Massively Parallel Processors: A Hands-on Approach shows both student and professional alike the basic concepts of parallel programming and GPU architecture. Various techniques for constructing parallel programs are explored in detail. Case studies demonstrate the development process, which begins with computational thinking and ends with effective and efficient parallel programs. Topics of performance, floating-point format, parallel patterns, and dynamic parallelism are covered in depth. For this new edition, the authors are updating their coverage of CUDA, including the concept of unified memory, and expanding content in areas such as threads, while still retaining its concise, intuitive, practical approach based on years of road-testing in the authors' own parallel computing courses.-- |
Assigning source |
Source other than the Library of Congess. |
650 #7 - SUBJECT ADDED ENTRY--TOPICAL TERM |
Topical term or geographic name entry element |
Parallel programming (Computer science). |
Source of heading or term |
NGU-sh |
9 (RLIN) |
5976 |
650 #7 - SUBJECT ADDED ENTRY--TOPICAL TERM |
Topical term or geographic name entry element |
Parallel processing (Electronic computers). |
Source of heading or term |
NGU-sh |
9 (RLIN) |
5977 |
650 #7 - SUBJECT ADDED ENTRY--TOPICAL TERM |
Topical term or geographic name entry element |
Multiprocessors. |
Source of heading or term |
NGU-sh |
9 (RLIN) |
5978 |
650 #7 - SUBJECT ADDED ENTRY--TOPICAL TERM |
Topical term or geographic name entry element |
Computer architecture. |
Source of heading or term |
NGU-sh |
9 (RLIN) |
4432 |
700 1# - ADDED ENTRY--PERSONAL NAME |
Personal name |
Kirk, David, |
Dates associated with a name |
1960-, |
Relator term |
author. |
9 (RLIN) |
5979 |
700 1# - ADDED ENTRY--PERSONAL NAME |
Personal name |
El Hajj, Izzat, |
Relator term |
author. |
9 (RLIN) |
5980 |