Compressing large neural networks (Record no. 767485)

MARC details
000 -LEADER
fixed length control field 03931ntm a22003377a 4500
003 - CONTROL NUMBER IDENTIFIER
control field AT-ISTA
005 - DATE AND TIME OF LATEST TRANSACTION
control field 20250704134144.0
008 - FIXED-LENGTH DATA ELEMENTS--GENERAL INFORMATION
fixed length control field 250704s2024 au ||||| m||| 00| 0 eng d
040 ## - CATALOGING SOURCE
Transcribing agency ISTA
100 ## - MAIN ENTRY--PERSONAL NAME
Personal name Frantar, Elias
9 (RLIN) 1083072
245 ## - TITLE STATEMENT
Title Compressing large neural networks
Remainder of title : Algorithms, systems and scaling laws
260 ## - PUBLICATION, DISTRIBUTION, ETC. (IMPRINT)
Name of publisher, distributor, etc. Institute of Science and Technology Austria
Date of publication, distribution, etc. 2024
500 ## - GENERAL NOTE
General note Thesis
505 ## - FORMATTED CONTENTS NOTE
Formatted contents note Abstract
505 ## - FORMATTED CONTENTS NOTE
Formatted contents note Acknowledgments
505 ## - FORMATTED CONTENTS NOTE
Formatted contents note About the Author
505 ## - FORMATTED CONTENTS NOTE
Formatted contents note List of Collaborators and Publications
505 ## - FORMATTED CONTENTS NOTE
Formatted contents note Table of Contents
505 ## - FORMATTED CONTENTS NOTE
Formatted contents note 1 Introduction
505 ## - FORMATTED CONTENTS NOTE
Formatted contents note 2 Background
505 ## - FORMATTED CONTENTS NOTE
Formatted contents note 3 Algorithms
505 ## - FORMATTED CONTENTS NOTE
Formatted contents note 4 Systems
505 ## - FORMATTED CONTENTS NOTE
Formatted contents note 5 Scaling Laws
505 ## - FORMATTED CONTENTS NOTE
Formatted contents note 6 Conclusion
505 ## - FORMATTED CONTENTS NOTE
Formatted contents note Bibliography
505 ## - FORMATTED CONTENTS NOTE
Formatted contents note A Algorithms
505 ## - FORMATTED CONTENTS NOTE
Formatted contents note B Scaling Laws
520 ## - SUMMARY, ETC.
Summary, etc. Large language models (LLMs) have made tremendous progress in the past few years, from being able to generate coherent text to matching or surpassing humans in a wide variety of creative, knowledge or reasoning tasks. Much of this can be attributed to massively increased scale, both in the size of the model as well as the amount of training data, from 100s of millions to 100s of billions, or even trillions. This trend is expected to continue, which, although exciting, also raises major practical concerns. Already today's 100+ billion parameter LLMs require top-of-the-line hardware just to run. Hence, it is clear that sustaining these developments will require significant efficiency advances. Historically, one of the most practical ways of improving model efficiency has been compression, especially in the form of sparsity or quantization. While this has been studied extensively in the past, existing accurate methods are all designed for models around 100 million parameters; scaling them up to ones literally 1000x larger is highly challenging. In this thesis, we introduce a new unified sparsification and quantization approach OBC, which through additional algorithmic enhancements leads to GPTQ and SparseGPT, the first techniques fast and accurate enough to compress 100+ billion parameter models to 4- or even 3-bit precision and 50% weight-sparsity, respectively. Additionally, we show how weight-only quantizion does not just bring space savings but also up to 4.5x faster generation speed, via custom GPU kernels. In fact, we show for the first time that it is possible to develop an FP16 times INT4 mixed-precision matrix multiplication kernel, called Marlin, which comes close to simultaneously maximizing both memory and compute utilization, making weight-only quantization highly practical even for multi-user serving. Further, we demonstrate that GPTQ can be scaled to widely overparametrized trillion-parameter models, where extreme sub-1-bit compression rates can be achieved without any inference slow-down, by co-designing a bespoke entropy coding scheme together with an efficient kernel. Finally, we also study compression from the perspective of someone with access to massive amounts of compute resources for training large models completely from scratch. Here the key questions evolve around the joint scaling behavior between compression, model size, and amount of training data used. Based on extensive experimental results for both vision and text models, we introduce the first scaling law which accurately captures the relationship between weight-sparsity, number of non-zero weights and data. This further allows us to characterize the optimal sparsity, which we find to increase the longer a fixed cost model is being trained. Overall, this thesis presents contributions to three different angles of large model efficiency: affordable but accurate algorithms, highly efficient systems implementations, and fundamental scaling laws for compressed training.
856 ## - ELECTRONIC LOCATION AND ACCESS
Uniform Resource Identifier <a href="https://doi.org/10.15479/at:ista:17485">https://doi.org/10.15479/at:ista:17485</a>
942 ## - ADDED ENTRY ELEMENTS (KOHA)
Source of classification or shelving scheme Dewey Decimal Classification
Holdings
Withdrawn status Lost status Source of classification or shelving scheme Damaged status Not for loan Home library Current library Date acquired Total Checkouts Full call number Barcode Date last seen Price effective from Koha item type
  Not Lost Dewey Decimal Classification     Library Library 04/07/2025   Quiet Room AT-ISTA#003250 16/09/2025 04/07/2025 Book

Powered by Koha