Skip to content

记录平时阅读的有关大模型的一些材料

Notifications You must be signed in to change notification settings

JustQJ/LLM_Materials

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

17 Commits
 
 

Repository files navigation

LLM Materials

Recording some materials about LLM.

LLM Training

Techniques

Datasets

LLM Evaluation

LLM Architecture

LLM Compression

Survey

  • [Arxiv 2024] A Survey of Resource-efficient LLM and Multimodal Foundation Models. This survey delves into the critical importance of such research, examining both algorithmic and systemic aspects. It offers a comprehensive analysis and valuable insights gleaned from existing literature, encompassing a broad array of topics from cutting-edge model architectures and training/serving algorithms to practical system designs and implementations.
  • [Arxiv 2024] LLM Inference Unveiled: Survey and Roofline Model Insights. not only summarizing the current state of research but also by introducing a framework based on roofline model for systematic analysis of LLM inference techniques. This framework identifies the bottlenecks when deploying LLMs on hardware devices and provides a clear understanding of practical problems, such as why LLMs are memory-bound, how much memory and computation they need, and how to choose the right hardware. Code
  • [Arxiv 2023] A Survey on Model Compression for Large Language Models.

Low-rank Decomposition

Pruning

Sparsity

Quantization

LLM System

Survey

  • [Arxiv 2024] A Survey of Resource-efficient LLM and Multimodal Foundation Models. This survey delves into the critical importance of such research, examining both algorithmic and systemic aspects. It offers a comprehensive analysis and valuable insights gleaned from existing literature, encompassing a broad array of topics from cutting-edge model architectures and training/serving algorithms to practical system designs and implementations.
  • [Arxiv 2024] LLM Inference Unveiled: Survey and Roofline Model Insights. not only summarizing the current state of research but also by introducing a framework based on roofline model for systematic analysis of LLM inference techniques. This framework identifies the bottlenecks when deploying LLMs on hardware devices and provides a clear understanding of practical problems, such as why LLMs are memory-bound, how much memory and computation they need, and how to choose the right hardware. Code

Training System

Inference System

Some Collections

Projects

Multimodel LLM

Survey

About

记录平时阅读的有关大模型的一些材料

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published