Data Engineering for Scaling Language Models to 128K Context
Are All Large Language Models Really in 1.58 Bits?
FrugalGPT: How to Use Large Language Models While Reducing Cost and Improving Performance
4-bit LLM Quantization with GPTQ
Large-scale Multi-Modal Pre-trained Models: A Comprehensive Survey
OpenELM:An Efficient Language Model Family with Open Training and Inference Framework
Will infinite context windows kill LLM fine-tuning and RAG?
Dynamic Memory Compression: Retrofitting LLMs for Accelerated Inference
'Daily-Trend-Review' 카테고리의 다른 글
24/05/29: MS build 2024 (0) | 2024.05.29 |
---|---|
24/05/12: LLM pricing (0) | 2024.05.12 |
24/04/16: Are All Large Language Models Really in 1.58 Bits? (0) | 2024.04.16 |
24/04/13: LLM cost vs. Performance (0) | 2024.04.13 |
24/03/31: Transformer math 101 (0) | 2024.03.31 |