Daily-Trend-Review

2023/09/27: Speed up Inference

hellcat 2023. 9. 27. 09:47

7 Ways To Speed Up Inference of Your Hosted LLMs

 

7 Ways to Speed Up Inference of Your Hosted LLMs

TLDR; techniques to speed up inference of LLMs to increase token generation speed and reduce memory consumption

betterprogramming.pub

 

Fixing Hallucinations in LLMs

'Daily-Trend-Review' 카테고리의 다른 글

2023/10/16: RAG  (0) 2023.10.16
2023/10/06: long context llms  (0) 2023.10.06
2023/09/24: 상품화를 위한 LLM 최적화  (0) 2023.09.24
2023/09/21: 언어모델링=압축  (0) 2023.09.21
2023/09/18: Textbooks Are All You Need 등  (0) 2023.09.18