LLM Agents - Intuitively and Exhaustively Explained
DPO, Open-Source’s New Weapon in the AI War
Hardware-aware Algorithms for Sequence Modeling - Tri Dao | Stanford MLSys #87
Democratizing LLMs: 4-bit Quantization for Optimal LLM Inference
A Detailed Explanation of Mixtral 8x7B Model
'Daily-Trend-Review' 카테고리의 다른 글
2024/01/26: Leading with open Models, frameworks, and systems (0) | 2024.01.26 |
---|---|
2024/01/20: 스터디 내용 정리 (0) | 2024.01.20 |
2024/01/05: Decoding Strategies in Large Language Models (1) | 2024.01.05 |
2024/01/02: Transformer inference tricks (0) | 2024.01.02 |
2023/12/25: Towards 100x Speedup: Full Stack Transformer Inference Optimization (0) | 2023.12.25 |