Daily-Trend-Review 107

2023/05/07: Single GPU로 LLM 추론하기, 효율적인 Transformers 등

1. High-throughput Generative Inference of Large Language Models with a Single GPU source: https://arxiv.org/pdf/2303.06865.pdf 2. Deploying Large NLP Models: Infrastructure Cost Optimization source: https://neptune.ai/blog/nlp-models-infrastructure-cost-optimization 3. What Are Transformer Models and How Do They Work? source: https://txt.cohere.com/what-are-transformer-models/ 4. Efficient Tran..

Daily-Trend-Review 2023.05.07

2023/04/22: AI 모델 데이터 문제, StableLM 등

1. AI 모델의 데이터를 다룰 때 발생하는 세 가지 문제 (Data! Data! Data!) source: https://www.upstage.ai/blog/business/why-is-data-labeling-important-for-ai 2. StableLM: Stability AI Language Models source: https://github.com/Stability-AI/StableLM?fbclid=IwAR3jtwsESYvHx5C5Q9Ru1VpU1lPG-akoZlxkTdtIhmvLkEgJrbKIV-XANZw 3. ChatGPT Beyond English: Towards a Comprehensive Evaluation of Large Language Models in Multilingual..

Daily-Trend-Review 2023.04.22

2023/04/17: 지식 베이스와 Q&A 플랫폼을 ChatGPT에 통합하기 등

1. Integrating ChatGPT with internal knowledge base and question-answer platform source: https://medium.com/singapore-gds/integrating-chatgpt-with-internal-knowledge-base-and-question-answer-platform-36a3283d6334 2. Transformer Taxonomy (the last lit reveiw) source: https://kipp.ly/blog/transformer-taxonomy 3. Generative Agents: Interactive Simulacra of Human Behavior source: https://arxiv.org/p..

Daily-Trend-Review 2023.04.17