📑

AI Paper Research

AI 논문 조사 및 정리

Foundations
대규모 언어모델Large Language Models
Mixtral of ExpertsGemma: Open Models Based on Gemini Resea...
LLaMA: Open and Efficient Foundation Lan...GPT-4 Technical ReportRetrieval-Augmented Generation for Knowl...Llama 2: Open Foundation and Fine-Tuned ...
Training Compute-Optimal Large Language ...Chain-of-Thought Prompting Elicits Reaso...PaLM: Scaling Language Modeling with Pat...
Evaluating Large Language Models Trained...
홈/대규모 언어모델/2022

대규모 언어모델 — 2022

3편의 논문

NeurIPS 20223,000+

Training Compute-Optimal Large Language Models

계산 최적 대규모 언어 모델 학습

Jordan Hoffmann, Sebastian Borgeaud, Arthur Mensch et al. (2022)

NeurIPS 2022Outstanding5,000+

Chain-of-Thought Prompting Elicits Reasoning in Large Language Models

연쇄 사고 프롬프팅은 대규모 언어 모델에서 추론을 이끌어낸다

Jason Wei, Xuezhi Wang, Dale Schuurmans et al. (2022)

arXiv7,000+

PaLM: Scaling Language Modeling with Pathways

PaLM: Pathways를 활용한 언어 모델링 스케일링

Aakanksha Chowdhery, Sharan Narang, Jacob Devlin et al. (2022)

← 대규모 언어모델 전체