论文标题:Direct Preference Optimization: Your Language Model is Secretly a Reward Model论文链接...
论文标题:Direct Preference Optimization: Your Language Model is Secretly a Reward Model论文链接...
论文标题:Propagation Tree Is Not Deep: Adaptive Graph Contrastive Learning Approach for Rum...
一、概述 大语言模型(LLMs)在预训练的过程中通常会捕捉数据的特征,而这些训练数据通常既包含高质量的也包含低质量的,因此模型有时会产生不被期望的行为,如编造事实,生成有偏见...
论文标题:LoRA: Low-Rank Adaptation of Large Language Models论文链接:https://arxiv.org/abs/2106....
论文标题:Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallel...
论文标题:Tree of Thoughts: Deliberate Problem Solving with Large Language Models论文链接:https:...
论文标题:LIMA: Less Is More for Alignment论文链接:https://arxiv.org/abs/2305.11206[https://arxi...
论文标题:Self-Consistency Improves Chain of Thought Reasoning in Language Models论文链接:https:...
论文标题:GPipe: Easy Scaling with Micro-Batch Pipeline Parallelism论文链接:https://arxiv.org/ab...
论文标题:Training language models to follow instructions with human feedback论文链接:https://ar...
论文标题:LLaMA: Open and Efficient Foundation Language Models论文链接:https://arxiv.org/abs/230...
论文标题:Pure Transformers are Powerful Graph Learners论文链接:https://arxiv.org/abs/2207.02505...
论文标题:Chain-of-Thought Prompting Elicits Reasoning in Large Language Models论文链接:https://...
论文标题:Language Models are Few-Shot Learners论文链接:https://arxiv.org/abs/2005.14165[https:/...
论文标题:Language Models are Unsupervised Multitask Learners论文链接:https://d4mucfpksywv.cloud...
论文标题:Improving Language Understanding by Generative Pre-Training论文链接:https://www.cs.ubc...