-
Large Language Model Alignment: A Survey
Paper • 2309.15025 • Published • 2 -
Aligning Large Language Models with Human: A Survey
Paper • 2307.12966 • Published • 1 -
Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Paper • 2305.18290 • Published • 50 -
SteerLM: Attribute Conditioned SFT as an (User-Steerable) Alternative to RLHF
Paper • 2310.05344 • Published • 1
David Vaughn
davidsvaughn
AI & ML interests
ML,NLP
Recent Activity
updated
a model
2 days ago
davidsvaughn/llama-siam-3
liked
a model
about 1 month ago
ministral/Ministral-3b-instruct
Organizations
None yet
Collections
1
models
2
datasets
None public yet