AI Alignment and Social Choice: Fundamental Limitations and Policy Implications
This paper by Abhilash Mishra investigates the challenges of aligning AI systems with human intentions and values using Reinforcement Learning with Human Feedback (RLHF). It discusses the issue of determining…
Continue reading