The monster behind the LLM
I just spent some time exploring the site at Systemic Misalignment: Exposing Key Failures of Surface-Level AI Alignment Methods, and it’s a thought-provoking place. In the context of AI, “alignment is the process of encoding human values and goals into large language models to make them as helpful, safe, and reliable as possible.[1]” Researchers at […]