Discussion about this post

User's avatar
Yatharth Arora's avatar

"For LLM developers, this raises deeper questions. How do we know what’s in the training data? How do we filter out subtle narrative manipulation that isn’t easily flagged? And what happens when a model starts to confidently explain a biased version of history?"

Maybe that's where human feedback comes in? We should not forget that at the end of the day, LLMs just generate text.

Expand full comment

No posts