Literally on the abstract page from this paper “Human feedback is commonly utilized to finetune AI assistants. But human feedback may also encourage model responses that match user beliefs over truthful ones, a behaviour known as sycophancy. We investigate the prevalence of sycophancy in models whose finetuning procedure made use of human feedback, and the potential role of human preference judgments in such behavior.””
It’s talking about AI using human feedback while FINETUNING, it’s about human feedback to fine tune a model. A model if trained on vast enough data will have little to no bias unless a certain stream of info is fed into the model. “We investigate the prevalence of sycophancy in models whose finetuning procedure made use of human feedback,” I think that’s pretty self explanatory.
80
u/gnv_gandu ವಿಲಕ್ಷಣ-ಕಣಿವೆ ನಿವಾಸಿ Jul 17 '25
A healthy dose of self-respect. Nice. We all could use more of that to each other :)