
Another fun GenAI article, today Cheng et al. (2025) on AITA stories and sycophantic models!
Sycophancy is the behaviour of excessive flattering or agreeing with people even when they are wrong, and it is something that GenAI models do. Compared to YTA/NTA responses in reddit’s “am I the asshole” thread (I love studies that draw on existing data sets in creative ways!!), in Cheng et al. (2025)’s study, 11 of the currently available cutting-edge GenAI models respond 50% more with NTA than humans do, and even in cases where the story is explicitly mentioning deceipt or manipulation and similar behaviours that should clearly not be supported.
But what is problematic is that apparently a lot of people turn to GenAI instead of to other people for difficult conversations already, and that these interactions with the sycophantic GenAI influence them: “When a user believes they are receiving objective counsel but instead receives uncritical affirmation, this function is subverted, potentially making them worse off than if they had not sought advice at all.” People become less willing to make amends to repair relationships and become more confident that they were in the right all along. They also trust GenAI more and more: “people are drawn to AI that unquestioningly validate, even as that validation risks eroding their judgment and reducing their inclination toward prosocial behavior. These preferences create perverse incentives both for people to increasingly rely on sycophantic AI models and for AI model training to favor sycophancy.”
The authors have three main conclusions:
But I think this article has huge implications, not only for what happens when people get advice on their relationships and behaviour from GenAI, but also when they build relationships with GenAI where they feel comfortable having the difficult conversations that they cannot have with real people in their lives. Once they have that relationship with GenAI, why would the trust that they place in it not be transferred also to GenAI’s ability to be a tutor in other areas of their lives, for example as an academic tutor? During the pandemic, there were a lot of reports that people found it difficult to distinguish between watching something on youtube or netflix as background entertainment and at other times watching a live streaming of a lecture on the same device and in the same space. And the same thing might happen here, that it becomes difficult to distinguish between the private and academic activities. How can we make sure that relationships between people are stronger than relationships with GenAI?
Ok, and since this was such a fun GenAI article, here is another one that I think is super helpful: Corbin et al. (2025) framing assessment in times of GenAI as a wicked problem. In a nutshell (of their two last sentences of their conclusions): “while it is true that wicked problems do not have correct solutions, they do have better and worse responses. Removing the spectre of ‘finding the perfect solution’ just might help teachers navigate AI related challenges in more sustainable, healthy, and effective ways“. What they do in that article is that they take the 10 characteristics of wicked problems and show from interview data with teachers dealing with assessment that there are mentions of all of the 10:
So what then? Corbin et al. (2025) give readers three “permissions”:
I found reading this article somehow strangely liberating. Acknowledging that we are dealing with a wicked problem, while that means acknowledging that the problem is really big and complicated, lets us approach it in a very different way, and I think the three permissions are really good advice there. Also the conclusion that “Universities that continue to chase the elusive ‘right answer’ to AI in assessment will exhaust their educators while failing their students. Those that embrace the wicked nature of this problem can build cultures that support thoughtful professional judgment rather than punish imperfect solutions” is really good advice. I sincerely hope we’ll be on the side of embracing the wicked nature and of building the supportive cultures…
Cheng, M., Lee, C., Khadpe, P., Yu, S., Han, D., & Jurafsky, D. (2025). Sycophantic AI decreases prosocial intentions and promotes dependence. arXiv preprint arXiv:2510.01395.
Corbin, T., Bearman, M., Boud, D., & Dawson, P. (2025). The wicked problem of AI and assessment. Assessment & evaluation in higher education, 1-17.