1 code implementation • 19 May 2024 • William M. Hayes, Nicolas Yax, Stefano Palminteri
Given their potential use as (autonomous) decision-making agents, it is important to understand how these models perform such reinforcement learning (RL) tasks and the extent to which they are susceptible to biases.
no code implementations • 6 Apr 2024 • Nicolas Yax, Pierre-Yves Oudeyer, Stefano Palminteri
This paper introduces PhyloLM, a method adapting phylogenetic algorithms to Large Language Models (LLMs) to explore whether and how they relate to each other and to predict their performance characteristics.
no code implementations • 25 Jan 2024 • William M. Hayes, Nicolas Yax, Stefano Palminteri
Studies of reinforcement learning in humans and animals have demonstrated a preference for options that yielded relatively better outcomes in the past, even when those options are associated with lower absolute reward.
no code implementations • 21 Sep 2023 • Nicolas Yax, Hernan Anlló, Stefano Palminteri
In the present study, we investigate and compare reasoning in large language models (LLM) and humans using a selection of cognitive psychology tools traditionally dedicated to the study of (bounded) rationality.