A University of Washington study shows that AI can learn culture-specific values, like altruism, by watching people play a cooperative video game. The work points toward future AI systems that better reflect the communities they serve.
Artificial intelligence may one day learn to “fit in” with different cultures the way children do — by watching and absorbing how people treat each other.
In a new study published in PLOS One, University of Washington researchers showed that AI systems can pick up culture-specific values such as altruism simply by observing people play a cooperative video game. The AI then carried those learned values into a completely different situation.
The work addresses a growing concern in AI: most large systems are trained on data scraped from across the internet, which tends to reflect the values of some groups more than others. That can lead to tools and chatbots that feel out of step with many communities.
“We shouldn’t hard code a universal set of values into AI systems, because many cultures have their own values,” senior author Rajesh Rao, a UW professor in the Paul G. Allen School of Computer Science & Engineering and co-director of the Center for Neurotechnology, said n a news release. “So we wanted to find out if an AI system can learn values the way children do, by observing people in their culture and absorbing their values.”
To explore that idea, the UW team turned to a surprising training ground: a modified version of the popular cooperative cooking game Overcooked.
In the experiment, 190 adults who identified as white and 110 who identified as Latino played a special version of the game. Players had to cook and deliver as much onion soup as possible. On the screen, they could see another kitchen where a second player had to walk farther to complete the same tasks, putting that player at a clear disadvantage.
What the human participants did not know was that the second player was actually a computer-controlled bot. The bot was programmed to ask for help, and the humans had a choice: they could give up some of their onions to help the disadvantaged partner, sacrificing their own score, or keep the onions and focus on maximizing their own success.
Each cultural group — white and Latino — was paired with its own AI “agent,” a software system designed to learn from behavior. Instead of being told what to do or given explicit rewards, these agents used a technique called inverse reinforcement learning.
In standard reinforcement learning, an AI is given a goal and rewarded when it gets closer to that goal, like a robot that learns to play tennis by scoring points. In inverse reinforcement learning, the AI watches humans or other agents and tries to infer what goals and values must be driving their actions.
This approach, the researchers argue, is closer to how people, especially children, actually learn.
Co-author Andrew Meltzoff, a UW psychology professor and co-director of the Institute for Learning & Brain Sciences, drew a direct parallel to parenting.
“Parents don’t simply train children to do a specific task over and over. Rather, they model or act in the general way they want their children to act. For example, they model sharing and caring towards others,” he said in the news release. “Kids learn almost by osmosis how people act in a community or culture. The human values they learn are more ‘caught’ than ‘taught.’”
By feeding the Overcooked gameplay data into the AI agents, the team let each system “watch” how people from its assigned cultural group behaved when faced with a choice between self-interest and helping someone at a disadvantage.
On average, participants in the Latino group chose to help the disadvantaged partner more often than participants in the white group. The AI agents picked up on this pattern. When they later played the game themselves, the agent trained on Latino players’ data gave away more onions than the agent trained on white players’ data, mirroring the more altruistic behavior it had observed.
To see whether the AI had truly learned a general value — not just a game-specific trick — the researchers ran a second test. This time, the agents faced a different kind of moral decision: whether to donate some of their money to someone in need.
Again, the agent trained on data from Latino participants in Overcooked behaved more altruistically, choosing to give away more of its resources. That suggested the system had internalized a broader preference for helping others, not just a strategy tied to onions and soup.
The results are an early but promising sign that AI can be tuned to better reflect the values of specific communities by learning from their everyday behavior, according to Rao.
“We think that our proof-of-concept demonstrations would scale as you increase the amount and variety of culture-specific data you feed to the AI agent. Using such an approach, an AI company could potentially fine-tune their model to learn a specific culture’s values before deploying their AI system in that culture,” Rao added.
The study builds on earlier UW research showing that 19-month-old children raised in Latino and Asian households were more prone to altruism than those from other cultural backgrounds. By echoing that pattern in AI, the new work suggests that machines can, in a limited way, mirror the value-learning process seen in human development.
At the same time, the researchers emphasize that this is just a first step. The experiments involved only two cultural groups, a simplified game environment and a relatively narrow focus on altruism. Real-world settings are far more complex, with many overlapping cultures, conflicting values and high-stakes decisions.
Future research will need to test how inverse reinforcement learning performs when AI systems are exposed to richer, messier data from daily life, and when they must navigate trade-offs between competing values such as fairness, loyalty, privacy and efficiency.
“Creating culturally attuned AI is an essential question for society,” added Meltzoff. “How do we create systems that can take the perspectives of others into account and become civic minded?”
The UW team’s answer, at least for now, is to let AI watch us more closely — not just what we say we value, but how we actually behave when helping others comes at a cost. If that approach scales, tomorrow’s AI might be less about imposing a single global rulebook and more about learning, community by community, what it means to be a good neighbor.
Source: University of Washington

