A new study shows that humans can learn new things from artificial intelligence systems and pass them to other humans, in ways that could potentially influence wider human culture.
The study, published on Monday by a group of researchers at the Center for Human and Machines at the Max Planck Institute for Human Development, suggests that while humans can learn from algorithms how to better solve certain problems, human biases prevented performance improvements from lasting as long as expected. Humans tended to prefer solutions from other humans over those proposed by algorithms, because they were more intuitive, or were less costly upfront—even if they paid off more, later.
“Digital technology already influences the processes of social transmission among people by providing new and faster means of communication and imitation,” the researchers write in the study. “Going one step further, we argue that rather than a mere means of cultural transmission (such as books or the Internet), algorithmic agents and AI may also play an active role in shaping cultural evolution processes online where humans and algorithms routinely interact.”
The crux of this research rests on a relatively simple question: If social learning, or the ability of humans to learn from one another, forms the basis of how humans transmit culture or solve problems collectively, what would social learning look like between humans and algorithms? Considering scientists don’t always know and often can’t reproduce how their own algorithms work or improve, the idea that machine learning could influence human learning—and culture itself—throughout generations is a frightening one.
“There’s a concept called cumulative cultural evolution, where we say that each generation is always pulling up on the next generation, all throughout human history,” Levin Brinkmann, one of the researchers who worked on the study, told Motherboard. “Obviously, AI is pulling up on human history—they’re trained on human data. But we also found it interesting to think about the other way around: that maybe in the future our human culture would be built up on solutions which have been found originally by an algorithm.”
One early example cited in the research is Go, a Chinese strategy board game that saw an algorithm—AlphaGo—beat the human world champion Lee Sedol in 2016. AlphaGo made moves that were extremely unlikely to be made by human players and were learned via self-play instead of analyzing human gameplay data. The algorithm was made public in 2017 and such moves have become more common among human players, suggesting that a hybrid form of social learning between humans and algorithms was not only possible but durable.
We already know that algorithms can and do significantly affect humans. They’re not only used to control workers and citizens in physical workplaces, but also control workers on digital platforms and influence the behavior of individuals who use them. Even studies of algorithms have previewed the worrying ease with which these systems can be used to dabble in phrenology and physiognomy. A federal review of facial recognition algorithms in 2019 found that they were rife with racial biases. One 2020 Nature paper used machine learning to track historical changes in how “trustworthiness” has been depicted in portraits, but created diagrams indistinguishable from well-known phrenology booklets and offered universal conclusions from a dataset limited to European portraits of wealthy subjects.
“I don’t think our work can really say a lot about the formation of norms or how much AI can interfere with that,” Brinkmann said. “We’re focused on a different type of culture, what you could call the culture of innovation, right? A measurable value or peformance where you can clearly say, ‘Okay this paradigm—like with AlphaGo—is maybe more likely to lead to success or less likely.’”
See FULL STORY at Vice.com.
Source link
Author Editor