Bernstein Network Computational Neuroscience
  • Home
  • Network
    • The Bernstein Network
    • Bernstein Centers
      • Berlin
      • Freiburg
      • Göttingen
      • Munich
      • Tübingen
      • Heidelberg-Mannheim
    • Bernstein Nodes
      • Bernstein Node Bochum
      • Bernstein Node Bonn-Cologne
      • Bernstein Node Chemnitz
      • Bernstein Node Hamburg
      • Bernstein Node Rhine-Main Region
      • Bernstein Node Taiwan
    • Research Infrastructure
      • High Performance Simulation and Data Analysis
      • Research Data Management
      • Science Communication
      • Scientific Coordination
    • Awards and Initiatives
      • Valentin Braitenberg Award
      • Brains for Brains Young Researcher Award
      • Bernstein SmartSteps
    • Committees
    • Mission Statement
    • Statutes
    • Membership
    • History
    • Donation
    • Contact
  • Newsroom
    • Newsroom
    • News
    • Events
    • Calls
    • Media Coverage
    • Network Publications
    • Bernstein Bulletin
    • Press
  • Teaching and Research
    • Teaching and Research
    • Meet the Scientist
    • Find a Scientist
    • Degree Programs
      • Master Programs
      • PhD Programs
    • Study and Training
      • Bernstein Student Workshop Series
      • Online Learning
      • Advanced Courses
      • Internships and Master theses
      • Podcasts
  • Career
    • Career
    • Job Pool
    • Calls
    • Internships and Master theses
  • Bernstein Conference
    • Bernstein Conference
    • Program
      • Schedule
      • Satellite Workshops
      • Conference Dinner
    • Early Career Scientists
      • PhD Symposium
      • Postdoc Meeting
      • Travel Grants
      • Buddy Program
    • General Information
      • Important Dates & FAQ
      • Plan Your Visit
      • Press
      • Code of Conduct
      • PR Media Policy
      • Data Policy
    • Past and future Bernstein Conferences
  • DE
  • EN
  • Click to open the search input field Click to open the search input field Search
  • Menu Menu
You are here: Home1 / Newsroom2 / News3 / When AI “thinks” like us
Berlin, Germany – November 13, 2025

When AI “thinks” like us

Even though so-called Vision Foundation models, computer models for automated image recognition, have made enormous progress in recent years, they still differ significantly from human visual understanding. For example, they generally do not capture multi-level semantic hierarchies and have difficulty with relationships between semantically related but visually dissimilar objects. In a joint project with Google DeepMind, scientists from TU Berlin, Max-Planck-Institute for Human Cognitive and Brain Sciences, and MPI for Human Development have generated a new approach called ‘AligNet,’ which integrates human semantic structures into neural image processing models for the first time, thereby bringing the visual understanding of computer models closer to that of humans. The results have now been published in the renowned journal Nature under the title ‘Aligning Machine and Human Visual Representations across Abstraction Levels’.

The eyes of AI have become increasingly sophisticated in recent years, but they still cannot match the visual understanding of humans.

Bernstein member involved: Klaus-Robert Müller

The scientists are investigating how visual representations in modern deep neural networks are structured compared to human perception and conceptual knowledge, and how these can be better aligned. Although artificial intelligence (AI) today achieves impressive performance in image processing, machines often generalize less robustly than humans, for instance, when faced with new types of images or unfamiliar relations.

“The central question of our study is: what do modern machine learning systems lack to show human-like behavior, not only in terms of performance, but also in how they organize and form representations?” explains lead author Lukas Muttenthaler, scientist at the MPI CBS, the BIFOLD institute of TU Berlin, and former employee at Google DeepMind.

The researchers show that human knowledge is typically organized hierarchically, from fine-grained distinctions (e.g., “pet dog”) to coarse ones (e.g., “animal”). Machine learning systems, on the other hand, often fail to capture these different levels of abstraction and semantics. To align the models with human conceptual knowledge, the scientists first trained a teacher model to imitate human similarity judgments. This teacher model thus learned a representational structure that can be considered “human-like.” The learned structure was then used to improve already pretrained, high-performing Vision Foundation Models, the so-called student models, through a process called soft alignment. This fine-tuning requires several orders of magnitude less computational cost than retraining the models from scratch.

A crucial step toward interpretable, cognitively grounded AI

The student models were fine-tuned using „AligNet“, a large image dataset synthetically generated through the teacher model that incorporates similarity judgments corresponding to human perceptions. To evaluate the fine-tuned student models, the researchers used a specially collected dataset known as the „Levels“ dataset.

“For this dataset, around 500 participants performed an image-similarity task that covered multiple levels of semantic abstraction, from very coarse categorizations to fine-grained distinctions and category boundaries. For each judgment, we recorded both full response distributions and reaction times to capture potential links with human decision uncertainty. The resulting dataset represents a newly established benchmark for human-machine alignment, which we open-scourced,” reports Frieda Born, PhD student at BIFOLD and the MPI for Human Development.

The models trained with „AligNet“ show significant improvements in alignment with human judgments, including up to a 93.5% relative improvement in coarse semantic evaluations. In some cases, they even surpass the reliability of human ratings. Moreover, these models exhibit no loss in performance; on the contrary, they demonstrate consistent performance increases (25% to 150% relative improvement) across various complex real-world machine learning tasks, all at minimal computational cost.

Klaus-Robert Müller, Co-director at BIFOLD: “Our research methodologically bridges cognitive science (human levels of abstraction) and modern deep-learning practice (Vision Foundation Models), thus forming a link between the concept of representation in humans and in machines. This represents an important step toward more interpretable, cognitively grounded AI.”

AligNet demonstrates that hierarchical conceptual structures can be transferred to neural networks without explicit hierarchical training, with reorganisation visible across network layers. These results suggest that AligNet achieves fundamental improvements in visual representations that better reflect the human level of conceptual understanding, thereby making AI less of a ‘black box’.

Andrew K. Lampinen from Google DeepMind adds: “For the first time, researchers have found an efficient way to teach computer vision models about the hierarchical structure of human conceptual knowledge. We show that this not only makes the representations of these models more human-like, and therefore more interpretable, but also improves their predictive power and robustness across a wide range of vision tasks.”

Further links

Original press release

> more

Original publication

> more

When AI “thinks” like us

15. December 2025/in Ausgewählter Aktuelles-Post für die Startseite /by Elena Reiriz Martinez

Kontakt Aktuelles

Contact

Prof. Dr. Klaus-Robert Müller

Berlin Institute for the Foundations of Learning and Data (BIFOLD)
Department of Machine Learning
Faculty IV – Electrical Engineering and Computer Science
Technical University of Berlin

klaus-robert.mueller@tu-berlin.de

Bernstein Netzwerk Computational Neuroscience Logo

Become a member
Statutes
Donation
Subscribe to Newsletter

 

Follow us on

LinkedIn
Bluesky
Vimeo
X
© 2026 Bernstein Network Computational Neuroscience
  • Contact
  • Imprint
  • Privacy Policy
Scroll to top Scroll to top Scroll to top
Cookie-Zustimmung verwalten
We use cookies to optimize our website and our service.
Functional Always active
Der Zugriff oder die technische Speicherung ist unbedingt für den rechtmäßigen Zweck erforderlich, um die Nutzung eines bestimmten Dienstes zu ermöglichen, der vom Abonnenten oder Nutzer ausdrücklich angefordert wurde, oder für den alleinigen Zweck der Übertragung einer Nachricht über ein elektronisches Kommunikationsnetz.
Vorlieben
Die technische Speicherung oder der Zugriff ist für den rechtmäßigen Zweck der Speicherung von Voreinstellungen erforderlich, die nicht vom Abonnenten oder Nutzer beantragt wurden.
Statistics
Die technische Speicherung oder der Zugriff, der ausschließlich zu statistischen Zwecken erfolgt. Die technische Speicherung oder der Zugriff, der ausschließlich zu anonymen statistischen Zwecken verwendet wird. Ohne eine Aufforderung, die freiwillige Zustimmung Ihres Internetdienstanbieters oder zusätzliche Aufzeichnungen von Dritten können die zu diesem Zweck gespeicherten oder abgerufenen Informationen allein in der Regel nicht zu Ihrer Identifizierung verwendet werden.
Marketing
Die technische Speicherung oder der Zugriff ist erforderlich, um Nutzerprofile zu erstellen, um Werbung zu versenden oder um den Nutzer auf einer Website oder über mehrere Websites hinweg zu ähnlichen Marketingzwecken zu verfolgen.
Manage options Manage services Manage {vendor_count} vendors Read more about these purposes
Settings
{title} {title} {title}