Skip to content
AI's Ability to Comprehend Related Concepts: Insights from a New Technique

AI's Ability to Comprehend Related Concepts: Insights from a New Technique

There is no doubt that humans have a distinctive ability to comprehend and apply new concepts in a variety of connected contexts. This function is immediately apparent in everyday actions like grasping the concept of 'skipping', then intuitively understanding the instruction to 'skip twice around the room' or 'skip with your hands up'. Despite tremendous advancements in the artificial intelligence (AI) and machine learning spheres, it remained seemingly plausible for quite some time that AI could not duplicate this unique human ability, often referred to as 'compositional generalizations'.

This debate, initiated by philosophers and cognitive scientists Jerry Fodor and Zenon Pylyshyn in the late 1980s, argued that artificial neural networks, crucial engines of AI, could not form these types of connections. However, the plot may be about to thicken, as researchers continue to unveil strategies that could potentially instill this understanding into neural networks and related technologies, thus reigniting this long-standing discussion.

A collaborative effort from researchers at both New York University and Spain's Pompeu Fabra University has resulted in a groundbreaking technique. The new procedure, known as Meta-learning for Compositionality (MCL), seeks to better these tools' capacity for compositional generalizations. Notably, in cases, it has demonstrated a performance surpassing human capabilities.

The MLC framework powerfully demonstrates the untapped potential present in explicitly practicing these abilities, opposed to expecting this understanding to emerge from standard training protocols. This technique could be critical for developers of current AI systems, especially those employing large language models.

The researchers' creation, MLC, is a unique learning procedure whereby a neural network is continually updated to increase its skills over a series of episodes. In an episode, MLC integrates a new word and uses it compositionally, generating new combinations like 'jump twice' or 'jump around right twice'. Then, with each consecutive episode featuring a different word, the network continues to enhance its compositional skills.

Stringent experiments have validated MLC's efficacy. These trials, in which human participants aimed to perform the same tasks set by MLC, showcased a remarkable result; in some cases, MLC outperformed the human counterparts. It's also noteworthy that both MLC and humans performed better than ChatGPT and GPT-4 on these tasks, despite their impressive general abilities.

Larger language models like ChatGPT, though they have seen improvements within compositional generalization in recent years, still struggle to master this skill. However, the introduction of MLC may potentially further enhance these systems' compositional prowess.

Disclaimer: This content was created with the assistance of an AI tool. For more information, please refer to the source of this article here.