Skip to main content

Can AI Actually Discover New Knowledge?

A common question about modern AI systems is whether they are actually capable of producing something new, or whether everything they generate is ultimately a rearrangement of what they have already seen. When a model writes an essay, proposes an idea, or even contributes to scientific research, the result can feel original, as if something genuinely new has been created. At the same time, these systems are trained on existing data, which makes it unclear whether anything they produce can truly go beyond it.

This tension between creation and recombination leads to a deeper question about what it means to generate knowledge at all, and whether novelty depends on the source of an idea or on the structure it takes when it appears.

What AI Systems Are Trained to Do

Most modern AI systems learn by identifying patterns in large datasets and using those patterns to generate outputs that follow similar structures. In language models, this involves predicting the next word in a sequence, while in other systems it may involve recognizing relationships, optimizing decisions, or mapping inputs to outputs in a consistent way.

Because of this, it is easy to assume that every output must be derived from something already present in the training data. However, this assumption becomes less clear when we consider how combining patterns can produce results that were never explicitly shown during training. A model can connect ideas from different contexts, generate new analogies, or solve problems it has never encountered before, which suggests that recombination can lead to outcomes that feel genuinely new.

When Recombination Becomes Novelty

This raises an important question about where novelty actually comes from. If an idea has never appeared before in any dataset, but can still be traced back to underlying patterns within that data, should it count as new knowledge?

Human creativity already operates in a similar way. Scientific theories build on earlier work by connecting ideas that were previously separate. Mathematical discoveries often extend existing structures into unfamiliar domains. Even creative writing draws from language and concepts that already exist, while arranging them in ways that feel original.

From this perspective, novelty does not require complete independence from prior knowledge. It can emerge from the way existing elements are extended or reinterpreted. In this sense, recombination is not the opposite of creation, but one of the processes through which it happens.

This visualization shows how books cluster in a high-dimensional embedding space based on learned similarities in content and style. Even without explicit labels, related genres and themes form natural groupings, suggesting that AI systems organize knowledge as structured relationships rather than fixed symbolic definitions.
"Neural Network Embeddings Explained", Will Koehrsen, Oct 1 2018, towards data science, https://towardsdatascience.com/neural-network-embeddings-explained-4d028e6f0526/

The Limits of AI-Generated Novelty

Even if AI systems can produce new outputs, there are reasons to question whether this counts as discovery in a deeper sense. These systems don't choose their own problems, and they don't decide which directions are worth exploring. Their behavior is shaped by training objectives and external prompts, rather than by curiosity or intention.

Therefore, the results they generate may lack a sense of significance that is often associated with human discovery. A model can produce a correct or useful answer, yet it does not recognize why that answer matters or how it fits into a broader understanding of the world. The process remains tied to pattern generation, rather than to the pursuit of meaning.

This suggests that novelty alone may not be enough to define discovery. The context in which an idea is produced, and the way it is understood, may also play an important role.

Knowledge as Structure and Knowledge as Understanding

The question can also be framed in terms of what knowledge itself consists of. If knowledge is understood as structured information, then AI systems may already be capable of generating it, since they can produce new configurations that reveal relationships between ideas. However, if knowledge involves understanding, interpretation, or awareness, then the absence of subjective experience in these systems becomes more significant.

Philosophical views differ on this point. Some traditions treat knowledge as something objective and independent, waiting to be uncovered through the right methods. Others emphasize the role of perspective, suggesting that it matters how that knowledge is experienced and interpreted.

The distinction between these views becomes especially important when considering AI, since it highlights the gap between generating correct outputs and grasping their meaning.

AI systems have contributed to areas like protein folding and mathematical discovery, producing results that extend beyond their training data, which raises the question of whether these outputs should be understood as genuine discoveries or as structured recombinations.
"
AlphaFold Proves That AI Can Crack Fundamental Scientific Problems", Payal Dhar, Dec 7 2020, IEEE Spectrum, https://spectrum.ieee.org/alphafold-proves-that-ai-can-crack-fundamental-scientific-problems

So Can AI Discover New Knowledge?

The answer depends on how discovery is defined. If discovery involves producing something new, useful, and not explicitly present in prior data, then AI systems may already meet that standard in certain cases. If discovery requires understanding, intention, or awareness of significance, then the situation becomes less clear.

This question ultimately reflects back on human knowledge as well. Much of what is considered discovery may itself emerge from recombination, guided by intuition and existing frameworks. The difference may lie less in the structure of the output and more in the way it is interpreted and integrated into a broader understanding.

Comments

Popular posts from this blog

Does String Theory Count as Science?

String theory is one of the most ambitious and imaginative ideas in modern physics. It aims to do something no other theory has done: unify all the fundamental forces of nature ( gravity, electromagnetism, the strong nuclear force, and the weak nuclear force) into a single framework. It replaces point-like particles with tiny vibrating strings , whose vibrations determine the type of particle you observe. But despite its promise, string theory is also one of the most controversial theories, because right now, it can't be tested . So this leads to a deep philosophical question: If a theory explains everything but can’t be tested, does it still count as science? In string theory, fundamental particles like electrons, protons, and quarks are represented as tiny vibrating strings. The type of particle is determined by the string’s vibrational pattern, similar to how different notes come from the same guitar string. Tripathi, A. (2024, March 24). String Theory: Dimensional Implicatio...

Can Intelligence Emerge from Simplicity?

Recently, I joined BrightStar Labs  ( https://brightstarlabs.ai/ ) as an affiliate researcher, where I’ve been working on a class of systems known as Emergent Models (EMs). They’re not built like neural networks or traditional programs. Instead, they evolve from very simple beginnings: just a line or grid of colored cells following a small set of update rules. Over time, patterns emerge. Some of those patterns perform tasks that resemble basic forms of computation. At first, EMs may seem abstract or even mechanical. But the more I work with them, the more they raise interesting questions. These questions aren't just about computation, but about how we understand concepts like memory, intelligence, and structure. Intelligence Without Intent? One of the main features of an EM is that it’s not designed in the usual sense. There's no optimizer telling it what to do, and no fixed model architecture. Instead, its behavior is shaped by initial conditions and a rule table that dete...

Is Consciousness Just a Feedback Loop?

One of the most persistent mysteries in philosophy and neuroscience is this: Why does the brain feel like anything at all? Why does some neural activity come with the experience of being , while other processes like digestion or circulation do not? A compelling proposal gaining traction is Recurrent Processing Theory (RPT) , which offers a specific answer: Consciousness arises when information isn’t just processed forward, but also loops back. This post explores a brief overview of the core ideas behind RPT, why this kind of recurrence matters, and what it could mean for how we understand the structure of conscious experience. What Is Recurrent Processing Theory? At its heart, Recurrent Processing Theory claims that consciousness depends not just on the feedforward sweep of information through the brain (say, from your retina through the visual cortex), but on recurrent interactions,  where higher areas send signals back to earlier areas, creating feedback loops. This idea wa...