Gardner predicts that by 2030, synthetic data will outweigh real data in AI model training, meaning more than 50% (likely well over) of data used in AI systems will be artificially generated
Artificial intelligence creates synthetic data – information that doesn’t exist in the real world. Unlike data drawn from actual events, people, or objects, synthetic data is entirely machine-made. It’s not rooted in human experience. In short, it’s artificial.
Yet, this fabricated data is infiltrating our lives. From training AI models and powering virtual assistants to generating images, voices, and videos, synthetic data is everywhere. Thanks to technological leaps, it can now look and sound astonishingly lifelike. Deepfake videos, for instance, can convincingly depict someone saying or doing something they never did. The boundary between real and fake is growing hazy.
This prompts a troubling question: are we losing our grip on reality?
The Rise of Artificial Worlds
Generative AI makes it effortless to craft virtual environments populated by synthetic characters, stories, and interactions. These digital realms may feel vivid and immersive, but they’re entirely fabricated. The Metaverse is a prime example – a space where AI-generated content constructs an alternate version of existence.
For some, these worlds are magnetic. They offer comfort, excitement, or control that the real world often lacks. What begins as curiosity can spiral into habit, even obsession. In extreme cases, people immerse themselves so deeply in these synthetic spaces that they neglect their actual lives.
They forget to eat. They stop socialising. They skip work. Self-care fades. The digital world becomes their primary reality, while the physical one slips away.
A Digital Addiction?
This behaviour mirrors addiction. Like gambling or substance misuse, the pull of artificial worlds can be relentless. People return repeatedly, chasing the thrill of a flawless image, a convincing deepfake, or an idealised avatar crafted by algorithms.
These spaces aren’t just entertainment. They offer a sense of limitless control, where users can shape experiences to their desires. But the deeper they dive, the more they detach from reality.
This is a serious concern. We risk nurturing a generation ill-equipped to navigate the real world – a world that’s messy, unpredictable, and imperfect.
Why Real Data Matters
Real data, captured from human observation or experience, has a vital edge over its synthetic counterpart: it reflects life. It’s rich with noise, errors, context, and emotion. It captures how people genuinely think, feel, and act in their environment.
Synthetic data, by contrast, is a derivative. Often built from real data patterns, it lacks the depth and spontaneity of lived experience. While useful for tasks like training AI models – especially where privacy or scale is an issue – it’s no substitute for reality. Confusing the two is a mistake.
The Threat to Human Connection
Overreliance on synthetic data risks creating a world that mirrors our desires rather than our truths. It’s tempting to surround ourselves with AI-generated voices that echo our views, images that flatter us, or information that reinforces our beliefs.
This may feel reassuring, but it’s harmful. It stifles personal growth, erodes empathy, and weakens our resilience to real-world challenges. We drift not only from reality but also from each other.
Human relationships thrive on trust, shared experiences, and authentic history. No matter how lifelike, artificial replicas can’t replicate these. A chat with an AI bot may simulate warmth, but it lacks true understanding. It’s no replacement for a friend, family member, or loved one.
Where Do We Go From Here?
Synthetic data has undeniable value. Used thoughtfully, it can enhance systems, solve problems, and drive innovation. But we must maintain a clear distinction between real and artificial. Education, awareness, and robust ethical guidelines are essential to help people – especially younger generations – forge a balanced relationship with technology.
We must also prioritise real-world experiences: face-to-face conversations, direct observation of our surroundings, and opinions grounded in what we see and feel, not just what machines generate.
The more we inhabit artificial worlds, the more we must question: are we living our own lives, or merely watching a simulation? Is technology our tool, or our trap?
Synthetic data is here to stay. Let’s ensure we stay, too – firmly rooted in the real world.
About the author:
Dr Kitty Hung, a BCS Fellow, BCS CITP and member of IIBA, gained her PhD in Computer Science from the University of Sheffield, UK in 1999. She has over 26 years of experience with a career spanning business analysis, project management and business consultancy roles at various prestigious organisations including Cell Structures, London Business School, London Metropolitan Police, Capita, Raytheon Technologies, and AtkinsRéalis having successfully delivered business critical and large-scale projects and programmes for Central Government and Commercial Sectors covering policing, national security, telecommunications, defence, transportation, aviation, disruptive and emerging technologies. Her recent publication of her book titled: “Business Analysis in the era of Generative AI” which has reached the Amazon #1 Best Seller chart has added particular credibility to her future-focused insights.