10 mistakes startups make when building with GenAI

Insights from a GenAI expert on the most common mistakes teams make when building with AI.

Anton Vasiliev
March 3, 2025
Table of contents

When you’re building an AI startup or teams that are expected to be AI-native, one question keeps coming up in the background: how to avoid the mistakes that seem almost inevitable, from poor data quality and missing guardrails to hidden costs at scale, frustrating user experiences, and risks around bias, security, and legal exposure.

I spoke with Yuval Machlin, CEO at hiFred and an influential voice in the GenAI space who has guided many teams in building and scaling AI products. We discussed the most common mistakes that teams make when building with GenAI and now I’m eager to share all the insights with you.

Disclaimer: We're not going to teach you how to run your startup. This is just our take on what’s happening in the industry. But hey, maybe it’s time to pause on the GenAI rush and think a bit.

The ‘GenAI or die’ trap

In the race to integrate AI into products, many startups jump in simply because everyone else is doing it. It creates the illusion of progress but in reality, it often leads to scattered use cases, unclear priorities, and products that don’t meaningfully improve.

And this gap between expectation and reality is where many AI startups fail.

Teams invest in GenAI without defining what success looks like, how it connects to the product, or what needs to be in place to support it long-term. As a result, promising ideas turn into wasted budget, stalled initiatives, and features that never fully deliver.

However, the companies that get it right approach AI differently. They build around well-defined use cases, approach trends thoughtfully, invest in the right custom software development capabilities early, and treat AI as part of the product.

Now, let’s get into the real problems and look at the most common and costly GenAI mistakes startups keep making.

Mistake 1. Garbage in, garbage out

One of the most fundamental mistakes is underestimating data quality.

The point is that GenAI is only as good as the data it’s trained on. Feed them junk, and they’ll spit out junk. Simple.

But startups trying to build their own models often lack enough quality data to train them. As a result, they get a system that learns from bad examples and produces unreliable, biased, or totally useless outputs.

Some startups skip the headache by using existing models via APIs, sidestepping the data nightmare. But for those determined to build their own GenAI, high-quality and diverse training data is non-negotiable. For example, if you’re training a model to identify plants, random internet pictures won’t cut it. You need thousands of images in different lighting, angles, and conditions. Sometimes far more, depending on complexity. Otherwise, the model will fail in real-world scenarios. 

Synthetic data can be used here as a workaround. Microsoft’s Phi models, Google’s Gemma, and Nvidia’s latest AI models all incorporate synthetic data. By 2030, synthetic data could be a $2.5 billion market but it’s still AI-generated guesswork. Overrelying on it creates a feedback loop detached from reality.

Mistake 2. No guardrails, no control

AI without guardrails is one of the fastest ways AI failures turn into real problems.

When safeguards are weak or missing, GenAI systems can generate biased, offensive, or harmful content. And once that reaches users, it quickly becomes a legal and reputational risk.

Even leading players struggle here. In early 2024, a vulnerability in OpenAI’s DALL·E 3 allowed users to bypass safety filters and generate inappropriate images.

The broader issue is systemic. The 2025 AI Safety Index by the Future of Life Institute points out that safety still lags behind development. Companies are moving fast but their safeguards often lack depth and consistency, falling short of emerging standards like the EU AI Code of Practice.

Startups using off-the-shelf models believe they’re inheriting the same level of safety as industry giants. But in fact they’re not. Many open-source models come with zero built-in protections, exposing founders to regulatory backlash, reputational damage, and potential security disasters.

On top of that, failing to lock down data is another ticking time bomb. If an AI model is training on user inputs by default, like ChatGPT before OpenAI rolled out privacy controls, startups risk leaking private or proprietary data. And once it’s in the model, there’s no undo button.

Mistake 3. Ignoring GenAI’s limits

AI models have strengths but they also have significant weaknesses that can wreck a product if founders don’t test them properly.

Many assume a GenAI model will handle any task they throw at it. Need a marketing copy? Just ask the AI. But in reality this is just another AI implementation mistake.

A model might spit out catchy headlines but completely miss the brand’s tone. A chatbot might sound smart in a demo but fail with real customer queries. And more interestingly, different models behave differently. Stanford’s 2024 AI Index Report highlighted that startups relying on a single GenAI system often hit performance ceilings. Using multiple models in tandem can considerably improve accuracy, but it’s a strategy few founders take into account.

Mistake 4. High cost at scale

Some AI startups fail because they underestimate how expensive GenAI becomes at scale.

Scaling costs can indeed get out of hand. Generating content isn’t free and when AI startups start processing massive amounts of text, images, or video, expenses enter the stratosphere. A single AI-generated image costs significantly more in compute than a text query, and large-scale models drain resources fast.

The numbers don’t lie. Developing a generative AI product can cost anywhere from $5 million to $20 million, depending on model complexity, data processing, and infrastructure. And that’s before factoring in ongoing cloud expenses. According to Epoch AI, the compute cost of training frontier AI models has been doubling every 10 months, putting even deep-pocketed startups under pressure.

Cloud compute cost to train frontier AI models over time
Cloud compute cost to train frontier AI models over time 
Source: Epoch AI

While GenAI unlocks exciting capabilities, many founders forget that cheaper, traditional methods often work fine. Basic automation and rule-based systems can handle tasks that don’t need AI.

Mistake 5. User frustration

Too many startups rush to launch AI-powered chatbots, search tools, and assistants without real-world testing, which results in a frustrating and clunky experience that drives users away.

When customers receive irrelevant or unhelpful responses, they lose trust and patience. A 2025 Gartner survey, for instance, found that 50% of customers would prefer companies that didn't use AI for customer service at all. That’s not only a trust problem; it directly translates into lost revenue.

Mistake 6. The illusion of stability

Many startups expect GenAI to behave like traditional software. But it doesn’t and this is why many AI startups fail.

So, for instance, ask the same question twice, and you may get different answers. Sometimes the variation is minor but in some situations, there might be a complete contradiction. Thus, you always need to remember that AI responses may include mistakes.

Yet teams rarely test for this properly and wrongly assume consistency instead of validating it.

A study led by Professor Mesut Cicek and his team at Washington State University highlights the issue. Researchers tested over 700 hypotheses by submitting each one 10 times to ChatGPT, asking it to determine whether the statements were supported by research. The results showed clear inconsistencies across responses.

What you need to remember is that GenAI is not static. It shifts with phrasing, context, and even silent model updates. Ignore that, and your AI will be unpredictably wrong.

Mistake 7. Legal issues

Just because something is online doesn’t mean it is free to use.

Many founders feed datasets into GenAI without thinking about ownership, licensing, or consent. That assumption is already leading to real AI failures and lawsuits.

In 2025, The New York Times sued Perplexity AI over the use of copyrighted content. Earlier, Getty Images filed a case against Stability AI for training on copyrighted images. While the court later limited the scope of liability in that case, it still raised important questions around how training data is sourced and how AI outputs can reproduce protected material.

What’s more, data privacy regulations are getting stricter and compliance is becoming a moving target.

Ignore this, and the consequences go beyond technical issues and impact your product, funding, and reputation.

Mistake 8. Overlooking bias until it’s too late

Bias is one of the biggest AI mistakes that startups routinely ignore until it fails. Train a model on biased data, and it will reflect those biases, whether that means racial, gender, or cultural distortions.

Researchers at Anthropic tried reducing bias in AI responses but found that doing so sometimes made the model less capable overall. Meanwhile, Google now conducts fairness reviews before launching AI-driven products, recognizing that unchecked bias can damage trust and invite regulatory scrutiny.

Feature steering showing on and off-target effects
How feature steering increase or decrease specific social biases
Source: Anthropic

The trade-off between fairness and accuracy is messy, and there’s no universal solution. Adversarial training and algorithmic audits show promise but startups rarely invest in them early. Instead, they wait until users or regulators call them out. 

Mistake 9. Security as an afterthought

Security gaps are behind some of the most damaging and costly AI mistakes.

In January 2025, for instance, Chinese AI startup DeepSeek faced a large-scale cyberattack that forced the company to limit new user registrations. This attack coincided with DeepSeek's AI assistant becoming the top-rated free app on the U.S. Apple App Store, leading to significant website outages. 

This isn’t new. AI models have been tricked into leaking internal logic, private prompts, and even personal user information. And many founders don’t check where their AI tools send or store data. Some free API versions quietly reserve the right to user inputs for future training. The fine print? Most don’t even read it.

Mistake 10. Blind trust in AI results

Security gaps also lead to the most costly and damaging AI failures.

Many founders don’t validate results before shipping AI-driven products, which ​​we see across the most common AI pitfalls. AI hallucinations happen between 3% and 10% of the time, depending on the model. GPT-4’s hallucination rate is around 3%, while Google’s PaLM 2 reportedly reaches 27%.

AI model perfromance: accuracy vs hallucination rate
AI hallucination leaderboard
Source: Prompt Engineering & AI Institute

Yet many founders don’t validate results before shipping AI-driven products. They think the model "mostly works" and let users figure out the mistakes. In high-stakes industries like healthcare or finance, even a small error rate can lead to serious consequences. And users won’t forgive it.

Wrapping up

Is GenAI a revolution? Absolutely! Is it helpful for many industries? Without a doubt. But is it a magic bullet for everyone? Not really. Before you incorporate GenAI into your products, take a step back and think about your users first. Will it solve their problems? If the answer is yes, then move on. Otherwise resist following hype blindly. 

If you're a founder struggling with AI integration or simply an early-stage startup in need of a fractional CTO, we're happy to help. On The Spot provides CTO as a Service, assisting startups without a permanent CTO in areas like strategic technology planning, technology roadmap development, and overcoming scalability issues. If that sounds like your situation, drop us a line.

Stay in tune
with On The Spot

Monthly digest of the latest open positions, tech events, and podcast episodes.
Thank you!

You've successfully signed up for our newsletter. Keep an eye on your inbox.

Oops! Something went wrong while submitting the form.

Discover what else is happening at On The Spot

TechSpot

Tech events in Poland and online for everyone interested in software architecture and systems design. Like-minded people, top experts from leading companies, and a big engineering community.

137 Podcast

Conversations with top engineers, CTOs, and founders about day-to-day engineering and everything that excites us about tech.