Matt Asay
Contributing Writer

Your generative AI project is going to fail

opinion
Jul 1, 20245 mins
Data ScienceGenerative AIMachine Learning

Fueled by vibes and with stars in their eyes, enterprises are not taking the time to understand generative AIโ€™s limitations and to create their own rules-based approach.

Your generative AI project is almost certainly going to fail. But take heart: You probably shouldnโ€™t have been using AI to solve your business problem, anyway. This seems to be an accepted fact among the data science crowd, but that wisdom has been slow to reach enterprise executives. For example, data scientist Noah Lorang once suggested, โ€œThere is a very small subset of business problems that are best solved by machine learning; most of them just need good data and an understanding of what it means.โ€

And yet 87% of companies surveyed by Bain & Company said theyโ€™re developing generative AI applications. For some, thatโ€™s the exactly right approach. For many others, itโ€™s not.

We have collectively gotten so far ahead of ourselves with generative AI that weโ€™re setting ourselves up for failure. That failure comes from a variety of sources, including data governance or data quality issues, but the primary problem right now is expectations. People dabble with ChatGPT for an afternoon and expect it to be able to resolve their supply chain issues or customer support questions. It wonโ€™t. But AI isnโ€™t the problem, we are.

โ€˜Expectations set purely based on vibesโ€™

Shreya Shankar, a machine learning engineer at Viaduct, arguesย that one of the blessings and curses of genAIย is that it seemingly eliminates the need for data preparation, which has long been one of the hardest aspects of machine learning. โ€œBecause youโ€™ve put in such little effort into data preparation, itโ€™s very easy to get pleasantly surprised by initial results,โ€ she says, which then โ€œpropels the next stage of experimentation, also known as prompt engineering.โ€

Rather than do the hard, dirty work of data preparation, with all the testing and retraining to get a model to yield even remotely useful results, people are jumping straight to dessert, as it were. This, in turn, leads to unrealistic expectations: โ€œGenerative AI and LLMs are a little more interesting in that most people donโ€™t have any form of systematic evaluation before they ship (why would they be forced to, if they didnโ€™t collect a training dataset?), so their expectations are set purely based on vibes,โ€ Shankar says.

Vibes, as it turns out, are not a good data set for successful AI applications.

The real key to machine learning success is something that is mostly missing from generative AI: the constant tuning of the model. โ€œIn ML and AI engineering,โ€ Shankar writes, โ€œteams often expect too high of accuracy or alignment with their expectations from an AI application right after itโ€™s launched, and often donโ€™t build out the infrastructure to continually inspect data, incorporate new tests, and improve the end-to-end system.โ€ Itโ€™s all the work that happens before and after the prompt, in other words, that delivers success. For generative AI applications, partly because of how fast it is to get started, much of this discipline is lost.

Things also get more complicated with generative AI because there is no consistency between prompt and response. I love the way Amol Ajgaonkar, CTO of product innovation at Insight, put it. Sometimes we think our interactions with LLMs are like having a mature conversation with an adult. Itโ€™s not, he says, but rather, โ€œItโ€™s like giving my teenage kids instructions. Sometimes you have to repeat yourself so it sticks.โ€ Making it more complicated, โ€œSometimes the AI listens, and other times it wonโ€™t follow instructions. Itโ€™s almost like a different language.โ€

Learning how to converse with generative AI systems is both art and science and requires considerable experience to do it well. Unfortunately, many gain too much confidence from their casual experiments with ChatGPT and set expectations much higher than the tools can deliver, leading to disappointing failure.

Put down the shiny new toy

Many are sprinting into generative AI without first considering whether there are simpler, better ways of accomplishing their goals. Santiago Valdarrama, founder of Tideily,ย recommends starting with simple heuristics, or rules. He offers two advantages to this approach: โ€œFirst, youโ€™ll learn much more about the problem you need to solve. Second, youโ€™ll have a baseline to compare against any future machine-learning solution.โ€

As with software development, where the hardest work isnโ€™t coding but rather figuring out which code to write, the hardest thing in AI is figuring out how or if to apply AI. When simple rules need to yield to more complicated rules, Valdarrama suggests switching to a simple model. Note the continued stress on โ€œsimple.โ€ As he says, โ€œsimplicity always winsโ€ and should dictate decisions until more complicated models are absolutely necessary.

So, back to generative AI. Yes, genAI might be what your business needs to deliver customer value in a given scenario. Maybe. Itโ€™s more likely that solid analysis and rules-based approaches will give the desired yields. For those who are determined to use the shiny new thing, well, even then itโ€™s still best to start small and simple and learn how to use generative AI successfully.

Matt Asay

Matt Asay runs developer marketing at Oracle. Previously Asay ran developer relations at MongoDB, and before that he was a Principal at Amazon Web Services and Head of Developer Ecosystem for Adobe. Prior to Adobe, Asay held a range of roles at open source companies: VP of business development, marketing, and community at MongoDB; VP of business development at real-time analytics company Nodeable (acquired by Appcelerator); VP of business development and interim CEO at mobile HTML5 start-up Strobe (acquired by Facebook); COO at Canonical, the Ubuntu Linux company; and head of the Americas at Alfresco, a content management startup. Asay is an emeritus board member of the Open Source Initiative (OSI) and holds a JD from Stanford, where he focused on open source and other IP licensing issues. The views expressed in Mattโ€™s posts are Mattโ€™s, and donโ€™t represent the views of his employer.

More from this author