Fueled by vibes and with stars in their eyes, enterprises are not taking the time to understand generative AIโs limitations and to create their own rules-based approach.
Your generative AI project is almost certainly going to fail. But take heart: You probably shouldnโt have been using AI to solve your business problem, anyway. This seems to be an accepted fact among the data science crowd, but that wisdom has been slow to reach enterprise executives. For example, data scientist Noah Lorang once suggested, โThere is a very small subset of business problems that are best solved by machine learning; most of them just need good data and an understanding of what it means.โ
And yet 87% of companies surveyed by Bain & Company said theyโre developing generative AI applications. For some, thatโs the exactly right approach. For many others, itโs not.
We have collectively gotten so far ahead of ourselves with generative AI that weโre setting ourselves up for failure. That failure comes from a variety of sources, including data governance or data quality issues, but the primary problem right now is expectations. People dabble with ChatGPT for an afternoon and expect it to be able to resolve their supply chain issues or customer support questions. It wonโt. But AI isnโt the problem, we are.
โExpectations set purely based on vibesโ
Shreya Shankar, a machine learning engineer at Viaduct, arguesย that one of the blessings and curses of genAIย is that it seemingly eliminates the need for data preparation, which has long been one of the hardest aspects of machine learning. โBecause youโve put in such little effort into data preparation, itโs very easy to get pleasantly surprised by initial results,โ she says, which then โpropels the next stage of experimentation, also known as prompt engineering.โ
Rather than do the hard, dirty work of data preparation, with all the testing and retraining to get a model to yield even remotely useful results, people are jumping straight to dessert, as it were. This, in turn, leads to unrealistic expectations: โGenerative AI and LLMs are a little more interesting in that most people donโt have any form of systematic evaluation before they ship (why would they be forced to, if they didnโt collect a training dataset?), so their expectations are set purely based on vibes,โ Shankar says.
Vibes, as it turns out, are not a good data set for successful AI applications.
The real key to machine learning success is something that is mostly missing from generative AI: the constant tuning of the model. โIn ML and AI engineering,โ Shankar writes, โteams often expect too high of accuracy or alignment with their expectations from an AI application right after itโs launched, and often donโt build out the infrastructure to continually inspect data, incorporate new tests, and improve the end-to-end system.โ Itโs all the work that happens before and after the prompt, in other words, that delivers success. For generative AI applications, partly because of how fast it is to get started, much of this discipline is lost.
Things also get more complicated with generative AI because there is no consistency between prompt and response. I love the way Amol Ajgaonkar, CTO of product innovation at Insight, put it. Sometimes we think our interactions with LLMs are like having a mature conversation with an adult. Itโs not, he says, but rather, โItโs like giving my teenage kids instructions. Sometimes you have to repeat yourself so it sticks.โ Making it more complicated, โSometimes the AI listens, and other times it wonโt follow instructions. Itโs almost like a different language.โ
Learning how to converse with generative AI systems is both art and science and requires considerable experience to do it well. Unfortunately, many gain too much confidence from their casual experiments with ChatGPT and set expectations much higher than the tools can deliver, leading to disappointing failure.
Put down the shiny new toy
Many are sprinting into generative AI without first considering whether there are simpler, better ways of accomplishing their goals. Santiago Valdarrama, founder of Tideily,ย recommends starting with simple heuristics, or rules. He offers two advantages to this approach: โFirst, youโll learn much more about the problem you need to solve. Second, youโll have a baseline to compare against any future machine-learning solution.โ
As with software development, where the hardest work isnโt coding but rather figuring out which code to write, the hardest thing in AI is figuring out how or if to apply AI. When simple rules need to yield to more complicated rules, Valdarrama suggests switching to a simple model. Note the continued stress on โsimple.โ As he says, โsimplicity always winsโ and should dictate decisions until more complicated models are absolutely necessary.
So, back to generative AI. Yes, genAI might be what your business needs to deliver customer value in a given scenario. Maybe. Itโs more likely that solid analysis and rules-based approaches will give the desired yields. For those who are determined to use the shiny new thing, well, even then itโs still best to start small and simple and learn how to use generative AI successfully.


