Developers are tired of hearing about AI as a panacea. The backlash may be just what organizations need to effectively implement the technology.
At a developers conference I attended not too long ago, attendees did little to hide their disdain every time the term โAIโ was bandied about. (And it was bandied about a lot!)ย So I was careful on a recent call attended by about 250 engineers to preface the AI portion of the discussion with, โI know this will make you cringe, butโฆโ That got a lot of knowing laughs and thumbs-up emojis.
Whatโs going on here? Itโs not that these developers and engineers are against the use of AI; itโs that they are tired of hearing about artificial intelligence as a panacea without pragmatism. They want to hear about how they can pragmatically and easily harness it โ now โ for real-life use cases.ย
Indeed, weโve spent the last few years bombarded by hyperbolic talk about AI (Robotaxis anyone?). How itโs going to transform life as we know it. How itโs going to take our jobs. When it will become sentientโฆ
Meanwhile, AI has kind of quietly become part of the fabric of our lives โ not by changing our lives or taking our jobs or becoming sentient, but by making our lives and our jobs easier. For example, when I Googled โWhen will AI become sentient?โ (and โWhen did Skynet become self-aware,โ for comparison purposes), I didnโt have to comb through results one at a time but instead read the AI-generated summary of the most relevant content at the top, with sources. (Spoiler alert: Opinions are mixed.)
There are hundreds of other examples of AI applications that are, well, pretty boring but really useful. Whatโs a lot less boring right now is scaling and integrating AI across the organization. And thatโs where the AI backlash can be leveraged.
Making AI usefully boring
Developers, engineers, operations personnel, enterprise architects, IT managers, and others need AI to be as boring for them as it has become for consumers. They need it not to be a โthing,โ but rather something that is managed and integrated seamlessly into โ and supported by โ the infrastructure stack and the tools they use to do their jobs. They donโt want to endlessly hear about AI; they just want AI to seamlessly work for them so it just works for customers.
Organizations can support that by using tools that are open, transparent, easy to use, compatible with existing systems, and scalable. In other words, boring.
The open source RamaLama projectโs stated goal, for example, is to make AI boring through the use of OCI containers. RamaLama facilitates the ability to locally discover, test, learn about, and serve generative AI models โ in containers. It first inspects your system for GPU support, defaulting to CPU support if no GPUs are present. It then uses either Podman or Docker (or runs on the local system if neither are present) to pull the OCI image you want with all the software needed to run an AI model with your systemโs setup. This eliminates the need for users to perform complex configurations on their systems for AI.ย
The Ollama project similarly helps users get up and running with AI models locally, but it doesnโt help you run in production. RamaLama goes a step further by helping you push that model into a container image and then push that container image out to a registry. Once you have a container image, you can ship it off, fine-tune it, and bring it back. It gives you the portability of containers for model development.
(My colleagues Dan Walsh and Eric Curtin posted a great video on YouTube that puts RamaLama in perspective, along with a demo.)
RamaLama isnโt the only project or product that can support AI humdrumness, but itโs a great example of the kinds of things to look for when adopting AI systems across the organization.
Right-sizing the models
The models themselves are also, rightly, growing more mainstream. A year ago they were anything but, with talk of potentially gazillions of parameters and fears about the legal, privacy, financial, and even environmental challenges such a data abyss would create.
Those LLLMs (literally large language models) are still out there, and still growing, but many organizations are looking for their models to be far less extreme. They donโt need (or want) a model that includes everything anyone ever learned about anything; rather, they need models that are fine-tuned with data that is relevant to the business, that donโt necessarily require state-of-the-art GPUs, and that promote transparency and trust. As Matt Hicks, CEO of Red Hat, put it, โSmall models unlock adoption.โ
Similarly, organizations are looking for ways to move AI from the rarefied air of data science to a place where stakeholders across the organization can understand and make use of it as part of their day-to-day work. For developers, this kind of democratization requires tools that enable safe spaces for experimentation with building, testing, and running intelligent applications.
Hereโs a provocative premise: LLMs and models are just software. Theyโre just files and processes, processes which run on CPUs and GPUs.
It just so happens that we have a technology that can help with files and processes โ Linux containers. Linux is the default platform for AI development, so it makes sense to use Linux containers, which give developers a safe place to experiment without necessarily putting their data in the cloud. Containers also give developers an easy way to move their applications from those safe spaces to production, without having to worry about infrastructure.
A home in containers
Organizations already have container infrastructure like registry servers, CI/CD testing, and production tools like Kubernetes, with key capabilities like scalability, security, and Linux optimizations for running predictive and generative AI workloads. This approach empowers teams to effectively leverage AI capabilities while maintaining flexibility and control over their data across diverse environments.
The AI wave is no different than that surrounding other transformative technologies. (And, yes, I said it, it is transformative.) Think about how we used to say web-based or cloud-based before everything just became web-based or cloud-based and we didnโt need to add the modifier anymore. But that happened only after similar backlash and efforts to make web- and cloud-based technology more usable. The same will (and is) happening with AI.ย
โ
Generative AI Insights provides a venue for technology leadersโincluding vendors and other outside contributorsโto explore and discuss the challenges and opportunities of generative artificial intelligence. The selection is wide-ranging, from technology deep dives to case studies to expert opinion, but also subjective, based on our judgment of which topics and treatments will best serve InfoWorldโs technically sophisticated audience. InfoWorld does not accept marketing collateral for publication and reserves the right to edit all contributed content. Contact doug_dineley@foundryco.com.


