How to "AI" - LLMs as a Commodity

Johannes Stiehler
Random Rant
#ProjectExecution
#LargeLanguageModels
#ProductDevelopment
Cover Image for How to "AI" - LLMs as a Commodity

Some software is very expensive, some is cheap and some is free. And some types of software cost as much as several million in 2005 and can now be obtained for absolutely nothing.

Enterprise search has been on this journey. In the beginning, the sheer power of putting together a scalable system that you could stuff data into and, ideally, get it out of was worth hundreds of thousands of dollars to customers. And tens of thousands more in professional services to keep it all running (at least some of the time).

Solr, OpenSearch and Vespa are now (and have been for a long time) free in the true sense of the word and can do the same or more.

Is that why people don't make money from search anymore? On the contrary, but the focus has shifted drastically. Customization, scaling the system in heterogeneous environments or smart solutions for business problems are in the spotlight and are paid for. Core software development is the responsibility of the open source community.

The moral of the story: the actual software is (almost) never a revenue generator in the long run, unless it creates a quasi-monopoly for itself. Research, open source, disruptive competitors, all this will sooner or later cause such licensing models to collapse.

In the same way, even today, one should wisely refrain from building a business model on the premise of using LLMs (i.e. Large Language Models). An API wrapper around ChatGPT or a service that in reality only prompted Vicuna cannot last long. Even owning the most wonderful LLM in the world and making it available for money is probably not a sustainable business model in the long run. There is too much competition from different models, more and more of which are freely available or at least cheap to use.

Obviously, this does not bother the infrastructure providers like Microsoft and Amazon very much, they can earn a lot of money with overpriced GPU instances anyway.
But how can you create lasting and unique value in the generative AI space?

I argue that it works very similarly to search engines, document management systems, and other enterprise software: the added value comes from the knowledge and data that users provide to the software and the way you handle it as a service provider.

My top 4 ingredients for unique solutions that will still be worth their money in 5 years:

  1. Technical understanding
    Yes, you have to know what you are doing, how to use, adapt, roll out the appropriate technologies. Even this is lacking in regards to LLMs especially among those who are the loudest on LinkedIn.
  2. Understanding the problem
    What is the problem we are solving? What business process is being improved? How does the customer save money or make money? Don't look for the problem to the technology, look for the technology to the problem is the name of the game here.
  3. The right data in the right form
    "Data is the new oil." This is true insofar as data is the fuel and lubricant of such technologies. What data can be used to fine-tune an LLM so that it becomes an expert in European regulations? What are the best text sections for sample prompts? All these treasures exist at the customer's site and need to be put into the right form for the LLM technology to benefit from them.
  4. User interactions
    A big opportunity of machine learning algorithms in general is that you can build feedback loops to the user. The user doesn't like an answer? His "thumbs down" can directly become a negative example for the next fine-tuning. A user edits the output of the LLM a bit? Great, now we have a positive example or an extension for the prompt.

In "service provider circles," people like to talk a lot about "innovating with the customer." This integration of technical expertise with data assets and process understanding on the part of the customer is how we think of this at NEOMO.

Johannes Stiehler
CO-Founder NEOMO GmbH
Johannes has spent his entire professional career working on software solutions that process, enrich and surface textual information.

There's more where this came from!

Subscribe to our newsletter

If you want to disconnect from the Twitter madness and LinkedIn bubble but still want our content, we are honoured and we got you covered: Our Newsletter will keep you posted on all that is noteworthy.

Please use the form below to subscribe.

NEOMO is committed to protecting and respecting your privacy and will only use your personal data to manage your account and provide the information you request. In order to provide you with the requested content, we need to store and process your personal data. If you consent to us storing your personal data for this purpose, please check the box below.

Follow us for insights, updates and random rants!

Whenever new content is available or something noteworthy is happening in the industry, we've got you covered.

Follow us on LinkedIn and Twitter to get the news and on YouTube for moving pictures.

Sharing is caring

If you like what we have to contribute, please help us get the word out by activating your own network.

More blog posts

Image

"Digitale Wissensbissen": The future of data analysis – A conversation with Christian Schömmer

Data Warehouse, Data Lake, Data Lakehouse - the terms are constantly escalating. But what do I really need for which purpose? Is my old (and expensive) database sufficient or would a “Data Lakehouse” really help my business? Especially in combination with Generative AI, the possibilities are as diverse as they are confusing. Together with Christian Schömmer, we sit down in front of the data house by the lake and get to the bottom of it.

Image

"Digitale Wissensbissen": Generative AI in Business-Critical Processes

After the somewhat critical view of generative AI in the last episode, this time we are looking at the specific application: can generative AI already be integrated into business processes and, if so, how exactly does it work? It turns out that if you follow two or three basic rules, most of the problems fade into the background and the cool possibilities of generative AI can be exploited with relatively little risk. We discuss in detail how we built a compliance application that maximizes the benefits of large language models without sacrificing human control and accountability. (Episode in German)

Image

Bleeding Edge - curse or blessing?

We rely on bleeding edge technologies to drive companies forward with innovative solutions. That's why, in discussions with customers and partners or in our webinars, we are always keen to explain the benefits and possibilities of modern technologies to companies. But we also use AI ourselves: by automating tendering processes, we have been able to save valuable resources and increase efficiency.