What will be the most important trends in AI in 2025? Well, I'm going to share my own educated guesses. I don't have any kind of top secret classified information or anything.
But also, this isn't my first rodeo. I did take a shot at predicting AI trends for 2024 and well, I think I did alright. Although a little confession about that video.
I waited until March of 2024 to shoot it. So I already had like a quarter of the year to go on. But that's not the case this time.
So let's get cracking with eight important AI trends in 2025. Let's start with an obvious one. Number one.
Agenetic AI. Every time we post the video about agents to this channel, viewership spikes. So there's clearly an appetite for understanding this tech.
So what are AI agents? Well, the intelligence systems that can reason they can plan and they can take action. An agent can break down complex problems to create multi step plans and that can interact with tools and databases to achieve goals.
And I think most people are on board with the utility of a well-performing AI agent. Trouble is, today's models, well, they struggle with consistent logical reasoning. They can usually execute simple plans, but when it comes to handling complex scenarios with multiple variables, they have to lose track and they make decisions that don't quite add up.
So we'll need better models in 2025. Speaking of which, trends number two is inference time compute. Now, during inference and our model goes to work on real time data, comparing the user's query with information processed during training and stored in its weights.
New AI models are extending inference processing to essentially spend some time thinking before giving you an answer, and the amount of time it spends. Thinking is variable based on how much reasoning it needs to do so. A simple request that might take a second or two or something larger and harder might take several minutes.
And what makes inference time compute models interesting is the inference reasoning is something that can be tuned and improved without having to train and tweak the underlying model. So there are now two places in the development of an LLM where reasoning can be improved at training time with better quality training data, but now also inference time with better chain of thought training, which could ultimately lead to smarter AI agents. All right.
Trend number three is very large models. Large language models consist of many parameters which are refined over the training process. Now, the frontier models in 2024.
They're in the range of like 1 to 2 trillion parameters in size. The next generation of models are expected to be many times larger than that, perhaps upwards of 50 trillion parameters. But if 2025 is the year of enormous models, it may also be the year of number four, very small models, models that are only a few billion parameters in size.
And yet you don't hear the phrase only a few billion very often, but there you go, and these models, they don't need huge data centers loaded with stacks of GPUs to operate. They can run on your laptop or even on your phone. Actually, I have the 2 billion parameter IBM Granite three model running on my laptop, and my device doesn't even have to break a sweat to run it.
So expect to see more models of this size tuned to complete specific tasks without requiring large compute overhead. Now, do you know what the most common enterprise use cases were for AI in 2024? Well, according to a Harris poll, it's improving customer experience, IT operations and automation, virtual assistants and cyber security.
Looking ahead to 2025, we will see more advanced use cases. So think customer service bots that can actually solve complex problems instead of just routing ticket. So think about AI systems that can proactively optimize entire IT networks, or think about security tools that can adapt to new threats in real time.
Now, when I first used generative AI back in the day to help me build a beer recipe, the context window for the LLM was a mere 2000 tokens. Today's models have context when those measured in the hundreds of thousands or even the millions of tokens. We are getting close to number six near infinite memory where bots can keep everything they know about us in memory at all times.
We'll soon be in the era of customer service chat bots that can recall every conversation it has ever had with us, which hopefully we'll consider a good thing. Okay. Trend number seven.
That is human in the loop augmentation. Now, perhaps you heard about the study where a chat bot outperformed physicians in clinical reasoning. So 50 doctors were asked to diagnose medical conditions from examining case reports.
A chat bot presented with the same cases actually scored higher than the doctors, but where this gets really interesting is some doctors were randomly assigned to use a chat bot to help them in this study. Now the doctor plus chat bot group also scored lower than when the chat bot was asked to solve the cases alone. And that is a failing of AI and human augmentation.
An expert paired with an effective AI system should be smarter together than either of those two entities operating by themselves. But look, prompting LLM chat bots can be hard. You got to tailor the right prompts.
You've got to ask for things in the right way. So we need better systems that allow professionals to augment AI tools into their workflow without those professionals needing to be experts in how to use AI. So expect more to come in this area.
Now, the final trend in my 2024 Trends video, I actually turn this one over to the audience asking which AI trend do you think will be important in the year ahead? And I'm so glad I did. Hundreds of viewers shared their thoughts.
Well, I know when I'm on to a good thing. So trend number eight, that one is over to you. What do you think will be an important AI trend in 2025?
Let me know in the comments.