24.11.06
If you're involved in the space of LLMs, you're probably hearing more and more about agents. After all, in a recent Reddit AMA, Sam Altman said "we will have better and better models, but i think the thing that will feel like the next giant breakthrough will be agents". You can believe this or not, but it's been over 1.5 years since GPT-4 released, and hence unclear whether we will see big jumps in foundational models from OpenAI or other companies in the future.
So instead, many people are thinking more and more about LLM agents. Several agentic frameworks already exist which take different approaches to try and enable LLMs to do more things in the real-world more effectively. To name just a few:
But ask yourself, have you actually heard of, or seen, an "agent" being used for a real-world purpose, that couldn't be done just almost easily without the "agent" abstraction?
I'm serious by the way - if you have - I would be very curious to hear about it. Email me at anders@langur.ai
Don't get me wrong - these libraries can have their uses - but I think we're still a long ways away from really having the kinds of agents we can actually rely on for complex real-world tasks.
Most "agents" you hear about are state machines doing things like RAG for Q&A chatbots. We already know LLMs are good at this stuff, so these platforms might serve as useful frameworks to build certain agentic behaviors, but we aren't really opening any "new" capabilities here. When you use an agentic framework for something, you either (1) give up too much control to the point where the agent becomes too unpredictable, or (2) restrict the agent into a specific behavior pattern, which might be handy when your task fits into a neat box - but this is not always the case.
If we're going to have an effective agent abstraction layer on top of LLMs/prompting, it needs to be reliable while retaining behavioral dynamism. You should be able to give an arbitrarily complex task to an agent, and trust that the agent will work towards that task based on reasonable assumptions.
It's not completely clear how to achieve this right now, but I do think there are identifiable areas that we can work on to get much closer:
I'm building Langur, which is an agentic framework that attempts to address all of these limitations. I want to build a platform that takes LLM agents from being an interesting gimmick that's useful on occasion, to dependable actors that you can trust to operate in production systems to solve real problems.
If you're curious to learn more, or want to share your own thoughts on the topic, visit https://langur.ai or shoot me an email at anders@langur.ai. Thanks for reading!