
Because I say so. That was easy.
No, serious, we just close June and I was lucky enough to participate in both Sapphire in Orlando and Barcelona, brilliant discussions around AI. I will summarize and structure my conversations in this blog.
As already discussed in this community, AI Agents are applications we build on top of an LLM in which we give instructions and expect an outcome. Thats it.
It’s also valuable to highlight what an AI agent is not:
Successful AI Agents are gaining traction of the industry, like search agents (Perplexity). In the SAP world, an Agent is built to complete a specific outcome, like Maintenance Planning Agent .
Agents are the coolest wrappers on top of LLMs these days because, yes, Agents are limited by their model(s). The underlying model(s) you use is the brain of your agent’s body. If the model sucks at making decisions, there is no way to end up in anything good.
In fact, there is a trade-off between over spending our time in AI Agents vs Fine Tuning small models on specific tasks.
Back to my topic, if you look for documentation, there is a lot, but for simplicity, Agents are programmed with Planning, Tools, and Memory (I am missing Instructions because its not coded, and I am oversimplifying)
For each of the blocks described above, we have a lot of coding, and that's why I deliberately removed Instructions because Instructions are not coded.
This means in order to build our agents, we use a Framework.
An AI Framework (like CrewAI, Amazon Bedrock, Microsoft AutoGPT, or SAP Generative AI Hub) provide us the mechanism to develop the Agents, but if you look at the differences between the Frameworks, essentially we can summarize the differences in two blocks;
Back to the origin of this blog, why we should be building our SAP AI agents on BTP and BTP only, Memory and Tools have a fundamental weight in the success of an SAP Agent scenario.
Langchain, LlamaIndex or CrewAI provide sophisticated memory systems for AI agents that includes Short term and long term memory, as well as shared memory or contextual memory, but all of them are FIFO queues based on Function Calling, summarizing the memory management there are two options in discussion
Langchain, LlamaIndex or CrewAI provide sophisticated memory systems for AI agents that includes Short term and long term memory, as well as shared memory or contextual memory, but all of them are FIFO queues based on Function Calling for memory updates
So as long as the model can read Timestamps, it will be Ok. Only AutoGPT's memory uses vector databases like Weaviate, Milvus, Redis or Pinecone for efficient persistent memory retrieval, it saves agent actions for future recall and learning.
But the common of all the frameworks I presented, including AutoGPT, is they are standalone frameworks, they dont belong to our Enterprise Application scenarios which could be the case of Amazon Bedrock or SAP Generative AI Hub.
Amazon Bedrock can leverage the ultra fast DynamoDB for this, while at SAP Generative AI Hub still does not possess memory, we can leverage the langchain.memory class combined with HANA Cloud Vector, for data memory. If you're familiar with word embeddings and text embeddings, leveraging memory type stores vector representations of the conversation, enabling efficient retrieval of relevant context using vector similarity calculations.
I have described tools quite extensively in the past so I will talk about some new conversations about what is the difference between RPAs and AI Agents.
In the architecture of agentic AI, the primary building blocks are the agents, and the business environments they interact with. Each agent operates [semi-]autonomously, perceiving its environment, reasoning about its circumstances, making decisions, and taking appropriate actions, making it really challenging to imagine we will build our agents in standalone frameworks, they must belong to an Enterprise Platform like BTP.
These agents interact with multiple digital environments to achieve specific goals. Central to this system is as we discussed previously, the memory, a repository that allows for seamless communication and coordination among all agents. This shared memory serves as the hub where information, plans, and goals are exchanged, ensuring that each agent can contribute to and benefit from the collective knowledge and strategies, but also, the Data Stores. Data Stores are our data repositories, with a sort of Unstructured Data like text, images, Vector Stores, Structured Data or Knowledge Graphs.
Our journey of building AI agents over the past year has been a roller coaster ride, and we're undoubtedly still in the early stages of this new wave of technology. Here's a brief overview of what I've learned so far.
You must be a registered user to add a comment. If you've already registered, sign in. Otherwise, register and sign in.