As the world continues to digest the rise of generative AI, agentic AI lies waiting on the bleeding edge. While few accounting firms are using it at the moment, major players in the space have already made significant investments in what they believe to be the next step in the AI revolution.
Very broadly, an AI agent is software that is capable of at least some degree of autonomy to make decisions and interact with things outside itself in order to achieve some sort of goal—whether booking a flight, sending a bill or buying a gift—without needing constant human guidance.
The concept of an AI agent is not new, as computer scientists and software engineers have been using the term for years. Such agents are already used in commercial applications.
Given this history, one might wonder why interest in
“Agents have been around, but we had to program them in a logical format. But because of large language models who can understand natural language, it makes it much more flexible to create very sophisticated systems without having to know so much coding, and it’s much easier to implement on a larger scale,” he said.
He contrasted this with classic AI models.
“If you look at a machine learning thing like recommendations, those are pretty sophisticated. They’re pretty useful in today’s market in entertainment, but those are not really doing a task. They make up a menu like if you need to find Christmas movies. They don’t make a decision by themselves. You make the final decision that I am going to see this. They propose, but you make the decision. [In contrast] an AI agent can accomplish a task,” he said.
Beyond this, advances in generative AI have also made AI agents themselves more effective in the field. Pascal Finette, co-founder and “chief heretic” at tech advisory firm Be Radical, contrasted this with robotic process automation. While he said RPA is not to be underestimated, even today, it tends to be very rigid in its setup, operating mostly on if/then/else principles that work very well for defined use cases but struggles in the face of unstructured data or unusual edge cases. Agents, bolstered with generative AI, become much more flexible.
“The reason I think why this is happening is we now have this superpower of an LLM, which allows us to look at the world and look at data in a much more unstructured way and still get some really interest insight from it, which we can then use to automate stuff, to execute on our behalf. … The beauty of LLMs and gen AI is it has the flexibility to be able to actually create meaningful interactions,” he said.
David Wood, an accounting professor at Brigham Young University whose research also heavily involves AI, noted that “agents” can be thought of as a framework for applying technology. Agents are programmed to do a task, and they can use other tools to accomplish that task, so rather than being some sort of evolution from traditional RPA or generative AI, an agent can be thought of as something that will use RPA and generative AI.
“This is a different framework for how we do programming. We program an agent to do something. It could be to use generative AI, it could be to do a machine learning algorithm, it could be to simply change the color of the font. You can program an agent to do what you want, and agents can work together or even compete against each other to do something, so it is not just a generative AI topic but highly valuable now because agents can use generative AI,” he said.
This increased flexibility has led to major investments in the technology from significant players. Big Four firm KPMG, for example, announced in
Around the same time, accounting solutions provider Thomson Reuters
That same month, Microsoft
Despite these high profile announcements, though, the field is very young, with many applications still in the experimental phase. Finette said it isn’t even necessarily bleeding edge so much as jagged edge. However, based on announcements like these, it appears this is the direction the AI community wants to go next.
Wood agreed, saying there are not a lot of agentic AI solutions right now that are fully production ready, but he sees great potential in the technology once it grows to maturity. For example, many accounting firms bill according to how time is spent, which can be very time consuming to effectively track. Agentic AI would be able to observe an accountant work on company A for 45 minutes and company B for 60 minutes and bill accordingly. He said this might lead to people getting rid of all timekeeping because a computer can do it for them.
He also raised the idea that it could greatly increase efficiency for audits. Imagine, he said, if an agentic AI bot could automatically do most audit confirmations, send them to the humans for approval, and flag the things it couldn’t do itself “so you could build tools for end to end processes to do full tasks together.”
Finette also saw great potential, saying it could act like a full AI worker capable of complex tasks. He said people eventually should be able to go to their AI and say they’re having a meeting in two days with someone and they need a flight and a hotel within their preferred parameters (e.g., cost, distance, etc.) The AI would then perform all the research, compare prices, maybe even generate its own spreadsheet to aggregate all the options, then make a judgment call on which flight to book and which hotel to reserve and actually do it. While an agent might struggle with novel tasks, for the most part it should be able to handle most of the routine work.
“You can translate that into a tax practice, where you have these complex workflows which a human breaks down into individual steps, each influencing the next: you take a document, extract the info from the document, put it into your accounting system, classify it, and do the booking in the system. All of this in theory agentic AI should be able to do for you,” said Finette.
Other accounting-specific applications he could envision include anything that has to do with data entry, reconciliation of accounts and classification of information in systems, as well as expense management, which he said is already semi-automated already.
“Right now it is semi-automated where you upload something into Expensify or something and it does image recognition on those expenses and pulls them in, but in the future it should do the report for me, there’s no reason why it should not take all this information and put the report together and submit it on my behalf,” he said.
However, Wood warned that agentic AI will still carry many of the risks that generative AI has today, especially the risk of making up information or being inconsistent with its outputs. While companies might promise a genie-like wish fulfillment, it will be especially important for people to understand the limitations of this technology.
“These systems interact, you wont always get a deterministic outcome like they think computers will generate, its not a calculator, so if you give an agent the ability to be creative, sometimes it might produce output A and sometimes produce output B and in accounting and business that can be a great strength, like in marketing, but when you do a tax form you don’t want that, you want income to be correct every single time. So the risk is that everyone gets hyped up and excited and applies it in the wrong place, you gotta use these tools and understand their strengths and weaknesses,” he said. “It’s sort of like gen AI right now, they think it will solve everything, but it solves these specific sets of issues and problems, so knowing where to use it and how to use it will be important.”
Finette agreed that the tendency of generative AI to make up information would still be a risk, and that there will likely be a lot of hype trying to minimize this risk as well. But he also noted that the fact that agents can actually act semi-autonomously and make decisions means the consequences of these risks can be bigger.
“In the flight booking use case, do you really trust the AI to actually book the flight for you? Will you be on the right flight at the right time? This is a silly example but a very real one,” he said. “The other [risk] is when you let AI make ‘moral’ decisions like letting AI do promotion decisions or suggesting out of 100 people here are the people who are top performers, where you get into issues like bias, which we know exists in AI. So all the issues we have with AI will be amplified with agentic AI.”
While theoretically these AI agents will be supervised by humans, Finette wondered about the degree to which people will actually do so, especially when AI can be so convincing in its reasoning even when wrong.
“These systems are so overly confident in their responses it is hard for some humans to step back and say don’t trust it. We all have experiences where you use ChatGPT and it tells you something wrong but they tell it to you in such a convincing way that if you didn’t have the knowledge you’d take it as gospel. … It is amplified if you let the system execute on this information. The human challenge is, and there are a bunch of research papers showing AIs are as convincing or even more so than humans, we need to get our workforce to understand that they should tread with caution and not let the AI bully you into a corner,” he said.
Credit: Source link