AI implementation at a CAS practice is hard enough, but it becomes even more so when people don’t fully understand what AI can and cannot do.
Speaking during the Information Technology Alliance’s spring collaborative in Memphis, Tennessee, Jessica Barnas, the partner leading the finance and accounting solutions advisory group for top 25 firm Wipfli, lamented that public discourse around AI has given people the impression it’s some sort of magic wand that can fix anything, which then leads to unrealistic expectations around its capabilities.
“I talked to a lot of clients, I think they think that AI is like an elf that jumps out of the box and does things magically. They just say, ‘Can’t AI do that?’ I even had one of our partners [tell me this recently], we’re working on a five year revenue prediction—he said, ‘Well, can’t you just upload that to Copilot and have it spin up the business plan and everything?’ and I’m like, ‘Do you have any idea how generative AI works? It doesn’t do that.’ But I think that there’s just this misconception [that], oh, technology it is just this magic wand that’s going to make all of my accounting problems disappear,” Barnas said.
Chris Gallo, director of outsourced business accounting services with Kansas-based firm Creative Planning and another one of the panelists, made a similar point, saying that it’s important to be realistic about what technology can do. While it can do a lot, he echoed Barnas in saying that some people seem to think it is magic.
“If we believed everything that everybody told us you would be flying around in flying cars right now. I think we need to kind of take it with a grain of salt at some point. Because why wouldn’t we just say ‘ChatGPT build me a flying car,’ and then the bot people that you know Tesla’s building will just go do that. Right? It becomes a little bit ridiculous at some point too… There’s a lot of expectation, or unaligned expectations,” said Gallo.
Misconceptions about AI capabilities also serve to drive fear on the part of accountants. Barnas said that a big part of the change management process when it comes to implementing AI is allaying fears from staff that they’re not going to fire everyone and replace them with bots. While there have been major improvements in AI over the years, she does not believe it is in the position to wholesale replace human accountants just yet. Instead, it has become a great way to augment those humans and make them more competitive against the humans who are not using AI.
“They think ‘AI will eliminate my job!’ So we talk about our philosophy. We’re looking to adopt these tools to help you get bigger and better and embrace the advisory role, but the only way AI will replace you is if a person using AI will replace you. You need to give that level of comfort to your teams so that everyone knows we’re just trying to get better, we’re just picking up new tools, this is not a replacement for you,” Barnas said.
There is a similar fear when it comes to billable hours, also explored in another panel (see other story), of what happens when a process that normally takes 8 hours now only takes 1. Barnas first described the billable hour as “the enemy of all of us here in the room” but also conceded it is a real anxiety for practices that have built their foundation on it. She suggested, in response to this concern, to take a page from Google and encourage people to develop pet projects using AI and rewarding them if it turns into something useful for the entire team; and if it really does lead to a reduction in billable hours, don’t punish people with less money when they’ve done what you wanted them do in the first place. Overall, a firm’s business model should not be one that punishes efficiency: a practice should value results, not burning hours. She conceded that, for certain firms set in their ways, this might need retraining.
“Okay, I took this process down from seven hours to half hour every week. Now what? Teach me how to do advisory. Because being a CFO, doing modeling and projections, it is not something [you learn] from reading a book or sitting in on one webinar. We would all be doing that if that were the case. So how can we train our teams on what to do next? All of that is involved in change management: being a guide and providing the safety for each step,” she said.
Gregg Landers, the last panelist and managing director of client accounting and advisory services and internal control services with Top 10 firm CBIZ, talked about how a lot of the misunderstandings and misconceptions regarding AI can be allayed from people just experimenting with it themselves, which not only lets them get a better impression of its current capabilities but will train them in using those capabilities to their fullest potential.
“I’ve been encouraging some of my teams to use their personal generative AI a little Black Mirror-like, [where you] keep talking to it, and it talks back. You get accustomed to how to give a context, how to get better answers. Sometimes, if you’re nice to it, [you get] a tighter answer than if you’re not. So experiment around with it.
He gave an example from his own life, where he needed to learn more about digital services taxes. Through an extended conversation with an LLM he was able to understand what the DST is and how it works and how accountants manage it. He was able to get good outputs from the model, though, because previous experience taught him that he needs to provide more context and information for a decent answer, because these models can get tripped up by ambiguities. He compared it to a fortune cookie that could be interrupted in many ways, people should be clear and concise when prompting AIs.
“We’ve become a society of fortune cookies. I may ask ‘how is that project going’ and you tell me ‘it’s going good’ but what I mean is ‘is it on time?’ and what you might mean is ‘I had this hiccup that put me two weeks behind but now it is resolved so it is good.’ We can’t have fortune cookies when interacting with generative AI. You need clear, concise, contextual communication,” he said.