Digital disruption and climate change are the two fastest-growing risk areas for organizations across various industries, according to a new report.
The report, issued last week by the Institute of Internal Auditors’ Internal Audit Foundation’s the latest in its Risk in Focus research series. The report was based on feedback from over 3,500 internal audit leaders around the world, global risk levels for digital disruption and climate change are projected to increase 20 percentage points and 16 percentage points, respectively, over the next three years — far outpacing other risk areas.
The risks from climate change can be seen from the devastating impact of Hurricane Helene on the Southeast in recent days.
The IIA noted that despite the growing intensity of these risks, most audit plans don’t currently prioritize them. Neither digital disruption nor climate change were named among the top five areas where internal audit functions allocate the most time and effort, with both ranked in the lower half of audit priorities. Globally, internal audit functions focus predominantly on cybersecurity, governance/corporate reporting, and business continuity, indicating a perceived gap between evolving threats and current areas of attention.
“Our latest research tells us cybersecurity, business continuity and human capital continue to hold the top three spots in risk ratings,” said IIA president and CEO Anthony Pugliese in a statement last week. “However, respondents anticipate significant changes as risks related to climate change and digital disruption accelerate in the coming years. To ensure both short-term success and long-term sustainability, organizations and their internal audit functions must adapt risk management practices to keep pace with the changing risk landscape.”
Institute of Internal Auditors president and CEO Anthony Pugliese
Approximately 39% of the survey respondents worldwide ranked digital disruption as a top five risk, with that number expected to jump to 59% in three years. For North America, these figures were higher at 48% and 70%, respectively. Survey respondents worldwide anticipate digital disruption to rise from the fourth to the second highest ranked risk area in three years.
Artificial intelligence introduced new risks to watch, especially related to cybersecurity, according to 75% of the survey respondents. AI has also affected many other risk areas, including human capital, fraud, communications, reputation, and more. AI challenges for internal auditors include upskilling and adopting new tools, as well as global disparities in access to and knowledge of emerging technology.
Climate-related risks are currently ranked relatively low, but they are expected to rise substantially soon. About one in four (23%) of global respondents view climate change as a top five risk today. However, nearly 40% of respondents anticipate it will reach the top five in the next three years, climbing from 13th place to 5th.
Globally, roundtable participants agree that sustainability reporting and compliance requirements are the main drivers for boards, management and internal audit functions to allocate resources to climate change. The report revealed significant regional differences in climate-related risk perceptions. For instance, 33% of European audit leaders and 30% of Canadian audit leaders rate climate change as a top five risk, compared to 9% for the U.S. Despite the U.S. position, North American respondents expect ratings for climate change as a top 5 risk will double from 13% to 27% in three years.
“While climate change has long been recognized as a growing risk for organizations, these findings reveal the extent to which climate-related risks are expected to surge in the near term,” said Pugliese. “It is imperative for organizations, stakeholders, and internal audit leaders to objectively assess the short-term and longer-term risks to their organizations beyond basic compliance with regulations.”
Extreme weather can cause supply chain disruptions, higher operational costs, flooding, famine and more. While some consumers and investors are calling on organizations to implement more sustainability initiatives, these initiatives need to be reported accurately to avoid greenwashing and reputational damage.
States are looking beyond the 150-hour requirement for CPA licensure and adding alternative pathways amid the profession’s pervasive ongoing talent shortage.
Ohio and Virginia were the first two states to pass legislation establishing new pathways to licensure, with others following suit by introducing similar bills to their state legislatures, including Iowa this week. The bill language varies slightly by state, but one requirement that firmly remains is passing the Uniform CPA Examination.
See below for where things stand in those states that are moving forward on the issue, and read our feature here.
Registration is now open for this year’s Accounting MOVE Project, with the theme “The Cost of Losing Talent — Why Belonging is Essential for Firm Resilience and Growth.”
The Accounting MOVE Project is an annual survey of accounting and financial firms that measures demographic data and workplace culture to help advance women and other underrepresented groups in the profession. Firms can participate in the 2025 MOVE Project by registering on or after April 10. The survey opens May 7 and closes July 11. Research results and the 2025 Best CPA Firms for Women and Best CPA Firms for Equity Leadership will be announced at the annual Women Who Count conference in October and published online.
The key research areas are on how firms can retain talent and safeguard for long-term growth, why diverse and inclusive teams are essential for competitiveness and client service, how firm leaders are responding to political, legal and cultural pressures, and actionable strategies for building engaged and resilient workforces. This year, MOVE is partnering with the Center for Accounting Transformation.
“Our research shows that recruiting diverse talent isn’t enough; firms must invest in retention to truly thrive,” Donny Shimamoto, founder of the Center for Accounting Transformation and founder and managing director of IntrapriseTechKnowlogies, said in a statement. “Ignoring this issue isn’t an option for firms that want to remain competitive.”
“Partnering with the Center allows us to deepen our research and provide firms with concrete strategies to build sustainable, people-centered cultures,” said Bonnie Buol Ruszczyk, president of the Accounting MOVE Project.
The MOVE Project will also research its core areas of “money,” “opportunity,” “vital supports” and “entrepreneurship.” Firms can join MOVE by either participating in the annual MOVE survey for no fee, or by upgrading to receive a firm-specific report.
As AI works its way into more and more business processes, it has become increasingly important for auditors to understand where, why, when and how organizations use it and what impact it is having not only on the entity itself but its various stakeholders as well.
Speaking at a virtual conference on AI and finance hosted by Financial Executives International, Ryan Hittner, an audit and assurance principal with Big Four firm Deloitte, noted that since the technology is still relatively new it has not yet had time to significantly impact the audit process. However, given AI’s rapid rate of development and adoption throughout the economy, he expects this will change soon, and it won’t be long before auditors are routinely examining AI systems as a natural part of the engagement. As auditors are preparing for this future, he recommended that companies do as well.
“We expect lots of AI tools to inject themselves into multiple areas. We think most companies should be getting ready for this. If you’re using AI and doing it in a way where no one is aware it is being used, or without controls on top of it, I think there is some risk for audits, both internal and external,” he said.
Elevated View Of Robotic Hand Examining Financial Data With Magnifying Glass
Andrey Popov/stock.adobe.com
There are several risks that are especially relevant to the audit process. The primary risk, he said, is accuracy. While models are improving in this area, they still have the tendency to make things up, which might be fine for creative writing but terrible for financial data reporting. Second, AI tends to lack transparency, which is especially problematic for auditors, as their decision making process is often opaque, so unlike a human, an AI may not necessarily be able to explain why it classified an invoice a particular way, or how it decided on this specific chart of accounts for that invoice. Finally, there is the fact that AI can be unpredictable. Auditors, he said, are used to processes with consistent steps and consistent results that can be reviewed and tested; AI, however, can produce wildly inconsistent outputs even from the same prompt, making it difficult to test.
This does not mean auditors are helpless, but that they need to adjust their approach. Hittner said that an auditor will likely need to consider the impact of AI on the entity and its internal controls over financial reporting; assess the impact of AI on their risk assessment procedures; consider an entity’s use of AI when identifying relevant controls and AI technologies or applications; and assess the impact of AI on their audit response.
In order to best assist auditors evaluating AI, management should be able to answer relevant questions when it comes to their AI systems. Hittner said auditors might want to know how the entity assesses the appropriate of AI for the intended purpose, what governance controls are in place around the use of AI, how the entity measures and monitors AI performance metrics, whether or how often they backtest the AI system, and what is the level of human oversight over the model and what approach does the entity take for overriding outputs when necessary.
“Management should really be able to answer these kinds of questions,” he said, adding that one of the biggest questions an auditor might ask is “how did the organization get comfortable with the result of what is coming out of this box. Is it a low risk area with lots of review levels? … How do you measure the risk and how do you measure whether something is acceptable for use or not, and what is your threshold? If it’s 100% accurate, that’s pretty good, but no backtesting, no understanding of performance would give auditors pause.”
He also said that it’s important that organizations be transparent about their AI use not just with auditors but stakeholders as well. He said cases are already starting to appear where people unaware that generative AI was producing the information they were reviewing.
Morgan Dove, a Deloitte senior manager within the AI & Algorithmic Assurance practice, stressed the importance of human review and oversight of AI systems, as well as documenting how that oversight works for auditors. When should there be human review? Anywhere in the AI lifecycle, according to Dove.
“Even the most powerful AIs can make mistakes, which is why human review is essential for accuracy and reliability. Depending on use case and model, human review may be incorporated in any stage of the AI lifecycle, starting with data processing and feature selection to development and training, validation and testing, to ongoing use,” she said.
But how does one perform this oversight? Dove said data control is a big part of it, as the quality and accuracy of a model hinges on its data stores. Organizations need to verify the quality, completeness, relevance and accuracy of any data they put into an AI, not just the training data but also what is fed into the AI in its day to day functions.
She also said that organizations need to archive the inputs and outputs of their AI models, without this documentation it becomes very difficult for auditors to review the system because it allows them to trace the inputs to the outputs to test consistency and reliability. When archiving data she said organizations should include details like the name and title of the dataset, and its source. They should also document the prompts fed into the system, with timestamps, so they can possibly be linked with related outputs.
Dove added that effective change management is also essential, as even little changes in model behaviors can create large variations in performance and outputs. It is therefore important to document any changes to the model, along with the rationale for the change, the expected impact and the results of testing, all of which supports a robust audit trail. She said this should be done regardless of whether the organization is using its own proprietary models or a third party vendor model.
“There are maybe two nuances. One is, as you know, vendor solutions are proprietary so that contributes to the black box lack of transparency, and consequently does not provide users with the appropriate visibility … into the testing and how the given model makes decisions. So organizations may need to arrange for additional oversight in outputs made by the AI system in question. The second point is around the integration and adoption of a chosen solution, they need to figure out how they process data from existing systems, they also need to devote necessary resources to train personnel in using the solution and making sure there’s controls at the input and output levels as well as pertinent data integration points,” she said.
When monitoring an AI, what exactly should people be looking for? Dove said people have already developed many different metrics for AI performance. Some include what’s called a SemScore, which measures how similar the meaning of the generated text is to the reference text, BLEU (bilingual evaluation understudy), which measures how many words or phrases in the generated text match the reference text, or ROC-AUC (Receiver Operating Characteristic Area Under the Curve) which measures the overall ability of an AI model to distinguish between positive and negative classes.
Mark Hughes, an audit and assurance consultant with Deloitte, added that humans can also monitor the Character Error Rate, which measures the exact accuracy of an output down to the character (important for processes like calculating the exact dollar amount of an invoice), Word Error Rate, which is similar but does the evaluation at the word level, and the “Levenshtein distance,” defined as the number of single character edits needed to fix an extracted text to see how far away the output is from the ground truth text.
Hittner said that even if an organization is only just experimenting with AI now, it is critical to understand where AI is used, what tools the finance and accounting function have at their disposal to use, and how it will impact the financial statement process.
“Are they just drafting emails, or are they drafting actual parts of the financial statements or management estimates or [are] replacing a control? All these are questions we have to think about,” he said.