For the last year or two, every CFO conversation eventually drifts into AI. Sometimes it’s framed as excitement, sometimes as anxiety, and sometimes as an awkward silence followed by, “Well, we’re looking at it.” What’s striking is that most of the tension around AI in finance has very little to do with the technology itself. The models work. The tools are improving fast. The vendors all have slick demos.
The real issue is governance.
Finance teams are wired around controls, auditability, and repeatability. AI systems, by contrast, are probabilistic, opaque, and constantly evolving. That mismatch is where most CFO discomfort comes from — and it’s why “let’s just automate this” often stalls once it hits a real finance process.
The first mistake I see is treating AI like just another system implementation. ERP projects taught us how painful that mindset can be. AI requires a different framing: not “what can this tool do?” but “what decisions are we willing to delegate, and under what constraints?” That sounds abstract. It isn’t.
Over the past year I’ve pushed AI tools on real finance questions: revenue recognition edge cases, SEC disclosure interpretations, covenant calculations, and technical accounting memos. The patterns that show up are not technology failures. They are governance failures waiting to happen.
1. AI doesn’t fight back.
If you have ever debated an accounting position with a strong controller or technical accounting lead, you know what conviction feels like. You push. They push back. You test assumptions. They defend them with chapter and verse. That friction is healthy. Same thing for a forecast analysis. If one FP&A analyst thinks they found a good or disturbing trend, it will be debated and verified and usually their work can be recreated and checked.
AI does not behave that way.
If you tell it, “I think you’re wrong,” it often apologizes and produces a different answer. Sometimes an entirely opposite answer. The confidence level remains high. The tone remains polished. The data is processed inside the model, and the AI often struggles to explain — or even remain consistent in — its answers.
In a live finance organization, that would be a red flag. If a manager flipped their view that quickly under mild pressure, you would question the depth of analysis. With AI, the flip can look like responsiveness rather than fragility.
That is a governance issue. It means you cannot treat an AI output as a position that has survived adversarial testing. It hasn’t. It has survived prompt engineering. And the prompt may have been poor.
2. The praise problem.
Most AI agents are relentlessly deferential. “Great question.” “Excellent point.” “You’re absolutely right to focus on that.” In a consumer context, that feels pleasant. In a finance context, it is dangerous.
Finance works because of tension — between risk and growth, between conservatism and disclosure clarity, between what management wants and what GAAP allows. When the “advisor” in the room is constantly affirming the user, it subtly reinforces bias.
I’ve seen this firsthand when asking an AI to pressure-test a disclosure approach. Rather than aggressively identifying weaknesses, it often validates the framing of the question. The tone can make a marginal position sound well-supported. In other words, the user’s confidence can rise faster than the quality of the analysis.
Governance must assume that AI will not naturally challenge you the way a seasoned audit partner or skeptical board member will.
3. The citation illusion.
This one should make every CFO uncomfortable.
Ask an AI to provide citations to accounting guidance or SEC commentary, and it will often comply — confidently. Paragraph numbers. Codification references. Even plausible-sounding excerpts.
The problem is that some of them are fabricated. They look right. They read right. They are formatted correctly. But they do not exist.
In finance, citations are not decorative. They are the backbone of defensibility. When you write a technical memo on revenue recognition or stock-based compensation, the citation is the bridge between your judgment and the authoritative literature.
If an AI invents that bridge, and a team relies on it without independent verification, the failure is not the model’s. It is the control environment’s. Any AI-assisted accounting memo must include a verification step where a human independently confirms the authoritative source. Not “glances at it.” Confirms it.
4. Rule changes and historical drift.
Accounting rules change. Constantly.
Revenue recognition under ASC 606 replaced a patchwork of legacy guidance. Lease accounting under ASC 842 upended decades of practice. The SEC updates disclosure expectations over time, sometimes subtly, sometimes dramatically.
Meanwhile, the SEC’s EDGAR archive goes back decades. There are scanned paper filings from eras when the rules were materially different. There are thousands of examples built under superseded guidance.
AI models trained on broad corpuses struggle here. They can blend old and new regimes. They can cite legacy practice as if it were current. They can rely heavily on the abundance of historical examples rather than the correctness of modern policy.
I have seen AI answers that lean on pre-606 revenue language as though nothing changed. Or that reference lease accounting concepts that no longer apply post-842. To a non-expert, the answer looks sophisticated. To someone who lived through the transition, the seams are obvious.
Governance means you assume the model does not instinctively know the effective date of your accounting framework. You have to constrain it.
5. Finance is not plain English.
Financial reporting language is precise. “Probable” does not mean “likely” in a colloquial sense. “Material” is not a synonym for “important.” “Reasonably possible” has a defined meaning.
AI systems are trained on massive volumes of plain English. That is a strength in many domains. In accounting, it can be a weakness.
I’ve seen answers where the model drifts into narrative explanations that sound sensible but subtly misapply defined terms. In a board deck, that might pass. In a 10-K, that is a problem.
When language itself carries regulatory weight, small deviations matter.
So what does governance look like in practice?
It is not banning AI. That is neither realistic nor wise. The productivity gains are real. Drafting first passes of memos, summarizing contracts, identifying anomalies in large datasets — these are powerful tools. AI can be properly trained on your data and become more accurate. Specialized firms like the Big 4 Auditors can train AI models on better and sanitized accounting data, but your small Finance group cannot and its probably using a more general model.
But they must sit inside a control framework.
At a minimum:
- AI outputs that influence external reporting require documented human review.
- AI conclusions about trends must be independently tested and verified. Don’t order another $1M of a part because a model suggested it.
- Authoritative citations must be independently verified.
- Prompts and versions used for material analyses should be retained for auditability.
- Use cases must be categorized: drafting support is different from judgment replacement.
- Responsibility for the final position must be clearly assigned to a human owner.
Most importantly, the CFO has to set the tone.
Let me make a direct observation: most leadership team members are not finance experts, but AI can create the illusion that they are. You need to make sure they understand the risk.
If AI is positioned as an infallible oracle, teams will over-rely on it. If it is positioned as a junior analyst — fast, helpful, occasionally wrong, and requiring supervision — behavior adjusts appropriately.
The question is not whether AI will be used in finance. It already is.
The question is whether it will be used inside a governance framework that protects credibility.
Investors do not care how you produced your numbers. Auditors do not care how you drafted your memo. Regulators certainly do not care that a model was “usually right.” They care that your disclosures are accurate, supportable, and controlled.
AI in finance is not a technology problem. It is a governance problem. And like most governance problems, it lands squarely on the CFO’s desk.
I don’t want to sound like Cassandra warning of inevitable doom. Nor do I want to be the boy who cried wolf while your competitor quietly figures this out and gains an advantage.
In future posts, I will outline where I believe AI can genuinely add value inside a disciplined finance organization.
