CurateSuite
Framework10 min read

What AI Can and Cannot Do in Your Accounting Practice Today

An honest look at where AI saves real time in accounting firms and where it falls short. For every claim that works, a matched limitation that vendors will not mention first.

By CurateSuite
A person at a desk leaning forward with one hand on their forehead, staring at a laptop screen with a frustrated expression, warm natural daylight from the right

AI is genuinely useful in accounting right now for three things: pulling data out of documents, matching transactions to the right accounts, and flagging patterns a human would miss in a stack of working papers. Those three tasks are high-volume, repetitive, and follow predictable rules, which is exactly the shape of problem that machine learning handles well. Outside that shape, AI in accounting is unreliable, overpromised, or simply not ready. The client conversation, the materiality call, the instinct that something in the numbers feels wrong before you can say exactly what: those are still human work, and no product on the market today changes that.

This article walks through both sides. For every area where AI tools deliver measurable time savings, there is a matched area where the technology falls short or where vendor marketing stretches further than the product does. The goal is not to talk you out of buying anything. It is to help you buy with accurate expectations.

What AI does well right now

Three categories of accounting work produce consistent, real time savings with current AI tools. The savings are measurable in hours per client per month, not in vague "efficiency gains."

Document capture and data extraction. This is where the strongest results sit. Dext reads receipts, invoices, and bank statements and posts them into the ledger as coded transactions. The model learns on a per-supplier basis, so accuracy improves with volume. For a firm with twenty or more bookkeeping clients, Dext or a comparable capture tool typically saves one to two hours per client per month in manual keying. The AI is not doing anything conceptually complex here. It is reading structured documents and mapping fields. That is exactly why it works reliably: the task is narrow, the patterns are consistent, and the consequences of a wrong guess are caught in the next review cycle.

Transaction matching and approval routing. BILL and Stampli use AI to match incoming invoices against purchase orders, route them for approval, and flag exceptions. BILL works from the approval chain outward: invoices enter, get matched, get routed, get paid. Stampli works from the invoice inward: each invoice becomes a conversation thread with context, suggested coding, and approval routing attached. Both approaches save time because the volume of invoices in a mid-sized firm is high enough that manual matching and routing is a full day of someone's week. The AI handles the 80 percent of invoices that follow predictable patterns and surfaces the 20 percent that need human attention.

Pattern-based review and cross-referencing. DataSnipper sits inside Excel and cross-references working-paper cells against source documents. It flags mismatches, missing references, and inconsistencies that a human reviewer would need twenty minutes per client to catch manually. Karbon automates recurring workflow patterns: if a client's bank statements arrive, trigger the reconciliation task; if a document request goes unanswered for five days, send the follow-up. These are not deep reasoning tasks. They are pattern recognition applied to a consistent data environment, and that is the sweet spot for current AI.

The common thread across all three: the AI is doing something narrow, repetitive, and pattern-based on structured or semi-structured data. When a task fits that description, current tools deliver. When it does not, they struggle.

What AI does badly or not at all

Three categories of accounting work resist AI automation, and buying tools that promise to solve them usually leads to disappointment.

Professional judgement. A $400 miscoding might matter this month or it might not, depending on context the tool does not have. Whether to capitalize an expense, how to treat an unusual related-party transaction, when to flag something to a client versus fixing it silently: these are judgement calls shaped by experience, client history, and professional standards. No AI tool on the market today can make them. Some tools flag potential issues (DataSnipper flags mismatches, Karbon surfaces overdue items), but the decision about what to do with that flag is entirely human.

Context that lives outside the data. Why did revenue drop 30 percent this month? Is it a seasonal pattern, a lost contract, or a data-entry error? The answer usually comes from a phone call, not from the ledger. AI tools see the numbers. They do not see the conversation with the client last Tuesday where the client mentioned losing a major customer. They do not know that the bookkeeper coded three invoices to the wrong period because the client changed their billing cycle without telling anyone. The gap between what the data shows and what actually happened is where most of the hard accounting work sits, and AI cannot bridge it.

Client relationships. The check-in call about cash flow. The difficult conversation about late filings. The trust that keeps a client from shopping around during busy season. These interactions are the reason clients stay with a firm, and no AI product replaces them. Some tools (Karbon, TaxDome) automate the scheduling and tracking of client communication, which is useful. But automating the logistics of a conversation is not the same as automating the conversation itself.

The claims that are technically true but misleading

Vendor marketing in AI accounting tends to stretch in predictable ways. Recognizing the patterns helps you read product pages without either dismissing them or taking them at face value.

"AI-powered." This label covers everything from a genuine machine-learning model trained on millions of invoices to a set of if-then rules with a chatbot wrapper. When a vendor says "AI-powered," the useful follow-up question is: what specifically does the AI do, and what would change if you turned it off? If the answer is "the product would still work but you would need to code transactions manually," the AI is doing something real. If the answer is vague, the label is probably decoration.

"Automates your workflow." Partially true, often overstated. Karbon automates recurring task sequences and client communication triggers. That is workflow automation. But the phrase implies the entire workflow runs without human input, which it does not. The AI handles the scheduling, assignment, and reminders. The actual work, preparing the return, reviewing the reconciliation, writing the advisory note, still requires a person. "Automates parts of the steps around your workflow" is more accurate. It is also less appealing on a landing page.

"Saves X hours per week." Usually an average across a customer base that includes firms much larger than yours. A tool that saves a fifty-person firm ten hours a week may save a five-person firm forty-five minutes. The per-firm savings depend on volume, on how manual the firm's current process is, and on how well the tool integrates with the existing stack. When a vendor quotes a time-savings number, ask: at what firm size, on what ledger, and compared to what baseline? If they cannot answer those questions, the number is marketing, not measurement.

Where AI is heading versus where it is today

Blue J is the clearest example of what the next generation of AI in accounting looks like. It applies natural language processing to tax research: a user describes a fact pattern, and the system returns relevant case law, rulings, and risk assessments with confidence scores. This is closer to reasoning than pattern matching. It is also limited to tax research in specific jurisdictions, expensive, and aimed at firms with a dedicated tax research function.

The gap between Blue J and the document-capture tools is worth noticing. Most AI in accounting today is pattern-based: find the date, match the supplier, flag the mismatch. Blue J is working on a harder problem: interpret the question, find the relevant authority, and assess the risk. That harder problem will eventually reach more areas of accounting, but "eventually" is not "this year," and buying tools based on a vendor's two-year roadmap rather than their current product is how firms end up in the 44 percent that regret the purchase.

Buy for what the tool does today. If the roadmap features arrive and they work, that is a bonus. If the roadmap is the main reason you are buying, wait.

A practical way to think about it

The simplest mental model: divide your firm's work into two columns.

Column one is pattern-based and high-volume. Receipt coding, bank categorization, invoice matching, working-paper cross-referencing, recurring task scheduling. AI handles this column well today. Current tools deliver measurable savings here, and the technology is mature enough that the savings are reliable month to month.

Column two is judgement-based and context-dependent. Materiality calls, unusual transactions, client conversations, advisory recommendations, the instinct that something does not add up. AI does not handle this column, and products that claim to are overstating their capabilities.

The firms that get the most from AI tools are the ones that buy for column one and protect column two. They automate the data handling so their people have more time for the judgement work, rather than trying to automate both.

The firms that waste money on AI tools are usually trying to buy their way out of column two, or assuming that a tool marketed as "AI-powered advisory" will eventually replace the advisory work itself. It will not. Not this year, probably not next year, and the vendors who imply otherwise are selling a roadmap.

What to do with this

If you have read this far and decided that AI is worth trying for a specific part of your practice, the next step is to narrow the field to a shortlist of two or three tools that fit your bottleneck, your ledger, and your budget. The CurateSuite matchmaker does a version of that in about a minute.

If you have read this far and decided to wait, that is a reasonable position too. The tools will still be there in six months, and they will probably be better.

Find the right AI tools for your firm

Six questions, one minute, a short list of the five tools that fit your firm best. Your results, instantly. No sign-up needed.

Take the matchmaker

Tools referenced in this article

More articles

Last updated 2026-05-11. Tool comparisons are based on vendor-published specs. See our methodology.