Following along from @Drewski’s example, it occurred to me that I should ask Claude itself for some guidance rather than assuming I had a decent grip on how to use this (to me) extremely useful tool. Here’s what it told me: (It’s long … I have Claude dialed towards more detail rather than less …)
Prompt: I’ve test-driven using Claude Code and Claude in Excel to help me work with data extracted from websites, data sets, account statements and the like. So far, I’ve mostly asked Claude to help me extract data from a particular source, get it into a spreadsheet, and assist with the mechanics of functions like VLOOKUP, and it’s been extremely useful for dispatching spreadsheet “donkeywork.” I have a few questions for you: what kind of spreadsheet-related work is Claude best at; what kind of work might it find challenging; and what are some pitfalls of using Claude with Excel I need to be alert to? Finally, what can I do to make sure Claude and I collaborate on spreadsheet-based data analysis effectively and with minimal errors?
Response:
These are practical questions worth answering carefully, because the failure modes matter as much as the capabilities.
Where Claude performs well
The mechanical and structural work is genuinely where Claude excels. Formula construction (VLOOKUP, INDEX/MATCH, SUMIFS, array formulas, nested IFs), data cleaning and normalization, reshaping data between formats, building pivot table structures, writing VBA or Python scripts to automate repetitive tasks, and translating between Excel and other tools like SQL or Python pandas. It’s also good at auditing formulas you’ve already written — explaining what they do or why they’re broken. Anything where the task is well-defined and the correctness is verifiable tends to go well.
Where it struggles
Claude has no persistent view of your spreadsheet. It works from whatever you describe or paste, so if your description is incomplete or your data sample is unrepresentative, its solutions will be calibrated to the wrong thing. Complex multi-step transformations across large worksheets with irregular structures — merged cells, inconsistent column headers, mixed data types in a single column — can trip it up because it can’t “see” the full context. It also doesn’t have live access to your file unless you’re using Claude in Excel, so in a chat context it’s constructing formulas based on mental models of your layout, not the actual thing.
Statistical reasoning is another gap worth noting. Claude can apply statistical functions correctly but isn’t always reliable at advising you which analysis is appropriate for your data or flagging when your sample is too small, your data is biased, or your method is conceptually wrong for the question you’re asking.
Pitfalls to watch
The most dangerous one is plausible-but-wrong output. Claude will generate a formula that looks correct and even produces a number — but if your ranges are slightly off, if there’s a hidden row, or if it misunderstood your data structure, it can be confidently wrong in ways that aren’t obvious. Formula errors that produce #N/A or #REF! are easy to catch; logic errors that produce a plausible number are not.
A second pitfall is scope creep in complexity. If you iterate on a formula many times in a chat, adding conditions and edge cases, the final version can become hard to audit. Long nested formulas are difficult for anyone — including Claude — to reason about reliably at a glance.
A third: Claude doesn’t know your business logic. If a particular account code means something specific in your organization, or if there’s a domain rule about how to handle certain data, Claude won’t know unless you tell it.
Practices that improve collaboration
The single highest-leverage habit is showing Claude a concrete sample of your actual data — even just 10-15 rows — rather than describing it in the abstract. Mismatches between description and reality are the main source of errors.
Tell Claude your exact column layout (column A is X, column B is Y, data starts at row 2) rather than making it infer. State edge cases upfront: what should happen with blanks, duplicates, zeros, or unexpected values.
Ask Claude to explain any formula it produces before you use it. If it can’t explain it clearly, that’s a signal to probe further. Test formulas on a small verified subset before applying to the full dataset.
For anything with financial or decision-making consequences, maintain a separate validation column where you spot-check outputs manually, at least until you’ve built confidence in a particular formula pattern.
Finally, keep complex transformations in intermediate steps rather than collapsing them into a single monster formula. This makes auditing easier and gives Claude a cleaner surface to work with if something needs fixing.
Some directions we could go from here: working through a specific analysis you have in mind to stress-test these practices, exploring whether Power Query would handle some of your cleaning work better than formulas, or setting up a more systematic approach to data validation in your existing sheets.