Skip to content
Prompt Stack

5 AI prompts for earnings call analysis: what the numbers don't tell you

Five AI prompts for earnings call analysis that find what the numbers don't: omissions, hedge patterns, and language drift from last quarter.

// TL;DR

Problem: every AI earnings post covers data extraction — numbers, ratios, headlines. Fix: language analysis is the harder half — what management didn't say, hedged, or reframed from last quarter. Payoff: five structured prompts that find it.

// On this page

Most AI earnings content stops at extraction. The numbers, the headlines, the ratios — every general-purpose model handles that adequately, and the comparison piece on the best AI tools for earnings analysis sorts out which one to use. These five AI prompts for earnings call analysis cover the harder half: language, not data.

The harder half is reading the language. Not what management said — every transcript service summarises that. What they committed to, what they hedged, what they conspicuously didn’t address, and how their framing has drifted since last quarter. I held META into the Q1 2026 print. Revenue came in at $56.3bn, up 33% year on year. The headline was strong. The capex guidance was $125–145bn for the full year, given without a payback timeline. The stock fell roughly 8% in the session that followed.

The numbers weren't the problem; the omission was.

What these prompts cover: omission detection (Prompt 1), confidence calibration (Prompt 2), gap to consensus (Prompt 3), quarter-on-quarter language shift (Prompt 4), and the hostile analyst question you wish someone had asked (Prompt 5). Each runs against pasted text — no live data access required.

The discipline for earnings call analysis

Every prompt below assumes you have copied the management commentary into your AI session. The model does not have access to the live filing, and any prompt that lets it pretend otherwise produces confident fiction. I learned that the hard way the first time I asked Perplexity to “analyse the last META call” without pasting anything. Paste the text. Then ask the question.

All five prompts follow the Prompt Stack — ROLE, FILTER, RISK, VERDICT — for the reasons every other post on this site uses it: ROLE keeps the model out of cheerleader mode, FILTER forces real text in, RISK is the bit retail readers skip, VERDICT prevents an “it depends” non-answer.


Prompt 1 — What did they not say?

Before reading what management said, list what a careful investor would expect to hear addressed. Then check whether each item was covered, hedged, or absent. Management can’t hide what they don’t mention, but a casual reader notices presences, not absences. I ran this on the META Q1 2026 commentary: the model flagged quantified AI ROI as the most conspicuous absence — Zuckerberg used the word “return” without producing one while adding $10bn to an already-record capex line — and marked tariff and supply-chain exposure ABSENT in a quarter where the capex revision was pinned on “higher component costs”.

// Prompt 1 — Conspicuous omissions

ROLE: Act as a sceptical analyst preparing to read [COMPANY NAME]‘s [QUARTER] results. You have not seen the transcript yet. Predict what should be in it, then check what is.

FILTER: Based on what you know about [COMPANY NAME]‘s business, the prior quarter’s commentary, and the sector context, list four to six topics or metrics that a careful investor would most expect management to address this quarter. Be specific — “margin trajectory in segment X” not “performance”.

I will then paste the management commentary below. For each of your four to six topics, mark it as ADDRESSED (with a specific number or commitment), MENTIONED (raised but without specificity), or ABSENT (not raised at all).

[PASTE COMMENTARY HERE]

RISK: For any absent topic, suggest one reason management might have chosen not to address it. Distinguish “we don’t have news” from “we have news we’d rather not lead with”.

VERDICT: Name the single most conspicuous absence and one sentence on whether it changes how the headline numbers should read.


Prompt 2 — Genuine versus performative confidence

Confident-sounding language splits into two categories. Genuine confidence has at least one specific number, timeline, or falsifiable commitment. Performative confidence has none. Both sound similar in a transcript; only one is a commitment. AI is good at this specific task — there is a clean linguistic signature on each side. I ran this on Apple’s most recent services commentary: List A came in at 13 statements, List B at 11. Strip out the backward-looking numbers any CFO has to report and the forward-looking content is dominated by List B — management committed to exactly two specific forward-looking things, next-quarter services growth around 13% ex-FX and Apple Maps ads launching this summer in the US and Canada. Everything else about the future was mood music.

// Prompt 2 — Confidence calibration

ROLE: Act as a language analyst, not a financial analyst. You are assessing whether the language reflects genuine confidence or performative confidence, not whether the business is doing well.

FILTER: Below is the management commentary section from [COMPANY NAME]‘s [QUARTER] results. Produce two lists.

LIST A — GENUINE CONFIDENCE: statements with at least one of a specific number, a specific timeline, a falsifiable commitment, or a comparison to a prior stated target. Example structure: “We delivered X% growth in segment Y, ahead of the Z% guidance given in [prior period].”

LIST B — PERFORMATIVE CONFIDENCE: statements that sound confident but contain no specific commitment, no timeline, and no falsifiable claim. Example structure: “We remain well-positioned to capitalise on the opportunities ahead.”

[PASTE COMMENTARY HERE]

Count both lists. If List B is materially longer than List A, name the topics where confidence is performative — those are the topics management is uncertain about and is choosing not to say so.

VERDICT: One sentence — does this commentary commit management to anything specific, or does it preserve their optionality?


Prompt 3 — Commentary against consensus

The gap between management's language and consensus's mid-point is where stock reactions live.

A reader who doesn’t track consensus loses the magnitude of any surprise — they hear “increased investment level” and miss that it means the top end of a guide sits well above what analysts modelled. Give the AI the consensus number and the language reads differently. I was holding META into the Q1 2026 print: FactSet consensus going in was around $122.6bn of full-year capex. The $125–145bn band came in above that, and the high end was the part the market punished. The language framed the raise as “investment in AI capability” without a payback horizon — the gap to consensus was the thing the prompt could have surfaced before the stock moved.

// Prompt 3 — Consensus vs. reality

ROLE: Act as a sell-side analyst checking whether your model assumptions need to change after this commentary.

FILTER: Before reading the commentary below, here is the consensus picture for [COMPANY NAME] going into this print:

  • Consensus revenue [PERIOD]: [VALUE]
  • Consensus margin: [VALUE]%
  • Consensus capex / opex / [KEY LINE]: [VALUE]
  • Consensus guidance midpoint: [VALUE] Source: [BLOOMBERG / VISIBLE ALPHA / IBES / etc.] as of [DATE].

[PASTE MANAGEMENT COMMENTARY HERE]

For each consensus assumption, identify any phrase in the commentary that supports it, contradicts it, or implies a different number. Quote the phrase. Quantify the implied gap where you can.

RISK: Identify the single consensus assumption most likely to need a revision after this commentary, and the direction.

VERDICT: One sentence — does the commentary read as ahead of, in line with, or behind consensus on the line item that matters most for the stock?


Prompt 4 — Language shift since last quarter

This is the move only AI can do at speed.

A human reader can hold one prior transcript in their head; the model can hold four.

Same activity, different framing usually signals something — confidence rising or falling, with the language as the early tell. BMNR's framing around its ETH buying programme shifted between Q4 2025 and Q1 2026. What read as aggressive accumulation in the earlier call became something closer to disciplined positioning by the next one, around the time the buying pace was publicly described as moderating toward the 5% supply target. Same programme, different sentence. I have been reading every BMNR release for this kind of shift since the NYSE uplisting.
// Prompt 4 — Prior call comparison

ROLE: Act as a language pattern analyst. You will compare how [COMPANY NAME] management framed a specific topic across two consecutive quarters.

FILTER: Topic to compare: [TOPIC — e.g. “margin trajectory in segment X” / “AI capex strategy” / “subscriber growth”].

[PRIOR QUARTER LABEL — e.g. “Q4 2025 commentary”]: [PASTE PRIOR QUARTER LANGUAGE ON THE TOPIC]

[CURRENT QUARTER LABEL — e.g. “Q1 2026 commentary”]: [PASTE CURRENT QUARTER LANGUAGE ON THE TOPIC]

Produce a side-by-side analysis covering:

  • Specificity: did the language become more or less specific?
  • Time horizon: did the timeline shorten, lengthen, or vanish?
  • Commitment level: hedge words added or removed?
  • Frame: did the topic shift from achievement (past) to intention (future), or the other way?

RISK: Name the single most material shift and one observable signal in the next quarter that would confirm whether the shift reflects a real change in trajectory or just a different draft of the same script.

VERDICT: Rising confidence, falling confidence, or essentially unchanged framing? One sentence.


Prompt 5 — The hostile analyst question

The hardest question management didn’t have to answer on the call — usually because the analysts weren’t aggressive enough, or because the question got politely deflected. That’s the question worth holding in your head before deciding what to do with the position. I ran this on Apple’s most recent services commentary: the model identified the Google search licensing payment — reportedly around $20bn a year, flowing largely to gross profit — as the structural question, the single number that most explains why services gross margin sits at 76.7% and the one management won’t quantify on the call. Predicted response from the CFO: total non-disclosure on Google economics. Which is itself the tell.

// Prompt 5 — Hostile analyst

ROLE: Act as the most sceptical short-side analyst covering [COMPANY NAME]. You believe the bull case is overstated and you have one question on this call. Make it the right one.

FILTER: Below is the management commentary and any Q&A excerpts you have. Read everything once before formulating your question.

[PASTE COMMENTARY AND Q&A]

Your question must:

  • Refer to a specific claim or number management made on this call, not a generic concern
  • Identify the assumption the claim depends on, not the claim itself
  • Be a question a fund manager would ask on the next call — not theatre, not gotcha

RISK: Predict how management would most likely respond to your question, in their style. Then identify the part of their response that would itself be evasive.

VERDICT: State your question in one paragraph. Then state, in one sentence, what their answer would need to contain for you to leave the position alone — and what would make you trim.


Where these prompts fall short

Pasted text is the limit. The model can only assess what you give it; if the commentary section runs long and you copy the highlights, you have already done the sceptical sort — and I have caught myself doing exactly that. Tone classifiers on financial language are unreliable — “we are taking a conservative approach” reads as bearish to a generic NLP model but is often a positive capital-allocation signal in context. None of these prompts predict the next print or the share-price reaction; they read what is in front of you. And the model is bad at distinguishing rehearsed restraint from real reticence — a CEO who is naturally careful with language will look hedged on Prompt 2 even when the business is performing well.

The fix on all four limits is the same: run more than one prompt and weigh the outputs against the numbers. Language reading is the second pass, not the first.

What I run on results morning

Numbers first — Perplexity does the fastest live retrieval, Claude reads the deck more carefully (see the tools comparison for the per-task breakdown). Then Prompt 1, before reading what management said, while my expectations are still my own. Prompts 2 and 3 on the prepared remarks once the commentary is open. Prompt 4 if I have the prior transcript handy — for META, BMNR, and Apple I always do. Prompt 5 last, before deciding whether to act — which feeds into the five questions I ask AI before any trade.

The whole sequence takes about 20 minutes if the prior transcript is open in another tab. The point isn’t speed. The point is that the numbers tell you what happened and the language tells you what they think happens next. The Field Guide has the underlying method these prompts apply.

Verdict

//

What worked

Five prompts structured around specific tells — omissions, performative confidence, gap to consensus, language shift, the unasked question. Each is copy-paste ready and each runs against pasted text, not invented numbers.

What didn't

No prompt here predicts the share-price reaction. None replaces reading the document. Tone-classification on financial language is unreliable, and naturally restrained CEOs will look hedged on Prompt 2 even when they shouldn't.

Confidence

Medium-high — these are the prompts I run on real prints I have positions in. They have caught omissions and shifts I would have missed reading the transcript once.

Ben Dixon
// Written by Ben Dixon

Ben documents AI experiments against his own investment portfolio — real money, human analysis, sceptical use. About Ben →

// Keep reading
Prompt Stack

The single prompt change that made AI analysis worth using

One AI prompt separates observable facts from assumptions in any stock analysis. What that changes, why it matters for investing, and how to apply it.

Prompt Stack

7 AI prompts for covered calls

AI prompts for covered calls that work: seven prompts built around real chain data, covering position fit, IV regime, strike selection, and roll decisions.

Decision Process

What AI gets right (and wrong) about options trading

AI for options trading: a working taxonomy from real covered-call trades. Four uses where it helps. Six failure modes where it invents numbers.

← All posts More in Prompt Stack →