Audit Yourself to Get More From GenAI
The people who get the best results from AI aren’t the most technical: They’re the most disciplined. Apply this self-audit tool to keep improving your GenAI sessions.
News
- Trump Administration Eyes Regulatory Shift on AI as Cyber Threats Intensify
- Jensen Huang Rebuts AI Fears, Frames It as Industrial-Scale Job Creator
- Dubai Moves to Embed Agentic AI Across Private Sector Operations
- AWS and HUMAIN to Build a Full-Stack AI Ecosystem in Saudi Arabia
- AI Delivers More Accurate ER Diagnoses Than Doctors, Harvard Study Finds
- Oman to Scale AI Ecosystem With New Special Economic Zone
Carolyn Geason-Beissel/MIT SMR | Getty Images
More than a year into using generative AI daily, I wondered whether I was getting the most out of my AI use. There was no benchmark or feedback loop, and no one was grading my sessions with ChatGPT and Claude — until I created a self-audit.
I did what I’ve always done when faced with a process that lacked measurement. I studied every method I could find — prompting guides, conversations with colleagues, my own session patterns. I used AI to help me use AI better. Over time, I built a single self-audit prompt — one that encapsulates more than 30 habits for getting the most from AI.
Each time I ran the self-audit prompt, the output got sharper. The discipline became reflexive for me. That’s the real value of the self-audit: It made me better at using AI, in every session.
Now, at the end of any significant AI session, I simply prompt: “Review this session and assess it against my AI habits guide. Score how I did, identify what I missed, and guide me to apply missed habits.” Within a few minutes, I get a diagnostic that is uncomfortably specific about what I missed. I now have an answer to a key question: whether my process was good, not just the GenAI output.
A recent field experiment confirmed what I found through my experience. A research team that included MIT Sloan professor Jackson Lu randomly assigned 250 employees at a technology consulting firm in China to either use ChatGPT to assist with their work or to work without it.1 The employees with ChatGPT access were judged as significantly more creative by both their supervisors and outside evaluators. But the gains showed up exclusively among employees with strong metacognitive strategies — those who reflected on their own thinking, recognized knowledge gaps, and refined their approach when results were weak. That finding underscores that metacognition — thinking about your thinking — is the missing link between simply using AI and using it well.
AI widens the gap between disciplined and undisciplined professionals. People who skip the discipline generate more volume without more insight — a pattern consistent with what researchers at the University of California, Berkeley’s Haas School of Business called “unsustainable intensity” in findings published in early 2026.2
Knowing how to use AI is good — but to get the most value from the tool, you need to know whether you’re using it well. The self-audit gives you that.
A Self-Audit That Measures Five Key Goals
My self-audit prompt is organized across five goals: set up, refine, verify, own, and systematize. These goals represent a practice that experienced professionals have instinctively followed for years, long before generative AI’s arrival. You don’t need technical training to score well on this audit. You need to replicate the thinking and brainstorming process that you are likely already good at when conducting competitive research, responding to requests for proposals (RFPs), engaging in acquisition analysis, and planning a sales presentation, for example. It is your skill in the application of AI, not the AI itself, that makes the difference.
The self-audit assesses each generative AI session with five questions linked to each of the goals:
- Set up: Did you prepare the AI before asking it to work?
- Refine: Did you iterate on your own thinking, or just reprompt?
- Verify: Did you verify before trusting?
- Own: Did you make the output yours, or accept the default?
- Systematize: Did you build something reusable, or close the chat and start over?
You won’t score well on all five goals in every session — nor should you. But knowing which ones you missed, and why, enables you to change your next session. Think of it as AI holding a mirror to your own ability. It gets sharper every time you make it your own.
To illustrate what strong performance looks like at each goal, and what the self-audit is measuring, I applied the audit to an actual competitive due diligence analysis on a $5 billion global services company. Details have been modified for confidentiality, but the habits, prompts, and results are drawn from actual chat sessions. I’ll focus on the impact one goal at a time.
1. Set Up: Pass the Intern Test
What the self-audit measures: Did you prepare the AI with sufficient role, context, constraints, and materials before asking it to work — or did you jump straight to a question?
The most consequential decision in any AI interaction happens before the first prompt. It’s the decision to prepare.
I tell the AI who it should be, what it has to work with, and what I need it to produce. “You are an elite research analyst specializing in competitive intelligence. Here are the target company’s last two annual reports and its most recent earnings-call transcript. Assess this company’s ability to disrupt our core business within 18 months and recommend our strategic response.” That prompt will produce far better output than “Tell me about this competitor.”
I call this the “intern test.” If you handed your prompt to a brand-new hire with no context about your company, your industry, or your priorities, would they know what to do? If not, why would you expect your AI to?
Most readers will likely pass this test. Any GenAI prompting guide or video covers the basics of setup.
What gets overlooked is making clear what setup should not do — the negative constraint. I specify what I do not want: “Do not give me a generic SWOT. Do not hedge every statement. Do not define terms I already know.” And upload your materials. The more context you provide, the more accurate the output. It’s like telling a new team member “Figure out our competitive position” versus handing them your last three strategy decks and customer feedback.
Two additional practices make setup more effective. Before a significant AI chat, I run a preflight check: “What does a great outcome look like? What are the three most important things to get right?” After the first good draft, I generate a bridge summary so context carries forward, especially when I’ll be taking a long break between prompts or need to transition to a new chat. You might not have considered using this tactic before. A bridge summary is especially valuable if you tend to have long, multipart exchanges over days or even weeks. (In one case, Claude suggested doing so at time intervals to avoid having the conversation get too complicated.)
In the due diligence scenario, the difference in outputs before and after the self-audit was stark. While my first prompt was solid, the negative constraints and a preflight check were missing. The variable was me. What made the biggest difference? The negative constraint. Once I told the AI what not to do — no generic SWOT, no hedging, no defining terms I already know — the output became richer in insight and started reading like a briefing, not a book report.
2. Refine: Pass the Rethink Test
What the self-audit measures: Did you truly iterate on your own instructions and thinking, or did you simply reprompt for a better answer?
The first output from any AI session is a draft, not a deliverable. The real value comes from iteration. But the most productive iteration improves your own instructions, not the AI’s answer.
That’s metacognition in action. The person who pauses to ask, “What did I fail to specify? What assumption did the AI make that I should have preempted?” is exercising exactly the reflective discipline that separates high performers from the rest. AI rewards those who rethink their own instructions — not those who rephrase the same request.
AI rewards those who rethink their own instructions — not those who rephrase the same request.
I started catching my own patterns. Sometimes the output sounded right, but I couldn’t explain why — so I’d ask the AI to walk me through its reasoning, and the gaps would surface. Other times, I’d catch myself reprompting the same request with slightly different words and realize that the real problem was that I hadn’t broken the task down. The hardest one to admit: When I still couldn’t get what I wanted, it was usually because I couldn’t describe the desired goal clearly enough. Pasting in an example of output that showed what I was after worked better than trying to describe it.
One of the most powerful refining habits is embarrassingly simple: Ask the AI what you should be asking. “What question should I be asking that I am not currently asking?” That one prompt has produced more valuable insights than any other, in my experience.
When I applied these habits to the due diligence, they surfaced a critical insight I’d overlooked: The competitor’s employee sentiment data contradicted its public narrative of a thriving digital transformation. That disconnect between external messaging and internal reality changed my entire threat assessment. I never would have discovered that if I hadn’t challenged my own assumptions.
3. Verify: Pass the Trust Test
What the self-audit measures: Did you independently verify the AI’s claims, check its sources, and stress-test its confidence — or did you trust fluent output at face value?
AI output typically reads well — which can be a problem. It’s linguistically fluent and structurally polished, even when the underlying claims are fabricated, outdated, or mathematically wrong. This is a new kind of quality risk, and it misleads experienced professionals more often than they’d like to admit.
I once asked AI to summarize the regulatory history of the credit card industry, which I know well. The response was beautifully written, logically structured, and completely wrong on two key regulatory revisions. It read like an A-minus term paper from a student who’d skipped the reading. I almost didn’t catch it — because it sounded right. That’s what worried me. I knew the domain well, and I still nearly walked into a committee meeting with hallucinated data.
Since then, I’ve built verification into my routine. I ask the AI to surface and rank every assumption behind its answer. I request verifiable sources and note when the model can’t provide them. For anything involving numbers, I ask for step-by-step calculations. I’ve found two habits particularly effective: the temporal awareness check (“What is the date of the most recent information you’re drawing on?”) and the confidence stress test (“Rate your confidence in each factual claim as high, medium, or low”).
It’s the same discipline we’ve always followed: Verify before you trust; trust before you share.
During the due diligence, the AI flagged that its revenue figures were nine months old and rated its confidence in the regulatory settlement details as medium. When I verified the output independently, I discovered a $42 million enforcement action that the AI had understated. That single verification changed the risk profile of the entire analysis.
4. Own: Pass the Signature Test
What the self-audit measures: Did you actively impose your voice, your position, and your audience on the output — or did you accept AI’s default?
The real work starts here. I used to stop too early. Most of us do.
AI models default to hedged, tonally generic output. Left unguided, they produce content that is competent but indistinct — written by a smart person who seems to have an opinion about everything yet commits to nothing. That’s fine for a rough research summary, but it doesn’t reflect your voice or your style, and it’s not something you’d want to put your name on.
The first complete draft was exactly that: well organized, factually grounded, and thoroughly researched. But it was hedged throughout and read like a report designed to avoid being wrong rather than to help someone make a decision. When I forced the AI to take a clear position on the competitive threat, pushed it for unconventional strategic responses, and asked it to apply champion-challenger lenses, the analysis became richer and something I would stake my reputation on.
The habit most people underuse is calibrating AI to their personality: how they think, how they argue, and what they won’t tolerate in a deliverable.
One technique I use at this stage is running a draft by a virtual personal board of directors that I built. These distinct personas help push my thinking and the AI’s analysis away from the default path toward the edges. I built AI-powered personas modeled on real personalities: v_SunTzu for power dynamics, v_Indra (Nooyi) for the human dimension, v_Mark (Cuban) for commercial realism, and v_Meg (Whitman) for operational rigor. What survives that gauntlet of virtual advisers is sharper and more defensible.
The habit most people underuse is calibrating AI to their own personality: how they think, how they argue, and what they won’t tolerate in a deliverable. Take ownership of the thinking, not just the editing. That’s when the output starts sounding like you.
5. Systematize: Pass the Reuse Test
What the self-audit measures: Did you build systems that make your next session better — or did you close the chat and leave yourself having to start from scratch next time?
Nearly everyone treats each AI session as a stand-alone thread — which may be productive in isolation, but the value doesn’t compound. Here, the discipline shifts from improving sessions to building systems.
Building repeatable processes out of one-off successes is what I do. Yet, early on in my GenAI use, I spent two hours building a detailed competitive analysis that delivered exceptional output — and then I closed the chat. I’d produced a great deliverable but captured none of the thinking that made it great. I should have known better. When I needed to run a similar analysis a month later, I had to start from scratch — the same role definition, the same constraints, the same verification steps, all rebuilt from memory.
Three habits make the difference. These are not habits you apply at the end of the conversation but throughout — after every prompt, at every logical checkpoint, or after a break.
First, maintain continuity. During any significant working session, I ask the AI to maintain a running summary of what we’ve accomplished, what’s still open, and what I will need to copy and paste to resume the conversation in another chat. This produces a bridge summary that makes it easy for you to pick up the discussion in a new session without losing continuity, especially if you run out of tokens on one chat.
Second, be a coeditor. Review the AI’s output after every prompt, or at logical break points, and feed your own judgment back in. You read what the AI produced. Some of it is good; some of it is wrong. Some of it is vague in ways you didn’t notice until you tried to use it. You fix it, mark it up, and hand it back: “Here’s my revised version. Use this as our new baseline and continue from here.”
Third, “templatize” what works. Every time you craft a session that produces exceptional output — a due diligence workflow, an RFP evaluation, a customer analysis — convert it into a reusable template. Replace the specifics with [variable] placeholders and save the session as what I call a macro-prompt — a single structured prompt that combines the entire session’s workflow so anyone can run it without having to start from scratch. Individual expertise becomes organizational capability.
That single due diligence session became a reusable macro-prompt I’ve now used for partnership evaluations, board position assessments, and acquisition analyses — each time just pasting it in the chat to start the conversation. From there, AI guides me step-by-step — instead of me guiding the AI — with all of the thinking intensity captured from the original session. After every use, I run a prompt to improve this macro-prompt for the next session.
How to Start Auditing and Improving
Below, I’ve shared the self-audit macro-prompt that includes all 30 habits to audit oneself. Think of it as a companion resource. You can just copy and paste it into an existing conversation you’ve been having with AI on a significant, extended topic. See what it tells you about your use of AI across all five goals and 30 habits. The self-audit will show you exactly where to refocus.
Then, start practicing these habits in your GenAI conversations wherever you see the opportunity.
Generative AI technology has already proved its capabilities and will keep getting better. The discipline is what unlocks real value — and that discipline will always be needed, regardless of which AI tool you use.
There’s one last thing I didn’t expect when I started this journey: The better I got at working with AI, the better I got at thinking without it.
Run the self-audit. See what it tells you about your critical thinking.
Self-Audit Prompt
Copy and paste the prompt below during or after any significant AI working session. The AI will autonomously review your entire conversation, evaluate it against 30 habits spanning five goals, and deliver a structured diagnostic with scores, specific gaps it identified, and the exact prompts you should have used.
During or after a session: Paste it at any point in a conversation — midsession to course-correct, or at the end, to score what you did against the five goals.
Retroactively: Paste it into any past conversation you’ve had with an AI to learn from your history.
This macro-prompt includes micro-prompts or checks for every habit so the AI will know exactly what to look for and will be able to show you precisely what you should have said.
SELF-AUDIT MACRO-PROMPT — COPY AND PASTE BELOW
SELF-AUDIT OF AI SESSION
Review the entire conversation we just had. Evaluate how effectively I used AI in this session by assessing my performance against the 30 habits below.
For each goal, check whether I applied the habits listed. For each habit I missed, show me the EXACT PROMPT I should have used — written specifically for the content of this session, not as a generic template.
Work through the five goals in order. After all five, deliver the scorecard.
=================================
GOAL 1: SET UP — Did I prepare the AI before asking it to work?
=================================
Habit 1 — The preflight
Did I define what a great outcome looks like before starting?
Micro-prompt: “Before we begin, help me define: What does a great outcome look like for this task? What are the three most important things to get right? What mistakes do people typically make?”
Habit 2 — The mission
Did I assign a clear role, context, and mission?
Micro-prompt: “You are [specific expert role with years of experience in relevant domain]. Here is what I need: [specific deliverable]. Here is the context: [situation, constraints, timeline]. Your mission: [clear objective].”
Habit 3 — The negative constraint
Did I state what I did NOT want?
Micro-prompt: “Do not [produce generic output]. Do not [hedge every statement]. Do not [define terms I already know]. Do not [give balanced, ‘on the other hand’ analysis].”
Habit 4 — The context upload
Did I provide relevant documents, data, or prior work?
Micro-prompt: “Here are the attachments: [list files]. Use these as the primary basis for your analysis. Flag where you are drawing on general knowledge versus the documents I provided.”
Habit 5 — The session bridge
Did I provide or request a bridge summary for continuity?
Micro-prompt: “This is a continuation of our previous work on [topic]. Here is where we left off: [paste summary]. Confirm your understanding, flag anything unclear, and suggest where to pick up.”
=================================
GOAL 2: REFINE — Did I iterate on my own thinking, not just reprompt?
=================================
Habit 6 — The iteration
Did I challenge assumptions and explore alternative scenarios?
Micro-prompt: “Your analysis assumes [X]. Surface that assumption. What changes if [alternative scenario A]? What changes if [alternative scenario B]?”
Habit 7 — The reasoning request
Did I ask the AI to show its reasoning step-by-step?
Micro-prompt: “Think step-by-step through your reasoning for [conclusion]. Show me the logic chain before restating your conclusion. I want to see how you got there, not just where you landed.”
Habit 8 — The prompt self-critique
Did I ask the AI to critique or improve my prompt?
Micro-prompt: “How would you improve my original prompt? Rate it 1-10 for clarity, specificity, and completeness. Show me what a 10 would look like.”
Habit 9 — The strategic question
Did I ask what question I should be asking but haven’t?
Micro-prompt: “Step back. What question should I be asking about [topic] that I haven’t asked? What blind spots does my framing have?”
Habit 10 — The decomposition
Did I break complex tasks into sequential subtasks?
Micro-prompt: “Before writing the full [deliverable], (1) list the top three [dimensions], (2) rank them by [criteria], and (3) draft only the highest-priority one with supporting evidence.”
Habit 11 — The expert thinking
Did I request an expert or alternative perspective?
Micro-prompt: “How would a [specific expert role] evaluate this? What would they focus on that [my current perspective] might miss?”
Habit 12 — The few-shot example
Did I provide concrete examples of desired output?
Micro-prompt: “Here is an example of the depth and structure I want: [paste excerpt]. Match this level of specificity and directness.”
Habit 13 — The diagnosis
Did I diagnose and fix vague or generic responses?
Micro-prompt: “Your [section] feels generic. Identify the assumptions you made and the context that was missing. Then revise with more specificity about [specific aspect].”
=================================
GOAL 3: VERIFY — Did I verify before trusting?
=================================
Habit 14 — The assumption surface
Did I ask the AI to surface and rank its assumptions?
Micro-prompt: “List every assumption underlying your [analysis/recommendation]. Which ones are weakest? Which would change your conclusion entirely if wrong?”
Habit 15 — The source demand
Did I demand verifiable sources?
Micro-prompt: “Provide sources I can independently verify for [specific claims]. If you cannot provide a verifiable source, say so explicitly.”
Habit 16 — The counterargument
Did I request the strongest opposing case?
Micro-prompt: “Make the strongest possible case that [opposite of your conclusion]. What evidence supports that view?”
Habit 17 — The math audit
Did I ask for step-by-step math on calculations?
Micro-prompt: “Recalculate [specific figures]. Show your math step-by-step.”
Habit 18 — The confidence stress test
Did I request confidence ratings on factual claims?
Micro-prompt: “For each factual claim in this [output], rate your confidence as high, medium, or low. Flag anything below high and explain why.”
Habit 19 — The freshness check
Did I check the recency of the data?
Micro-prompt: “What is the date of the most recent information you drew on? Flag anything that may be outdated.”
Habit 20 — The hallucination stress test
Did I stress-test which claims are most likely wrong?
Micro-prompt: “Which specific factual claims in this [output] are you least certain about? If I fact-checked every statement, which ones are most likely to be wrong?”
=================================
GOAL 4: OWN — Did I make this mine, or accept the AI’s default?
=================================
Habit 21 — The position forcer
Did I force a clear position rather than accepting hedged output?
Micro-prompt: “Do not hedge. Take a clear position: [specific question]. Defend your position, then address the strongest counterargument.”
Habit 22 — The originality push
Did I push for unconventional or nonobvious angles?
Micro-prompt: “Generate three unconventional [responses/strategies/angles] that most [consultants/analysts/writers] would not recommend. Label one as high risk, high reward.”
Habit 23 — The specificity demand
Did I require specific data points instead of abstract claims?
Micro-prompt: “Support every claim with a specific data point from the documents I provided or a verifiable source. Remove anything abstract.”
Habit 24 — The narrative shaper
Did I shape output into narrative rather than accepting lists?
Micro-prompt: “Rewrite this as a strategic narrative: What is the one thing [audience] needs to understand, why does it matter, and what is the decision we need to make now? No lists. End with a clear recommendation.”
Habit 25 — The audience calibration
Did I calibrate output for a specific audience?
Micro-prompt: “Rewrite this for [specific audience]. Assume they are [smart but not immersed in details]. Lead with [what matters to them].”
Habit 26 — The multi-persona workflow
Did I use multiple perspectives to challenge the output?
Micro-prompt: “Now review this from three perspectives: (1) [strategist role]: What are we failing to anticipate? (2) [empathetic leader role]: What human factors are missing? (3) [editor role]: Tighten and cut.”
=================================
GOAL 5: SYSTEMATIZE — Did I build systems, not just outputs?
=================================
Habit 27 — The coeditor
Did I feed my own edits back in as a coeditor THROUGHOUT the session?
Check: Did this happen at multiple points during the conversation — not just once at the end? Count how many times I revised and handed back my own version. More is better. Flag any stretch of three or more prompts where I accepted output without coediting.
Micro-prompt: “Here is my revised version with my edits. Use this as our new baseline. Incorporate my changes, flag anything you disagree with, and continue from here.”
Habit 28 — The session debrief
Did I request bridge summaries THROUGHOUT the session?
Check: Did this happen at logical break points, before long breaks, or when approaching token limits — not just at the end? Count how many bridge summaries were requested. Flag any point where continuity was lost because a bridge summary was missing.
Micro-prompt: “Summarize what we accomplished, what’s still open, and what I should bring to our next session to pick up where we left off.”
Habit 29 — The self-audit
Did I run self-audit checkpoints THROUGHOUT the session?
Check: Did I pause at logical milestones to assess session quality before moving on — or did I audit only at the very end? Flag any major transition between goals or phases where a midsession audit would have caught a gap earlier.
(You’re running the final self-audit now.)
Habit 30 — The macro maker
Did I convert the session into a reusable macro-prompt?
Micro-prompt: “Convert this session into a reusable macro-prompt with [variable] placeholders. Format it so anyone can copy, paste, and follow the steps to produce [deliverable type].”
=================================
SCORECARD — Deliver this after evaluating all five goals
=================================
For each goal (1-5), provide:
- Score (1-5, where 5 = all habits demonstrated, 1 = none)
- Habits demonstrated well (with specific examples from our conversation)
- Habits missed (with the EXACT prompt I should have used, written for the specific content of THIS session)
- How each missed prompt would have improved the output
Then provide:
- Overall session score (average of five goals)
- The single highest-impact habit I missed
- Top three habits to focus on in my next session
Be specific and direct. Reference actual moments in our conversation.
Do not soften the assessment.
=================================
SESSION CLOSE
=================================
After delivering the scorecard, ask me: “Would you like me to (1) go back and apply the missed habits now to improve the work we just did, (2) generate a bridge summary for your next session, or (3) suggest improvements to this self-audit macro-prompt based on what we learned in this session?”
Apply these tips to get the most from the self-audit:
- Run it at the end of every significant AI session, not just occasionally. The habit of measuring is itself the discipline.
- Don’t stop at the scorecard. When the AI asks, “Would you like me to go back and apply the missed habits?” say yes. Then run the self-audit again. Repeat until you’re satisfied you’ve extracted the most value from the session.
- Track your scores over time. You’ll notice patterns — goals you consistently score well on and goals you consistently skip. Those patterns are your development road map.
- Improve the prompt itself. When the AI suggests improvements to this macro-prompt based on your session, review them and update your saved copy. The self-audit gets sharper each time you use it.
- Make it yours. Add habits that matter to your work, remove ones that don’t, or build in your own techniques. The 30 habits here are a starting point, not a ceiling.
- Share it with your team. When everyone runs the same self-audit, you build a shared language for AI session quality across the organization.
References (2)
1. S. Sun, Z.A. Li, M.-D. Foo, et al. “How and for Whom Using Generative AI Affects Creativity: A Field Experiment,” Journal of Applied Psychology 110, no. 12 (December 2025): 1561-1573, https://psycnet.apa.org/doi/10.1037/apl0001296.
2. A. Ranganathan and X.M. Ye, “AI Doesn’t Reduce Work — It Intensifies It,” Harvard Business Review, Feb. 9, 2026, https://hbr.org.