Claude Prompt Library

20 Claude Prompts for Data Analysis That Replace Hours of Spreadsheet Work

20 copy-paste prompts

XML-structured prompts for Excel formulas, SQL queries, data cleaning, visualization, and statistical analysis — built for how Claude actually processes data.

Excel & Spreadsheets

4 prompts

Complex Formula Builder

1/20

<context> Spreadsheet tool: [Excel / Google Sheets] Data layout: [describe your columns, e.g. "Column A = dates, B = product names, C = revenue, D = region"] Current row count: [approximate number of rows] Skill level: [beginner / intermediate / advanced] </context> <task> I need a formula that does the following: [DESCRIBE WHAT YOU WANT THE FORMULA TO DO — e.g. "Sum revenue for Product X in Q1 2026, but only for the East region, and show $0 if no data exists"] Provide: 1. The exact formula I can paste into cell [TARGET CELL] 2. A plain-English breakdown of what each part of the formula does 3. How to modify it if my column layout changes 4. A simpler alternative if one exists </task> <constraints> - Use only functions available in [Excel / Google Sheets] - Prefer XLOOKUP over VLOOKUP, FILTER over helper columns - If the formula exceeds 100 characters, break it into named ranges or helper columns and explain why - Handle edge cases: empty cells, text-in-number-columns, #N/A errors - Wrap in IFERROR with a meaningful fallback, not just 0 </constraints>

Generates exact spreadsheet formulas with plain-English explanations and error handling built in.

💡

Pro tip: Paste a few rows of sample data into the prompt. Claude writes more accurate formulas when it can see your actual column structure.

Pivot Table Design & Setup

2/20

<context> Tool: [Excel / Google Sheets] Dataset: [describe your data — e.g. "12 months of sales transactions with columns: Date, Salesperson, Product, Region, Amount, Quantity"] Row count: [approximate] Goal: [what question you want the pivot table to answer] </context> <task> Design a pivot table configuration to answer: "[YOUR QUESTION — e.g. Which salesperson has the highest average deal size by region?]" Provide: 1. Rows, Columns, Values, and Filters — exactly what to drag where 2. Value field settings (Sum vs. Average vs. Count, and why) 3. Any calculated fields I should add 4. Recommended sort order and top-N filtering 5. A second pivot table layout if the data tells a richer story with a different angle </task> <constraints> - Assume I know how to create a pivot table but not how to configure it optimally - If the question requires multiple pivot tables, say so and design each one - Recommend slicers or timelines if they'd make the analysis interactive - Flag if my data needs cleaning before the pivot table will work correctly </constraints>

Designs the exact pivot table configuration to answer your business question, with field placement and settings.

💡

Pro tip: Use Claude artifacts to get a visual mockup of the pivot table layout alongside the instructions.

Data Cleanup & Standardization

3/20

<context> Tool: [Excel / Google Sheets] Problem columns: [list columns that need cleaning — e.g. "Column B has mixed date formats, Column D has inconsistent company names, Column F has numbers stored as text"] Row count: [approximate] Deadline pressure: [I need this done in minutes / I can spend an hour] </context> <task> Create a step-by-step cleanup plan for my spreadsheet: For each problem column, provide: 1. A formula-based fix I can apply in a helper column (no VBA unless I ask) 2. The exact formula with cell references assuming data starts in row 2 3. Find-and-replace patterns for bulk fixes 4. A validation check to confirm the cleanup worked (e.g. COUNTIF for remaining issues) 5. Instructions to paste-values and remove helper columns when done Common issues to address: - Leading/trailing spaces - Inconsistent capitalization - Mixed date formats - Numbers stored as text - Duplicate variations of the same entity (e.g. "IBM" vs "I.B.M." vs "International Business Machines") </task> <constraints> - Formulas only — no macros unless the dataset exceeds 50K rows - Preserve original data in a backup sheet first - Each step must be reversible - If entity matching requires fuzzy logic, provide a VLOOKUP + helper table approach </constraints>

Creates a formula-based cleanup plan for messy spreadsheets with validation checks at every step.

💡

Pro tip: Paste 10-15 examples of the messy data directly into the prompt. Claude identifies patterns and edge cases you might miss.

VBA Macro Generator

4/20

<context> Excel version: [Excel 365 / 2021 / 2019 / Google Apps Script] Task frequency: [one-time / daily / weekly / triggered by event] Current manual process: [describe what you do by hand — e.g. "Every Monday I open 4 regional CSV files, consolidate into one sheet, add a pivot table, and email the summary to my manager"] VBA experience: [none / can edit existing macros / comfortable writing macros] </context> <task> Write a VBA macro (or Google Apps Script) that automates this process: [DESCRIBE THE FULL WORKFLOW TO AUTOMATE] Requirements: 1. Complete, runnable code with comments explaining each section 2. Error handling — what happens if a file is missing, a sheet doesn't exist, or data is empty 3. A status message or log so I know what the macro did 4. Instructions for where to paste the code and how to run it 5. How to assign it to a button or keyboard shortcut </task> <constraints> - Code must be copy-paste ready — no placeholder functions - Use early binding for known libraries, late binding if the user might not have references set - Add a confirmation dialog before any destructive action (deleting sheets, overwriting data) - If the task involves email, use Outlook object model (not CDO) - Keep the code under 150 lines — break into sub-procedures if longer - Add a comment header with: purpose, author, date, and instructions </constraints>

Generates complete, commented VBA macros with error handling and setup instructions for any repetitive Excel task.

💡

Pro tip: Save your common macro patterns in a Claude Project. Claude will maintain consistent coding style and variable naming across all your macros.

XML tags are just the start. Learn the full Claude workflow.

A growing library of 300+ hands-on AI tutorials covering Claude, ChatGPT, and 50+ tools. New tutorials added every week.

Start 14-Day Free Trial

SQL & Databases

4 prompts

Write SQL Query from Plain English

5/20

<context> Database: [PostgreSQL / MySQL / SQL Server / SQLite / BigQuery] Tables involved: [list tables with key columns — e.g. "orders(id, customer_id, amount, created_at, status), customers(id, name, email, region, signup_date)"] Relationships: [e.g. "orders.customer_id -> customers.id"] Data volume: [approximate row counts per table] </context> <task> Write a SQL query that answers this question: "[YOUR QUESTION IN PLAIN ENGLISH — e.g. Show me the top 10 customers by total spend in the last 90 days who haven't placed an order in the last 30 days]" Provide: 1. The complete query, formatted and commented 2. Expected output columns and what each represents 3. Any assumptions you made about the data 4. Performance notes if the query touches large tables 5. How to modify the query for common variations (different date ranges, additional filters) </task> <constraints> - Use syntax specific to [DATABASE TYPE] - Use CTEs over subqueries for readability when nesting exceeds 2 levels - Include appropriate indexes to suggest if the query is slow - Handle NULLs explicitly — don't let them silently distort results - If the question is ambiguous, state the assumption and provide both interpretations - Add comments for any non-obvious logic </constraints>

Translates plain-English business questions into production-ready SQL with comments and performance notes.

💡

Pro tip: Paste your actual CREATE TABLE statements into the prompt. Claude writes perfectly accurate JOINs when it sees real schema definitions.

Query Performance Optimizer

6/20

<context> Database: [PostgreSQL / MySQL / SQL Server / BigQuery] Table sizes: [list tables and approximate row counts] Current indexes: [list existing indexes, or say "unknown"] Query execution time: [current runtime — e.g. "takes 45 seconds"] Target execution time: [acceptable runtime — e.g. "under 2 seconds"] </context> <task> Optimize this slow query: [PASTE YOUR SLOW SQL QUERY] Provide: 1. Why it's slow — the specific bottleneck (full table scan, missing index, correlated subquery, etc.) 2. The optimized query with changes highlighted in comments 3. Indexes to create (with exact CREATE INDEX statements) 4. EXPLAIN plan interpretation — what to look for to confirm the fix worked 5. Alternative approaches if the optimization still isn't enough (materialized views, denormalization, caching) </task> <constraints> - Don't change the business logic — the optimized query must return identical results - If you suggest an index, estimate the storage cost and write impact - Prefer query restructuring over index creation when both solve the problem - If the query can't be optimized further, explain why and suggest architectural alternatives - Include the EXPLAIN ANALYZE command to run before and after </constraints>

Diagnoses slow SQL queries and provides optimized rewrites, index recommendations, and EXPLAIN plan guidance.

💡

Pro tip: Paste the output of EXPLAIN ANALYZE alongside the query. Claude pinpoints the exact operation causing the bottleneck.

SQL Query Explainer

7/20

<context> My SQL skill level: [beginner / intermediate / advanced] Database: [PostgreSQL / MySQL / SQL Server / other] Context: [why I'm reading this query — e.g. "inherited codebase", "debugging a report", "reviewing a PR"] </context> <task> Explain this SQL query in plain English: [PASTE THE SQL QUERY] Provide: 1. One-sentence summary: what this query does in business terms 2. Step-by-step walkthrough — what each clause does, in the order the database executes them (FROM -> WHERE -> GROUP BY -> HAVING -> SELECT -> ORDER BY -> LIMIT) 3. A visual diagram of how the tables join (text-based) 4. Edge cases — what data scenarios might produce unexpected results 5. If this query has bugs or inefficiencies, flag them </task> <constraints> - Explain at my stated skill level — [beginner: no jargon / intermediate: use standard terms / advanced: focus on execution plan implications] - Don't just translate syntax — explain the "why" behind each clause - If window functions are used, explain the partitioning and ordering in plain language - Highlight any parts that are unnecessarily complex and suggest simpler alternatives </constraints>

Breaks down complex SQL into plain-English explanations at your skill level, with join diagrams and edge case warnings.

💡

Pro tip: Enable extended thinking for deeply nested queries. Claude traces the execution order and catches logic errors that are invisible in a flat read.

Database Migration Script

8/20

<context> Source database: [type, version, and relevant schema] Target database: [type, version] Migration scope: [full database / specific tables / schema only / data only] Downtime tolerance: [zero downtime required / maintenance window of X hours / flexible] Data volume: [approximate total size] </context> <task> Create a migration plan from [SOURCE] to [TARGET]: 1. Schema conversion — DDL statements for the target database, handling type differences 2. Data migration — INSERT/COPY scripts or ETL approach 3. Sequence/auto-increment handling 4. Foreign key and constraint recreation order 5. Index recreation 6. Stored procedures / views / triggers that need rewriting 7. Validation queries — how to confirm row counts and data integrity match 8. Rollback plan — how to revert if something goes wrong </task> <constraints> - Handle data type incompatibilities explicitly (e.g. MySQL ENUM to PostgreSQL, SQL Server IDENTITY to PostgreSQL SERIAL) - Migration scripts must be idempotent — safe to re-run - Include transaction boundaries for data safety - Flag any features in the source that have no direct equivalent in the target - Provide estimated migration time based on stated data volume - Test queries should compare checksums, not just row counts </constraints>

Generates a complete database migration plan with schema conversion, data scripts, validation, and rollback procedures.

💡

Pro tip: Paste your source schema (pg_dump --schema-only output) into a Claude Project. Claude generates accurate type mappings when it sees real DDL.

Data Cleaning & Transformation

4 prompts

Deduplication Strategy

9/20

<context> Data source: [CSV file / database table / spreadsheet] Record count: [approximate] Key columns for matching: [e.g. "name, email, phone, address"] Known duplicate patterns: [e.g. "same person with different email addresses", "company name variations like LLC vs L.L.C."] Tool preference: [Excel formulas / Python pandas / SQL / any] </context> <task> Build a deduplication strategy for my dataset: 1. Exact match pass — identify rows that are identical on [KEY COLUMNS] 2. Fuzzy match pass — catch near-duplicates using: - Name variations (Bob/Robert, Inc/Incorporated) - Typos (Levenshtein distance threshold) - Phone/email normalization 3. Merge rules — when duplicates are found, which record wins? (most recent, most complete, specific source priority) 4. The complete code/formulas to execute this 5. A report showing: total records, duplicates found, records after dedup, confidence scores </task> <constraints> - Preserve all original data in a backup before any dedup - Assign a confidence score to each match: Definite (100%), Likely (80%+), Review (60-80%) - "Review" matches should be flagged for human decision, not auto-merged - If using Python, use only pandas and fuzzywuzzy/rapidfuzz (no paid APIs) - If using SQL, provide the query for each pass separately - Output a merge log showing which records were combined and why </constraints>

Creates a multi-pass deduplication strategy with exact matching, fuzzy matching, confidence scores, and merge rules.

💡

Pro tip: Paste 20-30 rows including known duplicates. Claude calibrates the fuzzy matching threshold to your actual data quality.

Data Normalization Pipeline

10/20

<context> Data source: [CSV / API response / database export] Columns to normalize: [list each column and its current state — e.g. "phone: mixed formats (555-1234, (555) 123-4567, +1 5551234567)", "state: mix of abbreviations and full names"] Target format for each: [e.g. "phone: +1XXXXXXXXXX", "state: 2-letter abbreviation"] Tool: [Python / SQL / Excel / Google Sheets] </context> <task> Write a normalization pipeline that transforms each column to its target format: For each column: 1. Detection — identify all current format variations in the data 2. Transformation rules — map each variation to the target format 3. The complete code/formula 4. Edge cases that can't be auto-fixed (flag for manual review) 5. Validation — a check that confirms 100% of values now match the target format </task> <constraints> - Handle international formats where applicable (phone numbers, dates, addresses) - Don't silently drop data that doesn't match any known pattern — flag it - Preserve the original column and create a new normalized column alongside it - If using regex, comment each pattern explaining what it matches - Provide a before/after sample for each transformation - If any transformation is lossy (reduces information), warn explicitly </constraints>

Builds a complete normalization pipeline with detection, transformation, edge case handling, and validation for any column format.

💡

Pro tip: Use Claude artifacts to preview the before/after transformation on your sample data before running it on the full dataset.

Unstructured Text Parser

11/20

<context> Data format: [emails / PDF extracts / web scrapes / log files / free-text survey responses] Sample data: [paste 3-5 representative examples] Target structure: [describe the structured output you want — e.g. "columns: sender, date, subject, action_items, deadline"] Volume: [number of records to process] Tool: [Python / Excel / Google Sheets / regex only] </context> <task> Create a parser that extracts structured data from this unstructured text: 1. Pattern analysis — identify the consistent and variable elements across my samples 2. Extraction rules — for each target field, the exact regex or parsing logic 3. Complete code that processes all records 4. Confidence score for each extraction (did the pattern match cleanly?) 5. Exception handling — what to do when the text doesn't match expected patterns 6. Output as [CSV / JSON / database INSERT statements] </task> <constraints> - The parser must handle variations in formatting, not just the exact samples provided - Build in a "fallback" column that captures the full original text for failed parses - Test the parser against the provided samples and show the output - If using regex, keep patterns readable — prefer multiple simple patterns over one complex one - Flag records where extraction confidence is below 80% - Include a summary: X records parsed successfully, Y flagged for review </constraints>

Converts unstructured text (emails, PDFs, logs) into structured data with extraction rules, confidence scores, and exception handling.

💡

Pro tip: Paste real samples with all their messiness. Claude builds more robust parsers from imperfect examples than from idealized ones.

Data Validation Rule Builder

12/20

<context> Dataset: [describe the dataset — e.g. "customer records with 15 columns"] Critical columns: [list columns and expected data types/formats — e.g. "email: valid email format, age: integer 18-120, revenue: positive decimal, country: ISO 3166-1 alpha-2"] Business rules: [any cross-column rules — e.g. "if status = 'active', email must not be null", "end_date must be after start_date"] Tool: [Python / SQL / Excel / Google Sheets] </context> <task> Build a comprehensive data validation suite: For each column: 1. Type check — is the data the correct type? 2. Format check — does it match the expected pattern? 3. Range check — is it within acceptable bounds? 4. Completeness check — are required fields populated? 5. Cross-column check — do related columns pass business rules? Deliverables: - Complete validation code/formulas - A validation report: column, rule, records passing, records failing, failure rate - A "data quality score" (0-100) for the dataset - Specific rows to investigate (the worst offenders) </task> <constraints> - Don't just count NULLs — check for disguised missing data ("N/A", "none", "-", "0" used as placeholder) - Flag statistical outliers (values beyond 3 standard deviations) separately from rule violations - Validation must be re-runnable — design it as a repeatable check, not a one-time script - Output a clean/dirty split: validated records vs. records needing attention - Include a "fix suggestion" for each common failure pattern </constraints>

Creates a complete validation suite with type checks, business rules, outlier detection, and a data quality score.

💡

Pro tip: Add your validation rules to a Claude Project. Claude will apply the same rules consistently every time you upload a new dataset.

These prompts give you the what. Tutorials give you the why.

Learn when to use extended thinking, how to build Claude Projects, and workflows that compound. 300+ tutorials and growing.

Try AI Academy Free

Statistical Analysis

4 prompts

Regression Analysis Setup

13/20

<context> Dataset: [describe your data — e.g. "monthly marketing spend (5 channels) and revenue over 3 years"] Dependent variable (Y): [what you're trying to predict — e.g. "monthly revenue"] Independent variables (X): [potential predictors — e.g. "Google Ads spend, Facebook spend, email campaigns sent, seasonality, price changes"] Tool: [Python / R / Excel / Google Sheets] Statistical background: [none / took a stats class / use regression regularly] </context> <task> Set up a regression analysis to answer: "[YOUR QUESTION — e.g. Which marketing channels have the strongest impact on revenue, and what's the expected ROI of increasing spend by 20%?]" Provide: 1. Data preparation steps — handling missing values, encoding categoricals, checking for multicollinearity (VIF scores) 2. The right type of regression for this problem (linear, multiple, logistic, polynomial) and why 3. Complete code to run the regression 4. How to interpret the output — R-squared, p-values, coefficients in plain English 5. Diagnostic checks — residual plots, normality test, heteroscedasticity 6. The business answer to my question, based on hypothetical but realistic output </task> <constraints> - Explain every statistical concept at my stated level — no unexplained jargon - If assumptions are violated (non-linearity, non-normal residuals), suggest the appropriate alternative - Include confidence intervals, not just point estimates - Warn about correlation vs. causation if the question implies causal claims - If the sample size is too small for reliable regression, say so and suggest what to do </constraints>

Sets up a complete regression analysis with data prep, model selection, diagnostic checks, and plain-English interpretation.

💡

Pro tip: Enable extended thinking for statistical analysis. Claude checks assumptions more thoroughly and catches issues like multicollinearity that invalidate results.

Hypothesis Test Selector & Runner

14/20

<context> Research question: [e.g. "Did our new pricing page increase conversion rates?"] Data: [describe your two groups — e.g. "Group A (old page): 1,200 visitors, 84 conversions. Group B (new page): 1,350 visitors, 118 conversions"] Sample sizes: [n per group] Significance level: [0.05 unless you have a reason for different] Tool: [Python / R / Excel / Google Sheets] </context> <task> 1. Identify the correct statistical test for this question (t-test, chi-square, Mann-Whitney, ANOVA, etc.) and explain why 2. State the null and alternative hypotheses in plain English 3. Check assumptions (normality, equal variance, independence, minimum sample size) 4. Run the test — provide complete code 5. Interpret the result: - p-value and what it means for my decision - Effect size (Cohen's d, odds ratio, etc.) and whether the difference is practically significant - Confidence interval for the difference 6. One-sentence recommendation: what should we do based on this result? </task> <constraints> - If assumptions are violated, run the appropriate non-parametric alternative and compare results - Don't just report "statistically significant" — always report effect size and practical significance - If the sample size is too small for reliable inference, calculate the minimum sample needed - Warn about multiple testing if I'm running several tests (Bonferroni correction) - Use two-tailed tests unless I explicitly state a directional hypothesis </constraints>

Selects the right hypothesis test, checks assumptions, runs the analysis, and interprets results with effect sizes and practical significance.

💡

Pro tip: Paste your raw data or summary statistics directly into the prompt. Claude selects and runs the correct test without you needing to know which one to use.

Correlation & Pattern Discovery

15/20

<context> Dataset: [describe variables — e.g. "20 columns of customer behavior metrics: visit frequency, pages viewed, time on site, support tickets, NPS score, churn (yes/no)"] Goal: [e.g. "find which behaviors predict churn"] Tool: [Python / R / Excel] Number of variables: [count] Number of records: [count] </context> <task> Perform a correlation and pattern discovery analysis: 1. Correlation matrix — all pairwise correlations with significance flags 2. Top 10 strongest correlations (excluding trivially obvious ones) 3. Surprising findings — correlations that are unexpectedly strong, weak, or reversed from intuition 4. Multivariate patterns — combinations of variables that together predict the target variable better than any single one 5. Visualization recommendations — which charts best reveal the patterns you found 6. Actionable insights — what these correlations suggest we should investigate or act on </task> <constraints> - Use Pearson for linear relationships, Spearman for non-linear — report both where they diverge - Flag spurious correlations that are likely driven by a confounding variable - Don't claim causation from correlation — suggest experiments to test causal hypotheses - If variables are on different scales, normalize before analysis - For high-dimensional data (>20 variables), suggest PCA or feature selection before exhaustive correlation - Include the code for a publication-quality correlation heatmap </constraints>

Discovers meaningful correlations and patterns in your data with significance testing, spurious correlation warnings, and actionable insights.

💡

Pro tip: Use Claude artifacts to generate the correlation heatmap. You can iterate on the visualization without re-running the analysis.

Time Series Forecasting

16/20

<context> Metric to forecast: [e.g. "monthly recurring revenue (MRR)"] Historical data: [time period and granularity — e.g. "36 months of monthly data"] Known factors: [seasonality, trends, events — e.g. "strong Q4 seasonality, launched new product in month 18, COVID dip in months 10-14"] Forecast horizon: [how far ahead — e.g. "next 6 months"] Tool: [Python / R / Excel] Accuracy requirement: [directional guidance / budget planning accuracy / financial reporting precision] </context> <task> Build a time series forecast: 1. Exploratory analysis — trend, seasonality, and stationarity assessment 2. Model selection — compare at least 3 approaches: - Simple (moving average, exponential smoothing) - Intermediate (SARIMA/ARIMA) - Advanced (Prophet, if applicable) 3. The winning model with complete code 4. Forecast with confidence intervals (80% and 95%) 5. Accuracy metrics on a held-out test set (MAPE, RMSE, MAE) 6. Plain-English summary: "We expect [metric] to be [range] over [period], with [confidence level]" </task> <constraints> - Split data into train/test (last [N] periods as test) before selecting the model - Account for known anomalies (COVID, one-time events) — don't let them distort the forecast - If seasonality exists, use seasonal decomposition and show the components - Confidence intervals are mandatory — point forecasts alone are dangerous for decisions - If the data is too short or noisy for reliable forecasting, say so and suggest minimum data requirements - Include a "what could go wrong" section — scenarios where the forecast would break down </constraints>

Builds a multi-model time series forecast with confidence intervals, accuracy metrics, and plain-English interpretation.

💡

Pro tip: Save your historical data as a CSV in a Claude Project. Claude can re-run the forecast each month as you append new data points.

Data Visualization & Reporting

4 prompts

Chart Type Recommender

17/20

<context> Data: [describe your dataset and what each column represents] Audience: [executive team / technical team / clients / board / public] Question to answer: [what story should the chart tell — e.g. "How has our market share changed relative to competitors over the past 8 quarters?"] Presentation medium: [PowerPoint / dashboard / printed report / email] Tool: [Excel / Google Sheets / Tableau / Python matplotlib / Power BI] </context> <task> Recommend the best chart type and design: 1. Primary recommendation — chart type with reasoning (why this chart and not alternatives) 2. Alternative option — for a different angle on the same data 3. Exact specification: - X-axis, Y-axis, color/size encodings - Title (concise, insight-driven — not "Sales by Quarter" but "Sales grew 23% YoY despite Q3 dip") - Labels, annotations, and reference lines to include - Color palette (accessible, print-safe) 4. Common mistakes to avoid with this chart type 5. Step-by-step creation instructions in [TOOL] 6. If the data needs reshaping before charting, provide the transformation steps </task> <constraints> - Never recommend pie charts for more than 5 categories or 3D charts for any reason - Optimize for the stated audience: executives want the takeaway, analysts want the detail - Title must state the insight, not just describe the data - All charts must pass colorblind accessibility (no red/green only encoding) - If the data has too many dimensions for a single chart, recommend a small multiples approach - Include axis formatting (start Y-axis at 0 unless there's a justified reason not to) </constraints>

Recommends the optimal chart type with exact specifications, insight-driven titles, and step-by-step creation instructions.

💡

Pro tip: Describe your audience first. Claude optimizes the chart complexity, annotation density, and title style for who will actually read it.

Dashboard Design Blueprint

18/20

<context> Business context: [what the dashboard monitors — e.g. "SaaS metrics for a B2B startup"] Key metrics: [list 5-10 KPIs — e.g. "MRR, churn rate, CAC, LTV, trial-to-paid conversion, NPS, support ticket volume, feature adoption"] Primary users: [who views this dashboard and what decisions they make with it] Update frequency: [real-time / daily / weekly / monthly] Tool: [Tableau / Power BI / Looker / Google Data Studio / Excel] Data sources: [where the data lives — e.g. "Stripe, Mixpanel, Zendesk, PostgreSQL"] </context> <task> Design a dashboard blueprint: 1. Layout — where each element goes (header, KPI cards, charts, filters) 2. KPI hierarchy — which metrics are primary (large, top), secondary (medium), and detail (accessible but not prominent) 3. Chart specifications for each visualization: - Chart type, data mapping, color encoding - Drill-down behavior (what happens when you click) 4. Filter and interactivity design — date range, segments, comparison toggles 5. Alert thresholds — when should a metric turn red/yellow/green 6. Mobile responsiveness — what to show vs. hide on smaller screens </task> <constraints> - Follow the "5-second rule" — the most important insight must be visible within 5 seconds - Maximum 8 visualizations per dashboard view — use tabs for additional detail - Every chart must answer a specific question (state the question above each chart) - Include a "compared to what?" benchmark for every metric (last period, target, industry average) - Design for the decision, not the data — if a metric doesn't drive action, remove it - Specify exact dimensions in a 12-column grid layout </constraints>

Creates a complete dashboard blueprint with layout, KPI hierarchy, chart specs, interactivity, and alert thresholds.

💡

Pro tip: Use Claude artifacts to generate a wireframe-style layout. Iterate on the arrangement before building in your BI tool.

Executive Summary Generator

19/20

<context> Report type: [monthly business review / quarterly board report / weekly ops update / ad-hoc analysis] Audience: [CEO / board / VP-level / cross-functional team] Key data points: [paste your numbers, or describe the analysis results — e.g. "Revenue: $2.4M (+12% MoM), Churn: 3.2% (up from 2.8%), New customers: 145, CAC: $340 (down 8%), NPS: 72"] Notable events: [anything that explains the numbers — e.g. "launched enterprise tier, lost 2 large accounts, hired 3 new sales reps"] Previous period context: [what happened last period for comparison] </context> <task> Write an executive summary: 1. TL;DR — 2 sentences capturing the overall narrative (are we winning or losing, and why?) 2. Key metrics table — this period vs. last period vs. target, with directional arrows 3. What went well — top 2-3 wins with supporting data 4. What needs attention — top 2-3 concerns with severity and proposed action 5. Outlook — what to expect next period based on current trajectory 6. Decisions needed — specific questions for leadership to resolve (not vague "we should discuss") </task> <constraints> - Total length: under 500 words (executives won't read more) - Lead with the narrative, not the numbers — numbers support the story - Every concern must have a proposed action (not just "churn increased") - Use plain language — no jargon, no acronyms without definition on first use - Include one "early warning" signal — something not yet a problem but trending toward one - Format for skimmability: bold key numbers, use bullet points, clear headers </constraints>

Generates a concise executive summary with narrative framing, metrics comparison, and specific decisions needed from leadership.

💡

Pro tip: Save your company metrics format and previous summaries in a Claude Project. Claude maintains consistent structure and tracks trends across periods.

Automated Report Template

20/20

<context> Report purpose: [e.g. "weekly sales performance report sent to the sales team every Monday"] Data source: [where the data comes from — e.g. "Salesforce export CSV", "database query output", "Google Analytics"] Distribution: [email / Slack / shared drive / printed] Current process: [how you create this report now — e.g. "manually copy data into a template, update charts, write commentary, takes 2 hours"] Tool: [Python / Excel VBA / Google Apps Script / R Markdown] </context> <task> Build an automated report generation system: 1. Data ingestion — code to read and validate the source data 2. Calculations — all derived metrics, comparisons, and rankings 3. Narrative generation — template-based commentary that adapts to the data: - If metric is up: "[Metric] increased [X%] to [value], driven by [top contributor]" - If metric is down: "[Metric] declined [X%] to [value]. Investigate [likely cause]" - If metric is flat: "[Metric] held steady at [value], within [X%] of target" 4. Visualization — auto-generated charts (no manual chart updates) 5. Output — formatted report in [PDF / HTML email / Excel / Google Slides] 6. Scheduling — how to run this automatically (cron job / Windows Task Scheduler / Cloud Function) </task> <constraints> - The script must handle missing data gracefully (skip a metric rather than crash) - Commentary must reference actual numbers from the data, not hardcoded text - Charts must auto-scale axes based on data range - Include a "data freshness" timestamp so readers know when the data was pulled - The entire pipeline must run in under 60 seconds - Include a dry-run mode that generates the report but doesn't send it </constraints>

Creates a complete automated report pipeline from data ingestion to formatted output with adaptive narrative commentary.

💡

Pro tip: Start with the Python version even if you normally use Excel. Claude generates cleaner automation in Python, and you can always export the final output to Excel.

Frequently Asked Questions

Claude cannot open .xlsx files directly, but it excels at working with spreadsheet data you paste into the conversation. Copy your column headers and a sample of rows, and Claude will write exact formulas, VBA macros, pivot table configurations, and data cleaning steps for your specific layout. For large datasets, describe the structure and Claude generates the formulas and code you paste back into Excel.
Claude handles larger data contexts (200K tokens — roughly 500 pages of data), follows multi-step analytical instructions more precisely, and produces more cautious statistical interpretations. Claude will warn you about violated assumptions, spurious correlations, and insufficient sample sizes rather than just running the numbers. It is particularly strong at writing production-quality SQL and Python pandas code.
Use both. Claude is best as an analyst assistant, not a replacement for your tools. It writes the SQL queries you run in your database, the Python scripts you execute in Jupyter, the formulas you paste into Excel, and the chart specifications you build in Tableau. Claude accelerates the thinking and coding — your tools handle the execution on real data.
Never paste confidential data into Claude unless your organization has approved it. Instead, anonymize your data first (replace company names with Company A/B/C, use representative sample data), or describe the schema and ask Claude to generate code you run locally. Claude Projects with API access offer additional privacy controls for enterprise use.

Prompts are the starting line. Tutorials are the finish.

A growing library of 300+ hands-on tutorials on ChatGPT, Claude, Midjourney, and 50+ AI tools. New tutorials added every week.

14-day free trial. Cancel anytime.