TL;DR: You can analyze CSV, Excel, and Parquet files with AI assistance — without uploading anything to the cloud. This guide walks through the browser-based approach using DuckDB WebAssembly and schema-only AI prompts.
Why you'd want to avoid the cloud
You export a CSV from your CRM. It has customer names, revenue figures, maybe contract terms. You need a quick answer: "What's our average deal size by region?"
The fastest path used to be: upload to ChatGPT, get an answer in 30 seconds.
The problem: your file is now on OpenAI's servers. For personal side projects, that's fine. For client data, healthcare records, financial reports, or anything under NDA — it's a risk that many compliance teams won't accept.
But avoiding the cloud doesn't mean going back to the dark ages. Modern browser technology lets you run full SQL analytics locally, with AI that never touches your data.
Step 1: Load your file into a browser-based SQL engine
The key technology is DuckDB WebAssembly — a full analytical SQL engine compiled to run inside your browser tab.
When you drop a CSV into a tool built on DuckDB WASM:
- The file is read by JavaScript's
File API— it goes from your disk into browser memory - DuckDB parses the file, infers column types, and creates a virtual table
- No network request is made. You can verify this by opening DevTools > Network tab
This isn't a "lite" SQL engine. DuckDB supports JOINs, window functions, CTEs, GROUP BY, and handles millions of rows. It's the same engine data engineers use locally, running in your browser sandbox.
Step 2: Explore the schema
Before asking questions, understand what you're working with. A good tool auto-profiles your data:
- Column types: Is "revenue" stored as a number or text?
- Null rates: Which columns have missing data?
- Cardinality: How many unique values in each column?
- Distributions: Min/max/median for numbers, top values for strings
This profiling happens entirely in the browser. The raw data never leaves.
Step 3: Ask questions with AI — schema only
Here's where privacy data analytics gets interesting. You want AI help, but you don't want to send your data to an AI provider.
The solution: schema-only prompts. The AI receives:
- Column names and types (e.g.,
region VARCHAR,revenue DOUBLE,close_date DATE) - Lightweight statistics (value ranges, top categories)
- Your question in natural language
The AI does not receive:
- Any actual data rows
- Individual values
- PII or sensitive content
From the schema, the AI generates SQL:
SELECT region, AVG(revenue) as avg_deal_size
FROM deals
WHERE close_date >= '2025-01-01'
GROUP BY region
ORDER BY avg_deal_size DESC
This SQL runs locally in DuckDB. The AI wrote the question; your browser answered it.
Step 4: Run multi-step analysis
One question leads to another. "Average deal size by region" → "Which reps are above/below average?" → "Is there a seasonal pattern?"
Good tools support iterative exploration:
- Follow-up questions that build on previous context
- Auto-generated charts when results should be visualized
- Drill-down by clicking on results
All of this runs locally. The AI agent may run multiple SQL queries in sequence, but the data processing always happens in your browser.
Step 5: Export insights, not data
When you're done analyzing, you share the output — charts, summaries, reports — not the underlying dataset. The raw data stays on your machine. The insight travels.
Export options for browser-based tools typically include:
- CSV export of query results
- Chart images (PNG/SVG)
- HTML reports
- PDF for formal deliverables
When you need full offline mode
For maximum privacy, some tools offer complete offline operation:
- Pre-download the AI model (WebLLM runs in-browser via WebGPU)
- Disconnect from the internet
- Analyze as normal — SQL, AI, charts all work without any network
This is useful for:
- Working at secure client sites with restricted networks
- Air-gapped environments
- Travel without reliable connectivity
- Maximum peace of mind
The tradeoffs
Let's be honest about what you give up:
Schema-only AI handles well:
- Aggregations, filtering, grouping, joins
- "Show me X by Y" questions
- Statistical analysis and anomaly detection
- Data profiling and quality checks
Where full-data AI is better:
- Free-text analysis ("summarize these customer comments")
- Pattern recognition across individual row values
- Tasks that require reading and interpreting specific content
For structured tabular data — which is most business analytics — schema-only AI covers the vast majority of use cases.
Tools that support cloud-free analysis
| Tool | Engine | AI | Offline | Free |
|---|---|---|---|---|
| QueryVeil | DuckDB WASM | Schema-only + local AI | Yes | Yes |
| DuckDB CLI | DuckDB native | None built-in | Yes | Open source |
| Evidence | DuckDB | None | Dev mode | Open source |
| Jupyter + DuckDB | DuckDB Python | None built-in | Yes | Open source |
The bottom line
Analyzing data without uploading to the cloud is not a compromise in 2026 — it's a better workflow for sensitive data. Browser-based SQL engines are fast, capable, and verifiable. AI can help by looking at your schema instead of your data.
Stop uploading sensitive CSVs. Analyze them locally.
QueryVeil lets you analyze data without uploading it anywhere. Try the demo with sample data — no signup required.
Learn more: What is privacy data analytics? | How to analyze sensitive data locally