Create NotebookLM-ready Markdown source packs from Gmail MBOX exports, with local SQLite, JSONL, and CSV outputs for search, RAG, archive review, and LLM workflows.
mboxer is a local-first email archive processor designed around a common problem:
You can export Gmail as an MBOX file,
but a raw MBOX archive is not useful for NotebookLM, RAG, review, or analysis.
mboxer turns that raw archive into organized, structured, reusable knowledge assets.
Gmail / Google Takeout
→ MBOX file
→ local SQLite index
→ organized Markdown source packs
→ NotebookLM, RAG, search, review, JSONL, and future tools
Gmail archives often contain years of valuable personal, professional, legal, financial, operational, project, and organizational history.
Google Takeout makes it possible to export that history as an .mbox file, but the exported file is not immediately useful for modern AI workflows.
NotebookLM works best with readable, focused, well-organized source documents.
RAG systems work best with structured, chunkable records.
Spreadsheets work best with clean rows and metadata.
Local review works best when everything is inspectable before anything is uploaded.
mboxer bridges that gap.
The main selling point of mboxer is converting Gmail MBOX exports into clean, category-organized Markdown files that can be used as NotebookLM sources.
Instead of uploading one giant raw archive, mboxer creates structured source packs like:
exports/notebooklm/
finance/
invoices/
2024/
finance-invoices-2024-001.md
legal/
contracts/
2023-2024/
legal-contracts-2023-2024-001.md
projects/
product-launch/
2026/
projects-product-launch-2026-001.md
operations/
vendor-correspondence/
2025/
operations-vendor-correspondence-2025-001.md
The goal is to make exported Gmail content easier to:
- upload into NotebookLM
- organize by topic or category
- review before upload
- split into useful source packs
- preserve context from email threads
- exclude sensitive or irrelevant material
- reuse later for RAG, search, or analysis
Markdown is the primary output format.
Each exported file preserves useful email context:
- subject
- sender and recipients
- date
- thread hints
- category
- source account
- cleaned body text
- attachment references
Export output is split by category, year, and size band to respect NotebookLM source limits. A JSON manifest is written alongside each export run.
SQLite is the durable local project index.
The schema tracks:
- accounts and MBOX sources
- messages with normalized metadata and body text
- Gmail label associations
- thread groupings with participant and date ranges
- ingest runs with resumable checkpoint keys
- ingest errors per run
- attachments with SHA-256, content type, and extraction status
- classifications per message and per thread
- category taxonomy with locked/global flags
- category proposals for review and approval
- export items and export run records
- security findings per message
JSONL is intended for RAG pipelines, embeddings, local LLM tools, and structured downstream processing.
Each line represents one message with clean body text, metadata, and classification context. Account key is injected into the output path automatically to keep multi-account exports separated.
CSV export is planned for spreadsheet review, filtering, auditing, and manual cleanup.
The core pipeline is implemented and working.
Implemented:
- MBOX ingest into SQLite using Python's
mailboxstdlib - resumable ingest with per-run checkpoint keys and batch commits
- deduplication via
INSERT OR IGNOREon message identity - multi-account separation with per-account keyed storage
- message normalization: subject, sender, recipients, dates, body text, body hash, word count
- Gmail label parsing and storage
- thread grouping with participant aggregation and date ranges
- attachment extraction to disk with SHA-256 and content-type tracking
- rule-based classification at both message and thread level
- thread-level rule classification with message inheritance
assign(confidence 1.0) andassign_hint(confidence 0.75) rule actions- category taxonomy with locked categories and proposal workflow
- category review, approval, and rejection via CLI
- security scan and scrub hooks
- NotebookLM Markdown export with category directories, year bands, and size-limit profiles
- export dry-run mode
- JSONL export
- JSON export manifests
- five NotebookLM limit profiles:
standard,plus,pro,ultra,ultra_safe - CLI with subcommands for all pipeline stages
- YAML config loading with deep key access and environment variable support
pyproject.tomlpackaging with optionalpdfanddevextras
In progress / planned:
- CSV export
- LLM-based classification via Ollama (config shape is present, wiring is not complete)
- local web UI for category review
- incremental export tracking
- scrub profiles for PII redaction before cloud upload
Project name: mboxer
Python package: mboxer
CLI command: mboxer
Default database: var/mboxer.sqlite
Entry point: mboxer.cli:main
Python requires: >=3.11
src/mboxer/
cli.py # argparse CLI: all subcommands
config.py # YAML config loading, path helpers
ingest.py # MBOX ingest pipeline
normalize.py # message normalization and body extraction
classify.py # rule-based classification (message + thread)
taxonomy.py # category management and proposal workflow
accounts.py # account CRUD and resolution
attachments.py # attachment extraction and storage
limits.py # NotebookLM limit profiles and validation
naming.py # slugify and category path normalization
db/
schema.sql # full SQLite schema
schema.py # init_db helper
migrations/ # future migration scripts
exporters/
notebooklm.py # Markdown source pack exporter
jsonl.py # JSONL exporter
manifest.py # JSON manifest writer
security/
scan.py # security scan runner
scrub.py # scrub hooks
policy.py # export policy helpers
config/
mboxer.example.yaml # full annotated config example
tests/
test_accounts.py
test_classify.py
test_config.py
test_db.py
test_export.py
test_first_run.py
test_ingest.py
test_limits.py
test_manifest.py
test_migration.py
test_naming.py
test_normalize.py
test_scrub_export.py
test_taxonomy.py
test_thread_classify.py
python -m venv .venv
source .venv/bin/activate
pip install -e .
mboxer --helpCopy and customize the example config:
cp config/mboxer.example.yaml config/mboxer.yamlComplete walkthrough from a fresh checkout to a dry-run export.
1. Initialize the database
mboxer init-db --config config/mboxer.yaml2. Register your account
mboxer account add primary-gmail \
--display-name "Primary Gmail" \
--email user@example.com \
--config config/mboxer.yaml3. Verify the account was registered
mboxer account list --config config/mboxer.yaml4. Ingest a small test archive first (see warning below)
mboxer ingest data/mboxes/primary-gmail/sample.mbox \
--config config/mboxer.yaml \
--account primary-gmail \
--source-name "Sample" \
--extract-attachments \
--resume5. Classify with rules
mboxer classify \
--config config/mboxer.yaml \
--account primary-gmail6. Review categories
mboxer review-categories \
--config config/mboxer.yaml \
--account primary-gmail7. Run a security scan
mboxer security-scan \
--config config/mboxer.yaml \
--account primary-gmail8. Dry-run export to verify output shape
mboxer export notebooklm \
--config config/mboxer.yaml \
--account primary-gmail \
--profile ultra_safe \
--dry-run9. Real export when ready
mboxer export notebooklm \
--config config/mboxer.yaml \
--account primary-gmail \
--profile ultra_safe \
--out exports/notebooklmWarning: test with a small MBOX before ingesting large archives.
Gmail MBOX exports can exceed several gigabytes for long-lived accounts. Before ingesting a full archive:
- Extract a small slice of messages into a separate
.mboxfile and ingest that first.- Run
mboxer export notebooklm --dry-runto verify the output shape.- Review the generated exports locally before uploading anything to a cloud service.
--resumemakes ingest restartable, but a full ingest of a large archive still takes significant time and disk space. Running--dry-runon exports is free and fast.
You can export Gmail data from Google Takeout / Google Data Request.
The typical flow is:
- Request an export of your Gmail data.
- Download the archive from Google.
- Extract the downloaded archive locally.
- Locate the
.mboxfile. - Ingest the
.mboxfile withmboxer. - Export organized Markdown files for NotebookLM.
Example:
mboxer ingest data/mboxes/archive.mbox \
--config config/mboxer.yaml \
--source-name "Primary Gmail Archive" \
--account primary-gmail \
--extract-attachments \
--resumemboxer ingest data/mboxes/archive.mbox \
--config config/mboxer.yaml \
--source-name "Primary Gmail Archive" \
--account primary-gmail \
--extract-attachments \
--resume
mboxer classify \
--config config/mboxer.yaml \
--account primary-gmail \
--level thread
mboxer review-categories \
--config config/mboxer.yaml \
--account primary-gmail
mboxer security-scan \
--config config/mboxer.yaml \
--account primary-gmail
mboxer export notebooklm \
--config config/mboxer.yaml \
--account primary-gmail \
--profile ultra_safe \
--out exports/notebooklm
mboxer export jsonl \
--config config/mboxer.yaml \
--account primary-gmail \
--out exports/rag/messages.jsonlmboxer supports multiple separate Gmail accounts and archives in the same local project.
Each ingested source is tracked by account, source name, import run, and original MBOX file.
Example account keys:
primary-account
work-account
business-archive
organization-archive
project-archive
To export multiple accounts into a single NotebookLM run:
mboxer export notebooklm \
--config config/mboxer.yaml \
--accounts primary-account,work-account \
--profile ultra_safe \
--out exports/notebooklmNotebookLM exports are Markdown-first and organized by category directories.
Filenames remain meaningful even if the folder hierarchy is flattened during upload.
category-topic-year-sequence.md
Examples:
finance-invoices-2024-001.md
legal-contracts-2023-2024-001.md
projects-product-launch-2026-001.md
operations-vendor-correspondence-2025-001.md
research-literature-review-2024-001.md
support-customer-requests-2025-001.md
NotebookLM limit profiles are defined in config/mboxer.example.yaml.
| Profile | Max sources | Reserved | Target sources | Target words/source |
|---|---|---|---|---|
standard |
50 | 10 | 40 | 300,000 |
plus |
100 | 20 | 80 | 300,000 |
pro |
300 | 50 | 250 | 300,000 |
ultra |
600 | 75 | 525 | 300,000 |
ultra_safe |
600 | 100 | 450 | 225,000 |
Use ultra_safe as the default for large NotebookLM-oriented workflows where you want to preserve headroom for manual sources, attachments, PDFs, and later additions.
Limit overrides can be passed directly on the CLI:
mboxer export notebooklm \
--profile ultra_safe \
--max-sources 400 \
--target-words 200000Classification runs in two passes.
Rule-based (deterministic, no network required):
Rules match on sender domain, sender address fragment, and subject keywords.
Each rule assigns a category_path, sensitivity, notebooklm_priority, and export_profile.
At thread level, a matching rule is applied to the whole thread and then inherited down to all messages in the thread.
Rules support two assignment modes:
assign— confident match, confidence 1.0assign_hint— soft match, confidence 0.75
LLM-based (optional, local-first):
Config accepts an Ollama endpoint and model name. LLM classification is wired in the config shape and CLI but is not yet fully connected to the pipeline.
Classification can be scoped by account and run at message or thread level:
mboxer classify --level thread --account primary-gmailCategories are slash-delimited paths that become directory hierarchies in exports.
medical
medical/hospital-billing
medical/pharmacy
legal
legal/law-firm-correspondence
finance
household/utilities
postal/usps-informed-delivery
noise/marketing
noise/spam
Locked categories are defined in config and cannot be deleted.
The classifier can propose new categories. Proposals appear in review-categories and require explicit approval before being used in exports:
mboxer approve-category <proposal_id>
mboxer reject-category <proposal_id>mboxer assumes mail archives contain sensitive material.
Raw exports are local-only by default.
The security pipeline:
ingest
→ normalize
→ classify
→ security-scan
→ scrub
→ review
→ export
The config controls which redaction passes are applied before export:
security:
redact_phone_numbers: true
redact_ssn_like_numbers: true
redact_credit_card_like_numbers: true
redact_email_addresses: false
redact_physical_addresses: false
scan_attachments: true
quarantine_unsafe_attachments: trueCloud-oriented exports should use reviewed, scrubbed, or metadata-only profiles.
pip install -e ".[dev]"
pytestLinting and type checking:
ruff check src/
mypy src/Optional PDF support:
pip install -e ".[pdf]"mboxer should be:
- NotebookLM-friendly
- Gmail MBOX-focused
- local-first
- privacy-conscious
- resumable
- inspectable
- useful without a cloud service
- useful with local LLMs
- useful with future RAG systems
- safe for sensitive archives
- flexible enough for multiple Gmail accounts
- structured enough to support future application features
mboxer is not intended to be:
- a Gmail client
- a replacement for Gmail search
- a hosted SaaS product
- a tool that uploads raw email archives by default
- a black-box AI classifier
- a cloud-first archive processor
Releases are automated via .github/workflows/release.yml.
On every push to master (which is effectively every merged PR when branch protection requires PR-only changes), the workflow:
- Finds the highest existing
v*tag (by semantic version order). - Increments its patch component to produce the next version. If no prior
v*tag exists, starts fromv0.1.0. - Creates and pushes the new tag.
- Publishes a GitHub Release from that tag with autogenerated release notes.
Assumptions and caveats:
- Only the patch component is bumped automatically. To release a minor or major version, push the tag manually before merging to
master. - The workflow runs on any direct push to
master, not only merged PRs. If direct pushes are enabled, each push creates a new release. - The automated tag is pushed using
GITHUB_TOKEN. Due to a GitHub Actions limitation, tag pushes made byGITHUB_TOKENdo not re-trigger other workflows. This means thebuild-and-publish(PyPI) job in the same file will not run automatically for auto-created tags; PyPI publishing requires a manually pushedv*.*.*tag.
MIT