Features, Limitations & Tips

Everything EQBook can do, where its boundaries lie, and how to get the most from it.

Built for deep, private research

EQBook combines a local LLM, a RAG pipeline, and a focused research UI — all running offline on your Apple Silicon device.

Multi-source workspaces

Organise research into workspaces. Each workspace holds unlimited PDFs, DOCX, TXT, MD, audio, and video files.

Conversational Q&A

Chat against all your sources at once. Answers stream in real time with inline citations you can click to view the exact source passage.

Semantic search

Search across every chunk in the workspace using vector similarity. Finds conceptually related passages, not just keyword matches.

Whisper transcription

Drop in MP3 or MP4 files. EQBook transcribes them on-device using CoreML Whisper Small, then indexes the transcript for Q&A.

Inline citations

Every AI answer includes dotted-underline citations. Click any citation to see the exact source paragraph it was drawn from.

Audio Overview

Generate a podcast-style two-host conversation from your workspace sources. Absorb research on the go — produced entirely on-device.

Adaptive model selection

Hardware is auto-detected at launch. The right Gemma 4 variant is recommended based on your Mac's unified memory.

Source grouping & sidebar

Sources are grouped by type (PDF, Audio, Video, Text) in a collapsible sidebar. Rename, delete, or view chunk counts per source.

Persistent chat history

All conversations are saved locally. Sessions persist across restarts, giving you a running record of every research question you've asked.

Full macOS experience

Native macOS app with a collapsible source sidebar, keyboard shortcuts, and Spotlight integration — built for focused, distraction-free research.

EQBook Audio Overview — podcast-style two-host conversation generated from your workspace

Audio Overview — a podcast-style summary generated entirely on-device from your sources

The RAG Pipeline

EQBook's Retrieval-Augmented Generation pipeline runs entirely on-device in five stages.

StageWhat happensTechnology
1 — IngestText extracted from PDF, DOCX, TXT, MD, or Whisper transcriptPDFKit, XMLCoder
2 — ChunkText split into 512-token chunks with 50-token overlap, respecting paragraph breaksCustom Swift chunker
3 — EmbedEach chunk converted to a 384-dimension vectorMLX all-MiniLM-L6-v2
4 — RetrieveTop-8 semantically closest chunks fetched for the user's queryOn-device vector search (cosine)
5 — GenerateRetrieved chunks + question passed to Gemma 4; response streamed with citations extractedMLX-LM Gemma 4
EQBook quiz view — test your understanding of workspace sources

Quiz mode — test your comprehension of any workspace with AI-generated questions

Known Limitations

EQBook is powerful but not without trade-offs. Here's what to expect.

⚠️

Apple Silicon only

All inference uses Apple's MLX framework, which requires Apple Silicon (M1 or later). Intel Macs are not supported.

⚠️

Context window caps

The E4B model tops out at 128K tokens; the 26B MoE model at 256K. Very large document collections may exceed the context window — use focused workspaces to stay within limits.

⚠️

Document size limit (200 MB)

Documents larger than 200 MB prompt you to choose between copying or referencing. Files referenced from disk may cause issues if moved or deleted.

Get the most out of EQBook

Practical tips from the development team and early access users.

1

Create focused workspaces

Don't dump all your documents into one workspace. Use separate workspaces per project or topic to keep the retrieval context clean and relevant.

2

Let transcription finish before asking questions

Audio and video transcription happens asynchronously. Wait for a source's status badge to show "Ready" before running queries against it.

3

Ask specific questions

The RAG retrieval works best with specific, focused questions. Vague prompts retrieve scattered chunks and produce weaker answers.

4

Use the Insights panel first

Open the Insights tab after ingesting sources to get a quick summary. This helps you know what's in the corpus before writing chat prompts.

5

Click citations to verify answers

Always click inline citations to verify the source passage. LLMs can occasionally misrepresent or conflate information — check primary sources.

6

Upgrade to the 26B model if you can

If your Mac has ≥16 GB RAM, download the 26B MoE variant. It has a 256K context window and significantly better reasoning on complex queries.

7

Pre-process scanned PDFs

Scanned PDFs without embedded text will have poor extraction results. Run them through an OCR tool (e.g. Apple's Preview, Adobe) before importing.

Ready to try it?

Download EQBook for Mac. Free, zero cloud.