← Back to AI Insights
Gemini Executive Synthesis

CyberWriter, a Markdown editor leveraging Apple's on-device AI stack (Foundation Models, NLContextualEmbedding, SFSpeechRecognizer/SpeechAnalyzer) for features like semantic search, AI-powered writing assistance, and voice notes.

Technical Positioning
A privacy-focused Markdown editor that integrates Apple's on-device AI for advanced features, offering an alternative to cloud-based AI solutions, especially for sensitive data like health information.
SaaS Insight & Market Implications
CyberWriter demonstrates the significant B2B potential of Apple's on-device AI stack, particularly for privacy-sensitive industries like healthcare. The "no API key, no cloud call, no per-token cost" model fundamentally alters the cost and security landscape for AI integration. By keeping all data and processing local, it eliminates major compliance and data sovereignty hurdles associated with cloud AI. The semantic search, RAG, and AI writing assistance features enhance knowledge worker productivity within a secure environment. This product highlights a critical trend: the shift towards edge AI for enterprise applications where data privacy, low latency, and cost predictability are paramount. It positions Apple Silicon as a viable platform for secure, powerful local AI.
Proprietary Technical Taxonomy
Markdown editor Apple's on-device AI stack Foundation Models (macOS 26) ~3B-parameter LLM streaming, structured output, tool use no API key, no cloud call, no per-token cost NLContextualEmbedding (Natural Language framework, macOS 14+) BERT-style 512-dim text embedder

Raw Developer Origin & Technical Request

Source Icon Hacker News Apr 20, 2026
Show HN: CyberWriter – a .md editor built on Apple's (barely-used) on-device AI

Apple has quietly shipped a pretty complete on-device AI stack into macOS, with these features first getting API access in MacOS 26. There are multiple components in the foundation model, but the skills it shipped with actually make this ~3b parameter model useful. The API to hit the model is super easy, and no one is really wiring them together yet.- Foundation Models (macOS 26) - a ~3B-parameter LLM with an API. Streaming, structured output, tool use. No API key, no cloud call, no per-token cost.
- NLContextualEmbedding (Natural Language framework, macOS 14+) -- a BERT-style 512-dim text embedder. Exactly what OpenAI and Cohere sell, sitting in Apple's SDKs since iOS 17.
- SFSpeechRecognizer / SpeechAnalyzer - on-device speech-to-text including live dictation. Solid accuracy on Apple Silicon.I built cyberWriter, a Markdown editor, on top of all three, mostly as a test and showcase to see what it can do. I actually integrated local and cloud AI first, and then Apple shipped the foundation model, it stacked on super easy, and now users with no local or API AI knowledge can use it with just a click or two. Well the real reason is because most markdown editors need plugins that run with full system access, and I work on health data and can't have that.Vault chat / semantic search. The app indexes your Markdown folder via NLContextualEmbedding (around 50 seconds for 1000 chunks on an M1). The search bar gets a "Related Ideas" section that matches by meaning - typing "orbital mechanics" surfaces notes about rockets and launch windows even when those exact words never appear. Ask the AI a question and it retrieves the top 5 chunks as context. Plain RAG, but the embedder, retrieval, chat model, and search all run locally.AI Workspace. Command+Shift+A opens a chat panel, Command+J triggers inline quick actions (rewrite, summarize, change tone, fix grammar, continue). Apple Intelligence is the default; Claude, OpenAI, Ollama, and LM Studio all work if you prefer. The same context layer - document selection, attached files, retrieved vault chunks - feeds every provider through the same system-message path. Because the vault context is file and filename aware, it can create backlinks to the referenced file if it writes or edits a doc for you.Voice notes and dictation. Record a voice note directly into your doc, transcribe it with SpeechAnalyzer, or just dictate into the editor while you think. Audio never leaves the Mac.The privacy story is straightforward because the primitives are already private. Vectors live in a `.vault.embeddings.json` file next to your vault, never sent anywhere. If you use Apple Intelligence, even the retrieved text stays on-device. For cloud models there is a clear toggle and an inline warning before any filenames or snippets leave the machine.Honest limitations:- 512-dim embeddings are solid mid-tier. A GPT-4-class embedder catches subtler relationships this will miss.
- 256-token chunks can split long paragraphs mid-argument.
- Foundation Models caps its context window around 6K characters, so vault context is budgeted to 3K with truncation markers on the rest.
- Multilingual support is English-only right now. NLContextualEmbedding has Latin, Cyrillic, and CJK model variants; wiring the language detector across chunks is Phase 2.The developer experience for these APIs is genuinely good. Foundation Models streams cleanly, NLContextualEmbedding downloads assets on demand and gives you mean-poolable token vectors in a handful of lines. Curious what others here are building on this stack - feels like low-hanging fruit that has been sitting there for a while.imgur.com/a/HyhHLv2The Apple AI embedding feature is going live today. I'm honestly surprised it even works out of the box.

Developer Debate & Comments

No active discussions extracted for this entry yet.

Engagement Signals

4
Upvotes
3
Comments

Cross-Market Term Frequency

Quantifies the cross-market adoption of foundational terms like Apple Silicon and Markdown editor by tracking occurrence frequency across active SaaS architectures and enterprise developer debates.