toolkit-chat-turbo4B dense
Base: Qwen3-4B-Instruct + our LoRA
Training: LoRA r=32, bf16, ~60K samples daily
Retrains: Daily (4am + 7pm UTC)
Real-time conversational model trained on fresh information — the only model we retrain on our own infra.
Conversational responsesCurrent eventsLocal informationFast inferencePersona awareness
Data: News, sports, markets, weather, jobs, dining, travel
Freshness weighting: today = 100%, this week = 50%, >30 days = dropped. This model reads the news every day.
toolkit-chat9B dense
Base: Qwen3.5-9B + our LoRA (our pod)
Training: LoRA r=32, bf16, full dataset
Retrains: Weekly
Premium conversational model with stronger reasoning. Our fine-tuned 9B with tk_ personality + domain tuning.
Tool callingStructured outputsMulti-step reasoningBusiness conversationsPlanning
Served from our pod. Daily briefing context injected as system prompt.
toolkit-code-turbo9B dense
Base: Qwen3.5-9B + our code LoRA (our pod)
Training: LoRA r=32, bf16, curated code dataset
Retrains: Weekly
Fast coding model for autocomplete, small refactors, unit tests.
Fast code generationSQLAPIsFunctionsIDE integration
toolkit-voicevoice system
Base: Toolkit voice stack
Training: Private voice tuning
Retrains: Daily voice updates
Full-duplex voice conversation for real-time assistant workflows.
Full-duplex voiceNatural conversationSpeaker profilesReal-time interactionSub-second knowledge lookup
Data: ~13K voice conversation pairs + live briefing context
toolkit-cam9B multimodal
Base: MiniCPM-o 4.5
Training: LoRA on .llm backbone
Retrains: Not yet live
Phone camera and screenshot understanding.
Phone camera inputScreenshot understandingVisual reasoningMultimodal knowledge
Text-only v1 trained, vision-grounded v2 planned.
toolkit-basehosted model
Base: Toolkit hosted base rail
Training: Hosted — no weight updates by us
Retrains: Daily briefing via system prompt
General workhorse for business analysis and long-document tasks (200K runtime context).
200K contextTool orchestrationBusiness logicStructured reasoning
Data: Our daily briefing: news, markets, weather, regional data — injected as system prompt
toolkit-code-backendhosted model
Base: Toolkit backend code rail
Training: Hosted — no weight updates by us
Retrains: —
Agent workflows + repo-aware coding with cached-prompt reuse.
65K repo contextSession-affinity cacheMulti-turn agent workflowsTool callingSystem design
toolkit-codehosted model
Base: Toolkit hosted code rail
Training: LoRA r=32, bf16 on curated code dataset
Retrains: Biweekly
Multi-file code + architecture with our domain tuning.
Multi-file refactorsRepo-wide changesAgentic codingTool calling131K contextDebugging across files
toolkit-think / toolkit-visionhosted models
Base: Toolkit reasoning and vision rails
Training: Hosted — no weight updates by us
Retrains: Daily briefing via system prompt
Deep reasoning and multimodal understanding through separate routed rails.
Deep multi-step reasoningResearchImage + document OCR128K vision contextChart comprehension
Think mode and vision mode are routed separately; public card values follow runtime limits.