Inline CLI: Total API Control Inside the Prompt

Drop in one-liner commands to control everything from model selection to memory, integrity, shaping, and real-time search β right from the userβs prompt.
π§ͺ Examples:β
:deepsearch :becreative :answerwithclaude :setmodel:openai/gpt-4o :setmemoryon
βοΈ Enable via:β
"inline_cli": "all" // (default)
π¨οΈ User Prompt Example:β
{ "content": "Summarize today's AI news :deepsearch :setsearchlang:en :setsearchgeo:US" }
β Supports:β
- Model switching & state-maintained overrides (multi-tenancy-ready)
- Search grounding with region/language targeting
- Prompt shaping (
:becreative
,:beprecise
, etc.) - Memory control
- User state management & more
π‘ Just send :help
as a user prompt to see inline help
π« Disable with: "inline_cli": false
π Docs: Inline CLI Reference
π§ Multi-Tenant Memory + Observabilityβ
We now support per-user state, usage tracking, and long-term memory across all requests.
π§ Just include:
"user": "your_user_id"
π This Unlocks:β
- π Sub-user usage tracking
- πΎ Memory tied to user + sub-user
- π Persistent CLI settings
- π Deep observability by user/session
Weβre launching one of the most powerful multi-tenant AI backends out there.
π§± Works Seamlessly with OpenAI-Compatible APIsβ
Everything above works instantly with your existing OpenAI-based stack using our /v1/chat/completions
route.
β
Real-time web grounding
β
Inline API controls
β
Persistent memory
β
CLI shaping & model overrides
π§ͺ Try it now: Docs
Weβre not just leveling up β
Weβre launching a whole new way to AI.
Modular, memory-aware, and real-world grounded β out of the box.
Letβs build the future β smarter, faster, and together.
β Team APIpie βοΈ