Skip to main content

Inline CLI: Total API Control Inside the Prompt

Β· 2 min read
Alexander Carrington
COO of Neuronic AI
Inline CLI Feature

Drop in one-liner commands to control everything from model selection to memory, integrity, shaping, and real-time search β€” right from the user’s prompt.

πŸ§ͺ Examples:​

:deepsearch :becreative :answerwithclaude :setmodel:openai/gpt-4o :setmemoryon

βš™οΈ Enable via:​

"inline_cli": "all" // (default)

πŸ—¨οΈ User Prompt Example:​

{ "content": "Summarize today's AI news :deepsearch :setsearchlang:en :setsearchgeo:US" }

βœ… Supports:​

  • Model switching & state-maintained overrides (multi-tenancy-ready)
  • Search grounding with region/language targeting
  • Prompt shaping (:becreative, :beprecise, etc.)
  • Memory control
  • User state management & more

πŸ’‘ Just send :help as a user prompt to see inline help
🚫 Disable with: "inline_cli": false

πŸ“˜ Docs: Inline CLI Reference


🧠 Multi-Tenant Memory + Observability​

We now support per-user state, usage tracking, and long-term memory across all requests.

πŸ”§ Just include:

"user": "your_user_id"

πŸš€ This Unlocks:​

  • πŸ” Sub-user usage tracking
  • πŸ’Ύ Memory tied to user + sub-user
  • πŸ” Persistent CLI settings
  • πŸ“Š Deep observability by user/session

We’re launching one of the most powerful multi-tenant AI backends out there.


🧱 Works Seamlessly with OpenAI-Compatible APIs​

Everything above works instantly with your existing OpenAI-based stack using our /v1/chat/completions route.

βœ… Real-time web grounding
βœ… Inline API controls
βœ… Persistent memory
βœ… CLI shaping & model overrides


πŸ§ͺ Try it now: Docs


We’re not just leveling up β€”
We’re launching a whole new way to AI.

Modular, memory-aware, and real-world grounded β€” out of the box.
Let’s build the future β€” smarter, faster, and together.

β€” Team APIpie βš™οΈ