Toolhouse
Overview
Toolhouse is a Backend-as-a-Service (BaaS) to build, run, and manage AI agents. Toolhouse simplifies the process of building agents in a local environment and running them in production.
With Toolhouse, you define agents as code and deploy them as APIs using a single command. Toolhouse agents are automatically connected to the Toolhouse MCP Server; it gives agents access to RAG, memory, code execution, browser use, and all the functionality agents need to perform actions. You can add MCP Servers and even define custom code that the agent can use to perform actions not covered by public MCP Servers. Toolhouse has built-in eval, prompt optimization, and agentic orchestration.
For further information about the framework, please check the official documentation:
Integration via Python
Installation
pip install toolhouse openai python-dotenvOptionally add:
pip install groqor other SDKs, depending on the target LLM platform.
Connection Setup
You should obtain our API key first.
Create a
.envfile in your project:
Example Python integration (
toolhouse_example.py):
GUI Integration
The Toolhouse GUI (https://app.toolhouse.ai) supports:
API key management
Tool selection via Bundles
Agent execution & history
Monitoring tool calls in logs
Tool configuration is managed entirely through their GUI and reflected in tool discovery (th.get_tools()).
Integration via TypeScript
Installation
Install the required dependencies:
Connection Setup
Create a
.envfile in the project root:
Create a TypeScript file (
toolhouse.ts) with the following content:
Run the script:
GUI Integration
Toolhouse provides a browser-based GUI at app.toolhouse.ai where you can:
Manage API keys
Add and organize tools via Bundles
Monitor execution logs
Trigger and test agents visually
β Toolhouse integration with AIMLAPI is fully supported via
baseURLoverride in the OpenAI-compatible SDK.
β
Supported AIMLAPI Models
All chat-compatible models served by AIMLAPI are supported, including:
Mistralai β
Mistral-7B-Instruct,Mixtral-8x7BMeta β
Meta-LLaMA-3.1,LLaMA-3.3Anthropic β
Claude-3.5-HaikuNVIDIA β
Nemotron-70BGoogle, xAI, Alibaba, Cohere, DeepSeek β all supported through the unified
https://api.aimlapi.com/v1endpoint.
π View our full text (chat) model catalog.
βοΈ Supported Parameters
No AIMLAPI-specific parameter differences were found. Use standard OpenAI-compatible parameters:
modelmessagestemperaturemax_tokensstreamtools(Toolhouse integration)
π§ Supported Call Features
Synchronous calls
β
β
Asynchronous use
π‘ (manual)
β (via Promises)
Tool Calling
β
β
Streaming
β
β
Threads
β
β
Local tools
β
β
via registerLocalTool()
Last updated
Was this helpful?