Self Hosted Models
Connect and manage self hosted models to your AI agents.
What Are Self-Hosted Models?
Self-hosted models let you run any OpenAI-API-compatible checkpoint on your own hardware (or cloud) and plug it directly into MindStudio. This is ideal if you:
Are a hobbyist experimenting with cutting-edge open-source models (e.g., Llama 3.2).
Work at an organization that needs private, fine-tuned, or on-premise models.
How to Connect a Self-Hosted Model in MindStudio
Prepare Your Model Endpoint
Run your model locally or on a server in an OpenAI-compatible way (for example, serving Llama 3.2 via
ncro
).
Open the Self-Hosted Models Panel
In your MindStudio workspace, click Self-Hosted Models in the sidebar.
Add a New Model
Click Add New Model.
Name: Enter a friendly label (e.g., “Llama 3.2”).
API Name: Choose an identifier you’ll use in blocks.
API Endpoint URL: Paste the URL where your model is served (e.g.,
http://localhost:8000/v1
).Leave all other settings at their defaults and click Save.
Use Your Self-Hosted Model in an Agent
Create or open an existing agent.
Add a Generate Text block (or any block that uses a model).
In the block’s settings, switch to the Self-Hosted category and select your model (“Llama 3.2”).
Enter your prompt (e.g., “What is your name?”) and run.
Watch MindStudio route the request to your endpoint and return the result in the chat.
Compatibility & Notes
Any API-compatible service works—open-source forks, private enterprise models, or on-premise deployments.
No extra setup in your prompts or blocks; MindStudio handles routing transparently.
Visible requests: You can observe incoming requests in your model logs to verify connectivity.
Last updated