Data-Enriched AI Pipelines

Your AI. Your Data. Your Infra.

Launch open-source models on European GPUs and enrich them with your documents and data. Integrate via our OpenAI-compatible API into your own product, or use our chatbots, agents and workflows to get started right away.

main.py
from openai import OpenAI

client = OpenAI(
    base_url="https://api.hostyour.ai/v1",
    api_key="hyai_..."
)

response = client.chat.completions.create(
    model="llama-3.2-70b",
    messages=[{"role": "user", "content": "Hello!"}]
)

Trusted by teams at

Rijksuniversiteit Groningen Hanzehogeschool Provincie Drenthe Frisius AI Jumbo
4 simple steps

How it works

From sign-up to a live AI pipeline in four steps.

1

Create an account

Sign up for free with your email. No credit card required.

2

Launch your model

Pick an open-source model and deploy on a European GPU, or connect your own API key.

3

Enrich with your data

Upload documents, crawl your website or connect a data source. Your AI now knows your context.

Go live

Integrate via the OpenAI-compatible API, build a chatbot, or start a workflow.

zsh — python
$ pip install openai
Successfully installed openai-1.x
$ python
>>> from openai import OpenAI
>>> client = OpenAI(
... base_url="https://api.hostyour.ai/v1"
... )
>>> response = client.chat.completions.create(
... model="llama-3.3-70b",
... messages=[{"role": "user", "content": "Hi!"}]
... )
>>> print(response.choices[0].message.content)
"Hello! How can I help you?"
Features

One platform for your entire AI pipeline

From data to deployment. Everything you need to build data-enriched AI.

LLM & Image Generation

Deploy text models (Llama, Qwen, DeepSeek) or image models (FLUX, SDXL) on dedicated GPUs. Or connect your own API key.

OpenAI-compatible

Same SDK you already know. One endpoint for everything, whether GPU or BYOK.

Knowledge base (RAG)

Upload documents, crawl websites, connect email. Your AI answers based on your own data.

Chatbots & agents

Build chatbots with custom branding. Embed on your site or share via link.

EU hosted & GDPR

4 EU datacenters. End-to-end encryption. Full data sovereignty.

Teams & workflows

Collaborate with roles and permissions. Automate multi-step AI pipelines.

Instances

Your own AI instance

Each instance is a dedicated AI model running on its own GPU. Choose from 100+ text models or image generation models like FLUX and SDXL, select your European region, and deploy with one click. You can also deploy any custom HuggingFace model. You get a private OpenAI-compatible API endpoint with no shared resources and no rate limits from others.

  • Text generation (Llama, Qwen, DeepSeek) and image generation (FLUX, SDXL) on dedicated GPUs
  • Deploy any custom HuggingFace model or choose from our curated list
  • OpenAI-compatible API endpoint ready in ~10 minutes, works with any SDK
  • Start, stop and scale on demand. Pay only when running
Already have an API key?

Prefer to use your own key from OpenAI, Groq, or another provider? Connect it in seconds and get the same proxy endpoint, knowledge base, chatbots, and all other features. No GPU needed.

OpenAI Groq Mistral DeepSeek Together Custom
Instances
DeepSeek R1 70B
A100 80GB · Amsterdam
running
Llama 3.3 70B
A100 80GB · Frankfurt
running
FLUX.1 Schnell IMAGE
RTX 4090 · Amsterdam
running
gpt-4o-mini BYOK
OpenAI · own key
running
Qwen 2.5 72B
H100 80GB · Helsinki
stopped
3 active · 1 stopped
+ New instance
Templates
Customer Service
Answer customer questions
Code Review
Analyze & improve code
Translation
Translate between 50+ languages
Summarization
Summarize long documents
Prompt Templates

Start in seconds with prompt templates

Don't start from scratch. Choose a ready-made prompt template for your use case and deploy a fully configured AI agent in seconds. Each template includes a pre-written system prompt, recommended model, and optional knowledge base, ready to customize.

  • Pre-built prompt templates for customer service, code review, translation, summarization and more
  • Each template includes a pre-written system prompt, recommended model, and sample configuration
  • Fully customizable. Use as starting point and adjust to your needs
Knowledge Base

Give your AI knowledge about your business

Upload documents, crawl your website, or connect your email. Everything is automatically split into chunks, converted to vector embeddings, and linked to your AI. When a user asks a question, the most relevant context is found and injected into the prompt so your AI answers based on your own data.

Upload or connect sources
PDF, DOCX, website crawl via sitemap, email (IMAP), REST APIs
Automatic chunking & vectorization
Documents are split into chunks and converted to vector embeddings for semantic search
AI answers with context
Relevant chunks are automatically injected into every API response, including external OpenAI-compatible calls
Knowledge Base
product-docs.pdf
42 chunks · 128 vectors
synced
hostyourai.com
62 pages · sitemap crawl
synced
support@company.com
IMAP · 1,240 emails
synced
AI response with context
AI
Based on 3 sources
Based on the product docs, the platform supports real-time streaming, batch processing and...
Bot Preview
AI
Support Bot
online
AI
Hi! I'm the Acme support bot. How can I help?
How do I reset my password?
AI
Go to Settings → Account → Change password. Click "Reset" and follow the steps.
Source: help-center.pdf
Bots

Deploy chatbots for your customers

Create AI-powered chatbots that you can embed on your website or share via a link. Each bot gets its own system prompt, knowledge base, and branding. Your customers chat with your AI, powered by your own model and data, without seeing any technical details.

  • Custom system prompt, name, avatar, and welcome message
  • Connect a knowledge base so the bot answers based on your documentation
  • Embed on your website with a single script tag, or share via unique link
  • Full conversation history and analytics in your dashboard
Workflows

Automate with AI workflows

Build multi-step AI pipelines that run automatically. Combine LLM calls, conditions, and actions into powerful workflows. For example: classify an incoming message, search your knowledge base for relevant info, generate a response, and escalate if the AI is unsure, all without writing code.

  • Chain multiple AI steps: classify, search, generate, decide
  • Add conditions and branches to route based on intent, sentiment, or keywords
  • Trigger automatically on new messages, or run manually from your dashboard
Workflow Builder
Trigger
New message
LLM
Classify intent
Condition
intent == "support"?
yes
Search KB + respond
no
Forward to agent
Team: Engineering
MV
Martijn de Vries
martijn@company.com
admin
SB
Sophie Bakker
sophie@company.com
member
JK
Jan Koster
jan@company.com
member
3 instances
2 knowledge bases
+ Invite
Teams

Collaborate with your team

Invite colleagues to your workspace and work together on AI projects. Share instances, knowledge bases, and bots across your team. Each member gets a role (admin or member) so you control who can deploy, edit, or only view.

  • Invite team members by email. They join instantly with one click
  • Share instances, knowledge bases, and bots across the team
  • Role-based access: admin (full control) or member (use & view)
Try it live

API Playground

Test our OpenAI-compatible API directly in your browser. Same interface, your own infrastructure.

curl https://api.hostyour.ai/v1/chat/completions \
  -H "Authorization: Bearer hyai_..." \
  -H "Content-Type: application/json" \
  -d '{
    "model": "deepseek-r1-distill-llama-70b",
    "messages": [
      {"role": "system", "content": "You are a helpful assistant."},
      {"role": "user", "content": "What is the capital of the Netherlands?"}
    ],
    "temperature": 0.7,
    "max_tokens": 1024
  }'
Playground
U
User
What is the capital of the Netherlands?
AI
DeepSeek R1 127 tokens · 342ms
The capital of the Netherlands is Amsterdam. However, The Hague (Den Haag) is the seat of government where the parliament is located.
Type your message...
100%
OpenAI Compatible
<100ms
Time to first token
0
Code changes needed
100+ models

Text & Image Models

From Llama to FLUX. Deploy text or image generation models with one click, or bring your own HuggingFace model.

DeepSeek R1 32B
DeepSeek R1 70B
DeepSeek R1 7B
DeepSeek Coder V2
Qwen 2.5 72B
Qwen 2.5 32B
Qwen Coder 32B
Llama 3.3 70B
Llama 3.1 70B
Llama 3.1 8B
Mixtral 8x22B
Mixtral 8x7B
Mistral Small 22B
Mistral Nemo 12B
Mistral 7B
Codestral 22B
Gemma 2 27B
Gemma 2 9B
Phi 3.5 MoE
Phi 3 Medium
CodeLlama 70B
CodeLlama 34B
StarCoder2 15B
Command R+
Command R
Yi 1.5 34B
InternLM 2.5 20B
Vicuna 13B
FLUX.1 Schnell
SDXL 1.0
FLUX.1 Dev
SD 3.5 Medium
+ 50 more

We set up your GPU

No CUDA drivers, no Docker, no ML ops. We ensure your model runs optimally. Works for text and image models.

Custom HuggingFace Models

Don't see your model? Deploy any HuggingFace model directly. Just enter the model ID and required VRAM.

0
DevOps needed
100+
AI models
6
Providers (BYOK)
4
EU datacenters
EU Sovereign

Your data, safe in Europe

Complete data sovereignty. No American cloud, no CLOUD Act, no worries.

EU Datacenters

Amsterdam, Frankfurt, Paris, Helsinki

GDPR Compliant

Full compliance with EU privacy legislation

No CLOUD Act

Beyond reach of American legislation

Dedicated Hardware

Your model on your own GPU, no sharing

GDPR

No vendor lock-in, no unexpected price increases, no content policies limiting you, no data being used for training purposes. Open-source models, European infrastructure.

Testimonials

What our customers say

Teams across Europe are building with HostYourAI.

"Finally a platform where we don't have to manage GPUs ourselves. Deploy in 10 minutes, OpenAI-compatible API, and everything runs in the EU."

MV
Martijn de Vries
CTO, DataFlow AI

"We switched from AWS Bedrock. Costs are 40% lower and we now have full control over which model we run."

SB
Sophie Bakker
Lead Developer, TechNL

"For our research, GDPR compliance was essential. HostYourAI offers dedicated instances in Amsterdam with complete data sovereignty."

JK
Dr. Jan Koster
AI Researcher, RUG
Pricing

Simple and transparent

GPU instances pay-as-you-go. BYOK is free.

Pay as you go
Credits for GPU instances. BYOK at no platform cost
From €0.10 /hour
GPU price varies per type • BYOK instances are free
  • GPU deploy or bring your own API key
  • All models, agents & knowledge bases
  • Top up with iDEAL or credit card
  • Teams, workflows & templates included
Create account

Need enterprise? Contact us

Ready to build your AI pipeline?

Connect your data, deploy your model, and go live. Up and running in minutes.