Deploy Dedicated GPU server to run AI models

Deploy Model
Skip to main content
Enterprise Open-Source Models

50 open-source models you can deploy on dedicated GPUs

This hub is the crawlable catalog for the enterprise model pages. It covers popular open-source image, video, audio, 3D, and LLM deployments that teams want to run with private infrastructure, code access, and bring-your-own S3 storage.

Dedicated GPU deployment with no shared queue contention

100% private workloads, prompts, and generated outputs

Code access for custom runtimes, adapters, and optimization

Bring-your-own S3 storage for assets, checkpoints, and outputs

Featured deployment pages

These are the pages with the strongest current commercial and product fit for ModelsLab Enterprise, including the exact models you called out like FLUX, Qwen, Stable Diffusion, DeepSeek, Whisper, and 3D stacks.

Stable Diffusion sample output
ImageDedicated GPU

Stable Diffusion

Stable Diffusion is still the broadest open image generation family for teams that want checkpoint flexibility, custom fine-tunes, adapters, and private asset pipelines.

Text to imageImage to image
FLUX.1 Dev sample output
ImageDedicated GPU

FLUX.1 Dev

FLUX.1 Dev is a strong open image generation baseline for teams that want modern prompt performance and private inference without shared platform bottlenecks.

Text to imageImage to image
FLUX 2 Dev sample output
ImageDedicated GPU

FLUX 2 Dev

FLUX 2 Dev is already wired into the repo for enterprise-class text generation and multi-image editing flows, making it a strong dedicated GPU target for advanced image products.

Text to imageMulti-image img2img
FLUX Kontext Dev sample output
ImageDedicated GPU

FLUX Kontext Dev

FLUX Kontext Dev is positioned for prompt-guided image transformation where teams want tighter control over edits, references, and enterprise runtime behavior.

Image to imageReference-guided editing
FLUX Klein sample output
ImageDedicated GPU

FLUX Klein

FLUX Klein is a lighter FLUX-family option for teams that want the FLUX visual stack in a smaller dedicated deployment footprint.

Text to imageDedicated FLUX-family hosting
Qwen Edit sample output
ImageDedicated GPU

Qwen Edit

Qwen Edit is a strong fit for teams that want a Qwen-branded image editing deployment with private prompt handling and dedicated enterprise infrastructure.

Image editingReference-based changes
Qwen Image Edit 2511 character consistency example
ImageDedicated GPU

Qwen Image Edit 2511

Qwen Image Edit 2511 is the strongest repo-backed example of the enterprise open-model approach: it supports multi-image editing, text-guided transformations, and production fetch/webhook flows on dedicated infrastructure.

Up to 4 input images2048px max width and height
DeepSeek R1 sample output
LLMDedicated GPU

DeepSeek R1

DeepSeek R1 is one of the clearest enterprise deployment wins in the open LLM landscape because teams want its reasoning ability without exposing prompts or internal context to third-party shared providers.

Chat completionsPrivate prompt handling
Llama 3.3 70B sample output
LLMDedicated GPU

Llama 3.3 70B

Llama 3.3 70B remains a high-intent enterprise model page because teams actively compare private open-weight Llama deployments against shared hosted APIs.

Chat completionsPrivate context handling
Whisper Large V3 sample output
AudioDedicated GPU

Whisper Large V3

Whisper Large V3 is still the obvious enterprise speech page because teams repeatedly need transcription that keeps private audio off shared infrastructure.

Speech to textDedicated audio processing
HunyuanVideo sample output
VideoDedicated GPU

HunyuanVideo

HunyuanVideo is a strong enterprise target for teams that want an open video generation stack without routing prompts, frames, and outputs through shared systems.

Dedicated video generationPrivate prompt handling
Hunyuan3D 2 sample output
3DDedicated GPU

Hunyuan3D 2

Hunyuan3D 2 is a good dedicated enterprise page because private 3D generation often involves proprietary product imagery and design workflows.

Text to 3DImage to 3D

Image Models

Dedicated deployment pages for 20 image models and adjacent enterprise use cases.

20 pages

Stable Diffusion sample output
ImageDedicated GPU

Stable Diffusion

Stable Diffusion is still the broadest open image generation family for teams that want checkpoint flexibility, custom fine-tunes, adapters, and private asset pipelines.

Text to imageImage to image
Stable Diffusion XL sample output
ImageDedicated GPU

Stable Diffusion XL

SDXL is the default open model choice for teams that want strong prompt adherence and broad ecosystem support without giving up deployment control.

Text to imageImage to image
Stable Diffusion 3.5 Large sample output
ImageDedicated GPU

Stable Diffusion 3.5 Large

Stable Diffusion 3.5 Large is positioned for higher-quality image generation teams that want a modern Stable Diffusion stack on infrastructure they fully control.

Text to imageControlled image workflows
Stable Diffusion 3.5 Medium sample output
ImageDedicated GPU

Stable Diffusion 3.5 Medium

Stable Diffusion 3.5 Medium is a lighter entry point for teams that want newer Stable Diffusion quality with more practical dedicated GPU cost envelopes.

Text to imageImage-to-image pipelines
Stable Diffusion 1.5 sample output
ImageDedicated GPU

Stable Diffusion 1.5

SD 1.5 still matters for legacy fine-tunes, mature community checkpoints, and teams that have existing prompt libraries they do not want to migrate yet.

Text to imageImage to image
SDXL Turbo sample output
ImageDedicated GPU

SDXL Turbo

SDXL Turbo is useful when speed matters more than maximal quality and teams want a fast, private image generation runtime on their own dedicated GPU envelope.

Fast text to imageInteractive generation loops
FLUX.1 Dev sample output
ImageDedicated GPU

FLUX.1 Dev

FLUX.1 Dev is a strong open image generation baseline for teams that want modern prompt performance and private inference without shared platform bottlenecks.

Text to imageImage to image
FLUX.1 Schnell sample output
ImageDedicated GPU

FLUX.1 Schnell

FLUX.1 Schnell is the speed-focused FLUX variant for teams that want private inference with faster feedback loops and tighter latency targets.

Fast text to imagePrivate prompt handling
FLUX 2 Dev sample output
ImageDedicated GPU

FLUX 2 Dev

FLUX 2 Dev is already wired into the repo for enterprise-class text generation and multi-image editing flows, making it a strong dedicated GPU target for advanced image products.

Text to imageMulti-image img2img
FLUX Kontext Dev sample output
ImageDedicated GPU

FLUX Kontext Dev

FLUX Kontext Dev is positioned for prompt-guided image transformation where teams want tighter control over edits, references, and enterprise runtime behavior.

Image to imageReference-guided editing
FLUX Klein sample output
ImageDedicated GPU

FLUX Klein

FLUX Klein is a lighter FLUX-family option for teams that want the FLUX visual stack in a smaller dedicated deployment footprint.

Text to imageDedicated FLUX-family hosting
Qwen Image sample output
ImageDedicated GPU

Qwen Image

Qwen Image gives teams a Qwen-native multimodal image stack that works well for text generation, reference-aware edits, and private enterprise creative systems.

Text to imageImage to image
Qwen Edit sample output
ImageDedicated GPU

Qwen Edit

Qwen Edit is a strong fit for teams that want a Qwen-branded image editing deployment with private prompt handling and dedicated enterprise infrastructure.

Image editingReference-based changes
Qwen Image Edit 2511 character consistency example
ImageDedicated GPU

Qwen Image Edit 2511

Qwen Image Edit 2511 is the strongest repo-backed example of the enterprise open-model approach: it supports multi-image editing, text-guided transformations, and production fetch/webhook flows on dedicated infrastructure.

Up to 4 input images2048px max width and height
Z Image Turbo sample output
ImageDedicated GPU

Z Image Turbo

Z Image Turbo is already modeled in the repo as a dedicated image deployment target for teams prioritizing faster image generation and enterprise control.

Text to imageImage to image
Z Image Base sample output
ImageDedicated GPU

Z Image Base

Z Image Base is a practical dedicated deployment target for teams that want the Z Image family on private infrastructure instead of shared inference.

Text to imageCustom deployment control
ControlNet sample output
ImageDedicated GPU

ControlNet

ControlNet is still one of the most useful open add-on stacks for structure-guided image generation when teams need more control than prompt-only workflows can provide.

Conditioned generationCustom checkpoint stacks
IP-Adapter sample output
ImageDedicated GPU

IP-Adapter

IP-Adapter is a useful enterprise deployment target when teams want image-conditioned generation without exposing private brand or product references to shared infrastructure.

Image-conditioned generationPrivate reference handling
InstantID sample output
ImageDedicated GPU

InstantID

InstantID is useful for identity-preserving generation when teams need strict privacy around reference photos and consistent dedicated throughput.

Identity conditioningPrivate reference inputs
Real-ESRGAN sample output
ImageDedicated GPU

Real-ESRGAN

Real-ESRGAN is still one of the easiest SEO and enterprise wins in the open model stack because teams constantly need private upscaling and restoration pipelines.

Image upscalingRestoration

Video Models

Dedicated deployment pages for 10 video models and adjacent enterprise use cases.

10 pages

HunyuanVideo sample output
VideoDedicated GPU

HunyuanVideo

HunyuanVideo is a strong enterprise target for teams that want an open video generation stack without routing prompts, frames, and outputs through shared systems.

Dedicated video generationPrivate prompt handling
Wan 2.1 sample output
VideoDedicated GPU

Wan 2.1

Wan 2.1 is a high-interest video model target for teams exploring dedicated private video generation infrastructure beyond closed video APIs.

Text-to-video workloadsPrivate prompt handling
CogVideoX 5B sample output
VideoDedicated GPU

CogVideoX 5B

CogVideoX 5B is a recognizable open video model for teams that want private generation infrastructure rather than handing prompts and assets to shared providers.

Video generationPrivate runtime control
Stable Video Diffusion sample output
VideoDedicated GPU

Stable Video Diffusion

Stable Video Diffusion is an important open model target when teams want image-to-video capability in a private dedicated environment.

Image to videoPrivate asset handling
LTX Video sample output
VideoDedicated GPU

LTX Video

LTX Video is relevant for teams evaluating newer open video stacks that need dedicated enterprise hosting, runtime control, and private media handling.

Dedicated video generationPrivate storage integration
Mochi 1 sample output
VideoDedicated GPU

Mochi 1

Mochi 1 is a practical video model target for teams that want open video experimentation on infrastructure they own instead of shared video services.

Video generationPrivate prompt handling
Open-Sora sample output
VideoDedicated GPU

Open-Sora

Open-Sora is a natural enterprise deployment target for teams that want a recognizable open video stack they can run inside private infrastructure.

Dedicated video generationCode access
SkyReels V2 sample output
VideoDedicated GPU

SkyReels V2

SkyReels V2 is attractive for teams looking at newer open video systems but still needing enterprise-grade privacy, dedicated throughput, and runtime control.

Video generationPrivate storage integration
Pyramid Flow sample output
VideoDedicated GPU

Pyramid Flow

Pyramid Flow is a useful open video deployment target for teams exploring dedicated private motion generation stacks across enterprise workloads.

Dedicated video inferencePrivate data handling
AnimateDiff sample output
VideoDedicated GPU

AnimateDiff

AnimateDiff is still relevant for teams building controlled motion pipelines around open image models and needing dedicated private GPU infrastructure.

Motion generationAdapter-based workflows

LLM and Multimodal Models

Dedicated deployment pages for 12 llm and multimodal models and adjacent enterprise use cases.

12 pages

DeepSeek R1 sample output
LLMDedicated GPU

DeepSeek R1

DeepSeek R1 is one of the clearest enterprise deployment wins in the open LLM landscape because teams want its reasoning ability without exposing prompts or internal context to third-party shared providers.

Chat completionsPrivate prompt handling
DeepSeek V3 sample output
LLMDedicated GPU

DeepSeek V3

DeepSeek V3 is a strong dedicated enterprise target when teams want a cost-aware open LLM stack for private production inference.

Chat completionsPrivate prompt flow
DeepSeek Coder V2 sample output
LLMDedicated GPU

DeepSeek Coder V2

DeepSeek Coder V2 is a natural fit for private engineering copilots where source code and developer prompts should stay inside dedicated infrastructure.

Coding chatPrivate code context
Llama 3.3 70B sample output
LLMDedicated GPU

Llama 3.3 70B

Llama 3.3 70B remains a high-intent enterprise model page because teams actively compare private open-weight Llama deployments against shared hosted APIs.

Chat completionsPrivate context handling
Llama 3.1 8B sample output
LLMDedicated GPU

Llama 3.1 8B

Llama 3.1 8B is attractive for teams that want a smaller dedicated LLM footprint while keeping prompts, retrieval context, and code-level runtime changes private.

ChatPrivate inference
Qwen 3 32B sample output
LLMDedicated GPU

Qwen 3 32B

Qwen 3 32B is a strong open LLM candidate for private multilingual and reasoning workloads that need enterprise-grade control instead of shared hosted endpoints.

Chat completionsPrivate prompt flow
Qwen 2.5 72B sample output
LLMDedicated GPU

Qwen 2.5 72B

Qwen 2.5 72B is a high-intent dedicated deployment target for teams that need stronger open-model performance with private enterprise hosting.

ChatPrivate context handling
Qwen 2.5 VL sample output
LLMDedicated GPU

Qwen 2.5 VL

Qwen 2.5 VL is a strong enterprise deployment candidate for multimodal apps that want private image understanding and dedicated runtime control.

Multimodal reasoningImage understanding
Mixtral 8x7B sample output
LLMDedicated GPU

Mixtral 8x7B

Mixtral 8x7B remains one of the most recognizable open MoE models for teams comparing dedicated open LLM hosting options.

ChatDedicated private inference
Mistral Nemo sample output
LLMDedicated GPU

Mistral Nemo

Mistral Nemo is useful when teams want a smaller open Mistral-family deployment with dedicated privacy, code access, and infrastructure control.

ChatPrivate prompt handling
Phi-4 sample output
LLMDedicated GPU

Phi-4

Phi-4 is a strong fit for smaller dedicated enterprise deployments where teams want a compact model footprint without leaving shared hosted services in the loop.

ChatPrivate inference
Gemma 3 27B sample output
LLMDedicated GPU

Gemma 3 27B

Gemma 3 27B is relevant for enterprise teams comparing Google-origin open-weight models with fully dedicated private deployment options.

ChatPrivate prompt handling

Audio and Voice Models

Dedicated deployment pages for 6 audio and voice models and adjacent enterprise use cases.

6 pages

Whisper Large V3 sample output
AudioDedicated GPU

Whisper Large V3

Whisper Large V3 is still the obvious enterprise speech page because teams repeatedly need transcription that keeps private audio off shared infrastructure.

Speech to textDedicated audio processing
Kokoro 82M sample output
AudioDedicated GPU

Kokoro 82M

Kokoro 82M is a compact open TTS deployment target for teams that want private voice generation without relying on closed hosted voice APIs.

Text to speechPrivate content handling
F5-TTS sample output
AudioDedicated GPU

F5-TTS

F5-TTS is a strong page for enterprise audio buyers because it maps directly to private TTS infrastructure and custom voice pipeline control.

Text to speechDedicated hosting
XTTS v2 sample output
AudioDedicated GPU

XTTS v2

XTTS v2 is attractive when teams want open multilingual TTS inside dedicated infrastructure instead of sending voice content to shared providers.

Text to speechMultilingual output
OpenVoice V2 sample output
AudioDedicated GPU

OpenVoice V2

OpenVoice V2 is a natural dedicated enterprise target when teams want private voice cloning and speech transformation workloads.

Voice generationVoice cloning
CosyVoice 2 sample output
AudioDedicated GPU

CosyVoice 2

CosyVoice 2 is useful for teams that want a modern open speech stack with private enterprise hosting and code-level runtime control.

Speech generationDedicated hosting

3D Models

Dedicated deployment pages for 2 3d models and adjacent enterprise use cases.

2 pages

Hunyuan3D 2 sample output
3DDedicated GPU

Hunyuan3D 2

Hunyuan3D 2 is a good dedicated enterprise page because private 3D generation often involves proprietary product imagery and design workflows.

Text to 3DImage to 3D
TRELLIS sample output
3DDedicated GPU

TRELLIS

TRELLIS is a useful SEO and enterprise target for teams that want modern 3D generation on infrastructure they fully control.

3D generationPrivate asset handling
Enterprise Deployment

Deploy open-source model stacks with enterprise control

Whether you need Stable Diffusion checkpoints, FLUX editing runtimes, private reasoning models, or speech pipelines, the enterprise plan gives you dedicated GPUs, code access, and storage control.

100% privacy for prompts, inputs, and outputs
Code access for custom runtimes and adapters
Bring-your-own S3 for checkpoints and generated assets
Dedicated GPU throughput with no shared queue

Pricing

$1999/month

Starting price for enterprise dedicated GPU plans. Move to higher GPU tiers when you need more VRAM, throughput, or concurrency.

Get Expert Support in Seconds

We're Here to Help.

Want to know more? You can email us anytime at support@modelslab.com

View Docs