Deploy Dedicated GPU server to run AI models

Deploy Model
Skip to main content
RedCraft | 红潮 CADS | UPdated-May11 | Commercial & Advertising Design System - RED-UNO(In-Context) thumbnail

RedCraft | 红潮 CADS | UPdated-May11 | Commercial & Advertising Design System - RED-UNO(In-Context)

by ModelsLab

REDAIGC FT Model used to match UNO In-Context Generation

(with improved quality compared to F.1 dev)

---

Diffusers 脚本:

https://github.com/bytedance/UNO

Dit-LoRA 权重:

bytedance-research/UNO · Hugging Face

ComfyUI-nodes 组件:

HM-RunningHub/ComfyUI_RH_UNO: This is a UNO ComfyUI plugin

---

propose a highly-consistent data synthesis pipeline to tackle this challenge. This pipeline harnesses the intrinsic in-context generation capabilities of diffusion transformers and generates high-consistency multi-subject paired data. Additionally, we introduce UNO, which consists of progressive cross-modal alignment and universal rotary position embedding. It is a multi-image conditioned subject-to-image model iteratively trained from a text-to-image model. Extensive experiments show that our method can achieve high consistency while ensuring controllability in both single-subject and multi-subject driven generation.

redcraftcadsupdatedmay11commercialadvertisingdesignsystem-redunoincontext
Open Source ModelUnlimited UsageLLMs.txt
API PlaygroundAPI Documentation

Input

Select models...

Per image generation will cost 0.0047$
For premium plan image generation will cost 0.00$ i.e Free.

Output

Idle

Unknown content type

RedCraft | 红潮 CADS | UPdated-May11 | Commercial & Advertising Design System - RED-UNO(In-Context) Readme

REDAIGC FT Model used to match UNO In-Context Generation

(with improved quality compared to F.1 dev)

---

Diffusers 脚本:

https://github.com/bytedance/UNO

Dit-LoRA 权重:

bytedance-research/UNO · Hugging Face

ComfyUI-nodes 组件:

HM-RunningHub/ComfyUI_RH_UNO: This is a UNO ComfyUI plugin

---

propose a highly-consistent data synthesis pipeline to tackle this challenge. This pipeline harnesses the intrinsic in-context generation capabilities of diffusion transformers and generates high-consistency multi-subject paired data. Additionally, we introduce UNO, which consists of progressive cross-modal alignment and universal rotary position embedding. It is a multi-image conditioned subject-to-image model iteratively trained from a text-to-image model. Extensive experiments show that our method can achieve high consistency while ensuring controllability in both single-subject and multi-subject driven generation.