Все модели
Полный список отсортирован по популярности на Replicate.
Stable Diffusion fine tuned on Midjourney v4 images.
Run any image through the Stable Diffusion content filter
An SDXL fine-tune based on Apple Emojis
Proteus v0.2 shows subtle yet significant improvements over Version 0.1. It demonstrates enhanced prompt understanding that surpasses MJ6, while also approaching its stylistic capabilities.
Private instance of real-esrgan
Remove images background
multilingual text2image latent diffusion model
A 70 billion parameter language model from Meta, fine tuned for chat completions
Modify images using M-LSD line detection
A premium text-based image editing model that delivers maximum performance and improved typography generation for transforming images through natural language prompts
Anime-themed text-to-image stable diffusion model
A model for text, audio, and image embeddings in one space
The latest Qwen-Image’s iteration with improved multi-image editing, single-image consistency, and native support for ControlNet
Robust face restoration algorithm for old photos/AI-generated faces
Text-to-Audio (T2A) that offers voice synthesis, emotional expression, and multilingual capabilities. Designed for real-time applications with low latency
chameleonn: one-click face swap (formerly roop)
Create photos, paintings and avatars for anyone in any style within seconds.
GFPGAN aims at developing Practical Algorithms for Real-world Face and Object Restoration
Hyper FLUX 16-step by ByteDance
moondream2 is a small vision language model designed to run efficiently on edge devices
This is an optimised version of the hidream-l1 model using the pruna ai optimisation toolkit!
Generate Pokémon from a text description
Turbo is the fastest and cheapest Ideogram v3. v3 creates images with stunning realism, creative designs, and consistent styles
Recraft V3 (code-named red_panda) is a text-to-image model with the ability to generate long texts, and images in a wide list of styles. As of today, it is SOTA in image generation, proven by the Text-to-Image Benchmark by Artificial Analysis
Practical Image Restoration Algorithms for General/Anime Images
Run any ComfyUI workflow. Guide: https://github.com/replicate/cog-comfyui
Kokoro v1.0 - text-to-speech (82M params, based on StyleTTS2)
Control diffusion models
Google's Imagen 4 flagship model
A very fast and cheap PrunaAI optimized version of Wan 2.2 A14B image-to-video
Meta's flagship 405 billion parameter language model, fine-tuned for chat completions
Meina Mix V11 Model (Text2Img, Img2Img and Inpainting)
Open-weight version of FLUX.1 Kontext
SDXL based text-to-image model applying Distribution Matching Distillation, supporting zero-shot identity generation in 2-5s. https://ai-visionboard.com
Image Restoration Using Swin Transformer
Accelerated transcription, word-level timestamps and diarization with whisperX large-v3
text2img model trained on LAION HighRes and fine-tuned on internal datasets
controlnet 1.1 lineart x realistic-vision-v2.0 (updated to v5)
⚡️ Blazing fast audio transcription with speaker diarization | Whisper Large V3 Turbo | word & sentence level timestamps | prompt
A version of flux-dev, a text to image model, that supports fast fine-tuned lora inference
A background removal model enhanced with better matting
ProteusV0.3: The Anime Update
Coqui XTTS-v2: Multilingual Text To Speech Voice Cloning
LLaVA v1.6: Large Language and Vision Assistant (Mistral-7B)
DeepSeek-V3-0324 is the leading non-reasoning model, a milestone for open source
A 13 billion parameter language model from Meta, fine tuned for chat completions
The CLIP Interrogator is a prompt engineering tool that combines OpenAI's CLIP and Salesforce's BLIP to optimize text prompts to match a given image. Use the resulting prompts with text-to-image models like Stable Diffusion to create cool art!