How to use from the
Use from the
Diffusers library
pip install -U diffusers transformers accelerate
import torch
from diffusers import DiffusionPipeline
from diffusers.utils import load_image

# switch to "mps" for apple devices
pipe = DiffusionPipeline.from_pretrained("lavinal712/sd-control-lora-segmentation", dtype=torch.bfloat16, device_map="cuda")

prompt = "Turn this cat into a dog"
input_image = load_image("https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/diffusers/cat.png")

image = pipe(image=input_image, prompt=prompt).images[0]

Model Card for lavinal712/sd-control-lora-segmentation

Model Description

This is controlnet weight trained on runwayml/stable-diffusion-v1-5 with segmentaion.

Training

This model was trained using a Segmented dataset based on the SAM-LLaVA-Captions10M Dataset. Stable Diffusion v1.5 checkpoint was used as the base model for the controlnet.

Training Method

Downloads last month
11
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train lavinal712/sd-control-lora-segmentation