pink hair guy in glasses, photograph, sporty body, cinematic lighting, clear eyes, perfect face, blush, beautiful nose, beautiful eyes, detailed eyes
Anime Diffusion2 is a latent text-to-image diffusion model based on Vintedois (22h) Diffusion model trained on BLIP captions for Danbou set + Demon slayer + arts from 4ch
This model is open access and available to all, with a CreativeML OpenRAIL-M license further specifying rights and usage. The CreativeML OpenRAIL License specifies:
This model can be used for entertainment purposes and as a generative art assistant.
import torch
from torch import autocast
from diffusers import StableDiffusionPipeline
pipe = StableDiffusionPipeline.from_pretrained(
'AlexWortega/AnimeDiffuion2',
torch_dtype=torch.float32
).to('cuda')
negative_prompt = """low-res, duplicate, poorly drawn face, ugly, undetailed face"""
d = 'white girl'
prompt = f"pink hair guy in glasses, photograph, sporty body, cinematic lighting, clear eyes, perfect face, blush, beautiful nose, beautiful eyes, detailed eyes"
num_samples = 1
with torch.inference_mode():
images = pipe([prompt] * num_samples,
negative_prompt = [negative_prompt]*num_samples,
height=512, width=512,
num_inference_steps=50,
guidance_scale=8,
).images
images[0].save("test.png")
This project would not have been possible without the incredible work by the CompVis Researchers .
In order to reach me, here is my blog: