--- base_model: Alpha-VLLM/Lumina-Image-2.0 library_name: diffusers license: apache-2.0 instance_prompt: a puppy, yarn art style widget: - text: a puppy in a pond, yarn art style output: url: yarn_lora.png - text: a puppy in a pond, yarn art style (dark env) output: url: yarn_lora_You_are_an_assistant_designed_to_generate_superior_images_with_a_dark_overall_theme.png - text: a puppy in a pond, yarn art style (shiny env) output: url: yarn_lora_You_are_an_assistant_designed_to_generate_superior_images_with_a_bright_and_shiny_overall_.png tags: - text-to-image - diffusers-training - diffusers - lora - lumina2 - lumina2-diffusers - template:sd-lora --- # Lumina2 DreamBooth LoRA - trained-lumina2-lora-yarn ## Model description These are `trained-lumina2-lora-yarn` DreamBooth LoRA weights for [Alpha-VLLM/Lumina-Image-2.0](https://hf.co/Alpha-VLLM/Lumina-Image-2.0). The weights were trained using [DreamBooth](https://dreambooth.github.io/) with the [Lumina2 diffusers trainer](https://github.com/huggingface/diffusers/blob/main/examples/dreambooth/README_lumina2.md). ## Trigger words You should use `yarn art style` to trigger the image generation. The following `system_prompt` was also used used during training (ignore if `None`): None. ## Download model [Download the *.safetensors LoRA]({repo_id}/tree/main) in the Files & versions tab. ## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers) ```py import torch from diffusers import Lumina2Text2ImgPipeline pipe = Lumina2Text2ImgPipeline.from_pretrained( "Alpha-VLLM/Lumina-Image-2.0", torch_dtype=torch.bfloat16 ).to("cuda") pipe.load_lora_weights("trained-lumina2-lora-yarn") prompt = "a puppy in a pond, yarn art style" image = pipe( prompt, negative_prompt="bad quality, worse quality, degenerate quality", guidance_scale=6, num_inference_steps=35, generator=torch.manual_seed(0) ).images[0] ``` For more details, including weighting, merging and fusing LoRAs, check the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters). ## Results The model benefits from `system_prompt`. Here is a comparison across different system prompts:
No system prompt "Dark surrounding"
system prompt
"Sunny surrounding"
system prompt
No system prompt image Dark surrounding image Sunny surrounding image
Original prompt: a puppy in a pond, yarn art style
Code ```py import torch from diffusers import Lumina2Text2ImgPipeline pipe = Lumina2Text2ImgPipeline.from_pretrained( "Alpha-VLLM/Lumina-Image-2.0", torch_dtype=torch.bfloat16 ).to("cuda") system_prompts = [ None, "You are an assistant designed to generate superior images with a dark overall theme.", "You are an assistant designed to generate superior images with a bright and shiny overall theme." ] pipe.load_lora_weights("trained-lumina2-lora-yarn") prompt = "a puppy in a pond, yarn art style" for sp in system_prompts: filename = "yarn_lora" image = pipe( prompt, negative_prompt="bad quality, worse quality, degenerate quality", system_prompt=sp, guidance_scale=6, num_inference_steps=35, generator=torch.manual_seed(0) ).images[0] if sp: filename += "_" + "_".join(sp.split(" ")).replace(",", "").replace(".", "") filename = filename[:100] image.save(f"{filename}.png") ```