Convert Figma logo to code with AI

divamgupta logodiffusionbee-stable-diffusion-ui

Diffusion Bee is the easiest way to run Stable Diffusion locally on your M1 Mac. Comes with a one-click installer. No dependencies or technical knowledge needed.

12,379
607
12,379
369

Top Related Projects

Stable Diffusion web UI

23,395

Invoke is a leading creative engine for Stable Diffusion models, empowering professionals, artists, and enthusiasts to generate and create visual media using the latest AI-driven technologies. The solution offers an industry leading WebUI, and serves as the foundation for multiple commercial products.

High-Resolution Image Synthesis with Latent Diffusion Models

25,061

🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch and FLAX.

Quick Overview

DiffusionBee is a user-friendly macOS application for running Stable Diffusion locally on Apple Silicon Macs. It provides an easy-to-use interface for generating images using AI, without requiring technical knowledge or complex setup processes.

Pros

  • Easy installation and setup for non-technical users
  • Runs locally on Apple Silicon Macs, ensuring privacy and offline usage
  • User-friendly interface for generating and managing AI-generated images
  • Regular updates and improvements based on user feedback

Cons

  • Limited to macOS and Apple Silicon devices
  • May have performance limitations compared to more powerful GPU setups
  • Lacks some advanced features found in other Stable Diffusion implementations
  • Dependent on the development team for updates and new features

Getting Started

  1. Visit the DiffusionBee website
  2. Download the latest version of DiffusionBee for your macOS device
  3. Open the downloaded file and drag the DiffusionBee application to your Applications folder
  4. Launch DiffusionBee from your Applications folder
  5. Follow the on-screen instructions to start generating AI images

Note: This is not a code library, so code examples and detailed quick start instructions are not applicable.

Competitor Comparisons

Stable Diffusion web UI

Pros of stable-diffusion-webui

  • More extensive features and customization options
  • Supports a wider range of models and extensions
  • Active development with frequent updates and community contributions

Cons of stable-diffusion-webui

  • Requires more technical knowledge to set up and use
  • May have higher system requirements for optimal performance
  • Not as user-friendly for beginners compared to DiffusionBee

Code Comparison

DiffusionBee (Python):

def generate_image(prompt, num_inference_steps=50, guidance_scale=7.5):
    with torch.no_grad():
        image = pipeline(prompt, num_inference_steps=num_inference_steps, guidance_scale=guidance_scale).images[0]
    return image

stable-diffusion-webui (Python):

def generate(prompt, steps=50, cfg_scale=7.0, sampler_name='Euler a', batch_size=1):
    p = StableDiffusionProcessing(
        sd_model=shared.sd_model,
        outpath_samples=opts.outdir_samples or opts.outdir_txt2img_samples,
        prompt=prompt,
        steps=steps,
        cfg_scale=cfg_scale,
        sampler_name=sampler_name,
        batch_size=batch_size,
    )
    processed = processing.process_images(p)
    return processed.images

The code comparison shows that stable-diffusion-webui offers more granular control over the generation process, with additional parameters and options available for fine-tuning the output.

23,395

Invoke is a leading creative engine for Stable Diffusion models, empowering professionals, artists, and enthusiasts to generate and create visual media using the latest AI-driven technologies. The solution offers an industry leading WebUI, and serves as the foundation for multiple commercial products.

Pros of InvokeAI

  • More advanced features and customization options
  • Supports multiple AI models and pipelines
  • Active development with frequent updates and community contributions

Cons of InvokeAI

  • More complex setup and installation process
  • Steeper learning curve for beginners
  • Higher system requirements

Code Comparison

DiffusionBee:

def generate_image(prompt, num_inference_steps=50, guidance_scale=7.5):
    with autocast("cuda"):
        image = pipe(prompt, num_inference_steps=num_inference_steps, guidance_scale=guidance_scale).images[0]
    return image

InvokeAI:

def generate_image(prompt, steps=50, cfg_scale=7.5, width=512, height=512, sampler_name='k_lms'):
    generator = Generator(model, device)
    result = generator.generate(prompt, sampler_name=sampler_name, steps=steps, cfg_scale=cfg_scale,
                                width=width, height=height)
    return result.image

The code snippets show that InvokeAI offers more parameters and flexibility in image generation, while DiffusionBee provides a simpler interface. InvokeAI's code allows for specifying dimensions and sampler type, reflecting its more advanced feature set.

High-Resolution Image Synthesis with Latent Diffusion Models

Pros of stablediffusion

  • More comprehensive and flexible codebase for advanced users and researchers
  • Supports a wider range of models and architectures
  • Offers more customization options and fine-tuning capabilities

Cons of stablediffusion

  • Steeper learning curve and more complex setup process
  • Requires more technical knowledge to use effectively
  • May have higher hardware requirements for optimal performance

Code Comparison

diffusionbee-stable-diffusion-ui:

def generate_image(prompt, num_inference_steps=50, guidance_scale=7.5):
    with torch.no_grad():
        latents = torch.randn((1, 4, 64, 64))
        scheduler.set_timesteps(num_inference_steps)
        for t in scheduler.timesteps:
            # Simplified generation process
            latents = pipeline(prompt, latents, t, guidance_scale)
    return latents_to_image(latents)

stablediffusion:

@torch.no_grad()
def sample(model, x, t, c, steps, eta):
    alphas, sigmas = make_schedule(steps, device=x.device)
    for i in trange(steps):
        with torch.enable_grad():
            x = x.detach().requires_grad_()
            t_enc = torch.tensor([t]).to(x.device)
            c_enc = model.encode_first_stage(c)
            e_t = model.apply_model(x, t_enc, c_enc)
        # More advanced sampling process
        x = x - alphas[i] * e_t + sigmas[i] * torch.randn_like(x)
    return x

The code comparison shows that stablediffusion offers more advanced and customizable sampling processes, while diffusionbee-stable-diffusion-ui provides a simpler, more user-friendly approach to image generation.

25,061

🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch and FLAX.

Pros of diffusers

  • More comprehensive library with support for multiple diffusion models and techniques
  • Actively maintained by Hugging Face with frequent updates and improvements
  • Extensive documentation and examples for various use cases

Cons of diffusers

  • Steeper learning curve for beginners compared to DiffusionBee's user-friendly interface
  • Requires more setup and configuration to get started
  • May be overkill for users who only need basic image generation capabilities

Code Comparison

DiffusionBee (simplified usage):

from diffusionbee import DiffusionBee

bee = DiffusionBee()
image = bee.text_to_image("A beautiful sunset over the ocean")
image.save("sunset.png")

diffusers:

from diffusers import StableDiffusionPipeline
import torch

pipe = StableDiffusionPipeline.from_pretrained("runwayml/stable-diffusion-v1-5", torch_dtype=torch.float16)
pipe = pipe.to("cuda")
image = pipe("A beautiful sunset over the ocean").images[0]
image.save("sunset.png")

The diffusers library offers more flexibility and control over the generation process, while DiffusionBee provides a simpler interface for quick image generation. diffusers is better suited for developers and researchers who need advanced features, while DiffusionBee is ideal for users who want a straightforward, GUI-based solution for Stable Diffusion image generation.

Convert Figma logo designs to code with AI

Visual Copilot

Introducing Visual Copilot: A new AI model to turn Figma designs to high quality code using your components.

Try Visual Copilot

README

Diffusion Bee - Stable Diffusion GUI App for MacOS

Twitter

Diffusion Bee is the easiest way to run Stable Diffusion locally on your Intel / M1 Mac. Comes with a one-click installer. No dependencies or technical knowledge needed.

  • Runs locally on your computer no data is sent to the cloud ( other than request to download the weights or unless you chose to upload an image ).
  • If you like Diffusion Bee, consider checking https://Liner.ai , a one-click tool to train machine learning models

Download

Download at https://diffusionbee.com/


For prompt ideas visit https://arthub.ai

Join discord server : https://discord.gg/t6rC5RaJQn

Features

  • Full data privacy - nothing is sent to the cloud ( unless you chose to upload an image )
  • Clean and easy to use UI with one-click installer
  • Image to image
  • Supported models : - SD 1.x, SD 2.x, SD XL, Inpainting, ControlNet, LoRA
  • Download models from the app
  • In-painting
  • Out-painting
  • Generation history
  • Upscaling images
  • Multiple image sizes
  • Optimized for M1/M2 Chips
  • Runs locally on your computer
  • Negative prompts
  • Advanced prompt options
  • ControlNet

How to use

  1. Download and start the application

image

  1. Enter a prompt and click generate

Text to image:

Screenshot 2023-12-13 at 1 19 38 PM

Image to image:

Screenshot 2023-12-13 at 1 14 35 PM

Multiple Apps:

Screenshot 2023-12-13 at 1 11 14 PM

Image to image with mask:

image

Inpainting:

image

image

Advanced AI Canvas:

Screenshot 2023-12-13 at 1 13 59 PM

image

ControlNet:

Screenshot 2023-12-13 at 1 12 43 PM

Screenshot 2023-12-13 at 1 12 22 PM

Download Models:

Screenshot 2023-12-13 at 1 13 08 PM

History:

image

To learn more, visit the documentation.

Requirements

  • Mac with Intel or M1/M2 CPU
  • For Intel : MacOS 12.3.1 or later
  • For M1/M2 : MacOS 11.0.0 or later

License : Stable Diffusion is released under the CreativeML OpenRAIL M license : https://github.com/CompVis/stable-diffusion/blob/main/LICENSE Diffusion Bee is just a GUI wrapper on top of Stable Diffusion, so all the term of Stable Diffusion are applied on the outputs.

References

  1. https://github.com/CompVis/stable-diffusion

  2. https://github.com/madebyollin/maple-diffusion

  3. https://github.com/divamgupta/stable-diffusion-tensorflow

  4. https://github.com/liuliu/swift-diffusion (big thanks to Liu Liu)

  5. https://github.com/huggingface/diffusers