Opendream brings much needed and familiar features, such as layering, non-destructive editing, portability, and easy-to-write extensions, to your Stable Diffusion workflows.
- Clone this repository.
- Navigate to this project within your terminal and run
sh ./run_opendream.sh
. After ~30 seconds, both the frontend and backend of the Opendream system should be up and running.
Diffusion models have emerged as powerful tools in the world of image generation and manipulation. While they offer significant benefits, these models are often considered black boxes due to their inherent complexity. The current diffusion image generation ecosystem is defined by tools that allow one-off image manipulation tasks to control these models - text2img, in-painting, pix2pix, among others.
For example, popular interfaces like Automatic1111, Midjourney, and Stability.AI's DreamStudio only support destructive editing: each edit "consumes" the previous image. This means users cannot easily build off of previous images or run multiple experiments on the same image, limiting their options for creative exploration.
Non-destructive editing is a method of image manipulation that preserves the original image data while allowing users to make adjustments and modifications without overwriting previous work. This approach facilitates experimentation and provides more control over the editing process by using layers and masks. When you delete a layer, all layers after it also get deleted. This guarantees that all layers currently on the canvas are a product of other existing layers. This also allows one to deterministically "replay" a workflow.
Like Photoshop, Opendream supports non-destructive editing out of the box. Learn more about the principles of non-destructive editing in Photoshop here.
Users can also save their current workflows into a portable file format that can be opened up at a later time or shared with collaborators. In this context, a "state" is just a JSON file describing all of the current layers and how they were created.
As the open-source ecosystem flourishes around these models and tools, extensibility has also become a major concern. While Automatic1111 does offer extensions, they are often difficult to program, use, and install. It is far from being as full-featured as an application like Adobe Photoshop.
As new features for Stable Diffusion, like ControlNet, are released, users should be able to seamlessly integrate them into their artistic workflows with minimal overload and time.
Opendream makes writing and using new diffusion features as simple as writing a Python function. Keep reading to learn how.
From the get-go, Opendream supports two key primitive operations baked into the core system: dream
and mask_and_inpaint
. In this repository, extensions for instruct_pix2pix
, controlnet_canny
, controlnet_openpose
, and sam
(Segment Anything) are provided.
Any image manipulation logic can be easily written as an extension. With extensions, you can also decide how certain operations work. For example, you can override the dream
operation to use OpenAI's DALL-E instead or call a serverless endpoint on a service like AWS or Replicate. Here's an example using Baseten.
There are two ways to load extensions.
- Install a pre-written one through the Web UI.
- (Manual) Download a valid extension file (or write one yourself!) and add it to the
opendream/extensions
folder. Instructions for writing your own extension are below.
Here is a sampling of currently supported extensions. You can use the links to install any given extension through the Web UI.
Extension | Link |
---|---|
OpenAI's DALL-E | File |
Serverless Stable Diffusion | File |
Instruct Pix2Pix | File |
ControlNet Canny | File |
ControlNet Openpose | File |
Segment Anything | File |
PhotoshopGPT | Gist |
Feel free to make a PR if you create a useful extension!
Users can write their own extensions as follows:
- Create a new Python file in the
opendream/extensions
folder. - Write a method with type hints and a
@opendream.define_op
decorator. This decorator registers this method with the Opendream backend.
The method has a few requirements:
- Parameters must have type hints. These enable the backend to generate a schema for the input which is parsed into form components on the frontend. Valid types include:
str
,int
,float
,Layer
,MaskLayer
, orImageLayer
. - The only valid return types are a
Layer
or a list ofLayer
objects.
Opendream was built by Varun Shenoy, Eric Lou, Shashank Rammoorthy, and Rahul Shiv as a part of Stanford's CS 348K.
Feel free to provide any contibutions you deem necessary or useful. This project is licensed under the MIT License.