
FLUX.1 Depth [dev] is a 12 billion parameter rectified flow transformer capable of generating an image based on a text description while following the structure of a given input image. For more information, please read our blog post.
FLUX.1 Depth [dev] more efficient.FLUX.1 [dev] Non-Commercial License.We provide a reference implementation of FLUX.1 Depth [dev], as well as sampling code, in a dedicated github repository.
Developers and creatives looking to build on top of FLUX.1 Depth [dev] are encouraged to use this as a starting point.
FLUX.1 Depth [pro] is available in our API bfl.ml

To use FLUX.1-Depth-dev with the 🧨 diffusers python library, first install or upgrade diffusers and image_gen_aux.
pip install -U diffusers pip install git+https://github.com/asomoza/image_gen_aux.git
Then you can use FluxControlPipeline to run the model
import torch
from diffusers import FluxControlPipeline, FluxTransformer2DModel
from diffusers.utils import load_image
from image_gen_aux import DepthPreprocessor
pipe = FluxControlPipeline.from_pretrained("black-forest-labs/FLUX.1-Depth-dev", torch_dtype=torch.bfloat16).to("cuda")
prompt = "A robot made of exotic candies and chocolates of different kinds. The background is filled with confetti and celebratory gifts."
control_image = load_image("https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/robot.png")
processor = DepthPreprocessor.from_pretrained("LiheYoung/depth-anything-large-hf")
control_image = processor(control_image)[0].convert("RGB")
image = pipe(
prompt=prompt,
control_image=control_image,
height=1024,
width=1024,
num_inference_steps=30,
guidance_scale=10.0,
generator=torch.Generator().manual_seed(42),
).images[0]
image.save("output.png")
To learn more check out the diffusers documentation
The model and its derivatives may not be used
This model falls under the FLUX.1 [dev] Non-Commercial License.