# How to Use JasperAI Depth Controlnet on Flux Dev

{% embed url="<https://youtu.be/zMSedAiXxhQ>" %}

[Check out the example Glif here!](https://glif.app/@araminta_k/glifs/cm1yy68lc000brws3tujhw07p)

#### Step by step

1. Click **Build** to start a new Glif project.
2. Add an image input block.
3. Add a text input block.
4. Add a ComfyUI block.
5. Copy the provided JSON into the Comfyui block.
6. (Optional) Adjust the LoRA input node in the Comfyui block.
7. (Optional) Adjust the controlnet strength nodes in the Comfyui block.
8. Test and publish!

{% tabs %}
{% tab title="ComfyUI JSON" %}
Insert this into the JSON field in the Comfyui block.

```
{
  "3": {
    "inputs": {
      "seed": 376385039400309,
      "steps": 28,
      "cfg": 1,
      "sampler_name": "dpmpp_2m",
      "scheduler": "beta",
      "denoise": 1,
      "model": [
        "78",
        0
      ],
      "positive": [
        "47",
        0
      ],
      "negative": [
        "47",
        1
      ],
      "latent_image": [
        "83",
        0
      ]
    },
    "class_type": "KSampler",
    "_meta": {
      "title": "KSampler"
    }
  },
  "6": {
    "inputs": {
      "text": "{input1}, L1d4r style",
      "clip": [
        "78",
        1
      ]
    },
    "class_type": "CLIPTextEncode",
    "_meta": {
      "title": "CLIP Text Encode (Prompt)"
    }
  },
  "7": {
    "inputs": {
      "text": "",
      "clip": [
        "78",
        1
      ]
    },
    "class_type": "CLIPTextEncode",
    "_meta": {
      "title": "CLIP Text Encode (Prompt)"
    }
  },
  "8": {
    "inputs": {
      "samples": [
        "3",
        0
      ],
      "vae": [
        "27",
        0
      ]
    },
    "class_type": "VAEDecode",
    "_meta": {
      "title": "VAE Decode"
    }
  },
  "9": {
    "inputs": {
      "filename_prefix": "ComfyUI",
      "images": [
        "8",
        0
      ]
    },
    "class_type": "SaveImage",
    "_meta": {
      "title": "Save Image"
    }
  },
  "11": {
    "inputs": {
      "guidance": 4,
      "conditioning": [
        "6",
        0
      ]
    },
    "class_type": "FluxGuidance",
    "_meta": {
      "title": "FluxGuidance"
    }
  },
  "27": {
    "inputs": {
      "vae_name": "ae.sft"
    },
    "class_type": "VAELoader",
    "_meta": {
      "title": "Load VAE"
    }
  },
  "30": {
    "inputs": {
      "image": "{image-input1}"
    },
    "class_type": "LoadImage",
    "_meta": {
      "title": "Load Image"
    }
  },
  "31": {
    "inputs": {
      "preprocessor": "Zoe-DepthMapPreprocessor",
      "resolution": 1024,
      "image": [
        "75",
        0
      ]
    },
    "class_type": "AIO_Preprocessor",
    "_meta": {
      "title": "AIO Aux Preprocessor"
    }
  },
  "47": {
    "inputs": {
      "strength": 0.6,
      "start_percent": 0,
      "end_percent": 0.3,
      "positive": [
        "11",
        0
      ],
      "negative": [
        "7",
        0
      ],
      "control_net": [
        "48",
        0
      ],
      "image": [
        "31",
        0
      ],
      "vae": [
        "27",
        0
      ]
    },
    "class_type": "ControlNetApplyAdvanced",
    "_meta": {
      "title": "Apply ControlNet"
    }
  },
  "48": {
    "inputs": {
      "control_net_name": "flux/jasperai_depth_flux_dev.safetensors"
    },
    "class_type": "ControlNetLoader",
    "_meta": {
      "title": "Load ControlNet Model"
    }
  },
  "72": {
    "inputs": {
      "unet_name": "flux1-dev-fp8.safetensors",
      "weight_dtype": "default"
    },
    "class_type": "UNETLoader",
    "_meta": {
      "title": "Load Diffusion Model"
    }
  },
  "73": {
    "inputs": {
      "clip_name1": "t5xxl_fp8_e4m3fn.safetensors",
      "clip_name2": "clip_l.safetensors",
      "type": "flux"
    },
    "class_type": "DualCLIPLoader",
    "_meta": {
      "title": "DualCLIPLoader"
    }
  },
  "75": {
    "inputs": {
      "width": 1024,
      "height": 1024,
      "upscale_method": "nearest-exact",
      "keep_proportion": false,
      "divisible_by": 2,
      "width_input": [
        "79",
        0
      ],
      "height_input": [
        "79",
        1
      ],
      "crop": "disabled",
      "image": [
        "30",
        0
      ]
    },
    "class_type": "ImageResizeKJ",
    "_meta": {
      "title": "Resize Image"
    }
  },
  "78": {
    "inputs": {
      "repo_id": "glif/LiDAR-Vision",
      "subfolder": "",
      "filename": "Lidar.safetensors",
      "strength_model": 1,
      "strength_clip": 1,
      "model": [
        "72",
        0
      ],
      "clip": [
        "73",
        0
      ]
    },
    "class_type": "HFHubLoraLoader",
    "_meta": {
      "title": "Load HF Lora"
    }
  },
  "79": {
    "inputs": {
      "image": [
        "30",
        0
      ]
    },
    "class_type": "SDXLAspectRatio",
    "_meta": {
      "title": "Image to SDXL compatible WH"
    }
  },
  "83": {
    "inputs": {
      "width": [
        "79",
        0
      ],
      "height": [
        "79",
        1
      ],
      "batch_size": 1
    },
    "class_type": "EmptyLatentImage",
    "_meta": {
      "title": "Empty Latent Image"
    }
  }
}
```

{% endtab %}
{% endtabs %}
