# Lotus-2
**Repository Path**: monkeycc/Lotus-2
## Basic Information
- **Project Name**: Lotus-2
- **Description**: No description available
- **Primary Language**: Python
- **License**: Not specified
- **Default Branch**: main
- **Homepage**: None
- **GVP Project**: No
## Statistics
- **Stars**: 0
- **Forks**: 0
- **Created**: 2025-12-03
- **Last Updated**: 2025-12-03
## Categories & Tags
**Categories**: Uncategorized
**Tags**: None
## README
#
Lotus-2: Advancing Geometric Dense Prediction with Powerful Image Generative Model
[](https://lotus-2.github.io/)
[](https://arxiv.org/abs/2512.01030)
[-yellow)](https://huggingface.co/spaces/haodongli/Lotus-2_Depth)
[-yellow)](https://huggingface.co/spaces/haodongli/Lotus-2_Normal)
[Jing He](https://scholar.google.com/citations?hl=en&user=RsLS11MAAAAJ)1,
[Haodong Li](https://haodong-li.com/)12✱,
[Mingzhi Sheng]()1✱,
[Ying-Cong Chen](https://www.yingcong.me/)13✉
1HKUST(GZ)
2UC San Diego
3HKUST
✱Both authors contributed equally.
✉Corresponding author.

**We present Lotus-2, a two-stage deterministic framework for monocular geometric dense prediction.** Our method leverages pre-trained generative model as a deterministic world prior to achieve **new state-of-the-art accuracy** while requiring **remarkably minimal data** (trained on only **0.66%** of the samples used by MoGe-2). This figure demonstrates Lotus-2's robust zero-shot generalization with sharp geometric details, especially in challenging cases like oil paintings and transparent objects.
đđđ **Please also check the** [**Project Page**](https://lotus3d.github.io/) **and** [**Github Repo**](https://github.com/EnVision-Research/Lotus) **our prior work: Lotus!** đđđ
## đĸ News
- 2025-12-01: [Paper](https://arxiv.org/abs/2512.01030) released!
- 2025-11-28: The inference code and HuggingFace demo ([Depth](https://huggingface.co/spaces/haodongli/Lotus-2_Depth) & [Normal](https://huggingface.co/spaces/haodongli/Lotus-2_Normal)) are available!
## đ ī¸ Setup
This installation was tested on: Ubuntu 20.04 LTS, Python 3.10, CUDA 12.3, NVIDIA A800-SXM4-80GB.
1. Be sure you have a GPU with at least **40GB** memory.
2. Clone the repository (requires git):
```
git clone https://github.com/EnVision-Research/Lotus-2.git
cd Lotus-2
```
3. Install dependencies (requires conda):
```
conda create -n lotus2 python=3.10 -y
conda activate lotus2
pip install -r requirements.txt
```
4. Be sure you have access to [`black-forest-labs/FLUX.1-dev`](https://huggingface.co/black-forest-labs/FLUX.1-dev).
5. Login your huggingface account via (if you want to switch account, run `hf auth logout` at first):
```
hf auth login
```
## đ¤ Gradio Demo
1. Online demo: [Depth](https://huggingface.co/spaces/haodongli/Lotus-2_Depth) & [Normal](https://huggingface.co/spaces/haodongli/Lotus-2_Normal)
2. Local demo:
- For **depth** estimation, run:
```
python app.py depth
```
- For **normal** estimation, run:
```
python app.py normal
```
## đšī¸ Inference
1. Place your images in a directory, for example, under `./assets/in-the-wild_example` (where we have already prepared several examples).
2. Run the inference command:
```
sh infer.sh
```
## đ Evaluation
1. Prepare benchmark datasets:
- For **depth** estimation, please download the [Marigold evaluation datasets](https://share.phys.ethz.ch/~pf/bingkedata/marigold/evaluation_dataset/) via:
```
cd datasets/eval/depth/
wget -r -np -nH --cut-dirs=4 -R "index.html*" -P . https://share.phys.ethz.ch/~pf/bingkedata/marigold/evaluation_dataset/
```
- For **normal** estimation, please (manually) download the [DSINE evaluation datasets](https://drive.google.com/drive/folders/1t3LMJIIrSnCGwOEf53Cyg0lkSXd3M4Hm?usp=drive_link) (`dsine_eval.zip`) under: `datasets/eval/normal/` and unzip it.
2. Run the evaluation command (modify the `TASK_NAME` in `eval.sh` to switch tasks):
```
sh eval.sh
```
## đ Citation
If you find our work useful in your research, please consider citing our paper:
```bibtex
@article{he2025lotus,
title={Lotus-2: Advancing Geometric Dense Prediction with Powerful Image Generative Model},
author={He, Jing and Li, Haodong and Sheng, Mingzhi and Chen, Ying-Cong},
journal={arXiv preprint arXiv:2512.01030},
year={2025}
}
```