--- license: cc-by-nc-4.0 --- [COLING 2025] Idea23D: Collaborative LMM Agents Enable 3D Model Generation from Interleaved Multimodal Inputs # Idea23D evaluation dataset Full inference code can be found at:
GitHub Repo StarsarXiv
## ✨Introduction ![idea23d](./page/idea23d.gif) Based on the LMM we developed Idea23D, a multimodal iterative self-refinement system that enhances any T2I model for automatic 3D model design and generation, enabling various new image creation functionalities togther with better visual qualities while understanding high level multimodal inputs. ## 📜Citations ``` @article{chen2024idea23d, title={Idea-2-3D: Collaborative LMM Agents Enable 3D Model Generation from Interleaved Multimodal Inputs}, author={Junhao Chen and Xiang Li and Xiaojun Ye and Chao Li and Zhaoxin Fan and Hao Zhao}, year={2024}, eprint={2404.04363}, archivePrefix={arXiv}, primaryClass={cs.CV} } ``` ## ⭐️ Star History [![Star History Chart](https://api.star-history.com/svg?repos=yisuanwang/Idea23D&type=Date)](https://star-history.com/#yisuanwang/Idea23D&Date)