Picture for Lu Sheng

Lu Sheng

From Parts to Whole: A Unified Reference Framework for Controllable Human Image Generation

Add code
Apr 23, 2024
Figure 1 for From Parts to Whole: A Unified Reference Framework for Controllable Human Image Generation
Figure 2 for From Parts to Whole: A Unified Reference Framework for Controllable Human Image Generation
Figure 3 for From Parts to Whole: A Unified Reference Framework for Controllable Human Image Generation
Figure 4 for From Parts to Whole: A Unified Reference Framework for Controllable Human Image Generation
Viaarxiv icon

Self-Supervised Monocular Depth Estimation in the Dark: Towards Data Distribution Compensation

Add code
Apr 22, 2024
Viaarxiv icon

RH20T-P: A Primitive-Level Robotic Dataset Towards Composable Generalization Agents

Add code
Mar 28, 2024
Figure 1 for RH20T-P: A Primitive-Level Robotic Dataset Towards Composable Generalization Agents
Figure 2 for RH20T-P: A Primitive-Level Robotic Dataset Towards Composable Generalization Agents
Figure 3 for RH20T-P: A Primitive-Level Robotic Dataset Towards Composable Generalization Agents
Figure 4 for RH20T-P: A Primitive-Level Robotic Dataset Towards Composable Generalization Agents
Viaarxiv icon

Assessment of Multimodal Large Language Models in Alignment with Human Values

Add code
Mar 26, 2024
Figure 1 for Assessment of Multimodal Large Language Models in Alignment with Human Values
Figure 2 for Assessment of Multimodal Large Language Models in Alignment with Human Values
Figure 3 for Assessment of Multimodal Large Language Models in Alignment with Human Values
Figure 4 for Assessment of Multimodal Large Language Models in Alignment with Human Values
Viaarxiv icon

MineDreamer: Learning to Follow Instructions via Chain-of-Imagination for Simulated-World Control

Add code
Mar 19, 2024
Figure 1 for MineDreamer: Learning to Follow Instructions via Chain-of-Imagination for Simulated-World Control
Figure 2 for MineDreamer: Learning to Follow Instructions via Chain-of-Imagination for Simulated-World Control
Figure 3 for MineDreamer: Learning to Follow Instructions via Chain-of-Imagination for Simulated-World Control
Figure 4 for MineDreamer: Learning to Follow Instructions via Chain-of-Imagination for Simulated-World Control
Viaarxiv icon

From GPT-4 to Gemini and Beyond: Assessing the Landscape of MLLMs on Generalizability, Trustworthiness and Causality through Four Modalities

Add code
Jan 29, 2024
Figure 1 for From GPT-4 to Gemini and Beyond: Assessing the Landscape of MLLMs on Generalizability, Trustworthiness and Causality through Four Modalities
Figure 2 for From GPT-4 to Gemini and Beyond: Assessing the Landscape of MLLMs on Generalizability, Trustworthiness and Causality through Four Modalities
Figure 3 for From GPT-4 to Gemini and Beyond: Assessing the Landscape of MLLMs on Generalizability, Trustworthiness and Causality through Four Modalities
Figure 4 for From GPT-4 to Gemini and Beyond: Assessing the Landscape of MLLMs on Generalizability, Trustworthiness and Causality through Four Modalities
Viaarxiv icon

Data-Free Generalized Zero-Shot Learning

Add code
Jan 28, 2024
Figure 1 for Data-Free Generalized Zero-Shot Learning
Figure 2 for Data-Free Generalized Zero-Shot Learning
Figure 3 for Data-Free Generalized Zero-Shot Learning
Figure 4 for Data-Free Generalized Zero-Shot Learning
Viaarxiv icon

Multi-modality Affinity Inference for Weakly Supervised 3D Semantic Segmentation

Add code
Dec 29, 2023
Viaarxiv icon

MP5: A Multi-modal Open-ended Embodied System in Minecraft via Active Perception

Add code
Dec 13, 2023
Viaarxiv icon

EpiDiff: Enhancing Multi-View Synthesis via Localized Epipolar-Constrained Diffusion

Add code
Dec 11, 2023
Viaarxiv icon