Skip to content
@hustvl

HUST Vision Lab

HUST Vision Lab of the School of EIC in HUST. Lab Lead @xinggangw

Welcome to the Vision Lab @ HUST!

🙋‍♀️ Introduction

Hello! This is the GitHub space for the Vision Lab led by Professor Xinggang Wang. We are based at the Artificial Intelligence Institute, School of Electronic Information and Communications, Huazhong University of Science and Technology (HUST).

Our research focuses on computer vision and deep learning. We are particularly interested in:

  • Multimodal Foundation Models
  • Visual Representation Learning
  • Object Detection, Segmentation, and Tracking

Our group strives to push the boundaries of visual intelligence and has produced highly influential works in the field, including CCNet, Mask Scoring R-CNN, FairMOT, ByteTrack, and EVA.

🌈 Contribution Guidelines & Collaboration

We actively contribute to the research community through publications and open-source projects.

  • Research Collaboration: We are open to collaborations in our areas of interest. Please feel free to reach out to Prof. Xinggang Wang (xgwang # hust.edu.cn).
  • Prospective Students: Our group has a strong track record of mentoring Ph.D. and Master's students who lead impactful publications. Interested students can find more information on Prof. Wang's faculty page.
  • Using Our Code: You are welcome to explore and use the code in our repositories. Please ensure you cite the corresponding publications appropriately. Specific details can usually be found in the README files of individual repositories.
  • Contributing to Projects: For guidelines on contributing to specific projects (e.g., bug reports, pull requests), please check the individual repositories.

👩‍💻 Useful Resources

Pinned Loading

  1. Vim Public

    [ICML 2024] Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model

    Python 3.4k 231

  2. 4DGaussians Public

    [CVPR 2024] 4D Gaussian Splatting for Real-Time Dynamic Scene Rendering

    Jupyter Notebook 2.6k 224

  3. GaussianDreamer Public

    [CVPR 2024] GaussianDreamer: Fast Generation from Text to 3D Gaussians by Bridging 2D and 3D Diffusion Models

    Python 756 39

  4. VAD Public

    [ICCV 2023] VAD: Vectorized Scene Representation for Efficient Autonomous Driving

    Python 921 105

  5. MapTR Public

    [ICLR'23 Spotlight & IJCV'24] MapTR: Structured Modeling and Learning for Online Vectorized HD Map Construction

    Python 1.3k 203

  6. SparseInst Public

    [CVPR 2022] SparseInst: Sparse Instance Activation for Real-Time Instance Segmentation

    Python 602 73

Repositories

Showing 10 of 104 repositories
  • ControlAR Public

    [ICLR 2025] ControlAR: Controllable Image Generation with Autoregressive Models

    Python 236 Apache-2.0 7 4 0 Updated Apr 24, 2025
  • .github Public
    0 0 0 0 Updated Apr 24, 2025
  • MaTVLM Public
    Python 39 Apache-2.0 2 1 0 Updated Apr 24, 2025
  • OmniMamba Public

    OmniMamba: Efficient and Unified Multimodal Understanding and Generation via State Space Models

    Python 121 MIT 4 3 1 Updated Apr 9, 2025
  • LightningDiT Public

    [CVPR 2025 Oral] Reconstruction vs. Generation: Taming Optimization Dilemma in Latent Diffusion Models

    Python 709 MIT 19 0 0 Updated Apr 8, 2025
  • WeakSAM Public

    [ACM MM 2024] WeakSAM: Segment Anything Meets Weakly-supervised Instance-level Recognition

    Python 50 3 5 0 Updated Apr 8, 2025
  • DiffusionDrive Public

    [CVPR 2025 Highlight] Truncated Diffusion Model for Real-Time End-to-End Autonomous Driving

    Python 666 MIT 43 5 0 Updated Apr 4, 2025
  • Python 56 1 2 0 Updated Apr 1, 2025
  • AlphaDrive Public

    Unleashing the Power of VLMs in Autonomous Driving via Reinforcement Learning and Reasoning

    Python 207 Apache-2.0 9 9 0 Updated Mar 26, 2025
  • GaussTR Public

    [CVPR 2025] GaussTR: Foundation Model-Aligned Gaussian Transformer for Self-Supervised 3D Spatial Understanding

    Python 131 MIT 6 1 0 Updated Mar 25, 2025