🏠
Working from home
PhD student, ShowLab @ NUS.
Video Multimodal.
-
National University of Singapore
- Singapore
- qinghonglin.github.io
- @KevinQHLin
Pinned Loading
-
showlab/ShowUI
showlab/ShowUI PublicRepository for ShowUI: One Vision-Language-Action Model for GUI Visual Agent
-
showlab/EgoVLP
showlab/EgoVLP Public[NeurIPS2022] Egocentric Video-Language Pretraining
-
showlab/UniVTG
showlab/UniVTG Public[ICCV2023] UniVTG: Towards Unified Video-Language Temporal Grounding
-
showlab/VLog
showlab/VLog PublicTransform Video as a Document with ChatGPT, CLIP, BLIP2, GRIT, Whisper, LangChain.
-
showlab/Awesome-GUI-Agent
showlab/Awesome-GUI-Agent Public💻 A curated list of papers and resources for multi-modal Graphical User Interface (GUI) agents.
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.