Vision and Video Dynamics Lab
Hyeongmin Lee · SeoulTech, Dept. of Electronic Engineering
Field
Image · Video
One agenda · four pillars
Low-level video enhancement
Enhancement
Compression
All tasks reduce to precise modeling of inter-frame motion.
Low-level video enhancement
Real-ESRGAN · Wang et al. ICCVW 2021
Low-level video enhancement
Low-level video enhancement
Temporal semantics & representation learning
From pixels to meaning — what changes, and why.
Obama is playfully pressing down on the scale with his foot to make the man weighing himself appear heavier. The aides nearby are laughing at the prank.
A group of men in suits are standing in a hallway, and one of them is using a weighing scale.
Same question, two attempts. Modern VLMs can do situational reasoning — but the difference between getting the joke and missing it is exactly what we work on.
The model says 5 fingers — but the emoji shows 4 fingers + 1 thumb. Counting and strict definitions still trip up state-of-the-art VLMs.
No model — and no human — can tell.
Direction of motion, intent, before-and-after — these only exist between frames.
Image understanding is not video understanding. We need temporal models that reason about change.
3D scene reconstruction & rendering
From 2D observations to 3D — and into time, when scenes move.
3D scene reconstruction & rendering
3D Gaussian Splatting for Real-Time Radiance Field Rendering · Kerbl et al. SIGGRAPH 2023
3D scene reconstruction & rendering
Predicting — and interacting with — the visual world
Don't just predict the world. Act in it, and watch it respond.
Generative simulation of the visual world
Generative simulation of the visual world
DreamDojo: A Generalist Robot World Model from Large-Scale Human Videos · Gao et al. NVIDIA 2026
Why one lab, four pillars
Where this leads
… and graduate school, startups,
or anywhere visual AI is used.
Why this field
All you need is an idea and a laptop. Research happens wherever you are.
The field moves fast. New questions open up every week — there is always something to chase.
A paper today can be a product people use within months. Your work reaches real users — quickly.
ViViD Lab
Research, course questions, or just curious — drop by anytime.
Scan