My research lies at the intersection of AI and real-world data, with a focus on computer vision, large language models (LLMs), and multimodal AI systems. I’m especially interested in developing intelligent methods that can understand complex signals, images, and video, while also integrating multiple data modalities to build systems that are both practical and scalable.
Current research projects include:
Computer vision and Signal and Image/Video Processing
- Sign language analysis
- Deep learning for real-time sport video analysis
- Medical signal/image processing using data-driven models
LLMs and Multimodal AI
- Multimodal learning combining text and vision
- Retrieval-augmented and context-aware AI systems