About Hui Wu
I am a senior research scientist and a manager at the Distributed MLOps group @ IBM Research.
I am passionate about developing new AI applications that drive real-world impact, particulary around Edge AI use cases.
My current interest includes vision and language, neural-symbolic learning, computer vision for fashion and more
recently, distributed MLOps.
I joined IBM Research since 2015, and worked in the computer vision and multimedia group at
IBM Research AI (2015-2020), and
MIT-IBM Watson AI Lab (2020-2021).
Prior to joining IBM, I received my PhD in Computer Science from UNC Charlotte,
with a thesis focus on machine learning
techniques applied to image set analysis problems.
Contact me: wuhu AT us.ibm.com
Updates
-
Mar. 2021: Two papers accepted at CVPR 2021!
-
Dec. 2020: Paper "NASTransfer: Analyzing Architecture Transferability in Large Scale Neural Architecture Search" accepted at AAAI 2021.
-
Jun. 2020: We are hosting Fashion IQ challenge at the third workshop on Computer Vision for Fashion, Art and Design
at CVPR 2020. Please see the article at CVPR Daily.
-
Sep. 2019: Drill-down: Interactive Retrieval of Complex Scenes using Natural Language Queries accepted at NeurIPS 2019.
-
Jul. 2019: We are hosting Fashion IQ challenge at ICCV 2019.
-
May. 2019: I am co-chairing the second workshop on Computer Vision for Fashion, Art and Design
at ICCV 2019.
I am also co-chairing Linguistics Meets Image and Video Retrieval Workshop
at ICCV 2019.
- Apr. 2019: Our demo on interactive fashion retrieval accepted at CVPR 2019 demo track.
- Sep. 2018: Dialog Based Interactive Image Retrieval accepted at NeurIPS 2018.
-
Sep. 2018: I was co-chairing the first workshop on Computer Vision for Fashion, Art and Design
at ECCV 2018.
-
Jan. 2018: Worked with fashion designers from Fashion Institute of Technology on exploring computer vision to enhance fashion design process. [Summary Video]
[Women's Wear Daily]
[Town & Country]
Selected Projects
Dialog-based Interactive Image Retrieval
Dialog-based Interactive Image Retrieval
Xiaoxiao Guo*, Hui Wu*, Yu Cheng, Steven J. Rennie, Gerald Tesauro and Rogério S. Feris (* equal contribution)
NeurIPS 2018
[PDF]
[CODE]
[DEMO]
Overview
We proposed a novel type of dialog agent for the task of interactive image retrieval.
Recently, there has been a rapid rise of research interest in visually grounded conversational
agents, driven by the progress of deep learning techniques for both image and natural
language understanding. A few interesting application scenarios have been explored by
recent work, such as collaborative drawing, visual dialog and object guessing game.
In this work, we tested the value of visually grounded dialog agents in a practical and yet
challenging context. Specifially, we proposed a novel framework of image retrieval system which learns to seek
natural and expressive dialog feedbacks from the user and iteratively refine the retrieval result.
Semantic-aware Food Visual Recognition
Learning to make better mistakes - Semantics-aware visual food recognition
Hui Wu, Michele Merler, Rosario Uceda-Sosa and John R. Smith
ACM Multimedia, 2016
[PDF]
[Watson API]
The growing popularity of fitness applications and people’s need for
easy logging of calorie consumption on mobile devices has
made accurate food visual recognition increasingly desireable.
In this project, we proposed a visual food recognition framework that integrates
the semantic relationships among fine-grained food classes.