Professor Elhoseiny’s research focuses on developing affective artificial intelligence that understands and generates novel visual content. He has contributed to and led numerous seminal works of affective AI art creation.

Biography

Mohamed Elhoseiny is an associate professor in the Computer Science (CS) Program at KAUST, a senior member honoree of AAAI and IEEE, and the principal investigator of the KAUST Vision-CAIR Research Group. Before joining KAUST, he was a visiting faculty member at Stanford Computer Science Department, a visiting faculty member at Baidu Research, and a postdoctoral researcher at Facebook AI Research.

Elhoseiny earned his Ph.D. in 2016 from Rutgers University. He received a B.Sc. degree in 2006 and an M.S. degree in 2010, both in computer systems from Ain Shams University.

His work has received numerous recognitions, including third place at the Data+AI Summit hackathon at San Francisco held in May 2024 (200 participants) with a multimodal LLM hack called HomeGPT. He was selected as an MIT 35 under 35 semi-finalist in 2020. He received the Best Paper Award at the 2018 European Conference on Computer Vision (ECCV) Workshop on Fashion, Art, and Design for his research "DesIGN: Design Inspiration from Generative Networks." He also received the Doctoral Consortium Award at the 2016 Conference on Computer Vision and Pattern Recognition (CVPR) and an NSF Fellowship for his "Write-a-Classifier Project" in 2014. His research on creative art generation has been featured in New Scientist Magazine and MIT Technology Review, which also highlighted his work on lifelong learning.

Professor Elhoseiny’s contributions include zero-shot learning, which was featured at the United Nations, and his creative AI work was featured in MIT Technology Review, New Scientist Magazine, Forbes Science, and HBO's Silicon Valley. He has served as an Area Chair at major CV/AI conferences, including CVPR21, ICCV21, IJCAI22, ECCV22, ICLR23, CVPR23, ICCV’23, NeurIPS23, ICLR’24, CVPR’24, ECCV’24, SG Asia’24, and has organized Closing the Loop Between Vision and Language workshops at ICCV’15, ICCV’17, ICCV’19, ICCV’21, ICCV’23.

He has been involved in several pioneering works in affective AI art creation and has authored or co-authored numerous award-winning papers.

Research Interests

His primary research interests are in computer vision, focusing on efficient multimodal learning with limited data in zero- and few-shot learning, and Vision and language. He is also interested in Affective AI and especially in understanding and generating novel visual content, such as art and fashion.

Awards and Distinctions

  • IEEE Senior Member, 2024
  • Program Chair and Co-founder of the SAAI Initiative (Art& AI Symposium and multi-stage Hackathon), 2024
  • Best paper award for his work on creative fashion generation at ECCV workshop, Tamara Berg of UNC chapel hill and sponsored by IBM Research and JD AI Research, 2018
  • Doctoral Consortium award, CVPR, 2016
  • NSF Fellowship for Write-a-Classifier project, 2014
  • Best Paper Award, ECCV workshop on Art and Fashion, 2018
  • Nominated to represent Facebook at the United Nations Biodiversity conference by Manohar Paluri , Director of Facebook AI and Jerome Presenti, VP of AI at Facebook, 2024
  • Doctorial Consortium award, International Conference on Computer Vision and Pattern Recognition (CVPR), 2016
  • Best Research-Intern Award (silver), Vision and Learning Group, Center of Vision Technologies, SRI International, 2014
  • NSF Fellowship, 2014
  • Nominated to represent Facebook AI Research (FAIR) at the “Best of AI meeting” that was held at “Disney Innovation Group” to demonstrate Creative Adversarial Networks (CAN), 2017

Education

Doctor of Philosophy (Ph.D.)
Computer Science, Rutgers University, United States, 2016
Master of Science (M.S.)
Computer Science, Rutgers University, United States, 2014
Master of Science (M.S.)
Computer Systems, Ain Shams University, Egypt, 2010
Bachelor of Science (B.S.)
Computer Systems, Ain Shams University, Egypt, 2006

Related Media

Patents

  • Adobe Research: Mohamed Elhoseiny, Scott Cohen, Wbalter Chang, Brian Price, Ahmed Elgammal, “Structured knowledge modeling, extraction and localization from images”, US Patent, 2019.
  • SRI International: Hui Cheng, Jingen Liu, Harpreet Sawhney, Mohamed Elhoseiny, “Zero-shot event detection using semantic embedding”, US, Patent, 2021.