default search action
Mohamed Elhoseiny
Person information
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
2020 – today
- 2024
- [j4]Deyao Zhu, Jun Chen, Kilichbek Haydarov, Xiaoqian Shen, Wenxuan Zhang, Mohamed Elhoseiny:
ChatGPT Asks, BLIP-2 Answers: Automatic Questioning Towards Enriched Visual Descriptions. Trans. Mach. Learn. Res. 2024 (2024) - [c82]Eslam Abdelrahman, Pengzhan Sun, Li Erran Li, Mohamed Elhoseiny:
ImageCaptioner2: Image Captioner for Image Captioning Bias Amplification Assessment. AAAI 2024: 20902-20911 - [c81]Kilichbek Haydarov, Aashiq Muhamed, Xiaoqian Shen, Jovana Lazarevic, Ivan Skorokhodov, Chamuditha Jayanga Galappaththige, Mohamed Elhoseiny:
Adversarial Text to Continuous Image Generation. CVPR 2024: 6316-6326 - [c80]Faizan Farooq Khan, Diana Kim, Divyansh Jha, Youssef Mohamed, Hanna H. Chang, Ahmed Elgammal, Luba Elliott, Mohamed Elhoseiny:
AI Art Neural Constellation: Revealing the Collective and Contrastive State of AI-Generated and Human Art. CVPR Workshops 2024: 7470-7478 - [c79]Habib Slim, Mohamed Elhoseiny:
ShapeWalk: Compositional Shape Editing Through Language-Guided Chains. CVPR 2024: 22574-22583 - [c78]Wenxuan Zhang, Paul Janson, Rahaf Aljundi, Mohamed Elhoseiny:
Overcoming Generic Knowledge Loss with Selective Parameter Update. CVPR 2024: 24046-24056 - [c77]Kilichbek Haydarov, Xiaoqian Shen, Avinash Madasu, Mahmoud Salem, Li-Jia Li, Gamaleldin Elsayed, Mohamed Elhoseiny:
Affective Visual Dialog: A Large-Scale Benchmark for Emotional Reasoning Based on Visually Grounded Conversations. ECCV (75) 2024: 18-36 - [c76]Sanjoy Chowdhury, Sayan Nag, Subhrajyoti Dasgupta, Jun Chen, Mohamed Elhoseiny, Ruohan Gao, Dinesh Manocha:
MEERKAT: Audio-Visual Large Language Model for Grounding in Space and Time. ECCV (64) 2024: 52-70 - [c75]Xiang Li, Jian Ding, Zhaoyang Chen, Mohamed Elhoseiny:
Uni3DL: A Unified Model for 3D Vision-Language Understanding. ECCV (23) 2024: 74-92 - [c74]Kirolos Ataallah, Xiaoqian Shen, Eslam Abdelrahman, Essam Sleiman, Mingchen Zhuge, Jian Ding, Deyao Zhu, Jürgen Schmidhuber, Mohamed Elhoseiny:
Goldfish: Vision-Language Understanding of Arbitrarily Long Videos. ECCV (29) 2024: 251-267 - [c73]Youssef Mohamed, Runjia Li, Ibrahim Ahmad, Kilichbek Haydarov, Philip Torr, Kenneth Church, Mohamed Elhoseiny:
No Culture Left Behind: ArtELingo-28, a Benchmark of WikiArt with Captions in 28 Languages. EMNLP 2024: 20939-20962 - [c72]Eslam Mohamed Bakr, Mohamed Ayman, Mahmoud Ahmed, Habib Slim, Mohamed Elhoseiny:
CoT3DRef: Chain-of-Thoughts Data-Efficient 3D Visual Grounding. ICLR 2024 - [c71]Wenxuan Zhang, Youssef Mohamed, Bernard Ghanem, Philip Torr, Adel Bibi, Mohamed Elhoseiny:
Continual Learning on a Diet: Learning from Sparsely Labeled Streams Under Constrained Computation. ICLR 2024 - [c70]Deyao Zhu, Jun Chen, Xiaoqian Shen, Xiang Li, Mohamed Elhoseiny:
MiniGPT-4: Enhancing Vision-Language Understanding with Advanced Large Language Models. ICLR 2024 - [c69]Xinliang Zhu, Arnab Dhua, Douglas Gray, I. Zeki Yalniz, Tan Yu, Mohamed Elhoseiny, Bryan A. Plummer:
Multimodal Representation and Retrieval [MRR 2024]. SIGIR 2024: 3047-3050 - [c68]Salman Khan, Izzeddin Teeti, Andrew Bradley, Mohamed Elhoseiny, Fabio Cuzzolin:
A Hybrid Graph Network for Complex Activity Detection in Video. WACV 2024: 6748-6758 - [i86]Faizan Farooq Khan, Diana Kim, Divyansh Jha, Youssef Mohamed, Hanna H. Chang, Ahmed Elgammal, Luba Elliott, Mohamed Elhoseiny:
AI Art Neural Constellation: Revealing the Collective and Contrastive State of AI-Generated and Human Art. CoRR abs/2402.02453 (2024) - [i85]Kirolos Ataallah, Xiaoqian Shen, Eslam Abdelrahman, Essam Sleiman, Deyao Zhu, Jian Ding, Mohamed Elhoseiny:
MiniGPT4-Video: Advancing Multimodal LLMs for Video Understanding with Interleaved Visual-Textual Tokens. CoRR abs/2404.03413 (2024) - [i84]Wenxuan Zhang, Youssef Mohamed, Bernard Ghanem, Philip H. S. Torr, Adel Bibi, Mohamed Elhoseiny:
Continual Learning on a Diet: Learning from Sparsely Labeled Streams Under Constrained Computation. CoRR abs/2404.12766 (2024) - [i83]Junjie Fei, Mahmoud Ahmed, Jian Ding, Eslam Mohamed Bakr, Mohamed Elhoseiny:
Kestrel: Point Grounding Multimodal LLM for Part-Aware 3D Vision-Language Understanding. CoRR abs/2405.18937 (2024) - [i82]Abdulwahab Felemban, Eslam Mohamed Bakr, Xiaoqian Shen, Jian Ding, Abduallah A. Mohamed, Mohamed Elhoseiny:
iMotion-LLM: Motion Prediction Instruction Tuning. CoRR abs/2406.06211 (2024) - [i81]Xiang Li, Jian Ding, Mohamed Elhoseiny:
VRSBench: A Versatile Vision-Language Benchmark Dataset for Remote Sensing Image Understanding. CoRR abs/2406.12384 (2024) - [i80]Kirolos Ataallah, Chenhui Gou, Eslam Abdelrahman, Khushbu Pahwa, Jian Ding, Mohamed Elhoseiny:
InfiniBench: A Comprehensive Benchmark for Large Multimodal Models in Very Long Video Understanding. CoRR abs/2406.19875 (2024) - [i79]Sanjoy Chowdhury, Sayan Nag, Subhrajyoti Dasgupta, Jun Chen, Mohamed Elhoseiny, Ruohan Gao, Dinesh Manocha:
Meerkat: Audio-Visual Large Language Model for Grounding in Space and Time. CoRR abs/2407.01851 (2024) - [i78]Asma Alkhaldi, Raneem Alnajim, Layan Alabdullatef, Rawan Alyahya, Jun Chen, Deyao Zhu, Ahmed Alsinan, Mohamed Elhoseiny:
MiniGPT-Med: Large Language Model as a General Interface for Radiology Diagnosis. CoRR abs/2407.04106 (2024) - [i77]Kirolos Ataallah, Xiaoqian Shen, Eslam Abdelrahman, Essam Sleiman, Mingchen Zhuge, Jian Ding, Deyao Zhu, Jürgen Schmidhuber, Mohamed Elhoseiny:
Goldfish: Vision-Language Understanding of Arbitrarily Long Videos. CoRR abs/2407.12679 (2024) - [i76]Zilyu Ye, Jinxiu Liu, Ruotian Peng, Jinjin Cao, Zhiyang Chen, Yiyang Zhang, Ziwei Xuan, Mingyuan Zhou, Xiaoqian Shen, Mohamed Elhoseiny, Qi Liu, Guo-Jun Qi:
Openstory++: A Large-scale Dataset and Benchmark for Instance-aware Open-domain Visual Storytelling. CoRR abs/2408.03695 (2024) - [i75]Chenhui Gou, Abdulwahab Felemban, Faizan Farooq Khan, Deyao Zhu, Jianfei Cai, Hamid Rezatofighi, Mohamed Elhoseiny:
How Well Can Vision Language Models See Image Details? CoRR abs/2408.03940 (2024) - [i74]Wenxuan Zhang, Philip H. S. Torr, Mohamed Elhoseiny, Adel Bibi:
Bi-Factorial Preference Optimization: Balancing Safety-Helpfulness in Language Models. CoRR abs/2408.15313 (2024) - 2023
- [c67]Xiaoqian Shen, Xiang Li, Mohamed Elhoseiny:
MoStGAN-V: Video Generation with Temporal Motion Styles. CVPR 2023: 5652-5661 - [c66]Jun Chen, Ming Hu, Darren J. Coker, Michael L. Berumen, Blair R. Costelloe, Sara Beery, Anna Rohrbach, Mohamed Elhoseiny:
MammalNet: A Large-Scale Video Benchmark for Mammal Recognition and Behavior Understanding. CVPR 2023: 13052-13061 - [c65]Jun Chen, Deyao Zhu, Guocheng Qian, Bernard Ghanem, Zhicheng Yan, Chenchen Zhu, Fanyi Xiao, Sean Chang Culatana, Mohamed Elhoseiny:
Exploring Open-Vocabulary Semantic Segmentation from CLIP Vision Encoder Distillation Only. ICCV 2023: 699-710 - [c64]Wenxuan Zhang, Paul Janson, Kai Yi, Ivan Skorokhodov, Mohamed Elhoseiny:
Continual Zero-Shot Learning through Semantically Guided Generative Random Walks. ICCV 2023: 11540-11551 - [c63]Eslam Mohamed Bakr, Pengzhan Sun, Xiaoqian Shen, Faizan Farooq Khan, Li Erran Li, Mohamed Elhoseiny:
HRS-Bench: Holistic, Reliable and Scalable Benchmark for Text-to-Image Models. ICCV 2023: 19984-19996 - [c62]Runjia Li, Shuyang Sun, Mohamed Elhoseiny, Philip H. S. Torr:
OxfordTVG-HIC: Can Machine Make Humorous Captions from Images? ICCV 2023: 20236-20246 - [c61]Faizan Farooq Khan, Xiang Li, Andrew J. Temple, Mohamed Elhoseiny:
FishNet: A Large-scale Dataset and Benchmark for Fish Recognition, Detection, and Functional Trait Prediction. ICCV 2023: 20439-20449 - [c60]Deyao Zhu, Li Erran Li, Mohamed Elhoseiny:
Value Memory Graph: A Graph-Structured World Model for Offline Reinforcement Learning. ICLR 2023 - [c59]Hang Xu, Wenxuan Zhang, Jiawei Fei, Yuzhe Wu, Tingwen Xie, Jun Huang, Yuchen Xie, Mohamed Elhoseiny, Panos Kalnis:
SLAMB: Accelerated Large Batch Training with Sparse Communication. ICML 2023: 38801-38825 - [i73]Deyao Zhu, Yuhui Wang, Jürgen Schmidhuber, Mohamed Elhoseiny:
Guiding Online Reinforcement Learning with Action-Free Offline Pretraining. CoRR abs/2301.12876 (2023) - [i72]Deyao Zhu, Jun Chen, Kilichbek Haydarov, Xiaoqian Shen, Wenxuan Zhang, Mohamed Elhoseiny:
ChatGPT Asks, BLIP-2 Answers: Automatic Questioning Towards Enriched Visual Descriptions. CoRR abs/2303.06594 (2023) - [i71]Xiaoqian Shen, Xiang Li, Mohamed Elhoseiny:
MoStGAN-V: Video Generation with Temporal Motion Styles. CoRR abs/2304.02777 (2023) - [i70]Jun Chen, Deyao Zhu, Kilichbek Haydarov, Xiang Li, Mohamed Elhoseiny:
Video ChatCaptioner: Towards Enriched Spatiotemporal Descriptions. CoRR abs/2304.04227 (2023) - [i69]Eslam Mohamed Bakr, Pengzhan Sun, Li Erran Li, Mohamed Elhoseiny:
ImageCaptioner2: Image Captioner for Image Captioning Bias Amplification Assessment. CoRR abs/2304.04874 (2023) - [i68]Eslam Mohamed Bakr, Pengzhan Sun, Xiaoqian Shen, Faizan Farooq Khan, Li Erran Li, Mohamed Elhoseiny:
HRS-Bench: Holistic, Reliable and Scalable Benchmark for Text-to-Image Models. CoRR abs/2304.05390 (2023) - [i67]Deyao Zhu, Jun Chen, Xiaoqian Shen, Xiang Li, Mohamed Elhoseiny:
MiniGPT-4: Enhancing Vision-Language Understanding with Advanced Large Language Models. CoRR abs/2304.10592 (2023) - [i66]Jun Chen, Deyao Zhu, Guocheng Qian, Bernard Ghanem, Zhicheng Yan, Chenchen Zhu, Fanyi Xiao, Mohamed Elhoseiny, Sean Chang Culatana:
Exploring Open-Vocabulary Semantic Segmentation without Human Labels. CoRR abs/2306.00450 (2023) - [i65]Jun Chen, Ming Hu, Darren J. Coker, Michael L. Berumen, Blair R. Costelloe, Sara M. Beery, Anna Rohrbach, Mohamed Elhoseiny:
MammalNet: A Large-scale Video Benchmark for Mammal Recognition and Behavior Understanding. CoRR abs/2306.00576 (2023) - [i64]Runjia Li, Shuyang Sun, Mohamed Elhoseiny, Philip H. S. Torr:
OxfordTVG-HIC: Can Machine Make Humorous Captions from Images? CoRR abs/2307.11636 (2023) - [i63]Wenxuan Zhang, Paul Janson, Kai Yi, Ivan Skorokhodov, Mohamed Elhoseiny:
Continual Zero-Shot Learning through Semantically Guided Generative Random Walks. CoRR abs/2308.12366 (2023) - [i62]Wenxuan Zhang, Paul Janson, Rahaf Aljundi, Mohamed Elhoseiny:
Overcoming General Knowledge Loss with Selective Parameter Finetuning. CoRR abs/2308.12462 (2023) - [i61]Kilichbek Haydarov, Xiaoqian Shen, Avinash Madasu, Mahmoud Salem, Jia Li, Gamaleldin Elsayed, Mohamed Elhoseiny:
Affective Visual Dialog: A Large-Scale Benchmark for Emotional Reasoning Based on Visually Grounded Conversations. CoRR abs/2308.16349 (2023) - [i60]Eslam Mohamed Bakr, Mohamed Ayman, Mahmoud Ahmed, Habib Slim, Mohamed Elhoseiny:
CoT3DRef: Chain-of-Thoughts Data-Efficient 3D Visual Grounding. CoRR abs/2310.06214 (2023) - [i59]Jun Chen, Deyao Zhu, Xiaoqian Shen, Xiang Li, Zechun Liu, Pengchuan Zhang, Raghuraman Krishnamoorthi, Vikas Chandra, Yunyang Xiong, Mohamed Elhoseiny:
MiniGPT-v2: large language model as a unified interface for vision-language multi-task learning. CoRR abs/2310.09478 (2023) - [i58]Salman Khan, Izzeddin Teeti, Andrew Bradley, Mohamed Elhoseiny, Fabio Cuzzolin:
A Hybrid Graph Network for Complex Activity Detection in Video. CoRR abs/2310.17493 (2023) - [i57]Habib Slim, Xiang Li, Yuchen Li, Mahmoud Ahmed, Mohamed Ayman, Ujjwal Upadhyay, Ahmed Abdelreheem, Arpit Prajapati, Suhail Pothigara, Peter Wonka, Mohamed Elhoseiny:
3DCoMPaT++: An improved Large-scale 3D Vision Dataset for Compositional Recognition. CoRR abs/2310.18511 (2023) - [i56]Eslam Mohamed Bakr, Liangbing Zhao, Vincent Tao Hu, Matthieu Cord, Patrick Pérez, Mohamed Elhoseiny:
ToddlerDiffusion: Flash Interpretable Controllable Diffusion Model. CoRR abs/2311.14542 (2023) - [i55]Botos Csaba, Wenxuan Zhang, Matthias Müller, Ser-Nam Lim, Mohamed Elhoseiny, Philip H. S. Torr, Adel Bibi:
Label Delay in Continual Learning. CoRR abs/2312.00923 (2023) - [i54]Xiaoqian Shen, Mohamed Elhoseiny:
StoryGPT-V: Large Language Models as Consistent Story Visualizers. CoRR abs/2312.02252 (2023) - [i53]Xiang Li, Jian Ding, Zhaoyang Chen, Mohamed Elhoseiny:
Uni3DL: Unified Model for 3D and Language Understanding. CoRR abs/2312.03026 (2023) - 2022
- [c58]Ivan Skorokhodov, Sergey Tulyakov, Mohamed Elhoseiny:
StyleGAN-V: A Continuous Video Generator with the Price, Image Quality and Perks of StyleGAN2. CVPR 2022: 3616-3626 - [c57]Jun Chen, Han Guo, Kai Yi, Boyang Li, Mohamed Elhoseiny:
VisualGPT: Data-efficient Adaptation of Pretrained Language Models for Image Captioning. CVPR 2022: 18009-18019 - [c56]Jun Chen, Aniket Agarwal, Sherif Abdelkarim, Deyao Zhu, Mohamed Elhoseiny:
RelTransformer: A Transformer-Based Long-Tail Visual Relationship Recognition. CVPR 2022: 19485-19495 - [c55]Youssef Mohamed, Faizan Farooq Khan, Kilichbek Haydarov, Mohamed Elhoseiny:
It is Okay to Not Be Okay: Overcoming Emotional Bias in Affective Image Captioning by Contrastive Data Collection. CVPR 2022: 21231-21240 - [c54]Yuchen Li, Ujjwal Upadhyay, Habib Slim, Ahmed Abdelreheem, Arpit Prajapati, Suhail Pothigara, Peter Wonka, Mohamed Elhoseiny:
3D CoMPaT: Composition of Materials on Parts of 3D Things. ECCV (8) 2022: 110-127 - [c53]Kai Yi, Xiaoqian Shen, Yunhao Gou, Mohamed Elhoseiny:
Exploring Hierarchical Graph Representation for Large-Scale Zero-Shot Image Classification. ECCV (20) 2022: 116-132 - [c52]Abduallah A. Mohamed, Deyao Zhu, Warren Vu, Mohamed Elhoseiny, Christian G. Claudel:
Social-Implicit: Rethinking Trajectory Prediction Evaluation and The Effectiveness of Implicit Maximum Likelihood Estimation. ECCV (22) 2022: 463-479 - [c51]Youssef Mohamed, Mohamed Abdelfattah, Shyma Alhuwaider, Feifan Li, Xiangliang Zhang, Kenneth Church, Mohamed Elhoseiny:
ArtELingo: A Million Emotion Annotations of WikiArt with Emphasis on Diversity over Language and Culture. EMNLP 2022: 8770-8785 - [c50]Divyansh Jha, Kai Yi, Ivan Skorokhodov, Mohamed Elhoseiny:
Creative Walk Adversarial Networks: Novel Art Generation with Probabilistic Random Walk Deviation from Style Norms. ICCC 2022: 195-204 - [c49]Eslam Mohamed Bakr, Yasmeen Alsaedy, Mohamed Elhoseiny:
Look Around and Refer: 2D Synthetic Semantics Knowledge Distillation for 3D Visual Grounding. NeurIPS 2022 - [c48]Guocheng Qian, Yuchen Li, Houwen Peng, Jinjie Mai, Hasan Hammoud, Mohamed Elhoseiny, Bernard Ghanem:
PointNeXt: Revisiting PointNet++ with Improved Training and Scaling Strategies. NeurIPS 2022 - [c47]Ahmed Abdelreheem, Ujjwal Upadhyay, Ivan Skorokhodov, Rawan Al Yahya, Jun Chen, Mohamed Elhoseiny:
3DRefTransformer: Fine-Grained Object Identification in Real-World Scenes Using Natural Language. WACV 2022: 607-616 - [i52]Yuanpeng Li, Joel Hestness, Mohamed Elhoseiny, Liang Zhao, Kenneth Church:
Efficiently Disentangle Causal Representations. CoRR abs/2201.01942 (2022) - [i51]Kai Yi, Xiaoqian Shen, Yunhao Gou, Mohamed Elhoseiny:
Exploring Hierarchical Graph Representation for Large-Scale Zero-Shot Image Classification. CoRR abs/2203.01386 (2022) - [i50]Abduallah A. Mohamed, Deyao Zhu, Warren Vu, Mohamed Elhoseiny, Christian G. Claudel:
Social-Implicit: Rethinking Trajectory Prediction Evaluation and The Effectiveness of Implicit Maximum Likelihood Estimation. CoRR abs/2203.03057 (2022) - [i49]Youssef Mohamed, Faizan Farooq Khan, Kilichbek Haydarov, Mohamed Elhoseiny:
It is Okay to Not Be Okay: Overcoming Emotional Bias in Affective Image Captioning by Contrastive Data Collection. CoRR abs/2204.07660 (2022) - [i48]Jun Chen, Ming Hu, Boyang Li, Mohamed Elhoseiny:
Efficient Self-supervised Vision Pretraining with Local Masked Reconstruction. CoRR abs/2206.00790 (2022) - [i47]Deyao Zhu, Li Erran Li, Mohamed Elhoseiny:
Value Memory Graph: A Graph-Structured World Model for Offline Reinforcement Learning. CoRR abs/2206.04384 (2022) - [i46]Guocheng Qian, Yuchen Li, Houwen Peng, Jinjie Mai, Hasan Abed Al Kader Hammoud, Mohamed Elhoseiny, Bernard Ghanem:
PointNeXt: Revisiting PointNet++ with Improved Training and Scaling Strategies. CoRR abs/2206.04670 (2022) - [i45]Paul Janson, Wenxuan Zhang, Rahaf Aljundi, Mohamed Elhoseiny:
A Simple Baseline that Questions the Use of Pretrained-Models in Continual Learning. CoRR abs/2210.04428 (2022) - [i44]Youssef Mohamed, Mohamed Abdelfattah, Shyma Alhuwaider, Feifan Li, Xiangliang Zhang, Kenneth Ward Church, Mohamed Elhoseiny:
ArtELingo: A Million Emotion Annotations of WikiArt with Emphasis on Diversity over Language and Culture. CoRR abs/2211.10780 (2022) - [i43]Eslam Mohamed Bakr, Yasmeen Alsaedy, Mohamed Elhoseiny:
Look Around and Refer: 2D Synthetic Semantics Knowledge Distillation for 3D Visual Grounding. CoRR abs/2211.14241 (2022) - 2021
- [c46]Ahmed Ayyad, Yuchen Li, Raden Muaz, Shadi Albarqouni, Mohamed Elhoseiny:
Semi-Supervised Few-Shot Learning with Prototypical Random Walks. MetaDL@AAAI 2021: 45-57 - [c45]Deyao Zhu, Mohamed Zahran, Li Erran Li, Mohamed Elhoseiny:
Motion Forecasting with Unlikelihood Training in Continuous Space. CoRL 2021: 1003-1012 - [c44]Ivan Skorokhodov, Savva Ignatyev, Mohamed Elhoseiny:
Adversarial Generation of Continuous Images. CVPR 2021: 10753-10764 - [c43]Panos Achlioptas, Maks Ovsjanikov, Kilichbek Haydarov, Mohamed Elhoseiny, Leonidas J. Guibas:
ArtEmis: Affective Language for Visual Art. CVPR 2021: 11569-11579 - [c42]Divyansh Jha, Hanna H. Chang, Mohamed Elhoseiny:
Wölfflin's Affective Generative Analysis for Visual Art. ICCC 2021: 429-433 - [c41]Ivan Skorokhodov, Grigorii Sotnikov, Mohamed Elhoseiny:
Aligning Latent and Image Spaces to Connect the Unconnectable. ICCV 2021: 14124-14133 - [c40]Sherif Abdelkarim, Aniket Agarwal, Panos Achlioptas, Jun Chen, Jiaji Huang, Boyang Li, Kenneth Church, Mohamed Elhoseiny:
Exploring Long Tail Visual Relationship Recognition with Large Vocabulary. ICCV 2021: 15901-15910 - [c39]Ivan Skorokhodov, Mohamed Elhoseiny:
Class Normalization for (Continual)? Generalized Zero-Shot Learning. ICLR 2021 - [c38]Deyao Zhu, Mohamed Zahran, Li Erran Li, Mohamed Elhoseiny:
HalentNet: Multimodal Trajectory Forecasting with Hallucinative Intents. ICLR 2021 - [i42]Mohamed Elhoseiny, Kai Yi, Mohamed Elfeki:
CIZSL++: Creativity Inspired Generative Zero-Shot Learning. CoRR abs/2101.00173 (2021) - [i41]Panos Achlioptas, Maks Ovsjanikov, Kilichbek Haydarov, Mohamed Elhoseiny, Leonidas J. Guibas:
ArtEmis: Affective Language for Visual Art. CoRR abs/2101.07396 (2021) - [i40]Jun Chen, Han Guo, Kai Yi, Boyang Li, Mohamed Elhoseiny:
VisualGPT: Data-efficient Image Captioning by Balancing Visual Input and Linguistic Knowledge from Pretraining. CoRR abs/2102.10407 (2021) - [i39]Ivan Skorokhodov, Grigorii Sotnikov, Mohamed Elhoseiny:
Aligning Latent and Image Spaces to Connect the Unconnectable. CoRR abs/2104.06954 (2021) - [i38]Mohamed Elhoseiny, Divyansh Jha, Kai Yi, Ivan Skorokhodov:
Imaginative Walks: Generative Random Walk Deviation Loss for Improved Unseen Learning Representation. CoRR abs/2104.09757 (2021) - [i37]Jun Chen, Aniket Agarwal, Sherif Abdelkarim, Deyao Zhu, Mohamed Elhoseiny:
RelTransformer: Balancing the Visual Relationship Detection from Local Context, Scene and Memory. CoRR abs/2104.11934 (2021) - [i36]Kai Yi, Mohamed Elhoseiny:
Domain-Aware Continual Zero-Shot Learning. CoRR abs/2112.12989 (2021) - [i35]Ivan Skorokhodov, Sergey Tulyakov, Mohamed Elhoseiny:
StyleGAN-V: A Continuous Video Generator with the Price, Image Quality and Perks of StyleGAN2. CoRR abs/2112.14683 (2021) - 2020
- [c37]Abduallah A. Mohamed, Kun Qian, Mohamed Elhoseiny, Christian G. Claudel:
Social-STGCNN: A Social Spatio-Temporal Graph Convolutional Neural Network for Human Trajectory Prediction. CVPR 2020: 14412-14420 - [c36]Panos Achlioptas, Ahmed Abdelreheem, Fei Xia, Mohamed Elhoseiny, Leonidas J. Guibas:
ReferIt3D: Neural Listeners for Fine-Grained 3D Object Identification in Real-World Scenes. ECCV (1) 2020: 422-440 - [c35]Sayna Ebrahimi, Mohamed Elhoseiny, Trevor Darrell, Marcus Rohrbach:
Uncertainty-guided Continual Learning with Bayesian Neural Networks. ICLR 2020 - [c34]Yuanpeng Li, Liang Zhao, Kenneth Church, Mohamed Elhoseiny:
Compositional Language Continual Learning. ICLR 2020 - [c33]Uchenna Akujuobi, Jun Chen, Mohamed Elhoseiny, Michael Spranger, Xiangliang Zhang:
Temporal Positive-unlabeled Learning for Biomedical Hypothesis Generation via Risk Estimation. NeurIPS 2020 - [i34]Abduallah A. Mohamed, Kun Qian, Mohamed Elhoseiny, Christian G. Claudel:
Social-STGCNN: A Social Spatio-Temporal Graph Convolutional Neural Network for Human Trajectory Prediction. CoRR abs/2002.11927 (2020) - [i33]Sherif Abdelkarim, Panos Achlioptas, Jiaji Huang, Boyang Li, Kenneth Church, Mohamed Elhoseiny:
Long-tail Visual Relationship Recognition with a Visiolinguistic Hubless Loss. CoRR abs/2004.00436 (2020) - [i32]Jun Chen, Robert Hoehndorf, Mohamed Elhoseiny, Xiangliang Zhang:
Efficient long-distance relation extraction with DG-SpanBERT. CoRR abs/2004.03636 (2020) - [i31]Abduallah A. Mohamed, Muhammed Mohaimin Sadiq, Ehab AlBadawy, Mohamed Elhoseiny, Christian G. Claudel:
Inner Ensemble Nets. CoRR abs/2006.08305 (2020) - [i30]Ivan Skorokhodov, Mohamed Elhoseiny:
Normalization Matters in Zero-Shot Learning. CoRR abs/2006.11328 (2020) - [i29]Uchenna Akujuobi, Jun Chen, Mohamed Elhoseiny, Michael Spranger, Xiangliang Zhang:
Temporal Positive-unlabeled Learning for Biomedical Hypothesis Generation via Risk Estimation. CoRR abs/2010.01916 (2020) - [i28]Ivan Skorokhodov, Savva Ignatyev, Mohamed Elhoseiny:
Adversarial Generation of Continuous Images. CoRR abs/2011.12026 (2020)
2010 – 2019
- 2019
- [c32]Ji Zhang, Yannis Kalantidis, Marcus Rohrbach, Manohar Paluri, Ahmed Elgammal, Mohamed Elhoseiny:
Large-Scale Visual Relationship Understanding. AAAI 2019: 9185-9194 - [c31]Sayna Ebrahimi, Mohamed Elhoseiny, Trevor Darrell, Marcus Rohrbach:
Uncertainty-Guided Continual Learning in Bayesian Neural Networks - Extended Abstract. CVPR Workshops 2019: 75-78 - [c30]Mohamed Elhoseiny, Mohamed Elfeki:
Creativity Inspired Zero-Shot Learning. ICCV 2019: 5783-5792 - [c29]Arslan Chaudhry, Marc'Aurelio Ranzato, Marcus Rohrbach, Mohamed Elhoseiny:
Efficient Lifelong Learning with A-GEM. ICLR (Poster) 2019 - [c28]Mohamed Elfeki, Camille Couprie, Morgane Rivière, Mohamed Elhoseiny:
GDPP: Learning Diverse Generations using Determinantal Point Processes. ICML 2019: 1774-1783 - [c27]Mennatullah Siam, Chen Jiang, Steven Weikai Lu, Laura Petrich, Mahmoud Gamal, Mohamed Elhoseiny, Martin Jägersand:
Video Object Segmentation using Teacher-Student Adaptation in a Human Robot Interaction (HRI) Setting. ICRA 2019: 50-56 - [i27]Arslan Chaudhry, Marcus Rohrbach, Mohamed Elhoseiny, Thalaiyasingam Ajanthan, Puneet Kumar Dokania, Philip H. S. Torr, Marc'Aurelio Ranzato:
Continual Learning with Tiny Episodic Memories. CoRR abs/1902.10486 (2019) - [i26]Ahmed Ayyad, Nassir Navab, Mohamed Elhoseiny, Shadi Albarqouni:
Semi-Supervised Few-Shot Learning with Local and Global Consistency. CoRR abs/1903.02164 (2019) - [i25]Mohamed Elhoseiny, Mohamed Elfeki:
Creativity Inspired Zero-Shot Learning. CoRR abs/1904.01109 (2019) - [i24]Sayna Ebrahimi, Mohamed Elhoseiny, Trevor Darrell, Marcus Rohrbach:
Uncertainty-guided Continual Learning with Bayesian Neural Networks. CoRR abs/1906.02425 (2019) - 2018
- [c26]Ahmed Elgammal, Bingchen Liu, Diana Kim, Mohamed Elhoseiny, Marian Mazzone:
The Shape of Art History in the Eyes of the Machine. AAAI 2018: 2183-2191 - [c25]Mohamed Elhoseiny, Francesca Babiloni, Rahaf Aljundi, Marcus Rohrbach, Manohar Paluri, Tinne Tuytelaars:
Exploring the Challenges Towards Lifelong Fact Learning. ACCV (6) 2018: 66-84 - [c24]Yizhe Zhu, Mohamed Elhoseiny, Bingchen Liu, Xi Peng, Ahmed Elgammal:
A Generative Adversarial Approach for Zero-Shot Learning From Noisy Texts. CVPR 2018: 1004-1013 - [c23]Othman Sbai, Mohamed Elhoseiny, Antoine Bordes, Yann LeCun, Camille Couprie:
DesIGN: Design Inspiration from Generative Networks. ECCV Workshops (3) 2018: 37-44 - [c22]Rahaf Aljundi, Francesca Babiloni, Mohamed Elhoseiny, Marcus Rohrbach, Tinne Tuytelaars:
Memory Aware Synapses: Learning What (not) to Forget. ECCV (3) 2018: 144-161 - [c21]Ramprasaath R. Selvaraju, Prithvijit Chattopadhyay, Mohamed Elhoseiny, Tilak Sharma, Dhruv Batra, Devi Parikh, Stefan Lee:
Choose Your Neuron: Incorporating Domain Knowledge Through Neuron-Importance. ECCV (13) 2018: 540-556 - [i23]Ahmed M. Elgammal, Marian Mazzone, Bingchen Liu, Diana Kim, Mohamed Elhoseiny:
The Shape of Art History in the Eyes of the Machine. CoRR abs/1801.07729 (2018) - [i22]Othman Sbai, Mohamed Elhoseiny, Antoine Bordes, Yann LeCun, Camille Couprie:
DeSIGN: Design Inspiration from Generative Networks. CoRR abs/1804.00921 (2018) - [i21]Ji Zhang, Yannis Kalantidis, Marcus Rohrbach, Manohar Paluri, Ahmed M. Elgammal, Mohamed Elhoseiny:
Large-Scale Visual Relationship Understanding. CoRR abs/1804.10660 (2018) - [i20]Ramprasaath R. Selvaraju, Prithvijit Chattopadhyay, Mohamed Elhoseiny, Tilak Sharma, Dhruv Batra, Devi Parikh, Stefan Lee:
Choose Your Neuron: Incorporating Domain Knowledge through Neuron-Importance. CoRR abs/1808.02861 (2018) - [i19]Mennatullah Siam, Chen Jiang, Steven Weikai Lu, Laura Petrich, Mahmoud Gamal, Mohamed Elhoseiny, Martin Jägersand:
Video Segmentation using Teacher-Student Adaptation in a Human Robot Interaction (HRI) Setting. CoRR abs/1810.07733 (2018) - [i18]Mohamed Elfeki, Camille Couprie, Morgane Rivière, Mohamed Elhoseiny:
GDPP: Learning Diverse Generations Using Determinantal Point Process. CoRR abs/1812.00068 (2018) - [i17]Arslan Chaudhry, Marc'Aurelio Ranzato, Marcus Rohrbach, Mohamed Elhoseiny:
Efficient Lifelong Learning with A-GEM. CoRR abs/1812.00420 (2018) - [i16]Mohamed Elhoseiny, Francesca Babiloni, Rahaf Aljundi, Marcus Rohrbach, Manohar Paluri, Tinne Tuytelaars:
Exploring the Challenges towards Lifelong Fact Learning. CoRR abs/1812.10524 (2018) - 2017
- [j3]Mohamed Elhoseiny, Ahmed M. Elgammal, Babak Saleh:
Write a Classifier: Predicting Visual Classifiers from Unstructured Text. IEEE Trans. Pattern Anal. Mach. Intell. 39(12): 2539-2553 (2017) - [c20]Mohamed Elhoseiny, Scott Cohen, Walter Chang, Brian L. Price, Ahmed M. Elgammal:
Sherlock: Scalable Fact Learning in Images. AAAI 2017: 4016-4024 - [c19]Ji Zhang, Mohamed Elhoseiny, Scott Cohen, Walter Chang, Ahmed M. Elgammal:
Relationship Proposal Networks. CVPR 2017: 5226-5234 - [c18]Mohamed Elhoseiny, Yizhe Zhu, Han Zhang, Ahmed M. Elgammal:
Link the Head to the "Beak": Zero Shot Learning from Noisy Text Description at Part Precision. CVPR 2017: 6288-6297 - [c17]Ahmed M. Elgammal, Bingchen Liu, Mohamed Elhoseiny, Marian Mazzone:
CAN: Creative Adversarial Networks, Generating "Art" by Learning About Styles and Deviating from Style Norms. ICCC 2017: 96-103 - [i15]Mohamed Elhoseiny, Ahmed M. Elgammal:
Overlapping Cover Local Regression Machines. CoRR abs/1701.01218 (2017) - [i14]Ahmed M. Elgammal, Bingchen Liu, Mohamed Elhoseiny, Marian Mazzone:
CAN: Creative Adversarial Networks, Generating "Art" by Learning About Styles and Deviating from Style Norms. CoRR abs/1706.07068 (2017) - [i13]Mohamed Elhoseiny, Yizhe Zhu, Han Zhang, Ahmed M. Elgammal:
Link the head to the "beak": Zero Shot Learning from Noisy Text Description at Part Precision. CoRR abs/1709.01148 (2017) - [i12]Rahaf Aljundi, Francesca Babiloni, Mohamed Elhoseiny, Marcus Rohrbach, Tinne Tuytelaars:
Memory Aware Synapses: Learning what (not) to forget. CoRR abs/1711.09601 (2017) - [i11]Yizhe Zhu, Mohamed Elhoseiny, Bingchen Liu, Ahmed M. Elgammal:
Imagine it for me: Generative Adversarial Approach for Zero-Shot Learning from Noisy Texts. CoRR abs/1712.01381 (2017) - 2016
- [j2]Mohamed Elhoseiny, Ahmed M. Elgammal:
Text to multi-level MindMaps - A novel method for hierarchical visual abstraction of natural language text. Multim. Tools Appl. 75(8): 4217-4244 (2016) - [c16]Mohamed Elhoseiny, Jingen Liu, Hui Cheng, Harpreet S. Sawhney, Ahmed M. Elgammal:
Zero-Shot Event Detection by Multimodal Distributional Semantic Embedding of Videos. AAAI 2016: 3478-3486 - [c15]Mohamed Elhoseiny, Scott Cohen, Walter Chang, Brian L. Price, Ahmed M. Elgammal:
Automatic Annotation of Structured Facts in Images. VL@ACL 2016 - [c14]Han Zhang, Tao Xu, Mohamed Elhoseiny, Xiaolei Huang, Shaoting Zhang, Ahmed M. Elgammal, Dimitris N. Metaxas:
SPDA-CNN: Unifying Semantic Part Detection and Abstraction for Fine-Grained Recognition. CVPR 2016: 1143-1152 - [c13]Mohamed Elhoseiny, Tarek El-Gaaly, Amr Bakry, Ahmed M. Elgammal:
A Comparative Analysis and Study of Multiview CNN Models for Joint Object Categorization and Pose Estimation. ICML 2016: 888-897 - [c12]Amr Bakry, Tarek El-Gaaly, Mohamed Elhoseiny, Ahmed M. Elgammal:
Joint object recognition and pose estimation using a nonlinear view-invariant latent generative model. WACV 2016: 1-9 - [c11]Amr Bakry, Mohamed Elhoseiny, Tarek El-Gaaly, Ahmed M. Elgammal:
Digging Deep into the Layers of CNNs: In Search of How CNNs Achieve View Invariance. ICLR (Poster) 2016 - [i10]Mohamed Elhoseiny, Ahmed M. Elgammal, Babak Saleh:
Write a Classifier: Predicting Visual Classifiers from Unstructured Text Descriptions. CoRR abs/1601.00025 (2016) - [i9]Mohamed Elhoseiny, Scott Cohen, Walter Chang, Brian L. Price, Ahmed M. Elgammal:
Automatic Annotation of Structured Facts in Images. CoRR abs/1604.00466 (2016) - 2015
- [j1]Mohamed Elhoseiny, Ahmed M. Elgammal:
Generalized Twin Gaussian processes using Sharma-Mittal divergence. Mach. Learn. 100(2-3): 399-424 (2015) - [c10]Mohamed Elhoseiny, Ahmed M. Elgammal:
Visual Classifier Prediction by Distributional Semantic Embedding of Text Descriptions. VL@EMNLP 2015: 48-50 - [c9]Mohamed Elhoseiny, Ahmed M. Elgammal:
Overlapping Domain Cover for Scalable and Accurate Regression Kernel Machines. BMVC 2015: 94.1-94.12 - [c8]Sheng Huang, Mohamed Elhoseiny, Ahmed M. Elgammal, Dan Yang:
Learning Hypergraph-regularized Attribute Predictors. CVPR 2015: 409-417 - [c7]Mohamed Elhoseiny, Sheng Huang, Ahmed M. Elgammal:
Weather classification with deep convolutional neural networks. ICIP 2015: 3349-3353 - [i8]Sheng Huang, Mohamed Elhoseiny, Ahmed M. Elgammal, Dan Yang:
Learning Hypergraph-regularized Attribute Predictors. CoRR abs/1503.05782 (2015) - [i7]Mohamed Elhoseiny, Ahmed M. Elgammal, Babak Saleh:
Tell and Predict: Kernel Classifier Prediction for Unseen Visual Classes from Unstructured Text Descriptions. CoRR abs/1506.08529 (2015) - [i6]Mohamed Elhoseiny, Scott Cohen, Walter Chang, Brian L. Price, Ahmed M. Elgammal:
Sherlock: Modeling Structured Knowledge in Images. CoRR abs/1511.04891 (2015) - [i5]Mohamed Elhoseiny, Tarek El-Gaaly, Amr Bakry, Ahmed M. Elgammal:
Convolutional Models for Joint Object Categorization and Pose Estimation. CoRR abs/1511.05175 (2015) - [i4]Mohamed Elhoseiny, Jingen Liu, Hui Cheng, Harpreet S. Sawhney, Ahmed M. Elgammal:
Zero-Shot Event Detection by Multimodal Distributional Semantic Embedding of Videos. CoRR abs/1512.00818 (2015) - 2014
- [c6]Sheng Huang, Mohamed Elhoseiny, Ahmed M. Elgammal, Dan Yang:
Improving non-negative matrix factorization via ranking its bases. ICIP 2014: 5951-5955 - [c5]Hui Cheng, Jingen Liu, Ishani Chakraborty, Guang Chen, Qiguang Liu, Mohamed Elhoseiny, Gary Gan, Ajay Divakaran, Harpreet S. Sawhney, James Allan, John Foley, Mubarak Shah, Afshin Dehghan, Michael Witbrock, Jon Curtis:
SRI-Sarnoff AURORA System at TRECVID 2014 Multimedia Event Detection and Recounting. TRECVID 2014 - [i3]Mohamed Elhoseiny, Ahmed M. Elgammal:
Text to Multi-level MindMaps: A New Way for Interactive Visualization and Summarization of Natural Language Text. CoRR abs/1408.1031 (2014) - [i2]Mohamed Elhoseiny, Ahmed M. Elgammal:
Generalized Twin Gaussian Processes using Sharma-Mittal Divergence. CoRR abs/1409.7480 (2014) - 2013
- [c4]Mohamed Elhoseiny, Amr Bakry, Ahmed M. Elgammal:
MultiClass Object Classification in Video Surveillance Systems - Experimental Study. CVPR Workshops 2013: 788-793 - [c3]Mohamed Elhoseiny, Babak Saleh, Ahmed M. Elgammal:
Write a Classifier: Zero-Shot Learning Using Purely Textual Descriptions. ICCV 2013: 2584-2591 - [c2]Mohamed Elhoseiny, Bing Song, Jeremi Sudol, David McKinnon:
Low-bitrate benefits of JPEG compression on sift recognition. ICIP 2013: 3657-3661 - [i1]Mohamed Elhoseiny, Hossam El Deen Mostafa Faheem, Taymour Nazmy, Eman Shaaban:
GPU-Framework for Teamwork Action Recognition. CoRR abs/1310.3322 (2013) - 2012
- [c1]Mohamed Elhoseiny, Ahmed M. Elgammal:
English2MindMap: An Automated System for MindMap Generation from English Text. ISM 2012: 326-331
Coauthor Index
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
Add open access links from to the list of external document links (if available).
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy.
Archived links via Wayback Machine
For web page which are no longer available, try to retrieve content from the of the Internet Archive (if available).
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy.
Reference lists
Add a list of references from , , and to record detail pages.
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org, opencitations.net, and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy, as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
Add a list of citing articles from and to record detail pages.
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
Load additional information about publications from .
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex.
last updated on 2024-11-15 20:39 CET by the dblp team
all metadata released as open data under CC0 1.0 license
see also: Terms of Use | Privacy Policy | Imprint