Skip to content
View NicholasMoratelli's full-sized avatar

Highlights

  • Pro

Block or report NicholasMoratelli

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don’t include any personal information such as legal names or email addresses. Markdown is supported. This note will only be visible to you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
NicholasMoratelli/README.md

Nicholas Moratelli

AI Researcher in Vision–Language & Multimodal AI
PhD in Artificial Intelligence (Defense April 2026) @ AImageLab, UNIMORE

About Me

🤖 I am an AI researcher working on Vision–Language Models and Multimodal Large Language Models at AImageLab, University of Modena and Reggio Emilia.

👀 My research focuses on knowledge-intensive multimodal reasoning and retrieval-augmented generation (RAG).

I work on the design, training, and evaluation of large-scale multimodal systems, combining generative models with retrieval, grounding, and structured reasoning mechanisms.

Alongside academic research, I worked as an Applied Scientist Intern at Amazon Science (Cambridge, UK), conducting research on large-scale multimodal models within the Amazon AGI organization.


Research Topics 🔭

  • Vision–Language Models (VLMs)
  • Multimodal Large Language Models (MLLMs)
  • Retrieval-Augmented Generation (RAG)
  • Knowledge-based Visual Question Answering
  • Multimodal reasoning and evaluation

Publications 📚

My work has been published at top-tier venues including CVPR, ICLR, ACL, and BMVC.
👉 Full publication list available on Google Scholar.


Links & Contact 📫

Pinned Loading

  1. aimagelab/ReflectiVA aimagelab/ReflectiVA Public

    [CVPR 2025] Augmenting Multimodal LLMs with Self-Reflective Tokens for Knowledge-based Visual Question Answering

    Python 55 1

  2. aimagelab/COGT aimagelab/COGT Public

    [ICLR 2025] Causal Graphical Models for Vision-Language Compositional Understanding

    Python 10

  3. aimagelab/DiCO aimagelab/DiCO Public

    [BMVC 2024 Oral ✨] Revisiting Image Captioning Training Paradigm via Direct CLIP-based Optimization

    Python 20 1

  4. aimagelab/LLaVA-MORE aimagelab/LLaVA-MORE Public

    [ICCVW 25] LLaVA-MORE: A Comparative Study of LLMs and Visual Backbones for Enhanced Visual Instruction Tuning

    Python 159 14

  5. aimagelab/pacscore aimagelab/pacscore Public

    [CVPR 2023 & IJCV 2025] Positive-Augmented Contrastive Learning for Image and Video Captioning Evaluation

    Python 68 9