I'm a PhD candidate in the National PhD in Artificial Intelligence at the University of Pisa, conducting my research at AImageLab, the research laboratory of the Department of Engineering "Enzo Ferrari" at the University of Modena and Reggio Emilia. I recently participated in an internship at Amazon Prime Video in Seattle. My specialization areas include AI Safety, Responsible AI, and Trustworthy AI.
My research focuses on advancing Generative AI and Multimodal Architectures, aiming to bridge the gap between cutting-edge deep learning technologies and ethical alignment with human values.
Passionate about leveraging AI to solve complex societal challenges, I am dedicated to ensuring AI systems are safe, transparent, and aligned with responsible principles, driving innovation with a human-centric approach. Some papers are highlighted.
CounterVid introduces a framework leveraging diffusion models for image editing and video generation combined with the reasoning capabilities of video-language models to generate counterfactual videos that mitigate hallucinations, developed during an internship at Amazon Prime Video.
We propose the prefilling attack, a structured natural-language prefix prepended to the model output, to steer the model to respond with a clean, valid option in multiple-choice question answering tasks, significantly improving accuracy and calibration.
HySAC, Hyperbolic Safety-Aware CLIP, models hierarchical safety relations to enable effective retrieval of unsafe content, dynamically redirecting it to safer alternatives for enhanced content moderation.
Vision-and-language model designed to mitigate the risks associated with NSFW content in AI applications. Safe-CLIP is fine-tuned to serve the association between linguistic and visual concepts, ensuring safer outputs in text-to-image and image-to-text retrieval and generation tasks.
A study on aligning Large Language Models (LLMs) for offensive content removal, introducing the CAiA dataset and Safe-CLIP to ensure ethical and responsible AI-driven content moderation.
An investigation into background-induced bias in 6-DoF object pose estimation, revealing how ArUco markers in datasets like Linemod influence model predictions and proposing mitigation strategies using dataset augmentation and saliency map analysis.
Deep Learning for DeepFake Detection
Tobia Poppi,
Elena Govi, Fabio Marinelli
2022
pdf
A deep learning pipeline for DeepFake detection, leveraging custom datasets and an optimized Xception-based CNN to identify manipulated facial images with high accuracy."
Il labelling: i migliori tool disponibili e la loro applicazione per il rilevamento di persone e cartelli stradali e per la pose estimation di oggetti 3D
Tobia Poppi
2021
arXiv
A comprehensive study of data labeling tools and techniques, evaluating their efficiency in person and traffic sign detection, as well as 3D pose estimation for robotic applications.
The FAIR project is a national scale, multidisciplinary initiative aimed at reimagining and developing
large-scale foundational models. It explores research questions, methodologies, models, technologies,
as well as ethical and legal frameworks for creating Artificial Intelligence systems capable of interacting
and collaborating with humans.