I'm an Applied Scientist II at Amazon AGI, where I develop large-scale information retrieval systems for RAG and pre-train and fine-tune multimodal language models (VLMs) for multimedia analysis and processing at scale. My work spans text, images, and videos, leveraging LLMs to extract insights from diverse data sources. I hold a PhD in Information and Communication Technology and an MS in Computer Science from the University of Trento, Italy. My research centers on Large Language Models for Natural Language Processing and Information Retrieval, with broad interests in Computer Vision, multimodal learning, and constrained generative models.
FrameworksAmazon AGI
Amazon AGI
Amazon Alexa AI
Amazon Alexa AI
SpazioDati SRL
University of Trento
University of Trento
University of Trento
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics: ACL 2023
This paper proposes three pre-training objectives designed to mimic the downstream fine-tuning task of contextual Answer Sentence Selection, and shows that their pre- training approaches (applied to RoBERTa and ELECTRA) can improve baseline contextual AS2 accuracy by up to 8% on some datasets.
Notes:
Conference on Empirical Methods in Natural Language Processing: EMNLP 2022
This paper proposes three novel sentence-level transformer pre-training objectives that incorporate paragraph-level semantics within and across documents, to improve the performance of transformers for Answer Sentence Selection, and mitigate the requirement of large labeled datasets.
North American Chapter of the Association for Computational Linguistics: NAACL 2022
This paper shows that popular pre-trained transformers perform poorly when used for fine-tuning on multi-candidate inference tasks, and proposes a new pre-training objective that models the paragraph-level semantics across multiple input sentences.
PhD Thesis: 2023
Self-supervised pre-training tasks that align structurally with downstream applications, reducing the need for labeled data and achieving state-of-the-art results on various benchmark datasets are proposed.
Findings of the Association for Computational Linguistics: EMNLP 2022
It is proved that eliminating the MASK token and considering the whole output during the loss computation are essential choices to improve performance and it is shown that ELECTRA benefits heavily from a state-of-the-art hyper-parameters search.
The Journal of Open Source Software: JOSS 2022
A main problem with reproducing machine learning publications is the variance of metric implementations across papers. A lack of standardization leads to different behavior in mechanisms such as checkpointing, learning rate schedulers or early stopping, that will influence the reported results. For example, a complex metric such as Fréchet inception distance (FID) for synthetic image quality evaluation (Heusel et al., 2017) will differ based on the specific interpolation method used.
Notes:
The Thirty-Sixth Annual Conference on Neural Information Processing Systems: NeurIPS 2020
Constrained Adversarial Networks are proposed, an extension of GANs in which the constraints are embedded into the model during training and which efficiently generate valid structures that are both high-quality and novel.
Book: Compendium of Neurosymbolic Artificial Intelligence
Series: Frontiers in Artificial Intelligence and Applications
This chapter introduces the semantic loss, a training method that incorporates symbolic structural constraints into neural networks to ensure outputs respect underlying dependencies (like valid graph paths). We enhance it with entropy minimization to prefer simpler solutions and demonstrate its versatility by integrating it with both discriminative and generative models, enabling efficient generation of complex structured objects.