Publications

Explore our research publications: papers, articles, and conference proceedings from AImageLab.

Tip: type @ to pick an author and # to pick a keyword.

M-VAD Names: a Dataset for Video Captioning with Naming

Authors: Pini, Stefano; Cornia, Marcella; Bolelli, Federico; Baraldi, Lorenzo; Cucchiara, Rita

Published in: MULTIMEDIA TOOLS AND APPLICATIONS

Current movie captioning architectures are not capable of mentioning characters with their proper name, replacing them with a generic "someone" … (Read full abstract)

Current movie captioning architectures are not capable of mentioning characters with their proper name, replacing them with a generic "someone" tag. The lack of movie description datasets with characters' visual annotations surely plays a relevant role in this shortage. Recently, we proposed to extend the M-VAD dataset by introducing such information. In this paper, we present an improved version of the dataset, namely M-VAD Names, and its semi-automatic annotation procedure. The resulting dataset contains 63k visual tracks and 34k textual mentions, all associated with character identities. To showcase the features of the dataset and quantify the complexity of the naming task, we investigate multimodal architectures to replace the "someone" tags with proper character names in existing video captions. The evaluation is further extended by testing this application on videos outside of the M-VAD Names dataset.

2019 Articolo su rivista

Manual Annotations on Depth Maps for Human Pose Estimation

Authors: D'Eusanio, Andrea; Pini, Stefano; Borghi, Guido; Vezzani, Roberto; Cucchiara, Rita

Few works tackle the Human Pose Estimation on depth maps. Moreover, these methods usually rely on automatically annotated datasets, and … (Read full abstract)

Few works tackle the Human Pose Estimation on depth maps. Moreover, these methods usually rely on automatically annotated datasets, and these annotations are often imprecise and unreliable, limiting the achievable accuracy using this data as ground truth. For this reason, in this paper we propose an annotation refinement tool of human poses, by means of body joints, and a novel set of fine joint annotations for the Watch-n-Patch dataset, which has been collected with the proposed tool. Furthermore, we present a fully-convolutional architecture that performs the body pose estimation directly on depth maps. The extensive evaluation shows that the proposed architecture outperforms the competitors in different training scenarios and is able to run in real-time.

2019 Relazione in Atti di Convegno

METODO DI VALUTAZIONE DI UNO STATO DI SALUTE DI UN ELEMENTO ANATOMICO, RELATIVO DISPOSITIVO DI VALUTAZIONE E RELATIVO SISTEMA DI VALUTAZIONE

Authors: Giuseppe, Marrucchella; Bergamini, Luca; Porrello, Angelo; Del Negro, Ercole; Capobianco Dondona, Andrea; Di Tondo, Francesco; Calderara, Simone

Sistema in grado di rilevare le lesioni delle mezzene al macello attraverso l'utilizzo di tecniche di deep learning per individuazioni … (Read full abstract)

Sistema in grado di rilevare le lesioni delle mezzene al macello attraverso l'utilizzo di tecniche di deep learning per individuazioni del tipo di lesioni presenti

2019 Brevetto

Novel and Rare Fusion Transcripts Involving Transcription Factors and Tumor Suppressor Genes in Acute Myeloid Leukemia

Authors: Padella And, Antonella; Simonetti And, Giorgia; Paciello And, Giulia; Giotopoulos And, George; Baldazzi And, Carmen; Righi And, Simona; Ghetti And, Martina; Stengel And, Anna; Guadagnuolo And, Viviana; De Tommaso And, Rossella; Papayannidis And, Cristina; Robustelli And, Valentina; Franchini And, Eugenia; Ghelli Luserna Di Rorà And, Andrea; Ferrari And, Anna; Fontana And Maria, Chiara; Bruno And, Samantha; Ottaviani And, Emanuela; Soverini And, Simona; Storlazzi And Clelia, Tiziana; Haferlach And, Claudia; Sabattini And, Elena; Testoni And, Nicoletta; Iacobucci And, Ilaria; Huntly And Brian, J. P.; Ficarra, Elisa; Martinelli And, Giovanni

Published in: CANCERS

Approximately 18% of acute myeloid leukemia (AML) cases express a fusion transcript. However, few fusions are recurrent across AML and … (Read full abstract)

Approximately 18% of acute myeloid leukemia (AML) cases express a fusion transcript. However, few fusions are recurrent across AML and the identification of these rare chimeras is of interest to characterize AML patients. Here, we studied the transcriptome of 8 adult AML patients with poorly described chromosomal translocation(s), with the aim of identifying novel and rare fusion transcripts. We integrated RNA-sequencing data with multiple approaches including computational analysis, Sanger sequencing, fluorescence in situ hybridization and in vitro studies to assess the oncogenic potential of the ZEB2-BCL11B chimera. We detected 7 different fusions with partner genes involving transcription factors (OAZ-MAFK, ZEB2-BCL11B), tumor suppressors (SAV1-GYPB, PUF60-TYW1, CNOT2-WT1) and rearrangements associated with the loss of NF1 (CPD-PXT1, UTP6-CRLF3). Notably, ZEB2-BCL11B rearrangements co-occurred with FLT3 mutations and were associated with a poorly differentiated or mixed phenotype leukemia. Although the fusion alone did not transform murine c-Kit+ bone marrow cells, 45.4% of 14q32 non-rearranged AML cases were also BCL11B-positive, suggesting a more general and complex mechanism of leukemogenesis associated with BCL11B expression. Overall, by combining different approaches, we described rare fusion events contributing to the complexity of AML and we linked the expression of some chimeras to genomic alterations hitting known genes in AML.

2019 Articolo su rivista

OpenFACS: An Open Source FACS-Based 3D Face Animation System

Authors: Cuculo, V.; D'Amelio, A.

Published in: LECTURE NOTES IN COMPUTER SCIENCE

We present OpenFACS, an open source FACS-based 3D face animation system. OpenFACS is a software that allows the simulation of … (Read full abstract)

We present OpenFACS, an open source FACS-based 3D face animation system. OpenFACS is a software that allows the simulation of realistic facial expressions through the manipulation of specific action units as defined in the Facial Action Coding System. OpenFACS has been developed together with an API which is suitable to generate real-time dynamic facial expressions for a three-dimensional character. It can be easily embedded in existing systems without any prior experience in computer graphics. In this note, we discuss the adopted face model, the implemented architecture and provide additional details of model dynamics. Finally, a validation experiment is proposed to assess the effectiveness of the model.

2019 Relazione in Atti di Convegno

Precision computation of wind turbine power upgrades: An aerodynamic and control optimization test case

Authors: Astolfi, D.; Castellani, F.; Fravolini, M. L.; Cascianelli, S.; Terzi, L.

Published in: JOURNAL OF ENERGY RESOURCES TECHNOLOGY

Wind turbine upgrades have recently been spreading in the wind energy industry for optimizing the efficiency of the wind kinetic … (Read full abstract)

Wind turbine upgrades have recently been spreading in the wind energy industry for optimizing the efficiency of the wind kinetic energy conversion. These interventions have material and labor costs; therefore, it is fundamental to estimate the production improvement realistically. Furthermore, the retrofitting of the wind turbines sited in complex environments might exacerbate the stress conditions to which those are subjected and consequently might affect the residual life. In this work, a two-step upgrade on a multimegawatt wind turbine is considered from a wind farm sited in complex terrain. First, vortex generators and passive flow control devices have been installed. Second, the management of the revolutions per minute has been optimized. In this work, a general method is formulated for assessing the wind turbine power upgrades using operational data. The method is based on the study of the residuals between the measured power output and a judicious model of the power output itself, before and after the upgrade. Therefore, properly selecting the model is fundamental. For this reason, an automatic feature selection algorithm is adopted, based on the stepwise multivariate regression. This allows identifying the most meaningful input variables for a multivariate linear model whose target is the power of the upgraded wind turbine. For the test case of interest, the adopted upgrade is estimated to increase the annual energy production to 2.660.1%. The aerodynamic and control upgrades are estimated to be 1.8% and 0.8%, respectively, of the production improvement.

2019 Articolo su rivista

Predicting the Driver's Focus of Attention: the DR(eye)VE Project

Authors: Palazzi, Andrea; Abati, Davide; Calderara, Simone; Solera, Francesco; Cucchiara, Rita

Published in: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE

Predicting the Driver's Focus of Attention: the DR(eye)VE Project Andrea Palazzi, Davide Abati, Simone Calderara, Francesco Solera, Rita Cucchiara (Submitted … (Read full abstract)

Predicting the Driver's Focus of Attention: the DR(eye)VE Project Andrea Palazzi, Davide Abati, Simone Calderara, Francesco Solera, Rita Cucchiara (Submitted on 10 May 2017 (v1), last revised 6 Jun 2018 (this version, v3)) In this work we aim to predict the driver's focus of attention. The goal is to estimate what a person would pay attention to while driving, and which part of the scene around the vehicle is more critical for the task. To this end we propose a new computer vision model based on a multi-branch deep architecture that integrates three sources of information: raw video, motion and scene semantics. We also introduce DR(eye)VE, the largest dataset of driving scenes for which eye-tracking annotations are available. This dataset features more than 500,000 registered frames, matching ego-centric views (from glasses worn by drivers) and car-centric views (from roof-mounted camera), further enriched by other sensors measurements. Results highlight that several attention patterns are shared across drivers and can be reproduced to some extent. The indication of which elements in the scene are likely to capture the driver's attention may benefit several applications in the context of human-vehicle interaction and driver attention analysis.

2019 Articolo su rivista

Predictive Sampling of Facial Expression Dynamics Driven by a Latent Action Space

Authors: Boccignone, G.; Bodini, M.; Cuculo, V.; Grossi, G.

We present a probabilistic generative model for tracking by prediction the dynamics of affective spacial expressions in videos. The model … (Read full abstract)

We present a probabilistic generative model for tracking by prediction the dynamics of affective spacial expressions in videos. The model relies on Bayesian filter sampling of facial landmarks conditioned on motor action parameter dynamics; namely, trajectories shaped by an autoregressive Gaussian Process Latent Variable state-space. The analysis-by-synthesis approach at the heart of the model allows for both inference and generation of affective expressions. Robustness of the method to occlusions and degradation of video quality has been assessed on a publicly available dataset.

2019 Relazione in Atti di Convegno

Problems with Saliency Maps

Authors: Boccignone, Giuseppe; Cuculo, Vittorio; D’Amelio, Alessandro

Published in: LECTURE NOTES IN COMPUTER SCIENCE

Despite the popularity that saliency models have gained in the computer vision community, they are most often conceived, exploited and … (Read full abstract)

Despite the popularity that saliency models have gained in the computer vision community, they are most often conceived, exploited and benchmarked without taking heed of a number of problems and subtle issues they bring about. When saliency maps are used as proxies for the likelihood of fixating a location in a viewed scene, one such issue is the temporal dimension of visual attention deployment. Through a simple simulation it is shown how neglecting this dimension leads to results that at best cast shadows on the predictive performance of a model and its assessment via benchmarking procedures.

2019 Relazione in Atti di Convegno

Recognizing social relationships from an egocentric vision perspective

Authors: Alletto, Stefano; Cornia, Marcella; Baraldi, Lorenzo; Serra, Giuseppe; Cucchiara, Rita

In this chapter we address the problem of partitioning social gatherings into interacting groups in egocentric scenarios. People in the … (Read full abstract)

In this chapter we address the problem of partitioning social gatherings into interacting groups in egocentric scenarios. People in the scene are tracked, their head pose and 3D location are estimated. Following the formalism of the f-formation, we define with the orientation and distance an inherently social pairwise feature capable of describing how two people stand in relation to one another. We present a Structural SVM based approach to learn how to weight each component of the feature vector depending on the social situation is applied to. To better understand the social dynamics, we also estimate what we call social relevance of each subject in a group using a saliency attentive model. Extensive tests on two publicly available datasets show that our solution achieves encouraging results when detecting social groups and their relevant subjects in the challenging egocentric scenarios.

2019 Capitolo/Saggio

Page 46 of 106 • Total publications: 1054