Publications by Vittorio Cuculo

Explore our research publications: papers, articles, and conference proceedings from AImageLab.

Tip: type @ to pick an author and # to pick a keyword.

Active filters (Clear): Author: Vittorio Cuculo

Anomaly detection from log files using unsupervised deep learning

Authors: Bursic, S.; Cuculo, V.; D'Amelio, A.

Published in: LECTURE NOTES IN COMPUTER SCIENCE

Computer systems have grown in complexity to the point where manual inspection of system behaviour for purposes of malfunction detection … (Read full abstract)

Computer systems have grown in complexity to the point where manual inspection of system behaviour for purposes of malfunction detection have become unfeasible. As these systems output voluminous logs of their activity, machine led analysis of them is a growing need with already several existing solutions. These largely depend on having hand-crafted features, require raw log preprocessing and feature extraction or use supervised learning necessitating having a labeled log dataset not always easily procurable. We propose a two part deep autoencoder model with LSTM units that requires no hand-crafted features, no preprocessing of data as it works on raw text and outputs an anomaly score for each log entry. This anomaly score represents the rarity of a log event both in terms of its content and temporal context. The model was trained and tested on a dataset of HDFS logs containing 2 million raw lines of which half was used for training and half for testing. While this model cannot match the performance of a supervised binary classifier, it could be a useful tool as a coarse filter for manual inspection of log files where a labeled dataset is unavailable.

2020 Relazione in Atti di Convegno

Gender recognition in the wild with small sample size : A dictionary learning approach

Authors: D'Amelio, A.; Cuculo, V.; Bursic, S.

Published in: LECTURE NOTES IN COMPUTER SCIENCE

In this work we address the problem of gender recognition from facial images acquired in the wild. This problem is … (Read full abstract)

In this work we address the problem of gender recognition from facial images acquired in the wild. This problem is particularly difficult due to the presence of variations in pose, ethnicity, age and image quality. Moreover, we consider the special case in which only a small sample size is available for the training phase. We rely on a feature representation obtained from the well known VGG-Face Deep Convolutional Neural Network (DCNN) and exploit the effectiveness of a sparse-driven sub-dictionary learning strategy which has proven to be able to represent both local and global characteristics of the train and probe faces. Results on the publicly available LFW dataset are provided in order to demonstrate the effectiveness of the proposed method.

2020 Relazione in Atti di Convegno

How to look next? A data-driven approach for scanpath prediction

Authors: Boccignone, G.; Cuculo, V.; D'Amelio, A.

Published in: LECTURE NOTES IN COMPUTER SCIENCE

By and large, current visual attention models mostly rely, when considering static stimuli, on the following procedure. Given an image, … (Read full abstract)

By and large, current visual attention models mostly rely, when considering static stimuli, on the following procedure. Given an image, a saliency map is computed, which, in turn, might serve the purpose of predicting a sequence of gaze shifts, namely a scanpath instantiating the dynamics of visual attention deployment. The temporal pattern of attention unfolding is thus confined to the scanpath generation stage, whilst salience is conceived as a static map, at best conflating a number of factors (bottom-up information, top-down, spatial biases, etc.). In this note we propose a novel sequential scheme that consists of a three-stage processing relying on a center-bias model, a context/layout model, and an object-based model, respectively. Each stage contributes, at different times, to the sequential sampling of the final scanpath. We compare the method against classic scanpath generation that exploits state-of-the-art static saliency model. Results show that accounting for the structure of the temporal unfolding leads to gaze dynamics close to human gaze behaviour.

2020 Relazione in Atti di Convegno

On Gaze Deployment to Audio-Visual Cues of Social Interactions

Authors: Boccignone, G.; Cuculo, V.; D'Amelio, A.; Grossi, G.; Lanzarotti, R.

Published in: IEEE ACCESS

Attention supports our urge to forage on social cues. Under certain circumstances, we spend the majority of time scrutinising people, … (Read full abstract)

Attention supports our urge to forage on social cues. Under certain circumstances, we spend the majority of time scrutinising people, markedly their eyes and faces, and spotting persons that are talking. To account for such behaviour, this article develops a computational model for the deployment of gaze within a multimodal landscape, namely a conversational scene. Gaze dynamics is derived in a principled way by reformulating attention deployment as a stochastic foraging problem. Model simulation experiments on a publicly available dataset of eye-tracked subjects are presented. Results show that the simulated scan paths exhibit similar trends of eye movements of human observers watching and listening to conversational clips in a free-viewing condition

2020 Articolo su rivista

Give Ear to My Face: Modelling Multimodal Attention to Social Interactions

Authors: Boccignone, Giuseppe; Cuculo, Vittorio; D’Amelio, Alessandro; Grossi, Giuliano; Lanzarotti, Raffaella

Published in: LECTURE NOTES IN COMPUTER SCIENCE

We address the deployment of perceptual attention to social interactions as displayed in conversational clips, when relying on multimodal information … (Read full abstract)

We address the deployment of perceptual attention to social interactions as displayed in conversational clips, when relying on multimodal information (audio and video). A probabilistic modelling framework is proposed that goes beyond the classic saliency paradigm while integrating multiple information cues. Attentional allocation is determined not just by stimulus-driven selection but, importantly, by social value as modulating the selection history of relevant multimodal items. Thus, the construction of attentional priority is the result of a sampling procedure conditioned on the potential value dynamics of socially relevant objects emerging moment to moment within the scene. Preliminary experiments on a publicly available dataset are presented.

2019 Relazione in Atti di Convegno

OpenFACS: An Open Source FACS-Based 3D Face Animation System

Authors: Cuculo, V.; D'Amelio, A.

Published in: LECTURE NOTES IN COMPUTER SCIENCE

We present OpenFACS, an open source FACS-based 3D face animation system. OpenFACS is a software that allows the simulation of … (Read full abstract)

We present OpenFACS, an open source FACS-based 3D face animation system. OpenFACS is a software that allows the simulation of realistic facial expressions through the manipulation of specific action units as defined in the Facial Action Coding System. OpenFACS has been developed together with an API which is suitable to generate real-time dynamic facial expressions for a three-dimensional character. It can be easily embedded in existing systems without any prior experience in computer graphics. In this note, we discuss the adopted face model, the implemented architecture and provide additional details of model dynamics. Finally, a validation experiment is proposed to assess the effectiveness of the model.

2019 Relazione in Atti di Convegno

Predictive Sampling of Facial Expression Dynamics Driven by a Latent Action Space

Authors: Boccignone, G.; Bodini, M.; Cuculo, V.; Grossi, G.

We present a probabilistic generative model for tracking by prediction the dynamics of affective spacial expressions in videos. The model … (Read full abstract)

We present a probabilistic generative model for tracking by prediction the dynamics of affective spacial expressions in videos. The model relies on Bayesian filter sampling of facial landmarks conditioned on motor action parameter dynamics; namely, trajectories shaped by an autoregressive Gaussian Process Latent Variable state-space. The analysis-by-synthesis approach at the heart of the model allows for both inference and generation of affective expressions. Robustness of the method to occlusions and degradation of video quality has been assessed on a publicly available dataset.

2019 Relazione in Atti di Convegno

Problems with Saliency Maps

Authors: Boccignone, Giuseppe; Cuculo, Vittorio; D’Amelio, Alessandro

Published in: LECTURE NOTES IN COMPUTER SCIENCE

Despite the popularity that saliency models have gained in the computer vision community, they are most often conceived, exploited and … (Read full abstract)

Despite the popularity that saliency models have gained in the computer vision community, they are most often conceived, exploited and benchmarked without taking heed of a number of problems and subtle issues they bring about. When saliency maps are used as proxies for the likelihood of fixating a location in a viewed scene, one such issue is the temporal dimension of visual attention deployment. Through a simple simulation it is shown how neglecting this dimension leads to results that at best cast shadows on the predictive performance of a model and its assessment via benchmarking procedures.

2019 Relazione in Atti di Convegno

Robust single-sample face recognition by sparsity-driven sub-dictionary learning using deep features

Authors: Cuculo, Vittorio; D'Amelio, Alessandro; Grossi, Giuliano; Lanzarotti, Raffaella; Lin, Jianyi

Published in: SENSORS

Face recognition using a single reference image per subject is challenging, above all when referring to a large gallery of … (Read full abstract)

Face recognition using a single reference image per subject is challenging, above all when referring to a large gallery of subjects. Furthermore, the problem hardness seriously increases when the images are acquired in unconstrained conditions. In this paper we address the challenging Single Sample Per Person (SSPP) problem considering large datasets of images acquired in the wild, thus possibly featuring illumination, pose, face expression, partial occlusions, and low-resolution hurdles. The proposed technique alternates a sparse dictionary learning technique based on the method of optimal direction and the iterative ℓ 0 -norm minimization algorithm called k-LIMAPS. It works on robust deep-learned features, provided that the image variability is extended by standard augmentation techniques. Experiments show the effectiveness of our method against the hardness introduced above: first, we report extensive experiments on the unconstrained LFW dataset when referring to large galleries up to 1680 subjects; second, we present experiments on very low-resolution test images up to 8 × 8 pixels; third, tests on the AR dataset are analyzed against specific disguises such as partial occlusions, facial expressions, and illumination problems. In all the three scenarios our method outperforms the state-of-the-art approaches adopting similar configurations.

2019 Articolo su rivista

Social traits from stochastic paths in the core affect space

Authors: Boccignone, Giuseppe; Cuculo, Vittorio; D'Amelio, Alessandro; Lanzarotti, Raffaella

We discuss a preliminary investigation on the feasibility of inferring traits of social participation from the observable behaviour of individuals … (Read full abstract)

We discuss a preliminary investigation on the feasibility of inferring traits of social participation from the observable behaviour of individuals involved in dyadic interactions. Trait inference relies on a stochastic model of the dynamics occurring in the individual core affect state-space. Results obtained on a publicly available interaction dataset are presented and examined.

2019 Relazione in Atti di Convegno

Page 3 of 4 • Total publications: 39