Meneghetti, Laura
A Continuous Convolutional Trainable Filter for Modelling Unstructured Data
Coscia, Dario, Meneghetti, Laura, Demo, Nicola, Stabile, Giovanni, Rozza, Gianluigi
In the deep learning field, a convolutional neural network (CNN) [28] is one of the most important architectures, widely used in academia and industrial research. For an overview of the topic, the interested reader might refer to [30, 16, 2, 5, 52]. Despite the great success in many fields including, but not limited, to computer vision [26, 40, 22] or natural language processing [50, 11], current CNNs are constrained to structural data. Indeed, the basic building block of a CNN is a trainable filter, represented by a discrete grid, which performs cross-correlation, also known as convolution, on a discrete domain. Nevertheless, the idea behind convolution can be easily extended mathematically to unstructured domains, for reference see [18]. One possible approach for this kind of problem is the graph neural networks (GNN) [24, 49], where a graph is built starting from the topology of the discretized space.
A Proper Orthogonal Decomposition approach for parameters reduction of Single Shot Detector networks
Meneghetti, Laura, Demo, Nicola, Rozza, Gianluigi
As a major breakthrough in artificial intelligence and deep learning, Convolutional Neural Networks have achieved an impressive success in solving many problems in several fields including computer vision and image processing. Real-time performance, robustness of algorithms and fast training processes remain open problems in these contexts. In addition object recognition and detection are challenging tasks for resource-constrained embedded systems, commonly used in the industrial sector. To overcome these issues, we propose a dimensionality reduction framework based on Proper Orthogonal Decomposition, a classical model order reduction technique, in order to gain a reduction in the number of hyperparameters of the net. We have applied such framework to SSD300 architecture using PASCAL VOC dataset, demonstrating a reduction of the network dimension and a remarkable speedup in the fine-tuning of the network in a transfer learning context.