Goto

Collaborating Authors

 cell structure


What is the best RNN-cell structure to forecast each time series behavior?

Khaldi, Rohaifa, Afia, Abdellatif El, Chiheb, Raddouane, Tabik, Siham

arXiv.org Artificial Intelligence

It is unquestionable that time series forecasting is of paramount importance in many fields. The most used machine learning models to address time series forecasting tasks are Recurrent Neural Networks (RNNs). Typically, those models are built using one of the three most popular cells: ELMAN, Long Short-Term Memory (LSTM), or Gated Recurrent Unit (GRU) cells. Each cell has a different structure and implies a different computational cost. However, it is not clear why and when to use each RNN-cell structure. Actually, there is no comprehensive characterization of all the possible time series behaviors and no guidance on what RNN cell structure is the most suitable for each behavior. The objective of this study is twofold: it presents a comprehensive taxonomy of almost all time series behaviors and provides insights into the best RNN cell structure for each time series behavior. We conducted two experiments: (1) We evaluate and analyze the role of each component in the LSTM-Vanilla cell by creating 11 variants based on one alteration in its basic architecture (removing, adding, or substituting one cell component). (2) We evaluate and analyze the performance of 20 possible RNN-cell structures. To evaluate, compare, and select the best model, different statistical metrics were used: error-based metrics, information criterion-based metrics, naive-based metrics, and direction change-based metrics. To further improve our confidence in the models interpretation and selection, the Friedman Wilcoxon-Holm signed-rank test was used. Our results advocate the usage and exploration of the newly created RNN variant, named SLIM, in time series forecasting thanks to its high ability to accurately predict the different time series behaviors, as well as its simple structural design that does not require expensive temporal and computing resources.


A Polarization and Radiomics Feature Fusion Network for the Classification of Hepatocellular Carcinoma and Intrahepatic Cholangiocarcinoma

Dong, Jia, Yao, Yao, Lin, Liyan, Dong, Yang, Wan, Jiachen, Peng, Ran, Li, Chao, Ma, Hui

arXiv.org Machine Learning

Classifying hepatocellular carcinoma (HCC) and intrahepatic cholangiocarcinoma (ICC) is a critical step in treatment selection and prognosis evaluation for patients with liver diseases. Traditional histopathological diagnosis poses challenges in this context. In this study, we introduce a novel polarization and radiomics feature fusion network, which combines polarization features obtained from Mueller matrix images of liver pathological samples with radiomics features derived from corresponding pathological images to classify HCC and ICC. Our fusion network integrates a two-tier fusion approach, comprising early feature-level fusion and late classification-level fusion. By harnessing the strengths of polarization imaging techniques and image feature-based machine learning, our proposed fusion network significantly enhances classification accuracy. Notably, even at reduced imaging resolutions, the fusion network maintains robust performance due to the additional information provided by polarization features, which may not align with human visual perception. Our experimental results underscore the potential of this fusion network as a powerful tool for computer-aided diagnosis of HCC and ICC, showcasing the benefits and prospects of integrating polarization imaging techniques into the current image-intensive digital pathological diagnosis. We aim to contribute this innovative approach to top-tier journals, offering fresh insights and valuable tools in the fields of medical imaging and cancer diagnosis. By introducing polarization imaging into liver cancer classification, we demonstrate its interdisciplinary potential in addressing challenges in medical image analysis, promising advancements in medical imaging and cancer diagnosis.


Kohonen Feature Maps and Growing Cell Structures - a Performance Comparison

Neural Information Processing Systems

A performance comparison of two self-organizing networks, the Ko(cid:173) honen Feature Map and the recently proposed Growing Cell Struc(cid:173) tures is made. For this purpose several performance criteria for self-organizing networks are proposed and motivated. The models are tested with three example problems of increasing difficulty. The Kohonen Feature Map demonstrates slightly superior results only for the simplest problem. Additional advantages of the new model are that all parameters are constant over time and that size as well as structure of the network are determined auto(cid:173) matically.


Supervised Learning with Growing Cell Structures

Neural Information Processing Systems

We present a new incremental radial basis function network suit(cid:173) able for classification and regression problems. Center positions are continuously updated through soft competitive learning. The width of the radial basis functions is derived from the distance to topological neighbors. During the training the observed error is accumulated locally and used to determine where to insert the next unit. This leads (in case of classification problems) to the placement of units near class borders rather than near frequency peaks as is done by most existing methods.


A Survey on Evaluation Metrics for Synthetic Material Micro-Structure Images from Generative Models

Shah, Devesh, Suresh, Anirudh, Admasu, Alemayehu, Upadhyay, Devesh, Deb, Kalyanmoy

arXiv.org Artificial Intelligence

The evaluation of synthetic micro-structure images is an emerging problem as machine learning and materials science research have evolved together. Typical state of the art methods in evaluating synthetic images from generative models have relied on the Fr\'echet Inception Distance. However, this and other similar methods, are limited in the materials domain due to both the unique features that characterize physically accurate micro-structures and limited dataset sizes. In this study we evaluate a variety of methods on scanning electron microscope (SEM) images of graphene-reinforced polyurethane foams. The primary objective of this paper is to report our findings with regards to the shortcomings of existing methods so as to encourage the machine learning community to consider enhancements in metrics for assessing quality of synthetic images in the material science domain.


Using artificial intelligence to build a better organoid

#artificialintelligence

The next breakthrough for artificial intelligence may help scientists design and print smarter cell structures that mimic human organs - thereby reducing trial-and-error and costs while contributing to new methods to fight disease and improve human health. These three-dimensional cell structures are called organoids - they can be designed to exhibit unique organ function and are used to conduct complex research on human tissue physiology, genetic diseases, organ-specific infectious diseases and cancer. Organoids can replicate development of native tissue in the architecture, function, cellular composition and transcriptional profile of almost any organ. Current methods of manufacturing organoids have yet to demonstrate consistent and robust extraction of mature organoids from renewable cells. This is where AI comes into the picture - designing and testing organoids utilizing computers rather than traditional lab methods.


Differentiable Neural Architecture Search with Morphism-based Transformable Backbone Architectures

Jie, Renlong, Gao, Junbin

arXiv.org Artificial Intelligence

This study aims at making the architecture search process more adaptive for one-shot or online training. It is extended from the existing study on differentiable neural architecture search, and we made the backbone architecture transformable rather than fixed during the training process. As is known, differentiable neural architecture search (DARTS) requires a pre-defined over-parameterized backbone architecture, while its size is to be determined manually. Also, in DARTS backbone, Hadamard product of two elements is not introduced, which exists in both LSTM and GRU cells for recurrent nets. This study introduces a growing mechanism for differentiable neural architecture search based on network morphism. It enables growing of the cell structures from small size towards large size ones with one-shot training. Two modes can be applied in integrating the growing and original pruning process. We also implement a recently proposed two-input backbone architecture for recurrent neural networks. Initial experimental results indicate that our approach and the two-input backbone structure can be quite effective compared with other baseline architectures including LSTM, in a variety of learning tasks including multi-variate time series forecasting and language modeling. On the other hand, we find that dynamic network transformation is promising in improving the efficiency of differentiable architecture search.


AI spots cell structures that humans can't

#artificialintelligence

Susanne Rafelski and her colleagues had a deceptively simple goal. "We wanted to be able to label many different structures in the cell, but do live imaging," says the quantitative cell biologist and deputy director of the Allen Institute for Cell Science in Seattle, Washington. "And we wanted to do it in 3D." That kind of goal normally relies on fluorescence microscopy -- problematic in this case because, with only a handful of colours to use, the scientists would run out of labels well before they ran out of structures. Also problematic is that these reagents are pricey and laborious to use.


Using Neural Architecture Search for Improving Software Flaw Detection in Multimodal Deep Learning Models

Cooper, Alexis, Zhou, Xin, Heidbrink, Scott, Dunlavy, Daniel M.

arXiv.org Artificial Intelligence

Software flaw detection using multimodal deep learning models has been demonstrated as a very competitive approach on benchmark problems. In this work, we demonstrate that even better performance can be achieved using neural architecture search (NAS) combined with multimodal learning models. We adapt a NAS framework aimed at investigating image classification to the problem of software flaw detection and demonstrate improved results on the Juliet Test Suite, a popular benchmarking data set for measuring performance of machine learning models in this problem domain.


NASGEM: Neural Architecture Search via Graph Embedding Method

Cheng, Hsin-Pai, Zhang, Tunhou, Li, Shiyu, Yan, Feng, Li, Meng, Chandra, Vikas, Li, Hai, Chen, Yiran

arXiv.org Artificial Intelligence

Neural Architecture Search (NAS) automates and prospers the design of neural networks. Recent studies show that mapping the discrete neural architecture search space into a continuous space which is more compact, more representative, and easier to optimize can significantly reduce the exploration cost. However, existing differentiable methods cannot preserve the graph information when projecting a neural architecture into a continuous space, causing inaccuracy and/or reduced representation capability in the mapped space. Moreover, existing methods can explore only a very limited inner-cell search space due to the cell representation limitation or poor scalability. To enable quick search of more sophisticated neural architectures while preserving graph information, we propose NASGEM which stands for Neural Architecture Search via Graph Embedding Method. NASGEM is driven by a novel graph embedding method integrated with similarity estimation to capture the inner-cell information in the discrete space. Thus, NASGEM is able to search a wider space (e.g., 30 nodes in a cell). By precisely estimating the graph distance, NASGEM can efficiently explore a large amount of candidate cells to enable a more flexible cell design while still keeping the search cost low. GEMNet, which is a set of networks discovered by NASGEM, has higher accuracy while less parameters (up to 62% less) and Multiply-Accumulates (up to 20.7% less) compared to networks crafted by existing differentiable search methods. Our ablation study on NASBench-101 further validates the effectiveness of the proposed graph embedding method, which is complementary to many existing NAS approaches and can be combined to achieve better performance.