Not enough data to create a plot.
Try a different view from the menu above.
Ye, Xiucai
MoFormer: Multi-objective Antimicrobial Peptide Generation Based on Conditional Transformer Joint Multi-modal Fusion Descriptor
Wang, Li, Fu, Xiangzheng, Yang, Jiahao, Zhang, Xinyi, Ye, Xiucai, Liu, Yiping, Sakurai, Tetsuya, Zeng, Xiangxiang
Deep learning holds a big promise for optimizing existing peptides with more desirable properties, a critical step towards accelerating new drug discovery. Despite the recent emergence of several optimized Antimicrobial peptides(AMP) generation methods, multi-objective optimizations remain still quite challenging for the idealism-realism tradeoff. Here, we establish a multi-objective AMP synthesis pipeline (MoFormer) for the simultaneous optimization of multi-attributes of AMPs. MoFormer improves the desired attributes of AMP sequences in a highly structured latent space, guided by conditional constraints and fine-grained multi-descriptor.We show that MoFormer outperforms existing methods in the generation task of enhanced antimicrobial activity and minimal hemolysis. We also utilize a Pareto-based non-dominated sorting algorithm and proxies based on large model fine-tuning to hierarchically rank the candidates. We demonstrate substantial property improvement using MoFormer from two perspectives: (1) employing molecular simulations and scoring interactions among amino acids to decipher the structure and functionality of AMPs; (2) visualizing latent space to examine the qualities and distribution features, verifying an effective means to facilitate multi-objective optimization AMPs with design constraints.
HMAMP: Hypervolume-Driven Multi-Objective Antimicrobial Peptides Design
Wang, Li, Li, Yiping, Fu, Xiangzheng, Ye, Xiucai, Shi, Junfeng, Yen, Gary G., Zeng, Xiangxiang
Antimicrobial peptides (AMPs) have exhibited unprecedented potential as biomaterials in combating multidrug-resistant bacteria. Despite the increasing adoption of artificial intelligence for novel AMP design, challenges pertaining to conflicting attributes such as activity, hemolysis, and toxicity have significantly impeded the progress of researchers. This paper introduces a paradigm shift by considering multiple attributes in AMP design. Presented herein is a novel approach termed Hypervolume-driven Multi-objective Antimicrobial Peptide Design (HMAMP), which prioritizes the simultaneous optimization of multiple attributes of AMPs. By synergizing reinforcement learning and a gradient descent algorithm rooted in the hypervolume maximization concept, HMAMP effectively expands exploration space and mitigates the issue of pattern collapse. This method generates a wide array of prospective AMP candidates that strike a balance among diverse attributes. Furthermore, we pinpoint knee points along the Pareto front of these candidate AMPs. Empirical results across five benchmark models substantiate that HMAMP-designed AMPs exhibit competitive performance and heightened diversity. A detailed analysis of the helical structures and molecular dynamics simulations for ten potential candidate AMPs validates the superiority of HMAMP in the realm of multi-objective AMP design. The ability of HMAMP to systematically craft AMPs considering multiple attributes marks a pioneering milestone, establishing a universal computational framework for the multi-objective design of AMPs.
LSEC: Large-scale spectral ensemble clustering
Li, Hongmin, Ye, Xiucai, Imakura, Akira, Sakurai, Tetsuya
Ensemble clustering is a fundamental problem in the machine learning field, combining multiple base clusterings into a better clustering result. However, most of the existing methods are unsuitable for large-scale ensemble clustering tasks due to the efficiency bottleneck. In this paper, we propose a large-scale spectral ensemble clustering (LSEC) method to strike a good balance between efficiency and effectiveness. In LSEC, a large-scale spectral clustering based efficient ensemble generation framework is designed to generate various base clusterings within a low computational complexity. Then all based clustering are combined through a bipartite graph partition based consensus function into a better consensus clustering result. The LSEC method achieves a lower computational complexity than most existing ensemble clustering methods. Experiments conducted on ten large-scale datasets show the efficiency and effectiveness of the LSEC method. The MATLAB code of the proposed method and experimental datasets are available at https://github.com/Li- Hongmin/MyPaperWithCode.
Multiclass spectral feature scaling method for dimensionality reduction
Matsuda, Momo, Morikuni, Keiichi, Imakura, Akira, Ye, Xiucai, Sakurai, Tetsuya
Dimensionality reduction is a technique for reducing the number of variables of data samples and has been successfully applied in many fields to make machine learning algorithms faster and more accurate, including the pathological diagnoses of gene expression data [26], the analysis of chemical sensor data [16], the community detection in social networks [27], the analyses of neural spike sorting [1], and others [22]. Due to their dependence on label information, dimensionality reduction methods can be divided into supervised and unsupervised methods. Typical unsupervised dimensionality reduction methods are the principal component analysis (PCA) [12, 15], the classical multidimensional scaling (MDS) [4], the locality preserving projections (LPP) [11], and the t-distributed stochastic neighbor embedding (t-SNE) [28]. To make use of prior knowledge on the labels, we focus on supervised dimensionality reduction methods. Supervised dimensionality reduction methods map data samples into an optimal low-dimensional space for satisfactory classification while incorporating the label information. One of the most popular supervised dimensionality reduction methods is the linear discriminant analysis (LDA) [3], which maximizes the between-class scatter and reduces the within-class scatter in a low-dimensional space.
Global Discriminant Analysis for Unsupervised Feature Selection with Local Structure Preservation
Ye, Xiucai (University of Tsukuba) | Ji, Kaiyang (University of Tsukuba) | Sakurai, Tetsuya (University of Tsukuba)
Feature selection is an efficient technique for data dimension reduction in data mining and machine learning. Unsupervised feature selection is much more difficult than supervised feature selection due to the lack of label information. Discriminant analysis is powerful to select discriminative features, while local structure preservation is important to unsupervised feature selection. In this paper, we incorporate discriminant analysis, local structure preservation and l2,1-norm regularization into a joint framework for unsupervised feature selection. The global structure of data is captured by the discriminant analysis, while the local manifold structure is revealed by the locality preserving projections. By imposing row sparsity on the transformation matrix, the resultant formulation optimizes for selecting the most discriminative features which can better capture both the global and local structure of data. We develop an efficient algorithm to solve the l2,1-norm-based optimization problem in our method. Experimental results on different types of real-world data demonstrate the effectiveness of the proposed method.