Chen, Hung-Hsuan
From Structured Prompts to Open Narratives: Measuring Gender Bias in LLMs Through Open-Ended Storytelling
Chen, Evan, Zhan, Run-Jun, Lin, Yan-Bai, Chen, Hung-Hsuan
Large Language Models (LLMs) have revolutionized natural language processing, yet concerns persist regarding their tendency to reflect or amplify social biases present in their training data. This study introduces a novel evaluation framework to uncover gender biases in LLMs, focusing on their occupational narratives. Unlike previous methods relying on structured scenarios or carefully crafted prompts, our approach leverages free-form storytelling to reveal biases embedded in the models. Systematic analyses show an overrepresentation of female characters across occupations in six widely used LLMs. Additionally, our findings reveal that LLM-generated occupational gender rankings align more closely with human stereotypes than actual labor statistics. These insights underscore the need for balanced mitigation strategies to ensure fairness while avoiding the reinforcement of new stereotypes.
Dynamic DropConnect: Enhancing Neural Network Robustness through Adaptive Edge Dropping Strategies
Yang, Yuan-Chih, Chen, Hung-Hsuan
Dropout and DropConnect are well-known techniques that apply a consistent drop rate to randomly deactivate neurons or edges in a neural network layer during training. This paper introduces a novel methodology that assigns dynamic drop rates to each edge within a layer, uniquely tailoring the dropping process without incorporating additional learning parameters. We perform experiments on synthetic and openly available datasets to validate the effectiveness of our approach. The results demonstrate that our method outperforms Dropout, DropConnect, and Standout, a classic mechanism known for its adaptive dropout capabilities. Furthermore, our approach improves the robustness and generalization of neural network training without increasing computational complexity. The complete implementation of our methodology is publicly accessible for research and replication purposes at https://github.com/ericabd888/Adjusting-the-drop-probability-in-DropConnect-based-on-the-magnitude-of-the-gradient/.
Flexible Bivariate Beta Mixture Model: A Probabilistic Approach for Clustering Complex Data Structures
Hsu, Yung-Peng, Chen, Hung-Hsuan
This unsupervised learning method is widely used in various applications, including image analysis, information retrieval, text analysis, bioinformatics, and many more [1, 2, 3, 4]. Clustering helps uncover the underlying structure of the data, facilitates data summarization, and sometimes serves as a preprocessing step for other algorithms [2]. Despite its widespread use, one of the primary challenges many traditional clustering algorithms face is that they often assume that the data points form clusters with convex shapes. For example, centroid-based algorithms like k -means and distribution-based models like Gaussian Mixture Models (GMM) typically produce clusters that are hyperspherical or ellipsoidal [5]. Although this assumption simplifies the clustering process, it restricts the flexibility of these models to handle complex data distributions that do not conform to convex shapes.
Understanding Gradient Boosting Classifier: Training, Prediction, and the Role of $\gamma_j$
Chen, Hung-Hsuan
The Gradient Boosting Classifier (GBC) is a widely used machine learning algorithm for binary classification, which builds decision trees iteratively to minimize prediction errors. This document explains the GBC's training and prediction processes, focusing on the computation of terminal node values $\gamma_j$, which are crucial to optimizing the logistic loss function. We derive $\gamma_j$ through a Taylor series approximation and provide a step-by-step pseudocode for the algorithm's implementation. The guide explains the theory of GBC and its practical application, demonstrating its effectiveness in binary classification tasks. We provide a step-by-step example in the appendix to help readers understand.
Multivariate Beta Mixture Model: Probabilistic Clustering With Flexible Cluster Shapes
Hsu, Yung-Peng, Chen, Hung-Hsuan
Data clustering groups data points into components so that similar points are within the same component. Data clustering is commonly used for data exploration and is sometimes used as a preprocessing step for later analysis [1]. In this paper, the multivariate beta mixture model (MBMM), a new probabilistic model for soft clustering, is proposed. As the MBMM is a mixture model, it shares many properties with the Gaussian mixture model (GMM), including its soft cluster assignment and parametric modeling. In addition, the MBMM allows the generation of new (synthetic) instances based on a generative process. Because the beta distribution is highly flexible (e.g., unimodal, bimodal, straight line, or exponentially increasing or decreasing), MBMM can fit data with versatile shapes.
TTSWING: a Dataset for Table Tennis Swing Analysis
Chou, Che-Yu, Chen, Zheng-Hao, Sheu, Yung-Hoh, Chen, Hung-Hsuan, Wu, Sheng K.
Based on the collected dataset, we perform pilot studies on swing analyses using various machine learning models. We introduce TTSWING, a novel dataset designed Specifically, we investigate the feasibility of using machine for table tennis swing analysis. This dataset comprises learning techniques to automatically predict a player's age, comprehensive swing information obtained gender, playing experience in years, racket-holding hand, and through 9-axis sensors integrated into custom-made swing mode, which includes swinging in the air, full power racket grips, accompanied by anonymized demographic stroke, and stable hitting. Our results demonstrate that the data of the players. We detail the data collection dataset is suitable for training and evaluating machine learning and annotation procedures.
Associated Learning: Decomposing End-to-end Backpropagation based on Auto-encoders and Target Propagation
Kao, Yu-Wei, Chen, Hung-Hsuan
Backpropagation has been widely used in deep learning approaches, but it is inefficient and sometimes unstable because of backward locking and vanishing/exploding gradient problems, especially when the gradient flow is long. Additionally, updating all edge weights based on a single objective seems biologically implausible. In this paper, we introduce a novel biologically motivated learning structure called Associated Learning, which modularizes the network into smaller components, each of which has a local objective. Because the objectives are mutually independent, Associated Learning can learn the parameters independently and simultaneously when these parameters belong to different components. Surprisingly, training deep models by Associated Learning yields comparable accuracies to models trained using typical backpropagation methods, which aims at fitting the target variable directly. Moreover, probably because the gradient flow of each component is short, deep networks can still be trained with Associated Learning even when some of the activation functions are sigmoid-a situation that usually results in the vanishing gradient problem when using typical backpropagation. We also found that the Associated Learning generates better metafeatures, which we demonstrated both quantitatively (via inter-class and intra-class distance comparisons in the hidden layers) and qualitatively (by visualizing the hidden layers using t-SNE).
CiteSeerX: AI in a Digital Library Search Engine
Wu, Jian (Pennsylvania State University) | Williams, Kyle Mark (Pennsylvania State University) | Chen, Hung-Hsuan (Industrial Technology Research Institute) | Khabsa, Madian (Pennsylvania State University) | Caragea, Cornelia (University of North Texas) | Tuarob, Suppawong (Pennsylvania State University) | Ororbia, Alexander G. (Pennsylvania State University) | Jordan, Douglas (Pennsylvania State University) | Mitra, Prasenjit (Pennsylvania State University) | Giles, C. Lee (Pennsylvania State University)
CiteSeerX is a digital library search engine providing access to more than five million scholarly documents with nearly a million users and millions of hits per day. These AI technologies have been developed by CiteSeerX group members over the past 5–6 years. We also present AI technologies implemented in table and algorithm search, which are special search modes in CiteSeerX. While it is challenging to rebuild a system like CiteSeerX from scratch, many of these AI technologies are transferable to other digital libraries and/or search engines.
CiteSeerX: AI in a Digital Library Search Engine
Wu, Jian (Pennsylvania State University) | Williams, Kyle Mark (Pennsylvania State University) | Chen, Hung-Hsuan (Industrial Technology Research Institute) | Khabsa, Madian (Pennsylvania State University) | Caragea, Cornelia (University of North Texas) | Tuarob, Suppawong (Pennsylvania State University) | Ororbia, Alexander G. (Pennsylvania State University) | Jordan, Douglas (Pennsylvania State University) | Mitra, Prasenjit (Pennsylvania State University) | Giles, C. Lee (Pennsylvania State University)
Since then, the project has been directed by C. Lee Giles. While it is challenging to rebuild a system like Cite-SeerX from scratch, many of these AI technologies are transferable to other digital libraries and search engines. This is different from arXiv, Harvard ADS, and machine cluster to a private cloud using virtualization PubMed, where papers are submitted by authors or techniques (Wu et al. 2014). CiteSeerX extensively pushed by publishers. Unlike Google Scholar and leverages open source software, which significantly Microsoft Academic Search, where a significant portion reduces development effort. Red Hat of documents have only metadata (such as titles, Enterprise Linux (RHEL) 5 and 6 are the operating authors, and abstracts) available, users have full-text systems for all servers. Tomcat 7 is CiteSeerX keeps its own repository, which used for web service deployment on web and indexing serves cached versions of papers even if their previous servers. MySQL is used as the database management links are not alive any more. In additional to system to store metadata. Apache Solr is used paper downloads, CiteSeerX provides automatically for the index, and the Spring framework is used in extracted metadata and citation context, which the web application. In this section, we highlight four AI solutions that are Document metadata download service is not available leveraged by CiteSeerX and that tackle different challenges from Google Scholar and only recently available in metadata extraction and ingestion modules from Microsoft Academic Search. Finally, CiteSeerX (tagged by C, E, D, and A in figure 1).
CiteSeerX: AI in a Digital Library Search Engine
Wu, Jian (The Pennsylvania State University) | Williams, Kyle (The Pennsylvania State University) | Chen, Hung-Hsuan (The Pennsylvania State University) | Khabsa, Madian (The Pennsylvania State University) | Caragea, Cornelia (University of North Texas) | Ororbia, Alexander (The Pennsylvania State University) | Jordan, Douglas (The Pennsylvania State University) | Giles, C. Lee (The Pennsylvania State University)
CiteSeerX is a digital library search engine that provides access to more than 4 million academic documents with nearly a million users and millions of hits per day. Artificial intelligence (AI) technologies are used in many components of CiteSeerX, e.g. to accurately extract metadata, intelligently crawl the web, and ingest documents. We present key AI technologies used in the following components: document classification and deduplication, document and citation clustering, automatic metadata extraction and indexing, and author disambiguation. These AI technologies have been developed by CiteSeerX group members over the past 5–6 years. We also show the usage status, payoff, development challenges, main design concepts, and deployment and maintenance requirements. While it is challenging to rebuild a system like CiteSeerX from scratch, many of these AI technologies are transferable to other digital libraries and/or search engines.