weight
Train-by-Reconnect: Decoupling Locations of Weights from Their Values
What makes untrained deep neural networks (DNNs) different from the trained performant ones? By zooming into the weights in well-trained DNNs, we found that it is the location of weights that holds most of the information encoded by the training. Motivated by this observation, we hypothesized that weights in DNNs trained using stochastic gradient-based methods can be separated into two dimensions: the location of weights, and their exact values. To assess our hypothesis, we propose a novel method called lookahead permutation (LaPerm) to train DNNs by reconnecting the weights. We empirically demonstrate LaPerm's versatility while producing extensive evidence to support our hypothesis: when the initial weights are random and dense, our method demonstrates speed and performance similar to or better than that of regular optimizers, e.g., Adam. When the initial weights are random and sparse (many zeros), our method changes the way neurons connect, achieving accuracy comparable to that of a well-trained dense network. When the initial weights share a single value, our method finds a weight agnostic neural network with far-better-than-chance accuracy.
corrected_LSF
CART loss given by Equation 5. The reason for the name "theoretical cuts" is that by the strong law of large numbers CART loss We first remark that in [32] their proof of Lemma 1 in fact proves more than its statement: Lemma 1. P ( Y y | X) is bounded, we may omit the truncation operators appearing in the original statements.) B.2.1 The approximation error goes to 0 (uniformly in y) B.2.2 The estimation error goes to 0 (uniformly in y) Theorem 4, we have that for all "> 0, " The classical conformalized prediction algorithm transforms a point prediction algorithm into an algorithm that outputs prediction intervals. Even with this adjustment, it took considerably longer than the other methods. In Section 6.1, we use We have re-run the tabular experiments from Section 6.1 five times to get confidence intervals.
The best powered bookshelf speakers for 2025, tested and reviewed
Spin records, stream playlists, or score your next session--these all-in-one systems keep up wherever you set up, from your desktop to a bookshelf. We may earn revenue from the products available on this page and participate in affiliate programs. Powered bookshelf speakers are having a moment. Once considered a middle ground between bulky component systems and wimpy desktop speakers, today's powered models pack serious performance into compact cabinets. They don't need a receiver, they cut cable clutter, and many double as all-in-one multimedia hubs for music, movies, and gaming. Whether you want a sleek hi-fi upgrade, a small-space solution, or something that straddles studio projects and casual listening, we've tested the best powered bookshelf speakers released in 2025--like our best overall, the Fluance Ri71 --to help you find the best fit. We chose the best powered bookshelf speakers by combining hands-on testing with research from trusted peers.
- Leisure & Entertainment (1.00)
- Media > Music (0.46)
Real-Time Cell Sorting with Scalable In Situ FPGA-Accelerated Deep Learning
Islam, Khayrul, Forelli, Ryan F., Han, Jianzhong, Bhadane, Deven, Huang, Jian, Agar, Joshua C., Tran, Nhan, Ogrenci, Seda, Liu, Yaling
Precise cell classification is essential in biomedical diagnostics and therapeutic monitoring, particularly for identifying diverse cell types involved in various diseases. Traditional cell classification methods such as flow cytometry depend on molecular labeling which is often costly, time-intensive, and can alter cell integrity. To overcome these limitations, we present a label-free machine learning framework for cell classification, designed for real-time sorting applications using bright-field microscopy images. This approach leverages a teacher-student model architecture enhanced by knowledge distillation, achieving high efficiency and scalability across different cell types. Demonstrated through a use case of classifying lymphocyte subsets, our framework accurately classifies T4, T8, and B cell types with a dataset of 80,000 preprocessed images, accessible via an open-source Python package for easy adaptation. Our teacher model attained 98\% accuracy in differentiating T4 cells from B cells and 93\% accuracy in zero-shot classification between T8 and B cells. Remarkably, our student model operates with only 0.02\% of the teacher model's parameters, enabling field-programmable gate array (FPGA) deployment. Our FPGA-accelerated student model achieves an ultra-low inference latency of just 14.5~$\mu$s and a complete cell detection-to-sorting trigger time of 24.7~$\mu$s, delivering 12x and 40x improvements over the previous state-of-the-art real-time cell analysis algorithm in inference and total latency, respectively, while preserving accuracy comparable to the teacher model. This framework provides a scalable, cost-effective solution for lymphocyte classification, as well as a new SOTA real-time cell sorting implementation for rapid identification of subsets using in situ deep learning on off-the-shelf computing hardware.
- North America > United States > Pennsylvania > Northampton County > Bethlehem (0.04)
- North America > United States > Pennsylvania > Philadelphia County > Philadelphia (0.04)
- North America > United States > New Jersey > Camden County > Camden (0.04)
- (4 more...)
- Health & Medicine > Therapeutic Area > Immunology (0.88)
- Health & Medicine > Therapeutic Area > Hematology (0.55)
A Flexible Defense Against the Winner's Curse
Zrnic, Tijana, Fithian, William
Across science and policy, decision-makers often need to draw conclusions about the best candidate among competing alternatives. For instance, researchers may seek to infer the effectiveness of the most successful treatment or determine which demographic group benefits most from a specific treatment. Similarly, in machine learning, practitioners are often interested in the population performance of the model that performs best empirically. However, cherry-picking the best candidate leads to the winner's curse: the observed performance for the winner is biased upwards, rendering conclusions based on standard measures of uncertainty invalid. We introduce the zoom correction, a novel approach for valid inference on the winner. Our method is flexible: it can be employed in both parametric and nonparametric settings, can handle arbitrary dependencies between candidates, and automatically adapts to the level of selection bias. The method easily extends to important related problems, such as inference on the top k winners, inference on the value and identity of the population winner, and inference on "near-winners."
- North America > United States > Tennessee > Davidson County > Nashville (0.04)
- North America > United States > New York (0.04)
- North America > United States > California > Alameda County > Berkeley (0.04)
- Asia > Middle East > Jordan (0.04)
Survival regression with accelerated failure time model in XGBoost
Barnwal, Avinash, Cho, Hyunsu, Hocking, Toby Dylan
Survival regression is used to estimate the relation between time-to-event and feature variables, and is important in application domains such as medicine, marketing, risk management and sales management. Nonlinear tree based machine learning algorithms as implemented in libraries such as XGBoost, scikit-learn, LightGBM, and CatBoost are often more accurate in practice than linear models. However, existing implementations of tree-based models have offered limited support for survival regression. In this work, we propose and implement loss functions for learning accelerated failure time (AFT) models in XGBoost, to increase the support for survival modeling for different kinds of label censoring. The AFT model assumes effects that directly accelerate or decelerate the survival time for different kinds of censored data sets. We demonstrate with real and simulated experiments the effectiveness of AFT in XGBoost with respect to a number of baselines, in two respects: generalization performance and training speed. Furthermore, we take advantage of the support for NVIDIA GPUs in XGBoost to achieve substantial speedup over multi-coreCPUs. To our knowledge, our work is the first implementation of AFT that utilizes the processing power of NVIDIA GPUs.
- North America > United States > California > San Francisco County > San Francisco (0.14)
- North America > United States > New York > Suffolk County > Stony Brook (0.04)
- North America > United States > District of Columbia > Washington (0.04)
- (4 more...)
- Information Technology (1.00)
- Health & Medicine > Therapeutic Area > Oncology (0.46)
Machine Learning for Unbalanced Datasets using Neural Networks
There are a few ways to address unbalanced datasets: from built-in class_weight in a logistic regression and sklearn estimators to manual oversampling, and SMOTE. We will look at whether neural networks can serve as a reliable out-of-the-box solution and what parameters can be tweaked to achieve a better performance. Code is available on GitHub. We'll use the Framingham Heart Study data set from Kaggle for this exercise. It presents a binary classification problem in which we need to predict a value of the variable "TenYearCHD" (zero or one) that shows whether a patient will develop a heart disease.
Tuning xgboost in R: Part II
In this previous post I discussed some of the parameters we have to tune to estimate a boosting model using the xgboost package. In this post I will discuss the two parameters that were left out in part I, which are the gamma and the min_child_weight. These two parameters are much less obvious to understand but they can significantly change the results. Unfortunately, the best way to set them changes from dataset to dataset and we have to test a few values to select the best model. Note that there are many other parameters in the xgboost package.