idc
IDC warns of major PC market downturn due to memory crunch
The AI industry's RAM needs could lead to higher PC prices and fewer sales in 2026. The demand building out AI infrastructure has placed on PC component makers has already led to the death of one consumer-facing RAM brand, but a new report from the International Data Corporation (IDC) suggests it could have an even worse impact on the PC industry at large. In its worst-case-scenario model, the IDC predicts PC shipments could shrink by up to 8.9 percent in 2026 because of the high cost of memory. Instead of expanding conventional DRAM and NAND used in smartphones, PCs and other consumer electronics, major memory makers have shifted production toward memory used in AI data centers, such as high-bandwidth (HBM) and high-capacity DDR5, IDC writes. That's continued to drive up the price of the RAM that is available for PC makers, which has naturally led to them to raise the price of their own products to stay above water.
- Information Technology > Communications > Mobile (0.79)
- Information Technology > Artificial Intelligence (0.73)
- Information Technology > Cloud Computing (0.57)
EH-MAM: Easy-to-Hard Masked Acoustic Modeling for Self-Supervised Speech Representation Learning
Seth, Ashish, Selvakumar, Ramaneswaran, Sakshi, S, Kumar, Sonal, Ghosh, Sreyan, Manocha, Dinesh
In this paper, we present EH-MAM (Easy-to-Hard adaptive Masked Acoustic Modeling), a novel self-supervised learning approach for speech representation learning. In contrast to the prior methods that use random masking schemes for Masked Acoustic Modeling (MAM), we introduce a novel selective and adaptive masking strategy. Specifically, during SSL training, we progressively introduce harder regions to the model for reconstruction. Our approach automatically selects hard regions and is built on the observation that the reconstruction loss of individual frames in MAM can provide natural signals to judge the difficulty of solving the MAM pre-text task for that frame. To identify these hard regions, we employ a teacher model that first predicts the frame-wise losses and then decides which frames to mask. By learning to create challenging problems, such as identifying harder frames and solving them simultaneously, the model is able to learn more effective representations and thereby acquire a more comprehensive understanding of the speech. Quantitatively, EH-MAM outperforms several state-of-the-art baselines across various low-resource speech recognition and SUPERB benchmarks by 5%-10%. Additionally, we conduct a thorough analysis to show that the regions masked by EH-MAM effectively capture useful context across speech frames.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- South America > Chile > Santiago Metropolitan Region > Santiago Province > Santiago (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- (2 more...)
- Information Technology > Artificial Intelligence > Speech > Acoustic Processing (0.90)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (0.68)
- Information Technology > Artificial Intelligence > Speech > Speech Recognition (0.67)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.46)
Nonlinear Gaussian process tomography with imposed non-negativity constraints on physical quantities for plasma diagnostics
We propose a novel tomographic method, nonlinear Gaussian process tomography (nonlinear GPT) that employs the Laplace approximation to ensure the non-negative physical quantity, such as the emissivity of plasma optical diagnostics. This new method implements a logarithmic Gaussian process (log-GP) to model plasma distribution more naturally, thereby expanding the limitations of standard GPT, which are restricted to linear problems and may yield non-physical negative values. The effectiveness of the proposed log-GP tomography is demonstrated through a case study using the Ring Trap 1 (RT-1) device, where log-GPT outperforms existing methods, standard GPT, and the Minimum Fisher Information (MFI) methods in terms of reconstruction accuracy. The result highlights the effectiveness of nonlinear GPT for imposing physical constraints in applications to an inverse problem.
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.04)
- North America > United States > Massachusetts (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Information Technology > Modeling & Simulation (0.95)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Uncertainty > Bayesian Inference (0.93)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models > Directed Networks > Bayesian Learning (0.46)
Supervised Contrastive Vision Transformer for Breast Histopathological Image Classification
Shiri, Mohammad, Reddy, Monalika Padma, Sun, Jiangwen
Invasive ductal carcinoma (IDC) is the most prevalent form of breast cancer. Breast tissue histopathological examination is critical in diagnosing and classifying breast cancer. Although existing methods have shown promising results, there is still room for improvement in the classification accuracy and generalization of IDC using histopathology images. We present a novel approach, Supervised Contrastive Vision Transformer (SupCon-ViT), for improving the classification of invasive ductal carcinoma in terms of accuracy and generalization by leveraging the inherent strengths and advantages of both transfer learning, i.e., pre-trained vision transformer, and supervised contrastive learning. Our results on a benchmark breast cancer dataset demonstrate that SupCon-Vit achieves state-of-the-art performance in IDC classification, with an F1-score of 0.8188, precision of 0.7692, and specificity of 0.8971, outperforming existing methods. In addition, the proposed model demonstrates resilience in scenarios with minimal labeled data, making it highly efficient in real-world clinical settings where labelled data is limited. Our findings suggest that supervised contrastive learning in conjunction with pre-trained vision transformers appears to be a viable strategy for an accurate classification of IDC, thus paving the way for a more efficient and reliable diagnosis of breast cancer through histopathological image analysis.
- North America > United States > Virginia > Norfolk City County > Norfolk (0.05)
- North America > United States > Pennsylvania (0.04)
- North America > United States > New Jersey (0.04)
- Europe > Spain > Andalusia > Granada Province > Granada (0.04)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.70)
- Information Technology > Artificial Intelligence > Machine Learning > Performance Analysis > Accuracy (0.66)
- Information Technology > Artificial Intelligence > Vision > Image Understanding (0.64)
Text2Cohort: Facilitating Intuitive Access to Biomedical Data with Natural Language Cohort Discovery
Kulkarni, Pranav, Kanhere, Adway, Yi, Paul H., Parekh, Vishwa S.
The Imaging Data Commons (IDC) is a cloud-based database that provides researchers with open access to cancer imaging data, with the goal of facilitating collaboration. However, cohort discovery within the IDC database has a significant technical learning curve. Recently, large language models (LLM) have demonstrated exceptional utility for natural language processing tasks. We developed Text2Cohort, a LLM-powered toolkit to facilitate user-friendly natural language cohort discovery in the IDC. Our method translates user input into IDC queries using grounding techniques and returns the query's response. We evaluate Text2Cohort on 50 natural language inputs, from information extraction to cohort discovery. Our toolkit successfully generated responses with an 88% accuracy and 0.94 F1 score. We demonstrate that Text2Cohort can enable researchers to discover and curate cohorts on IDC with high levels of accuracy using natural language in a more intuitive and user-friendly way.
The NCI Imaging Data Commons as a platform for reproducible research in computational pathology
Schacherer, Daniela P., Herrmann, Markus D., Clunie, David A., Höfener, Henning, Clifford, William, Longabaugh, William J. R., Pieper, Steve, Kikinis, Ron, Fedorov, Andrey, Homeyer, André
Background and Objectives: Reproducibility is a major challenge in developing machine learning (ML)-based solutions in computational pathology (CompPath). The NCI Imaging Data Commons (IDC) provides >120 cancer image collections according to the FAIR principles and is designed to be used with cloud ML services. Here, we explore its potential to facilitate reproducibility in CompPath research. Methods: Using the IDC, we implemented two experiments in which a representative ML-based method for classifying lung tumor tissue was trained and/or evaluated on different datasets. To assess reproducibility, the experiments were run multiple times with separate but identically configured instances of common ML services. Results: The AUC values of different runs of the same experiment were generally consistent. However, we observed small variations in AUC values of up to 0.045, indicating a practical limit to reproducibility. Conclusions: We conclude that the IDC facilitates approaching the reproducibility limit of CompPath research (i) by enabling researchers to reuse exactly the same datasets and (ii) by integrating with cloud ML services so that experiments can be run in identically configured computing environments.
- North America > United States > Massachusetts > Suffolk County > Boston (0.04)
- North America > United States > Washington > King County > Seattle (0.04)
- North America > United States > Pennsylvania (0.04)
- (5 more...)
- Health & Medicine > Therapeutic Area > Oncology (1.00)
- Health & Medicine > Diagnostic Medicine (1.00)
- Information Technology > Sensing and Signal Processing (1.00)
- Information Technology > Cloud Computing (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.68)
- Information Technology > Artificial Intelligence > Machine Learning > Performance Analysis > Accuracy (0.46)
A Comprehensive Study on Dataset Distillation: Performance, Privacy, Robustness and Fairness
Chen, Zongxiong, Geng, Jiahui, Zhu, Derui, Woisetschlaeger, Herbert, Li, Qing, Schimmler, Sonja, Mayer, Ruben, Rong, Chunming
The aim of dataset distillation is to encode the rich features of an original dataset into a tiny dataset. It is a promising approach to accelerate neural network training and related studies. Different approaches have been proposed to improve the informativeness and generalization performance of distilled images. However, no work has comprehensively analyzed this technique from a security perspective and there is a lack of systematic understanding of potential risks. In this work, we conduct extensive experiments to evaluate current state-of-the-art dataset distillation methods. We successfully use membership inference attacks to show that privacy risks still remain. Our work also demonstrates that dataset distillation can cause varying degrees of impact on model robustness and amplify model unfairness across classes when making predictions. This work offers a large-scale benchmarking framework for dataset distillation evaluation.
- North America > United States > Virginia (0.04)
- Europe > Norway > Western Norway > Rogaland > Stavanger (0.04)
- Europe > Germany > North Rhine-Westphalia > Upper Bavaria > Munich (0.04)
- Research Report > New Finding (0.46)
- Research Report > Promising Solution (0.34)
Iterative Double Clustering for Unsupervised and Semi-Supervised Learning
We present a powerful meta-clustering technique called Iterative Dou- ble Clustering (IDC). The IDC method is a natural extension of the recent Double Clustering (DC) method of Slonim and Tishby that ex- hibited impressive performance on text categorization tasks [12]. Us- ing synthetically generated data we empirically flnd that whenever the DC procedure is successful in recovering some of the structure hidden in the data, the extended IDC procedure can incrementally compute a signiflcantly more accurate classiflcation. IDC is especially advan- tageous when the data exhibits high attribute noise. Our simulation results also show the efiectiveness of IDC in text categorization prob- lems.
Software Eats The World, And AI Eats Software
We know, as you do, that artificial intelligence is driving a lot of spending at IT organizations and is probably the fundamental driver of spending by the hyperscalers and cloud builders that have, thus far, benefitted most from the machine learning revolution. But just how much money are companies sinking into AI, and how will that grow over time? We have not seen a lot of good data on this, and the market researchers at IDC, as usual, have been the most vocal about how they dice and slice the AI market in their public statements, which dribble out some insight here and there to keep their name out there and to drive deeper engagements with AI startups, their investors, IT suppliers who are chasing this market, and large enterprises that are on the forefront of commercializing AI in their application stacks. What sent us down this AI spending rathole was some numbers that IDC released on March 7, which talks about worldwide spending on AI-Centric systems, including hardware, software, and services. AI-centric means that without the AI component, an application will not function.
Worldwide Spending on AI-Centric Systems Forecast to Reach $154 Billion in 2023, According to IDC
NEEDHAM, Mass., March 7, 2023 – A new forecast from the International Data Corporation (IDC) Worldwide Artificial Intelligence Spending Guide shows that global spending on artificial intelligence (AI), including software, hardware, and services for AI-centric systems*, will reach $154 billion in 2023, an increase of 26.9% over the amount spent in 2022. The ongoing incorporation of AI into a wide range of products will result in a compound annual growth rate (CAGR) of 27.0% over the 2022-2026 forecast with spending on AI-centric systems expected to surpass $300 billion in 2026. "Companies that are slow to adopt AI will be left behind – large and small. AI is best used in these companies to augment human abilities, automate repetitive tasks, provide personalized recommendations, and make data-driven decisions with speed and accuracy," said Mike Glennon, senior market research analyst with IDC's Customer Insights & Analysis team. "Suppliers of AI technologies need to know which are the largest and fastest growing opportunities, but without data they become just another opinion. IDC's AI Spending Guide provides the foundation for marketing strategy through its comprehensive coverage of AI opportunities and gives a robust basis for a market focus that ties with companies' capabilities."
- North America > United States > Massachusetts > Norfolk County > Needham (0.25)
- Asia > China (0.06)
- Europe > Western Europe (0.05)