dsi
- North America > United States (0.14)
- North America > Dominican Republic (0.04)
DivideandContrast: Source-freeDomainAdaptation viaAdaptiveContrastiveLearning (SupplementaryMaterial)
Consideringa C-wayclassification task, our model consists of source classifier and feature extractor h = gs ϕ, which maps input spaceRI topredictionvector spaceRC,andh(x) = argmaxc h(x)[c]. Following in[25,26,27,28],wedenoteDTc astheconditional distribution (probability measure) ofDT given the ground truthy = c, and also assume that the supports ofDTi andDTj aredisjointforalli = j. Following [25, 27, 26], we study target domain relies on theexpansion property, which implies the continuity of data distributions in each class-wise subpopulations. Thus, x DS,x B(x) DS, the network predictions are consistent, i.e.RDS(h)=0. Theorem A.2. Suppose the condition of Claim 3.1 holds andDT,DS satisfies (q,γ)-constant expansion.
- Asia > China > Guangdong Province > Shenzhen (0.05)
- Asia > China > Zhejiang Province > Hangzhou (0.04)
- Asia > China > Jiangsu Province > Nanjing (0.04)
- Asia > China > Guangdong Province > Guangzhou (0.04)
Inducing Diversity in Differentiable Search Indexing
Phatak, Abhijeet, Sachdev, Jayant, Rosario, Sean D, Kirti, Swati, Tripathy, Chittaranjan
Differentiable Search Indexing (DSI) is a recent paradigm for information retrieval which uses a transformer-based neural network architecture as the document index to simplify the retrieval process. A differentiable index has many advantages enabling modifications, updates or extensions to the index. In this work, we explore balancing relevance and novel information content (diversity) for training DSI systems inspired by Maximal Marginal Relevance (MMR), and show the benefits of our approach over the naive DSI training. We present quantitative and qualitative evaluations of relevance and diversity measures obtained using our method on NQ320K and MSMARCO datasets in comparison to naive DSI. With our approach, it is possible to achieve diversity without any significant impact to relevance. Since we induce diversity while training DSI, the trained model has learned to diversify while being relevant. This obviates the need for a post-processing step to induce diversity in the recall set as typically performed using MMR. Our approach will be useful for Information Retrieval problems where both relevance and diversity are important such as in sub-topic retrieval. Our work can also be easily be extended to the incremental DSI settings which would enable fast updates to the index while retrieving a diverse recall set.
- North America > United States > New York > New York County > New York City (0.05)
- North America > United States > California > Santa Clara County > Sunnyvale (0.04)
- North America > United States > California > Sacramento County > Sacramento (0.04)
- North America > United States > California > Los Angeles County > Los Angeles (0.04)
Originality in scientific titles and abstracts can predict citation count
Culbert, Jack H., Kenett, Yoed N., Mayr, Philipp
In this research-in-progress paper, we apply a computational measure correlating with originality from creativity science: Divergent Semantic Integration (DSI), to a selection of 99,557 scientific abstracts and titles selected from the Web of Science. We observe statistically significant differences in DSI between subject and field of research, and a slight rise in DSI over time. We model the base 10 logarithm of the citation count after 5 years with DSI and find a statistically significant positive correlation in all fields of research with an adjusted $R^2$ of 0.13.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- Europe > Germany > North Rhine-Westphalia > Cologne Region > Cologne (0.04)
- Asia > Middle East > Israel > Haifa District > Haifa (0.04)
Distributed Speculative Inference of Large Language Models
Timor, Nadav, Mamou, Jonathan, Korat, Daniel, Berchansky, Moshe, Pereg, Oren, Wasserblat, Moshe, Galanti, Tomer, Gordon, Michal, Harel, David
Accelerating the inference of large language models (LLMs) is an important challenge in artificial intelligence. This paper introduces distributed speculative inference (DSI), a novel distributed inference algorithm that is provably faster than speculative inference (SI) [Leviathan et al., 2023, Chen et al., 2023, Miao et al., 2023] and traditional autoregressive inference (non-SI). Like other SI algorithms, DSI works on frozen LLMs, requiring no training or architectural modifications, and it preserves the target distribution. Prior studies on SI have demonstrated empirical speedups (compared to non-SI) but require a fast and accurate drafter LLM. In practice, off-the-shelf LLMs often do not have matching drafters that are sufficiently fast and accurate. We show a gap: SI gets slower than non-SI when using slower or less accurate drafters. We close this gap by proving that DSI is faster than both SI and non-SI--given any drafters. By orchestrating multiple instances of the target and drafters, DSI is not only faster than SI but also supports LLMs that cannot be accelerated with SI. Our simulations show speedups of off-the-shelf LLMs in realistic settings: DSI is 1.29-1.92x
- North America > United States > Massachusetts (0.04)
- North America > United States > California > Santa Clara County > Palo Alto (0.04)
- North America > Dominican Republic (0.04)
- (2 more...)
Dataset Structural Index: Leveraging a machine's perspective towards visual data
But when it came to visual datasets, the field immediately stepped towards the algorithmic side. One of the fundamental reasons was the amount of information needed to translate from an image. But with the introduction of convolutional networks and transfer learning [1], [2], [3], it is possible to convert an image or a visual object into feature vectors without losing too much information about the entity under concern. It defined a way to use feature maps to compare and distinguish one visual object from another [4]. There has been a lot of work in using these feature vector conversions in systems like content-based image retrievals [5], using feature vectors as representations of different scenarios [6], [7]. It is critical to understand that there is a difference between the way a machine looks at the data and the way we do. There are scenarios in which the interpretation through features is a little different from the interpretation of humans. DSI is there to bridge the gap and understand the machine's perspective before molding it to shape better architectures, in turn, better model performances. I think two concepts could be linked together to understand a machine's viewpoint while working with visual
- Europe > United Kingdom > England > Staffordshire (0.04)
- Oceania > New Zealand > South Island > Marlborough District > Blenheim (0.04)
- North America > United States > Virginia (0.04)
- (3 more...)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Supervised Learning > Representation Of Examples (0.75)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.69)
An Internal Cluster Validity Index Based on Distance-based Separability Measure
To evaluate clustering results is a significant part in cluster analysis. Usually, there is no true class labels for clustering as a typical unsupervised learning. Thus, a number of internal evaluations, which use predicted labels and data, have been created. They also named internal cluster validity indices (CVIs). Without true labels, to design an effective CVI is not simple because it is similar to create a clustering method. And, to have more CVIs is crucial because there is no universal CVI that can be used to measure all datasets, and no specific method for selecting a proper CVI for clusters without true labels. Therefore, to apply more CVIs to evaluate clustering results is necessary. In this paper, we propose a novel CVI - called Distance-based Separability Index (DSI), based on a data separability measure. We applied the DSI and eight other internal CVIs including early studies from Dunn (1974) to most recent studies CVDD (2019) as comparison. We used an external CVI as ground truth for clustering results of five clustering algorithms on 12 real and 97 synthetic datasets. Results show DSI is an effective, unique, and competitive CVI to other compared CVIs. In addition, we summarized the general process to evaluate CVIs and created a new method - rank difference - to compare the results of CVIs.
- North America > United States > District of Columbia > Washington (0.04)
- North America > United States > Wisconsin (0.04)
Data Separability for Neural Network Classifiers and the Development of a Separability Index
Guan, Shuyue, Loew, Murray, Ko, Hanseok
In machine learning, the performance of a classifier depends on both the classifier model and the dataset. For a specific neural network classifier, the training process varies with the training set used; some training data make training accuracy fast converged to high values, while some data may lead to slowly converged to lower accuracy. To quantify this phenomenon, we created the Distance-based Separability Index (DSI), which is independent of the classifier model, to measure the separability of datasets. In this paper, we consider the situation where different classes of data are mixed together in the same distribution is most difficult for classifiers to separate, and we show that the DSI can indicate whether data belonging to different classes have similar distributions. When comparing our proposed approach with several existing separability/complexity measures using synthetic and real datasets, the results show the DSI is an effective separability measure. We also discussed possible applications of the DSI in the fields of data science, machine learning, and deep learning.
- North America > United States > Iowa > Story County > Ames (0.04)
- North America > United States > District of Columbia > Washington (0.04)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
- (3 more...)