model number
Infinite forecast combinations based on Dirichlet process
Ren, Yinuo, Li, Feng, Kang, Yanfei, Wang, Jue
Forecast combination integrates information from various sources by consolidating multiple forecast results from the target time series. Instead of the need to select a single optimal forecasting model, this paper introduces a deep learning ensemble forecasting model based on the Dirichlet process. Initially, the learning rate is sampled with three basis distributions as hyperparameters to convert the infinite mixture into a finite one. All checkpoints are collected to establish a deep learning sub-model pool, and weight adjustment and diversity strategies are developed during the combination process. The main advantage of this method is its ability to generate the required base learners through a single training process, utilizing the decaying strategy to tackle the challenge posed by the stochastic nature of gradient descent in determining the optimal learning rate. To ensure the method's generalizability and competitiveness, this paper conducts an empirical analysis using the weekly dataset from the M4 competition and explores sensitivity to the number of models to be combined. The results demonstrate that the ensemble model proposed offers substantial improvements in prediction accuracy and stability compared to a single benchmark model.
- Research Report > New Finding (0.48)
- Research Report > Experimental Study (0.47)
Analysis of frequent trading effects of various machine learning models
In recent years, high-frequency trading has emerged as a crucial strategy in stock trading. This study aims to develop an advanced high-frequency trading algorithm and compare the performance of three different mathematical models: the combination of the cross-entropy loss function and the quasi-Newton algorithm, the FCNN model, and the vector machine. The proposed algorithm employs neural network predictions to generate trading signals and execute buy and sell operations based on specific conditions. By harnessing the power of neural networks, the algorithm enhances the accuracy and reliability of the trading strategy. To assess the effectiveness of the algorithm, the study evaluates the performance of the three mathematical models. The combination of the cross-entropy loss function and the quasi-Newton algorithm is a widely utilized logistic regression approach. The FCNN model, on the other hand, is a deep learning algorithm that can extract and classify features from stock data. Meanwhile, the vector machine is a supervised learning algorithm recognized for achieving improved classification results by mapping data into high-dimensional spaces. By comparing the performance of these three models, the study aims to determine the most effective approach for high-frequency trading. This research makes a valuable contribution by introducing a novel methodology for high-frequency trading, thereby providing investors with a more accurate and reliable stock trading strategy.
- Asia > China > Hubei Province (0.04)
- North America > Trinidad and Tobago > Trinidad > Arima > Arima (0.04)
- Asia > China > Shanghai > Shanghai (0.04)
- Asia > China > Guangdong Province > Shenzhen (0.04)
Revisiting Design Choices in Model-Based Offline Reinforcement Learning
Lu, Cong, Ball, Philip J., Parker-Holder, Jack, Osborne, Michael A., Roberts, Stephen J.
Offline reinforcement learning enables agents to leverage large pre-collected datasets of environment transitions to learn control policies, circumventing the need for potentially expensive or unsafe online data collection. Significant progress has been made recently in offline model-based reinforcement learning, approaches which leverage a learned dynamics model. This typically involves constructing a probabilistic model, and using the model uncertainty to penalize rewards where there is insufficient data, solving for a pessimistic MDP that lower bounds the true MDP. Existing methods, however, exhibit a breakdown between theory and practice, whereby pessimistic return ought to be bounded by the total variation distance of the model from the true dynamics, but is instead implemented through a penalty based on estimated model uncertainty. This has spawned a variety of uncertainty heuristics, with little to no comparison between differing approaches. In this paper, we compare these heuristics, and design novel protocols to investigate their interaction with other hyperparameters, such as the number of models, or imaginary rollout horizon. Using these insights, we show that selecting these key hyperparameters using Bayesian Optimization produces superior configurations that are vastly different to those currently used in existing hand-tuned state-of-the-art methods, and result in drastically stronger performance.
- North America > United States (0.04)
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.04)
- Europe > France > Hauts-de-France > Nord > Lille (0.04)
- Research Report > New Finding (0.46)
- Research Report > Promising Solution (0.34)
Interpretable Methods for Identifying Product Variants
West, Rebecca, Jadda, Khalifeh Al, Ahsan, Unaiza, Qu, Huiming, Cui, Xiquan
For e-commerce companies with large product selections, the organization and grouping of products in meaningful ways is important for creating great customer shopping experiences and cultivating an authoritative brand image. One important way of grouping products is to identify a family of product variants, where the variants are mostly the same with slight and yet distinct differences (e.g. color or pack size). In this paper, we introduce a novel approach to identifying product variants. It combines both constrained clustering and tailored NLP techniques (e.g. extraction of product family name from unstructured product title and identification of products with similar model numbers) to achieve superior performance compared with an existing baseline using a vanilla classification approach. In addition, we design the algorithm to meet certain business criteria, including meeting high accuracy requirements on a wide range of categories (e.g. appliances, decor, tools, and building materials, etc.) as well as prioritizing the interpretability of the model to make it accessible and understandable to all business partners.
- Asia > Taiwan > Taiwan Province > Taipei (0.06)
- North America > United States > Georgia > Fulton County > Atlanta (0.05)
- North America > United States > New York > New York County > New York City (0.04)
- Research Report (0.84)
- Overview (0.66)
- Materials > Construction Materials (0.54)
- Information Technology > Services > e-Commerce Services (0.35)
- Information Technology > Information Management (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Natural Language (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning > Clustering (1.00)
Accelerating Multi-Model Inference by Merging DNNs of Different Weights
Jeong, Joo Seong, Kim, Soojeong, Yu, Gyeong-In, Lee, Yunseong, Chun, Byung-Gon
Standardized DNN models that have been proved to perform well on machine learning tasks are widely used and often adopted as-is to solve downstream tasks, forming the transfer learning paradigm. However, when serving multiple instances of such DNN models from a cluster of GPU servers, existing techniques to improve GPU utilization such as batching are inapplicable because models often do not share weights due to fine-tuning. We propose NetFuse, a technique of merging multiple DNN models that share the same architecture but have different weights and different inputs. NetFuse is made possible by replacing operations with more general counterparts that allow a set of weights to be associated with only a certain set of inputs. Experiments on ResNet-50, ResNeXt-50, BERT, and XLNet show that NetFuse can speed up DNN inference time up to 3.6x on a NVIDIA V100 GPU, and up to 3.0x on a TITAN Xp GPU when merging 32 model instances, while only using up a small additional amount of GPU memory.
Approximate Model Counting by Partial Knowledge Compilation
Model counting is the problem of computing the number of satisfying assignments of a given propositional formula. Although exact model counters can be naturally furnished by most of the knowledge compilation (KC) methods, in practice, they fail to generate the compiled results for the exact counting of models for certain formulas due to the explosion in sizes. Decision-DNNF is an important KC language that captures most of the practical compilers. We propose a generalized Decision-DNNF (referred to as partial Decision-DNNF) via introducing a class of new leaf vertices (called unknown vertices), and then propose an algorithm called PartialKC to generate randomly partial Decision-DNNF formulas from the given formulas. An unbiased estimate of the model number can be computed via a randomly partial Decision-DNNF formula. Each calling of PartialKC consists of multiple callings of MicroKC, while each of the latter callings is a process of importance sampling equipped with KC technologies. The experimental results show that PartialKC is more accurate than both SampleSearch and SearchTreeSampler, PartialKC scales better than SearchTreeSampler, and the KC technologies can obviously accelerate sampling.
Will Apple's new iPhone SE have a notch?
Apple's hugely popular iPhone SE is overdue and overhaul - and it could see the end of the headphone jack. The latest renders claiming to show the next generation phone reveal the jack has gone - but that a new'notch' has appeared. The notch, first seen in the iPhone X, would give the phone FaceID capabilities. The latest renders claiming to show the next generation phone reveal the headphone jack and the home button have gone - but that a new'notch' has appeared However, the latest leaks also reveal the home button and headphone jack are gone, bringing the iPhone SE into line with the rest of Apple's line. The images were posted by @onleaks, although even he admitted they could be fake, tweeting'Now that u aware I can't confirm if this one is partially or completely accurate or even exists but despite of that decided to share it for discussion purposes only.'
- Information Technology > Communications > Mobile (1.00)
- Information Technology > Artificial Intelligence > Vision > Face Recognition (0.54)
Singlecue Gen 2 review: This gesture-recognition device nearly provoked us into making some rude gestures
I have an Amazon Echo and a Logitech Harmony Remote in my bedroom, but my goal is to eliminate as many remotes as possible so I can control the TV, cable box, Amazon Fire TV, and other gadgets in my house as quickly and efficiently as possible. I want to be able to do things like dim the lights, turn on the TV, and tune to my favorite program in a single step, without needing to reach for a switch or fumble with multiple remotes. It's against that backdrop--and a desire to simplify my life--that I anxiously broke out Singlecue gesture-control device from the box and plugged it in. I was hoping that Singlecue's promise to enable me to control my home's smart devices with a wave of my hand would further my mission to eliminate remotes altogether. At first blush, Singlecue is a compelling device.
Constructing Reference Sets from Unstructured, Ungrammatical Text
Michelson, M., Knoblock, C. A.
Vast amounts of text on the Web are unstructured and ungrammatical, such as classified ads, auction listings, forum postings, etc. We call such text posts. Despite their inconsistent structure and lack of grammar, posts are full of useful information. This paper presents work on semi-automatically building tables of relational information, called reference sets, by analyzing such posts directly. Reference sets can be applied to a number of tasks such as ontology maintenance and information extraction. Our reference-set construction method starts with just a small amount of background knowledge, and constructs tuples representing the entities in the posts to form a reference set. We also describe an extension to this approach for the special case where even this small amount of background knowledge is impossible to discover and use. To evaluate the utility of the machine-constructed reference sets, we compare them to manually constructed reference sets in the context of reference-set-based information extraction. Our results show the reference sets constructed by our method outperform manually constructed reference sets. We also compare the reference-set-based extraction approach using the machine-constructed reference set to supervised extraction approaches using generic features. These results demonstrate that using machine-constructed reference sets outperforms the supervised methods, even though the supervised methods require training data.
- North America > United States > California > San Francisco County > San Francisco (0.04)
- North America > United States > California > San Diego County > San Diego (0.04)
- North America > United States > California > Los Angeles County > El Segundo (0.04)
- Automobiles & Trucks > Manufacturer (1.00)
- Transportation > Passenger (0.93)
- Transportation > Ground > Road (0.93)
- Government > Regional Government > North America Government > United States Government (0.67)
Exploiting Background Knowledge to Build Reference Sets for Information Extraction
Michelson, Matthew (Fetch Technologies) | Knoblock, Craig A. (University of Southern California / Information Sciences Institute)
Previous work on information extraction from unstructured, ungrammatical text (e.g. classified ads) showed that exploiting a set of background knowledge, called a "reference set," greatly improves the precision and recall of the extractions. However, finding a source for this reference set is often difficult, if not impossible. Further, even if a source is found, it might not overlap well with the text for extraction. In this paper we present an approach to building the reference set directly from the text itself. Our approach eliminates the need to find the source for the reference set, and ensures better overlap between the text and reference set. Starting with a small amount of background knowledge, our technique constructs tuples representing the entities in the text to form a reference set. Our results show that our method outperforms manually constructed reference sets, since hand built reference sets may not overlap with the entities in the unstructured, ungrammatical text. We also ran experiments comparing our method to the supervised approach of Conditional Random Fields (CRFs) using simple, generic features. These results show our method achieves an improvement in F1-measure for 6/9 attributes and is competitive in performance on the others, and this is without training data.
- Automobiles & Trucks > Manufacturer (1.00)
- Transportation > Passenger (0.94)
- Transportation > Ground > Road (0.94)
- Government (0.68)