Goto

Collaborating Authors

 Country


Gradient Descent Can Take Exponential Time to Escape Saddle Points

Neural Information Processing Systems

Although gradient descent (GD) almost always escapes saddle points asymptotically [Lee et al., 2016], this paper shows that even with fairly natural random initialization schemes and non-pathological functions, GD can be significantly slowed down by saddle points, taking exponential time to escape. On the other hand, gradient descent with perturbations [Ge et al., 2015, Jin et al., 2017] is not slowed down by saddle points--it can find an approximate local minimizer in polynomial time. This result implies that GD is inherently slower than perturbed GD, and justifies the importance of adding perturbations for efficient non-convex optimization. While our focus is theoretical, we also present experiments that illustrate our theoretical findings.


A framework for Multi-A(rmed)/B(andit) Testing with Online FDR Control

Neural Information Processing Systems

We propose an alternative framework to existing setups for controlling false alarms when multiple A/B tests are run over time. This setup arises in many practical applications, e.g. when pharmaceutical companies test new treatment options against control pills for different diseases, or when internet companies test their default webpages versus various alternatives over time. Our framework proposes to replace a sequence of A/B tests by a sequence of best-arm MAB instances, which can be continuously monitored by the data scientist. When interleaving the MAB tests with an online false discovery rate (FDR) algorithm, we can obtain the best of both worlds: low sample complexity and any time online FDR control. Our main contributions are: (i) to propose reasonable definitions of a null hypothesis for MAB instances; (ii) to demonstrate how one can derive an always-valid sequential p-value that allows continuous monitoring of each MAB test; and (iii) to show that using rejection thresholds of online-FDR algorithms as the confidence levels for the MAB algorithms results in both sample-optimality, high power and low FDR at any point in time. We run extensive simulations to verify our claims, and also report results on real data collected from the New Yorker Cartoon Caption contest.


Dynamic Revenue Sharing

Neural Information Processing Systems

Many online platforms act as intermediaries between a seller and a set of buyers. Examples of such settings include online retailers (such as Ebay) selling items on behalf of sellers to buyers, or advertising exchanges (such as AdX) selling pageviews on behalf of publishers to advertisers. In such settings, revenue sharing is a central part of running such a marketplace for the intermediary, and fixed-percentage revenue sharing schemes are often used to split the revenue among the platform and the sellers. In particular, such revenue sharing schemes require the platform to (i) take at most a constant fraction \alpha of the revenue from auctions and (ii) pay the seller at least the seller declared opportunity cost c for each item sold. A straightforward way to satisfy the constraints is to set a reserve price at c / (1 - \alpha) for each item, but it is not the optimal solution on maximizing the profit of the intermediary.


Kernel Feature Selection via Conditional Covariance Minimization

Neural Information Processing Systems

We propose a method for feature selection that employs kernel-based measures of independence to find a subset of covariates that is maximally predictive of the response. Building on past work in kernel dimension reduction, we show how to perform feature selection via a constrained optimization problem involving the trace of the conditional covariance operator. We prove various consistency results for this procedure, and also demonstrate that our method compares favorably with other state-of-the-art algorithms on a variety of synthetic and real data sets.


Watch: Iranians show daily life under air strikes and regime crackdown

BBC News

The BBC has obtained footage and interviews from the Iranian capital Tehran which evoke a city of strained nerves, of constant waiting for the next air strike and relentless fear of the state security apparatus. The identities of the people in this report have been protected. While independent journalists still try to gather testimony that offers a credible alternative view, they run the risk of arrest, torture and possibly worse. Displaced Palestinians were told to secure their tents to prevent them being blown away as a storm swept through the enclave. Video filmed by a witness and verified by the BBC shows a drone crashing close to the airport.


Dyson's New PencilWash Is Here

WIRED

Dyson's Newest Wet Floor Cleaner Is Available as of Today The debut follows the release of Dyson's newest robot vacuum and larger wet cleaner last week. Welcome to a new world of mopping options from Dyson. After announcing several new models last year at IFA Berlin, Dyson has begun rolling out its latest suite of vacuums and wet floor cleaners to the public. Last week, Dyson's newest robot vacuum, the Spot+Scrub Ai ($1,200), became available for purchase online, along with the Clean+Wash Hygiene ($500), one of the brand's new wet floor cleaners. The recently announced Dyson PencilWash ($350) is available as of today.


Tomography of the London Underground: a Scalable Model for Origin-Destination Data

Neural Information Processing Systems

The paper addresses the classical network tomography problem of inferring local traffic given origin-destination observations. Focussing on large complex public transportation systems, we build a scalable model that exploits input-output information to estimate the unobserved link/station loads and the users path preferences. Based on the reconstruction of the users' travel time distribution, the model is flexible enough to capture possible different path-choice strategies and correlations between users travelling on similar paths at similar times. The corresponding likelihood function is intractable for medium or large-scale networks and we propose two distinct strategies, namely the exact maximum-likelihood inference of an approximate but tractable model and the variational inference of the original intractable model. As an application of our approach, we consider the emblematic case of the London Underground network, where a tap-in/tap-out system tracks the start/exit time and location of all journeys in a day. A set of synthetic simulations and real data provided by Transport For London are used to validate and test the model on the predictions of observable and unobservable quantities.


Deep Learning for Precipitation Nowcasting: A Benchmark and A New Model

Neural Information Processing Systems

With the goal of making high-resolution forecasts of regional rainfall, precipitation nowcasting has become an important and fundamental technology underlying various public services ranging from rainstorm warnings to flight safety. Recently, the Convolutional LSTM (ConvLSTM) model has been shown to outperform traditional optical flow based methods for precipitation nowcasting, suggesting that deep learning models have a huge potential for solving the problem. However, the convolutional recurrence structure in ConvLSTM-based models is location-invariant while natural motion and transformation (e.g., rotation) are location-variant in general. Furthermore, since deep-learning-based precipitation nowcasting is a newly emerging area, clear evaluation protocols have not yet been established. To address these problems, we propose both a new model and a benchmark for precipitation nowcasting. Specifically, we go beyond ConvLSTM and propose the Trajectory GRU (TrajGRU) model that can actively learn the location-variant structure for recurrent connections. Besides, we provide a benchmark that includes a real-world large-scale dataset from the Hong Kong Observatory, a new training loss, and a comprehensive evaluation protocol to facilitate future research and gauge the state of the art.


A very serious guide to buying your own humanoid robot butler

New Scientist

You can now buy a humanoid robot housekeeper for less than the price of a second-hand car. But before splashing out, there's something you need to know Science fiction is strewn with humanoid robots, from bad-tempered Bender in to cunning Ava in . And it has long seemed like that's the natural home for such robots - on the screen and in books. The idea of a walking, talking, functioning robot with two arms and two legs has appeared to be a distant dream. Last year, machines ran, boxed and even played football at China's World Humanoid Robot Games, albeit sometimes falling over in the process . Meanwhile, companies have been readying their own range of humanoids that promise to do something a bit more useful: help around the house .


Google makes Gemini personalization available to free users

Engadget

After AI Pro and Ultra subscribers first got to first try the feature, now anyone in the US can enable it. Gemini's Personal Intelligence feature is now rolling out to more users in the US. At the start of the year, Google introduced Personal Intelligence, a Gemini feature that allows the chatbot to pull information from the user's other Google apps and services to generate personalized responses. After making the feature first available to Google AI Pro and Ultra subscribers, the company is expanding availability to more users in the US. Google is kicking off the expansion with AI Mode.