Not enough data to create a plot.
Try a different view from the menu above.
Takeno, Shion
Dose-finding design based on level set estimation in phase I cancer clinical trials
Seno, Keiichiro, Matsui, Kota, Iwazaki, Shogo, Inatsu, Yu, Takeno, Shion, Matsui, Shigeyuki
Dose-finding design based on level set estimation in phase I cancer clinical trials Keiichiro Seno 1 a, Kota Matsui 2b, Shogo Iwazaki 3, Yu Inatsu 4, Shion Takeno 5, 6 and Shigeyuki Matsui 2, 7 1 Department of Biostatistics, Nagoya University 2 Department of Biostatistics, Kyoto University 3 MI-6 Ltd. 4 Department of Computer Science, Nagoya Institute of Technology 5 Department of Mechanical Systems Engineering, Nagoya University 6 Center for Advanced Intelligence Project, RIKEN 7 Research Center for Medical and Health Data Science, The Institute of Statistical Mathematics Abstract The primary objective of phase I cancer clinical trials is to evaluate the safety of a new experimental treatment and to find the maximum tolerated dose (MTD). We show that the MTD estimation problem can be regarded as a level set estimation (LSE) problem whose objective is to determine the regions where an unknown function value is above or below a given threshold. Then, we propose a novel ...
Distributionally Robust Active Learning for Gaussian Process Regression
Takeno, Shion, Okura, Yoshito, Inatsu, Yu, Tatsuya, Aoyama, Tanaka, Tomonari, Satoshi, Akahane, Hanada, Hiroyuki, Hashimoto, Noriaki, Murayama, Taro, Lee, Hanju, Kojima, Shinya, Takeuchi, Ichiro
Gaussian process regression (GPR) or kernel ridge regression is a widely used and powerful tool for nonlinear prediction. Therefore, active learning (AL) for GPR, which actively collects data labels to achieve an accurate prediction with fewer data labels, is an important problem. However, existing AL methods do not theoretically guarantee prediction accuracy for target distribution. Furthermore, as discussed in the distributionally robust learning literature, specifying the target distribution is often difficult. Thus, this paper proposes two AL methods that effectively reduce the worst-case expected error for GPR, which is the worst-case expectation in target distribution candidates. We show an upper bound of the worst-case expected squared error, which suggests that the error will be arbitrarily small by a finite number of data labels under mild conditions. Finally, we demonstrate the effectiveness of the proposed methods through synthetic and real-world datasets.
Improved Regret Analysis in Gaussian Process Bandits: Optimality for Noiseless Reward, RKHS norm, and Non-Stationary Variance
Iwazaki, Shogo, Takeno, Shion
The Gaussian process (GP) bandits [Srinivas et al., 2010] is a powerful framework for sequential decision-making tasks to minimize regret defined by a black-box reward function, which belongs to known reproducing kernel Hilbert space (RKHS). The applications include many fileds such as robotics Berkenkamp et al. [2021], experimental design Lei et al. [2021], and hyperparameter tuning task Snoek et al. [2012]. Many existing studies have been conducted to obtain the theoretical guarantee for the regret. Establised work by Srinivas et al. [2010] has shown the upper bounds of the cumulative regret for the GP upper confidence bound (GP-UCB) algorithm. Furthermore, Valko et al. [2013] have shown the tighter regret upper bound for the SupKernelUCB algorithm. Scarlett et al. [2017] have shown the lower bound of the regret, which implies that the regret upper bound from [Valko et al., 2013] is near-optimal; that is, the regret upper bound matches the lower bound except for the poly-logarithmic factor. Then, several studies further tackled obtaining a near-optimal GP-bandit algorithm. Vakili et al. [2021] have proposed maximum variance reduction (MVR), which is shown to be near-optimal for the simple regret incurred by the last recommended action.
Near-Optimal Algorithm for Non-Stationary Kernelized Bandits
Iwazaki, Shogo, Takeno, Shion
Kernelized bandit (KB) problem [Srinivas et al., 2010], also called Gaussian process bandit or Bayesian optimization, is one of the important sequential decision-making problems where one seeks to minimize the regret under an unknown reward function via sequentially acquiring function evaluations. As the name suggests, in the KB problem, the underlying reward function is assumed to be an element of reproducing kernel Hilbert space (RKHS) induced by a known fixed kernel function. KB has been applied in many applications, such as materials discovery [Ueno et al., 2016], drug discovery [Korovina et al., 2020], and robotics [Berkenkamp et al., 2023]. In addition, the near-optimal KB algorithms, whose regret upper bound matches the regret lower bound derived in Scarlett et al. [2017], have been shown [Camilleri et al., 2021, Salgia et al., 2021, Li and Scarlett, 2022, Salgia et al., 2024]. Non-stationary KB [Bogunovic et al., 2016] considers the optimization under a non-stationary environment; that is, the reward function may change over time within some RKHS. This modification is crucial in many practical applications where an objective function varies over time, such as financial markets [Heaton and Lucas, 1999] and recommender systems [Hariri et al., 2015]. For example, Zhou and Shroff [2021], Deng et al. [2022] have proposed upper confidence bound (UCB)-based algorithms for the non-stationary KB problem and derived the upper bound of the cumulative regret. Recently, Hong et al. [2023] have proposed an optimization-based KB
Distributionally Robust Safe Sample Screening
Hanada, Hiroyuki, Tatsuya, Aoyama, Satoshi, Akahane, Tanaka, Tomonari, Okura, Yoshito, Inatsu, Yu, Hashimoto, Noriaki, Takeno, Shion, Murayama, Taro, Lee, Hanju, Kojima, Shinya, Takeuchi, Ichiro
In this study, we propose a machine learning method called Distributionally Robust Safe Sample Screening (DRSSS). DRSSS aims to identify unnecessary training samples, even when the distribution of the training samples changes in the future. To achieve this, we effectively combine the distributionally robust (DR) paradigm, which aims to enhance model robustness against variations in data distribution, with the safe sample screening (SSS), which identifies unnecessary training samples prior to model training. Since we need to consider an infinite number of scenarios regarding changes in the distribution, we applied SSS because it does not require model training after the change of the distribution. In this paper, we employed the covariate shift framework to represent the distribution of training samples and reformulated the DR covariate-shift problem as a weighted empirical risk minimization problem, where the weights are subject to uncertainty within a predetermined range. By extending the existing SSS technique to accommodate this weight uncertainty, the DRSSS method is capable of reliably identifying unnecessary samples under any future distribution within a specified range. We provide a theoretical guarantee for the DRSSS method and validate its performance through numerical experiments on both synthetic and real-world datasets.
Bounding Box-based Multi-objective Bayesian Optimization of Risk Measures under Input Uncertainty
Inatsu, Yu, Takeno, Shion, Hanada, Hiroyuki, Iwata, Kazuki, Takeuchi, Ichiro
In this study, we propose a novel multi-objective Bayesian optimization (MOBO) method to efficiently identify the Pareto front (PF) defined by risk measures for black-box functions under the presence of input uncertainty (IU). Existing BO methods for Pareto optimization in the presence of IU are risk-specific or without theoretical guarantees, whereas our proposed method addresses general risk measures and has theoretical guarantees. The basic idea of the proposed method is to assume a Gaussian process (GP) model for the black-box function and to construct high-probability bounding boxes for the risk measures using the GP model. Furthermore, in order to reduce the uncertainty of non-dominated bounding boxes, we propose a method of selecting the next evaluation point using a maximin distance defined by the maximum value of a quasi distance based on bounding boxes. As theoretical analysis, we prove that the algorithm can return an arbitrary-accurate solution in a finite number of iterations with high probability, for various risk measures such as Bayes risk, worst-case risk, and value-at-risk. We also give a theoretical analysis that takes into account approximation errors because there exist non-negligible approximation errors (e.g., finite approximation of PFs and sampling-based approximation of bounding boxes) in practice. We confirm that the proposed method outperforms compared with existing methods not only in the setting with IU but also in the setting of ordinary MOBO through numerical experiments.
Multi-Objective Bayesian Optimization with Active Preference Learning
Ozaki, Ryota, Ishikawa, Kazuki, Kanzaki, Youhei, Suzuki, Shinya, Takeno, Shion, Takeuchi, Ichiro, Karasuyama, Masayuki
There are a lot of real-world black-box optimization problems that need to optimize multiple criteria simultaneously. However, in a multi-objective optimization (MOO) problem, identifying the whole Pareto front requires the prohibitive search cost, while in many practical scenarios, the decision maker (DM) only needs a specific solution among the set of the Pareto optimal solutions. We propose a Bayesian optimization (BO) approach to identifying the most preferred solution in the MOO with expensive objective functions, in which a Bayesian preference model of the DM is adaptively estimated by an interactive manner based on the two types of supervisions called the pairwise preference and improvement request. To explore the most preferred solution, we define an acquisition function in which the uncertainty both in the objective functions and the DM preference is incorporated. Further, to minimize the interaction cost with the DM, we also propose an active learning strategy for the preference estimation. We empirically demonstrate the effectiveness of our proposed method through the benchmark function optimization and the hyper-parameter optimization problems for machine learning models.
Posterior Sampling-Based Bayesian Optimization with Tighter Bayesian Regret Bounds
Takeno, Shion, Inatsu, Yu, Karasuyama, Masayuki, Takeuchi, Ichiro
Among various acquisition functions (AFs) in Bayesian optimization (BO), Gaussian process upper confidence bound (GP-UCB) and Thompson sampling (TS) are well-known options with established theoretical properties regarding Bayesian cumulative regret (BCR). Recently, it has been shown that a randomized variant of GP-UCB achieves a tighter BCR bound compared with GP-UCB, which we call the tighter BCR bound for brevity. Inspired by this study, this paper first shows that TS achieves the tighter BCR bound. On the other hand, GP-UCB and TS often practically suffer from manual hyperparameter tuning and over-exploration issues, respectively. To overcome these difficulties, we propose yet another AF called a probability of improvement from the maximum of a sample path (PIMS). We show that PIMS achieves the tighter BCR bound and avoids the hyperparameter tuning, unlike GP-UCB. Furthermore, we demonstrate a wide range of experiments, focusing on the effectiveness of PIMS that mitigates the practical issues of GP-UCB and TS.
Towards Practical Preferential Bayesian Optimization with Skew Gaussian Processes
Takeno, Shion, Nomura, Masahiro, Karasuyama, Masayuki
We study preferential Bayesian optimization (BO) where reliable feedback is limited to pairwise comparison called duels. An important challenge in preferential BO, which uses the preferential Gaussian process (GP) model to represent flexible preference structure, is that the posterior distribution is a computationally intractable skew GP. The most widely used approach for preferential BO is Gaussian approximation, which ignores the skewness of the true posterior. Alternatively, Markov chain Monte Carlo (MCMC) based preferential BO is also proposed. In this work, we first verify the accuracy of Gaussian approximation, from which we reveal the critical problem that the predictive probability of duels can be inaccurate. This observation motivates us to improve the MCMC-based estimation for skew GP, for which we show the practical efficiency of Gibbs sampling and derive the low variance MC estimator. However, the computational time of MCMC can still be a bottleneck in practice. Towards building a more practical preferential BO, we develop a new method that achieves both high computational efficiency and low sample complexity, and then demonstrate its effectiveness through extensive numerical experiments.
Randomized Gaussian Process Upper Confidence Bound with Tighter Bayesian Regret Bounds
Takeno, Shion, Inatsu, Yu, Karasuyama, Masayuki
Gaussian process upper confidence bound (GP-UCB) is a theoretically promising approach for black-box optimization; however, the confidence parameter $\beta$ is considerably large in the theorem and chosen heuristically in practice. Then, randomized GP-UCB (RGP-UCB) uses a randomized confidence parameter, which follows the Gamma distribution, to mitigate the impact of manually specifying $\beta$. This study first generalizes the regret analysis of RGP-UCB to a wider class of distributions, including the Gamma distribution. Furthermore, we propose improved RGP-UCB (IRGP-UCB) based on a two-parameter exponential distribution, which achieves tighter Bayesian regret bounds. IRGP-UCB does not require an increase in the confidence parameter in terms of the number of iterations, which avoids over-exploration in the later iterations. Finally, we demonstrate the effectiveness of IRGP-UCB through extensive experiments.