bigg
Minimax-Optimal Multi-Agent RL in Markov Games With a Generative Model
All prior results suffer from at least one of the two obstacles: the curse of multiple agents and the barrier of long horizon, regardless of the sampling protocol in use. We take a step towards settling this problem, assuming access to a flexible sampling mechanism: the generative model. Focusing on non-stationary finite-horizon Markov games, we develop a fast learning algorithm called Q-FTRL and an adaptive sampling scheme that leverage the optimism principle in online adversarial learning (particularly the Follow-the-Regularized-Leader (FTRL) method). Our algorithm learns an $\varepsilon$-approximate CCE in a general-sum Markov game using $$ \widetilde{O}\bigg( \frac{H^4 S \sum_{i=1}^m A_i}{\varepsilon^2} \bigg) $$ samples, where $m$ is the number of players, $S$ indicates the number of states, $H$ is the horizon, and $A_i$ denotes the number of actions for the $i$-th player. This is minimax-optimal (up to log factor) when $m$ is fixed. When applied to two-player zero-sum Markov games, our algorithm provably finds an $\varepsilon$-approximate Nash equilibrium with a minimal number of samples. Along the way, we derive a refined regret bound for FTRL that makes explicit the role of variance-type quantities, which might be of independent interest.
America's most renowned 'prophet' makes startling prediction about alien 'mothership'
Florida's housing market is flashing a warning for the rest of the US Now scientists redefine'obese' - and they've made up to 60% more people'fat' Skip Bayless claims Travis Hunter has'mentally checked out' after he opted to get baptized on morning of a game'You will DIE if you do not remove your breasts', doctors screamed at me. I refused and tried a new experimental therapy instead... now I'm cancer-free Police say they have FOUND woman seen in viral'kidnapping' video and reveal what happened to her after harrowing footage emerged Bella Hadid's health battle takes dark turn: Loved ones reveal hellish new details about'missing' model... as ominous texts emerge The world's most powerful passport revealed - as UK and USA both drop to record lows Unmasked after 80 years - the Nazi executioner in infamous WWII photo: Historian uses AI to uncover identity of killer in'The Last Jew of Vinnytsia' image Will Trump's Gaza peace deal fail? Policy expert MARK DUBOWITZ breaks down all the forces at play... and how the president can actually pull this off America's most renowned'prophet' makes startling prediction about alien'mothership' Kim Kardashian says she wasn't'emotionally or financially safe' during'toxic' marriage to Kanye West as she claims rapper hasn't contacted their children for MONTHS and has destroyed her dating life Every woman I date has the same repulsive bedroom kink... it feels so wrong, but I always say yes: DEAR JANE'Pathetic' JD Vance slammed for'cheap' reaction to racist texts as Young Republicans spark Trump world crisis Ugly divorce war between Mitt Romney's wealthy brother and estranged wife before she was found dead Full horrors of torture suffered by Noa Argamani's commando boyfriend are revealed - including how 6ft 5in hostage was beaten and kept chained in 6ft cell for a year after he tried to escape from Hamas Mother, 52, and daughter, 21, die after eating'poisoned birthday cake delivered by relative who owed them money' in Brazil I had 30 debilitating symptoms but doctors dismissed me. America's most renowned'prophet' makes startling prediction about alien'mothership' READ MORE: Precise date interstellar visitor will reveal itself as'alien mothership' or comet confirmed A Christian pastor who accurately foresaw the assassination attempt on Donald Trump three months prior has shared a new vision about a threat in the sky. Brandon Biggs claimed that God showed him a vision of an ' alien ' ship flying over the Vatican and Mayan temples in Mexico .
- North America > Mexico (0.25)
- Europe > Holy See (0.24)
- South America > Brazil (0.24)
- (17 more...)
- Media > Music (1.00)
- Leisure & Entertainment (1.00)
- Law Enforcement & Public Safety > Crime Prevention & Enforcement (1.00)
- (2 more...)
Minimax-Optimal Multi-Agent RL in Markov Games With a Generative Model
All prior results suffer from at least one of the two obstacles: the curse of multiple agents and the barrier of long horizon, regardless of the sampling protocol in use. We take a step towards settling this problem, assuming access to a flexible sampling mechanism: the generative model. Focusing on non-stationary finite-horizon Markov games, we develop a fast learning algorithm called Q-FTRL and an adaptive sampling scheme that leverage the optimism principle in online adversarial learning (particularly the Follow-the-Regularized-Leader (FTRL) method). Our algorithm learns an \varepsilon -approximate CCE in a general-sum Markov game using \widetilde{O}\bigg( \frac{H 4 S \sum_{i 1} m A_i}{\varepsilon 2} \bigg) samples, where m is the number of players, S indicates the number of states, H is the horizon, and A_i denotes the number of actions for the i -th player. This is minimax-optimal (up to log factor) when m is fixed.
- Information Technology > Artificial Intelligence > Representation & Reasoning > Agents (0.77)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Search (0.64)
- Information Technology > Artificial Intelligence > Natural Language > Generation (0.64)
- Information Technology > Artificial Intelligence > Machine Learning > Reinforcement Learning (0.59)
CANITA: Faster Rates for Distributed Convex Optimization with Communication Compression
Due to the high communication cost in distributed and federated learning, methods relying on compressed communication are becoming increasingly popular. Besides, the best theoretically and practically performing gradient-type methods invariably rely on some form of acceleration/momentum to reduce the number of communications (faster convergence), e.g., Nesterov's accelerated gradient descent [31, 32] and Adam [14]. In order to combine the benefits of communication compression and convergence acceleration, we propose a \emph{compressed and accelerated} gradient method based on ANITA [20] for distributed optimization, which we call CANITA. Our results show that as long as the number of devices n is large (often true in distributed/federated learning), or the compression \omega is not very high, CANITA achieves the faster convergence rate O\Big(\sqrt{\frac{L}{\epsilon}}\Big), i.e., the number of communication rounds is O\Big(\sqrt{\frac{L}{\epsilon}}\Big) (vs. As a result, CANITA enjoys the advantages of both compression (compressed communication in each round) and acceleration (much fewer communication rounds).
House GOP lawmaker proposes using AI to cut federal red tape, streamline services
FIRST ON FOX: House Rep. Andy Biggs is eyeing artificial intelligence (AI) technology as a way to cut unnecessary government red tape. The Arizona Republican is introducing a bill on Tuesday that would mandate federal agencies use AI to review regulations under their purview with the aim of cutting rules that fail to meet certain standards. "American businesses must be given the opportunity to thrive without overbearing, costly, contradictory, and duplicative regulations mandated by the DC Swamp," Biggs told Fox News Digital. "Federal overregulation takes a colossal toll on the U.S. economy. Thousands of new regulations go into effect every year, and there simply isn't enough manpower or existing technology to sift through previously issued regulations. AI technology is an effective tool that can save taxpayer dollars, benefit American business owners, and promote economic growth."
- Law > Statutes (0.74)
- Government > Regional Government > North America Government > United States Government (0.72)
- Banking & Finance > Economy (0.58)
Scalable Deep Generative Modeling for Sparse Graphs
Dai, Hanjun, Nazi, Azade, Li, Yujia, Dai, Bo, Schuurmans, Dale
Learning graph generative models is a challenging task for deep learning and has wide applicability to a range of domains like chemistry, biology and social science. However current deep neural methods suffer from limited scalability: for a graph with $n$ nodes and $m$ edges, existing deep neural methods require $\Omega(n^2)$ complexity by building up the adjacency matrix. On the other hand, many real world graphs are actually sparse in the sense that $m\ll n^2$. Based on this, we develop a novel autoregressive model, named BiGG, that utilizes this sparsity to avoid generating the full adjacency matrix, and importantly reduces the graph generation time complexity to $O((n + m)\log n)$. Furthermore, during training this autoregressive model can be parallelized with $O(\log n)$ synchronization stages, which makes it much more efficient than other autoregressive models that require $\Omega(n)$. Experiments on several benchmarks show that the proposed approach not only scales to orders of magnitude larger graphs than previously possible with deep autoregressive graph generative models, but also yields better graph generation quality.
Kapil Sharma
Least squares estimates are often not very satisfactory due to their poor out-of-sample performance, especially when the model is overly complex with a lot of features. We can attribute this to low bias and large variance in least squares estimates. Additionally, when we have a lot of features in our model, it is harder to explain the features with the strongest effect or what we call the Big Picture. Hence, we might want to choose fewer features in order to trade a worse in-sample variance for a better out-of-sample prediction. Regularization is a method to shrink or drop coefficients/parameters from a model by imposing a penalty on their size.
LogisticRegression - mlxtend
Related to the Perceptron and'Adaline', a Logistic Regression model is a linear model for binary classification. However, instead of minimizing a linear cost function such as the sum of squared errors (SSE) in Adaline, we minimize a sigmoid function, i.e., the logistic function: Here, p(y 1 \mid \mathbf{x}) is the conditional probability that a particular sample belongs to class 1 given its features \mathbf{x} . The logit function takes inputs in the range [0, 1] and transform them to values over the entire real number range. In contrast, the logistic function takes input values over the entire real number range and transforms them to values in the range [0, 1]. In other words, the logistic function is the inverse of the logit function, and it lets us predict the conditional probability that a certain sample belongs to class 1 (or class 0).