nullh
KAN/H: Kolmogorov-Arnold Network using Haar-like bases
This paper proposes KAN/H, a variant of Kolmogorov-Arnold Network (KAN) that uses a Haar-variant basis system having both global and local bases instead of B-spline. The resulting algorithm is applied to function approximation problems and MNIST. We show that it does not require most of the problem-specific hyper-parameter tunings.
- Asia > Japan > Kyūshū & Okinawa > Kyūshū > Miyazaki Prefecture > Miyazaki (0.04)
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.04)
Duality-Induced Regularizer for Tensor Factorization Based Knowledge Graph Completion Supplementary Material
Theorem 1. Suppose that ˆ X In DB models, the commonly used p is either 1 or 2. When p = 2, DURA takes the form as the one in Equation (8) in the main text. If p = 1, we cannot expand the squared score function of the associated DB models as in Equation (4). Therefore, we choose p = 2 . 2 Table 2: Hyperparameters found by grid search. Suppose that k is the number of triplets known to be true in the knowledge graph, n is the embedding dimension of entities. That is to say, the computational complexity of weighted DURA is the same as the weighted squared Frobenius norm regularizer.
A Convergence on Two-Layer Nonlinear Networks We consider the family of neural networks f (x) = 1 p p null
Lemma A.2. Assume W (0), β (0) and b have i.i.d. The proof for (A.5) is similar since V ar( To prove (A.6), since | y With a union bound argument, we can show (A.6). Finally, (A.7) followed from standard Gaussian tail bounds and union bound argument, yielding P(max Under the conditions of Theorem 3.2, we define matrices G(0),H (0) R Under the conditions of Theorem 3.2, if the error bound (3.1) holds for all t = 1, 2,...,t From the feedback alignment updates (A.3), we have for all t T | β Lemma A.5. Assume all the inequalities from Lemma A.2 hold. Under the conditions of Theorem 3.2, if the bound for the weights difference (3.2) holds for all t t We prove the inequality (3.1) by induction. Suppose (3.1) and (3.2) hold for all t = 1, 2,...,t Assume all the inequalities from Lemma A.2 hold.
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Oceania > Australia > New South Wales > Sydney (0.04)
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.04)
Duality-Induced Regularizer for Tensor Factorization Based Knowledge Graph Completion Supplementary Material
Theorem 1. Suppose that ˆ X In DB models, the commonly used p is either 1 or 2. When p = 2, DURA takes the form as the one in Equation (8) in the main text. If p = 1, we cannot expand the squared score function of the associated DB models as in Equation (4). Therefore, we choose p = 2 . 2 Table 2: Hyperparameters found by grid search. Suppose that k is the number of triplets known to be true in the knowledge graph, n is the embedding dimension of entities. That is to say, the computational complexity of weighted DURA is the same as the weighted squared Frobenius norm regularizer.
Splitting Answer Set Programs with respect to Intensionality Statements (Extended Version)
Fandinno, Jorge, Lierler, Yuliya
Splitting a logic program allows us to reduce the task of computing its stable models to similar tasks for its subprograms. This can be used to increase solving performance and prove program correctness. We generalize the conditions under which this technique is applicable, by considering not only dependencies between predicates but also their arguments and context. This allows splitting programs commonly used in practice to which previous results were not applicable.
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- North America > United States > Connecticut > New Haven County > New Haven (0.04)
- Europe > Denmark (0.04)