inaddition
- North America > United States > Illinois (0.04)
- North America > United States > District of Columbia > Washington (0.04)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
3d779cae2d46cf6a8a99a35ba4167977-AuthorFeedback.pdf
Our approach is purely based on 2D convolutions. Nevertheless, it3 outperforms or performs comparably to many more costly 3D models. We thank the reviewers for pointing out some related (or missing) references. The12 Timeception layers involve group convolutions at different time scales while our TAM layers only use depthwise13 convolution. As a result, the Timeception has significantly more parameters than the TAM (10% vs. 0.1% of the14 totalmodelparameters).
6 SupplementaryMaterial
The original CLUTRR data generation framework made sure that each testproof is not in the training set in order to test whether a model is able to generalize to unseen proofs. Initial results on the original CLUTRR test sets resulted in strong model performance ( 99%) on levels seen during training (2, 4, 6) but no generalization at all ( 0%) to other levels. The models are given as input "
- North America > United States > New York > New York County > New York City (0.05)
- Europe > Sweden > Stockholm > Stockholm (0.05)
- Asia > Middle East > Jordan (0.05)
- (6 more...)
- North America > United States > Virginia (0.04)
- Europe > United Kingdom > Scotland > City of Edinburgh > Edinburgh (0.04)
- Europe > France (0.04)
- North America > United States > California > Santa Clara County > Palo Alto (0.04)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.04)
- Asia > South Korea > Seoul > Seoul (0.05)
- North America > United States > California (0.04)
- Asia > South Korea > Daejeon > Daejeon (0.04)
- North America > United States > Colorado > Denver County > Denver (0.04)
- North America > United States > California > Alameda County > Berkeley (0.04)
- North America > United States > Arizona > Maricopa County > Phoenix (0.04)
- (4 more...)
GameSolvingwithOnlineFine-Tuning
A.1 PCNtraining We basically follow the same PCN training method by Wu et al.[1] but replace the AlphaZero algorithm with the Gumbel AlphaZero algorithm [2], where the simulation count is set to 322 in self-play and starts by sampling 16 actions. The architecture of the PCN contains three residual blocks with 256 hidden channels. Atotal of400,000 self-play games are generated for the whole training. During optimization, the learning rate is fixed at 0.02, and the batch size is set to 1,024. A.3 Workerdesign The worker is itself a Killall-Go solver. Thus,tofullyutilize GPU resources, we implement batch GPU inferencing to accelerate PCN evaluations for workers.
- Asia > Taiwan (0.05)
- North America > Canada (0.04)