Goto

Collaborating Authors

 interaction vertex


Sequential Learning in the Dense Associative Memory

McAlister, Hayden, Robins, Anthony, Szymanski, Lech

arXiv.org Artificial Intelligence

Sequential learning involves learning tasks in a sequence, and proves challenging for most neural networks. Biological neural networks regularly conquer the sequential learning challenge and are even capable of transferring knowledge both forward and backwards between tasks. Artificial neural networks often totally fail to transfer performance between tasks, and regularly suffer from degraded performance or catastrophic forgetting on previous tasks. Models of associative memory have been used to investigate the discrepancy between biological and artificial neural networks due to their biological ties and inspirations, of which the Hopfield network is perhaps the most studied model. The Dense Associative Memory, or modern Hopfield network, generalizes the Hopfield network, allowing for greater capacities and prototype learning behaviors, while still retaining the associative memory structure. We investigate the performance of the Dense Associative Memory in sequential learning problems, and benchmark various sequential learning techniques in the network. We give a substantial review of the sequential learning space with particular respect to the Hopfield network and associative memories, as well as describe the techniques we implement in detail. We also draw parallels between the classical and Dense Associative Memory in the context of sequential learning, and discuss the departures from biological inspiration that may influence the utility of the Dense Associative Memory as a tool for studying biological neural networks. We present our findings, and show that existing sequential learning methods can be applied to the Dense Associative Memory to improve sequential learning performance.


Graph Neural Networks for Low-Energy Event Classification & Reconstruction in IceCube

Abbasi, R., Ackermann, M., Adams, J., Aggarwal, N., Aguilar, J. A., Ahlers, M., Ahrens, M., Alameddine, J. M., Alves, A. A. Jr., Amin, N. M., Andeen, K., Anderson, T., Anton, G., Argüelles, C., Ashida, Y., Athanasiadou, S., Axani, S., Bai, X., V., A. Balagopal, Baricevic, M., Barwick, S. W., Basu, V., Bay, R., Beatty, J. J., Becker, K. -H., Tjus, J. Becker, Beise, J., Bellenghi, C., Benda, S., BenZvi, S., Berley, D., Bernardini, E., Besson, D. Z., Binder, G., Bindig, D., Blaufuss, E., Blot, S., Bontempo, F., Book, J. Y., Borowka, J., Meneguolo, C. Boscolo, Böser, S., Botner, O., Böttcher, J., Bourbeau, E., Braun, J., Brinson, B., Brostean-Kaiser, J., Burley, R. T., Busse, R. S., Campana, M. A., Carnie-Bronca, E. G., Chen, C., Chen, Z., Chirkin, D., Choi, K., Clark, B. A., Classen, L., Coleman, A., Collin, G. H., Connolly, A., Conrad, J. M., Coppin, P., Correa, P., Countryman, S., Cowen, D. F., Cross, R., Dappen, C., Dave, P., De Clercq, C., DeLaunay, J. J., López, D. Delgado, Dembinski, H., Deoskar, K., Desai, A., Desiati, P., de Vries, K. D., de Wasseige, G., DeYoung, T., Diaz, A., Díaz-Vélez, J. C., Dittmer, M., Dujmovic, H., DuVernois, M. A., Ehrhardt, T., Eller, P., Engel, R., Erpenbeck, H., Evans, J., Evenson, P. A., Fan, K. L., Fazely, A. R., Fedynitch, A., Feigl, N., Fiedlschuster, S., Fienberg, A. T., Finley, C., Fischer, L., Fox, D., Franckowiak, A., Friedman, E., Fritz, A., Fürst, P., Gaisser, T. K., Gallagher, J., Ganster, E., Garcia, A., Garrappa, S., Gerhardt, L., Ghadimi, A., Glaser, C., Glauch, T., Glüsenkamp, T., Goehlke, N., Gonzalez, J. G., Goswami, S., Grant, D., Gray, S. J., Grégoire, T., Griswold, S., Günther, C., Gutjahr, P., Haack, C., Hallgren, A., Halliday, R., Halve, L., Halzen, F., Hamdaoui, H., Minh, M. Ha, Hanson, K., Hardin, J., Harnisch, A. A., Hatch, P., Haungs, A., Helbing, K., Hellrung, J., Henningsen, F., Heuermann, L., Hickford, S., Hill, C., Hill, G. C., Hoffman, K. D., Hoshina, K., Hou, W., Huber, T., Hultqvist, K., Hünnefeld, M., Hussain, R., Hymon, K., In, S., Iovine, N., Ishihara, A., Jansson, M., Japaridze, G. S., Jeong, M., Jin, M., Jones, B. J. P., Kang, D., Kang, W., Kang, X., Kappes, A., Kappesser, D., Kardum, L., Karg, T., Karl, M., Karle, A., Katz, U., Kauer, M., Kelley, J. L., Kheirandish, A., Kin, K., Kiryluk, J., Klein, S. R., Kochocki, A., Koirala, R., Kolanoski, H., Kontrimas, T., Köpke, L., Kopper, C., Koskinen, D. J., Koundal, P., Kovacevich, M., Kowalski, M., Kozynets, T., Krupczak, E., Kun, E., Kurahashi, N., Lad, N., Gualda, C. Lagunas, Larson, M. J., Lauber, F., Lazar, J. P., Lee, J. W., Leonard, K., Leszczyńska, A., Lincetto, M., Liu, Q. R., Liubarska, M., Lohfink, E., Love, C., Mariscal, C. J. Lozano, Lu, L., Lucarelli, F., Ludwig, A., Luszczak, W., Lyu, Y., Ma, W. Y., Madsen, J., Mahn, K. B. M., Makino, Y., Mancina, S., Sainte, W. Marie, Mariş, I. C., Marka, S., Marka, Z., Marsee, M., Martinez-Soler, I., Maruyama, R., McElroy, T., McNally, F., Mead, J. V., Meagher, K., Mechbal, S., Medina, A., Meier, M., Meighen-Berger, S., Merckx, Y., Micallef, J., Mockler, D., Montaruli, T., Moore, R. W., Morse, R., Moulai, M., Mukherjee, T., Naab, R., Nagai, R., Naumann, U., Nayerhoda, A., Necker, J., Neumann, M., Niederhausen, H., Nisa, M. U., Nowicki, S. C., Pollmann, A. Obertacke, Oehler, M., Oeyen, B., Olivas, A., Orsoe, R., Osborn, J., O'Sullivan, E., Pandya, H., Pankova, D. V., Park, N., Parker, G. K., Paudel, E. N., Paul, L., Heros, C. Pérez de los, Peters, L., Petersen, T. C., Peterson, J., Philippen, S., Pieper, S., Pizzuto, A., Plum, M., Popovych, Y., Porcelli, A., Rodriguez, M. Prado, Pries, B., Procter-Murphy, R., Przybylski, G. T., Raab, C., Rack-Helleis, J., Rameez, M., Rawlins, K., Rechav, Z., Rehman, A., Reichherzer, P., Renzi, G., Resconi, E., Reusch, S., Rhode, W., Richman, M., Riedel, B., Roberts, E. J., Robertson, S., Rodan, S., Roellinghoff, G., Rongen, M., Rott, C., Ruhe, T., Ruohan, L., Ryckbosch, D., Cantu, D. Rysewyk, Safa, I., Saffer, J., Salazar-Gallegos, D., Sampathkumar, P., Herrera, S. E. Sanchez, Sandrock, A., Santander, M., Sarkar, S., Sarkar, S., Schaufel, M., Schieler, H., Schindler, S., Schlueter, B., Schmidt, T., Schneider, J., Schröder, F. G., Schumacher, L., Schwefer, G., Sclafani, S., Seckel, D., Seunarine, S., Sharma, A., Shefali, S., Shimizu, N., Silva, M., Skrzypek, B., Smithers, B., Snihur, R., Soedingrekso, J., Søgaard, A., Soldin, D., Spannfellner, C., Spiczak, G. M., Spiering, C., Stamatikos, M., Stanev, T., Stein, R., Stezelberger, T., Stürwald, T., Stuttard, T., Sullivan, G. W., Taboada, I., Ter-Antonyan, S., Thompson, W. G., Thwaites, J., Tilav, S., Tollefson, K., Tönnis, C., Toscano, S., Tosi, D., Trettin, A., Tung, C. F., Turcotte, R., Twagirayezu, J. P., Ty, B., Elorrieta, M. A. Unland, Upshaw, K., Valtonen-Mattila, N., Vandenbroucke, J., van Eijndhoven, N., Vannerom, D., van Santen, J., Vara, J., Veitch-Michaelis, J., Verpoest, S., Veske, D., Walck, C., Wang, W., Watson, T. B., Weaver, C., Weigel, P., Weindl, A., Weldert, J., Wendt, C., Werthebach, J., Weyrauch, M., Whitehorn, N., Wiebusch, C. H., Willey, N., Williams, D. R., Wolf, M., Wrede, G., Wulff, J., Xu, X. W., Yanez, J. P., Yildizci, E., Yoshida, S., Yu, S., Yuan, T., Zhang, Z., Zhelnin, P.

arXiv.org Artificial Intelligence

IceCube, a cubic-kilometer array of optical sensors built to detect atmospheric and astrophysical neutrinos between 1 GeV and 1 PeV, is deployed 1.45 km to 2.45 km below the surface of the ice sheet at the South Pole. The classification and reconstruction of events from the in-ice detectors play a central role in the analysis of data from IceCube. Reconstructing and classifying events is a challenge due to the irregular detector geometry, inhomogeneous scattering and absorption of light in the ice and, below 100 GeV, the relatively low number of signal photons produced per event. To address this challenge, it is possible to represent IceCube events as point cloud graphs and use a Graph Neural Network (GNN) as the classification and reconstruction method. The GNN is capable of distinguishing neutrino events from cosmic-ray backgrounds, classifying different neutrino event types, and reconstructing the deposited energy, direction and interaction vertex. Based on simulation, we provide a comparison in the 1-100 GeV energy range to the current state-of-the-art maximum likelihood techniques used in current IceCube analyses, including the effects of known systematic uncertainties. For neutrino event classification, the GNN increases the signal efficiency by 18% at a fixed false positive rate (FPR), compared to current IceCube methods. Alternatively, the GNN offers a reduction of the FPR by over a factor 8 (to below half a percent) at a fixed signal efficiency. For the reconstruction of energy, direction, and interaction vertex, the resolution improves by an average of 13%-20% compared to current maximum likelihood techniques in the energy range of 1-30 GeV. The GNN, when run on a GPU, is capable of processing IceCube events at a rate nearly double of the median IceCube trigger rate of 2.7 kHz, which opens the possibility of using low energy neutrinos in online searches for transient events.


Dense Associative Memory is Robust to Adversarial Inputs

Krotov, Dmitry, Hopfield, John J

arXiv.org Machine Learning

Deep neural networks (DNN) trained in a supervised way suffer from two known problems. First, the minima of the objective function used in learning correspond to data points (also known as rubbish examples or fooling images) that lack semantic similarity with the training data. Second, a clean input can be changed by a small, and often imperceptible for human vision, perturbation, so that the resulting deformed input is misclassified by the network. These findings emphasize the differences between the ways DNN and humans classify patterns, and raise a question of designing learning algorithms that more accurately mimic human perception compared to the existing methods. Our paper examines these questions within the framework of Dense Associative Memory (DAM) models. These models are defined by the energy function, with higher order (higher than quadratic) interactions between the neurons. We show that in the limit when the power of the interaction vertex in the energy function is sufficiently large, these models have the following three properties. First, the minima of the objective function are free from rubbish images, so that each minimum is a semantically meaningful pattern. Second, artificial patterns poised precisely at the decision boundary look ambiguous to human subjects and share aspects of both classes that are separated by that decision boundary. Third, adversarial images constructed by models with small power of the interaction vertex, which are equivalent to DNN with rectified linear units (ReLU), fail to transfer to and fool the models with higher order interactions. This opens up a possibility to use higher order models for detecting and stopping malicious adversarial attacks. The presented results suggest that DAM with higher order energy functions are closer to human visual perception than DNN with ReLUs.