Goto

Collaborating Authors

 conn


Silicon Valley Tech Workers Are Campaigning to Get ICE Out of US Cities

WIRED

Even as Big Tech CEOs curry favor with President Trump, Silicon Valley employees are calling on their bosses to use their influence to help stop his immigration policies. The first Trump administration, and the tech industry that stood up to it, are both looking quainter by the day. Here's one example: In 2017, when President Trump issued a series of executive orders instituting a travel ban on foreigners from certain countries (predominantly Muslim-majority ones), people from across the United States vigorously protested the policy. They included some of tech's most elite: Google cofounder Sergey Brin, who joined a demonstration at the San Francisco airport; Amazon founder Jeff Bezos, who wrote a company-wide email outlining "legal options" that Amazon was considering to fight the ban; and Facebook founder Mark Zuckerberg, who took to Instagram to describe his own family's immigrant roots. On Saturday, hours after federal agents shot and killed ICU nurse Alex Pretti in the streets of Minneapolis, several prominent tech executives attended a private White House screening of, a documentary being released by (of course) Amazon MGM Studios. The timing was not lost on the group of Silicon Valley workers who recently launched ICEout.tech The letter, posted following Renee Nicole Good's killing earlier this month, has now been signed by more than 1,000 tech employees. Those workers, who come from across the spectrum of Big Tech companies and startups, are asking that executives use their clout to demand Immigration and Customs Enforcement agents leave American cities, that they cancel company contracts with the agency, and that they speak publicly about ICE's violent and deadly tactics. Worker-led demands like those were commonplace during Trump 1.0, when tech employees at the world's biggest companies often spoke out--internally and externally--about the cruelty of the US administration and the industry's role in facilitating or tempering its most craven policies. Meanwhile, the executives leading those companies have been busy kissing the ring-- over dinner at the White House or with outlandishly expensive documentaries nobody's watching--at every opportunity. Is the dam finally breaking? This week, Silicon Valley leaders including Anthropic heads Dario and Daniela Amodei, OpenAI CEO Sam Altman, and Apple CEO Tim Cook finally spoke out about ICE's outrageous overreach.


Scalable Satellite Swarm Deployment via Distance-based Orbital Transition Under $J_2$ Perturbation

Takahashi, Yuta, Sakai, Shin-ichiro

arXiv.org Artificial Intelligence

This paper presents an autonomous guidance and control strategy for a satellite swarm that enables scalable distributed space structures for innovative science and business opportunities. The averaged $J_2$ orbital parameters that describe the drift and periodic orbital motion were derived along with their target values to achieve a distributed space structure in a decentralized manner. This enabled the design of a distance-based orbital stabilizer to ensure autonomous deployment into a monolithic formation of a coplanar equidistant configuration on a user-defined orbital plane. Continuous formation control was assumed to be achieved through fuel-free actuation, such as satellite magnetic field interaction and differential aerodynamic forces, thereby maintaining long-term formation stability without thruster usage. A major challenge for such actuation systems is the potential loss of control capability due to increasing inter-satellite distances resulting from unstable orbital dynamics, particularly for autonomous satellite swarms. To mitigate this risk, our decentralized deployment controller minimized drift distance during unexpected communication outages. As a case study, we consider the deployment of palm-sized satellites into a coplanar equidistant formation in a $J_2$-perturbed orbit. Moreover, centralized grouping strategies are presented.


Discursive Circuits: How Do Language Models Understand Discourse Relations?

Miao, Yisong, Kan, Min-Yen

arXiv.org Artificial Intelligence

Which components in transformer language models are responsible for discourse understanding? We hypothesize that sparse computational graphs, termed as discursive circuits, control how models process discourse relations. Unlike simpler tasks, discourse relations involve longer spans and complex reasoning. To make circuit discovery feasible, we introduce a task called Completion under Discourse Relation (CuDR), where a model completes a discourse given a specified relation. To support this task, we construct a corpus of minimal contrastive pairs tailored for activation patching in circuit discovery. Experiments show that sparse circuits ($\approx 0.2\%$ of a full GPT-2 model) recover discourse understanding in the English PDTB-based CuDR task. These circuits generalize well to unseen discourse frameworks such as RST and SDRT. Further analysis shows lower layers capture linguistic features such as lexical semantics and coreference, while upper layers encode discourse-level abstractions. Feature utility is consistent across frameworks (e.g., coreference supports Expansion-like relations).


Reviews: Cooperative neural networks (CoNN): Exploiting prior independence structure for improved classification

Neural Information Processing Systems

Summary: The authors propose a new method that combines a latent Dirichlet allocation (LDA) model with a neural network architecture for the application of supervised text classification –– a model that can be trained end-to-end. In particular, they use a network structure to approximate the intractable inference equations that solve the KL-divergence between the LDA posterior and its approximation which is based on marginal distributions. The authors show that an embedding in a Hilbert space can allow for the approximation of the inference equations, and they choose neural networks to parametrize the functional mapping. Finally, based on two applications, the authors demonstrate an incremental advancement over previous models. Clarity: The overall writing is good, especially as it is a very technical paper with many mathematical details.


Beyond Text-to-SQL for IoT Defense: A Comprehensive Framework for Querying and Classifying IoT Threats

Pavlich, Ryan, Ebadi, Nima, Tarbell, Richard, Linares, Billy, Tan, Adrian, Humphreys, Rachael, Das, Jayanta Kumar, Ghandiparsi, Rambod, Haley, Hannah, George, Jerris, Slavin, Rocky, Choo, Kim-Kwang Raymond, Dietrich, Glenn, Rios, Anthony

arXiv.org Artificial Intelligence

Recognizing the promise of natural language interfaces to databases, prior studies have emphasized the development of text-to-SQL systems. While substantial progress has been made in this field, existing research has concentrated on generating SQL statements from text queries. The broader challenge, however, lies in inferring new information about the returned data. Our research makes two major contributions to address this gap. First, we introduce a novel Internet-of-Things (IoT) text-to-SQL dataset comprising 10,985 text-SQL pairs and 239,398 rows of network traffic activity. The dataset contains additional query types limited in prior text-to-SQL datasets, notably temporal-related queries. Our dataset is sourced from a smart building's IoT ecosystem exploring sensor read and network traffic data. Second, our dataset allows two-stage processing, where the returned data (network traffic) from a generated SQL can be categorized as malicious or not. Our results show that joint training to query and infer information about the data can improve overall text-to-SQL performance, nearly matching substantially larger models. We also show that current large language models (e.g., GPT3.5) struggle to infer new information about returned data, thus our dataset provides a novel test bed for integrating complex domain-specific reasoning into LLMs.


Collaborative Graph Neural Networks for Attributed Network Embedding

Tan, Qiaoyu, Zhang, Xin, Huang, Xiao, Chen, Hao, Li, Jundong, Hu, Xia

arXiv.org Artificial Intelligence

Graph neural networks (GNNs) have shown prominent performance on attributed network embedding. However, existing efforts mainly focus on exploiting network structures, while the exploitation of node attributes is rather limited as they only serve as node features at the initial layer. This simple strategy impedes the potential of node attributes in augmenting node connections, leading to limited receptive field for inactive nodes with few or even no neighbors. Furthermore, the training objectives (i.e., reconstructing network structures) of most GNNs also do not include node attributes, although studies have shown that reconstructing node attributes is beneficial. Thus, it is encouraging to deeply involve node attributes in the key components of GNNs, including graph convolution operations and training objectives. However, this is a nontrivial task since an appropriate way of integration is required to maintain the merits of GNNs. To bridge the gap, in this paper, we propose COllaborative graph Neural Networks--CONN, a tailored GNN architecture for attribute network embedding. It improves model capacity by 1) selectively diffusing messages from neighboring nodes and involved attribute categories, and 2) jointly reconstructing node-to-node and node-to-attribute-category interactions via cross-correlation. Experiments on real-world networks demonstrate that CONN excels state-of-the-art embedding algorithms with a great margin.


Mastering Symbolic Operations: Augmenting Language Models with Compiled Neural Networks

Weng, Yixuan, Zhu, Minjun, Xia, Fei, Li, Bin, He, Shizhu, Liu, Kang, Zhao, Jun

arXiv.org Artificial Intelligence

Language models (LMs) proficiency in handling deterministic symbolic reasoning and rule-based tasks remains limited due to their dependency implicit learning on textual data. To enable fully rule comprehension ability, we explore how to incorporate compiled neural networks (CoNNs) which weight is specially designed into the architecture of LMs, to achieve high accuracy and robust performance. CoNNs are transformer-based neural networks that execute rules through artificially generated attention weights. Our method, which call "Neural Comprehension", by incorporating CoNN modules into the LM, the framework effectively tackles rule-intensive challenges. Our experiments on symbolic reasoning tasks and real-world arithmetic reasoning tasks demonstrate the superior performance of our method compared to existing techniques. Furthermore, our LM achieves flawless execution on symbolic operations tasks, highlighting the potential of our method in enabling LMs to possess true symbolic comprehension capabilities. Our code is publicly available at: https://github.com/WENGSYX/Neural-Comprehension.


Approximate spectral clustering density-based similarity for noisy datasets

Alshammari, Mashaan, Takatsuka, Masahiro

arXiv.org Artificial Intelligence

Approximate spectral clustering (ASC) was developed to overcome heavy computational demands of spectral clustering (SC). It maintains SC ability in predicting non-convex clusters. Since it involves a preprocessing step, ASC defines new similarity measures to assign weights on graph edges. Connectivity matrix (CONN) is an efficient similarity measure to construct graphs for ASC. It defines the weight between two vertices as the number of points assigned to them during vector quantization training. However, this relationship is undirected, where it is not clear which of the vertices is contributing more to that edge. Also, CONN could be tricked by noisy density between clusters. We defined a directed version of CONN, named DCONN, to get insights on vertices contributions to edges. Also, we provided filtering schemes to ensure CONN edges are highlighting potential clusters. Experiments reveal that the proposed filtering was highly efficient when noise cannot be tolerated by CONN.


Fulltime Data Architect openings in Portland on September 05, 2022

#artificialintelligence

Let s immerse on a journey that shapes the future of coffee, transforming one person, one cup and one neighborhood at a time. Primary Responsibilities: • Own the technical relationship and discovery with cross-functional leaders to create product-centric solutions to advance the sustainability cause.


An Efficient Federated Distillation Learning System for Multi-task Time Series Classification

Xing, Huanlai, Xiao, Zhiwen, Qu, Rong, Zhu, Zonghai, Zhao, Bowen

arXiv.org Artificial Intelligence

This paper proposes an efficient federated distillation learning system (EFDLS) for multi-task time series classification (TSC). EFDLS consists of a central server and multiple mobile users, where different users may run different TSC tasks. EFDLS has two novel components, namely a feature-based student-teacher (FBST) framework and a distance-based weights matching (DBWM) scheme. Within each user, the FBST framework transfers knowledge from its teacher's hidden layers to its student's hidden layers via knowledge distillation, with the teacher and student having identical network structure. For each connected user, its student model's hidden layers' weights are uploaded to the EFDLS server periodically. The DBWM scheme is deployed on the server, with the least square distance used to measure the similarity between the weights of two given models. This scheme finds a partner for each connected user such that the user's and its partner's weights are the closest among all the weights uploaded. The server exchanges and sends back the user's and its partner's weights to these two users which then load the received weights to their teachers' hidden layers. Experimental results show that the proposed EFDLS achieves excellent performance on a set of selected UCR2018 datasets regarding top-1 accuracy.