Goto

Collaborating Authors

 neighborhood


Two Literal Crypto Bros Built a Real Estate Empire. Then the Homes Started to Fall Apart

WIRED

Two Literal Crypto Bros Built a Real Estate Empire. In 2019, two Canadian brothers blew into Detroit with an irresistible pitch: For $50, almost anyone could become a property owner. When houses decayed and the city intervened, the blame games began. A fire broke out at 10410 Cadieux in March 2025, burning a hole in the roof. The smell hit me first: damp brick, stagnant water, mold, and bleach. I was partway down a flight of wooden stairs that led to the basement of a 1920s duplex in east Detroit, Michigan. Leading the way was Cornell Dorris, a tenant in the building for nearly a decade. Dorris is in his early forties, has two daughters who visit on weekends, and makes a living smoking meat and cooking for events. As my eyes adjusted, I made out rodent droppings and a black puddle that spread across the basement floor. "Anytime it rains, the water comes down," Dorris said. The air was unnaturally heavy, and I felt a nagging urge to leave. Dorris doesn't have a typical landlord. Almost four years ago, his building was acquired by a startup called RealToken, or RealT.


Discriminative Gaifman Models

Neural Information Processing Systems

Gaifman models learn feature representations bottom up from representations of locally connected and bounded-size regions of knowledge bases (KBs). Considering local and bounded-size neighborhoods of knowledge bases renders logical inference and learning tractable, mitigates the problem of overfitting, and facilitates weight sharing. Gaifman models sample neighborhoods of knowledge bases so as to make the learned relational models more robust to missing objects and relations which is a common situation in open-world KBs. We present the core ideas of Gaifman models and apply them to large-scale relational learning problems. We also discuss the ways in which Gaifman models relate to some existing relational machine learning approaches.


Non-Local Recurrent Network for Image Restoration

Neural Information Processing Systems

Many classic methods have shown non-local self-similarity in natural images to be an effective prior for image restoration. However, it remains unclear and challenging to make use of this intrinsic property via deep networks. In this paper, we propose a non-local recurrent network (NLRN) as the first attempt to incorporate non-local operations into a recurrent neural network (RNN) for image restoration. The main contributions of this work are: (1) Unlike existing methods that measure self-similarity in an isolated manner, the proposed non-local module can be flexibly integrated into existing deep networks for end-to-end training to capture deep feature correlation between each location and its neighborhood.


KONG: Kernels for ordered-neighborhood graphs

Neural Information Processing Systems

We present novel graph kernels for graphs with node and edge labels that have ordered neighborhoods, i.e. when neighbor nodes follow an order. Graphs with ordered neighborhoods are a natural data representation for evolving graphs where edges are created over time, which induces an order.


Sketching Method for Large Scale Combinatorial Inference

Neural Information Processing Systems

We present computationally efficient algorithms to test various combinatorial structures of large-scale graphical models. In order to test the hypotheses on their topological structures, we propose two adjacency matrix sketching frameworks: neighborhood sketching and subgraph sketching. The neighborhood sketching algorithm is proposed to test the connectivity of graphical models. This algorithm randomly subsamples vertices and conducts neighborhood regression and screening. The global sketching algorithm is proposed to test the topological properties requiring exponential computation complexity, especially testing the chromatic number and the maximum clique. This algorithm infers the corresponding property based on the sampled subgraph. Our algorithms are shown to substantially accelerate the computation of existing methods.






Efficient Discrepancy Testing for Learning with Distribution Shift Gautam Chandrasekaran UT Austin Adam R. Klivans UT Austin Vasilis Kontonis UT Austin Konstantinos Stavropoulos

Neural Information Processing Systems

Our approach generalizes and improves all prior work on TDS learning: (1) we obtain universal learners that succeed simultaneously for large classes of test distributions, (2) achieve near-optimal error rates, and (3) give exponential improvements for constant depth circuits.