Goto

Collaborating Authors

LinkNet: Relational Embedding for Scene Graph

Neural Information Processing Systems

Objects and their relationships are critical contents for image understanding. A scene graph provides a structured description that captures these properties of an image. However, reasoning about the relationships between objects is very challenging and only a few recent works have attempted to solve the problem of generating a scene graph from an image. In this paper, we present a novel method that improves scene graph generation by explicitly modeling inter-dependency among the entire object instances. We design a simple and effective relational embedding module that enables our model to jointly represent connections among all related objects, rather than focus on an object in isolation. Our novel method significantly benefits two main parts of the scene graph generation task: object classification and relationship classification. Using it on top of a basic Faster R-CNN, our model achieves state-of-the-art results on the Visual Genome benchmark.



Automating Bayesian optimization with Bayesian optimization

Neural Information Processing Systems

Bayesian optimization is a powerful tool for global optimization of expensive functions. One of its key components is the underlying probabilistic model used for the objective function f. In practice, however, it is often unclear how one should appropriately choose a model, especially when gathering data is expensive. In this work, we introduce a novel automated Bayesian optimization approach that dynamically selects promising models for explaining the observed data using Bayesian Optimization in the model space. Crucially, we account for the uncertainty in the choice of model; our method is capable of using multiple models to represent its current belief about f and subsequently using this information for decision making. We argue, and demonstrate empirically, that our approach automatically finds suitable models for the objective function, which ultimately results in more-efficient optimization.


A Model Architecture

Neural Information Processing Systems

In this section, we provide comprehensive details about the Transformer model architectures considered in this work. Experiments conducted on both DMLab and RoboMimic include RGB image observations. Detailed model parameters are listed in Table A.1. We learn 16-dim embedding vectors for all discrete actions. To encode proprioceptive measurement in RoboMimic, we follow Mandlekar et al.





Posha vs. Thermomix: Kitchen Robots Face Off on Thanksgiving Sides

WIRED

The Posha and the Thermomix TM7 are the closest things to a home robot chef that mere mortals can afford. The catch is that you're the prep cook. The holiday is still almost a week away, and I'm sick of Thanksgiving. I've already made four rounds of mashed potatoes, three of mac and cheese, and three turkeys (with more still waiting in my fridge) as part of testing smart probes to help smoke turkeys outside and preparing seven-course holiday meal kits for friends and family. I was eager to finally outsource some of the cooking by testing two very different robo-chef devices, the Thermomix TM7 and the Posha kitchen robot . Both promise to plan my meals and also do most of the cooking, which sounds pretty good to me. The Thermomix descends from a German device launched in 1968--a time when the best-known robot chef was cartoon Rosie on --that was essentially a blender with a heater. It's since caught on big in countries from Italy to Portugal to Australia, and over the years it's added multi-tier steaming, baking, proofing, a touchscreen, an encyclopedic recipe app, and a whole lot of smart features.