Goto

Collaborating Authors

 technique


Techniques for Symbol Grounding with SATNet

Neural Information Processing Systems

Many experts argue that the future of artificial intelligence is limited by the field's ability to integrate symbolic logical reasoning into deep learning architectures. The recently proposed differentiable MAXSAT solver, SATNet, was a breakthrough in its capacity to integrate with a traditional neural network and solve visual reasoning problems. For instance, it can learn the rules of Sudoku purely from image examples. Despite its success, SATNet was shown to succumb to a key challenge in neurosymbolic systems known as the Symbol Grounding Problem: the inability to map visual inputs to symbolic variables without explicit supervision (label leakage). In this work, we present a self-supervised pre-training pipeline that enables SATNet to overcome this limitation, thus broadening the class of problems that SATNet architectures can solve to include datasets where no intermediary labels are available at all. We demonstrate that our method allows SATNet to attain full accuracy even with a harder problem setup that prevents any label leakage. We additionally introduce a proofreading method that further improves the performance of SATNet architectures, beating the state-of-the-art on Visual Sudoku.


Conditioning and Processing: Techniques to Improve Information-Theoretic Generalization Bounds

Neural Information Processing Systems

Obtaining generalization bounds for learning algorithms is one of the main subjects studied in theoretical machine learning. In recent years, information-theoretic bounds on generalization have gained the attention of researchers. This approach provides an insight into learning algorithms by considering the mutual information between the model and the training set. In this paper, a probabilistic graphical representation of this approach is adopted and two general techniques to improve the bounds are introduced, namely conditioning and processing. In conditioning, a random variable in the graph is considered as given, while in processing a random variable is substituted with one of its children. These techniques can be used to improve the bounds by either sharpening them or increasing their applicability. It is demonstrated that the proposed framework provides a simple and unified way to explain a variety of recent tightening results. New improved bounds derived by utilizing these techniques are also proposed.


Value Imprint: A Technique for Auditing the Human Values Embedded in RLHF Datasets

Neural Information Processing Systems

LLMs are increasingly fine-tuned using RLHF datasets to align them with human preferences and values. However, very limited research has investigated which specific human values are operationalized through these datasets. In this paper, we introduce Value Imprint, a framework for auditing and classifying the human values embedded within RLHF datasets. To investigate the viability of this framework, we conducted three case study experiments by auditing the Anthropic/hh-rlhf, OpenAI WebGPT Comparisons, and Alpaca GPT-4-LLM datasets to examine the human values embedded within them. Our analysis involved a two-phase process.


Working with Text -Part 4. Techniques in handling text data

#artificialintelligence

Example: 'I want to read a book' In the above example there are 6 tokens which are- ('I', 'want, 'to', 'read', 'a' and'book') A type is the class of all tokens containing the same character sequence. In the above example, there are only 5 types which are - 'can, 'you', 'a, 'as' and'canner' as'can', 'as' and'a' are being repeated. In the above example, by deleting period and hyphens between the characters and words we are normalizing the type by making it a term. So the term in the above example is: 'USA' and'antiinflammatory' Example: "Hello everyone.Welcome to the course." The tokens for the given sentence will be -- ['Hello','everyone', 'Welcome', 'to', 'the', 'course'] Welcome to the Natural Language Processing course.


Techniques for Training Large Neural Networks

#artificialintelligence

Large neural networks are at the core of many recent advances in AI, but training them is a difficult engineering and research challenge which requires orchestrating a cluster of GPUs to perform a single synchronized calculation. As cluster and model sizes have grown, machine learning practitioners have developed an increasing variety of techniques to parallelize model training over many GPUs. At first glance, understanding these parallelism techniques may seem daunting, but with only a few assumptions about the structure of the computation these techniques become much more clear--at that point, you're just shuttling around opaque bits from A to B like a network switch shuttles around packets. Each color refers to one layer and dashed lines separate different GPUs. Training a neural network is an iterative process.


If Your Company Isn't Good at Analytics, It's Not Ready for AI

#artificialintelligence

Management teams often assume they can leapfrog best practices for basic data analytics by going directly to adopting artificial intelligence and other advanced technologies. But companies that rush into sophisticated artificial intelligence before reaching a critical mass of automated processes and structured analytics can end up paralyzed. They can become saddled with expensive start-up partnerships, impenetrable black-box systems, cumbersome cloud computational clusters, and open-source toolkits without programmers to write code for them. By contrast, companies with strong basic analytics -- such as sales data and market trends -- make breakthroughs in complex and critical areas after layering in artificial intelligence. For example, one telecommunications company we worked with can now predict with 75 times more accuracy whether its customers are about to bolt using machine learning.


Everything You Ever Wanted to Know About Artificial Intelligence

#artificialintelligence

Artificial intelligence is overhyped--there, we said it. Superintelligent algorithms aren't about to take all the jobs or wipe out humanity. But software has gotten significantly smarter of late. It's why you can talk to your friends as an animated poop on the iPhone X using Apple's Animoji, or ask your smart speaker to order more paper towels. Tech companies' heavy investments in AI are already changing our lives and gadgets, and laying the groundwork for a more AI-centric future.


Everything You Ever Wanted to Know About Artificial Intelligence

#artificialintelligence

Artificial intelligence is overhyped--there, we said it. Superintelligent algorithms aren't about to take all the jobs or wipe out humanity. But software has gotten significantly smarter of late. It's why you can talk to your friends as an animated poop on the iPhone X using Apple's Animoji, or ask your smart speaker to order more paper towels. Tech companies' heavy investments in AI are already changing our lives and gadgets, and laying the groundwork for a more AI-centric future.


How an A.I. 'Cat-and-Mouse Game' Generates Believable Fake Photos

#artificialintelligence

The woman in the photo seems familiar. She looks like Jennifer Aniston, the "Friends" actress, or Selena Gomez, the child star turned pop singer. But not exactly. She appears to be a celebrity, one of the beautiful people photographed outside a movie premiere or an awards show. And yet, you cannot...


Progress in AI seems like it's accelerating, but here's why it could be plateauing

#artificialintelligence

I'm standing in what is soon to be the center of the world, or is perhaps just a very large room on the seventh floor of a gleaming tower in downtown Toronto. Showing me around is Jordan Jacobs, who cofounded this place: the nascent Vector Institute, which opens its doors this fall and which is aiming to become the global epicenter of artificial intelligence. We're in Toronto because Geoffrey Hinton is in Toronto, and Geoffrey Hinton is the father of "deep learning," the technique behind the current excitement about AI. "In 30 years we're going to look back and say Geoff is Einstein--of AI, deep learning, the thing that we're calling AI," Jacobs says. Of the researchers at the top of the field of deep learning, Hinton has more citations than the next three combined. His students and postdocs have gone on to run the AI labs at Apple, Facebook, and OpenAI; Hinton himself is a lead scientist on the Google Brain AI team.