Goto

Collaborating Authors

 symbolic distillation


Symbolic Distillation for Learned TCP Congestion Control

Neural Information Processing Systems

Recent advances in TCP congestion control (CC) have achieved tremendous success with deep reinforcement learning (RL) approaches, which use feedforward neural networks (NN) to learn complex environment conditions and make better decisions. However, such ``black-box'' policies lack interpretability and reliability, and often, they need to operate outside the traditional TCP datapath due to the use of complex NNs. This paper proposes a novel two-stage solution to achieve the best of both worlds: first to train a deep RL agent, then distill its (over-)parameterized NN policy into white-box, light-weight rules in the form of symbolic expressions that are much easier to understand and to implement in constrained environments. At the core of our proposal is a novel symbolic branching algorithm that enables the rule to be aware of the context in terms of various network conditions, eventually converting the NN policy into a symbolic tree. The distilled symbolic rules preserve and often improve performance over state-of-the-art NN policies while being faster and simpler than a standard neural network. We validate the performance of our distilled symbolic rules on both simulation and emulation environments.


Appendix: Symbolic Distillation for Learned TCP Congestion Control S P Sharan

Neural Information Processing Systems

We now specify how we build the DRL behavior dataset and process into a symbolic regression friendly format. It is an indicator of the population of genetic programs' performances. The fitness metric driving our evolution is simply the MSE between the predicted action and the "expert" action (teacher model's action). We specifically follow 5 different evolution schemes, either one picked stochastically. This mutant variant carries forth genetic material from both its sources.


Appendix: Symbolic Distillation for Learned TCP Congestion Control S P Sharan

Neural Information Processing Systems

We now specify how we build the DRL behavior dataset and process into a symbolic regression friendly format. It is an indicator of the population of genetic programs' performances. The fitness metric driving our evolution is simply the MSE between the predicted action and the "expert" action (teacher model's action). We specifically follow 5 different evolution schemes, either one picked stochastically. This mutant variant carries forth genetic material from both its sources.


Symbolic Distillation for Learned TCP Congestion Control

Neural Information Processing Systems

Recent advances in TCP congestion control (CC) have achieved tremendous success with deep reinforcement learning (RL) approaches, which use feedforward neural networks (NN) to learn complex environment conditions and make better decisions. However, such black-box'' policies lack interpretability and reliability, and often, they need to operate outside the traditional TCP datapath due to the use of complex NNs. This paper proposes a novel two-stage solution to achieve the best of both worlds: first to train a deep RL agent, then distill its (over-)parameterized NN policy into white-box, light-weight rules in the form of symbolic expressions that are much easier to understand and to implement in constrained environments. At the core of our proposal is a novel symbolic branching algorithm that enables the rule to be aware of the context in terms of various network conditions, eventually converting the NN policy into a symbolic tree. The distilled symbolic rules preserve and often improve performance over state-of-the-art NN policies while being faster and simpler than a standard neural network. We validate the performance of our distilled symbolic rules on both simulation and emulation environments.


MetaIE: Distilling a Meta Model from LLM for All Kinds of Information Extraction Tasks

Peng, Letian, Wang, Zilong, Yao, Feng, Wang, Zihan, Shang, Jingbo

arXiv.org Artificial Intelligence

Information extraction (IE) is a fundamental area in natural language processing where prompting large language models (LLMs), even with in-context examples, cannot defeat small LMs tuned on very small IE datasets. We observe that IE tasks, such as named entity recognition and relation extraction, all focus on extracting important information, which can be formalized as a label-to-span matching. In this paper, we propose a novel framework MetaIE to build a small LM as meta-model by learning to extract "important information", i.e., the meta-understanding of IE, so that this meta-model can be adapted to all kind of IE tasks effectively and efficiently. Specifically, MetaIE obtains the small LM via a symbolic distillation from an LLM following the label-to-span scheme. We construct the distillation dataset via sampling sentences from language model pre-training datasets (e.g., OpenWebText in our implementation) and prompting an LLM to identify the typed spans of "important information". We evaluate the meta-model under the few-shot adaptation setting. Extensive results on 13 datasets from 6 IE tasks confirm that MetaIE can offer a better starting point for few-shot tuning on IE datasets and outperform other meta-models from (1) vanilla language model pre-training, (2) multi-IE-task pre-training with human annotations, and (3) single-IE-task symbolic distillation from LLM. Moreover, we provide comprehensive analyses of MetaIE, such as the size of the distillation dataset, the meta-model architecture, and the size of the meta-model.