General-Purpose Pre-Trained Models in Robotics
The impressive generalization capabilities of large neural network models hinge on the ability to integrate enormous quantities of training data. This presents a major challenge for most downstream tasks where data is scarce. As a result, we have seen a transformation over the years away from training large models entirely from scratch, and toward methods that utilize finetuning or few-shot learning. Classically, models might be pre-trained on a large-scale supervised or self-supervised task (e.g., pre-training a large ResNet model on ImageNet), and then the last few layers of the model might be fine-tuned on a much smaller dataset for the task of interest. More recently, open-vocabulary vision-language models and promptable language models have made it possible to avoid fine-tuning, and instead define new tasks by constructing a textual prompt, potentially containing a few examples of input-output pairs.
Nov-8-2022, 21:30:35 GMT
- Industry:
- Leisure & Entertainment > Sports > Tennis (0.70)
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning > Neural Networks (0.86)
- Natural Language (1.00)
- Robots (1.00)
- Information Technology > Artificial Intelligence