Sagare, Shivprasad
Audio-visual training for improved grounding in video-text LLMs
Sagare, Shivprasad, S, Hemachandran, Sarabhai, Kinshuk, Ullegaddi, Prashant, SA, Rajeshkumar
Recent advances in multimodal LLMs, have led to several video-text models being proposed for critical video-related tasks. However, most of the previous works support visual input only, essentially muting the audio signal in the video. Few models that support both audio and visual input, are not explicitly trained on audio data. Hence, the effect of audio towards video understanding is largely unexplored. To this end, we propose a model architecture that handles audio-visual inputs explicitly. We train our model with both audio and visual data from a video instruction-tuning dataset. Comparison with vision-only baselines, and other audiovisual models showcase that training on audio data indeed leads to improved grounding of responses. Figure 1: An example of improved grounding in the For better evaluation of audio-visual video-text LLM outputs, due to the additional audio models, we also release a human-annotated signal as input.
XWikiGen: Cross-lingual Summarization for Encyclopedic Text Generation in Low Resource Languages
Taunk, Dhaval, Sagare, Shivprasad, Patil, Anupam, Subramanian, Shivansh, Gupta, Manish, Varma, Vasudeva
Lack of encyclopedic text contributors, especially on Wikipedia, makes automated text generation for low resource (LR) languages a critical problem. Existing work on Wikipedia text generation has focused on English only where English reference articles are summarized to generate English Wikipedia pages. But, for low-resource languages, the scarcity of reference articles makes monolingual summarization ineffective in solving this problem. Hence, in this work, we propose XWikiGen, which is the task of cross-lingual multi-document summarization of text from multiple reference articles, written in various languages, to generate Wikipedia-style text. Accordingly, we contribute a benchmark dataset, XWikiRef, spanning ~69K Wikipedia articles covering five domains and eight languages. We harness this dataset to train a two-stage system where the input is a set of citations and a section title and the output is a section-specific LR summary. The proposed system is based on a novel idea of neural unsupervised extractive summarization to coarsely identify salient information followed by a neural abstractive model to generate the section-specific text. Extensive experiments show that multi-domain training is better than the multi-lingual setup on average.