Goto

Collaborating Authors

 koala


First koala chlamydia vaccine approved

Popular Science

The disease has been plaguing the endangered marsupials since the 1990s. Breakthroughs, discoveries, and DIY tips sent every weekday. The first vaccine to protect endangered koalas from chlamydia has officially been approved in Australia. The vaccine was developed by the University of the Sunshine Coast's (UniSC) veterinary medicine division in eastern Australia and is another step towards ensuring the survival of the marsupial. The team spent over a decade developing the vaccine to protect the animals from the disease, which can cause urinary tract infections, blindness, infertility, and in some cases, death.


Secret koala population discovered near Australian city

Popular Science

Breakthroughs, discoveries, and DIY tips sent every weekday. When you think of koalas (Phascolarctos cinereus), chances are that words like cute or fluffy come to mind--not cryptic or stealthy. And yet, researchers in southeastern Australia have just discovered hundreds of previously undocumented koalas living surprisingly close to the city of Newcastle. The team conducted what they claim to be the largest and most accurate peer-reviewed koala survey to date. As detailed in a study published this month in the journal Biological Conversation, the survey estimates that a population of 4,357 koalas across 166,302 acres of land is living in the state of New South Wales.


KOALA: Enhancing Speculative Decoding for LLM via Multi-Layer Draft Heads with Adversarial Learning

Zhang, Kaiqi, Zhao, Jing, Chen, Rui

arXiv.org Artificial Intelligence

Large Language Models (LLMs) exhibit high inference latency due to their autoregressive decoding nature. While the draft head in speculative decoding mitigates this issue, its full potential remains unexplored. In this paper, we introduce KOALA (K-layer Optimized Adversarial Learning Architecture), an orthogonal approach to the draft head. By transforming the conventional single-layer draft head into a multi-layer architecture and incorporating adversarial learning into the traditional supervised training, KOALA significantly improves the accuracy of the draft head in predicting subsequent tokens, thus more closely mirroring the functionality of LLMs. Although this improvement comes at the cost of slightly increased drafting overhead, KOALA substantially unlocks the draft head's potential, greatly enhancing speculative decoding. We conducted comprehensive evaluations of KOALA, including both autoregressive and non-autoregressive draft heads across various tasks, demonstrating a latency speedup ratio improvement of 0.24x-0.41x, which is 10.57%-14.09% faster than the original draft heads.


Federated Knowledge Transfer Fine-tuning Large Server Model with Resource-Constrained IoT Clients

Chen, Shaoyuan, You, Linlin, Liu, Rui, Yu, Shuo, Abdelmoniem, Ahmed M.

arXiv.org Artificial Intelligence

The training of large models, involving fine-tuning, faces the scarcity of high-quality data. Compared to the solutions based on centralized data centers, updating large models in the Internet of Things (IoT) faces challenges in coordinating knowledge from distributed clients by using their private and heterogeneous data. To tackle such a challenge, we propose KOALA (Federated Knowledge Transfer Fine-tuning Large Server Model with Resource-Constrained IoT Clients) to impel the training of large models in IoT. Since the resources obtained by IoT clients are limited and restricted, it is infeasible to locally execute large models and also update them in a privacy-preserving manner. Therefore, we leverage federated learning and knowledge distillation to update large models through collaboration with their small models, which can run locally at IoT clients to process their private data separately and enable large-small model knowledge transfer through iterative learning between the server and clients. Moreover, to support clients with similar or different computing capacities, KOALA is designed with two kinds of large-small model joint learning modes, namely to be homogeneous or heterogeneous. Experimental results demonstrate that compared to the conventional approach, our method can not only achieve similar training performance but also significantly reduce the need for local storage and computing power resources.


Japanese-English Sentence Translation Exercises Dataset for Automatic Grading

Miura, Naoki, Funayama, Hiroaki, Kikuchi, Seiya, Matsubayashi, Yuichiroh, Iwase, Yuya, Inui, Kentaro

arXiv.org Artificial Intelligence

This paper proposes the task of automatic assessment of Sentence Translation Exercises (STEs), that have been used in the early stage of L2 language learning. We formalize the task as grading student responses for each rubric criterion pre-specified by the educators. We then create a dataset for STE between Japanese and English including 21 questions, along with a total of 3, 498 student responses (167 on average). The answer responses were collected from students and crowd workers. Using this dataset, we demonstrate the performance of baselines including finetuned BERT and GPT models with few-shot in-context learning. Experimental results show that the baseline model with finetuned BERT was able to classify correct responses with approximately 90% in F1, but only less than 80% for incorrect responses. Furthermore, the GPT models with few-shot learning show poorer results than finetuned BERT, indicating that our newly proposed task presents a challenging issue, even for the stateof-the-art large language models.


A Comparative Study of Open-Source Large Language Models, GPT-4 and Claude 2: Multiple-Choice Test Taking in Nephrology

Wu, Sean, Koo, Michael, Blum, Lesley, Black, Andy, Kao, Liyo, Scalzo, Fabien, Kurtz, Ira

arXiv.org Artificial Intelligence

In recent years, there have been significant breakthroughs in the field of natural language processing, particularly with the development of large language models (LLMs). These LLMs have showcased remarkable capabilities on various benchmarks. In the healthcare field, the exact role LLMs and other future AI models will play remains unclear. There is a potential for these models in the future to be used as part of adaptive physician training, medical co-pilot applications, and digital patient interaction scenarios. The ability of AI models to participate in medical training and patient care will depend in part on their mastery of the knowledge content of specific medical fields. This study investigated the medical knowledge capability of LLMs, specifically in the context of internal medicine subspecialty multiple-choice test-taking ability. We compared the performance of several open-source LLMs (Koala 7B, Falcon 7B, Stable-Vicuna 13B, and Orca Mini 13B), to GPT-4 and Claude 2 on multiple-choice questions in the field of Nephrology. Nephrology was chosen as an example of a particularly conceptually complex subspecialty field within internal medicine. The study was conducted to evaluate the ability of LLM models to provide correct answers to nephSAP (Nephrology Self-Assessment Program) multiple-choice questions. The overall success of open-sourced LLMs in answering the 858 nephSAP multiple-choice questions correctly was 17.1% - 25.5%. In contrast, Claude 2 answered 54.4% of the questions correctly, whereas GPT-4 achieved a score of 73.3%. We show that current widely used open-sourced LLMs do poorly in their ability for zero-shot reasoning when compared to GPT-4 and Claude 2. The findings of this study potentially have significant implications for the future of subspecialty medical training and patient care.


Can Instruction Fine-Tuned Language Models Identify Social Bias through Prompting?

Dige, Omkar, Tian, Jacob-Junqi, Emerson, David, Khattak, Faiza Khan

arXiv.org Artificial Intelligence

As the breadth and depth of language model applications continue to expand rapidly, it is increasingly important to build efficient frameworks for measuring and mitigating the learned or inherited social biases of these models. In this paper, we present our work on evaluating instruction fine-tuned language models' ability to identify bias through zero-shot prompting, including Chain-of-Thought (CoT) prompts. Across LLaMA and its two instruction fine-tuned versions, Alpaca 7B performs best on the bias identification task with an accuracy of 56.7%. We also demonstrate that scaling up LLM size and data diversity could lead to further performance gain. This is a work-in-progress presenting the first component of our bias mitigation framework. We will keep updating this work as we get more results.


Koala: A dialogue model for academic research

AIHub

In this post, we introduce Koala, a chatbot trained by fine-tuning Meta's LLaMA on dialogue data gathered from the web. We describe the dataset curation and training process of our model, and also present the results of a user study that compares our model to ChatGPT and Stanford's Alpaca. Our results show that Koala can effectively respond to a variety of user queries, generating responses that are often preferred over Alpaca, and at least tied with ChatGPT in over half of the cases. We hope that these results contribute further to the discourse around the relative performance of large closed-source models to smaller public models. In particular, it suggests that models that are small enough to be run locally can capture much of the performance of their larger cousins if trained on carefully sourced data.


Koala: An Index for Quantifying Overlaps with Pre-training Corpora

Vu, Thuy-Trang, He, Xuanli, Haffari, Gholamreza, Shareghi, Ehsan

arXiv.org Artificial Intelligence

In very recent years more attention has been placed on probing the role of pre-training data in Large Language Models (LLMs) downstream behaviour. Despite the importance, there is no public tool that supports such analysis of pre-training corpora at large scale. To help research in this space, we launch Koala, a searchable index over large pre-training corpora using compressed suffix arrays with highly efficient compression rate and search support. In its first release we index the public proportion of OPT 175B pre-training data. Koala provides a framework to do forensic analysis on the current and future benchmarks as well as to assess the degree of memorization in the output from the LLMs. Koala is available for public use at https://koala-index.erc.monash.edu/.


Flinders University scientists use biology from insects to build robots with a brain - ABC News

#artificialintelligence

Scientists at a South Australian university are using biology from insects to build robots with a brain – technology that could become a game changer for police, defence and national security. "I'm giving a robot a brain so it can understand its environment," said Flinders University associate professor for autonomous systems, Dr Russell Brinkworth. His biologically-inspired robots have the ability to not just take a picture of the world, but interpret the surrounding environment and adapt accordingly. "Our current robots work well in structured environments that don't change. That sounds complex – but they're all the same," Dr Brinkworth said.