Goto

Collaborating Authors

 licensee


EXAONE Path 2.0: Pathology Foundation Model with End-to-End Supervision

Pyeon, Myeongjang, Lee, Janghyeon, Lee, Minsoo, Yun, Juseung, Choi, Hwanil, Kim, Jonghyun, Kim, Jiwon, Hu, Yi, Jang, Jongseong, Lee, Soonyoung

arXiv.org Artificial Intelligence

In digital pathology, whole-slide images (WSIs) are often difficult to handle due to their gigapixel scale, so most approaches train patch encoders via self-supervised learning (SSL) and then aggregate the patch-level embeddings via multiple instance learning (MIL) or slide encoders for downstream tasks. However, patch-level SSL may overlook complex domain-specific features that are essential for biomarker prediction, such as mutation status and molecular characteristics, as SSL methods rely only on basic augmentations selected for natural image domains on small patch-level area. Moreover, SSL methods remain less data efficient than fully supervised approaches, requiring extensive computational resources and datasets to achieve competitive performance. To address these limitations, we present EXAONE Path 2.0, a pathology foundation model that learns patch-level representations under direct slide-level supervision. Using only 37k WSIs for training, EXAONE Path 2.0 achieves state-of-the-art average performance across 10 biomarker prediction tasks, demonstrating remarkable data efficiency.


EXAONE Deep: Reasoning Enhanced Language Models

Research, LG AI, Bae, Kyunghoon, Choi, Eunbi, Choi, Kibong, Choi, Stanley Jungkyu, Choi, Yemuk, Hong, Seokhee, Hwang, Junwon, Jeon, Hyojin, Jeon, Kijeong, Jo, Gerrard Jeongwon, Jo, Hyunjik, Jung, Jiyeon, Kim, Hyosang, Kim, Joonkee, Kim, Seonghwan, Kim, Soyeon, Kim, Sunkyoung, Kim, Yireun, Kim, Yongil, Kim, Youchul, Lee, Edward Hwayoung, Lee, Haeju, Lee, Honglak, Lee, Jinsik, Lee, Kyungmin, Park, Sangha, Park, Yongmin, Yang, Sihoon, Yeen, Heuiyeen, Yi, Sihyuk, Yun, Hyeongu

arXiv.org Artificial Intelligence

We present EXAONE Deep series, which exhibits superior capabilities in various reasoning tasks, including math and coding benchmarks. We train our models mainly on the reasoning-specialized dataset that incorporates long streams of thought processes. Evaluation results show that our smaller models, EXAONE Deep 2.4B and 7.8B, outperform other models of comparable size, while the largest model, EXAONE Deep 32B, demonstrates competitive performance against leading open-weight models. All EXAONE Deep models are openly available for research purposes and can be downloaded from https://huggingface.co/LGAI-EXAONE


EXAONE 3.5: Series of Large Language Models for Real-world Use Cases

Research, LG AI, An, Soyoung, Bae, Kyunghoon, Choi, Eunbi, Choi, Kibong, Choi, Stanley Jungkyu, Hong, Seokhee, Hwang, Junwon, Jeon, Hyojin, Jo, Gerrard Jeongwon, Jo, Hyunjik, Jung, Jiyeon, Jung, Yountae, Kim, Hyosang, Kim, Joonkee, Kim, Seonghwan, Kim, Soyeon, Kim, Sunkyoung, Kim, Yireun, Kim, Yongil, Kim, Youchul, Lee, Edward Hwayoung, Lee, Haeju, Lee, Honglak, Lee, Jinsik, Lee, Kyungmin, Lim, Woohyung, Park, Sangha, Park, Sooyoun, Park, Yongmin, Yang, Sihoon, Yeen, Heuiyeen, Yun, Hyeongu

arXiv.org Artificial Intelligence

This technical report introduces the EXAONE 3.5 instruction-tuned language models, developed and released by LG AI Research. The EXAONE 3.5 language models are offered in three configurations: 32B, 7.8B, and 2.4B. These models feature several standout capabilities: 1) exceptional instruction following capabilities in real-world scenarios, achieving the highest scores across seven benchmarks, 2) outstanding long-context comprehension, attaining the top performance in four benchmarks, and 3) competitive results compared to state-of-the-art open models of similar sizes across nine general benchmarks. The EXAONE 3.5 language models are open to anyone for research purposes and can be downloaded from https://huggingface.co/LGAI-EXAONE. For commercial use, please reach out to the official contact point of LG AI Research: contact_us@lgresearch.ai.


EXAONEPath 1.0 Patch-level Foundation Model for Pathology

Yun, Juseung, Hu, Yi, Kim, Jinhyung, Jang, Jongseong, Lee, Soonyoung

arXiv.org Artificial Intelligence

Recent advancements in digital pathology have led to the development of numerous foundational models that utilize self-supervised learning on patches extracted from gigapixel whole slide images (WSIs). While this approach leverages vast amounts of unlabeled data, we have discovered a significant issue: features extracted from these self-supervised models tend to cluster by individual WSIs, a phenomenon we term WSI-specific feature collapse. This problem can potentially limit the model's generalization ability and performance on various downstream tasks. To address this issue, we introduce EXAONEPath, a novel foundational model trained on patches that have undergone stain normalization. Stain normalization helps reduce color variability arising from different laboratories and scanners, enabling the model to learn more consistent features. EXAONEPath is trained using 285,153,903 patches extracted from a total of 34,795 WSIs. Our experiments demonstrate that EXAONEPath significantly mitigates the feature collapse problem, indicating that the model has learned more generalized features rather than overfitting to individual WSI characteristics. We compared EXAONEPath with state-of-the-art models across six downstream task datasets, and our results show that EXAONEPath achieves superior performance relative to the number of WSIs used and the model's parameter count. This suggests that the application of stain normalization has substantially improved the model's efficiency and generalization capabilities.


EXAONE 3.0 7.8B Instruction Tuned Language Model

Research, LG AI, :, null, An, Soyoung, Bae, Kyunghoon, Choi, Eunbi, Choi, Stanley Jungkyu, Choi, Yemuk, Hong, Seokhee, Hong, Yeonjung, Hwang, Junwon, Jeon, Hyojin, Jo, Gerrard Jeongwon, Jo, Hyunjik, Jung, Jiyeon, Jung, Yountae, Kim, Euisoon, Kim, Hyosang, Kim, Joonkee, Kim, Seonghwan, Kim, Soyeon, Kim, Sunkyoung, Kim, Yireun, Kim, Youchul, Lee, Edward Hwayoung, Lee, Haeju, Lee, Honglak, Lee, Jinsik, Lee, Kyungmin, Lee, Moontae, Lee, Seungjun, Lim, Woohyung, Park, Sangha, Park, Sooyoun, Park, Yongmin, Seo, Boseong, Yang, Sihoon, Yeen, Heuiyeen, Yoo, Kyungjae, Yun, Hyeongu

arXiv.org Artificial Intelligence

We introduce EXAONE 3.0 instruction-tuned language model, the first open model in the family of Large Language Models (LLMs) developed by LG AI Research. Among different model sizes, we publicly release the 7.8B instruction-tuned model to promote open research and innovations. Through extensive evaluations across a wide range of public and in-house benchmarks, EXAONE 3.0 demonstrates highly competitive real-world performance with instruction-following capability against other state-of-the-art open models of similar size. Our comparative analysis shows that EXAONE 3.0 excels particularly in Korean, while achieving compelling performance across general tasks and complex reasoning. With its strong real-world effectiveness and bilingual proficiency, we hope that EXAONE keeps contributing to advancements in Expert AI. Our EXAONE 3.0 instruction-tuned model is available at https://huggingface.co/LGAI-EXAONE/EXAONE-3.0-7.8B-Instruct


Minimalist Grammar: Construction without Overgeneration

Maier, Isidor Konrad, Kuhn, Johannes, Beisegel, Jesse, Huber-Liebl, Markus, Wolff, Matthias

arXiv.org Artificial Intelligence

In this paper we give instructions on how to write a minimalist grammar (MG). In order to present the instructions as an algorithm, we use a variant of context free grammars (CFG) as an input format. We can exclude overgeneration, if the CFG has no recursion, i.e. no non-terminal can (indirectly) derive to a right-hand side containing itself. The constructed MGs utilize licensors/-ees as a special way of exception handling. A CFG format for a derivation $A\_eats\_B\mapsto^* peter\_eats\_apples$, where $A$ and $B$ generate noun phrases, normally leads to overgeneration, e.\,g., $i\_eats\_apples$. In order to avoid overgeneration, a CFG would need many non-terminal symbols and rules, that mainly produce the same word, just to handle exceptions. In our MGs however, we can summarize CFG rules that produce the same word in one item and handle exceptions by a proper distribution of licensees/-ors. The difficulty with this technique is that in most generations the majority of licensees/-ors is not needed, but still has to be triggered somehow. We solve this problem with $\epsilon$-items called \emph{adapters}.


Home

#artificialintelligence

With cutting edge technologies such as computer vision, blockchain and machine learning, Atlas Navi is set to become the leading navigation choice for drivers around the world. Atlas Navi is a navigation app that uses A.I. and the smartphone camera to avoid traffic by detecting road conditions, accidents, traffic in each lane, police vehicles and many more. It analyses the road 25 times per second which generates 100 times better data than other navigation apps; other drivers are routed to avoid areas with problems. Purchase Brand Ambassador NFTs for lifetime royalties when users buy NFT cars from that brand. Brand Ambassador positions are VERY RARE and will likely increase in value.


Artificial Intelligence Briefing: Feds Take Aim at Algorithmic Bias

#artificialintelligence

The Federal Trade Commission delivered a report to Congress warning about the use of artificial intelligence to combat online harms. The June 16 report lays out the FTC's latest thinking on AI, and any organization that uses algorithmic decision-making in a way that impacts consumers should take heed. Key takeaways include: The importance (and limitations) of having a human in the loop. The need for AI to be "meaningfully transparent, which includes the need for it to be explainable and contestable, especially when people's rights are involved or when personal data is being collected or used." Companies that use AI "must be accountable both for their data practices and for their results" and should consider independent audits and algorithmic impact assessments.


Nvidia data center sales grew 55% on demand for artificial intelligence chips

#artificialintelligence

Kress said customers are using the chips for tasks such as understanding human speech and crunching data to offer customer recommendations. Gaming, Nvidia's biggest market, reported $3.2 billion in sales, up 42% from $2.27 billion in the same quarter last year. The company said it was primarily due to increased sales of its GeForce consumer graphics processors, but the company said supply remained limited. Nvidia's gaming graphics cards now have software that prevents them from being used for cryptocurrency mining, the company said. Nvidia introduced dedicated graphics cards for crypto mining earlier this year to help meet some of the demand.


How NVIDIA's Arm acquisition will drive AI to every edge

#artificialintelligence

NVIDIA is sitting pretty in AI (artificial intelligence) right now. For the next few years, most AI systems will continue to be trained on NVIDIA GPUs and specialized hardware and cloud services that incorporate these processors. However, NVIDIA has been frustrated in its attempts to become a dominant provider of AI chips for deployment into smartphones, embedded systems, and other edge devices. To address that strategic gap, NVIDIA this past week announced that it is acquiring processor architecture firm Arm Holdings from SoftBank Group and the SoftBank Vision Fund. Once the acquisition closes in the expected 18 months, NVIDIA will retain Arm's name, brand identity, management team, and base of operations in Cambridge, United Kingdom.