licensee
EXAONE Path 2.0: Pathology Foundation Model with End-to-End Supervision
Pyeon, Myeongjang, Lee, Janghyeon, Lee, Minsoo, Yun, Juseung, Choi, Hwanil, Kim, Jonghyun, Kim, Jiwon, Hu, Yi, Jang, Jongseong, Lee, Soonyoung
In digital pathology, whole-slide images (WSIs) are often difficult to handle due to their gigapixel scale, so most approaches train patch encoders via self-supervised learning (SSL) and then aggregate the patch-level embeddings via multiple instance learning (MIL) or slide encoders for downstream tasks. However, patch-level SSL may overlook complex domain-specific features that are essential for biomarker prediction, such as mutation status and molecular characteristics, as SSL methods rely only on basic augmentations selected for natural image domains on small patch-level area. Moreover, SSL methods remain less data efficient than fully supervised approaches, requiring extensive computational resources and datasets to achieve competitive performance. To address these limitations, we present EXAONE Path 2.0, a pathology foundation model that learns patch-level representations under direct slide-level supervision. Using only 37k WSIs for training, EXAONE Path 2.0 achieves state-of-the-art average performance across 10 biomarker prediction tasks, demonstrating remarkable data efficiency.
- North America > United States (0.14)
- Asia > South Korea > Seoul > Seoul (0.04)
- Health & Medicine > Therapeutic Area > Oncology (1.00)
- Health & Medicine > Diagnostic Medicine (0.83)
EXAONE Deep: Reasoning Enhanced Language Models
Research, LG AI, Bae, Kyunghoon, Choi, Eunbi, Choi, Kibong, Choi, Stanley Jungkyu, Choi, Yemuk, Hong, Seokhee, Hwang, Junwon, Jeon, Hyojin, Jeon, Kijeong, Jo, Gerrard Jeongwon, Jo, Hyunjik, Jung, Jiyeon, Kim, Hyosang, Kim, Joonkee, Kim, Seonghwan, Kim, Soyeon, Kim, Sunkyoung, Kim, Yireun, Kim, Yongil, Kim, Youchul, Lee, Edward Hwayoung, Lee, Haeju, Lee, Honglak, Lee, Jinsik, Lee, Kyungmin, Park, Sangha, Park, Yongmin, Yang, Sihoon, Yeen, Heuiyeen, Yi, Sihyuk, Yun, Hyeongu
We present EXAONE Deep series, which exhibits superior capabilities in various reasoning tasks, including math and coding benchmarks. We train our models mainly on the reasoning-specialized dataset that incorporates long streams of thought processes. Evaluation results show that our smaller models, EXAONE Deep 2.4B and 7.8B, outperform other models of comparable size, while the largest model, EXAONE Deep 32B, demonstrates competitive performance against leading open-weight models. All EXAONE Deep models are openly available for research purposes and can be downloaded from https://huggingface.co/LGAI-EXAONE
- Asia > South Korea > Seoul > Seoul (0.04)
- Asia > North Korea > Hwanghae-namdo > Haeju (0.04)
- Law (1.00)
- Information Technology (0.68)
EXAONE 3.5: Series of Large Language Models for Real-world Use Cases
Research, LG AI, An, Soyoung, Bae, Kyunghoon, Choi, Eunbi, Choi, Kibong, Choi, Stanley Jungkyu, Hong, Seokhee, Hwang, Junwon, Jeon, Hyojin, Jo, Gerrard Jeongwon, Jo, Hyunjik, Jung, Jiyeon, Jung, Yountae, Kim, Hyosang, Kim, Joonkee, Kim, Seonghwan, Kim, Soyeon, Kim, Sunkyoung, Kim, Yireun, Kim, Yongil, Kim, Youchul, Lee, Edward Hwayoung, Lee, Haeju, Lee, Honglak, Lee, Jinsik, Lee, Kyungmin, Lim, Woohyung, Park, Sangha, Park, Sooyoun, Park, Yongmin, Yang, Sihoon, Yeen, Heuiyeen, Yun, Hyeongu
This technical report introduces the EXAONE 3.5 instruction-tuned language models, developed and released by LG AI Research. The EXAONE 3.5 language models are offered in three configurations: 32B, 7.8B, and 2.4B. These models feature several standout capabilities: 1) exceptional instruction following capabilities in real-world scenarios, achieving the highest scores across seven benchmarks, 2) outstanding long-context comprehension, attaining the top performance in four benchmarks, and 3) competitive results compared to state-of-the-art open models of similar sizes across nine general benchmarks. The EXAONE 3.5 language models are open to anyone for research purposes and can be downloaded from https://huggingface.co/LGAI-EXAONE. For commercial use, please reach out to the official contact point of LG AI Research: contact_us@lgresearch.ai.
- Asia > Vietnam > Long An Province (0.05)
- North America > United States > California > San Francisco County > San Francisco (0.04)
- Asia > Middle East > Jordan (0.04)
- (9 more...)
- Overview (0.46)
- Research Report (0.40)
- Law (1.00)
- Government (1.00)
- Information Technology > Security & Privacy (0.67)
EXAONEPath 1.0 Patch-level Foundation Model for Pathology
Yun, Juseung, Hu, Yi, Kim, Jinhyung, Jang, Jongseong, Lee, Soonyoung
Recent advancements in digital pathology have led to the development of numerous foundational models that utilize self-supervised learning on patches extracted from gigapixel whole slide images (WSIs). While this approach leverages vast amounts of unlabeled data, we have discovered a significant issue: features extracted from these self-supervised models tend to cluster by individual WSIs, a phenomenon we term WSI-specific feature collapse. This problem can potentially limit the model's generalization ability and performance on various downstream tasks. To address this issue, we introduce EXAONEPath, a novel foundational model trained on patches that have undergone stain normalization. Stain normalization helps reduce color variability arising from different laboratories and scanners, enabling the model to learn more consistent features. EXAONEPath is trained using 285,153,903 patches extracted from a total of 34,795 WSIs. Our experiments demonstrate that EXAONEPath significantly mitigates the feature collapse problem, indicating that the model has learned more generalized features rather than overfitting to individual WSI characteristics. We compared EXAONEPath with state-of-the-art models across six downstream task datasets, and our results show that EXAONEPath achieves superior performance relative to the number of WSIs used and the model's parameter count. This suggests that the application of stain normalization has substantially improved the model's efficiency and generalization capabilities.
- Asia > South Korea > Seoul > Seoul (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- Asia > China > Guangxi Province > Nanning (0.04)
- Law (1.00)
- Health & Medicine > Therapeutic Area > Oncology (1.00)
- Health & Medicine > Diagnostic Medicine > Imaging (0.68)
- Information Technology > Artificial Intelligence > Machine Learning > Inductive Learning (0.89)
- Information Technology > Artificial Intelligence > Machine Learning > Unsupervised or Indirectly Supervised Learning (0.66)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.47)
EXAONE 3.0 7.8B Instruction Tuned Language Model
Research, LG AI, :, null, An, Soyoung, Bae, Kyunghoon, Choi, Eunbi, Choi, Stanley Jungkyu, Choi, Yemuk, Hong, Seokhee, Hong, Yeonjung, Hwang, Junwon, Jeon, Hyojin, Jo, Gerrard Jeongwon, Jo, Hyunjik, Jung, Jiyeon, Jung, Yountae, Kim, Euisoon, Kim, Hyosang, Kim, Joonkee, Kim, Seonghwan, Kim, Soyeon, Kim, Sunkyoung, Kim, Yireun, Kim, Youchul, Lee, Edward Hwayoung, Lee, Haeju, Lee, Honglak, Lee, Jinsik, Lee, Kyungmin, Lee, Moontae, Lee, Seungjun, Lim, Woohyung, Park, Sangha, Park, Sooyoun, Park, Yongmin, Seo, Boseong, Yang, Sihoon, Yeen, Heuiyeen, Yoo, Kyungjae, Yun, Hyeongu
We introduce EXAONE 3.0 instruction-tuned language model, the first open model in the family of Large Language Models (LLMs) developed by LG AI Research. Among different model sizes, we publicly release the 7.8B instruction-tuned model to promote open research and innovations. Through extensive evaluations across a wide range of public and in-house benchmarks, EXAONE 3.0 demonstrates highly competitive real-world performance with instruction-following capability against other state-of-the-art open models of similar size. Our comparative analysis shows that EXAONE 3.0 excels particularly in Korean, while achieving compelling performance across general tasks and complex reasoning. With its strong real-world effectiveness and bilingual proficiency, we hope that EXAONE keeps contributing to advancements in Expert AI. Our EXAONE 3.0 instruction-tuned model is available at https://huggingface.co/LGAI-EXAONE/EXAONE-3.0-7.8B-Instruct
- North America > United States (0.04)
- Europe > Spain > Catalonia > Barcelona Province > Barcelona (0.04)
- Asia > South Korea > Seoul > Seoul (0.04)
- (4 more...)
- Law (1.00)
- Information Technology > Security & Privacy (1.00)
Minimalist Grammar: Construction without Overgeneration
Maier, Isidor Konrad, Kuhn, Johannes, Beisegel, Jesse, Huber-Liebl, Markus, Wolff, Matthias
In this paper we give instructions on how to write a minimalist grammar (MG). In order to present the instructions as an algorithm, we use a variant of context free grammars (CFG) as an input format. We can exclude overgeneration, if the CFG has no recursion, i.e. no non-terminal can (indirectly) derive to a right-hand side containing itself. The constructed MGs utilize licensors/-ees as a special way of exception handling. A CFG format for a derivation $A\_eats\_B\mapsto^* peter\_eats\_apples$, where $A$ and $B$ generate noun phrases, normally leads to overgeneration, e.\,g., $i\_eats\_apples$. In order to avoid overgeneration, a CFG would need many non-terminal symbols and rules, that mainly produce the same word, just to handle exceptions. In our MGs however, we can summarize CFG rules that produce the same word in one item and handle exceptions by a proper distribution of licensees/-ors. The difficulty with this technique is that in most generations the majority of licensees/-ors is not needed, but still has to be triggered somehow. We solve this problem with $\epsilon$-items called \emph{adapters}.
- Europe > Germany > Brandenburg > Potsdam (0.04)
- North America > United States > Texas > Borden County (0.04)
- North America > United States > Oregon > Multnomah County > Portland (0.04)
- (2 more...)
Home
With cutting edge technologies such as computer vision, blockchain and machine learning, Atlas Navi is set to become the leading navigation choice for drivers around the world. Atlas Navi is a navigation app that uses A.I. and the smartphone camera to avoid traffic by detecting road conditions, accidents, traffic in each lane, police vehicles and many more. It analyses the road 25 times per second which generates 100 times better data than other navigation apps; other drivers are routed to avoid areas with problems. Purchase Brand Ambassador NFTs for lifetime royalties when users buy NFT cars from that brand. Brand Ambassador positions are VERY RARE and will likely increase in value.
- Information Technology (1.00)
- Transportation > Ground > Road (0.31)
Artificial Intelligence Briefing: Feds Take Aim at Algorithmic Bias
The Federal Trade Commission delivered a report to Congress warning about the use of artificial intelligence to combat online harms. The June 16 report lays out the FTC's latest thinking on AI, and any organization that uses algorithmic decision-making in a way that impacts consumers should take heed. Key takeaways include: The importance (and limitations) of having a human in the loop. The need for AI to be "meaningfully transparent, which includes the need for it to be explainable and contestable, especially when people's rights are involved or when personal data is being collected or used." Companies that use AI "must be accountable both for their data practices and for their results" and should consider independent audits and algorithmic impact assessments.
- North America > United States > Indiana > Marion County > Indianapolis (0.07)
- North America > United States > California > Los Angeles County > Los Angeles (0.06)
- North America > United States > District of Columbia (0.05)
- Law (1.00)
- Government > Regional Government > North America Government > United States Government (1.00)
- Banking & Finance (1.00)
- Health & Medicine (0.98)
Nvidia data center sales grew 55% on demand for artificial intelligence chips
Kress said customers are using the chips for tasks such as understanding human speech and crunching data to offer customer recommendations. Gaming, Nvidia's biggest market, reported $3.2 billion in sales, up 42% from $2.27 billion in the same quarter last year. The company said it was primarily due to increased sales of its GeForce consumer graphics processors, but the company said supply remained limited. Nvidia's gaming graphics cards now have software that prevents them from being used for cryptocurrency mining, the company said. Nvidia introduced dedicated graphics cards for crypto mining earlier this year to help meet some of the demand.
- Information Technology > Hardware (1.00)
- Information Technology > Artificial Intelligence (1.00)
- Information Technology > e-Commerce > Financial Technology (0.97)
- Information Technology > Graphics (0.85)
How NVIDIA's Arm acquisition will drive AI to every edge
NVIDIA is sitting pretty in AI (artificial intelligence) right now. For the next few years, most AI systems will continue to be trained on NVIDIA GPUs and specialized hardware and cloud services that incorporate these processors. However, NVIDIA has been frustrated in its attempts to become a dominant provider of AI chips for deployment into smartphones, embedded systems, and other edge devices. To address that strategic gap, NVIDIA this past week announced that it is acquiring processor architecture firm Arm Holdings from SoftBank Group and the SoftBank Vision Fund. Once the acquisition closes in the expected 18 months, NVIDIA will retain Arm's name, brand identity, management team, and base of operations in Cambridge, United Kingdom.
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.25)
- North America > United States (0.05)
- Europe > Switzerland (0.05)
- Asia > China (0.05)
- Information Technology > Artificial Intelligence (1.00)
- Information Technology > Communications > Mobile (0.38)