Goto

Collaborating Authors

 deepseek


What's next for Chinese open-source AI

MIT Technology Review

Chinese open models are spreading fast, from Hugging Face to Silicon Valley. In this photo illustration, the DeepSeek apps is seen on a phone in front of a flag of China on January 28, 2025 in Hong Kong, China. The past year has marked a turning point for Chinese AI. Since DeepSeek released its R1 reasoning model in January 2025, Chinese companies have repeatedly delivered AI models that match the performance of leading Western models at a fraction of the cost. Just last week the Chinese firm Moonshot AI released its latest open-weight model, Kimi K2.5, which came close to top proprietary systems such as Anthropic's Claude Opus on some early benchmarks. The difference: K2.5 is roughly one-seventh Opus's price.


6 Graphs That Show Where the U.S. Leads China on AI--and Where It Doesn't

TIME - Tech

Two important things happened on January 20, 2025. In Washington, D.C., Donald Trump was inaugurated as President of the United States. In Hangzhou, China, a little-known Chinese firm called DeepSeek released R1, an AI model that industry watchers called a "Sputnik moment" for the country's AI industry. "Whether we like it or not, we're suddenly engaged in a fast-paced competition to build and define this groundbreaking technology that will determine so much about the future of civilization," said Trump later that year, as he announced his administration's AI action plan, which was titled "Winning the Race." There are many interpretations of what AI companies and their governments are racing towards, says AI policy researcher Lennart Heim: to deploy AI systems in the economy, to build robots, to create human-like artificial general intelligence.


How China Caught Up on AI--and May Now Win the Future

TIME - Tech

He Xiaopeng launches Xpeng's next-gen Iron humanoid robot during a press conference at the company's headquarters in Guangzhou on November 5, 2025. He Xiaopeng launches Xpeng's next-gen Iron humanoid robot during a press conference at the company's headquarters in Guangzhou on November 5, 2025. It was a controversy laced with pride for He Xiaopeng. In November, He, the founder and CEO of Chinese physical AI firm XPeng, had just debuted his new humanoid robot, IRON, whose balance, posture shifts, and coquettish swagger mirrored human motion with such eerie precision that a slew of netizens accused him of faking the demonstration by putting a human in a bodysuit. To silence the naysayers, He boldly cut open the robot's leg live on stage to reveal the intricate mechanical systems that allow it to adapt to uneven surfaces and maintain stability just like the human body. "At first, it made me sad," He tells TIME in his Guangzhou headquarters.


Thousands of Companies Are Driving China's AI Boom. A Government Registry Tracks Them All

WIRED

Thousands of Companies Are Driving China's AI Boom. How the Cyberspace Administration of China inadvertently made a guide to the country's homegrown AI revolution. When DeepSeek burst onto the global stage in January 2025, it seemed to appear out of nowhere. But the large language model was just one of the thousands of generative AI tools that have been released in China since 2023--and there's a public archive of every single one of them. Here are 23 ways China is rewiring the future .


The Race to Build the DeepSeek of Europe Is On

WIRED

As Europe's longstanding alliance with the US falters, its push to become a self-sufficient AI superpower has become more urgent. As the relationship between the US and its European allies shows signs of strain, AI labs across the continent are searching for inventive ways to close the gap with American rivals that have so far dominated the field. With rare exceptions, US-based firms outstrip European competitors across the AI production line--from processor design and manufacturing, to datacenter capacity, to model and application development. Likewise, the US has captured a massive proportion of the money pouring into AI, reflected in the performance last year of its homegrown stocks and the growth of its econonmy . The belief in some quarters is that the US-based leaders --Nvidia, Google, Meta, OpenAI, Anthropic, and the like--are already so entrenched as to make it impossible for European nations to break their dependency on American AI, mirroring the pattern in cloud services.


So Long, GPT-5. Hello, Qwen

WIRED

In the AI boom, chatbots and GPTs come and go quickly. On a drizzly and windswept afternoon this summer, I visited the headquarters of Rokid, a startup developing smart glasses in Hangzhou, China. As I chatted with engineers, their words were swiftly translated from Mandarin to English, and then transcribed onto a tiny translucent screen just above my right eye using one of the company's new prototype devices. Rokid's high-tech spectacles use Qwen, an open-weight large language model developed by the Chinese ecommerce giant Alibaba. OpenAI's GPT-5, Google's Gemini 3, and Anthropic's Claude often score higher on benchmarks designed to gauge different dimensions of machine cleverness.


The Download: why 2025 has been the year of AI hype correction, and fighting GPS jamming

MIT Technology Review

When OpenAI released a free web app called ChatGPT in late 2022, it changed the course of an entire industry--and several world economies. Millions of people started talking to their computers, and their computers started talking back. We were enchanted, and we expected more. Well, 2025 has been a year of reckoning. For a start, the heads of the top AI companies made promises they couldn't keep. At the same time, updates to the core technology are no longer the step changes they once were.


Edge Deployment of Small Language Models, a comprehensive comparison of CPU, GPU and NPU backends

Prieto, Pablo, Abad, Pablo

arXiv.org Artificial Intelligence

Edge computing processes data where it is generated, enabling faster decisions, lower bandwidth usage, and improved privacy. However, edge devices typically operate under strict constraints on processing power, memory, and energy consumption, making them unsuitable for large language models (LLMs). Fortunately, Small Language Models (SLMs) offer lightweight alternatives that bring AI inference to resource-constrained environments by significantly reducing computational cost while remaining suitable for specialization and customization. In this scenario, selecting the hardware platform that best balances performance and efficiency for SLM inference is challenging due to strict resource limitations. To address this issue, this study evaluates the inference performance and energy efficiency of commercial CPUs (Intel and ARM), GPUs (NVIDIA), and NPUs (RaiderChip) for running SLMs. GPUs, the usual platform of choice, are compared against commercial NPUs and recent multi-core CPUs. While NPUs leverage custom hardware designs optimized for computation, modern CPUs increasingly incorporate dedicated features targeting language-model workloads. Using a common execution framework and a suite of state-of-the-art SLMs, we analyze both maximum achievable performance and processing and energy efficiency across commercial solutions available for each platform. The results indicate that specialized backends outperform general-purpose CPUs, with NPUs achieving the highest performance by a wide margin. Bandwidth normalization proves essential for fair cross-architecture comparisons. Although low-power ARM processors deliver competitive results when energy usage is considered, metrics that combine performance and power (such as EDP) again highlight NPUs as the dominant architecture. These findings show that designs optimized for both efficiency and performance offer a clear advantage for edge workloads.


Large Language Models for Education and Research: An Empirical and User Survey-based Analysis

Rahman, Md Mostafizer, Shiplu, Ariful Islam, Amin, Md Faizul Ibne, Watanobe, Yutaka, Peng, Lu

arXiv.org Artificial Intelligence

Pretrained Large Language Models (LLMs) have achieved remarkable success across diverse domains, with education and research emerging as particularly impactful areas. Among current state-of-the-art LLMs, ChatGPT and DeepSeek exhibit strong capabilities in mathematics, science, medicine, literature, and programming. In this study, we present a comprehensive evaluation of these two LLMs through background technology analysis, empirical experiments, and a real-world user survey. The evaluation explores trade-offs among model accuracy, computational efficiency, and user experience in educational and research affairs. We benchmarked these LLMs performance in text generation, programming, and specialized problem-solving. Experimental results show that ChatGPT excels in general language understanding and text generation, while DeepSeek demonstrates superior performance in programming tasks due to its efficiency-focused design. Moreover, both models deliver medically accurate diagnostic outputs and effectively solve complex mathematical problems. Complementing these quantitative findings, a survey of students, educators, and researchers highlights the practical benefits and limitations of these models, offering deeper insights into their role in advancing education and research.


Explainable Melanoma Diagnosis with Contrastive Learning and LLM-based Report Generation

Zheng, Junwen, Xu, Xinran, Wang, Li Rong, Cai, Chang, Tan, Lucinda Siyun, Wang, Dingyuan, Tey, Hong Liang, Fan, Xiuyi

arXiv.org Artificial Intelligence

Deep learning has demonstrated expert-level performance in melanoma classification, positioning it as a powerful tool in clinical dermatology. However, model opacity and the lack of interpretability remain critical barriers to clinical adoption, as clinicians often struggle to trust the decision-making processes of black-box models. To address this gap, we present a Cross-modal Explainable Framework for Melanoma (CEFM) that leverages contrastive learning as the core mechanism for achieving interpretability. Specifically, CEFM maps clinical criteria for melanoma diagnosis-namely Asymmetry, Border, and Color (ABC)-into the Vision Transformer embedding space using dual projection heads, thereby aligning clinical semantics with visual features. The aligned representations are subsequently translated into structured textual explanations via natural language generation, creating a transparent link between raw image data and clinical interpretation. Experiments on public datasets demonstrate 92.79% accuracy and an AUC of 0.961, along with significant improvements across multiple interpretability metrics. Qualitative analyses further show that the spatial arrangement of the learned embeddings aligns with clinicians' application of the ABC rule, effectively bridging the gap between high-performance classification and clinical trust.