Goto

Collaborating Authors

 succession


A Framework for Processing Textual Descriptions of Business Processes using a Constrained Language -- Technical Report

Burattin, Andrea, Grama, Antonio, Sima, Ana-Maria, Rivkin, Andrey, Weber, Barbara

arXiv.org Artificial Intelligence

This report explores how (potentially constrained) natural language can be used to enable non-experts to develop process models by simply describing scenarios in plain text. To this end, a framework, called BeePath, is proposed. It allows users to write process descriptions in a constrained pattern-based language, which can then be translated into formal models such as Petri nets and DECLARE. The framework also leverages large language models (LLMs) to help convert unstructured descriptions into this constrained language.


HyperCLOVA X THINK Technical Report

NAVER Cloud HyperCLOVA X Team, null

arXiv.org Artificial Intelligence

We introduce HyperCLOVA X THINK, the first reasoning-focused large language model in the HyperCLOVA X family, pre-trained on roughly $6$ trillion high-quality Korean, and English tokens, augmented with targeted synthetic Korean data. It was implemented as a compute-memory-balanced Peri-LN Transformer scaled with $μ$P, pre-trained through a three-stage curriculum that expands the context window to $128$K tokens, and post-trained via supervised fine-tuning with Reinforcement Learning from Verifiable Rewards supports both detailed rationale and concise-answer modes. It delivers competitive performance against similarly sized models on Korea-focused benchmarks such as KMMLU, CSAT, KoBALT-700, HAERAE-1.0, and KoBigBench, while preserving robust bilingual consistency and translation quality. In addition, a vision-augmented variant matches or exceeds GPT-4.1 on the KCSAT STEM benchmark, all of which are achieved with substantially lower training compute than existing models of similar sizes. We also present a pruning and distillation technique that will soon be applied to HyperCLOVA X THINK for an open-source and business-friendly foundation model. Altogether, these capabilities position HyperCLOVA X THINK as a robust foundation for Korean AI innovation and a valuable resource for the global research community.


"Mountainhead" Channels the Absurdity of the Tech Bro

The New Yorker

Four tech billionaires walk into a mansion. It sounds like the setup for a punch line, but it also forms nearly the entire conceit behind "Mountainhead," a savagely entertaining but somewhat shallow new satire written and directed by Jesse Armstrong, the creator of "Succession." The film, which is streaming on HBO's Max, is a sort of chamber play, its stage a modernist castle in Utah--the Mountainhead of the title--overlooking snowy peaks. The players are a quartet of friends, or, more accurately, frenemies, who resemble a mishmash of real-world Silicon Valley founders. Steve Carell plays Randall Garrett, the group's Peter Thiel-esque mentor who, not unlike the late Steve Jobs, has cancer that his doctor tells him is incurable.


The Creator of em Succession /em Is Back With a Movie. There's a Reason He Rushed to Make It Right Away.

Slate

Outside an opulent retreat in the mountains of Utah, the world is going to hell. Thanks to disinformation-spreading tools on the world's largest social media platform, people are being executed by bloodthirsty mobs and machine-gunned by their neighbors, politicians assassinated and governments crumbling. But inside Mountainhead, the billionaire tech moguls responsible for the chaos are smoking cigars and shooting the breeze, debating whether the eruption of global chaos is a crisis to be managed or a surge of "creative destruction" that will help usher humanity into a brighter future. If the fictional setting of Mountainhead, the debut feature by Jesse Armstrong, seems a little too close to reality, that's because it's meant to be. The movie, which stars Steve Carell, Jason Schwartzman, Ramy Youssef, and Cory Michael Smith, was conceived, written, cast, shot, edited, and released in about six months, an astonishingly short timeline for any director, let alone a first-timer.

  Country:
  Genre: Personal (0.47)
  Industry:

The New Movie From the Creator of em Succession /em Is Less a Satire Than a Documentary

Slate

For the quartet of tech billionaires in Jesse Armstrong's Mountainhead, ideas are so powerful that nothing else seems real. Holed up in a resplendent snowy retreat built by meditation-app developer Hugo Van Yalk (Jason Schwartzman), they're glued to their phones as the outside world is erupting into chaos, thanks in no small part to the wildfire spread of A.I. deepfakes on the social media platform owned by the world's richest man, Venis Parish (Cory Michael Smith). People in Gujarat are being burned alive after being falsely accused of desecrating religious symbols, and Midwestern Americans are machine-gunning each other over minor disagreements, but for these four men, the widespread devastation is in some ways proof of concept that they're as important as they believe themselves to be. And besides, those bodies going up in flames are just images on a tiny screen, so distant they might as well be theoretical. As he trudges through the snow with Randall (Steve Carell), the venture capitalist who serves as the group's self-appointed philosopher king, Venis asks him, "Do you … believe in other people?"


It's the End of the World (And It's Their Fault)

The Atlantic - Technology

It's late morning on a Monday in March and I am, for reasons I will explain momentarily, in a private bowling alley deep in the bowels of a 65 million mansion in Utah. Jesse Armstrong, the showrunner of HBO's hit series Succession, approaches me, monitor headphones around his neck and a wide grin on his face. "I take it you've seen the news," he says, flashing his phone and what appears to be his X feed in my direction. Everyone had: An hour earlier, my boss Jeffrey Goldberg had published a story revealing that U.S. national-security leaders had accidentally added him to a Signal group chat where they discussed their plans to conduct then-upcoming military strikes in Yemen. "Incredibly fucking depressing," Armstrong said.


OmAgent: A Multi-modal Agent Framework for Complex Video Understanding with Task Divide-and-Conquer

Zhang, Lu, Zhao, Tiancheng, Ying, Heting, Ma, Yibo, Lee, Kyusong

arXiv.org Artificial Intelligence

Recent advancements in Large Language Models (LLMs) have expanded their capabilities to multimodal contexts, including comprehensive video understanding. However, processing extensive videos such as 24-hour CCTV footage or full-length films presents significant challenges due to the vast data and processing demands. Traditional methods, like extracting key frames or converting frames to text, often result in substantial information loss. To address these shortcomings, we develop OmAgent, efficiently stores and retrieves relevant video frames for specific queries, preserving the detailed content of videos. Additionally, it features an Divide-and-Conquer Loop capable of autonomous reasoning, dynamically invoking APIs and tools to enhance query processing and accuracy. This approach ensures robust video understanding, significantly reducing information loss. Experimental results affirm OmAgent's efficacy in handling various types of videos and complex tasks. Moreover, we have endowed it with greater autonomy and a robust tool-calling system, enabling it to accomplish even more intricate tasks.


Generating Feasible and Plausible Counterfactual Explanations for Outcome Prediction of Business Processes

Stevens, Alexander, Ouyang, Chun, De Smedt, Johannes, Moreira, Catarina

arXiv.org Artificial Intelligence

In recent years, various machine and deep learning architectures have been successfully introduced to the field of predictive process analytics. Nevertheless, the inherent opacity of these algorithms poses a significant challenge for human decision-makers, hindering their ability to understand the reasoning behind the predictions. This growing concern has sparked the introduction of counterfactual explanations, designed as human-understandable what if scenarios, to provide clearer insights into the decision-making process behind undesirable predictions. The generation of counterfactual explanations, however, encounters specific challenges when dealing with the sequential nature of the (business) process cases typically used in predictive process analytics. Our paper tackles this challenge by introducing a data-driven approach, REVISEDplus, to generate more feasible and plausible counterfactual explanations. First, we restrict the counterfactual algorithm to generate counterfactuals that lie within a high-density region of the process data, ensuring that the proposed counterfactuals are realistic and feasible within the observed process data distribution. Additionally, we ensure plausibility by learning sequential patterns between the activities in the process cases, utilising Declare language templates. Finally, we evaluate the properties that define the validity of counterfactuals.


Byte Pair Encoding for Symbolic Music

Fradet, Nathan, Gutowski, Nicolas, Chhel, Fabien, Briot, Jean-Pierre

arXiv.org Artificial Intelligence

When used with deep learning, the symbolic music modality is often coupled with language model architectures. To do so, the music needs to be tokenized, i.e. converted into a sequence of discrete tokens. This can be achieved by different approaches, as music can be composed of simultaneous tracks, of simultaneous notes with several attributes. Until now, the proposed tokenizations rely on small vocabularies of tokens describing the note attributes and time events, resulting in fairly long token sequences, and a sub-optimal use of the embedding space of language models. Recent research has put efforts on reducing the overall sequence length by merging embeddings or combining tokens. In this paper, we show that Byte Pair Encoding, a compression technique widely used for natural language, significantly decreases the sequence length while increasing the vocabulary size. By doing so, we leverage the embedding capabilities of such models with more expressive tokens, resulting in both better results and faster inference in generation and classification tasks. The source code is shared on Github, along with a companion website. Finally, BPE is directly implemented in MidiTok, allowing the reader to easily benefit from this method.


About some compression algorithms

Lecian, Orchidea Maria, Tirozzi, Brunello

arXiv.org Artificial Intelligence

We use neural network algorithms for finding compression methods of images in the framework of iterated function systems which is a collection of the transformations of the interval $(0, 1)$ satisfying suitable properties.