Goto

Collaborating Authors

 ai pipeline


Generative AI Pipeline for Interactive Prompt-driven 2D-to-3D Vascular Reconstruction for Fontan Geometries from Contrast-Enhanced X-Ray Fluoroscopy Imaging

Menon, Prahlad G

arXiv.org Artificial Intelligence

Fontan palliation for univentricular congenital heart disease progresses to hemodynamic failure with complex flow patterns poorly characterized by conventional 2D imaging. Current assessment relies on fluoroscopic angiography, providing limited 3D geometric information essential for computational fluid dynamics (CFD) analysis and surgical planning. A multi-step AI pipeline was developed utilizing Google's Gemini 2.5 Flash (2.5B parameters) for systematic, iterative processing of fluoroscopic angiograms through transformer-based neural architecture. The pipeline encompasses medical image preprocessing, vascular segmentation, contrast enhancement, artifact removal, and virtual hemodynamic flow visualization within 2D projections. Final views were processed through Tencent's Hunyuan3D-2mini (384M parameters) for stereolithography file generation. The pipeline successfully generated geometrically optimized 2D projections from single-view angiograms after 16 processing steps using a custom web interface. Initial iterations contained hallucinated vascular features requiring iterative refinement to achieve anatomically faithful representations. Final projections demonstrated accurate preservation of complex Fontan geometry with enhanced contrast suitable for 3D conversion. AI-generated virtual flow visualization identified stagnation zones in central connections and flow patterns in branch arteries. Complete processing required under 15 minutes with second-level API response times. This approach demonstrates clinical feasibility of generating CFD-suitable geometries from routine angiographic data, enabling 3D generation and rapid virtual flow visualization for cursory insights prior to full CFD simulation. While requiring refinement cycles for accuracy, this establishes foundation for democratizing advanced geometric and hemodynamic analysis using readily available imaging data.


Visual Analytics for Explainable and Trustworthy Artificial Intelligence

Chatzimparmpas, Angelos

arXiv.org Artificial Intelligence

Our society increasingly depends on intelligent systems to solve complex problems, ranging from recommender systems suggesting the next movie to watch to AI models assisting in medical diagnoses for hospitalized patients. With the iterative improvement of diagnostic accuracy and efficiency, AI holds significant potential to mitigate medical misdiagnoses by preventing numerous deaths and reducing an economic burden of approximately 450 EUR billion annually. However, a key obstacle to AI adoption lies in the lack of transparency: many automated systems function as "black boxes," providing predictions without revealing the underlying processes. This opacity can hinder experts' ability to trust and rely on AI systems. Visual analytics (VA) provides a compelling solution by combining AI models with interactive visualizations. These specialized charts and graphs empower users to incorporate their domain expertise to refine and improve the models, bridging the gap between AI and human understanding. In this work, we define, categorize, and explore how VA solutions can foster trust across the stages of a typical AI pipeline. We propose a design space for innovative visualizations and present an overview of our previously developed VA dashboards, which support critical tasks within the various pipeline stages, including data processing, feature engineering, hyperparameter tuning, understanding, debugging, refining, and comparing models.


FAIRTOPIA: Envisioning Multi-Agent Guardianship for Disrupting Unfair AI Pipelines

Vakali, Athena, Dimitriadis, Ilias

arXiv.org Artificial Intelligence

AI models have become active decision makers, often acting without human supervision. The rapid advancement of AI technology has already caused harmful incidents that have hurt individuals and societies and AI unfairness in heavily criticized. It is urgent to disrupt AI pipelines which largely neglect human principles and focus on computational biases exploration at the data (pre), model(in), and deployment (post) processing stages. We claim that by exploiting the advances of agents technology, we will introduce cautious, prompt, and ongoing fairness watch schemes, under realistic, systematic, and human-centric fairness expectations. We envision agents as fairness guardians, since agents learn from their environment, adapt to new information, and solve complex problems by interacting with external tools and other systems. To set the proper fairness guardrails in the overall AI pipeline, we introduce a fairness-by-design approach which embeds multi-role agents in an end-to-end (human to AI) synergetic scheme. Our position is that we may design adaptive and realistic AI fairness frameworks, and we introduce a generalized algorithm which can be customized to the requirements and goals of each AI decision making scenario. Our proposed, so called FAIRTOPIA framework, is structured over a three-layered architecture, which encapsulates the AI pipeline inside an agentic guardian and a knowledge-based, self-refining layered scheme. Based on our proposition, we enact fairness watch in all of the AI pipeline stages, under robust multi-agent workflows, which will inspire new fairness research hypothesis, heuristics, and methods grounded in human-centric, systematic, interdisciplinary, socio-technical principles.


A Framework for Cryptographic Verifiability of End-to-End AI Pipelines

Balan, Kar, Learney, Robert, Wood, Tim

arXiv.org Artificial Intelligence

The increasing integration of Artificial Intelligence across multiple industry sectors necessitates robust mechanisms for ensuring transparency, trust, and auditability of its development and deployment. This topic is particularly important in light of recent calls in various jurisdictions to introduce regulation and legislation on AI safety. In this paper, we propose a framework for complete verifiable AI pipelines, identifying key components and analyzing existing cryptographic approaches that contribute to verifiability across different stages of the AI lifecycle, from data sourcing to training, inference, and unlearning. This framework could be used to combat misinformation by providing cryptographic proofs alongside AI-generated assets to allow downstream verification of their provenance and correctness. Our findings underscore the importance of ongoing research to develop cryptographic tools that are not only efficient for isolated AI processes, but that are efficiently `linkable' across different processes within the AI pipeline, to support the development of end-to-end verifiable AI technologies.


ALPACA -- Adaptive Learning Pipeline for Comprehensive AI

Torka, Simon, Albayrak, Sahin

arXiv.org Artificial Intelligence

In the rapidly evolving landscape of artificial intelligence (AI), the ability to comprehensively analyze and understand complex data is paramount. However, the ever-increasing complexity of real data requires sophisticated AI pipelines that seamlessly integrate different stages such as data collection, preparation, model generation, and evaluation. Such a pipeline can be illustrated as a chain of distinct yet interdependent stages, each contributing to the overarching goal of turning data into actionable intelligence. Like a well-coordinated symphony, these stages require harmonious collaboration to achieve optimal results. The concept of AI pipelines therefore represents more than just a linear progression; it signifies the orchestration of diverse processes to accomplish a larger purpose. At the core of this revolution lie AI pipelines, intricate networks of interconnected data processing and analysis steps designed to transform raw data into meaningful insights or outcomes using AI techniques. The evolution of simple AI models to adaptive, systematic AI pipelines has ushered in a new era of data-driven decision-making by solving complex tasks in an ever-changing environment. Making AI understandable, accessible and usable by everyone in every domain requires a domain-independent, easy-to-use pipeline architecture that can be integrated into a complex ecosystem of experts and non-experts. However, the design and implementation of such pipelines often prove challenging due to the intricate interplay of technical components and the diverse requirements of different application domains.


Reducing Hyperparameter Tuning Costs in ML, Vision and Language Model Training Pipelines via Memoization-Awareness

Essofi, Abdelmajid, Salahuddeen, Ridwan, Nwadike, Munachiso, Zhalieva, Elnura, Zhang, Kun, Xing, Eric, Neiswanger, Willie, Ho, Qirong

arXiv.org Machine Learning

The training or fine-tuning of machine learning, vision, and language models is often implemented as a pipeline: a sequence of stages encompassing data preparation, model training and evaluation. In this paper, we exploit pipeline structures to reduce the cost of hyperparameter tuning for model training/fine-tuning, which is particularly valuable for language models given their high costs in GPU-days. We propose a "memoization-aware" Bayesian Optimization (BO) algorithm, EEIPU, that works in tandem with a pipeline caching system, allowing it to evaluate significantly more hyperparameter candidates per GPU-day than other tuning algorithms. The result is better-quality hyperparameters in the same amount of search time, or equivalently, reduced search time to reach the same hyperparameter quality. In our benchmarks on machine learning (model ensembles), vision (convolutional architecture) and language (T5 architecture) pipelines, we compare EEIPU against recent BO algorithms: EEIPU produces an average of $103\%$ more hyperparameter candidates (within the same budget), and increases the validation metric by an average of $108\%$ more than other algorithms (where the increase is measured starting from the end of warm-up iterations).


AI for Low-Code for AI

Rao, Nikitha, Tsay, Jason, Kate, Kiran, Hellendoorn, Vincent J., Hirzel, Martin

arXiv.org Artificial Intelligence

Low-code programming allows citizen developers to create programs with minimal coding effort, typically via visual (e.g. drag-and-drop) interfaces. In parallel, recent AI-powered tools such as Copilot and ChatGPT generate programs from natural language instructions. We argue that these modalities are complementary: tools like ChatGPT greatly reduce the need to memorize large APIs but still require their users to read (and modify) programs, whereas visual tools abstract away most or all programming but struggle to provide easy access to large APIs. At their intersection, we propose LowCoder, the first low-code tool for developing AI pipelines that supports both a visual programming interface (LowCoder_VP) and an AI-powered natural language interface (LowCoder_NL). We leverage this tool to provide some of the first insights into whether and how these two modalities help programmers by conducting a user study. We task 20 developers with varying levels of AI expertise with implementing four ML pipelines using LowCoder, replacing the LowCoder_NL component with a simple keyword search in half the tasks. Overall, we find that LowCoder is especially useful for (i) Discoverability: using LowCoder_NL, participants discovered new operators in 75% of the tasks, compared to just 32.5% and 27.5% using web search or scrolling through options respectively in the keyword-search condition, and (ii) Iterative Composition: 82.5% of tasks were successfully completed and many initial pipelines were further successfully improved. Qualitative analysis shows that AI helps users discover how to implement constructs when they know what to do, but still fails to support novices when they lack clarity on what they want to accomplish. Overall, our work highlights the benefits of combining the power of AI with low-code programming.


Strategies for Optimizing End-to-End Artificial Intelligence Pipelines on Intel Xeon Processors

Arunachalam, Meena, Sanghavi, Vrushabh, Yao, Yi A, Zhou, Yi A, Wang, Lifeng A, Wen, Zongru, Ammbashankar, Niroop, Wang, Ning W, Mohammad, Fahim

arXiv.org Artificial Intelligence

End-to-end (E2E) artificial intelligence (AI) pipelines are composed of several stages including data preprocessing, data ingestion, defining and training the model, hyperparameter optimization, deployment, inference, postprocessing, followed by downstream analyses. To obtain efficient E2E workflow, it is required to optimize almost all the stages of pipeline. Intel Xeon processors come with large memory capacities, bundled with AI acceleration (e.g., Intel Deep Learning Boost), well suited to run multiple instances of training and inference pipelines in parallel and has low total cost of ownership (TCO). To showcase the performance on Xeon processors, we applied comprehensive optimization strategies coupled with software and hardware acceleration on variety of E2E pipelines in the areas of Computer Vision, NLP, Recommendation systems, etc. We were able to achieve a performance improvement, ranging from 1.8x to 81.7x across different E2E pipelines. In this paper, we will be highlighting the optimization strategies adopted by us to achieve this performance on Intel Xeon processors with a set of eight different E2E pipelines.


Google Cloud launches new AI-enabled imaging technologies

#artificialintelligence

Google Cloud on Tuesday announced Medical Imaging Suite, new technology it says can help with accessibility and interoperability of radiology and other imaging data. WHY IT MATTERS The new suite includes components focused on storage, lab, datasets, dashboards and AI pipelines for imaging, according to Google Cloud. It's designed to offer flexible options for cloud, on-prem or edge deployment to allow organizations to meet diverse sovereignty, data security, and privacy requirements, officials say, while providing centralized management and policy enforcement with Google Distributed Cloud, enabled by Anthos. Google's Cloud Healthcare API enables secure data exchange using the international DICOMweb standard for imaging and offers a scalable, enterprise-grade development environment and includes automated DICOM de-identification. Other technology partners include NetApp for seamless on-prem to cloud data management, and Change Healthcare's cloud-native enterprise imaging PACS.


IKEA launches AI-powered design experience (no Swedish meatballs included)

#artificialintelligence

We are excited to bring Transform 2022 back in-person July 19 and virtually July 20 - 28. Join AI and data leaders for insightful talks and exciting networking opportunities. For IKEA, the latest in digital transformation is all about home design driven by artificial intelligence (AI) – minus the home furnishing and decor retailer's famous Swedish meatballs. Today, it launched IKEA Kreativ, a design experience meant to bridge the ecommerce and in-store customer journeys, powered by the latest AI developments in spatial computing, machine learning and 3D mixed reality technologies. Available in-app and online, IKEA Kreativ's core technology was developed by Geomagical Labs, an IKEA retail company, which Ingka Group (the holding company that controls 367 stores of 422 IKEA stores) acquired in April 2020. IKEA Kreativ is the next step in IKEA's long journey towards digital transformation.