extrusion
VideoCAD: A Dataset and Model for Learning Long-Horizon 3D CAD UI Interactions from Video
Man, Brandon, Nehme, Ghadi, Alam, Md Ferdous, Ahmed, Faez
Computer-Aided Design (CAD) is a time-consuming and complex process, requiring precise, long-horizon user interactions with intricate 3D interfaces. While recent advances in AI-driven user interface (UI) agents show promise, most existing datasets and methods focus on short, low-complexity tasks in mobile or web applications, failing to capture the demands of professional engineering tools. In this work, we introduce VideoCAD, the first attempt to model UI interactions for precision engineering tasks. Specifically, VideoCAD is a large-scale synthetic dataset consisting of over 41K annotated video recordings of CAD operations, generated using an automated framework for collecting high-fidelity UI action data from human-made CAD designs. Compared to existing datasets, VideoCAD offers an order-of-magnitude increase in complexity for real-world engineering UI tasks, with time horizons up to 20x longer than those in other datasets. We show two important downstream applications of VideoCAD: (1) learning UI interactions from professional 3D CAD tools for precision tasks and (2) a visual question-answering (VQA) benchmark designed to evaluate multimodal large language models (LLMs) on spatial reasoning and video understanding. To learn the UI interactions, we propose VideoCADFormer, a state-of-the-art model for learning CAD interactions directly from video, which outperforms existing behavior cloning baselines. Both VideoCADFormer and the VQA benchmark derived from VideoCAD reveal key challenges in the current state of video-based UI understanding, including the need for precise action grounding, multi-modal and spatial reasoning, and long-horizon dependencies.
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.14)
- North America > Canada > Manitoba > Westman Region > Brandon (0.04)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
- Information Technology > Artificial Intelligence > Vision (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
Defect Mitigation for Robot Arm-based Additive Manufacturing Utilizing Intelligent Control and IOT
Ali, Matsive, Gassen, Blake, Liu, Sen
This paper presents an integrated robotic fused deposition modeling additive manufacturing system featuring closed-loop thermal control and intelligent in-situ defect correction using a 6-degree of freedom robotic arm and an Oak-D camera. The robot arm end effector was modified to mount an E3D hotend thermally regulated by an IoT microcontroller, enabling precise temperature control through real-time feedback. Filament extrusion system was synchronized with robotic motion, coordinated via ROS2, ensuring consistent deposition along complex trajectories. A vision system based on OpenCV detects layer-wise defects position, commanding autonomous re-extrusion at identified sites. Experimental validation demonstrated successful defect mitigation in printing operations. The integrated system effectively addresses challenges real-time quality assurance. Inverse kinematics were used for motion planning, while homography transformations corrected camera perspectives for accurate defect localization. The intelligent system successfully mitigated surface anomalies without interrupting the print process. By combining real-time thermal regulation, motion control, and intelligent defect detection & correction, this architecture establishes a scalable and adaptive robotic additive manufacturing framework suitable for aerospace, biomedical, and industrial applications.
- North America > United States > Tennessee > Shelby County > Memphis (0.40)
- North America > United States > Louisiana > Lafayette Parish > Lafayette (0.04)
CAD-Tokenizer: Towards Text-based CAD Prototyping via Modality-Specific Tokenization
Wang, Ruiyu, Sun, Shizhao, Ma, Weijian, Bian, Jiang
Computer-Aided Design (CAD) is a foundational component of industrial prototyping, where models are defined not by raw coordinates but by construction sequences such as sketches and extrusions. This sequential structure enables both efficient prototype initialization and subsequent editing. Text-guided CAD prototyping, which unifies Text-to-CAD generation and CAD editing, has the potential to streamline the entire design pipeline. However, prior work has not explored this setting, largely because standard large language model (LLM) tokenizers decompose CAD sequences into natural-language word pieces, failing to capture primitive-level CAD semantics and hindering attention modules from modeling geometric structure. We conjecture that a multimodal tokenization strategy, aligned with CAD's primitive and structural nature, can provide more effective representations. To this end, we propose CAD-Tokenizer, a framework that represents CAD data with modality-specific tokens using a sequence-based VQ-VAE with primitive-level pooling and constrained decoding. This design produces compact, primitive-aware representations that align with CAD's structural nature. Applied to unified text-guided CAD prototyping, CAD-Tokenizer significantly improves instruction following and generation quality, achieving better quantitative and qualitative performance over both general-purpose LLMs and task-specific baselines.
- North America > Canada > Ontario > Toronto (0.14)
- Asia > Middle East > Qatar > Ad-Dawhah > Doha (0.04)
3D Extended Object Tracking based on Extruded B-Spline Side View Profiles
Han, Longfei, Kefferpütz, Klaus, Beyerer, Jürgen
Object tracking is an essential task for autonomous systems. With the advancement of 3D sensors, these systems can better perceive their surroundings using effective 3D Extended Object Tracking (EOT) methods. Based on the observation that common road users are symmetrical on the right and left sides in the traveling direction, we focus on the side view profile of the object. In order to leverage of the development in 2D EOT and balance the number of parameters of a shape model in the tracking algorithms, we propose a method for 3D extended object tracking (EOT) by describing the side view profile of the object with B-spline curves and forming an extrusion to obtain a 3D extent. The use of B-spline curves exploits their flexible representation power by allowing the control points to move freely. The algorithm is developed into an Extended Kalman Filter (EKF). For a through evaluation of this method, we use simulated traffic scenario of different vehicle models and realworld open dataset containing both radar and lidar data.
Toward Fully Autonomous Flexible Chunk-Based Aerial Additive Manufacturing: Insights from Experimental Validation
Stamatopoulos, Marios-Nektarios, Haluska, Jakub, Small, Elias, Marroush, Jude, Banerjee, Avijit, Nikolakopoulos, George
A novel autonomous chunk-based aerial additive manufacturing framework is presented, supported with experimental demonstration advancing aerial 3D printing. An optimization-based decomposition algorithm transforms structures into sub-components, or chunks, treated as individual tasks coordinated via a dependency graph, ensuring sequential assignment to UA Vs considering inter-dependencies and printability constraints for seamless execution. A specially designed hexacopter equipped with a pressurized canister for lightweight expandable foam extrusion is utilized to deposit the material in a controlled manner. To further enhance precise execution of the printing, an offset-free Model Predictive Control mechanism is considered compensating reactively for disturbances and ground effect during execution. Additionally, an interlocking mechanism is introduced in the chunking process to enhance structural cohesion and improve layer adhesion. Extensive experiments demonstrate the framework's effectiveness in constructing precise structures of various shapes, while seamlessly adapting to practical challenges, proving its potential for a transformative leap in aerial robotic capability for autonomous construction. A video with the overall demonstration can be found here: https://youtu.be/WC1rLMLKEg4. Preprint submitted to Journal of Automation In Construction February 27, 2025 1. Introduction In recent times, ground breaking advancement in additive manufacturing, seamlessly integrated with autonomous robotics, are unlocking an exciting frontier in next generation construction and manufacturing process. Additive manufacturing has demonstrated a paradigm shift impact, addressing complex manufacturing processes with unprecedented precision and efficiency. Its transformative potential is becoming increasingly evident as it evolves and finds applications across a wide range of industries [1, 2, 3], while simultaneously paving the way for further innovations in the future. An intriguing development is its recent integration into the construction industry, capitalizing on its ability to automate construction processes, provide extensive design flexibility, and construct intricate structures designed using Computer-Aided Design (CAD) software [4, 5]. Numerous studies have demonstrated the design and deployment of large-scale robotic arms and gantry systems for printing building components and even entire houses using a variety of base materials [6]. A key advantage of such methods is their ability to adapt with high level of automation throughout the construction process, making them particularly well-suited for deployment in remote, inaccessible, and harsh environments[7, 8]. Notable examples include disaster-stricken areas, such as regions impacted by fires and earthquakes, where the rapid construction of shelters and basic infrastructure is imperative.
- Research Report > New Finding (0.74)
- Research Report > Promising Solution (0.48)
- Machinery > Industrial Machinery (1.00)
- Construction & Engineering (1.00)
- Energy > Oil & Gas > Upstream (0.66)
- Information Technology > Artificial Intelligence > Robots (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Search (0.46)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Agents (0.46)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Planning & Scheduling (0.46)
FlexCAD: Unified and Versatile Controllable CAD Generation with Fine-tuned Large Language Models
Zhang, Zhanwei, Sun, Shizhao, Wang, Wenxiao, Cai, Deng, Bian, Jiang
Recently, there is a growing interest in creating computer-aided design (CAD) models based on user intent, known as controllable CAD generation. Existing work offers limited controllability and needs separate models for different types of control, reducing efficiency and practicality. To achieve controllable generation across all CAD construction hierarchies, such as sketch-extrusion, extrusion, sketch, face, loop and curve, we propose FlexCAD, a unified model by fine-tuning large language models (LLMs). First, to enhance comprehension by LLMs, we represent a CAD model as a structured text by abstracting each hierarchy as a sequence of text tokens. Second, to address various controllable generation tasks in a unified model, we introduce a hierarchy-aware masking strategy. Specifically, during training, we mask a hierarchy-aware field in the CAD text with a mask token. This field, composed of a sequence of tokens, can be set flexibly to represent various hierarchies. Subsequently, we ask LLMs to predict this masked field. During inference, the user intent is converted into a CAD text with a mask token replacing the part the user wants to modify, which is then fed into FlexCAD to generate new CAD models. Comprehensive experiments on public dataset demonstrate the effectiveness of FlexCAD in both generation quality and controllability. Code will be available at https://github.com/microsoft/CADGeneration/FlexCAD.
- North America > United States > New York > New York County > New York City (0.04)
- Asia (0.04)
Robotic Barrier Construction through Weaved, Inflatable Tubes
Kim, H. J., Abdel-Raziq, H., Liu, X., Siskovic, A. Y., Patil, S., Petersen, K. H., Kao, H. L.
In this article, we present a mechanism and related path planning algorithm to construct light-duty barriers out of extruded, inflated tubes weaved around existing environmental features. Our extruded tubes are based on everted vine-robots and in this context, we present a new method to steer their growth. We characterize the mechanism in terms of accuracy resilience, and, towards their use as barriers, the ability of the tubes to withstand distributed loads. We further explore an algorithm which, given a feature map and the size and direction of the external load, can determine where and how to extrude the barrier. Finally, we showcase the potential of this method in an autonomously extruded two-layer wall weaved around three pipes. While preliminary, our work indicates that this method has the potential for barrier construction in cluttered environments, e.g. shelters against wind or snow. Future work may show how to achieve tighter weaves, how to leverage weave friction for improved strength, how to assess barrier performance for feedback control, and how to operate the extrusion mechanism off of a mobile robot.
Reconstructing editable prismatic CAD from rounded voxel models
Lambourne, Joseph G., Willis, Karl D. D., Jayaraman, Pradeep Kumar, Zhang, Longfei, Sanghi, Aditya, Malekshan, Kamal Rahimi
Reverse Engineering a CAD shape from other representations is an important geometric processing step for many downstream applications. In this work, we introduce a novel neural network architecture to solve this challenging task and approximate a smoothed signed distance function with an editable, constrained, prismatic CAD model. During training, our method reconstructs the input geometry in the voxel space by decomposing the shape into a series of 2D profile images and 1D envelope functions. These can then be recombined in a differentiable way allowing a geometric loss function to be defined. During inference, we obtain the CAD data by first searching a database of 2D constrained sketches to find curves which approximate the profile images, then extrude them and use Boolean operations to build the final CAD model. Our method approximates the target shape more closely than other methods and outputs highly editable constrained parametric sketches which are compatible with existing CAD software.
- Asia > South Korea > Daegu > Daegu (0.05)
- North America > United States > New York > New York County > New York City (0.04)
- North America > Canada > Quebec > Montreal (0.04)
- (3 more...)
Kartta Labs: Collaborative Time Travel
Tavakkol, Sasan, Han, Feng, Mayer, Brandon, Phillips, Mark, Shahabi, Cyrus, Chiang, Yao-Yi, Kiveris, Raimondas
We introduce the modular and scalable design of Kartta Labs, an open source, open data, and scalable system for virtually reconstructing cities from historical maps and photos. Kartta Labs relies on crowdsourcing and artificial intelligence consisting of two major modules: Maps and 3D models. Each module, in turn, consists of sub-modules that enable the system to reconstruct a city from historical maps and photos. The result is a spatiotemporal reference that can be used to integrate various collected data (curated, sensed, or crowdsourced) for research, education, and entertainment purposes. The system empowers the users to experience collaborative time travel such that they work together to reconstruct the past and experience it on an open source and open data platform.
- North America > United States > New York > New York County > New York City (0.04)
- North America > United States > Hawaii (0.04)
- North America > United States > California > Los Angeles County > Los Angeles (0.04)
- (3 more...)
- Information Technology > Services (0.69)
- Health & Medicine (0.68)
- Education (0.68)
Future Of Healthcare Through Deep Learning & 3D-Printed Organoids
Organoids 3D printing has quickly become one of the leading segments of the 3D printing industry in terms of innovation. Until recently, the market was primarily focused on North America, however many companies, laboratories, and universities around the world are exploring this field as well. Thanks to 3D printing techniques, cells and biomaterials can be combined and deposited layer by layer to create biomedical developments that have the same properties as living tissues. During this process, various bio-links can be used to create these tissue-like structures, which have applications in the fields of medical and tissue engineering. Of course, it is more than knowing that the goal of all these developments is to successfully bioprint a fully functional human organ.
- North America > United States > Texas (0.05)
- North America > United States > Pennsylvania (0.05)
- Machinery > Industrial Machinery (0.94)
- Health & Medicine > Health Care Technology (0.59)