contour
Revisiting 3D Object Detection From an Egocentric Perspective
For these applications, we care most about how the detections affect the ego-agent's behavior and safety (the egocentric perspective). Intuitively, we seek more accurate descriptions of object geometry when it's more likely to interfere with the ego-agent's motion trajectory. However, current detection metrics, based on box Intersection-over-Union (IoU), are object-centric and aren't designed to capture the spatio-temporal relationship between objects and the ego-agent. To address this issue, we propose a new egocentric measure to evaluate 3D object detection, namely Support Distance Error (SDE). Our analysis based on SDE reveals that the egocentric detection quality is bounded by the coarse geometry of the bounding boxes. Given the insight that SDE would benefit from more accurate geometry descriptions, we propose to represent objects as amodal contours, specifically amodal star-shaped polygons, and devise a simple model, StarPoly, to predict such contours. Our experiments on the large-scale Waymo Open Dataset show that SDE better reflects the impact of detection quality on the ego-agent's safety compared to IoU; and the estimated contours from StarPoly consistently improve the egocentric detection quality over recent 3D object detectors.
ShadowDraw: From Any Object to Shadow-Drawing Compositional Art
Luo, Rundong, Snavely, Noah, Ma, Wei-Chiu
We introduce ShadowDraw, a framework that transforms ordinary 3D objects into shadow-drawing compositional art. Given a 3D object, our system predicts scene parameters, including object pose and lighting, together with a partial line drawing, such that the cast shadow completes the drawing into a recognizable image. To this end, we optimize scene configurations to reveal meaningful shadows, employ shadow strokes to guide line drawing generation, and adopt automatic evaluation to enforce shadow-drawing coherence and visual quality. Experiments show that ShadowDraw produces compelling results across diverse inputs, from real-world scans and curated datasets to generative assets, and naturally extends to multi-object scenes, animations, and physical deployments. Our work provides a practical pipeline for creating shadow-drawing art and broadens the design space of computational visual art, bridging the gap between algorithmic design and artistic storytelling. Check out our project page https://red-fairy.github.io/ShadowDraw/ for more results and an end-to-end real-world demonstration of our pipeline!
Automatic Multi-View X-Ray/CT Registration Using Bone Substructure Contours
Flepp, Roman, Nissen, Leon, Sigrist, Bastian, Nieuwland, Arend, Cavalcanti, Nicola, Fürnstahl, Philipp, Dreher, Thomas, Calvet, Lilian
Purpose: Accurate intraoperative X-ray/CT registration is essential for surgical navigation in orthopedic procedures. However, existing methods struggle with consistently achieving sub-millimeter accuracy, robustness under broad initial pose estimates or need manual key-point annotations. This work aims to address these challenges by proposing a novel multi-view X-ray/CT registration method for intraoperative bone registration. Methods: The proposed registration method consists of a multi-view, contour-based iterative closest point (ICP) optimization. Unlike previous methods, which attempt to match bone contours across the entire silhouette in both imaging modalities, we focus on matching specific subcategories of contours corresponding to bone substructures. This leads to reduced ambiguity in the ICP matches, resulting in a more robust and accurate registration solution. This approach requires only two X-ray images and operates fully automatically. Additionally, we contribute a dataset of 5 cadaveric specimens, including real X-ray images, X-ray image poses and the corresponding CT scans. Results: The proposed registration method is evaluated on real X-ray images using mean reprojection error (mRPD). The method consistently achieves sub-millimeter accuracy with a mRPD 0.67mm compared to 5.35mm by a commercial solution requiring manual intervention. Furthermore, the method offers improved practical applicability, being fully automatic. Conclusion: Our method offers a practical, accurate, and efficient solution for multi-view X-ray/CT registration in orthopedic surgeries, which can be easily combined with tracking systems. By improving registration accuracy and minimizing manual intervention, it enhances intraoperative navigation, contributing to more accurate and effective surgical outcomes in computer-assisted surgery (CAS).
- Health & Medicine > Therapeutic Area (1.00)
- Health & Medicine > Diagnostic Medicine > Imaging (1.00)
Three-Dimensional Anatomical Data Generation Based on Artificial Neural Networks
Müller, Ann-Sophia, Jeong, Moonkwang, Zhang, Meng, Tian, Jiyuan, Miernik, Arkadiusz, Speidel, Stefanie, Qiu, Tian
Surgical planning and training based on machine learning requires a large amount of 3D anatomical models reconstructed from medical imaging, which is currently one of the major bottlenecks. Obtaining these data from real patients and during surgery is very demanding, if even possible, due to legal, ethical, and technical challenges. It is especially difficult for soft tissue organs with poor imaging contrast, such as the prostate. To overcome these challenges, we present a novel workflow for automated 3D anatomical data generation using data obtained from physical organ models. We additionally use a 3D Generative Adversarial Network (GAN) to obtain a manifold of 3D models useful for other downstream machine learning tasks that rely on 3D data. We demonstrate our workflow using an artificial prostate model made of biomimetic hydrogels with imaging contrast in multiple zones. This is used to physically simulate endoscopic surgery. For evaluation and 3D data generation, we place it into a customized ultrasound scanner that records the prostate before and after the procedure. A neural network is trained to segment the recorded ultrasound images, which outperforms conventional, non-learning-based computer vision techniques in terms of intersection over union (IoU). Based on the segmentations, a 3D mesh model is reconstructed, and performance feedback is provided.
- North America > United States (0.05)
- Europe > Germany > Saxony > Dresden (0.05)
- Europe > Germany > Baden-Württemberg > Freiburg (0.04)
- (2 more...)
- Workflow (0.74)
- Research Report > New Finding (0.46)
- Health & Medicine > Therapeutic Area (0.95)
- Health & Medicine > Surgery (0.68)
- Health & Medicine > Diagnostic Medicine > Imaging (0.34)
A new kid on the block: Distributional semantics predicts the word-specific tone signatures of monosyllabic words in conversational Taiwan Mandarin
Jin, Xiaoyun, Ernestus, Mirjam, Baayen, R. Harald
We present a corpus-based investigation of how the pitch contours of monosyllabic words are realized in spontaneous conversational Mandarin, focusing on the effects of words' meanings. We used the generalized additive model to decompose a given observed pitch contour into a set of component pitch contours that are tied to different control variables and semantic predictors. Even when variables such as word duration, gender, speaker identity, tonal context, vowel height, and utterance position are controlled for, the effect of word remains a strong predictor of tonal realization. We present evidence that this effect of word is a semantic effect: word sense is shown to be a better predictor than word, and heterographic homophones are shown to have different pitch contours. The strongest evidence for the importance of semantics is that the pitch contours of individual word tokens can be predicted from their contextualized embeddings with an accuracy that substantially exceeds a permutation baseline. For phonetics, distributional semantics is a new kid on the block. Although our findings challenge standard theories of Mandarin tone, they fit well within the theoretical framework of the Discriminative Lexicon Model.
- Europe > Germany > Baden-Württemberg > Tübingen Region > Tübingen (0.14)
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.14)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- (17 more...)
- Health & Medicine (0.67)
- Education (0.67)
- Leisure & Entertainment (0.46)
- North America > United States > Texas > Travis County > Austin (0.14)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.05)
- North America > United States > California > Santa Clara County > Stanford (0.04)
- North America > Canada > Quebec > Montreal (0.04)
- Asia > Japan > Honshū > Chūbu > Ishikawa Prefecture > Kanazawa (0.04)
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- North America > United States > Louisiana > Orleans Parish > New Orleans (0.04)
- North America > United States > Iowa (0.04)
- Research Report (0.67)
- Overview (0.46)
- Machinery > Industrial Machinery (0.68)
- Information Technology (0.46)
Unitho: A Unified Multi-Task Framework for Computational Lithography
Jin, Qian, Liu, Yumeng, Jiang, Yuqi, Sun, Qi, Zhuo, Cheng
Abstract--Reliable, generalizable data foundations are critical for enabling large-scale models in computational lithography. However, essential tasks--mask generation, rule violation detection, and layout optimization--are often handled in isolation, hindered by scarce datasets and limited modeling approaches. T o address these challenges, we introduce Unitho, a unified multi-task large vision model built upon the Transformer architecture. Trained on a large-scale industrial lithography simulation dataset with hundreds of thousands of cases, Unitho supports end-to-end mask generation, lithography simulation, and rule violation detection. As process nodes continue to shrink, geometric distortions induced by photolithography, such as optical proximity effects (OPE), pose a growing challenge to device performance and manufacturing yield. To ensure that design layouts are transferred to the wafer with high fidelity, optical proximity correction (OPC) and subsequent lithography verification have become indispensable steps in the chip design workflow [1]. However, the industry-standard physics-based simulation, while accurate, is computationally intensive and time-consuming, as shown in Figure 1 This bottleneck is severely exacerbated during process window (PW) analysis, which requires validating design robustness under variations in focus and exposure dose. Since simulations must be repeated across the entire process matrix, the resulting computational overhead significantly prolongs design iteration cycles and severely impedes early-stage Design-Technology Co-Optimization (DTCO), as shown in Figure 1.
- Europe (0.04)
- Asia > China > Zhejiang Province > Hangzhou (0.04)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.14)
- North America > United States > Texas > Travis County > Austin (0.14)
- North America > United States > California > Santa Clara County > Stanford (0.04)
- North America > Canada > Quebec > Montreal (0.04)