Goto

Collaborating Authors

 manning


Tree-to-tree Neural Networks for Program Translation

Xinyun Chen, Chang Liu, Dawn Song

Neural Information Processing Systems

Program translation isanimportant tool tomigrate legacycode inone language into an ecosystem built in a different language. In this work, we are the first to employ deep neural networks toward tackling this problem.


Rare 19th-century coins found after fire at historic tavern

Popular Science

Last December, flames engulfed Ohio's Overfield Tavern Museum. Now archaeologists get to dig beneath the floorboards. A 50 cent coin from 1817 (left) and an 1846 Liberty head one-cent coin (right) are just two of the numerous objects uncovered in a dig in Troy, Ohio. Breakthroughs, discoveries, and DIY tips sent every weekday. Out of the ashes of a devastating fire, archeologists are uncovering exciting insights into Ohio history.


Second language Korean Universal Dependency treebank v1.2: Focus on data augmentation and annotation scheme refinement

Sung, Hakyung, Shin, Gyu-Ho

arXiv.org Artificial Intelligence

We expand the second language (L2) Korean Universal Dependencies (UD) treebank with 5,454 manually annotated sentences. The annotation guidelines are also revised to better align with the UD framework. Using this enhanced treebank, we fine-tune three Korean language models and evaluate their performance on in-domain and out-of-domain L2-Korean datasets. The results show that fine-tuning significantly improves their performance across various metrics, thus highlighting the importance of using well-tailored L2 datasets for fine-tuning first-language-based, general-purpose language models for the morphosyntactic analysis of L2 data.


Scientific Machine Learning of Flow Resistance Using Universal Shallow Water Equations with Differentiable Programming

Liu, Xiaofeng, Song, Yalan

arXiv.org Artificial Intelligence

Shallow water equations (SWEs) are the backbone of most hydrodynamics models for flood prediction, river engineering, and many other water resources applications. The estimation of flow resistance, i.e., the Manning's roughness coefficient $n$, is crucial for ensuring model accuracy, and has been previously determined using empirical formulas or tables. To better account for temporal and spatial variability in channel roughness, inverse modeling of $n$ using observed flow data is more reliable and adaptable; however, it is challenging when using traditional SWE solvers. Based on the concept of universal differential equation (UDE), which combines physics-based differential equations with neural networks (NNs), we developed a universal SWEs (USWEs) solver, Hydrograd, for hybrid hydrodynamics modeling. It can do accurate forward simulations, support automatic differentiation (AD) for gradient-based sensitivity analysis and parameter inversion, and perform scientific machine learning for physics discovery. In this work, we first validated the accuracy of its forward modeling, then applied a real-world case to demonstrate the ability of USWEs to capture model sensitivity (gradients) and perform inverse modeling of Manning's $n$. Furthermore, we used a NN to learn a universal relationship between $n$, hydraulic parameters, and flow in a real river channel. Unlike inverse modeling using surrogate models, Hydrograd uses a two-dimensional SWEs solver as its physics backbone, which eliminates the need for data-intensive pretraining and resolves the generalization problem when applied to out-of-sample scenarios. This differentiable modeling approach, with seamless integration with NNs, provides a new pathway for solving complex inverse problems and discovering new physics in hydrodynamics.


MorphNLI: A Stepwise Approach to Natural Language Inference Using Text Morphing

Negru, Vlad Andrei, Vacareanu, Robert, Lemnaru, Camelia, Surdeanu, Mihai, Potolea, Rodica

arXiv.org Artificial Intelligence

We introduce MorphNLI, a modular step-by-step approach to natural language inference (NLI). When classifying the premise-hypothesis pairs into {entailment, contradiction, neutral}, we use a language model to generate the necessary edits to incrementally transform (i.e., morph) the premise into the hypothesis. Then, using an off-the-shelf NLI model we track how the entailment progresses with these atomic changes, aggregating these intermediate labels into a final output. We demonstrate the advantages of our proposed method particularly in realistic cross-domain settings, where our method always outperforms strong baselines with improvements up to 12.6% (relative). Further, our proposed approach is explainable as the atomic edits can be used to understand the overall NLI label.


Scaling Graph-Based Dependency Parsing with Arc Vectorization and Attention-Based Refinement

Floquet, Nicolas, Roux, Joseph Le, Tomeh, Nadi, Charnois, Thierry

arXiv.org Artificial Intelligence

We propose a novel architecture for graph-based dependency parsing that explicitly constructs vectors, from which both arcs and labels are scored. Our method addresses key limitations of the standard two-pipeline approach by unifying arc scoring and labeling into a single network, reducing scalability issues caused by the information bottleneck and lack of parameter sharing. Additionally, our architecture overcomes limited arc interactions with transformer layers to efficiently simulate higher-order dependencies. Experiments on PTB and UD show that our model outperforms state-of-the-art parsers in both accuracy and efficiency.


Reviews: Collaborative Recurrent Autoencoder: Recommend while Learning to Fill in the Blanks

Neural Information Processing Systems

Making recommendations by exploiting the plethora of text data that accompanies products seems like an under-explored area, especially using recent advances in deep language models. I commend the authors for contributing to this research direction. The CRAE seems to work well (at least in recall at k), performing at a high level on two real-world datasets. However, I think this paper would be a better fit for a more applied conference, such as KDD or RecSys, because there is little novelty to the model's core components. I'll address each individually, in order of (my perceived) importance: 1) Robust Recurrent Networks (RRN): The proposed RRN uses distributional activations that are backpropagated through directly.


Semgrex and Ssurgeon, Searching and Manipulating Dependency Graphs

Bauer, John, Kiddon, Chloe, Yeh, Eric, Shan, Alex, Manning, Christopher D.

arXiv.org Artificial Intelligence

Searching dependency graphs and manipulating them can be a time consuming and challenging task to get right. We document Semgrex, a system for searching dependency graphs, and introduce Ssurgeon, a system for manipulating the output of Semgrex. The compact language used by these systems allows for easy command line or API processing of dependencies. Additionally, integration with publicly released toolkits in Java and Python allows for searching text relations and attributes over natural text.


Deep Neural Networks with 3D Point Clouds for Empirical Friction Measurements in Hydrodynamic Flood Models

Haces-Garcia, Francisco, Kotzamanis, Vasileios, Glennie, Craig, Rifai, Hanadi

arXiv.org Artificial Intelligence

Friction is one of the cruxes of hydrodynamic modeling; flood conditions are highly sensitive to the Friction Factors (FFs) used to calculate momentum losses. However, empirical FFs are challenging to measure because they require laboratory experiments. Flood models often rely on surrogate observations (such as land use) to estimate FFs, introducing uncertainty. This research presents a laboratory-trained Deep Neural Network (DNN), trained using flume experiments with data augmentation techniques, to measure Manning's n based on Point Cloud data. The DNN was deployed on real-world lidar Point Clouds to directly measure Manning's n under regulatory and extreme storm events, showing improved prediction capabilities in both 1D and 2D hydrodynamic models. For 1D models, the lidar values decreased differences with regulatory models for in-channel water depth when compared to land cover values. For 1D/2D coupled models, the lidar values produced better agreement with flood extents measured from airborne imagery, while better matching flood insurance claim data for Hurricane Harvey. In both 1D and 1D/2D coupled models, lidar resulted in better agreement with validation gauges. For these reasons, the lidar measurements of Manning's n were found to improve both regulatory models and forecasts for extreme storm events, while simultaneously providing a pathway to standardize the measurement of FFs. Changing FFs significantly affected fluvial and pluvial flood models, while surge flooding was generally unaffected. Downstream flow conditions were found to change the importance of FFs to fluvial models, advancing the literature of friction in flood models. This research introduces a reliable, repeatable, and readily-accessible avenue to measure high-resolution FFs based on 3D point clouds, improving flood prediction, and removing uncertainty from hydrodynamic modeling.


Texas' Arch Manning will not opt to appear in EA Sports college football video game: report

FOX News

Texas Longhorns quarterback Arch Manning may have to wait at least one more season before he gets his chance to start for the team. With EA Sports releasing College Football 25 this summer, fans may have gotten the chance to make him their QB1 in the video game -- if he opted in. But it doesn't appear that will be the case. Manning has not opted in to have his name, image and likeness used in the game, Orangebloods.com He is reportedly "focused on playing football on the field."