snowflake
Snow isn't actually white
Winter wonderlands are only possible thanks to a sparkly light trick. Few languages have as many distinct words for snow as Japanese, which has words like miyuki or beautiful snow. Breakthroughs, discoveries, and DIY tips sent six days a week. When someone says " as white as snow," it's easy to envision what they're talking about. We often think of snow as a dazzling white, the same way we immediately conjure up a color when someone says "blood red" or "ocean blue."
- North America > United States > California > San Francisco County > San Francisco (0.15)
- North America > United States > Nevada (0.05)
- North America > United States > Colorado > Boulder County > Boulder (0.05)
- (4 more...)
Snowflake: Scaling GNNs to high-dimensional continuous control via parameter freezing
Recent research has shown that graph neural networks (GNNs) can learn policies for locomotion control that are as effective as a typical multi-layer perceptron (MLP), with superior transfer and multi-task performance. However, results have so far been limited to training on small agents, with the performance of GNNs deteriorating rapidly as the number of sensors and actuators grows. A key motivation for the use of GNNs in the supervised learning setting is their applicability to large graphs, but this benefit has not yet been realised for locomotion control. We show that poor scaling in GNNs is a result of increasingly unstable policy updates, caused by overfitting in parts of the network during training. To combat this, we introduce Snowflake, a GNN training method for high-dimensional continuous control that freezes parameters in selected parts of the network. Snowflake significantly boosts the performance of GNNs for locomotion control on large agents, now matching the performance of MLPs while offering superior transfer properties.
Swatch MoonSwatch Mission To Earthphase Moonshine Gold Cold Moon: Price, Specs, Availability
Swatch will laser unique gold snowflakes on every new Cold Moon MoonSwatch, but there's a catch--you'll only be able to buy one when it's snowing in Switzerland. First a confession: I own more MoonSwatches than I care to admit. Never let it be said that WIRED does not walk the walk when it comes to recommending products--Swatch has assiduously extracted a considerable amount of cash from me, all in $285 increments. This was no doubt the Swiss company's dastardly plan all along, to lure us in, then, oh so gently, get watch fans hooked. It's worked, too--Swatch has, so far, netted hundreds of millions of dollars from MoonSwatch sales.
- Europe > Switzerland (0.25)
- North America > United States > California (0.05)
- Europe > United Kingdom (0.05)
- (2 more...)
AI-Driven Generation of Data Contracts in Modern Data Engineering Systems
Data contracts formalize agreements between data producers and consumers regarding schema, semantics, and quality expectations. As data pipelines grow in complexity, manual authoring and maintenance of contracts becomes error-prone and labor-intensive. We present an AI-driven framework for automatic data contract generation using large language models (LLMs). Our system leverages parameter-efficient fine-tuning methods, including LoRA and PEFT, to adapt LLMs to structured data domains. The models take sample data or schema descriptions and output validated contract definitions in formats such as JSON Schema and Avro. We integrate this framework into modern data platforms (e.g., Databricks, Snowflake) to automate contract enforcement at scale. Experimental results on synthetic and real-world datasets demonstrate that the fine-tuned LLMs achieve high accuracy in generating valid contracts and reduce manual workload by over 70%. We also discuss key challenges such as hallucination, version control, and the need for continuous learning. This work demonstrates that generative AI can enable scalable, agile data governance by bridging the gap between intent and implementation in enterprise data management.
DropletVideo: A Dataset and Approach to Explore Integral Spatio-Temporal Consistent Video Generation
Zhang, Runze, Du, Guoguang, Li, Xiaochuan, Jia, Qi, Jin, Liang, Liu, Lu, Wang, Jingjing, Xu, Cong, Guo, Zhenhua, Zhao, Yaqian, Gong, Xiaoli, Li, Rengang, Fan, Baoyu
Spatio-temporal consistency is a critical research topic in video generation. A qualified generated video segment must ensure plot plausibility and coherence while maintaining visual consistency of objects and scenes across varying viewpoints. Prior research, especially in open-source projects, primarily focuses on either temporal or spatial consistency, or their basic combination, such as appending a description of a camera movement after a prompt without constraining the outcomes of this movement. However, camera movement may introduce new objects to the scene or eliminate existing ones, thereby overlaying and affecting the preceding narrative. Especially in videos with numerous camera movements, the interplay between multiple plots becomes increasingly complex. This paper introduces and examines integral spatio-temporal consistency, considering the synergy between plot progression and camera techniques, and the long-term impact of prior content on subsequent generation. Our research encompasses dataset construction through to the development of the model. Initially, we constructed a DropletVideo-10M dataset, which comprises 10 million videos featuring dynamic camera motion and object actions. Each video is annotated with an average caption of 206 words, detailing various camera movements and plot developments. Following this, we developed and trained the DropletVideo model, which excels in preserving spatio-temporal coherence during video generation. The DropletVideo dataset and model are accessible at https://dropletx.github.io.
- Media > Television (1.00)
- Media > Photography (1.00)
- Media > Film (1.00)
- Information Technology > Artificial Intelligence > Vision (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (0.67)
Snowflake: Scaling GNNs to high-dimensional continuous control via parameter freezing
Recent research has shown that graph neural networks (GNNs) can learn policies for locomotion control that are as effective as a typical multi-layer perceptron (MLP), with superior transfer and multi-task performance. However, results have so far been limited to training on small agents, with the performance of GNNs deteriorating rapidly as the number of sensors and actuators grows. A key motivation for the use of GNNs in the supervised learning setting is their applicability to large graphs, but this benefit has not yet been realised for locomotion control. We show that poor scaling in GNNs is a result of increasingly unstable policy updates, caused by overfitting in parts of the network during training. To combat this, we introduce Snowflake, a GNN training method for high-dimensional continuous control that freezes parameters in selected parts of the network. Snowflake significantly boosts the performance of GNNs for locomotion control on large agents, now matching the performance of MLPs while offering superior transfer properties.
Improved Large Language Model Jailbreak Detection via Pretrained Embeddings
Galinkin, Erick, Sablotny, Martin
The adoption of large language models (LLMs) in many applications, from customer service chat bots and software de - velopment assistants to more capable agentic systems neces - sitates research into how to secure these systems. Attacks l ike prompt injection and jailbreaking attempt to elicit respon ses and actions from these models that are not compliant with the safety, privacy, or content policies of organizations u sing the model in their application. In order to counter abuse of LLMs for generating potentially harmful replies or taking u n-desirable actions, LLM owners must apply safeguards during training and integrate additional tools to block the LLM fro m generating text that abuses the model. Jailbreaking prompt s play a vital role in convincing an LLM to generate potentially harmful content, making it important to identify jai l-breaking attempts to block any further steps. In this work, w e propose a novel approach to detect jailbreak prompts based on pairing text embeddings well-suited for retrieval with t ra-ditional machine learning classification algorithms. Our a p-proach outperforms all publicly available methods from ope n source LLM security applications.
- North America > United States > California > Santa Clara County > Santa Clara (0.04)
- North America > Canada > Alberta > Census Division No. 15 > Improvement District No. 9 > Banff (0.04)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Performance Analysis > Accuracy (0.98)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models > Directed Networks > Bayesian Learning (0.46)
Efficient Probabilistic Modeling of Crystallization at Mesoscopic Scale
Timmer, Pol, Minartz, Koen, Menkovski, Vlado
Crystallization processes at the mesoscopic scale, where faceted, dendritic growth, and multigrain formation can be observed, are of particular interest within materials science and metallurgy. These processes are highly nonlinear, stochastic, and sensitive to small perturbations of system parameters and initial conditions. Methods for the simulation of these processes have been developed using discrete numerical models, but these are computationally expensive. This work aims to scale crystal growth simulation with a machine learning emulator. Specifically, autoregressive latent variable models are well suited for modeling the joint distribution over system parameters and the crystallization trajectories. However, successfully training such models is challenging due to the stochasticity and sensitivity of the system. Existing approaches consequently fail to produce diverse and faithful crystallization trajectories. In this paper, we introduce the Crystal Growth Neural Emulator (CGNE), a probabilistic model for efficient crystal growth emulation at the mesoscopic scale that overcomes these challenges. We validate CGNE results using the morphological properties of the crystals produced by numerical simulation. CGNE delivers a factor of 11 improvement in inference time and performance gains compared with recent state-of-the-art probabilistic models for dynamical systems.
- Europe > Netherlands > North Brabant > Eindhoven (0.04)
- Europe > Germany > Bavaria > Middle Franconia > Nuremberg (0.04)
LILO: Learning Interpretable Libraries by Compressing and Documenting Code
Grand, Gabriel, Wong, Lionel, Bowers, Matthew, Olausson, Theo X., Liu, Muxin, Tenenbaum, Joshua B., Andreas, Jacob
While large language models (LLMs) now excel at code generation, a key aspect of software development is the art of refactoring: consolidating code into libraries of reusable and readable programs. In this paper, we introduce LILO, a neurosymbolic framework that iteratively synthesizes, compresses, and documents code to build libraries tailored to particular problem domains. LILO combines LLM-guided program synthesis with recent algorithmic advances in automated refactoring from Stitch: a symbolic compression system that efficiently identifies optimal lambda abstractions across large code corpora. To make these abstractions interpretable, we introduce an auto-documentation (AutoDoc) procedure that infers natural language names and docstrings based on contextual examples of usage. In addition to improving human readability, we find that AutoDoc boosts performance by helping LILO's synthesizer to interpret and deploy learned abstractions. We evaluate LILO on three inductive program synthesis benchmarks for string editing, scene reasoning, and graphics composition. Compared to existing neural and symbolic methods - including the state-of-the-art library learning algorithm DreamCoder - LILO solves more complex tasks and learns richer libraries that are grounded in linguistic knowledge.
- North America > United States > California > San Francisco County > San Francisco (0.14)
- North America > United States > California > Los Angeles County > Los Angeles (0.14)
- North America > Canada > Quebec > Montreal (0.04)
- (19 more...)
- Health & Medicine (0.46)
- Education > Educational Setting > Continuing Education (0.34)
Snowflake Introduces Manufacturing Data Cloud to Empower Industries with Data and AI
Data and AI technology for manufacturing is having a moment. As AI has expanded its lengthy reach into manufacturing, there have been several purpose-built releases lately from companies like Nvidia and Databricks that are helping companies make sense of the deluge of data collected on everything from physical operations to the supply chain. Snowflake is now part of this action with the debut of its Manufacturing Data Cloud. The company says this new offering will enable companies in the automotive, technology, energy, and industrial sectors to tap into the value of siloed industrial data by leveraging Snowflake's data platform, partner solutions, and industry-specific datasets. The Snowflake Data Cloud provides a platform for data warehousing, SQL analytics, machine learning, data engineering, and monetization of third-party data.