ple
Incorporating Legal Logic into Deep Learning: An Intelligent Approach to Probation Prediction
Wang, Qinghua, Zhang, Xu, Yang, Lingyan, Shao, Rui, Wang, Bonan, Wang, Fang, Qu, Cunquan
Probation is a crucial institution in modern criminal law, embodying the principles of fairness and justice while contributing to the harmonious development of society. Despite its importance, the current Intelligent Judicial Assistant System (IJAS) lacks dedicated methods for probation prediction, and research on the underlying factors influencing probation eligibility remains limited. In addition, probation eligibility requires a comprehensive analysis of both criminal circumstances and remorse. Much of the existing research in IJAS relies primarily on data-driven methodologies, which often overlooks the legal logic underpinning judicial decision-making. To address this gap, we propose a novel approach that integrates legal logic into deep learning models for probation prediction, implemented in three distinct stages. First, we construct a specialized probation dataset that includes fact descriptions and probation legal elements (PLEs). Second, we design a distinct probation prediction model named the Multi-Task Dual-Theory Probation Prediction Model (MT-DT), which is grounded in the legal logic of probation and the \textit{Dual-Track Theory of Punishment}. Finally, our experiments on the probation dataset demonstrate that the MT-DT model outperforms baseline models, and an analysis of the underlying legal logic further validates the effectiveness of the proposed approach.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- Asia > China (0.05)
- North America > United States > Texas > Travis County > Austin (0.04)
- (9 more...)
- Law > Criminal Law (1.00)
- Law Enforcement & Public Safety > Crime Prevention & Enforcement (0.93)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.94)
- Information Technology > Artificial Intelligence > Representation & Reasoning (0.69)
- Information Technology > Artificial Intelligence > Natural Language (0.68)
Latent Embedding Adaptation for Human Preference Alignment in Diffusion Planners
Ng, Wen Zheng Terence, Chen, Jianda, Xu, Yuan, Zhang, Tianwei
This work addresses the challenge of personalizing trajectories generated in automated decision-making systems by introducing a resource-efficient approach that enables rapid adaptation to individual users' preferences. Our method leverages a pretrained conditional diffusion model with Preference Latent Embeddings (PLE), trained on a large, reward-free offline dataset. The PLE serves as a compact representation for capturing specific user preferences. By adapting the pretrained model using our proposed preference inversion method, which directly optimizes the learnable PLE, we achieve superior alignment with human preferences compared to existing solutions like Reinforcement Learning from Human Feedback (RLHF) and Low-Rank Adaptation (LoRA). To better reflect practical applications, we create a benchmark experiment using real human preferences on diverse, high-reward trajectories.
- Europe > Netherlands > North Holland > Amsterdam (0.04)
- Europe > Germany > Bavaria > Upper Bavaria > Munich (0.04)
- Asia > Singapore (0.04)
Inside the Black Box: Detecting Data Leakage in Pre-trained Language Encoders
Xin, Yuan, Li, Zheng, Yu, Ning, Chen, Dingfan, Fritz, Mario, Backes, Michael, Zhang, Yang
Despite being prevalent in the general field of Natural Language Processing (NLP), pre-trained language models inherently carry privacy and copyright concerns due to their nature of training on large-scale web-scraped data. In this paper, we pioneer a systematic exploration of such risks associated with pre-trained language encoders, specifically focusing on the membership leakage of pre-training data exposed through downstream models adapted from pre-trained language encoders-an aspect largely overlooked in existing literature. Our study encompasses comprehensive experiments across four types of pre-trained encoder architectures, three representative downstream tasks, and five benchmark datasets. Intriguingly, our evaluations reveal, for the first time, the existence of membership leakage even when only the black-box output of the downstream model is exposed, highlighting a privacy risk far greater than previously assumed. Alongside, we present in-depth analysis and insights toward guiding future researchers and practitioners in addressing the privacy considerations in developing pre-trained language models.
- North America > United States > New York > Tompkins County > Ithaca (0.04)
- North America > United States > Oregon > Jackson County > Medford (0.04)
- Asia > China (0.04)
- Information Technology > Security & Privacy (1.00)
- Law (0.93)
- Transportation (0.85)
Enhancing User Interest based on Stream Clustering and Memory Networks in Large-Scale Recommender Systems
Liu, Peng, Wang, Nian, Xu, Cong, Zhao, Ming, Wang, Bin, Ren, Yi
Recommender Systems (RSs) provide personalized recommendation Recommender Systems (RSs) [1, 2] which provide personalized service based on user interest, which are widely used in various recommendation service based on user interest are widely used in platforms. However, there are lots of users with sparse interest various platforms such as short video platforms [3, 7, 14], video due to lacking consumption behaviors, which leads to poor recommendation platforms [4, 5], E-commerce platforms [6, 8-11] and social networks results for them. This problem is widespread in [12, 13], serving billions of users. In RSs, Ranking typically large-scale RSs and is particularly difficult to address. To solve uses a Multi-Task Learning model (MTL) [4, 8, 16-21] and lots this problem, we propose a novel solution named User Interest of features to finely predict the scores of various user behaviors Enhancement (UIE) which enhances user interest including user such as click, watching time, fast slide, like and sharing for thousands profile and user history behavior sequences using the enhancement of candidates. The accuracy of the scores outputted by MTL vectors and personalized enhancement vector generated with is crucial for RSs [4]. In RSs, user interest includes user profile the help of other similar users and relevant items based on stream and user history behavior sequences, as shown in Figure 1 and clustering and memory networks from different perspectives. UIE Figure 2, which determines the upper limit of ranking model's not only remarkably improves model performance on the users performance. However, lots of users only have sparse interest due with sparse interest but also significantly enhance model performance to lacking consumption behaviors.
- Asia > China > Beijing > Beijing (0.05)
- Asia > Myanmar > Tanintharyi Region > Dawei (0.04)
- North America > United States > District of Columbia > Washington (0.04)
- (17 more...)
Removing Bias from Maximum Likelihood Estimation with Model Autophagy
Mayer, Paul, Luzi, Lorenzo, Siahkoohi, Ali, Johnson, Don H., Baraniuk, Richard G.
We propose autophagy penalized likelihood estimation (PLE), an unbiased alternative to maximum likelihood estimation (MLE) which is more fair and less susceptible to model autophagy disorder (madness). Model autophagy refers to models trained on their own output; PLE ensures the statistics of these outputs coincide with the data statistics. This enables PLE to be statistically unbiased in certain scenarios where MLE is biased. When biased, MLE unfairly penalizes minority classes in unbalanced datasets and exacerbates the recently discovered issue of self-consuming generative modeling. Theoretical and empirical results show that 1) PLE is more fair to minority classes and 2) PLE is more stable in a self-consumed setting. Furthermore, we provide a scalable and portable implementation of PLE with a hypernetwork framework, allowing existing deep learning architectures to be easily trained with PLE. Finally, we show PLE can bridge the gap between Bayesian and frequentist paradigms in statistics.
- North America > United States > Texas > Harris County > Houston (0.05)
- North America > United States > New York > New York County > New York City (0.04)
- North America > United States > California > Santa Clara County > Palo Alto (0.04)
- (3 more...)
- Research Report > Experimental Study (0.46)
- Research Report > New Finding (0.34)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Uncertainty > Bayesian Inference (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models > Directed Networks > Bayesian Learning (1.00)
On Embeddings for Numerical Features in Tabular Deep Learning
Gorishniy, Yury, Rubachev, Ivan, Babenko, Artem
Recently, Transformer-like deep architectures have shown strong performance on tabular data problems. Unlike traditional models, e.g., MLP, these architectures map scalar values of numerical features to high-dimensional embeddings before mixing them in the main backbone. In this work, we argue that embeddings for numerical features are an underexplored degree of freedom in tabular DL, which allows constructing more powerful DL models and competing with GBDT on some traditionally GBDT-friendly benchmarks. We start by describing two conceptually different approaches to building embedding modules: the first one is based on a piecewise linear encoding of scalar values, and the second one utilizes periodic activations. Then, we empirically demonstrate that these two approaches can lead to significant performance boosts compared to the embeddings based on conventional blocks such as linear layers and ReLU activations. Importantly, we also show that embedding numerical features is beneficial for many backbones, not only for Transformers. Specifically, after proper embeddings, simple MLP-like models can perform on par with the attention-based architectures. Overall, we highlight embeddings for numerical features as an important design aspect with good potential for further improvements in tabular DL.
A Bayesian Hyperprior Approach for Joint Image Denoising and Interpolation, with an Application to HDR Imaging
Aguerrebere, Cecilia, Almansa, Andrés, Delon, Julie, Gousseau, Yann, Musé, Pablo
Recently, impressive denoising results have been achieved by Bayesian approaches which assume Gaussian models for the image patches. This improvement in performance can be attributed to the use of per-patch models. Unfortunately such an approach is particularly unstable for most inverse problems beyond denoising. In this work, we propose the use of a hyperprior to model image patches, in order to stabilize the estimation procedure. There are two main advantages to the proposed restoration scheme: Firstly it is adapted to diagonal degradation matrices, and in particular to missing data problems (e.g. inpainting of missing pixels or zooming). Secondly it can deal with signal dependent noise models, particularly suited to digital cameras. As such, the scheme is especially adapted to computational photography. In order to illustrate this point, we provide an application to high dynamic range imaging from a single image taken with a modified sensor, which shows the effectiveness of the proposed scheme.
- Europe > France > Île-de-France > Paris > Paris (0.14)
- Europe > France > Île-de-France > Val-de-Marne > Cachan (0.04)
- South America > Uruguay > Montevideo > Montevideo (0.04)
- (5 more...)
- Information Technology > Sensing and Signal Processing > Image Processing (1.00)
- Information Technology > Artificial Intelligence > Vision (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Uncertainty > Bayesian Inference (0.66)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models > Directed Networks > Bayesian Learning (0.66)
ntasfi/PyGame-Learning-Environment
PyGame Learning Environment (PLE) is a learning environment, mimicking the Arcade Learning Environment interface, allowing a quick start to Reinforcement Learning in Python. The goal of PLE is allow practitioners to focus design of models and experiments instead of environment design. PLE hopes to eventually build an expansive library of games. Docs for the project can be found here. A PLE instance requires a game exposing a set of control methods.