Well File:


Appendix A Data and Code Availability 17 A.1 Code 17 A.2 Data 17 A.3 Result 17 B Dataset Documentation

Neural Information Processing Systems

The robust ability of LLMs to generate and acquire domain-specific knowledge has been a significant factor in this potential [17]. While researchers have explored the use of LLMs in answering agriculture-related exams [55], their performance in certain crop cultivation scenarios, such as pest management, has been less than satisfactory [66]. Moreover, there remains a considerable gap between the ability to answer exam questions and the application of this knowledge in real-world situations. To bridge the gap and thoroughly assess LLMs in supporting the crop science field, we introduce CROP. CROP comprises an instruction tuning dataset that equips LLMs with the necessary skills to aid tasks in crop production, along with a carefully designed benchmark to evaluate the extent to which LLMs fulfill the demands of real-world agricultural applications. We anticipate that CROP will serve the research community and also provide practical benefits to industry practitioners. E.2 LLM-based Multi-turn Dialogue Generation In recent research, several LLM-based approaches have emerged for constructing multi-turn dialogues.


Empowering and Assessing the Utility of Large Language Models in Crop Science 1

Neural Information Processing Systems

Large language models (LLMs) have demonstrated remarkable efficacy across knowledge-intensive tasks. Nevertheless, their untapped potential in crop science presents an opportunity for advancement.



Frequency-aware Generative Models for Multivariate Time Series Imputation Xinyu Yang

Neural Information Processing Systems

Missing data in multivariate time series are common issues that can affect the analysis and downstream applications. Although multivariate time series data generally consist of the trend, seasonal and residual terms, existing works mainly focus on optimizing the modeling for the first two items. However, we find that the residual term is more crucial for getting accurate fillings, since it is more related to the diverse changes of data and the biggest component of imputation errors. Therefore, in this study, we introduce frequency-domain information and design Frequency-aware Generative Models for Multivariate Time Series Imputation (FGTI). Specifically, FGTI employs a high-frequency filter to boost the residual term imputation, supplemented by a dominant-frequency filter for the trend and seasonal imputation.


Density-based User Representation using Gaussian Process Regression for Multi-interest Personalized Retrieval

Neural Information Processing Systems

Accurate modeling of the diverse and dynamic interests of users remains a significant challenge in the design of personalized recommender systems. Existing user modeling methods, like single-point and multi-point representations, have limitations w.r.t.


Appendix In our proofs, we use c, c and kAk

Neural Information Processing Systems

For a set S, we use S to denote the complement of the set. Since we only analyze a single iteration, for simplicity we drop the superscript that indicates the iteration counter. Probability of erroneous cluster identity estimation We begin with the analysis of the probability of incorrect cluster identity estimation. Now we proceed to analyze the gradient descent step. Without loss of generality, we only analyze the first cluster.


An Efficient Framework for Clustered Federated Learning

Neural Information Processing Systems

We address the problem of Federated Learning (FL) where users are distributed and partitioned into clusters. This setup captures settings where different groups of users have their own objectives (learning tasks) but by aggregating their data with others in the same cluster (same learning task), they can leverage the strength in numbers in order to perform more efficient Federated Learning. We propose a new framework dubbed the Iterative Federated Clustering Algorithm (IFCA), which alternately estimates the cluster identities of the users and optimizes model parameters for the user clusters via gradient descent. We analyze the convergence rate of this algorithm first in a linear model with squared loss and then for generic strongly convex and smooth loss functions. We show that in both settings, with good initialization, IFCA converges at an exponential rate, and discuss the optimality of the statistical error rate. When the clustering structure is ambiguous, we propose to train the models by combining IFCA with the weight sharing technique in multi-task learning. In the experiments, we show that our algorithm can succeed even if we relax the requirements on initialization with random initialization and multiple restarts. We also present experimental results showing that our algorithm is efficient in non-convex problems such as neural networks. We demonstrate the benefits of IFCA over the baselines on several clustered FL benchmarks.