Plotting

 Menzies, Tim


What is Wrong with Topic Modeling? (and How to Fix it Using Search-based Software Engineering)

arXiv.org Artificial Intelligence

Context: Topic modeling finds human-readable structures in unstructured textual data. A widely used topic modeler is Latent Dirichlet allocation. When run on different datasets, LDA suffers from "order effects" i.e. different topics are generated if the order of training data is shuffled. Such order effects introduce a systematic error for any study. This error can relate to misleading results;specifically, inaccurate topic descriptions and a reduction in the efficacy of text mining classification results. Objective: To provide a method in which distributions generated by LDA are more stable and can be used for further analysis. Method: We use LDADE, a search-based software engineering tool that tunes LDA's parameters using DE (Differential Evolution). LDADE is evaluated on data from a programmer information exchange site (Stackoverflow), title and abstract text of thousands ofSoftware Engineering (SE) papers, and software defect reports from NASA. Results were collected across different implementations of LDA (Python+Scikit-Learn, Scala+Spark); across different platforms (Linux, Macintosh) and for different kinds of LDAs (VEM,or using Gibbs sampling). Results were scored via topic stability and text mining classification accuracy. Results: In all treatments: (i) standard LDA exhibits very large topic instability; (ii) LDADE's tunings dramatically reduce cluster instability; (iii) LDADE also leads to improved performances for supervised as well as unsupervised learning. Conclusion: Due to topic instability, using standard LDA with its "off-the-shelf" settings should now be depreciated. Also, in future, we should require SE papers that use LDA to test and (if needed) mitigate LDA topic instability. Finally, LDADE is a candidate technology for effectively and efficiently reducing that instability.


Why is Differential Evolution Better than Grid Search for Tuning Defect Predictors?

arXiv.org Machine Learning

Context: One of the black arts of data mining is learning the magic parameters which control the learners. In software analytics, at least for defect prediction, several methods, like grid search and differential evolution (DE), have been proposed to learn these parameters, which has been proved to be able to improve the performance scores of learners. Objective: We want to evaluate which method can find better parameters in terms of performance score and runtime cost. Methods: This paper compares grid search to differential evolution, which is an evolutionary algorithm that makes extensive use of stochastic jumps around the search space. Results: We find that the seemingly complete approach of grid search does no better, and sometimes worse, than the stochastic search. When repeated 20 times to check for conclusion validity, DE was over 210 times faster than grid search to tune Random Forests on 17 testing data sets with F-Measure Conclusions: These results are puzzling: why does a quick partial search be just as effective as a much slower, and much more, extensive search? To answer that question, we turned to the theoretical optimization literature. Bergstra and Bengio conjecture that grid search is not more effective than more randomized searchers if the underlying search space is inherently low dimensional. This is significant since recent results show that defect prediction exhibits very low intrinsic dimensionality-- an observation that explains why a fast method like DE may work as well as a seemingly more thorough grid search. This suggests, as a future research direction, that it might be possible to peek at data sets before doing any optimization in order to match the optimization algorithm to the problem at hand.


A deep learning model for estimating story points

arXiv.org Machine Learning

Although there has been substantial research in software analytics for effort estimation in traditional software projects, little work has been done for estimation in agile projects, especially estimating user stories or issues. Story points are the most common unit of measure used for estimating the effort involved in implementing a user story or resolving an issue. In this paper, we offer for the \emph{first} time a comprehensive dataset for story points-based estimation that contains 23,313 issues from 16 open source projects. We also propose a prediction model for estimating story points based on a novel combination of two powerful deep learning architectures: long short-term memory and recurrent highway network. Our prediction system is \emph{end-to-end} trainable from raw input data to prediction outcomes without any manual feature engineering. An empirical evaluation demonstrates that our approach consistently outperforms three common effort estimation baselines and two alternatives in both Mean Absolute Error and the Standardized Accuracy.


Reports on the AAAI 1999 Workshop Program

AI Magazine

The AAAI-99 Workshop Program (a part of the sixteenth national conference on artificial intelligence) was held in Orlando, Florida. Each workshop was limited to approximately 25 to 50 participants. Participation was by invitation from the workshop organizers. The workshops were Agent-Based Systems in the Business Context, Agents' Conflicts, Artificial Intelligence for Distributed Information Networking, Artificial Intelligence for Electronic Commerce, Computation with Neural Systems Workshop, Configuration, Data Mining with Evolutionary Algorithms: Research Directions (Jointly sponsored by GECCO-99), Environmental Decision Support Systems and Artificial Intelligence, Exploring Synergies of Knowledge Management and Case-Based Reasoning, Intelligent Information Systems, Intelligent Software Engineering, Machine Learning for Information Extraction, Mixed-Initiative Intelligence, Negotiation: Settling Conflicts and Identifying Opportunities, Ontology Management, and Reasoning in Context for AI Applications.


Reports on the AAAI 1999 Workshop Program

AI Magazine

The AAAI-99 Workshop Program (a part of the sixteenth national conference on artificial intelligence) was held in Orlando, Florida. The program included 16 workshops covering a wide range of topics in AI. Each workshop was limited to approximately 25 to 50 participants. Participation was by invitation from the workshop organizers. The workshops were Agent-Based Systems in the Business Context, Agents' Conflicts, Artificial Intelligence for Distributed Information Networking, Artificial Intelligence for Electronic Commerce, Computation with Neural Systems Workshop, Configuration, Data Mining with Evolutionary Algorithms: Research Directions (Jointly sponsored by GECCO-99), Environmental Decision Support Systems and Artificial Intelligence, Exploring Synergies of Knowledge Management and Case-Based Reasoning, Intelligent Information Systems, Intelligent Software Engineering, Machine Learning for Information Extraction, Mixed-Initiative Intelligence, Negotiation: Settling Conflicts and Identifying Opportunities, Ontology Management, and Reasoning in Context for AI Applications.


An Investigation of AI and Expert Systems Literature: 1980-1984

AI Magazine

This article records the results of an experiment in which a survey of AI and expert systems (ES) literature was attempted using Science Citation Indexes. The survey identified a sample of authors and institutions that have had a significant impact on the historical development of AI and ES. However, it also identified several glaring problems with using Science Citation Indexes as a method of comprehensively studying a body of scientific research. Accordingly, the reader is cautioned against using the results presented here to conclude that author A is a better or worse AI researcher than author B.


An Investigation of AI and Expert Systems Literature: 1980-1984

AI Magazine

This article records the results of an experiment in which a survey of AI and expert systems (ES) literature was attempted using Science Citation Indexes. The survey identified a sample of authors and institutions that have had a significant impact on the historical development of AI and ES. However, it also identified several glaring problems with using Science Citation Indexes as a method of comprehensively studying a body of scientific research. Accordingly, the reader is cautioned against using the results presented here to conclude that author A is a better or worse AI researcher than author B.