If you are looking for an answer to the question What is Artificial Intelligence? and you only have a minute, then here's the definition the Association for the Advancement of Artificial Intelligence offers on its home page: "the scientific understanding of the mechanisms underlying thought and intelligent behavior and their embodiment in machines."
However, if you are fortunate enough to have more than a minute, then please get ready to embark upon an exciting journey exploring AI (but beware, it could last a lifetime) …
Conditions which affect our mental health often affect the way we use language; and treatment often involves linguistic interaction. This talk will present work on three related projects investigating the use of computational natural language processing (NLP) to help understand and improve diagnosis and treatment for such conditions. We will look at clinical dialogue between patient and doctor or therapist, in cases involving schizophrenia, depression and dementia; in each case, we find that diagnostic information and/or treatment outcomes are related to observable features of a patient's language and interaction with their conversational partner. We discuss the nature of these phenomena and the suitability and accuracy of NLP techniques for detecting them automatically.
Conditions which affect our mental health often affect the way we use language; and treatment often involves linguistic interaction. In his talk Dr Matthew Purver from Queen Mary University of London presents work on three related projects investigating the use of computational natural language processing (NLP) to help understand and improve diagnosis and treatment for such conditions.
RavenPack's prestigious annual event has experienced growing interest, with attendance exceeding 260 buy-side professionals. Word on the street is RavenPack's research symposium is a "must attend event" for quantitative investors and financial professionals that are serious about Big Data. An excellent set of senior finance professionals shared their latest research and experience with big data and machine learning.
The growth of the Web is a success story that has spurred much research in knowledge discovery and data mining. Data mining over Web domains that are unusual is an even harder problem. There are several factors that make a domain unusual. In particular, such domains have significant long tails and exhibit concept drift, and are characterized by high levels of heterogeneity. Notable examples of unusual Web domains include both illicit domains, such as human trafficking advertising, illegal weapons sales, counterfeit goods transactions, patent trolling and cyberattacks, and also non-illicit domains such as humanitarian and disaster relief.
Deep learning has been widely successful in solving complex tasks such as image recognition (ImageNet), speech recognition, machine translation, etc. In the area of personalized recommender systems, deep learning has started showing promising advances in recent years. The key to success of deep learning in personalized recommender systems is its ability to learn distributed representations of users' and items' attributes in low dimensional dense vector space and combine these to recommend relevant items to users. To address scalability, the implementation of a recommendation system at web scale often leverages components from information retrieval systems, such as inverted indexes where a query is constructed from a user's attribute and context, learning to rank techniques. Additionally, it relies on machine learning models to predict the relevance of items, such as collaborative filtering.
Over the last decade, crowdsourcing has been used to harness the power of human computation to solve tasks that are notoriously difficult to solve with computers alone, such as determining whether or not an image contains a tree, rating the relevance of a website, or verifying the phone number of a business. The machine learning and natural language processing communities were early to embrace crowdsourcing as a tool for quickly and inexpensively obtaining the vast quantities of labeled data needed to train systems. Once this data is collected, it can be handed off to algorithms that learn to make autonomous predictions or actions. Usually this handoff is where interaction with the crowd ends. The crowd provides the data, but the ultimate goal is to eventually take humans out of the loop.
Human activity recognition (HAR) plays an important role in people's daily life by learning and identifying high-level knowledge about human activity from raw sensor inputs. Conventional pattern recognition approaches have made tremendous progress on HAR tasks by adopting machine learning algorithms such as decision tree, random forest or support vector machine, but the fast development and advancement of deep learning have overpass the accuracy of traditional machine learning results. This seminar is focused on Deep learning applied to HAR using wearable sensors. Current architectures used and how to implement them for achieving good results will be explained. Limitations and new challenges will be also discussed.
Recent years have seen a dramatic growth of natural language text data, including web pages, news articles, scientific literature, emails, enterprise documents, and social media such as blog articles, forum posts, product reviews, and tweets. This has led to an increasing demand for powerful software tools to help people manage and analyze vast amount of text data effectively and efficiently. Unlike data generated by a computer system or sensors, text data are usually generated directly by humans for humans. First, since text data are generated by people, they are especially valuable for discovering knowledge about human opinions and preferences, in addition to many other kinds of knowledge that we encode in text. Second, since text is written for consumption by humans, humans play a critical role in any text data application system, and a text management and analysis system must involve them in the loop of text analysis.
This repo contains an incremental sequence of notebooks designed to teach deep learning, Apache MXNet (incubating), and the gluon interface. Our goal is to leverage the strengths of Jupyter notebooks to present prose, graphics, equations, and code together in one place. If we're successful, the result will be a resource that could be simultaneously a book, course material, a prop for live tutorials, and a resource for plagiarising (with our blessing) useful code. To our knowledge there's no source out there that teaches either (1) the full breadth of concepts in modern deep learning or (2) interleaves an engaging textbook with runnable code. We'll find out by the end of this venture whether or not that void exists for a good reason.