Goto

Collaborating Authors

Machine Learning


Why and how should you learn "Productive Data Science"? - KDnuggets

#artificialintelligence

Data science and machine learning can be practiced with varying degrees of efficiency and productivity. Let's imagine somebody is teaching a "Productive Data Science" course or writing a book about it -- using Python as the language framework. What should the typical expectations be from such a course or book? The course/book should be intended for those who wish to leapfrog beyond the standard way of performing data science and machine learning tasks and utilize the full spectrum of the Python data science ecosystem for a much higher level of productivity. Readers should be taught how to look out for inefficiencies and bottlenecks in the standard process and how to think beyond the box.


Artificial intelligence 'is gonna ruin the world... and then we adapt'

#artificialintelligence

"AI will pull away some of the reliance on creative agencies," said Fan. Once an agency creative department crafts an original concept, "they're used for all these minor tweaks in copy, images and other things. And may in fact take into account other factors to ensure it's a compelling headline or image that's automatically selected." Fan pointed to AI-driven technology such as GPT-3 (generative pre-trained transformer 3, which is an autoregressive language model that uses deep learning to produce human-like text) as key to such advancements. With GPT-3, "you can get it to respond the way you want it to. It's almost creepy how you can have AI talk to you like a human does," he added.


Demoting Outdated 'Truth' With Machine Learning

#artificialintelligence

Sometimes the truth has an expiry date. When a time-limited claim (such as'masks are obligatory on public transport') emerges in search engine rankings, its apparent'authoritative' solution can outstay its welcome even by many years, outranking later and more accurate content on the same topic. This is a by-product of search engine algorithms' determination to identify and promote'long-term' definitive solutions, and of their proclivity to prioritize well-linked content that maintains traffic over time – and of an increasingly circumspect attitude to newer content in the emerging age of fake news. Alternately, devaluing valuable web content simply because the timestamp associated with it has passed an arbitrary'validity window' risks that a generation of genuinely useful content will be automatically demoted in favor of subsequent material that may be of a lower standard. Towards redressing this syndrome, a new paper from researchers in Italy, Belgium and Denmark has used a variety of machine learning techniques to develop a methodology for time-aware evidence ranking.


Machine Learning Consensus Clustering of Hospitalized Patients with Admission Hyponatremia

#artificialintelligence

Background: The objective of this study was to characterize patients with hyponatremia at hospital admission into clusters using an unsupervised machine learning approach, and to evaluate the short- and long-term mortality risk among these distinct clusters. Methods: We performed consensus cluster analysis based on demographic information, principal diagnoses, comorbidities, and laboratory data among 11,099 hospitalized adult hyponatremia patients with an admission serum sodium below 135 mEq/L. The standardized mean difference was utilized to identify each cluster’s key features. We assessed the association of each hyponatremia cluster with hospital and one-year mortality using logistic and Cox proportional hazard analysis, respectively. Results: There were three distinct clusters of hyponatremia patients: 2033 (18%) in cluster 1, 3064 (28%) in cluster 2, and 6002 (54%) in cluster 3. Among these three distinct clusters, clusters 3 patients were the youngest, had lowest comorbidity burden, and highest kidney function. Cluster 1 patients were more likely to be admitted for genitourinary disease, and have diabetes and end-stage kidney disease. Cluster 1 patients had the lowest kidney function, serum bicarbonate, and hemoglobin, but highest serum potassium and prevalence of acute kidney injury. In contrast, cluster 2 patients were the oldest and were more likely to be admitted for respiratory disease, have coronary artery disease, congestive heart failure, stroke, and chronic obstructive pulmonary disease. Cluster 2 patients had lowest serum sodium and serum chloride, but highest serum bicarbonate. Cluster 1 patients had the highest hospital mortality and one-year mortality, followed by cluster 2 and cluster 3, respectively. Conclusion: We identified three clinically distinct phenotypes with differing mortality risks in a heterogeneous cohort of hospitalized hyponatremic patients using an unsupervised machine learning approach.


Pinecone CEO on bringing vector similarity search to dev teams

#artificialintelligence

All the sessions from Transform 2021 are available on-demand now. The traditional way for a database to answer a query is with a list of rows that fit the criteria. If there's any sorting, it's done by one field at a time. Vector similarity search looks for matches by comparing the likeness of objects, as captured by machine learning models. Vector similarity search is particularly useful with real-world data because that data is often unstructured and contains similar yet not identical items.


What is Artificial Intelligence? -- Suffixtree

#artificialintelligence

Artificial Intelligence (AI) is the field of computer science dedicated to solving cognitive problems commonly associated with human intelligence, such as learning, problem solving, and pattern recognition. Artificial Intelligence, often abbreviated as "AI", may connote robotics or futuristic scenes, AI goes well beyond the automatons of science fiction, into the non-fiction of modern day advanced computer science. Professor Pedro Domingos, a prominent researcher in this field, describes "five tribes" of machine learning, comprised of symbolists, with origins in logic and philosophy; connectionists, stemming from neuroscience; evolutionaries, relating to evolutionary biology; Bayesians, engaged with statistics and probability; and analogizers with origins in psychology. Recently, advances in the efficiency of statistical computation have led to Bayesians being successful at furthering the field in a number of areas, under the name "machine learning". Similarly, advances in network computation have led to connectionists furthering a subfield under the name "deep learning".


NSF Artificial Intelligence Grant

#artificialintelligence

Working together, the universities will form The Institute for Learning-enabled Optimization at Scale (TILOS). The five-year research partnership will focus on the optimization of artificial intelligence (AI) and machine learning. They will work closely with industry leaders to develop optimization tools that will enable real-world improvements in key industries, including chip design, robotics, and communications networks. National University faculty will work with TILOS partners in workforce development and education to update and create career-relevant courses and modules in AI aimed at the growing population of adult learners served by the San Diego-based nonprofit institution, including career professionals, members of the military community, and underserved students in STEM). "We are thrilled to be collaborating with such incredible institutions on an initiative that directly serves our greater mission as an organization to prepare students today for the jobs of tomorrow," said National University System Chancellor and Interim President of National University Dr. Michael R. Cunningham.


GitHub Copilot Open Source Alternatives - KDnuggets

#artificialintelligence

Recently, GitHub publicly unveiled Copilot, the preview of its "AI pair programmer," a code completion style tool designed to provide line or function suggestions in your IDE. It has certainly made waves in the world of programming and beyond, and you have likely heard at least something about it. But Copilot is more than simple autocomplete and is more context aware than other code assistants. Powered by OpenAI's Codex AI system, Copilot contextualizes a situation using docstrings, function names, comments, and preceding code to best generate and suggest what it determines to be the most appropriate code. Copilot is designed to improve over time, "learning" from how developers use it.


Machine Learning Strategies: Why Businesses are Failing at it?

#artificialintelligence

In the 21st century, it is hard to imagine a life without artificial intelligence. From improving efficiencies to augment human capabilities, AI is intertwined to do anything and everything. Artificial intelligence refers to multiple technologies like machine learning, algorithm, deep learning, etc. that helps in providing significant development opportunities to businesses. Cost-effective, better customer experience, and all new features are some of the major benefits of machine learning strategies. Still, many companies are failing to develop working AI strategies because there are certain barriers, one needs to overcome before you apply the power of machine learning to your business and operations.


The Machine Learning Schools Championed by the Biggest AI Labs in the World

#artificialintelligence

I recently started an AI-focused educational newsletter, that already has over 80,000 subscribers. TheSequence is a no-BS (meaning no hype, no news etc) ML-oriented newsletter that takes 5 minutes to read. The goal is to keep you up to date with machine learning projects, research papers and concepts. Recently, one of my students asked me a question as of whether DeepMind was solely working in reinforcement learning applications. The answer is obviously no but the question is still valid as it rooted in the fact that most of DeepMind's highly publicized work such as AlphaGo, MuZero or AlphaFold are based in reinforcement learning.