Progressive digitalization is changing the game of many industrial sectors. Focus-ing on product quality the main profitability driver of this so-called Industry 4.0 will be the horizontal integration of information over the complete supply chain. Therefore, the European RFCS project 'Quality4.0' aims in developing an adap-tive platform, which releases decisions on product quality and provides tailored information of high reliability that can be individually exchanged with customers. In this context Machine Learning will be used to detect outliers in the quality data. This paper discusses the intermediate project results and the concepts developed so far for this horizontal integration of quality information.
This paper presents a novel approach for automated analysis of process models discovered using process mining techniques. Process mining explores underlying processes hidden in the event data generated by various devices. Our proposed Inductive machine learning method was used to build business process models based on actual event log data obtained from a hotel's Property Management System (PMS). The PMS can be considered as a Multi Agent System (MAS) because it is integrated with a variety of external systems and IoT devices. Collected event log combines data on guests stay recorded by hotel staff, as well as data streams captured from telephone exchange and other external IoT devices. Next, we performed automated analysis of the discovered process models using formal methods. Spin model checker was used to simulate process model executions and automatically verify the process model. We proposed an algorithm for the automatic transformation of the discovered process model into a verification model. Additionally, we developed a generator of positive and negative examples. In the verification stage, we have also used Linear temporal logic (LTL) to define requested system specifications. We find that the analysis results will be well suited for process model repair.
Mining frequent episodes aims at recovering sequential patterns from temporal data sequences, which can then be used to predict the occurrence of related events in advance. On the other hand, gradual patterns that capture co-variation of complex attributes in the form of " when X increases/decreases, Y increases/decreases" play an important role in many real world applications where huge volumes of complex numerical data must be handled. Recently, these patterns have received attention from the data mining community exploring temporal data who proposed methods to automatically extract gradual patterns from temporal data. However, to the best of our knowledge, no method has been proposed to extract gradual patterns that regularly appear at identical time intervals in many sequences of temporal data, despite the fact that such patterns may add knowledge to certain applications, such as e-commerce. In this paper, we propose to extract co-variations of periodically repeating attributes from the sequences of temporal data that we call seasonal gradual patterns. For this purpose, we formulate the task of mining seasonal gradual patterns as the problem of mining periodic patterns in multiple sequences and then we exploit periodic pattern mining algorithms to extract seasonal gradual patterns. We discuss specific features of these patterns and propose an approach for their extraction based on mining periodic frequent patterns common to multiple sequences. We also propose a new anti-monotonous support definition associated to these seasonal gradual patterns. The illustrative results obtained from some real world data sets show that the proposed approach is efficient and that it can extract small sets of patterns by filtering numerous nonseasonal patterns to identify the seasonal ones.
The current knowledge system of macroeconomics is built on interactions among a small number of variables, since traditional macroeconomic models can mostly handle a handful of inputs. Recent work using big data suggests that a much larger number of variables are active in driving the dynamics of the aggregate economy. In this paper, we introduce a knowledge graph (KG) that consists of not only linkages between traditional economic variables but also new alternative big data variables. We extract these new variables and the linkages by applying advanced natural language processing (NLP) tools on the massive textual data of academic literature and research reports. As one example of the potential applications, we use it as the prior knowledge to select variables for economic forecasting models in macroeconomics. Compared to statistical variable selection methods, KG-based methods achieve significantly higher forecasting accuracy, especially for long run forecasts.
The TriRhenaTech alliance presents a collection of accepted papers of the cancelled tri-national 'Upper-Rhine Artificial Inteeligence Symposium' planned for 13th May 2020 in Karlsruhe. The TriRhenaTech alliance is a network of universities in the Upper-Rhine Trinational Metropolitan Region comprising of the German universities of applied sciences in Furtwangen, Kaiserslautern, Karlsruhe, and Offenburg, the Baden-Wuerttemberg Cooperative State University Loerrach, the French university network Alsace Tech (comprised of 14 'grandes \'ecoles' in the fields of engineering, architecture and management) and the University of Applied Sciences and Arts Northwestern Switzerland. The alliance's common goal is to reinforce the transfer of knowledge, research, and technology, as well as the cross-border mobility of students.
Digital transformation is one of the top priorities for industrial companies. The largest players are already moving in this direction, for many years continuously working to improve production efficiency and launching large-scale optimisation programs. They're called advanced analytics or digital innovation, and at their core, the technology could be summarised under artificial intelligence. In all cases, the efforts to utilise AI models or data analytics systems are part of a bigger digital transformation effort of the progressing companies. In an industrial context, such strategies for cost-saving and process optimisation often start from pilot projects, or top management directives for digital change guide them. In general, changes in processes or investments in capital-intensive and competitive industries require large sums of money. Traditional capital expenditures usually stretch over a long period, so a current financial standing may not allow for a complete physical overhaul of the plants or facilities. These high costs lead to the search for cheaper alternatives.
Amazon's 2019 Climate Pledge calls for a commitment to net zero carbon across their businesses by 2040. Since then, the company has reduced the weight of their outbound packaging by 33%, eliminating 915,000 tons of packaging material worldwide, or the equivalent of over 1.5 billion shipping boxes. With less packaging used throughout the supply chain, volume per shipment is reduced and transportation becomes more efficient. The cumulative impact across Amazon's enormous network is a dramatic reduction in carbon emissions. To make this happen, the customer packaging experience team partnered with AWS to build a machine learning solution powered by Amazon SageMaker.
Data mining is considered to be one of the popular terms of machine learning as it extracts meaningful information from the large pile of datasets and is used for decision-making tasks. It is a technique to identify patterns in a pre-built database and is used quite extensively by organisations as well as academia. The various aspects of data mining include data cleaning, data integration, data transformation, data discretisation, pattern evaluation and more. Below, we have listed the top eight data mining techniques in machine learning that is most used by data scientists. Association Rule Learning is one of the unsupervised data mining techniques in which an item set is defined as a collection of one or more items.
According to a Nielsen report, brick-and-mortar alcohol dollar sales were up 21% in April 2020 compared to the same period a year ago. Online alcohol sales skyrocketed by 234% over the same period in 2019. However, despite the increase, global sales are decreasing due to the shutdowns in restaurants, bars, live events and travel. Next Century Spirits is a liquor technology startup with $9.6 M in funding. The company uses big data and machine learning to create and filter bespoke distilled spirits.
The implementation of robust, stable, and user-centered data analytics and machine learning models is confronted by numerous challenges in production and manufacturing. Therefore, a systematic approach is required to develop, evaluate, and deploy such models. The data-driven knowledge discovery framework provides an orderly partition of the data-mining processes to ensure the practical implementation of data analytics and machine learning models. However, the practical application of robust industry-specific data-driven knowledge discovery models faces multiple data-- and model-development--related issues. These issues should be carefully addressed by allowing a flexible, customized, and industry-specific knowledge discovery framework; in our case, this takes the form of the cross-industry standard process for data mining (CRISP-DM). This framework is designed to ensure active cooperation between different phases to adequately address data- and model-related issues. In this paper, we review several extensions of CRISP-DM models and various data-robustness-- and model-robustness--related problems in machine learning, which currently lacks proper cooperation between data experts and business experts because of the limitations of data-driven knowledge discovery models.