Information Fusion
Automatically Utilizing Secondary Sources to Align Information Across Sources
Michalowski, Martin, Thakkar, Snehal, Knoblock, Craig A.
XML, web services, and the semantic web have opened the door for new and exciting informationintegration applications. Information sources on the web are controlled by different organizations or people, utilize different text formats, and have varying inconsistencies. Therefore, any system that integrates information from different data sources must identify common entities from these sources. Data from many data sources on the web does not contain enough information to link the records accurately using state-of-the-art record-linkage systems. However, it is possible to exploit secondary data sources on the web to improve the recordlinkage process. We present an approach to accurately and automatically match entities from various data sources by utilizing a state-of-the-art record-linkage system in conjunction with a data-integration system. The data-integration system is able to automatically determine which secondary sources need to be queried when linking records from various data sources. In turn, the record-linkage system is then able to utilize this additional information to improve the accuracy of the linkage between datasets.
An In-Depth Look at Information Fusion Rules & the Unification of Fusion Theories
This paper may look like a glossary of the fusion rules and we also introduce new ones presenting their formulas and examples: Conjunctive, Disjunctive, Exclusive Disjunctive, Mixed Conjunctive-Disjunctive rules, Conditional rule, Dempster's, Yager's, Smets' TBM rule, Dubois-Prade's, Dezert-Smarandache classical and hybrid rules, Murphy's average rule, Inagaki-Lefevre-Colot-Vannoorenberghe Unified Combination rules [and, as particular cases: Iganaki's parameterized rule, Weighting Average Operator, minC (M. Daniel), and newly Proportional Conflict Redistribution rules (Smarandache-Dezert) among which PCR5 is the most exact way of redistribution of the conflicting mass to non-empty sets following the path of the conjunctive rule], Zhang's Center Combination rule, Convolutive x-Averaging, Consensus Operator (Josang), Cautious Rule (Smets), ?-junctions rules (Smets), etc. and three new T-norm & T-conorm rules adjusted from fuzzy and neutrosophic sets to information fusion (Tchamova-Smarandache). Introducing the degree of union and degree of inclusion with respect to the cardinal of sets not with the fuzzy set point of view, besides that of intersection, many fusion rules can be improved. There are corner cases where each rule might have difficulties working or may not get an expected result.
The CIDOC Conceptual Reference Module: An Ontological Approach to Semantic Interoperability of Metadata
This article presents the methodology that has been successfully used over the past seven years by an interdisciplinary team to create the International Committee for Documentation of the International Council of Museums (CIDOC) CONCEPTUAL REFERENCE MODEL (CRM), a high-level ontology to enable information integration for cultural heritage data and their correlation with library and archive information. The CIDOC CRM is now in the process to become an International Organization for Standardization (ISO) standard. The CIDOC CRM analyzes the common conceptualizations behind data and metadata structures to support data transformation, mediation, and merging. It is assumed that the presented methodology and the upper level of the ontology are applicable in a far wider domain.
An approach to identify design and manufacturing features from a data exchanged part model
Due to the large variety of CAD systems in the market, data exchange between different CAD systems is indispensable. Currently, data exchange standards such as STEP and IGES, etc. provide a unique approach for interfacing among different CAD platforms. Once the feature-based CAD model created in one CAD system is input into another via data exchange standards, many of the original features and the feature-related information may not exist any longer. The identification of the design features and their further decomposition into machining features for the downstream activities from a data exchanged part model is a bottleneck in integrated product and process design and development. In this paper, the feature panorama is succinctly articulated from the viewpoint of product design and manufacturing.
Workshop on Intelligent Information Integration (III-99)
Fensel, Dieter, Knoblock, Craig, Kushmerick, Nicholas, Rousset, Marie-Christine
The Workshop on Intelligent Information Integration (III), organized in conjunction with the Sixteenth International Joint Conference on Artificial Intelligence, was held on 31 July 1999 in Stockholm, Sweden. Approximately 40 people participated, and nearly 20 papers were presented. This packed workshop schedule resulted from a large number of submissions that made it difficult to reserve discussion time without rejecting an unproportionately large number of papers. Participants included scientists and practitioners from industry and academia.
Workshop on Intelligent Information Integration (III-99)
Fensel, Dieter, Knoblock, Craig, Kushmerick, Nicholas, Rousset, Marie-Christine
The Workshop on Intelligent Information Integration (III), organized in conjunction with the Sixteenth International Joint Conference on Artificial Intelligence, was held on 31 July 1999 in Stockholm, Sweden. Approximately 40 people participated, and nearly 20 papers were presented. This packed workshop schedule resulted from a large number of submissions that made it difficult to reserve discussion time without rejecting an unproportionately large number of papers. Participants included scientists and practitioners from industry and academia. Topics included query planning, applications of III, mediator architectures, and the use of ontologies for III.
Probability Concepts for an Expert System Used for Data Fusion
Probability concepts for ruled-based expert systems are developed that are compatible with probability used in data fusion of imprecise information. Confidence limits are defined as being proportional to root-mean-square errors in estimates, and a method is outlined that allows the confidence limits in the probability estimate of the hypothesis to be expressed in terms of the confidence limits in the estimate of the evidence. Procedures are outlined for weighting and combining multiple reports that pertain to the same item of evidence. The illustrative examples apply to tactical data fusion, but the same probability procedures can be applied to other expert systems.