Machine Learning Techniques in Data Fusion: A Review

  • Conference paper
  • First Online: 25 July 2023
  • Cite this conference paper

thesis data fusion

  • Muskan Sharma 13 ,
  • Priyanka Kushwaha 13 ,
  • Pragati Kumari 13 ,
  • Pushpanjali Kumari 13 &
  • Richa Yadav 13  

Part of the book series: Lecture Notes in Networks and Systems ((LNNS,volume 686))

Included in the following conference series:

  • International Conference on Communication and Intelligent Systems

448 Accesses

In data fusion, various data sources are unified to get more data than a single source. Due to the increasing amount of real-time data and increasing data complexity, the traditional methods are not enough in today’s world, therefore the dependency on machine learning algorithms is increasing for data fusion. Many machine learning methods have been studied before but a structural review of machine learning algorithms in data fusion is yet to come. Therefore, it is vital to analyze and sum up the various machine learning techniques used in data fusion. This paper provides a structural analysis of machine learning techniques used in data fusion. Firstly, the paper explains the framework of machine learning and data fusion followed by various standards as well as essential machine learning techniques which are used in most of the paper, and at last, a quick introduction to all typical machine learning techniques are given for better understanding.

Supported by organization IGDTUW.

Insights and solutions that were hidden before and assist teams in making better informed, and often quicker, decisions that benefit many stakeholders. Whatever the application for data fusion, one of the key challenges is ensuring that data sources speak the same language, otherwise they cannot fuse together and you do not get the benefits of one holistic image. In data fusion, various data sources are unified to get more data than a single source. In this paper data fusion and machine learning would be our main focus areas. There are many data fusion methods for separate areas, which include wireless sensor networks, image processing, radar systems, object tracking, target detection and identification, intrusion detection, situation assessment, etc. The contribution of this paper includes:

• Classification of data fusion techniques.

• Deep learning and machine learning importance for data fusion.

• Structural format of machine learning techniques in data fusion.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
  • Available as EPUB and PDF
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Elmenreich W (2022) An introduction to sensor fusion

Google Scholar  

Castanedo F (2013) A review of data fusion techniques. Sci World J. 2013:704504. https://doi.org/10.1155/2013/704504

Beddar-Wiesing S, Bieshaar M (2020) Multi-sensor data and knowledge fusion a proposal for a terminology definition. https://doi.org/10.13140/RG.2.2.26028.82567

Chen G, Liu Z, Yu G, Liang J (2021) A new view of multisensory data fusion: research on generalized fusion. Math Prob Eng 2021:1–21. https://doi.org/10.1155/2021/5471242

Maimaitijiang M, Ghulam A, Sidike P, Hartling S, Maimaitiyiming M, Peterson K, Shavers E, Fishman J, Peterson J, Kadam S, Burken J, Fritschi F (2017) Unmanned aerial system (UAS) based phenotyping of soybean using multi-sensor data fusion and extreme learning machine 43–58. 0924-2716

Blank S, Föhst T, Berns K (2010) A case study towards evaluation of redundant multi-sensor data fusion

Bouain M, Ali K, Berdjag D, Fakhfakh N, Atitallah R (2018) An embedded multi-sensor data fusion design for vehicle perception tasks. J Commun 13:18–14. https://doi.org/10.12720/jcm.13.1.8-14

Labayrade R, Perrollaz M, Gruyer D, Aubert D (2010) Sensor data fusion for road obstacle detection: a validation framework 375 X sensor data fusion for road obstacle detection: a validation framework

Sawyer G, Mason D, Hindley N, Johnson DG, Jones-Parry, IH, Oddy CJ, Pike TK, Plassard T, Rye AJ, Salabert A, Serpico S, Wielogorski AL (1992) MuSIP multi-sensor image processing system. Image Vis Comput 10:589–609. https://doi.org/10.1016/0262-8856(92)90034-Z

Xia Y, Leung H, Bossé É (2002) Neural data fusion algorithms based ona linearly constrained least square method. Neural Netw, IEEE Trans. 13:320–329. https://doi.org/10.1109/72.991418

Lau PLB, Marakkalage H, Zhou Y, UL Hassan N, Yuen C, Zhang M, Tan U-X (2019) A survey of data fusion in smart city applications. Inf Fusion 52 https://doi.org/10.1016/j.inffus.2019.05.004

Zhang X, Fan W, Guo X (2022) Urban landscape design based on data fusion and computer virtual reality technology. Wirel Commun Mob Comput 2022. https://doi.org/10.1155/2022/7207585

Chen S-Q, Ronghui Z, Hu J-M, Zhang J (2017) Feature fusion based on convolutional neural network for SAR ATR. ITM Web Conf 12:05001. https://doi.org/10.1051/itmconf/20171205001

Zhang S, Li C (2022) Research on feature fusion speech emotion recognition technology for smart teaching. Mob Inf Syst 2022:1–9. https://doi.org/10.1155/2022/7785929

Rajangam A, Rajavel R, Jacob S (2019) Fusion feature selection: new insights into feature subset detection in biological data mining. Stud Inf Control. 28:327–336. https://doi.org/10.24846/v28i3y201909

Mangai U, Samanta S, Das S, Roy Chowdhury P (2010) A survey of decision fusion and feature fusion strategies for pattern classification. IETE Tech Rev 27. https://doi.org/10.4103/0256-4602.64604

Zoghlami F, Kaden M, Villmann T, Schneider G, Heinrich H (2021) AI-based multi sensor fusion for smart decision making: a bi-functional system for single sensor evaluation in a classification task. Sensors. 21:4405. https://doi.org/10.3390/s21134405

Bigdeli B, Samadzadegan F, Reinartz P (2014) A decision fusion method based on multiple support vector machine system for fusion of hyperspectral and LIDAR data. Int J Image Data Fusion. 5:196–209. https://doi.org/10.1080/19479832.2014.919964

Biagetti G, Crippa P, Falaschetti L, Focante E, Martínez MN, Seepold R (2020) Machine learning and data fusion techniques applied to physical activity classification using photoplethysmographic and accelerometric signals. Procedia Comput Sci 176:3103–3111. https://doi.org/10.1016/j.procs.2020.09.178

AbdElRahman M, El-Bastawissy A, Hegazy O (2013) A statistical datafusion technique in virtual data integration environment. Int J Data Min Knowl Manag Process 3. https://doi.org/10.5121/ijdkp.2013.3503

Dong X, Gabrilovich E, Heitz G, Horn W, Murphy K, Sun S, Zhang W (2015) From data fusion to knowledge fusion. Proc VLDB Endowment 7. https://doi.org/10.14778/2732951.2732962

Salustiano RE, Reis C (2006) Signal-level sensor fusion applied to monitoring environment conditions, 261–265. https://doi.org/10.1109/ICCDCS.2006.250871

Zheng S (2022) Pixel-level image fusion algorithms for multi-camera imaging system

Wang J, Gao Y, Kumar A, Kim H-J (2019). An intelligent data gathering schema with data fusion supported for mobile sink in wireless sensor networks. Int J Distrib Sens Netw 15:155014771983958. https://doi.org/10.1177/1550147719839581

Khan M, Ghazal T, Lee S-W, Sakhawat A (2021) Data fusion-based machine learning architecture for intrusion detection. Comput Mater Continua 70:3399–3413. https://doi.org/10.32604/cmc.2022.020173

Liggins ME, Chong C-Y, Kadar I, Alford M, Vannicola V, Thomopoulos S (1997) Distributed fusion architectures and algorithms for target tracking. Proc IEEE 85:95–107. https://doi.org/10.1109/JPROC.1997.554211

Dautov R, Distefano S (2017) Distributed data fusion for the Internet of Things. 427–432. https://doi.org/10.1007/978-3-319-62932-2_41

Zhang L, Tamminedi T, Ganguli A, Yosiphon G, Yadegar J (2010) Hierarchical multiple sensor fusion using structurally learned Bayesian network. 174183. https://doi.org/10.1145/1921081.1921102

Taylor C, Bishop A (2018) Homogeneous functionals and Bayesian data fusion with unknown correlation. Inf Fusion 45. https://doi.org/10.1016/j.inffus.2018.02.002

Kashinath S, Mostafa S, Mustapha A, Mahdin H, Lim D, Mahmoud M, Mohammed M, Al-rimy B, Md Fudzee Mohd Farhan, Yang T (2021) Review of data fusion methods for real time and multi-sensor traffic flow analysis. IEEE Access, p 1. https://doi.org/10.1109/AC-CESS.2021.3069770

Elmenreich W, Leidenfrost R (2008) Fusion of heterogeneous sensors data. In: Proceedings of the 6th workshop on intelligent solutions in embedded systems, WISES’08, pp 1–10. https://doi.org/10.1109/WISES.2008.4623307

Khan Md, Anwar S (2019) Time-domain data fusion using weighted evidence and Dempster–Shafer combination rule: application in object classification. Sensors 19:5187. https://doi.org/10.3390/s19235187

Chen H, Sun Q, Xu L, Xiong Z (2013) Application of spatial data fusion in the production and updating of spatial data. In: ISPRS—international archives of the photogrammetry, remote sensing and spatial information sciences. XL-7/W1. 510. https://doi.org/10.5194/isprsarchives-XL-7-W1-5-2013

Nguyen H, Cressie N, Braverman A (2017) Multivariate spatial data fusion for very large remote sensing datasets. Remote Sens. 9:142. https://doi.org/10.3390/rs9020142

Wiemann S, Bernard L (2015) Spatial data fusion in spatial data infrastructures using linked data. Int J Geogr Inf Sci 30:1–24. https://doi.org/10.1080/13658816.2015.1084420

Wang X, Wang X (2020) Spatiotemporal fusion of remote sensing image based on deep learning. J Sens 2020:1–11. https://doi.org/10.1155/2020/8873079

Fawzy D, Moussa S, Badr N (2021) The spatiotemporal data fusion (STDF) approach: IoT-based data fusion using Big Data analytics. Sensors. 21:7035. https://doi.org/10.3390/s21217035

Foo PH, Ng G-W (2013) High-level information fusion: an overview. J Adv Inf Fusion 8:33–72

Lytrivis P, Thomaidis G, Amditis A (2009) Sensor data fusion in automotive applications. https://doi.org/10.5772/6574

Dhamodharavadhani S, Ramalingam R (2019). Enhanced-logistic-regression (ELR)-model-for-Big-Data. https://doi.org/10.4018/978-1-7998-0106-1.ch008

Evgeniou T, Pontil M (2001) Support vector machines: theory and applications. 2049:249–257. https://doi.org/10.1007/3-540-44673-7_12

Miura K (2011) An introduction to maximum likelihood estimation and information geometry. Interdisc Inf Sci (IIS). 17. https://doi.org/10.4036/iis.2011.155

Lalaoui L, Mohamadi T (2013) A comparative study of image region based segmentation algorithms. Int J Adv Comput Sci Appl 4. https://doi.org/10.14569/IJACSA.2013.040627

Balakrishna S, Solanki V, Kumar R, Thirumaran M (2020) Survey on machine learning-based clustering algorithms for IoT data cluster analysis. 1195–1204. https://doi.org/10.1007/978-981-15-2780-7_123

Bondarenko A, Borisov A (2013) Research on deep belief networks classification abilities on small and medium datasets. Int J Inf Technol Manag 16:60–65. https://doi.org/10.2478/itms-2013-0009

Tong Z, Xu P, Denœux T (2021) An evidential classifier based on Dempster-Shafer theory and deep learning

Natekin A, Knoll A (2013) Gradient boosting machines, a tutorial. Front Neurorobot 7:21. https://doi.org/10.3389/fnbot.2013.00021

Müller M (2004). Generalized Linear Models. https://doi.org/10.1007/978-3-642-21551-3_24

Article   Google Scholar  

Rezatofighi H, Milan A, Zhang Z, Shi Q, Dick A, Reid I (2015) Joint probabilistic data association revisited. https://doi.org/10.1109/ICCV.2015.349

Guo G, Wang H, Bell D, Bi Y (2004) KNN model-based approach in classification

Tharwat A, Gaber T, Ibrahim A, Hassanien AE (2017) Linear discriminant analysis: a detailed tutorial. Ai Commun 30:169–190,. https://doi.org/10.3233/AIC-170729

Vardeman SB, Morris M (2013) Majority voting by independent classifiers can increase error rates. Am Stat 67:94–96. https://doi.org/10.1080/00031305.2013.778788

Alrammal M, Naveed Munir (2020) Monte-Carlo based reinforcement learning (MCRL). Int J Mach Learn Comput 10:227–232. https://doi.org/10.18178/ijmlc.2020.10.2.924

Chen C, Cao X, Tian L (2019) Partial least squares regression performs well in MRI-based individualized estimations. Front Neurosci 13. https://doi.org/10.3389/fnins.2019.01282

Villmann T, Engelsberger A, Ravichandran J, Villmann A, Kaden M (2022) Quantum-inspired learning vector quantizers for prototypebased classification. Neural Comput Appl 34:1–10. https://doi.org/10.1007/s00521020-05517-y

Ali J, Khan R, Ahmad N, Maqsood I (2012) Random forests and decision trees. Int J Comput Sci Issues (IJCSI) 9

Montella C (2011) The Kalman filter and related algorithms: a literature review

Howley T, Madden M, O’Connell M-L, Ryder A (2005) The effect of principal component analysis on machine learning accuracy with high dimensional spectral data, 209–222. https://doi.org/10.1007/1-84628-224-1_16

Bianchi G, Heo C (2021) A Bayesian statistics approach to hospitality research. Curr Issues Tourism 24:1–10. https://doi.org/10.1080/13683500.2021.1896486

Sharkawy A-N (2020) Principle of neural network and its main types: review. J Adv Appl Comput Math 7:8–19. https://doi.org/10.15377/2409-5761.2020.07.2

Download references

Acknowledgements

We are grateful to our renowned college Indira Gandhi Delhi Technical University for Women for motivating us and providing us with all the much-needed guidance and mentorship.

Author information

Authors and affiliations.

Indira Gandhi Delhi Technical University for Women, New Delhi, India

Muskan Sharma, Priyanka Kushwaha, Pragati Kumari, Pushpanjali Kumari & Richa Yadav

You can also search for this author in PubMed   Google Scholar

Corresponding author

Correspondence to Muskan Sharma .

Editor information

Editors and affiliations.

Department of Computer Science and Engineering, Rajasthan Technical University, Kota, India

Harish Sharma

National Institute of Technology Delhi, New Delhi, Delhi, India

Vivek Shrivastava

Indian Institute of Information Technology, Design and Manufacturing, Jabalpur, Madhya Pradesh, India

Kusum Kumari Bharti

School of Electrical and Electronic Engineering, Nanyang Technological University, Singapore, Singapore

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Cite this paper.

Sharma, M., Kushwaha, P., Kumari, P., Kumari, P., Yadav, R. (2023). Machine Learning Techniques in Data Fusion: A Review. In: Sharma, H., Shrivastava, V., Bharti, K.K., Wang, L. (eds) Communication and Intelligent Systems. ICCIS 2022. Lecture Notes in Networks and Systems, vol 686. Springer, Singapore. https://doi.org/10.1007/978-981-99-2100-3_31

Download citation

DOI : https://doi.org/10.1007/978-981-99-2100-3_31

Published : 25 July 2023

Publisher Name : Springer, Singapore

Print ISBN : 978-981-99-2099-0

Online ISBN : 978-981-99-2100-3

eBook Packages : Intelligent Technologies and Robotics Intelligent Technologies and Robotics (R0)

Share this paper

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

  • Publish with us

Policies and ethics

  • Find a journal
  • Track your research

MIT Libraries home DSpace@MIT

  • DSpace@MIT Home
  • MIT Libraries
  • Doctoral Theses

Multimodal Data Fusion for Estimating Electricity Access and Demand

Thumbnail

Terms of use

Date issued, collections.

Restricted to current U-M faculty, staff, and students

  •   Home
  • Research Collections
  • Dissertations and Theses (Ph.D. and Master's)

Informative Data Fusion: Beyond Canonical Correlation Analysis

PDF file

Collections

Remediation of harmful language.

The University of Michigan Library aims to describe library materials in a way that respects the people and communities who create, use, and are represented in our collections. Report harmful or offensive language in catalog records, finding aids, or elsewhere in our collections anonymously through our metadata feedback form . More information at Remediation of Harmful Language .

Accessibility

If you are unable to use this file in its current format, please select the Contact Us link and we can modify it to make it more accessible to you.

  • Northeastern University
  • College of Engineering
  • Department of Electrical and Computer Engineering
  • Electrical and Computer Engineering Theses and Dissertations
  • Electrical and Computer Engineering Dissertations
  • Electrical Engineering Dissertations

Bayesian data fusion for distributed learning

Your file is being processed

Return to Electrical Engineering Dissertations

Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.

  • View all journals
  • My Account Login
  • Explore content
  • About the journal
  • Publish with us
  • Sign up for alerts
  • Review Article
  • Open access
  • Published: 16 October 2020

Fusion of medical imaging and electronic health records using deep learning: a systematic review and implementation guidelines

  • Shih-Cheng Huang   ORCID: orcid.org/0000-0001-9882-833X 1 , 2   na1 ,
  • Anuj Pareek   ORCID: orcid.org/0000-0002-1526-3685 2 , 3   na1 ,
  • Saeed Seyyedi 2 , 3 ,
  • Imon Banerjee   ORCID: orcid.org/0000-0002-3327-8004 2 , 4 , 5 &
  • Matthew P. Lungren 1 , 2 , 3  

npj Digital Medicine volume  3 , Article number:  136 ( 2020 ) Cite this article

41k Accesses

256 Citations

82 Altmetric

Metrics details

  • Data integration
  • Machine learning
  • Medical imaging

Advancements in deep learning techniques carry the potential to make significant contributions to healthcare, particularly in fields that utilize medical imaging for diagnosis, prognosis, and treatment decisions. The current state-of-the-art deep learning models for radiology applications consider only pixel-value information without data informing clinical context. Yet in practice, pertinent and accurate non-imaging data based on the clinical history and laboratory data enable physicians to interpret imaging findings in the appropriate clinical context, leading to a higher diagnostic accuracy, informative clinical decision making, and improved patient outcomes. To achieve a similar goal using deep learning, medical imaging pixel-based models must also achieve the capability to process contextual data from electronic health records (EHR) in addition to pixel data. In this paper, we describe different data fusion techniques that can be applied to combine medical imaging with EHR, and systematically review medical data fusion literature published between 2012 and 2020. We conducted a systematic search on PubMed and Scopus for original research articles leveraging deep learning for fusion of multimodality data. In total, we screened 985 studies and extracted data from 17 papers. By means of this systematic review, we present current knowledge, summarize important results and provide implementation guidelines to serve as a reference for researchers interested in the application of multimodal fusion in medical imaging.

Similar content being viewed by others

thesis data fusion

Multimodal fusion with deep neural networks for leveraging CT imaging and electronic health record: a case-study in pulmonary embolism detection

thesis data fusion

Artificial intelligence-based methods for fusion of electronic health records and imaging data

thesis data fusion

Designing clinically translatable artificial intelligence systems for high-dimensional medical imaging

Introduction.

The practice of modern medicine relies heavily on synthesis of information and data from multiple sources; this includes imaging pixel data, structured laboratory data, unstructured narrative data, and in some cases, audio or observational data. This is particularly true in medical image interpretation where substantial clinical context is often essential to provide diagnostic decisions. For example, it has repeatedly been shown that a lack of access to clinical and laboratory data during image interpretation results in lower performance and decreased clinical utility for the referring provider 1 , 2 . In a survey of radiologists, the majority (87%) stated that clinical information had a significant impact on interpretation 3 . The importance of clinical context for accurate interpretation of imaging data is not limited to radiology; instead many other imaging-based medical specialties such as pathology, ophthalmology, and dermatology, also rely on clinical data to guide image interpretation in practice 4 , 5 , 6 . Pertinent and accurate information regarding the current symptoms and past medical history enables physicians to interpret imaging findings in the appropriate clinical context, leading to a more relevant differential diagnosis, a more useful report for the physicians, and optimal outcome for the patient.

In the current digital era, the volume of radiological imaging exams is growing. To meet this increased workload demand, an average radiologist may have to interpret an image every 3–4 s over an 8-h workday which contributes to fatigue, burnout, and increased error-rate 7 . Deep learning in healthcare is proliferating due to the potential for successful automated systems to either augment or offload cognitive work from busy physicians 8 , 9 , 10 . One class of deep learning, namely convolutional neural networks (CNN) has proven very effective for image recognition and classification tasks, and are therefore often applied to medical images. Early applications of CNNs for image analysis in medicine include diabetic retinopathy, skin cancer, and chest X-rays 11 , 12 , 13 , 14 , 15 , 16 , 17 , 18 . Yet, these models consider only the pixel data as a single modality for input and cannot contextualize other clinical information as would be done in medical practice, therefore may ultimately limit clinical translation.

As an example consider the “simple” task in radiology of identifying pneumonia on a chest radiograph, something that has been achieved by many investigators training deep learning models for automated detection and classification of pathologies on chest X-rays 19 , 20 . Yet without clinical context such as patient history, chief complaint, prior diagnoses, laboratory values, such applications may ultimately have limited impact on clinical practice. The imaging findings on chest X-rays consistent with pneumonia, despite having imaging features that can generally differentiate alternative diagnoses, are nonspecific and accurate diagnosis requires the context of clinical and laboratory data. In other words, the chest X-ray findings that suggest pneumonia would be accurate in one person with fever and an elevated white blood cell count but in another patient without those supporting clinical characteristics and laboratory values, similar imaging finding may instead represent other etiologies such as atelectasis, pulmonary edema, or even lung cancer. There are countless examples across different medical fields in which clinical context, typically in the form of structured and unstructured clinical data from the electronic health record (EHR), is critical for accurate and clinically relevant medical imaging interpretation. As with human physicians, automated detection and classification systems that can successfully utilize both medical imaging data together with clinical data from the EHR, such as patient demographics, previous diagnoses and laboratory values, may lead to better performing and more clinically relevant models.

Multimodal deep learning models that can ingest pixel data along with other data types (fusion) have been successful in applications outside of medicine, such as autonomous driving and video classification. As an example, a multimodal fusion detection system for autonomous vehicles, that combines visual features from cameras along with data from Light Detection and Ranging (LiDAR) sensors, is able to achieve significantly higher accuracy (3.7% improvement) than a single-modal CNN detection model 21 . Similarly, a multimodal social media video classification pipeline leveraging both visual and textual features increased the classification accuracy to 88.0%, well above single modality neural networks such as Google’s InceptionV3 which reached an accuracy of 76.4% on the same task 22 . The improvements in performance for these efforts not only echo the justification in medical applications, leveraging fusion strategies for medical imaging is also primarily motivated by the desire to integrate complementary contextual information and overcome the limitation of image-only models.

The recent medical imaging literature shows a similar trend where both EHR and pixel data are leveraged in a “fusion-paradigm” for solving complex tasks which cannot readily be tackled by a single modality (Fig. 1 ). The new fusion paradigm covers a wide range of methodologies and techniques with varying terms and model architectures that have not been studied systematically. The purpose of this review paper is to present a comprehensive analysis of deep learning models that leverage multiple modalities for medical imaging tasks, define and consolidate relevant terminology, and summarize the results from state-of-the-art models in relevant current literature. We hope this review can help inform future modeling frameworks and serve as a reference for researchers interested in the application of multimodal fusion in medical imaging.

figure 1

Timeline showing growth in publications on deep learning for medical imaging, found by using the same search criteria on PubMed and Scopus. The figure shows that fusion has only constituted a small, but growing, subset of medical deep learning literature.

Terminology and strategies in fusion

Data fusion refers to the process of joining data from multiple modalities with the aim of extracting complementary and more complete information for better performing machine learning models as opposed to using a single data modality.

Figure 2 illustrates the three main different fusion strategies, namely early, joint, and late fusion. Here we define and describe each fusion strategy in detail:

figure 2

Model architecture for different fusion strategies. Early fusion (left figure) concatenates original or extracted features at the input level. Joint fusion (middle figure) also joins features at the input level, but the loss is propagated back to the feature extracting model. Late fusion (right figure) aggregates predictions at the decision level.

Early fusion 23 , commonly known as feature level fusion, refers to the process of joining multiple input modalities into a single feature vector before feeding into one single machine learning model for training (Fig. 2 Early Fusion). Input modalities can be joined in many different ways, including concatenation, pooling or by applying a gated unit 23 , 24 . Fusing the original features represents early fusion type I, while fusing extracted features, either from manual extraction, imaging analysis software or learned representation from another neural network represents early fusion type II. We consider predicted probabilities to be extracted features, thus fusing features with predicted probabilities from different modalities is also early fusion type II.

Joint fusion (or intermediate fusion) is the process of joining learned feature representations from intermediate layers of neural networks with features from other modalities as input to a final model. The key difference, compared to early fusion, is that the loss is propagated back to the feature extracting neural networks during training, thus creating better feature representations for each training iteration (Fig. 2 Joint Fusion). Joint fusion is implemented with neural networks due to their ability to propagate loss from the prediction model to the feature extraction model(s). When feature representations are extracted from all modalities, we consider this joint fusion type I. However, not all input features require the feature extraction step to be defined as joint fusion (Fig. 2 Joint Fusion—Type II).

Late fusion 23 refers to the process of leveraging predictions from multiple models to make a final decision, which is why it is often known as decision-level fusion (Fig. 2 Late Fusion). Typically, different modalities are used to train separate models and the final decision is made using an aggregation function to combine the predictions of multiple models. Some examples of aggregation functions include: averaging, majority voting, weighted voting or a meta-classifier based on the predictions from each model. The choice of the aggregation function is usually empirical, and it varies depending on the application and input modalities.

A total of 985 studies were identified through our systematic search. After removing duplicates and excluding studies based on title and abstract using our study selection criteria (see Methods), 44 studies remained for full-text screening. A total of 17 studies fulfilled our eligibility criteria and were included for systematic review and data extraction. The studies were in English except for a single paper in Chinese. Figure 3 presents a flowchart of the study screening and selection process and Table 1 displays the included studies and extracted data.

figure 3

Two authors independently screened all records for eligibility. Seventeen studies were included in the systematic review.

Early fusion

The majority of the studies that remained after our full-text screening (11/17) used early fusion to join the multimodal input. Thung et al. 25 conducted image-image fusion of PET and MRI images using a joint fusion approach, but since they concatenated clinical and imaging features into one single feature vector before feeding into their neural network, we categorized their approach as early fusion. Six out of eleven early fusion studies extracted features from medical imaging using a CNN (Table 1 ). Four out of the six studies that applied neural networks for feature extraction simply concatenated the extracted imaging features with clinical features for their fusion strategy 26 , 27 , 28 , 29 . The remaining two studies by Liu et al. 30 and Nie et al. 31 applied dimensionality reduction techniques before concatenating the features. Five studies used software generated and/or manually extracted features from medical imaging before fusing with clinical data. Software-based feature extraction included radiomics features such as skewness and kurtosis 32 or volume and thickness quantification of the regions of interest 25 , 33 . Manually extracted features included radiological assessments such as size, angle, and morphology of anatomical structures 34 . Out of these five studies, two applied feature selection strategies to reduce the feature dimension and improve predictive performance. The employed feature selection strategies included a rank-based method using Gini coefficients 32 , a filter-based method based on mutual information of the features 35 , and a genetic-algorithm based method 35 . Seven of the early fusion studies compared the performance of their fusion models against single modality models (Table 1 ). Six of these studies showed an improvement in performance when using fusion 25 , 26 , 28 , 29 , 31 , 33 , and the remaining one achieved the same performance but reduced standard deviation 27 , alluding to a model with better stability.

Joint fusion

Joint fusion was used in four out of the seventeen studies. Spasov et al. 36 , Yala et al. 37 , and Yoo et al. 38 implemented CNNs to learn image features and fused these feature representations with clinical features before feeding them into a feed-forward neural network. Spasov et al. and Yala. et al. both used simple concatenation to fuse the learned imaging and clinical features. To cater to the differences between the dimensionality and dynamic range between the imaging and clinical features, Yoo et al. replicated and scaled their clinical features before fusion and they observed improvements in performances. Kawahara et al. 39 also used CNNs as feature extractors for imaging modalities but experimented with a unique multimodal multi-task loss function that considers multiple combinations of the input modalities. The predicted probabilities of these multi-task outputs were aggregated for prediction, but we do not consider this late fusion since the probabilities were not from separate models. Kawahara et al., Yala et al. and Yoo et al. reported an improvement in performance using fusion compared to image-only models (Table 1 ). Yoo et al. further compared their joint fusion model to a late fusion model and achieved a 0.02 increase in Area Under Receiver Operating Characteristic Curve (AUROC).

Late fusion

Late fusion was used in three out of the seventeen included studies (Table 1 ). Each of the three late fusion papers applied a different type of aggregation strategy. Yoo et al. 38 took the mean of the predicted probabilities from two single modality models as the final prediction. Reda et al. 40 built another classifier using the single modality models’ prediction probabilities as inputs. Qiu et al. 41 trained three independent imaging models that took as input a single MRI slice, each from a specific anatomical location. Max, mean and majority voting were applied to aggregate predictions from the three imaging models. The results from the three aggregation methods were combined again by majority voting before another round of late fusion with the clinical models. All late fusion models showed improvements in performances when compared to models that used only single modalities.

The purpose of this review is to aggregate the collective knowledge of prior work applying multimodal deep learning fusion techniques that combine medical imaging with clinical data. We propose consistent terminology for multimodal fusion techniques and categorize prior work by fusion strategy. Overall, we found that multimodality fusion models generally led to increased accuracy (1.2–27.7%) and AUROC (0.02–0.16) over traditional single modality models for the same task. However, no single fusion strategy consistently led to optimal performance across all domains. Since our literature review shows that additional patient information and clinical context can result in better model performance, and fusion methods better replicate the human expert interpretation workflow, it is recommended to always experiment with fusion strategies when multimodal data is available.

The deep learning fusion models reviewed represent a spectrum of medical applications ranging from radiology 31 to hematology 29 . For example, fusion strategies were often applied to the diagnosis and prediction of Alzheimer’s disease 25 , 28 , 33 , 36 , 41 . In clinical practice, neither imaging nor clinical data alone are sufficient for the diagnosis of Alzheimer’s disease. Leveraging deep learning fusion techniques consistently showed improvements in performance for Alzheimer’s disease diagnosis, while physicians struggle with accurate and reliable diagnostics even when multimodality is present, as proven by histopathological correlation 42 . This highlights the importance and utility of multimodal fusion techniques in clinical applications.

Fusion approaches in other less complex clinical applications also improved performance over single modality models, even those in which single modality models have been widely reported to achieve high performance, such as pixel-based models for automated skin cancer detection 43 . While the fusion approach varied widely, the consistent improvement in reported performance across a wide variety of clinical use cases suggests that model performance based on single-modal data may not represent state of the art for a given application when multimodal data are not considered.

The complexity of the non-imaging data in multimodal fusion work was limited, particularly in the context of available feature-rich and time-series data in the EHR. Instead, most studies focused primarily on basic demographic information such as age and gender 25 , 27 , 39 , a limited range of categorical clinical history such as hypertension or smoking status 32 , 34 or disease-specific clinical features known to be strongly associated with the disease of interest such as APOE4 for Alzheimer’s 25 , 28 , 33 , 36 or PSA blood test for prediction of prostate cancer 40 . While selecting features known to be associated with disease is meaningful, future work may further benefit from utilizing large volumes of feature-rich data, as seen in fields outside medicine such as autonomous driving 44 , 45 .

Implementation guidelines for fusion models

In most applications early fusion was used as the first attempt for multimodal learning, a straightforward approach that does not necessarily require training multiple models. However, when the input modalities are not in the same dimensions, which is typical when combining clinical data represented in 1D with imaging data in 2D or 3D, then high-level imaging features must be extracted as a 1D vector before fusing with the 1D clinical data. There were a variety of strategies used to accomplish this; including using manually extracted imaging features or software-generated features 25 , 32 , 33 , 34 , 35 . It is worth noting, that unless there is a compelling reason for using such an approach, outputs from linear layers of a CNN are usually effective feature representations of the original image 28 , 29 , 31 . This is because learned features representations often result in much better task-specific performance than can be obtained with manual or software extracted features 46 . Based on the reviewed papers, early fusion consistently improved performance over single modality models, and is supported by this review as an initial strategy to fuse multimodal data.

When using CNNs to extract features from imaging modalities, the same CNNs can also be used in joint fusion. However, joint fusion is implemented using neural networks which can be a limitation especially with smaller datasets better suited for traditional machine learning models. For example, if there are disproportionately few samples relative to the number of features in the dataset or if some of the input features are sparsely represented, early or late fusion is preferred because they can be implemented with traditional machine learning algorithms (e.g., Lasso and ElasticNet 47 ) that are better suited for this type of data 48 . Nevertheless, joint and early fusion neural networks are both able to learn shared representations, making it easier for the model to learn correlations across modalities, thereby resulting in better performance 49 . Studies have also shown that fusing highly correlated features in earlier layers and less correlated features in deeper layers improve model performance 50 , 51 . In addition, we suspect that joint fusion models have the potential to outperform other fusion strategies, as the technique iteratively updates its feature representations to better complement each modality through simultaneous propagation of the loss to all feature extracting models. Yet to date, there is insufficient evidence to systematically assess this effect in fusion for medical imaging and is an important area for future exploration.

When signals from different modalities do not complement each other, that is to say input modalities separately inform the final prediction and do not have inherent interdependency, then trying a late fusion approach is preferred. This is chiefly because when feature vectors from multiple modalities are concatenated, such as in early and joint fusion, high-dimensional vectors are generated which can be difficult for machine learning models to learn without overfitting, unless a large number of input samples are available. This is the so-called “curse of dimensionality” in machine learning 52 , 53 . Late fusion mitigates this problem by utilizing multiple models that are each specialized on a single modality, thus limiting the input feature vector size for each model. For example, the quantitative result of a Mini Mental State Examination and the pixel data obtained from a brain MRI (e.g., Qiu et al. 41 ) are largely independent data, and would therefore be suitable candidates for input into late fusion models.

Furthermore, in the common real-world scenario of missing or incomplete data, i.e. some patients have only clinical data available but no imaging data or vice-versa, late fusion retains the ability to make predictions. This is because late fusion employs separate models for separate modalities, and aggregation functions such as majority voting and averaging can be applied even when predictions from a modality is missing. When the different input modalities have very different numbers of features, predictions might be overly influenced by the most feature-rich modality (e.g., Reda et al. 40 ). Late fusion is favorable in this scenario as it considers each modality separately. Yoo et al. 38 also showed that repeating or scaling the modality that has fewer features before fusion achieved a boost in the model’s performance. Nonetheless, joint fusion can also be tuned to mitigate the difference in number of features, by setting feature producing linear layers of the feature extraction model to output a similar number of features as the other modalities. Our recommendations are summarized in Table 2 .

Ideally, researchers want to first build and optimize single modality models to dually serve as baseline models and provide inputs to fusion models. Multiple fusion strategies can then be implemented to compare model performance and guide subsequent fusion experiments. Since better performance is consistently achieved with multimodal fusion techniques, routine best practice should include reporting of the systematic investigation of various fusion strategies in addition to deep learning architectures and hyperparameters.

Limitations

We devised our search string to only consider papers after 2012. This constitutes a limitation as we excluded earlier papers that applied fusion using traditional machine learning techniques or simple feed-forward neural networks. Publication bias is an important limitation since positive results can be disproportionately reported in the published literature, which may have the aggregate effect of overrepresenting the advantages of fusion techniques. Furthermore, using our study selection criteria, we only looked at fusion techniques applied to clinical prediction and diagnosis, but we recognize that fusion can be applied to other interesting medical tasks such as segmentation and registration.

As the included studies investigate different objectives, use different input modalities, report different performance metrics, and not all papers provide confidence bounds, we are not able to aggregate or statistically compare the performance gains in a meta-analysis. In addition, the reported metrics cannot always be considered valid, since some studies didn’t use an independent test-set for an unbiased performance estimate 29 , 40 . The limited number of studies per medical field and the heterogeneity of each study also makes it difficult to compare the studies qualitatively. A few studies implemented fusion in unconventional ways, which may introduce subjectivity when we classify each study into early, late, and joint fusion.

Future research

This systematic review found that multimodal fusion in medicine is a promising yet nascent field that complements the clinical practice of medical imaging interpretation across all disciplines. We have defined and summarized key terminology, techniques, and evaluated the state of the art for multimodal fusion in medical imaging, honing in on key insights and unexplored questions to guide task and modality-specific strategies. The field of multimodal fusion for deep learning in medical imaging is expanding and novel fusion methods are expected to be developed. Future work should focus on shared terminology and metrics, including direct evaluation of different multimodal fusion approaches when applicable. We found that multimodal fusion for automated medical imaging tasks broadly improves the performance over single modality models, and further work may discover additional insights to inform optimal approaches.

This systematic review was conducted based on the PRISMA guidelines 54 .

Search strategy

A systematic literature search was implemented in PubMed and Scopus under the supervision of a licensed librarian. The key search terms included a combination of the three major themes: ‘deep learning’, ‘multimodality fusion’, and ‘medical imaging’. Terms for segmentation, registration, and reconstruction were used as exclusion criteria in the search. The search encompassed papers published between 2012 and 2020. This range was considered appropriate due to the rise in popularity in applying CNN on medical images since the 2012 ImageNet challenge. The complete search string for both databases is provided in Supplementary Methods . For potentially eligible studies cited by articles already included in this review, additional targeted free-text searches were conducted on Google Scholar if they did not appear in Scopus or PubMed.

We included all research articles in all languages that applied deep learning models for clinical outcome prediction or diagnosis using a combination of medical imaging modalities and EHR data. Studies specific to deep learning were included rather than the broader field of machine learning because deep learning has consistently shown superior performance in image-related tasks. We selected only studies that fused medical imaging with EHR data since, unlike image-image fusion, this is an exciting new technique that effectively merges heterogeneous data types and adds complementary rather than overlapping information to inform prediction and diagnosis. We defined medical imaging modalities as any type of medical images used in clinical care. Studies that used deep learning only for feature extractions were also included for our review. We excluded any study that combined extracted imaging features with the original imaging modality, as we still considered this a single modality. Articles that fused multimodal data for segmentation, registration or reconstruction were also excluded due to our criteria for outcome prediction and diagnosis. Articles from electronic preprint archives such as ArXiv were excluded in order to ensure only papers that passed peer-review were included. Lastly, papers with poor quality that hindered our ability to meaningfully extract data were also excluded.

Study selection

The Covidence software ( www.covidence.org ) was used for screening and study selection. After removal of duplicates, studies were screened based on title and abstract, and then full-texts were obtained and assessed for inclusion. Study selection was performed by two independent researchers (S.-C.H. and A.P.), and disagreements were resolved through discussion. In cases where consensus could not be achieved a third researcher was consulted (I.B.).

Data extraction

For benchmarking the existing approaches we extracted the following data from each of the selected articles: (a) fusion strategy, (b) year of publication, (c) authors, (d) clinical domain, (e) target outcome, (f) fusion details, (g) imaging modality, (h) non-imaging modality, (i) number of samples, and (j) model performance (Table 1 ). We classified the specific fusion strategy based on the definitions in the section “Terminology and strategies in fusion”. The number of samples reported is the full data-size including training, validation and testing data. For classification tasks we extracted AUROC whenever this metric was reported, otherwise we extracted accuracy. When the article contained several experiments, metrics from the experiment with the best performing fusion model were extracted. These items were extracted to enable researchers to find and compare current fusion studies in their medical field or input modalities of interest.

Data availability

The authors declare that all data supporting the findings of this study are available within the paper and its Supplementary information files.

Leslie, A., Jones, A. J. & Goddard, P. R. The influence of clinical information on the reporting of CT by radiologists. Br. J. Radiol. 73 , 1052–1055 (2000).

Article   CAS   PubMed   Google Scholar  

Cohen, M. D. Accuracy of information on imaging requisitions: does it matter? J. Am. Coll. Radiol. 4 , 617–621 (2007).

Article   PubMed   Google Scholar  

Boonn, W. W. & Langlotz, C. P. Radiologist use of and perceived need for patient data access. J. Digit. Imaging 22 , 357–362 (2009).

Comfere, N. I. et al. Provider-to-provider communication in dermatology and implications of missing clinical information in skin biopsy requisition forms: a systematic review. Int. J. Dermatol. 53 , 549–557 (2014).

Comfere, N. I. et al. Dermatopathologists’ concerns and challenges with clinical information in the skin biopsy requisition form: a mixed-methods study: Clinical information in dermatopathology. J. Cutan. Pathol. 42 , 333–345 (2015).

Article   PubMed   PubMed Central   Google Scholar  

Jonas, J. B. et al. Glaucoma. The Lancet 390 , 2183–2193 (2017).

Article   Google Scholar  

McDonald, R. J. et al. The effects of changes in utilization and technological advancements of cross-sectional imaging on radiologist workload. Acad. Radiol. 22 , 1191–1198 (2015).

Dean, N. C. et al. Impact of an electronic clinical decision support tool for emergency department patients with pneumonia. Ann. Emerg. Med. 66 , 511–520 (2015).

Banerjee, I. et al. Development and Performance of the Pulmonary Embolism Result Forecast Model (PERFORM) for Computed Tomography Clinical Decision Support. JAMA Netw. Open 2 , e198719 (2019).

Sandeep Kumar, E. & Jayadev, P. S. Deep Learning for Clinical Decision Support Systems: A Review from the Panorama of Smart Healthcare. In Deep Learning Techniques for Biomedical and Health Informatics. Studies in Big Data (eds. Dash, S. et al.) vol. 68, 79–99 (Springer, Cham, 2020).

Hinton, G. Deep learning—a technology with the potential to transform health care. JAMA 320 , 1101 (2018).

Stead, W. W. Clinical implications and challenges of artificial intelligence and deep learning. JAMA 320 , 1107 (2018).

Dunnmon, J. A. et al. Assessment of convolutional neural networks for automated classification of chest radiographs. Radiology 290 , 537–544 (2019).

Irvin, J. et al. CheXpert: a large chest radiograph dataset with uncertainty labels and expert comparison. Proc. AAAI Conf. Artif. Intell. 33 , 590–597 (2019).

Google Scholar  

Jaeger, S. et al. Two public chest X-ray datasets for computer-aided screening of pulmonary diseases. Quant. Imaging Med. Surg. 4 , 475–477 (2014).

PubMed   PubMed Central   Google Scholar  

Johnson, A. E. W. et al. MIMIC-CXR, a de-identified publicly available database of chest radiographs with free-text reports. Sci. Data 6 , 317 (2019).

Kallianos, K. et al. How far have we come? Artificial intelligence for chest radiograph interpretation. Clin. Radiol. 74 , 338–345 (2019).

Gulshan, V. et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA 316 , 2402 (2016).

Rajpurkar, P. et al. Deep learning for chest radiograph diagnosis: A retrospective comparison of the CheXNeXt algorithm to practicing radiologists. PLoS Med. 15 , e1002686 (2018).

Majkowska, A. et al. Chest radiograph interpretation with deep learning models: assessment with radiologist-adjudicated reference standards and population-adjusted evaluation. Radiology 294 , 421–431 (2020).

Person, M., Jensen, M., Smith, A. O. & Gutierrez, H. Multimodal fusion object detection system for autonomous vehicles. J. Dyn. Syst. Meas. Control 141 , 071017 (2019).

Trzcinski, T. Multimodal social media video classification with deep neural networks. In Photonics Applications in Astronomy, Communications, Industry, and High-Energy Physics Experiments 2018 (eds. Romaniuk, R. S. & Linczuk, M.) (SPIE, 2018).

Ramachandram, D. & Taylor, G. W. Deep Multimodal Learning: A Survey on Recent Advances and Trends. IEEE Signal Process. Mag. 34 , 96–108 (2017).

Kiela, D., Grave, E., Joulin, A. & Mikolov, T. Efficient large-scale multi-modal classification. In The Thirty-Second AAAI Conference on Artificial Intelligence (AAAI-18) , (2018).

Thung, K.-H., Yap, P.-T. & Shen, D. Multi-stage Diagnosis of Alzheimer’s Disease with Incomplete Multimodal Data via Multi-task Deep Learning. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support (eds. Cardoso, M. J. et al.) vol. 10553, 160–168 (Springer International Publishing, 2017).

Kharazmi, P., Kalia, S., Lui, H., Wang, Z. J. & Lee, T. K. A feature fusion system for basal cell carcinoma detection through data-driven feature learning and patient profile. Skin Res. Technol. 24 , 256–264 (2018).

Yap, J., Yolland, W. & Tschandl, P. Multimodal skin lesion classification using deep learning. Exp. Dermatol. 27 , 1261–1267 (2018).

Li, H. & Fan, Y. Early Prediction Of Alzheimer’s Disease Dementia Based On Baseline Hippocampal MRI and 1-Year Follow-Up Cognitive Measures Using Deep Recurrent Neural Networks. In 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019) 368–371 (IEEE, 2019).

Purwar, S., Tripathi, R. K., Ranjan, R. & Saxena, R. Detection of microcytic hypochromia using cbc and blood film features extracted from convolution neural network by different classifiers. Multimed. Tools Appl. 79 , 4573–4595 (2020).

Liu, Mingqian, Lan, Jun, Chen, Xu, Yu, Guangjun & Yang, Xiujun Bone age assessment model based on multi-dimensional feature fusion using deep learning. Acad. J. Second Mil. Med. Univ. 39 , 909–916 (2018).

Nie, D. et al. Multi-channel 3D deep feature learning for survival time prediction of brain tumor patients using multi-modal neuroimages. Sci. Rep. 9 , 1103 (2019).

Article   PubMed   PubMed Central   CAS   Google Scholar  

Hyun, S. H., Ahn, M. S., Koh, Y. W. & Lee, S. J. A machine-learning approach using PET-based radiomics to predict the histological subtypes of lung cancer. Clin. Nucl. Med. 44 , 956–960 (2019).

Bhagwat, N., Viviano, J. D., Voineskos, A. N., Chakravarty, M. M. & Alzheimer’s Disease Neuroimaging Initiative. Modeling and prediction of clinical symptom trajectories in Alzheimer’s disease using longitudinal data. PLOS Comput. Biol. 14 , e1006376 (2018).

Liu, J. et al. Prediction of rupture risk in anterior communicating artery aneurysms with a feed-forward artificial neural network. Eur. Radiol. 28 , 3268–3275 (2018).

An, G. et al. Comparison of Machine-Learning Classification Models for Glaucoma Management. J. Healthc. Eng. 2018 , 1–8 (2018).

Spasov, S. E., Passamonti, L., Duggento, A., Lio, P. & Toschi, N. A. Multi-modal Convolutional Neural Network Framework for the Prediction of Alzheimer’s Disease. In 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC) 1271–1274 (IEEE, 2018).

Yala, A., Lehman, C., Schuster, T., Portnoi, T. & Barzilay, R. A deep learning mammography-based model for improved breast cancer risk prediction. Radiology 292 , 60–66 (2019).

Yoo, Y. et al. Deep learning of brain lesion patterns and user-defined clinical and MRI features for predicting conversion to multiple sclerosis from clinically isolated syndrome. Comput. Methods Biomech. Biomed. Eng. Imaging Vis. 7 , 250–259 (2019).

Kawahara, J., Daneshvar, S., Argenziano, G. & Hamarneh, G. Seven-point checklist and skin lesion classification using multitask multimodal neural nets. IEEE J. Biomed. Health Inform. 23 , 538–546 (2019).

Reda, I. et al. Deep learning role in early diagnosis of prostate cancer. Technol. Cancer Res. Treat. 17 , 153303461877553 (2018).

Qiu, S. et al. Fusion of deep learning models of MRI scans, Mini–Mental State Examination, and logical memory test enhances diagnosis of mild cognitive impairment. Alzheimers Dement. Diagn. Assess. Dis. Monit. 10 , 737–749 (2018).

Beach, T. G., Monsell, S. E., Phillips, L. E. & Kukull, W. Accuracy of the Clinical Diagnosis of Alzheimer Disease at National Institute on Aging Alzheimer Disease Centers, 2005–2010. J. Neuropathol. Exp. Neurol. 71 , 266–273 (2012).

Esteva, A. et al. Dermatologist-level classification of skin cancer with deep neural networks. Nature 542 , 115–118 (2017).

Article   CAS   PubMed   PubMed Central   Google Scholar  

Hecker, S., Dai, D. & Van Gool, L. End-to-End Learning of Driving Models with Surround-View Cameras and Route Planners. In Computer Vision – ECCV 2018 (eds. Ferrari, V. et al.) vol. 11211, 449–468 (Springer International Publishing, 2018).

Jain, A., Singh, A., Koppula, H. S., Soh, S. & Saxena, A. Recurrent Neural Networks for driver activity anticipation via sensory-fusion architecture. In 2016 IEEE International Conference on Robotics and Automation (ICRA) 3118–3125 (IEEE, 2016).

Goodfellow, I., Bengio, Y. & Courville, C. Deep Learning (MIT Press, 2017).

Zou, H. & Hastie, T. Regularization and variable selection via the elastic net. J. R. Stat. Soc. Ser. B Stat. Methodol. 67 , 301–320 (2005).

Subramanian, V., Do, M. N. & Syeda-Mahmood, T. Multimodal Fusion of Imaging and Genomics for Lung Cancer Recurrence Prediction. In 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI) 804–808 (IEEE, 2020).

Ngiam, J. et al. Multimodal Deep Learning. In Proceedings of the 28th International Conference on Machine Learning (ICML) 689–696 (2011).

Karpathy, A. et al. Large-Scale Video Classification with Convolutional Neural Networks. In 2014 IEEE Conference on Computer Vision and Pattern Recognition 1725–1732 (IEEE, 2014).

Neverova, N., Wolf, C., Taylor, G. & Nebout, F. ModDrop: adaptive multi-modal gesture recognition. IEEE Trans. Pattern Anal. Mach. Intell. 38 , 1692–1706 (2016).

Bach, F. Breaking the curse of dimensionality with convex neural networks. J. Mach. Learn. Res. 18 , 1–53 (2017).

Mwangi, B., Tian, T. S. & Soares, J. C. A review of feature reduction techniques in neuroimaging. Neuroinformatics 12 , 229–244 (2014).

David, M., Alessandro, L., Jennifer, T. & Douglas, G. A. The PRISMA Group Preferred reporting items for systematic reviews and meta-analyses: The PRISMA statement. PLoS Med . 6 , e1000097 (2009).

Download references

Acknowledgements

The authors wish to thank John Alexander Borghi from Stanford Lane Medical Library for his help with creating the systematic search. The research reported in this publication was supported by the National Library of Medicine of the National Institutes of Health under Award Number R01LM012966. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.

Author information

These authors contributed equally: Shih-Cheng Huang, Anuj Pareek.

Authors and Affiliations

Department of Biomedical Data Science, Stanford University, Stanford, USA

Shih-Cheng Huang & Matthew P. Lungren

Center for Artificial Intelligence in Medicine & Imaging, Stanford University, Stanford, USA

Shih-Cheng Huang, Anuj Pareek, Saeed Seyyedi, Imon Banerjee & Matthew P. Lungren

Department of Radiology, Stanford University, Stanford, USA

Anuj Pareek, Saeed Seyyedi & Matthew P. Lungren

Department of Biomedical Informatics, Emory University, Atlanta, USA

Imon Banerjee

Department of Radiology, Emory University, Atlanta, USA

You can also search for this author in PubMed   Google Scholar

Contributions

S.-C.H. and A.P. are co-first authors who contributed equally to this study. Concept and design : S.-C.H., A.P., M.P.L., and I.B. Study selection: S.-C.H. and A.P. Data extraction : S.-C.H., A.P., and S.S. Drafting of the manuscript : S.-C.H., A.P., I.B., and M.P.L. Critical revision of the manuscript for important intellectual content : S.-C.H., A.P., I.B., and M.P.L. Supervision : I.B. and M.P.L.

Corresponding author

Correspondence to Shih-Cheng Huang .

Ethics declarations

Competing interests.

The authors declare no competing interests.

Additional information

Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary information

Supplementary information, rights and permissions.

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ .

Reprints and permissions

About this article

Cite this article.

Huang, SC., Pareek, A., Seyyedi, S. et al. Fusion of medical imaging and electronic health records using deep learning: a systematic review and implementation guidelines. npj Digit. Med. 3 , 136 (2020). https://doi.org/10.1038/s41746-020-00341-z

Download citation

Received : 22 April 2020

Accepted : 17 September 2020

Published : 16 October 2020

DOI : https://doi.org/10.1038/s41746-020-00341-z

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

This article is cited by

Deep learning based joint fusion approach to exploit anatomical and functional brain information in autism spectrum disorders.

  • Sara Saponaro
  • Francesca Lizzi
  • Alessandra Retico

Brain Informatics (2024)

AtPCa-Net: anatomical-aware prostate cancer detection network on multi-parametric MRI

  • Haoxin Zheng
  • Alex Ling Yu Hung
  • Kyunghyun Sung

Scientific Reports (2024)

Healthcare on the brink: navigating the challenges of an aging society in the United States

  • Charles H. Jones
  • Mikael Dolsten

npj Aging (2024)

Development of Medical Imaging Data Standardization for Imaging-Based Observational Research: OMOP Common Data Model Extension

  • Woo Yeon Park
  • Kyulee Jeon

Journal of Imaging Informatics in Medicine (2024)

Joint Representation Learning with Generative Adversarial Imputation Network for Improved Classification of Longitudinal Data

  • Sharon Torao Pingi
  • Duoyi Zhang
  • Richi Nayak

Data Science and Engineering (2024)

Quick links

  • Explore articles by subject
  • Guide to authors
  • Editorial policies

Sign up for the Nature Briefing newsletter — what matters in science, free to your inbox daily.

thesis data fusion

Feature Correlation-based Data Fusion using Dempster-Shafer Evidence Theory for WSN

Ieee account.

  • Change Username/Password
  • Update Address

Purchase Details

  • Payment Options
  • Order History
  • View Purchased Documents

Profile Information

  • Communications Preferences
  • Profession and Education
  • Technical Interests
  • US & Canada: +1 800 678 4333
  • Worldwide: +1 732 981 0060
  • Contact & Support
  • About IEEE Xplore
  • Accessibility
  • Terms of Use
  • Nondiscrimination Policy
  • Privacy & Opting Out of Cookies

A not-for-profit organization, IEEE is the world's largest technical professional organization dedicated to advancing technology for the benefit of humanity. © Copyright 2024 IEEE - All rights reserved. Use of this web site signifies your agreement to the terms and conditions.

Scholars' Mine

  • < Previous

Home > Student Works > TDs > Doctoral Dissertations > 2428

Doctoral Dissertations

Data fusion techniques for nondestructive evaluation and medical image analysis.

"Data fusion is a technique for combining data obtained from multiple sources for an enhanced detection or decision. Fusion of data can be done at the raw-data level, feature level or decision level. Applications of data fusion include defense (such as battlefield surveillance and autonomous vehicle control), medical diagnosis and structural health monitoring. Techniques for data fusion have been drawn from areas such as statistics, image processing, pattern recognition and computational intelligence. This dissertation includes investigation and development of methods to perform data fusion for nondestructive evaluation (NDE) and medical imaging applications. The general framework for these applications includes region-of-interest (ROI) detection followed by feature extraction and classification of the detected ROI. Image processing methods such as edge detection and projection-based methods were used for ROI detection. The features extracted from the detected ROIs include texture, color, shape/geometry and profile-based correlation. Analysis and classification of the detected ROIs was performed using feature- and decision-level data fusion techniques such as fuzzy-logic, statistical methods and voting algorithms"--Abstract, page iv.

Stanley, R. Joe

Committee Member(s)

Moss, Randy Hays, 1953- Stoecker, William V. Xiao, Hai, Dr. Adekpedjou, Akim

Department(s)

Electrical and Computer Engineering

Degree Name

Ph. D. in Electrical Engineering

Missouri University of Science and Technology

Publication Date

Journal article titles appearing in thesis/dissertation.

  • A comprehensive structural analysis process for failure assessment in aircraft lap-joint mimics using intra-modal fusion of eddy current data
  • A comprehensive multi-modal NDE data fusion approach for failure assessment in aircraft lap-joint mimics
  • Automated biomedical text detection method in support of multi-scale content-based image retrieval systems
  • A data fusion-based approach for uterine cervical cancer histology image classification

xiv, 187 pages

Note about bibliography

Includes bibliographical references.

© 2013 Soumya De, All rights reserved.

Document Type

Dissertation - Open Access

Subject Headings

Multisensor data fusion Nondestructive testing Image processing -- Analysis -- Technique

Thesis Number

Print oclc #, electronic oclc #, recommended citation.

De, Soumya, "Data fusion techniques for nondestructive evaluation and medical image analysis" (2013). Doctoral Dissertations . 2428. https://scholarsmine.mst.edu/doctoral_dissertations/2428

Since September 29, 2015

Included in

Electrical and Computer Engineering Commons

Advanced Search

  • Notify me via email or RSS
  • Collections
  • Disciplines
  • All Authors
  • Faculty Authors

Author Corner

  • Share Your Dissertation

Useful Links

  • Library Resources

S&T logo

Dissertation Locations

  • View dissertations on map
  • View dissertations in Google Earth

Home | About | FAQ | My Account | Accessibility Statement

Privacy Copyright

Help | Advanced Search

High Energy Physics - Theory

Title: the fusion categorical diagonal.

Abstract: We define a Frobenius algebra over fusion categories of the form Rep$(G)\boxtimes$Rep$(G)$ which generalizes the diagonal subgroup of $G\times G$. This allows us to extend field theoretical constructions which depend on the existence of a diagonal subgroup to non-invertible symmetries. We give explicit calculations for theories with Rep$(S_3)\boxtimes$Rep$(S_3)$ symmetry, applying the results to gauging topological quantum field theories which carry this non-invertible symmetry. Along the way, we also discuss how Morita equivalence is implemented for algebras in symmetry categories.

Submission history

Access paper:.

  • HTML (experimental)
  • Other Formats

References & Citations

  • INSPIRE HEP
  • Google Scholar
  • Semantic Scholar

BibTeX formatted citation

BibSonomy logo

Bibliographic and Citation Tools

Code, data and media associated with this article, recommenders and search tools.

  • Institution

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs .

Business Wire

RIYADH, Arabie saoudite--( BUSINESS WIRE )-- Cognite , un leader mondialement réputé du domaine des logiciels industriels, a annoncé aujourd’hui que Cognite Data Fusion ® , son produit phare d’opérations de données (DataOps), est maintenant disponible sur Google Cloud en Arabie saoudite. Cette intégration stratégique, qui associe la gestion des données de pointe et l’intelligence artificielle sécurisée et fiable (IA) de Cognite avec la puissance de l’infrastructure et des capacités d’analyse des données de Google Cloud, permet aux clients de migrer en douceur leurs données vers le cloud et garantit l’évolutivité, la flexibilité et la sécurité tout en donnant aux clients saoudiens un accès aisé à leurs données industrielles complexes.

En plus des outils avancés d’IA et d’apprentissage automatique de Google Cloud, Cognite Data Fusion déploiera ses propres capacités de contextualisation et d’IA générative spécialement conçues pour des données industrielles complexes, une combinaison puissante qui produira des informations exploitables, accroîtra l’efficacité opérationnelle et favorisera la transformation numérique dans des secteurs tels que l’énergie, la fabrication et la logistique.

« Nous sommes ravis d’apporter Cognite Data et des solutions d’IA aux clients de Google Cloud en Arabie saoudite », a déclaré François Laborie, vice-président exécutif, Cognite. « Il s’agit d’une étape importante pour notre partenariat et pour notre engagement à l’égard du royaume saoudien. La puissance des données et de l’IA fournies par Cognite et Google Cloud facilite la transformation de l’industrie grâce à des solutions qui créent de la valeur pour nos clients. »

« Nous sommes heureux de nous associer à Cognite pour mettre sa plateforme avancée de données industrielles à la disposition des clients saoudiens de Google Cloud », a déclaré Abdul Rahman Al Thehaiban, directeur général Moyen-Orient, Turquie et Afrique de Google Cloud. « En s’appuyant sur l’infrastructure évolutive et sécurisée de Google Cloud, Cognite Data Fusion permettra aux clients d’exploiter la puissance de leurs données pour prendre des décisions d’affaires plus intelligentes et accélérer leur parcours de transformation numérique. »

Adaptées aux cas d’utilisation industriels, les capacités avancées d’IA générative de Cognite Data Fusion permettent aux clients de générer des données synthétiques, de simuler des scénarios et d’optimiser leurs opérations et, finalement, d’améliorer leurs prises de décision et leurs résultats opérationnels.

En Arabie saoudite et dans l’ensemble de la région MENA, Cognite Data Fusion est commercialisé par CNTXT, une co-entreprise Saudi Aramco et Cognite. Ce partenariat associe la compréhension approfondie du paysage industriel régional de CNTXT avec la technologie de pointe de Cognite pour fournir aux clients des solutions sur mesure qui répondent à leurs besoins et défis spécifiques. « Cognite Data Fusion est sur le point de révolutionner la manière dont les industries saoudiennes tirent parti des informations issues de leurs données. Cognite Data Fusion offre au niveau mondial des capacités de pointe en matière d’IA générative adaptées aux secteurs industriels », a déclaré Abdullah Jarwan, PDG de CNTXT.

Cognite Data Fusion sur Google Cloud est disponible dès maintenant pour les clients saoudiens. Pour plus d’informations, visiter le site Web : www.cntxt.com .

À propos de CNTXT :

CNTXT, une joint-venture entre Saudi Aramco et Cognite basée à Riyad, en Arabie saoudite, est une entreprise de premier plan dédiée à la transformation numérique. Dans le cadre de sa mission de soutien des entreprises en matière de technologies et d’innovation, CNTXT s’appuie sur des partenariats stratégiques et un réseau de revendeurs pour accélérer la transformation numérique dans les secteurs public et privé du royaume saoudien.

À propos de Cognite

Cognite met l’IA générative au service de l’industrie. Les entreprises leaders dans les domaines de l’énergie, de la fabrication, de l’électricité et des énergies renouvelables choisissent Cognite pour fournir des données sécurisées, fiables et en temps réel afin de transformer leurs opérations exigeantes en termes d’actifs pour qu’elles soient plus sûres, plus durables et plus rentables. Cognite propose une plateforme de DataOps industrielles conviviale, sécurisée et évolutive, Cognite Data Fusion®, qui permet à tous les décideurs, du terrain aux centres d’opérations distants, d’accéder et de comprendre des données industrielles complexes, de collaborer en temps réel et de construire un avenir meilleur. Rendez-vous sur notre site www.cognite.ai et suivez-nous sur LinkedIn et X (Twitter).

Le texte du communiqué issu d’une traduction ne doit d’aucune manière être considéré comme officiel. La seule version du communiqué qui fasse foi est celle du communiqué dans sa langue d’origine. La traduction devra toujours être confrontée au texte source, qui fera jurisprudence.

Giedre Malinauskaite [email protected] +966 57 0489987

thesis data fusion

thesis data fusion

Journal of Analytical Atomic Spectrometry

Air pressure prediction model based on laser-induced plasma image and spectra fusion.

The vacuum switch uses vacuum as the insulation and arc-extinguishing medium. It is the core equipment in high-voltage transmission and distribution power systems. Online detection of vacuum level for vacuum switch is an international challenge that has not been effectively addressed for over seventy years. In the previous work, our team first proposed an online detection method of vacuum level for vacuum switches based on laser-induced plasma (LIP), which holds the potential for achieving safe and reliable vacuum level detection. However, the accuracy of the vacuum level detection based on a single variable is low and difficult to meet the requirements of high-voltage transmission power systems. To address this issue, this paper proposes a vacuum level prediction model based on the fusion of laser-induced plasma images and spectra. The proposed model comprises two modules: Image module and Spectra module. The Image module fuses plasma and spectral images to extract features, while the Spectra module extracts features from 1D spectral data. Ultimately, features from both images and 1D spectral data are combined for air pressure prediction. Experimental results reveal that the fusion of image and spectra improves the Macro-Precision (MacP) of air pressure prediction, achieving an accuracy of 96.83%. In comparison to utilizing only images or solely 1D spectral data for air pressure prediction, the MacP increases by approximately 1.41% and 6.83%, respectively. Compared to other baseline models, this model has superior predictive performance. These findings underscore the effectiveness of the proposed air pressure prediction model that fuses image and spectra. This paper establishes a foundation for the implementation of the online detection of vacuum level based on laser-induced plasma in vacuum switches.

Article information

Download citation, permissions.

thesis data fusion

W. Ke, H. Luo, S. Lv, H. Yuan, X. Wang, A. Yang, J. Chu, D. Liu and M. Rong, J. Anal. At. Spectrom. , 2024, Accepted Manuscript , DOI: 10.1039/D4JA00040D

To request permission to reproduce material from this article, please go to the Copyright Clearance Center request page .

If you are an author contributing to an RSC publication, you do not need to request permission provided correct acknowledgement is given.

If you are the author of this article, you do not need to request permission to reproduce figures and diagrams provided correct acknowledgement is given. If you want to reproduce the whole article in a third-party publication (excluding your thesis/dissertation for which permission is not required) please go to the Copyright Clearance Center request page .

Read more about how to correctly acknowledge RSC content .

Social activity

Search articles by author.

This article has not yet been cited.

Advertisements

IMAGES

  1. thesis data fusion

    thesis data fusion

  2. Presentation of PhD thesis on Location Data Fusion

    thesis data fusion

  3. Schematic of the data fusion process: (a) low-level data fusion and (b

    thesis data fusion

  4. General multimodal data fusion approach for EMD/LA

    thesis data fusion

  5. Data fusion and knowledge fusion research framework for multimodal

    thesis data fusion

  6. Summary of Data Fusion Methods

    thesis data fusion

VIDEO

  1. KinfuSeg: a Dynamic SLAM Approach Based on Kinect Fusion

  2. Data Contextualization Masterclass 2023

  3. Data fusion with Gaussian processes

  4. WTAB HAL1

  5. Respect #viral #viralshorts #SSwriters123 #shorts #short

  6. Enterprise Data Integration with Cloud Data Fusion

COMMENTS

  1. A survey on machine learning for data fusion

    In the Luo and Kay's architecture, raw data come from sensors, and are fused in the nodes of an information system. For example, data from sensor 1 and 2 can be fused as data x 1, 2.After that, the output data x 1, 2 will be further fused in the next fusion node with data from sensor 3, turning into data x 1, 2, 3.Similarly, data x 1, 2, … n from the last fusion node is the highest fusion ...

  2. Data Fusion: Theory, Methods, and Applications

    Data fusion is a little bit similar to data aggregation. Data aggregation is essentially used to provide information in a synthetic form (summary of data) ... In this thesis, we propose a stateful ...

  3. Machine Learning Techniques in Data Fusion: A Review

    Here, every feasible machine learning technique employed in the various research publications on data fusion has been reviewed. According to our observations, the most popular machine learning algorithms include SVM, CNN, KF, ANN, Bayesian, fuzzy theory, clus-tering, maximum likelihood, DSET, and KNN.

  4. Advances in Hierarchical Probabilistic Multimodal Data Fusion

    Abstract. Multimodal data fusion is the process of integrating disparate data sources into a shared representation suitable for complex reasoning. As a result, one can make more precise inferences about the underlying phenomenon than is possible with each data source used in isolation. In the thesis we adopt a Bayesian view of multimodal data ...

  5. Multisensor data fusion: A review of the state-of-the-art

    Introduction. Multisensor data fusion is a technology to enable combining information from several sources in order to form a unified picture. Data fusion systems are now widely used in various areas such as sensor networks, robotics, video and image processing, and intelligent system design, to name a few.

  6. Multimodal Data Fusion: An Overview of Methods, Challenges, and

    This paper deals with two key issues: "why we need data fusion" and "how we perform it." The first issue is motivated by numerous examples in science and technology, followed by a mathematical framework that showcases some of the benefits that data fusion provides. In order to address the second issue, "diversity" is introduced as a ...

  7. Multimodal Data Fusion for Estimating Electricity Access and Demand

    The IEA's "Net Zero by 2050" roadmap for net zero emissions models that electricity demand for "emerging market and developing economies" will need to exceed double the electricity demand in "advanced economies" by mid-century. Our development and climate imperatives both rest upon electricity demand in low- and middle-income ...

  8. PDF Application of Data Fusion and Machine Learning to The Analysis of The

    on projects with aeronautical companies such as this thesis with the Federal Aviation Administration. Thank you very much Doc, my experience at Georgia Tech would have been ... 4.3.1 Data Fusion to assess the relevancy of recommended reroutes 38: 4.4 Step #4: Data Analysis and Results 39: 4.4.1 Tracking Flights Approach 39

  9. Informative Data Fusion: Beyond Canonical Correlation Analysis

    Multi-modal data fusion is a challenging but common problem arising in fields such as economics, statistical signal processing, medical imaging, and machine learning. ... CCA fails to reliably detect correlation between the datasets. This thesis explores the the problem of detecting correlations from data modeled by the ubiquitous signal-plus ...

  10. Data fusion algorithm based on fuzzy sets and D-S theory of evidence

    In cyber-physical systems, multidimensional data fusion is an important method to achieve comprehensive evaluation decisions and reduce data redundancy. In this paper, a data fusion algorithm based on fuzzy set theory and Dempster-Shafer (D-S) evidence theory is proposed to overcome the shortcomings of the existing decision-layer multidimensional data fusion algorithms. The basic probability ...

  11. PDF An Architectural Selection Framework for Data Fusion in Sensor Platforms

    by Atif R. Mirza. Submitted to the System Design and Management Program in Partial Fulfillment of the Requirements for the Degree of Master of Science in Engineering and Management. The role of data fusion in sensor platforms is becoming increasingly important in various domains of science, technology and business.

  12. Bayesian data fusion for distributed learning

    Data fusion typically includes raw data fusion, feature fusion, and decision fusion. In this thesis, we will concentrate on feature fusion. Distributed data fusion involves merging sensor data from different sources to estimate an unknown process. Bayesian framework is often used because it can provide an optimal and explainable feature by ...

  13. PDF Bayesian Approaches to Multi-Sensor Data Fusion

    of the method allows the generalisation of the algorithm to the case of centralized fusion of several signals from di erent sources. The thesis concludes with an analysis of synthetic data \obtained" from three di erent sources (multiple simple steps, ramps and piecewise constant AR process), to illustrate the proposed technique.

  14. Wiley Online Library

    Wiley Online Library

  15. Fusion of medical imaging and electronic health records using deep

    In this paper, we describe different data fusion techniques that can be applied to combine medical imaging with EHR, and systematically review medical data fusion literature published between 2012 ...

  16. Feature Correlation-based Data Fusion using Dempster-Shafer Evidence

    In wireless sensor networks, data fusion is widely used to merge data obtained from heterogeneous sensors. Since sensors provide uncertain information and containing redundant and irrelevant features, it is imperative to avoid undesired features and select the subset of relevant features for use in data fusion. In this paper, a novel data fusion scheme is proposed based on Pearson correlation ...

  17. DATA FUSION: MODELS AND PROCEDURES

    The integration of data, recorded from a multiple sensor system, together with knowledge, is known as data fusion. Data fusion first appeared in the literature in the 1960s, as mathematical models for data. manipulation. It was implemented in the US in the 1970s in the fields of robotics and. defence.

  18. PDF Data Fusion Methodologies for Multisensor Aircraft Navigation Systems

    The thesis covers data fusion for aircraft navigation systems in distributed sensor systems. Data fusion methodologies are developed for the design, development, analysis and simulation of multisensor aircraft navigation systems. The problems of sensor failure detection and isolation (FDI), distributed data fusion algorithms and

  19. Multi-sensor integrated navigation/positioning systems using data

    Data fusion, which fuses diverse information from various sources, is the most important part of integrated systems, a widely accepted definition of which was provided by the Joint Directors of the Laboratories (JDL) workshop [16]: "A multi-level process dealing with the association, correlation, combination of data, and information from ...

  20. PDF Data Fusion & Diversity

    the speci c pieces of research endeavor throughout this thesis, has applied the adaptive ltering framework to problems that employ the techniques of data diversity and fusion which includes quaternions, tensors and graphs. At the rst glance, all these structures share one common important feature: invertible isomorphism. In other words, they are

  21. PhD Thesis: Airborne Hyperspectral Imaging for Multisensor Data Fusion

    The main benefit of multisensor integration of airborne HS and magnetic data is a. possibility. (green), peridotite (blue), and py roxenite (red). ( B, C) show the calculated pseudo-magnetic ...

  22. PDF A Low Cost Localization Solution Using a Kalman Filter for Data Fusion

    1.1 Thesis Overview This thesis presents a way to provide a drivable and accurate position solution using inexpensive sensors. A Kalman filter is used to fuse that data from these sensors to produce an output that is better than that of any single sensor. The goal is to provide a cost-effective solution accurate enough

  23. "Data fusion techniques for nondestructive evaluation and medical image

    De, Soumya, "Data fusion techniques for nondestructive evaluation and medical image analysis" (2013). Doctoral Dissertations. 2428. "Data fusion is a technique for combining data obtained from multiple sources for an enhanced detection or decision. Fusion of data can be done at the raw-data level, feature level or decision level.

  24. [2405.07845] Multi-Task Learning for Fatigue Detection and Face

    In driving scenarios, automobile active safety systems are increasingly incorporating deep learning technology. These systems typically need to handle multiple tasks simultaneously, such as detecting fatigue driving and recognizing the driver's identity. However, the traditional parallel-style approach of combining multiple single-task models tends to waste resources when dealing with similar ...

  25. [2405.07655] Quality-aware Selective Fusion Network for V-D-T Salient

    Therefore, we propose a quality-aware selective fusion network (QSF-Net) to conduct VDT salient object detection, which contains three subnets including the initial feature extraction subnet, the quality-aware region selection subnet, and the region-guided selective fusion subnet. Firstly, except for extracting features, the initial feature ...

  26. Expressway Vehicle Trajectory Prediction Based on Fusion Data of ...

    Research on vehicle trajectory prediction based on road monitoring video data often utilizes a global map as an input, disregarding the fact that drivers rely on the road structures observable from their own positions for path planning. This oversight reduces the accuracy of prediction. To address this, we propose the CVAE-VGAE model, a novel trajectory prediction approach. Initially, our ...

  27. [2405.08058] The Fusion Categorical Diagonal

    The Fusion Categorical Diagonal. Daniel Robbins, Thomas Vandermeulen. We define a Frobenius algebra over fusion categories of the form Rep (G)⊠ Rep (G) which generalizes the diagonal subgroup of G × G. This allows us to extend field theoretical constructions which depend on the existence of a diagonal subgroup to non-invertible symmetries.

  28. Cognite Data Fusion® Now Available on Google Cloud in Saudi Arabia

    Cognite Data Fusion will deploy its own comprehensive contextualization and generative AI (gen AI) capabilities purpose-built for complex industrial data on top of Google Cloud's advanced AI and ...

  29. Cognite Data Fusion® est désormais disponible sur Google Cloud en

    Cognite Data Fusion ® sur Google Cloud apportera des outils avancés d'IA et d'apprentissage automatique pour créer des informations exploitables, accroître l'efficacité opérationnelle ...

  30. Air Pressure Prediction Model Based on Laser-Induced Plasma Image and

    To address this issue, this paper proposes a vacuum level prediction model based on the fusion of laser-induced plasma images and spectra. The proposed model comprises two modules: Image module and Spectra module. The Image module fuses plasma and spectral images to extract features, while the Spectra module extracts features from 1D spectral data.