Applications of machine learning techniques for enhancing nondestructive food quality and safety det
TL;DR Summary
This review highlights machine learning's role in enhancing nondestructive food quality and safety detection, contrasting traditional and deep learning techniques integrated with acoustic, vision, and electronic nose technologies. Deep learning shows superior potential for real-t
Abstract
Critical Reviews in Food Science and Nutrition ISSN: 1040-8398 (Print) 1549-7852 (Online) Journal homepage: www.tandfonline.com/journals/bfsn20 Applications of machine learning techniques for enhancing nondestructive food quality and safety detection Yuandong Lin, Ji Ma, Qijun Wang & Da-Wen Sun To cite this article: Yuandong Lin, Ji Ma, Qijun Wang & Da-Wen Sun (2023) Applications of machine learning techniques for enhancing nondestructive food quality and safety detection, Critical Reviews in Food Science and Nutrition, 63:12, 1649-1669, DOI: 10.1080/10408398.2022.2131725 To link to this article: https://doi.org/10.1080/10408398.2022.2131725 Published online: 12 Oct 2022. Submit your article to this journal Article views: 5243 View related articles View Crossmark data Citing articles: 129 View citing articles Full Terms & Conditions of access and use can be found at https://www.tandfonline.com/action/journalInformation?journalCode=bfsn20
Mind Map
In-depth Reading
English Analysis
1. Bibliographic Information
1.1. Title
Applications of machine learning techniques for enhancing nondestructive food quality and safety detection
1.2. Authors
Yuandong Lin, Ji Ma, Qijun Wang & Da-Wen Sun
1.3. Journal/Conference
Critical Reviews in Food Science and Nutrition. This journal is a highly reputable publication in the field of food science, known for publishing comprehensive review articles that critically evaluate advancements in various aspects of food research, including quality, safety, processing, and nutrition. Its reviews are often influential in shaping future research directions.
1.4. Publication Year
2023
1.5. Abstract
The paper addresses the growing global demand for high-quality food, highlighting the increasing interest in nondestructive and rapid detection technologies within the food industry. It acknowledges that data analysis from most nondestructive techniques is often complex, time-consuming, and requires specialized skills. Traditional chemometric or statistical methods face limitations due to noise, sample variability, and data complexity under diverse testing conditions. The review positions machine learning (ML) techniques as a powerful solution due to their capabilities in handling irrelevant information, extracting feature variables, and building calibration models, particularly in nondestructive technology and equipment intelligence.
The paper introduces and compares ML techniques, categorizing them into traditional machine learning (TML) and deep learning (DL). It then presents applications of several novel nondestructive technologies—acoustic analysis, machine vision (MV), electronic nose (e-nose), and spectral imaging—combined with advanced ML techniques for food quality assessment, including variety identification and classification, safety inspection, and processing control. Challenges and future prospects are also discussed. The review concludes that the integration of nondestructive testing and state-of-the-art ML techniques holds significant potential for monitoring food quality and safety, with different ML algorithms having distinct characteristics and applicability. Deep learning, characterized by its feature learning nature, is identified as one of the most promising and powerful techniques for real-time applications, warranting further extensive research and wider adoption in the food industry.
1.6. Original Source Link
https://doi.org/10.1080/10408398.2022.2131725 (Officially published)
2. Executive Summary
2.1. Background & Motivation
The global population's increasing demand for high-quality food necessitates robust and efficient methods for ensuring food quality and safety. Traditional detection methods, such as gas chromatography (GC), high-performance liquid chromatography (HPLC), and polymerase chain reaction (PCR), are destructive, labor-intensive, expensive, and time-consuming, making them unsuitable for rapid, online, or industrial applications.
This limitation led to the development of nondestructive techniques like spectral imaging, acoustic vibration methods, machine vision (MV), and electronic nose (e-nose). While these technologies offer advantages in speed and non-invasiveness, they generate complex, high-dimensional, noisy, and often redundant datasets. Analyzing these datasets with conventional chemometric or statistical methods is challenging, particularly for nonlinear and large-scale data, requiring highly skilled operators and significant time. This complexity limits their real-time industrial application.
The core problem the paper aims to solve is the need for efficient and effective data analysis methods to unlock the full potential of nondestructive food quality and safety detection technologies. The paper's innovative idea and entry point is to leverage the capabilities of machine learning (ML) techniques, including traditional machine learning (TML) and deep learning (DL), to overcome the analytical challenges posed by nondestructive sensor data. ML can handle complex, nonlinear, and large datasets, automate feature extraction, and build robust predictive models, thereby enhancing the efficiency and applicability of these detection technologies.
2.2. Main Contributions / Findings
The paper makes several primary contributions:
-
Comprehensive Overview of ML Techniques: It provides a structured introduction and comparison of
traditional machine learning (TML)anddeep learning (DL)algorithms relevant to nondestructive food analysis, outlining their principles, advantages, and limitations. -
Integration with Novel Nondestructive Technologies: It systematically summarizes the applications of advanced ML techniques when combined with various novel nondestructive technologies, specifically
acoustic analysis,machine vision (MV),electronic nose (e-nose), andspectral imaging. -
Diverse Application Scenarios: The review showcases the utility of these integrated approaches across key areas of the food industry, including
food quality assessment(variety identification, classification),safety inspection(chemical and microbial contamination), andprocessing control. -
Identification of Challenges and Future Directions: It critically discusses the existing challenges in applying ML to nondestructive food detection, such as the
lack of labeled data,data standardization, andcomputational demands, while also proposing promising future research avenues liketransfer learning,lifelong learning, and the development oflightweight DL models.The key conclusions and findings reached by the paper are:
-
Nondestructive testing technologies, when combined with state-of-the-art ML techniques, demonstrate significant potential for monitoring the quality and safety of food products.
-
Different ML algorithms possess unique characteristics and are suited for specific application scenarios.
-
Deep learning, owing to its powerfulfeature learningcapabilities and ability to handle complex data, is identified as a particularly promising technique forreal-time applications. However, its widespread adoption requires further research to address challenges related todata volume,computational resources, andmodel complexity. -
The integration of ML simplifies
data analysis,feature extraction, andmodel building, thereby enhancing the efficiency and effectiveness of nondestructive detection systems and moving towardsdevice intellectualizationin the food industry.
3. Prerequisite Knowledge & Related Work
3.1. Foundational Concepts
To fully understand this paper, a beginner should grasp several foundational concepts in food science, sensing technology, and artificial intelligence.
-
Nondestructive Technologies: These are methods for testing materials, components, or systems without causing damage to the item being tested. In food science, they are crucial for quality control without altering the food product.
- Acoustic Analysis / Ultrasound: Utilizes sound waves (audible or ultrasonic) to probe material properties. Changes in sound propagation (velocity, attenuation, resonance frequencies) are correlated with physical characteristics (e.g., firmness, moisture content, internal defects).
- Machine Vision (MV): Also known as
computer vision system (CVS), it involves equipping computers with the ability to "see" and interpret images. In food, this often usesRGB camerasto capture visual data (color, shape, texture, size, surface defects) for automated inspection and grading. - Electronic Nose (E-nose): An array of gas sensors designed to mimic the human olfactory system. It detects and analyzes
volatile organic compounds (VOCs)emitted by food, generating unique "odor fingerprints" that can be correlated with freshness, spoilage, or specific aromas. - Spectral Imaging (e.g., Hyperspectral Imaging - HSI): Combines traditional imaging with spectroscopy. Instead of just
RGB(red, green, blue) bands,HSIcollects light across a wide range of continuous electromagnetic spectrum bands for each pixel, creating ahypercube. This allows for simultaneous acquisition ofspatial(image) andspectral(chemical/physical properties) data, enabling the identification and quantification of chemical compositions.
-
Chemometrics / Statistical Methods: These are mathematical and statistical techniques applied to chemical data. In food analysis, they are used to extract meaningful information from complex analytical measurements. Examples include
Principal Component Analysis (PCA),Partial Least Squares (PLS),Linear Discriminant Analysis (LDA). The paper notes their limitations in handlingnonlinearandhigh-dimensionaldatasets. -
Artificial Intelligence (AI), Machine Learning (ML), and Deep Learning (DL):
- Artificial Intelligence (AI): A broad field of computer science aimed at creating intelligent machines that can reason, learn, and act autonomously.
- Machine Learning (ML): A subfield of AI where algorithms learn patterns from data and make predictions or decisions without being explicitly programmed. The machine "learns" from training data.
-
Traditional Machine Learning (TML): Refers to ML algorithms that typically require manual
feature extraction(i.e., human expertise to define relevant characteristics in the data) and often perform well onsmaller datasets. -
Deep Learning (DL): A subfield of ML inspired by the structure and function of the human brain's
neural networks. It usesdeep neural networkswith multiple layers (hence "deep") to automatically learn complexfeature representationsfrom raw data, often performing exceptionally well onlarge datasets. It minimizes the need for manual feature engineering. The relationship can be visualized as a hierarchy, with AI encompassing ML, and ML encompassing DL, as shown in the following figure (Figure 2 from the original paper).
该图像是一个示意图,展示了人工智能、机器学习和深度学习之间的包含关系,及其与无损检测技术的关联,突出深度学习在机器学习和人工智能中的核心地位。The above figure (Figure 2 from the original paper) illustrates the hierarchical relationship, showing
Deep Learningas a subset ofMachine Learning, which in turn is a subset ofArtificial Intelligence.
-
-
Types of Machine Learning:
- Supervised Learning: Algorithms learn from
labeled data, where both the input and the desired output are provided. The goal is to learn a mapping from inputs to outputs forpredictionorclassification.-
Support Vector Machine (SVM): A powerful algorithm for
classificationandregression. It finds an optimalhyperplanethat best separates different classes in the feature space, maximizing the margin between them. It can handlenonlinearrelationships usingkernel functions(e.g.,radial basis function (RBF)). -
Logistic Regression (LR): A
statistical modelused for binaryclassification(predicting a probability of 0 or 1). It models the probability of a certain class or event existing. -
K-Nearest Neighbor (KNN): A simple,
non-parametricalgorithm used for bothclassificationandregression. It classifies a data point based on the majority class of its nearest data points in the feature space. -
Decision Tree (DT): A
tree-like modelwhere each internal node represents a test on an attribute, each branch represents the outcome of the test, and each leaf node represents a class label or a predicted value.Random Forest (RF)andGradient Boostingare ensemble methods built upon decision trees. -
Naïve Bayes (NB): A
probabilistic classifierbased onBayes' theoremwith the "naïve" assumption ofconditional independencebetween features. -
Artificial Neural Networks (ANNs) / Multilayer Perceptrons (MLPs): Inspired by biological neural networks, ANNs consist of interconnected
nodes(neurons) organized in layers (input, hidden, output). They learn by adjustingweightsandbiasesthrough training (e.g.,backpropagation) to map inputs to outputs, capable of modelingnonlinear systems.
该图像是三种神经网络结构的示意图,包括(a)卷积神经网络结构流程,(b)多层感知器结构,及(c)采用自编码器进行特征编码和解码的流程图。图中用表示感知器函数。The above figure (Figure 4 from the original paper) illustrates different neural network architectures. Specifically, Figure 4(b) depicts a
Multilayer Perceptron (MLP), also known as a typicalArtificial Neural Network (ANN), with multiple hidden layers. The figure also shows theconvolutional neural networkstructure (a) and theautoencoderprocess (c), which aredeep learningarchitectures.
-
- Unsupervised Learning: Algorithms learn from
unlabeled datato discover hidden patterns, structures, or relationships within the data.- Dimensionality Reduction: Techniques to reduce the number of
features(variables) in a dataset while retaining most of the important information.- Principal Component Analysis (PCA): A linear dimensionality reduction technique that transforms data into a new coordinate system, where the greatest variance by some projection comes to lie on the first
principal component, the second greatest variance on the second principal component, and so on.
- Principal Component Analysis (PCA): A linear dimensionality reduction technique that transforms data into a new coordinate system, where the greatest variance by some projection comes to lie on the first
- Clustering: Grouping similar data points into clusters without prior knowledge of their labels.
- K-means Clustering: An iterative algorithm that partitions observations into clusters, where each observation belongs to the cluster with the nearest mean (centroid).
- Dimensionality Reduction: Techniques to reduce the number of
- Semi-supervised Learning: Combines a small amount of
labeled datawith a large amount ofunlabeled dataduring training. It is useful when obtaining labeled data is expensive or time-consuming.
- Supervised Learning: Algorithms learn from
-
Deep Learning Architectures:
- Autoencoder (AE): A type of
neural networkused for unsupervised learning of efficient data encodings. It has anencoderthat compresses the input into alatent-space representationand adecoderthat reconstructs the input from this representation. It's often used fordimensionality reductionorfeature learning. - Convolutional Neural Network (CNN): A specialized type of
neural networkparticularly effective for processinggrid-like datasuch as images. Key components includeconvolutional layers(which apply filters to detect features),pooling layers(which reduce spatial dimensions), andfully connected layers(for classification). - Restricted Boltzmann Machine (RBM): A
generative stochastic artificial neural networkthat can learn a probability distribution over its set of inputs. It consists of a visible layer and a hidden layer, with symmetric connections between them, but no connections within a layer.Deep Belief Networks (DBNs)are often built by stacking RBMs.
- Autoencoder (AE): A type of
3.2. Previous Works
The paper acknowledges several relevant prior reviews but distinguishes its contribution by highlighting their specific focus:
-
Rehman et al. (2019): Focused on current applications of
TML techniquesinMV systems. -
Saha and Manickavasagan (2021): Outlined applications of different
machine learning techniquesinhyperspectral image analysis. -
Liu, Pu, and Sun (2021): Summarized applications of
CNN modelsinfood quality detectionand discussedfeature extraction methodsbased on 1D, 2D, and 3D models. -
Sun, Zhang, and Mujumdar (2019): Reviewed
AI technologiesand their applications specifically infood drying. -
D. Y. Wang et al. (2022): Reviewed
AI-based methodsfor detection and prediction insorting,drying,disinfecting,sterilizing, andfreezing of berries.While these reviews cover important aspects, the current paper observes that they either concentrate on a
single nondestructive technologyor asingle food process. This implies a gap in a systematic, comprehensive review that compares the performances of bothTMLandDLacross various novel nondestructive technologies for enhancing efficiency and effectiveness in food quality and safety detection.
3.3. Technological Evolution
The evolution of technologies in this field has been driven by advancements in several interconnected areas:
-
Sensing Technology: Development of more sophisticated and affordable sensors for
spectral imaging,acoustic analysis,machine vision, ande-noseshas enabled richer and more diverse data collection from food products. This includes higher resolution cameras, more sensitive gas sensors, and more precise acoustic transducers. -
Computer Science and Hardware: Significant progress in
computational power(especially withGraphics Processing Units - GPUs),data storage, andefficient algorithmshas made it feasible to process themassive amounts of informationgenerated by nondestructive techniques. This includes the development of user-friendlysoftware packages(e.g., Scikit-learn, TensorFlow, PyTorch). -
Data Science and AI: The maturation of
machine learninganddeep learningalgorithms has provided powerful tools for extracting meaningful patterns from complex, high-dimensional, and noisy datasets, moving beyond the limitations of traditionalchemometricandstatistical methods. The shift frommanual feature engineeringin TML toautomatic feature learningin DL represents a significant leap.This paper's work fits into the current technological timeline by showcasing how the convergence of advanced sensing, powerful computing, and sophisticated AI algorithms is leading to the
intellectualization of food detection equipment, enabling more accurate, rapid, and automated quality and safety assessment.
3.4. Differentiation Analysis
Compared to the main methods in related work, this paper's core differentiation and innovation lie in its comprehensive and comparative approach:
-
Holistic ML Coverage: Unlike reviews focusing solely on TML or DL, this paper introduces and contrasts both categories (
TMLandDL), providing a broader perspective on the available ML toolkit. -
Multi-Technology Integration: While other reviews often target a single nondestructive technology (e.g., only MV or only HSI), this paper integrates
acoustic analysis,machine vision,electronic nose, andspectral imagingwithin a single framework. This cross-technology perspective is crucial for understanding the diverse applications and challenges. -
Systematic Comparison: By presenting
Table 1(Comparison of DL against TML) and discussing specific algorithm choices, the paper aims to provide insights for better selection of ML techniques for different nondestructive technologies and application scenarios, a level of comparative analysis that was less explicit in prior single-focus reviews. -
Emphasis on "Device Intellectualization": The paper explicitly frames the combination of nondestructive techniques with advanced ML as a trend towards
device intellectualization, going beyond mere automation to intelligent decision-making.In essence, this review fills a gap by offering a systematic, side-by-side analysis of various ML paradigms (
TMLandDL) applied across a spectrum of novel nondestructive food sensing technologies, providing a more integrated and comparative understanding of the field's current state and future potential.
4. Methodology
As a review paper, the "methodology" is not an experimental procedure but rather the structured approach taken by the authors to survey, categorize, and synthesize existing research. The paper's methodology involves a systematic organization of knowledge, an introduction to foundational concepts, a comparative analysis of techniques, and a summary of applications, followed by a discussion of challenges and future directions.
4.1. Principles
The core principle guiding this review is to provide a comprehensive and insightful overview of how machine learning (ML) techniques are being utilized to enhance nondestructive food quality and safety detection. It aims to clarify the landscape of ML algorithms (categorized into Traditional Machine Learning (TML) and Deep Learning (DL)) and demonstrate their synergistic application with various nondestructive technologies across different food industry needs. The theoretical basis is that ML can overcome the limitations of traditional data analysis by effectively processing complex, high-dimensional, and noisy data generated by advanced sensors, thereby enabling more efficient and accurate real-time food assessment.
4.2. Core Methodology In-depth (Layer by Layer)
The paper's methodology can be broken down into the following structural and conceptual layers:
4.2.1. Introduction to the Problem and Scope (Section 1)
The review begins by establishing the necessity for nondestructive and rapid detection technologies in the food industry due to the global demand for high-quality food and the shortcomings of traditional destructive methods. It then identifies the key challenge: the complexity, time-consuming nature, and skill requirement for analyzing data from nondestructive technologies, alongside the limitations of conventional chemometric or statistical methods when faced with noise, sample variability, and data complexity.
Machine learning (ML) is introduced as a powerful solution for handling irrelevant information, extracting feature variables, and building calibration models, specifically highlighting its role in nondestructive technology and equipment intelligence. The scope of the review is then defined: an introduction and comparison of TML and DL, a summary of their applications with acoustic analysis, machine vision (MV), electronic nose (e-nose), and spectral imaging for food quality assessment, safety inspection, and processing control, concluding with a discussion of challenges and prospects.
4.2.2. Overview and Comparison of Machine Learning Techniques (Section 2)
This section forms the backbone of the review's conceptual methodology, systematically detailing ML approaches.
4.2.2.1. Traditional Machine Learning (TML) (Section 2.1)
The paper defines TML as algorithms that typically involve manual feature extraction on smaller sample sets, balancing validity with interpretability. It outlines the statistical foundations and categorizes TML into:
-
Supervised Learning (Section 2.1.1):
-
Support Vector Machine (SVM): Explained as a technique for
classificationandregressionthat constructs amaximal marginal hyperplaneto separate data points (e.g., "red star" and "blue star" categories in Figure 3(a)). It addressesoverfittingusingkernel functions(linear,radial basis function (RBF),polynomial,sigmoid).Support Vector Regression (SVR)is mentioned as its regression counterpart.
该图像是一个示意图,展示了图(a)支持向量机(SVM)分类,图(b) k近邻(k-NN)分类,图(c)单棵决策树分类,以及图(d)随机森林多棵决策树集成的基本原理和流程。The above figure (Figure 3(a) from the original paper) illustrates the core principle of a
Support Vector Machine (SVM). It shows how anSVMidentifies an optimalhyperplane(represented by the dashed line) that maximizes the separation margin between two classes of data points (labeled as "red stars" and "blue stars"). The points closest to the hyperplane are calledsupport vectors. -
Logistic Regression (LR): Described as a method for
binary classification(0 or 1) and estimating probabilities. It focuses on theodds(ratio of occurrence to non-occurrence probability).Multinomial Logistic Regression (MLR)is noted as an extension for more than two classes. -
K-Nearest Neighbor (KNN): Explained as a
classificationmethod based ondistanceorsimilaritymeasurement. An unclassified point is assigned the class of the majority of its nearest neighbors (illustrated in Figure 3(b)). The choice of is crucial for accuracy.
该图像是一个示意图,展示了图(a)支持向量机(SVM)分类,图(b) k近邻(k-NN)分类,图(c)单棵决策树分类,以及图(d)随机森林多棵决策树集成的基本原理和流程。The above figure (Figure 3(b) from the original paper) demonstrates the
K-Nearest Neighbor (KNN)algorithm. It shows an unclassified data point (black star) and how its classification depends on the value of . For , the black star is classified as "blue" due to two blue neighbors and one red neighbor. For , it is classified as "red" due to three red neighbors and two blue neighbors, illustrating the importance of the parameter. -
Algorithms based on Decision Trees (DT) (Section 2.1.1.4): Described as
tree-like structureswhere internal nodes areattribute judgmentsand leaf nodes areclassification resultsorcontinuous values(Figure 3(d)).Random Forest (RF)builds multiple deep, potentially overfitted trees and combines their outputs.Gradient Boosting(e.g.,XGBoost,CatBoost) builds shallow trees sequentially to reduce classification error.
该图像是一个示意图,展示了图(a)支持向量机(SVM)分类,图(b) k近邻(k-NN)分类,图(c)单棵决策树分类,以及图(d)随机森林多棵决策树集成的基本原理和流程。The above figure (Figure 3(c) from the original paper) depicts a single
Decision Tree, showing how it makes decisions based on attributes at each node to reach a classification at the leaf nodes. Figure 3(d) illustrates aRandom Forest (RF), an ensemble method that combines multiple such decision trees to make a more robust prediction. -
Naïve Bayes (NB): A
probabilistic classifierthat predicts values based onBayes' theoremand the assumption ofindependence between variables. It hasMultinomial,Poisson, andBernoullimodels. Feature selection is critical due to the independence assumption. -
Artificial Neural Networks (ANNs): Presented as
biologically inspired modelsforpredictionandclassification. They consist ofartificial neuronsin layers (Figure 4(b)), usingtransfer functions(e.g.,sigmoid,linear,hyperbolic tangent,logistic).Backpropagationis the typical training method to minimize theloss function. ANNs excel atnonlinear systemsbut requirelarge datasetsand carefulhyperparameter tuning.
该图像是三种神经网络结构的示意图,包括(a)卷积神经网络结构流程,(b)多层感知器结构,及(c)采用自编码器进行特征编码和解码的流程图。图中用表示感知器函数。The above figure (Figure 4(b) from the original paper) illustrates the typical structure of an
Artificial Neural Network (ANN)orMultilayer Perceptron (MLP). It shows an input layer, multiple hidden layers, and an output layer, with neurons (nodes) in each layer connected to neurons in adjacent layers. The perceptron function represents the weighted sum of inputs plus bias before activation.
-
-
Unsupervised Learning (Section 2.1.2):
- Dimensionality Reduction: Techniques to simplify
messyorhigh-dimensional datasets.- Principal Component Analysis (PCA): A
linear projectionmethod fordata decompositionandvisualizationthat maximizesvariancein lower dimensions. It identifies dominant patterns and can reduce data by selecting components explaining a high cumulative variance (e.g., 90%). - Other methods mentioned:
Minimum Noise Fraction (MNF)andIndependent Component Analysis (ICA). These are noted for theirlinear assumption, which may not suit inherentnonlinear structuresin nondestructive data.
- Principal Component Analysis (PCA): A
- Clustering: Groups
unlabeled data itemsintosimilar groupsbased on mathematical similarity.- K-means Clustering: A common algorithm that partitions data into clusters by iteratively assigning points to the nearest cluster centroid and updating centroids. The number of clusters is a crucial parameter.
- Dimensionality Reduction: Techniques to simplify
-
Semi-supervised Learning (Section 2.1.3): Addresses the challenge of limited
labeled databy combining small labeled sets with largeunlabeled data. It generally includesgenerative models,self-learning models,co-training models,transductive SVM (TSVM) learning models, andgraph-based learning models.Graph-based learningis highlighted for its better classification accuracy but higher computational complexity.
4.2.2.2. Deep Learning (DL) (Section 2.2)
DL is defined as representational learning using deep ANNs to extract features from raw data automatically for detection, classification, or regression.
-
Auto-encoder (AE): An
unsupervised neural networkusingbackpropagation. It acts as a feature extractor, mapping input to afeature vector(encoding) and reconstructing the input (decoding) (Figure 4(c)). Deep AEs stack layers for encoding and decoding. Various types exist (de-noising,sparse,variational,contractive). AE offers more flexibility than PCA by allowinglinearandnonlinearrepresentations.
该图像是三种神经网络结构的示意图,包括(a)卷积神经网络结构流程,(b)多层感知器结构,及(c)采用自编码器进行特征编码和解码的流程图。图中用表示感知器函数。The above figure (Figure 4(c) from the original paper) illustrates the architecture of an
Autoencoder (AE). It consists of anencoderthat compresses the input data into a lower-dimensionallatent-space representationand adecoderthat attempts to reconstruct the original input from this compressed representation. The goal is to learn an efficient encoding of the input. -
Convolutional Neural Network (CNN): A flourishing ML model since 1998, highly successful for image applications. Its architecture includes
convolutional layers(for feature asset representation through 2D symmetric operations and nonlinear transfer functions),pooling layers(to reduce dimensions and parameters), andfully connected layers(for final prediction) (Figure 4(a)). CNNs requirelarge datasetsbut can leveragetransfer learningwithpre-trained models(e.g.,ResNet-34,VGG-16/19,AlexNet,MobileNetv2) to improve performance on smaller datasets.
该图像是三种神经网络结构的示意图,包括(a)卷积神经网络结构流程,(b)多层感知器结构,及(c)采用自编码器进行特征编码和解码的流程图。图中用表示感知器函数。The above figure (Figure 4(a) from the original paper) presents the typical architecture of a
Convolutional Neural Network (CNN). It shows the sequential arrangement ofconvolutional layers(which extract features using filters),pooling layers(which reduce the dimensionality of feature maps), andfully connected layers(which perform classification based on the learned features). -
Restricted Boltzmann Machine (RBM): An
undirected graphical modelrepresentinghiddenandvisible layerswith symmetric connections but no intra-layer connections.Deep Belief Networks (DBNs)are multi-layer architectures incorporating RBMs.
4.2.2.3. Comparison (Section 2.3)
The paper provides a comparative analysis between DL and TML (summarized in Table 1, which will be presented in the Results section). Key aspects include data dependency (DL needs big data, TML for small/medium), hardware dependency (DL on GPUs, TML on CPUs), feature engineering (DL automatic, TML hand-crafted), training time (DL longer, TML shorter), logical explanatory (DL black-box, TML mathematical foundation), and multiple tasks (DL simultaneous, TML step-by-step).
4.2.3. Nondestructive Analytical Techniques (Section 3)
This section describes the principles of the four main nondestructive technologies and how ML integrates with them.
- Acoustic Techniques (Section 3.1): Based on extracting signal characteristics (propagation velocity, resonance frequencies, damping ratios) after sound contact or passage through food, relating to
mechanicalandphysical properties. ML (PCA,NB,LR,RF,ANNs,SVM) is used to buildfood quality inspection modelsfrom these acoustic parameters. - Electronic Nose (E-nose) (Section 3.2): Mimics human olfaction to detect
volatilesusing agas sensor array. ML is crucial forpattern recognitionfrom thetime-varying intensitiesof electric signals, performingfeature extraction,dimensionality reduction(PCA,LDA), andclassification/prediction(PLSR,SVM,RF,ANNs). - Machine Vision (MV) (Section 3.3): A
computer vision system (CVS)usingRGB camerasto capturecolorandtexture information. ML (TMLandDL) is applied toimage processingandanalysisto extract features (color, texture, shape, size, surface defects) and makeautomated decisionsforquality inspectionandidentification. - Spectral Imaging (Section 3.4): Combines
spectroscopyandimaging(e.g.,HSI) to acquirespatialandspectral data(hypercube). ML plays an essential role indata analysisforquality determination, particularly inwavelength selectionandfeature extractionto reduce redundancy. Calibration models are built usingTML(PLSR,SVM,RF,ANNs) andDL(CNN) to determinechemical compositions.
4.2.4. Recent Applications in the Food Industry (Section 4)
This section surveys practical applications, categorizing them by purpose and food type, and correlating them with specific nondestructive technologies and ML algorithms. This section uses Table 2 (presented in the Results section) to summarize findings.
- Recognition and Classification (Section 4.1): Examples include
food adulteration(e.g., sorghum, minced beef) andfood classification(e.g., fresh/frozen pork, Chinese liquors, bulk raisins), showing how ML improves accuracy, sometimes throughfeature-levelordecision-level fusionof multisensor data. - Quality Detection (Section 4.2):
- Fruits:
Ripeness classification,defect discrimination,prediction of soluble solids content (SSC)andfirmness(e.g., apples, loquat, kiwifruit, blueberries) using HSI, MV, e-nose, and acoustic methods combined with TML (PLSR, SVM, RF, ANNs) and DL (CNN, AE). - Meat and Aquatic Products:
DHA/EPA prediction,TVB-N,WHC,pH(e.g., salmon, crucian carp, cod),bone residues detection(e.g., salmon),texture detection(fish meat). Utilizes HSI, MV, Ultrasound with TML (PLSR, KNN, SVM, BP-ANN) and DL (CNN). - Other Products: Quality assessment for
nuts,tea,seeds,edible oil,ginger slices,mushroomusing various combinations of nondestructive tech and ML.
- Fruits:
- Safety Detection (Section 4.3):
- Chemical Contaminations: Detection of
pesticide residues(e.g., apples, mulberry fruit) andheavy metals(e.g., lettuce) using e-nose, LIBS, HSI, with TML (PCA, LDA, SVM, PLSR) and DL (WT-SCAE-SVR). - Microbial Contaminations: Identification of
fungal growthandpathogenic bacteria(e.g., peanuts, rice kernels, Agaricus bisporus) using HSI, e-nose, MV, with TML (KNN, RF, RBF-SVM, BP-ANN, PLSR, SVM, LVQ) and DL (CNN).
- Chemical Contaminations: Detection of
4.2.5. Challenges and Future Work (Section 5)
This section outlines the current limitations and proposes future research directions, serving as a forward-looking aspect of the methodology. It focuses on:
Lack of labeled dataand the potential oftransfer learning.- Need for
benchmark and open datasetsandStandard Operating Procedures (SOPs). - Addressing
environmental effectsandvariabilitythroughlifelong learning (LL)andreinforcement learning. - Optimizing
DLforlarge datasetsandcomputational demands, suggestinglightweight models(e.g.,SqueezeNet,ShuffleNet,MobileNet) for real-time, on-site detection. - Expanding
DL applicationstoe-noseandacoustic analysis, includingautomatic drift compensation.
4.2.6. Conclusions (Section 6)
A summary reiterating the potential of ML with nondestructive techniques and emphasizing the promise of DL for real-time applications.
5. Experimental Setup
As a review paper, this article does not present its own experimental setup in the traditional sense (i.e., no specific datasets, metrics, or baselines were generated by the authors of this paper for new experiments). Instead, it synthesizes and reports the experimental setups and results from numerous other research papers. Therefore, this section will describe the types of datasets, evaluation metrics, and baselines commonly reported in the reviewed literature, as presented in the paper.
5.1. Datasets
The paper discusses a wide array of datasets based on the specific food products and nondestructive technologies used in the reviewed studies. These datasets vary significantly in their characteristics, scale, and domain.
-
Source and Characteristics:
- Food Products: The datasets originated from diverse food items, including:
- Grains: Sorghum, maize kernels, wheat kernels, rice kernels, barley, peanuts, seeds.
- Meats & Aquatic Products: Minced beef, pork (fresh, frozen-thawed), salmon fillet, crucian carp fillets, cod, fish meat.
- Fruits & Vegetables: Gooseberry, loquat, apples, kiwifruit, yellow peach, blueberries, mulberry fruit, lettuce, potatoes, dry black goji berries, Agaricus bisporus (mushroom).
- Beverages & Other: Chinese liquors, bulk raisin, black tea, tea, olive oil.
- Data Types based on Nondestructive Technologies:
- Machine Vision (MV): Typically
RGB imagescontainingcolorandtexture information. For example,RGBimages of apples for defect detection, or images of bulk raisins for classification. - Hyperspectral Imaging (HSI):
Hypercubes(3D data) containing bothspatialandspectralinformation across hundreds of narrow wavelength bands. Examples include hyperspectral images of sorghum for adulteration, pork for quality, or blueberries forSSCandfirmness. - Electronic Nose (E-nose):
Time-varying intensitiesof electric signals from sensor arrays, formingvectorsorodor fingerprints. Used for detectingmycotoxin contaminationin maize,kiwifruit ripeness, orpesticide residueson apples. - Acoustic Analysis / Ultrasound:
Response signals(e.g.,acoustic emissions,ultrasonic velocity,attenuation coefficient,acoustic impedance) which can be analyzed in thetime domainorfrequency domain. Used formealiness detectionin apples orfish meat texture. - Laser-Induced Breakdown Spectroscopy (LIBS): Spectral data (intensity vs. wavelength) from plasma generated by a laser. Used for
thiophanate-methyl residuedetection on mulberry fruit.
- Machine Vision (MV): Typically
- Food Products: The datasets originated from diverse food items, including:
-
Scale and Domain: The paper notes that imaging technologies like MV and HSI generate
massive amounts of information. DL methods, in particular, are highlighted asdata-hungry, performing well withlarge datasets, while TML can work withsmall or medium datasets. The domain is consistentlyfood quality and safety detection. -
Example Data Sample (Conceptual):
-
For
Machine Vision, a data sample might be aJPEG imageof an apple, from which features likeaverage red color intensity,circularity of shape, ortexture variancecould be extracted. -
For
Hyperspectral Imaging, a data sample would be ahypercubefor a specific pixel or region on a food item, providing aspectrum(light intensity across wavelengths) that reveals its chemical composition. -
For
E-nose, a data sample is avectorofsensor responsesat a given time point, representing anodor fingerprint. -
For
Acoustic Analysis, a data sample could be afrequency spectrumof a vibration signal from a fruit, indicating itsfirmness.The choice of these datasets in the reviewed literature is driven by the specific
food quality attribute(e.g., ripeness, defect, adulteration, chemical composition) orsafety concern(e.g., pesticide, microbial contamination) being investigated, and the suitability of the chosen nondestructive technology to acquire relevant information.
-
5.2. Evaluation Metrics
The paper, through its summary of applications in Table 2, refers to several common evaluation metrics used in classification and regression tasks within machine learning.
-
Accuracy:
- Conceptual Definition: Measures the proportion of correctly predicted instances (both true positives and true negatives) out of the total number of instances. It is a general measure of correctness for classification models.
- Mathematical Formula: $ \text{Accuracy} = \frac{\text{Number of Correct Predictions}}{\text{Total Number of Predictions}} = \frac{TP + TN}{TP + TN + FP + FN} $
- Symbol Explanation:
TP: True Positives (correctly predicted positive instances).TN: True Negatives (correctly predicted negative instances).FP: False Positives (incorrectly predicted positive instances, type I error).FN: False Negatives (incorrectly predicted negative instances, type II error).
-
(Coefficient of Determination):
- Conceptual Definition: A statistical measure that represents the proportion of the variance in the dependent variable that can be predicted from the independent variables. It indicates how well the model explains the variability of the response data around its mean. A higher value generally indicates a better fit.
- Mathematical Formula: $ R^2 = 1 - \frac{\sum_{i=1}^{n} (y_i - \hat{y}i)^2}{\sum{i=1}^{n} (y_i - \bar{y})^2} $
- Symbol Explanation:
- : The actual value of the dependent variable for the -th observation.
- : The predicted value of the dependent variable for the -th observation.
- : The mean of the actual values of the dependent variable.
- : The total number of observations.
-
RMSEP (Root Mean Squared Error of Prediction):
- Conceptual Definition: A measure of the average magnitude of the errors between predicted values and actual values. It's the square root of the average of the squared differences between prediction and actual observation. It gives a relatively high weight to large errors. Lower RMSEP indicates better predictive performance.
- Mathematical Formula: $ \text{RMSEP} = \sqrt{\frac{1}{n} \sum_{i=1}^{n} (y_i - \hat{y}_i)^2} $
- Symbol Explanation:
- : The actual value for the -th observation.
- : The predicted value for the -th observation.
- : The total number of observations.
-
RPD (Ratio of Performance to Deviation):
- Conceptual Definition: Used in spectroscopy, RPD is a measure of the predictive performance of a calibration model, indicating the ratio of the standard deviation of the reference values to the standard error of prediction. A higher RPD value (typically > 2.0) indicates a robust and reliable model for prediction.
- Mathematical Formula: $ \text{RPD} = \frac{SD}{\text{RMSEP}} $
- Symbol Explanation:
SD: Standard Deviation of the reference (actual) values.- : Root Mean Squared Error of Prediction.
-
Recall (Sensitivity or True Positive Rate):
- Conceptual Definition: Measures the proportion of actual positive instances that were correctly identified by the model. It's important when the cost of false negatives is high (e.g., missing a defect).
- Mathematical Formula: $ \text{Recall} = \frac{TP}{TP + FN} $
- Symbol Explanation:
TP: True Positives.FN: False Negatives.
-
Specificity (True Negative Rate):
- Conceptual Definition: Measures the proportion of actual negative instances that were correctly identified by the model. It's important when the cost of false positives is high (e.g., incorrectly identifying a good product as defective).
- Mathematical Formula: $ \text{Specificity} = \frac{TN}{TN + FP} $
- Symbol Explanation:
TN: True Negatives.FP: False Positives.
-
F1-score:
- Conceptual Definition: The harmonic mean of
PrecisionandRecall. It is a balanced metric that considers both false positives and false negatives, especially useful when class distribution is uneven.Precision=
- Mathematical Formula: $ \text{F1-score} = 2 \times \frac{\text{Precision} \times \text{Recall}}{\text{Precision} + \text{Recall}} $
- Symbol Explanation:
- : Proportion of positive identifications that were actually correct.
- : Proportion of actual positives that were correctly identified.
- Conceptual Definition: The harmonic mean of
-
PR-AUC (Area Under Precision-Recall Curve):
- Conceptual Definition: The area under the
Precision-Recall Curve, which plotsPrecisionon the y-axis andRecallon the x-axis for differentthresholds. It is particularly useful for evaluatingimbalanced datasetswhere the positive class is rare, as it focuses on the performance on the positive class. A higher PR-AUC indicates better performance. - Mathematical Formula: There is no single explicit formula for PR-AUC as it's calculated by integrating the area under the curve formed by various precision and recall points. It is typically approximated using numerical integration methods.
- Conceptual Definition: The area under the
-
Correlation Coefficient ():
- Conceptual Definition: A statistical measure that expresses the extent to which two variables are linearly related. It ranges from -1 to +1. A value of +1 indicates a perfect positive linear relationship, -1 a perfect negative linear relationship, and 0 no linear relationship.
- Mathematical Formula (Pearson correlation coefficient): $ R = \frac{n(\sum xy) - (\sum x)(\sum y)}{\sqrt{[n\sum x^2 - (\sum x)^2][n\sum y^2 - (\sum y)^2]}} $
- Symbol Explanation:
- : Number of data points.
- : Values of the first variable (e.g., predicted values).
- : Values of the second variable (e.g., actual values).
- : Sum of the product of and values.
- : Sum of values.
- : Sum of values.
- : Sum of squared values.
- : Sum of squared values.
5.3. Baselines
The reviewed papers commonly compare the proposed ML methods against:
- Traditional Chemometric/Statistical Methods:
PCA,PLS,LDA, which are often linear and require more manual feature engineering. For instance,FDPCAwas compared againstPCAandDPCA. - Other TML Algorithms: Different
TMLalgorithms are often benchmarked against each other (e.g.,SVMvs.KNNvs.RFvs.ANNsvs.NBvs.LR) to find the best performing model for a specific task. - Deep Learning Models:
CNNsare often compared againstTMLclassifiers (e.g.,KNN,RF,RBF-SVM,BP-ANN) to demonstrate the superiority ofDLinfeature extractionandaccuracyfor image-based tasks. DifferentCNN architectures(e.g.,AlexNet,VGG-16/19,ResNet) are also compared. - Ablation Studies/Feature Engineering Variants: Comparisons are also made regarding the impact of
feature selection methods(e.g.,CARS,SPA,UVE,GA) orpreprocessing methodson model performance, often implicitly serving as baselines for the full proposed method. - Human Inspection/Conventional Methods: Although not explicitly shown in tables, the underlying motivation for nondestructive ML techniques is to surpass the limitations of
destructive,labor-intensive, orsubjective human inspectionmethods.
6. Results & Analysis
The paper synthesizes a vast array of results from various studies, demonstrating the effectiveness of combining machine learning (ML) techniques with nondestructive technologies for food quality and safety detection. The results generally indicate superior performance of ML-enhanced systems compared to traditional methods, with Deep Learning (DL) often outperforming Traditional Machine Learning (TML) in complex, data-rich scenarios, especially for vision-based tasks.
6.1. Core Results Analysis
The paper highlights how ML algorithms significantly improve the accuracy and robustness of nondestructive detection.
-
Food Adulteration, Recognition, and Classification:
- Studies show high accuracy in detecting food adulteration (e.g.,
sorghum adulterationusing HSI+PLS-DA with 91% accuracy;minced beef recognitionusing HSI+PLSR with ). - For classification tasks, the fusion of multiple sensor data (e.g.,
E-noseandMVfortea quality) can boost accuracy (from 78% to 92% formulti-sensor datawithSVM). - DL models, particularly
CNNswithtransfer learning, showed enhanced performance inbarley classification(from 88.15% to 92.23% accuracy), demonstrating their ability to extract relevantspatialandspectralinformation from hyperspectral images.
- Studies show high accuracy in detecting food adulteration (e.g.,
-
Food Quality Detection (Fruits, Meat, Aquatic, Other Products):
- Fruits: ML models successfully predict
ripeness, detectdefects, and quantifychemical compositions(SSC,firmness,total phenolics/flavonoids/anthocyanins). For instance,PLSRmodels withHSIfor blueberries achieved forSSCand forfirmness.CNNsanddeep AEused as feature extraction methods withHSIforblack goji berriesshowed high (0.897 for total anthocyanins).CNNsalso showed highaccuracy(92%) fordefective apple detectionusingMV. - Meat and Aquatic Products:
HSIcombined withPLSRaccurately predictedliquid lossincod(, RMSEP=0.58%). DL models likeAlexNetandVGGachieved high accuracy (0.75-0.87 F1-score) forbone residue detectionin salmon, despite some vulnerability in industrial application. - General Performance: TML algorithms (
SVM,BP-ANN,KNN) are powerful tools forquality classificationandnutrient prediction. The paper notes thatSVMis a good choice for nondestructive technologies due to its ability to handlenonlinear problemsand avoidover-learning.
- Fruits: ML models successfully predict
-
Food Safety Detection (Chemical and Microbial Contaminations):
-
Chemical Contaminations:
E-nosewithTMLalgorithms (PCA,LDA,SVM) could detectpesticide residueson apples with 94% overall accuracy.LIBSandHSIwithPLSRdetectedthiophanate-methyl residueon mulberry fruit (). Advanced DL methods (e.g.,WT-SCAE-SVR) demonstrated high values (0.9319 for Cd, 0.9418 for Pb) forheavy metal detectionon lettuce. -
Microbial Contaminations:
HSIcombined withCNNachieved high recognition rates (over 96% pixel-level, over 90% kernel-level) foraflatoxin detectionin peanuts, outperforming traditional TML models.E-nosewithBP-ANNaccurately classified and predictedAspergillus spp. contaminationin rice (96.4% accuracy).Overall, the results consistently support the premise that integrating ML, particularly DL for complex imaging data, significantly enhances the capabilities of nondestructive technologies. However, the paper also implicitly highlights that the choice of ML algorithm (TML vs. DL) and specific variant depends heavily on the
data characteristics,computational resources, andspecific application requirements.
-
6.2. Data Presentation (Tables)
The following are the results from Table 1 and Table 2 of the original paper:
The following are the results from Table 1 of the original paper:
| Features | Deep learning (DL) | Traditional machine learning (TML) |
|---|---|---|
| Data dependency | Requirement of big data for training | Requirement of small or medium datasets for training |
| Hardware dependency | Dependency of graphics processing units (GPUs) and storage rooms for accelerating the training process | Dependency of low-end processors such as central processing units (CPUs) |
| Feature engineering | An end-to-end mapping from the input to output automatically | Requirement of hand-crafted features, human expertise and complicated task-specific |
| Training time | Taking much time | optimization Taking a relatively short time |
| Logical explanatory | A black-box method and confusion of the hyperparameters for designing the complex | Requirement of a certain mathematical theoretical |
| Multiple tasks | network Completing simultaneously | foundation Completing step by step Higher accuracy in a |
| Others | Higher accuracy in vision-based tasks such as objection detection, classification, and segmentation Having the premier flexibility and generalization ability | simple task with a small dataset Applications in a specific domain |
The above table (Table 1 from the original paper) provides a comparison between Deep Learning (DL) and Traditional Machine Learning (TML) across several key features, highlighting their differences in data requirements, hardware needs, feature engineering processes, training time, interpretability, and task handling capabilities.
The following are the results from Table 2 of the original paper:
| Applications of nondestructive qualitative analysis in food. | |||||||
| Product | Purpose | Tools | Data processing | Feature selection | ML classifiers | Best results | Reference |
| Food adulteration | |||||||
| Sorghum | Adulteration recognition | HSI | MSC | PCA | PLS-DA | Accuracy > 90% | Bai et al. (2020) |
| Minced beef | Adulteration recognition | HSI | MSC, SNV, 1st and 2nd derivatives, and SG | Stepwise regression | PLSR | R2= 0.97, RMSEP = 2.61%, RPD = 5.86 | Kamruzzaman, Makino, and Oshita (2016) |
| Pork | Adulteration recognition | E-nose | The DWT and several mother wavelets, PCA Box plot and | Standard deviation, mean, kurtosis, skewness Time domain | SVM, ANNs, LDA, KNN NB, SVM, | Accuracy = 98.10% | Sarno et al. (2020) |
| Food classification | |||||||
| Thawed pork | Fresh and frozen pork classification | HSI | PCA | HS, GLCM, GLGCM, SPA, | PNN | Accuracy = 94.14% | Pu et al. (2015) |
| Meat product | Category classification | HSI | Graph-based post-processing | UVE 3D-CNN | 3D-CNN, PLS-DA, | Accuracy = 97.1% | Al-Sarayreh et al. (2020) |
| Chinese liquors | Category classification | E-nose | method Data correction, average value, normalization | FDPCA, DPCA, PCA | SVM KNN | Accuracy = 98.78% | X. H. Wu et al. (2019) |
| Bulk raisin | Category classification | MV | Image thesholding | GLCM, GLRM, LBP | PCA, SVM, LDA | Accuracy = 85.55% | Khojastehnazhand and Ramezani (2020) |
| Fruit quality detection | |||||||
| Gooseberry | Ripeness classification | MV | PCA | Color feature extraction | ANNs, DT, SVM, KNN | Accuracy = 93.02% | Castro et al. (2019) |
| Loquat | Defect discrimination | HSI | SNV, SG, 1st and 2nd derivatives, gap segment derivative | 200 random models, Monte Carlo method | PLS-DA, RF, XGBoost | Accuracy > 95.9% | Munera et al. (2021) |
| Apples | Defect detection | MV | Otsu thresholding | CNN, RGB color, GLCM | CNN, SVM | Recall = 91%, Specificity = 93%, Accuracy = 92% | Fan et al. (2020) |
| Kiwifruit | Ripeness prediction | E-nose | Time domain features | PLSR, SVM, RF | Accuracy > 99.4%, R² > 0.9143 | Du et al. (2019) | |
| Apple | Mealiness detection | Acoustic | STFT | CNN | AlexNet, VGG | Accuracy = 91.11% | Lashgari, Imanmehr, and Tavakoli (2020) |
| Aquatic and meat products quality detection | |||||||
| Salmon fillet | DHA and EPA prediction | HSI | Image thesholding | GA, PNN | PLSR, MLR, PCR | DHA:R2 = 0.829, RMSEP = 0.585, EPA: R2 = 0.843, RMSEP = 0.195 | Cheng et al. (2019) |
| Crucian carp fillets | TVB-N, WHC, pH | HSI | Image thesholding | SPA, K-means | PLSR | TVB-N: R2 = 0.79, WHC: R2 = 0.71, pH: R2 = 0.88 | Wang et al. (2019) |
| Cod | Liquid loss of cod | HSI | PCA, 1st derivative | GA | PLSR, KNN | R2 = 0.88, RMSEP = 0.62% | Anderssen et al. (2020) |
| Salmon | Bone residues detection | MV | Data augmentation, JEPG lossy compression | Faster-RCNN | AlexNet, VGG | F1-score = 0.87, PR-AUC = 0.76 | Xie et al. (2021) |
| Fish meat | Fast content and texture detection | Ultrasound | Low-pass filter, pre-emphasis, STFT, | SOM | RBF network | R2 = 0.89 (p < 0.05) | Tokunaga et al. (2020) |
| Other products quality detection | |||||||
| Nuts | HSI | normalization PCA | CNN | CNN-SVM | Accuracy = | Han et al. (2021) | |
| Black tea | Quality grades classification | MV, MNIR | SNV, PCA | Color and texture feature extraction | PCA-SVM | 94.29% | Li et al. (2021) |
| Tea | Quality identification | MV, E-nose | Color and texture feature extraction | KNN, SVM, MLR | Accuracy = 100% | Xu, Wang, and Gu (2019) | |
| Seed | Viability prediction | HSI | PCA, DWT | CNN | CNN, SVM | Accuracy > 90% | Ma, Tsuchikawa, and Inagaki (2020) |
| Chemical contamination | |||||||
| Mulberry fruit | Thiophanate-methyl residue detection | LIBS, HSI | SNV, PCA | CARS | PLSR, PCA | Correlation coefficient = 0.921 | D. Wu et al. (2019) |
| Chlorella pyrenoidosa | Pesticide residues detection | HSI | WT, SG smoothing, and baseline correction | SPA | PLS-DA, LDA | Accuracy = 90% | Shao et al. (2016) |
| Apples | Pesticide residue detection | E-nose | Radar map and bar char | Heat map, PCA, LDA | PCA, LDA, SVM | Accuracy = 99.37% | Tang et al. (2021) |
| Microbial contamination | |||||||
| Agaricus bisporus | Fungal growth detection | HSI, NIR, MIR, E-nose | MSC, SNV, 1st and 2nd derivatives, normalization | PCA | PLSR, PLS-DA | R2 = 0.670 ~ 0.821, Accuracy = 99% | Wang et al. (2020) |
| Peanuts | Mouldy identification | HSI | Image thresholding | KNN, RF, RBF-SVM, BP-ANNs, CNN | Accuracy = 97.26% | Han and Gao (2019) | |
| Rice kernels | Aspergillus spp. discrimination | E-nose | PCA | PLSR, BP-ANN, SVM, LVQ | Accuracy = 96.4%, R2 = 0.917 | Gu, Wang, and Wang (2019) | |
| Bacteria | Species classifi cation | MV | PCA | SVM | Accuracy = 93.3% | Kim et al. (2021) | |
The above table (Table 2 from the original paper) comprehensively lists various applications of nondestructive technologies combined with machine learning (ML) for food quality and safety detection. It details the product, purpose of detection, specific tools (nondestructive technologies), data processing techniques, feature selection methods, ML classifiers used, the best results achieved (often in terms of accuracy or ), and the corresponding references. This table serves as a concrete summary of the empirical evidence supporting the paper's claims.
6.3. Ablation Studies / Parameter Analysis
The review paper does not present new ablation studies or parameter analyses, but it references how various studies incorporate these concepts:
-
Feature Selection Methods: Many studies mentioned in Table 2 or the text utilize different
feature selectionorfeature extractionmethods (PCA,SPA,UVE,CARS,GA,FDPCA,DWT,CNNas feature extractor) to reducedata dimensionalityand improve model performance. Comparing models with and without these steps, or with different feature sets, serves as an implicit form of ablation. For example, the use ofGenetic Algorithms (GA)withPLSRforcod liquid lossprediction showed modest improvements and reduction of necessary components, indicating the value of feature selection. -
Hyperparameter Optimization: For algorithms like
KNN, the parameter needs optimization. ForSVM,regularization parameter (C)and are optimized, especially forRBF kernels.Random Forestmodels optimize thenumber of decision treesandfeatures in each tree. While the paper doesn't detail specific optimization processes, it highlights their importance for achieving the reported accuracies. -
Preprocessing Methods: The impact of
data preprocessing methods(e.g.,MSC,SNV,SG,first/second derivatives,normalization,baseline correction,image thresholding) on model performance is frequently investigated in the referenced studies, which can be seen as an analysis of component contribution to the overall system. -
Comparison of ML Algorithms: The explicit comparisons between
TMLandDLalgorithms, or between differentTMLalgorithms (e.g.,KNN,RF,RBF-SVM,BP-ANNvs.CNNfor aflatoxin detection), effectively demonstrate the relative strengths of different modeling components for specific tasks and data types.These analyses, embedded within the cited works, confirm that individual components like specific
feature engineeringtechniques,preprocessing steps, or the choice ofML architecturesignificantly influence the finalpredictive accuracyandrobustnessof the nondestructive food detection systems.
7. Conclusion & Reflections
7.1. Conclusion Summary
The paper effectively concludes that the integration of machine learning (ML) techniques with nondestructive inspection technologies offers substantial potential for advancing food quality and safety assessment. It systematically reviews traditional machine learning (TML) and deep learning (DL) algorithms, detailing their principles and diverse applications across acoustic analysis, machine vision, electronic nose, and spectral imaging. The review highlights that ML is crucial for efficiently extracting useful information from complex, high-dimensional data generated by these sensing technologies, overcoming limitations of conventional chemometric methods. While acknowledging the value of various ML algorithms in different contexts, the paper particularly emphasizes deep learning as a powerful and promising technique for real-time applications due to its inherent feature learning capabilities, although it notes the need for further research to enable its full and wide applications in the food industry.
7.2. Limitations & Future Work
The authors identify several critical challenges and propose future research directions:
- Lack of Labeled Data: A significant barrier for
MLapplications, especiallyDL, is the scarcity oflabeled data. Manually labeling optimal features from complex data likehyperspectral imagesis time-consuming and challenging.- Future Work: Focus on
transfer learning(inductive, transductive, unsupervised) to leverage existing knowledge with minimal new labeled data, by re-weighting data, finding suitable feature representations, or building relational knowledge between domains.
- Future Work: Focus on
- Lack of Benchmark Datasets and Standardization: The absence of publicly available, standardized
benchmark datasetsacross various nondestructive technologies hinders large-scale application and comparative analysis of different techniques.- Future Work: Establish
Standard Operating Procedures (SOPs)for creating benchmark and open datasets for each nondestructive technology.
- Future Work: Establish
- Environmental and Dynamic Variability: Factors like
variety differences,instrument effects, andenvironmental conditions(e.g.,lighting,humidity,temperature) dynamically alter food quality patterns, necessitating costly and time-consumingmodel recalibration.- Future Work: Explore
lifelong learning (LL)withneural networksandreinforcement learningto build robust models that accumulate knowledge over time, akin to human intelligence. This involves integratingregularization,ensembling,rehearsal, anddual-memorymethodologies to account for various influencing factors.
- Future Work: Explore
- Computational Demands of Deep Learning:
DLisdata-hungryand requires significanttraining timeand expensiveGPUsandstorage rooms, posing challenges for further development, especially forHSIandMV.- Future Work: Focus on shortening
training timesand simplifyingDL model architectures. Developlightweightandefficient models(e.g.,SqueezeNet,ShuffleNet,MobileNet,GhostNet) that offer better accuracy with fewer parameters, suitable foron-siteandreal-time detectionon portable devices.
- Future Work: Focus on shortening
- Limited DL Application in E-nose and Acoustic Analysis: While
MVandHSIextensively useDL, its application ine-noseandacoustic analysisis less developed.- Future Work: Expand
DL techniquestoe-noseandacoustic analysis, for example, as anautomatic drift compensation methodto addresssensor driftcaused by environmental conditions without manual rule-setting.
- Future Work: Expand
7.3. Personal Insights & Critique
This review serves as an excellent entry point for anyone interested in the intersection of machine learning and nondestructive food quality/safety detection. Its strength lies in its comprehensive structure, clearly differentiating TML from DL and mapping their applications across diverse sensing technologies and food products. The comparison table (Table 1) is particularly helpful for beginners to grasp the fundamental trade-offs between DL and TML.
One key insight drawn from this paper is the undeniable shift towards deep learning for vision-based tasks (MV, HSI), where its ability to automatically extract complex features from high-dimensional raw data offers significant performance advantages over traditional methods requiring manual feature engineering. This highlights the increasing importance of raw data quality and volume as direct inputs to DL models.
A potential area for improvement, though typical for a high-level review, is the relative lack of detail on how certain ML models are implemented within specific food contexts. For a truly beginner-friendly guide, elaborating on the typical data preprocessing pipelines, feature engineering steps for TML, or specific CNN architectures used for a given task would be beneficial. For instance, explaining the general workflow of a CNN for image classification in the context of apple defect detection with concrete examples of convolutional layers and pooling operations would deepen understanding.
From a critical perspective, while the paper emphasizes DL's power, it also implicitly reveals its primary practical hurdle: the data-hungry nature and computational expense. The call for lightweight models and lifelong learning is critical, as real-time industrial applications demand not just accuracy but also efficiency and adaptability to changing conditions without constant, costly retraining. The black-box nature of DL also remains a concern for regulatory environments where interpretability and explainability are highly valued.
The paper's emphasis on standardized datasets and SOPs is a crucial point that often gets overlooked. Without these, comparing research, reproducing results, and deploying models at scale remain challenging. The concepts of transfer learning and lifelong learning are vital for overcoming the cold-start problem (lack of initial labeled data) and the catastrophic forgetting problem (loss of old knowledge when learning new tasks) in real-world, dynamic food processing environments.
Overall, this review provides a robust framework for understanding the current state of the art and clearly charts a path for future research, particularly in making DL more accessible, efficient, and interpretable for the complex and dynamic domain of food quality and safety detection.
Similar papers
Recommended via semantic vector search.