Movatterモバイル変換


[0]ホーム

URL:


Next Article in Journal
The Adoption and Openness of Livestreaming on the Retail Platform with Third-Party Sellers
Next Article in Special Issue
Optimal Recommendation Strategies for AI-Powered E-Commerce Platforms: A Study of Duopoly Manufacturers and Market Competition
Previous Article in Journal
A Recommendation System in E-Commerce with Profit-Support Fuzzy Association Rule Mining (P-FARM)
Previous Article in Special Issue
The Impact of Return Shipping Insurance on a Retailer Based on Restricting Rights
 
 
Search for Articles:
Title / Keyword
Author / Affiliation / Email
Journal
Article Type
 
 
Section
Special Issue
Volume
Issue
Number
Page
 
Logical OperatorOperator
Search Text
Search Type
 
add_circle_outline
remove_circle_outline
 
 
Journals
JTAER
Volume 18
Issue 2
10.3390/jtaer18020044
Font Type:
ArialGeorgiaVerdana
Font Size:
AaAaAa
Line Spacing:
Column Width:
Background:
Article

An Explainable Artificial Intelligence Approach for Multi-Criteria ABC Item Classification

1
MIS Department, College of Business, University of Jeddah, Jeddah 23218, Saudi Arabia
2
LARODEC Lab, ISG Tunis, University of Tunis, Le Bardo 2000, Tunisia
3
Knowledge Discovery & Web Mining Lab, University of Louisville, Louisville, KY 40292, USA
*
Authors to whom correspondence should be addressed.
J. Theor. Appl. Electron. Commer. Res.2023,18(2), 848-866;https://doi.org/10.3390/jtaer18020044
Submission received: 19 January 2023 /Revised: 9 March 2023 /Accepted: 13 March 2023 /Published: 11 April 2023

Abstract

:
Multi-criteria ABC classification is a useful model for automatic inventory management and optimization. This model enables a rapid classification of inventory items into three groups, having varying managerial levels. Several methods, based on different criteria and principles, were proposed to build the ABC classes. However, existing ABC classification methods operate as black-box AI processes that only provide assignments of the items to the different ABC classes without providing further managerial explanations. The multi-criteria nature of the inventory classification problem makes the utilization and the interpretation of item classes difficult, without further information. Decision makers usually need additional information regarding important characteristics that were crucial in determining the managerial classes of the items because such information can help managers better understand the inventory groups and make inventory management decisions more transparent. To address this issue, we propose a two-phased explainable approach based on eXplainable Artificial Intelligence (XAI) capabilities. The proposed approach provides both local and global explanations of the built ABC classes at the item and class levels, respectively. Application of the proposed approach in inventory classification of a firm, specialized in retail sales, demonstrated its effectiveness in generating accurate and interpretable ABC classes. Assignments of the items to the different ABC classes were well-explained based on the item’s criteria. The results in this particular application have shown a significant impact of the sales, profit, and customer priority as criteria that had an impact on determining the item classes.

    1. Introduction

    Efficient inventory management plays an essential role in the management optimization of the overall supply chain. The procurement of raw materials and the manufacturing of products with the lowest stock-levels significantly contribute to cost reduction. Given that hundreds to thousands of items must be monitored and managed, it would be unrealistic to focus on each item separately. To address this issue, managers often attempt to use machine learning and data-driven methods to automatically classify items, commodities, and products into inventory classes in order to effectively manage each inventory group of items. This automatic process is referred to as inventory classification or item classification.
    To effectively control the inventory and optimize the management of the supply chain, accurate inventory classification is required. Several methods and techniques for performing automatic inventory classification have been proposed [1,2,3]. ABC classification is one of the well used methods which divides inventory items into three managerial classes based on item description criteria. The first class, referred to as class A, contains a relatively small number of items that mostly contribute to the activity of the company. The second class, referred to as class B, includes items that are considered important, but having lesser importance than those of class A. The third class, referred to as class C, contains a large number of items having the least importance.
    Although the classification of items was previously carried out based on one criterion, the “annual dollar usage” [4], several more recent works [5] in the literature consider other criteria as a basis for the classification of the items, including inventory holding unit cost, variability of replenishment lead time, scarcity, dependence to other items, substitutability, etc. The multi-criteria nature of the inventory classification problem makes the utilization and the interpretation of items classes, when these are derived from a black box process, difficult to understand. This is because knowing only which item belongs to which class is insufficient in real life business applications. Managers need to also understand which criterion and/or criteria has really contributed to the decision of assigning an item to a specific class rather than assigning it to the other classes. Unfortunately, when the automatic classification process operates as a black-box, it only gives a final organization of the items into three managerial classes without providing explanations about the reasons for assigning the items to a class. Yet, such explanations, when available, may help decision makers to transparently and effectively determine the right inventory management strategies for the items.
    To solve this issue, we propose an eXplainable Artificial Intelligence (XAI) approach for multi-criteria ABC item classification. The proposed approach is based on the explainable artificial intelligence framework, SHape Additive exPplanations (SHAP), that provides an easy schematizing of the contribution of each criterion when building the inventory classes. It also allows to explain reasons behind the assignment of each item to any class. Such explanations make the resulting ABC inventory classes more transparent for decision makers. The rest of the paper is organized as follows:Section 1 gives an overview of inventory classification methods while describing the different criteria used to organize the inventory. Then,Section 2 describes the theoretical background for explainable artificial intelligence and the SHAP framework.Section 3 andSection 4 then describe the theoretical framework and empirical experiments for the proposed explainable clustering method for inventory organization. Finally,Section 5 gives the conclusion and future directions.

    1.1. Literature Review of Inventory Classification Methods

    Given the complexity of enterprise operations and architectures, an effective management of the inventory requires intelligent tools, techniques and methods to better increase service efficiency. In this context, ABC inventory classification is widely used to automatically organize the items into three groups of different managerial-levels and sizes. Several methods, based on different criteria and principles, have been proposed for ABC inventory classification [6,7,8]. Existing methods can be classified based on the used approach to building the different classes. These approaches can be categorized into: (i) Decision making, (ii) Mathematical programming and (iii) Machine learning and soft computing approaches. We review these approaches below.
    The first category of item classification methods, decision-making methods, are based on probabilistic or conceptual data modeling techniques to look for an optimal classification of the items. Usually, these methods consist of several phases to solve the multi-criteria problem. An example of such methods is the Analytic Network Process (ANP) method [9] for evaluating the logistic strategies and the production speed. This method models the multi-criteria problem by a network involving the internal dependency of criteria to make the decision-making more efficient. Other existing works [10,11] proposed to extend Annual Dollar Usage, and Scoring methods to build an ABC classification of the items. Examples of such methods are those proposed by Liu and Hung [10] who proposed a Data Envelopment Analysis (DEA) model to define the ABC classes and Onwubolu and Dube [11] who proposed using spreadsheets to solve the multi-criteria problem. More recent decision-making methods have been proposed to build inventory classes such as the works of Zheng et al. [12] who applied Shannon’s entropy to look for optimal ABC classes and the works of Wu et al. [13] who used the weighted least-squares dissimilarity approach to obtain a set of local item weights which are then aggregated in the overall evaluation score function. Another interesting decision-making system is proposed in the works of Eraslan and Ic [3] who developed an efficient decision support system to help decision-makers build the ABC classes of the items. This system gives high flexibility to decision-makers by introducing subjective criteria.
    The second category of item classification methods, mathematical programming methods, are essentially based on linear and nonlinear programming models to build a vector or a matrix that optimizes weighted item scores. An overall score is usually calculated by an objective function that would be optimized. Examples of conventional mathematical programming methods include the weighted linear optimization model [14] and the Ng-model [15]. Extensions of the Ng-model were presented in the works of [16] that proposed both the H-model and F-model approaches. More recent mathematical-based methods incorporate a cross-efficiency evaluation method into a weighted linear optimization model for ranking the items, as presented in the works of [2] which proposed a Cross-Evaluation methodology to optimize the weighted linear optimization problem. In addition, the work in [17] integrated an alternative overall measure within the Ng-model to optimize the descending ordering of the items.
    The third category of existing methods comes from the machine learning and soft computing domains. A learning process is integrated that allows estimating the final assignment of the items while optimizing a learning model or an objective function. Several machine learning classification methods are used to build the ABC classes such as k-means, Fuzzy c-means (FCM), K-nearest neighbors and Support Vector Machines (SVM) algorithms. In [18], Chu et al. proposed the ABC-Fuzzy Classification (ABC-FC) method which is based on the Fuzzy c-means (FCM) algorithm and incorporates decision makers’ judgment of inventory classification. In the same way, Keskin and Ozkan [1] and Cebi et al. [19] proposed to design an FCM-based process to solve the multi-criteria classification problem in order to help managers make better decisions under fuzzy circumstances. In the works of [5], the authors proposed the AHP–FCM–Rveto method that is based on three phases to build the final inventory classes which are AHP phase, FCM phase and Revised-Veto phase. Other machine learning techniques have been used to build the ABC classes such as Artificial Neural Networks. Partovi and Anandarajan [20] proposed a genetic algorithm (GA)-based learning method to develop an artificial neural network for inventory classification. The obtained results have shown that neural network-based inventory classification can give higher predictive accuracy than conventional inventory classification methods. An interesting comparative study is presented in the works of [21]. The authors compared the performance of three machine learning-based classification models, namely SVM, Back Propagation Networks and KNN when building inventory classes.

    1.2. ABC Inventory Classification: Problem Definition and Challenges

    LetX =x1xn be a dataset ofn inventory items, also called Stock Keeping Units (SKU), where each itemxi is described overm quantitative criteriaj1,j2,...,jm. We also assume that them criteria are of benefit-type and have a positive relation to the importance level of the items. If one criterion is not positively related to the importance level of the items, the reciprocal can be used to make it positively related. The objective of the ABC inventory classification process is to organize then items into three groups (classes) of different sizes namely A, B and C. The distribution of the items over the three different groups must follow as follows: class A items will be a small percent of inventory, but contribute most in terms of activity and revenue, class C items will form a large percent of inventory, but contribute the least to activity and revenue, and finally, class B items will fall in the middle. Theoretically, ABC aims to build an assignment matrixU(n,3) where each rowi corresponds to an itemxi, and each columnt represents one class. The matrix valuesuit,i[1..n],t[1...3] are binary where the valueuit=1 corresponds to the categorization of an itemxi as belonging to the classt. To meet the ABC classification model principles, the sizes of class A, B and C must be in the intervals[10%...20%],[20%...30%] and[50%...70%], respectively.Figure 1 gives an overview of the inputs and outputs of the multi-criteria ABC inventory classification process.
    In this work, we focus onlearning-based multi-criteria ABC inventory classification methods which are usually based on the application of conventional clustering techniques to build the different ABC classes. Although this category of methods has been widely used and has shown good performance, it raises the issue of non-explainability and non-transparency of results in real-life applications. In fact, the clustering algorithm works like a black-box process that ends by giving a final categorization of the items into three groups without giving sufficient explanations relative to the decision of assigning an item to any particular class. In practice, given the multi-criteria nature of the problem, managers usually need additional information regarding the criterion (or criteria) that was crucial in determining the assignment of an item to the A, B or C class. Such information allows a better analysis of the items, allows easy detection of miss-classified items, helps managers to better understand the inventory groups and makes inventory management decisions more flexible. To the best of our knowledge, there is no existing work that studied the explainability of ABC inventory classes.
    To deal with the non-explainability of ABC inventory classes, we propose a new multi-criteria ABC inventory classification approach based on Explainable Artificial Intelligence (XAI) capabilities. We present, in the next section, the background and basic concepts of the used approach.

    2. Background on Explainable Clustering

    2.1. Explainable Clustering

    Clustering offers a solution to analyze a large amount of real-world unlabeled data. However, most of the existing clustering methods do not provide a way for decision-makers to understand their results especially for non-domain experts. Hence, these methods act as “black-boxes” and build clusters without providing any explanations regarding the reasoning behind their creation. This issue reduces user trust and makes interpretations and choices less transparent.
    Explainable clustering, a branch of Explainable Artificial Intelligence (XAI), attempts to address this problem by allowing decision-makers to interpret and evaluate the clusters by providing additional explanations of used features. XAI methods were firstly designed to interpret complex supervised methods by focusing on feature importance in black-box models [22]. Several XAI methods were proposed in the literature [22,23,24] which can be classified intomodel-specific andmodel-agnostic. Model-agnostic methods can be applied to any machine-learning model. Typically, model agnostic models require labels on data records to achieve interpretability. Model-specific methods are restricted to a particular model. Small decision trees are one such example of model specific interpretable model, as the splitting criteria used to explain decision trees are restricted to decision tree algorithms. All of these methods provide eitherLocal orglobal explanations, or both at once. Local explainability can be used to explain why a specific data point belongs to a given class or how to change the label of a data point by changing its feature values. Global explainability can be defined as generating explanations on why a set of data points belongs to a specific class, the important features that decide the similarities between points within a class and the feature value differences between different classes.
    In the same way as for supervised models, explainability was also studied for unsupervised models, especially clustering methods. Existing explainable clustering methods try to generate explanations in terms of the underlying features used in the black-box clustering process. Some existing explainable methods [25,26,27] were proposed, to visualize the resulting clusters across two or three Principal Component Analysis (PCA) axes. The main limitation of these methods is the dimensionality reduction, which does not show relationships between the clusters and the original features. Furthermore, cluster interpretations become difficult since PCA axes cannot be readily interpreted in the case of input features. Other existing explainable clustering methods propose to generate cluster representatives to facilitate the understanding of clusters by either computing cluster-centroids or by choosing a small subset of data objects [28,29,30]. These methods may make it easier to analyze each cluster through a simple interpretation of the cluster representatives, but they are particularly sensitive to the geometry of clusters and assume a strict feature similarity between data objects within a cluster, which is not the case in real life-applications.
    Recently, a two-step process was proposed for explainable clustering [31,32,33] to explain the clusters using recent XAI agnostic-method. The first step is devoted to the label assignments (i.e., clusters), then in the second step, these labels are used as target variables in a classification task. Explanations are then built based on the obtained supervised learning model. For example, researchers in [33] have used existing supervised XAI methods for interpreting clustering approaches (EXPLAIN-IT). First, they cluster the input data using existing clustering methods such as K-Means or DBSCAN. A classifier is then trained on input data using the generated cluster labels as class labels for the classifier. Finally, the classifier is explained using existing model agnostic methods such as LIME [34]. LIME is one of the most popular interpretability methods, which can generate interpretations for a single prediction produced by any classifier. Although LIME is powerful and straightforward, it was designed for mainly providing local explanations. So, we propose in this paper to use the SHAP method to improve the explanations of the ABC item classification. SHAP can provide local and global explanations at the same time, and it has a solid theoretical foundation compared to other XAI methods [35].

    2.2. SHAP (Shapley Additive Explanations)

    SHAP is a model-agnostic XAI method, used to interpret predictions of machine learning models [36]. It is based on ideas from game theory and provides explanations by detecting how much each feature contributes to the accuracy of the predictions. SHAP also provides the most important features and their impact on model prediction. It uses the Shapley values to measure each feature’s impact on the machine learning prediction model. Shapley values are defined as the (weighted) average of marginal contributions. It is characterized by the impact of feature value on the prediction across all possible feature coalitions. The Shapley value for an instancex is defined as follows:
    ϕjr(x)=S{j1jm}{jr}ηm!×ωjr
    with
    ωjr=[φSjr(x)φS(x)]
    η=(mS1)!×S!
    whereϕjr(x) denotes the Shapley value for the feature value having the indexjr[1..m],S is a subset of the features used in the prediction model,|S| is the cardinality ofS,m denotes the number of features,φSjr(x) andφS(x) are the prediction function for the set of feature values inS with and without including the featurejr, respectively. The Shapley valueϕjr(x) measures how much the featurejr contributes to the prediction model, either positively or negatively. For that reason, the model is trained with and without including this feature and then predictions from the two models are compared for all feature subsetS{j1,...,jm}{jr}. Whenϕjr(x) is a large positive value, it means that featurejr has a large positive impact on the prediction model. However, when this value is negative, it means that featurejr has a large negative impact on the prediction model.

    3. Proposed Explainable Clustering Method for Multi-Criteria ABC Inventory Classification

    In order to deal with the complexity of explaining ABC inventory classes, we propose a new explainable clustering approach based on both k-means based clustering and XAI capabilities. The proposed approach, referred to as Explainable k-means (Ex-k-means), includes two main phases: item classification and explanation generation. In the first phase, a k-means-based clustering is applied to guide the item classification process to build three classes having varying sizes respecting the ABC distribution of the items. A second phase, the ABC-inventory-interpretation phase, is devoted to improving the transparency of the obtained ABC classes based on the XAI SHAP method capabilities. Because the transparency and the easy-interpretation of inventory classes are almost as important as item classification accuracy, SHAP-based process is designed to have as outputs detailed explanations of the global structure of inventory classes as well as local explanations of the assignment of each item to any specific ABC class.Figure 2 gives an overview of the two main phases of the proposed explainable clustering approach as well as the inputs and outputs of the proposed approach.

    3.1. Phase 1: Item Classification

    This phase uses, as inputs, the numerical description of the set of items and returns, as outputs, the built ABC inventory classes by using a k-means clustering process. This process aims to build three ABC inventory classes based on an alternating iterative process.
    Given a datasetX =x1xn containingn items, described bym criteria, the aim of Ex-k-means is to findk=3 clusters{U1A,U2B,U3C} by minimizing the overall within-cluster-sum-of-squares, denoted (J), as follows:
    J=i=1nt=1kuitd(xi,rt),
    whereuit takes value 1 if itemxi is assigned to clustert or 0 otherwise,rt{rA,rB,rC} is the representative of clustert andd(xi,rt) is the Euclidean distance betweenxi and cluster representativert. The cluster Representativert is calculated for each clustert by:
    rt=i=1rtxirt,
    wherert the number of items assigned to clustert.
    The minimization of the overall within-cluster-sum-of-squares (J) is performed by using an alternating optimization of two independent sub-steps: item assignment and update of the set of clusters representatives{rA,rB,rC}. The first sub-step assigns each item to the nearest cluster representative. After assigning all items, the next sub-step updates the new cluster representative of each of the three ABC clusters. These two sub-steps are repeated until convergence is reached. The convergence is characterized by a maximal number of iterations or no improvement in the objective criterion between two repeated iterations.

    3.2. Phase 2: ABC-Inventory-Interpretation Phase

    The goal of this phase is to generate the explanations for the ABC classes which are obtained in the previous phase. These explanations help decision makers to understand and interpret the final ABC item classification in terms of local and global built structures. Local explanations have the objective to explain the reasons behind the assignment of each item to any ABC class in terms of feature values while global explanations try to explain important feature values for each ABC class. The proposed explanation phase is based on the SHAP method. First, the resulting ABC clusters are configured as the target supervised variables of the explainable process. Then, local and global explanations are built by using the SHAP functionalities.
    For local explanations of the items, we compute the Shapley value at the item level for each criterion with respect to the assigned ABC class. These Shapley values objectively quantify the contribution of each criterion when deciding to assign any items to the ABC classes. These local explanations allow decision makers to understand the reasons behind assigning an item to the A, B or C class. The decision-makers can also evaluate the impacts of increasing or decreasing the criterion value of each item on the predicted class label.
    For global explanations, we provide detailed explanations regarding the importance and contributions of the features when building each ABC class. Local Shapley values of data items are summarized and used as “atomic units” for building the global explanations as follows:
    Gj=1ni=1nϕj(xi)
    whereϕj(xi) denotes the Shapley value of criterionj for itemxi,Gj refers to the overall Shapley value of criterionj andn is the total number of items in the dataset. Global Shapley values are then sorted in decreasing order to show the most important feature when building ABC classes.
    We give, in the following, an illustrative example of local and global explanations. Let us consider a small dataset including 7 items that are described by 4 criteriaj1,j2,j3, andj4, we want to interpret the multi-criteria ABC classification of these items in terms of the local and global explanations.Figure 3 illustrates the outputs of the proposed explanation process based on the SHAP method. First, the Shapley value of each data item and each criterion is calculated with respect to the class label using Equation (1). Then, an average of Shapley values from each criterion are summarized using Equation (6).
    For local explanations, criteria with high Shapley values are interpreted as having high impact on the decision of item assignments, whereas criteria with low Shapley values have lower impact. In fact, it is important to note that criteria with high Shapley values are interpreted as pushing toward one class and low Shapley values as pushing towards the other classes. For example, for itemx4, we can see that the criteria with the highest Shapley values arej1 andj4 while for the itemx2, criteria with the highest Shapley values arej2 andj3. For global explanations, the criteria with the highest values are interpreted as largely contributing to the decision of class labels of the items. For example, we can observe that the first criterion (j2) has the largest contribution in the decision of building the overall ABC classes while the third criterion (j3) has the lowest contribution.

    4. Experiments and Results

    We use a dataset from a company specializing in retail sales. An extract of the first tenth raw data is reported inFigure 4. The dataset contains 301 inventory items described by 10 numerical criteria as follows:Sales,Profit,Lead time,Synergy,Customer priority,Customer sensitivity,Substitution,Expiry or obsolescence risk,Competition for the supplier,Dangerous good classification.Figure 5 reports histograms of data values and their respective frequencies for each criterion. A multi-criteria ABC inventory classification seems mandatory for the company to help decision-makers automatically define different managerial levels for the set of items.
    We begin by studying the data variability for all the item criteria.Table 1 shows minimum, 1st quartile, median, mean, 3rd quartile and maximum values for each criterion. This table shows good data centrality around the median and mean except for those obtained for Profit, Sales and Lead Time. These three criteria also have large scales compared to the other criteria, which require normalized data before building the ABC clusters. A min-max normalization[0...1] was performed using the following formula:
    xˇij=xijxjminxjmaxxjmin
    wherexˇij is the normalized value of the data itemxi for criterionj,xjmin is the minimum value of criterionj andxjmax is the maximum value of criterionj. Once data are normalized, we used the proposed Ex-k-means algorithm, to build the different ABC classes.

    4.1. Evaluation of the Clustering Performance

    We evaluate the performance of the proposed explainable clustering method compared to existing ones AHP-k-means [37], AHP-k-means-Veto [37], AHP-FCM [5] and AHP-FCM-Rveto [5].Table 2 shows the cluster-sizes obtained by each method by fixing the number of clustersk=3 for all methods. We assigned the A, B and C classes respecting the sizes of the built clusters from lowest to highest, respectively, (class A for the cluster with the minimal size).Table 2 shows that AHP–FCM, AHP–FCM–Rveto and Ex-k-means methods gives a good distribution of the items over the three classes. We show large differences in the sizes of the obtained ABC classes. AHP-k-means and AHP-k-means-Veto build a cluster (Class A items) with very small size. The AHP-FCM-Rveto method build the best distributed clusters, based on the ABC principles, given its integrated Rveto phase.
    Next, we evaluate the quality of the obtained ABC clusters by using the internal validation measures, Silhouette Coefficient (SC) [38], Davies–Bouldin Index (DBI) [39], and Calinski–Harabasz Index (CHI) [40]. These validation measures aim to measure the compactness and separation of the clusters. Compactness measures how closely related the items in a cluster are while separation measures how distinct or well-separated a cluster is from other clusters. Higher values of the SC and DBI measures indicate a good quality partition, while lower values of the CHI measure indicate the best partition. The results are reported inTable 3. This table shows that the proposed Ex-k-means method outperforms the compared methods in terms of DBI and CHI measures.

    4.2. Local Explanations of Item Assignment to the ABC Classes

    The local explanation has the objective to build an easy item-interpretable model explaining the decision behind the assignment of each item to any ABC class. We used the force plot which schematizes how features contributed to the cluster assignment of each item separately. This plot provides a deep understanding and easy interpretation of the classification of any particular item. We randomly selected a set of 3 items from each class and then the force plot of Shapley values of each item is reported.Figure 6,Figure 7 andFigure 8 show the contribution of each criterion, at the item level, when building the ABC inventory classes. The bold numeric values on the right side of these figures evaluate the probability of the decision of assigning one item to its current class rather than any other class. We obtain a high probability for the selected items from all classes (higher than97%) except for the probability reported for the third item assigned to class A (Figure 6c) which reports a probability of85%. High probability indicates the facility of identifying the respective class of the item by using the proposed classification model. However, any item with a low probability value, or a probability converging to the base value, indicates the difficulty of determining its correct class given its dissimilarity with all built classes. The base value is the average value that would be predicted if we did not know any criteria for the current item. The base value is the average of the model labels in the training dataset. We show inFigure 6,Figure 7 andFigure 8 that base values correspond to the considered cluster-sizes (20%,30% and50%) for the A, B and C classes, respectively.
    The reported numbers below the plot arrow represent the value of each criterion for that item. Two colors are then used to schematize criteria value impact: red for criteria values that pushed the model classification probability higher and blue for criteria values that pushed the model classification probability lower. The bigger the arrow size, the bigger the impact of the criterion on the model classification.Figure 6 shows that the decision of assigning the selected set of items to class A was based on a very high value ofsales (average ≃1,100,000 for the three selected items), highProfit (average ≃92,000 for the three selected items) and highCustomer priority (≃10). All these criteria values have positively contributed (red color) to increase the probability of assigning the respective items to class A.Figure 6a,c show that the decision of assigning the respective items to class A has been negatively (blue color) affected by the low value ofprofit (≃139) for the first item and the high value ofsubstitution (≃10) for the second item. This fact explains the low reported probabilities inFigure 6a,c compared to the reported probability inFigure 6b.
    Concerning the items of class B,Figure 7 shows that the decision of assigning these items was based on moderate volumes ofSales (Average ≃77,000 for the three selected items), high profitProfit (average ≃7000 for the three selected items), high values of bothSynergy (≃6) andCompetitionForTheSupplier (≃6). These criteria values increase the probability to judge the item as belonging to class B. All these criteria values lie between the average and the third quartile values. However,Figure 7a,b show that a low value ofExpiry Obsolescence Risk (≃2) and a high value ofCompetition For The Supplier (≃9) slightly decrease (blue color) the probability of assigning these two items to class B.
    Items of the third class, class C, are reported inFigure 8. This figure shows that criteria values which mostly contributed to the assignment of these items are very lowSales Volume (average ≃5700 for the three selected items), very lowProfit (average ≃500 for the three selected items) and moderate values of bothExpiry obsolescence risk (≃5) andCompetition For the Supplier (≃5). All these low criteria values increase the probability of assigning the items to class C, while it is not the case for the high value ofCustomer priority (≃10) (reported inFigure 8a) that highly reduces the probability of assigning the respective item to class C.
    These locally built explanations can be very useful to decision-makers for a deep understanding of the assignment of each item separately. In some cases, further analysis would be required to decide the class of some items, especially those having low classification probability. The local explanations increase the classification model transparency and improve the quality of determining the right managerial classes for the items.

    4.3. Global Explanations of ABC Classes

    Global explanations have the objective of easily interpreting the ABC classification model at the level of classes rather than the local item level. For each ABC class, the criteria that mostly contributed to the decision of assigning items to a specific class are stacked and reported. We used the feature-importance plot, in which items Shapley values are summarized for each class separately.Figure 9 reports the most important criteria after summarizing shape values for items assigned to A, B and C classes separately. These criteria are schematized according to their contribution to the classification model from highest (at the top of the plot) to lowest (at the bottom of the plot). The contribution of each criterion within each class is reported with a different color: navy-green for class A, pink for class B and blue for class C.Figure 9 shows thatSales,Profit,Customer priority, andCompetion for the supplier criteria have the largest impact on the classification model compared to the other criteria. In addition, we can see that the criterionCustomer priority largely contributes to determining class B items compared to the other classes. We can also see thatExpiry or obsolescence risk andSynergy criteria have the same low contribution for all classes. These findings are broadly in accordance with the literature on inventory classification.
    The feature-importance plot only shows the importance of each criterion when building the ABC classes, without explaining its impact (positively or negatively) on the classification model. In order to show criteria impacts for each class, we build the feature-summary-plot that combines feature importance and impact.Figure 10 shows the positive and negative relationships of criteria values with the target class. Every single point in these sub-figures is a Shapley value of one criterion for a specific item. The position on the y-axis is determined by the criterion, and the x-axis is determined by the Shapley value. The color represents the criterion value from low (blue) to high (red). First, this figure confirms the finding thatSales andProfit criteria have the highest importance when deciding the inventory class of any item and for the three classes A, B, and C. On the other hand,Lead Time,Expiry obsolescence risk andSynergy have the lowest importance which tends to be negligible. Besides,Figure 10a shows that the larger the values ofSales andProfit (red points), the larger the Shapley value is (points are located at the right of the x-axis). This shows the positive impact of high values ofSales, andProfit on the decision of assigning the items to class A. In the opposite way,Figure 10c shows the negative impacts of the same criteria on the decision of assigning one item to class C. Items with high Shapley values in terms ofSales andProfit, which are located at the right of the x-axis, are characterized by low criteria values (blue points) in terms ofSales andProfit. Concerning the decision of assigning items to class B,Figure 10b shows that high values of criteriaSales andProfit were also crucial in assigning the items to this class, but also items of this class are characterized by highCustomer priority and highCompetition for the supplier compared to the other classes.
    Another important finding that can be interpreted fromFigure 10a,c is that high values ofSales andProfit (relatively low) cannot be by itself good criteria to decide whether an item belongs to A or C classes. For example, we show inFigure 10a some items with low values ofSales andProfit (blue points), but having high values of bothSubstitution andLead time. This finding demonstrates the multi-criteria nature of the item classification task.

    5. Conclusions

    We proposed a multi-criteria ABC classification approach that deals with the issue of the non-explainability of ABC inventory classes. The proposed approach is based on alternating clustering and explainable artificial intelligence capabilities to generate micro and macro explanations. It includes two independent phases: item classification and model interpretation. The first phase aims to build ABC item classes based on a rapid kmeans-based clustering while the second phase is devoted to the generation of local and global explanations. The generated explanations allow decision-makers to better interpret the built ABC classes at both the item and the class levels.
    The application of explainable artificial intelligence to the classification of products of a retail company has shown high transparency in explaining the unsupervised ABC classes. The generated global explanation has shown a large impact of the criteriaSales,Profit, andCustomer priority, compared to the other criteria, in determining the ABC inventory class of the items. Concerning local explanations, the results showed that assigning any items to class A was mainly based on high values ofSales,Profit, andCustomer priority. On the other hand, the decision to assign items to class C was mainly based on low values ofSales,Profit, and also smaller values ofExpiry obsolescence risk andCompetition For the Supplier. Local explanations have also shown that items having high values ofCustomer priority have a large probability to be assigned to the A or B classes rather than the C class. These findings are relative to the data of the studied company and cannot be generalized to any company type (industrial, commercial, retail, ..., etc.) or any business field (chemicals, technology, foods, ..., etc). However, the proposed model is generic and can easily be applied to explain the ABC inventory classification of any company type. It would be interesting to evaluate this proposed model in explaining ABC classes of other company types and business domains.
    In this work, we only considered the explainability of ABC classes based on numerical item criteria. The proposed approach can be improved by exploring the possibility of integrating other data types such as categorical and symbolic criteria, unless the data gets transformed to numerical type. This direction may require adopting other explainable artificial intelligence models designed to deal with such diverse types of data. Besides, equal feature weights were considered to build the ABC classes. A more advanced ABC classification models may assign different weights for the input criteria. In that case, it would be interesting to investigate the adjustment of the XAI-SHAP model to take into account differences in initial feature weights.

    Author Contributions

    Conceptualization, A.A.Q.; Methodology, C.-E.B.N.; Software, M.-A.B.H.; Validation, O.N.; Formal analysis, A.A.Q. and C.-E.B.N.; Resources, A.A.Q.; Writing—original draft, A.A.Q. and M.-A.B.H.; Writing—review & editing, C.-E.B.N. and O.N.; Supervision, O.N.; Project administration, C.-E.B.N. All authors have read and agreed to the published version of the manuscript.

    Funding

    This work was funded by the university of Jeddah, Jeddah, Saudi Arabia, under grant No (UJ—21—ICL—3). The authors, therefore, acknowledge with thanks the University of Jeddah technical and financial support. Olfa Nasraoui was partially supported by NSF-EPSCoR-RII Track-1: Kentucky Advanced Manufacturing Partnership for Enhanced Robotics and Structures (Award IIP#1849213) and by NSF DRL-2026584.

    Institutional Review Board Statement

    Not applicable.

    Informed Consent Statement

    Not applicable.

    Data Availability Statement

    Not applicable.

    Conflicts of Interest

    The authors declare no conflict of interest.

    References

    1. Keskin, G.; Ozkan, C. Multiple criteria ABC analysis with FCM clustering.J. Ind. Eng.2013,2013, 827274. [Google Scholar] [CrossRef] [Green Version]
    2. Park, J.; Bae, H.; Bae, J. Cross-Evaluation-Based Weighted Linear Optimization for Multi-Criteria ABC Inventory Classification.Comput. Ind. Eng.2014,76, 40–48. [Google Scholar] [CrossRef]
    3. Ergün, E.; Ic, Y. An improved decision support system for ABC inventory classification.Evol. Syst.2020,11, 683–696. [Google Scholar] [CrossRef]
    4. Dickie, H.F. ABC Inventory Analysis Shoots for Dollars Not Pennies.Fact. Manag. Maint.1951,109, 92–94. [Google Scholar]
    5. Yiğit, F.; Esnaf, Ş. A new Fuzzy C-Means and AHP-based three-phased approach for multiple criteria ABC inventory classification.J. Intell. Manuf.2021,32, 1517–1528. [Google Scholar] [CrossRef]
    6. Chen, Y.; Li, K.W.; Marc Kilgour, D.; Hipel, K.W. A case-based distance model for multiple criteria ABC analysis.Comput. Oper. Res.2008,35, 776–796. [Google Scholar] [CrossRef]
    7. Saaty, T.The Analytic Hierarchy Process: Planning, Priority Setting, Resource Allocation; Advanced Book Program; McGraw-Hill International Book Company: New York, NY, USA, 1980. [Google Scholar]
    8. Xu, R.; Zhai, X. Fuzzy logarithmic least squares ranking method in analytic hierarchy process.Fuzzy Sets Syst.1996,77, 175–190. [Google Scholar] [CrossRef]
    9. Meade, L.; Sarkis, J. Analyzing organizational project alternatives for agile manufacturing processes: An analytical network approach.Int. J. Prod. Res.1999,37, 241–261. [Google Scholar] [CrossRef]
    10. Liu, Q.; Huang, D. Classifying ABC Inventory with Multicriteria Using a Data Envelopment Analysis Approach. In Proceedings of the Sixth International Conference on Intelligent Systems Design and Applications, Jinan, China, 16–18 October 2006; Volume 1, pp. 1185–1190. [Google Scholar] [CrossRef]
    11. Onwubolu, G.; Dube, B. Implementing an improved inventory control system in a small company: A case study.Prod. Plan. Control2006,17, 67–76. [Google Scholar] [CrossRef]
    12. Zheng, S.; Fu, Y.; Lai, K.K.; Liang, L. An improvement to multiple criteria ABC inventory classification using Shannon entropy.J. Syst. Sci. Complex.2017,30, 857–865. [Google Scholar] [CrossRef]
    13. Wu, S.; Fu, Y.; Lai, K.K.; Leung, J. A Weighted Least-Square Dissimilarity Approach for Multiple Criteria ABC Inventory Classification.Asia-Pac. J. Oper. Res.2018,35, 1850025. [Google Scholar] [CrossRef]
    14. Ramanathan, R. ABC inventory classification with multiple-criteria using weighted linear optimization.Comput. Oper. Res.2006,33, 695–700. [Google Scholar] [CrossRef]
    15. Ng, W.L. A simple classifier for multiple criteria ABC analysis.Eur. J. Oper. Res.2007,177, 344–353. [Google Scholar] [CrossRef]
    16. Hadi-Vencheh, A. An improvement to multiple criteria ABC inventory classification.Eur. J. Oper. Res.2010,201, 962–965. [Google Scholar] [CrossRef]
    17. Karagiannis, G. Partial average cross-weight evaluation for ABC inventory classification.Int. Trans. Oper. Res.2021,28, 1526–1549. [Google Scholar] [CrossRef]
    18. Chu, C.W.; Liang, G.S.; Liao, C.T. Controlling inventory by combining ABC analysis and fuzzy classification.Comput. Ind. Eng.2008,55, 841–851. [Google Scholar] [CrossRef]
    19. Çebi, F.; Kahraman, C.; Bolat, B. A multiattribute ABC classification model using fuzzy AHP. In Proceedings of the 40th International Conference on Computers and Indutrial Engineering, Awaji City, Japan, 25–28 July 2010; pp. 1–6. [Google Scholar] [CrossRef]
    20. Partovi, F.Y.; Anandarajan, M. Classifying inventory using an artificial neural network approach.Comput. Ind. Eng.2002,41, 389–404. [Google Scholar] [CrossRef]
    21. Yu, M.C. Multi-criteria ABC analysis using artificial-intelligence-based classification techniques.Expert Syst. Appl.2011,38, 3416–3421. [Google Scholar] [CrossRef]
    22. Arrieta, A.B.; Díaz-Rodríguez, N.; Del Ser, J.; Bennetot, A.; Tabik, S.; Barbado, A.; García, S.; Gil-López, S.; Molina, D.; Benjamins, R.; et al. Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI.Inf. Fusion2020,58, 82–115. [Google Scholar] [CrossRef] [Green Version]
    23. Das, A.; Rad, P. Opportunities and challenges in explainable artificial intelligence (xai): A survey.arXiv2020, arXiv:2006.11371. [Google Scholar]
    24. Linardatos, P.; Papastefanopoulos, V.; Kotsiantis, S. Explainable ai: A review of machine learning interpretability methods.Entropy2020,23, 18. [Google Scholar] [CrossRef] [PubMed]
    25. Arias-Castro, E.; Lerman, G.; Zhang, T. Spectral clustering based on local PCA.J. Mach. Learn. Res.2017,18, 253–309. [Google Scholar]
    26. Ding, C.; He, X. K-means clustering via principal component analysis. In Proceedings of the Twenty-First International Conference on Machine Learning, Banff, AB, Canada, 4–8 July 2004; p. 29. [Google Scholar]
    27. Jafarzadegan, M.; Safi-Esfahani, F.; Beheshti, Z. Combining hierarchical clustering approaches using the PCA method.Expert Syst. Appl.2019,137, 1–10. [Google Scholar] [CrossRef]
    28. Kacem, M.A.B.H.; N’cir, C.E.B.; Essoussi, N. MapReduce-based k-prototypes clustering method for big data. In Proceedings of the 2015 IEEE International Conference on Data Science and Advanced Analytics (DSAA), Paris, France, 19–21 October 2015; pp. 1–7. [Google Scholar]
    29. Mahmud, M.S.; Rahman, M.M.; Akhtar, M.N. Improvement of K-means clustering algorithm with better initial centroids based on weighted average. In Proceedings of the 2012 7th International Conference on Electrical and Computer Engineering, Dhaka, Bangladesh, 20–22 December 2012; pp. 647–650. [Google Scholar]
    30. Yedla, M.; Pathakota, S.R.; Srinivasa, T. Enhancing K-means clustering algorithm with improved initial center.Int. J. Comput. Sci. Inf. Technol.2010,1, 121–125. [Google Scholar]
    31. Bandyapadhyay, S.; Fomin, F.; Golovach, P.A.; Lochet, W.; Purohit, N.; Simonov, K. How to Find a Good Explanation for Clustering? In Proceedings of the AAAI-2022, Virtually, 22 February–1 March 2022; Volume 36. [Google Scholar]
    32. Dasgupta, S.; Nave Frost, M.M.; Rashtchian, C. Explainable k-Means and k-Medians Clustering. In Proceedings of the 37 th International Conference on Machine Learning, Virtually, 13–18 July 2020. [Google Scholar]
    33. Morichetta, A.; Casas, P.; Mellia, M. EXPLAIN-IT: Towards explainable AI for unsupervised network traffic analysis. In Proceedings of the 3rd ACM CoNEXT Workshop on Big Data, Machine Learning and Artificial Intelligence for Data Communication Networks, Orlando, FL, USA, 9 December 2019; pp. 22–28. [Google Scholar]
    34. Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why should i trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1135–1144. [Google Scholar]
    35. Wang, M.; Zheng, K.; Yang, Y.; Wang, X. An explainable machine learning framework for intrusion detection systems.IEEE Access2020,8, 73127–73141. [Google Scholar] [CrossRef]
    36. Lundberg, S.M.; Lee, S.I. A unified approach to interpreting model predictions.Adv. Neural Inf. Process. Syst.2017,30. [Google Scholar]
    37. Lolli, F.; Ishizaka, A.; Gamberini, R. New AHP-based approaches for multi-criteria inventory classification.Int. J. Prod. Econ.2014,156, 62–74. [Google Scholar] [CrossRef] [Green Version]
    38. Rousseeuw, P.J. Silhouettes: A graphical aid to the interpretation and validation of cluster analysis.J. Comput. Appl. Math.1987,20, 53–65. [Google Scholar] [CrossRef] [Green Version]
    39. Davies, D.L.; Bouldin, D.W. A cluster separation measure.IEEE Trans. Pattern Anal. Mach. Intell.1979, 224–227. [Google Scholar] [CrossRef]
    40. Calinski, T. A dendrite method for cluster analysis.Commun. Stat.1974,3, 1–27. [Google Scholar]
    Jtaer 18 00044 g001 550
    Figure 1. Inputs and outputs of the multi-criteria ABC inventory classification process.
    Figure 1. Inputs and outputs of the multi-criteria ABC inventory classification process.
    Jtaer 18 00044 g001
    Jtaer 18 00044 g002 550
    Figure 2. The main phases of the proposed explainable clustering approach for multi-criteria ABC items classification. Phase 1: item classification and phase 2: ABC item interpretations.
    Figure 2. The main phases of the proposed explainable clustering approach for multi-criteria ABC items classification. Phase 1: item classification and phase 2: ABC item interpretations.
    Jtaer 18 00044 g002
    Jtaer 18 00044 g003 550
    Figure 3. An illustrative example of the SHAP method on a set of 7 items described by 4 criteria.
    Figure 3. An illustrative example of the SHAP method on a set of 7 items described by 4 criteria.
    Jtaer 18 00044 g003
    Jtaer 18 00044 g004 550
    Figure 4. Extract of the first tenth raw data of the dataset.
    Figure 4. Extract of the first tenth raw data of the dataset.
    Jtaer 18 00044 g004
    Jtaer 18 00044 g005a 550Jtaer 18 00044 g005b 550
    Figure 5. Histograms schematizing intervals and frequencies of item numerical criteria.
    Figure 5. Histograms schematizing intervals and frequencies of item numerical criteria.
    Jtaer 18 00044 g005aJtaer 18 00044 g005b
    Jtaer 18 00044 g006 550
    Figure 6. Local explanations of three items assigned to class A: (a) force plot of Shapley values of item 1 (b) force plot of Shapley values of item 2 (c) force plot of Shapley values of item 3.
    Figure 6. Local explanations of three items assigned to class A: (a) force plot of Shapley values of item 1 (b) force plot of Shapley values of item 2 (c) force plot of Shapley values of item 3.
    Jtaer 18 00044 g006
    Jtaer 18 00044 g007 550
    Figure 7. Local explanations of three items assigned to class B: (a) force plot of Shapley values of item 1 (b) force plot of Shapley values of item 2 (c) force plot of Shapley values of item 3.
    Figure 7. Local explanations of three items assigned to class B: (a) force plot of Shapley values of item 1 (b) force plot of Shapley values of item 2 (c) force plot of Shapley values of item 3.
    Jtaer 18 00044 g007
    Jtaer 18 00044 g008 550
    Figure 8. Local explanations of three items assigned to class C: (a) force plot of Shapley values of item 1 (b) force plot of Shapley values of item 2 (c) force plot of Shapley values of item 3.
    Figure 8. Local explanations of three items assigned to class C: (a) force plot of Shapley values of item 1 (b) force plot of Shapley values of item 2 (c) force plot of Shapley values of item 3.
    Jtaer 18 00044 g008
    Jtaer 18 00044 g009 550
    Figure 9. Average criteria importance of the proposed inventory classification method.
    Figure 9. Average criteria importance of the proposed inventory classification method.
    Jtaer 18 00044 g009
    Jtaer 18 00044 g010 550
    Figure 10. Relationships of the criteria values with the target class.
    Figure 10. Relationships of the criteria values with the target class.
    Jtaer 18 00044 g010
    Table
    Table 1. Statistics of the item descriptive criteria: minimum, first quartile, median, mean, third quartile and maximum values for each criterion.
    Table 1. Statistics of the item descriptive criteria: minimum, first quartile, median, mean, third quartile and maximum values for each criterion.
    Min1st Qu.MedianMean3rd Qu.Max
    Profit1150197529,91013,4601,582,146
    Sales90660028,213173,467103,4016,790,487
    Lead Time3202046.413060
    Synergy1254.85710
    Customer Priority1556.481010
    Customers Sensitivity1355.21810
    Substitution1355.31810
    Expiry or Obsolescence Risk1354.96710
    Competition for the Supplier1154.779
    Dangerous Good Classification1355.12710
    Table
    Table 2. Cluster-sizes build by the proposed Ex-k-means compared to those obtained by AHP-k-means, AHP-k-means-Veto, AHP-FCM and AHP-FCM-Rveto. We considered for all methods that the smallest cluster is A, the biggest one is C and the last one is B.
    Table 2. Cluster-sizes build by the proposed Ex-k-means compared to those obtained by AHP-k-means, AHP-k-means-Veto, AHP-FCM and AHP-FCM-Rveto. We considered for all methods that the smallest cluster is A, the biggest one is C and the last one is B.
    MethodABC
    AHP-k-means1%16%83%
    AHP-k-means-Veto1%10%89%
    AHP-FCM9%15%76%
    AHP-FCM-Rveto21%30%49%
    Ex-K-means8%12%80%
    Table
    Table 3. Comparison of the clustering performance of Ex K-means with existing methods using three internal validation measures (Silhouette Coefficient (SC), Davies–Bouldin Index (DBI), and Calinski–Harabasz Index (CHI)).
    Table 3. Comparison of the clustering performance of Ex K-means with existing methods using three internal validation measures (Silhouette Coefficient (SC), Davies–Bouldin Index (DBI), and Calinski–Harabasz Index (CHI)).
    MethodSCDBICHI
    AHP-k-means0.87630.270.51
    AHP-k-means-Veto0.0312.802.40
    AHP-FCM0.59434.430.66
    Ex-k-means0.861203.030.28
    Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

    © 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).

    Share and Cite

    MDPI and ACS Style

    Qaffas, A.A.; Ben HajKacem, M.-A.; Ben Ncir, C.-E.; Nasraoui, O. An Explainable Artificial Intelligence Approach for Multi-Criteria ABC Item Classification.J. Theor. Appl. Electron. Commer. Res.2023,18, 848-866. https://doi.org/10.3390/jtaer18020044

    AMA Style

    Qaffas AA, Ben HajKacem M-A, Ben Ncir C-E, Nasraoui O. An Explainable Artificial Intelligence Approach for Multi-Criteria ABC Item Classification.Journal of Theoretical and Applied Electronic Commerce Research. 2023; 18(2):848-866. https://doi.org/10.3390/jtaer18020044

    Chicago/Turabian Style

    Qaffas, Alaa Asim, Mohamed-Aymen Ben HajKacem, Chiheb-Eddine Ben Ncir, and Olfa Nasraoui. 2023. "An Explainable Artificial Intelligence Approach for Multi-Criteria ABC Item Classification"Journal of Theoretical and Applied Electronic Commerce Research 18, no. 2: 848-866. https://doi.org/10.3390/jtaer18020044

    APA Style

    Qaffas, A. A., Ben HajKacem, M.-A., Ben Ncir, C.-E., & Nasraoui, O. (2023). An Explainable Artificial Intelligence Approach for Multi-Criteria ABC Item Classification.Journal of Theoretical and Applied Electronic Commerce Research,18(2), 848-866. https://doi.org/10.3390/jtaer18020044

    Article Metrics

    No
    No

    Article Access Statistics

    For more information on the journal statistics, clickhere.
    Multiple requests from the same IP address are counted as one view.
    J. Theor. Appl. Electron. Commer. Res., EISSN 0718-1876, Published by MDPI
    RSSContent Alert

    Further Information

    Article Processing Charges Pay an Invoice Open Access Policy Contact MDPI Jobs at MDPI

    Guidelines

    For Authors For Reviewers For Editors For Librarians For Publishers For Societies For Conference Organizers

    MDPI Initiatives

    Sciforum MDPI Books Preprints.org Scilit SciProfiles Encyclopedia JAMS Proceedings Series

    Follow MDPI

    LinkedIn Facebook X
    MDPI

    Subscribe to receive issue release notifications and newsletters from MDPI journals

    © 1996-2025 MDPI (Basel, Switzerland) unless otherwise stated
    Terms and Conditions Privacy Policy
    We use cookies on our website to ensure you get the best experience.
    Read more about our cookieshere.
    Accept
    Back to TopTop
    [8]ページ先頭

    ©2009-2025 Movatter.jp