Skip to main content
Erschienen in: Production Engineering 1/2024

Open Access 01.08.2023 | Computer Assisted Approaches in Production Engineering

Supervised learning to support the process planning of contract logistics projects

verfasst von: Marius Veigt, Michael Freitag

Erschienen in: Production Engineering | Ausgabe 1/2024

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Due to the outsourcing trend, contract logistics is a constantly growing industry. Especially for the essential and time-consuming planning of logistics processes in a contract logistics project, experienced planners are required. However, the growing shortage of skilled workers makes recruiting these planners increasingly difficult. Hence, a supervised learning approach will be investigated to support especially inexperienced planners in process planning. This article explores how supervised learning can extract the process knowledge contained in legacy contract logistics project documentation to suggest process steps during a new project process planning. The investigation results in boosted decision trees predicting the next process step correctly in 81% of the cases. In addition, the article guides what data should be collected today for even better results in future applications.
Hinweise

Supplementary Information

The online version contains supplementary material available at https://​doi.​org/​10.​1007/​s11740-023-01217-3.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

1 Introduction

For decades, there has been a trend for companies in industry and retail to focus on their core competencies and outsource other tasks—especially logistical task packages [1]. As a result, the logistics market volume, especially contract logistics, is growing almost steadily [2].
Contract logistics is characterized as the bundling of several logistics functions into a service package of increased complexity. Several traditional logistics services, such as transport, handling, and warehousing, are usually combined with logistics-related value-added services, e.g., minor assembly and production activities, packaging, and others. The logistics service provider (LSP) adapts its processes to the customer’s requirements, creating longer-term contracts and business relationships [3].
The reason for outsourcing the service bundles is essentially a cost reduction [4]. When complex logistics services are being tendered, customers ensure that the proposals are comparable in terms of results and price structure [5]. Thus, LSPs must plan the logistics processes precisely to not overprice on the one hand and operate at a loss on the other hand. This would be risky due to the long-term contracts. On this background, a reliable and high-quality proposal calculation is a critical success factor for LSP [6]. It requires the commitment of experienced and highly qualified employees [7]. Due to the emerging shortage of skilled professionals and the fluctuating number of contract logistics tenders, it is unavoidable that less skilled employees have to take over the calculation of the proposal. This increases the time required, reduces the quality of the logistics concept, and thus reduces the chances of winning the contract [8].
This contribution presents results on how supervised learning can support planners of contract logistics offers during the planning process by suggesting appropriate MTM codes for subsequent process steps. The aim is to enable inexperienced planners in particular to benefit from the knowledge gained from previous planning and accelerate the work of all planners.
The research was conducted by using the Cross Industry Standard Process for Data Mining (CRISP-DM) [9]. The CRISP-DM is still the most popular framework for executing data science [10] and a de facto standard for applying data science projects [11]. This contribution summarizes the key findings structured according to the IMRAD standard. Thus, Sects. 1 and 2 present the introduction, the state of the art and identify the current research gap. These topics can be seen as part of the business understanding phase of CRISP-DM. The Sect. 3 describes the data set, the supervised learning approach, and the methods used. Which are the key findings of the data understanding and prepration as well as the modelling phases. The Sect. 4 evaluates and checks the plausibility of the results. The Sect. 5 discusses the results and possible deployment options. Finally, this contribution concludes with a summary and an outlook for further research in Sect. 6.
Machine learning (ML) is already being researched in production engineering, e.g., for self-optimizing process planning in polishing [12], for process planning in metal and planking forming [13], or for wear prediction in sheet metal forming [14].
However, there is a lack of machine learning research in contract logistics, although these technologies have great potential for this sector, e.g., for assisting humans in making decisions [15]. Recent research has analyzed tender management in contract logistics and identified the demand and potential for ML application in this domain. As a result, support during the planning of logistics processes would be a valuable improvement, especially for inexperienced logistics planners [8].
Tender managers often use the Methods-Time Measurement (MTM) for process planning, which enables the precise determination of process times, which is the fundament for an accurate and comprehensible quotation [4]. The MTM method divides processes into basic movements (e.g., grasping, walking) and assigns base times to these movements, which are labeled with an MTM code [16]. However, this method is time-consuming and requires a high degree of empirical knowledge [8].
Table 1 presents an example of a process description using MTM codes for unloading pallets from a truck using a forklift. The knowledge of the logistics planner is required in the process design in two aspects: Which process steps are to be planned, and which MTM code is to be selected for this process step. Supervised learning should take over these two outputs by searching for analogies in past contract logistics projects and making suggestions to the planners for the following process step with the associated MTM code. The planners can accept a suggestion or plan a different process step. The algorithm should learn from such confirmations and adapt the suggestions in the future. After planning the process steps, the process time can be calculated using the base times assigned to the MTM code, and critically checked by the planner [8].
Table 1
Exemplary process description
#
Process step description
MTM code
1
Drive from the goods receiving area to truck
SFISF
2
Pick up pallet
SABAFM
3
Drive from truck to goods receiving area
SFISF
4
Set down pallet
SABAFM
5
Getting off the driver’s seat
AZA
6
Walk to the pallet
KA
7
Visual inspection of the goods for obvious damage
VA
8
Pick up pen
EH3
9
Pick up the delivery paper
IAHO3
10
Note the number of the receipt area on the delivery paper
IAKW
Initial research uses the N-gram method [17] to predict process steps and achieves a probability of up to 90% to find the next MTM code correctly if five predecessors are considered [18]. However, this method predicts a process step only based on its predecessor process steps. Other influencing factors are not considered. Consequently, the disadvantage of this method is that it cannot be used at the beginning of the process, where either no or just a few (e.g. less than five) predecessors are available. Thus, the above mentioned 90% accuracy considered only the part of the data set where at least five predecessors are available.
Classification methods are an alternative approach to predict process steps. The latest publication elaborates on decision factors and compares different classifiers [19]. As a result, decision trees using the C4.5-algorithm [20] obtain an accuracy of up to 79% in predicting process steps [19]. The suitability of tree-based classifiers seems plausible, as these classifiers represent the planners’ decision rules well.
This contribution combines the N-gram approach with the tree-based classification by examining a different number of predecessor process steps as a decision feature for the classifier. Furthermore, this contribution investigates the classification approach in more depth. On the one hand, the hyper-parameters of the classifier are optimized, and the plausibility of the results is checked. On the other hand, ways to improve and to deploy the results are discussed.

3 Material and methods

This section describes the data set and data preparation, the modelling methods, the metrics to assess performance and the approach to check the plausibility of the results.

3.1 Data set description

Planning data of 13 contract logistics projects are available for this research work. The data set consists of six customers from the automotive and construction industries.
The data set includes the description of 13,707 process steps, each identified with an MTM code. The process steps are grouped in 440 processes. For example, one process describes the goods receipt of a pallet, while the storage of this pallet represents a new process. The shortest process includes only one MTM code, and the longest contains 155 MTM codes. The average is about 31 MTM codes per process. The MTM codes can be divided into 256 unique MTM codes. The aim of the supervised learning method is to suggest the most probable next process step with its associated MTM code to the planer. In this way, the planer should be supported during contract logistics process planning.
To determine what information is used to plan contract logistics processes, workshops were conducted with planners. In these workshops, the planners explained which information they use to plan the processes. This information must also be provided to the machine learning classifier so that the classifier can learn the dependencies between this information and the MTM code sequences. This information is referred to as features in the following. Table 2 lists these features and highlights those, which are included in the data set and, thus, are available for supervised learning. For example, information on load carrier dimensions, item specifications, or process flows (e.g., storage method) is missing either due to this information being unknown at the time of planning or because this information is implicitly available to the planners but is not documented. So it could not be captured automatically when creating the data set. The discrepancy of the comparison reveals the missing information and consequently complicates the application of supervised learning in this case.
Table 2
List of features (independent variables)
#
Description
Availability
Amount of different categories
1
Process specification
1.1
 Main process
9
1.2
 Equipment
26
1.3
 Employee group
16
1.4
 IT system
1.5
 Supply means
7
1.6
 Pick-up and storage situation during stacking
1.7
 Storage type
1.8
 Storage levels
1.9
 Storage method
2
 Layout
3
Article specification
3.1
 Article name
203
3.2
 Piece goods/Bulk goods
1
3.3
 Food
1
3.4
 Hazardous material
1
3.5
 Dimensions (length × width × height)
3.6
 Weight
3.7
 Custom Clearance
4
Load carrier specification
4.1
 Type
19
4.2
 Filling
3
4.3
 Dimensions (length × width × height)
4.4
 Weight
4.5
 Stackability
5
 Customer company name
6
6
 Location designation
7
7
 Industry
2
8
Previous MTM code sequence
8.1
 Up to 1 previous MTM code
235
8.2
 Up to 2 previous MTM codes
1434
8.3
 Up to 3 previous MTM codes
2721
8.4
 Up to 4 previous MTM codes
3599
8.5
 Up to 5 previous MTM codes
4189
8.6
 Up to 6 previous MTM codes
4626
8.7
 Up to 7 previous MTM codes
4978
8.8
 Up to 8 previous MTM codes
5263
8.9
 Up to 9 previous MTM codes
5498
The process steps and MTM codes to be planned depend on the process specification (see #1 in Table 2). In the data set, nine categories of main process (1.1) could be distinguished. These main process categories are: Goods receipt, storage, removal from storage, conveying, commissioning, sequencing, pre-assembly, repacking, and dispatching.
In addition, it is relevant which equipment (1.2) is used in the process. For example, different MTM codes are used for a forklift than for a pallet jack. In the data set were 26 different categories of equipment counted. The employee group (e.g., forklift driver, booker, packer, assembly operator) listed as 1.3 specifies what is to be done in a process (e.g., booker books, packer packs). The IT system (1.4) specifies how goods are to be booked. Supply means (1.5) defines the means of delivery by which the goods are received or shipped. For example, the unloading of a truck differs significantly from the delivery of a courier and express provider service provider. Storage and removal from storage (1.6–1.9) depend on the stacking situation, which storage type (e.g., block storage, high-rack storage) is used, whether the goods are stored directly (1-level-storage) or transferred to a stacker crane (2-level-storage), and which storage method (e.g., first in, first out) is used. The features mentioned above are process specific. To avoid missing data, the value “irrelevant” is inserted when a feature is irrelevant for decision-making in a process, e.g., in storage processes, the supply means are irrelevant.
The factory or warehouse layout (2) influences whether and how many drives around curves must be planned, which is to be planned in the MTM code sequence.
The article specifications (3) also influence the MTM code selection. The handling of piece goods and bulk goods (3.2) is different. Particular goods, such as food (3.3) or hazardous material (3.4), have special safety regulations that must be considered. The dimensions (3.5) and weight (3.6) of an item influence the handling during picking and pre-assembly. In addition, customs clearance (3.7) affect the process of goods receipt and goods issue. Since some of these characteristics are unavailable, the article name (e.g., axle, curing part 182, or set 112 engine cover) was used as category (3.1).
The load carriers (4) also influence the process. The load carrier type (4.1) describes which load carrier it is (e.g., pallet, mesh box, AKL container, and package). If no load carrier is specified, a single piece is handled. In addition, processes partly differ in whether a load carrier is loaded with full goods, mixed goods, or bulky goods or whether it is empty (4.2). The dimensions (4.3) and weight (4.4) of the load carrier and the possibility of stacking it (4.5) influence the process design as well.
Further influences arise from special customer requirements, location specifications, and industry specifications. For this reason, the customer name (5), location (6), and industry (7) industry were taken into account in the present data set.
In addition, it is essential at which position of the process a new process step is to be added. Hence, previous process steps (8) must be taken into account. This contribution explores how many previous process steps should be considered. Since there are no or only a few predecessors for the first process steps, the features “Up to n previous MTM codes” (pMTMC) were generated (8.1–8.9). Table 3 shows what “Up to n” means based on the example process mentioned in Table 1. This analysis examines up to nine predecessors.
Table 3
Exemplary description of the features “Up to n MTMC”
#
MTM code
Up to 1 pMTMC
Up to 2 pMTMC
Up to 3 pMTMC
1
SFISF
No predecessor
No predecessor
No predecessor
2
SABAFM
SFISF
SFISF
SFISF
3
SFISF
SABAFM
SFISF-SABAFM
SFISF-SABAFM
4
SABAFM
SFISF
SABAFM-SFISF
SFISF-SABAFM-SFISF
5
AZA
SABAFM
SFISF-SABAFM
SABAFM-SFISF-SABAFM
6
KA
AZA
SABAFM-AZA
SFISF-SABAFM-AZA
7
VA
KA
AZA-KA
SABAFM-AZA-KA
8
EH3
VA
KA-VA
AZA-KA-VA
9
IAHO3
EH3
VA-EH3
KA-VA-EH3
10
IAKW
IAHO3
EH3-IAHO3
VA-EH3-IAHO3
Figure 1 displays the histogram of the 256 unique MTM codes used to identify the process steps, with logarithmic scaling of the y-axis. For better readability, only every sixth MTM code is written on the abscissa. The figure shows a strong imbalance of the MTM code classes. The most frequent MTM code is used 1371 times. The rarest MTM code is used only once. This leads to the problem that rare MTM codes can either be learned but not used, or not learned at all. To have enough samples for train the classifier, all MTM codes occurring currently less than seven times were removed from the data set. This reduces the amount of different MTM codes to 156. The data set decreases from 13,707 samples to 13,432 samples.
The aim is to predict the most probable next MTM code, so the classifier has to choose from the 256 possible MTM codes (multi-class). Since the classes vary in frequency (see Fig. 1) and only categorical features are currently available (see Table 2), predicting the next process step is an imbalanced multi-class classification problem with categorical data. The following section describes the method used to deal with this problem.

3.2 Description of used methods

The used methods were coded in Python using the scikit-learn library [21], version 1.2.1, if not otherwise noted. Figure 2 illustrates the proceeding of supervised learning in this study.

3.2.1 Metrics

A suitable metric is required to address the data set imbalance. The ROC AUC score is insensitive to class imbalance and, thus, an appropriate metric to evaluate the quality of the model’s predictions. ROC stands for Receiver Operating Characteristic, which is a curve that plots the true positive rate against the false positive rate at various classification thresholds. The AUC (Area Under the Curve) score is the area under this curve, which ranges from 0 to 1. A perfect classifier has an AUC score of 1, while a random classifier has an AUC score of 0.5. In the following, the macro ROC AUC is calculated using the one-vs-rest-method.
In addition, the accuracy (ACC) is calculated. The ACC determines in how many cases the prediction was correct. In this context, correct means that the exact MTM code was predicted. For example, if the MTM code EH3 was used, but the classifier predicted EH2, this prediction is incorrect. Accuracy is the metric of most interest for the later application. However, accuracy can be misleading in imbalanced data sets since a model that only predicts the majority classes and ignores the rare classes can still achieve high accuracy. Therefore, attention should be paid to false negative and false positive predictions, which is why most attention is paid to ROC AUC during the development of the classification model.
3.2.1.1 Preprocessing
During the preprocessing, the rare classes are removed as mentioned above. Furthermore, missing data are imputed by introducing a category “unknown” for each feature. To handle the categorical data, it is necessary to encode the categories. There are many ways to encode categorical data. A common practice is to use a one-hot-encoder, which transforms each categorical value into a single feature with 0 or 1. 0 means the value is absent, 1 means the value is present. A one-hot-encoder is suited for a wide range of classifiers. However, preliminary work found that tree-based classifiers are best suited for the presented data [19]. This kind of classifier can also handle categorical values if label encoded. A label encoder transforms just each categorical feature value to an integer (0 to amount of categories-1). In contrast to one-hot-encoding, this saves memory and computational energy. Thus, label encoding was conducted by using the label encoder from scikit-learn.
3.2.1.2 Data split
For the investigation, the data are split into training and test data in a four-to-one ratio. A stratified fold ensures that the class distribution is the same in the training and test data set. A cross-validation (CV) with a stratified 5-fold is performed for feature selection and hyper-parameter optimization in the training set. The selected features and optimized model are used to predict the process steps of the test data set.
3.2.1.3 Model selection
In preliminary work, tree-based classifiers were evaluated as suitable models for the presented use case [19]. Consequently, this investigation starts with the Decision Tree Classifier (DT) [22] and the Random Forest Classifier (RF) [23] from the scikit-learn library as well as the XG Boost Classifier (XG) from XG Boost Library [24]. Due to its tree boosting the XG Boost Library has been shown to give state-of-the-art results on many standard classification benchmarks [25].
3.2.1.4 Feature selection
The features of piece/bulk goods, food, and hazardous materials cannot be considered in this analysis because they are only available in one value and, therefore, are irrelevant to decision-making in this case. Thus, the decision-making features ‘Customer’, ‘Location’, ‘Industry’, ‘Article’, ‘Load carrier type’, ‘Filling’, ‘Main process’, ‘Supply means’, ‘Equipment’, and ‘Employee group’ remain in the data set. These ten features are summarized as s0. In addition, there are nine constructed “up to” features. Assuming only one “up to” feature should be selected, the pre-selected s0 features will be evaluated with any “up to” to find the best combination.
3.2.1.5 Hyper-parameter optimization
For the best feature combination and the best model, a hyper-parameter optimization is performed in a grid search with cross-validation applying a stratified 5-fold to get the best result in the training set.
3.2.1.6 Evaluation
Finally, the optimized classifier is evaluated using the selected features on the previously unused test data set.

3.3 Plausibility of the results

To check the plausibility of the results, the feature combinations (FC) in the data set are analyzed in more detail. The analysis takes advantage of the categorical nature of all features used. Due to this categorical nature, a classifier can only make a meaningful prediction based on a known feature combination. A transfer to unknown feature combinations is not possible with the methods used. An example shall clarify this: The classifier knows how unloading pallets from a truck proceeds. It cannot transfer this knowledge to unloading packages from a truck.
The following cases may occur in the data set:
Case (a)
A FC leads to a class in 100% of the cases and occurs multiple times in the data set. A classifier must recognize this pattern. It is a must-have.
 
Case (b)
A FC leads to a class in 100% of the cases because the FC is unique in the data set. These patterns can either be learned by a classifier but not applied or not learned because the pattern occurs only in the test data set.
 
Cases (c)
The same FC can lead to different classes. The classifier should predict the most likely class (c1). However, the unlikely classes are expected to be mispredicted because they are non-distinguishable (c2). Case (c) can occur if relevant decision features are unavailable or different planners assessed the same situation (the same FC) differently and had planned other process steps.
 
A large fraction of case (a) and case (c1) is useful for classification because these cases are well classifiable. Since the FC corresponding to case (c2) are not classifiable, the fraction of (c2) must be kept as low as possible. Similarly, the fraction of cases (b) should be kept as low as possible. However, in these cases, a correct classification can be achieved by chance or if the classifier consider only a feature subset.

4 Results

This section presents the results of the feature and model selection, followed by the results of the plausibility check and the hyper-parameter optimization. Finally, the selected and fine-tuned model is evaluated using the unseen test data set.

4.1 Feature and model selection

This section examines the models and the predecessor features. The complete training data set with 10,745 samples and 156 different MTM codes is used. The nine different “up to” features are tested together with the previously selected features s0. For this purpose, a cross-validation was performed on the training data set for each of the nine feature combinations. Tables 4, 5 and 6 list the exact values for the ROC AUC score as well as for the ACC including the standard deviations (SD) across the 5 folds in the cross-validation for the three different models.
Table 4
Evaluation of the Decision Tree classifier depending on feature combination
Feature Combination
ROC AUC
SD ROC AUC
ACC
SD ACC
s0 + Up to 1 pMTMC
0.7848
0.1341
0.5163
0.0109
s0 + Up to 2 pMTMC
0.8224
0.1321
0.7231
0.0097
s0 + Up to 3 pMTMC
0.8267
0.1330
0.7575
0.0092
s0 + Up to 4 pMTMC
0.8260
0.1301
0.7662
0.0108
s0 + Up to 5 pMTMC
0.8221
0.1316
0.7670
0.0108
s0 + Up to 6 pMTMC
0.8222
0.1314
0.7698
0.0095
s0 + Up to 7 pMTMC
0.8195
0.1329
0.7675
0.0105
s0 + Up to 8 pMTMC
0.8193
0.1326
0.7658
0.0107
s0 + Up to 9 pMTMC
0.8194
0.1317
0.7642
0.0091
The best scores of each metric (ROC AUC and ACC) are highlighted in bold
Table 5
Evaluation of the Random Forest classifier depending on feature combination
Feature Combination
ROC AUC
SD ROC AUC
ACC
SD ACC
s0 + Up to 1 pMTMC
0.8997
0.1046
0.4775
0.0060
s0 + Up to 2 pMTMC
0.9184
0.0931
0.6040
0.0073
s0 + Up to 3 pMTMC
0.9269
0.0786
0.6282
0.0059
s0 + Up to 4 pMTMC
0.9223
0.0862
0.6338
0.0056
s0 + Up to 5 pMTMC
0.9217
0.0852
0.6366
0.0050
s0 + Up to 6 pMTMC
0.9209
0.0848
0.6350
0.0055
s0 + Up to 7 pMTMC
0.9219
0.0870
0.6378
0.0063
s0 + Up to 8 pMTMC
0.9242
0.0823
0.6344
0.0053
s0 + Up to 9 pMTMC
0.9234
0.0840
0.6331
0.0031
The best scores of each metric (ROC AUC and ACC) are highlighted in bold
Table 6
Evaluation XG Boost classifier depending on feature combination.
Feature Combination
ROC AUC
SD ROC AUC
ACC
SD ACC
s0 + Up to 1 pMTMC
0.9738
0.0284
0.5259
0.0117
s0 + Up to 2 pMTMC
0.9770
0.0281
0.7391
0.0104
s0 + Up to 3 pMTMC
0.9769
0.0286
0.7743
0.0126
s0 + Up to 4 pMTMC
0.9772
0.0287
0.7819
0.0120
s0 + Up to 5 pMTMC
0.9770
0.0293
0.7833
0.0139
s0 + Up to 6 pMTMC
0.9768
0.0293
0.7848
0.0152
s0 + Up to 7 pMTMC
0.9769
0.0286
0.7847
0.0144
s0 + Up to 8 pMTMC
0.9768
0.0287
0.7813
0.0156
s0 + Up to 9 pMTMC
0.9767
0.0289
0.7810
0.0148
The best scores of each metric (ROC AUC and ACC) are highlighted in bold
The DT classifier and the RF achieve the best ROC AUC score with up to 3 pMTMC and the best ACC with up to 6 (DT) or up to 7 pMTMC (RF), see Tables 4 and 5. The XG classifier requires up to 4 pMTMC to achieve its best ROC AUC score and up to 7 pMTMC to reach its best ACC, see Table 6.
Each classifier achieves higher ROC AUC scores than ACC scores. This is quite common, as the ROC AUC score is usually between 0.5 and 1 and the ACC score between 0 and 1. In this case, the high amount of classes and the imbalance of the classes increase the difference.
Figure 3 compares the ROC AUC score of the three classifiers depending on the feature combination. It depicts that the XG classifier achieves the best and most stable ROC AUC scores of the three compared classifiers independent of the feature combination. This implies using the XG classifier and the feature up to 4 pMTMC for further steps.
As with the ROC AUC score, the XG classifier also achieves the highest scores for ACC. Comparing the ACC scores over all classifiers, it is noticeable that the ACC increases more strongly with the addition of predecessors as the ROC AUC score and reaches its maximum at up to 6 or 7 pMTMC.

4.2 Plausibility check

To check the plausibility of these results, the feature combinations are evaluated. Figure 4 illustrates the grouping of the FC according to the approach described in Sect. 3.2.
Figure 4 shows that as the number of predecessors increases, the size of the must-have group initially increases, and the size of the non-distinguishable group decreases. This effect is positive and results in increasing scores. From up to 3 pMTMC, however, the must-have group starts to shrink. Although the non-distinguishable group continues to shrink, only the group of unique FC group grows. This fragmentation is disadvantageous due to the categorical nature of the features. The trend of growing unique FC levels out similar to the scores of ROC AUC and ACC. This emphasizes the plausibility of the results.
The conclusion of these results is, on the one hand, that the XG Classifier achieves the best results for this data set. On the other hand, considering up to 4 pMTMC is sufficient. Considering more predecessors may improve the ACC even slightly. However, this effect is probably due to the imbalance of the data set. Taking too many predecessors into account leads to a higher fragmentation of the FC and, thus, to lower scores.

4.3 Hyper-parameter optimization

In the next step, a grid search with cross-validation is performed to optimize the hyper-parameters of the XG Boost classifier. Table 7 lists the grid and the best parameters.
Table 7
Grid search for hyper-parameter optimization
Parameter
Grid
Best
Maximal depth
3, 5, 7, 9, 11, 13, 15
11
N estimators
50, 100, 200, 300
200
The best configuration (maximal depth = 11 and 200 estimators) leads to the best ROC AUC score of 0.9812 and an ACC of 0.7901 in the training data set. Consequently, a minor improvement could be achieved through optimization.

4.4 Test result

Finally, the XG classifier is evaluated with the selected feature combination and the optimized hyper-parameters on the previously unused test data set. Figure 5 plots the result of this evaluation and compares the test results with the training results. The ACC reaches a score of 0.8091 and the ROC AUC score is 0.9752 which means a good model fit. Furthermore, the test results are very similar to the scores achieved in training. Surprisingly, the ACC score on the test data is a bit higher than on the training data. Thus, a suitable generalization of the model can be assumed.
Assuming that the classifier suggests a list of probable MTM codes, the top k accuracy is calculated, where k is the length of the list. For example, assuming the XG classifier should suggest a list of the five most probable MTM Codes, the top 5 accuracy indicates if the correct MTM code is in this list. Figure 6 shows the top k accuracy from k = 1 to 10.
The top one accuracy is precisely the accuracy mentioned before, with 0.809. The more extensive the list of proposed MTM codes, the higher the probability that the correct MTM code is listed. For example, the top five accuracy is 0.933.
Machine learning was performed on a Lenovo Think Pad P16s with an Intel Core i7-1260P and 32 GB of RAM. Training the optimized model took 77.39 seconds. Predicting all the test data took 0.28 seconds, giving 0.0001 seconds per prediction.

5 Discussion

With an accuracy of 80.9%, the XG classifier suggests well. This is notable since contract logistics projects are very individual. However, it must be taken into account that rarely used MTM codes were initially rejected. Thus, these individual process steps are not fully considered in this study.
The results were discussed with logistics planners from the practice. The feedback was positive. In particular, the suggestion of a list of probable MTM codes with a generic process step description (e.g., for MTM code “SFISF” a generic description would be “drive with fork lifter from X to Y”) and a categorized probability (e.g., high, medium, low) supports inexperienced planners. This is because selecting from a list of suggestions is more accessible than planning a process without guidance. Based on the current results, standard processes or frequently occurring process segments can thus be planned very well with AI support. For individual processes or process segments, experienced planners are indispensable.
Figure 7 displays the feature’s importance obtained from the trained XG classifier, where the weight is the number of times a feature appears in the trees. The consideration of the predecessors plays a significant role. This can be explained by the fact that some information about the other features is already encoded in the MTM code. For example, MTM codes are partially used in an equipment-specific way. Furthermore, the sequence of MTM codes allows for specifying the activity more precisely, which supports the prediction of the next process step.
However, the importance of these features cannot be generalized. It is only valid for the present data set. For example, the feature “Industry” only exists in two variants in the presented data set. Hence, no tree uses this feature, thus it is not listed in Fig. 7. If the data set is extended to contain more industries, the relevance of this feature can increase. This aspect argues for using tree-based classifiers in this application. Their inherent evaluation of the feature importance by calculating the information gain can streamline the pre-processing in a later use with a constantly growing data set. Other classifiers need more preprocessing, e.g., the feature selection.
The result of this study outperforms the accuracy of 78.5% achieved in the previous work [19]. Nevertheless, it is notable that the result of previous work could not be exceeded much despite feature evaluation of the most crucial feature of pMTMC and hyper-parameter optimization. This leads to the question of whether the achieved ACC score is already the maximum for the current data set or whether a different pre-processing or another classifier can achieve a better score.
The result of the FC analysis indicates an answer to this question. More than half of the feature combinations are unique in the data set. Consequently, they can be learned but not applied or not learned at all. Knowledge transfer from one feature combination to another is impossible with the current method due to the categorical nature of the features.
Improvements could be reached by using another encoding method. Using word embeddings for the categorical data can encode similarities, enabling the classifier to transfer knowledge from one category to another. In addition, the domain of deep learning classifiers still needs to be tested for the presented data set.
However, a more appropriate solution is the gathering of further data. This is even more relevant since unique and rare classes (MTM codes) already had to be removed from the data set at the beginning of the pre-processing. As the number of samples increases, the number of unique FCs will likely decrease, and the currently unique classes will also become more common. However, there is also a risk that the number of non-distinguishable FC will increase. All decision-relevant data (see Table 2) should be recorded to avoid non-distinguishable FC. If these data are available, the relevant features can be extracted. Consequently, more data should be gathered to improve the accuracy of process step predictions.
There are two possible ways for the deployment. Since the MTM calculation is nowadays performed via a spreadsheet, the spreadsheet could be extended by the classification model, e.g. via PyXLL or xlwings. Spreadsheets offer a lot of freedom in planning, but are error-prone and not designed for ML applications. An implementation in existing professional MTM planning environments or AI-based assistance systems seems to make more sense. In particular, such software systems allow companies to build up a central database of all contract logistics projects, so that over time a volume of data and, thus, knowledge is built up, which further increases the advantage of such an ML application.
Predicting the next MTM code or the selection list is performed in milliseconds, so that the MTM code suggestion can be rapidly presented to the planner. After the completion of a new process planning, this planning should extend the training data set and the model should be trained again to learn the new knowledge. As this is not time-critical, the learning phase can be performed as batch training overnight.

6 Conclusion

This contribution describes how supervised learning can support process planning in contract logistics. This has the advantage that knowledge from past planning can be extracted and suggested for new planning. In particular, young and inexperienced planners benefit from this. Due to the upcoming shortage of skilled workers, this approach is auspicious.
The current data set can be used to suggest the correct process step in 80% of the cases. Since the supervised learning approach is intended to support planers, not to replace planers, this score can be rated as suitable. Nevertheless, this article discusses several ideas to enhance the results. The current results suggest that first of all a more comprehensive data set should be acquired to improve the accuracy. Until now, only some decision-relevant features have been included in the data set. The article provides information on which data should be collected and stored in planning today so that a machine supervised learner can use it in the future.
In addition, the article provides indications that recurring MTM sequences can be found in the process plans. This is shown by the fact that the previous MTM code sequence is the most vital feature. A suggestion of MTM code sequences (process step groups) would significantly accelerate planning, compared to single-step planning, and, therefore, it is a rewarding goal. Further research is required to find meaningful MTM code sequences, preprocessing steps, and classification models to predict these sequences.

Acknowledgments

This work results from the research project “INSERT—AI-based assistance system for concept planning in production and logistics”, funded by the BAB—Bremer Aufbaubank under the reference number FUE0626B and with funds from the European Regional Development Fund (ERDF).

Declarations

Conflict of interest

The authors declare that they have no conflict of interest.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://​creativecommons.​org/​licenses/​by/​4.​0/​.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Anhänge

Supplementary Information

Below is the link to the electronic supplementary material.
Literatur
1.
Zurück zum Zitat Prahalad CK, Hamel G (2006) The core competence of the corporation. In: Hahn D, Taylor B (eds) Strategische Unternehmungsplanung/Strategische Unternehmungsführung. Springer-Verlag, Berlin/Heidelberg, pp 275–292CrossRef Prahalad CK, Hamel G (2006) The core competence of the corporation. In: Hahn D, Taylor B (eds) Strategische Unternehmungsplanung/Strategische Unternehmungsführung. Springer-Verlag, Berlin/Heidelberg, pp 275–292CrossRef
2.
Zurück zum Zitat Schwemmer M, Klaus P (2021) TOP 100 in European transport and logistics services: Market sizes, market segments and market leaders in the European logistics industry, 2021/2022. DVZ Top 100. DVV Media Group, Hamburg Schwemmer M, Klaus P (2021) TOP 100 in European transport and logistics services: Market sizes, market segments and market leaders in the European logistics industry, 2021/2022. DVZ Top 100. DVV Media Group, Hamburg
4.
Zurück zum Zitat Mühlencoert T (2012) Kontraktlogistik-management. Gabler Verlag, WiesbadenCrossRef Mühlencoert T (2012) Kontraktlogistik-management. Gabler Verlag, WiesbadenCrossRef
5.
Zurück zum Zitat Geiger C, Eikelau M (2013) Kontraktlogistik. In: Clausen U, Geiger C (eds) Verkehrs- und Transportlogistik. Springer, Berlin Heidelberg, pp 95–119CrossRef Geiger C, Eikelau M (2013) Kontraktlogistik. In: Clausen U, Geiger C (eds) Verkehrs- und Transportlogistik. Springer, Berlin Heidelberg, pp 95–119CrossRef
6.
Zurück zum Zitat Straube F (2014) Entwicklung und Implementierung eines praxistauglichen Instrumentariums zur Produktivitätsmessung und Wertdarstellung für Dienstleister in der Kontraktlogistik: Schlussbericht des Verbundprojektes ProKoLog, Berlin Straube F (2014) Entwicklung und Implementierung eines praxistauglichen Instrumentariums zur Produktivitätsmessung und Wertdarstellung für Dienstleister in der Kontraktlogistik: Schlussbericht des Verbundprojektes ProKoLog, Berlin
7.
Zurück zum Zitat Straube F, Ouyeder O, Siegmann J et al. (2011) Empirical Analysis of the tender management process of contract logistics service providers. In: Blecker T (ed) Maritime logistics in the global economy: Current trends and approaches, 1. Aufl. Eul, Lohmar, Köln, pp 393–407 Straube F, Ouyeder O, Siegmann J et al. (2011) Empirical Analysis of the tender management process of contract logistics service providers. In: Blecker T (ed) Maritime logistics in the global economy: Current trends and approaches, 1. Aufl. Eul, Lohmar, Köln, pp 393–407
8.
Zurück zum Zitat Veigt M, Steinbacher L, Freitag M (2021) Planungsassistenz in der Kontraktlogistik: Ein Konzept zur KI-basierten Planungsunterstützung innerhalb einer digitalen Plattform. Ind 4.0 Manag 37:11–15 Veigt M, Steinbacher L, Freitag M (2021) Planungsassistenz in der Kontraktlogistik: Ein Konzept zur KI-basierten Planungsunterstützung innerhalb einer digitalen Plattform. Ind 4.0 Manag 37:11–15
16.
Zurück zum Zitat Karger DW, Bayha FH (1987) Engineered work measurement: The principles, techniques, and data of methods-time measurement, background and foundations of work measurement and methods-time measurement, plus other related material, 4th edn. Industrial Press Inc, New York Karger DW, Bayha FH (1987) Engineered work measurement: The principles, techniques, and data of methods-time measurement, background and foundations of work measurement and methods-time measurement, plus other related material, 4th edn. Industrial Press Inc, New York
17.
Zurück zum Zitat Kochmar E (2022) Getting started with natural language processing. Manning Publications Co. LLC, New York Kochmar E (2022) Getting started with natural language processing. Manning Publications Co. LLC, New York
18.
Zurück zum Zitat Veigt M, Steinbacher L, Freitag M (2022) Using supervised learning to predict process steps for process planning of third-party logistics. In: Freitag M, Kinra A, Kotzab H et al (eds) Dynamics in logistics. Springer International Publishing, Cham, pp 423–434CrossRef Veigt M, Steinbacher L, Freitag M (2022) Using supervised learning to predict process steps for process planning of third-party logistics. In: Freitag M, Kinra A, Kotzab H et al (eds) Dynamics in logistics. Springer International Publishing, Cham, pp 423–434CrossRef
19.
Zurück zum Zitat Veigt M, Steinbacher L, Freitag M (2023) KI als Planungsassistenz in der Kontraktlogistik: Überwachtes Lernen zur Vorhersage von Prozessschritten in der Planung von Logistikprozessen. Industrie 4.0 Management 39:9–13 Veigt M, Steinbacher L, Freitag M (2023) KI als Planungsassistenz in der Kontraktlogistik: Überwachtes Lernen zur Vorhersage von Prozessschritten in der Planung von Logistikprozessen. Industrie 4.0 Management 39:9–13
20.
Zurück zum Zitat Quinlan JR (2014) C4.5: Programs for machine learning. Ebrary online. Morgan Kaufmann Publishers, San Mateo Quinlan JR (2014) C4.5: Programs for machine learning. Ebrary online. Morgan Kaufmann Publishers, San Mateo
21.
Zurück zum Zitat Pedregosa F, Varoquaux G, Gramfort A et al (2011) Scikit-learn: machine learning in python. J Mach Learn Res 12:2825–2830MathSciNet Pedregosa F, Varoquaux G, Gramfort A et al (2011) Scikit-learn: machine learning in python. J Mach Learn Res 12:2825–2830MathSciNet
22.
Zurück zum Zitat Breiman L, Friedman JH, Olshen RA et al (2017) Classification and regression trees. RoutledgeCrossRef Breiman L, Friedman JH, Olshen RA et al (2017) Classification and regression trees. RoutledgeCrossRef
25.
Zurück zum Zitat Chen T, Guestrin C (2016) XGBoost: A Scalable Tree Boosting System. In: Krishnapuram B, Shah M, Smola A et al. (eds) Proceedings of the 22nd ACM SIGKDD Int. Conf. on Knowledge Discovery and Data Mining. ACM, New York, NY, USA, pp 785–794 Chen T, Guestrin C (2016) XGBoost: A Scalable Tree Boosting System. In: Krishnapuram B, Shah M, Smola A et al. (eds) Proceedings of the 22nd ACM SIGKDD Int. Conf. on Knowledge Discovery and Data Mining. ACM, New York, NY, USA, pp 785–794
Metadaten
Titel
Supervised learning to support the process planning of contract logistics projects
verfasst von
Marius Veigt
Michael Freitag
Publikationsdatum
01.08.2023
Verlag
Springer Berlin Heidelberg
Erschienen in
Production Engineering / Ausgabe 1/2024
Print ISSN: 0944-6524
Elektronische ISSN: 1863-7353
DOI
https://doi.org/10.1007/s11740-023-01217-3

Weitere Artikel der Ausgabe 1/2024

Production Engineering 1/2024 Zur Ausgabe

    Marktübersichten

    Die im Laufe eines Jahres in der „adhäsion“ veröffentlichten Marktübersichten helfen Anwendern verschiedenster Branchen, sich einen gezielten Überblick über Lieferantenangebote zu verschaffen.