07062020  Production Management  Issue 3/2020 Open Access
Reinforcement learning for an intelligent and autonomous production control of complex jobshops under time constraints
 Journal:
 Production Engineering > Issue 3/2020
Important notes
Publisher’s Note
Springer Nature remains neutral with regard to
jurisdictional claims in published maps and institutional affiliations.
1 Introduction
Manufacturing companies are subjected to constant transformations of their internal processes and their environment [
24]. Globalization and competition through emerging industries in developing nations are already wellknown [
11]. Especially, with the ongoing digitalization accelerating markets are becoming unpredictable, and companies are in need to react quickly and decisively [
1]. Optimally exploiting the operational abilities and resources is of utmost importance under these conditions.
In the semiconductor industry, complex job shops are facing similar challenges [
23]. Job shops are widely evaluated in other industries, since the requirements for flexible and rapidly changeable production systems are increasing [
4]. Complexity, as well as opportunities and the pressure to make use of the latter, are therefore essential endeavours to ensure a competitive position [
23].
Advertisement
The emergence of artificial intelligence methods such as deep learning, reinforcement learning (RL) and other machine learning approaches bear the potential to encounter these challenges with new quantitative methods. The methods are backed by significantly decreased computational times, the development of easy to use open source libraries and achievements like beating the human experts in strategic games [
19].
The objective of this paper is the development and implementation of an autonomous and selflearning algorithm addressing order dispatching with strict time constraints in complex job shops. This is the first publication which utilizes RL to manage an environment with strict time constraints.
The research is structured as follows: Sect.
2 introduces the fundamentals of time constrained production planning and control, the essential characteristics of job shops in the semiconductor industry, and the basics of RL. Section
3 describes the method and RLalgorithm that has been developed in the present work, leading to the computational result in Sect.
4.
2 Fundamentals and literature review
2.1 Production planning and control
Production planning and control (PPC) focuses on organizing and optimizing the internal processes within a manufacturing system [
3]. Production planning predefines the production program, including the production process setup information. Production control takes these predefined inputs to schedule all processes necessary to fulfill the production plan and, thereby, optimally utilize the available production factors given changing circumstances such as shortages of material, machines or workforce [
12,
18]. The order release is the interface between production planning and production control.
Advertisement
Order dispatching, which is the focus of this work, is one PPC task and considers the assignment of orders to the next processing machine given a set of available machines and orders. Other tasks are, for instance, order sequencing which defines the sequences in which individual machines process orders [
5,
17].
Optimization and decision support methods that are used for PPC are often categorized into three classes [
14]: first, there is mathematical optimization, which allows finding the optimal solution but requires high computational effort. Second, heuristics are in widespread use to overcome the computational drawbacks in large realworld problems. They achieve acceptable results in much less time but fall short of the optimal solution in most cases. Learningbased techniques, the third class of methods, are considered in the present work.
2.2 Complex job shops in semiconductor manufacturing
Certain features are typical for the wafer fabrication [
14]. Reentrant flows let wafers run through the same machine group more than once. Some machine types are relatively unreliable by nature of the physical manufacturing process and need frequent maintenance measures. Within the same process chain, there are not only serial but also batch processes. As a consequence, massive queues are caused by the succession of parallel batching tools with serial tools. Moreover, dynamic bottlenecks occur. A multifaceted product portfolio and changing product mix ensure constantly varying boundary conditions. For some tools, setup times and multiple processing times are further challenges. Finally, highly competitive due dates prevail.
All in all, wafer fabs belong to the category of complex job shops as a distinct job shop type with the features described above [
23].
2.3 Handling of time constraints in PPC
Time constraints, also called timecoupling constraints, describe the maximum time that is allowed to pass between the end of one process step and the start of another one (see [
14]). Time constraints are a significant challenge in modern manufacturing systems, but in particular present in the semiconductor industry [
12]. Hence, time constraints are an essential part of realistic manufacturing process modelling. Some research focused on the incorporation of time constraints in PPC methods.
Klemmt and Mönch model time constraints and present an approach for solving a scheduling problem with time constraints [
6]. Besides customer due dates, time constraints can also exist between two or more consecutive process steps. The authors classify time constraints in five different types, depending on the consecutiveness, adjacency, and overlapping of time constraints. Overlapping time constraints occur in complex job shops, which makes the computation of an optimal solution infeasible for realworld problems.
Sun et al. introduce delay time constraints in the application domain of semiconductor manufacturing [
2]. Delay time constraints refer to a period within a consecutive process needs to be started based on a start process step. The authors present objective function and constraint formulations that are required to consider such time constraints in existing manufacturing control systems. The approach is based on mixedinteger programming and constrained programming models. However, experimental results from an application are not presented.
Knopp addresses scheduling problems in complex semiconductor job shops with heuristic and metaheuristic approaches [
7]. The focus of his work is on batching processes and time constraints. The time constraints are considered as maximum time lag constraints and implemented in the model as soft constraints with violation costs attached. Reworkable and nonreworkable time lags are distinguished. Reworkable time lags lead to a lot requiring rework to be done in case of violation. This results in increased cycle time for the lot itself and additional machine capacity, which raises its relevance for the overall fab performance. Nonreworkable time lags describe a point in time at which the risk of the wafers to become defective starts to increase for the violation.
2.4 Basics of RL
Machine learning techniques cover a variety of algorithms that can process large amounts of data and identify patterns that can be transferred to new situations [
16]. They are suitable to meet the demand for dynamic, realtime production control applications [
21].
In RL, being one type of machine learning algorithms, agents constantly adapt and learn strategies in a known or unknown environment through feedback received: the agent perceives the state of the environment
\(s_{t} \in S\) from a finite set of possible states at each time
t. On this basis, it selects an action
\(a_{t} \in A\) from the set of possible actions. The environment responds with the resulting state
\(s_{t+1}\) and a feedback, the reward
\(r_{t} \in {\mathbb {R}}\), before the next iteration starts [
21]. This iteration can be described by a Markov decision process, i.e.
\(s_{t+1}\) depends only on
\(s_{t}\), not on previous states. The agent’s goal is to optimize its strategy to maximize the cumulative reward. The strategy is represented by a probability distribution for choosing a specific action in a given state.
3 Use case description and RL modelling
This section introduces the use case taken from a realworld waferfab. The RL modelling with respect to the state and action space, reward function, optimization algorithm, as well as the time constraint consideration are outlined. Overall, the following questions are to be answered:

Is the proposed singleagent based RL approach able to learn and, therefore, improve over time its ability in optimizing the adherence to a time constrained schedule?

Can the proposed approach perform better than established competitive heuristics?
3.1 Description of the waferfab use case
The considered use case represents a production process with ten specialized machines suborganized in five machine groups. Every machine group, also called work center, has a buffer stock with a fixed number of twenty buffer slots. Challenges arise from many timecoupled process steps ranging over multiple work centers. The high share of time coupling constraints for around 30% of the operations, increases the complexity of the order process flow and the internal logistics. Moreover, machines also suffer from reduced availability. Furthermore, order flow is reentrant in earlier visited machine groups. Finally, machines require a product specific setup. Hence, the use case incorporates the following complexity drivers introduced in Sect.
1: nonlinear process flows, reentrance flows, sequencedependent setup times, and time constraints. This renders an appropriate scenario for the application of RL to analyse the influence of time constraints within a complex environment.
Here, two product variants are modelled with an equal product mix proportion but different process times and lengths of time constraint intervals (Table
1). The first product variant is faster but has more stringent time constraints than the second. Both product variants can be processed on all machines and have the same basic recipes and, thus, identical process chains. So, on average, orders of both product variants basically have to wait for the same time depending on what orders are waiting in front of the machine. However, the agent’s decisions could favor one product variant over the other (Fig.
1).
Table 1
Parametrization of the simulation in the waferfab use case
System parameters

Waferfab use case


Number of machine groups

5

Number of machines per machine group

41212

Number of product types

2

Probability of product type being generated

50%, 50%

Size of buffers before machine groups

20 order slots

WIP restriction

80

Order release time interval

70

Recipe (machine group sequence)

(123145) x2

Average RPT (per machine)

Between 20 and 80

Machine failures (per machine)

MTBF between 1700 and 4000, MTTR between 15 and 60

Setup times (per machine)

Between 0 and 15

Target flow factor (per product type)

12, 8

Timecoupling constraint (per product type)

Between 225 and 600

×
3.2 State space modelling
The state space depicts the information given to the agent as basis for decisionmaking on which action to select.
The state space defined in this work is quite rich and has 210 entries consisting of the following elements, which are described in the remainder of this section:
The state values are mostly binary and ranging from 0 to 1. In order to normalize categorical data, onehot encoding is applied. The state values are determined based on the state of the environment at the point of decisionmaking.

Current machine where the action is asked for (10 entries, binary, onehot coding for 10 work groups)

Loading status of all machines (10 entries, binary, idle or busy)

Product setup per machine (50 entries, binary, onehot coding for 5 product variants and ten machines)

Product variant in the current machine’s buffer slots (100 entries, binary, onehot coding for up to twenty buffers and 5 product variants)

Order status per buffer slot (20 entries, realvalue, 20 buffer slots)

Full or empty status per buffer slot (20 entries, realvalue, 20 buffer slots)
In this use case, the order status observed by the RLagent is given by an order’s rank of “urgency ratios” (explained in the next section, taking a value between − 1 and 1). The three orders with the highest urgency are given the rank values 3, 2, and 1. All other orders are labelled with a 0. Blank slots in the buffer of a machine group are indicated by a value of − 5. Thus, jobs of different product variants with different time constraints and different processing times are made comparable. In addition to the order’s urgency ratio, the benchmark heuristics use a due date deviation based on planned cycle time of a job as second prioritization rule, determined by the raw process time multiplied with a target flow factor. The raw process times are given by the technological process parameters and the product variant. The target flow factors is a predetermined input values, that considers strategic objectives such as customer lead times.
It is important to note that the observation of the orders waiting at a work center is shuffled with respect to buffer slot position before communicating the order status per buffer slot to the agent. This procedure leads to a random allocation of slots for every order and every empty slot. This is a crucial element in preventing the agent from learning a biased policy to pick slotoriented and not orderspecific.
Further, the setup status type of every tool is included, in order to enable the agent to optimize setup sequences of the machines. Combining the machines’ setups with respect to the product variants should result in a better setup and load management.
3.3 Action space modelling
Whenever the agent is requested, it chooses one of 21 options:
Each of the twenty slots of a buffer is either occupied or empty. Only selecting an occupied slot is considered a valid action. Invalid actions are not executed (i.e. the simulation state is not changed) but used as feedback for learning: the agent is penalized with a negative reward and requested to select another action.

Selecting an order from a buffer slot (20 actions)

Idle and choose no order (1 action)
When experimenting with global rewards (see Sect.
3.5) the agent is given the option to increase or decrease the rate of new order starts to allow control and optimization of the WIP level in the line. This leads to two more options for action:

Increase loading by reducing current inter arrival times at start buffer

Decrease loading by increasing current inter arrival times at start buffer
3.4 Modelling extensions for time constraints
Timecoupling constraints are measured for every order individually. In the present work, time constraints are only considered between two consecutive process steps. The length of the time constraint is product variantspecific and depends on the next process step. If a constraint is violated nothing directly happens to the affected order, however, it influences the agent’s reward. The number of time constraint violations is a key indicator for the performance of a learned strategy or heuristics.
The state space observed by the agent reflects the state with respect to time constraintspecific urgency ratio
UR. The urgency ratio is also considered for the time constraintspecific reward function and for the benchmark heuristics. The urgency ratio
UR indicates how well an order performs according to its time constraint (
\(T_{C}\)) and is calculated via:
The time
t refers to the current simulation time and
\(t_{\text {finished last step}}\) is the point in time when the last process step was completed. Hence, no distinction is made for orders that are delayed at least
\(100\%\) of their time constraint. If an order does not have a time constraint for the currently next process step, it is set to
\(UR = 1\) per default. Therefore, the following cases can occur:
We use as metrics to evaluate the performance of the system with respect to time constraint violation avoidance the total number of time constraint violation events by one job at any machine stock during the production process. Additionally, a cumulative variant of the urgency ratio best described as cumulated delay
\(D_{cum}(p)\) is considered:
which is the average over all lots of product type
p of the cumulated delays measured at the exit buffer.
$$\begin{aligned} UR := \max {\left\{ \frac{T_{C}  (t  t_{\text {finished last step}})}{T_{C}} , 1 \right\} } \end{aligned}$$
(1)

\(UR > 0\): The order is considered to be on time.

\(UR = 0\): If the order is not selected in this time step, it will violate its time constraint.

\(UR < 0\): The order has violated its time constraint.
$$\begin{aligned} D_{cum}(p) ~ := ~ \left\langle \sum _{j \in \text {production steps}} \left( 1UR_j \right) \right\rangle _{\text {finished lots}(p)} \end{aligned}$$
(2)
3.5 Reward function modelling
The reward function
r is the key element that directs the agents behavior and determines the optimal policy. As introduced above, invalid actions, i.e. picking an empty slot, are rewarded with
\(r = 1\) and a reward of
\(r = 0\) is given for the idleaction, i.e. doing nothing. The exponential time constraint reward function used is modelled as a reward continuously following the
UR ratio:
local reward
\(r_{\text {local}}\) is different from the global reward
\(r_{\text {global}}\), which will be introduced afterwards. As the reward is just based on the status at the specific machine for which the action had been selected, it is called “local”. The local reward is bound between 2 and 10. Contrary to the approach used by heuristics, the agent is rewarded for picking orders with a low urgency (i.e.
\(UR \ge 0\)), a reward only sustainable achievable if the agent finds a strategy that removes unfavorable, i.e. high urgency orders, from the system.
$$\begin{aligned} r_{\text {local}} = 2 \cdot 5^{(UR+1)/2} \end{aligned}$$
(3)
In addition to the local reward, two global rewards are investigated. The first directly awards a low number of average time constraint violations measured at the end of the production line:
Here,
\(V_{p}\) is the average number of time constraint violations over the last 50 orders for product type
p . The notation
\(\left<~\right>_p\) refers to the average over all products
p in the exit buffer. The exponential function dampens the reward for increasing
\(V_{p}\).
$$\begin{aligned} r_{\text {global}} = 10 \cdot \left< \exp \left( \frac{V_{p}}{3}\right) \right>_{p} \end{aligned}$$
(4)
Furthermore, experiments showed a tendency of the agent to reduce the order release and reduce new order starts. This leads to an overall less loaded system where time constraints are easier to handle. Therefore, two additional actions are included to increase and decrease order starts and combine this option with a second global reward component
\(r_{\text {WIP}}\) awarding adherence to a predefined WIP level over all buffers
i:
\(\text {WIP}^{\text {target}}_i\) refers to the target WIP level in buffer
i and
\(\beta\) represents a balancing parameter which controls exceeding the target WIP level. Hence,
\(\beta\) controls the tradeoff between a production line prioritizing high WIP levels at the cost of cycle time and time constraint violations, on the one hand, and a line favoring cycle time and time constraint violation avoidance over high WIP levels, on the other hand.
$$\begin{aligned} r_{\text {WIP}} = \prod _{i \in \text {buffers}} 1+\max \left\{ \beta \cdot (\text {WIP}_{i}\text {WIP}^{\text {target}}_i),0\right\} \end{aligned}$$
(5)
The effective total reward transits over the course of the exploration phase from
\(r_{\text {local}}\) to
\(r_{\text {global}}\) to smoothly aim at global learning:
$$\begin{aligned} r_{\text {total}} = r_{\text {WIP}}\cdot r_{\text {local}} \longrightarrow {} r_{\text {WIP}}\cdot r_{\text {global}} \end{aligned}$$
(6)
3.6 RL algorithm and simulation of environment
The DQNagent [
13] provided by the library KerasRL [
15] is the RLagent used in this research. Hereinafter, a short overview of some of the most important settings and hyperparameters is given (see Table
2), based on a review of other DQN applications (e.g. [
22]). A sequential replay memory is used with a limit of one million experiences. The policy deployment uses linear annealing with an
\(\epsilon\) starting at
\(100\%\) and linearly decreasing to
\(1\%\) over time. Depending on
\(\epsilon\) the policy chooses either to “explore” by taking a random action with probability
\(\epsilon\) or to “exploit” by going for the best learned action with probability
\(1\epsilon\). The agent warms up for the first 10,000 steps, as the simulation first needs to fill up and pass the transient phase. The target network is updated every 10,000 steps. The discounting factor is set to
\(\gamma =0.9\) and mean absolute error is used as metric. The adaptive moment estimation optimizer (Adam) is used with a learning rate of
\(\alpha = 0.00025\).
Table 2
The settings for the DQNagent used in this work for the waferfab use case scenario
Parameter

Default value


Number of decisions by the agent

1.2 million steps

Learning until

600,000 steps

Inner policy

Epsilon greedy Q policy

Policy

Linear annealed policy

Exploration/exploitation

\(\epsilon = 100\% \rightarrow 1\%\) /
\(1\%\)

Target model update

Every 10,000 steps

Metric

Mean absolute error (MAE)

Warmup phase

10,000 steps

Size of sequential memory

1,000,000

Discount factor

\(\gamma = 0.9\)

Optimizer

Adaptive moment estimation (Adam)

Learning rate

\(\alpha = 0.00025\)

3.7 Benchmark heuristics
The performance of the RL approach is compared to a benchmark, stateoftheart time constraint control heuristics. A combined time constraint heuristic (short,
TC) chooses the order with the lowest urgency ratio
\(\min {UR}\), if
\(\min {UR} \le 0.5\) and, otherwise, the order with the largest due date deviation. The idea behind the combined rule is to model a more realistic approach with time constraints only being considered if an order comes close to violating its time constraint.
Additionally, a firstinfirstout (FIFO) heuristics is included as benchmark heuristic that is wellestablished in production control application but not directly considering time constraints and due dates.
4 Results
×
×
The computational results are based on simulation runs for the above presented scenario. In every experiment, 1.2 million simulation steps are performed, whereby a simulation step is defined as an agent’s action selection and execution. The exploration value
\(\epsilon\), which refers to the share of explorative actions, is initially set to 1 and decreases linearly to 0.01 over the first 600,000 steps. Hence, for the last 600,000 steps the agent just exploits the best learned action so far. However, it is important to note that learning will not stop after the exploitation phase because random machine failures and order sequence will introduce randomness into the learning process.
4.1 RLagent and heuristic performance analysis—time constraint violations
The benchmark heuristics results are shown in Fig.
2 in the first two charts to the left, the two charts to the right show the results of the RLagent for local and global rewards. The charts show the frequency of time constraint violations
\(V_{p}\) per product type after finishing processing. The entire data set of 1.2 million simulation steps is split into six equalsized phases with a length of 200,000 simulation steps that are shown separately in order to see the learning process for the learningbased agents.
First, one can see that the values spread significantly, especially for the rulebased heuristics on the left due to the system immanent stochastic processes. As expected, the heuristic performance does not change over the course of the phases, as no learning process applies. More interestingly, the TC benchmark achieves results which are similar to FIFO (see Sect.
3.7). For both products, the two heuristics achieve an average number and standard deviation of time constraint violations of
\(3.7 \pm 1.6\) for product 0 and
\(3.1 \pm 1.5\) for product 1 (FIFO), respectively
\(3.7 \pm 1.6\) and
\(3.7 \pm 1.6\) (TC). Focusing only on time constraints, one would expect the TC heuristic to perform better in avoiding violations.
On the one hand, the FIFO heuristic does not specifically consider time constraints, it is following waiting times. Hence, it is profiting from the correlation between the waiting time of an order in the current buffer and the probability of an order violating its time constraint. As shown in Table
1, orders of product type 0 have
\(25\%\) less time to finish than orders of type 1. The heterogeneous product types are not addressed accordingly by FIFO and, hence, the violations are about
\(20\%\) more frequent for product type 0. On the other hand, TC levels the number of violations for both product types better than FIFO. As it always picks the order with the lowest urgency ratio, neither product type is preferred.
Next, the two charts to the right in Fig.
2 show the result of the RLagents (DQNagent). The first represents the agent just with the local reward function
\(r_{\text {local}}\) and the second the reward of the second agent transitioning between Phase 4 and Phase 6 from a purely local reward to an entirely global reward
\(r_{\text {total}}\). After Phase 3, nodeliberate random actions due to
\(\epsilon\) are chosen and the agent decisions are taken as is.
Overall, one can say that the developed reward functions succeed in minimizing the violations against time constraints and clearly outperform both benchmark heuristics. Moreover, the trend over the six phases reveals that the agent is successfully minimizing the number of violations and the average sum of deviations is reduced. For both products, a local reward trained agent achieves an average number of time constraint violations of
\(3.1 \pm 1.0\) and
\(2.1 \pm 0.8\), respectively, and a global reward trained agent even
\(2.4 \pm 0.5\) and
\(1.7 \pm 0.5\), respectively. Note, that in Eq. (
4)
\(V_{p}\) is explicitly entering the global reward function, while in Eq. (
3) the urgency ratio
UR is taken into account for the local reward.
Figure
3 shows the same simulation experiments but with respect to the summed delay
\(D_{cum}(p)\) per product type. This parameter measures the cumulative URperformance and not only the number of TC constraint violation events. Distribution averages and standard deviations for product types 0/1 are: for FIFO
\(7.9 \pm 2.6\)/
\(6.6 \pm 2.3\), TC
\(8.0 \pm 2.6\)/
\(8.0 \pm 2.6\), RL local
\(6.8 \pm 1.8\)/
\(5.2 \pm 1.6\), and RL global
\(5.6 \pm 0.9\)/
\(4.1 \pm 0.9\), respectively.
The RLagent learns based on a cumulative reward that confronts it with the consequences of past actions, especially relevant in a reentrant flow scenario like it is used here. This fact makes the difference between the agent and heuristic approaches, as the latter are limited to a rulebased procedure with local information that is included in the decision rule. In contrast to this, the agent captures during the training phase the relationship between the product type, the order’s urgency ratio, and the combined information of product types and urgency ratios of the other orders in the buffer. Moreover, in the case of the local reward, the agent is rewarded for picking orders with a low criticality (i.e., high urgency ratio), which leads to a strategy that ultimately avoids the occurrence of critical orders. In the case of the global reward, the agent is awarded higher if the final count of time constraint violations after processing is as low as feasible. All this results in agents that are not simply picking the order with the lowest urgency ratio, but instead the order most endangered to a negative urgency ratio. In other words, it is able to create a farsighted control strategy which is more successful than the myopic and greedy approach used by heuristics.
4.2 Investigation of RLagent’s action selection policy
On the left of Fig.
4 it is shown how the local reward approaches over the course of the exploratory learning phase during the first 600,000 steps an average value of about 7.5 by increasingly eliminating invalid actions and neutral actions as well as selecting positive
UR orders, as any
\(UR>0\) leads to a positive reward larger than
\(2 \cdot \sqrt{5}\). The upper and lower bounds of the local reward according Eq. (
3), the zero reward for idle actions and the punishment for invalid actions at
\(1\), are also clearly visible.
In contrast to that, the global reward directly awards the desired overall result, i.e. a small number of time constraint violations
\(V_{p}\) (see Eq.
4). Moreover, it creates an even more complex behaviour due to two additional actions, which increase or lower the number of order releases at the start of the production line. As discussed above, these actions are balanced by a WIPsensitive multiplicative reward component as defined in Eq. (
5) to prevent an underloaded system.
The resulting reward is displayed on the right in Fig.
4. Note the difference in the yaxis scaling due to the multiplication with the WIP reward component. It is interesting to observe how fluctuations induced by WIP imbalances are narrowed down as a balance between order starts, time constraint violations, and predefined WIP target is reached over the learning process.
It should also be emphasized how a minor extension of the action space to include order release actions balanced by a corresponding WIP reward allows to broaden the problem scope and thus solution space. This demonstrates the superior versatility of RL approaches in complex realworld manufacturing settings.
Figure
5 compares the agent’s action selection with two graphs showing the actual buffer slot selected by the agent. Recall that the actions are shuffled in every step and it is, therefore, prevented that the agent memorizes single slots instead of actually learning a control policy. Recall also, that FIFO strictly picks the first slot, i.e. slot 0, which is also preferred but not strictly adhered to by the TC heuristic. However, the two RLagent strategies vary this pattern by including also late entries to the buffer, i.e. higher slot numbers. While the local reward function favors orders that are well in time compared to their time constraint, possibly letting a certain amount of orders run entirely out of time, the global reward function focuses on minimization of time constraint violations, which also means taking care of potential time constraint violations more holistically.
Most of the time, human decision makers would choose orders that can be avoided from running out of time. But this is apparently not the case for the locally rewarded agent, which goes with the two “extremes” of reaping high rewards for many orders and no rewards for late orders. Interestingly, the global reward mitigates this drastic local strategy by continuously smoothing the order picking between the first entered lot all the way down to the last lot.
5 Summary and outlook
5.1 Summary
The opportunities and challenges of manufacturing in times of pervasive digitalization lead to a new era of operations management. Intelligent and autonomous control approaches of complex jobshops are enhanced by reinforcement learning. This work contributes to this research by an investigation of order dispatching under timeconstraints in complex jobshops. The computational results are based on a discreteevent simulation of a semiconductor wafer fab and reveal the high potential of a modelfree temporaldifference
Qlearning algorithm (DQN).
The results show that RLagents can be successfully applied to control order dispatching in a realworldsized application use case. Moreover, time constraints are managed better than industrywide established benchmarks, i.e. a heuristic that optimizes time constraints or a FIFOoriented heuristic.
Furthermore, the experiments show that modelling of rewards is a major and critical part for successful RL applications. This is, in particular, relevant for constraint problems such as orderbased time constraints. Thereby, this research contributes to the research question of how to best enforce RLagents to optimize action validity. Moreover, it is shown how a minor extension of the action space in conjunction with a complementing reward allows to easily increase the scope of action and policy space.
5.2 Outlook
Elaborating and optimizing reward functions as well as the state space seem to offer further potentials for even more improved results. For instance, including the buffer utilization or expected residual machine life time in the state space would allow the agent to optimize the WIP flow and even preventive maintenance schedules [
8]. With respect to the deep learning capabilities, inserting additional layers and more sophisticated layer structures should be envisaged. Moreover, the numerous opportunities for hyperparametrizing RLagents are so far exploited only to a certain degree. Finally, alternative RLagents, e.g. advanced policybased RL algorithms, might achieve good results, too, as they especially prove to be robust given changing problem characteristics.
The longterm vision of selflearning manufacturing systems in realworld is certainly still a long way down the road. For this to happen, improvements need to be made in multiple areas: data availability and quality is still a core issue and a decisive constraint for many companies. Additionally, cybersecurity and management of anomalies are major concerns when implementing autonomous systems. Humans need to get convinced to understand and accept autonomous decisions. Therefore, RL, being the most promising way to go, still needs to be understood deeper and developed further in order to be used as single decisionmaking controller in manufacturing systems.
Acknowledgement
Open Access funding provided by Projekt DEAL.
Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit
http://creativecommons.org/licenses/by/4.0/.
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.