Login| Sign Up| Help| Contact|

Patent Searching and Data


Title:
PREDICTING A BATCH QUALITY VALUE BY A MACHINE LEARNING MODEL FOR A DEDICATED BATCH OF MATERIAL IN A PRODUCTION LINE
Document Type and Number:
WIPO Patent Application WO/2024/041904
Kind Code:
A1
Abstract:
Computer-implemented method for generating a trained ML model predicting a batch quality value for a batch of material, produced in a production entity of a production stage, wherein the batch of material is mixed in a subsequent production entity of at least one further production stage of a production line with other batches of material produced in other production entities of the same production stage, comprising the steps : - collecting sensor data (S1) measured at each production entity of the production stages and a time stamp providing a data collection time at which the sensor data (X j) was collected, - collecting a multiplicity of lab samples (S2 ), each lab sample consisting of a set of products output at the production entities of a last production stage of the production line, wherein the lab sample provides a quality measure of the batch of products, and a time stamp providing a lab sample time at which the lab sample was collected, - establishing (S3 ) a weight matrix comprising weight coefficients of non-zero value for each of the batches of material which contribute to the lab samples, - establishing (S4 ) a ML model which receives the collected sensor data as input and outputs the predicted quality value for each of the batch of material, and - generating (S5 ) a trained ML model (Mt) by jointly optimizing predicted batch quality labels of the ML model and the weight matrix with respect to the collected lab samples.

Inventors:
BIERWEILER THOMAS (DE)
BUSCH JULIAN (US)
LABISCH DANIEL (DE)
LEBACHER MICHAEL (DE)
Application Number:
PCT/EP2023/072165
Publication Date:
February 29, 2024
Filing Date:
August 10, 2023
Export Citation:
Click for automatic bibliography generation   Help
Assignee:
SIEMENS AG (DE)
International Classes:
G05B19/418; G05B13/02; G05B17/02
Domestic Patent References:
WO2022058412A12022-03-24
WO2022136144A12022-06-30
Foreign References:
US20210264332A12021-08-26
US20190332101A12019-10-31
Attorney, Agent or Firm:
SIEMENS PATENT ATTORNEYS (DE)
Download PDF:
Claims:
Claims

1. Computer-implemented method for generating a trained ML model predicting a batch quality value (yi) for a batch of material (Bll, B12, B13, B2n, B3n, B4n) , produced in a production entity (Ell, E12, E13, E2n, E3n, E4n) of a production stage (PSI, PS2, PS3, PS4) , wherein the batch of material (Bll, B12, B13, B2n, B3n, B4n) is mixed in a subsequent production entity (E2n, E3n, E4n) of at least one further production stage (PS2, PS3, PS4) of a production line (10) with other batches of material (B12, B13) produced in other production entities (E12, E13) of the same production stage (PSI) , comprising the steps:

- collecting sensor data (SI) measured at each production entity (Ell, E12, E13, E2n, E3n, E4n) of the production stages (PSI, PS2, PS3, PS4) and a time stamp providing a data collection time at which the sensor data (Xj) was collected,

- collecting a multiplicity of lab samples (S2) , each lab sample (Sj) consisting of a set of products output at the production entities (E4n) of a last production stage (PS4) of the production line (10) , wherein the lab sample provides a quality measure of the batch of products, and a time stamp providing a lab sample time at which the lab sample (S ) was collected,

- establishing (S3) a weight matrix (W) comprising weight coefficients of non-zero value for each of the batches of material which contribute to the lab samples (S ) ,

- establishing (S4) a ML model which receives the collected sensor data (Xj) as input and outputs the predicted quality value (yj) for each of the batch of material (Bll, B12, B13, B2n, B3n, B4n) , and

- generating (S5) a trained ML model (Mt) by jointly optimizing predicted batch quality labels (y^) of the ML model and the weight matrix (W) with respect to the collected lab samples (sz) .

2 . Method according to claim 1 , wherein the collected sensor data (Xi) is provided by a real-valued vector for one specific point in time or a time-series of real-valued vectors .

3 . Method according to any of the preceding claims , wherein those production entities , which contribute to the lap sample (Si) , are determined from a temporal relationship between the lab sample time of the lap sample and the data collection time of the sensor data (Xj) measured at each production entity of the production stages .

4 . Method according to any of the preceding claims , wherein the optimization is performed by minimizing a mean square error between mixed predicted batch quality labels of the ML model and the collected lab samples (Sj ) .

5 . Method according to claim 4 , wherein adding a regularization loss of the weight matrix (W) , while the production entities (Ell , E12 , E13 , E2n, E3n, E4n) have fixed batch quality labels for the lab sample (Sj ) .

6 . Method according to claim 5 , wherein the regularization loss is a L2-regularization which provides uniform weight distributions .

7 . Method according to claim 5 , wherein the regularization loss is similarity based, preferably a gradientregularization, which focusses on batch quality labels (y^) most similar to the lab sample (S ) .

8 . Method according to claim 4 , wherein the optimization applies at least the same number of lab samples (S ) as the number of batches of material exist in the production line ( 10 ) , and the weight matrix W is full rank, and the weight coefficients are fixed . 9. Method according to any of the preceding claims, wherein the ML model is a neural network, comprising an encoder structured as a Multilayer perceptron MLP, convolutional neural network CNN or a Long short-term memory LSTM.

10. Computer-implemented method for predicting a batch quality value (yj) for a batch of material (Bll, B12, B13, B2n, B3n, B4n) produced in a production entity (Ell, E12, E13, E2n, E3n, E4n) of a production stage (PSI, PS2, PS3, PS4) , wherein the batch of material (Bl) is mixed in a subsequent production entity (E2n, E3n, E4n) of at least one further production stage (BS2, BS3, BS4) of a production line (10) with other batches of material (B12, B13) produced in other production entities (E12, E13) of the same production stage (PSI) , comprising the steps:

- collecting (S10) actual sensor data (xact) measured at each production entity (Ell, E12, E13, E2n, E3n, E4n) of the production stages (PSI, PS2, PS3, PS4) and a time stamp providing a data collection time at which the sensor data (xact) was collected,

- inputting (Sil) the actual sensor data (xact) into a trained ML model (Mt) , which was trained according to claim 1 to 9, wherein the trained ML model (Mt) is generated by jointly optimizing predicted batch quality labels (yi) of the ML model and the weight matrix (W) with respect to the collected lab samples (Sj) , and

- outputting (S12) predicted batch quality values (y^) for each of the batches of material (Bll, B12, B13, B2n, B3n, B4n) produced at the batch production entities (Ell, E12, E13, E2n, E3n, E4n) .

11. Computer-implemented method according to claim 10, comprising : - generating (S13) control instructions based on the predicted batch quality values (y and input into the production line (10) .

12. Assistance apparatus for generating a trained ML model predicting a batch quality value (y^) for a batch of material (Bll, B12, B13, B2n, B3n, B4n) produced in a production entity (Ell, E12, E13, E2n, E3n, E4n) of a production stage (PSI, PS2, PS3, PS4) , wherein the batch of material (Bll, B12, B13, B2n, B3n, B4n) is mixed in a subsequent production entity (E2n, E3n, E4n) of at least one further production stage (PS2, PS3, PS4) of a production line (10) with other batches of material (B12, B13) produced in other production entities (E12, E13) of the same production stage (PSI) , comprising at least one processor, configured to perform the steps:

- collecting sensor data ( j) measured at each batch production entity (Ell, E12, E13, E2n, E3n, E4n) of the production stages (PSI, PS2, PS3, PS4) and a time stamp providing a data collection time at which the sensor data was collected,

- collecting a multiplicity of lab samples (Sj) , each lab sample (Sj) consisting of a set of products output at the production entities (E4n) of a last production stage (PS4) of the production line (10) , wherein the lab sample (Sj) provides a quality measure of the batch of products, and a time stamp providing a lab sample time at which the lab sample

(Si) was collected,

- establishing a weight matrix (W) comprising weight coefficients of non-zero value for each of the batches of material which contribute to the lab samples (S ) ,

- establishing a ML model which receives the collected sensor data (xi) as input and outputs the predicted quality value (yi) for each of the batch of material (Bll, B12, B13, B2n, B3n, B4n) , and - generating a trained ML model (Mt) by jointly optimizing predicted batch quality labels (y of the ML model and the weight matrix with respect to the collected lab samples (S ) .

13. Monitoring apparatus for predicting a batch quality value ( i) for a batch of material (Bll, B12, B13, B2n, B3n, B4n) produced in a production entity (Ell, E12, E13, E2n, E3n, E4n) of a production stage (PSI, PS2, PS3, PS4) , wherein the batch of material (Bl) is mixed in a subsequent production entity (E2n, E3n, E4n) of at least one further production stage (BS2, BS3, BS4) of a production line (10) with other batches of material (B12, B13) produced in other production entities (E12, E13) of the same production stage (PSI) , comprising at least one processor, configured to perform the steps :

- collecting actual sensor data (xact) measured at each production entity (Ell, E12, E13, E2n, E3n, E4n) of the production stages (PSI, PS2, PS3, PS4) and a time stamp providing a data collection time at which the sensor data was collected,

- inputting the actual sensor data (xact) into a trained ML model (Mt) , which was trained according to claim 1 to 9, wherein the trained ML model (Mt) is generated by jointly optimizing predicted batch quality labels (y^) of the ML model and the weight matrix (W) with respect to the collected lab samples (Sj) , and

- outputting predicted batch quality values (y^) for each of the batches of material (Bll, B12, B13, B2n, B3n, B4n) produced at the production entities (Ell, E12, E13, E2n, E3n, E4n) .

14. A computer program product directly loadable into the internal memory of a digital computer, comprising software code portions for performing the steps of claims 1-9 when said product is run on said digital computer. 15 . A computer program product directly loadable into the internal memory of a digital computer, comprising software code portions for performing the steps of claims 10 and 11 when said product is run on said digital computer .

Description:
Description

Predicting a batch quality value by a Machine Learning Model for a dedicated batch of material in a production line

The present disclosure relates to an assistance apparatus and a computer-implemented method for generating a trained ML model predicting a batch quality value for a dedicated batch of material produced in a dedicated production entity of a production stage, wherein the dedicated batch of material is mixed in a subsequent production entity of at least one further production stage of a production line with other batches of material produced in other production entities of the same production stage .

In process plants materials flow through production units continuously or they are treated batchwise in reactors and undergo certain processing steps , such as chemical reactions , heating, cooling, drying, or other mechanical or chemical treatments . Examples include chemical and pharmaceutical production processes , oil and gas processing, and food and beverage production . Monitoring of processes and data-driven product quality prediction is a central task that enables more complex tasks such as production monitoring and optimization, predictive maintenance of the production line, and identifying, analysing and mitigating anomalous production .

Quality control for production processes is commonly performed at the end of the production where product samples are taken at certain points in time . A central challenge consists of attributing the results of these quality measurements to specific batches of material that were produced prior to the specific sample being taken and were mixed later along the production line . Assigning quality labels to specific batches of material is necessary for being able to treat different production units such as multiple chemical reactors producing material batches in parallel separately for monitoring, optimization, maintenance, etc . In summary, mixing and unknown time delays obscure a direct relationship between lab samples and produced batches .

Therefore, it is the obj ect of the present application to provide an assistance apparatus and method that enables a prediction of end-quality per production stage and not only at the end of the whole production where the results of intermediate production stages are mixed .

This obj ect is solved by the features of the independent claims . The dependent claims contain further developments of the invention .

A first aspect concerns a computer-implemented method for generating a trained ML model predicting a batch quality value for a batch of material produced in a production entity of a production stage, wherein the batch of material is mixed in a subsequent production entity of at least one further production stage of a production line with other batches of material produced in other production entities of the same production stage, comprising the steps ,

- collecting sensor data measured at each production entity of the production stages and a time stamp providing a data collection time at which the sensor data was collected,

- collecting at least one lab sample consisting of a set of products output at the batch entities of a last production stage of the production line, wherein the lab sample provides a quality measure of the batch of products , and a time stamp providing a lab sample time at which the lab sample was collected,

- establishing a weight matrix comprising weight coefficients of non-zero value for each of the batches of material which contribute to the lab samples ,

- establishing a ML model which receives the collected sensor data as input and outputs the predicted quality value for each of the batch of material , and - generating a trained ML model by j ointly optimizing predicted batch quality labels of the ML model and the weight matrix with respect to the collected lab samples .

The weight matrix provides the information how a certain lab sample is influenced by different produced batches of material . The ML model provides a suitable relationship between the lab samples and the quality of the batch of material produced in a production entity by j ointly optimizing the ML model and the weight matrix and allows to conclude on the quality of each production entities in the production line . The method learns optimal mixing coefficients together with the ML model .

In an embodiment of the method, the collected sensor data is provided by a real-valued vector for one specific point in time or a time-series of real-valued vectors .

This enables to use all available sensor data and lab samples by the method to infer the batch quality labels and mixing weights as accurately as possible .

In an embodiment of the method, those batch entities , which contribute to the lap sample, are determined from a temporal relationship between the lab sample time of the lap sample and the data collection time of the sensor data measured at each production entity of the production stages .

This allows to relate potential production entities with the considered lap sample and thus allows to conclude from the lap sample to the quality of the batch of material produced by production entities and in consequence the condition of the production entities as such . In contrast, if lab samples would only be associated with windows of this time series rather than batches of material and production entities respectively, the ML model does not allow for treating different production entities , such as multiple chemical reactors pro- ducing batches of material in parallel , separately for tasks such as monitoring, optimization, maintenance, etc .

In an embodiment of the method, the optimization is performed by minimizing a mean square error (MSE) between mixed predicted batch quality labels of the ML model and the collected lab samples .

The mixed predicted batch quality labels are provided by the matrix-vector product between the weight matrix and the predicted batch quality labels . The MSE has the advantage that it penalizes strong deviations , i . e . , batch quality labels with low deviations are prefered . Further it is differentiable which allows its usage in different ML approaches , especially in neural networks .

In an embodiment of the method, a regularization loss of the weight matrix (W) is added, while the production entities have fixed batch quality labels .

Introducing the regularization loss of the weight matrix provides a unique solution of the optimization task . In this case the batch quality labels have a fixed value .

In an embodiment of the method, the regularization loss is a L2-regularization which provides uniform weight distributions .

Applying a L2-regularization favors more uniform weight distributions . This embodiment should be applied if it can be assumed that the batches of material contribute uniformly to the lab sample .

In an embodiment of the method, the regularization loss is similarity based, preferably a gradient-regularization, which focusses on batch quality labels most similar to the lab sample . The gradient-regularization is applied if it is assumed that the batches of material having batch quality labels most similar to the lap sample contribute most to the lap sample .

In an embodiment of the method, the optimization applies at least the same number of lab samples as the number of batches of material existing in the production line, and the weight matrix W is full rank, and the weight coefficients are fixed .

This provides a unique solution of the optimization when it is assumed that the weights in the weighting matrix are fixed .

A second aspect concerns a computer-implemented method for predicting a batch quality value for a batch of material produced in a production entity of a production stage, wherein the batch of material is mixed in a subsequent production entity of at least one further production stage of a production line with other batches of material produced in other production entities of the same production stage, comprising the steps :

- collecting actual sensor data measured at each production entity of the production stages and a time stamp providing a data collection time at which the sensor data was collected,

- inputting the actual sensor data into a trained ML model , which was trained according to claim 1 to 9 , and

- outputting predicted batch quality values of the batches of material produced at production entity .

This method provides batch quality values for a batch of material produced in a respective production entity without taking (lab) samples of the set of products output at the batch entities of a last production stage of the production line . The batch quality values provide information which is necessary to treat different production units separately for monitoring, optimization, maintenance, etc . In an embodiment of the method, the method comprises the step of generating control instructions based on the predicted batch quality values and input into the production line .

This enables to control the production entities of the production line based on the output of the trained ML model processing actual sensor data of the production entities .

A third aspect concerns an assistance apparatus for generating a trained ML model predicting a batch quality value for a batch of material produced in a production entity of a production stage, wherein the batch of material is mixed in a subsequent production entity of at least one further production stage of a production line with other batches of material produced in other production entities of the same production stage, comprising at least one processor, configured to perform the steps :

- collecting sensor data measured at each production entity of the production stages and a time stamp providing a data collection time at which the sensor data was collected,

- collecting a multiplicity of lab samples , each lab sample consisting of a set of products output at the batch entities of a last production stage of the production line, wherein the lab sample provides a quality measure of the batch of products , and a time stamp providing a lab sample time at which the lab sample was collected,

- establishing a weight matrix comprising weight coefficients of non-zero value for each of the batches of mate-rial which contribute to the lab samples ,

- establishing a ML model which receives the collected sensor data ( j) as input and outputs the predicted quality value for each of the batch of material , and

- generating a trained ML model by j ointly optimizing predicted batch quality labels of the ML model and the weight matrix with respect to the collected lab samples . The assistance apparatus provides the trained ML model based on historical sensor data and lap sample of the production line .

A fourth aspect concerns a monitoring apparatus for predicting a batch quality value for a batch of material produced in a production entity of a production stage, wherein the batch of material is mixed in a subsequent production entity of at least one further production stage of a production line with other batches of material produced in other production entities of the same production stage, comprising at least one processor, configured to perform the steps

- collecting actual sensor data measured at each production entity of the production stages and a time stamp providing a data collection time at which the sensor data was collected,

- inputting the actual sensor data into a trained ML model , which was trained according to the method described above, and

- outputting predicted batch quality values of the batches of material produced at production entity .

A fifth aspect concerns a computer program product directly loadable into the internal memory of a digital computer, comprising software code portions for performing the steps of the method for generating a trained ML model , as described before, when said product is run on said digital computer .

A sixth aspect concerns a computer program product directly loadable into the internal memory of a digital computer, comprising software code portions for performing the steps of the method for predicting a batch quality value, as described before, when said product is run on said digital computer .

The invention will be explained in more detail by reference to accompanying figures . Similar obj ects will be marked by the same reference signs . Figure 1 schematically illustrates an example of a production line with production entities in several production stages .

Figure 2 illustrates an embodiment of the inventive method for generating a trained ML model as a flow chart .

Figure 3 schematically illustrates a determination of the weight coefficients .

Figure 4 illustrates various options for selecting possible batch quality labels during generation of an embodiment of a trained ML model .

Figure 5 illustrates the determination of a batch quality values in an embodiment of the trained ML model .

Figure 6 schematically illustrates an embodiment of the inventive assistance apparatus .

Figure 7 illustrates , as a flow chart, an embodiment of the inventive method for predicting a batch quality value in a production line applying the trained ML model .

Figure 8 schematically illustrates an embodiment of the inventive monitoring apparatus .

It is noted that in the following detailed description of embodiments , the accompanying drawings are only schematic, and the illustrated elements are not necessarily shown to scale . Rather, the drawings are intended to illustrate functions and the co-operation of components . Here, it is to be understood that any connection or coupling of functional blocks , devices , components or other physical or functional elements could also be implemented by an indirect connection or coupling, e . g . , via one or more intermediate elements . A connection or a coupling of elements or components or nodes can for example be implemented by a wire-based, a wireless connection and/or a combination of a wire-based and a wireless connection . Functional units can be implemented by dedicated hardware, e . g . , processor, firmware or by software, and/or by a combination of dedicated hardware and firmware and software . It is further noted that each functional unit described for an apparatus can perform a functional step of the related method and vice versa .

In process plants materials flow through production units continuously or they are treated batchwise in reactors and undergo certain processing stages , such as chemical reactions , heating, cooling, drying, or other mechanical or chemical treatments . Quality control for such production processes is commonly performed at the end of the production where product samples are taken at certain points in time .

A central challenge consists of attributing the results of these quality measurements to specific batches of material that were produced prior to the specific sample being taken and were mixed later along the production . Assigning quality values/labels , further called batch quality labels , to specific batches of material is necessary for being able to treat different production entities , such as multiple chemical reactors producing batches of material in parallel , separately for monitoring, optimization, maintenance, and further . I . e . , there is a strong need to provide a prediction of end-quality per production stage and not only at the end of the whole production line, where the results of intermediate production stages are mixed .

Figure 1 shows an example process executed in production line 10 of a process plant . The batch production line 10 consists of various production stages PS1-PS4 , each production stage PSI , PS2 , PS3 , PS4 comprising at least one but mostly several production entities E11-E14 , E2n, E3n, E4n . The main batch process , e . g . , batch production stage PSI , runs in parallel but not synchronized in production entities Ell, E12, E13. In each of the production entities Ell, E12, E13 a mixture of material Cll, C12, C13 is process into batches of material Bll, B12, B13. Storage tank Ml mixes all produced batches of material Bll, B12, B13 produced in production entities Ell- E13. Also, production stage PS2 operates in batch mode. Process stages PS3 and PS4 are continuous processes. A silo M3 again leads to a mixture of batches especially by delivering material to production entities E4n. Quality relevant lab samples Sn are taken prior to the silo M3 and after the filtering, indicated by diamonds .

In consequence, mixing and unknown time delays obscure a direct relationship between lab samples Sn and produced batches of material Bll, B12, B13 and the production entities Ell, E12, E13 respectively.

One possible approach would be to not consider individual production batches, i.e., batches of material Bln, B2n, B3n, B4, and the respective production entities, but rather represent the whole production line 10 as a single time series. Then, lab samples Sn would be associated with windows of this time series rather than batches. While this approach might save data pre-processing effort compared to extracting batch data, it does not allow for treating different production entities, e.g., Ell, E12, E13 producing material batches in parallel, separately for tasks such as monitoring, optimization, maintenance, etc. Also, the task of matching the individual time series window labels to the lab samples would remain .

Another approach would be to first estimate fixed mixing weights, e.g., assuming uniform weights or time-based weighting, and then train a regression model and predict batch quality labels. However, the estimated mixing coefficients might not correspond to the actual weights and could potentially be misleading and lead to wrong batch quality labels .

The method disclosed in the following generates a trained machine learning (ML) model , which provides batch quality labels for each of the batches of material Bln, B2n, B3n, B4n from which a quality value/label of the respective production entities Eln, E2n, E3n, E4n can be derived based on sensor data measured by sensors S located at the production entities Ell , E12 , E13 , and at each of the production entities E2n, E3n, E4n for which batch quality values shall be predicted by the trained ML model . The resulting trained ML model can predict batch quality values for the same production entities of the same production line based on actually collected sensor data collected at the production entities .

The main steps of the method are illustrated in Fig . 2 and described in more detail below relating to the production line shown in Fig . 1 . To generate a trained ML model which provides a prediction of a batch quality value/label for a batch of material Bll , B12 , B13 , B2n, B3n, B4n produced in production entities Ell , E12 , E13 , E2n, E3n, E4n sensor data are collected which are measured at each production entity of the production stages . Further a time stamp is collected for each of the collected sensor data providing a data collection time at which the sensor data was collected, see step SI .

For production entities Ell , E12 , E13 , E2n, E3n, E4n which do not provide sensor data, there will be no quality value label evaluated . The sensor data represent production conditions and settings of the production entity . The collected sensor data is structured in a real-valued vector for one specific point in time or a time-series of real-valued vectors .

In step S2 , a multiplicity of lab samples Sn is collected, each lab sample Sn consisting of a set of products output at the batch entities E4n of a last production stage PS4 of the production line 10 , wherein the lab sample Sn provides a quality measure of the batch of products . Further, to each of the collected lab sample Sn a time stamp is collected providing a lab sample time at which the lab sample (Sf ) was collected .

A weight matrix comprising weight coefficients of non-zero value is established for each of the batches of material Bll , B12 , B13 , B2n, B3n, B4n which contribute to the respective lab samples Sn, see step S3 . Those batch entities , which contribute to the lap sample, are determined from a temporal relationship between the lab sample time of the lap sample and the data collection time of the sensor data measured at each production entity of the production stages .

Further a ML model is established, see step S4 , which receives the collected sensor data as input and outputs the predicted quality value for each of the batch of material Bll , B12 , B13 , B2n, B3n, B4n . The ML model is a neural network, comprising an encoder structured as a Multilayer perceptron MLP, a convolutional neural net CNN, or a Long shortterm memory LSTM .

A trained ML model is generated, see step S5, by j ointly optimizing predicted batch quality labels of the ML model and the weight matrix with respect to the collected lab samples . The optimization is performed by minimizing a mean square error between mixed predicted batch quality labels of the ML model and the collected lab samples Sn .

The technical problem is formalized as follows . A set of batches of material bi is given by where Xi denotes a feature representation, i . e . , measured sensor data representing production conditions of the batch of material . y t G IR. denotes the true quality of the batch of material that is unknown and needs to be determined . The time stamp providing the data collection time is known, at which production of the batch material, e.g., Bll, completed in the production entity, here Ell, and it was forwarded in the production line 10 and potentially mixed with other batches of material, e.g., B12, B13. The feature representation, i.e., the collected sensor data optionally in a pre-processed form, can either be a real-valued vector or a time series of real- valued vectors, depending on the available data and contains information on setpoints configured at the production entity and actual sensor measurements taken during production of the batch .

It is assumed that there exists a functional relationship f that allows for predicting the batch quality y t from the batch feature representations, i.e., the sensor data Xt to a certain degree. Further, lab samples S( G IR. are taken at certain points in time. We assume that each lab sample St is a convex combination of a set of potentially relevant batches of material bj for that lab sample, i.e., s< = / W l‘ y l where w tJ > 0 for all l,j and for all i.

Given the batch quality label vector y and the lab sample vector s, this results in matrix form as s = W y , where W G R nxm denotes the mixing coefficient of a weight matrix. We assume weight matrix IV to be sparse, i.e., not all batches of material Bll, B2i, B3i, B4i, the index i indicating any of the batches of material of production stages PS2, PS3, PS4, potentially related to all lab samples s { . While it is assumed that the sparsity structure of IV to be given we assume that the non-zero mixing coefficients are unknown. The non-zero mixing coefficients are, e.g., derived from the temporal difference between the time stamp at which the processing of the batch of material in the production entity completed, i.e., the data collection time and the time stamp at which the lab sample was taken, i.e., the lap sample time.

The sparsity structure of weight matrix W can be represented as a bipartite graph with adjacency matrix

A e {0, l} nxm and W o A = W where the left-hand set of nodes correspond to the batches and the right-hand set of nodes corresponds to the lab samples. The problem is illustrated in Fig. 3. The task is to derive the batch quality labels y given the batch features representation X, the adjacency matrix A and the lab samples s. The weight matrix W and the weight coefficients W are unknown .

The main idea of our solution is to learn the regression function and the mixing coefficients jointly in an end-to-end fashion. The major advantage of this approach is that we can integrate all available information (batch features, adjacency matrix, lab samples) and thus learn more appropriate mixing coefficients that are better compatible with the batch features compared to estimating a fixed weight matrix beforehand. We formulate our general solution as the following optimization problem :

The obj ective function is the same as above, i . e . , the mean squared error (MSE) between the mixed predicted batch quality labels and the true lab samples is minimized . However, instead of assuming a fixed given weight matrix W, we optimize W together with parameters 0 of the ML model , especially a regression model f . Constraints are added to ensure that W is stochastic, i . e . , that the lab sample is a convex combination of the batches contributing to the lab sample, and a constraint ensuring that W has the same sparsity structure as A where ° denotes the Hadamard product .

In general , the above optimization problem does not have a unique solution . Given fixed batch quality labels y, the optimization problem w . r . t . W, does in general not allow for a unique solution . Note that this problem can be solved independently for each lab sample St, since the batch quality labels y are fixed . For each St, a unique solution exists if at most two batches of material are connected to St and if these batches of material have unique batch quality labels. Otherwise, infinitely many possible convex combinations of the batch quality labels could lead to the same lab sample value.

An illustration is provided in Fig. 4. At the left hand, see part A of Fig. 4, a lab sample Si is shown with connected batch quality labels. The horizontal axis denoted the quality value. Two batch quality labels 20, 21 are smaller than the lab sample S( and two batch quality labels 22, 23 are larger than the lab sample S(. The lab sample value can be expressed as infinitely many possible convex combinations of the smaller 20, 21,... and larger batch quality labels 22, 23,.... Ambiguities could be resolved by favouring specific combinations.

Part B of Fig. 4 shows a solution where a most uniform weight distribution is favoured. Here the batch quality labels 20, 21, 22, 23 have a non-zero weight. Part C of Fig.4 shows a combination where only the batch quality labels 21, 22 which are most similar to the lab sample St are considered, i.e., only the two batch quality labels closest to the lab sample Sf have a non-zero weight.

These ambiguities are resolved by introducing a regularization loss on W. One possible variant is L2-regularization, which favors more uniform weight distributions, see part B in Fig.4 :

Note that the feasible set of solutions, i.e., the standard simplex, is closed and convex and that the loss function is strictly convex, i.e., the Hessian yy T + l is positive definite. Thus, the solution is unique. Another possibility is to focus instead on the batch quality labels most similar to the lab sample, see part C of Fig . 4 : w tj > 0 Vy,

If a weight is Wjy = 0, the regularization term for yy is also zero . If a weight w^y > 0, the squared distance between Sj and yy needs to be minimal . Thus , given unique batch quality labels for Si, the solution will uniquely focus on the batch quality labels most similar to the lab sample . Gradients are not propagated from the regularizer through the regression model to avoid potential side effects .

For fixed weights W, the optimal batch quality labels are in general also not unique . A unique solution exists if there are at least as many lab samples Sj as batches of material and if the weight matrix W is full rank . These conditions can in general not be satisfied easily, e . g . , if lab samples are not taken regularly enough . Further, the regression model for predicting will usually be non-convex . However, our approach of learning to predict the batch quality labels y t j ointly with the mixing coefficients Wp potentially allows for resolving ambiguities based on the given batch features Xt and learned regression function f . An example is illustrated in Figure 4 . Without further knowledge, infinitely many solutions (y l f y 2 ) exist, e . g . , ( 100 , 100 ) , (200 , 0 ) , or ( 0 , 200 ) , see left hand part of Fig . 5 , where the contributing batch quality labels y l f y 2 contribute each by 50% . With additional knowledge about the batch features and x 2 and assuming a relatively simple functional relationship f between x and y, a specific solution becomes more likely than others , see right hand part of Fig . 5 . In this example, identical batch features indicate that both batches of material should have the same label .

That is , the solution (y l f y 2 ) = ( 100 , 100 ) is more likely than other possible solutions .

Fig . 6 shows an embodiment of the inventive assistance apparatus . Assistance apparatus 30 consists at least one processor which is configured to perform the steps as described before . Especially the assistance apparatus 30 comprises a data interface unit 31 , a training unit 32 and an output interface unit 33 . The data interface unit 31 is configured to receive the measured sensor data x and the collected lab samples S true . The output interface is configured to output the batch quality labels relating to the input the measured sensor data x and the collected lab samples S true and/or a trained ML model Mt .

The training unit 32 is configured to provide a model architecture to train the ML model Mt . The parameter / components learned are marked with bold lines in training unit 32 . These are a feature extraction network Enc and a dense layer Den of the ML model Mt and the weight matrix W . The encoder Enc extracts features from the sensor data . The encoder Enc denotes a generic neural network that could be, depending on the structure of the collected sensor data X, e . g . , a Multilayer perceptron MLP, convolutional neural network CNN, or Long short-term memory LSTM module . The dense layer Den predicts the batch quality values depending on the learned features . The constraints on W can be ensured by first applying a softplus function softp to avoid negative weights , multiply- ing elementwise with adj acency matrix A to ensure the sparsity structure and by finally applying row-wise Ll- normalization norm . The model can be trained end-to- end with gradient-based optimization algorithms , such as Adam . A data flow, i . e . , a forward pass , is shown training unit 32 in indicated by solid arrows . The gradient flow, i . e . , a backward pass , is shown training unit 32 in indicated by dashed arrows .

An embodiment of the method for predicting a batch quality value for a batch of material produced in a production entity of a production stage in a production line is described with flow chart in Fig . 7 .

In a first step S10 , actual sensor data measured are collected at each production entity of the production stages . A time stamp providing a data collection time at which the sensor data is collected preferably together with the sensor data . These actual sensor data are input into a trained ML model , which was trained as described before, see step Sil . Predicted batch quality values of the batches of material produced at a production entity are output based on the collected sensor data and the ML method trained for this production line, see step S12 . Optionally, control instructions , e . g . , settings of any of the production entities involved in the production or even changing the production entities involved in the production line, are generated based on the predicted batch quality values and input into the batch production, see step S13 . Thus , batch quality values are output without taking any lab samples from the series of products actually produced at the end of the production line .

Fig . 8 shows an embodiment of a monitoring apparatus 50 applied in a scenario with an industrial process plant 40 . The monitoring apparatus comprises at least one processor configured as a data interface unit 51 , a predicting unit 52 and an output interface 53 . The data interface unit 51 is configured to receive measured sensor data x act , also named feature rep- resentations of the batch of material comprising not only information on the material but also on the production entity which is processing the batch of material . The prediction unit 52 is configured to perform the trained ML model generated for the monitored production line in the process plant 40 . The output unit interface 53 provides the process plant and/or comprises a user interface, e . g . , a graphical user interface, to present the data to a user, e . g . , operating personnel .

In batch production, a more straightforward approach which is trying to map a specific lab sample to a specific batch is not feasible due to the complex material flow . The presented generation of a ML method solves the problem of predicting the batch quality under the constraint of a many- to-many relationship between lab samples and batch production . This is achieved by learning the regression function together with the mixing weights in an end-to-end fashion .

The ML model considers individual production batches to allow for a localized analysis . It allows for inferring distinctive and highly accurate batch quality labels . The assumptions on the mixing weights are as minimal as possible . This is important since in real-world use cases with complex industrial processes the information about the process itself and the uncertainty related to it is usually very high . All available data, i . e . , batch features and lab samples , should be used by the algorithm to infer the batch quality labels and mixing weights must be inferred from the data as accurately as possible . Any method that does not consider this requirement leads to potentially seriously biased results .

It is to be understood that the above description of examples is intended to be illustrative and that the illustrated components are susceptible to various modifications . For example, the illustrated concepts could be applied for different technical systems and especially for different sub-types of the respective technical system with only minor adaptions .