Soft Sensor Development for Hydrocracker Product Quality Prediction Paula Barbosa, Carla Pinheiro, Dora Nogueira The main goal of this work is to maximize the productive capacity, and the revenue from each oil barrel, GalpEnergia Sines Refinery has invested in an hydrocracking unit. Given that all fuels are subject to strict regulation, it is necessary to have tight control over their quality. Therefore, in order to implement future advanced control on the unit, we proceeded to a first approach of the prediction of a quality variable of the diesel produced by making use of a soft- sensor. To develop the soft sensor for quality prediction, variables of interest and their historical data were collected and analyzed. Step-tests were performed in the real industrial plant in order to better understand the dynamic behaviour of the fractionator. Subsequently, four soft sensors were developed using Principal Components Analysis followed by a Partial Least Squares regression to obtain linear models able of quality prediction. The soft-sensors developed were also good detectors of process faults because they included the faulty variables for prediction. Keywords: PCA, PLS, Multivariate Analysis, Hydrocracking, Quality prediction, Soft-Sensor. 1. Introduction [1,2] better way , providing real-time necessary for effective quality control Chemical plants are usually [3] information . highly instrumented and have a large number of sensors The span of tasks performed by Soft-Sensors that collect measured data for process control and is quite broad but the most common use is the monitoring. About two decades ago researchers prediction of process variables that can only be began using the large amount of data to build known either at low sampling rates or through off-line predictive models, and these models are called Soft- analysis Sensors. The term soft sensor is a combination of important for process control because they are the words ‘software’ (mainly because models are usually related to the process output quality and it is developed in computer programs) and ‘sensors’, naturally because additional information about these variables at higher these models are providing similar [1,4] . These important variables and are usually necessary to [1,5] very deliver information as hardware sensors. These soft-sensors sampling rate or lower financial burden are often divided into two categories: model-driven field of application of soft-sensors is of process and data-driven [1,2] . Another . Model-driven sensors (also monitoring and process fault detection by finding the called white-box models) are most commonly based state of the process and identification of the deviation on First Principle Models that describe the physical source. and chemical properties of the process[9,10], are developed primarily for the planning of the plants and usually only describe ideal process steady-states and not real process dynamics, describing a simplified theoretical background rather than real-life process conditions measured [1] . Data-driven models are based on data within the processing plants, thus describing the true conditions of the process in a Measuring variables that define product quality is a major problem in process industries. These variables are called primary or quality variables and quantify the productivity or the specifications upon which the product is sold, like purity or physical or chemical properties, and these are the most difficult to measure online . The online variables that are easy to access and measure are often called secondary variables and can be temperature, pressure and flow rate and can be used pure PCA is that it can only effectively handle linear to infer primary variables. Because of the nature of relationships of the data and cannot deal with data chemical and processing engineering systems, the non-linearity. Another disadvantage is the selection dynamics and state of the secondary variables of the optimal number of principal components (that reflects the dynamics and state of the primary can variables, meaning that changes in secondary techniques). Another problem is that the principal variables are indicative of changes in product quality. components describe very well the input space but The technique of using secondary variables to do not explain the relations between the input and generate estimates of product quality is usually output data, that is usually what has to modelled . be addressed by using cross-validation [5] called ‘soft-sensing’ and these inferential estimators 2.2 Partial Least Squares Regression are usually in place of direct on-line measurement of controlled variables if direct measurements are (PLS) The regression problem, that is, the modelling [5] expensive, unreliable or add large lag . of primary variables, by means of a set of predictor 2. Data-driven methods for soft-sensing variables, the secondary variables, is one of the most The product quality in Hydrocracking units is common problems in data-analysis in science and significantly influenced by operating conditions and technology, and one example of such problems may the cracking yield is reduced with time by catalyst include deactivation. Therefore, the continuous monitoring of manufactured products to the conditions of the product quality is very important especially to avoid manufacturing process .PLS can be seen as an off-spec petroleum fractions, that usually cause extension of PCA. The algorithm pays attention to [2] relating the quality and quantity of [8] covariance matrix that brings together the input and problems downstream at the blending stage . output data space. This method decomposes the It is usually difficult to get precise and reliable product composition measurements without time delay because most composition analysers have significant time lags and their reliability is usually input and output simultaneously while keeping the orthogonality constraint, having the model focussed on the relation between the input and output [1] variables . quite low. Due to the strong correlation between secondary variables, Principal Components Analysis (PCA) or Partial Least Squares (PLS) methods will be applied [6]. The objective of this method is to fit a linear relationship between the dependent secondary variables and independent primary variables. PLS is a simple and powerful approach for data-analysis for 2.1 Principal Components Analysis (PCA) Noise can be found in almost all variables of the majority of datasets. Latent variable models like PCA and PLS estimate the relevant part and the noise of each variable and therefore are used in the [7] complex problems because of its flexibility and ability to deal with incomplete and noisy data with multiple variables and observations (measurements). The disadvantage of PLS is that like PCA, it can only model linear relations between the data [1] . present work . Principal Component Analysis (PCA) was used for analysing the data so that only the secondary variables important to the determination of [5] 3. Implementation of Step Tests Performing step tests in the unit was of great product quality were selected . Using PCA, the data importance because the unit is new and had never can be described using far fewer variables than the been submitted to step tests. Therefore these tests original of were planned and performed, in order to better information, and also, PCA often produces linear understand its response and behaviour. By better combinations of variables that are useful predictors understanding the process performance, and by variables with [4] no significant loss of particular processes . One of the limitations of submitting the unit to step tests, we hoped to use these data to develop a soft-sensor that could of 3%, 5%, 7%, 10%, 13% and 15% on each of the explain and predict the behaviour of the quality chosen secondary variables. variable even in the case of the unit operating out of 3.3 Step Tests Results the specified operating temperature, pressure and The scheduling and the sequence of the flow values. variables testing was organized to accommodate the 3.1 Historical Data Analysis and Variable Selection Refinery conveniences, in order to minimize the impact into the production profile and quality. For Since step tests had never been performed in each particular test a sample was taken and the time this unit the first approach was to select the variables stamp of the sample was annotated. Samples were that could be tested. This first step included the study only collected after the calculations using the of the fractionator together with the insight and preliminary model, indicated that the quality variable experience of the Refinery Team and the Thesis Y had stabilized after a given step test. As expected, Supervisors, and after some exchange of ideas and the step tests planned had a clear effect on the suggestions it was agreed that the secondary quality variable (Fig.1). Having the quality variable variables X3, X8, X9, X10, X12, X13 and X22 were to covering a wide range of values allows the dynamic be tested. The next step was to build a preliminary data to accommodate the influence of a wider range model using the historical data available using PCA of process conditions on the quality variable. Most of followed by PLS (obtained in a similar fashion as the variables tested end up appearing in the models described in chapter 4). This model was to be used developed in the next chapter. Variables X3, X10, only for assessing if a given step test would indeed X13, X8 and X22 appear in Model B, variables X3 be expected to influence the quality variable Y, and if and X13 appear in Model C, and variable X13 it did, how long would it take the quality variable to appears in Model A. stabilize. Then we looked into the historical data and 4. checked if there were disturbances in the secondary variables selected previously that could In this section the quality variable Y is be considered as ‘step-test’ (like a sudden decline or increase of rate or temperature). Using those ‘steptest’ values, we calculated the Y results, and evaluated and estimated the quality variable settling time for each step test. The sequence for the variables testing was agreed with the Refinery Team in order to reduce the overall impact in the operating conditions. Soft Sensor Development predicted using 25 online secondary variables available in the database. These variables include flowmeters, temperature and pressure sensors and all are online measured variables. The selection of which variables should be included in the soft-sensor is a complex task and the strategy consists in finding a good variable subset capable of making accurate predictions. In this study two methods are used to obtain the soft-sensors to predict the quality variable: 3.2 Sensitivity Analysis Partial Least Squares (PLS) as a linear modelling To evaluate the impact that the tests could tool, and Principal Component Analysis (PCA) as a have on the quality variable Y, the fractionator was tool to select a good model variable set and to strip modelled using the simulator Petro-SIM™ version down the models from outliers and noise. Datasets 4.1. After modelling the unit in Petro-SIM™, the were collected directly from the Digital Control amplitudes of the step tests were tested to access System (DCS) and the Real Time Database (RTDB) their influence in Y. Based on previous tests of the Refinery and were used to build four models. performed in other units of the Refinery, it was The soft-sensors obtained from these data were decided to test the impact of the following steps of - labelled as Model A, B, C and D, and the datasets 3%, -5%, -7%, -10%, -13% and -15%, and if changes are: Pre-tests samples X3 1,035 1,03 1,025 X10 Y 1,02 1,015 X12 1,01 X9 1,005 X13 1 X8 0,995 0,99 X22 0 5 10 15 20 25 Sample number 30 35 40 45 Figure 1: Step-tests results. Model A: The soft-sensor is obtained from could be collected and if the results could be training data collected during the week of the step compared to the other dataset’s results and if we tests, in 2013 from October 27 th st to October 31 , using the same data to calibrate the model. could extract a better soft-sensor from it. The purpose of developing Model D was to access the importance of step tests in the modelling of the Model B: The soft-sensor is obtained from training data collected in 2013 between the August 1 st st and the October 31 , using the same data to calibrate the model. quality variable, therefore this model was developed st using only the historical data from 1 of February to st October the 26 of 2013. These soft-sensors were all st validated with data collected between the 1 and 27 Model C: The soft-sensor is obtained using st training data collected in 2013, from 1 of February of November. The software used in PCA and PLS modelling was PLS Toolbox advanced chemometrics to October the 31 , using the same data to calibrate software the model. environment. st Model D: The soft-sensor is obtained using st training data collected in 2013, from 1 of February st to October the 26 , using the same data to calibrate the model. th used within MATLAB® computational As there isn’t a universally accepted method to obtain the best models for each set of data, each one of the models presented here was obtained by firstly using PCA to select outliers, excluding them from the dataset and immediately using the PLS to The justification for this procedure is based in build the soft-sensor and compare the Variance the fact that we wanted to understand thoroughly the Accounted For (VAF) obtained with the VAF of information the step tests could provide (and of preliminary soft-sensors. After that step, some course extracting the maximum of knowledge of such outliers were selected by fine-tuning the model: if the testing, using it in Model A), and we wanted to model revealed decreased quality, then these data understand what kind of information we could extract samples were not considered as outliers; if not, those from all the data collected in the DCS since the data samples were excluded from the dataset. beginning of operation (Model C). Furthermore, we were aware of the substantial difference of size 4.1 Principal Components Analysis between these two datasets, so it was decided to Choosing the number of principal components build another dataset with an intermediate size has no universal rules or procedures , but a (Model B) to understand what kind of information principal component with an eigenvalue equal or [9] Figure 2: Correlations map for Model A,B.C and D greater than one is believed to be of statistical [10] first eight PC’s have eigenvalues greater than one . After selecting the number of PC’s, the having the first two PC’s captured 44.19% of next step was to detect the presence of outliers. The variance and eight PC’s captured 82% of variance, relevance 2 Hotelling’s T statistic measures the distance of a therefore a PCA model was constructed with that given score sample from the origin and the Q number of principal components. In the case of residuals statistic is the squared projection error and Model A no sample was taken off of the original data, measures the error from a given sample to the the data had no ouliers. For the development of principal component model. So for instance, if after Model B, the sampling interval was increased. calculation of the limits of these two statistics with Instead of analysing only the week of step tests, the significant level of 95%, a sample shows a high value training data from August the 1 in the Q statistic, then this sample does not comply October of 2013 (in a total of 194 samples) was fed with the principal component model. If a sample into the PLS Toolbox, autoscaled and cross-validated 2 st th to the 31 of shows a high value in the Hotelling’s T statistic, it is with venetian blinds, with 10 data splits. After outlier assumed that maybe the unit is not operating at its removal, a PCA model was to be built with 6 PC’s, usual operating rates (with high deviations from the explaining 68,49% of the data variance (having the [11] In the case of a given sample first two PC’s explaining 36,26% of variance). Model showing high values in both Q residuals and C was set up using all the data available since mean values) 2 Hotelling’s T statistics, the principal component st February the 1 to October 31 st of 2013, having a model is not valid for this sample, and it was grand total of 549 samples. As on the previous considered an outlier. model, the data were autoscaled and venetian blinds cross-validated using 40 splits. After outlier removal, Model A was derived based on the 42 laboratory analysis obtained in the week of the steptests performed at Sines Refinery. This dataset was loaded to PLS Toolbox, autoscaled, leave-one-out cross-validated and a PCA model was obtained. The a PCA model with 6PC’s was obtained, having 80,97% of variance explained (having the first two PC’s explained 58,41% of variance). To understand how much the step-tests influenced model Figure 3: Loadings plots for Model A, B, C and D. development and their predictions, Model D was developed based only in data from the 1 st secondary variables. In the loadings plot, the of variables located farther from the origin have a large of October of 2013, impact in the PCA model, and variables located in having a total of 507 samples as its dataset. The the vicinity of the origin have negligible impact. The data were autoscaled and venetian-blinds cross- variables nearest to Diesel D86 show high positive validated using 40 splits. After outlier removal a PCA correlation with it, and the variables farther from model with 5 PC’s was built, having 77,24% of Diesel D86 show great negative correlation with it. variance explained (the first two PC’s explained Geometric interpretation of partial correlation is made 57,22% of variance). using vectors that start at the origin of the plot (Fig. th February of 2013 and the 26 3) and end at each variable point. Correlation After selecting the number of PC’s the next step was to analyse the correlation map and the loadings plot of each model. The correlation maps in Fig. 2 show the pair-wise correlation coefficients between the 25 variables and the quality variable of each model. The intensity of the colours mirror the amount of correlation, so, higher positive correlations show intense shades of red and higher negative correlations show intense shades of blue. The loadings plots on Fig. 3 also show the relations between the quality variable Diesel D86 and the between all the secondary variables and Diesel D86 is obtained using the cosine of the angle between their corresponding vectors. Angles between 0 and 90º correspond to positive correlations and angles between 90º and 180º correspond to negative correlations. Vectors having angles close to 90º correspond to small correlations between their variables and angles closer to 0º or 180º correspond to substantial correlation. Using both the correlations map and the loading plots for each Model, the secondary variables selected to build the soft sensor are in table 1. Table 1: Variables selected Model B: 0,0055 1 1,0277 3 0,0792 2 0,0060 5 0,0048 6 0,0050 7 0,3131 for PLS 8 0,0684 10 0,0039 11 0,6706 13 0,6637 15 0,2244 17 2,0526 18 modelling for each model. 0,2754 20 12,6802 21 10,8107 22 Model A B C D 27,9043 23 22,1807 24 15,6703 25 Variables X1, X2, X5, X7, X13, X14, X15, X17, X21, X23, X24, X25 X1, X3, X2, X4, X6, X7, X8, X10, X11, X13, X15, X17, X18, X20, X21, X22, X23, X24, X25 X3, X2, X11, X13, X14, X15, X17, X18, X19, X20, X23, X24,X25 X2, X11, X13, X14, X17, X18, X19, X20, X23, X24, X25 196,0344 (2) Model C: 2,2609 3 0,0029 2 0,0334 11 0,4001 13 0,3324 14 0,1006 15 0,1226 17 2,7232 18 0,2527 X19 0,1815 X20 34,2095 X23 155,7240 X24 For Model A, variables X4 and X16 were not 10,8638 X25 202,3602 used in the model because they degraded the results (3) Model D: in the model calibration. For Model C, in later stages of model development, it was found that X7 would 0,0118 2 0,0114 11 0,1203 13 deteriorate 0,2316 14 0,4581 17 2,0338 18 model quality and therefore this secondary variable was left out of the model. As for 0,7490 19 2,2141 20 27,5397 23 Model D, X16 would deteriorate model quality and 88,0357 24 224,7459 25 69,4008 (4) was left out of it. 4.3 Model Calibration 4.2 Partial Lest Squares Model The calibration consisted of using equations After variable selection using PCA, the data 1, 2, 3 and 4 to predict each models training data, for each model were loaded to the PLS Toolbox, and the results of model performance are shown in autoscaled and leave-one-out cross-validated for Table 2. Model A, venetian blinds cross validated using 10 splits for Model B and venetian blinds cross-validated using 40 splits for Models C and D. For Model A and B, 10 Table 2: Performance criteria VAF for the modelling results. Latent Variables (LV) were selected for developing the soft model, for Model C, 9 LV’s were selected and finally, for Model D, 7 LV’s were A selected. B The soft sensors obtained were: C Model A: 0,0012 1 0,0386 2 0,1298 5 0,0217 7 1,3417 13 3,2987 14 0,8355 15 0,7202 17 23,3623 21 5,8127 23 71,1457 24 51,0246 25 147,7861 Model (1) D Dataset th st 27 to 31 October st 1 August to st 31 October st 1 February to st 31 October st 1 February to th 26 October Calibration VAF (%) 77,8 Validation VAF (%) 20,4 61,3 56,3 68,1 78,9 17,2 70,9 4.4 Model Validation samples 8 and 9). The range of Y temperatures of The validation consisted of using an unseen st th Model B calibration data was larger than Model A dataset (from the 1 to the 27 of November 2013) and closer to the range of the validation data Y as input data in equations 1, 2, 3, 4 and the results values, enabling Model B to present a better fit than obtained can be seen in the plots of figure 4 and Model A when comparing VAF only, since most of table 2. the predictions fall out of experimental laboratory error. Also, the better fit in validation (when 4.5 Model results discussion comparing VAF) can be explained because Model B The performance criteria Variance Accounted For (VAF) was used to assess the quality of the models by comparing the model’s fit to the real data, as shown in equation 5. includes more secondary online variables than Model A. However Model B shows a clear upper bias in its predictions (Fig 4). Model C was calibrated and validated outperforming all models in capturing the 1 ̂ ! 100 (5) Where "# is the real Y sample laboratory result, "̂# is the model predicted value for Y. The range of the calibration dataset of the Y values is [0,994; 1.031] for Model A, [0.973; 1.034] for Model B, and [0.921; 1.133] for both Models C and D. The range of the validation dataset of Y is [0.941; 1.083] nonlinear data dynamics. The VAF of validation, with a value of 78.9 is considerably higher than the previous three models. Figure 4 show that even not predicting the values of samples 7 to 14, the model follows data dynamics and for all other samples the model makes In the calibration step, Model D provides the worst performance (with a VAF of 17.2), a fact that could not be explained especially because models C and D differ only by two secondary for all models. variables and have a difference of one week of data Model A fit to calibration data has a VAF of (44 samples) for their development. However, Model 77.8, showing that the model could fit the data very D validation shows second best fitting, having a VAF well. As to validation, the model VAF is the worst of of 70.9. Figure 4 shows that the model describes the all the four models (having a VAF of 20.4), because data dynamics but some of the predictions (excluding the range of Y values used for model development is those from the 7 sample to the 14 ) fall out of the narrower than the range of calibration Y data. Also, laboratory analysis error. Model D was developed to the size of the dataset used for model development understand if the step tests data were important in and calibration was only of one week of unit the model development. The results show that operation. The model could not fit data outside the without considering the step tests samples in model values (sometimes even not describing the dynamics D derived in poorer quality predictions compared to th th of the data, as seen in predictions of the 8 and 9 th th Model C predictions. samples – figure 4) used for its development, but data within that range was very well fitted as seen in figure 4: all predictions except those ranging from the 7 th sample to the 14 th sample are very good since most of them fall within the laboratory experimental analysis error. Model B fit to calibration data has a VAF of 61.3, a value smaller than Model A calibration. Model B validation (with a VAF of 56.3) was substantially better than that for Model A validation (VAF of 20.4), especially because the model followed the dynamics of the data as seen in figure 4 (which Model A wouldn’t in the case of 5. Soft-Sensor fault detection During the step of model validation, the first st dataset selected spanned from the 1 of November 2013 to January 13 th of 2014. The four models proposed where tested with this data set and a very clear bias emerged for a specific period (as seen in the left hand plot of Fig. 5), from samples 53 to 135. In this article, we will only show results in fault detection for Model C. The fact was reported back to the Refinery Team in search of an explanation. The Refinery Team advised that an exploratory look Figure 4: Validation results for all Models should be done on the dataset because they too comparison to the historical data. That meant that the could not find an explanation for such evident bias in positive contribution of the variable to de prediction all model’s predictions. The Hydrocracker unit had would indeed an operational problem and the unit was shut predicted values of the quality variable. Having down, but they were fully operating only a few days reported these conclusions to the Refinery Team, we after, and so there was no explanation why the were informed that indeed a pipe had clogged when model’s predictions were systematically biased. the unit was shut down and that affected the Following the Refinery Team advice, the dataset was measuring of X18. substantially decrease, decreasing the thoroughly analysed and it was found that one 6. measured variable, X18, was having values very different of nominal those historically operating measured conditions and substantial deviations from their previous normal operating measurements. Variable X18 was used in Model C. To confirm or discard the suspicion that X18 was effectively influencing the model’s predictions, the model’s predictions for the quality variable were again obtained using the average of the historical past values before the shut down for X18, instead of the measured values. The results obtained for the predicted quality variable are presented in the right hand plot of figure 5, showing that the approach was successful in proving that the measurements of variable X18 were affected by any equipment failure that changed their values to a range different from the normal operating range. In Model C, the variable’s coefficient is positive but the measured values decreased about 20% in The aim of this study was to develop a soft under reported Conclusions sensor capable of predicting the quality variable, Y. Two methods were used to obtain the soft-sensors to predict this quality variable: PLS as a linear modelling tool and PCA as a tool to select a good model variable set and to strip down the models from outliers and noise. Based in both historical and step tests data, four soft-sensors were developed, calibrated with plant data and validated with new data that had not been used during the calibration step. The different datasets used in each model development affected their prediction results but all the models follow the same dynamics behaviour for the quality variable and feature good predictions, but the use of a linear prediction tool in a clearly nonlinear unit is cause of error. Model C seems to be the best choice to be built into the DCS system as an inferential sensor, to provide real time information of the Y prediction to the operators and also to be used Figure 5: Fault detection for Model C for control purposes. Model C was built with most of the historical data available. Having a larger dataset to build the model is of great advantage since it surely covers most of the process conditions and quality variable range. In calibration the model fits better to the real data than Model B and D but worse than Model A. In validation Model C has the best behaviour of the four models, since the dataset used for its development and calibration had a larger Y value range than the dataset used for validation, joining the best validation result with the advantage of a smaller number of model variables. During the validation step of this study, we were able to make use of one of the advantages of soft-sensors: process monitoring and fault detection. By using three models to fit the validation data, we discovered a clear bias between predicted and real values of Diesel D86 during a certain time period. This prompted a careful analysis of the dataset and the detection of faulty measures on two variables measured online in the plant. The soft sensors developed were good detectors of process faults because they included the faulty variables for [3] B. Lin, B. Recke, J.K.H. Knudsen, S.B. Jorgensen. “A systematic approache for soft-sensor development.” Computers and Chemical Engineering. vol 31. pp 419-425. 2007. [4] W. Zong, J. Yu. “Mimo soft-sensors for estimating product quality with on-line correction”. Institution of Chemical Engineers. Trans IChemE. vol 78. Part A. May 2000. [5] S. Abdul-Wahab, A. Elkamel, C.R. Madhuranthakam, M.B. Al-Otaibi. “Building Inferential Estimators for Modelling Product Quality in a Crude Oil Desalting and Dehydration Process”.Chemical Engineering and Processing. vol 45. pp 568-577. 2006. [6] J. Zang. “Off-set inferential feddback control of distillation compositions based on PCR and PLS Models.” Chemical Engineering Technology. vol 29, No 5. 2006. [7] N. Kettaneh, A. Berglund, S. Wold. “PCA and PLS with very large data sets.” Computational Statistics & Data Analysis. vol 48. pp 69-85. 2005. [8] S.Wold, regression: M. a Sjostrom, basic tool L. of Eriksson. “PLS- chemometrics.” Chemometrics and Intelligent Laboratorys Systems. vol 58. pp 109-130. 2001. [1] P. Kadlec, B. Gabrys, S. Strandt. “Data-Driven [9] Barry M. Wise, Neal B. Gallagher, Rasmus Bro, Jeremy M. Shaver, Willem Windig, R. Scott Koch. Chemometrics Tutorial for PLS_Toolbox and Solo. Eigenvector Research, Inc., 3905 West Eaglerock Drive, Wenatchee, WA 98801 USA. [10] R. Gonzalez. “Co-Processing Feedstocks in Soft-sensors in the Process Industry”. Computers single-stage or two-stage hydrocrackers.” Refinery and Chemical Engineering. vol 33. pp 795-814. 2009. Tracker. vol 1. Issue 4, pp. 2. October 15, 2009. [2] H.M.S. Lababidi, D. Chedabeh, M.R. Riazi, A. Al- [11] V.M.Vieira Lopes. “Data-Driven Techniques for Modelling Petrochemical Processes”. PhD. Universidade Técnica de Lisboa – Instituto Superior Técnico. Portugal. 2004. prediction. 7. Bibliography Quattan, H. A. Al-Adwani. “Prediction of product quality for catalytic hydrocracking of vacuum gas oil”. Fuel. vol. 90. pp 719-727. 2011.
© Copyright 2026 Paperzz