1. Summary
Traditional experimentation requires designing an experimental array, stopping production, setting up the equipment/machinery for each run, and running all the runs of the array. This process inevitably results in lost production time, scrap, and the use of extraordinary resources. In many real manufacturing scenarios, this is difficult to do due to the complexity of production demand trade-offs and manufacturing process dynamics. The main idea of Bo
x’s Evolutive Operations [
1], Spendley and Himsworth Simplex [
2], and Nelder and Mea
d’s Variable Simplex [
3] is to find the best combination of process parameters by introducing small changes in the process variables. These small changes allow the experimenter to run the process continuously because the response variable does not move dramatically away from its target, thereby eliminating or minimizing the generation of nonconforming parts (scrap).
These direct search methods (EVOPS, Simplex, Variable Simplex, etc.) have some disadvantages that have limited their use in a manufacturing process:
- They are very sensitive to noise (internal noise).
- The complexity of the computation increases dramatically with the number of variables considered.
- The number of iterations to find a local optimum grows exponentially with the number of variables in the algorithm.
These drawbacks make these methods not very practical to apply to a manufacturing process.
On the other hand, these procedures are based on heuristic/stochastic algorithms, i.e., a series of iterations is developed by introducing small changes in the input variables and evaluating the observed response. The process continues until the response is located on the desired nominal value.
The use of logic algorithms is widely used in the engineering field, from genetic algorithms applied to analyze and improve the location, assignment and routing of transportation systems [
4], to many other applications; these authors documented another few heuristic and stochastic methods and their objectives. It can be seen that there are many designs of algorithms, mainly, they can be classified as
- Deterministic. Those that work with a known function
- Heuristic: Algorithms based on test and error procedures.
- Stochastic: They use successive iterations, taking into account changes in the input variables and measuring the change observed in the output variables, with the help of probabilistic theories.
One of the algorithms that seem to work like the Nelder and Mead algorithm presented in this paper is the Genetic Algorithm, widely used to analyze and improve many different systems. These algorithms evolve a population of individuals by subjecting it to random actions, similar to those that act in biological evolution (mutations and genetic recombination), as well as natural selection. According to some criterion, it is decided which are the most adapted individuals, which survive, and which are the least adapted, which are discarded. An algorithm is a set of organized steps that describe the process to be followed to solve a particular problem, such as an experimental design for optimizing process parameters. There is extensive literature on these algorithm applications, [
5,
6,
7,
8].
The problem with the Genetic Algorithm is that it requires a declared function of the system to be optimized, making it a deterministic algorithm. The direct search methods are better suited to purely heuristic situations where the system function is unknown, such as a production process or machine situations. In these cases, the logical way to improve the operating parameters is to introduce small changes in the input variables and evaluate the changes in the response or responses.
An exhaustive review of the literature shows evidence of the loss of efficiency of these direct search algorithms when the number of dimensions (variables) is increased, as an example, [
9] compares different algorithms (heuristic and deterministic) and their efficiency; our attention is focused on the number of runs required to obtain a local or global optimum. As shown in
Table 1, the number of runs (iterations of the algorithm) increases dramatically when the number of dimensions (variables) is increased (n).
As another example, [
10] compared the NMS algorithm with 5 typical problems.
Table 2 shows the comparison of results under different sample sizes and number of variables for five typical problems; it can be seen that the number of iterations increases dramatically as the number of variables in the geometric array (dimensions) increases.
To give the reader a better idea, these algorithms are essentially different from those used to make decisions with multiple objectives, such as those used as multicriteria. To understand these, you can see [
11,
12,
13] and others.
The algorithm we discuss here is used to find the best combination of parameters for a process, whatever it may be, from a design, a machine, or a manufacturing process. It has to be considered that the increase in the number of iterations for these algorithms is directly related to the number of variables analyzed. If two variables are considered, the geometric array corresponds to a triangle, if three variables are considered, the array corresponds to a square, if four variables are considered, the geometric array corresponds to a pentagon, and so on. As the number of variables increases, so does the number of edges and vertices, which slows down the algorithm and increases the complexity of the computation. In many cases, the high number of variables leads to a non-convergence situation (the algorithm gets “lost”).
The proposed methodology tries to overcome exactly this situation, that is, to search for the optimal response (local or global) when the number of input variables increases, trying to minimize the number of iterations without the algorithm “getting lost”, that is, it converges quickly. Again, these algorithms, by introducing small changes in the input variables, allow them to be executed during the normal production process. Every process/manufacturing engineer knows the value of this.
2. Methods
Box [
1] proposed an iterative experimentation process, basically his proposal consisted in the introduction of a two-level factorial design that is modified in search of the best response [
1]. Small changes in the levels of the design are estimated to minimize the generation of nonconforming product. This way of managing the experimentation process is known as operative evolutions (EVOP). EVOP is mainly used in the 3rd phase of experimentation, optimization, after the process has been characterized.
On the other hand, [
14], questions about the sample size, the Box algorithm does not consider whether the sample size is sufficient to generate acceptable statistical properties. It is clear that the proposal of Box has many advantages and that it is innovative because it searches for the best response with small changes, during production, minimizing the nonconforming product.
Spendley [
2], proposed a big change in the algorithm of Box, adding a simplex arrangement that replaces the factorial design [
2]. It is through this simplex that the levels of each iteration are calculated. The algorithm of [
2] is known as simplex EVOP, it iteratively moves the polyhedron (with n+1 vertices) looking for the optimization of the response. Here is a brief review of the operation of the variable simplex of Nelder and Mead [
3].
In the classical array, the levels of the variables considered are pre-designed, that is, once the array is defined, these levels remain fixed during the process of experimentation. In the Nelder and Mead Variable Simplex (NMS), the algorithm starts with pre-designed levels, then these levels are modified by the iterations of the algorithm according to a set of rules (operations of the simplex). The algorithm consists in using a simplex (a geometric arrangement), generally a polyhedron with n-1 vertices (for two input variables it will be a triangle). The search mechanism consists in using one of the vertices as a pivot (called the worse vertex) to estimate the next vertices (variable levels), always moving towards the best vertex (the one closest to the response objective: maximum, minimum, target). The algorithm moves through the response, calculating new vertexes in each iteration, through four basic operations: Reflection, Expansion, Contraction, and Shrinkage;
Figure 1 illustrates these operations.
There is plenty of literature available to understand the logistics of variable simplex operations, in this section a brief explanation of the logistics of the algorithm of NMS is resumed.
2.1. Basic Operations of the Nelder and Mead Algorithm
For the first iteration, k=0, the algorithm needs an initial simplex S0= {} because it prefers to start with a regular simplex. This characteristic changes immediately, since the successive iterations replace the worst vertex, , of the Sk-th simplex with another vertex. This worse vertex will be replaced by one of the new points (vertexes) . The selection of one of these three points, at any k≥0 iteration, is defined by one of the three operations of the algorithm: Reflection, Expansion, Contraction.
2.1.1. Reflection
Considering that
for at least one
i, where
i≠h or
i≠l; α is the coefficient of reflection.
Figure 2 shows the reflection operation in two dimensions (two variables of the process) for the n=2 case. The distance between the points
and
is equal to the distance between
and
. This is obtained when α=1.
2.1.2. Expansion
The worse vertex
is replaced by:
Considering that (i)
fr<fe, and (ii)
fe<fl. If (i) is kept but (ii) no, then the worse vertex
is replaced by
. γ is the expansion coefficient.
Figure 3 shows the expansion operation for two dimensions, n=2. The distance between the points
and
is equivalent to the distances between the points
and
, and
and
. This is obtained for a γ= 2.
2.1.3. Contraction
In the contraction operation, the worse vertex
is replaced by:
Considering that
fr ≥ fh, or:
When fr < fh ; β is the coefficient of contraction. In Figure 4 the operation of contraction is shown for two dimensions n=2. The distance between the points and corresponds to two times the distances between and , or and . This is obtained for a β= ½.
In the case studied and presented in this paper, they were considered six variables, so the geometrical array is to be a heptagon, if the traditional NMS algorithm was used. As an alternative to these mathematical and practical problems the named Parallel Simplex is proposed.
3. The Parallel Simplex Proposal
This document concentrates on the proposed alternative of the parallel simplex. On
Figure 5, it is shown the movements of the simplex (geometrical array), with the basic operations of Reflection, Expansion and Contraction, always moving towards the objective of the response variable. As shown on
Figure 5, the
parallel simplex approaches towards the objective of the response variable with three simplexes, independent between them, but pursuing the same response and objective. After an exhaustive literature review, nothing similar was found. Many authors have tried to minimize the problems of non-convergency and the high number of iterations adding or modifying the original Nelder Mead simplex, never the less, the results are very questionable because of:
- -
The complexity of the algorithm still increases with the number of input variables
- -
Almost all the evidence found is generated through known test functions (thus, deterministic approximations)
- -
Almost all the cases presented are computer simulated situations
Maintaining the three simplexes independent between them, assures the efficiency of the algorithm, avoiding the increase of the complexity of the calculus, thus reducing the number of iterations, as it will happen if the traditional array of a heptagon (n+1 dimensions) is used. Because the three simplexes chase the same objective, the simplex operations are the same for every single simplex.
4. The Analyzed Process
The process analyzed corresponds to the manufacturing of plastic profiles, to be applied as seals for refrigerators, washing machines and other domestic equipment. Plastic is melted through a simple plasticizing screw and extruded through a nozzle with the designed profile. On
Figure 6 the selected profile is shown. This profile was chosen because it presented a high rejection rate, actually, 100% rejection of lots. 100% inspection and rework were used on a regular basis.
This seal showed a very high rate of defects, mainly, incomplete shots, on
Figure 7 it is shown this defect.
The main control variables of this process are the cubic feet per minute of air (at a constant 4 Bar pressure) injected by four nozzles, this is done to keep the profile of the extrusion, as shown on
Figure 8. The fifth variable is the height of the puller. This height is the distance between the rollers that pull the extruded material. The sixth variable is the speed of the extruder and the puller (that are synchronized). The output variable is measured as defects per 1,000 parts as λ (The continuous extruded material is cut to the design length).
At this point is very important to consider that, if a traditional/classic array is used, let’s say, a two-level factorial design, it will imply:
The design, set up and run of 26 runs, that is 64 runs
If more replicates are used for better fit estimations, the number of runs will increase
The majority of product will not be conformant to specifications
Production will be lost because the process must stop (to set up each run)
Extra resources (time, Technicians, operators, materials, etc.) will be needed
The principle of Evolutive Operations of these direct search algorithms, as the NMS, allows the investigator to search for the best response while the process is running. The small changes induced in the process allows to reduce or even eliminate the generation of non-conformant parts.
4.1. The Experimental Array
The Experimental Array
A specific software was developed to run the Parallel Simplex, named Armentum
®. The original algorithm of Nelder and Mead was modified to add a response known as dual response or simultaneous optimization, [
15]. This way of evaluating the response considers the natural response, the objective pursued and the variation in the process at the same time, increasing the efficiency of the search. The expected value of the response variable is considered as:
Leading to a practical standard, considering 3 standard deviations for a normal distributed process:
Where is the dual response, is the average of the runs (in the same iteration), and is the standard deviation of the runs. In this case the objective of the response variable y is cero defects per 1,000 parts.
The control variables and the initial vertexes of the simplexes are set up in the software. This is shown on
Figure 9.
The first set of geometrical arrays; three triangles, 3 vertexes each one, are set up; 3 combinations for each pair of variables,
and
;
and
;
and
. The process is started and the feedback of the response is input to the software. The software calculates, on every new iteration, a new combination (vertex) for each simplex (in classical experimentation, these combinations -runs- are determined in advance, luckily some runs will be useful to estimate changes in the response, many other will not).
Table 3 shows the final results of the experimentation process.
The NMS algorithm categorizes the response as Best (B), Near the Worse (N
W) and Worse (W), considering the distance of the response from the objective (in this case a minimum of cero is intended), this is documented as the Ranking of the response. There are a lot of stop criteria of the algorithm on the literature, never the less, the Teta parameter developed [
10] was included because its simplicity and accuracy.
This is a comparison of the standard error of the response against a predetermined value, typically θ=1x10-6, for test functions. In real practical processes, a value of Teta of θ=1x10-3 is more than acceptable.
5. Results Analysis
On Table 3 it is shown that the algorithm stopped at the 24th iteration, obtaining a desired 0.002 on the dual response and a Teta value of 2.22 X10-7. Before any other comparison, a two proportions analysis was made to compare the process with the new set up. Historical data of three months against lots of 1,000 pieces where utilized. The analysis showed a p-value of 0.0000 for the probability of the alfa error, for the hypothesis that ; thus, the difference is statistically significant.
Utilizing Minitab® it can be seen that, for a full factorial design needed to estimate a Response Surface, 90 runs are required for six variables (with the minimum of central, axial and radial points), and, at minimum, 53 runs are required for a half design.
These runs were not executed because it implied to many runs, to much material, to many scrap, etcetera; never the less, data of the Parallel Simplex results were utilized to feed the Response Surface array and compare the results. This comparison is shown on
Table 4.
The internal operation of the Parallel Simplex is based on the continuous iteration of three independent simplexes, with two variables each in this case. As a further comparison, a Response Surface analysis was made for each pair of variables (simplex), utilizing the algorithm results (again, these runs were not executed). A full model with principal and interactions effects were designed and analyzed.
As part of the comparison, a canonical analysis was made to assure that the optimum found with the Poly Simplex algorithm corresponds to a local minimum. The regression analysis for the first simplex resulted on the equation:
With an adjusted and p-values for of 0.000 and 0.003
The stationary point estimated for this pair of variables is:
Considering matrix B and b
and Thus
; the canonical conversion shows that
for as both roots are positive, this stationary point is a minimum.
The regression analysis for the second simplex resulted on the equation:
With an adjusted and p-values for of 0.000 and 0.723
The stationary point estimated for this pair of variables is:
Considering matrix B and b
and Thus
; the canonical conversion shows that
for as both roots are positive, this stationary point is a minimum.
The canonical analysis demonstrates that the final combinations of each pair of variables is a local minimum. On Table 5 is shown a resume of the regression coefficients, for the six variables at the same time and for each pair of variables.
6. Conclusions
The proposed algorithm has demonstrated its efficiency by finding the global minimum, for the six variables included in the model, in only 24 iterations. A traditional model using the original Nelder and Mead algorithm, which requires a heptagon for the analysis, would require a very large number of iterations, and, as already demonstrated, a traditional factorial design would also require a considerable number of experimental runs. Among the most important contributions is the fact that this process was performed during a normal production run, production was not stopped, no extraordinary resources were used and the generation of nonconforming parts was minimal.
On the other hand, the canonical analysis shows that a local minimum for the response variable has been found for every pair of variables. It is worth mentioning that with the new operating parameters, the process currently has a quality efficiency of 98%.
7. Discussion
The case study presented evidences that the Parallel Simplex is an efficient alternative to classical experimentation. It is an experimentation process that does not require to stop production, generate scrap and use additional resources.
For many years these Direct Search Algorithms were concealed to theoretical approaches, with test functions and so. One of the main reasons is because the mathematical complexity of the calculus and the sensibility of the algorithms to the internal noise, present in any industrial process. As a matter of fact, it is difficult to find literature of successful approaches on real processes, In addition, an approach such as the one proposed here was not found in the literature reviewed: three independent simplexes seeking to optimize the same response in a synchronous manner. A few examples can be revised on [
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30], among many others. This alternative is a powerful way to optimize processes, overcoming the restrictions of the original algorithms.
Applying the Parallel concept to a higher dimension (variables) needs to be developed to investigate the efficiency of the algorithm. Further investigations of this algorithm are being developed to accumulate evidence of its practical benefit to the average industrial Engineer.
Acknowledgments
“This work was published with the support of the Institute of Innovation and Competitiveness of the Ministry of Innovation and Economic Development of the State of Chihuahua, México.”. “Esta obra fue publicada con el apoyo del Instituto de Innovación y Competitividad de la Secretaría de Innovación y Desarrollo Económico del Estado de Chihuahua, México”.
Conflicts of Interest
“The authors declare no conflicts of interest.”
References
- Box, G.E.P. (1957). Evolutionary Operation: A Method for Increasing Industrial Productivity. Applied Statistics, Vol. 6, No. 2, pp. 110-115.
- Spendely, W., G.R., and Himsworth F.R. (1962). Sequential Applications of Simplex Designs in Optimization and EVOP. Technometrics, Vol. 4, Num. 4, pp. 441-461.
- Nelder, J. A. and Mead, R. (1965). A Simplex Method for Function Minimization. Computer Journal, Vol. 7, pp. 308-313.
- Mrad, M., Bamatraf, K., Alkahtani, M., & Hidri, L. (2023). A GENETIC ALGORITHM FOR THE INTEGRATED WAREHOUSE LOCATION, ALLOCATION AND VEHICLE ROUTING PROBLEM IN A POOLED TRANSPORTATION SYSTEM. International Journal of Industrial Engineering: Theory, Applications and Practice, 30 (3). [CrossRef]
- Vineetha, G. R., & Shiyas, C. R. (2023). MACHINE LEARNING-ENHANCED GENETIC ALGORITHM FOR ROBUST LAYOUT DESIGN IN DYNAMIC FACILITY LAYOUT PROBLEMS: Implementation of Dynamic Facility Layout Problems. International Journal of Industrial Engineering: Theory, Applications and Practice, 30 (6).
- Xie, Y., Jiang, H., Wang, L., & Wang, C. (2023). SPORTS ECOTOURISM DEMAND PREDICTION USING IMPROVED FRUIT FLY OPTIMIZATION ALGORITHM. International Journal of Industrial Engineering: Theory, Applications and Practice, 30 (6). [CrossRef]
- Praga-Alejo, R. J. ., de León-Delgado, H., González-González, D. S., Cantú-Sifuentes, M., & Tahaei, A. (2022). MANUFACTURING PROCESSES MODELING USING MULTIVARIATE INFORMATION CRITERIA FOR RADIAL BASIS FUNCTION SELECTION. International Journal of Industrial Engineering: Theory, Applications and Practice, 29 (1). [CrossRef]
- Farajzadeh Bardeji, S., Fakoor Saghih, A. mohammad, & Mirghaderi, S.-H. (2022). MULTI-OBJECTIVE INVENTORY AND ROUTING MODEL FOR A MULTI-PRODUCT AND MULTI-PERIOD PROBLEM OF VETERINARY DRUGS. International Journal of Industrial Engineering: Theory, Applications and Practice, 29(4). [CrossRef]
- Drazic, Milan; Drazic, Zorica; Dragan, Urosevic (2014). Continuous variable neighborhood search with modified Nelder-Mead for non-differentiable optimization. IMA Journal of Management Mathematics. 27, pp. 75-78. [CrossRef]
- Fan, S.; Zahara E. (2006). Stochastic response surface optimization via an Enhanced Nelder Mead simplex search procedure. Engineering Optimization, 38:1, pp. 15-36.
- Park, H., Shin, Y., & Moon, I. (2024). MULTIPLE-OBJECTIVE SCHEDULING FOR BATCH PROCESS SYSTEMS USING STOCHASTIC UTILITY EVALUATION. International Journal of Industrial Engineering: Theory, Applications and Practice, 31(2). [CrossRef]
- Uluskan, M., & Beki, B. (2024). PROJECT SELECTION REVISITED: CUSTOMIZED TYPE-2 FUZZY ORESTE APPROACH FOR PROJECT PRIORITIZATION. International Journal of Industrial Engineering: Theory, Applications and Practice, 31(2). [CrossRef]
- Sharma, J., Tyagi, M., & Bhardwaj, A. (2023). CONTEMPLATION OF OPERATIONS PERSPECTIVES ENCUMBERING THE PRODUCTION-CONSUMPTION AVENUES OF THE PROCESSED FOOD SUPPLY CHAIN. International Journal of Industrial Engineering: Theory, Applications and Practice, 30(1). [CrossRef]
- Ryan, T. P. (1989). Statistical Methods for Quality Improvement. John Wiley. New York.
- Montgomery, D.C. (1997). Design and Analysis of Experiments, 4th. Edition. Jhon Wiley & Sons. New York.
- Kaelo, P.; Ali, M. (2006). Some Variants of the Controlled Random Search Algorithm for Global Optimization. Journal of Optimization Theory and Applications. Vol. 130. No. 2, pp. 253-264. August 2006.
- Lewis, R.; Shepherd A.; Torczon, V. (2007). Implementing Generating Set Search Methods for Linearly Constrained Minimization. Society for Industrial and Applied Mathematics.Vol.29, No. 6, pp. 2507-2530.
- S. Fan Shu-Kai y Zahara Erwie (2007). Stochastic response surface optimization via an enhanced Nelder-Mead simplex search procedure. Engineering Optimization, Vol.38, No. 1, pp. 15-36.
- Lewis, R.; and Torczon V. (2009). Active set identification for linearly constrained minimization without explicit derivatives. Society for Industrial and Applied Mathematics. Vol. 20, No. 3, pp. 1378-1405.
- Bogani, C. ,Gasparo M.G. y Papini A. (2009). GENERATING SET SEARCH METHODS FOR PIECE WISE SMOOTH PROBLEMS. Society for Industrial and Applied Mathematics, SIAM J. OPTIM., vol. 20, No 1, pp. 321-335.
- Khorsandi, A. , Alimardani A., Vahidir B. y Hosseinian S.H. (2010). Hybrid shuffled frog leaping algorithm and Nelder-Mead simplex search for optimal reactive power dispatch. Published in IET Generation, Transmission and Distribution, vol.5,Iss. 2, pp. 249-256.
- Bera S. y Mukherjee I. (2010). A Mahalanobis Distance-based Diversification and Nelder-Mead Simplex Intensification Search Scheme for Continuous Ant Colony Optimization. World Academy of Science, Engineering and Technology 69.
- Griffin J. y Kolda T. (2010). Asynchronous parallel hybrid optimization combining DIRECT and GSS. Optimization Methods y Software , Vol. 25, No. 5, pp. 797-817.
- Gao, X.K., Low, T.S., Liu, Z.J., Chen, S.X. (2002). Robust Design for Torque Optimization Using Response Surface Methodology. IEEE Transactions on Magnetics. Vol. 38, No. 2.
- Luo C. y Yu B. (2011). Low dimensional simplex evolution: a new heuristic for global optimization. Journal of Global Optimization 52: pp. 45-55.
- Barzinpour F., Noorossana R.,Akhavan Niaki S. y Javad Ershadi M. (2012). A hybrid Nelder- Mead simplex and PSO approach on economic and economic-statistical designs of MEWMA control charts. Int. J. Adv.Manuf. Technol 65: pp. 1339-1348.
- Hossein Gandomi A., Yang X., Hossein Alavi A. (2011). Cuckoo search algorithm: a metaheuristic approach to solve structural optimization problems. Engineering with Computers 29: pp. 17-35.
- Hussain S. y Gabbar H. (2013). Fault Diagnosis In Gearbox Using Adaptive Wavelet Filtering and Shock Response Spectrum Features Extraction. Structural Health Monitoring, pp. 1-12.
- McCormick, Matthew, and Verghese, Tomy. (2013). An Approach to Unbiased Subsample Interpolation for Motion Tracking. Ultrasonic Imaging, pp. 35: 76.
- Ren Jian, y Duan Jinqiao. (2013). A parameter estimation method based on random slow manifolds. Cornell University Library. arXiv:1303.4600.
- Tippayawannakorn Noocharin y Pichitlamken Juta (2013).NELDER- MEAD METHOD WITH LOCAL SELECTION USING NEIGHBORHOOD AND MEMORY FOR STOCHASTIC OPTIMIZATION. Journal of Computer Science, 9 (4): pp. 463-476.
|
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).