.. _appx-smathry-doe: DOE Methods ----------- The sampling methods implemented in the :ref:`sma-doe` node do not call external python libraries and are directly coded within Nodeworks. Therefore, a brief overview of the different hard-coded methods are provided here. All these methods return an array of samples between 0 and 1. These samples are then converted to the appropriate ranges as specified. .. _appx-smathry-doe-fact: Factorial +++++++++ Factorial constructs a full factorial design for an arbitrary number of factors at an arbitrary number of levels. This typically results in the largest number of samples since every possible combination of factors is produced. The code here was inspired by `pyDOE: Factorial <https://github.com/tisimst/pyDOE/blob/master/pyDOE/doe_factorial.py>`_ .. image:: images/factorial.png :align: center .. _appx-smathry-doe-cov: Covary ++++++ Covary simply varies one or more factors linearly together with a slope of 1 and an intercept of 0, i.e. :math:`y=x`. .. image:: images/covary.png :align: center .. _appx-smathry-doe-mcs: Monte Carlo +++++++++++ Numpy's `random.uniform <https://docs.scipy.org/doc/numpy/reference/generated/numpy.random.uniform.html>`_ method is used to create the random samples. If a seed is used to fix the internal random generator, then `random.seed <https://docs.scipy.org/doc/numpy/reference/generated/numpy.random.seed.html>`_ is called before calling `random.uniform <https://docs.scipy.org/doc/numpy/reference/generated/numpy.random.uniform.html>`_. .. image:: images/montecarlo.png :align: center .. _appx-smathry-doe-lhs: Latin Hypercube +++++++++++++++ The sample space is first divided into equal intervals. In each interval, random points are selected using Numpy's, `random.rand <https://docs.scipy.org/doc/numpy/reference/generated/numpy.random.rand.html>`_. Finally, points are randomly paired to make samples using Numpy's `random.permutation <https://docs.scipy.org/doc/numpy/reference/generated/numpy.random.permutation.html>`_. Similarly to the :ref:`appx-smathry-doe-mcs`, if a seed is used to fix the internal random generator, then `random.seed <https://docs.scipy.org/doc/numpy/reference/generated/numpy.random.seed.html>`_ is called before entering the routine. This allows for reproducible designs. The code here was inspired by `pyDOE: LHS <https://github.com/tisimst/pyDOE/blob/master/pyDOE/doe_lhs.py>`_ .. image:: images/latin_hypercube.png :align: center Generally, the basic Latin Hypercube algorithm does a better job of spacing filling than :ref:`appx-smathry-doe-mcs`. However, there are techniques to improve the spacing filling of the basic Latin Hypercube. Three are implemented here. maximin ....... If ``maximin`` is selected, then a user defined number (``iterations``) of basic Latin Hypercubes are constructed. The minimum distance between any two points is determined for each design using `Scipy's pairwise distance <https://docs.scipy.org/doc/scipy/reference/generated/scipy.spatial.distance.pdist.html>`_ function. The design with the maximum minimum distance is returned. ratio ..... Similar to ``maximin``, ``ratio`` constructs a user defined number (``iterations``) of basic Latin Hypercubes. A ratio of the maximum distance between any two points over the minimum distance between any two points (:math:`max(dist)/min(dist)`) is calculated using `Scipy's pairwise distance <https://docs.scipy.org/doc/scipy/reference/generated/scipy.spatial.distance.pdist.html>`_ function for each design. The design with the minimum ration is returned. genetic ....... The ``genetic`` option follows the methodology described by Jin et al. [Jin2005]_ for taking a design and using an enhanced stochastic evolutionary (ESE) algorithm to optimize the design by columnwise operations, called element-exchange, which interchanges two distinct elements in a column. This technique allows for variables in samples to be exchanged to improve the space filling of the original design without changing the original variables. This guarantees that the resulting "optimized" design will still be a latin hypercube. The number of exchanges (:math:`J`), inner iterations (:math:`M`), as well as the warming and cooling parameters are all calculated based on the recommendations made by Jin et al. [Jin2005]_. The number of outer iterations is user selected (``iterations``). The objective function is the Wrap-Around L2-Discrepancy from Eq.5 of [Fang2001]_. .. math:: WD^2(D) = -(4/3)^K + 1/N^2 \Sigma_{i,j=1}^{N} Pi_{k=1}^K [3/2 - |x_k^1 - x_k^2| * (1 - |x_k^1 - x_k^2|)] The actual python implementation of the L2-Discrepancy was extracted from `damar-wicaksono/gsa-module <https://github.com/damar-wicaksono/gsa-module>`_ (MIT license). .. figure:: images/latin_hypercube_opt.png :align: center Latin Hypercube techniques with 2 variables, 30 samples, and 1000 iterations .. _appx-smathry-doe-ccd: Central Composite +++++++++++++++++ Central composite design is a combination of a 2-level full-factorial and a star. There are three different options for the relation between the star points and the corner points: * circumscribed - star points are at some distance ``alpha`` from the center. * inscribed - uses the factors settings as the star points and creates a factorial or fractional factorial design within those limits. * faced - the star points are at the center of each face of the factorial space. ``alpha`` can be: +------------+------------------------------------------------------------------------+ | orthogonal | :math:`\sqrt{n_{variables}{{1+ 1/n_{axial}}\over{1+1/k_{factorial}}}}` | +------------+------------------------------------------------------------------------+ | rotatable | :math:`\sqrt[4]{k_{factorial}}` | +------------+------------------------------------------------------------------------+ The code here was inspired by `pyDOE: Composite <https://github.com/tisimst/pyDOE/blob/master/pyDOE/doe_composite.py>`_ .. image:: images/central_composite.png :align: center .. _appx-smathry-doe-sobol: Sobol +++++ The sobol implementation was adapted from `jburkardt <people.sc.fsu.edu/~jburkardt/py_src/sobol/sobol_lib.py>`_ and can handle a maximum of 40 variables. .. image:: images/sobol.png :align: center .. _appx-smathry-doe-hammersly: Hammersly +++++++++ The Hammersly method is a sampling technique based on the low discrepancy Hammersly sequence. It provides more uniform samples as compared to :ref:`appx-smathry-doe-mcs`. The code here was adapted from `PhaethonPrime <https://libraries.io/github/PhaethonPrime/hammersley>`_ which is based on `Sampling with Hammersley and Halton Points <http://www.cse.cuhk.edu.hk/%7Ettwong/papers/udpoint/udpoint.pdf>`_ .. image:: images/hammersly.png :align: center .. _appx-smathry-doe-halton: Halton ++++++ Similar to the :ref:`appx-smathry-doe-hammersly` method, the Halton method is a sampling technique based on the low discrepancy Halton sequence. It provides more uniform samples as compared to :ref:`appx-smathry-doe-mcs`. The code here was adapted from `PhaethonPrime <https://libraries.io/github/PhaethonPrime/hammersley>`_ which is based on `Sampling with Hammersley and Halton Points <http://www.cse.cuhk.edu.hk/%7Ettwong/papers/udpoint/udpoint.pdf>`_ .. image:: images/halton.png :align: center References ++++++++++ .. [Fang2001] K.T. Fang and C.X. Ma, "Wrap-Around L2-Discrepancy of Random Sampling, Latin Hypercube, and Uniform Designs," Journal of Complexity, vol. 17, pp. 608-624, 2001. .. [Jin2005] Ruichen Jin, Wei Chen, Agus Sudjianto, "An efficient algorithm for constructing optimal design of computer experiments", Journal of Statistical Planning and Inference, vol. 134, pp. 268-287, 2005, ISSN 0378-3758, https://doi.org/10.1016/j.jspi.2004.02.014.