.. _appx-smathry-doe:


DOE Methods
-----------

The sampling methods implemented in the :ref:`sma-doe` node do not call external
python libraries and are directly coded within Nodeworks. Therefore, a brief
overview of the different hard-coded methods are provided here.

All these methods return an array of samples between 0 and 1. These samples
are then converted to the appropriate ranges as specified.

.. _appx-smathry-doe-fact:


Factorial
+++++++++

Factorial constructs a full factorial design for an arbitrary number of factors
at an arbitrary number of levels. This typically results in the largest number
of samples since every possible combination of factors is produced.

The code here was inspired by `pyDOE: Factorial <https://github.com/tisimst/pyDOE/blob/master/pyDOE/doe_factorial.py>`_

.. image:: images/factorial.png
    :align: center

.. _appx-smathry-doe-cov:


Covary
++++++

Covary simply varies one or more factors linearly together with a slope of 1 and
an intercept of 0, i.e. :math:`y=x`.

.. image:: images/covary.png
    :align: center

.. _appx-smathry-doe-mcs:


Monte Carlo
+++++++++++

Numpy's `random.uniform <https://docs.scipy.org/doc/numpy/reference/generated/numpy.random.uniform.html>`_
method is used to create the random samples. If a seed is used to fix the
internal random generator, then `random.seed <https://docs.scipy.org/doc/numpy/reference/generated/numpy.random.seed.html>`_
is called before calling `random.uniform <https://docs.scipy.org/doc/numpy/reference/generated/numpy.random.uniform.html>`_.

.. image:: images/montecarlo.png
    :align: center

.. _appx-smathry-doe-lhs:


Latin Hypercube
+++++++++++++++

The sample space is first divided into equal intervals. In each interval, random
points are selected using Numpy's, `random.rand <https://docs.scipy.org/doc/numpy/reference/generated/numpy.random.rand.html>`_.
Finally, points are randomly paired to make samples using Numpy's `random.permutation <https://docs.scipy.org/doc/numpy/reference/generated/numpy.random.permutation.html>`_.

Similarly to the :ref:`appx-smathry-doe-mcs`,  if a seed is used to fix the
internal random generator, then `random.seed <https://docs.scipy.org/doc/numpy/reference/generated/numpy.random.seed.html>`_
is called before entering the routine. This allows for reproducible designs.

The code here was inspired by `pyDOE: LHS <https://github.com/tisimst/pyDOE/blob/master/pyDOE/doe_lhs.py>`_

.. image:: images/latin_hypercube.png
    :align: center

Generally, the basic Latin Hypercube algorithm does a better job of spacing
filling than :ref:`appx-smathry-doe-mcs`. However, there are techniques to
improve the spacing filling of the basic Latin Hypercube. Three are implemented
here.

maximin
.......

If ``maximin`` is selected, then a user defined number (``iterations``) of basic
Latin Hypercubes are constructed. The minimum distance between any two points is
determined for each design using `Scipy's pairwise distance <https://docs.scipy.org/doc/scipy/reference/generated/scipy.spatial.distance.pdist.html>`_
function. The design with the maximum minimum distance is returned.

ratio
.....

Similar to ``maximin``, ``ratio`` constructs a user defined number
(``iterations``) of basic Latin Hypercubes. A ratio of the maximum distance
between any two points over the minimum distance between any two points
(:math:`max(dist)/min(dist)`) is calculated using
`Scipy's pairwise distance <https://docs.scipy.org/doc/scipy/reference/generated/scipy.spatial.distance.pdist.html>`_
function for each design. The design with the minimum ration is returned.

genetic
.......

The ``genetic`` option follows the methodology described by Jin et al.
[Jin2005]_ for taking a design and using an enhanced stochastic evolutionary
(ESE) algorithm to optimize the design by columnwise operations, called
element-exchange, which interchanges two distinct elements in a column. This
technique allows for variables in samples to be exchanged to improve the
space filling of the original design without changing the original variables.
This guarantees that the resulting "optimized" design will still be a latin
hypercube.

The number of exchanges (:math:`J`), inner iterations (:math:`M`), as well as
the warming and cooling parameters are all calculated based on the
recommendations made by Jin et al. [Jin2005]_. The number of outer iterations
is user selected (``iterations``). The objective function is the Wrap-Around
L2-Discrepancy from Eq.5 of [Fang2001]_.

.. math::

    WD^2(D) = -(4/3)^K + 1/N^2 \Sigma_{i,j=1}^{N} Pi_{k=1}^K [3/2 - |x_k^1 - x_k^2| * (1 - |x_k^1 - x_k^2|)]

The actual python implementation of the L2-Discrepancy was extracted from
`damar-wicaksono/gsa-module <https://github.com/damar-wicaksono/gsa-module>`_
(MIT license).

.. figure:: images/latin_hypercube_opt.png
    :align: center

    Latin Hypercube techniques with 2 variables, 30 samples, and 1000 iterations

.. _appx-smathry-doe-ccd:


Central Composite
+++++++++++++++++

Central composite design is a combination of a 2-level full-factorial and a
star. There are three different options for the relation between the star points
and the corner points:

* circumscribed - star points are at some distance ``alpha`` from the center.
* inscribed - uses the factors settings as the star points and creates a
  factorial or fractional factorial design within those limits.
* faced - the star points are at the center of each face of the factorial space.

``alpha`` can be:

+------------+------------------------------------------------------------------------+
| orthogonal | :math:`\sqrt{n_{variables}{{1+ 1/n_{axial}}\over{1+1/k_{factorial}}}}` |
+------------+------------------------------------------------------------------------+
| rotatable  | :math:`\sqrt[4]{k_{factorial}}`                                        |
+------------+------------------------------------------------------------------------+

The code here was inspired by `pyDOE: Composite <https://github.com/tisimst/pyDOE/blob/master/pyDOE/doe_composite.py>`_

.. image:: images/central_composite.png
    :align: center

.. _appx-smathry-doe-sobol:


Sobol
+++++

The sobol implementation was adapted from `jburkardt <people.sc.fsu.edu/~jburkardt/py_src/sobol/sobol_lib.py>`_
and can handle a maximum of 40 variables.

.. image:: images/sobol.png
    :align: center

.. _appx-smathry-doe-hammersly:


Hammersly
+++++++++

The Hammersly method is a sampling technique based on the low discrepancy
Hammersly sequence. It provides more uniform samples as compared to
:ref:`appx-smathry-doe-mcs`.

The code here was adapted from `PhaethonPrime <https://libraries.io/github/PhaethonPrime/hammersley>`_
which is based on `Sampling with Hammersley and Halton Points <http://www.cse.cuhk.edu.hk/%7Ettwong/papers/udpoint/udpoint.pdf>`_

.. image:: images/hammersly.png
    :align: center

.. _appx-smathry-doe-halton:


Halton
++++++

Similar to the :ref:`appx-smathry-doe-hammersly` method, the Halton method is a
sampling technique based on the low discrepancy Halton sequence. It provides
more uniform samples as compared to :ref:`appx-smathry-doe-mcs`.

The code here was adapted from `PhaethonPrime <https://libraries.io/github/PhaethonPrime/hammersley>`_
which is based on `Sampling with Hammersley and Halton Points <http://www.cse.cuhk.edu.hk/%7Ettwong/papers/udpoint/udpoint.pdf>`_

.. image:: images/halton.png
    :align: center



References
++++++++++

.. [Fang2001] K.T. Fang and C.X. Ma, "Wrap-Around L2-Discrepancy of Random
    Sampling, Latin Hypercube, and Uniform Designs," Journal of Complexity,
    vol. 17, pp. 608-624, 2001.

.. [Jin2005] Ruichen Jin, Wei Chen, Agus Sudjianto, "An efficient algorithm for
    constructing optimal design of computer experiments", Journal of Statistical
    Planning and Inference, vol. 134, pp. 268-287, 2005,
    ISSN 0378-3758, https://doi.org/10.1016/j.jspi.2004.02.014.