Agreed! You should be less defensive when someone finds a higher-performance solution to a cherry-picked function.Give this discussion a chance to run its course. Going all defensive is a bit silly.

And less cherry picking, por favor

- Traden4Alpha
**Posts:**23951**Joined:**

Agreed! You should be less defensive when someone finds a higher-performance solution to a cherry-picked function.Give this discussion a chance to run its course. Going all defensive is a bit silly.

And less cherry picking, por favor

- Cuchulainn
**Posts:**57054**Joined:****Location:**Amsterdam-
**Contact:**

Are you that 'someone'? As I said, using Mathematica is cheating in this case. So tell us what's the best solution? It's also OK to say you don't know.Agreed! You should be less defensive when someone finds a higher-performance solution to a cherry-picked function.Give this discussion a chance to run its course. Going all defensive is a bit silly.

And less cherry picking, por favor

- Traden4Alpha
**Posts:**23951**Joined:**

So now accessing Wolfram Alpha is out? (But accessing old math papers is still permitted?) Moving the goal posts once again?Are you that 'someone'? As I said, using Mathematica is cheating in this case. So tell us what's the best solution? It's also OK to say you don't know.

And less cherry picking, por favor

How can I know the "best solution" if the requirements are kept hidden (e.g., solution can't use calculus) or are capriciously revised (e.g., solution must now work on a second equation but apparently other cherry-picked functions are out of scope)? Is performance still important? It was specifically required for the first function but that requirement seems to have been dropped for the second one. If performance is important in the case of your two functions, then calculus leads to a higher performance calculation of the first derivative with no worries about the accuracy issues associated with using a numerical differential variable.

- Cuchulainn
**Posts:**57054**Joined:****Location:**Amsterdam-
**Contact:**

If you don't like the style, why don't you start your own thread? Save both of us all the hassle. BTW in all these years I cannot remember your ever having done that.

I will not respond forthwith if you keep going in English bla bla.

I will not respond forthwith if you keep going in English bla bla.

- Cuchulainn
**Posts:**57054**Joined:****Location:**Amsterdam-
**Contact:**

Cool.The use of the complex-step in calculating the derivative of an analytic function was introduced by Lyness & Moler.Icomplexify[$]f(z)[$] where [$]z = x + ih[$] and h = 0.001, e.g.

Then compute theimaginary partof [$]f(z)/h[$] and you are done.(*)

For x = 1.87 I get 1.3684..... e+289 for both exact and this method.(12 digits accuracy)

No subtraction cancellation as with FD classic.

(*) Squire and Trapp 1998.

Numerical Differentiation of Analytics Functions, SIAM Vol 4, N2, 1967 available here: link to the pdf paper

PS: In a previous life, I used the complex-step gradients to compute the sensittivity derivatives of the aerodynamic cost function.

The Lyness/Moler paper looks more computationally intensive than S&T for f'(x). It uses a series solution + numerical integration(?)

I am examining the wider applicability of this method for gradients compared to AD method (I exclude exact methods and classic fdm for various reason). I tested first order call option greeks and bond sensitivities and the results so far agree with exact and fdm solutions. I even tried option speed by applying S&T to gamma.

For functions with several complex variables z1, z2, ...as arguments I can now perturb each z_j (j =1, 2,..) independendently and compute partial derivatives with ease. Is that how the work?

I feel the method is most applicable with a small number of parameters, like MLE(?) (your aerodynamics have 18 parameters?)

Would the approach work for ML and more general optimisation, i.e. does it scale?

//

It was mentioned that the compiler needs to support erfc(z) for z complex. Not in C++11 but I found that it can be written in terms of the Faddeeva function (See package "Faddeeva" and good to go).

http://ab-initio.mit.edu/wiki/index.php ... va_Package

- FaridMoussaoui
**Posts:**232**Joined:**

For the AD method for greeks, have to look to Mike Giles work.

- Cuchulainn
**Posts:**57054**Joined:****Location:**Amsterdam-
**Contact:**

+ DiffSharp?For the AD method for greeks, have to look to Mike Giles work.

- FaridMoussaoui
**Posts:**232**Joined:**

I stay with universal languages, e.g C++/Fortran

- Cuchulainn
**Posts:**57054**Joined:****Location:**Amsterdam-
**Contact:**

There's a lot to be said for this approach for this computationally intensive problem.I stay with universal languages, e.g C++/Fortran

I reckon Fortran is the better of the two in this case?

- Cuchulainn
**Posts:**57054**Joined:****Location:**Amsterdam-
**Contact:**

Computing Dates of Rosh Hasbonah and Passover.

https://www.sciencedirect.com/science/a ... 2177900931

https://www.sciencedirect.com/science/a ... 2177900931

- Cuchulainn
**Posts:**57054**Joined:****Location:**Amsterdam-
**Contact:**

Computing derivatives without the pain of 1) differentiation, 2) catastrophic round-off errors. Sample code.

The scalar (almost a 1-liner). I leave the vector case as an exercise.

The scalar (almost a 1-liner). I leave the vector case as an exercise.

Code: Select all

```
// TestComplexStep.cpp
//
// Complex-step method to compute approximate derivatives.
// Example is scalar-valued function of a scalar argument.
//
// https://pdfs.semanticscholar.org/3de7/e8ae217a4214507b9abdac66503f057aaae9.pdf
//
// http://mdolab.engin.umich.edu/sites/default/files/Martins2003CSD.pdf
//
// (C) Datasim Education BV 2018
//
#include <functional>
#include <complex>
#include <iostream>
#include <iomanip>
#include <cmath>
// Notation and function spaces
using value_type = double;
template <typename T>
using FunctionType = std::function < T(const T& c)>;
using CFunctionType = FunctionType<std::complex<value_type>>;
// Test case from Squire&Trapp 1998
template <typename T> T func(const T& t)
{
T n1 = std::exp(t);
T d1 = std::sin(t);
T d2 = std::cos(t);
return n1 / (d1*d1*d1 + d2*d2*d2);
}
template <typename T> T func2(const T& t)
{ // Derivative of e^t, sanity check
return std::exp(std::pow(t,1));
// return std::exp(std::pow(t, 5));
}
value_type Derivative(const CFunctionType& f, value_type x, value_type h)
{ // df/dx at x using tbe Complex step method
std::complex<value_type> z(x, h); // x + ih, i = sqrt(-1)
return std::imag(f(z)) / h;
}
int main()
{
// Squire Trapp
double x = 1.5; double h = 0.1;
do
{
std::cout << std::setprecision(12) << Derivative(func<std::complex<value_type>>, x, h) << '\n';
h *= 0.1;
} while (h > 1.0e-300);
// Exponential function (101 sanity check)
x = 5.0;
h = 1.0e-10;
std::cout << "Exponential 1: " << std::setprecision(12) << Derivative(func2<std::complex<value_type>>, x, h) << '\n';
return 0;
}
```

GZIP: On