I was just saying that I can design finite difference schemes with NNs approach. But I sincerely hope that this bridge between NNs and finite difference methods has already been noticed before.
Remind us again. I missed it. The last time the details were missing AFAIR.
Are you referring to approximating the gradient by divided differences?
// I have never seen FD mentioned in the same breath as NN.
Yes, I am saying that with NNs-type kernels I can design any type of finite difference schemes, in particular coinciding with forward / backward differences. I did not give any examples, lets give it : consider a simple 1D example over [0,1], N points x^n = (n+0.5)/N. Consider the rectifier function \varphi(x) = max(x,0) and the kernel K(x,y) = 1 - \varphi(y-x). Define the NxN matrix K = K(x^n,x^m)_{n,m}. Its inverse is (...,0,-N, 2N,-N,0,...) (Cuchullain, you should here have a tickle in your head). Takes now the derivative : K' = -\varphi'(x^n-x^m)_{n,m}. Do you want to express the derivative of a function with this kernel ? you would find the following NxN matrix operator (...,0, -N,+N,0,...), coinciding with the classical finite forward difference operator. You want to compute the Laplacian of a function associated with this kernel ? This is the very classical finite difference NxN matrix operator (...0,-N^2, 2N^2,-N^2,0,...). You want to move the x^n's because your algorithms is learning something ? it won't change these operators so much.