Finding a minimum in a flat neighborhood¶
An excercise of finding minimum. This excercise is hard because the function is very flat around the minimum (all its derivatives are zero). Thus gradient information is unreliable.
The function admits a minimum in [0, 0]. The challenge is to get within 1e-7 of this minimum, starting at x0 = [1, 1].
The solution that we adopt here is to give up on using gradient or information based on local differences, and to rely on the Powell algorithm. With 162 function evaluations, we get to 1e-8 of the solution.
スクリプトの出力:
Optimization terminated successfully.
Current function value: 0.000000
Iterations: 2
Function evaluations: 162
Python ソースコード: plot_exercise_flat_minimum.py
import numpy as np
from scipy import optimize
import pylab as pl
def f(x):
return np.exp(-1/(.01*x[0]**2 + x[1]**2))
# A well-conditionned version of f:
def g(x):
return f([10*x[0], x[1]])
# The gradient of g. We won't use it here for the optimization.
def g_prime(x):
r = np.sqrt(x[0]**2 + x[1]**2)
return 2/r**3*g(x)*x/r
x_min = optimize.fmin_powell(g, [1, 1], xtol=1e-10)
###############################################################################
# Some pretty plotting
pl.figure(0)
pl.clf()
t = np.linspace(-1.1, 1.1, 100)
pl.plot(t, f([0, t]))
pl.figure(1)
pl.clf()
X, Y = np.mgrid[-1.5:1.5:100j, -1.1:1.1:100j]
pl.imshow(f([X, Y]).T, cmap=pl.cm.gray_r, extent=[-1.5, 1.5, -1.1, 1.1],
origin='lower')
pl.contour(X, Y, f([X, Y]), cmap=pl.cm.gnuplot)
# Plot the gradient
dX, dY = g_prime([.1*X[::5, ::5], Y[::5, ::5]])
# Adjust for our preconditioning
dX *= .1
pl.quiver(X[::5, ::5], Y[::5, ::5], dX, dY, color='.5')
# Plot our solution
pl.plot(x_min[0], x_min[1], 'r+', markersize=15)
pl.show()
Total running time of the example: 0.12 seconds ( 0 minutes 0.12 seconds)