I'm trying to optimize only two parameters in a computationally expensive model, and I'm using lsqcurvefit with the Levenberg-Marquardt Algorithm.
I want to limit the minimum step sizes (or change magnitude in parameter values) to certain values; because the change in parameter values is very small by default, the gain of the loss functions is very small which lead the optimizer of getting stuck in local minima.
I tried to set a minimum variable difference, but the "MinDiffChange" can only be one scalar applied to both, while the two parameter values are not normalized to each other. This means that one parameter is 1, while another parameter is 10000. So changing the first parameter by 1 is a big deal, changing the second parameter by 1 is not!
How can I make sure the optimizer takes minimum step sizes proportional to each parameter used?
What is the "ScaleProblem" parameter in the optimization options of lsqcurvefit? Is it related to my issue?