MathGroup Archive 2011

[Date Index] [Thread Index] [Author Index]

Search the Archive

Re: Findminimum too slow for iterative reweighted least squares

  • To: mathgroup at smc.vnet.net
  • Subject: [mg123529] Re: Findminimum too slow for iterative reweighted least squares
  • From: "Oleksandr Rasputinov" <oleksandr_rasputinov at hmamail.com>
  • Date: Sat, 10 Dec 2011 07:29:05 -0500 (EST)
  • Delivered-to: l-mathgroup@mail-archive0.wolfram.com
  • References: <jbspn1$3rf$1@smc.vnet.net>

On Fri, 09 Dec 2011 10:59:45 -0000, Alberto Maydeu-Olivares  
<amaydeu at gmail.com> wrote:

> Hi, I'm trying to use FindMinimum to minimize a weighted least squares
> function where the weight matrix needs to be updated at each
> iteration. The weight matrix involves the inverse of a symbolic
> matrix. This should be fast if FindMinimum first evaluates the
> symbolic matrix, then takes the inverse. Given how long it takes, it's
> obviously not doing that. I wonder if anyone can suggest a way to
> force FindMinimum to do that. I obviously tried Evaluate, but does not
> work. At attach a toy example so that you can try. Finding the minimum
> of this toy example takes a full second. It takes minutes to do a
> realistic size job, It should not take more than a couple of seconds.
>
> Thanks a lot for your help.
>
>
> model = {l1^2 + ps, l2*l1, l2^2 + ps, l3*l2, l3*l1, l3^2 + ps};
> theta = {l1, l2, l3, ps};
> j = Outer[D, model, theta];
> data = {2., .42, 3., .56, .48, 1.};
> startval = Transpose[{theta, {.5, .5, .5, .5}}];
> e = data - model;
> mat = {{2 (l1^2 + ps)^2, 2 l1 l2 (l1^2 + ps), 2 l1^2 l2^2,
>     2 l2 l3 (l1^2 + ps), 2 l1 l2^2 l3,
>     2 l2^2 l3^2}, {2 l1 l2 (l1^2 + ps),
>     ps (l2^2 + ps) + l1^2 (2 l2^2 + ps), 2 l1 l2 (l2^2 + ps),
>     l1 l3 (l1^2 + l2^2 + ps), l2 l3 (l1^2 + l2^2 + ps),
>     2 l1 l2 l3^2}, {2 l1^2 l2^2, 2 l1 l2 (l2^2 + ps), 2 (l2^2 + ps)^2,
>      2 l1^2 l2 l3, 2 l1 l3 (l2^2 + ps),
>     2 l1^2 l3^2}, {2 l2 l3 (l1^2 + ps), l1 l3 (l1^2 + l2^2 + ps),
>     2 l1^2 l2 l3, l2^2 l3^2 + (l1^2 + ps) (l3^2 + ps),
>     l1 l2 (2 l3^2 + ps), 2 l2 l3 (l3^2 + ps)}, {2 l1 l2^2 l3,
>     l2 l3 (l1^2 + l2^2 + ps), 2 l1 l3 (l2^2 + ps),
>     l1 l2 (2 l3^2 + ps), l1^2 l3^2 + (l2^2 + ps) (l3^2 + ps),
>     2 l1 l3 (l3^2 + ps)}, {2 l2^2 l3^2, 2 l1 l2 l3^2, 2 l1^2 l3^2,
>     2 l2 l3 (l3^2 + ps), 2 l1 l3 (l3^2 + ps), 2 (l3^2 + ps)^2}};
>
> (*brute force approach to benchmark*)
> w = Inverse[mat];
> irls = FindMinimum[e. w  . e, Evaluate[Sequence @@ startval], Gradient
> -> -2 e.   w .j]; // Timing
>
> (*this should work, in fact, it takes almost twice as long*)
> w := Inverse[mat];
> irls = FindMinimum[Evaluate[e. w  . e],  Evaluate[Sequence @@
> startval], Gradient -> Evaluate[-2 e.   w .j]]; // Timing
>

The reason for the poor performance is simply that w is a huge expression.  
In the second case you've not only inserted Evaluate but also changed Set  
for SetDelayed on w so that the inverse has to be re-calculated at each  
iteration. You can get better performance like this:

w = Simplify@Inverse[mat];
irls = FindMinimum[
  Evaluate[e.w.e], Evaluate[Sequence @@ startval],
  Gradient -> Evaluate[-2 e.w.j]
]; // Timing

If FindMinimum needs to perform a lot of iterations (which is not the case  
in the example above), you may like to try:

w = Simplify@Inverse[mat];
ewe = Compile[Evaluate@Thread[{theta, _Real, 0}],  
Evaluate at Simplify[e.w.e]];
irls = FindMinimum[
  ewe[Sequence @@ theta], Evaluate[Sequence @@ startval],
  Gradient -> Evaluate[-2 e.w.j]
]; // Timing



  • Prev by Date: Re: Can FindFit take two (or more) equations?
  • Next by Date: Re: Complex diagram
  • Previous by thread: Re: Findminimum too slow for iterative reweighted least squares
  • Next by thread: Re: Findminimum too slow for iterative reweighted least squares