[SciPy-user] constrained optimization
Robert Kern
robert.kern at gmail.com
Mon Apr 28 16:19:55 EDT 2008
On Mon, Apr 28, 2008 at 3:03 PM, Ondrej Certik <ondrej at certik.cz> wrote:
> On Mon, Apr 28, 2008 at 8:57 PM, Robert Kern <robert.kern at gmail.com> wrote:
> > On Mon, Apr 28, 2008 at 1:34 PM, John Hunter <jdh2358 at gmail.com> wrote:
> > > I need to do a N dimensional constrained optimization over a weight w
> > > vector with the constraints:
> > >
> > > * w[i] >=0
> > >
> > > * w.sum() == 1.0
> > >
> > > Scanning through the scipy.optimize docs, I see a number of examples
> > > where parameters can be bounded by a bracketing interval, but none
> > > where constraints can be placed on combinations of the parameters, eg
> > > the sum of them. One approach I am considering is doing a bracketed
> > > [0,1] constrained optimization over N-1 weights (assigning the last
> > > weight to be 1-sum others) and modifying my cost function to punish
> > > the optimizer when the N-1 input weights sum to more than one.
> > >
> > > Is there a better approach?
> >
> > Transform the coordinates to an unconstrained N-1-dimensional space.
> > One such transformation is the Aitchison (or "additive log-ratio")
> > transform:
> >
> > y = log(x[:-1] / x[-1])
> >
> > And to go back:
> >
> > tmp = hstack([exp(y), 1.0])
> > x = tmp / tmp.sum()
> >
> > Searching for "compositional data analysis" should yield similar
> > transformations, but this one should be sufficient for maintaining
> > constraints. For doing statistics, the other have better properties.
>
> Wow, that is very clever. Just today I was thinking how to do it and
> it didn't occur to me I should read scipy-user. :)
>
> The exp/log transform is clear, but I didn't figure out that in order
> to maintain
> the norm, I can maintain it in the last element, so it's enough to do:
>
> y = x[:-1]/x[-1]
>
> tmp = hstack([y, 1.0])
> x = tmp / tmp.sum()
>
> Very cool, thanks. However, the transform is not one to one, e.g. both
>
> x = [1, 2, 1, 4]
> x = [2, 4, 2, 8]
>
> represent the same thing:
>
> y = [0.25, 0.5, 0.25]
Yes, that is by design. With compositional data, only the ratios
between components matter. They are unique only up to a scaling
factor, and typically, you normalize them such that they sum to 1.
--
Robert Kern
"I have come to believe that the whole world is an enigma, a harmless
enigma that is made terrible by our own mad attempt to interpret it as
though it had an underlying truth."
-- Umberto Eco
More information about the SciPy-User
mailing list